"""List of Lists sparse matrix class """ __docformat__ = "restructuredtext en" __all__ = ['lil_array', 'lil_matrix', 'isspmatrix_lil'] from bisect import bisect_left import numpy as np from ._matrix import spmatrix from ._base import _spbase, sparray, issparse from ._index import IndexMixin, INT_TYPES, _broadcast_arrays from ._sputils import (getdtype, isshape, isscalarlike, upcast_scalar, check_shape, check_reshape_kwargs) from . import _csparsetools class _lil_base(_spbase, IndexMixin): _format = 'lil' def __init__(self, arg1, shape=None, dtype=None, copy=False): _spbase.__init__(self) self.dtype = getdtype(dtype, arg1, default=float) # First get the shape if issparse(arg1): if arg1.format == "lil" and copy: A = arg1.copy() else: A = arg1.tolil() if dtype is not None: A = A.astype(dtype, copy=False) self._shape = check_shape(A.shape) self.dtype = A.dtype self.rows = A.rows self.data = A.data elif isinstance(arg1,tuple): if isshape(arg1): if shape is not None: raise ValueError('invalid use of shape parameter') M, N = arg1 self._shape = check_shape((M, N)) self.rows = np.empty((M,), dtype=object) self.data = np.empty((M,), dtype=object) for i in range(M): self.rows[i] = [] self.data[i] = [] else: raise TypeError('unrecognized lil_array constructor usage') else: # assume A is dense try: A = self._ascontainer(arg1) except TypeError as e: raise TypeError('unsupported matrix type') from e if isinstance(self, sparray) and A.ndim != 2: raise ValueError(f"LIL arrays don't support {A.ndim}D input. Use 2D") A = self._csr_container(A, dtype=dtype).tolil() self._shape = check_shape(A.shape) self.dtype = A.dtype self.rows = A.rows self.data = A.data def __iadd__(self,other): self[:,:] = self + other return self def __isub__(self,other): self[:,:] = self - other return self def __imul__(self,other): if isscalarlike(other): self[:,:] = self * other return self else: return NotImplemented def __itruediv__(self,other): if isscalarlike(other): self[:,:] = self / other return self else: return NotImplemented # Whenever the dimensions change, empty lists should be created for each # row def _getnnz(self, axis=None): if axis is None: return sum([len(rowvals) for rowvals in self.data]) if axis < 0: axis += 2 if axis == 0: out = np.zeros(self.shape[1], dtype=np.intp) for row in self.rows: out[row] += 1 return out elif axis == 1: return np.array([len(rowvals) for rowvals in self.data], dtype=np.intp) else: raise ValueError('axis out of bounds') def count_nonzero(self): return sum(np.count_nonzero(rowvals) for rowvals in self.data) _getnnz.__doc__ = _spbase._getnnz.__doc__ count_nonzero.__doc__ = _spbase.count_nonzero.__doc__ def __str__(self): val = '' for i, row in enumerate(self.rows): for pos, j in enumerate(row): val += f" {str((i, j))}\t{str(self.data[i][pos])}\n" return val[:-1] def getrowview(self, i): """Returns a view of the 'i'th row (without copying). """ new = self._lil_container((1, self.shape[1]), dtype=self.dtype) new.rows[0] = self.rows[i] new.data[0] = self.data[i] return new def getrow(self, i): """Returns a copy of the 'i'th row. """ M, N = self.shape if i < 0: i += M if i < 0 or i >= M: raise IndexError('row index out of bounds') new = self._lil_container((1, N), dtype=self.dtype) new.rows[0] = self.rows[i][:] new.data[0] = self.data[i][:] return new def __getitem__(self, key): # Fast path for simple (int, int) indexing. if (isinstance(key, tuple) and len(key) == 2 and isinstance(key[0], INT_TYPES) and isinstance(key[1], INT_TYPES)): # lil_get1 handles validation for us. return self._get_intXint(*key) # Everything else takes the normal path. return IndexMixin.__getitem__(self, key) def _asindices(self, idx, N): # LIL routines handle bounds-checking for us, so don't do it here. try: x = np.asarray(idx) except (ValueError, TypeError, MemoryError) as e: raise IndexError('invalid index') from e if x.ndim not in (1, 2): raise IndexError('Index dimension must be <= 2') return x def _get_intXint(self, row, col): v = _csparsetools.lil_get1(self.shape[0], self.shape[1], self.rows, self.data, row, col) return self.dtype.type(v) def _get_sliceXint(self, row, col): row = range(*row.indices(self.shape[0])) return self._get_row_ranges(row, slice(col, col+1)) def _get_arrayXint(self, row, col): row = row.squeeze() return self._get_row_ranges(row, slice(col, col+1)) def _get_intXslice(self, row, col): return self._get_row_ranges((row,), col) def _get_sliceXslice(self, row, col): row = range(*row.indices(self.shape[0])) return self._get_row_ranges(row, col) def _get_arrayXslice(self, row, col): return self._get_row_ranges(row, col) def _get_intXarray(self, row, col): row = np.array(row, dtype=col.dtype, ndmin=1) return self._get_columnXarray(row, col) def _get_sliceXarray(self, row, col): row = np.arange(*row.indices(self.shape[0])) return self._get_columnXarray(row, col) def _get_columnXarray(self, row, col): # outer indexing row, col = _broadcast_arrays(row[:,None], col) return self._get_arrayXarray(row, col) def _get_arrayXarray(self, row, col): # inner indexing i, j = map(np.atleast_2d, _prepare_index_for_memoryview(row, col)) new = self._lil_container(i.shape, dtype=self.dtype) _csparsetools.lil_fancy_get(self.shape[0], self.shape[1], self.rows, self.data, new.rows, new.data, i, j) return new def _get_row_ranges(self, rows, col_slice): """ Fast path for indexing in the case where column index is slice. This gains performance improvement over brute force by more efficient skipping of zeros, by accessing the elements column-wise in order. Parameters ---------- rows : sequence or range Rows indexed. If range, must be within valid bounds. col_slice : slice Columns indexed """ j_start, j_stop, j_stride = col_slice.indices(self.shape[1]) col_range = range(j_start, j_stop, j_stride) nj = len(col_range) new = self._lil_container((len(rows), nj), dtype=self.dtype) _csparsetools.lil_get_row_ranges(self.shape[0], self.shape[1], self.rows, self.data, new.rows, new.data, rows, j_start, j_stop, j_stride, nj) return new def _set_intXint(self, row, col, x): _csparsetools.lil_insert(self.shape[0], self.shape[1], self.rows, self.data, row, col, x) def _set_arrayXarray(self, row, col, x): i, j, x = map(np.atleast_2d, _prepare_index_for_memoryview(row, col, x)) _csparsetools.lil_fancy_set(self.shape[0], self.shape[1], self.rows, self.data, i, j, x) def _set_arrayXarray_sparse(self, row, col, x): # Fall back to densifying x x = np.asarray(x.toarray(), dtype=self.dtype) x, _ = _broadcast_arrays(x, row) self._set_arrayXarray(row, col, x) def __setitem__(self, key, x): if isinstance(key, tuple) and len(key) == 2: row, col = key # Fast path for simple (int, int) indexing. if isinstance(row, INT_TYPES) and isinstance(col, INT_TYPES): x = self.dtype.type(x) if x.size > 1: raise ValueError("Trying to assign a sequence to an item") return self._set_intXint(row, col, x) # Fast path for full-matrix sparse assignment. if (isinstance(row, slice) and isinstance(col, slice) and row == slice(None) and col == slice(None) and issparse(x) and x.shape == self.shape): x = self._lil_container(x, dtype=self.dtype) self.rows = x.rows self.data = x.data return # Everything else takes the normal path. IndexMixin.__setitem__(self, key, x) def _mul_scalar(self, other): if other == 0: # Multiply by zero: return the zero matrix new = self._lil_container(self.shape, dtype=self.dtype) else: res_dtype = upcast_scalar(self.dtype, other) new = self.copy() new = new.astype(res_dtype) # Multiply this scalar by every element. for j, rowvals in enumerate(new.data): new.data[j] = [val*other for val in rowvals] return new def __truediv__(self, other): # self / other if isscalarlike(other): new = self.copy() new.dtype = np.result_type(self, other) # Divide every element by this scalar for j, rowvals in enumerate(new.data): new.data[j] = [val/other for val in rowvals] return new else: return self.tocsr() / other def copy(self): M, N = self.shape new = self._lil_container(self.shape, dtype=self.dtype) # This is ~14x faster than calling deepcopy() on rows and data. _csparsetools.lil_get_row_ranges(M, N, self.rows, self.data, new.rows, new.data, range(M), 0, N, 1, N) return new copy.__doc__ = _spbase.copy.__doc__ def reshape(self, *args, **kwargs): shape = check_shape(args, self.shape) order, copy = check_reshape_kwargs(kwargs) # Return early if reshape is not required if shape == self.shape: if copy: return self.copy() else: return self new = self._lil_container(shape, dtype=self.dtype) if order == 'C': ncols = self.shape[1] for i, row in enumerate(self.rows): for col, j in enumerate(row): new_r, new_c = np.unravel_index(i * ncols + j, shape) new[new_r, new_c] = self[i, j] elif order == 'F': nrows = self.shape[0] for i, row in enumerate(self.rows): for col, j in enumerate(row): new_r, new_c = np.unravel_index(i + j * nrows, shape, order) new[new_r, new_c] = self[i, j] else: raise ValueError("'order' must be 'C' or 'F'") return new reshape.__doc__ = _spbase.reshape.__doc__ def resize(self, *shape): shape = check_shape(shape) new_M, new_N = shape M, N = self.shape if new_M < M: self.rows = self.rows[:new_M] self.data = self.data[:new_M] elif new_M > M: self.rows = np.resize(self.rows, new_M) self.data = np.resize(self.data, new_M) for i in range(M, new_M): self.rows[i] = [] self.data[i] = [] if new_N < N: for row, data in zip(self.rows, self.data): trunc = bisect_left(row, new_N) del row[trunc:] del data[trunc:] self._shape = shape resize.__doc__ = _spbase.resize.__doc__ def toarray(self, order=None, out=None): d = self._process_toarray_args(order, out) for i, row in enumerate(self.rows): for pos, j in enumerate(row): d[i, j] = self.data[i][pos] return d toarray.__doc__ = _spbase.toarray.__doc__ def transpose(self, axes=None, copy=False): return self.tocsr(copy=copy).transpose(axes=axes, copy=False).tolil(copy=False) transpose.__doc__ = _spbase.transpose.__doc__ def tolil(self, copy=False): if copy: return self.copy() else: return self tolil.__doc__ = _spbase.tolil.__doc__ def tocsr(self, copy=False): M, N = self.shape if M == 0 or N == 0: return self._csr_container((M, N), dtype=self.dtype) # construct indptr array if M*N <= np.iinfo(np.int32).max: # fast path: it is known that 64-bit indexing will not be needed. idx_dtype = np.int32 indptr = np.empty(M + 1, dtype=idx_dtype) indptr[0] = 0 _csparsetools.lil_get_lengths(self.rows, indptr[1:]) np.cumsum(indptr, out=indptr) nnz = indptr[-1] else: idx_dtype = self._get_index_dtype(maxval=N) lengths = np.empty(M, dtype=idx_dtype) _csparsetools.lil_get_lengths(self.rows, lengths) nnz = lengths.sum(dtype=np.int64) idx_dtype = self._get_index_dtype(maxval=max(N, nnz)) indptr = np.empty(M + 1, dtype=idx_dtype) indptr[0] = 0 np.cumsum(lengths, dtype=idx_dtype, out=indptr[1:]) indices = np.empty(nnz, dtype=idx_dtype) data = np.empty(nnz, dtype=self.dtype) _csparsetools.lil_flatten_to_array(self.rows, indices) _csparsetools.lil_flatten_to_array(self.data, data) # init csr matrix return self._csr_container((data, indices, indptr), shape=self.shape) tocsr.__doc__ = _spbase.tocsr.__doc__ def _prepare_index_for_memoryview(i, j, x=None): """ Convert index and data arrays to form suitable for passing to the Cython fancy getset routines. The conversions are necessary since to (i) ensure the integer index arrays are in one of the accepted types, and (ii) to ensure the arrays are writable so that Cython memoryview support doesn't choke on them. Parameters ---------- i, j Index arrays x : optional Data arrays Returns ------- i, j, x Re-formatted arrays (x is omitted, if input was None) """ if i.dtype > j.dtype: j = j.astype(i.dtype) elif i.dtype < j.dtype: i = i.astype(j.dtype) if not i.flags.writeable or i.dtype not in (np.int32, np.int64): i = i.astype(np.intp) if not j.flags.writeable or j.dtype not in (np.int32, np.int64): j = j.astype(np.intp) if x is not None: if not x.flags.writeable: x = x.copy() return i, j, x else: return i, j def isspmatrix_lil(x): """Is `x` of lil_matrix type? Parameters ---------- x object to check for being a lil matrix Returns ------- bool True if `x` is a lil matrix, False otherwise Examples -------- >>> from scipy.sparse import lil_array, lil_matrix, coo_matrix, isspmatrix_lil >>> isspmatrix_lil(lil_matrix([[5]])) True >>> isspmatrix_lil(lil_array([[5]])) False >>> isspmatrix_lil(coo_matrix([[5]])) False """ return isinstance(x, lil_matrix) # This namespace class separates array from matrix with isinstance class lil_array(_lil_base, sparray): """ Row-based LIst of Lists sparse array. This is a structure for constructing sparse arrays incrementally. Note that inserting a single item can take linear time in the worst case; to construct the array efficiently, make sure the items are pre-sorted by index, per row. This can be instantiated in several ways: lil_array(D) where D is a 2-D ndarray lil_array(S) with another sparse array or matrix S (equivalent to S.tolil()) lil_array((M, N), [dtype]) to construct an empty array with shape (M, N) dtype is optional, defaulting to dtype='d'. Attributes ---------- dtype : dtype Data type of the array shape : 2-tuple Shape of the array ndim : int Number of dimensions (this is always 2) nnz size data LIL format data array of the array rows LIL format row index array of the array T Notes ----- Sparse arrays can be used in arithmetic operations: they support addition, subtraction, multiplication, division, and matrix power. Advantages of the LIL format - supports flexible slicing - changes to the array sparsity structure are efficient Disadvantages of the LIL format - arithmetic operations LIL + LIL are slow (consider CSR or CSC) - slow column slicing (consider CSC) - slow matrix vector products (consider CSR or CSC) Intended Usage - LIL is a convenient format for constructing sparse arrays - once an array has been constructed, convert to CSR or CSC format for fast arithmetic and matrix vector operations - consider using the COO format when constructing large arrays Data Structure - An array (``self.rows``) of rows, each of which is a sorted list of column indices of non-zero elements. - The corresponding nonzero values are stored in similar fashion in ``self.data``. """ class lil_matrix(spmatrix, _lil_base): """ Row-based LIst of Lists sparse matrix. This is a structure for constructing sparse matrices incrementally. Note that inserting a single item can take linear time in the worst case; to construct the matrix efficiently, make sure the items are pre-sorted by index, per row. This can be instantiated in several ways: lil_matrix(D) where D is a 2-D ndarray lil_matrix(S) with another sparse array or matrix S (equivalent to S.tolil()) lil_matrix((M, N), [dtype]) to construct an empty matrix with shape (M, N) dtype is optional, defaulting to dtype='d'. Attributes ---------- dtype : dtype Data type of the matrix shape : 2-tuple Shape of the matrix ndim : int Number of dimensions (this is always 2) nnz size data LIL format data array of the matrix rows LIL format row index array of the matrix T Notes ----- Sparse matrices can be used in arithmetic operations: they support addition, subtraction, multiplication, division, and matrix power. Advantages of the LIL format - supports flexible slicing - changes to the matrix sparsity structure are efficient Disadvantages of the LIL format - arithmetic operations LIL + LIL are slow (consider CSR or CSC) - slow column slicing (consider CSC) - slow matrix vector products (consider CSR or CSC) Intended Usage - LIL is a convenient format for constructing sparse matrices - once a matrix has been constructed, convert to CSR or CSC format for fast arithmetic and matrix vector operations - consider using the COO format when constructing large matrices Data Structure - An array (``self.rows``) of rows, each of which is a sorted list of column indices of non-zero elements. - The corresponding nonzero values are stored in similar fashion in ``self.data``. """