/external/eigen/Eigen/src/SparseCore/ |
SparseTranspose.h | 34 inline const StorageIndex* innerIndexPtr() const { return derived().nestedExpression().innerIndexPtr(); } 39 inline StorageIndex* innerIndexPtr() { return derived().nestedExpression().innerIndexPtr(); }
|
SparseCompressedBase.h | 70 * \sa innerIndexPtr(), outerIndexPtr() */ 74 * \sa innerIndexPtr(), outerIndexPtr() */ 80 inline const StorageIndex* innerIndexPtr() const { return derived().innerIndexPtr(); } 84 inline StorageIndex* innerIndexPtr() { return derived().innerIndexPtr(); } 89 * \sa valuePtr(), innerIndexPtr() */ 94 * \sa valuePtr(), innerIndexPtr() */ 158 : m_values(mat.valuePtr()), m_indices(mat.innerIndexPtr()), m_outer(outer) 176 : m_values(mat.valuePtr()), m_indices(mat.innerIndexPtr()), m_outer(0), m_id(0), m_end(mat.nonZeros() [all...] |
SparseMap.h | 96 /** \copydoc SparseMatrix::innerIndexPtr */ 97 inline const StorageIndex* innerIndexPtr() const { return m_innerIndices; } 124 inline SparseMapBase(Index rows, Index cols, Index nnz, IndexPointer outerIndexPtr, IndexPointer innerIndexPtr, 127 m_innerIndices(innerIndexPtr), m_values(valuePtr), m_innerNonZeros(innerNonZerosPtr) 131 inline SparseMapBase(Index size, Index nnz, IndexPointer innerIndexPtr, ScalarPointer valuePtr) 133 m_innerIndices(innerIndexPtr), m_values(valuePtr), m_innerNonZeros(0) 166 using Base::innerIndexPtr; 171 /** \copydoc SparseMatrix::innerIndexPtr */ 172 inline StorageIndex* innerIndexPtr() { return Base::m_innerIndices; } 195 inline SparseMapBase(Index rows, Index cols, Index nnz, StorageIndex* outerIndexPtr, StorageIndex* innerIndexPtr, [all...] |
MappedSparseMatrix.h | 42 inline MappedSparseMatrix(Index rows, Index cols, Index nnz, StorageIndex* outerIndexPtr, StorageIndex* innerIndexPtr, Scalar* valuePtr, StorageIndex* innerNonZeroPtr = 0) 43 : Base(rows, cols, nnz, outerIndexPtr, innerIndexPtr, valuePtr, innerNonZeroPtr)
|
SparseBlock.h | 151 internal::smart_copy(m_matrix.innerIndexPtr(), m_matrix.innerIndexPtr() + start, newdata.indexPtr()); 154 internal::smart_copy(tmp.innerIndexPtr() + tmp_start, tmp.innerIndexPtr() + tmp_start + nnz, newdata.indexPtr() + start); 157 internal::smart_copy(matrix.innerIndexPtr()+end, matrix.innerIndexPtr()+end + tail_size, newdata.indexPtr()+start+nnz); 173 internal::smart_memmove(matrix.innerIndexPtr()+end, matrix.innerIndexPtr() + end+tail_size, matrix.innerIndexPtr() + start+nnz); 179 internal::smart_copy(tmp.innerIndexPtr() + tmp_start, tmp.innerIndexPtr() + tmp_start + nnz, matrix.innerIndexPtr() + start) [all...] |
SparseSelfAdjointView.h | 492 dest.innerIndexPtr()[k] = StorageOrderMatch ? ip : jp; 498 dest.innerIndexPtr()[k] = ip; 506 dest.innerIndexPtr()[k] = ip; 509 dest.innerIndexPtr()[k] = jp; 572 dest.innerIndexPtr()[k] = int(DstMode)==int(Lower) ? (std::max)(ip,jp) : (std::min)(ip,jp);
|
SparseVector.h | 89 EIGEN_STRONG_INLINE const StorageIndex* innerIndexPtr() const { return m_data.indexPtr(); } 90 EIGEN_STRONG_INLINE StorageIndex* innerIndexPtr() { return m_data.indexPtr(); }
|
SparseRef.h | 101 ::new (static_cast<Base*>(this)) Base(expr.size(), expr.nonZeros(), expr.innerIndexPtr(), expr.valuePtr()); 103 ::new (static_cast<Base*>(this)) Base(expr.rows(), expr.cols(), expr.nonZeros(), expr.outerIndexPtr(), expr.innerIndexPtr(), expr.valuePtr(), expr.innerNonZeroPtr());
|
SparseAssign.h | 195 Map<ArrayXI>(dst.innerIndexPtr(), size).setLinSpaced(0,StorageIndex(size)-1);
|
SparseMatrix.h | 147 * \sa innerIndexPtr(), outerIndexPtr() */ 151 * \sa innerIndexPtr(), outerIndexPtr() */ 157 inline const StorageIndex* innerIndexPtr() const { return m_data.indexPtr(); } 161 inline StorageIndex* innerIndexPtr() { return m_data.indexPtr(); } 165 * \sa valuePtr(), innerIndexPtr() */ 169 * \sa valuePtr(), innerIndexPtr() */ [all...] |
/external/eigen/unsupported/Eigen/src/SparseExtra/ |
RandomSetter.h | 278 while ( (i >= posStart) && (mp_target->innerIndexPtr()[i] > inner) ) 281 mp_target->innerIndexPtr()[i+1] = mp_target->innerIndexPtr()[i]; 284 mp_target->innerIndexPtr()[i+1] = inner;
|
BlockSparseMatrix.h | [all...] |
/external/eigen/Eigen/src/UmfPackSupport/ |
UmfPackSupport.h | 372 mp_matrix.outerIndexPtr(), mp_matrix.innerIndexPtr(), mp_matrix.valuePtr(), 385 m_fact_errorCode = umfpack_numeric(mp_matrix.outerIndexPtr(), mp_matrix.innerIndexPtr(), mp_matrix.valuePtr(), 455 umfpack_get_numeric(m_l.outerIndexPtr(), m_l.innerIndexPtr(), m_l.valuePtr(), 456 m_u.outerIndexPtr(), m_u.innerIndexPtr(), m_u.valuePtr(), 493 mp_matrix.outerIndexPtr(), mp_matrix.innerIndexPtr(), mp_matrix.valuePtr(),
|
/external/eigen/Eigen/src/OrderingMethods/ |
Ordering.h | 144 for(StorageIndex i=0; i < nnz; i++) A(i) = mat.innerIndexPtr()[i];
|
Amd.h | 125 StorageIndex* Ci = C.innerIndexPtr();
|
/external/eigen/Eigen/src/PardisoSupport/ |
PardisoSupport.h | 270 m_matrix.valuePtr(), m_matrix.outerIndexPtr(), m_matrix.innerIndexPtr(), 291 m_matrix.valuePtr(), m_matrix.outerIndexPtr(), m_matrix.innerIndexPtr(), 311 m_matrix.valuePtr(), m_matrix.outerIndexPtr(), m_matrix.innerIndexPtr(), 358 m_matrix.valuePtr(), m_matrix.outerIndexPtr(), m_matrix.innerIndexPtr(),
|
/external/eigen/Eigen/src/PaStiXSupport/ |
PaStiXSupport.h | 105 ++mat.innerIndexPtr()[i]; 120 --mat.innerIndexPtr()[i]; 321 internal::eigen_pastix(&m_pastixdata, MPI_COMM_WORLD, m_size, mat.outerIndexPtr(), mat.innerIndexPtr(), 346 internal::eigen_pastix(&m_pastixdata, MPI_COMM_WORLD, m_size, mat.outerIndexPtr(), mat.innerIndexPtr(),
|
/external/eigen/Eigen/src/SparseCholesky/ |
SimplicialCholesky_impl.h | 112 StorageIndex* Li = m_matrix.innerIndexPtr();
|
/external/eigen/Eigen/src/SuperLUSupport/ |
SuperLUSupport.h | 214 res.storage.innerInd = mat.innerIndexPtr(); 273 res.storage.innerInd = mat.innerIndexPtr(); 726 int* Lrow = m_l.innerIndexPtr(); 730 int* Urow = m_u.innerIndexPtr(); 806 eigen_assert(m_u.innerIndexPtr()[lastId]<=j); 807 if (m_u.innerIndexPtr()[lastId]==j) [all...] |
/external/eigen/Eigen/src/IterativeLinearSolvers/ |
IncompleteCholesky.h | 219 Map<VectorIx> rowIdx(m_L.innerIndexPtr(), nnz); //Row indices 330 rowIdx = Map<const VectorIx>(L_save.innerIndexPtr(), nnz);
|
/external/eigen/test/ |
sparse_solvers.cpp | 64 //Index rows, Index cols, Index nnz, Index* outerIndexPtr, Index* innerIndexPtr, Scalar* valuePtr 65 MappedSparseMatrix<Scalar> mm2(rows, cols, cm2.nonZeros(), cm2.outerIndexPtr(), cm2.innerIndexPtr(), cm2.valuePtr());
|
sparse_basic.cpp | 401 Map<SparseMatrixType> mapMat2(m2.rows(), m2.cols(), m2.nonZeros(), m2.outerIndexPtr(), m2.innerIndexPtr(), m2.valuePtr(), m2.innerNonZeroPtr()); 402 Map<SparseMatrixType> mapMat3(m3.rows(), m3.cols(), m3.nonZeros(), m3.outerIndexPtr(), m3.innerIndexPtr(), m3.valuePtr(), m3.innerNonZeroPtr()); 407 MappedSparseMatrix<Scalar,SparseMatrixType::Options,StorageIndex> mapMat2(m2.rows(), m2.cols(), m2.nonZeros(), m2.outerIndexPtr(), m2.innerIndexPtr(), m2.valuePtr(), m2.innerNonZeroPtr()); 408 MappedSparseMatrix<Scalar,SparseMatrixType::Options,StorageIndex> mapMat3(m3.rows(), m3.cols(), m3.nonZeros(), m3.outerIndexPtr(), m3.innerIndexPtr(), m3.valuePtr(), m3.innerNonZeroPtr()); 418 Map<SparseMatrixType> mapMat2(rows, cols, m2.nonZeros(), m2.outerIndexPtr(), m2.innerIndexPtr(), m2.valuePtr(), m2.innerNonZeroPtr());
|
sparse_solver.h | 86 MappedSparseMatrix<Scalar,Mat::Options,StorageIndex> Am(A.rows(), A.cols(), A.nonZeros(), const_cast<StorageIndex*>(A.outerIndexPtr()), const_cast<StorageIndex*>(A.innerIndexPtr()), const_cast<Scalar*>(A.valuePtr()));
|
/external/eigen/Eigen/src/CholmodSupport/ |
CholmodSupport.h | 65 res.i = mat.innerIndexPtr();
|