/external/eigen/test/ |
mapstride.cpp | 79 if(k==1 && (m.innerSize()+1)*m.outerSize() > maxsize2) 83 Map<MatrixType, Alignment, OuterStride<Dynamic> > map(array, rows, cols, OuterStride<Dynamic>(m.innerSize()+1)); 85 VERIFY(map.outerStride() == map.innerSize()+1); 87 for(int j = 0; j < m.innerSize(); ++j) 101 if(k==1 && (m.innerSize()+4)*m.outerSize() > maxsize2) 106 InnerSize = MatrixType::InnerSizeAtCompileTime, 107 OuterStrideAtCompileTime = InnerSize==Dynamic ? Dynamic : InnerSize+4 110 map(array, rows, cols, OuterStride<OuterStrideAtCompileTime>(m.innerSize()+4)); 112 VERIFY(map.outerStride() == map.innerSize()+4) [all...] |
sparse.h | 72 for(Index i=0; i<sparseMat.innerSize(); i++) 125 for(int i=0; i<sparseMat.innerSize(); i++)
|
sparse_basic.cpp | 24 //const Index inner = ref.innerSize(); 130 VectorXi r(VectorXi::Constant(m2.outerSize(), ((mode%2)==0) ? int(m2.innerSize()) : std::max<int>(1,int(m2.innerSize())/8))); 279 std::vector<Scalar> ref_value(m2.innerSize()); 280 std::vector<Index> ref_index(m2.innerSize()); 326 m2.reserve(VectorXi::Constant(m2.outerSize(), int(m2.innerSize())));
|
sparse_block.cpp | 30 const Index inner = ref.innerSize();
|
/external/eigen/Eigen/src/Core/ |
CoreIterators.h | 42 : m_eval(xpr), m_iter(m_eval, outerId, xpr.innerSize()) 82 EIGEN_STRONG_INLINE inner_iterator_selector(const EvaluatorType &eval, const Index &outerId, const Index &innerSize) 83 : m_eval(eval), m_inner(0), m_outer(outerId), m_end(innerSize) 118 EIGEN_STRONG_INLINE inner_iterator_selector(const EvaluatorType &eval, const Index &outerId, const Index &/*innerSize*/)
|
Redux.h | 195 for(Index i = 1; i < mat.innerSize(); ++i) 198 for(Index j = 0; j < mat.innerSize(); ++j) 277 const Index innerSize = mat.innerSize(); 282 const Index packetedInnerSize = ((innerSize)/packetSize)*packetSize; 293 for(Index i=packetedInnerSize; i<innerSize; ++i) 360 EIGEN_DEVICE_FUNC Index innerSize() const { return m_xpr.innerSize(); }
|
AssignEvaluator.h | 48 InnerSize = int(Dst::IsVectorAtCompileTime) ? int(Dst::SizeAtCompileTime) 60 typedef typename find_best_packet<DstScalar,InnerSize>::type InnerPacketType; 82 && int(InnerSize)!=Dynamic && int(InnerSize)%int(InnerPacketSize)==0 93 indicated by InnerMaxSize rather than InnerSize, think of the case of a dynamic block 121 MayUnrollInner = int(InnerSize) != Dynamic 122 && int(InnerSize) * (int(DstEvaluator::CoeffReadCost)+int(SrcEvaluator::CoeffReadCost)) <= int(UnrollingLimit) 162 EIGEN_DEBUG_VAR(InnerSize) 325 for(Index inner = 0; inner < kernel.innerSize(); ++inner) { 453 const Index innerSize = kernel.innerSize() [all...] |
Assign_MKL.h | 49 InnerSize = int(Dst::IsVectorAtCompileTime) ? int(Dst::SizeAtCompileTime) 98 VMLOP( dst.innerSize(), (const VMLTYPE*)src_ptr, \ 160 VMLOP( dst.innerSize(), (const VMLTYPE*)src_ptr, exponent, \
|
Array.h | 248 EIGEN_DEVICE_FUNC inline Index outerStride() const { return this->innerSize(); }
|
Matrix.h | 383 EIGEN_DEVICE_FUNC inline Index outerStride() const { return this->innerSize(); }
|
DenseBase.h | 230 Index innerSize() const
|
/external/eigen/unsupported/Eigen/src/SparseExtra/ |
RandomSetter.h | 184 const Index outerSize = SwapStorage ? target.innerSize() : target.outerSize(); 185 const Index innerSize = SwapStorage ? target.outerSize() : target.innerSize(); 191 Index aux = innerSize - 1;
|
DynamicSparseMatrix.h | 84 inline Index innerSize() const { return m_innerSize; } 213 const Index innerSize = IsRowMajor ? cols : rows; 214 if (m_innerSize>innerSize) 216 // remove all coefficients with innerCoord>=innerSize 231 eigen_assert(innerSize()==0 && outerSize()==0);
|
BlockSparseMatrix.h | 741 return (IsColMajor ? innerSize() : outerSize()); 750 return (IsColMajor ? outerSize() : innerSize()); 753 inline Index innerSize() const 793 eigen_assert(inner < innerSize() && "OUTER INDEX OUT OF BOUNDS"); [all...] |
/external/eigen/Eigen/src/SparseCore/ |
SparseView.h | 61 inline Index innerSize() const { return m_matrix.innerSize(); } 153 : m_sve(sve), m_inner(0), m_outer(outer), m_end(sve.m_view.innerSize())
|
SparseTriangularView.h | 106 : Base(xprEval.m_argImpl,outer), m_returnOne(false), m_containsDiag(Base::outer()<xprEval.m_arg.innerSize())
|
SparseSparseProductWithPruning.h | 27 // make sure to call innerSize/outerSize since we fake the storage order. 28 Index rows = lhs.innerSize(); 31 eigen_assert(lhs.outerSize() == rhs.innerSize());
|
ConservativeSparseSparseProduct.h | 22 // make sure to call innerSize/outerSize since we fake the storage order. 23 Index rows = lhs.innerSize(); 25 eigen_assert(lhs.outerSize() == rhs.innerSize()); 268 eigen_assert(lhs.outerSize() == rhs.innerSize());
|
SparseVector.h | 83 EIGEN_STRONG_INLINE Index innerSize() const { return m_size; } 443 eigen_internal_assert(src.innerSize()==src.size());
|
SparseMap.h | 82 /** \copydoc SparseMatrixBase::innerSize() */ 83 inline Index innerSize() const { return m_innerSize; }
|
SparseMatrixBase.h | 187 Index innerSize() const { return (int(Flags)&RowMajorBit) ? this->cols() : this->rows(); }
|
SparseCwiseBinaryOp.h | 170 : m_lhsEval(aEval.m_lhsImpl), m_rhsIter(aEval.m_rhsImpl,outer), m_functor(aEval.m_functor), m_value(0), m_id(-1), m_innerSize(aEval.m_expr.rhs().innerSize()) 258 : m_lhsIter(aEval.m_lhsImpl,outer), m_rhsEval(aEval.m_rhsImpl), m_functor(aEval.m_functor), m_value(0), m_id(-1), m_innerSize(aEval.m_expr.lhs().innerSize())
|
SparseMatrix.h | 141 inline Index innerSize() const { return m_innerSize; } 229 * mode while reserving room for 2 x this->innerSize() non zeros if reserve(Index) has not been called earlier. [all...] |
/external/eigen/unsupported/Eigen/src/Skyline/ |
SkylineMatrix.h | 82 inline Index innerSize() const { 151 eigen_assert(inner < innerSize()); 200 eigen_assert(inner < innerSize()); 236 eigen_assert(idx < innerSize()); 245 eigen_assert(inner < innerSize()); 269 eigen_assert(inner < innerSize()); 289 eigen_assert(idx < innerSize()); 298 eigen_assert(inner < innerSize()); 317 eigen_assert(inner < innerSize()); 334 eigen_assert(inner < innerSize()); [all...] |
SkylineMatrixBase.h | 134 Index innerSize() const {
|