/external/eigen/failtest/ |
map_nonconst_ctor_on_const_ptr_3.cpp | 12 Map<MatrixXf, Aligned, InnerStride<2> > m(ptr, rows, cols, InnerStride<2>());
|
/external/eigen/bench/tensors/ |
tensor_benchmarks.h | 56 const TensorMap<Tensor<int, 2, 0, TensorIndex>, Eigen::Aligned> A((int*)a_, sizes); 57 TensorMap<Tensor<T, 2, 0, TensorIndex>, Eigen::Aligned> B(b_, sizes); 72 TensorMap<Tensor<T, 2>, Eigen::Aligned> C(c_, sizes); 87 const TensorMap<Tensor<T, 2>, Eigen::Aligned> A(a_, sizes); 88 const TensorMap<Tensor<T, 2>, Eigen::Aligned> B(b_, sizes); 89 TensorMap<Tensor<T, 2>, Eigen::Aligned> C(c_, sizes); 117 const TensorMap<Tensor<T, 2, 0, TensorIndex>, Eigen::Aligned> B(b_, input_size); 120 TensorMap<Tensor<T, 1, 0, TensorIndex>, Eigen::Aligned> C(c_, output_size); 134 const TensorMap<Tensor<T, 2, 0, TensorIndex>, Eigen::Aligned> B(b_, input_size); 137 TensorMap<Tensor<T, 1, 0, TensorIndex>, Eigen::Aligned> C(c_, output_size) [all...] |
/external/eigen/test/ |
mapstride.cpp | 25 if(Alignment!=Aligned) 65 if(Alignment!=Aligned) 70 if(Alignment!=Aligned) 153 CALL_SUBTEST_1( map_class_vector<Aligned>(Matrix<float, 1, 1>()) ); 155 CALL_SUBTEST_2( map_class_vector<Aligned>(Vector4d()) ); 157 CALL_SUBTEST_3( map_class_vector<Aligned>(RowVector4f()) ); 159 CALL_SUBTEST_4( map_class_vector<Aligned>(VectorXcf(internal::random<int>(1,maxn))) ); 161 CALL_SUBTEST_5( map_class_vector<Aligned>(VectorXi(internal::random<int>(1,maxn))) ); 164 CALL_SUBTEST_1( map_class_matrix<Aligned>(Matrix<float, 1, 1>()) ); 166 CALL_SUBTEST_2( map_class_matrix<Aligned>(Matrix4d()) ) [all...] |
geo_quaternion.cpp | 177 typedef Map<Quaternion<Scalar>, Aligned> MQuaternionA; 178 typedef Map<const Quaternion<Scalar>, Aligned> MCQuaternionA; 270 VERIFY( !(internal::traits<Map<ConstPlainObjectType, Aligned> >::Flags & LvalueBit) ); 272 VERIFY( !(Map<ConstPlainObjectType, Aligned>::Flags & LvalueBit) );
|
sparse_ref.cpp | 39 VERIFY( !(internal::traits<Ref<ConstPlainObjectType, Aligned> >::Flags & LvalueBit) ); 41 VERIFY( !(Ref<ConstPlainObjectType, Aligned>::Flags & LvalueBit) );
|
ref.cpp | 147 VERIFY( !(internal::traits<Ref<ConstPlainObjectType, Aligned> >::Flags & LvalueBit) ); 149 VERIFY( !(Ref<ConstPlainObjectType, Aligned>::Flags & LvalueBit) );
|
permutationmatrices.cpp | 126 typedef Map<MatrixType, Aligned, S> MapType;
|
/external/eigen/Eigen/src/Core/products/ |
GeneralMatrixVector.h | 40 * - perform half aligned and half unaligned loads (-> EvenAligned case) 43 * - one over 4 consecutive columns is guaranteed to be aligned with the result vector, 44 * perform simple aligned loads for this column and aligned loads plus re-alignment for the other. (-> FirstAligned case) 53 * One might also wonder why in the EvenAligned case we perform unaligned loads instead of using the aligned-loads plus re-alignment 126 // How many coeffs of the result do we have to skip to be aligned. 127 // Here we assume data are at least aligned on the base scalar type. 137 // we cannot assume the first element is aligned because of sub-matrices 140 // find how many columns do we have to skip to be aligned with the result (if possible) 142 // if the data cannot be aligned (TODO add some compile time tests when possible, e.g. for floats [all...] |
/device/linaro/bootloader/edk2/UefiCpuPkg/CpuIoPei/ |
CpuIoPei.c | 216 BOOLEAN Aligned;
230 Aligned = (BOOLEAN)(((UINTN)Buffer & (mInStride[OperationWidth] - 1)) == 0x00);
235 if (Aligned) {
241 if (Aligned) {
247 if (Aligned) {
290 BOOLEAN Aligned;
304 Aligned = (BOOLEAN)(((UINTN)Buffer & (mInStride[OperationWidth] - 1)) == 0x00);
309 if (Aligned) {
315 if (Aligned) {
321 if (Aligned) {
[all...] |
/external/eigen/Eigen/src/Core/ |
ForceAlignedAccess.h | 18 * \brief Enforce aligned packet loads and stores regardless of what is requested 20 * \param ExpressionType the type of the object of which we are forcing aligned packet access 72 return m_expression.template packet<Aligned>(row, col); 78 m_expression.const_cast_derived().template writePacket<Aligned>(row, col, x); 84 return m_expression.template packet<Aligned>(index); 90 m_expression.const_cast_derived().template writePacket<Aligned>(index, x); 102 /** \returns an expression of *this with forced aligned access 112 /** \returns an expression of *this with forced aligned access 122 /** \returns an expression of *this with forced aligned access if \a Enable is true. 133 /** \returns an expression of *this with forced aligned access if \a Enable is true [all...] |
/external/valgrind/memcheck/tests/amd64/ |
sh-mem-vec256-plo-yes.stderr.exp | 62 ------ PL Aligned case with 0 leading acc+def bytes ------ 71 ------ PL Aligned case with 1 leading acc+def bytes ------ 80 ------ PL Aligned case with 2 leading acc+def bytes ------ 89 ------ PL Aligned case with 3 leading acc+def bytes ------ 98 ------ PL Aligned case with 4 leading acc+def bytes ------ 107 ------ PL Aligned case with 5 leading acc+def bytes ------ 116 ------ PL Aligned case with 6 leading acc+def bytes ------ 125 ------ PL Aligned case with 7 leading acc+def bytes ------ 134 ------ PL Aligned case with 8 leading acc+def bytes ------ 143 ------ PL Aligned case with 9 leading acc+def bytes ----- [all...] |
sh-mem-vec256-plo-no.stderr.exp | 62 ------ PL Aligned case with 0 leading acc+def bytes ------ 78 ------ PL Aligned case with 1 leading acc+def bytes ------ 94 ------ PL Aligned case with 2 leading acc+def bytes ------ 110 ------ PL Aligned case with 3 leading acc+def bytes ------ 126 ------ PL Aligned case with 4 leading acc+def bytes ------ 142 ------ PL Aligned case with 5 leading acc+def bytes ------ 158 ------ PL Aligned case with 6 leading acc+def bytes ------ 174 ------ PL Aligned case with 7 leading acc+def bytes ------ 190 ------ PL Aligned case with 8 leading acc+def bytes ------ 206 ------ PL Aligned case with 9 leading acc+def bytes ----- [all...] |
/external/eigen/bench/ |
quatmul.cpp | 17 c = internal::quat_product<0, Quat, Quat, typename Quat::Scalar, Aligned>::run(a,b);
|
/external/google-breakpad/src/common/dwarf/ |
bytereader_unittest.cc | 629 class Aligned: public AlignedFixture, public Test { }; 631 TEST_F(Aligned, DW_EH_PE_aligned0) { 639 TEST_F(Aligned, DW_EH_PE_aligned1) { 647 TEST_F(Aligned, DW_EH_PE_aligned2) { 655 TEST_F(Aligned, DW_EH_PE_aligned3) { 663 TEST_F(Aligned, DW_EH_PE_aligned11) { 672 TEST_F(Aligned, DW_EH_PE_aligned30) { 681 TEST_F(Aligned, DW_EH_PE_aligned23) { 690 TEST_F(Aligned, DW_EH_PE_aligned03) {
|
/external/eigen/Eigen/src/Geometry/ |
Quaternion.h | 313 struct traits<Map<Quaternion<_Scalar>, _Options> > : traits<Quaternion<_Scalar, (int(_Options)&Aligned)==Aligned ? AutoAlign : DontAlign> > 321 struct traits<Map<const Quaternion<_Scalar>, _Options> > : traits<Quaternion<_Scalar, (int(_Options)&Aligned)==Aligned ? AutoAlign : DontAlign> > 324 typedef traits<Quaternion<_Scalar, (int(_Options)&Aligned)==Aligned ? AutoAlign : DontAlign> > TraitsBase; 359 * If the template parameter _Options is set to #Aligned, then the pointer coeffs must be aligned. */ 396 * If the template parameter _Options is set to #Aligned, then the pointer coeffs must be aligned. * [all...] |
/external/eigen/unsupported/Eigen/CXX11/src/Tensor/ |
TensorAssign.h | 139 const int LhsStoreMode = TensorEvaluator<LeftArgType, Device>::IsAligned ? Aligned : Unaligned; 140 const int RhsLoadMode = TensorEvaluator<RightArgType, Device>::IsAligned ? Aligned : Unaligned;
|
TensorEvalTo.h | 137 internal::pstoret<CoeffReturnType, PacketReturnType, Aligned>(m_buffer + i, m_impl.template packet<TensorEvaluator<ArgType, Device>::IsAligned ? Aligned : Unaligned>(i));
|
TensorContractionMapper.h | 201 // matrix-vector product relies on the stride when dealing with aligned inputs. 202 return (Alignment == Aligned) && (side == Lhs) && inner_dim_contiguous ? 0 : size; 412 const int ActualAlignment = (AlignmentType == Aligned) && (Alignment == Aligned) ? Aligned : Unaligned; 420 EIGEN_DEVICE_FUNC EIGEN_ALWAYS_INLINE bool aligned(Index) const { function in class:Eigen::internal::TensorContractionSubMapper
|
TensorMap.h | 53 IsAligned = ((int(Options_)&Aligned)==Aligned),
|
/external/clang/lib/Sema/ |
SemaTemplateInstantiateDecl.cpp | 80 const AlignedAttr *Aligned, Decl *New, bool IsPackExpansion) { 81 if (Aligned->isAlignmentExpr()) { 84 ExprResult Result = S.SubstExpr(Aligned->getAlignmentExpr(), TemplateArgs); 86 S.AddAlignedAttr(Aligned->getLocation(), New, Result.getAs<Expr>(), 87 Aligned->getSpellingListIndex(), IsPackExpansion); 89 TypeSourceInfo *Result = S.SubstType(Aligned->getAlignmentType(), 90 TemplateArgs, Aligned->getLocation(), 93 S.AddAlignedAttr(Aligned->getLocation(), New, Result, 94 Aligned->getSpellingListIndex(), IsPackExpansion); 100 const AlignedAttr *Aligned, Decl *New) [all...] |
/external/eigen/Eigen/src/Core/util/ |
Constants.h | 154 * means the first coefficient packet is guaranteed to be aligned. 156 * In other words, this means we are allow to perform an aligned packet access to the first element regardless 159 * expression.packet<Aligned>(0); 226 * Enum for indicating whether a buffer is aligned or not. */ 229 Aligned8=8, /**< Data pointer is aligned on a 8 bytes boundary. */ 230 Aligned16=16, /**< Data pointer is aligned on a 16 bytes boundary. */ 231 Aligned32=32, /**< Data pointer is aligned on a 32 bytes boundary. */ 232 Aligned64=64, /**< Data pointer is aligned on a 64 bytes boundary. */ 233 Aligned128=128, /**< Data pointer is aligned on a 128 bytes boundary. */ 235 Aligned=16, /**< \deprecated Synonym for Aligned16. * [all...] |
/external/eigen/bench/btl/libs/eigen3/ |
eigen3_interface.hh | 141 dst.template copyPacket<Src, Aligned, Unaligned>(j, index, src); 143 dst.template copyPacket<Src, Aligned, Unaligned>(index, j, src);
|
/external/eigen/Eigen/src/SparseLU/ |
SparseLU_panel_bmod.h | 105 Map<ScalarMatrix, Aligned, OuterStride<> > U(tempv.data(), u_rows, u_cols, OuterStride<>(ldu));
|
/external/llvm/lib/Target/X86/ |
X86FastISel.cpp | 92 MachineMemOperand *MMO = nullptr, bool Aligned = false); 95 MachineMemOperand *MMO = nullptr, bool Aligned = false); 503 MachineMemOperand *MMO, bool Aligned) { 551 if (Aligned) { 560 if (Aligned) { 572 if (Aligned) { 582 if (Aligned) 589 if (Aligned) { 599 if (Aligned) 606 if (Aligned) [all...] |
/external/eigen/Eigen/src/Core/arch/CUDA/ |
PacketMath.h | 198 EIGEN_DEVICE_FUNC EIGEN_ALWAYS_INLINE float4 ploadt_ro<float4, Aligned>(const float* from) { 206 EIGEN_DEVICE_FUNC EIGEN_ALWAYS_INLINE double2 ploadt_ro<double2, Aligned>(const double* from) {
|