/external/eigen/test/ |
vectorization_logic.cpp | 102 PacketSize = internal::unpacket_traits<PacketType>::size, 108 typedef Matrix<Scalar,PacketSize,1> Vector1; 111 typedef Matrix<Scalar,PacketSize,PacketSize> Matrix11; 112 typedef Matrix<Scalar,2*PacketSize,2*PacketSize> Matrix22; 113 typedef Matrix<Scalar,(Matrix11::Flags&RowMajorBit)?16:4*PacketSize,(Matrix11::Flags&RowMajorBit)?4*PacketSize:16> Matrix44; 114 typedef Matrix<Scalar,(Matrix11::Flags&RowMajorBit)?16:4*PacketSize,(Matrix11::Flags&RowMajorBit)?4*PacketSize:16,DontAlign|EIGEN_DEFAULT_MATRIX_STORAGE_ORDER_OPTION> Matrix44u [all...] |
packetmath.cpp | 65 for (int i=0; i<PacketSize; ++i) \ 68 VERIFY(areApprox(ref, data2, PacketSize) && #POP); \ 93 for (int i=0; i<PacketSize; ++i) \ 96 VERIFY(areApprox(ref, data2, PacketSize) && #POP); \ 101 for (int i=0; i<PacketSize; ++i) \ 102 ref[i] = REFOP(data1[i], data1[i+PacketSize]); \ 103 h.store(data2, POP(h.load(data1),h.load(data1+PacketSize))); \ 104 VERIFY(areApprox(ref, data2, PacketSize) && #POP); \ 117 const int PacketSize = PacketTraits::size; 120 const int max_size = PacketSize > 4 ? PacketSize : 4 [all...] |
adjoint.cpp | 78 const Index PacketSize = internal::packet_traits<Scalar>::size; 124 if(PacketSize<m3.rows() && PacketSize<m3.cols()) 127 Index i = internal::random<Index>(0,m3.rows()-PacketSize); 128 Index j = internal::random<Index>(0,m3.cols()-PacketSize); 129 m3.template block<PacketSize,PacketSize>(i,j).transposeInPlace(); 130 VERIFY_IS_APPROX( (m3.template block<PacketSize,PacketSize>(i,j)), (m1.template block<PacketSize,PacketSize>(i,j).transpose()) ) [all...] |
/external/eigen/unsupported/Eigen/CXX11/src/Tensor/ |
TensorExecutor.h | 61 const int PacketSize = unpacket_traits<typename TensorEvaluator<Expression, DefaultDevice>::PacketReturnType>::size; 65 const Index UnrolledSize = (size / (4 * PacketSize)) * 4 * PacketSize; 66 for (Index i = 0; i < UnrolledSize; i += 4*PacketSize) { 68 evaluator.evalPacket(i + j * PacketSize); 71 const Index VectorizedSize = (size / PacketSize) * PacketSize; 72 for (Index i = UnrolledSize; i < VectorizedSize; i += PacketSize) { 104 static const int PacketSize = unpacket_traits<typename Evaluator::PacketReturnType>::size; 110 if (last - first >= PacketSize) { [all...] |
TensorStriding.h | 109 static const int PacketSize = internal::unpacket_traits<PacketReturnType>::size; 167 EIGEN_STATIC_ASSERT((PacketSize > 1), YOU_MADE_A_PROGRAMMING_MISTAKE) 168 eigen_assert(index+PacketSize-1 < dimensions().TotalSize()); 171 Index indices[] = {index, index + PacketSize - 1}; 195 if (inputIndices[1] - inputIndices[0] == PacketSize - 1) { 200 EIGEN_ALIGN_MAX typename internal::remove_const<CoeffReturnType>::type values[PacketSize]; 202 values[PacketSize-1] = m_impl.coeff(inputIndices[1]); 203 for (int i = 1; i < PacketSize-1; ++i) { 222 TensorOpCost(0, 0, compute_cost, vectorized, PacketSize); 282 static const int PacketSize = internal::unpacket_traits<PacketReturnType>::size [all...] |
TensorReverse.h | 110 static const int PacketSize = internal::unpacket_traits<PacketReturnType>::size; 198 EIGEN_STATIC_ASSERT((PacketSize > 1), YOU_MADE_A_PROGRAMMING_MISTAKE) 199 eigen_assert(index+PacketSize-1 < dimensions().TotalSize()); 204 values[PacketSize]; 205 for (int i = 0; i < PacketSize; ++i) { 222 TensorOpCost(0, 0, compute_cost, false /* vectorized */, PacketSize); 261 static const int PacketSize = internal::unpacket_traits<PacketReturnType>::size; 272 EIGEN_STATIC_ASSERT((PacketSize > 1), YOU_MADE_A_PROGRAMMING_MISTAKE) 273 eigen_assert(index+PacketSize-1 < dimensions().TotalSize()); 276 EIGEN_ALIGN_MAX CoeffReturnType values[PacketSize]; [all...] |
TensorShuffling.h | 109 static const int PacketSize = internal::unpacket_traits<PacketReturnType>::size; 169 EIGEN_STATIC_ASSERT((PacketSize > 1), YOU_MADE_A_PROGRAMMING_MISTAKE) 170 eigen_assert(index+PacketSize-1 < dimensions().TotalSize()); 172 EIGEN_ALIGN_MAX typename internal::remove_const<CoeffReturnType>::type values[PacketSize]; 173 for (int i = 0; i < PacketSize; ++i) { 185 TensorOpCost(0, 0, compute_cost, false /* vectorized */, PacketSize); 231 static const int PacketSize = internal::unpacket_traits<PacketReturnType>::size; 251 EIGEN_STATIC_ASSERT((PacketSize > 1), YOU_MADE_A_PROGRAMMING_MISTAKE) 253 EIGEN_ALIGN_MAX typename internal::remove_const<CoeffReturnType>::type values[PacketSize]; 255 for (int i = 0; i < PacketSize; ++i) [all...] |
TensorBroadcasting.h | 106 static const int PacketSize = internal::unpacket_traits<PacketReturnType>::size; 250 EIGEN_STATIC_ASSERT((PacketSize > 1), YOU_MADE_A_PROGRAMMING_MISTAKE) 251 eigen_assert(index+PacketSize-1 < dimensions().TotalSize()); 286 if (innermostLoc + PacketSize <= m_impl.dimensions()[0]) { 289 EIGEN_ALIGN_MAX typename internal::remove_const<CoeffReturnType>::type values[PacketSize]; 291 for (int i = 1; i < PacketSize; ++i) { 302 EIGEN_STATIC_ASSERT((PacketSize > 1), YOU_MADE_A_PROGRAMMING_MISTAKE) 303 eigen_assert(index+PacketSize-1 < dimensions().TotalSize()); 338 if (innermostLoc + PacketSize <= m_impl.dimensions()[NumDims-1]) { 341 EIGEN_ALIGN_MAX typename internal::remove_const<CoeffReturnType>::type values[PacketSize]; [all...] |
TensorChipping.h | 139 static const int PacketSize = internal::unpacket_traits<PacketReturnType>::size; 205 EIGEN_STATIC_ASSERT((PacketSize > 1), YOU_MADE_A_PROGRAMMING_MISTAKE) 206 eigen_assert(index+PacketSize-1 < dimensions().TotalSize()); 213 EIGEN_ALIGN_MAX typename internal::remove_const<CoeffReturnType>::type values[PacketSize]; 214 for (int i = 0; i < PacketSize; ++i) { 228 if (rem + PacketSize <= m_stride) { 233 EIGEN_ALIGN_MAX typename internal::remove_const<CoeffReturnType>::type values[PacketSize]; 234 for (int i = 0; i < PacketSize; ++i) { 263 TensorOpCost(0, 0, cost, vectorized, PacketSize); 324 static const int PacketSize = internal::unpacket_traits<PacketReturnType>::size [all...] |
TensorPatch.h | 90 static const int PacketSize = internal::unpacket_traits<PacketReturnType>::size; 187 EIGEN_STATIC_ASSERT((PacketSize > 1), YOU_MADE_A_PROGRAMMING_MISTAKE) 188 eigen_assert(index+PacketSize-1 < dimensions().TotalSize()); 191 Index indices[2] = {index, index + PacketSize - 1}; 232 if (inputIndices[1] - inputIndices[0] == PacketSize - 1) { 237 EIGEN_ALIGN_MAX CoeffReturnType values[PacketSize]; 239 values[PacketSize-1] = m_impl.coeff(inputIndices[1]); 240 for (int i = 1; i < PacketSize-1; ++i) { 253 TensorOpCost(0, 0, compute_cost, vectorized, PacketSize);
|
TensorInflation.h | 87 static const int PacketSize = internal::unpacket_traits<PacketReturnType>::size; 192 EIGEN_STATIC_ASSERT((PacketSize > 1), YOU_MADE_A_PROGRAMMING_MISTAKE) 193 eigen_assert(index+PacketSize-1 < dimensions().TotalSize()); 195 EIGEN_ALIGN_MAX typename internal::remove_const<CoeffReturnType>::type values[PacketSize]; 196 for (int i = 0; i < PacketSize; ++i) { 213 compute_cost, vectorized, PacketSize);
|
TensorPadding.h | 93 static const int PacketSize = internal::unpacket_traits<PacketReturnType>::size; 260 EIGEN_STATIC_ASSERT((PacketSize > 1), YOU_MADE_A_PROGRAMMING_MISTAKE) 261 eigen_assert(index+PacketSize-1 < dimensions().TotalSize()); 267 const Index last = index + PacketSize - 1; 292 const Index last = index + PacketSize - 1; 317 EIGEN_STATIC_ASSERT((PacketSize > 1), YOU_MADE_A_PROGRAMMING_MISTAKE) 318 eigen_assert(index+PacketSize-1 < dimensions().TotalSize()); 325 const Index last = index + PacketSize - 1; 350 const Index last = index + PacketSize - 1; 375 EIGEN_ALIGN_MAX typename internal::remove_const<CoeffReturnType>::type values[PacketSize]; [all...] |
/external/eigen/bench/ |
benchVecAdd.cpp | 94 const int PacketSize = internal::packet_traits<Scalar>::size; 97 for (int i=0; i<size; i+=PacketSize*8) 101 // a1 = internal::pload(&a[i+1*PacketSize]); 102 // b1 = internal::pload(&b[i+1*PacketSize]); 103 // a2 = internal::pload(&a[i+2*PacketSize]); 104 // b2 = internal::pload(&b[i+2*PacketSize]); 105 // a3 = internal::pload(&a[i+3*PacketSize]); 106 // b3 = internal::pload(&b[i+3*PacketSize]); 108 // a0 = internal::pload(&a[i+4*PacketSize]); 109 // b0 = internal::pload(&b[i+4*PacketSize]); [all...] |
/external/tensorflow/tensorflow/core/kernels/ |
sparse_matmul_op_test.cc | 268 : PacketSize(Eigen::internal::packet_traits<float>::size) { 272 data1[i] = internal::random<float>() / RealFloat(PacketSize); 273 data2[i] = internal::random<float>() / RealFloat(PacketSize); 274 data3[i] = internal::random<float>() / RealFloat(PacketSize); 277 data3[i] = internal::random<float>() / RealFloat(PacketSize); 317 const int PacketSize; 330 for (int i = 0; i < PacketSize; ++i) ref[i] = data1[0]; 333 ASSERT_TRUE(areApprox(ref, data2, PacketSize)); 334 if (PacketSize > 1) { 335 for (int i = 0; i < PacketSize; ++i) ref[i] = data1[1] [all...] |
/external/eigen/Eigen/src/SparseLU/ |
SparseLU_gemm_kernel.h | 33 PacketSize = packet_traits<Scalar>::size, 38 SM = PM*PacketSize // step along M 44 eigen_internal_assert(((lda%PacketSize)==0) && ((ldc%PacketSize)==0) && (i0==internal::first_default_aligned(C,m))); 62 Index actual_b_end2 = (actual_b/PacketSize)*PacketSize; // actual number of rows suitable for vectorization 109 c0 = pload<Packet>(C0+i+(I)*PacketSize); \ 110 c1 = pload<Packet>(C1+i+(I)*PacketSize); \ 113 a0 = pload<Packet>(A0+i+(I+1)*PacketSize); \ 116 a1 = pload<Packet>(A1+i+(I+1)*PacketSize); \ [all...] |
SparseLU_kernel_bmod.h | 65 const Index PacketSize = internal::packet_traits<Scalar>::size; 66 Index ldl = internal::first_multiple(nrow, PacketSize); 68 Index aligned_offset = internal::first_default_aligned(tempv.data()+segsize, PacketSize); 69 Index aligned_with_B_offset = (PacketSize-internal::first_default_aligned(B.data(), PacketSize))%PacketSize;
|
SparseLU_panel_bmod.h | 69 const Index PacketSize = internal::packet_traits<Scalar>::size; 104 Index ldu = internal::first_multiple<Index>(u_rows, PacketSize); 147 Index ldl = internal::first_multiple<Index>(nrow, PacketSize); 148 Index offset = (PacketSize-internal::first_default_aligned(B.data(), PacketSize)) % PacketSize;
|
/external/eigen/Eigen/src/Jacobi/ |
Jacobi.h | 306 PacketSize = packet_traits<Scalar>::size, 328 (PacketSize == OtherPacketSize) && 329 ((incrx==1 && incry==1) || PacketSize == 1)) 335 Index alignedEnd = alignedStart + ((size-alignedStart)/PacketSize)*PacketSize; 355 for(Index i=alignedStart; i<alignedEnd; i+=PacketSize) 361 px += PacketSize; 362 py += PacketSize; 367 Index peelingEnd = alignedStart + ((size-alignedStart)/(Peeling*PacketSize))*(Peeling*PacketSize); [all...] |
/prebuilts/gcc/linux-x86/host/x86_64-w64-mingw32-4.8/x86_64-w64-mingw32/include/ |
dhcpssdk.h | 75 typedef DWORD (WINAPI *LPDHCP_NEWPKT)(LPBYTE *Packet,DWORD *PacketSize,DWORD IpAddress,LPVOID Reserved,LPVOID *PktContext,LPBOOL ProcessIt); 76 typedef DWORD (WINAPI *LPDHCP_DROP_SEND)(LPBYTE *Packet,DWORD *PacketSize,DWORD ControlCode,DWORD IpAddress,LPVOID Reserved,LPVOID PktContext); 77 typedef DWORD (WINAPI *LPDHCP_PROB)(LPBYTE Packet,DWORD PacketSize,DWORD ControlCode,DWORD IpAddress,DWORD AltAddress,LPVOID Reserved,LPVOID PktContext); 78 typedef DWORD (WINAPI *LPDHCP_GIVE_ADDRESS)(LPBYTE Packet,DWORD PacketSize,DWORD ControlCode,DWORD IpAddress,DWORD AltAddress,DWORD AddrType,DWORD LeaseTime,LPVOID Reserved,LPVOID PktContext); 79 typedef DWORD (WINAPI *LPDHCP_HANDLE_OPTIONS)(LPBYTE Packet,DWORD PacketSize,LPVOID Reserved,LPVOID PktContext,LPDHCP_SERVER_OPTIONS ServerOptions);
|
/device/linaro/bootloader/edk2/NetworkPkg/IpSecDxe/ |
IpSecDebug.c | 204 UINTN PacketSize;
210 PacketSize = Packet->PayloadTotalSize + sizeof (IKE_HEADER);
249 DEBUG ((DEBUG_INFO, "\n>>>Sending %d bytes %a to ", PacketSize, TypeStr));
251 DEBUG ((DEBUG_INFO, "\n>>>Receiving %d bytes %a from ", PacketSize, TypeStr));
|
/external/eigen/Eigen/src/Core/ |
Redux.h | 32 PacketSize = unpacket_traits<PacketType>::size, 42 MaySliceVectorize = bool(MightVectorize) && int(InnerMaxSize)>=3*PacketSize 56 UnrollingLimit = EIGEN_UNROLLING_LIMIT * (int(Traversal) == int(DefaultTraversal) ? 1 : int(PacketSize)) 72 EIGEN_DEBUG_VAR(PacketSize) 141 PacketSize = redux_traits<Func, Derived>::PacketSize, 160 index = Start * redux_traits<Func, Derived>::PacketSize, 219 const Index packetSize = redux_traits<Func, Derived>::PacketSize; 226 const Index alignedSize2 = ((size-alignedStart)/(2*packetSize))*(2*packetSize) [all...] |
Reverse.h | 75 PacketSize = internal::packet_traits<Scalar>::size, 79 OffsetRow = ReverseRow && IsColMajor ? PacketSize : 1, 80 OffsetCol = ReverseCol && IsRowMajor ? PacketSize : 1,
|
/external/eigen/Eigen/src/Core/products/ |
SelfadjointMatrixMatrix.h | 48 enum { PacketSize = packet_traits<Scalar>::size }; 53 const Index peeled_mc3 = Pack1>=3*PacketSize ? (rows/(3*PacketSize))*(3*PacketSize) : 0; 54 const Index peeled_mc2 = Pack1>=2*PacketSize ? peeled_mc3+((rows-peeled_mc3)/(2*PacketSize))*(2*PacketSize) : 0; 55 const Index peeled_mc1 = Pack1>=1*PacketSize ? (rows/(1*PacketSize))*(1*PacketSize) : 0 [all...] |
SelfadjointMatrixVector.h | 48 const Index PacketSize = sizeof(Packet)/sizeof(Scalar); 89 Index alignedEnd = alignedStart + ((endi-alignedStart)/(PacketSize))*(PacketSize); 116 for (Index i=alignedStart; i<alignedEnd; i+=PacketSize) 118 Packet A0i = ploadu<Packet>(a0It); a0It += PacketSize; 119 Packet A1i = ploadu<Packet>(a1It); a1It += PacketSize; 120 Packet Bi = ploadu<Packet>(rhsIt); rhsIt += PacketSize; // FIXME should be aligned in most cases 126 pstore(resIt,Xi); resIt += PacketSize;
|
GeneralBlockPanelKernel.h | [all...] |