/external/eigen/test/ |
packetmath.cpp | 202 Packet A0, A1, A2, A3; 203 internal::pbroadcast4<Packet>(data1, A0, A1, A2, A3); 204 internal::pstore(data2+0*PacketSize, A0); 214 Packet A0, A1; 215 internal::pbroadcast2<Packet>(data1, A0, A1); 216 internal::pstore(data2+0*PacketSize, A0);
|
/external/llvm/lib/Transforms/Scalar/ |
MergedLoadStoreMotion.cpp | 300 auto *A0 = dyn_cast<Instruction>(L0->getPointerOperand()); 302 if (A0 && A1 && A0->isIdenticalTo(A1) && isSafeToHoist(A0) && 303 A0->hasOneUse() && (A0->getParent() == L0->getParent()) && 305 isa<GetElementPtrInst>(A0)) { 309 hoistInstruction(BB, A0, A1); 426 auto *A0 = dyn_cast<Instruction>(S0->getPointerOperand()); 428 if (A0 && A1 && A0->isIdenticalTo(A1) && A0->hasOneUse() & [all...] |
/external/webp/src/dsp/ |
common_sse2.h | 167 const __m128i A0 = _mm_unpacklo_epi8(*in0, *in1); 171 const __m128i B0 = _mm_unpacklo_epi8(A0, A1); 172 const __m128i B1 = _mm_unpackhi_epi8(A0, A1); 176 // C1 = a7 a6 ... a1 a0 | r7 r6 ... r1 r0
|
filters_neon.c | 218 const uint8x16_t A0 = vld1q_u8(&in[i]); 219 const uint8x16_t A1 = vaddq_u8(A0, last);
|
filters_sse2.c | 41 const __m128i A0 = _mm_loadu_si128((const __m128i*)&src[i + 0]); 45 const __m128i C0 = _mm_sub_epi8(A0, B0); 59 const __m128i A0 = _mm_loadu_si128((const __m128i*)(src + i + 0 )); 63 const __m128i C0 = _mm_sub_epi8(A0, B0); 152 const __m128i A0 = _mm_loadl_epi64((const __m128i*)&row[i - 1]); 156 const __m128i A1 = _mm_unpacklo_epi8(A0, zero); 230 const __m128i A0 = _mm_loadl_epi64((const __m128i*)(in + i)); 231 const __m128i A1 = _mm_add_epi8(A0, last); 253 const __m128i A0 = _mm_loadu_si128((const __m128i*)&in[i + 0]); 257 const __m128i C0 = _mm_add_epi8(A0, B0) [all...] |
rescaler_neon.c | 47 LOAD_32x4(frow, A0); 49 const uint64x2_t C0 = vmull_n_u32(vget_low_u32(A0), A); 50 const uint64x2_t C1 = vmull_n_u32(vget_high_u32(A0), A); 74 LOAD_32x4(frow + x_out + 0, A0); 76 const uint32x4_t B0 = MULT_FIX(A0, fy_scale_half); 134 const uint32x4_t A0 = MULT_FIX(in0, yscale_half); 136 const uint32x4_t B0 = vqsubq_u32(in2, A0); 144 STORE_32x8(A0, A1, irow + x_out); 156 const uint32x4_t A0 = MULT_FIX(in0, fxy_scale_half); 158 const uint16x4_t B0 = vmovn_u32(A0); [all...] |
alpha_processing_sse2.c | 42 const __m128i a0 = _mm_loadl_epi64((const __m128i*)&alpha[i]); local 43 const __m128i a1 = _mm_unpacklo_epi8(a0, zero); 59 all_alphas = _mm_and_si128(all_alphas, a0); 83 const __m128i a0 = _mm_loadu_si128((const __m128i*)&alpha[i]); local 84 const __m128i a1 = _mm_unpacklo_epi8(zero, a0); // note the 'zero' first! 85 const __m128i b1 = _mm_unpackhi_epi8(zero, a0); 121 const __m128i a0 = _mm_loadu_si128(src + 0); local 123 const __m128i b0 = _mm_and_si128(a0, a_mask); 166 /* alpha2 = [ff a0 a0 a0][ff a1 a1 a1] */ 238 const __m128i a0 = _mm_loadu_si128((const __m128i*)(src + i + 0)); local 254 const __m128i a0 = _mm_loadu_si128((const __m128i*)(src + i + 0)); local 308 const __m128i a0 = _mm_loadl_epi64((const __m128i*)&alpha[x]); local [all...] |
rescaler_sse2.c | 149 const __m128i A0 = _mm_mullo_epi16(sum, mult0); 151 const __m128i B0 = _mm_unpacklo_epi16(A0, A1); // sum * x_sub 178 const __m128i A0 = _mm_loadu_si128((const __m128i*)(src + 0)); 180 const __m128i A2 = _mm_srli_epi64(A0, 32); 183 *out0 = _mm_mul_epu32(A0, *mult); 188 *out0 = A0; 195 static WEBP_INLINE void ProcessRow_SSE2(const __m128i* const A0, 203 const __m128i B0 = _mm_mul_epu32(*A0, *mult); 242 __m128i A0, A1, A2, A3; 243 LoadDispatchAndMult_SSE2(frow + x_out, NULL, &A0, &A1, &A2, &A3) [all...] |
lossless_neon.c | 154 static WEBP_INLINE uint8x8_t Average2_u8_NEON(uint32_t a0, uint32_t a1) { 155 const uint8x8_t A0 = LOAD_U32_AS_U8(a0); 157 return vhadd_u8(A0, A1); 177 static WEBP_INLINE uint32_t Average2_NEON(uint32_t a0, uint32_t a1) { 178 const uint8x8_t avg_u8x8 = Average2_u8_NEON(a0, a1); 183 static WEBP_INLINE uint32_t Average3_NEON(uint32_t a0, uint32_t a1, 185 const uint8x8_t avg0 = Average2_u8_NEON(a0, a2);
|
lossless_enc_sse2.c | 104 const __m128i A0 = _mm_slli_epi16(in0, 8); // r 0 | b 0 108 const __m128i C0 = _mm_mulhi_epi16(A0, mults_r); // x db | 0 0 152 const __m128i A0 = _mm_and_si128(in0, mask_g); // 0 0 | g 0 156 const __m128i C0 = _mm_mulhi_epi16(A0, mults_g); // 0 0 | x dr 186 const __m128i a0 = _mm_loadu_si128((const __m128i*)&a[i + 0]); local 198 _mm_storeu_si128((__m128i*)&out[i + 0], _mm_add_epi32(a0, b0)); 211 const __m128i a0 = _mm_loadu_si128((const __m128i*)&a[i + 0]); local 223 _mm_storeu_si128((__m128i*)&out[i + 0], _mm_add_epi32(a0, b0)); 341 __m128i A0 = _mm_loadu_si128((const __m128i*)&array1[0]); 347 const __m128i cmpA = _mm_cmpeq_epi32(A0, A1) [all...] |
/toolchain/binutils/binutils-2.27/gas/testsuite/gas/bfin/ |
arithmetic.s | 4 a0 = abs a0; define 5 A0 = ABS A1; 6 A1 = Abs a0; 8 A1 = abs a1, a0 = ABS A0; 82 A0 -= A1; 83 a0 -= a1 (w32); 92 a0 += a1; 93 A0 += A1 (w32) 128 a0 = r5.l * R7.H (w32); define 129 a0 = r0.h * r0.l; define 185 a0 = -a1; define [all...] |
load.s | 23 A0 = 0; 25 a1 = a0 = 0;
|
move2.s | 28 A0.X = A0.X; 29 A0.W = A0.W; 36 R2 = A0.W; 37 R3 = A0.X; 52 A0.X = R3; 53 A0.W = R2; 57 A0.X = A0.W [all...] |
/external/llvm/unittests/ADT/ |
SmallVectorTest.cpp | 795 EmplaceableArg<0> A0; 804 explicit Emplaceable(A0Ty &&A0) 805 : A0(std::forward<A0Ty>(A0)), State(ES_Emplaced) {} 808 Emplaceable(A0Ty &&A0, A1Ty &&A1) 809 : A0(std::forward<A0Ty>(A0)), A1(std::forward<A1Ty>(A1)), 813 Emplaceable(A0Ty &&A0, A1Ty &&A1, A2Ty &&A2) 814 : A0(std::forward<A0Ty>(A0)), A1(std::forward<A1Ty>(A1)) [all...] |
/art/runtime/arch/mips/ |
registers_mips.h | 35 A0 = 4, // Arguments.
|
/art/runtime/arch/mips64/ |
registers_mips64.h | 35 A0 = 4, // Arguments.
|
/external/clang/test/SemaTemplate/ |
temp_class_spec.cpp | 328 template<typename T, T N, typename U> class A0; 330 template<typename T, T N> class A0<T, N, int> { }; // expected-note{{here}} 331 template<typename T, T N> class A0<T, N, int>; 332 template<typename T, T N> class A0<T, N, int> { }; // expected-error{{redef}}
|
/external/llvm/lib/IR/ |
Attributes.cpp | 826 AttributeSetImpl *A0 = Attrs[0].pImpl; 827 if (A0) 828 AttrNodeVec.append(A0->getNode(0), A0->getNode(A0->getNumSlots())); [all...] |
/external/llvm/lib/Transforms/Instrumentation/ |
SanitizerCoverage.cpp | 495 Value *A0 = ICMP->getOperand(0); 497 if (!A0->getType()->isIntegerTy()) 499 uint64_t TypeSize = DL->getTypeStoreSizeInBits(A0->getType()); 500 // __sanitizer_cov_trace_cmp((type_size << 32) | predicate, A0, A1); 504 IRB.CreateIntCast(A0, Int64Ty, true),
|
/libcore/luni/src/test/java/libcore/java/lang/ |
ClassCastExceptionTest.java | 81 A0, B0, C0, D0, E0, F0, G0, H0, I0, J0, K0, L0, M0, N0, O0, P0, Q0, R0, S0, T0, U0, V0, W0, X0, Y0, Z0, 86 A0, B0, C0, D0, E0, F0, G0, H0, I0, J0, K0, L0, M0, N0, O0, P0, Q0, R0, S0, T0, U0, V0, W0, X0, Y0, Z0, 94 m.add(HugeF.A0);
|
/external/libvpx/libvpx/vpx_dsp/x86/ |
highbd_intrapred_intrin_ssse3.c | 99 const __m128i A0 = _mm_load_si128((const __m128i *)above); 103 const __m128i B0 = _mm_alignr_epi8(A1, A0, 2); 105 const __m128i C0 = _mm_alignr_epi8(A1, A0, 4); 107 __m128i avg3_0 = avg3_epu16(&A0, &B0, &C0); 134 const __m128i A0 = _mm_load_si128((const __m128i *)above); 140 const __m128i B0 = _mm_alignr_epi8(A1, A0, 2); 144 const __m128i C0 = _mm_alignr_epi8(A1, A0, 4); 148 __m128i avg3_0 = avg3_epu16(&A0, &B0, &C0); 216 const __m128i A0 = _mm_load_si128((const __m128i *)above); 219 const __m128i avg2_0 = _mm_avg_epu16(A0, B0) [all...] |
/external/opencv/cxcore/src/ |
cxmathfuncs.cpp | 912 A0 = .13369713757180123244806654839424e-2 / EXPPOLY_32F_A0; 915 #define EXPPOLY(x) (((((A0*(x) + A1)*(x) + A2)*(x) + A3)*(x) + A4)*(x) + A5) [all...] |
/external/eigen/Eigen/src/Core/products/ |
GeneralBlockPanelKernel.h | 348 * real*cplx : load lhs as (a0,a0,a1,a1), and mul as usual 1595 SLhsPacketHalf a0; local [all...] |
/external/libvpx/libvpx/vpx_dsp/arm/ |
highbd_intrapred_neon.c | 324 const uint16x8_t A0 = vld1q_u16(above); 325 const uint16x8_t above_right = vdupq_lane_u16(vget_high_u16(A0), 3); 328 const uint16x8_t avg1 = vhaddq_u16(A0, A2); [all...] |
intrapred_neon.c | 294 const uint8x8_t A0 = vld1_u8(above); 295 const uint8x8_t above_right = vdup_lane_u8(A0, 7); 296 const uint8x8_t A1 = vext_u8(A0, above_right, 1); 297 const uint8x8_t A2 = vext_u8(A0, above_right, 2); 298 const uint8x8_t avg1 = vhadd_u8(A0, A2); 322 const uint8x16_t A0 = vld1q_u8(above); 323 const uint8x16_t above_right = vdupq_lane_u8(vget_high_u8(A0), 7); 324 const uint8x16_t A1 = vextq_u8(A0, above_right, 1); 325 const uint8x16_t A2 = vextq_u8(A0, above_right, 2); 326 const uint8x16_t avg1 = vhaddq_u8(A0, A2) [all...] |