/external/eigen/Eigen/src/Geometry/arch/ |
Geometry_SSE.h | 48 __m128 mul1=_mm_mul_ps(vec4f_swizzle1(a,1,2,0,3),vec4f_swizzle1(b,2,0,1,3)); local 51 pstore(&res.x(),_mm_sub_ps(mul1,mul2));
|
/external/llvm/test/Transforms/Inline/ |
ignore-debug-info.ll | 18 %mul1 = fmul <4 x float> %b, <float 5.000000e+00, float 5.000000e+00, float 5.000000e+00, float 5.000000e+00> 20 %add = fadd <4 x float> %mul, %mul1
|
/external/llvm/test/Transforms/SLPVectorizer/X86/ |
compare-reduce.ll | 26 %mul1 = fmul double %conv, %1 27 %mul2 = fmul double %mul1, 7.000000e+00
|
in-tree-user.ll | 24 %mul1 = fmul double %conv, %1 25 %mul2 = fmul double %mul1, 7.000000e+00
|
reduction.ll | 28 %mul1 = fmul double %0, 7.000000e+00 33 %add5 = fadd double %mul1, %mul4
|
cse.ll | 59 %mul1 = fmul double %conv, %mul 60 %add = fadd double %mul1, 6.000000e+00 152 %mul1 = fmul double %conv, %mul 153 %add = fadd double %mul1, 6.000000e+00
|
/external/llvm/test/CodeGen/X86/ |
combine-multiplies.ll | 122 %mul1 = mul <4 x i32> %add1, <i32 22, i32 22, i32 22, i32 22> 125 store <4 x i32> %mul1, <4 x i32>* @v2, align 16 156 %mul1 = mul <4 x i32> %add1, <i32 22, i32 33, i32 44, i32 55> 159 store <4 x i32> %mul1, <4 x i32>* @v2, align 16
|
fmul-combines.ll | 108 %mul1 = fmul fast <4 x float> <float 2.0, float 2.0, float 2.0, float 2.0>, <float 3.0, float 3.0, float 3.0, float 3.0> 109 %mul2 = fmul fast <4 x float> <float 4.0, float 4.0, float 4.0, float 4.0>, %mul1 124 %mul1 = fmul fast <4 x float> <float 1.0, float 2.0, float 3.0, float 4.0>, <float 5.0, float 6.0, float 7.0, float 8.0> 125 %mul2 = fmul fast <4 x float> <float 9.0, float 10.0, float 11.0, float 12.0>, %mul1
|
/external/llvm/test/Transforms/LoopIdiom/AMDGPU/ |
popcnt.ll | 85 %mydata2.addr.012 = phi i32 [ %mul1, %while.body ], [ %mydata2, %entry ] 93 %mul1 = mul nsw i32 %conv, %mydata2.addr.012 99 %mydata2.addr.0.lcssa = phi i32 [ %mydata2, %entry ], [ %mul1, %while.body ]
|
/external/llvm/test/Analysis/LoopAccessAnalysis/ |
number-of-memchecks.ll | 43 %mul1 = mul i16 %mul, %loadC 46 store i16 %mul1, i16* %arrayidxD, align 2 52 store i16 %mul1, i16* %arrayidxF, align 2 133 %mul1 = mul i16 %mul, %loadB 136 store i16 %mul1, i16* %arrayidxC, align 2 205 %mul1 = mul i16 %mul, %loadB 208 store i16 %mul1, i16* %arrayidxC, align 2
|
/external/llvm/test/Analysis/DependenceAnalysis/ |
SymbolicSIV.ll | 77 %mul1 = mul i64 %n, 5 78 %add = add i64 %mul, %mul1 215 %mul1 = mul i64 %n, 3 216 %add = add i64 %mul, %mul1 260 %mul1 = shl i64 %n, 1 261 %sub = sub i64 %mul, %mul1 352 %mul1 = mul i64 %mul, %i.03 353 %add = add i64 %mul1, %M 401 %mul1 = mul i64 %mul, %i.03 402 %add = add i64 %mul1, % [all...] |
StrongSIV.ll | 271 %mul1 = shl i64 %i.02, 1 272 %arrayidx2 = getelementptr inbounds i32, i32* %A, i64 %mul1 308 %mul1 = shl i64 %i.02, 1 309 %arrayidx2 = getelementptr inbounds i32, i32* %A, i64 %mul1 424 %mul1 = mul i64 %i.02, %n 425 %add2 = add i64 %mul1, 5
|
/external/llvm/test/Transforms/LoopIdiom/X86/ |
popcnt.ll | 56 %mydata2.addr.012 = phi i32 [ %mul1, %while.body ], [ %mydata2, %entry ] 64 %mul1 = mul nsw i32 %conv, %mydata2.addr.012 70 %mydata2.addr.0.lcssa = phi i32 [ %mydata2, %entry ], [ %mul1, %while.body ]
|
/external/vulkan-validation-layers/libs/glm/detail/ |
intrinsic_common.inl | 259 __m128 mul1 = _mm_mul_ps(v2, a); 260 __m128 add0 = _mm_add_ps(mul0, mul1); 281 __m128 mul1 = _mm_mul_ps(clp0, clp0); 282 __m128 mul2 = _mm_mul_ps(mul1, sub2);
|
/external/llvm/test/CodeGen/SystemZ/ |
fp-mul-01.ll | 107 %mul1 = fmul float %mul0, %val1 108 %mul2 = fmul float %mul1, %val2
|
fp-mul-03.ll | 109 %mul1 = fmul double %mul0, %val1 110 %mul2 = fmul double %mul1, %val2
|
int-mul-04.ll | 128 %mul1 = mul i64 %mul0, %val1 129 %mul2 = mul i64 %mul1, %val2
|
int-mul-02.ll | 163 %mul1 = mul i32 %mul0, %val1 164 %mul2 = mul i32 %mul1, %val2
|
int-mul-03.ll | 169 %mul1 = mul i64 %mul0, %ext1 170 %mul2 = mul i64 %mul1, %ext2
|
/external/vulkan-validation-layers/libs/glm/gtx/ |
simd_quat.inl | 124 __m128 mul1 = _mm_mul_ps(q1.Data, _mm_shuffle_ps(q2.Data, q2.Data, _MM_SHUFFLE(1, 0, 3, 2))); 130 __m128 add1 = _mm_dp_ps(mul1, _mm_set_ps(1.0f, 1.0f, 1.0f, -1.0f), 0xff); 138 mul1 = _mm_mul_ps(mul1, _mm_set_ps(1.0f, 1.0f, 1.0f, -1.0f)); 139 __m128 add1 = _mm_add_ps(mul1, _mm_movehl_ps(mul1, mul1));
|
/external/libvpx/libvpx/vp8/common/mips/msa/ |
postproc_msa.c | 591 v4i32 mul0, mul1, mul2, mul3; local 625 mul0, mul1, mul2, mul3); 631 sum_sq1[0] = sum_sq0[3] + mul1[0]; 634 sum_sq1[cnt + 1] = sum_sq1[cnt] + mul1[cnt + 1]; 708 v4i32 mul1 = { 0 }; local [all...] |
/external/llvm/test/CodeGen/Hexagon/ |
expand-condsets-rm-segment.ll | 37 %mul1 = mul i32 %1, %call 38 store i32 %mul1, i32* @sysctl_sched_latency, align 4, !tbaa !1
|
/external/llvm/test/Transforms/BBVectorize/ |
cycle.ll | 13 define void @test1(double %a, double %b, double %c, double %add80, double %mul1, double %mul2.v.r1, double %mul73, double %sub, double %sub65, double %F.0, i32 %n.0, double %Bnm3.0, double %Bnm2.0, double %Bnm1.0, double %Anm3.0, double %Anm2.0, double %Anm1.0) { 90 %mul134 = fmul double %div115, %mul1
|
/external/llvm/test/Transforms/InstCombine/ |
fmul.ll | 147 %mul1 = fmul fast double %sqrt, %sqrt 148 %mul2 = fmul double %mul1, %sqrt
|
/external/webp/src/dsp/ |
dec.c | 37 #define MUL1(a) ((((a) * 20091) >> 16) + (a)) 47 const int c = MUL2(in[4]) - MUL1(in[12]); // [-3783, 3783] 48 const int d = MUL1(in[4]) + MUL2(in[12]); // [-3785, 3781] 68 const int c = MUL2(tmp[4]) - MUL1(tmp[12]); 69 const int d = MUL1(tmp[4]) + MUL2(tmp[12]); 83 const int d4 = MUL1(in[4]); 85 const int d1 = MUL1(in[1]); 91 #undef MUL1
|