/external/llvm/test/Transforms/StraightLineStrengthReduce/ |
slsr-mul.ll | 15 %mul1 = mul i32 %b1, %s 16 call void @foo(i32 %mul1) 36 %mul1 = mul i32 %b1, %s 37 call void @foo(i32 %mul1) 57 %mul1 = mul i32 %b1, %s 59 call void @foo(i32 %mul1) 81 %mul1 = mul i32 %a1, %b 86 call void @foo(i32 %mul1) 101 ; mul1 = mul0 + bump; // GVN ensures mul1 and mul2 use the same bump [all...] |
/external/llvm/test/CodeGen/AArch64/ |
arm64-fp-contract-zero.ll | 11 %mul1 = fmul double %b, 0.000000e+00 12 %sub = fsub double %mul, %mul1
|
/external/llvm/test/CodeGen/ARM/ |
2009-04-08-FloatUndef.ll | 7 %mul1 = fmul <4 x float> %shuffle7, zeroinitializer ; <<4 x float>> [#uses=1] 8 %add2 = fadd <4 x float> %mul1, %input2 ; <<4 x float>> [#uses=1]
|
shifter_operand.ll | 85 %mul1 = mul i32 %y, 127534 86 %or = or i32 %mul1, %x 97 %mul1 = mul i32 %y, 255068 98 %or = or i32 %mul1, %x 109 %mul1 = mul i32 %y, 510136 110 %or = or i32 %mul1, %x 121 %mul1 = mul i32 %y, 1020272 122 %or = or i32 %mul1, %x 159 %mul1 = mul i32 %y, 127534 160 %arrayidx = getelementptr inbounds i8, i8* %x, i32 %mul1 [all...] |
/external/llvm/test/CodeGen/PowerPC/ |
2007-11-19-VectorSplitting.ll | 13 %mul1 = fmul <4 x float> %shuffle7, zeroinitializer ; <<4 x 14 %add2 = fadd <4 x float> %mul1, %input2 ; <<4 x float>>
|
/external/llvm/test/Transforms/Reassociate/ |
mixed-fast-nonfast-fp.ll | 10 %mul1 = fmul fast float %a, %c 14 %add1 = fadd fast float %mul1, %mul3
|
canonicalize-neg-const.ll | 14 %mul1 = fmul double -1.234000e-01, %y 15 %add2 = fadd double %mul1, %x 30 %mul1 = fmul double %y, -1.234000e-01 31 %add2 = fadd double %mul1, %x 46 %mul1 = fmul double %y, -1.234000e-01 47 %add2 = fsub double %x, %mul1
|
/external/llvm/test/CodeGen/X86/ |
dag-fmf-cse.ll | 4 ; If fast-math-flags are propagated correctly, the mul1 expression 16 %mul1 = fmul fast float %a, %b 19 %abx2 = fsub fast float %mul1, %mul2
|
atom-sched.ll | 30 %mul1 = mul nsw i32 %2, %3 31 store i32 %mul1, i32* @d, align 4
|
i128-mul.ll | 16 ; CHECK: @mul1 20 define i64 @mul1(i64 %n, i64* nocapture %z, i64* nocapture %x, i64 %y) {
|
/external/llvm/test/Transforms/GVN/ |
2011-07-07-MatchIntrinsicExtract.ll | 34 %mul1 = mul i64 %a, %b 35 ret i64 %mul1 39 ; CHECK-NOT: mul1 70 %mul1 = mul i64 %a, %b 71 ret i64 %mul1 75 ; CHECK-NOT: mul1
|
/external/llvm/test/CodeGen/Mips/ |
2010-11-09-Mul.ll | 4 define i32 @mul1(i32 %a, i32 %b) nounwind readnone {
|
/external/vulkan-validation-layers/libs/glm/detail/ |
intrinsic_geometric.inl | 82 __m128 mul1 = _mm_mul_ps(swp1, swp2); 83 __m128 sub0 = _mm_sub_ps(mul0, mul1); 108 __m128 mul1 = _mm_mul_ps(N, mul0); 109 return mul1; 117 __m128 mul1 = _mm_mul_ps(mul0, glm::detail::two); 118 __m128 sub0 = _mm_sub_ps(I, mul1); 127 __m128 mul1 = _mm_mul_ps(dot0, dot0); 129 __m128 sub1 = _mm_sub_ps(glm::detail::one, mul1);
|
/external/llvm/test/Analysis/LoopAccessAnalysis/ |
safe-no-checks.ll | 17 ; CHECK-NEXT: store i16 %mul1, i16* %arrayidxA, align 2 41 %mul1 = mul i16 %mul, %loadC 44 store i16 %mul1, i16* %arrayidxA, align 2
|
unsafe-and-rt-checks.ll | 14 ; CHECK-NEXT: store i16 %mul1, i16* %arrayidxA_plus_2, align 2 53 %mul1 = mul i16 %mul, %loadC 57 store i16 %mul1, i16* %arrayidxA_plus_2, align 2
|
resort-to-memchecks-only.ll | 47 %mul1 = mul i16 %mul, %loadC 51 store i16 %mul1, i16* %arrayidxA2, align 2
|
/external/llvm/test/Transforms/SimplifyCFG/AArch64/ |
prefer-fma.ll | 27 ; CHECK: %mul1 = fmul fast double %1, %2 28 ; CHECK-NEXT: %sub1 = fsub fast double %mul1, %0 30 %mul1 = fmul fast double %1, %4 31 %sub1 = fsub fast double %mul1, %0
|
/external/llvm/test/Transforms/Inline/ |
vector-bonus.ll | 9 %mul1 = mul nsw i32 %i, %i 25 %retval.0 = phi i32 [ %mul1, %if.then ], [ %add7, %if.else ]
|
/external/llvm/test/Transforms/SLPVectorizer/X86/ |
crash_binaryop.ll | 21 %mul1 = fmul double %add1, 0.000000e+00 30 %binary_V = fadd double %mul1, %binaryop_B
|
barriercall.ll | 15 %mul1 = mul nsw i32 %n, 9 16 %add2 = add nsw i32 %mul1, 9
|
extractcost.ll | 14 %mul1 = mul nsw i32 %n, 9 15 %add2 = add nsw i32 %mul1, 9
|
/external/llvm/test/Transforms/InstCombine/ |
add2.ll | 218 %mul1 = mul nsw i16 %x, 8 219 %add2 = add nsw i16 %x, %mul1 227 %mul1 = mul nsw i16 %x, 8 228 %add2 = add nsw i16 %mul1, %x 236 %mul1 = mul i16 %a, 2 238 %add = add nsw i16 %mul1, %mul2 246 %mul1 = mul nsw i16 %a, 2 248 %add = add nsw i16 %mul1, %mul2 256 %mul1 = mul nsw i16 %a, 3 258 %add = add nsw i16 %mul1, %mul [all...] |
/art/test/455-checker-gvn/src/ |
Main.java | 67 int mul1 = mX * mY; local 68 int abs = Math.abs(mul1);
|
/external/libvpx/libvpx/vp8/common/mips/msa/ |
idct_msa.c | 182 v8i16 input0, input1, dequant_in0, dequant_in1, mul0, mul1; local 194 MUL2(input0, dequant_in0, input1, dequant_in1, mul0, mul1); 195 PCKEV_D2_SH(zero, mul0, zero, mul1, in0, in2); 196 PCKOD_D2_SH(zero, mul0, zero, mul1, in1, in3); 198 PCKEV_D2_SH(hz1_h, hz0_h, hz3_h, hz2_h, mul0, mul1); 200 UNPCK_SH_SW(mul1, hz2_w, hz3_w); 226 v8i16 mul0, mul1, mul2, mul3, dequant_in0, dequant_in1; local 236 mul0, mul1, mul2, mul3); 237 PCKEV_D2_SH(mul2, mul0, mul3, mul1, in0, in2); 238 PCKOD_D2_SH(mul2, mul0, mul3, mul1, in1, in3) [all...] |
/external/boringssl/src/crypto/bn/asm/ |
x86-mont.pl | 113 $mul1="mm5"; 132 &movd ($mul1,&DWP(0,$ap)); # ap[0] 135 &pmuludq($mul1,$mul0); # ap[0]*bp[0] 136 &movq ($car0,$mul1); 137 &movq ($acc0,$mul1); # I wish movd worked for 140 &pmuludq($mul1,$_n0q); # *=n0 142 &pmuludq($car1,$mul1); # "t[0]"*np[0]*n0 154 &pmuludq($acc1,$mul1); # np[j]*m1 172 &pmuludq($acc1,$mul1); # np[num-1]*m1 193 &movd ($mul1,&DWP(0,$ap)); # ap[0 [all...] |