Searched
full:tmp4 (Results
401 -
425 of
1043) sorted by null
<<11121314151617181920>>
| /external/swiftshader/third_party/LLVM/test/CodeGen/PowerPC/ |
| buildvec_canonicalize.ll | 12 %tmp4 = fmul <4 x float> %tmp, %tmp3 ; <<4 x float>> [#uses=1]
13 store <4 x float> %tmp4, <4 x float>* %P3
|
| return-val-i128.ll | 16 %tmp4 = load float* %a_addr, align 4 ; <float> [#uses=1] 17 %tmp5 = fsub float -0.000000e+00, %tmp4 ; <float> [#uses=1]
|
| rlwimi3.ll | 7 %tmp4 = and i32 %srcA, 31775 ; <i32> [#uses=1]
8 %tmp5 = or i32 %tmp2, %tmp4 ; <i32> [#uses=1]
|
| vcmp-fold.ll | 11 %tmp4 = load <4 x float>* %x ; <<4 x float>> [#uses=1]
13 %tmp.upgrd.2 = call <4 x i32> @llvm.ppc.altivec.vcmpbfp( <4 x float> %tmp4, <4 x float> %tmp6 ) ; <<4 x i32>> [#uses=1]
|
| /external/swiftshader/third_party/LLVM/test/CodeGen/Thumb2/ |
| thumb2-ldm.ll | 11 %tmp4 = call i32 @f1( i32 %tmp, i32 %tmp3 ) ; <i32> [#uses=1] 12 ret i32 %tmp4
|
| /external/swiftshader/third_party/LLVM/test/CodeGen/X86/ |
| 2006-05-02-InstrSched1.ll | 14 %tmp4 = getelementptr i8* %tmp.upgrd.2, i64 %gep.upgrd.3 ; <i8*> [#uses=2]
18 %tmp.upgrd.5 = tail call i32 @memcmp( i8* %tmp8, i8* %tmp4, i32 %tmp.upgrd.1 ) ; <i32> [#uses=1]
|
| 2006-05-08-InstrSched.ll | 13 %tmp4 = and i32 %tmp.upgrd.2, 16 ; <i32> [#uses=1]
15 %tmp6 = trunc i32 %tmp4 to i8 ; <i8> [#uses=2]
|
| 2007-11-06-InstrSched.ll | 12 %tmp4 = load i32* %tmp3, align 4 ; <i32> [#uses=1] 13 %tmp45 = sitofp i32 %tmp4 to float ; <float> [#uses=1]
|
| 2010-04-08-CoalescerBug.ll | 19 %tmp4 = getelementptr inbounds %struct.FC* %tmp3, i64 0, i32 1, i64 0 21 %tmp6 = bitcast i32* %tmp4 to i8*
|
| avoid-loop-align.ll | 33 %tmp4 = sub i32 %tmp3, %tmp ; <i32> [#uses=1] 34 %tmp5 = getelementptr [100 x i32]* @A, i32 0, i32 %tmp4 ; <i32*> [#uses=1]
|
| byval2.ll | 38 %tmp4 = getelementptr %struct.s* %d, i32 0, i32 2 39 store i64 %c, i64* %tmp4, align 16
|
| call-push.ll | 15 %tmp4 = load i32* %tmp23 ; <i32> [#uses=1] 16 %tmp514 = lshr i32 %tmp4, 24 ; <i32> [#uses=1]
|
| shl-anyext.ll | 32 %tmp4 = xor i64 %tmp2, 536870911 ; <i64> [#uses=1] 33 %tmp5 = and i64 %tmp3, %tmp4 ; <i64> [#uses=1]
|
| stride-nine-with-base-reg.ll | 22 %tmp4 = mul i8 %tmp3, 2 24 store i8 %tmp4, i8* %tmp5, align 4
|
| tailcallbyval64.ll | 40 %tmp4 = tail call fastcc i64 @tailcallee(%struct.s* byval %a , i64 %tmp3, i64 %b, i64 7, i64 13, i64 17) 41 ret i64 %tmp4
|
| use-add-flags.ll | 17 %tmp4 = add i32 %tmp2, %y ; <i32> [#uses=1] 18 %tmp5 = icmp slt i32 %tmp4, 0 ; <i1> [#uses=1]
|
| vec_shift3.ll | 20 %tmp4 = tail call <8 x i16> @llvm.x86.sse2.psrai.w( <8 x i16> %tmp2, i32 %bits ) nounwind readnone ; <<8 x i16>> [#uses=1] 21 %tmp5 = bitcast <8 x i16> %tmp4 to <2 x i64> ; <<2 x i64>> [#uses=1]
|
| /external/swiftshader/third_party/LLVM/test/Feature/ |
| ppcld.ll | 19 %tmp4 = fpext double %tmp3 to ppc_fp128 ; <ppc_fp128> [#uses=1] 20 store ppc_fp128 %tmp4, ppc_fp128* @ld
|
| x86ld.ll | 19 %tmp4 = fpext double %tmp3 to x86_fp80 ; <x86_fp80> [#uses=1] 20 store x86_fp80 %tmp4, x86_fp80* @ld
|
| /external/swiftshader/third_party/LLVM/test/Transforms/ArgumentPromotion/ |
| byval.ll | 21 %tmp4 = getelementptr %struct.ss* %S, i32 0, i32 1 ; <i64*> [#uses=1] 22 store i64 2, i64* %tmp4, align 4
|
| /external/swiftshader/third_party/LLVM/test/Transforms/IndVarSimplify/ |
| loop_evaluate11.ll | 25 %tmp4 = add i32 %order_start.0, 2 ; <i32> [#uses=1] 26 %tmp5 = add i32 %tmp4, undef ; <i32> [#uses=1]
|
| /external/swiftshader/third_party/LLVM/test/Transforms/InstCombine/ |
| 2008-04-29-VolatileLoadDontMerge.ll | 15 %tmp4 = add i32 %tmp3.reg2mem.0, 5 ; <i32> [#uses=1] 16 volatile store i32 %tmp4, i32* @g_1, align 4
|
| 2008-07-08-VolatileLoadMerge.ll | 16 %tmp4 = add i32 %tmp3.reg2mem.0, 5 ; <i32> [#uses=1] 17 volatile store i32 %tmp4, i32* @g_1, align 4
|
| 2010-11-01-lshr-mask.ll | 32 %tmp4 = and i8 %arg1, 33 35 %tmp7 = or i8 %tmp4, %tmp6
|
| bitcast.ll | 29 %tmp4 = bitcast i32 %tmp2 to float ; <float> [#uses=1] 31 %add = fadd float %tmp24, %tmp4 37 ; CHECK-NEXT: %tmp4 = extractelement <2 x float> {{.*}}, i32 0 38 ; CHECK-NEXT: %add = fadd float %tmp24, %tmp4 54 %tmp4 = bitcast i32 %tmp2 to float 56 %add = fadd float %tmp24, %tmp4 62 ; CHECK-NEXT: %tmp4 = extractelement <4 x float> {{.*}}, i32 2 63 ; CHECK-NEXT: %add = fadd float %tmp24, %tmp4
|
Completed in 594 milliseconds
<<11121314151617181920>>