HomeSort by relevance Sort by last modified time
    Searched full:tmp4 (Results 801 - 825 of 1043) sorted by null

<<31323334353637383940>>

  /external/swiftshader/third_party/LLVM/test/CodeGen/PowerPC/
2007-05-22-tailmerge-3.ll 26 %tmp4 = call i32 (...)* @baz( i32 5, i32 6 ) ; <i32> [#uses=0]
  /external/swiftshader/third_party/LLVM/test/CodeGen/X86/
sse2.ll 154 %tmp4 = fadd <4 x float> %tmp2, %tmp3 ; <<4 x float>> [#uses=1]
155 store <4 x float> %tmp4, <4 x float>* null
190 %tmp4 = shufflevector <4 x float> %tmp3, <4 x float> %tmp, <4 x i32> < i32 2, i32 3, i32 6, i32 7 > ; <<4 x float>> [#uses=1]
191 ret <4 x float> %tmp4
scalar_widen_div.ll 172 %tmp4 = load <3 x i32>* %arrayidx11 ; <<3 x i32>> [#uses=1]
175 %div = sdiv <3 x i32> %tmp4, %tmp8
  /external/swiftshader/third_party/LLVM/test/Transforms/GlobalOpt/
crash.ll 75 %tmp4 = load i32*** @g_52, align 8
  /external/clang/test/CodeGen/
aarch64-neon-vget.c 92 // CHECK: [[TMP4:%.*]] = bitcast i16* [[__REINT1_242]] to half*
93 // CHECK: [[TMP5:%.*]] = load half, half* [[TMP4]], align 2
185 // CHECK: [[TMP4:%.*]] = bitcast i16* [[__REINT1_244]] to half*
186 // CHECK: [[TMP5:%.*]] = load half, half* [[TMP4]], align 2
315 // CHECK: [[TMP4:%.*]] = load <4 x i16>, <4 x i16>* [[TMP3]], align 8
316 // CHECK: [[TMP5:%.*]] = bitcast <4 x i16> [[TMP4]] to <8 x i8>
412 // CHECK: [[TMP4:%.*]] = load <8 x i16>, <8 x i16>* [[TMP3]], align 16
413 // CHECK: [[TMP5:%.*]] = bitcast <8 x i16> [[TMP4]] to <16 x i8>
  /external/llvm/lib/Transforms/Utils/
IntegerDivision.cpp 129 // ; %tmp4 = xor i32 %q_mag, %q_sgn
130 // ; %q = sub i32 %tmp4, %q_sgn
139 Value *Tmp4 = Builder.CreateXor(Q_Mag, Q_Sgn);
140 Value *Q = Builder.CreateSub(Tmp4, Q_Sgn);
280 // ; %tmp4 = add i32 %divisor, -1
284 Value *Tmp4 = Builder.CreateAdd(Divisor, NegOne);
297 // ; %tmp9 = sub i32 %tmp4, %tmp7
315 Value *Tmp9 = Builder.CreateSub(Tmp4, Tmp7);
  /external/llvm/test/Analysis/TypeBasedAliasAnalysis/
dynamic-indices.ll 29 %tmp4 = load i16, i16* %arrayidx, align 2, !tbaa !0
30 %conv = zext i16 %tmp4 to i32
  /external/llvm/test/CodeGen/AArch64/
arm64-shrink-wrapping.ll 49 %tmp4 = call i32 @doSomething(i32 0, i32* %tmp)
53 %tmp.0 = phi i32 [ %tmp4, %true ], [ %a, %0 ]
660 %tmp4 = load i32, i32* %tmp
664 %tmp.0 = phi i32 [ %tmp4, %true ], [ %a, %0 ]
685 %tmp4 = lshr i32 %b, %a
690 %tmp9 = add i32 %tmp3, %tmp4
691 %tmp10 = add i32 %tmp4, %tmp5
704 store i32 %tmp4, i32* %ptr4, align 4
  /external/llvm/test/CodeGen/AMDGPU/
split-vector-memoperand-offsets.ll 40 %tmp4 = tail call i32 @llvm.amdgcn.workitem.id.z()
44 %tmp9 = add i32 %tmp11, %tmp4
  /external/llvm/test/CodeGen/ARM/
ldstrex.ll 26 %tmp4 = trunc i64 %val to i32
29 %strexd = tail call i32 @llvm.arm.strexd(i32 %tmp4, i32 %tmp7, i8* %ptr)
spill-q.ll 73 %tmp4 = fadd <4 x float> %tmp3, %ld6
74 %tmp5 = fadd <4 x float> %tmp4, %ld7
  /external/llvm/test/CodeGen/Thumb2/
thumb2-spill-q.ll 73 %tmp4 = fadd <4 x float> %tmp3, %ld6
74 %tmp5 = fadd <4 x float> %tmp4, %ld7
  /external/llvm/test/CodeGen/X86/
scalar_widen_div.ll 184 %tmp4 = load <3 x i32>, <3 x i32>* %arrayidx11 ; <<3 x i32>> [#uses=1]
187 %div = sdiv <3 x i32> %tmp4, %tmp8
  /external/llvm/test/Transforms/InstCombine/
simplify-libcalls.ll 51 %tmp4 = load i8*, i8** %start_addr, align 4 ; <i8*> [#uses=1]
52 %tmp5 = call i32 @memcmp( i8* %tmp4, i8* getelementptr ([5 x i8], [5 x i8]* @_2E_str, i32 0, i32 0), i32 4 ) nounwind readonly ; <i32> [#uses=1]
  /external/llvm/test/Transforms/LCSSA/
invoke-dest.ll 110 %tmp4 = invoke i8* @_Znaj(i32 124)
126 %tmp5 = icmp eq i8* %tmp4, null ; <i1> [#uses=1]
  /external/llvm/test/Transforms/LoopRotate/
PhiRename-1.ll 45 %tmp4 = load i8*, i8** %tmp3 ; <i8*> [#uses=1]
46 %tmp45 = bitcast i8* %tmp4 to %struct.operator* ; <%struct.operator*> [#uses=1]
  /external/llvm/test/Transforms/PGOProfile/
icp_covariant_invoke_return.ll 58 %tmp4 = bitcast %struct.Base* %call2 to i8*
59 call void @_ZdlPv(i8* %tmp4)
landingpad.ll 88 %tmp4 = call i8* @__cxa_begin_catch(i8* %tmp1)
89 %tmp5 = bitcast i8* %tmp4 to i32*
  /external/llvm/test/Transforms/StructurizeCFG/
post-order-traversal-bug.ll 46 %tmp4 = add nsw i64 %indvars.iv, -1
47 %arrayidx8 = getelementptr inbounds float, float* %nr, i64 %tmp4
  /external/swiftshader/third_party/LLVM/test/Analysis/ScalarEvolution/
nsw.ll 23 %tmp4 = load double* %tmp3, align 8 ; <double> [#uses=1]
24 %tmp5 = fmul double %tmp4, 9.200000e+00 ; <double> [#uses=1]
  /external/swiftshader/third_party/LLVM/test/Analysis/TypeBasedAliasAnalysis/
dynamic-indices.ll 29 %tmp4 = load i16* %arrayidx, align 2, !tbaa !0
30 %conv = zext i16 %tmp4 to i32
  /external/swiftshader/third_party/LLVM/test/CodeGen/ARM/
spill-q.ll 73 %tmp4 = fadd <4 x float> %tmp3, %ld6
74 %tmp5 = fadd <4 x float> %tmp4, %ld7
  /external/swiftshader/third_party/LLVM/test/CodeGen/Mips/
o32_cc_byval.ll 72 %tmp4 = load i16* %s, align 2, !tbaa !5
75 tail call void @callee4(i32 %tmp, double %tmp1, i64 %tmp2, i32 %tmp3, i16 signext %tmp4, i8 signext %tmp5, float %f) nounwind
  /external/swiftshader/third_party/LLVM/test/CodeGen/Thumb2/
thumb2-spill-q.ll 73 %tmp4 = fadd <4 x float> %tmp3, %ld6
74 %tmp5 = fadd <4 x float> %tmp4, %ld7
  /external/swiftshader/third_party/LLVM/test/DebugInfo/X86/
earlydup-crash.ll 32 %tmp4 = add i32 0, %tmp1
33 call void @foobar(i32 %tmp4)

Completed in 1575 milliseconds

<<31323334353637383940>>