|   /external/swiftshader/third_party/LLVM/test/CodeGen/PowerPC/ | 
| 2007-05-22-tailmerge-3.ll  | 26 	%tmp4 = call i32 (...)* @baz( i32 5, i32 6 )		; <i32> [#uses=0]
  | 
|   /external/swiftshader/third_party/LLVM/test/CodeGen/X86/ | 
| sse2.ll  | 154         %tmp4 = fadd <4 x float> %tmp2, %tmp3            ; <<4 x float>> [#uses=1] 155         store <4 x float> %tmp4, <4 x float>* null 190         %tmp4 = shufflevector <4 x float> %tmp3, <4 x float> %tmp, <4 x i32> < i32 2, i32 3, i32 6, i32 7 >           ; <<4 x float>> [#uses=1] 191         ret <4 x float> %tmp4
  | 
| scalar_widen_div.ll  | 172   %tmp4 = load <3 x i32>* %arrayidx11 ; <<3 x i32>> [#uses=1] 175   %div = sdiv <3 x i32> %tmp4, %tmp8
  | 
|   /external/swiftshader/third_party/LLVM/test/Transforms/GlobalOpt/ | 
| crash.ll  | 75   %tmp4 = load i32*** @g_52, align 8
  | 
|   /external/clang/test/CodeGen/ | 
| aarch64-neon-vget.c  | 92 // CHECK:   [[TMP4:%.*]] = bitcast i16* [[__REINT1_242]] to half* 93 // CHECK:   [[TMP5:%.*]] = load half, half* [[TMP4]], align 2 185 // CHECK:   [[TMP4:%.*]] = bitcast i16* [[__REINT1_244]] to half* 186 // CHECK:   [[TMP5:%.*]] = load half, half* [[TMP4]], align 2 315 // CHECK:   [[TMP4:%.*]] = load <4 x i16>, <4 x i16>* [[TMP3]], align 8 316 // CHECK:   [[TMP5:%.*]] = bitcast <4 x i16> [[TMP4]] to <8 x i8> 412 // CHECK:   [[TMP4:%.*]] = load <8 x i16>, <8 x i16>* [[TMP3]], align 16 413 // CHECK:   [[TMP5:%.*]] = bitcast <8 x i16> [[TMP4]] to <16 x i8>
  | 
|   /external/llvm/lib/Transforms/Utils/ | 
| IntegerDivision.cpp  | 129   // ;   %tmp4   = xor i32 %q_mag, %q_sgn 130   // ;   %q      = sub i32 %tmp4, %q_sgn 139   Value *Tmp4   = Builder.CreateXor(Q_Mag, Q_Sgn); 140   Value *Q      = Builder.CreateSub(Tmp4, Q_Sgn); 280   // ;   %tmp4 = add i32 %divisor, -1 284   Value *Tmp4 = Builder.CreateAdd(Divisor, NegOne); 297   // ;   %tmp9  = sub i32 %tmp4, %tmp7 315   Value *Tmp9  = Builder.CreateSub(Tmp4, Tmp7);
  | 
|   /external/llvm/test/Analysis/TypeBasedAliasAnalysis/ | 
| dynamic-indices.ll  | 29   %tmp4 = load i16, i16* %arrayidx, align 2, !tbaa !0 30   %conv = zext i16 %tmp4 to i32
  | 
|   /external/llvm/test/CodeGen/AArch64/ | 
| arm64-shrink-wrapping.ll  | 49   %tmp4 = call i32 @doSomething(i32 0, i32* %tmp) 53   %tmp.0 = phi i32 [ %tmp4, %true ], [ %a, %0 ] 660   %tmp4 = load i32, i32* %tmp 664   %tmp.0 = phi i32 [ %tmp4, %true ], [ %a, %0 ] 685   %tmp4 = lshr i32 %b, %a 690   %tmp9 = add i32 %tmp3, %tmp4 691   %tmp10 = add i32 %tmp4, %tmp5 704   store i32 %tmp4, i32* %ptr4, align 4
  | 
|   /external/llvm/test/CodeGen/AMDGPU/ | 
| split-vector-memoperand-offsets.ll  | 40   %tmp4 = tail call i32 @llvm.amdgcn.workitem.id.z() 44   %tmp9 = add i32 %tmp11, %tmp4
  | 
|   /external/llvm/test/CodeGen/ARM/ | 
| ldstrex.ll  | 26   %tmp4 = trunc i64 %val to i32 29   %strexd = tail call i32 @llvm.arm.strexd(i32 %tmp4, i32 %tmp7, i8* %ptr)
  | 
| spill-q.ll  | 73   %tmp4 = fadd <4 x float> %tmp3, %ld6 74   %tmp5 = fadd <4 x float> %tmp4, %ld7
  | 
|   /external/llvm/test/CodeGen/Thumb2/ | 
| thumb2-spill-q.ll  | 73   %tmp4 = fadd <4 x float> %tmp3, %ld6 74   %tmp5 = fadd <4 x float> %tmp4, %ld7
  | 
|   /external/llvm/test/CodeGen/X86/ | 
| scalar_widen_div.ll  | 184   %tmp4 = load <3 x i32>, <3 x i32>* %arrayidx11 ; <<3 x i32>> [#uses=1] 187   %div = sdiv <3 x i32> %tmp4, %tmp8
  | 
|   /external/llvm/test/Transforms/InstCombine/ | 
| simplify-libcalls.ll  | 51 	%tmp4 = load i8*, i8** %start_addr, align 4		; <i8*> [#uses=1] 52 	%tmp5 = call i32 @memcmp( i8* %tmp4, i8* getelementptr ([5 x i8], [5 x i8]* @_2E_str, i32 0, i32 0), i32 4 ) nounwind readonly 		; <i32> [#uses=1]
  | 
|   /external/llvm/test/Transforms/LCSSA/ | 
| invoke-dest.ll  | 110 	%tmp4 = invoke i8* @_Znaj(i32 124) 126 	%tmp5 = icmp eq i8* %tmp4, null		; <i1> [#uses=1]
  | 
|   /external/llvm/test/Transforms/LoopRotate/ | 
| PhiRename-1.ll  | 45 	%tmp4 = load i8*, i8** %tmp3		; <i8*> [#uses=1] 46 	%tmp45 = bitcast i8* %tmp4 to %struct.operator*		; <%struct.operator*> [#uses=1]
  | 
|   /external/llvm/test/Transforms/PGOProfile/ | 
| icp_covariant_invoke_return.ll  | 58   %tmp4 = bitcast %struct.Base* %call2 to i8* 59   call void @_ZdlPv(i8* %tmp4)
  | 
| landingpad.ll  | 88   %tmp4 = call i8* @__cxa_begin_catch(i8* %tmp1) 89   %tmp5 = bitcast i8* %tmp4 to i32*
  | 
|   /external/llvm/test/Transforms/StructurizeCFG/ | 
| post-order-traversal-bug.ll  | 46   %tmp4 = add nsw i64 %indvars.iv, -1 47   %arrayidx8 = getelementptr inbounds float, float* %nr, i64 %tmp4
  | 
|   /external/swiftshader/third_party/LLVM/test/Analysis/ScalarEvolution/ | 
| nsw.ll  | 23 	%tmp4 = load double* %tmp3, align 8		; <double> [#uses=1] 24 	%tmp5 = fmul double %tmp4, 9.200000e+00		; <double> [#uses=1]
  | 
|   /external/swiftshader/third_party/LLVM/test/Analysis/TypeBasedAliasAnalysis/ | 
| dynamic-indices.ll  | 29   %tmp4 = load i16* %arrayidx, align 2, !tbaa !0 30   %conv = zext i16 %tmp4 to i32
  | 
|   /external/swiftshader/third_party/LLVM/test/CodeGen/ARM/ | 
| spill-q.ll  | 73   %tmp4 = fadd <4 x float> %tmp3, %ld6 74   %tmp5 = fadd <4 x float> %tmp4, %ld7
  | 
|   /external/swiftshader/third_party/LLVM/test/CodeGen/Mips/ | 
| o32_cc_byval.ll  | 72   %tmp4 = load i16* %s, align 2, !tbaa !5 75   tail call void @callee4(i32 %tmp, double %tmp1, i64 %tmp2, i32 %tmp3, i16 signext %tmp4, i8 signext %tmp5, float %f) nounwind
  | 
|   /external/swiftshader/third_party/LLVM/test/CodeGen/Thumb2/ | 
| thumb2-spill-q.ll  | 73   %tmp4 = fadd <4 x float> %tmp3, %ld6 74   %tmp5 = fadd <4 x float> %tmp4, %ld7
  | 
|   /external/swiftshader/third_party/LLVM/test/DebugInfo/X86/ | 
| earlydup-crash.ll  | 32   %tmp4 = add i32 0, %tmp1 33   call void @foobar(i32 %tmp4)
  |