/external/llvm/test/Transforms/ObjCARC/ |
move-and-merge-autorelease.ll | 39 %tmp17 = bitcast %3* %arg3 to i8* 40 %tmp18 = call %4* bitcast (i8* (i8*, i8*, ...)* @objc_msgSend to %4* (i8*, i8*)*)(i8* %tmp17, i8* %tmp16) 42 %tmp20 = bitcast %4* %tmp18 to i8* 43 %tmp21 = call %5* bitcast (i8* (i8*, i8*, ...)* @objc_msgSend to %5* (i8*, i8*)*)(i8* %tmp20, i8* %tmp19) 44 %tmp22 = bitcast %5* %tmp21 to i8* 46 %tmp24 = bitcast i8* %tmp23 to %5* 52 %tmp30 = bitcast %1* %arg to i8* 53 %tmp31 = call i8* bitcast (i8* (i8*, i8*, ...)* @objc_msgSend to i8* (i8*, i8*)*)(i8* %tmp30, i8* %tmp29) 56 %tmp39 = call %0* bitcast (i8* (i8*, i8*, ...)* @objc_msgSend to %0* (i8*, i8*)*)(i8* %tmp34, i8* %tmp37) 57 %tmp40 = bitcast %0* %tmp39 to i8 [all...] |
/external/swiftshader/third_party/LLVM/test/Transforms/ObjCARC/ |
move-and-merge-autorelease.ll | 39 %tmp17 = bitcast %3* %arg3 to i8* 40 %tmp18 = call %4* bitcast (i8* (i8*, i8*, ...)* @objc_msgSend to %4* (i8*, i8*)*)(i8* %tmp17, i8* %tmp16) 42 %tmp20 = bitcast %4* %tmp18 to i8* 43 %tmp21 = call %5* bitcast (i8* (i8*, i8*, ...)* @objc_msgSend to %5* (i8*, i8*)*)(i8* %tmp20, i8* %tmp19) 44 %tmp22 = bitcast %5* %tmp21 to i8* 46 %tmp24 = bitcast i8* %tmp23 to %5* 52 %tmp30 = bitcast %1* %arg to i8* 53 %tmp31 = call i8* bitcast (i8* (i8*, i8*, ...)* @objc_msgSend to i8* (i8*, i8*)*)(i8* %tmp30, i8* %tmp29) 56 %tmp39 = call %0* bitcast (i8* (i8*, i8*, ...)* @objc_msgSend to %0* (i8*, i8*)*)(i8* %tmp34, i8* %tmp37) 57 %tmp40 = bitcast %0* %tmp39 to i8 [all...] |
/external/clang/test/CodeGenObjC/ |
arc-unsafeclaim.m | 28 // CHECK-NEXT: [[T1:%.*]] = bitcast [[A]]* [[T0]] to i8* 30 // CHECK-NEXT: [[T3:%.*]] = bitcast i8* [[T2]] to [[A]]* 31 // CHECK-NEXT: [[T4:%.*]] = bitcast [[A]]* [[T3]] to i8* 33 // CHECK-OPTIMIZED-NEXT: bitcast 40 // DISABLED-NEXT: [[T1:%.*]] = bitcast [[A]]* [[T0]] to i8* 52 // CHECK-NEXT: [[T1:%.*]] = bitcast [[A]]* [[T0]] to i8* 54 // CHECK-NEXT: [[T3:%.*]] = bitcast i8* [[T2]] to [[A]]* 55 // CHECK-NEXT: [[T4:%.*]] = bitcast [[A]]* [[T3]] to i8* 58 // CHECK-OPTIMIZED-NEXT: bitcast 60 // CHECK-OPTIMIZED-NEXT: bitcast [all...] |
arc-weak.m | 15 // CHECK-NEXT: [[T1:%.*]] = bitcast [[B]]** [[T0]] to [[A]]** 16 // CHECK-NEXT: [[T2:%.*]] = bitcast [[A]]** [[DEST]] to i8** 17 // CHECK-NEXT: [[T3:%.*]] = bitcast [[A]]** [[T1]] to i8** 19 // CHECK-NEXT: [[T0:%.*]] = bitcast [[A]]** [[DEST]] to i8**
|
related-result-type.m | 17 // CHECK: bitcast i8* 26 // CHECK: bitcast i8* 38 // CHECK-NEXT: bitcast i8* 51 // CHECK-NEXT: bitcast i8*
|
/external/llvm/test/Transforms/RewriteStatepointsForGC/ |
rematerialize-derived-pointers.ll | 17 ; CHECK: bitcast i8 addrspace(1)* %base.relocated to i32 addrspace(1)* 31 ; CHECK: %base.relocated.casted = bitcast i8 addrspace(1)* %base.relocated to i32 addrspace(1)* 41 %ptr = bitcast i32 addrspace(1)* %base to i64 addrspace(1)* 42 ; CHECK: bitcast i32 addrspace(1)* %base to i64 addrspace(1)* 45 ; CHECK: %base.relocated.casted = bitcast i8 addrspace(1)* %base.relocated to i32 addrspace(1)* 46 ; CHECK: bitcast i32 addrspace(1)* %base.relocated.casted to i64 addrspace(1)* 55 %ptr1 = bitcast i32 addrspace(1)* %base to i64 addrspace(1)* 56 %ptr2 = bitcast i64 addrspace(1)* %ptr1 to i16 addrspace(1)* 57 ; CHECK: bitcast i32 addrspace(1)* %base to i64 addrspace(1)* 58 ; CHECK: bitcast i64 addrspace(1)* %ptr1 to i16 addrspace(1) [all...] |
/external/clang/test/CodeGen/ |
x86-atomic-long_double.c | 8 // CHECK: [[INT_ADDR:%.+]] = bitcast x86_fp80* [[ADDR]] to i128* 10 // CHECK: [[INT_LOAD_ADDR:%.+]] = bitcast x86_fp80* [[LD_ADDR:%.+]] to i128* 17 // CHECK: [[OLD_VALUE_VOID_ADDR:%.+]] = bitcast x86_fp80* [[OLD_VALUE_ADDR:%.+]] to i8* 20 // CHECK: [[OLD_INT_ADDR:%.+]] = bitcast x86_fp80* [[OLD_VALUE_ADDR]] to i128* 22 // CHECK: [[NEW_VALUE_VOID_ADDR:%.+]] = bitcast x86_fp80* [[NEW_VALUE_ADDR:%.+]] to i8* 25 // CHECK: [[NEW_INT_ADDR:%.+]] = bitcast x86_fp80* [[NEW_VALUE_ADDR]] to i128* 27 // CHECK: [[OBJ_INT_ADDR:%.+]] = bitcast x86_fp80* [[ADDR]] to i128* 31 // CHECK: [[OLD_VALUE_RES_INT_PTR:%.+]] = bitcast x86_fp80* [[OLD_VALUE_RES_PTR:%.+]] to i128* 40 // CHECK32: [[VOID_PTR:%.+]] = bitcast x86_fp80* [[ADDR]] to i8* 41 // CHECK32: [[TEMP_LD_PTR:%.+]] = bitcast x86_fp80* [[TEMP_LD_ADDR:%.+]] to i8 [all...] |
arm_neon_intrinsics.c | 18 // CHECK: [[TMP0:%.*]] = bitcast <4 x i16> %b to <8 x i8> 19 // CHECK: [[TMP1:%.*]] = bitcast <4 x i16> %c to <8 x i8> 20 // CHECK: [[VABD_V_I_I:%.*]] = bitcast <8 x i8> [[TMP0]] to <4 x i16> 21 // CHECK: [[VABD_V1_I_I:%.*]] = bitcast <8 x i8> [[TMP1]] to <4 x i16> 23 // CHECK: [[VABD_V3_I_I:%.*]] = bitcast <4 x i16> [[VABD_V2_I_I]] to <8 x i8> 24 // CHECK: [[TMP2:%.*]] = bitcast <8 x i8> [[VABD_V3_I_I]] to <4 x i16> 32 // CHECK: [[TMP0:%.*]] = bitcast <2 x i32> %b to <8 x i8> 33 // CHECK: [[TMP1:%.*]] = bitcast <2 x i32> %c to <8 x i8> 34 // CHECK: [[VABD_V_I_I:%.*]] = bitcast <8 x i8> [[TMP0]] to <2 x i32> 35 // CHECK: [[VABD_V1_I_I:%.*]] = bitcast <8 x i8> [[TMP1]] to <2 x i32 [all...] |
pr4349.c | 19 // CHECK: @svars1 = global [1 x %struct.svar] [%struct.svar { i8* bitcast (%struct.cpu* @cpu to i8*) }] 24 // CHECK: @svars2 = global [1 x %struct.svar] [%struct.svar { i8* getelementptr (i8, i8* bitcast (%struct.cpu* @cpu to i8*), i64 1) }] 29 // CHECK: @svars3 = global [1 x %struct.svar] [%struct.svar { i8* getelementptr (i8, i8* bitcast (%struct.cpu* @cpu to i8*), i64 2) }] 34 // CHECK: @svars4 = global [1 x %struct.svar] [%struct.svar { i8* getelementptr (i8, i8* bitcast (%struct.cpu* @cpu to i8*), i64 3) }]
|
/external/llvm/test/CodeGen/SystemZ/ |
vec-combine-02.ll | 18 %hboth0 = bitcast i32 %elem0 to <2 x i16> 19 %hboth1 = bitcast i32 %elem1 to <2 x i16> 20 %hboth2 = bitcast i32 %elem2 to <2 x i16> 21 %hboth3 = bitcast i32 %elem3 to <2 x i16> 22 %hboth4 = bitcast i32 %elem4 to <2 x i16> 23 %hboth5 = bitcast i32 %elem5 to <2 x i16> 24 %hboth6 = bitcast i32 %elem6 to <2 x i16> 25 %hboth7 = bitcast i32 %elem7 to <2 x i16> 65 %hvec0 = bitcast <4 x i32> %wvec0 to <8 x i16> 66 %hvec1 = bitcast <4 x i32> %wvec1 to <8 x i16 [all...] |
/external/llvm/test/CodeGen/X86/ |
avx-logic.ll | 13 %0 = bitcast <4 x double> %x to <4 x i64> 14 %1 = bitcast <4 x double> %y to <4 x i64> 16 %2 = bitcast <4 x i64> %and.i to <4 x double> 30 %0 = bitcast <4 x double> %y to <4 x i64> 32 %1 = bitcast <4 x i64> %and.i to <4 x double> 44 %0 = bitcast <8 x float> %x to <8 x i32> 45 %1 = bitcast <8 x float> %y to <8 x i32> 47 %2 = bitcast <8 x i32> %and.i to <8 x float> 57 %0 = bitcast <8 x float> %y to <8 x i32> 59 %1 = bitcast <8 x i32> %and.i to <8 x float [all...] |
fp-logic.ll | 29 %bc1 = bitcast float %x to i32 43 %bc1 = bitcast float %x to i32 57 %bc1 = bitcast float %x to i32 71 %bc1 = bitcast float %x to i32 86 %bc1 = bitcast float %x to i32 88 %bc2 = bitcast i32 %and to float 102 %bc1 = bitcast float %x to i32 104 %bc2 = bitcast i32 %and to float 117 %bc1 = bitcast float %x to i32 119 %bc2 = bitcast i32 %and to floa [all...] |
x86-64-psub.ll | 18 %2 = bitcast <1 x i64> %__m1.0.insert.i to <8 x i8> 19 %3 = bitcast <8 x i8> %2 to x86_mmx 20 %4 = bitcast <1 x i64> %__m2.0.insert.i to <8 x i8> 21 %5 = bitcast <8 x i8> %4 to x86_mmx 23 %7 = bitcast x86_mmx %6 to <8 x i8> 24 %8 = bitcast <8 x i8> %7 to <1 x i64> 46 %2 = bitcast <1 x i64> %__m1.0.insert.i to <4 x i16> 47 %3 = bitcast <4 x i16> %2 to x86_mmx 48 %4 = bitcast <1 x i64> %__m2.0.insert.i to <4 x i16> 49 %5 = bitcast <4 x i16> %4 to x86_mm [all...] |
lower-bitcast.ll | 6 %1 = bitcast double %A to <2 x i32> 8 %2 = bitcast <2 x i32> %add to double 29 %1 = bitcast double %A to <2 x i32> 30 %2 = bitcast double %B to <2 x i32> 32 %3 = bitcast <2 x i32> %add to double 47 %1 = bitcast i64 %A to <2 x float> 49 %2 = bitcast <2 x float> %add to i64 66 %1 = bitcast i64 %A to <2 x i32> 68 %2 = bitcast <2 x i32> %add to i64 89 %1 = bitcast double %A to <2 x float [all...] |
/external/llvm/test/Transforms/Inline/ |
devirtualize.ll | 43 @_ZTV1D = linkonce_odr constant [6 x i8*] [i8* null, i8* bitcast (%2* @_ZTI1D to i8*), i8* bitcast (i32 (%struct.C*)* @_ZN1D1fEv to i8*), i8* inttoptr (i64 -8 to i8*), i8* bitcast (%2* @_ZTI1D to i8*), i8* bitcast (i32 (%struct.C*)* @_ZThn8_N1D1fEv to i8*)] ; <[6 x i8*]*> [#uses=2] 50 @_ZTI1A = linkonce_odr constant %0 { i8* bitcast (i8** getelementptr inbounds (i8*, i8** @_ZTVN10__cxxabiv117__class_type_infoE, i64 2) to i8*), i8* getelementptr inbounds ([3 x i8], [3 x i8]* @_ZTS1A, i32 0, i32 0) } ; <%0*> [#uses=1] 52 @_ZTI1B = linkonce_odr constant %0 { i8* bitcast (i8** getelementptr inbounds (i8*, i8** @_ZTVN10__cxxabiv117__class_type_infoE, i64 2) to i8*), i8* getelementptr inbounds ([3 x i8], [3 x i8]* @_ZTS1B, i32 0, i32 0) } ; <%0*> [#uses=1] 53 @_ZTI1C = linkonce_odr constant %1 { i8* bitcast (i8** getelementptr inbounds (i8*, i8** @_ZTVN10__cxxabiv121__vmi_class_type_infoE, i64 2) to i8*), i8* getelementptr inbounds ([3 x i8], [3 x i8]* @_ZTS1C, i32 0, i32 0), i32 0, i32 2, i8* bitcast (%0* @_ZTI1A to i8*), i64 2, i8* bitcast (%0* @_ZTI1B to i8*), i64 2050 } ; <%1*> [#uses=1] 54 @_ZTI1D = linkonce_odr constant %2 { i8* bitcast (i8** getelementptr inbounds (i8*, i8** @_ZTVN10__cxxabiv120__si_class_type_infoE, i64 2) to i8*), (…) [all...] |
/external/swiftshader/third_party/LLVM/test/Transforms/Inline/ |
devirtualize.ll | 43 @_ZTV1D = linkonce_odr constant [6 x i8*] [i8* null, i8* bitcast (%2* @_ZTI1D to i8*), i8* bitcast (i32 (%struct.C*)* @_ZN1D1fEv to i8*), i8* inttoptr (i64 -8 to i8*), i8* bitcast (%2* @_ZTI1D to i8*), i8* bitcast (i32 (%struct.C*)* @_ZThn8_N1D1fEv to i8*)] ; <[6 x i8*]*> [#uses=2] 50 @_ZTI1A = linkonce_odr constant %0 { i8* bitcast (i8** getelementptr inbounds (i8** @_ZTVN10__cxxabiv117__class_type_infoE, i64 2) to i8*), i8* getelementptr inbounds ([3 x i8]* @_ZTS1A, i32 0, i32 0) } ; <%0*> [#uses=1] 52 @_ZTI1B = linkonce_odr constant %0 { i8* bitcast (i8** getelementptr inbounds (i8** @_ZTVN10__cxxabiv117__class_type_infoE, i64 2) to i8*), i8* getelementptr inbounds ([3 x i8]* @_ZTS1B, i32 0, i32 0) } ; <%0*> [#uses=1] 53 @_ZTI1C = linkonce_odr constant %1 { i8* bitcast (i8** getelementptr inbounds (i8** @_ZTVN10__cxxabiv121__vmi_class_type_infoE, i64 2) to i8*), i8* getelementptr inbounds ([3 x i8]* @_ZTS1C, i32 0, i32 0), i32 0, i32 2, i8* bitcast (%0* @_ZTI1A to i8*), i64 2, i8* bitcast (%0* @_ZTI1B to i8*), i64 2050 } ; <%1*> [#uses=1] 54 @_ZTI1D = linkonce_odr constant %2 { i8* bitcast (i8** getelementptr inbounds (i8** @_ZTVN10__cxxabiv120__si_class_type_infoE, i64 2) to i8*), i8* (…) [all...] |
/external/llvm/test/Assembler/ |
2003-08-21-ConstantExprCast-Fold.ll | 5 @0 = global i32* bitcast ({ float }* @A to i32*) ; <i32**>:0 [#uses=0]
|
2007-09-10-AliasFwdRef.ll | 5 @__gthread_active_ptr.5335 = internal constant i8* bitcast (i32 (i32)* @__gthrw_pthread_cancel to i8*)
|
MultipleReturnValueType.ll | 12 store void ()* bitcast (%struct.S_102 ()* @f_102 to void ()*), void ()** @callthis, align 8
|
invalid_cast3.ll | 5 %bc = bitcast <4 x i32*> %c to <2 x i32*>
|
/external/llvm/test/Bitcode/ |
extractelement.ll | 6 %tmp24 = extractelement <2 x double> bitcast (<1 x i128> < i128 85070591730234615870450834276742070272 > to <2 x double>), i32 0
|
shuffle.ll | 12 <2 x double> bitcast ( 16 <3 x float> bitcast ( 18 i128 bitcast (<2 x double> bitcast (
|
/external/llvm/test/CodeGen/AArch64/ |
bitcast.ll | 11 ; Can't optimize the following bitcast to scalar_to_vector. 12 %2 = bitcast <2 x i32> %1 to <4 x i16> 23 ; Can't optimize the following bitcast to scalar_to_vector. 24 %2 = bitcast <2 x i32> %1 to <4 x i16>
|
/external/llvm/test/CodeGen/AMDGPU/ |
amdgpu-shader-calling-convention.ll | 8 %vi = bitcast float %v to i32 10 %xf = bitcast i32 %x to float 17 %vi = bitcast float %v to i32 19 %xf = bitcast i32 %x to float
|
/external/llvm/test/CodeGen/ARM/ |
2011-10-26-ExpandUnalignedLoadCrash.ll | 8 %1 = bitcast i16* %destValues to i8* 11 %4 = bitcast i8* %3 to <3 x i16>* 13 %6 = bitcast i16* %sourceA to i8* 15 %8 = bitcast i8* %7 to <3 x i16>*
|