HomeSort by relevance Sort by last modified time
    Searched full:bitcast (Results 226 - 250 of 4002) sorted by null

1 2 3 4 5 6 7 8 91011>>

  /external/llvm/test/Transforms/ObjCARC/
move-and-merge-autorelease.ll 39 %tmp17 = bitcast %3* %arg3 to i8*
40 %tmp18 = call %4* bitcast (i8* (i8*, i8*, ...)* @objc_msgSend to %4* (i8*, i8*)*)(i8* %tmp17, i8* %tmp16)
42 %tmp20 = bitcast %4* %tmp18 to i8*
43 %tmp21 = call %5* bitcast (i8* (i8*, i8*, ...)* @objc_msgSend to %5* (i8*, i8*)*)(i8* %tmp20, i8* %tmp19)
44 %tmp22 = bitcast %5* %tmp21 to i8*
46 %tmp24 = bitcast i8* %tmp23 to %5*
52 %tmp30 = bitcast %1* %arg to i8*
53 %tmp31 = call i8* bitcast (i8* (i8*, i8*, ...)* @objc_msgSend to i8* (i8*, i8*)*)(i8* %tmp30, i8* %tmp29)
56 %tmp39 = call %0* bitcast (i8* (i8*, i8*, ...)* @objc_msgSend to %0* (i8*, i8*)*)(i8* %tmp34, i8* %tmp37)
57 %tmp40 = bitcast %0* %tmp39 to i8
    [all...]
  /external/swiftshader/third_party/LLVM/test/Transforms/ObjCARC/
move-and-merge-autorelease.ll 39 %tmp17 = bitcast %3* %arg3 to i8*
40 %tmp18 = call %4* bitcast (i8* (i8*, i8*, ...)* @objc_msgSend to %4* (i8*, i8*)*)(i8* %tmp17, i8* %tmp16)
42 %tmp20 = bitcast %4* %tmp18 to i8*
43 %tmp21 = call %5* bitcast (i8* (i8*, i8*, ...)* @objc_msgSend to %5* (i8*, i8*)*)(i8* %tmp20, i8* %tmp19)
44 %tmp22 = bitcast %5* %tmp21 to i8*
46 %tmp24 = bitcast i8* %tmp23 to %5*
52 %tmp30 = bitcast %1* %arg to i8*
53 %tmp31 = call i8* bitcast (i8* (i8*, i8*, ...)* @objc_msgSend to i8* (i8*, i8*)*)(i8* %tmp30, i8* %tmp29)
56 %tmp39 = call %0* bitcast (i8* (i8*, i8*, ...)* @objc_msgSend to %0* (i8*, i8*)*)(i8* %tmp34, i8* %tmp37)
57 %tmp40 = bitcast %0* %tmp39 to i8
    [all...]
  /external/clang/test/CodeGenObjC/
arc-unsafeclaim.m 28 // CHECK-NEXT: [[T1:%.*]] = bitcast [[A]]* [[T0]] to i8*
30 // CHECK-NEXT: [[T3:%.*]] = bitcast i8* [[T2]] to [[A]]*
31 // CHECK-NEXT: [[T4:%.*]] = bitcast [[A]]* [[T3]] to i8*
33 // CHECK-OPTIMIZED-NEXT: bitcast
40 // DISABLED-NEXT: [[T1:%.*]] = bitcast [[A]]* [[T0]] to i8*
52 // CHECK-NEXT: [[T1:%.*]] = bitcast [[A]]* [[T0]] to i8*
54 // CHECK-NEXT: [[T3:%.*]] = bitcast i8* [[T2]] to [[A]]*
55 // CHECK-NEXT: [[T4:%.*]] = bitcast [[A]]* [[T3]] to i8*
58 // CHECK-OPTIMIZED-NEXT: bitcast
60 // CHECK-OPTIMIZED-NEXT: bitcast
    [all...]
arc-weak.m 15 // CHECK-NEXT: [[T1:%.*]] = bitcast [[B]]** [[T0]] to [[A]]**
16 // CHECK-NEXT: [[T2:%.*]] = bitcast [[A]]** [[DEST]] to i8**
17 // CHECK-NEXT: [[T3:%.*]] = bitcast [[A]]** [[T1]] to i8**
19 // CHECK-NEXT: [[T0:%.*]] = bitcast [[A]]** [[DEST]] to i8**
related-result-type.m 17 // CHECK: bitcast i8*
26 // CHECK: bitcast i8*
38 // CHECK-NEXT: bitcast i8*
51 // CHECK-NEXT: bitcast i8*
  /external/llvm/test/Transforms/RewriteStatepointsForGC/
rematerialize-derived-pointers.ll 17 ; CHECK: bitcast i8 addrspace(1)* %base.relocated to i32 addrspace(1)*
31 ; CHECK: %base.relocated.casted = bitcast i8 addrspace(1)* %base.relocated to i32 addrspace(1)*
41 %ptr = bitcast i32 addrspace(1)* %base to i64 addrspace(1)*
42 ; CHECK: bitcast i32 addrspace(1)* %base to i64 addrspace(1)*
45 ; CHECK: %base.relocated.casted = bitcast i8 addrspace(1)* %base.relocated to i32 addrspace(1)*
46 ; CHECK: bitcast i32 addrspace(1)* %base.relocated.casted to i64 addrspace(1)*
55 %ptr1 = bitcast i32 addrspace(1)* %base to i64 addrspace(1)*
56 %ptr2 = bitcast i64 addrspace(1)* %ptr1 to i16 addrspace(1)*
57 ; CHECK: bitcast i32 addrspace(1)* %base to i64 addrspace(1)*
58 ; CHECK: bitcast i64 addrspace(1)* %ptr1 to i16 addrspace(1)
    [all...]
  /external/clang/test/CodeGen/
x86-atomic-long_double.c 8 // CHECK: [[INT_ADDR:%.+]] = bitcast x86_fp80* [[ADDR]] to i128*
10 // CHECK: [[INT_LOAD_ADDR:%.+]] = bitcast x86_fp80* [[LD_ADDR:%.+]] to i128*
17 // CHECK: [[OLD_VALUE_VOID_ADDR:%.+]] = bitcast x86_fp80* [[OLD_VALUE_ADDR:%.+]] to i8*
20 // CHECK: [[OLD_INT_ADDR:%.+]] = bitcast x86_fp80* [[OLD_VALUE_ADDR]] to i128*
22 // CHECK: [[NEW_VALUE_VOID_ADDR:%.+]] = bitcast x86_fp80* [[NEW_VALUE_ADDR:%.+]] to i8*
25 // CHECK: [[NEW_INT_ADDR:%.+]] = bitcast x86_fp80* [[NEW_VALUE_ADDR]] to i128*
27 // CHECK: [[OBJ_INT_ADDR:%.+]] = bitcast x86_fp80* [[ADDR]] to i128*
31 // CHECK: [[OLD_VALUE_RES_INT_PTR:%.+]] = bitcast x86_fp80* [[OLD_VALUE_RES_PTR:%.+]] to i128*
40 // CHECK32: [[VOID_PTR:%.+]] = bitcast x86_fp80* [[ADDR]] to i8*
41 // CHECK32: [[TEMP_LD_PTR:%.+]] = bitcast x86_fp80* [[TEMP_LD_ADDR:%.+]] to i8
    [all...]
arm_neon_intrinsics.c 18 // CHECK: [[TMP0:%.*]] = bitcast <4 x i16> %b to <8 x i8>
19 // CHECK: [[TMP1:%.*]] = bitcast <4 x i16> %c to <8 x i8>
20 // CHECK: [[VABD_V_I_I:%.*]] = bitcast <8 x i8> [[TMP0]] to <4 x i16>
21 // CHECK: [[VABD_V1_I_I:%.*]] = bitcast <8 x i8> [[TMP1]] to <4 x i16>
23 // CHECK: [[VABD_V3_I_I:%.*]] = bitcast <4 x i16> [[VABD_V2_I_I]] to <8 x i8>
24 // CHECK: [[TMP2:%.*]] = bitcast <8 x i8> [[VABD_V3_I_I]] to <4 x i16>
32 // CHECK: [[TMP0:%.*]] = bitcast <2 x i32> %b to <8 x i8>
33 // CHECK: [[TMP1:%.*]] = bitcast <2 x i32> %c to <8 x i8>
34 // CHECK: [[VABD_V_I_I:%.*]] = bitcast <8 x i8> [[TMP0]] to <2 x i32>
35 // CHECK: [[VABD_V1_I_I:%.*]] = bitcast <8 x i8> [[TMP1]] to <2 x i32
    [all...]
pr4349.c 19 // CHECK: @svars1 = global [1 x %struct.svar] [%struct.svar { i8* bitcast (%struct.cpu* @cpu to i8*) }]
24 // CHECK: @svars2 = global [1 x %struct.svar] [%struct.svar { i8* getelementptr (i8, i8* bitcast (%struct.cpu* @cpu to i8*), i64 1) }]
29 // CHECK: @svars3 = global [1 x %struct.svar] [%struct.svar { i8* getelementptr (i8, i8* bitcast (%struct.cpu* @cpu to i8*), i64 2) }]
34 // CHECK: @svars4 = global [1 x %struct.svar] [%struct.svar { i8* getelementptr (i8, i8* bitcast (%struct.cpu* @cpu to i8*), i64 3) }]
  /external/llvm/test/CodeGen/SystemZ/
vec-combine-02.ll 18 %hboth0 = bitcast i32 %elem0 to <2 x i16>
19 %hboth1 = bitcast i32 %elem1 to <2 x i16>
20 %hboth2 = bitcast i32 %elem2 to <2 x i16>
21 %hboth3 = bitcast i32 %elem3 to <2 x i16>
22 %hboth4 = bitcast i32 %elem4 to <2 x i16>
23 %hboth5 = bitcast i32 %elem5 to <2 x i16>
24 %hboth6 = bitcast i32 %elem6 to <2 x i16>
25 %hboth7 = bitcast i32 %elem7 to <2 x i16>
65 %hvec0 = bitcast <4 x i32> %wvec0 to <8 x i16>
66 %hvec1 = bitcast <4 x i32> %wvec1 to <8 x i16
    [all...]
  /external/llvm/test/CodeGen/X86/
avx-logic.ll 13 %0 = bitcast <4 x double> %x to <4 x i64>
14 %1 = bitcast <4 x double> %y to <4 x i64>
16 %2 = bitcast <4 x i64> %and.i to <4 x double>
30 %0 = bitcast <4 x double> %y to <4 x i64>
32 %1 = bitcast <4 x i64> %and.i to <4 x double>
44 %0 = bitcast <8 x float> %x to <8 x i32>
45 %1 = bitcast <8 x float> %y to <8 x i32>
47 %2 = bitcast <8 x i32> %and.i to <8 x float>
57 %0 = bitcast <8 x float> %y to <8 x i32>
59 %1 = bitcast <8 x i32> %and.i to <8 x float
    [all...]
fp-logic.ll 29 %bc1 = bitcast float %x to i32
43 %bc1 = bitcast float %x to i32
57 %bc1 = bitcast float %x to i32
71 %bc1 = bitcast float %x to i32
86 %bc1 = bitcast float %x to i32
88 %bc2 = bitcast i32 %and to float
102 %bc1 = bitcast float %x to i32
104 %bc2 = bitcast i32 %and to float
117 %bc1 = bitcast float %x to i32
119 %bc2 = bitcast i32 %and to floa
    [all...]
x86-64-psub.ll 18 %2 = bitcast <1 x i64> %__m1.0.insert.i to <8 x i8>
19 %3 = bitcast <8 x i8> %2 to x86_mmx
20 %4 = bitcast <1 x i64> %__m2.0.insert.i to <8 x i8>
21 %5 = bitcast <8 x i8> %4 to x86_mmx
23 %7 = bitcast x86_mmx %6 to <8 x i8>
24 %8 = bitcast <8 x i8> %7 to <1 x i64>
46 %2 = bitcast <1 x i64> %__m1.0.insert.i to <4 x i16>
47 %3 = bitcast <4 x i16> %2 to x86_mmx
48 %4 = bitcast <1 x i64> %__m2.0.insert.i to <4 x i16>
49 %5 = bitcast <4 x i16> %4 to x86_mm
    [all...]
lower-bitcast.ll 6 %1 = bitcast double %A to <2 x i32>
8 %2 = bitcast <2 x i32> %add to double
29 %1 = bitcast double %A to <2 x i32>
30 %2 = bitcast double %B to <2 x i32>
32 %3 = bitcast <2 x i32> %add to double
47 %1 = bitcast i64 %A to <2 x float>
49 %2 = bitcast <2 x float> %add to i64
66 %1 = bitcast i64 %A to <2 x i32>
68 %2 = bitcast <2 x i32> %add to i64
89 %1 = bitcast double %A to <2 x float
    [all...]
  /external/llvm/test/Transforms/Inline/
devirtualize.ll 43 @_ZTV1D = linkonce_odr constant [6 x i8*] [i8* null, i8* bitcast (%2* @_ZTI1D to i8*), i8* bitcast (i32 (%struct.C*)* @_ZN1D1fEv to i8*), i8* inttoptr (i64 -8 to i8*), i8* bitcast (%2* @_ZTI1D to i8*), i8* bitcast (i32 (%struct.C*)* @_ZThn8_N1D1fEv to i8*)] ; <[6 x i8*]*> [#uses=2]
50 @_ZTI1A = linkonce_odr constant %0 { i8* bitcast (i8** getelementptr inbounds (i8*, i8** @_ZTVN10__cxxabiv117__class_type_infoE, i64 2) to i8*), i8* getelementptr inbounds ([3 x i8], [3 x i8]* @_ZTS1A, i32 0, i32 0) } ; <%0*> [#uses=1]
52 @_ZTI1B = linkonce_odr constant %0 { i8* bitcast (i8** getelementptr inbounds (i8*, i8** @_ZTVN10__cxxabiv117__class_type_infoE, i64 2) to i8*), i8* getelementptr inbounds ([3 x i8], [3 x i8]* @_ZTS1B, i32 0, i32 0) } ; <%0*> [#uses=1]
53 @_ZTI1C = linkonce_odr constant %1 { i8* bitcast (i8** getelementptr inbounds (i8*, i8** @_ZTVN10__cxxabiv121__vmi_class_type_infoE, i64 2) to i8*), i8* getelementptr inbounds ([3 x i8], [3 x i8]* @_ZTS1C, i32 0, i32 0), i32 0, i32 2, i8* bitcast (%0* @_ZTI1A to i8*), i64 2, i8* bitcast (%0* @_ZTI1B to i8*), i64 2050 } ; <%1*> [#uses=1]
54 @_ZTI1D = linkonce_odr constant %2 { i8* bitcast (i8** getelementptr inbounds (i8*, i8** @_ZTVN10__cxxabiv120__si_class_type_infoE, i64 2) to i8*), (…)
    [all...]
  /external/swiftshader/third_party/LLVM/test/Transforms/Inline/
devirtualize.ll 43 @_ZTV1D = linkonce_odr constant [6 x i8*] [i8* null, i8* bitcast (%2* @_ZTI1D to i8*), i8* bitcast (i32 (%struct.C*)* @_ZN1D1fEv to i8*), i8* inttoptr (i64 -8 to i8*), i8* bitcast (%2* @_ZTI1D to i8*), i8* bitcast (i32 (%struct.C*)* @_ZThn8_N1D1fEv to i8*)] ; <[6 x i8*]*> [#uses=2]
50 @_ZTI1A = linkonce_odr constant %0 { i8* bitcast (i8** getelementptr inbounds (i8** @_ZTVN10__cxxabiv117__class_type_infoE, i64 2) to i8*), i8* getelementptr inbounds ([3 x i8]* @_ZTS1A, i32 0, i32 0) } ; <%0*> [#uses=1]
52 @_ZTI1B = linkonce_odr constant %0 { i8* bitcast (i8** getelementptr inbounds (i8** @_ZTVN10__cxxabiv117__class_type_infoE, i64 2) to i8*), i8* getelementptr inbounds ([3 x i8]* @_ZTS1B, i32 0, i32 0) } ; <%0*> [#uses=1]
53 @_ZTI1C = linkonce_odr constant %1 { i8* bitcast (i8** getelementptr inbounds (i8** @_ZTVN10__cxxabiv121__vmi_class_type_infoE, i64 2) to i8*), i8* getelementptr inbounds ([3 x i8]* @_ZTS1C, i32 0, i32 0), i32 0, i32 2, i8* bitcast (%0* @_ZTI1A to i8*), i64 2, i8* bitcast (%0* @_ZTI1B to i8*), i64 2050 } ; <%1*> [#uses=1]
54 @_ZTI1D = linkonce_odr constant %2 { i8* bitcast (i8** getelementptr inbounds (i8** @_ZTVN10__cxxabiv120__si_class_type_infoE, i64 2) to i8*), i8* (…)
    [all...]
  /external/llvm/test/Assembler/
2003-08-21-ConstantExprCast-Fold.ll 5 @0 = global i32* bitcast ({ float }* @A to i32*) ; <i32**>:0 [#uses=0]
2007-09-10-AliasFwdRef.ll 5 @__gthread_active_ptr.5335 = internal constant i8* bitcast (i32 (i32)* @__gthrw_pthread_cancel to i8*)
MultipleReturnValueType.ll 12 store void ()* bitcast (%struct.S_102 ()* @f_102 to void ()*), void ()** @callthis, align 8
invalid_cast3.ll 5 %bc = bitcast <4 x i32*> %c to <2 x i32*>
  /external/llvm/test/Bitcode/
extractelement.ll 6 %tmp24 = extractelement <2 x double> bitcast (<1 x i128> < i128 85070591730234615870450834276742070272 > to <2 x double>), i32 0
shuffle.ll 12 <2 x double> bitcast (
16 <3 x float> bitcast (
18 i128 bitcast (<2 x double> bitcast (
  /external/llvm/test/CodeGen/AArch64/
bitcast.ll 11 ; Can't optimize the following bitcast to scalar_to_vector.
12 %2 = bitcast <2 x i32> %1 to <4 x i16>
23 ; Can't optimize the following bitcast to scalar_to_vector.
24 %2 = bitcast <2 x i32> %1 to <4 x i16>
  /external/llvm/test/CodeGen/AMDGPU/
amdgpu-shader-calling-convention.ll 8 %vi = bitcast float %v to i32
10 %xf = bitcast i32 %x to float
17 %vi = bitcast float %v to i32
19 %xf = bitcast i32 %x to float
  /external/llvm/test/CodeGen/ARM/
2011-10-26-ExpandUnalignedLoadCrash.ll 8 %1 = bitcast i16* %destValues to i8*
11 %4 = bitcast i8* %3 to <3 x i16>*
13 %6 = bitcast i16* %sourceA to i8*
15 %8 = bitcast i8* %7 to <3 x i16>*

Completed in 1480 milliseconds

1 2 3 4 5 6 7 8 91011>>