/external/llvm/test/Transforms/ScalarRepl/ |
AggregatePromote.ll | 7 define i64 @test1(i64 %X) { 8 %A = alloca i64 ; <i64*> [#uses=3] 9 store i64 %X, i64* %A 10 %B = bitcast i64* %A to i32* ; <i32*> [#uses=1] 13 %Y = load i64* %A ; <i64> [#uses=1] 14 ret i64 % [all...] |
/external/llvm/test/CodeGen/Alpha/ |
add.ll | 32 define signext i64 @aq(i64 signext %x.s, i64 signext %y.s) { 34 %tmp.3.s = add i64 %y.s, %x.s ; <i64> [#uses=1] 35 ret i64 %tmp.3.s 38 define i64 @aqi(i64 %x.s) { 40 %tmp.3.s = add i64 100, %x.s ; <i64> [#uses=1 [all...] |
/external/llvm/test/Transforms/InstCombine/ |
trunc.ll | 2 target datalayout = "e-p:64:64:64-i1:8:8-i8:8:8-i16:16:16-i32:32:32-i64:64:64-f32:32:32-f64:64:64-v64:64:64-v128:128:128-a0:0:64-s0:64:64-f80:128:128-n8:16:32:64" 8 define i64 @test1(i64 %a) { 9 %b = trunc i64 %a to i32 11 %d = zext i32 %c to i64 13 ret i64 %d 15 ; CHECK: %d = and i64 %a, 15 16 ; CHECK: ret i64 %d 18 define i64 @test2(i64 %a) [all...] |
exact.ll | 72 ; CHECK: %B = ashr exact i64 %A, 2 73 ; CHECK: ret i64 %B 74 define i64 @ashr1(i64 %X) nounwind { 75 %A = shl i64 %X, 8 76 %B = ashr i64 %A, 2 ; X/4 77 ret i64 %B 82 ; CHECK: %B = icmp eq i64 %X, 0 84 define i1 @ashr_icmp1(i64 %X) nounwind { 85 %A = ashr exact i64 %X, 2 ; X/ [all...] |
/external/clang/test/CodeGenCXX/ |
pointers-to-data-members.cpp | 13 // CHECK-GLOBAL: @_ZN8ZeroInit1aE = global i64 -1 16 // CHECK-GLOBAL: @_ZN8ZeroInit2aaE = global [2 x i64] [i64 -1, i64 -1] 19 // CHECK-GLOBAL: @_ZN8ZeroInit3aaaE = global [2 x [2 x i64]] {{\[}}[2 x i64] [i64 -1, i64 -1], [2 x i64] [i64 -1, i64 -1] [all...] |
/external/llvm/test/CodeGen/X86/ |
2009-04-scale.ll | 6 %struct.impl = type { %struct.vtable, i8, %struct.impl*, i32, i32, i64, i64 } 7 %struct.pair = type { i64, i64 } 16 %5 = getelementptr %struct.pair* %4, i32 0, i32 1 ; <i64*> [#uses=1] 17 %6 = load i64* %5, align 4 ; <i64> [#uses=1] 18 %7 = xor i64 0, %6 ; <i64> [#uses=1] 19 %8 = xor i64 %7, 0 ; <i64> [#uses=1 [all...] |
2004-04-09-SameValueCoalescing.ll | 8 define i64 @test(i64 %x) { 10 %tmp.1 = mul i64 %x, 4294967297 ; <i64> [#uses=1] 11 ret i64 %tmp.1
|
2009-01-31-BigShift2.ll | 4 define void @test(<8 x double>* %P, i64* %Q) nounwind { 8 %D = trunc i512 %C to i64 ; <i64> [#uses=1] 9 volatile store i64 %D, i64* %Q
|
alloca-align-rounding-32.ll | 3 declare void @bar(<2 x i64>* %n) 6 %p = alloca <2 x i64>, i32 %h 7 call void @bar(<2 x i64>* %p) 12 %p = alloca <2 x i64>, i32 %h, align 32 13 call void @bar(<2 x i64>* %p)
|
divrem.ll | 3 define void @si64(i64 %x, i64 %y, i64* %p, i64* %q) { 4 %r = sdiv i64 %x, %y 5 %t = srem i64 %x, %y 6 store i64 %r, i64* %p 7 store i64 %t, i64* % [all...] |
masked-iv-safe.ll | 14 define void @count_up(double* %d, i64 %n) nounwind { 19 %indvar = phi i64 [ 0, %entry ], [ %indvar.next, %loop ] 20 %indvar.i8 = and i64 %indvar, 255 21 %t0 = getelementptr double* %d, i64 %indvar.i8 25 %indvar.i24 = and i64 %indvar, 16777215 26 %t3 = getelementptr double* %d, i64 %indvar.i24 30 %t6 = getelementptr double* %d, i64 %indvar 34 %indvar.next = add i64 %indvar, 1 35 %exitcond = icmp eq i64 %indvar.next, 10 42 define void @count_down(double* %d, i64 %n) nounwind [all...] |
vec_set-C.ll | 5 define <2 x i64> @t1(i64 %x) nounwind { 6 %tmp8 = insertelement <2 x i64> zeroinitializer, i64 %x, i32 0 7 ret <2 x i64> %tmp8
|
vec_shuffle-27.ll | 4 target datalayout = "e-p:32:32:32-i1:8:8-i8:8:8-i16:16:16-i32:32:32-i64:32:64-f32:32:32-f64:32:64-v64:64:64-v128:128:128-a0:0:64-f80:32:32" 23 define void @test2(<4 x i64>* %ap, <4 x i64>* %bp) nounwind { 26 %a = load <4 x i64> * %ap 27 %b = load <4 x i64> * %bp 28 %mulaa = mul <4 x i64> %a, %a 29 %mulbb = mul <4 x i64> %b, %b 30 %mulab = mul <4 x i64> %a, %b 31 %vect1271 = shufflevector <4 x i64> %mulaa, <4 x i64> %mulbb, <4 x i32> <i32 0, i32 4, i32 undef, i32 undef [all...] |
optimize-max-1.ll | 6 target datalayout = "e-p:64:64:64-i1:8:8-i8:8:8-i16:16:16-i32:32:32-i64:64:64-f32:32:32-f64:64:64-v64:64:64-v128:128:128-a0:0:64-s0:64:64-f80:128:128" 8 define void @fs(double* nocapture %p, i64 %n) nounwind { 10 %tmp = icmp slt i64 %n, 1 ; <i1> [#uses=1] 11 %smax = select i1 %tmp, i64 1, i64 %n ; <i64> [#uses=1] 15 %i.0 = phi i64 [ 0, %entry ], [ %0, %bb ] ; <i64> [#uses=2] 16 %scevgep = getelementptr double* %p, i64 %i.0 ; <double*> [#uses=1] 18 %0 = add i64 %i.0, 1 ; <i64> [#uses=2 [all...] |
2009-02-12-DebugInfoVLA.ll | 4 target datalayout = "e-p:32:32:32-i1:8:8-i8:8:8-i16:16:16-i32:32:32-i64:32:64-f32:32:32-f64:32:64-v64:64:64-v128:128:128-a0:0:64-f80:128:128" 13 %1 = alloca i64 ; <i64*> [#uses=2] 14 %2 = alloca i64 ; <i64*> [#uses=1] 15 %3 = alloca i64 ; <i64*> [#uses=6] 23 %6 = call i64 @strlen(i8* %5) nounwind readonly, !dbg !13 ; <i64> [#uses=1] 24 %7 = add i64 %6, 1, !dbg !13 ; <i64> [#uses=1 [all...] |
2007-10-14-CoalescerCrash.ll | 5 define i32 @execute_stack_op(i8* %op_ptr, i8* %op_end, %struct._Unwind_Context* %context, i64 %initial) { 13 %tmp10 = load i64* null, align 8 ; <i64> [#uses=1] 14 %tmp16 = load i64* null, align 8 ; <i64> [#uses=1] 18 %tmp2526.cast = and i64 %tmp16, 4294967295 ; <i64> [#uses=1] 19 %tmp27 = ashr i64 %tmp10, %tmp2526.cast ; <i64> [#uses=1] 23 %result.0 = phi i64 [ %tmp27, %bb23 ], [ 0, %bb ], [ 0, %bb6 ] ; <i64> [#uses=0 [all...] |
commute-intrinsic.ll | 3 @a = external global <2 x i64> ; <<2 x i64>*> [#uses=1] 5 define <2 x i64> @madd(<2 x i64> %b) nounwind { 7 %tmp2 = load <2 x i64>* @a, align 16 ; <<2 x i64>> [#uses=1] 8 %tmp6 = bitcast <2 x i64> %b to <8 x i16> ; <<8 x i16>> [#uses=1] 9 %tmp9 = bitcast <2 x i64> %tmp2 to <8 x i16> ; <<8 x i16>> [#uses=1] 11 %tmp14 = bitcast <4 x i32> %tmp11 to <2 x i64> ; <<2 x i64>> [#uses=1 [all...] |
/external/llvm/test/ExecutionEngine/ |
test-cast.ll | 15 zext i1 true to i64 ; <i64>:8 [#uses=0] 16 zext i1 true to i64 ; <i64>:9 [#uses=0] 25 sext i8 4 to i64 ; <i64>:18 [#uses=0] 26 sext i8 4 to i64 ; <i64>:19 [#uses=0] 35 zext i8 4 to i64 ; <i64>:28 [#uses=0 [all...] |
/external/llvm/test/Analysis/ScalarEvolution/ |
2008-06-12-BinomialInt64.ll | 9 %accum.03 = phi i64 [ %tmp14, %bb10 ], [ 0, %entry ] ; <i64> [#uses=1] 11 %d.1.01 = phi i64 [ %tmp5.i, %bb10 ], [ 0, %entry ] ; <i64> [#uses=1] 12 %tmp5.i = add i64 %d.1.01, 1 ; <i64> [#uses=2] 13 %tmp14 = add i64 %accum.03, %tmp5.i ; <i64> [#uses=2] 19 %accum.0.lcssa = phi i64 [ 0, %entry ], [ %tmp14, %bb10 ] ; <i64> [#uses=0 [all...] |
trip-count4.ll | 6 define void @another_count_down_signed(double* %d, i64 %n) nounwind { 11 %indvar = phi i64 [ %n, %entry ], [ %indvar.next, %loop ] ; <i64> [#uses=4] 12 %s0 = shl i64 %indvar, 8 ; <i64> [#uses=1] 13 %indvar.i8 = ashr i64 %s0, 8 ; <i64> [#uses=1] 14 %t0 = getelementptr double* %d, i64 %indvar.i8 ; <double*> [#uses=2] 18 %indvar.next = sub i64 %indvar, 1 ; <i64> [#uses=2 [all...] |
/external/llvm/test/CodeGen/PowerPC/ |
LargeAbsoluteAddr.ll | 9 store i32 0, i32* inttoptr (i64 48725999 to i32*) 14 store i64 0, i64* inttoptr (i64 74560 to i64*)
|
/external/llvm/test/CodeGen/XCore/ |
unaligned_store_combine.ll | 7 define void @f(i64* %dst, i64* %src) nounwind { 9 %0 = load i64* %src, align 1 10 store i64 %0, i64* %dst, align 1
|
/external/llvm/test/Transforms/SimplifyCFG/ |
PR9946.ll | 7 br i1 icmp eq (i64 and (i64 ptrtoint (i32* @foo to i64), i64 15), i64 0), label %if.end, label %if.then
|
/external/llvm/test/Verifier/ |
2008-08-22-MemCpyAlignment.ll | 4 define void @x(i8* %a, i8* %src, i64 %len, i32 %align) nounwind { 6 tail call void @llvm.memcpy.i64( i8* %a, i8* %src, i64 %len, i32 %align) nounwind 10 declare void @llvm.memcpy.i64( i8* %a, i8* %src, i64 %len, i32)
|
/external/llvm/test/CodeGen/Generic/ |
2003-07-07-BadLongConst.ll | 7 define internal i64 @getL() { 9 ret i64 -5787213826675591005 14 %tmp.11 = call i64 @getL( ) ; <i64> [#uses=2] 15 %tmp.5 = trunc i64 %tmp.11 to i32 ; <i32> [#uses=2] 16 %tmp.23 = and i64 %tmp.11, -4294967296 ; <i64> [#uses=2] 17 %tmp.16 = call i32 (i8*, ...)* @printf( i8* getelementptr ([42 x i8]* @.str_1, i64 0, i64 0), i32 %tmp.5, i32 %tmp.5, i64 %tmp.23, i64 %tmp.23 ) ; <i32> [#uses=0 [all...] |