/external/llvm/test/CodeGen/X86/ |
memcpy.ll | 4 declare void @llvm.memcpy.p0i8.p0i8.i64(i8* nocapture, i8* nocapture, i64, i32, i1) nounwind 8 define i8* @test1(i8* %a, i8* %b, i64 %n) nounwind { 10 tail call void @llvm.memcpy.p0i8.p0i8.i64( i8* %a, i8* %b, i64 %n, i32 1, i1 0 ) 18 define i8* @test2(i64* %a, i64* %b, i64 %n) nounwind { 20 %tmp14 = bitcast i64* %a to i8* 21 %tmp25 = bitcast i64* %b to i8 [all...] |
memset64-on-x86-32.ll | 7 call void @llvm.memset.p0i8.i64(i8* null, i8 0, i64 80, i32 4, i1 false) 11 declare void @llvm.memset.p0i8.i64(i8* nocapture, i8, i64, i32, i1) nounwind
|
popcnt.ll | 27 define i64 @cnt64(i64 %x) nounwind readnone { 28 %cnt = tail call i64 @llvm.ctpop.i64(i64 %x) 29 ret i64 %cnt 38 declare i64 @llvm.ctpop.i64(i64) nounwind readnone
|
ret-mmx.ll | 4 @g_v1di = external global <1 x i64> 8 %call = call <1 x i64> @return_v1di() ; <<1 x i64>> [#uses=0] 9 store <1 x i64> %call, <1 x i64>* @g_v1di 17 declare <1 x i64> @return_v1di() 19 define <1 x i64> @t2() nounwind { 20 ret <1 x i64> <i64 1>
|
vec_shift2.ll | 3 define <2 x i64> @t1(<2 x i64> %b1, <2 x i64> %c) nounwind { 4 %tmp1 = bitcast <2 x i64> %b1 to <8 x i16> 6 %tmp3 = bitcast <8 x i16> %tmp2 to <2 x i64> 7 ret <2 x i64> %tmp3 10 define <4 x i32> @t2(<2 x i64> %b1, <2 x i64> %c) nounwind { 11 %tmp1 = bitcast <2 x i64> %b1 to <4 x i32>
|
change-compare-stride-1.ll | 19 %indvar19 = phi i64 [ 0, %bb4.thread ], [ %indvar.next29, %bb4 ] ; <i64> [#uses=3] 20 %indvar31 = trunc i64 %indvar19 to i16 ; <i16> [#uses=1] 24 %tmp21 = mul i64 %indvar19, -478 ; <i64> [#uses=1] 28 %indvar = phi i64 [ 0, %bb2.outer ], [ %indvar.next, %bb2 ] ; <i64> [#uses=3] 29 %indvar16 = trunc i64 %indvar to i16 ; <i16> [#uses=1] 30 %ctg2 = getelementptr i8* %out, i64 %tmp21 ; <i8*> [#uses=1] 31 %tmp22 = ptrtoint i8* %ctg2 to i64 ; <i64> [#uses=1 [all...] |
byval-align.ll | 13 %p = alloca i64 ; <i64*> [#uses=3] 18 %1 = ptrtoint i8* %0 to i64 ; <i64> [#uses=1] 19 store i64 %1, i64* %p, align 8 21 %3 = call i32 (i8*, ...)* @printf(i8* getelementptr inbounds ([10 x i8]* @.str, i64 0, i64 0), i8* %2) nounwind ; <i32> [#uses=0] 22 %4 = load i64* %p, align 8 ; <i64> [#uses=1 [all...] |
x86-64-jumps.ll | 2 target datalayout = "e-p:64:64:64-i1:8:8-i8:8:8-i16:16:16-i32:32:32-i64:64:64-f32:32:32-f64:64:64-v64:64:64-v128:128:128-a0:0:64-s0:64:64-f80:128:128" 19 @test.array = internal constant [3 x i32] [i32 trunc (i64 sub (i64 ptrtoint (i8* blockaddress(@test2, %foo) to i64), i64 ptrtoint (i8* blockaddress(@test2, %foo) to i64)) to i32), i32 trunc (i64 sub (i64 ptrtoint (i8* blockaddress(@test2, %bar) to i64), i64 ptrtoint (i8* blockaddress(@test2, %foo) to i64)) to i32), i32 trunc (i64 sub (i64 ptrtoint (i8* b (…) [all...] |
2011-06-12-FastAllocSpill.ll | 8 target datalayout = "e-p:64:64:64-i1:8:8-i8:8:8-i16:16:16-i32:32:32-i64:64:64-f32:32:32-f64:64:64-v64:64:64-v128:128:128-a0:0:64-s0:64:64-f80:128:128-n8:16:32:64" 11 %0 = type { i64, i64, i8*, i8* } 14 %3 = type <{ i8*, i32, i32, void (%4*)*, i8*, i64 }> 15 %4 = type { i8**, i32, i32, i8**, %5*, i64 } 16 %5 = type { i64, i64 } 23 %tmp5 = alloca i64, align 8 26 store i64 0, i64* %tmp5, align [all...] |
2011-01-24-DbgValue-Before-Use.ll | 3 target datalayout = "e-p:64:64:64-i1:8:8-i8:8:8-i16:16:16-i32:32:32-i64:64:64-f32:32:32-f64:64:64-v64:64:64-v128:128:128-a0:0:64-s0:64:64-f80:128:128-n8:16:32:64" 18 define i64 @gcd(i64 %a, i64 %b) nounwind readnone optsize noinline ssp { 20 tail call void @llvm.dbg.value(metadata !{i64 %a}, i64 0, metadata !10), !dbg !18 21 tail call void @llvm.dbg.value(metadata !{i64 %b}, i64 0, metadata !11), !dbg !19 25 %b.addr.0 = phi i64 [ %b, %entry ], [ %rem, %while.body ] 26 %a.addr.0 = phi i64 [ %a, %entry ], [ %b.addr.0, %while.body [all...] |
/external/llvm/test/ExecutionEngine/ |
test-ret.ll | 32 define i64 @test7() { 33 ret i64 0 36 define i64 @test8() { 37 ret i64 0
|
/external/llvm/test/Transforms/IndVarSimplify/ |
iv-zext.ll | 8 define void @foo(double* %d, i64 %n) nounwind { 13 %indvar = phi i64 [ 0, %entry ], [ %indvar.next, %loop ] 14 %indvar.i8 = and i64 %indvar, 255 15 %t0 = getelementptr double* %d, i64 %indvar.i8 19 %indvar.i24 = and i64 %indvar, 16777215 20 %t3 = getelementptr double* %d, i64 %indvar.i24 24 %t6 = getelementptr double* %d, i64 %indvar 28 %indvar.next = add i64 %indvar, 1 29 %exitcond = icmp eq i64 %indvar.next, 10
|
ashr-tripcount.ll | 7 target datalayout = "e-p:64:64:64-i1:8:8-i8:8:8-i16:16:16-i32:32:32-i64:64:64-f32:32:32-f64:64:64-v64:64:64-v128:128:128-a0:0:64-s0:64:64-f80:128:128-n32:64" 30 %.pn3 = sext i32 %.pn3.in to i64 ; <i64> [#uses=1] 31 %.pn2 = zext i32 %.pn2.in to i64 ; <i64> [#uses=1] 32 %.pn.in = getelementptr [0 x float]* %pow_2_tab.pn, i64 0, i64 %.pn3 ; <float*> [#uses=1] 33 %.pn1.in = getelementptr [0 x float]* %pow_2_025_tab.pn, i64 0, i64 %.pn2 ; <float*> [#uses=1] 48 %t6 = sext i32 %i.05 to i64 ; <i64> [#uses=1 [all...] |
/external/llvm/test/Transforms/InstCombine/ |
2006-12-05-fp-to-int-ext.ll | 7 define i64 @test(double %D) { 9 %B = zext i32 %A to i64 ; <i64> [#uses=1] 10 ret i64 %B
|
2011-10-07-AlignPromotion.ll | 3 target datalayout = "e-p:32:32:32-i1:8:32-i8:8:32-i16:16:32-i32:32:32-i64:32:64-f32:32:32-f64:32:64-v64:32:64-v128:32:128-a0:0:32-n32-S32" 11 %1 = bitcast %struct.CGPoint* %a to i64* 12 %2 = bitcast %struct.CGPoint* %Point to i64* 13 %3 = load i64* %1, align 4 14 store i64 %3, i64* %2, align 4 15 call void @foo(i64* %2) nounwind 17 ; CHECK: %Point = alloca i64, align 4 20 declare void @foo(i64*)
|
add2.ll | 3 define i64 @test1(i64 %A, i32 %B) { 4 %tmp12 = zext i32 %B to i64 5 %tmp3 = shl i64 %tmp12, 32 6 %tmp5 = add i64 %tmp3, %A 7 %tmp6 = and i64 %tmp5, 123 8 ret i64 %tmp6 10 ; CHECK-NEXT: and i64 %A, 123 11 ; CHECK-NEXT: ret i64
|
devirt.ll | 8 @_ZL1p = internal constant { i64, i64 } { i64 1, i64 0 }, align 8 12 %tmp = load { i64, i64 }* @_ZL1p, align 8 13 %memptr.adj = extractvalue { i64, i64 } %tmp, 1 15 %1 = getelementptr inbounds i8* %0, i64 %memptr.adj 17 %memptr.ptr = extractvalue { i64, i64 } %tmp, [all...] |
cast.ll | 3 target datalayout = "E-p:64:64:64-a0:0:8-f32:32:32-f64:64:64-i1:8:8-i8:8:8-i16:16:16-i32:32:32-i64:32:64-v64:64:64-v128:128:128-n8:16:32:64" 14 define i64 @test2(i8 %A) { 17 %Ret = zext i32 %c2 to i64 ; <i64> [#uses=1] 18 ret i64 %Ret 19 ; CHECK: %Ret = zext i8 %A to i64 20 ; CHECK: ret i64 %Ret 24 define i64 @test3(i64 %A) { 25 %c1 = trunc i64 %A to i8 ; <i8> [#uses=1 [all...] |
/external/llvm/test/Transforms/LoopStrengthReduce/ |
hoist-parent-preheader.ll | 9 %indvar419 = phi i64 [ %indvar.next420, %loop2.exit ], [ 0, %entry ] 10 %tmp474 = shl i64 %indvar419, 2 11 %tmp484 = add i64 %tmp474, 4 15 %indvar414 = phi i64 [ %indvar.next415, %loop2.backedge ], [ 0, %loop1 ] 16 %tmp473 = mul i64 %indvar414, -4 17 %tmp485 = add i64 %tmp484, %tmp473 18 %storemerge4 = getelementptr i8* %a, i64 %tmp485 23 %indvar.next415 = add i64 %indvar414, 1 27 %indvar.next420 = add i64 %indvar419, 1
|
/external/llvm/test/Transforms/SimplifyLibCalls/ |
2009-02-11-NotInitialized.ll | 8 %0 = tail call i64 @strlen(i8* %x) nounwind readonly ; <i64> [#uses=1] 9 %1 = trunc i64 %0 to i32 ; <i32> [#uses=1] 13 declare i64 @strlen(i8*) nounwind readonly
|
/external/llvm/test/Transforms/ConstProp/ |
loads.ll | 3 target datalayout = "e-p:64:64:64-i1:8:8-i8:8:8-i16:16:16-i32:32:32-i64:64:64-f32:32:32-f64:64:64-v64:64:64-v128:128:128-a0:0:64-s0:64:64-f80:128:128" 7 @g3 = constant {i64, i64} { i64 123, i64 112312312 } 45 define i64 @test6() { 46 %r = load i64* bitcast(double* @g2 to i64*) 47 ret i64 %r 50 ; CHECK: ret i64 460718241880001740 [all...] |
/external/llvm/test/Transforms/ObjCARC/ |
nested.ll | 3 %struct.__objcFastEnumerationState = type { i64, i8**, i64*, [5 x i64] } 6 @"\01L_OBJC_SELECTOR_REFERENCES_" = internal global i8* getelementptr inbounds ([43 x i8]* @"\01L_OBJC_METH_VAR_NAME_", i64 0, i64 0), section "__DATA, __objc_selrefs, literal_pointers, no_dead_strip" 15 declare void @llvm.memset.p0i8.i64(i8* nocapture, i8, i64, i32, i1) nounwind 34 call void @llvm.memset.p0i8.i64(i8* %tmp, i8 0, i64 64, i32 8, i1 false) 37 %call = call i64 bitcast (i8* (i8*, i8*, ...)* @objc_msgSend to i64 (i8*, i8*, %struct.__objcFastEnumerationState*, (…) [all...] |
/external/llvm/test/CodeGen/PowerPC/ |
2008-04-10-LiveIntervalCrash.ll | 3 define fastcc i64 @nonzero_bits1() nounwind { 58 ret i64 0 62 ret i64 0 64 ret i64 0 66 ret i64 0 68 ret i64 0 70 ret i64 0 72 ret i64 0 74 ret i64 0 76 ret i64 [all...] |
2007-04-30-InlineAsmEarlyClobber.ll | 10 target datalayout = "E-p:32:32:32-i1:8:8-i8:8:8-i16:16:16-i32:32:32-i64:32:64-f32:32:32-f64:32:64-v64:64:64-v128:128:128-a0:0:64" 21 define i64 @test(i32 %A, i32 %B, i32 %C) nounwind { 26 %tmp56 = zext i32 %tmp5 to i64 ; <i64> [#uses=1] 27 %tmp7 = shl i64 %tmp56, 32 ; <i64> [#uses=1] 28 %tmp89 = zext i32 %tmp4 to i64 ; <i64> [#uses=1] 29 %tmp10 = or i64 %tmp7, %tmp89 ; <i64> [#uses=1 [all...] |
/external/llvm/test/CodeGen/MSP430/ |
2009-11-20-NewNode.ll | 4 define i64 @_strtoll_r(i16 %base) nounwind { 12 %cond66 = select i1 undef, i64 -9223372036854775808, i64 9223372036854775807 ; <i64> [#uses=3] 13 %conv69 = sext i16 %base to i64 ; <i64> [#uses=1] 14 %div = udiv i64 %cond66, %conv69 ; <i64> [#uses=1] 24 %cmp123 = icmp ugt i64 undef, %div ; <i1> [#uses=1] 32 ret i64 %cond6 [all...] |