/external/llvm/test/CodeGen/X86/ |
vec_clear.ll | 8 %tmp15 = bitcast <4 x float> %tmp to <2 x i64> ; <<2 x i64>> [#uses=1] 9 %tmp24 = and <2 x i64> %tmp15, bitcast (<4 x i32> < i32 0, i32 0, i32 -1, i32 -1 > to <2 x i64>) ; <<2 x i64>> [#uses=1] 10 %tmp31 = bitcast <2 x i64> %tmp24 to <4 x float> ; <<4 x float>> [#uses=1]
|
vec_shuffle-17.ll | 8 define <2 x i64> @doload64(i64 %x) nounwind { 10 %tmp717 = bitcast i64 %x to double ; <double> [#uses=1] 13 %tmp11 = bitcast <2 x double> %tmp9 to <2 x i64> ; <<2 x i64>> [#uses=1] 14 ret <2 x i64> %tmp11
|
zext-sext.ll | 5 @llvm.used = appending global [1 x i8*] [i8* bitcast (void ([40 x i16]*, i32*, i16**, i64*)* @func to i8*)], section "llvm.metadata" 7 define void @func([40 x i16]* %a, i32* %b, i16** %c, i64* %d) nounwind { 9 %tmp103 = getelementptr inbounds [40 x i16]* %a, i64 0, i64 4 15 %tmp109 = sext i32 %tmp107 to i64 16 %tmp110 = getelementptr inbounds i16* %tmp108, i64 %tmp109 31 %tmp11 = sext i32 %tmp4 to i64 32 %tmp12 = add i64 %tmp11, 5089792279245435153 39 %tmp13 = sub i64 %tmp12, 2138875574 40 %tmp14 = zext i32 %tmp4 to i64 [all...] |
2006-11-12-CSRetCC.ll | 22 %real = getelementptr { double, double }* %tmp1, i64 0, i32 0 ; <double*> [#uses=1] 24 %real3 = getelementptr { double, double }* %tmp1, i64 0, i32 1 ; <double*> [#uses=1] 26 %tmp.upgrd.2 = getelementptr { double, double }* %tmp, i64 0, i32 0 ; <double*> [#uses=1] 27 %tmp4 = getelementptr { double, double }* %tmp1, i64 0, i32 0 ; <double*> [#uses=1] 30 %tmp6 = getelementptr { double, double }* %tmp, i64 0, i32 1 ; <double*> [#uses=1] 31 %tmp7 = getelementptr { double, double }* %tmp1, i64 0, i32 1 ; <double*> [#uses=1] 34 %tmp.upgrd.3 = bitcast { double, double }* %tmp to { i64, i64 }* ; <{ i64, i64 }*> [#uses=1 [all...] |
2009-02-03-AnalyzedTwice.ll | 4 target datalayout = "e-p:32:32:32-i1:8:8-i8:8:8-i16:16:16-i32:32:32-i64:32:64-f32:32:32-f64:32:64-v64:64:64-v128:128:128-a0:0:64-f80:32:32" 8 define void @bar(i64 %p_66) nounwind { 17 %1 = sext i32 %0 to i64 ; <i64> [#uses=1] 18 %2 = or i64 %1, %p_66 ; <i64> [#uses=1] 19 %3 = shl i64 %2, 0 ; <i64> [#uses=1] 20 %4 = and i64 %3, %p_66 ; <i64> [#uses=1 [all...] |
2009-03-05-burr-list-crash.ll | 3 target datalayout = "e-p:64:64:64-i1:8:8-i8:8:8-i16:16:16-i32:32:32-i64:64:64-f32:32:32-f64:64:64-v64:64:64-v128:128:128-a0:0:64-s0:64:64-f80:128:128" 7 declare i64 @strlen(i8* nocapture) nounwind readonly 16 %4 = call i64 @strlen(i8* %0) nounwind readonly ; <i64> [#uses=1] 17 %5 = trunc i64 %4 to i32 ; <i32> [#uses=2] 20 %8 = sext i32 %5 to i64 ; <i64> [#uses=1] 21 %9 = sext i32 %7 to i64 ; <i64> [#uses=1] 22 %10 = sub i64 %8, %9 ; <i64> [#uses=1 [all...] |
2010-08-04-MaskedSignedCompare.ll | 4 @g_16 = global i64 -3738643449681751625, align 8 ; <i64*> [#uses=1] 10 %tmp = load i64* @g_16 ; <i64> [#uses=1] 11 %not.lnot = icmp ne i64 %tmp, 0 ; <i1> [#uses=1] 12 %conv = sext i1 %not.lnot to i64 ; <i64> [#uses=1] 13 %and = and i64 %conv, 150 ; <i64> [#uses=1] 14 %conv.i = trunc i64 %and to i8 ; <i8> [#uses=1 [all...] |
fast-cc-merge-stack-adj.ll | 6 declare x86_fastcallcc void @func(i32*, i64) 8 define x86_fastcallcc void @caller(i32, i64) { 10 call x86_fastcallcc void @func( i32* %X, i64 0 )
|
mmx-bitcast-to-i64.ll | 3 define i64 @foo(x86_mmx* %p) { 6 %s = bitcast x86_mmx %u to i64 7 ret i64 %s 9 define i64 @goo(x86_mmx* %p) { 12 %s = bitcast x86_mmx %u to i64 13 ret i64 %s 15 define i64 @hoo(x86_mmx* %p) { 18 %s = bitcast x86_mmx %u to i64 19 ret i64 %s 21 define i64 @ioo(x86_mmx* %p) [all...] |
subreg-to-reg-0.ll | 6 define i64 @foo(i32* %p) nounwind { 9 %z = zext i32 %n to i64 10 ret i64 %z
|
subreg-to-reg-2.ll | 11 define internal fastcc %XXValue* @t(i64* %out, %"struct.XXC::ArrayStorage"* %tmp9) nounwind { 13 %array = load %XXValue** inttoptr (i64 11111111 to %XXValue**) ; <%XXValue*> [#uses=0] 14 %index = load %XXValue** inttoptr (i64 22222222 to %XXValue**) ; <%XXValue*> [#uses=1] 15 %tmp = ptrtoint %XXValue* %index to i64 ; <i64> [#uses=2] 16 store i64 %tmp, i64* %out 17 %tmp6 = trunc i64 %tmp to i32 ; <i32> [#uses=1] 21 %tmp10 = zext i32 %tmp6 to i64 ; <i64> [#uses=1 [all...] |
wide-integer-fold.ll | 6 define i64 @foo(i192 %a) nounwind { 10 %v = trunc i192 %u to i64 11 ret i64 %v
|
win64_vararg.ll | 25 define i8* @f5(i64 %a0, i64 %a1, i64 %a2, i64 %a3, i64 %a4, ...) nounwind { 36 define i8* @f4(i64 %a0, i64 %a1, i64 %a2, i64 %a3, ...) nounwind { 47 define i8* @f3(i64 %a0, i64 %a1, i64 %a2, ...) nounwind [all...] |
/external/llvm/test/Transforms/IPConstantProp/ |
user-with-multiple-uses.ll | 9 ; CHECK-NEXT: %call2 = tail call i32 @wwrite(i64 0) nounwind 14 %call2 = tail call i32 @wwrite(i64 0) nounwind 18 define internal i32 @wwrite(i64 %i) nounwind readnone { 20 switch i64 %i, label %sw.default [ 21 i64 3, label %return 22 i64 10, label %return
|
/external/llvm/test/Transforms/InstCombine/ |
2010-11-23-Distributed.ll | 13 define i1 @bar(i64 %x, i64 %y) { 15 %a = and i64 %y, %x 18 %not = xor i64 %a, -1 19 %b = and i64 %y, %not 20 %r = icmp eq i64 %b, 0
|
udiv-simplify-bug-1.ll | 9 define i64 @bar(i32 %x) nounwind { 12 %z = sext i32 %r to i64 13 ret i64 %z 15 define i64 @qux(i32 %x, i32 %v) nounwind { 18 %z = sext i32 %r to i64 19 ret i64 %z
|
/external/llvm/test/Transforms/LoopDeletion/ |
multiple-exit-conditions.ll | 13 define void @foo(i64 %n, i64 %m) nounwind { 18 %x.0 = phi i64 [ 0, %entry ], [ %t0, %bb ] 19 %t0 = add i64 %x.0, 1 20 %t1 = icmp slt i64 %x.0, %n 21 %t3 = icmp sgt i64 %x.0, %m
|
multiple-exits.ll | 11 define void @foo(i64 %n, i64 %m) nounwind { 16 %x.0 = phi i64 [ 0, %entry ], [ %t0, %bb2 ] 17 %t0 = add i64 %x.0, 1 18 %t1 = icmp slt i64 %x.0, %n 21 %t2 = icmp slt i64 %x.0, %m
|
/external/llvm/test/Transforms/MemCpyOpt/ |
memcpy-to-memset.ll | 5 declare void @llvm.memcpy.p0i8.p0i8.i64(i8* nocapture, i8* nocapture, i64, i32, i1) nounwind 11 call void @llvm.memcpy.p0i8.p0i8.i64(i8* %arr_i8, i8* bitcast ([3 x i32]* @cst to i8*), i64 12, i32 4, i1 false) 12 %arraydecay = getelementptr inbounds [3 x i32]* %arr, i64 0, i64 0
|
/external/llvm/test/Transforms/TailCallElim/ |
accum_recursion.ll | 43 define i64 @test3_fib(i64 %n) nounwind readnone { 47 ; CHECK: %accumulator.tr = phi i64 [ %n, %entry ], [ %3, %bb1 ] 48 ; CHECK: %n.tr = phi i64 [ %n, %entry ], [ %2, %bb1 ] 49 switch i64 %n, label %bb1 [ 50 ; CHECK: switch i64 %n.tr, label %bb1 [ 51 i64 0, label %bb2 52 i64 1, label %bb2 57 %0 = add i64 %n, -1 58 ; CHECK: %0 = add i64 %n.tr, - [all...] |
/external/llvm/test/Analysis/LoopDependenceAnalysis/ |
siv-strong.ll | 14 %i = phi i64 [ 0, %entry ], [ %i.next, %for.body ] 15 %y.addr = getelementptr [256 x i32]* @y, i64 0, i64 %i 16 %x.addr = getelementptr [256 x i32]* @x, i64 0, i64 %i 23 %i.next = add i64 %i, 1 24 %exitcond = icmp eq i64 %i.next, 256 39 %i = phi i64 [ 0, %entry ], [ %i.next, %for.body ] 40 %y.ld.addr = getelementptr [256 x i32]* @y, i64 0, i64 % [all...] |
alias.ll | 7 %x.ld.addr = getelementptr i32* %xptr, i64 6 8 %x.st.addr = getelementptr i32* %xptr, i64 5 12 %i = phi i64 [ 0, %entry ], [ %i.next, %for.body ] 16 %i.next = add i64 %i, 1 17 %exitcond = icmp eq i64 %i.next, 256 29 %x.ld.addr = getelementptr [256 x i32]* %xptr, i64 0, i64 6 30 %x.st.addr = getelementptr [256 x i32]* %xptr, i64 0, i64 5 34 %i = phi i64 [ 0, %entry ], [ %i.next, %for.body [all...] |
/external/llvm/test/CodeGen/ARM/ |
fmdrr-fmrrd.ll | 10 define i64 @test(double %X) { 11 %Y = bitcast double %X to i64 12 ret i64 %Y
|
/external/llvm/test/CodeGen/CBackend/ |
2003-05-13-VarArgFunction.ll | 6 define void @test(i64 %Ptr) { 7 %P = inttoptr i64 %Ptr to void (...)* ; <void (...)*> [#uses=1] 8 call void (...)* %P( i64 %Ptr )
|
/external/llvm/test/Feature/ |
vector-cast-constant-exprs.ll | 11 define <2 x i64> @gd() { 12 ret <2 x i64> zext (<2 x i32><i32 3, i32 4> to <2 x i64>) 14 define <2 x i64> @ge() { 15 ret <2 x i64> sext (<2 x i32><i32 3, i32 4> to <2 x i64>) 18 ret <2 x i32> trunc (<2 x i64><i64 3, i64 4> to <2 x i32>) 36 ret <2 x double> bitcast (<2 x i64><i64 4, i64 3> to <2 x double> [all...] |