HomeSort by relevance Sort by last modified time
    Searched full:i64 (Results 26 - 50 of 3211) sorted by null

12 3 4 5 6 7 8 91011>>

  /external/llvm/test/Transforms/InstCombine/
2003-05-27-ConstExprCrash.ll 5 define i64 @test() {
6 %C = add i64 1, 2 ; <i64> [#uses=1]
7 %V = add i64 ptrtoint (i32* @X to i64), %C ; <i64> [#uses=1]
8 ret i64 %V
srem.ll 3 define i64 @foo(i64 %x1, i64 %y2) {
4 %r = sdiv i64 %x1, %y2
5 %r7 = mul i64 %r, %y2
6 %r8 = sub i64 %x1, %r7
7 ret i64 %r8
urem.ll 3 define i64 @rem_unsigned(i64 %x1, i64 %y2) {
4 %r = udiv i64 %x1, %y2
5 %r7 = mul i64 %r, %y2
6 %r8 = sub i64 %x1, %r7
7 ret i64 %r8
constant-fold-gep.ll 2 target datalayout = "E-p:64:64:64-i1:8:8-i8:8:8-i16:16:16-i32:32:32-i64:32:64-f32:32:32-f64:64:64-v64:64:64-v128:128:128-a0:0:64"
12 ; CHECK: store i32 1, i32* getelementptr inbounds ([3 x %struct.X]* @Y, i64 0, i64 0, i32 0, i64 0), align 16
13 store i32 1, i32* getelementptr ([3 x %struct.X]* @Y, i64 0, i64 0, i32 0, i64 0), align 4
14 ; CHECK: store i32 1, i32* getelementptr inbounds ([3 x %struct.X]* @Y, i64 0, i64 0, i32 0, i64 1), align
    [all...]
  /external/llvm/test/CodeGen/X86/
inline-asm-2addr.ll 3 define i64 @t(i64 %a, i64 %b) nounwind ssp {
5 %asmtmp = tail call i64 asm "rorq $1,$0", "=r,J,0,~{dirflag},~{fpsr},~{flags},~{cc}"(i32 1, i64 %a) nounwind ; <i64> [#uses=1]
6 %asmtmp1 = tail call i64 asm "rorq $1,$0", "=r,J,0,~{dirflag},~{fpsr},~{flags},~{cc}"(i32 1, i64 %b) nounwind ; <i64> [#uses=1]
7 %0 = add i64 %asmtmp1, %asmtmp ; <i64> [#uses=1
    [all...]
sse42_64.ll 3 declare i64 @llvm.x86.sse42.crc32.64.8(i64, i8) nounwind
4 declare i64 @llvm.x86.sse42.crc32.64.64(i64, i64) nounwind
6 define i64 @crc32_64_8(i64 %a, i8 %b) nounwind {
7 %tmp = call i64 @llvm.x86.sse42.crc32.64.8(i64 %a, i8 %b)
8 ret i64 %tm
    [all...]
2009-06-02-RewriterBug.ll 4 define void @sha256_block1(i32* nocapture %arr, i8* nocapture %in, i64 %num) nounwind {
12 %indvar2787 = phi i64 [ 0, %bb.nph ], [ %indvar.next2788, %for.end ] ; <i64> [#uses=2]
13 %tmp2791 = mul i64 %indvar2787, 44 ; <i64> [#uses=0]
14 %ctg22996 = getelementptr i8* %in, i64 0 ; <i8*> [#uses=1]
15 %conv = zext i32 undef to i64 ; <i64> [#uses=1]
16 %conv11 = zext i32 undef to i64 ; <i64> [#uses=1
    [all...]
2009-04-16-SpillerUnfold.ll 6 %struct.SHA512_CTX = type { [8 x i64], i64, i64, %struct.anon, i32, i32 }
7 %struct.anon = type { [16 x i64] }
8 @K512 = external constant [80 x i64], align 32 ; <[80 x i64]*> [#uses=2]
10 define fastcc void @sha512_block_data_order(%struct.SHA512_CTX* nocapture %ctx, i8* nocapture %in, i64 %num) nounwind ssp {
15 %e.0489 = phi i64 [ 0, %entry ], [ %e.0, %bb349 ] ; <i64> [#uses=3]
16 %b.0472 = phi i64 [ 0, %entry ], [ %87, %bb349 ] ; <i64> [#uses=2
    [all...]
inline-asm-tied.ll 7 @llvm.used = appending global [1 x i8*] [i8* bitcast (i64 (i64)* @_OSSwapInt64 to i8*)], section "llvm.metadata" ; <[1 x i8*]*> [#uses=0]
9 define i64 @_OSSwapInt64(i64 %_data) nounwind {
11 %retval = alloca i64 ; <i64*> [#uses=2]
12 %_data.addr = alloca i64 ; <i64*> [#uses=4]
13 store i64 %_data, i64* %_data.add
    [all...]
2007-10-31-extractelement-i64.ll 3 target datalayout = "e-p:32:32:32-i1:8:8-i8:8:8-i16:16:16-i32:32:32-i64:32:64-f32:32:32-f64:32:64-v64:64:64-v128:128:128-a0:0:64-f80:128:128"
6 define <1 x i64> @a(<2 x i64> %__A) {
8 %__A_addr = alloca <2 x i64> ; <<2 x i64>*> [#uses=2]
9 %retval = alloca <1 x i64>, align 8 ; <<1 x i64>*> [#uses=3]
11 store <2 x i64> %__A, <2 x i64>* %__A_addr
12 %tmp = load <2 x i64>* %__A_addr, align 16 ; <<2 x i64>> [#uses=1
    [all...]
h-registers-1.ll 10 define i64 @foo(i64 %a, i64 %b, i64 %c, i64 %d,
11 i64 %e, i64 %f, i64 %g, i64 %h) {
12 %sa = lshr i64 %a,
    [all...]
2006-07-10-InlineAsmAConstraint.ll 4 define i64 @test() {
5 %tmp.i5 = call i64 asm sideeffect "rdtsc", "=A,~{dirflag},~{fpsr},~{flags}"( ) ; <i64> [#uses=1]
6 ret i64 %tmp.i5
2007-11-04-LiveVariablesBug.ll 4 define void @xor_sse_2(i64 %bytes, i64* %p1, i64* %p2) {
6 %p2_addr = alloca i64* ; <i64**> [#uses=2]
8 store i64* %p2, i64** %p2_addr, align 8
9 %tmp1 = lshr i64 %bytes, 8 ; <i64> [#uses=1]
10 %tmp12 = trunc i64 %tmp1 to i32 ; <i32> [#uses=2
    [all...]
  /external/llvm/test/CodeGen/ARM/
long.ll 3 define i64 @f1() {
6 ret i64 0
9 define i64 @f2() {
12 ret i64 1
15 define i64 @f3() {
19 ret i64 2147483647
22 define i64 @f4() {
26 ret i64 2147483648
29 define i64 @f5() {
34 ret i64 922337203685477580
    [all...]
2007-04-30-CombinerCrash.ll 3 target datalayout = "e-p:32:32:32-i1:8:8-i8:8:8-i16:16:16-i32:32:32-i64:32:32-f32:32:32-f64:32:32-v64:64:64-v128:128:128-a0:0:64"
5 %struct.CHESS_POSITION = type { i64, i64, i64, i64, i64, i64, i64, i64, i64, i64, i64, i64, i64, i64, i64, i64, i64, i64, i32, i32, i8, i8, [64 x i8], i8, i8, i8, i8, i8
    [all...]
formal.ll 3 declare void @bar(i64 %x, i64 %y)
6 call void @bar(i64 2, i64 3)
ret_i64_arg_split.ll 3 define i64 @test_i64_arg_split(i64 %a1, i32 %a2, i64 %a3) {
4 ret i64 %a3
  /external/llvm/test/CodeGen/Mips/
mips64shift.ll 3 define i64 @f0(i64 %a0, i64 %a1) nounwind readnone {
6 %shl = shl i64 %a0, %a1
7 ret i64 %shl
10 define i64 @f1(i64 %a0, i64 %a1) nounwind readnone {
13 %shr = ashr i64 %a0, %a1
14 ret i64 %sh
    [all...]
  /external/llvm/test/Transforms/GVN/
2011-07-07-MatchIntrinsicExtract.ll 4 %0 = type { i64, i1 }
6 define i64 @test1(i64 %a, i64 %b) nounwind ssp {
8 %uadd = tail call %0 @llvm.uadd.with.overflow.i64(i64 %a, i64 %b)
10 %add1 = add i64 %a, %b
11 ret i64 %add1
18 define i64 @test2(i64 %a, i64 %b) nounwind ssp
    [all...]
  /external/llvm/test/CodeGen/Thumb2/
thumb2-sbc.ll 3 define i64 @f1(i64 %a, i64 %b) {
6 %tmp = sub i64 %a, %b
7 ret i64 %tmp
11 define i64 @f2(i64 %a) {
15 %tmp = sub i64 %a, 734439407618
16 ret i64 %tmp
20 define i64 @f3(i64 %a)
    [all...]
thumb2-adc.ll 4 define i64 @f1(i64 %a) {
7 %tmp = add i64 %a, 734439407618
8 ret i64 %tmp
12 define i64 @f2(i64 %a) {
15 %tmp = add i64 %a, 5066626890203138
16 ret i64 %tmp
20 define i64 @f3(i64 %a)
    [all...]
  /external/llvm/test/CodeGen/Generic/
2008-02-04-Ctlz.ll 5 define i32 @main(i64 %arg) nounwind {
7 %tmp37 = tail call i64 @llvm.ctlz.i64( i64 %arg, i1 true ) ; <i64> [#uses=1]
8 %tmp47 = tail call i64 @llvm.cttz.i64( i64 %arg, i1 true ) ; <i64> [#uses=1]
9 %tmp57 = tail call i64 @llvm.ctpop.i64( i64 %arg ) ; <i64> [#uses=1
    [all...]
  /external/llvm/test/CodeGen/PowerPC/
atomic-2.ll 3 define i64 @exchange_and_add(i64* %mem, i64 %val) nounwind {
6 %tmp = atomicrmw add i64* %mem, i64 %val monotonic
8 ret i64 %tmp
11 define i64 @exchange_and_cmp(i64* %mem) nounwind {
14 %tmp = cmpxchg i64* %mem, i64 0, i64 1 monotoni
    [all...]
subc.ll 9 define i64 @sub_ll(i64 %a, i64 %b) {
11 %tmp.2 = sub i64 %a, %b ; <i64> [#uses=1]
12 ret i64 %tmp.2
15 define i64 @sub_l_5(i64 %a) {
17 %tmp.1 = sub i64 5, %a ; <i64> [#uses=1
    [all...]
  /external/llvm/test/Assembler/
vector-shift.ll 30 ; CHECK: ret <2 x i64> <i64 40, i64 192>
31 define <2 x i64> @foo_ce() nounwind {
32 ret <2 x i64> shl (<2 x i64> <i64 5, i64 6>, <2 x i64> <i64 3, i64 5>
    [all...]

Completed in 751 milliseconds

12 3 4 5 6 7 8 91011>>