/external/llvm/test/CodeGen/PowerPC/ |
shift128.ll | 3 define i128 @foo_lshr(i128 %x, i128 %y) { 4 %r = lshr i128 %x, %y 5 ret i128 %r 7 define i128 @foo_ashr(i128 %x, i128 %y) { 8 %r = ashr i128 %x, %y 9 ret i128 % [all...] |
return-val-i128.ll | 3 define i128 @__fixsfdi(float %a) { 6 %retval = alloca i128, align 16 ; <i128*> [#uses=2] 7 %tmp = alloca i128, align 16 ; <i128*> [#uses=3] 18 %tmp6 = call i128 @__fixunssfDI( float %tmp5 ) nounwind ; <i128> [#uses=1] 19 %tmp7 = sub i128 0, %tmp6 ; <i128> [#uses=1] 20 store i128 %tmp7, i128* %tmp, align 1 [all...] |
/external/llvm/test/CodeGen/ARM/ |
ret_i128_arg2.ll | 3 define i128 @test_i128(i128 %a1, i128 %a2, i128 %a3) { 4 ret i128 %a3
|
arguments6.ll | 4 define i128 @f(i32 %a, i128 %b) { 5 %tmp = call i128 @g(i128 %b) 6 ret i128 %tmp 9 declare i128 @g(i128)
|
/external/llvm/test/CodeGen/X86/ |
mul128.ll | 3 define i128 @foo(i128 %t, i128 %u) { 4 %k = mul i128 %t, %u 5 ret i128 %k
|
nocx16.ll | 2 define void @test(i128* %a) nounwind { 5 %0 = cmpxchg i128* %a, i128 1, i128 1 seq_cst seq_cst 7 %1 = atomicrmw xchg i128* %a, i128 1 seq_cst 9 %2 = atomicrmw add i128* %a, i128 1 seq_cst 11 %3 = atomicrmw sub i128* %a, i128 1 seq_cs [all...] |
2008-07-11-SHLBy1.ll | 2 define i128 @sl(i128 %x) { 3 %t = shl i128 %x, 1 4 ret i128 %t
|
shift-i128.ll | 4 define void @t(i128 %x, i128 %a, i128* nocapture %r) nounwind { 6 %0 = lshr i128 %x, %a 7 store i128 %0, i128* %r, align 16
|
extmul128.ll | 3 define i128 @i64_sext_i128(i64 %a, i64 %b) { 4 %aa = sext i64 %a to i128 5 %bb = sext i64 %b to i128 6 %cc = mul i128 %aa, %bb 7 ret i128 %cc 9 define i128 @i64_zext_i128(i64 %a, i64 %b) { 10 %aa = zext i64 %a to i128 11 %bb = zext i64 %b to i128 12 %cc = mul i128 %aa, %bb 13 ret i128 %c [all...] |
i128-sdiv.ll | 5 define i128 @test1(i128 %x) { 8 %tmp = sdiv i128 %x, 73786976294838206464 9 ret i128 %tmp 12 define i128 @test2(i128 %x) { 15 %tmp = sdiv i128 %x, -73786976294838206464 16 ret i128 %tmp 19 define i128 @test3(i128 %x) [all...] |
i128-immediate.ll | 3 define i128 @__addvti3() { 4 ret i128 -1
|
cmpxchg-i128-i1.ll | 3 define i1 @try_cmpxchg(i128* %addr, i128 %desired, i128 %new) { 9 %pair = cmpxchg i128* %addr, i128 %desired, i128 %new seq_cst seq_cst 10 %success = extractvalue { i128, i1 } %pair, 1 14 define void @cmpxchg_flow(i128* %addr, i128 %desired, i128 %new) [all...] |
mulx64.ll | 3 define i128 @f1(i64 %a, i64 %b) { 4 %x = zext i64 %a to i128 5 %y = zext i64 %b to i128 6 %r = mul i128 %x, %y 10 ret i128 %r 13 define i128 @f2(i64 %a, i64* %p) { 15 %x = zext i64 %a to i128 16 %y = zext i64 %b to i128 17 %r = mul i128 %x, %y 21 ret i128 % [all...] |
i128-ret.ll | 6 define i128 @test(i128 *%P) { 7 %A = load i128, i128* %P 8 ret i128 %A
|
2009-03-09-APIntCrash.ll | 10 %srcval18 = load i128, i128* null, align 8 ; <i128> [#uses=1] 11 %tmp15 = lshr i128 %srcval18, 64 ; <i128> [#uses=1] 12 %tmp9 = mul i128 %tmp15, 18446744073709551616000 ; <i128> [#uses=1] 16 %retval.0 = phi i128 [ %tmp9, %if.then ], [ undef, %entry ] ; <i128> [#uses=0] 20 define i128 @test(i128 %arg) nounwind [all...] |
/external/llvm/test/CodeGen/Generic/ |
i128-addsub.ll | 5 %tmp1 = zext i64 %AL to i128 ; <i128> [#uses=1] 6 %tmp23 = zext i64 %AH to i128 ; <i128> [#uses=1] 7 %tmp4 = shl i128 %tmp23, 64 ; <i128> [#uses=1] 8 %tmp5 = or i128 %tmp4, %tmp1 ; <i128> [#uses=1] 9 %tmp67 = zext i64 %BL to i128 ; <i128> [#uses=1 [all...] |
i128-arith.ll | 4 %tmp0 = zext i64 %x to i128 5 %tmp1 = sext i64 %y to i128 6 %tmp2 = or i128 %tmp0, %tmp1 7 %tmp7 = zext i32 13 to i128 8 %tmp3 = lshr i128 %tmp2, %tmp7 9 %tmp4 = trunc i128 %tmp3 to i64
|
/external/llvm/test/CodeGen/SystemZ/ |
int-add-09.ll | 7 define void @f1(i128 *%aptr) { 12 %a = load i128 , i128 *%aptr 13 %xor = xor i128 %a, 128 14 %add = add i128 %xor, 1 15 store i128 %add, i128 *%aptr 20 define void @f2(i128 *%aptr) { 25 %a = load i128 , i128 *%apt [all...] |
int-sub-05.ll | 6 declare i128 *@foo() 9 define void @f1(i128 *%ptr, i64 %high, i64 %low) { 14 %a = load i128 , i128 *%ptr 15 %highx = zext i64 %high to i128 16 %lowx = zext i64 %low to i128 17 %bhigh = shl i128 %highx, 64 18 %b = or i128 %bhigh, %lowx 19 %sub = sub i128 %a, %b 20 store i128 %sub, i128 *%pt [all...] |
int-add-08.ll | 6 declare i128 *@foo() 9 define void @f1(i128 *%ptr) { 14 %value = load i128 , i128 *%ptr 15 %add = add i128 %value, %value 16 store i128 %add, i128 *%ptr 22 define void @f2(i128 *%aptr, i64 %addr) { 27 %bptr = inttoptr i64 %addr to i128 * 28 %a = load volatile i128 , i128 *%apt [all...] |
int-add-10.ll | 7 define void @f1(i128 *%aptr, i32 %b) { 12 %a = load i128 , i128 *%aptr 13 %xor = xor i128 %a, 127 14 %bext = zext i32 %b to i128 15 %add = add i128 %xor, %bext 16 store i128 %add, i128 *%aptr 21 define void @f2(i128 *%aptr, i64 %b) { 26 %a = load i128 , i128 *%apt [all...] |
int-sub-06.ll | 7 define void @f1(i128 *%aptr, i32 %b) { 12 %a = load i128 , i128 *%aptr 13 %xor = xor i128 %a, 127 14 %bext = zext i32 %b to i128 15 %sub = sub i128 %xor, %bext 16 store i128 %sub, i128 *%aptr 21 define void @f2(i128 *%aptr, i64 %b) { 26 %a = load i128 , i128 *%apt [all...] |
/external/llvm/test/Transforms/LoopStrengthReduce/ |
pr2537.ll | 9 %y.0 = phi i128 [ 0, %entry ], [ %add, %dobody ] 10 %x.0 = phi i128 [ 0, %entry ], [ %add2, %dobody ] 11 %add = add i128 %y.0, shl (i128 1, i128 64) 12 %add2 = add i128 %x.0, shl (i128 1, i128 48) 13 call void @b( i128 %add ) 14 %cmp = icmp ult i128 %add2, shl (i128 1, i128 64 [all...] |
/external/llvm/test/CodeGen/NVPTX/ |
shift-parts.ll | 4 define void @shift_parts_left_128(i128* %val, i128* %amtptr) { 15 %amt = load i128, i128* %amtptr 16 %a = load i128, i128* %val 17 %val0 = shl i128 %a, %amt 18 store i128 %val0, i128* %val 23 define void @shift_parts_right_128(i128* %val, i128* %amtptr) [all...] |
/external/llvm/test/CodeGen/SPARC/ |
parts.ll | 10 define i128 @test(i128 %a, i128 %b) { 12 %tmp = shl i128 %b, %a 13 ret i128 %tmp
|