/prebuilts/misc/common/swig/include/2.0.11/mzscheme/ |
std_vector.i | 58 Scheme_Object *head, *tail; 60 tail = $input; 61 while (!SCHEME_NULLP(tail)) { 62 head = scheme_car(tail); 63 tail = scheme_cdr(tail); 91 Scheme_Object *head, *tail; 92 tail = $input; 93 while (!SCHEME_NULLP(tail)) { 94 head = scheme_car(tail); [all...] |
/frameworks/rs/driver/runtime/ll32/ |
allocation.ll | 37 %1 = tail call i8* @rsOffset([1 x i32] %a.coerce, i32 1, i32 %x, i32 %y, i32 %z) #2 43 %1 = tail call i8* @rsOffset([1 x i32] %a.coerce, i32 1, i32 %x, i32 %y, i32 %z) #2 50 %1 = tail call i8* @rsOffset([1 x i32] %a.coerce, i32 2, i32 %x, i32 %y, i32 %z) #2 57 %1 = tail call i8* @rsOffset([1 x i32] %a.coerce, i32 2, i32 %x, i32 %y, i32 %z) #2 65 %1 = tail call i8* @rsOffset([1 x i32] %a.coerce, i32 4, i32 %x, i32 %y, i32 %z) #2 73 %1 = tail call i8* @rsOffset([1 x i32] %a.coerce, i32 4, i32 %x, i32 %y, i32 %z) #2 82 %1 = tail call i8* @rsOffset([1 x i32] %a.coerce, i32 4, i32 %x, i32 %y, i32 %z) #2 89 %1 = tail call i8* @rsOffset([1 x i32] %a.coerce, i32 4, i32 %x, i32 %y, i32 %z) #2 97 %1 = tail call i8* @rsOffset([1 x i32] %a.coerce, i32 1, i32 %x, i32 %y, i32 %z) #2 103 %1 = tail call i8* @rsOffset([1 x i32] %a.coerce, i32 1, i32 %x, i32 %y, i32 %z) # [all...] |
/frameworks/rs/driver/runtime/ll64/ |
allocation.ll | 39 %1 = tail call i8* @rsOffset(%struct.rs_allocation* %a, i32 1, i32 %x, i32 %y, i32 %z) #2 45 %1 = tail call i8* @rsOffset(%struct.rs_allocation* %a, i32 1, i32 %x, i32 %y, i32 %z) #2 52 %1 = tail call i8* @rsOffset(%struct.rs_allocation* %a, i32 2, i32 %x, i32 %y, i32 %z) #2 60 %1 = tail call i8* @rsOffset(%struct.rs_allocation* %a, i32 2, i32 %x, i32 %y, i32 %z) #2 68 %1 = tail call i8* @rsOffset(%struct.rs_allocation* %a, i32 4, i32 %x, i32 %y, i32 %z) #2 77 %1 = tail call i8* @rsOffset(%struct.rs_allocation* %a, i32 4, i32 %x, i32 %y, i32 %z) #2 86 %1 = tail call i8* @rsOffset(%struct.rs_allocation* %a, i32 4, i32 %x, i32 %y, i32 %z) #2 93 %1 = tail call i8* @rsOffset(%struct.rs_allocation* %a, i32 4, i32 %x, i32 %y, i32 %z) #2 101 %1 = tail call i8* @rsOffset(%struct.rs_allocation* %a, i32 1, i32 %x, i32 %y, i32 %z) #2 107 %1 = tail call i8* @rsOffset(%struct.rs_allocation* %a, i32 1, i32 %x, i32 %y, i32 %z) # [all...] |
/external/llvm/test/Analysis/TypeBasedAliasAnalysis/ |
memcpyopt.ll | 9 ; CHECK-NEXT: tail call void @llvm.memcpy.p0i8.p0i8.i64(i8* %p, i8* %q, i64 16, i32 1, i1 false), !tbaa !0 13 tail call void @llvm.memcpy.p0i8.p0i8.i64(i8* %p, i8* %q, i64 16, i32 1, i1 false), !tbaa !2 15 tail call void @llvm.memcpy.p0i8.p0i8.i64(i8* %q, i8* %p, i64 16, i32 1, i1 false), !tbaa !2
|
/external/llvm/test/CodeGen/AArch64/ |
arm64-smaxv.ll | 9 %vmaxv.i = tail call i32 @llvm.aarch64.neon.smaxv.i32.v8i8(<8 x i8> %a1) 20 %vmaxv.i = tail call i32 @llvm.aarch64.neon.smaxv.i32.v4i16(<4 x i16> %a1) 32 %vmaxv.i = tail call i32 @llvm.aarch64.neon.smaxv.i32.v2i32(<2 x i32> %a1) 42 %vmaxv.i = tail call i32 @llvm.aarch64.neon.smaxv.i32.v16i8(<16 x i8> %a1) 53 %vmaxv.i = tail call i32 @llvm.aarch64.neon.smaxv.i32.v8i16(<8 x i16> %a1) 64 %vmaxv.i = tail call i32 @llvm.aarch64.neon.smaxv.i32.v4i32(<4 x i32> %a1) 74 %0 = tail call i32 @llvm.aarch64.neon.smaxv.i32.v8i8(<8 x i8> %a2) 86 %0 = tail call i32 @llvm.aarch64.neon.smaxv.i32.v4i16(<4 x i16> %a2) 98 %0 = tail call i32 @llvm.aarch64.neon.smaxv.i32.v2i32(<2 x i32> %a2) 109 %0 = tail call i32 @llvm.aarch64.neon.smaxv.i32.v16i8(<16 x i8> %a2 [all...] |
arm64-sminv.ll | 9 %vminv.i = tail call i32 @llvm.aarch64.neon.sminv.i32.v8i8(<8 x i8> %a1) 20 %vminv.i = tail call i32 @llvm.aarch64.neon.sminv.i32.v4i16(<4 x i16> %a1) 32 %vminv.i = tail call i32 @llvm.aarch64.neon.sminv.i32.v2i32(<2 x i32> %a1) 42 %vminv.i = tail call i32 @llvm.aarch64.neon.sminv.i32.v16i8(<16 x i8> %a1) 53 %vminv.i = tail call i32 @llvm.aarch64.neon.sminv.i32.v8i16(<8 x i16> %a1) 64 %vminv.i = tail call i32 @llvm.aarch64.neon.sminv.i32.v4i32(<4 x i32> %a1) 74 %0 = tail call i32 @llvm.aarch64.neon.sminv.i32.v8i8(<8 x i8> %a2) 86 %0 = tail call i32 @llvm.aarch64.neon.sminv.i32.v4i16(<4 x i16> %a2) 98 %0 = tail call i32 @llvm.aarch64.neon.sminv.i32.v2i32(<2 x i32> %a2) 109 %0 = tail call i32 @llvm.aarch64.neon.sminv.i32.v16i8(<16 x i8> %a2 [all...] |
tailcall-implicit-sret.ll | 6 ; Check that we don't try to tail-call with an sret-demoted return. 30 %a = tail call i1024 @test_sret() 42 %a = tail call i1024 %f()
|
/external/llvm/test/CodeGen/ARM/ |
arm-ttype-target2.ll | 16 %2 = tail call i32 @llvm.eh.typeid.for(i8* bitcast ({ i8*, i8* }* @_ZTI3Foo to i8*)) nounwind 24 %4 = tail call i8* @__cxa_begin_catch(i8* %3) nounwind 25 tail call void @__cxa_end_catch()
|
thumb_indirect_calls.ll | 10 tail call void %0(i32 %x) 27 tail call void %g(i32 %x) 28 tail call void %h(i32 %x)
|
/external/llvm/test/CodeGen/Hexagon/ |
combine_ir.ll | 9 tail call void @bar(i64 %1) nounwind 27 tail call void @bar(i64 %ins) nounwind 43 tail call void @bar(i64 %ins) nounwind
|
/external/llvm/test/CodeGen/MSP430/ |
2009-12-22-InlineAsm.ll | 11 %0 = tail call i8* asm "", "=r,0"(i8* getelementptr inbounds ([10 x i8], [10 x i8]* @buf, i16 0, i16 0)) nounwind ; <i8*> [#uses=1] 19 tail call void @abort() nounwind 23 tail call void @exit(i16 0) nounwind
|
/external/llvm/test/CodeGen/Mips/ |
i64arg.ll | 13 tail call void @ff1(i32 %i, i64 1085102592623924856) nounwind 18 tail call void @ff2(i64 %ll, double 3.000000e+00) nounwind 26 tail call void @ff3(i32 %i, i64 %ll, i32 %sub, i64 %ll1) nounwind
|
/external/llvm/test/CodeGen/PowerPC/ |
inlineasm-copy.ll | 6 %tmp = tail call i32 asm "foo $0", "=r"( ) ; <i32> [#uses=1] 12 %tmp1 = tail call i32 asm "foo $0, $1", "=r,r"( i32 %X ) ; <i32> [#uses=1] 19 %tmp1 = tail call { i32, i32, i32, i32, i32, i32, i32, i32, i32, i32, i32, i32, i32, i32, i32, i32, i32, i32, i32, i32 } asm sideeffect "foo $0, $1", "=r,=r,=r,=r,=r,=r,=r,=r,=r,=r,=r,=r,=r,=r,=r,=r,=r,=r,=r,=r,0,1,2,3,4,5,6,7,8,9,10,11,12,13,14,15,16,17,18,19"( i32 %X, i32 %Y, i32 %X, i32 %Y, i32 %X, i32 %Y, i32 %X, i32 %Y, i32 %X, i32 %Y, i32 %X, i32 %Y, i32 %X, i32 %Y, i32 %X, i32 %Y, i32 %X, i32 %Y, i32 %X, i32 %Y ) ; <i32> [#uses=1]
|
/external/llvm/test/CodeGen/Thumb/ |
2011-06-16-NoGPRs.ll | 19 %call14 = tail call i8* (i8*, i8*, ...)* (i8*, i8*) @f1(i8* undef, i8* %_cmd) optsize 21 tail call void %0(i8* %self, i8* %_cmd, %0* %inObjects, %0* %inIndexes) optsize 22 tail call void bitcast (i8* (i8*, i8*, ...)* @f2 to void (i8*, i8*, i32, %0*, %0*)*)(i8* %self, i8* undef, i32 2, %0* %inIndexes, %0* undef) optsize
|
inlineasm-imm-thumb.ll | 11 tail call void asm sideeffect ".word $0", "J"( i32 -255 ) nounwind 17 tail call void asm sideeffect ".word $0", "K"( i32 65280 ) nounwind 41 tail call void asm sideeffect "add sp, sp, $0; add sp, sp, $1", "O,O"( i32 -508, i32 508 ) nounwind
|
/external/llvm/test/CodeGen/X86/ |
ctpop-combine.ll | 6 %count = tail call i64 @llvm.ctpop.i64(i64 %x) 20 %count = tail call i64 @llvm.ctpop.i64(i64 %x) 32 %count = tail call i64 @llvm.ctpop.i64(i64 %x)
|
vec_shift3.ll | 7 %tmp3 = tail call <2 x i64> @llvm.x86.sse2.pslli.q( <2 x i64> %x1, i32 %bits ) nounwind readnone ; <<2 x i64>> [#uses=1] 13 %tmp3 = tail call <2 x i64> @llvm.x86.sse2.pslli.q( <2 x i64> %x1, i32 10 ) nounwind readnone ; <<2 x i64>> [#uses=1] 20 %tmp4 = tail call <8 x i16> @llvm.x86.sse2.psrai.w( <8 x i16> %tmp2, i32 %bits ) nounwind readnone ; <<8 x i16>> [#uses=1]
|
x86-upgrade-avx-vbroadcast.ll | 15 %1 = tail call <4 x float> @llvm.x86.avx.vbroadcast.ss(i8* %0) 24 %1 = tail call <4 x double> @llvm.x86.avx.vbroadcast.sd.256(i8* %0) 33 %1 = tail call <8 x float> @llvm.x86.avx.vbroadcast.ss.256(i8* %0)
|
/external/llvm/test/ExecutionEngine/OrcLazy/ |
hello.ll | 17 %puts.i = tail call i32 @puts(i8* getelementptr inbounds ([8 x i8], [8 x i8]* @str2, i64 0, i64 0)) 30 %puts.i.i.i = tail call i32 @puts(i8* getelementptr inbounds ([6 x i8], [6 x i8]* @str, i64 0, i64 0)) 31 %0 = tail call i32 @__cxa_atexit(void (i8*)* bitcast (void (%class.Foo*)* @_ZN3FooD1Ev to void (i8*)*), i8* getelementptr inbounds (%class.Foo, %class.Foo* @f, i64 0, i32 0), i8* @__dso_handle)
|
/external/llvm/test/Transforms/GVN/ |
calls-readonly.ll | 9 %0 = tail call i32 @strlen(i8* %P) ; <i32> [#uses=2] 19 %3 = tail call i8* @strchr(i8* %Q, i32 97) ; <i8*> [#uses=1] 20 %4 = tail call i32 @strlen(i8* %P) ; <i32> [#uses=1]
|
/external/llvm/test/Transforms/ObjCARC/ |
contract-storestrong-ivar.ll | 3 ; CHECK: tail call void @objc_storeStrong(i8** 25 %tmp4 = tail call i8* @objc_retain(i8* %tmp3) nounwind 27 tail call void @objc_release(i8* %tmp5) nounwind
|
/external/mdnsresponder/mDNSShared/ |
GenLinkedList.h | 28 *Tail; 48 *Tail; 70 Tail;
|
/frameworks/av/services/audioflinger/ |
FastMixerDumpState.cpp | 94 // the mean account for 99.73% of the population. So if we take each tail to be 1/1000 of the 97 uint32_t *tail = n >= kTailDenominator ? new uint32_t[n] : NULL; local 102 if (tail != NULL) { 103 tail[j] = wallNs; 144 if (tail != NULL) { 145 qsort(tail, n, sizeof(uint32_t), compare_uint32_t); 146 // assume same number of tail samples on each side, left and right 150 left.sample(tail[i]); 151 right.sample(tail[n - (i + 1)]); 155 " left tail: mean=%.2f min=%.2f max=%.2f stddev=%.2f\n [all...] |
/external/clang/test/CodeGen/ |
disable-tail-calls.c | 1 // RUN: %clang_cc1 -triple x86_64-apple-macosx10.9.0 -emit-llvm -O1 -mdisable-tail-calls -o - < %s | FileCheck %s
|
/external/clang/test/CodeGenCXX/ |
merge-functions.cpp | 13 // CHECK-NEXT: tail call i32 @_ZN1A1fEiPi
|