HomeSort by relevance Sort by last modified time
    Searched full:tail (Results 451 - 475 of 3492) sorted by null

<<11121314151617181920>>

  /prebuilts/misc/common/swig/include/2.0.11/mzscheme/
std_vector.i 58 Scheme_Object *head, *tail;
60 tail = $input;
61 while (!SCHEME_NULLP(tail)) {
62 head = scheme_car(tail);
63 tail = scheme_cdr(tail);
91 Scheme_Object *head, *tail;
92 tail = $input;
93 while (!SCHEME_NULLP(tail)) {
94 head = scheme_car(tail);
    [all...]
  /frameworks/rs/driver/runtime/ll32/
allocation.ll 37 %1 = tail call i8* @rsOffset([1 x i32] %a.coerce, i32 1, i32 %x, i32 %y, i32 %z) #2
43 %1 = tail call i8* @rsOffset([1 x i32] %a.coerce, i32 1, i32 %x, i32 %y, i32 %z) #2
50 %1 = tail call i8* @rsOffset([1 x i32] %a.coerce, i32 2, i32 %x, i32 %y, i32 %z) #2
57 %1 = tail call i8* @rsOffset([1 x i32] %a.coerce, i32 2, i32 %x, i32 %y, i32 %z) #2
65 %1 = tail call i8* @rsOffset([1 x i32] %a.coerce, i32 4, i32 %x, i32 %y, i32 %z) #2
73 %1 = tail call i8* @rsOffset([1 x i32] %a.coerce, i32 4, i32 %x, i32 %y, i32 %z) #2
82 %1 = tail call i8* @rsOffset([1 x i32] %a.coerce, i32 4, i32 %x, i32 %y, i32 %z) #2
89 %1 = tail call i8* @rsOffset([1 x i32] %a.coerce, i32 4, i32 %x, i32 %y, i32 %z) #2
97 %1 = tail call i8* @rsOffset([1 x i32] %a.coerce, i32 1, i32 %x, i32 %y, i32 %z) #2
103 %1 = tail call i8* @rsOffset([1 x i32] %a.coerce, i32 1, i32 %x, i32 %y, i32 %z) #
    [all...]
  /frameworks/rs/driver/runtime/ll64/
allocation.ll 39 %1 = tail call i8* @rsOffset(%struct.rs_allocation* %a, i32 1, i32 %x, i32 %y, i32 %z) #2
45 %1 = tail call i8* @rsOffset(%struct.rs_allocation* %a, i32 1, i32 %x, i32 %y, i32 %z) #2
52 %1 = tail call i8* @rsOffset(%struct.rs_allocation* %a, i32 2, i32 %x, i32 %y, i32 %z) #2
60 %1 = tail call i8* @rsOffset(%struct.rs_allocation* %a, i32 2, i32 %x, i32 %y, i32 %z) #2
68 %1 = tail call i8* @rsOffset(%struct.rs_allocation* %a, i32 4, i32 %x, i32 %y, i32 %z) #2
77 %1 = tail call i8* @rsOffset(%struct.rs_allocation* %a, i32 4, i32 %x, i32 %y, i32 %z) #2
86 %1 = tail call i8* @rsOffset(%struct.rs_allocation* %a, i32 4, i32 %x, i32 %y, i32 %z) #2
93 %1 = tail call i8* @rsOffset(%struct.rs_allocation* %a, i32 4, i32 %x, i32 %y, i32 %z) #2
101 %1 = tail call i8* @rsOffset(%struct.rs_allocation* %a, i32 1, i32 %x, i32 %y, i32 %z) #2
107 %1 = tail call i8* @rsOffset(%struct.rs_allocation* %a, i32 1, i32 %x, i32 %y, i32 %z) #
    [all...]
  /external/llvm/test/Analysis/TypeBasedAliasAnalysis/
memcpyopt.ll 9 ; CHECK-NEXT: tail call void @llvm.memcpy.p0i8.p0i8.i64(i8* %p, i8* %q, i64 16, i32 1, i1 false), !tbaa !0
13 tail call void @llvm.memcpy.p0i8.p0i8.i64(i8* %p, i8* %q, i64 16, i32 1, i1 false), !tbaa !2
15 tail call void @llvm.memcpy.p0i8.p0i8.i64(i8* %q, i8* %p, i64 16, i32 1, i1 false), !tbaa !2
  /external/llvm/test/CodeGen/AArch64/
arm64-smaxv.ll 9 %vmaxv.i = tail call i32 @llvm.aarch64.neon.smaxv.i32.v8i8(<8 x i8> %a1)
20 %vmaxv.i = tail call i32 @llvm.aarch64.neon.smaxv.i32.v4i16(<4 x i16> %a1)
32 %vmaxv.i = tail call i32 @llvm.aarch64.neon.smaxv.i32.v2i32(<2 x i32> %a1)
42 %vmaxv.i = tail call i32 @llvm.aarch64.neon.smaxv.i32.v16i8(<16 x i8> %a1)
53 %vmaxv.i = tail call i32 @llvm.aarch64.neon.smaxv.i32.v8i16(<8 x i16> %a1)
64 %vmaxv.i = tail call i32 @llvm.aarch64.neon.smaxv.i32.v4i32(<4 x i32> %a1)
74 %0 = tail call i32 @llvm.aarch64.neon.smaxv.i32.v8i8(<8 x i8> %a2)
86 %0 = tail call i32 @llvm.aarch64.neon.smaxv.i32.v4i16(<4 x i16> %a2)
98 %0 = tail call i32 @llvm.aarch64.neon.smaxv.i32.v2i32(<2 x i32> %a2)
109 %0 = tail call i32 @llvm.aarch64.neon.smaxv.i32.v16i8(<16 x i8> %a2
    [all...]
arm64-sminv.ll 9 %vminv.i = tail call i32 @llvm.aarch64.neon.sminv.i32.v8i8(<8 x i8> %a1)
20 %vminv.i = tail call i32 @llvm.aarch64.neon.sminv.i32.v4i16(<4 x i16> %a1)
32 %vminv.i = tail call i32 @llvm.aarch64.neon.sminv.i32.v2i32(<2 x i32> %a1)
42 %vminv.i = tail call i32 @llvm.aarch64.neon.sminv.i32.v16i8(<16 x i8> %a1)
53 %vminv.i = tail call i32 @llvm.aarch64.neon.sminv.i32.v8i16(<8 x i16> %a1)
64 %vminv.i = tail call i32 @llvm.aarch64.neon.sminv.i32.v4i32(<4 x i32> %a1)
74 %0 = tail call i32 @llvm.aarch64.neon.sminv.i32.v8i8(<8 x i8> %a2)
86 %0 = tail call i32 @llvm.aarch64.neon.sminv.i32.v4i16(<4 x i16> %a2)
98 %0 = tail call i32 @llvm.aarch64.neon.sminv.i32.v2i32(<2 x i32> %a2)
109 %0 = tail call i32 @llvm.aarch64.neon.sminv.i32.v16i8(<16 x i8> %a2
    [all...]
tailcall-implicit-sret.ll 6 ; Check that we don't try to tail-call with an sret-demoted return.
30 %a = tail call i1024 @test_sret()
42 %a = tail call i1024 %f()
  /external/llvm/test/CodeGen/ARM/
arm-ttype-target2.ll 16 %2 = tail call i32 @llvm.eh.typeid.for(i8* bitcast ({ i8*, i8* }* @_ZTI3Foo to i8*)) nounwind
24 %4 = tail call i8* @__cxa_begin_catch(i8* %3) nounwind
25 tail call void @__cxa_end_catch()
thumb_indirect_calls.ll 10 tail call void %0(i32 %x)
27 tail call void %g(i32 %x)
28 tail call void %h(i32 %x)
  /external/llvm/test/CodeGen/Hexagon/
combine_ir.ll 9 tail call void @bar(i64 %1) nounwind
27 tail call void @bar(i64 %ins) nounwind
43 tail call void @bar(i64 %ins) nounwind
  /external/llvm/test/CodeGen/MSP430/
2009-12-22-InlineAsm.ll 11 %0 = tail call i8* asm "", "=r,0"(i8* getelementptr inbounds ([10 x i8], [10 x i8]* @buf, i16 0, i16 0)) nounwind ; <i8*> [#uses=1]
19 tail call void @abort() nounwind
23 tail call void @exit(i16 0) nounwind
  /external/llvm/test/CodeGen/Mips/
i64arg.ll 13 tail call void @ff1(i32 %i, i64 1085102592623924856) nounwind
18 tail call void @ff2(i64 %ll, double 3.000000e+00) nounwind
26 tail call void @ff3(i32 %i, i64 %ll, i32 %sub, i64 %ll1) nounwind
  /external/llvm/test/CodeGen/PowerPC/
inlineasm-copy.ll 6 %tmp = tail call i32 asm "foo $0", "=r"( ) ; <i32> [#uses=1]
12 %tmp1 = tail call i32 asm "foo $0, $1", "=r,r"( i32 %X ) ; <i32> [#uses=1]
19 %tmp1 = tail call { i32, i32, i32, i32, i32, i32, i32, i32, i32, i32, i32, i32, i32, i32, i32, i32, i32, i32, i32, i32 } asm sideeffect "foo $0, $1", "=r,=r,=r,=r,=r,=r,=r,=r,=r,=r,=r,=r,=r,=r,=r,=r,=r,=r,=r,=r,0,1,2,3,4,5,6,7,8,9,10,11,12,13,14,15,16,17,18,19"( i32 %X, i32 %Y, i32 %X, i32 %Y, i32 %X, i32 %Y, i32 %X, i32 %Y, i32 %X, i32 %Y, i32 %X, i32 %Y, i32 %X, i32 %Y, i32 %X, i32 %Y, i32 %X, i32 %Y, i32 %X, i32 %Y ) ; <i32> [#uses=1]
  /external/llvm/test/CodeGen/Thumb/
2011-06-16-NoGPRs.ll 19 %call14 = tail call i8* (i8*, i8*, ...)* (i8*, i8*) @f1(i8* undef, i8* %_cmd) optsize
21 tail call void %0(i8* %self, i8* %_cmd, %0* %inObjects, %0* %inIndexes) optsize
22 tail call void bitcast (i8* (i8*, i8*, ...)* @f2 to void (i8*, i8*, i32, %0*, %0*)*)(i8* %self, i8* undef, i32 2, %0* %inIndexes, %0* undef) optsize
inlineasm-imm-thumb.ll 11 tail call void asm sideeffect ".word $0", "J"( i32 -255 ) nounwind
17 tail call void asm sideeffect ".word $0", "K"( i32 65280 ) nounwind
41 tail call void asm sideeffect "add sp, sp, $0; add sp, sp, $1", "O,O"( i32 -508, i32 508 ) nounwind
  /external/llvm/test/CodeGen/X86/
ctpop-combine.ll 6 %count = tail call i64 @llvm.ctpop.i64(i64 %x)
20 %count = tail call i64 @llvm.ctpop.i64(i64 %x)
32 %count = tail call i64 @llvm.ctpop.i64(i64 %x)
vec_shift3.ll 7 %tmp3 = tail call <2 x i64> @llvm.x86.sse2.pslli.q( <2 x i64> %x1, i32 %bits ) nounwind readnone ; <<2 x i64>> [#uses=1]
13 %tmp3 = tail call <2 x i64> @llvm.x86.sse2.pslli.q( <2 x i64> %x1, i32 10 ) nounwind readnone ; <<2 x i64>> [#uses=1]
20 %tmp4 = tail call <8 x i16> @llvm.x86.sse2.psrai.w( <8 x i16> %tmp2, i32 %bits ) nounwind readnone ; <<8 x i16>> [#uses=1]
x86-upgrade-avx-vbroadcast.ll 15 %1 = tail call <4 x float> @llvm.x86.avx.vbroadcast.ss(i8* %0)
24 %1 = tail call <4 x double> @llvm.x86.avx.vbroadcast.sd.256(i8* %0)
33 %1 = tail call <8 x float> @llvm.x86.avx.vbroadcast.ss.256(i8* %0)
  /external/llvm/test/ExecutionEngine/OrcLazy/
hello.ll 17 %puts.i = tail call i32 @puts(i8* getelementptr inbounds ([8 x i8], [8 x i8]* @str2, i64 0, i64 0))
30 %puts.i.i.i = tail call i32 @puts(i8* getelementptr inbounds ([6 x i8], [6 x i8]* @str, i64 0, i64 0))
31 %0 = tail call i32 @__cxa_atexit(void (i8*)* bitcast (void (%class.Foo*)* @_ZN3FooD1Ev to void (i8*)*), i8* getelementptr inbounds (%class.Foo, %class.Foo* @f, i64 0, i32 0), i8* @__dso_handle)
  /external/llvm/test/Transforms/GVN/
calls-readonly.ll 9 %0 = tail call i32 @strlen(i8* %P) ; <i32> [#uses=2]
19 %3 = tail call i8* @strchr(i8* %Q, i32 97) ; <i8*> [#uses=1]
20 %4 = tail call i32 @strlen(i8* %P) ; <i32> [#uses=1]
  /external/llvm/test/Transforms/ObjCARC/
contract-storestrong-ivar.ll 3 ; CHECK: tail call void @objc_storeStrong(i8**
25 %tmp4 = tail call i8* @objc_retain(i8* %tmp3) nounwind
27 tail call void @objc_release(i8* %tmp5) nounwind
  /external/mdnsresponder/mDNSShared/
GenLinkedList.h 28 *Tail;
48 *Tail;
70 Tail;
  /frameworks/av/services/audioflinger/
FastMixerDumpState.cpp 94 // the mean account for 99.73% of the population. So if we take each tail to be 1/1000 of the
97 uint32_t *tail = n >= kTailDenominator ? new uint32_t[n] : NULL; local
102 if (tail != NULL) {
103 tail[j] = wallNs;
144 if (tail != NULL) {
145 qsort(tail, n, sizeof(uint32_t), compare_uint32_t);
146 // assume same number of tail samples on each side, left and right
150 left.sample(tail[i]);
151 right.sample(tail[n - (i + 1)]);
155 " left tail: mean=%.2f min=%.2f max=%.2f stddev=%.2f\n
    [all...]
  /external/clang/test/CodeGen/
disable-tail-calls.c 1 // RUN: %clang_cc1 -triple x86_64-apple-macosx10.9.0 -emit-llvm -O1 -mdisable-tail-calls -o - < %s | FileCheck %s
  /external/clang/test/CodeGenCXX/
merge-functions.cpp 13 // CHECK-NEXT: tail call i32 @_ZN1A1fEiPi

Completed in 552 milliseconds

<<11121314151617181920>>