HomeSort by relevance Sort by last modified time
    Searched full:xmm3 (Results 126 - 150 of 367) sorted by null

1 2 3 4 56 7 8 91011>>

  /hardware/intel/common/omx-components/videocodec/libvpx_internal/libvpx/vp9/common/x86/
vp9_subpixel_8t_sse2.asm 25 pshuflw xmm3, xmm7, 11111111b ;k3
33 punpcklqdq xmm2, xmm3
53 punpckldq xmm2, xmm3
98 pshuflw xmm3, xmm7, 11111111b ;k3
107 punpcklwd xmm3, xmm3
116 movdqa k3, xmm3
137 movq xmm3, [rsi + rax * 2 + %1] ;3
149 punpcklbw xmm3, zero
158 pmullw xmm3, k
    [all...]
  /bionic/libm/x86/
e_pow.S 123 movapd %xmm0, %xmm3
140 psllq $12, %xmm3
143 psrlq $12, %xmm3
155 orpd %xmm1, %xmm3
163 andpd %xmm3, %xmm5
166 subsd %xmm5, %xmm3
172 mulsd %xmm0, %xmm3
182 unpcklpd %xmm3, %xmm5
184 addsd %xmm5, %xmm3
189 pshufd $68, %xmm3, %xmm
    [all...]
e_acos.S 107 movsd 6080(%ebx), %xmm3
125 subsd %xmm1, %xmm3
126 sqrtsd %xmm3, %xmm3
138 mulsd %xmm2, %xmm3
140 addsd %xmm3, %xmm6
144 subsd %xmm3, %xmm1
146 movapd %xmm1, %xmm3
151 mulsd %xmm1, %xmm3
154 mulsd %xmm3, %xmm
    [all...]
e_sinh.S 107 movsd 4272(%ebx), %xmm3
115 andpd %xmm0, %xmm3
124 subsd %xmm3, %xmm4
125 mulsd %xmm1, %xmm3
127 cvtsd2si %xmm3, %eax
130 movapd %xmm3, %xmm7
131 addsd %xmm6, %xmm3
134 subsd %xmm6, %xmm3
138 subsd %xmm3, %xmm7
158 pshufd $68, %xmm2, %xmm3
    [all...]
s_sin.S 214 movsd 2128(%ebx), %xmm3
219 mulsd %xmm1, %xmm3
229 subsd %xmm3, %xmm0
231 subsd %xmm3, %xmm4
234 movapd %xmm4, %xmm3
240 subsd %xmm4, %xmm3
243 subsd %xmm2, %xmm3
245 subsd %xmm3, %xmm1
246 movsd 24(%eax), %xmm3
247 addsd %xmm3, %xmm
    [all...]
  /bionic/libm/x86_64/
e_pow.S 104 movq %xmm0, %xmm3
120 psllq $12, %xmm3
123 psrlq $12, %xmm3
135 orpd %xmm1, %xmm3
140 andpd %xmm3, %xmm5
144 subsd %xmm5, %xmm3
151 mulsd %xmm0, %xmm3
162 unpcklpd %xmm3, %xmm5
164 addsd %xmm5, %xmm3
169 pshufd $68, %xmm3, %xmm
    [all...]
e_acos.S 89 movsd ONEMASK(%rip), %xmm3
106 subsd %xmm1, %xmm3
107 sqrtsd %xmm3, %xmm3
121 mulsd %xmm2, %xmm3
123 addsd %xmm3, %xmm6
127 subsd %xmm3, %xmm1
129 movq %xmm1, %xmm3
134 mulsd %xmm1, %xmm3
137 mulsd %xmm3, %xmm
    [all...]
s_sin.S 204 movd %r8, %xmm3
207 mulsd %xmm1, %xmm3
215 subsd %xmm3, %xmm4
217 subsd %xmm3, %xmm0
218 movddup %xmm4, %xmm3
225 subsd %xmm4, %xmm3
228 subsd %xmm6, %xmm3
230 subsd %xmm3, %xmm1
231 movq 24(%rax), %xmm3
232 addsd %xmm3, %xmm
    [all...]
  /external/libvpx/libvpx/vp8/common/x86/
subpixel_mmx.asm 276 paddw mm3, [GLOBAL(rd)] ; xmm3 += round value
277 psraw mm3, VP8_FILTER_SHIFT ; xmm3 /= 128
317 paddw mm3, [GLOBAL(rd)] ; xmm3 += round value
318 psraw mm3, VP8_FILTER_SHIFT ; xmm3 /= 128
334 paddw mm3, [GLOBAL(rd)] ; xmm3 += round value
335 psraw mm3, VP8_FILTER_SHIFT ; xmm3 /= 128
431 paddw mm3, [GLOBAL(rd)] ; xmm3 += round value
432 psraw mm3, VP8_FILTER_SHIFT ; xmm3 /= 128
472 paddw mm3, [GLOBAL(rd)] ; xmm3 += round value
473 psraw mm3, VP8_FILTER_SHIFT ; xmm3 /= 12
    [all...]
  /hardware/intel/common/omx-components/videocodec/libvpx_internal/libvpx/vp8/common/x86/
subpixel_mmx.asm 276 paddw mm3, [GLOBAL(rd)] ; xmm3 += round value
277 psraw mm3, VP8_FILTER_SHIFT ; xmm3 /= 128
317 paddw mm3, [GLOBAL(rd)] ; xmm3 += round value
318 psraw mm3, VP8_FILTER_SHIFT ; xmm3 /= 128
334 paddw mm3, [GLOBAL(rd)] ; xmm3 += round value
335 psraw mm3, VP8_FILTER_SHIFT ; xmm3 /= 128
431 paddw mm3, [GLOBAL(rd)] ; xmm3 += round value
432 psraw mm3, VP8_FILTER_SHIFT ; xmm3 /= 128
472 paddw mm3, [GLOBAL(rd)] ; xmm3 += round value
473 psraw mm3, VP8_FILTER_SHIFT ; xmm3 /= 12
    [all...]
  /external/llvm/test/CodeGen/X86/
fold-xmm-zero.ll 12 %0 = tail call %0 asm sideeffect "foo", "={xmm0},={xmm1},={xmm2},={xmm3},={xmm4},={xmm5},={xmm6},={xmm7},0,1,2,3,4,5,6,7,~{dirflag},~{fpsr},~{flags}"(float 1.000000e+00, float 2.000000e+00, float 3.000000e+00, float 4.000000e+00, float 5.000000e+00, float 6.000000e+00, float 7.000000e+00, float 8.000000e+00) nounwind
22 %1 = tail call %0 asm sideeffect "bar", "={xmm0},={xmm1},={xmm2},={xmm3},={xmm4},={xmm5},={xmm6},={xmm7},0,1,2,3,4,5,6,7,~{dirflag},~{fpsr},~{flags}"(float %div, float %asmresult8, float %asmresult9, float %asmresult10, float %asmresult11, float %asmresult12, float %asmresult13, float %asmresult14) nounwind
32 %2 = tail call %0 asm sideeffect "baz", "={xmm0},={xmm1},={xmm2},={xmm3},={xmm4},={xmm5},={xmm6},={xmm7},0,1,2,3,4,5,6,7,~{dirflag},~{fpsr},~{flags}"(float %div33, float %asmresult25, float %asmresult26, float %asmresult27, float %asmresult28, float %asmresult29, float %asmresult30, float %asmresult31) nounwind
stack-folding-fp-sse42.ll 14 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"()
22 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"()
30 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"()
38 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"()
47 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"()
55 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"()
64 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"()
73 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"()
82 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"()
96 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xm (…)
    [all...]
stack-folding-fp-avx1.ll 14 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"()
22 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"()
30 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"()
38 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"()
46 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"()
54 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"()
63 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"()
71 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"()
80 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"()
89 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xm (…)
    [all...]
stack-folding-int-avx2.ll 14 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm1},~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"()
23 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm1},~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"()
32 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm1},~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"()
44 %3 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm1},~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"()
51 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"()
61 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"()
70 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm1},~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"()
79 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm1},~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"()
88 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm1},~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"()
97 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xm (…)
    [all...]
fp-select-cmp-and.ll 169 ; CHECK: cmpless %xmm0, %xmm3
170 ; CHECK-NEXT: andps %xmm3, %xmm2
171 ; CHECK-NEXT: andnps %xmm1, %xmm3
172 ; CHECK-NEXT: orps %xmm2, %xmm3
181 ; CHECK: cmplesd %xmm0, %xmm3
182 ; CHECK-NEXT: andpd %xmm3, %xmm2
183 ; CHECK-NEXT: andnpd %xmm1, %xmm3
184 ; CHECK-NEXT: orpd %xmm2, %xmm3
vector-shuffle-256-v16.ll 544 ; AVX1-NEXT: vextractf128 $1, %ymm0, %xmm3
545 ; AVX1-NEXT: vpblendw {{.*#+}} xmm2 = xmm3[0],xmm2[1],xmm3[2],xmm2[3],xmm3[4],xmm2[5],xmm3[6],xmm2[7]
562 ; AVX1-NEXT: vextractf128 $1, %ymm1, %xmm3
563 ; AVX1-NEXT: vpblendw {{.*#+}} xmm2 = xmm3[0],xmm2[1],xmm3[2],xmm2[3],xmm3[4],xmm2[5],xmm3[6],xmm2[7
    [all...]
  /external/llvm/test/TableGen/
cast.td 49 def XMM3: Register<"xmm3">;
64 [XMM0, XMM1, XMM2, XMM3, XMM4, XMM5, XMM6, XMM7,
  /external/mesa3d/src/mesa/x86/
sse_xform3.S 80 MOVAPS ( REGOFF(48, EDX), XMM3 ) /* m12 | m13 | m14 | m15 */
98 ADDPS ( XMM3, XMM4 )
208 MOVSS ( M(10), XMM3 ) /* - | - | - | m10 */
220 MULSS ( XMM3, XMM0 ) /* sz*m10 */
273 MOVSS ( M(10), XMM3 ) /* m10 */
288 MULSS ( XMM3, XMM0 ) /* oz*m10 */
348 MOVSS ( S(0), XMM3 ) /* ox */
349 SHUFPS ( CONST(0x0), XMM3, XMM3 ) /* ox | ox */
350 MULPS ( XMM0, XMM3 ) /* ox*m1 | ox*m0 *
    [all...]
  /external/boringssl/win-x86/crypto/sha/
sha512-586.asm 396 movdqa xmm3,[ebp]
399 paddq xmm3,xmm0
401 movdqa [edx-128],xmm3
403 movdqa xmm3,xmm2
409 movdqa xmm4,xmm3
410 movdqu xmm3,[48+edi]
417 paddq xmm6,xmm3
449 movdqa [48+edx],xmm3
467 movdqa xmm3,xmm5
474 paddq xmm0,xmm3
    [all...]
  /external/libyuv/files/source/
scale.cc 247 movdqa xmm3, [eax + esi + 16] local
250 pavgb xmm1, xmm3
254 movdqa xmm3, xmm1 local
257 pand xmm3, xmm5 local
259 pavgw xmm1, xmm3
322 movdqu xmm3, [eax + esi + 16] local
325 pavgb xmm1, xmm3
329 movdqa xmm3, xmm1 local
332 pand xmm3, xmm5 local
334 pavgw xmm1, xmm3
400 movdqa xmm3, [eax + esi + 16] local
404 movdqa xmm3, [eax + esi * 2 + 16] local
409 pavgb xmm3, xmm5 local
415 movdqa xmm3, xmm1 local
418 pand xmm3, xmm7 local
494 movdqa xmm3, [eax + esi + 16] local
498 movdqa xmm3, [eax + esi * 2 + 16] local
504 pavgb xmm3, xmm5 local
509 movdqa xmm3, [ebp + 16] local
513 pavgb xmm3, xmm5 local
521 pavgb xmm3, xmm5 local
561 movdqa xmm3, kShuf0 local
613 movdqa xmm3, kShuf11 local
672 movdqa xmm3, kShuf11 local
768 movdqa xmm3, kShufAc3 local
834 movdqa xmm3, kShufAb1 local
905 movdqa xmm3, xmm2 local
907 punpckhbw xmm3, xmm4 local
964 movdqa xmm3, xmm2 local
966 punpckhbw xmm3, xmm4 local
970 psubw xmm3, xmm1 local
972 pmulhw xmm3, xmm5 local
    [all...]
compare.cc 68 // 44: 66 0F 38 40 DD pmulld xmm3,xmm5
93 movdqa xmm3, xmm2 local
94 punpcklwd xmm3, xmm7 // src[0-3] local
95 pmulld(0xdd) // pmulld xmm3, xmm5
108 paddd xmm3, xmm4 // add 16 results local
111 paddd xmm1, xmm3
174 "movdqa %%xmm2,%%xmm3 \n"
175 "punpcklwd %%xmm7,%%xmm3 \n"
176 "pmulld %%xmm5,%%xmm3 \n"
189 "paddd %%xmm4,%%xmm3 \n
271 movdqa xmm3, xmm1 \/\/ abs trick local
    [all...]
  /hardware/intel/common/omx-components/videocodec/libvpx_internal/libvpx/third_party/libyuv/source/
scale.c 684 movdqa xmm3, [eax + esi + 16] local
687 pavgb xmm1, xmm3
691 movdqa xmm3, xmm1 local
694 pand xmm3, xmm5 local
696 pavgw xmm1, xmm3
761 movdqa xmm3, [esi + ebx + 16] local
763 pavgb xmm1, xmm3
765 movdqa xmm3, [esi + ebx * 2 + 16] local
770 pavgb xmm3, xmm5 local
772 pavgb xmm1, xmm3
776 movdqa xmm3, xmm1 local
779 pand xmm3, xmm7 local
852 movdqa xmm3, [esi + ebx + 16] local
856 movdqa xmm3, [esi + ebx * 2 + 16] local
862 pavgb xmm3, xmm5 local
867 movdqa xmm3, [ebp + 16] local
871 pavgb xmm3, xmm5 local
879 pavgb xmm3, xmm5 local
918 movdqa xmm3, _shuf0 local
969 movdqa xmm3, _shuf11 local
1026 movdqa xmm3, _shuf11 local
1129 movhlps xmm3, xmm2 local
1133 punpcklbw xmm3, xmm7 local
1138 movhlps xmm3, xmm2 local
1140 punpcklbw xmm3, xmm7 local
1151 movdqa xmm3, xmm1 \/\/ 8 pixels -> 3,4,5 of xmm2 local
1153 paddusw xmm3, xmm1 local
1155 paddusw xmm3, xmm1 local
1156 pshufb xmm3, xmm5 local
1238 movhlps xmm3, xmm2 local
1241 punpcklbw xmm3, xmm5 local
1251 paddusw xmm3, xmm1 local
1302 movdqa xmm3, xmm2 local
1306 punpckhbw xmm3, xmm7 local
1310 pmullw xmm3, xmm6 local
1453 movdqa xmm3, _shuf11 local
    [all...]
  /bionic/libc/arch-x86/atom/string/
sse2-memchr-atom.S 157 movdqa 32(%edi), %xmm3
158 pcmpeqb %xmm1, %xmm3
160 pmovmskb %xmm3, %eax
189 movdqa 32(%edi), %xmm3
190 pcmpeqb %xmm1, %xmm3
192 pmovmskb %xmm3, %eax
215 movdqa 32(%edi), %xmm3
219 pcmpeqb %xmm1, %xmm3
222 pmaxub %xmm0, %xmm3
224 pmaxub %xmm3, %xmm
    [all...]
  /external/compiler-rt/lib/builtins/i386/
floatundisf.S 90 movd %edx, %xmm3
93 psrlq %xmm3, %xmm0 // (big input) ? input >> 12 : input
98 pslld $23, %xmm3
99 paddd %xmm3, %xmm0 // (float)input
  /external/llvm/test/MC/X86/
cfi_def_cfa-crash.s 28 movdqa %xmm3, -0x50(%rbp)
47 movdqa -0x50(%rbp), %xmm3

Completed in 143 milliseconds

1 2 3 4 56 7 8 91011>>