/external/llvm/test/MC/X86/ |
shuffle-comments.s | 3 palignr $8, %xmm0, %xmm1 4 # CHECK: xmm1 = xmm0[8,9,10,11,12,13,14,15],xmm1[0,1,2,3,4,5,6,7] 5 palignr $8, (%rax), %xmm1 6 # CHECK: xmm1 = mem[8,9,10,11,12,13,14,15],xmm1[0,1,2,3,4,5,6,7] 8 palignr $16, %xmm0, %xmm1 9 # CHECK: xmm1 = xmm1[0,1,2,3,4,5,6,7,8,9,10,11,12,13,14,15] 10 palignr $16, (%rax), %xmm1 [all...] |
x86_64-fma4-encoding.s | 4 // CHECK: vfmaddss (%rcx), %xmm1, %xmm0, %xmm0 6 vfmaddss (%rcx), %xmm1, %xmm0, %xmm0 8 // CHECK: vfmaddss %xmm1, (%rcx), %xmm0, %xmm0 10 vfmaddss %xmm1, (%rcx),%xmm0, %xmm0 12 // CHECK: vfmaddss %xmm2, %xmm1, %xmm0, %xmm0 14 vfmaddss %xmm2, %xmm1, %xmm0, %xmm0 16 // CHECK: vfmaddsd (%rcx), %xmm1, %xmm0, %xmm0 18 vfmaddsd (%rcx), %xmm1, %xmm0, %xmm0 20 // CHECK: vfmaddsd %xmm1, (%rcx), %xmm0, %xmm0 22 vfmaddsd %xmm1, (%rcx),%xmm0, %xmm [all...] |
x86_64-xop-encoding.s | 8 // CHECK: vphsubwd (%rcx,%rax), %xmm1 10 vphsubwd (%rcx,%rax), %xmm1 11 // CHECK: vphsubwd %xmm0, %xmm1 13 vphsubwd %xmm0, %xmm1 16 // CHECK: vphsubdq (%rcx,%rax), %xmm1 18 vphsubdq (%rcx,%rax), %xmm1 19 // CHECK: vphsubdq %xmm0, %xmm1 21 vphsubdq %xmm0, %xmm1 24 // CHECK: vphsubbw (%rax), %xmm1 26 vphsubbw (%rax), %xmm1 [all...] |
x86-32-fma3.s | 3 // CHECK: vfmadd132pd %xmm2, %xmm5, %xmm1 5 vfmadd132pd %xmm2, %xmm5, %xmm1 7 // CHECK: vfmadd132pd (%eax), %xmm5, %xmm1 9 vfmadd132pd (%eax), %xmm5, %xmm1 11 // CHECK: vfmadd132ps %xmm2, %xmm5, %xmm1 13 vfmadd132ps %xmm2, %xmm5, %xmm1 15 // CHECK: vfmadd132ps (%eax), %xmm5, %xmm1 17 vfmadd132ps (%eax), %xmm5, %xmm1 19 // CHECK: vfmadd213pd %xmm2, %xmm5, %xmm1 21 vfmadd213pd %xmm2, %xmm5, %xmm1 [all...] |
x86-32-avx.s | 274 // CHECK: vunpckhps %xmm1, %xmm2, %xmm4 276 vunpckhps %xmm1, %xmm2, %xmm4 278 // CHECK: vunpckhpd %xmm1, %xmm2, %xmm4 280 vunpckhpd %xmm1, %xmm2, %xmm4 282 // CHECK: vunpcklps %xmm1, %xmm2, %xmm4 284 vunpcklps %xmm1, %xmm2, %xmm4 286 // CHECK: vunpcklpd %xmm1, %xmm2, %xmm4 288 vunpcklpd %xmm1, %xmm2, %xmm4 306 // CHECK: vcmpps $0, %xmm0, %xmm6, %xmm1 308 vcmpps $0, %xmm0, %xmm6, %xmm1 [all...] |
x86_64-sse4a.s | 7 extrq %xmm1, %xmm0 8 # CHECK: extrq %xmm1, %xmm0 11 insertq $6, $5, %xmm1, %xmm0 12 # CHECK: insertq $6, $5, %xmm1, %xmm0 15 insertq %xmm1, %xmm0 16 # CHECK: insertq %xmm1, %xmm0
|
/external/compiler-rt/lib/i386/ |
ashrdi3.S | 18 movd 8(%esp), %xmm1 19 punpckldq %xmm1, %xmm0 // Load input 31 pcmpeqb %xmm1, %xmm1 // -1ULL 32 psrlq $58, %xmm1 // 0x3f 33 pandn %xmm1, %xmm2 // 63 - count 34 pcmpeqb %xmm1, %xmm1 // -1ULL 35 psubq %xmm1, %xmm2 // 64 - count 36 psllq %xmm2, %xmm1 // -1 << (64 - count) = leading sign bit [all...] |
floatdidf.S | 22 cvtsi2sd 8(%esp), %xmm1 26 mulsd REL_ADDR(twop32), %xmm1 // a_hi as a double (without rounding) 28 subsd %xmm2, %xmm1 // a_hi - 0x1p52 (no rounding occurs) 30 addsd %xmm1, %xmm0 // a_hi + a_lo (round happens here)
|
floatdisf.S | 22 movd 8(%esp), %xmm1 23 punpckldq %xmm1, %xmm0
|
floatdixf.S | 22 movd 8(%esp), %xmm1 23 punpckldq %xmm1, %xmm0
|
floatundidf.S | 34 movss 8(%esp), %xmm1 // high 32 bits of a 38 orpd REL_ADDR(twop84), %xmm1 // 0x1p84 + a_hi (no rounding occurs) 39 subsd REL_ADDR(twop84_plus_twop52), %xmm1 // a_hi - 0x1p52 (no rounding occurs) 41 addsd %xmm1, %xmm0 // a_hi + a_lo (round happens here)
|
floatundixf.S | 27 movss 4(%esp), %xmm1 // lo 32 bits of input 29 orpd REL_ADDR(twop52), %xmm1 // 2^52 + lo (as a double) 31 movsd %xmm1, 4(%esp)
|
floatundisf.S | 33 movd 8(%esp), %xmm1 35 punpckldq %xmm1, %xmm0 73 movd 8(%esp), %xmm1 75 punpckldq %xmm1, %xmm0 82 movsd STICKY, %xmm1 // (big input) ? 0xfff : 0 86 andpd %xmm0, %xmm1 // (big input) ? input & 0xfff : 0 89 orpd %xmm2, %xmm1 // 0x1.0p52 + ((big input) ? input & 0xfff : input) 90 orpd %xmm1, %xmm0 // 0x1.0p52 + ((big input) ? (input >> 12 | input & 0xfff) : input)
|
/dalvik/vm/mterp/x86/ |
OP_ADD_DOUBLE.S | 8 movq (rFP, %ecx, 4), %xmm1 # %xmm1<- vCC 10 addsd %xmm1, %xmm0
|
OP_MUL_DOUBLE.S | 9 movq (rFP, %ecx, 4), %xmm1 # %xmm1<- vCC 11 mulsd %xmm1, %xmm0
|
OP_SUB_DOUBLE.S | 9 movq (rFP, %ecx, 4), %xmm1 # %xmm1<- vCC 11 subsd %xmm1, %xmm0
|
OP_ADD_DOUBLE_2ADDR.S | 8 movq (rFP, rINST, 4), %xmm1 # %xmm1<- vB 11 addsd %xmm1, %xmm0 # %xmm0<- vA op vB
|
OP_MUL_DOUBLE_2ADDR.S | 9 movq (rFP, rINST, 4), %xmm1 # %xmm1<- vB 12 mulsd %xmm1, %xmm0 # %xmm0<- vA op vB
|
OP_SUB_DOUBLE_2ADDR.S | 9 movq (rFP, rINST, 4), %xmm1 # %xmm1<- vB 12 subsd %xmm1, %xmm0 # %xmm0<- vA op vB
|
/external/libvpx/libvpx/vp8/common/x86/ |
variance_impl_ssse3.asm | 68 movdqu xmm1, XMMWORD PTR [rsi+1] 71 punpcklbw xmm0, xmm1 72 punpckhbw xmm2, xmm1 92 movdqu xmm1, XMMWORD PTR [rsi] 94 movdqa xmm3, xmm1 96 punpcklbw xmm1, xmm2 98 pmaddubsw xmm1, [rax] 101 paddw xmm1, [GLOBAL(xmm_bi_rd)] 103 psraw xmm1, xmm_filter_shift 105 packuswb xmm1, xmm [all...] |
postproc_sse2.asm | 18 movdqa xmm5, xmm1 22 psubusb xmm4, xmm1 23 psubusb xmm1, xmm0 26 paddusb xmm4, xmm1 31 pxor xmm1, xmm1 37 pcmpeqb xmm2, xmm1 38 pcmpeqb xmm7, xmm1 45 movdqa xmm2, xmm1 46 pavgb xmm1, xmm [all...] |
/external/compiler-rt/lib/x86_64/ |
floatundidf.S | 38 movd %rdi, %xmm1 39 subsd REL_ADDR(twop84_plus_twop52), %xmm1 // a_hi - 0x1p52 (no rounding occurs) 40 addsd %xmm1, %xmm0 // a_hi + a_lo (round happens here)
|
/external/openssl/crypto/aes/asm/ |
aesni-x86.S | 14 movups 16(%edx),%xmm1 20 movups (%edx),%xmm1 38 movups 16(%edx),%xmm1 44 movups (%edx),%xmm1 56 movups 16(%edx),%xmm1 67 movups 16(%edx),%xmm1 87 movups 16(%edx),%xmm1 98 movups 16(%edx),%xmm1 117 movups 16(%edx),%xmm1 131 movups 16(%edx),%xmm1 [all...] |
/external/libvpx/libvpx/vp8/encoder/x86/ |
dct_sse2.asm | 72 movq xmm1, MMWORD PTR[input ] ;23 22 21 20 76 punpcklqdq xmm1, xmm3 ;33 32 31 30 23 22 21 20 79 punpckldq xmm0, xmm1 ;23 22 03 02 21 20 01 00 80 punpckhdq xmm2, xmm1 ;33 32 13 12 31 30 11 10 81 movdqa xmm1, xmm0 83 pshufhw xmm1, xmm1, 0b1h ;22 23 02 03 xx xx xx xx 86 punpckhdq xmm1, xmm2 ;32 33 22 23 12 13 02 03 88 paddw xmm0, xmm1 ;b1 a1 b1 a1 b1 a1 b1 a1 89 psubw xmm3, xmm1 ;c1 d1 c1 d1 c1 d1 c1 d [all...] |
/bionic/libc/arch-x86/string/ |
ssse3-strcmp-latest.S | 189 movlpd (%eax), %xmm1 191 movhpd 8(%eax), %xmm1 193 pcmpeqb %xmm1, %xmm0 194 pcmpeqb %xmm2, %xmm1 195 psubb %xmm0, %xmm1 196 pmovmskb %xmm1, %ecx 273 movdqa (%eax), %xmm1 275 pcmpeqb %xmm1, %xmm0 276 pcmpeqb (%edx), %xmm1 277 psubb %xmm0, %xmm1 [all...] |