/external/valgrind/main/none/tests/amd64/ |
pcmpxstrx64.stdout.exp | 3 istri $0x4A: xmm0 55555555555555555555555555555555 rcx 5555555555550006 flags 00000881 4 istri $0x0A: xmm0 55555555555555555555555555555555 rcx 5555555555550000 flags 00000881 5 istrm $0x4A: xmm0 000000000000000000ffffffffffffff rcx 5555555555555555 flags 00000881 6 istrm $0x0A: xmm0 0000000000000000000000000000007f rcx 5555555555555555 flags 00000881 7 estri $0x4A: xmm0 55555555555555555555555555555555 rcx 555555555555000f flags 000008c1 8 estri $0x0A: xmm0 55555555555555555555555555555555 rcx 5555555555550000 flags 000008c1 9 estrm $0x4A: xmm0 ffffffffffffffffffffffffffffffff rcx 5555555555555555 flags 000008c1 10 estrm $0x0A: xmm0 0000000000000000000000000000ffff rcx 5555555555555555 flags 000008c1 13 istri $0x4A: xmm0 55555555555555555555555555555555 rcx 555555555555000f flags 000000c1 14 istri $0x0A: xmm0 55555555555555555555555555555555 rcx 5555555555550007 flags 000000c [all...] |
/external/llvm/test/MC/X86/ |
x86_64-fma4-encoding.s | 4 // CHECK: vfmaddss (%rcx), %xmm1, %xmm0, %xmm0 6 vfmaddss (%rcx), %xmm1, %xmm0, %xmm0 8 // CHECK: vfmaddss %xmm1, (%rcx), %xmm0, %xmm0 10 vfmaddss %xmm1, (%rcx),%xmm0, %xmm0 12 // CHECK: vfmaddss %xmm2, %xmm1, %xmm0, %xmm0 [all...] |
/external/chromium/base/ |
cpu_unittest.cc | 27 __asm xorps xmm0, xmm0; local 32 __asm psrldq xmm0, 0; local 37 __asm addsubpd xmm0, xmm0; local 42 __asm psignb xmm0, xmm0; local 47 __asm pmuldq xmm0, xmm0; local 62 __asm__ __volatile__("xorps %%xmm0, %%xmm0\n" : : : "xmm0") [all...] |
/external/llvm/test/CodeGen/X86/ |
v2f32.ll | 14 ; X64-NEXT: pshufd $1, %xmm0, %xmm1 15 ; X64-NEXT: addss %xmm0, %xmm1 20 ; W64-NEXT: movdqa (%rcx), %xmm0 21 ; W64-NEXT: pshufd $1, %xmm0, %xmm1 22 ; W64-NEXT: addss %xmm0, %xmm1 27 ; X32-NEXT: pshufd $1, %xmm0, %xmm1 28 ; X32-NEXT: addss %xmm0, %xmm1 40 ; X64-NEXT: addps %xmm1, %xmm0 44 ; W64-NEXT: movaps (%rcx), %xmm0 45 ; W64-NEXT: addps (%rdx), %xmm0 [all...] |
vec_set-9.ll | 2 ; RUN: llc < %s -march=x86-64 | grep {movlhps.*%xmm0, %xmm0}
|
sse-minmax.ll | 19 ; CHECK-NEXT: maxsd %xmm1, %xmm0 22 ; UNSAFE-NEXT: maxsd %xmm1, %xmm0 25 ; FINITE-NEXT: maxsd %xmm1, %xmm0 34 ; CHECK-NEXT: minsd %xmm1, %xmm0 37 ; UNSAFE-NEXT: minsd %xmm1, %xmm0 40 ; FINITE-NEXT: minsd %xmm1, %xmm0 49 ; CHECK-NEXT: minsd %xmm0, %xmm1 50 ; CHECK-NEXT: movap{{[sd]}} %xmm1, %xmm0 53 ; UNSAFE-NEXT: minsd %xmm0, %xmm1 54 ; UNSAFE-NEXT: movap{{[sd]}} %xmm1, %xmm0 [all...] |
break-sse-dep.ll | 7 ; CHECK: movss ([[A0:%rdi|%rcx]]), %xmm0 8 ; CHECK: cvtss2sd %xmm0, %xmm0 18 ; CHECK: cvtsd2ss ([[A0]]), %xmm0 27 ; CHECK: movss ([[A0]]), %xmm0 28 ; CHECK: sqrtss %xmm0, %xmm0 37 ; CHECK: movsd ([[A0]]), %xmm0 38 ; CHECK: sqrtsd %xmm0, %xmm0 [all...] |
peep-vector-extract-concat.ll | 2 ; CHECK: pshufd $3, %xmm0, %xmm0 6 ; WIN64: movss 12(%rcx), %xmm0
|
/external/libvpx/vp8/encoder/x86/ |
subtract_sse2.asm | 95 movdqa xmm0, XMMWORD PTR [rsi] ; src 98 movdqa xmm2, xmm0 99 psubb xmm0, xmm1 105 movdqa xmm2, xmm0 107 punpcklbw xmm0, xmm1 ; put sign back to subtraction 110 movdqa XMMWORD PTR [rdi], xmm0 169 movq xmm0, MMWORD PTR [rsi] ; src 172 punpcklqdq xmm0, xmm2 174 movdqa xmm2, xmm0 175 psubb xmm0, xmm1 ; subtraction with sign misse [all...] |
sad_ssse3.asm | 16 movdqa xmm0, XMMWORD PTR [rsi] 21 psadbw xmm5, xmm0 22 psadbw xmm6, xmm0 23 psadbw xmm7, xmm0 25 movdqa xmm0, XMMWORD PTR [rsi] 30 psadbw xmm1, xmm0 31 psadbw xmm2, xmm0 32 psadbw xmm3, xmm0 38 movdqa xmm0, XMMWORD PTR [rsi+rax] 46 psadbw xmm1, xmm0 [all...] |
fwalsh_sse2.asm | 31 movq xmm0, MMWORD PTR [rsi] ; load input 37 punpcklwd xmm0, xmm1 40 movdqa xmm1, xmm0 41 punpckldq xmm0, xmm2 ; ip[1] ip[0] 44 movdqa xmm2, xmm0 45 paddw xmm0, xmm1 48 psllw xmm0, 2 ; d1 a1 51 movdqa xmm1, xmm0 52 punpcklqdq xmm0, xmm2 ; b1 a1 56 movq xmm6, xmm0 [all...] |
sad_sse4.asm | 16 movdqa xmm0, XMMWORD PTR [rsi] 24 mpsadbw xmm1, xmm0, 0x0 25 mpsadbw xmm2, xmm0, 0x5 27 psrldq xmm0, 8 30 mpsadbw xmm3, xmm0, 0x0 31 mpsadbw xmm4, xmm0, 0x5 37 movdqa xmm0, XMMWORD PTR [rsi] 45 mpsadbw xmm5, xmm0, 0x0 46 mpsadbw xmm2, xmm0, 0x5 48 psrldq xmm0, [all...] |
temporal_filter_apply_sse2.asm | 64 movd xmm0, arg(5) ; filter_weight 65 pshuflw xmm0, xmm0, 0 66 punpcklwd xmm0, xmm0 67 movdqa [rsp + filter_weight], xmm0 78 movq xmm0, [rsi] ; first row 80 punpcklbw xmm0, xmm7 ; src[ 0- 7] 87 movdqa xmm0, [rsi] ; src (frame1) 89 movdqa xmm1, xmm0 [all...] |
/bionic/libc/arch-x86/string/ |
sse2-memset5-atom.S | 283 /* Fill xmm0 with the pattern. */ 285 pxor %xmm0, %xmm0 287 movd %eax, %xmm0 288 pshufd $0, %xmm0, %xmm0 294 movdqu %xmm0, (%edx) 300 movd %xmm0, %eax 350 movdqa %xmm0, (%edx) 351 movdqa %xmm0, 0x10(%edx [all...] |
/external/valgrind/main/memcheck/tests/amd64/ |
bug279698.c | 11 ("movdqu (%0), %%xmm0 \n" 12 "packuswb %%xmm0, %%xmm0 \n" 13 "movdqu %%xmm0, 16(%0) \n" 15 :"memory","xmm0"
|
xor-undef-amd64.c | 66 "movups 0(%0), %%xmm0\n\t" 68 "xorps %%xmm8, %%xmm0\n\t" 69 "movups %%xmm0, 32(%0)\n\t" 73 : : "r"(junk) : "rax", "xmm8", "xmm0", "cc", "memory" 78 "movups 0(%0), %%xmm0\n\t" 80 "xorps %%xmm0, %%xmm0\n\t" 81 "movups %%xmm0, 32(%0)\n\t" 85 : : "r"(junk) : "rax", "xmm8", "xmm0", "cc", "memory" 93 "movups 0(%0), %%xmm0\n\t [all...] |
/system/core/libcutils/arch-x86/ |
sse2-memset32-atom.S | 230 pxor %xmm0, %xmm0 232 movd %eax, %xmm0 233 pshufd $0, %xmm0, %xmm0 239 movdqu %xmm0, (%edx) 245 movd %xmm0, %eax 295 movdqa %xmm0, (%edx) 296 movdqa %xmm0, 0x10(%edx) 297 movdqa %xmm0, 0x20(%edx [all...] |
sse2-memset16-atom.S | 314 pxor %xmm0, %xmm0 316 movd %eax, %xmm0 317 pshufd $0, %xmm0, %xmm0 323 movdqu %xmm0, (%edx) 329 movd %xmm0, %eax 381 movdqa %xmm0, (%edx) 382 movdqa %xmm0, 0x10(%edx) 383 movdqa %xmm0, 0x20(%edx [all...] |
/dalvik/vm/mterp/x86-atom/ |
binopF.S | 20 * specify an instruction that performs "%xmm0 = %xmm0 op %xmm1" 35 movss (rFP, %ecx, 4), %xmm0 # %xmm0<-vBB 37 $instr # %xmm0<- vBB op vCC 38 movss %xmm0, (rFP, rINST, 4) # vAA<- %xmm0; result
|
/external/compiler-rt/lib/x86_64/ |
floatundisf.S | 23 cvtsi2ssq %rdi, %xmm0 29 cvtsi2ssq %rdi, %xmm0 30 mulss REL_ADDR(two), %xmm0
|
/external/valgrind/main/memcheck/tests/x86/ |
xor-undef-x86.c | 68 "movups 0(%0), %%xmm0\n\t" 70 "xorps %%xmm7, %%xmm0\n\t" 71 "movups %%xmm0, 32(%0)\n\t" 77 : : "r"(junk) : "esi", "xmm7", "xmm0", "cc", "memory" 82 "movups 0(%0), %%xmm0\n\t" 84 "xorps %%xmm0, %%xmm0\n\t" 85 "movups %%xmm0, 32(%0)\n\t" 91 : : "r"(junk) : "esi", "xmm7", "xmm0", "cc", "memory" 99 "movups 0(%0), %%xmm0\n\t [all...] |
/frameworks/compile/libbcc/runtime/lib/x86_64/ |
floatundisf.S | 23 cvtsi2ssq %rdi, %xmm0 29 cvtsi2ssq %rdi, %xmm0 30 mulss REL_ADDR(two), %xmm0
|
/external/libvpx/vp8/common/x86/ |
recon_sse2.asm | 27 pxor xmm0, xmm0 30 punpcklbw xmm1, xmm0 32 packuswb xmm1, xmm0 ; pack and unpack to saturate 37 punpcklbw xmm2, xmm0 39 packuswb xmm2, xmm0 ; pack and unpack to saturate 44 punpcklbw xmm3, xmm0 46 packuswb xmm3, xmm0 ; pack and unpack to saturate 51 punpcklbw xmm4, xmm0 53 packuswb xmm4, xmm0 ; pack and unpack to saturat [all...] |
/external/openssl/crypto/aes/asm/ |
vpaes-x86.pl | 169 ## AES-encrypt %xmm0. 172 ## %xmm0 = input 176 ## Output in %xmm0 185 &pandn ("xmm1","xmm0"); 188 &pand ("xmm0","xmm6"); 189 &pshufb ("xmm2","xmm0"); 190 &movdqa ("xmm0",&QWP($k_ipt+16,$const)); 191 &pshufb ("xmm0","xmm1"); 193 &pxor ("xmm0","xmm2"); 204 &movdqa ("xmm0",&QWP($k_sb1+16,$const));# 0 : sb1 [all...] |
/frameworks/compile/libbcc/runtime/lib/i386/ |
floatdidf.S | 23 movss 4(%esp), %xmm0 // low 32 bits of a 29 orpd %xmm2, %xmm0 // 0x1p52 + a_lo (no rounding occurs) 30 addsd %xmm1, %xmm0 // a_hi + a_lo (round happens here) 31 movsd %xmm0, 4(%esp)
|