/external/libvpx/vp8/encoder/x86/ |
subtract_sse2.asm | 98 movdqa xmm2, xmm0 102 pxor xmm2, [GLOBAL(t80)] 103 pcmpgtb xmm1, xmm2 ; obtain sign information 105 movdqa xmm2, xmm0 108 punpckhbw xmm2, xmm3 ; put sign back to subtraction 111 movdqa XMMWORD PTR [rdi +16], xmm2 170 movq xmm2, MMWORD PTR [rsi+rdx] 172 punpcklqdq xmm0, xmm2 174 movdqa xmm2, xmm0 178 pxor xmm2, [GLOBAL(t80) [all...] |
variance_impl_ssse3.asm | 69 movdqa xmm2, xmm0 72 punpckhbw xmm2, xmm1 74 pmaddubsw xmm2, [rax] 77 paddw xmm2, [GLOBAL(xmm_bi_rd)] 79 psraw xmm2, xmm_filter_shift 81 packuswb xmm0, xmm2 93 movdqu xmm2, XMMWORD PTR [rsi+1] 96 punpcklbw xmm1, xmm2 97 punpckhbw xmm3, xmm2 107 movdqa xmm2, xmm [all...] |
fwalsh_sse2.asm | 34 movq xmm2, MMWORD PTR [rsi] 38 punpcklwd xmm2, xmm3 41 punpckldq xmm0, xmm2 ; ip[1] ip[0] 42 punpckhdq xmm1, xmm2 ; ip[3] ip[2] 44 movdqa xmm2, xmm0 46 psubw xmm2, xmm1 49 psllw xmm2, 2 ; c1 b1 52 punpcklqdq xmm0, xmm2 ; b1 a1 53 punpckhqdq xmm1, xmm2 ; c1 d1 61 movdqa xmm2, xmm [all...] |
quantize_sse2.asm | 84 movdqa xmm2, [rcx] 89 paddw xmm2, xmm7 93 psubw xmm1, xmm2 99 paddw xmm1, xmm2 102 movdqa xmm2, [rdx] 109 paddw xmm1, xmm2 170 movdqa xmm2, [rsp + qcoeff] 183 pxor xmm2, xmm0 186 psubw xmm2, xmm0 195 pmullw xmm0, xmm2 [all...] |
dct_sse2.asm | 68 movq xmm2, MMWORD PTR[input+ pitch] ;13 12 11 10 73 punpcklqdq xmm0, xmm2 ;13 12 11 10 03 02 01 00 76 movdqa xmm2, xmm0 78 punpckhdq xmm2, xmm1 ;33 32 13 12 31 30 11 10 80 punpckldq xmm0, xmm2 ;31 21 30 20 11 10 01 00 82 pshufhw xmm2, xmm2, 0b1h ;32 33 12 13 xx xx xx xx 84 punpckhdq xmm1, xmm2 ;32 33 22 23 12 13 02 03 109 movdqa xmm2, xmm0 111 punpckhqdq xmm2, xmm3 ;23 22 21 20 33 32 31 3 [all...] |
sad_sse4.asm | 19 movq xmm2, MMWORD PTR [rdi+16] 21 punpcklqdq xmm3, xmm2 23 movdqa xmm2, xmm1 25 mpsadbw xmm2, xmm0, 0x5 33 paddw xmm1, xmm2 40 movq xmm2, MMWORD PTR [rdi+16] 42 punpcklqdq xmm3, xmm2 44 movdqa xmm2, xmm5 46 mpsadbw xmm2, xmm0, 0x5 54 paddw xmm5, xmm2 [all...] |
variance_impl_sse2.asm | 39 movdqa xmm2, [rax+32] 43 pmaddwd xmm2, xmm2 47 paddd xmm2, xmm3 49 paddd xmm4, xmm2 130 movdqu xmm2, XMMWORD PTR [rdi] 136 movdqa xmm4, xmm2 142 punpcklbw xmm2, xmm0 146 psubw xmm1, xmm2 177 movdqa xmm2, xmm [all...] |
encodeopt.asm | 220 pxor xmm2, xmm2 223 por xmm1, xmm2 248 paddd xmm2, xmm5 250 paddd xmm2, xmm3 253 movdqa xmm0, xmm2 256 punpckhdq xmm2, xmm7 257 paddd xmm0, xmm2 350 movdqa xmm2, [rdi] 352 psubw xmm1, xmm2 [all...] |
/external/llvm/test/MC/X86/ |
x86-32-avx.s | 3 // CHECK: vaddss %xmm4, %xmm6, %xmm2 5 vaddss %xmm4, %xmm6, %xmm2 7 // CHECK: vmulss %xmm4, %xmm6, %xmm2 9 vmulss %xmm4, %xmm6, %xmm2 11 // CHECK: vsubss %xmm4, %xmm6, %xmm2 13 vsubss %xmm4, %xmm6, %xmm2 15 // CHECK: vdivss %xmm4, %xmm6, %xmm2 17 vdivss %xmm4, %xmm6, %xmm2 19 // CHECK: vaddsd %xmm4, %xmm6, %xmm2 21 vaddsd %xmm4, %xmm6, %xmm2 [all...] |
x86_64-xop-encoding.s | 27 // CHECK: vphsubbw %xmm2, %xmm1 29 vphsubbw %xmm2, %xmm1 35 // CHECK: vphaddwq %xmm6, %xmm2 37 vphaddwq %xmm6, %xmm2 59 // CHECK: vphadduwd %xmm2, %xmm1 61 vphadduwd %xmm2, %xmm1 67 // CHECK: vphaddudq %xmm6, %xmm2 69 vphaddudq %xmm6, %xmm2 83 // CHECK: vphaddubq %xmm2, %xmm2 [all...] |
x86-32-fma3.s | 3 // CHECK: vfmadd132pd %xmm2, %xmm5, %xmm1 5 vfmadd132pd %xmm2, %xmm5, %xmm1 11 // CHECK: vfmadd132ps %xmm2, %xmm5, %xmm1 13 vfmadd132ps %xmm2, %xmm5, %xmm1 19 // CHECK: vfmadd213pd %xmm2, %xmm5, %xmm1 21 vfmadd213pd %xmm2, %xmm5, %xmm1 27 // CHECK: vfmadd213ps %xmm2, %xmm5, %xmm1 29 vfmadd213ps %xmm2, %xmm5, %xmm1 35 // CHECK: vfmadd231pd %xmm2, %xmm5, %xmm1 37 vfmadd231pd %xmm2, %xmm5, %xmm [all...] |
/external/llvm/test/MC/Disassembler/X86/ |
enhanced.txt | 7 # CHECK: [o:xorps][w: ][2-r:%xmm1=r129][p:,][w: ][0-r:%xmm2=r130] 0:[XMM2/130]=0 1:[XMM2/130]=0 2:[XMM1/129]=0 9 # CHECK: [o:andps][w: ][2-r:%xmm1=r129][p:,][w: ][0-r:%xmm2=r130] 0:[XMM2/130]=0 1:[XMM2/130]=0 2:[XMM1/129]=0
|
/external/libvpx/vp8/common/x86/ |
idctllm_sse2.asm | 61 movq xmm2, [rax+2*rcx] 67 punpcklbw xmm2, xmm7 76 paddw xmm2, xmm4 82 packuswb xmm2, xmm7 91 movq [rax], xmm2 122 ; note the transpose of xmm1 and xmm2, necessary for shuffle 125 movdqa xmm2, [rax+16] 137 pmullw xmm2, [rdx+16] 149 movdqa xmm4, xmm2 150 punpckldq xmm2, xmm [all...] |
subpixel_ssse3.asm | 75 movq xmm2, MMWORD PTR [rsi + 3] ; 3 4 5 6 7 8 9 10 77 punpcklbw xmm0, xmm2 ; -2 3 -1 4 0 5 1 6 2 7 3 8 4 9 5 10 82 movdqa xmm2, xmm1 85 pshufb xmm2, [GLOBAL(shuf3bfrom1)] 89 pmaddubsw xmm2, xmm6 95 paddsw xmm2, xmm7 97 paddsw xmm0, xmm2 137 movdqa xmm2, xmm0 140 pshufb xmm2, xmm4 144 pmaddubsw xmm2, xmm [all...] |
recon_sse2.asm | 36 movq xmm2, MMWORD PTR [rsi+8] 37 punpcklbw xmm2, xmm0 38 paddsw xmm2, XMMWORD PTR [rdx+16] 39 packuswb xmm2, xmm0 ; pack and unpack to saturate 40 movq MMWORD PTR [rdi+rax], xmm2 91 movdqa xmm2, XMMWORD PTR [rsi+16] 92 movdqa xmm6, xmm2 93 punpcklbw xmm2, xmm0 95 paddsw xmm2, XMMWORD PTR [rdx+32] 97 packuswb xmm2, xmm6 ; pack and unpack to saturat [all...] |
loopfilter_sse2.asm | 19 movdqa xmm2, [rdi+2*rax] ; q3 25 movlps xmm2, [rsi + rcx*2] ; q3 30 movhps xmm2, [rdi + rcx*2] 45 psubusb xmm1, xmm2 ; q2-=q3 46 psubusb xmm2, xmm6 ; q3-=q2 52 por xmm1, xmm2 ; abs(q3-q2) 66 movdqa xmm2, [rsi+4*rax] ; p3 70 movlps xmm2, [rsi + rax] ; p3 74 movhps xmm2, [rdi + rax] 85 psubusb xmm4, xmm2 ; p2-=p [all...] |
subpixel_sse2.asm | 183 movq xmm2, MMWORD PTR [rsi +14] 184 pslldq xmm2, 8 186 por xmm2, xmm1 243 movdqa xmm3, xmm2 244 movdqa xmm4, xmm2 246 movdqa xmm5, xmm2 247 movdqa xmm6, xmm2 249 movdqa xmm7, xmm2 272 psrldq xmm2, 5 ; xx xx xx xx xx 0d 0c 0b 0a 09 08 07 06 05 04 03 276 punpcklbw xmm2, xmm0 ; xx0a xx09 xx08 xx07 xx06 xx05 xx04 xx0 [all...] |
postproc_sse2.asm | 48 movd xmm2, dword ptr arg(6) ;flimit 49 punpcklwd xmm2, xmm2 50 punpckldq xmm2, xmm2 51 punpcklqdq xmm2, xmm2 78 pcmpgtw xmm7, xmm2 89 pcmpgtw xmm6, xmm2 103 pcmpgtw xmm6, xmm2 [all...] |
/bionic/libc/arch-x86/string/ |
ssse3-strcmp-latest.S | 190 movlpd (%edx), %xmm2 192 movhpd 8(%edx), %xmm2 194 pcmpeqb %xmm2, %xmm1 291 movdqa (%edx, %ecx), %xmm2 294 pcmpeqb %xmm2, %xmm1 316 movdqa (%edx), %xmm2 319 pslldq $15, %xmm2 320 pcmpeqb %xmm1, %xmm2 321 psubb %xmm0, %xmm2 322 pmovmskb %xmm2, %ed [all...] |
ssse3-memcpy5.S | 305 movdqa 0x20(%eax), %xmm2 315 movdqa %xmm2, 0x20(%edx) 374 movdqa 0x20(%eax), %xmm2 384 movdqa %xmm2, 0x20(%edx) 447 movdqa 16(%eax, %edi), %xmm2 451 palignr $1, %xmm2, %xmm3 452 palignr $1, %xmm1, %xmm2 454 movdqa %xmm2, -32(%edx, %edi) 459 movdqa 16(%eax, %edi), %xmm2 463 palignr $1, %xmm2, %xmm [all...] |
/frameworks/compile/libbcc/runtime/lib/i386/ |
floatdidf.S | 27 movsd REL_ADDR(twop52), %xmm2 // 0x1.0p52 28 subsd %xmm2, %xmm1 // a_hi - 0x1p52 (no rounding occurs) 29 orpd %xmm2, %xmm0 // 0x1p52 + a_lo (no rounding occurs)
|
/external/openssl/crypto/aes/asm/ |
vpaes-x86.pl | 184 &movdqa ("xmm2",&QWP($k_ipt,$const)); 189 &pshufb ("xmm2","xmm0"); 192 &pxor ("xmm2","xmm5"); 193 &pxor ("xmm0","xmm2"); 202 &pshufb ("xmm4","xmm2"); # 4 = sb1u 208 &pshufb ("xmm5","xmm2"); # 4 = sb2u 210 &movdqa ("xmm2",&QWP($k_sb2+16,$const));# 2 : sb2t 211 &pshufb ("xmm2","xmm3"); # 2 = sb2t 212 &pxor ("xmm2","xmm5"); # 2 = 2A 217 &pxor ("xmm0","xmm2"); # 0 = 2A+ [all...] |
/external/libyuv/files/source/ |
row_posix.cc | 79 "movdqa %%xmm3,%%xmm2\n" 80 "palignr $0x8,%%xmm1,%%xmm2\n" // xmm2 = { xmm3[0:3] xmm1[8:15] } 81 "pshufb %%xmm6,%%xmm2\n" 82 "por %%xmm7,%%xmm2\n" 85 "movdqa %%xmm2,0x20(%1)\n" 116 "movdqa %%xmm3,%%xmm2\n" 117 "palignr $0x8,%%xmm1,%%xmm2\n" // xmm2 = { xmm3[0:3] xmm1[8:15] } 118 "pshufb %%xmm6,%%xmm2\n [all...] |
row_win.cc | 90 movdqa xmm2, [eax + 32] local 94 pmaddubsw xmm2, xmm7 local 98 phaddw xmm2, xmm3 local 100 psrlw xmm2, 7 local 101 packuswb xmm0, xmm2 123 movdqa xmm2, [eax + 32] local 127 pmaddubsw xmm2, xmm7 local 131 phaddw xmm2, xmm3 local 133 psrlw xmm2, 7 local 134 packuswb xmm0, xmm2 156 movdqa xmm2, [eax + 32] local 160 pmaddubsw xmm2, xmm7 local 164 phaddw xmm2, xmm3 local 166 psrlw xmm2, 7 local 197 movdqa xmm2, [eax + 32] local 201 pavgb xmm2, [eax + esi + 32] local 209 shufps xmm2, xmm3, 0x88 local 211 pavgb xmm2, xmm4 local 219 pmaddubsw xmm2, xmm7 local 261 movdqa xmm2, [eax + 32] local 265 pavgb xmm2, [eax + esi + 32] local 273 shufps xmm2, xmm3, 0x88 local 275 pavgb xmm2, xmm4 local 283 pmaddubsw xmm2, xmm7 local 325 movdqa xmm2, [eax + 32] local 329 pavgb xmm2, [eax + esi + 32] local 337 shufps xmm2, xmm3, 0x88 local 339 pavgb xmm2, xmm4 local 347 pmaddubsw xmm2, xmm7 local 384 movdqa xmm2, xmm3 local 385 palignr xmm2, xmm1, 8 \/\/ xmm2 = { xmm3[0:3] xmm1[8:15]} local 386 pshufb xmm2, xmm6 local 387 por xmm2, xmm7 local 423 movdqa xmm2, xmm3 local 424 palignr xmm2, xmm1, 8 \/\/ xmm2 = { xmm3[0:3] xmm1[8:15]} local 425 pshufb xmm2, xmm6 local 426 por xmm2, xmm7 local [all...] |
/external/llvm/test/CodeGen/X86/ |
illegal-vector-args-return.ll | 2 ; RUN: llc < %s -march=x86 -mattr=+sse2 | grep {mulpd %xmm2, %xmm0} 4 ; RUN: llc < %s -march=x86 -mattr=+sse2 | grep {addps %xmm2, %xmm0}
|