/external/libvpx/vp8/encoder/x86/ |
dct_sse2.asm | 33 movq xmm3, MMWORD PTR[rdi + rax] ;33 32 31 30 36 punpcklqdq xmm1, xmm3 ;33 32 31 30 23 22 21 20 49 movdqa xmm3, xmm0 51 psubw xmm3, xmm1 ;c1 d1 c1 d1 c1 d1 c1 d1 53 psllw xmm3, 3 ;c1 <<= 3 d1 <<= 3 57 movdqa xmm4, xmm3 58 pmaddwd xmm3, XMMWORD PTR[_5352_2217 GLOBAL] ;c1*2217 + d1*5352 61 paddd xmm3, XMMWORD PTR[_14500 GLOBAL] 63 psrad xmm3, 12 ;(c1 * 2217 + d1 * 5352 + 14500)>>12 67 packssdw xmm3, xmm4 ;op[3] op[1 [all...] |
variance_impl_sse2.asm | 40 movdqa xmm3, [rax+48] 44 pmaddwd xmm3, xmm3 47 paddd xmm2, xmm3 55 movdqa xmm3,xmm4 57 paddd xmm4,xmm3 58 movdqa xmm3,xmm4 60 paddd xmm4,xmm3 110 movdqa xmm3, xmm1 115 punpckhbw xmm3, xmm [all...] |
dct_mmx.asm | 255 movdqa xmm3, [rcx + rax] 262 psllw xmm3, 3 271 punpcklwd xmm4, xmm3 ; 20 30 21 31 22 32 23 33 272 punpckhwd xmm5, xmm3 ; 24 34 25 35 26 36 27 37 284 movdqa xmm3, xmm2 ; 02 12 22 32 03 13 23 33 286 punpckhqdq xmm3, xmm1 ; 03 13 23 33 07 17 27 37 297 ; xmm3 3 303 paddw xmm0, xmm3 ; a = 0 + 3 307 psubw xmm5, xmm3 ; d = 0 - 3 330 movdqa xmm3, xmm5 ; [all...] |
fwalsh_sse2.asm | 36 movdqa xmm3, xmm4 ; 13 12 11 10 03 02 01 00 38 punpckhwd xmm3, xmm0 ; 33 13 32 12 31 11 30 10 40 punpcklwd xmm4, xmm3 ; 31 21 11 01 30 20 10 00 41 punpckhwd xmm1, xmm3 ; 33 23 13 03 32 22 12 02 44 movdqa xmm3, xmm4 ;ip[4] ip[0] 47 psubw xmm3, xmm2 ;ip[4]-ip[8] ip[0]-ip[12] aka c1 d1 50 punpcklqdq xmm4, xmm3 ;d1 a1 51 punpckhqdq xmm5, xmm3 ;c1 b1 69 movdqa xmm3, xmm5 ;ip[4] ip[0] 72 psubw xmm3, xmm2 ;ip[4]-ip[8] ip[0]-ip[12] aka c1 d [all...] |
sad_ssse3.asm | 30 lddqu xmm3, [rdi+2] 34 psadbw xmm3, xmm0 38 paddw xmm7, xmm3 43 lddqu xmm3, QWORD PTR [rdi+rdx+2] 50 psadbw xmm3, xmm0 54 paddw xmm7, xmm3 77 movdqa xmm3, [rdi+16] 79 movdqa xmm1, xmm3 82 movdqa xmm2, xmm3 85 palignr xmm3, xmm4, (%2+2 [all...] |
quantize_mmx.asm | 309 movdqa xmm3, xmm0 312 pxor xmm3, xmm0 313 psubw xmm3, xmm0 ; abs 315 movdqa xmm2, xmm3 319 movdqa xmm3, xmm1 327 paddw xmm3, xmm2 328 pmulhuw xmm3, xmm1 330 pxor xmm3, xmm0 331 psubw xmm3, xmm0 ;gain the sign back 334 movdqa xmm0, xmm3 [all...] |
encodeopt.asm | 27 movdqa xmm3, [rsi] 33 psubw xmm3, xmm4 36 pmaddwd xmm3, xmm3 39 paddd xmm3, xmm5 42 movdqa xmm0, xmm3 45 punpckhdq xmm3, xmm7 47 paddd xmm0, xmm3 48 movdqa xmm3, xmm0 51 paddd xmm0, xmm3 [all...] |
preproc_mmx.c | 113 movq xmm3, QWORD PTR [edi] local 115 movq QWORD PTR [esi+8*eax], xmm3 117 punpcklbw xmm3, xmm2 // xmm3 source pixels local 124 psubsw xmm4, xmm3 // subtracted pixel values
|
quantize_sse2.asm | 67 movdqa xmm3, OWORD PTR[rcx + 16] ;load zbin_ptr 76 paddw xmm3, xmm7 79 psubw xmm5, xmm3 ;sub (zbin_ptr + zbin_oq_value) 88 paddw xmm5, xmm3 ;add (zbin_ptr + zbin_oq_value) back 91 movdqa xmm3, OWORD PTR[rsi] 99 pmulhw xmm1, xmm3 227 movdqa xmm3, OWORD PTR[rdi + 16] 233 pmullw xmm1, xmm3
|
sad_sse3.asm | 30 lddqu xmm3, [rdi+2] 34 psadbw xmm3, xmm0 38 paddw xmm7, xmm3 43 lddqu xmm3, QWORD PTR [rdi+rdx+2] 50 psadbw xmm3, xmm0 54 paddw xmm7, xmm3 122 lddqu xmm3, [rbx] 126 psadbw xmm3, xmm0 131 paddw xmm6, xmm3 139 lddqu xmm3, QWORD PTR [rbx+rbp [all...] |
/dalvik/vm/mterp/x86-atom/ |
OP_SHR_LONG.S | 45 movq .Lvalue64, %xmm3 # %xmm3<- 64 46 psubq %xmm0, %xmm3 # %xmm3<- 64 - shift amount 48 psllq %xmm3, %xmm4 # %xmm4<- correct mask for sign bits
|
OP_SHR_LONG_2ADDR.S | 46 movq .Lvalue64, %xmm3 # %xmm3<- 64 47 psubq %xmm0, %xmm3 # %xmm3<- 64 - shift amount 49 psllq %xmm3, %xmm4 # %xmm4<- correct mask for sign bits
|
/bionic/libc/arch-x86/string/ |
ssse3-memcmp3.S | 134 movdqu (%eax), %xmm3 138 pcmpeqb %xmm0, %xmm3 139 pmovmskb %xmm3, %edx 274 movdqa 32(%esi), %xmm3 275 palignr $1,%xmm2, %xmm3 276 pcmpeqb 16(%edi), %xmm3 278 pand %xmm1, %xmm3 279 pmovmskb %xmm3, %edx 299 movdqa 32(%esi), %xmm3 300 palignr $1,16(%esi), %xmm3 [all...] |
ssse3-memcpy5.S | 294 movdqa 0x30(%eax), %xmm3 304 movdqa %xmm3, 0x30(%edx) 363 movdqa 0x30(%eax), %xmm3 373 movdqa %xmm3, 0x30(%edx) 436 movdqa 32(%eax, %edi), %xmm3 437 movdqa %xmm3, %xmm4 438 palignr $1, %xmm2, %xmm3 442 movdqa %xmm3, -16(%edx, %edi) 448 movdqa 32(%eax, %edi), %xmm3 449 movdqa %xmm3, %xmm [all...] |
ssse3-strcmp.S | 320 movdqa (%edx), %xmm3 338 palignr $1, %xmm3, %xmm2 353 movdqa %xmm4, %xmm3 362 palignr $1, %xmm3, %xmm2 377 movdqa %xmm4, %xmm3 382 pcmpeqb %xmm3, %xmm0 399 psrldq $1, %xmm3 426 movdqa (%edx), %xmm3 444 palignr $2, %xmm3, %xmm2 459 movdqa %xmm4, %xmm3 [all...] |
/external/libvpx/vp8/common/x86/ |
idctllm_sse2.asm | 63 movq xmm3, [rax+rcx] 68 punpcklbw xmm3, xmm7 77 paddw xmm3, xmm4 83 packuswb xmm3, xmm7 92 movq [rax + rdx], xmm3 127 movdqa xmm3, [rax+48] 139 pmullw xmm3, [rdx+16] 150 punpckldq xmm2, xmm3 151 punpckhdq xmm4, xmm3 154 pshufd xmm3, xmm4, 11011000 [all...] |
subpixel_sse2.asm | 59 movq xmm3, MMWORD PTR [rsi - 2] 65 por xmm1, xmm3 73 punpcklbw xmm3, xmm0 ; xx05 xx04 xx03 xx02 xx01 xx01 xx-1 xx-2 76 pmullw xmm3, XMMWORD PTR [rdx] ; x[-2] * H[-2]; Tap 1 106 paddsw xmm4, xmm3 180 movq xmm3, MMWORD PTR [rsi - 2] 190 por xmm1, xmm3 198 punpcklbw xmm3, xmm0 ; xx05 xx04 xx03 xx02 xx01 xx01 xx-1 xx-2 201 pmullw xmm3, XMMWORD PTR [rdx] ; x[-2] * H[-2]; Tap 1 230 paddsw xmm4, xmm3 [all...] |
iwalsh_sse2.asm | 36 movdqa xmm3, xmm0 ;ip[4] ip[0] 39 psubw xmm3, xmm2 ;ip[4]-ip[8] ip[0]-ip[12] aka c1 d1 42 punpcklqdq xmm0, xmm3 ;d1 a1 43 punpckhqdq xmm4, xmm3 ;c1 b1 52 ;; movdqu [rdi + 16], xmm3 59 movdqa xmm3, xmm4 ; 13 12 11 10 03 02 01 00 61 punpckhwd xmm3, xmm0 ; 33 13 32 12 31 11 30 10 63 punpcklwd xmm4, xmm3 ; 31 21 11 01 30 20 10 00 64 punpckhwd xmm1, xmm3 ; 33 23 13 03 32 22 12 02 67 movdqa xmm3, xmm4 ;ip[4] ip[0 [all...] |
subpixel_ssse3.asm | 71 ;xmm3 free 110 movdqa xmm3, XMMWORD PTR [shuf2b GLOBAL] 121 ;xmm3 free 126 pshufb xmm0, xmm3 ;[shuf2b GLOBAL] 206 movdqu xmm3, XMMWORD PTR [rsi + 6] 210 movdqa xmm1, xmm3 211 pshufb xmm3, [shuf1b GLOBAL] 213 pmaddubsw xmm3, xmm4 224 paddsw xmm3, xmm1 225 paddsw xmm3, xmm [all...] |
loopfilter_sse2.asm | 25 movq xmm3, [rdi + rcx] 26 pslldq xmm3, 8 27 por xmm1, xmm3 48 movdqa xmm3, xmm4 ; q1 50 psubusb xmm6, xmm3 ; q2-=q1 66 psubusb xmm4, xmm3 ; q0-=q1 67 psubusb xmm3, xmm0 ; q1-=q0 68 por xmm4, xmm3 ; abs(q0-q1) 84 movq xmm3, [rdi + rax] 85 pslldq xmm3, [all...] |
postproc_sse2.asm | 64 movq xmm3, QWORD PTR [rsi] ; mm4 = r0 p0..p7 65 punpcklbw xmm3, xmm0 ; mm3 = p0..p3 66 movdqa xmm1, xmm3 ; mm1 = p0..p3 67 psllw xmm3, 2 ; 71 paddusw xmm3, xmm5 ; mm3 += mm6 82 paddusw xmm3, xmm5 ; mm3 += mm5 96 paddusw xmm3, xmm5 ; mm3 += mm5 108 paddusw xmm3, xmm4 ; mm3 += mm5 119 paddusw xmm3, RD42 ; mm3 += round value 120 psraw xmm3, 3 ; mm3 /= [all...] |
recon_sse2.asm | 43 movq xmm3, MMWORD PTR [rsi+16] 44 punpcklbw xmm3, xmm0 45 paddsw xmm3, XMMWORD PTR [rdx+32] 46 packuswb xmm3, xmm0 ; pack and unpack to saturate 47 movq MMWORD PTR [rdi+rax*2], xmm3 101 movdqa xmm3, XMMWORD PTR [rsi+32] 102 movdqa xmm7, xmm3 103 punpcklbw xmm3, xmm0 105 paddsw xmm3, XMMWORD PTR [rdx+64] 107 packuswb xmm3, xmm7 ; pack and unpack to saturat [all...] |
postproc_mmx.c | 440 movq xmm3, QWORD PTR [esi] ; local 443 punpcklbw xmm3, xmm0 ; 445 movdqa xmm1, xmm3 ; 447 psllw xmm3, 2 ; 453 paddusw xmm3, xmm5 ; 472 paddusw xmm3, xmm5 ; 495 paddusw xmm3, xmm5 ; 516 paddusw xmm3, xmm4 ; 534 paddusw xmm3, rd42 ; local 536 psraw xmm3, 3 541 pandn xmm7, xmm3 ; local 574 movdqa xmm3, xmm4 local 575 psrldq xmm3, 2 local 576 punpcklbw xmm3, xmm0 ; local 666 paddusw xmm3, rd42 ; local 673 pandn xmm7, xmm3 ; local 961 movdqa xmm3, xmm6 local 962 pslld xmm3, 4 local 964 psubd xmm3, xmm6 local 981 psubd xmm3, xmm1 local 984 psubd xmm3, flimit4 local 987 psrad xmm3, 31 local 990 packssdw xmm3, xmm4 local 991 packsswb xmm3, xmm0 local 1011 pandn xmm3, xmm2 local 1223 pshufd xmm3, xmm1, 3 \/\/ 0000 8--7 8--7 8--7 squared local 1229 pshufd xmm3, xmm1, 01011111b \/\/ 0000 0000 9--6 9--6 squared local 1235 pshufd xmm3, xmm1, 10111111b \/\/ 0000 0000 8--7 8--7 squared local 1241 movdqa xmm3, xmm6 local 1242 pmaddwd xmm3, xmm3 local [all...] |
subpixel_mmx.asm | 366 paddw mm3, [rd GLOBAL] ; xmm3 += round value 367 psraw mm3, VP8_FILTER_SHIFT ; xmm3 /= 128 407 paddw mm3, [rd GLOBAL] ; xmm3 += round value 408 psraw mm3, VP8_FILTER_SHIFT ; xmm3 /= 128 424 paddw mm3, [rd GLOBAL] ; xmm3 += round value 425 psraw mm3, VP8_FILTER_SHIFT ; xmm3 /= 128 521 paddw mm3, [rd GLOBAL] ; xmm3 += round value 522 psraw mm3, VP8_FILTER_SHIFT ; xmm3 /= 128 562 paddw mm3, [rd GLOBAL] ; xmm3 += round value 563 psraw mm3, VP8_FILTER_SHIFT ; xmm3 /= 12 [all...] |
/external/openssl/crypto/ |
x86_64cpuid.pl | 185 pxor %xmm3,%xmm3 218 pxor %xmm3,%xmm3
|