/external/libvpx/libvpx/vp8/common/x86/ |
loopfilter_sse2.asm | 32 movdqa xmm4, [rsi+rax] ; q1 38 movlps xmm4, [rsi] ; q1 43 movhps xmm4, [rdi] 50 movdqa [rsp+_q1], xmm4 ; store q1 55 movdqa xmm3, xmm4 ; q1 60 psubusb xmm4, xmm6 ; q1-=q2 63 por xmm4, xmm6 ; abs(q2-q1) 67 pmaxub xmm1, xmm4 79 movdqa xmm4, [rdi+4*rax] ; p2 83 movlps xmm4, [rsi] ; p [all...] |
loopfilter_block_sse2.asm | 203 movdqa xmm4, i4 207 LF_FILTER_HEV_MASK xmm0, xmm1, xmm2, xmm3, xmm4, xmm8, xmm9, xmm10 213 LF_FILTER xmm1, xmm2, xmm3, xmm8, xmm0, xmm4 224 movdqa xmm4, i9 227 LF_FILTER_HEV_MASK xmm3, xmm8, xmm0, xmm1, xmm2, xmm4, xmm10, xmm11, xmm9 231 movdqa xmm4, i8 233 LF_FILTER xmm0, xmm1, xmm4, xmm8, xmm3, xmm2 238 movdqa i8, xmm4 247 LF_FILTER_HEV_MASK xmm4, xmm8, xmm0, xmm1, xmm2, xmm3, xmm9, xmm11, xmm10 253 LF_FILTER xmm0, xmm1, xmm3, xmm8, xmm4, xmm [all...] |
subpixel_ssse3.asm | 61 movdqa xmm4, XMMWORD PTR [rax] ;k0_k5 81 pmaddubsw xmm0, xmm4 121 movdqa xmm4, XMMWORD PTR [GLOBAL(shuf3bfrom1)] 142 pshufb xmm2, xmm4 202 movdqa xmm4, XMMWORD PTR [rax] ;k0_k5 218 pmaddubsw xmm0, xmm4 235 pmaddubsw xmm3, xmm4 309 movdqa xmm4, XMMWORD PTR [rax] ;k0_k5 329 pmaddubsw xmm0, xmm4 458 movq xmm4, MMWORD PTR [rax + rdx * 2] ; [all...] |
mfqe_sse2.asm | 50 movdqa xmm4, [rdx] 61 movdqa xmm5, xmm4 62 punpcklbw xmm4, xmm6 64 pmullw xmm4, xmm1 68 paddw xmm2, xmm4 123 pxor xmm4, xmm4 131 punpcklbw xmm2, xmm4 135 punpcklbw xmm3, xmm4 143 packuswb xmm2, xmm4 [all...] |
variance_impl_ssse3.asm | 122 pxor xmm4, xmm4 123 punpcklbw xmm1, xmm4 125 punpcklbw xmm5, xmm4 188 pxor xmm4, xmm4 189 punpcklbw xmm3, xmm4 191 punpcklbw xmm5, xmm4 232 movq xmm4, QWORD PTR [rdi+8] 233 punpcklbw xmm4, xmm [all...] |
sad_ssse3.asm | 58 movdqa xmm4, XMMWORD PTR [rdi] 62 palignr xmm5, xmm4, %2 65 palignr xmm6, xmm4, (%2+1) 67 palignr xmm7, xmm4, (%2+2) 74 movdqa xmm4, XMMWORD PTR [rdi] 78 palignr xmm1, xmm4, %2 81 palignr xmm2, xmm4, (%2+1) 83 palignr xmm3, xmm4, (%2+2) 94 movdqa xmm4, XMMWORD PTR [rdi+rdx] 98 palignr xmm1, xmm4, % [all...] |
/external/libvpx/libvpx/vp9/common/x86/ |
vp9_postproc_sse2.asm | 106 movq xmm4, QWORD PTR [rsi+rax] ; mm4 = r-1 p0..p7 107 punpcklbw xmm4, xmm0 ; mm4 = r-1 p0..p3 108 paddusw xmm3, xmm4 ; mm3 += mm5 112 psubusw xmm6, xmm4 ; mm6 = p0..p3 - r-2 p0..p3 113 psubusw xmm4, xmm1 ; mm5 = r-1 p0..p3 - p0..p3 114 paddusw xmm6, xmm4 ; mm6 = abs(r0 p0..p3 - r-1 p0..p3) 147 movd xmm4, DWORD PTR [rdi +rdx +6] 149 pslldq xmm4, 8 150 por xmm4, xmm7 152 movdqa xmm3, xmm4 [all...] |
vp9_subpixel_8t_sse2.asm | 27 pshuflw xmm4, xmm7, 0b ;k4 34 punpcklqdq xmm5, xmm4 54 punpckldq xmm5, xmm4 99 pshufhw xmm4, xmm7, 0b ;k4 108 punpckhwd xmm4, xmm4 117 movdqa k4, xmm4 138 movq xmm4, [rsi + rdx + %1] ;4 150 punpcklbw xmm4, zero 159 pmullw xmm4, k [all...] |
/external/chromium_org/third_party/openssl/openssl/crypto/aes/asm/ |
bsaes-x86_64.S | 23 pxor %xmm8,%xmm4 38 pxor %xmm4,%xmm3 43 pxor %xmm3,%xmm4 62 movdqa %xmm4,%xmm9 63 psrlq $2,%xmm4 66 pxor %xmm6,%xmm4 68 pand %xmm8,%xmm4 70 pxor %xmm4,%xmm6 71 psllq $2,%xmm4 74 pxor %xmm9,%xmm4 [all...] |
aesni-x86_64.S | 52 xorps %xmm0,%xmm4 85 xorps %xmm0,%xmm4 118 xorps %xmm0,%xmm4 156 xorps %xmm0,%xmm4 195 pxor %xmm0,%xmm4 252 pxor %xmm0,%xmm4 309 pxor %xmm0,%xmm4 379 pxor %xmm0,%xmm4 458 movdqu 32(%rdi),%xmm4 475 movups %xmm4,32(%rsi [all...] |
/external/openssl/crypto/aes/asm/ |
bsaes-x86_64.S | 23 pxor %xmm8,%xmm4 38 pxor %xmm4,%xmm3 43 pxor %xmm3,%xmm4 62 movdqa %xmm4,%xmm9 63 psrlq $2,%xmm4 66 pxor %xmm6,%xmm4 68 pand %xmm8,%xmm4 70 pxor %xmm4,%xmm6 71 psllq $2,%xmm4 74 pxor %xmm9,%xmm4 [all...] |
aesni-x86_64.S | 52 xorps %xmm0,%xmm4 85 xorps %xmm0,%xmm4 118 xorps %xmm0,%xmm4 156 xorps %xmm0,%xmm4 195 pxor %xmm0,%xmm4 252 pxor %xmm0,%xmm4 309 pxor %xmm0,%xmm4 379 pxor %xmm0,%xmm4 458 movdqu 32(%rdi),%xmm4 475 movups %xmm4,32(%rsi [all...] |
/external/chromium_org/third_party/yasm/source/patched-yasm/modules/arch/x86/tests/ |
sse5-all.asm | 4 compd xmm1, xmm4, xmm7, 5 ; 0F 25 2D 347 10 05 8 comps xmm1, xmm4, xmm7, 5 ; 0F 25 2C 347 10 05 12 comsd xmm1, xmm4, xmm7, 5 ; 0F 25 2F 347 10 05 16 comss xmm1, xmm4, xmm7, 5 ; 0F 25 2E 347 10 05 20 cvtph2ps xmm1, xmm4 ; 0F 7A 30 314 24 cvtps2ph xmm1, xmm4 ; 0F 7A 31 341 231 pcomb xmm1, xmm4, xmm7, 5 ; 0F 25 4C 347 10 05 235 pcomd xmm1, xmm4, xmm7, 5 ; 0F 25 4E 347 10 05 239 pcomq xmm1, xmm4, xmm7, 5 ; 0F 25 4F 347 10 05 243 pcomub xmm1, xmm4, xmm7, 5 ; 0F 25 6C 347 10 0 [all...] |
/external/libvpx/libvpx/vp9/encoder/x86/ |
vp9_subpel_variance_impl_sse2.asm | 58 movdqa xmm4, xmm5 60 punpckhbw xmm4, xmm0 68 psubw xmm4, xmm3 71 paddw xmm6, xmm4 73 pmaddwd xmm4, xmm4 75 paddd xmm7, xmm4 166 movdqa xmm4, xmm5 168 punpckhbw xmm4, xmm0 175 psubw xmm4, xmm [all...] |
vp9_temporal_filter_apply_sse2.asm | 53 movq xmm4, rdx ; can't use rdx w/ shift 55 psrlw xmm5, xmm4 131 movdqa xmm4, [rax] 134 paddw xmm4, xmm2 137 movdqa [rax], xmm4 162 movdqa xmm4, [rdi] 167 paddd xmm4, xmm0 172 movdqa [rdi], xmm4
|
vp9_sad_ssse3.asm | 58 movdqa xmm4, XMMWORD PTR [rdi] 62 palignr xmm5, xmm4, %2 65 palignr xmm6, xmm4, (%2+1) 67 palignr xmm7, xmm4, (%2+2) 74 movdqa xmm4, XMMWORD PTR [rdi] 78 palignr xmm1, xmm4, %2 81 palignr xmm2, xmm4, (%2+1) 83 palignr xmm3, xmm4, (%2+2) 94 movdqa xmm4, XMMWORD PTR [rdi+rdx] 98 palignr xmm1, xmm4, % [all...] |
/external/libvpx/libvpx/vp8/encoder/x86/ |
temporal_filter_apply_sse2.asm | 53 movq xmm4, rdx ; can't use rdx w/ shift 55 psrlw xmm5, xmm4 131 movdqa xmm4, [rax] 134 paddw xmm4, xmm2 137 movdqa [rax], xmm4 162 movdqa xmm4, [rdi] 167 paddd xmm4, xmm0 172 movdqa [rdi], xmm4
|
subtract_sse2.asm | 90 movdqa xmm4, [GLOBAL(t80)] 102 pxor xmm1, xmm4 ;convert to signed values 103 pxor xmm2, xmm4 122 pxor xmm5, xmm4 ;convert to signed values 123 pxor xmm1, xmm4 159 movdqa xmm4, [GLOBAL(t80)] 184 pxor xmm1, xmm4 ;convert to signed values 185 pxor xmm2, xmm4 218 pxor xmm1, xmm4 ;convert to signed values 219 pxor xmm2, xmm4 [all...] |
/bionic/libc/arch-x86/string/ |
ssse3-memcpy-atom.S | 292 movdqa 0x40(%eax), %xmm4 302 movdqa %xmm4, 0x40(%edx) 360 movdqa 0x40(%eax), %xmm4 370 movdqa %xmm4, 0x40(%edx) 451 movaps 47(%eax), %xmm4 454 palignr $1, %xmm4, %xmm5 455 palignr $1, %xmm3, %xmm4 460 movaps %xmm4, 32(%edx) 496 movdqa %xmm3, %xmm4 509 palignr $1, %xmm4, %xmm [all...] |
ssse3-strcmp-atom.S | 351 movdqa %xmm2, %xmm4 368 movdqa %xmm4, %xmm3 375 movdqa %xmm2, %xmm4 392 movdqa %xmm4, %xmm3 457 movdqa %xmm2, %xmm4 474 movdqa %xmm4, %xmm3 481 movdqa %xmm2, %xmm4 498 movdqa %xmm4, %xmm3 564 movdqa %xmm2, %xmm4 581 movdqa %xmm4, %xmm [all...] |
ssse3-wcscpy-atom.S | 160 movaps 16(%ecx, %esi), %xmm4 162 pcmpeqd %xmm4, %xmm0 170 movaps %xmm4, (%edx, %esi) 208 movaps %xmm2, %xmm4 222 movaps %xmm4, -64(%edx) 229 pcmpeqd %xmm4, %xmm0 236 movaps %xmm4, -64(%edx) 326 movaps 44(%ecx), %xmm4 327 movaps %xmm4, %xmm7 335 palignr $4, %xmm4, %xmm [all...] |
ssse3-strcpy-atom.S | 295 movaps 16(%ecx, %esi), %xmm4 297 pcmpeqb %xmm4, %xmm0 308 movaps %xmm4, (%edx, %esi) 357 movaps %xmm2, %xmm4 374 movaps %xmm4, -64(%edx) 384 pcmpeqb %xmm4, %xmm0 394 movaps %xmm4, -64(%edx) 501 movaps 47(%ecx), %xmm4 502 movaps %xmm4, %xmm7 510 palignr $1, %xmm4, %xmm [all...] |
/external/llvm/test/MC/X86/ |
x86-32-avx.s | 3 // CHECK: vaddss %xmm4, %xmm6, %xmm2 5 vaddss %xmm4, %xmm6, %xmm2 7 // CHECK: vmulss %xmm4, %xmm6, %xmm2 9 vmulss %xmm4, %xmm6, %xmm2 11 // CHECK: vsubss %xmm4, %xmm6, %xmm2 13 vsubss %xmm4, %xmm6, %xmm2 15 // CHECK: vdivss %xmm4, %xmm6, %xmm2 17 vdivss %xmm4, %xmm6, %xmm2 19 // CHECK: vaddsd %xmm4, %xmm6, %xmm2 21 vaddsd %xmm4, %xmm6, %xmm [all...] |
/external/libyuv/files/source/ |
rotate.cc | 92 movq xmm4, qword ptr [eax] local 95 punpcklbw xmm4, xmm5 local 97 movdqa xmm5, xmm4 112 punpcklwd xmm4, xmm6 local 114 movdqa xmm6, xmm4 120 punpckldq xmm0, xmm4 122 movdqa xmm4, xmm0 local 123 palignr xmm4, xmm4, 8 local 124 movq qword ptr [edx + esi], xmm4 195 movdqa xmm4, [eax] local 199 punpcklbw xmm4, xmm5 local 223 punpcklwd xmm4, xmm6 local 237 movdqa xmm4, xmm6 local [all...] |
/packages/apps/Camera/jni/feature_stab/db_vlvm/ |
db_feature_detection.cpp | 536 cvtpi2ps xmm4,[eax+24] /*19 Convert two integers into floating point of low double-word*/ local 539 movlhps xmm4,xmm4 /*20 Move them to the high double-word*/ local 540 cvtpi2ps xmm4,[eax+16] /*21 Convert two integers into floating point of low double-word*/ local 556 movaps xmm3,xmm4 /*28 Copy Cxx*/ 557 mulps xmm4,xmm6 /*29 Multiply to get Gxx*Gyy*/ local 565 subps xmm4,xmm5 /*33 Subtract to get Gxx*Gyy-Gxy*Gxy*/ local 573 subps xmm4,xmm6 /*35 Subtract to get Gxx*Gyy-Gxy*Gxy-k*(Gxx+Gyy)*(Gxx+Gyy)*/ local 577 movaps [edx+16],xmm4 /*36 Store*/ 760 movaps xmm4,[eax+64 local 770 maxps xmm4,[eax+192] local 780 maxps xmm4,[eax+320] local 790 maxps xmm4,[eax+448] local 798 maxps xmm4,xmm5 local 801 maxps xmm4,xmm6 local 838 movaps xmm4,[eax+64] local 848 maxps xmm4,[eax+192] local 856 maxps xmm4,xmm5 local 859 maxps xmm4,xmm6 local 896 movaps xmm4,[eax+64] local 904 maxps xmm4,xmm5 local 907 maxps xmm4,xmm6 local 1090 movaps xmm4,[eax+64] local 1098 maxps xmm4,[ebx+64] local 1116 movaps xmm4,[eax+192] local 1124 maxps xmm4,[ebx+192] local 1142 movaps xmm4,[eax+320] local 1150 maxps xmm4,[ebx+320] local 1168 movaps xmm4,[eax+448] local 1176 maxps xmm4,[ebx+448] local 1216 movups xmm4,[eax+64] local 1224 maxps xmm4,[ebx+64] local 1242 movups xmm4,[eax+192] local 1250 maxps xmm4,[ebx+192] local 1268 movups xmm4,[eax+320] local 1276 maxps xmm4,[ebx+320] local 1294 movups xmm4,[eax+448] local 1302 maxps xmm4,[ebx+448] local [all...] |