HomeSort by relevance Sort by last modified time
    Searched refs:xmm4 (Results 1 - 25 of 32) sorted by null

1 2

  /external/libvpx/vp8/common/x86/
idctllm_sse2.asm 38 movd xmm4, [rax]
41 pinsrw xmm4, [rax+32], 4
44 pmullw xmm4, xmm5
50 pshuflw xmm4, xmm4, 00000000b
51 pshufhw xmm4, xmm4, 00000000b
54 paddw xmm4, [fours GLOBAL]
57 psraw xmm4, 3
74 paddw xmm0, xmm4
    [all...]
iwalsh_sse2.asm 41 movdqa xmm4, xmm0
43 punpckhqdq xmm4, xmm3 ;c1 b1
46 movdqa xmm1, xmm4 ;c1 b1
47 paddw xmm4, xmm0 ;dl+cl a1+b1 aka op[4] op[0]
51 ;; movdqu [rdi + 0], xmm4
59 movdqa xmm3, xmm4 ; 13 12 11 10 03 02 01 00
60 punpcklwd xmm4, xmm0 ; 23 03 22 02 21 01 20 00
62 movdqa xmm1, xmm4 ; 23 03 22 02 21 01 20 00
63 punpcklwd xmm4, xmm3 ; 31 21 11 01 30 20 10 00
67 movdqa xmm3, xmm4 ;ip[4] ip[0
    [all...]
subpixel_sse2.asm 67 movdqa xmm4, xmm1
74 psrldq xmm4, 1 ; xx 0d 0c 0b 0a 09 08 07 06 05 04 03 02 01 00 -1
77 punpcklbw xmm4, xmm0 ; xx06 xx05 xx04 xx03 xx02 xx01 xx00 xx-1
80 pmullw xmm4, XMMWORD PTR [rdx+16] ; x[-1] * H[-1]; Tap 2
103 paddsw xmm4, xmm7
104 paddsw xmm4, xmm5
106 paddsw xmm4, xmm3
107 paddsw xmm4, xmm6
109 paddsw xmm4, xmm1
110 paddsw xmm4, [rd GLOBAL
    [all...]
loopfilter_sse2.asm 39 movdqa xmm4, [rsi+rax] ; q1
42 movq xmm4, [rdi]
43 pslldq xmm4, 8
44 por xmm4, xmm0
45 movdqa XMMWORD PTR [rsp + 16], xmm4 ; store q1
48 movdqa xmm3, xmm4 ; q1
49 psubusb xmm4, xmm6 ; q1-=q2
51 por xmm4, xmm6 ; abs(q2-q1)
52 psubusb xmm4, xmm7
54 por xmm1, xmm4
    [all...]
postproc_sse2.asm 106 movq xmm4, QWORD PTR [rsi+rax] ; mm4 = r-1 p0..p7
107 punpcklbw xmm4, xmm0 ; mm4 = r-1 p0..p3
108 paddusw xmm3, xmm4 ; mm3 += mm5
112 psubusw xmm6, xmm4 ; mm6 = p0..p3 - r-2 p0..p3
113 psubusw xmm4, xmm1 ; mm5 = r-1 p0..p3 - p0..p3
114 paddusw xmm6, xmm4 ; mm6 = abs(r0 p0..p3 - r-1 p0..p3)
147 movd xmm4, DWORD PTR [rdi +rdx +6]
149 pslldq xmm4, 8
150 por xmm4, xmm7
152 movdqa xmm3, xmm4
    [all...]
subpixel_ssse3.asm 60 movdqa xmm4, XMMWORD PTR [rax] ;k0_k5
80 pmaddubsw xmm0, xmm4
111 movdqa xmm4, XMMWORD PTR [shuf3b GLOBAL]
127 pshufb xmm2, xmm4 ;[shuf3b GLOBAL]
187 movdqa xmm4, XMMWORD PTR [rax] ;k0_k5
201 pmaddubsw xmm0, xmm4
213 pmaddubsw xmm3, xmm4
332 movdqa xmm4, XMMWORD PTR [rax] ;k0_k5
352 pmaddubsw xmm0, xmm4
478 movq xmm4, MMWORD PTR [rax + rdx * 2] ;
    [all...]
recon_sse2.asm 50 movq xmm4, MMWORD PTR [rsi+24]
51 punpcklbw xmm4, xmm0
52 paddsw xmm4, XMMWORD PTR [rdx+48]
53 packuswb xmm4, xmm0 ; pack and unpack to saturate
54 movq MMWORD PTR [rdi+rax*2], xmm4
111 movdqa xmm4, XMMWORD PTR [rsi+48]
112 movdqa xmm5, xmm4
113 punpcklbw xmm4, xmm0
115 paddsw xmm4, XMMWORD PTR [rdx+96]
117 packuswb xmm4, xmm5 ; pack and unpack to saturat
    [all...]
postproc_mmx.c 512 movq xmm4, QWORD PTR [esi+eax] ; local
514 punpcklbw xmm4, xmm0 ;
516 paddusw xmm3, xmm4 ;
523 psubusw xmm6, xmm4 ;
525 psubusw xmm4, xmm1 ;
527 paddusw xmm6, xmm4 ;
569 movd xmm4, DWORD PTR [edi +edx +6] local
571 pslldq xmm4, 8 local
572 por xmm4, xmm7 local
574 movdqa xmm3, xmm4
624 movdqa xmm5, xmm4 ; local
645 psrldq xmm4, 1 ; local
943 movdqa xmm4, xmm2 local
946 punpckhwd xmm4, xmm0 local
967 movdqa xmm4, xmm5 local
970 pmulhw xmm4, xmm4 local
976 movdqa xmm4, xmm7 local
977 pslld xmm4, 4 local
979 psubd xmm4, xmm7 local
982 psubd xmm4, xmm2 local
985 psubd xmm4, flimit4 local
988 psrad xmm4, 31 local
1002 movdqu xmm4, vp8_rv[ecx*2] local
1224 pshufd xmm4, xmm2, 3 \/\/ 0000 8--7 8--7 8--7 squared local
1230 pshufd xmm4, xmm2, 01011111b \/\/ 0000 0000 9--6 9--6 squared local
1236 pshufd xmm4, xmm2, 10111111b \/\/ 0000 0000 8--7 8--7 squared local
    [all...]
  /dalvik/vm/mterp/x86-atom/
OP_SHR_LONG.S 47 movq .L64bits, %xmm4 # %xmm4<- lower 64 bits set
48 psllq %xmm3, %xmm4 # %xmm4<- correct mask for sign bits
49 por %xmm4, %xmm1 # %xmm1<- signed and shifted vBB
OP_SHR_LONG_2ADDR.S 48 movq .L64bits, %xmm4 # %xmm4<- lower 64 bits set
49 psllq %xmm3, %xmm4 # %xmm4<- correct mask for sign bits
50 por %xmm4, %xmm1 # %xmm1<- signed and shifted vBB
  /external/libvpx/vp8/encoder/x86/
fwalsh_sse2.asm 27 movdqu xmm4, [rsi + 0] ;ip[4] ip[0]
36 movdqa xmm3, xmm4 ; 13 12 11 10 03 02 01 00
37 punpcklwd xmm4, xmm0 ; 23 03 22 02 21 01 20 00
39 movdqa xmm1, xmm4 ; 23 03 22 02 21 01 20 00
40 punpcklwd xmm4, xmm3 ; 31 21 11 01 30 20 10 00
44 movdqa xmm3, xmm4 ;ip[4] ip[0]
46 paddw xmm4, xmm2 ;ip[4]+ip[8] ip[0]+ip[12] aka b1 a1
49 movdqa xmm5, xmm4
50 punpcklqdq xmm4, xmm3 ;d1 a1
54 paddw xmm5, xmm4 ;dl+cl a1+b1 aka op[4] op[0
    [all...]
dct_mmx.asm 254 movdqa xmm4, [rcx]
261 psllw xmm4, 3
266 movdqa xmm5, xmm4 ; 20 21 22 23 24 25 26 27
271 punpcklwd xmm4, xmm3 ; 20 30 21 31 22 32 23 33
275 punpckldq xmm0, xmm4 ; 00 10 20 30 01 11 21 31
277 punpckhdq xmm2, xmm4 ; 02 12 22 32 03 13 23 33
280 movdqa xmm4, xmm1 ; 04 14 05 15 06 16 07 17
281 punpckldq xmm4, xmm5 ; 04 14 24 34 05 15 25 35
290 punpcklqdq xmm0, xmm4 ; 00 10 20 30 04 14 24 34
292 punpckhqdq xmm1, xmm4 ; 01 11 21 32 05 15 25 3
    [all...]
preproc_mmx.c 121 movq xmm4, QWORD PTR [esi] // get frame buffer values local
122 punpcklbw xmm4, xmm7 // xmm4 frame buffer pixels local
123 movdqa xmm6, xmm4 // save the pixel values
124 psubsw xmm4, xmm3 // subtracted pixel values local
125 pmullw xmm4, xmm4 // square xmm4 local
127 psrlw xmm4, xmm5 // should be strength local
128 pmullw xmm4, threes // 3 * modifie local
131 movdqa xmm4, xmm5 \/\/ save the modifiers local
132 pmullw xmm4, xmm6 \/\/ multiplier values local
    [all...]
dct_sse2.asm 57 movdqa xmm4, xmm3
59 pmaddwd xmm4, XMMWORD PTR[_2217_neg5352 GLOBAL] ;d1*2217 - c1*5352
62 paddd xmm4, XMMWORD PTR[_7500 GLOBAL]
64 psrad xmm4, 12 ;(d1 * 2217 - c1 * 5352 + 7500)>>12
67 packssdw xmm3, xmm4 ;op[3] op[1]
100 pxor xmm4, xmm4 ;zero out for compare
103 pcmpeqw xmm2, xmm4
109 movdqa xmm4, xmm3
111 pmaddwd xmm4, XMMWORD PTR[_2217_neg5352 GLOBAL] ;d1*2217 - c1*535
    [all...]
sad_ssse3.asm 60 movdqa xmm4, [rdi]
64 palignr xmm5, xmm4, %2
67 palignr xmm6, xmm4, (%2+1)
69 palignr xmm7, xmm4, (%2+2)
76 movdqa xmm4, [rdi]
80 palignr xmm1, xmm4, %2
83 palignr xmm2, xmm4, (%2+1)
85 palignr xmm3, xmm4, (%2+2)
96 movdqa xmm4, QWORD PTR [rdi+rdx]
100 palignr xmm1, xmm4, %
    [all...]
quantize_mmx.asm 347 movdqa xmm4, [rsi+16]
352 movdqa xmm7, xmm4
353 psraw xmm4, 15
355 pxor xmm7, xmm4
356 psubw xmm7, xmm4 ; abs
371 pxor xmm7, xmm4
372 psubw xmm7, xmm4;gain the sign back
quantize_sse2.asm 55 movdqa xmm4, OWORD PTR[rdx + 16]
58 movdqa xmm5, xmm4
61 psraw xmm4, 15 ;sign of z (aka sz)
64 pxor xmm5, xmm4
73 psubw xmm5, xmm4 ;x = abs(z)
106 pxor xmm5, xmm4
109 psubw xmm5, xmm4
variance_impl_sse2.asm 34 pxor xmm4, xmm4
48 paddd xmm4, xmm0
49 paddd xmm4, xmm2
55 movdqa xmm3,xmm4
56 psrldq xmm4,8
57 paddd xmm4,xmm3
58 movdqa xmm3,xmm4
59 psrldq xmm4,4
60 paddd xmm4,xmm
    [all...]
encodeopt.asm 29 movdqa xmm4, [rdi]
33 psubw xmm3, xmm4
230 movdqa xmm4, [rdi]
239 psubw xmm3, xmm4
sad_sse2.asm 49 movq xmm4, QWORD PTR [rsi+rax]
63 punpcklbw xmm4, xmm6
66 psadbw xmm4, xmm5
69 paddw xmm7, xmm4
sad_sse3.asm 109 lddqu xmm4, [rcx]
114 psadbw xmm4, xmm0
128 paddw xmm4, xmm1
145 paddw xmm4, xmm1
619 movq xmm0, xmm4
620 psrldq xmm4, 8
622 paddw xmm0, xmm4
687 movq xmm0, xmm4
688 psrldq xmm4, 8
690 paddw xmm0, xmm4
    [all...]
  /bionic/libc/arch-x86/string/
ssse3-strcmp.S 336 movdqa %xmm2, %xmm4
353 movdqa %xmm4, %xmm3
360 movdqa %xmm2, %xmm4
377 movdqa %xmm4, %xmm3
442 movdqa %xmm2, %xmm4
459 movdqa %xmm4, %xmm3
466 movdqa %xmm2, %xmm4
483 movdqa %xmm4, %xmm3
549 movdqa %xmm2, %xmm4
566 movdqa %xmm4, %xmm
    [all...]
ssse3-memcpy5.S 295 movdqa 0x40(%eax), %xmm4
305 movdqa %xmm4, 0x40(%edx)
364 movdqa 0x40(%eax), %xmm4
374 movdqa %xmm4, 0x40(%edx)
437 movdqa %xmm3, %xmm4
451 palignr $1, %xmm4, %xmm2
480 movdqa %xmm3, %xmm4
494 palignr $2, %xmm4, %xmm2
523 movdqa %xmm3, %xmm4
537 palignr $3, %xmm4, %xmm
    [all...]
  /external/openssl/crypto/
x86_64cpuid.pl 186 pxor %xmm4,%xmm4
219 pxor %xmm4,%xmm4
  /external/libffi/src/x86/
unix64.S 198 movdqa 112(%r10), %xmm4
322 movdqa %xmm4, 112(%rsp)

Completed in 1528 milliseconds

1 2