HomeSort by relevance Sort by last modified time
    Searched full:xmm2 (Results 1 - 25 of 179) sorted by null

1 2 3 4 5 6 7 8

  /external/libvpx/vp8/encoder/x86/
subtract_sse2.asm 98 movdqa xmm2, xmm0
102 pxor xmm2, [GLOBAL(t80)]
103 pcmpgtb xmm1, xmm2 ; obtain sign information
105 movdqa xmm2, xmm0
108 punpckhbw xmm2, xmm3 ; put sign back to subtraction
111 movdqa XMMWORD PTR [rdi +16], xmm2
170 movq xmm2, MMWORD PTR [rsi+rdx]
172 punpcklqdq xmm0, xmm2
174 movdqa xmm2, xmm0
178 pxor xmm2, [GLOBAL(t80)
    [all...]
variance_impl_ssse3.asm 69 movdqa xmm2, xmm0
72 punpckhbw xmm2, xmm1
74 pmaddubsw xmm2, [rax]
77 paddw xmm2, [GLOBAL(xmm_bi_rd)]
79 psraw xmm2, xmm_filter_shift
81 packuswb xmm0, xmm2
93 movdqu xmm2, XMMWORD PTR [rsi+1]
96 punpcklbw xmm1, xmm2
97 punpckhbw xmm3, xmm2
107 movdqa xmm2, xmm
    [all...]
fwalsh_sse2.asm 34 movq xmm2, MMWORD PTR [rsi]
38 punpcklwd xmm2, xmm3
41 punpckldq xmm0, xmm2 ; ip[1] ip[0]
42 punpckhdq xmm1, xmm2 ; ip[3] ip[2]
44 movdqa xmm2, xmm0
46 psubw xmm2, xmm1
49 psllw xmm2, 2 ; c1 b1
52 punpcklqdq xmm0, xmm2 ; b1 a1
53 punpckhqdq xmm1, xmm2 ; c1 d1
61 movdqa xmm2, xmm
    [all...]
quantize_sse2.asm 84 movdqa xmm2, [rcx]
89 paddw xmm2, xmm7
93 psubw xmm1, xmm2
99 paddw xmm1, xmm2
102 movdqa xmm2, [rdx]
109 paddw xmm1, xmm2
170 movdqa xmm2, [rsp + qcoeff]
183 pxor xmm2, xmm0
186 psubw xmm2, xmm0
195 pmullw xmm0, xmm2
    [all...]
dct_sse2.asm 68 movq xmm2, MMWORD PTR[input+ pitch] ;13 12 11 10
73 punpcklqdq xmm0, xmm2 ;13 12 11 10 03 02 01 00
76 movdqa xmm2, xmm0
78 punpckhdq xmm2, xmm1 ;33 32 13 12 31 30 11 10
80 punpckldq xmm0, xmm2 ;31 21 30 20 11 10 01 00
82 pshufhw xmm2, xmm2, 0b1h ;32 33 12 13 xx xx xx xx
84 punpckhdq xmm1, xmm2 ;32 33 22 23 12 13 02 03
109 movdqa xmm2, xmm0
111 punpckhqdq xmm2, xmm3 ;23 22 21 20 33 32 31 3
    [all...]
sad_sse4.asm 19 movq xmm2, MMWORD PTR [rdi+16]
21 punpcklqdq xmm3, xmm2
23 movdqa xmm2, xmm1
25 mpsadbw xmm2, xmm0, 0x5
33 paddw xmm1, xmm2
40 movq xmm2, MMWORD PTR [rdi+16]
42 punpcklqdq xmm3, xmm2
44 movdqa xmm2, xmm5
46 mpsadbw xmm2, xmm0, 0x5
54 paddw xmm5, xmm2
    [all...]
variance_impl_sse2.asm 39 movdqa xmm2, [rax+32]
43 pmaddwd xmm2, xmm2
47 paddd xmm2, xmm3
49 paddd xmm4, xmm2
130 movdqu xmm2, XMMWORD PTR [rdi]
136 movdqa xmm4, xmm2
142 punpcklbw xmm2, xmm0
146 psubw xmm1, xmm2
177 movdqa xmm2, xmm
    [all...]
encodeopt.asm 220 pxor xmm2, xmm2
223 por xmm1, xmm2
248 paddd xmm2, xmm5
250 paddd xmm2, xmm3
253 movdqa xmm0, xmm2
256 punpckhdq xmm2, xmm7
257 paddd xmm0, xmm2
350 movdqa xmm2, [rdi]
352 psubw xmm1, xmm2
    [all...]
  /external/llvm/test/MC/X86/
x86-32-avx.s 3 // CHECK: vaddss %xmm4, %xmm6, %xmm2
5 vaddss %xmm4, %xmm6, %xmm2
7 // CHECK: vmulss %xmm4, %xmm6, %xmm2
9 vmulss %xmm4, %xmm6, %xmm2
11 // CHECK: vsubss %xmm4, %xmm6, %xmm2
13 vsubss %xmm4, %xmm6, %xmm2
15 // CHECK: vdivss %xmm4, %xmm6, %xmm2
17 vdivss %xmm4, %xmm6, %xmm2
19 // CHECK: vaddsd %xmm4, %xmm6, %xmm2
21 vaddsd %xmm4, %xmm6, %xmm2
    [all...]
x86_64-xop-encoding.s 27 // CHECK: vphsubbw %xmm2, %xmm1
29 vphsubbw %xmm2, %xmm1
35 // CHECK: vphaddwq %xmm6, %xmm2
37 vphaddwq %xmm6, %xmm2
59 // CHECK: vphadduwd %xmm2, %xmm1
61 vphadduwd %xmm2, %xmm1
67 // CHECK: vphaddudq %xmm6, %xmm2
69 vphaddudq %xmm6, %xmm2
83 // CHECK: vphaddubq %xmm2, %xmm2
    [all...]
x86-32-fma3.s 3 // CHECK: vfmadd132pd %xmm2, %xmm5, %xmm1
5 vfmadd132pd %xmm2, %xmm5, %xmm1
11 // CHECK: vfmadd132ps %xmm2, %xmm5, %xmm1
13 vfmadd132ps %xmm2, %xmm5, %xmm1
19 // CHECK: vfmadd213pd %xmm2, %xmm5, %xmm1
21 vfmadd213pd %xmm2, %xmm5, %xmm1
27 // CHECK: vfmadd213ps %xmm2, %xmm5, %xmm1
29 vfmadd213ps %xmm2, %xmm5, %xmm1
35 // CHECK: vfmadd231pd %xmm2, %xmm5, %xmm1
37 vfmadd231pd %xmm2, %xmm5, %xmm
    [all...]
  /external/llvm/test/MC/Disassembler/X86/
enhanced.txt 7 # CHECK: [o:xorps][w: ][2-r:%xmm1=r129][p:,][w: ][0-r:%xmm2=r130] 0:[XMM2/130]=0 1:[XMM2/130]=0 2:[XMM1/129]=0
9 # CHECK: [o:andps][w: ][2-r:%xmm1=r129][p:,][w: ][0-r:%xmm2=r130] 0:[XMM2/130]=0 1:[XMM2/130]=0 2:[XMM1/129]=0
  /external/libvpx/vp8/common/x86/
idctllm_sse2.asm 61 movq xmm2, [rax+2*rcx]
67 punpcklbw xmm2, xmm7
76 paddw xmm2, xmm4
82 packuswb xmm2, xmm7
91 movq [rax], xmm2
122 ; note the transpose of xmm1 and xmm2, necessary for shuffle
125 movdqa xmm2, [rax+16]
137 pmullw xmm2, [rdx+16]
149 movdqa xmm4, xmm2
150 punpckldq xmm2, xmm
    [all...]
subpixel_ssse3.asm 75 movq xmm2, MMWORD PTR [rsi + 3] ; 3 4 5 6 7 8 9 10
77 punpcklbw xmm0, xmm2 ; -2 3 -1 4 0 5 1 6 2 7 3 8 4 9 5 10
82 movdqa xmm2, xmm1
85 pshufb xmm2, [GLOBAL(shuf3bfrom1)]
89 pmaddubsw xmm2, xmm6
95 paddsw xmm2, xmm7
97 paddsw xmm0, xmm2
137 movdqa xmm2, xmm0
140 pshufb xmm2, xmm4
144 pmaddubsw xmm2, xmm
    [all...]
recon_sse2.asm 36 movq xmm2, MMWORD PTR [rsi+8]
37 punpcklbw xmm2, xmm0
38 paddsw xmm2, XMMWORD PTR [rdx+16]
39 packuswb xmm2, xmm0 ; pack and unpack to saturate
40 movq MMWORD PTR [rdi+rax], xmm2
91 movdqa xmm2, XMMWORD PTR [rsi+16]
92 movdqa xmm6, xmm2
93 punpcklbw xmm2, xmm0
95 paddsw xmm2, XMMWORD PTR [rdx+32]
97 packuswb xmm2, xmm6 ; pack and unpack to saturat
    [all...]
loopfilter_sse2.asm 19 movdqa xmm2, [rdi+2*rax] ; q3
25 movlps xmm2, [rsi + rcx*2] ; q3
30 movhps xmm2, [rdi + rcx*2]
45 psubusb xmm1, xmm2 ; q2-=q3
46 psubusb xmm2, xmm6 ; q3-=q2
52 por xmm1, xmm2 ; abs(q3-q2)
66 movdqa xmm2, [rsi+4*rax] ; p3
70 movlps xmm2, [rsi + rax] ; p3
74 movhps xmm2, [rdi + rax]
85 psubusb xmm4, xmm2 ; p2-=p
    [all...]
subpixel_sse2.asm 183 movq xmm2, MMWORD PTR [rsi +14]
184 pslldq xmm2, 8
186 por xmm2, xmm1
243 movdqa xmm3, xmm2
244 movdqa xmm4, xmm2
246 movdqa xmm5, xmm2
247 movdqa xmm6, xmm2
249 movdqa xmm7, xmm2
272 psrldq xmm2, 5 ; xx xx xx xx xx 0d 0c 0b 0a 09 08 07 06 05 04 03
276 punpcklbw xmm2, xmm0 ; xx0a xx09 xx08 xx07 xx06 xx05 xx04 xx0
    [all...]
postproc_sse2.asm 48 movd xmm2, dword ptr arg(6) ;flimit
49 punpcklwd xmm2, xmm2
50 punpckldq xmm2, xmm2
51 punpcklqdq xmm2, xmm2
78 pcmpgtw xmm7, xmm2
89 pcmpgtw xmm6, xmm2
103 pcmpgtw xmm6, xmm2
    [all...]
  /bionic/libc/arch-x86/string/
ssse3-strcmp-latest.S 190 movlpd (%edx), %xmm2
192 movhpd 8(%edx), %xmm2
194 pcmpeqb %xmm2, %xmm1
291 movdqa (%edx, %ecx), %xmm2
294 pcmpeqb %xmm2, %xmm1
316 movdqa (%edx), %xmm2
319 pslldq $15, %xmm2
320 pcmpeqb %xmm1, %xmm2
321 psubb %xmm0, %xmm2
322 pmovmskb %xmm2, %ed
    [all...]
ssse3-memcpy5.S 305 movdqa 0x20(%eax), %xmm2
315 movdqa %xmm2, 0x20(%edx)
374 movdqa 0x20(%eax), %xmm2
384 movdqa %xmm2, 0x20(%edx)
447 movdqa 16(%eax, %edi), %xmm2
451 palignr $1, %xmm2, %xmm3
452 palignr $1, %xmm1, %xmm2
454 movdqa %xmm2, -32(%edx, %edi)
459 movdqa 16(%eax, %edi), %xmm2
463 palignr $1, %xmm2, %xmm
    [all...]
  /frameworks/compile/libbcc/runtime/lib/i386/
floatdidf.S 27 movsd REL_ADDR(twop52), %xmm2 // 0x1.0p52
28 subsd %xmm2, %xmm1 // a_hi - 0x1p52 (no rounding occurs)
29 orpd %xmm2, %xmm0 // 0x1p52 + a_lo (no rounding occurs)
  /external/openssl/crypto/aes/asm/
vpaes-x86.pl 184 &movdqa ("xmm2",&QWP($k_ipt,$const));
189 &pshufb ("xmm2","xmm0");
192 &pxor ("xmm2","xmm5");
193 &pxor ("xmm0","xmm2");
202 &pshufb ("xmm4","xmm2"); # 4 = sb1u
208 &pshufb ("xmm5","xmm2"); # 4 = sb2u
210 &movdqa ("xmm2",&QWP($k_sb2+16,$const));# 2 : sb2t
211 &pshufb ("xmm2","xmm3"); # 2 = sb2t
212 &pxor ("xmm2","xmm5"); # 2 = 2A
217 &pxor ("xmm0","xmm2"); # 0 = 2A+
    [all...]
  /external/libyuv/files/source/
row_posix.cc 79 "movdqa %%xmm3,%%xmm2\n"
80 "palignr $0x8,%%xmm1,%%xmm2\n" // xmm2 = { xmm3[0:3] xmm1[8:15] }
81 "pshufb %%xmm6,%%xmm2\n"
82 "por %%xmm7,%%xmm2\n"
85 "movdqa %%xmm2,0x20(%1)\n"
116 "movdqa %%xmm3,%%xmm2\n"
117 "palignr $0x8,%%xmm1,%%xmm2\n" // xmm2 = { xmm3[0:3] xmm1[8:15] }
118 "pshufb %%xmm6,%%xmm2\n
    [all...]
row_win.cc 90 movdqa xmm2, [eax + 32] local
94 pmaddubsw xmm2, xmm7 local
98 phaddw xmm2, xmm3 local
100 psrlw xmm2, 7 local
101 packuswb xmm0, xmm2
123 movdqa xmm2, [eax + 32] local
127 pmaddubsw xmm2, xmm7 local
131 phaddw xmm2, xmm3 local
133 psrlw xmm2, 7 local
134 packuswb xmm0, xmm2
156 movdqa xmm2, [eax + 32] local
160 pmaddubsw xmm2, xmm7 local
164 phaddw xmm2, xmm3 local
166 psrlw xmm2, 7 local
197 movdqa xmm2, [eax + 32] local
201 pavgb xmm2, [eax + esi + 32] local
209 shufps xmm2, xmm3, 0x88 local
211 pavgb xmm2, xmm4 local
219 pmaddubsw xmm2, xmm7 local
261 movdqa xmm2, [eax + 32] local
265 pavgb xmm2, [eax + esi + 32] local
273 shufps xmm2, xmm3, 0x88 local
275 pavgb xmm2, xmm4 local
283 pmaddubsw xmm2, xmm7 local
325 movdqa xmm2, [eax + 32] local
329 pavgb xmm2, [eax + esi + 32] local
337 shufps xmm2, xmm3, 0x88 local
339 pavgb xmm2, xmm4 local
347 pmaddubsw xmm2, xmm7 local
384 movdqa xmm2, xmm3 local
385 palignr xmm2, xmm1, 8 \/\/ xmm2 = { xmm3[0:3] xmm1[8:15]} local
386 pshufb xmm2, xmm6 local
387 por xmm2, xmm7 local
423 movdqa xmm2, xmm3 local
424 palignr xmm2, xmm1, 8 \/\/ xmm2 = { xmm3[0:3] xmm1[8:15]} local
425 pshufb xmm2, xmm6 local
426 por xmm2, xmm7 local
    [all...]
  /external/llvm/test/CodeGen/X86/
illegal-vector-args-return.ll 2 ; RUN: llc < %s -march=x86 -mattr=+sse2 | grep {mulpd %xmm2, %xmm0}
4 ; RUN: llc < %s -march=x86 -mattr=+sse2 | grep {addps %xmm2, %xmm0}

Completed in 558 milliseconds

1 2 3 4 5 6 7 8