HomeSort by relevance Sort by last modified time
    Searched refs:movdqa (Results 1 - 25 of 84) sorted by null

1 2 3 4

  /external/libvpx/libvpx/vp8/common/x86/
loopfilter_block_sse2.asm 18 movdqa scratch1, %2 ; v2
31 movdqa scratch2, %3 ; save p1
68 movdqa scratch2, %6 ; save hev
72 movdqa scratch1, %1
78 movdqa scratch1, %3
85 movdqa scratch2, %5
90 movdqa scratch1, zero
101 movdqa scratch1, zero
113 movdqa scratch1, zero
198 movdqa xmm0, i
    [all...]
idctllm_sse2.asm 128 movdqa xmm0, [rax]
129 movdqa xmm2, [rax+16]
130 movdqa xmm1, [rax+32]
131 movdqa xmm3, [rax+48]
134 movdqa [rax], xmm7
135 movdqa [rax+16], xmm7
136 movdqa [rax+32], xmm7
137 movdqa [rax+48], xmm7
147 movdqa xmm4, xmm0
154 movdqa xmm4, xmm
    [all...]
loopfilter_sse2.asm 30 movdqa xmm2, [rdi+2*rax] ; q3
31 movdqa xmm1, [rsi+2*rax] ; q2
32 movdqa xmm4, [rsi+rax] ; q1
33 movdqa xmm5, [rsi] ; q0
49 movdqa [rsp+_q2], xmm1 ; store q2
50 movdqa [rsp+_q1], xmm4 ; store q1
52 movdqa xmm7, [rdx] ;limit
54 movdqa xmm6, xmm1 ; q2
55 movdqa xmm3, xmm4 ; q1
66 movdqa xmm0, xmm5 ; q
    [all...]
subpixel_sse2.asm 68 movdqa xmm4, xmm1
69 movdqa xmm5, xmm1
71 movdqa xmm6, xmm1
72 movdqa xmm7, xmm1
118 movdqa XMMWORD Ptr [rdi], xmm4
193 movdqa xmm4, xmm1
194 movdqa xmm5, xmm1
196 movdqa xmm6, xmm1
197 movdqa xmm7, xmm1
242 movdqa XMMWORD Ptr [rdi], xmm
    [all...]
  /external/openssl/crypto/aes/asm/
vpaes-x86.pl 161 &movdqa ("xmm7",&QWP($k_inv,$const));
162 &movdqa ("xmm6",&QWP($k_s0F,$const));
183 &movdqa ("xmm1","xmm6")
184 &movdqa ("xmm2",&QWP($k_ipt,$const));
190 &movdqa ("xmm0",&QWP($k_ipt+16,$const));
201 &movdqa ("xmm4",&QWP($k_sb1,$const)); # 4 : sb1u
204 &movdqa ("xmm0",&QWP($k_sb1+16,$const));# 0 : sb1t
207 &movdqa ("xmm5",&QWP($k_sb2,$const)); # 4 : sb2u
209 &movdqa ("xmm1",&QWP(-0x40,$base,$magic));# .Lk_mc_forward[]
210 &movdqa ("xmm2",&QWP($k_sb2+16,$const));# 2 : sb2
    [all...]
vpaes-x86.S 64 movdqa -48(%ebp),%xmm7
65 movdqa -16(%ebp),%xmm6
73 movdqa %xmm6,%xmm1
74 movdqa (%ebp),%xmm2
80 movdqa 16(%ebp),%xmm0
89 movdqa 32(%ebp),%xmm4
92 movdqa 48(%ebp),%xmm0
95 movdqa 64(%ebp),%xmm5
97 movdqa -64(%ebx,%ecx,1),%xmm1
98 movdqa 80(%ebp),%xmm
    [all...]
vpaes-x86_64.S 24 movdqa %xmm9,%xmm1
25 movdqa .Lk_ipt(%rip),%xmm2
31 movdqa .Lk_ipt+16(%rip),%xmm0
42 movdqa %xmm13,%xmm4
45 movdqa %xmm12,%xmm0
48 movdqa %xmm15,%xmm5
50 movdqa -64(%r11,%r10,1),%xmm1
51 movdqa %xmm14,%xmm2
54 movdqa (%r11,%r10,1),%xmm4
55 movdqa %xmm0,%xmm
    [all...]
bsaes-x86_64.S 11 movdqa (%rax),%xmm8
13 movdqa 80(%r11),%xmm7
31 movdqa 0(%r11),%xmm7
32 movdqa 16(%r11),%xmm8
33 movdqa %xmm5,%xmm9
35 movdqa %xmm3,%xmm10
47 movdqa %xmm1,%xmm9
49 movdqa %xmm15,%xmm10
61 movdqa 32(%r11),%xmm7
62 movdqa %xmm4,%xmm
    [all...]
vpaes-x86_64.pl 88 movdqa %xmm9, %xmm1
89 movdqa .Lk_ipt(%rip), %xmm2 # iptlo
95 movdqa .Lk_ipt+16(%rip), %xmm0 # ipthi
106 movdqa %xmm13, %xmm4 # 4 : sb1u
109 movdqa %xmm12, %xmm0 # 0 : sb1t
112 movdqa %xmm15, %xmm5 # 4 : sb2u
114 movdqa -0x40(%r11,%r10), %xmm1 # .Lk_mc_forward[]
115 movdqa %xmm14, %xmm2 # 2 : sb2t
118 movdqa (%r11,%r10), %xmm4 # .Lk_mc_backward[]
119 movdqa %xmm0, %xmm3 # 3 =
    [all...]
  /bionic/libc/arch-x86/string/
sse2-memset5-atom.S 350 movdqa %xmm0, (%edx)
351 movdqa %xmm0, 0x10(%edx)
352 movdqa %xmm0, 0x20(%edx)
353 movdqa %xmm0, 0x30(%edx)
354 movdqa %xmm0, 0x40(%edx)
355 movdqa %xmm0, 0x50(%edx)
356 movdqa %xmm0, 0x60(%edx)
357 movdqa %xmm0, 0x70(%edx)
363 movdqa %xmm0, (%edx)
364 movdqa %xmm0, 0x10(%edx
    [all...]
ssse3-memcpy5.S 225 movdqa (%eax, %edi), %xmm0
226 movdqa 16(%eax, %edi), %xmm1
228 movdqa %xmm0, (%edx, %edi)
229 movdqa %xmm1, 16(%edx, %edi)
233 movdqa (%eax, %edi), %xmm0
234 movdqa 16(%eax, %edi), %xmm1
236 movdqa %xmm0, (%edx, %edi)
237 movdqa %xmm1, 16(%edx, %edi)
241 movdqa (%eax, %edi), %xmm0
242 movdqa 16(%eax, %edi), %xmm
    [all...]
ssse3-strcmp-latest.S 273 movdqa (%eax), %xmm1
290 movdqa (%eax, %ecx), %xmm1
291 movdqa (%edx, %ecx), %xmm2
316 movdqa (%edx), %xmm2
317 movdqa (%eax), %xmm1
331 movdqa (%edx), %xmm3
345 movdqa (%eax, %ecx), %xmm1
346 movdqa (%edx, %ecx), %xmm2
347 movdqa %xmm2, %xmm4
364 movdqa %xmm4, %xmm
    [all...]
ssse3-memcmp3-new.S 227 movdqa (%esi), %xmm0
231 movdqa 16(%esi), %xmm2
237 movdqa %xmm0, %xmm1
238 movdqa 32(%esi), %xmm0
239 movdqa 48(%esi), %xmm2
257 movdqa %xmm0, %xmm1
277 movdqa 16(%esi), %xmm1
278 movdqa %xmm1, %xmm2
282 movdqa 32(%esi), %xmm3
303 movdqa 16(%esi), %xmm
    [all...]
  /external/libvpx/libvpx/vp8/encoder/x86/
quantize_ssse3.asm 55 movdqa xmm0, [rax]
56 movdqa xmm4, [rax + 16]
59 movdqa xmm2, [rcx]
60 movdqa xmm3, [rcx + 16]
62 movdqa xmm1, xmm0
63 movdqa xmm5, xmm4
83 movdqa xmm2, xmm1 ;store y for getting eob
84 movdqa xmm3, xmm5
91 movdqa [rax], xmm1
92 movdqa [rax + 16], xmm
    [all...]
temporal_filter_apply_sse2.asm 47 movdqa [rsp + strength], xmm6 ; where strength is used, all 16 bytes are read
54 movdqa xmm5, [GLOBAL(_const_top_bit)]
56 movdqa [rsp + rounding_bit], xmm5
67 movdqa [rsp + filter_weight], xmm0
87 movdqa xmm0, [rsi] ; src (frame1)
89 movdqa xmm1, xmm0
94 movdqa xmm2, [rdx] ; predictor (frame2)
95 movdqa xmm3, xmm2
121 movdqa xmm3, [GLOBAL(_const_16w)]
122 movdqa xmm2, [GLOBAL(_const_16w)
    [all...]
dct_sse2.asm 78 movdqa xmm2, xmm0
81 movdqa xmm1, xmm0
87 movdqa xmm3, xmm0
93 movdqa xmm1, xmm0
96 movdqa xmm4, xmm3
111 movdqa xmm2, xmm0
115 movdqa xmm3, xmm0
118 movdqa xmm2, xmm0
122 movdqa xmm5, XMMWORD PTR[GLOBAL(_7)]
124 movdqa xmm3, xmm
    [all...]
quantize_sse4.asm 61 movdqa xmm0, [rax]
62 movdqa xmm1, [rax + 16]
68 movdqa xmm2, xmm0
69 movdqa xmm3, xmm1
84 movdqa xmm4, [rcx]
85 movdqa xmm5, [rcx + 16]
91 movdqa xmm6, xmm2
92 movdqa xmm7, xmm3
99 movdqa xmm4, [rdx]
100 movdqa xmm5, [rdx + 16
    [all...]
quantize_sse2.asm 63 movdqa xmm0, [rdx]
64 movdqa xmm4, [rdx + 16]
70 movdqa xmm1, xmm0
71 movdqa xmm5, xmm4
85 movdqa xmm2, [rcx]
86 movdqa xmm3, [rcx + 16]
96 movdqa [rsp + abs_minus_zbin], xmm1
97 movdqa [rsp + abs_minus_zbin + 16], xmm5
103 movdqa xmm2, [rdx]
104 movdqa xmm6, [rdx + 16
    [all...]
fwalsh_sse2.asm 40 movdqa xmm1, xmm0
44 movdqa xmm2, xmm0
51 movdqa xmm1, xmm0
61 movdqa xmm2, xmm0
77 movdqa xmm2, xmm0
80 movdqa xmm3, xmm1
89 movdqa xmm0, xmm4
92 movdqa xmm1, xmm6
96 movdqa xmm2, xmm0
99 movdqa xmm3, xmm
    [all...]
subtract_sse2.asm 90 movdqa xmm4, [GLOBAL(t80)]
96 movdqa xmm0, [rsi] ; src
97 movdqa xmm1, [rax] ; pred
99 movdqa xmm2, xmm0
106 movdqa xmm2, xmm0
110 movdqa xmm3, [rsi + rdx]
111 movdqa xmm5, [rax + rbx]
116 movdqa [rdi], xmm0
117 movdqa [rdi +16], xmm2
119 movdqa xmm1, xmm
    [all...]
  /system/core/libcutils/arch-x86/
sse2-memset32-atom.S 295 movdqa %xmm0, (%edx)
296 movdqa %xmm0, 0x10(%edx)
297 movdqa %xmm0, 0x20(%edx)
298 movdqa %xmm0, 0x30(%edx)
299 movdqa %xmm0, 0x40(%edx)
300 movdqa %xmm0, 0x50(%edx)
301 movdqa %xmm0, 0x60(%edx)
302 movdqa %xmm0, 0x70(%edx)
308 movdqa %xmm0, (%edx)
309 movdqa %xmm0, 0x10(%edx
    [all...]
sse2-memset16-atom.S 381 movdqa %xmm0, (%edx)
382 movdqa %xmm0, 0x10(%edx)
383 movdqa %xmm0, 0x20(%edx)
384 movdqa %xmm0, 0x30(%edx)
385 movdqa %xmm0, 0x40(%edx)
386 movdqa %xmm0, 0x50(%edx)
387 movdqa %xmm0, 0x60(%edx)
388 movdqa %xmm0, 0x70(%edx)
394 movdqa %xmm0, (%edx)
395 movdqa %xmm0, 0x10(%edx
    [all...]
  /external/libyuv/files/source/
row_win.cc 139 movdqa xmm1, xmm0
144 movdqa [edx], xmm0 local
145 movdqa [edx + 16], xmm1 local
159 movdqa xmm5, kShuffleMaskBGRAToARGB
164 movdqa xmm0, [eax]
167 movdqa [eax + edx], xmm0 local
180 movdqa xmm5, kShuffleMaskABGRToARGB
185 movdqa xmm0, [eax]
188 movdqa [eax + edx], xmm0 local
201 movdqa xmm5, kShuffleMaskRGBAToARG
209 movdqa [eax + edx], xmm0 local
230 movdqa [eax + edx], xmm0 local
259 movdqa [edx + 32], xmm2 local
262 movdqa [edx], xmm0 local
266 movdqa [edx + 16], xmm1 local
269 movdqa [edx + 48], xmm3 local
299 movdqa [edx + 32], xmm2 local
302 movdqa [edx], xmm0 local
306 movdqa [edx + 16], xmm1 local
309 movdqa [edx + 48], xmm3 local
364 movdqa [eax * 2 + edx], xmm1 \/\/ store 4 pixels of ARGB local
365 movdqa [eax * 2 + edx + 16], xmm2 \/\/ store next 4 pixels of ARGB local
418 movdqa [eax * 2 + edx], xmm1 \/\/ store 4 pixels of ARGB local
419 movdqa [eax * 2 + edx + 16], xmm2 \/\/ store next 4 pixels of ARGB local
458 movdqa [eax * 2 + edx], xmm0 \/\/ store 4 pixels of ARGB local
459 movdqa [eax * 2 + edx + 16], xmm1 \/\/ store next 4 pixels of ARGB local
492 movdqa [edx], xmm0 \/\/ store 0 local
497 movdqa [edx + 16], xmm1 \/\/ store 1 local
498 movdqa [edx + 32], xmm2 \/\/ store 2 local
531 movdqa [edx], xmm0 \/\/ store 0 local
536 movdqa [edx + 16], xmm1 \/\/ store 1 local
537 movdqa [edx + 32], xmm2 \/\/ store 2 local
684 movdqa [edx], xmm0 local
752 movdqa [edx], xmm0 local
820 movdqa [edx], xmm0 local
888 movdqa [edx], xmm0 local
1645 movdqa [edx], xmm0 local
1646 movdqa [edx + 16], xmm1 local
1688 movdqa [edx], xmm0 local
1689 movdqa [edx + 16], xmm1 local
1732 movdqa [edx], xmm0 local
1733 movdqa [edx + 16], xmm1 local
1771 movdqa [edx], xmm0 local
1772 movdqa [edx + 16], xmm1 local
1809 movdqa [edx], xmm0 local
1810 movdqa [edx + 16], xmm1 local
2056 movdqa [edx], xmm5 local
2057 movdqa [edx + 16], xmm0 local
2138 movdqa [edx], xmm2 local
2139 movdqa [edx + 16], xmm1 local
2220 movdqa [edx], xmm5 local
2221 movdqa [edx + 16], xmm0 local
2310 movdqa [edx], xmm0 local
2311 movdqa [edx + 16], xmm1 local
2342 movdqa [edx], xmm0 local
2437 movdqa [edx], xmm0 local
2471 movdqa [edx], xmm0 local
2472 movdqa [edx + edi], xmm2 local
2497 movdqa [eax + edx], xmm0 local
2498 movdqa [eax + edx + 16], xmm1 local
2593 movdqa [edx], xmm0 local
2805 movdqa [edx], xmm0 local
3079 movdqa [edx], xmm0 local
3214 movdqa [edx], xmm0 local
3240 movdqa [edx], xmm0 local
3315 movdqa [eax + edx], xmm0 local
3364 movdqa [eax + edx], xmm0 local
3418 movdqa [eax + edx], xmm0 local
3466 movdqa [eax + edx], xmm0 local
3467 movdqa [eax + edx + 16], xmm1 local
3537 movdqa [eax], xmm0 local
3538 movdqa [eax + 16], xmm1 local
3600 movdqa [eax], xmm0 local
3601 movdqa [eax + 16], xmm1 local
3696 movdqa [eax], xmm0 local
3861 movdqa [edx], xmm2 local
3862 movdqa [edx + 16], xmm3 local
3863 movdqa [edx + 32], xmm4 local
3864 movdqa [edx + 48], xmm5 local
3921 movdqa [eax + edx], xmm0 local
4057 movdqa [esi + edi], xmm0 local
4069 movdqa [esi + edi], xmm0 local
4082 movdqa [esi + edi], xmm0 local
    [all...]
  /external/libffi/src/x86/
unix64.S 194 movdqa 48(%r10), %xmm0
195 movdqa 64(%r10), %xmm1
196 movdqa 80(%r10), %xmm2
197 movdqa 96(%r10), %xmm3
198 movdqa 112(%r10), %xmm4
199 movdqa 128(%r10), %xmm5
200 movdqa 144(%r10), %xmm6
201 movdqa 160(%r10), %xmm7
318 movdqa %xmm0, 48(%rsp)
319 movdqa %xmm1, 64(%rsp
    [all...]
  /external/libvpx/libvpx/third_party/libyuv/source/
scale.c 652 movdqa xmm0, [eax]
653 movdqa xmm1, [eax + 16]
658 movdqa [edx], xmm0 local
681 movdqa xmm0, [eax]
682 movdqa xmm1, [eax + 16]
683 movdqa xmm2, [eax + esi]
684 movdqa xmm3, [eax + esi + 16]
689 movdqa xmm2, xmm0 // average columns (32 to 16 pixels)
691 movdqa xmm3, xmm1
699 movdqa [edx], xmm local
1255 movdqa [edi], xmm2 local
1256 movdqa [edi + 16], xmm3 local
1316 movdqa [edi], xmm0 local
1330 movdqa [edi], xmm0 local
1346 movdqa [edi], xmm0 local
1398 movdqa [edi], xmm0 local
1412 movdqa [edi], xmm0 local
1428 movdqa [edi], xmm0 local
    [all...]

Completed in 213 milliseconds

1 2 3 4