HomeSort by relevance Sort by last modified time
    Searched refs:movdqu (Results 1 - 25 of 45) sorted by null

1 2

  /external/openssl/crypto/aes/asm/
bsaes-x86_64.pl 910 movdqu ($inp), %xmm7 # load round 0 key
919 movdqu ($inp), %xmm6 # load round 1 key
972 movdqu ($inp), %xmm6 # load next round key
1012 movdqu 0x00($inp), @XMM[0] # load input
1013 movdqu 0x10($inp), @XMM[1]
1014 movdqu 0x20($inp), @XMM[2]
1015 movdqu 0x30($inp), @XMM[3]
1016 movdqu 0x40($inp), @XMM[4]
1017 movdqu 0x50($inp), @XMM[5]
1018 movdqu 0x60($inp), @XMM[6
    [all...]
vpaes-x86.pl 186 &movdqu ("xmm5",&QWP(0,$key));
245 &movdqu ("xmm5",&QWP(0,$key));
275 &movdqu ("xmm5",&QWP(0,$key));
350 &movdqu ("xmm0",&QWP(0,$key));
372 &movdqu ("xmm0",&QWP(0,$inp)); # load key (unaligned)
386 &movdqu (&QWP(0,$key),"xmm0");
393 &movdqu (&QWP(0,$key),"xmm3");
436 &movdqu ("xmm0",&QWP(8,$inp)); # load key part 2 (very unaligned)
467 &movdqu ("xmm0",&QWP(16,$inp)); # load key part 2 (unaligned)
516 &movdqu (&QWP(0,$key),"xmm0"); # save last ke
    [all...]
aesni-x86.pl 378 &movdqu ($inout0,&QWP(0,$inp));
379 &movdqu ($inout1,&QWP(0x10,$inp));
380 &movdqu ($inout2,&QWP(0x20,$inp));
381 &movdqu ($inout3,&QWP(0x30,$inp));
382 &movdqu ($inout4,&QWP(0x40,$inp));
383 &movdqu ($inout5,&QWP(0x50,$inp));
390 &movdqu ($inout0,&QWP(0,$inp));
392 &movdqu ($inout1,&QWP(0x10,$inp));
394 &movdqu ($inout2,&QWP(0x20,$inp));
396 &movdqu ($inout3,&QWP(0x30,$inp))
    [all...]
aesni-x86_64.pl 543 movdqu ($inp),$inout0
544 movdqu 0x10($inp),$inout1
545 movdqu 0x20($inp),$inout2
546 movdqu 0x30($inp),$inout3
547 movdqu 0x40($inp),$inout4
548 movdqu 0x50($inp),$inout5
549 movdqu 0x60($inp),$inout6
550 movdqu 0x70($inp),$inout7
558 movdqu ($inp),$inout0
561 movdqu 0x10($inp),$inout
    [all...]
vpaes-x86_64.pl 90 movdqu (%r9), %xmm5 # round0 key
149 movdqu (%r9), %xmm5
181 movdqu (%r9), %xmm5 # round0 key
258 movdqu (%r9), %xmm0
288 movdqu (%rdi), %xmm0 # load key (unaligned)
301 movdqu %xmm0, (%rdx)
308 movdqu %xmm3, (%rdx)
352 movdqu 8(%rdi),%xmm0 # load key part 2 (very unaligned)
384 movdqu 16(%rdi),%xmm0 # load key part 2 (unaligned)
435 movdqu %xmm0, (%rdx) # save last ke
    [all...]
aesni-sha1-x86_64.pl 164 movdqu ($ivp),$iv # load IV
185 movdqu 0($inp),@X[-4&7] # load input to %xmm[0-3]
186 movdqu 16($inp),@X[-3&7]
187 movdqu 32($inp),@X[-2&7]
188 movdqu 48($inp),@X[-1&7]
423 &movdqu (@X[-4&7],"0($inp)"); # load input
424 &movdqu (@X[-3&7],"16($inp)");
425 &movdqu (@X[-2&7],"32($inp)");
426 &movdqu (@X[-1&7],"48($inp)");
  /external/libvpx/vp8/common/x86/
recon_sse2.asm 145 movdqu xmm0, [rsi]
150 movdqu xmm1, [rsi+rax]
151 movdqu xmm2, [rsi+rax*2]
163 movdqu xmm3, [rsi]
166 movdqu xmm4, [rsi+rax]
168 movdqu xmm5, [rsi+rax*2]
178 movdqu xmm0, [rsi]
181 movdqu xmm1, [rsi+rax]
183 movdqu xmm2, [rsi+rax*2]
192 movdqu xmm3, [rsi
    [all...]
iwalsh_sse2.asm 51 ;; movdqu [rdi + 0], xmm4
52 ;; movdqu [rdi + 16], xmm3
subpixel_sse2.asm     [all...]
subpixel_ssse3.asm 288 movdqu xmm1, XMMWORD PTR [rsi - 2]
295 movdqu xmm3, XMMWORD PTR [rsi + 6]
376 movdqu xmm0, XMMWORD PTR [rsi - 2]
427 movdqu xmm1, XMMWORD PTR [rsi - 2]
    [all...]
loopfilter_sse2.asm     [all...]
postproc_sse2.asm 397 movdqu xmm4, [rax + rcx*2] ;vp8_rv[rcx*2]
400 movdqu xmm4, [r8 + rcx*2] ;vp8_rv[rcx*2]
402 movdqu xmm4, [sym(vp8_rv) + rcx*2]
661 movdqu xmm1,[rsi+rax] ; get the source
667 movdqu xmm2,[rdi+rax] ; get the noise for this line
669 movdqu [rsi+rax],xmm1 ; store the result
  /external/zlib/contrib/amd64/
amd64-match.S 293 movdqu (%windowbestlen, %rdx), %xmm1
294 movdqu (%prev, %rdx), %xmm2
296 movdqu 16(%windowbestlen, %rdx), %xmm3
297 movdqu 16(%prev, %rdx), %xmm4
299 movdqu 32(%windowbestlen, %rdx), %xmm5
300 movdqu 32(%prev, %rdx), %xmm6
302 movdqu 48(%windowbestlen, %rdx), %xmm7
303 movdqu 48(%prev, %rdx), %xmm8
  /external/openssl/crypto/modes/asm/
ghash-x86.pl 926 &movdqu ($Hkey,&QWP(0,$Xip));
948 &movdqu (&QWP(0,$Htbl),$Hkey); # save H
949 &movdqu (&QWP(16,$Htbl),$Xi); # save H^2
963 &movdqu ($Xi,&QWP(0,$Xip));
972 &movdqu (&QWP(0,$Xip),$Xi);
988 &movdqu ($Xi,&QWP(0,$Xip));
990 &movdqu ($Hkey,&QWP(0,$Htbl));
1001 &movdqu ($T1,&QWP(0,$inp)); # Ii
1002 &movdqu ($Xn,&QWP(16,$inp)); # Ii+1
1016 &movdqu ($T1,&QWP(0,$inp)); # I
    [all...]
ghash-x86_64.pl 418 movdqu ($Xip),$Hkey
441 movdqu $Hkey,($Htbl) # save H
442 movdqu $Xi,16($Htbl) # save H^2
455 movdqu ($Xip),$Xi
457 movdqu ($Htbl),$Hkey
464 movdqu $Xi,($Xip)
497 movdqu ($Xip),$Xi
498 movdqu ($Htbl),$Hkey
504 movdqu 16($Htbl),$Hkey2
510 movdqu ($inp),$T1 # I
    [all...]
  /external/libvpx/vp8/encoder/x86/
variance_impl_ssse3.asm 67 movdqu xmm0, XMMWORD PTR [rsi]
68 movdqu xmm1, XMMWORD PTR [rsi+1]
92 movdqu xmm1, XMMWORD PTR [rsi]
93 movdqu xmm2, XMMWORD PTR [rsi+1]
163 movdqu xmm1, XMMWORD PTR [rsi]
173 movdqu xmm3, XMMWORD PTR [rsi]
264 movdqu xmm1, XMMWORD PTR [rsi]
265 movdqu xmm2, XMMWORD PTR [rsi+1]
variance_impl_sse2.asm 129 movdqu xmm1, XMMWORD PTR [rsi]
130 movdqu xmm2, XMMWORD PTR [rdi]
245 movdqu xmm1, XMMWORD PTR [rsi]
246 movdqu xmm2, XMMWORD PTR [rdi]
945 movdqu xmm5, XMMWORD PTR [rsi]
946 movdqu xmm3, XMMWORD PTR [rsi+1]
952 movdqu xmm1, XMMWORD PTR [rsi] ;
953 movdqu xmm2, XMMWORD PTR [rsi+1] ;
    [all...]
  /bionic/libc/arch-x86/string/
ssse3-memcpy5.S 204 movdqu (%eax), %xmm0
239 movdqu %xmm0, (%esi)
443 movdqu %xmm0, (%esi)
488 movdqu %xmm0, (%esi)
533 movdqu %xmm0, (%esi)
578 movdqu %xmm0, (%esi)
623 movdqu %xmm0, (%esi)
668 movdqu %xmm0, (%esi)
713 movdqu %xmm0, (%esi)
758 movdqu %xmm0, (%esi
    [all...]
  /external/libyuv/files/source/
convert.cc 380 movdqu xmm3, XMMWORD PTR [eax] ;in1
384 movdqu XMMWORD PTR [esi], xmm1 ;write to out1
386 movdqu xmm5, XMMWORD PTR [ecx] ;in2
390 movdqu XMMWORD PTR [edi], xmm2 ;write to out2
393 movdqu XMMWORD PTR [esi+16], xmm3 ;write to out1 again
398 movdqu XMMWORD PTR [edi+16], xmm5 ;write to out2 again
494 movdqu xmm3, XMMWORD PTR [eax] ;in1
497 movdqu XMMWORD PTR [esi], xmm1 ;write to out1
499 movdqu xmm5, XMMWORD PTR [ecx] ;in2
502 movdqu XMMWORD PTR [edi], xmm2 ;write to out
    [all...]
  /external/openssl/crypto/bn/asm/
modexp512-x86_64.pl 1132 movdqu (+16*0)(%rsi), %xmm0
1133 movdqu (+16*1)(%rsi), %xmm1
1134 movdqu (+16*2)(%rsi), %xmm2
1135 movdqu (+16*3)(%rsi), %xmm3
1146 movdqu (+16*0)(%rdx), %xmm0
1147 movdqu (+16*1)(%rdx), %xmm1
1148 movdqu (+16*2)(%rdx), %xmm2
1149 movdqu (+16*3)(%rdx), %xmm3
1297 movdqu (+16*0)(%rdx), %xmm0
1298 movdqu (+16*1)(%rdx), %xmm
    [all...]
x86_64-mont.pl 651 movdqu ($ap),%xmm1
653 movdqu %xmm1,($rp)
657 movdqu 16($ap,$i),%xmm2
658 movdqu 32($ap,$i),%xmm1
660 movdqu %xmm2,16($rp,$i)
662 movdqu %xmm1,32($rp,$i)
668 movdqu 16($ap,$i),%xmm2
670 movdqu %xmm2,16($rp,$i)
1461 movdqu ($tptr),%xmm1
1465 movdqu %xmm1,($rptr
    [all...]
x86_64-mont5.pl 786 movdqu ($ap),%xmm1
788 movdqu %xmm1,($rp)
792 movdqu 16($ap,$i),%xmm2
793 movdqu 32($ap,$i),%xmm1
795 movdqu %xmm2,16($rp,$i)
797 movdqu %xmm1,32($rp,$i)
803 movdqu 16($ap,$i),%xmm2
805 movdqu %xmm2,16($rp,$i)
  /external/v8/src/ia32/
codegen-ia32.cc 191 __ movdqu(xmm0, Operand(src, 0));
192 __ movdqu(Operand(dst, 0), xmm0);
240 __ movdqu(xmm0, Operand(src, count, times_1, -0x10));
241 __ movdqu(Operand(dst, count, times_1, -0x10), xmm0);
262 __ movdqu(xmm0, Operand(src, 0x00));
263 __ movdqu(xmm1, Operand(src, 0x10));
278 __ movdqu(xmm0, Operand(src, 0));
286 __ movdqu(xmm0, Operand(src, count, times_1, -0x10));
287 __ movdqu(Operand(dst, count, times_1, -0x10), xmm0);
  /external/openssl/crypto/sha/asm/
sha1-586.pl 487 &movdqu (@X[-4&7],&QWP(-64,$inp)); # load input to %xmm[0-3]
488 &movdqu (@X[-3&7],&QWP(-48,$inp));
489 &movdqu (@X[-2&7],&QWP(-32,$inp));
490 &movdqu (@X[-1&7],&QWP(-16,$inp));
700 &movdqu (@X[-4&7],&QWP(0,$inp)); # load input
701 &movdqu (@X[-3&7],&QWP(16,$inp));
702 &movdqu (@X[-2&7],&QWP(32,$inp));
703 &movdqu (@X[-1&7],&QWP(48,$inp));
    [all...]
sha1-x86_64.pl 341 movdqu 0($inp),@X[-4&7] # load input to %xmm[0-3]
342 movdqu 16($inp),@X[-3&7]
343 movdqu 32($inp),@X[-2&7]
344 movdqu 48($inp),@X[-1&7]
541 &movdqu (@X[-4&7],"0($inp)"); # load input
542 &movdqu (@X[-3&7],"16($inp)");
543 &movdqu (@X[-2&7],"32($inp)");
544 &movdqu (@X[-1&7],"48($inp)");

Completed in 495 milliseconds

1 2