/external/openssl/crypto/aes/asm/ |
bsaes-x86_64.pl | 910 movdqu ($inp), %xmm7 # load round 0 key 919 movdqu ($inp), %xmm6 # load round 1 key 972 movdqu ($inp), %xmm6 # load next round key 1012 movdqu 0x00($inp), @XMM[0] # load input 1013 movdqu 0x10($inp), @XMM[1] 1014 movdqu 0x20($inp), @XMM[2] 1015 movdqu 0x30($inp), @XMM[3] 1016 movdqu 0x40($inp), @XMM[4] 1017 movdqu 0x50($inp), @XMM[5] 1018 movdqu 0x60($inp), @XMM[6 [all...] |
vpaes-x86.pl | 186 &movdqu ("xmm5",&QWP(0,$key)); 245 &movdqu ("xmm5",&QWP(0,$key)); 275 &movdqu ("xmm5",&QWP(0,$key)); 350 &movdqu ("xmm0",&QWP(0,$key)); 372 &movdqu ("xmm0",&QWP(0,$inp)); # load key (unaligned) 386 &movdqu (&QWP(0,$key),"xmm0"); 393 &movdqu (&QWP(0,$key),"xmm3"); 436 &movdqu ("xmm0",&QWP(8,$inp)); # load key part 2 (very unaligned) 467 &movdqu ("xmm0",&QWP(16,$inp)); # load key part 2 (unaligned) 516 &movdqu (&QWP(0,$key),"xmm0"); # save last ke [all...] |
aesni-x86.pl | 378 &movdqu ($inout0,&QWP(0,$inp)); 379 &movdqu ($inout1,&QWP(0x10,$inp)); 380 &movdqu ($inout2,&QWP(0x20,$inp)); 381 &movdqu ($inout3,&QWP(0x30,$inp)); 382 &movdqu ($inout4,&QWP(0x40,$inp)); 383 &movdqu ($inout5,&QWP(0x50,$inp)); 390 &movdqu ($inout0,&QWP(0,$inp)); 392 &movdqu ($inout1,&QWP(0x10,$inp)); 394 &movdqu ($inout2,&QWP(0x20,$inp)); 396 &movdqu ($inout3,&QWP(0x30,$inp)) [all...] |
aesni-x86_64.pl | 543 movdqu ($inp),$inout0 544 movdqu 0x10($inp),$inout1 545 movdqu 0x20($inp),$inout2 546 movdqu 0x30($inp),$inout3 547 movdqu 0x40($inp),$inout4 548 movdqu 0x50($inp),$inout5 549 movdqu 0x60($inp),$inout6 550 movdqu 0x70($inp),$inout7 558 movdqu ($inp),$inout0 561 movdqu 0x10($inp),$inout [all...] |
vpaes-x86_64.pl | 90 movdqu (%r9), %xmm5 # round0 key 149 movdqu (%r9), %xmm5 181 movdqu (%r9), %xmm5 # round0 key 258 movdqu (%r9), %xmm0 288 movdqu (%rdi), %xmm0 # load key (unaligned) 301 movdqu %xmm0, (%rdx) 308 movdqu %xmm3, (%rdx) 352 movdqu 8(%rdi),%xmm0 # load key part 2 (very unaligned) 384 movdqu 16(%rdi),%xmm0 # load key part 2 (unaligned) 435 movdqu %xmm0, (%rdx) # save last ke [all...] |
aesni-sha1-x86_64.pl | 164 movdqu ($ivp),$iv # load IV 185 movdqu 0($inp),@X[-4&7] # load input to %xmm[0-3] 186 movdqu 16($inp),@X[-3&7] 187 movdqu 32($inp),@X[-2&7] 188 movdqu 48($inp),@X[-1&7] 423 &movdqu (@X[-4&7],"0($inp)"); # load input 424 &movdqu (@X[-3&7],"16($inp)"); 425 &movdqu (@X[-2&7],"32($inp)"); 426 &movdqu (@X[-1&7],"48($inp)");
|
/external/libvpx/vp8/common/x86/ |
recon_sse2.asm | 145 movdqu xmm0, [rsi] 150 movdqu xmm1, [rsi+rax] 151 movdqu xmm2, [rsi+rax*2] 163 movdqu xmm3, [rsi] 166 movdqu xmm4, [rsi+rax] 168 movdqu xmm5, [rsi+rax*2] 178 movdqu xmm0, [rsi] 181 movdqu xmm1, [rsi+rax] 183 movdqu xmm2, [rsi+rax*2] 192 movdqu xmm3, [rsi [all...] |
iwalsh_sse2.asm | 51 ;; movdqu [rdi + 0], xmm4 52 ;; movdqu [rdi + 16], xmm3
|
subpixel_sse2.asm | [all...] |
subpixel_ssse3.asm | 288 movdqu xmm1, XMMWORD PTR [rsi - 2] 295 movdqu xmm3, XMMWORD PTR [rsi + 6] 376 movdqu xmm0, XMMWORD PTR [rsi - 2] 427 movdqu xmm1, XMMWORD PTR [rsi - 2] [all...] |
loopfilter_sse2.asm | [all...] |
postproc_sse2.asm | 397 movdqu xmm4, [rax + rcx*2] ;vp8_rv[rcx*2] 400 movdqu xmm4, [r8 + rcx*2] ;vp8_rv[rcx*2] 402 movdqu xmm4, [sym(vp8_rv) + rcx*2] 661 movdqu xmm1,[rsi+rax] ; get the source 667 movdqu xmm2,[rdi+rax] ; get the noise for this line 669 movdqu [rsi+rax],xmm1 ; store the result
|
/external/zlib/contrib/amd64/ |
amd64-match.S | 293 movdqu (%windowbestlen, %rdx), %xmm1 294 movdqu (%prev, %rdx), %xmm2 296 movdqu 16(%windowbestlen, %rdx), %xmm3 297 movdqu 16(%prev, %rdx), %xmm4 299 movdqu 32(%windowbestlen, %rdx), %xmm5 300 movdqu 32(%prev, %rdx), %xmm6 302 movdqu 48(%windowbestlen, %rdx), %xmm7 303 movdqu 48(%prev, %rdx), %xmm8
|
/external/openssl/crypto/modes/asm/ |
ghash-x86.pl | 926 &movdqu ($Hkey,&QWP(0,$Xip)); 948 &movdqu (&QWP(0,$Htbl),$Hkey); # save H 949 &movdqu (&QWP(16,$Htbl),$Xi); # save H^2 963 &movdqu ($Xi,&QWP(0,$Xip)); 972 &movdqu (&QWP(0,$Xip),$Xi); 988 &movdqu ($Xi,&QWP(0,$Xip)); 990 &movdqu ($Hkey,&QWP(0,$Htbl)); 1001 &movdqu ($T1,&QWP(0,$inp)); # Ii 1002 &movdqu ($Xn,&QWP(16,$inp)); # Ii+1 1016 &movdqu ($T1,&QWP(0,$inp)); # I [all...] |
ghash-x86_64.pl | 418 movdqu ($Xip),$Hkey 441 movdqu $Hkey,($Htbl) # save H 442 movdqu $Xi,16($Htbl) # save H^2 455 movdqu ($Xip),$Xi 457 movdqu ($Htbl),$Hkey 464 movdqu $Xi,($Xip) 497 movdqu ($Xip),$Xi 498 movdqu ($Htbl),$Hkey 504 movdqu 16($Htbl),$Hkey2 510 movdqu ($inp),$T1 # I [all...] |
/external/libvpx/vp8/encoder/x86/ |
variance_impl_ssse3.asm | 67 movdqu xmm0, XMMWORD PTR [rsi] 68 movdqu xmm1, XMMWORD PTR [rsi+1] 92 movdqu xmm1, XMMWORD PTR [rsi] 93 movdqu xmm2, XMMWORD PTR [rsi+1] 163 movdqu xmm1, XMMWORD PTR [rsi] 173 movdqu xmm3, XMMWORD PTR [rsi] 264 movdqu xmm1, XMMWORD PTR [rsi] 265 movdqu xmm2, XMMWORD PTR [rsi+1]
|
variance_impl_sse2.asm | 129 movdqu xmm1, XMMWORD PTR [rsi] 130 movdqu xmm2, XMMWORD PTR [rdi] 245 movdqu xmm1, XMMWORD PTR [rsi] 246 movdqu xmm2, XMMWORD PTR [rdi] 945 movdqu xmm5, XMMWORD PTR [rsi] 946 movdqu xmm3, XMMWORD PTR [rsi+1] 952 movdqu xmm1, XMMWORD PTR [rsi] ; 953 movdqu xmm2, XMMWORD PTR [rsi+1] ; [all...] |
/bionic/libc/arch-x86/string/ |
ssse3-memcpy5.S | 204 movdqu (%eax), %xmm0 239 movdqu %xmm0, (%esi) 443 movdqu %xmm0, (%esi) 488 movdqu %xmm0, (%esi) 533 movdqu %xmm0, (%esi) 578 movdqu %xmm0, (%esi) 623 movdqu %xmm0, (%esi) 668 movdqu %xmm0, (%esi) 713 movdqu %xmm0, (%esi) 758 movdqu %xmm0, (%esi [all...] |
/external/libyuv/files/source/ |
convert.cc | 380 movdqu xmm3, XMMWORD PTR [eax] ;in1 384 movdqu XMMWORD PTR [esi], xmm1 ;write to out1 386 movdqu xmm5, XMMWORD PTR [ecx] ;in2 390 movdqu XMMWORD PTR [edi], xmm2 ;write to out2 393 movdqu XMMWORD PTR [esi+16], xmm3 ;write to out1 again 398 movdqu XMMWORD PTR [edi+16], xmm5 ;write to out2 again 494 movdqu xmm3, XMMWORD PTR [eax] ;in1 497 movdqu XMMWORD PTR [esi], xmm1 ;write to out1 499 movdqu xmm5, XMMWORD PTR [ecx] ;in2 502 movdqu XMMWORD PTR [edi], xmm2 ;write to out [all...] |
/external/openssl/crypto/bn/asm/ |
modexp512-x86_64.pl | 1132 movdqu (+16*0)(%rsi), %xmm0 1133 movdqu (+16*1)(%rsi), %xmm1 1134 movdqu (+16*2)(%rsi), %xmm2 1135 movdqu (+16*3)(%rsi), %xmm3 1146 movdqu (+16*0)(%rdx), %xmm0 1147 movdqu (+16*1)(%rdx), %xmm1 1148 movdqu (+16*2)(%rdx), %xmm2 1149 movdqu (+16*3)(%rdx), %xmm3 1297 movdqu (+16*0)(%rdx), %xmm0 1298 movdqu (+16*1)(%rdx), %xmm [all...] |
x86_64-mont.pl | 651 movdqu ($ap),%xmm1 653 movdqu %xmm1,($rp) 657 movdqu 16($ap,$i),%xmm2 658 movdqu 32($ap,$i),%xmm1 660 movdqu %xmm2,16($rp,$i) 662 movdqu %xmm1,32($rp,$i) 668 movdqu 16($ap,$i),%xmm2 670 movdqu %xmm2,16($rp,$i) 1461 movdqu ($tptr),%xmm1 1465 movdqu %xmm1,($rptr [all...] |
x86_64-mont5.pl | 786 movdqu ($ap),%xmm1 788 movdqu %xmm1,($rp) 792 movdqu 16($ap,$i),%xmm2 793 movdqu 32($ap,$i),%xmm1 795 movdqu %xmm2,16($rp,$i) 797 movdqu %xmm1,32($rp,$i) 803 movdqu 16($ap,$i),%xmm2 805 movdqu %xmm2,16($rp,$i)
|
/external/v8/src/ia32/ |
codegen-ia32.cc | 191 __ movdqu(xmm0, Operand(src, 0)); 192 __ movdqu(Operand(dst, 0), xmm0); 240 __ movdqu(xmm0, Operand(src, count, times_1, -0x10)); 241 __ movdqu(Operand(dst, count, times_1, -0x10), xmm0); 262 __ movdqu(xmm0, Operand(src, 0x00)); 263 __ movdqu(xmm1, Operand(src, 0x10)); 278 __ movdqu(xmm0, Operand(src, 0)); 286 __ movdqu(xmm0, Operand(src, count, times_1, -0x10)); 287 __ movdqu(Operand(dst, count, times_1, -0x10), xmm0);
|
/external/openssl/crypto/sha/asm/ |
sha1-586.pl | 487 &movdqu (@X[-4&7],&QWP(-64,$inp)); # load input to %xmm[0-3] 488 &movdqu (@X[-3&7],&QWP(-48,$inp)); 489 &movdqu (@X[-2&7],&QWP(-32,$inp)); 490 &movdqu (@X[-1&7],&QWP(-16,$inp)); 700 &movdqu (@X[-4&7],&QWP(0,$inp)); # load input 701 &movdqu (@X[-3&7],&QWP(16,$inp)); 702 &movdqu (@X[-2&7],&QWP(32,$inp)); 703 &movdqu (@X[-1&7],&QWP(48,$inp)); [all...] |
sha1-x86_64.pl | 341 movdqu 0($inp),@X[-4&7] # load input to %xmm[0-3] 342 movdqu 16($inp),@X[-3&7] 343 movdqu 32($inp),@X[-2&7] 344 movdqu 48($inp),@X[-1&7] 541 &movdqu (@X[-4&7],"0($inp)"); # load input 542 &movdqu (@X[-3&7],"16($inp)"); 543 &movdqu (@X[-2&7],"32($inp)"); 544 &movdqu (@X[-1&7],"48($inp)");
|