/external/libvpx/libvpx/vp9/common/x86/ |
vp9_intrapred_ssse3.asm | 241 palignr m5, m4, m0, 1 242 palignr m6, m4, m0, 2 259 palignr m3, m4, m5, 1 264 palignr m5, m4, m3, 1 269 palignr m3, m4, m5, 1 274 palignr m5, m4, m3, 1 423 palignr m6, m7, m0, 1 424 palignr m5, m7, m0, 2 436 palignr m3, m7, m0, 1 437 palignr m5, m4, m2, [all...] |
/hardware/intel/common/omx-components/videocodec/libvpx_internal/libvpx/vp9/common/x86/ |
vp9_intrapred_ssse3.asm | 241 palignr m5, m4, m0, 1 242 palignr m6, m4, m0, 2 259 palignr m3, m4, m5, 1 264 palignr m5, m4, m3, 1 269 palignr m3, m4, m5, 1 274 palignr m5, m4, m3, 1 423 palignr m6, m7, m0, 1 424 palignr m5, m7, m0, 2 436 palignr m3, m7, m0, 1 437 palignr m5, m4, m2, [all...] |
/bionic/libc/arch-x86/atom/string/ |
ssse3-memcpy-atom.S | 452 palignr $1, %xmm4, %xmm5 453 palignr $1, %xmm3, %xmm4 455 palignr $1, %xmm2, %xmm3 457 palignr $1, %xmm1, %xmm2 472 palignr $1, %xmm2, %xmm3 473 palignr $1, %xmm1, %xmm2 495 palignr $1, %xmm2, %xmm3 496 palignr $1, %xmm1, %xmm2 506 palignr $1, %xmm2, %xmm3 507 palignr $1, %xmm4, %xmm [all...] |
ssse3-strcpy-atom.S | 433 palignr $1, %xmm1, %xmm2 450 palignr $1, %xmm1, %xmm2 467 palignr $1, %xmm1, %xmm2 482 palignr $1, %xmm3, %xmm2 510 palignr $1, %xmm4, %xmm5 511 palignr $1, %xmm3, %xmm4 518 palignr $1, %xmm2, %xmm3 520 palignr $1, %xmm1, %xmm2 552 palignr $2, %xmm1, %xmm2 569 palignr $2, %xmm1, %xmm [all...] |
ssse3-wcscpy-atom.S | 271 palignr $4, %xmm1, %xmm2 284 palignr $4, %xmm3, %xmm2 297 palignr $4, %xmm1, %xmm2 309 palignr $4, %xmm3, %xmm2 335 palignr $4, %xmm4, %xmm5 336 palignr $4, %xmm3, %xmm4 340 palignr $4, %xmm2, %xmm3 342 palignr $4, %xmm1, %xmm2 382 palignr $8, %xmm1, %xmm2 395 palignr $8, %xmm3, %xmm [all...] |
ssse3-memcmp-atom.S | 323 palignr $1,(%esi), %xmm1 327 palignr $1,%xmm2, %xmm3 348 palignr $1,(%esi), %xmm0 352 palignr $1,16(%esi), %xmm3 362 palignr $1,48(%esi), %xmm3 365 palignr $1,32(%esi), %xmm0 409 palignr $2,(%esi), %xmm1 413 palignr $2,%xmm2, %xmm3 434 palignr $2,(%esi), %xmm0 438 palignr $2,16(%esi), %xmm [all...] |
ssse3-strcmp-atom.S | 353 palignr $1, %xmm3, %xmm2 377 palignr $1, %xmm3, %xmm2 459 palignr $2, %xmm3, %xmm2 483 palignr $2, %xmm3, %xmm2 566 palignr $3, %xmm3, %xmm2 590 palignr $3, %xmm3, %xmm2 672 palignr $4, %xmm3, %xmm2 697 palignr $4, %xmm3, %xmm2 781 palignr $5, %xmm3, %xmm2 805 palignr $5, %xmm3, %xmm [all...] |
/art/runtime/arch/x86/ |
memcmp16_x86.S | 191 palignr $2,(%esi), %xmm1 195 palignr $2,%xmm2, %xmm3 216 palignr $2,(%esi), %xmm0 220 palignr $2,16(%esi), %xmm3 230 palignr $2,48(%esi), %xmm3 233 palignr $2,32(%esi), %xmm0 274 palignr $4,(%esi), %xmm1 278 palignr $4,%xmm2, %xmm3 299 palignr $4,(%esi), %xmm0 303 palignr $4,16(%esi), %xmm [all...] |
/external/libvpx/libvpx/vp8/common/x86/ |
sad_ssse3.asm | 62 palignr xmm5, xmm4, %2 65 palignr xmm6, xmm4, (%2+1) 67 palignr xmm7, xmm4, (%2+2) 78 palignr xmm1, xmm4, %2 81 palignr xmm2, xmm4, (%2+1) 83 palignr xmm3, xmm4, (%2+2) 98 palignr xmm1, xmm4, %2 101 palignr xmm2, xmm4, (%2+1) 103 palignr xmm3, xmm4, (%2+2)
|
/external/libvpx/libvpx/vp9/encoder/x86/ |
vp9_sad_ssse3.asm | 62 palignr xmm5, xmm4, %2 65 palignr xmm6, xmm4, (%2+1) 67 palignr xmm7, xmm4, (%2+2) 78 palignr xmm1, xmm4, %2 81 palignr xmm2, xmm4, (%2+1) 83 palignr xmm3, xmm4, (%2+2) 98 palignr xmm1, xmm4, %2 101 palignr xmm2, xmm4, (%2+1) 103 palignr xmm3, xmm4, (%2+2)
|
/hardware/intel/common/omx-components/videocodec/libvpx_internal/libvpx/vp8/common/x86/ |
sad_ssse3.asm | 62 palignr xmm5, xmm4, %2 65 palignr xmm6, xmm4, (%2+1) 67 palignr xmm7, xmm4, (%2+2) 78 palignr xmm1, xmm4, %2 81 palignr xmm2, xmm4, (%2+1) 83 palignr xmm3, xmm4, (%2+2) 98 palignr xmm1, xmm4, %2 101 palignr xmm2, xmm4, (%2+1) 103 palignr xmm3, xmm4, (%2+2)
|
/hardware/intel/common/omx-components/videocodec/libvpx_internal/libvpx/vp9/encoder/x86/ |
vp9_sad_ssse3.asm | 62 palignr xmm5, xmm4, %2 65 palignr xmm6, xmm4, (%2+1) 67 palignr xmm7, xmm4, (%2+2) 78 palignr xmm1, xmm4, %2 81 palignr xmm2, xmm4, (%2+1) 83 palignr xmm3, xmm4, (%2+2) 98 palignr xmm1, xmm4, %2 101 palignr xmm2, xmm4, (%2+1) 103 palignr xmm3, xmm4, (%2+2)
|
/bionic/libc/arch-x86_64/string/ |
ssse3-strcmp-slm.S | 251 palignr $1, %xmm3, %xmm2 /* merge into one 16byte value */ 274 palignr $1, %xmm3, %xmm2 /* merge into one 16byte value */ 366 palignr $2, %xmm3, %xmm2 /* merge into one 16byte value */ 390 palignr $2, %xmm3, %xmm2 /* merge into one 16byte value */ 476 palignr $3, %xmm3, %xmm2 /* merge into one 16byte value */ 500 palignr $3, %xmm3, %xmm2 /* merge into one 16byte value */ 586 palignr $4, %xmm3, %xmm2 /* merge into one 16byte value */ 610 palignr $4, %xmm3, %xmm2 /* merge into one 16byte value */ 696 palignr $5, %xmm3, %xmm2 /* merge into one 16byte value */ 720 palignr $5, %xmm3, %xmm2 /* merge into one 16byte value * [all...] |
/external/libvpx/libvpx/third_party/libyuv/source/ |
rotate.cc | 103 palignr xmm1, xmm1, 8 109 palignr xmm3, xmm3, 8 115 palignr xmm5, xmm5, 8 120 palignr xmm7, xmm7, 8 126 palignr xmm2, xmm2, 8 127 palignr xmm3, xmm3, 8 132 palignr xmm6, xmm6, 8 133 palignr xmm7, xmm7, 8 139 palignr xmm4, xmm4, 8 144 palignr xmm6, xmm6, [all...] |
/external/libyuv/files/source/ |
rotate.cc | 87 palignr xmm1, xmm1, 8 93 palignr xmm3, xmm3, 8 99 palignr xmm5, xmm5, 8 104 palignr xmm7, xmm7, 8 110 palignr xmm2, xmm2, 8 111 palignr xmm3, xmm3, 8 116 palignr xmm6, xmm6, 8 117 palignr xmm7, xmm7, 8 123 palignr xmm4, xmm4, 8 128 palignr xmm6, xmm6, [all...] |
row_win.cc | 254 palignr xmm2, xmm1, 8 // xmm2 = { xmm3[0:3] xmm1[8:15]} 257 palignr xmm1, xmm0, 12 // xmm1 = { xmm3[0:7] xmm0[12:15]} 264 palignr xmm3, xmm3, 4 // xmm3 = { xmm3[4:15]} 294 palignr xmm2, xmm1, 8 // xmm2 = { xmm3[0:3] xmm1[8:15]} 297 palignr xmm1, xmm0, 12 // xmm1 = { xmm3[0:7] xmm0[12:15]} 304 palignr xmm3, xmm3, 4 // xmm3 = { xmm3[4:15]} [all...] |
/external/llvm/test/MC/X86/ |
shuffle-comments.s | 3 palignr $8, %xmm0, %xmm1 label 5 palignr $8, (%rax), %xmm1 label 8 palignr $16, %xmm0, %xmm1 label 10 palignr $16, (%rax), %xmm1 label 13 palignr $0, %xmm0, %xmm1 label 15 palignr $0, (%rax), %xmm1 label
|
/external/boringssl/src/crypto/aes/asm/ |
vpaes-x86.pl | 325 &palignr("xmm5","xmm5",12); 445 &palignr("xmm0","xmm6",8); 577 &palignr("xmm1","xmm2",15); 578 &palignr("xmm2","xmm2",15); 583 &palignr("xmm0","xmm0",1);
|
vpaes-x86_64.pl | 234 palignr \$12, %xmm5, %xmm5 362 palignr \$8,%xmm6,%xmm0 500 palignr \$15, %xmm8, %xmm1 501 palignr \$15, %xmm8, %xmm8 506 palignr \$1, %xmm0, %xmm0
|
/external/boringssl/src/crypto/sha/asm/ |
sha256-586.pl | 543 &palignr ($ABEF,$CDGH,8); # ABEF 579 &palignr ($TMP,@MSG[2],4); 591 &palignr ($TMP,@MSG[3],4); 604 &palignr ($TMP,@MSG[0],4); 618 &palignr ($TMP,@MSG[0],4); 647 &palignr ($CDGH,$TMP,8); # HGFE 725 &palignr ($t0,@X[0],4); # X[1..4] 729 &palignr ($t3,@X[2],4); # X[9..12] [all...] |
sha512-x86_64.pl | 563 palignr \$8,$CDGH,$ABEF # ABEF 600 palignr \$4,@MSG[2],$TMP 612 palignr \$4,@MSG[3],$TMP 626 palignr \$4,@MSG[0],$TMP 641 palignr \$4,@MSG[0],$TMP 670 palignr \$8,$TMP,$CDGH # HGFE 832 '&palignr ($t0,@X[0],$SZ)', # X[1..4] 833 '&palignr ($t3,@X[2],$SZ);', # X[9..12] 893 &palignr ($t0,@X[0],$SZ); # X[1..4] 896 &palignr ($t3,@X[2],$SZ); # X[9..12 [all...] |
sha512-586.pl | 574 &palignr ($t0,@X[0],8); # X[1..2] 576 &palignr ($t2,@X[4],8); # X[9..10]
|
/external/boringssl/src/crypto/perlasm/ |
x86asm.pl | 110 sub ::palignr 115 { &::generic("palignr",@_); }
|
x86_64-xlate.pl | 757 my $palignr = sub { [all...] |
/external/boringssl/src/crypto/modes/asm/ |
ghash-x86_64.pl | 531 palignr \$8,$T1,$T2 # low part is H.lo^H.hi... 549 palignr \$8,$T1,$T2 # low part is H^3.lo^H^3.hi... [all...] |