/external/libvpx/libvpx/vp9/common/x86/ |
vp9_intrapred_ssse3.asm | 241 palignr m5, m4, m0, 1 242 palignr m6, m4, m0, 2 259 palignr m3, m4, m5, 1 264 palignr m5, m4, m3, 1 269 palignr m3, m4, m5, 1 274 palignr m5, m4, m3, 1 423 palignr m6, m7, m0, 1 424 palignr m5, m7, m0, 2 436 palignr m3, m7, m0, 1 437 palignr m5, m4, m2, [all...] |
/bionic/libc/arch-x86/string/ |
ssse3-memcpy-atom.S | 454 palignr $1, %xmm4, %xmm5 455 palignr $1, %xmm3, %xmm4 457 palignr $1, %xmm2, %xmm3 459 palignr $1, %xmm1, %xmm2 474 palignr $1, %xmm2, %xmm3 475 palignr $1, %xmm1, %xmm2 497 palignr $1, %xmm2, %xmm3 498 palignr $1, %xmm1, %xmm2 508 palignr $1, %xmm2, %xmm3 509 palignr $1, %xmm4, %xmm [all...] |
ssse3-strcpy-atom.S | 433 palignr $1, %xmm1, %xmm2 450 palignr $1, %xmm1, %xmm2 467 palignr $1, %xmm1, %xmm2 482 palignr $1, %xmm3, %xmm2 510 palignr $1, %xmm4, %xmm5 511 palignr $1, %xmm3, %xmm4 518 palignr $1, %xmm2, %xmm3 520 palignr $1, %xmm1, %xmm2 552 palignr $2, %xmm1, %xmm2 569 palignr $2, %xmm1, %xmm [all...] |
ssse3-wcscpy-atom.S | 271 palignr $4, %xmm1, %xmm2 284 palignr $4, %xmm3, %xmm2 297 palignr $4, %xmm1, %xmm2 309 palignr $4, %xmm3, %xmm2 335 palignr $4, %xmm4, %xmm5 336 palignr $4, %xmm3, %xmm4 340 palignr $4, %xmm2, %xmm3 342 palignr $4, %xmm1, %xmm2 382 palignr $8, %xmm1, %xmm2 395 palignr $8, %xmm3, %xmm [all...] |
ssse3-memcmp-atom.S | 323 palignr $1,(%esi), %xmm1 327 palignr $1,%xmm2, %xmm3 348 palignr $1,(%esi), %xmm0 352 palignr $1,16(%esi), %xmm3 362 palignr $1,48(%esi), %xmm3 365 palignr $1,32(%esi), %xmm0 409 palignr $2,(%esi), %xmm1 413 palignr $2,%xmm2, %xmm3 434 palignr $2,(%esi), %xmm0 438 palignr $2,16(%esi), %xmm [all...] |
ssse3-strcmp-atom.S | 353 palignr $1, %xmm3, %xmm2 377 palignr $1, %xmm3, %xmm2 459 palignr $2, %xmm3, %xmm2 483 palignr $2, %xmm3, %xmm2 566 palignr $3, %xmm3, %xmm2 590 palignr $3, %xmm3, %xmm2 672 palignr $4, %xmm3, %xmm2 697 palignr $4, %xmm3, %xmm2 781 palignr $5, %xmm3, %xmm2 805 palignr $5, %xmm3, %xmm [all...] |
/external/libvpx/libvpx/vp8/common/x86/ |
sad_ssse3.asm | 62 palignr xmm5, xmm4, %2 65 palignr xmm6, xmm4, (%2+1) 67 palignr xmm7, xmm4, (%2+2) 78 palignr xmm1, xmm4, %2 81 palignr xmm2, xmm4, (%2+1) 83 palignr xmm3, xmm4, (%2+2) 98 palignr xmm1, xmm4, %2 101 palignr xmm2, xmm4, (%2+1) 103 palignr xmm3, xmm4, (%2+2)
|
/external/libvpx/libvpx/vp9/encoder/x86/ |
vp9_sad_ssse3.asm | 62 palignr xmm5, xmm4, %2 65 palignr xmm6, xmm4, (%2+1) 67 palignr xmm7, xmm4, (%2+2) 78 palignr xmm1, xmm4, %2 81 palignr xmm2, xmm4, (%2+1) 83 palignr xmm3, xmm4, (%2+2) 98 palignr xmm1, xmm4, %2 101 palignr xmm2, xmm4, (%2+1) 103 palignr xmm3, xmm4, (%2+2)
|
/external/chromium_org/third_party/yasm/source/patched-yasm/modules/arch/x86/tests/ |
ssse3.asm | 70 TEST_ALIGNR palignr, movq, mm0, mmx, mm1 71 TEST_ALIGNR palignr, movdqu, xmm0, xmm, xmm1
|
ssse3.c | 221 CHECK_FUNCTIONS( palignr );
|
/external/libyuv/files/source/ |
rotate.cc | 87 palignr xmm1, xmm1, 8 93 palignr xmm3, xmm3, 8 99 palignr xmm5, xmm5, 8 104 palignr xmm7, xmm7, 8 110 palignr xmm2, xmm2, 8 111 palignr xmm3, xmm3, 8 116 palignr xmm6, xmm6, 8 117 palignr xmm7, xmm7, 8 123 palignr xmm4, xmm4, 8 128 palignr xmm6, xmm6, [all...] |
/external/llvm/test/MC/X86/ |
shuffle-comments.s | 3 palignr $8, %xmm0, %xmm1 label 5 palignr $8, (%rax), %xmm1 label 8 palignr $16, %xmm0, %xmm1 label 10 palignr $16, (%rax), %xmm1 label 13 palignr $0, %xmm0, %xmm1 label 15 palignr $0, (%rax), %xmm1 label
|
/external/chromium_org/third_party/openssl/openssl/crypto/aes/asm/ |
vpaes-x86.pl | 327 &palignr("xmm5","xmm5",12); 445 &palignr("xmm0","xmm6",8); 577 &palignr("xmm1","xmm2",15); 578 &palignr("xmm2","xmm2",15); 583 &palignr("xmm0","xmm0",1);
|
vpaes-x86_64.pl | 236 palignr \$12, %xmm5, %xmm5 362 palignr \$8,%xmm6,%xmm0 500 palignr \$15, %xmm8, %xmm1 501 palignr \$15, %xmm8, %xmm8 506 palignr \$1, %xmm0, %xmm0
|
aesni-sha1-x86_64.pl | 262 &palignr(@X[0],@X[-4&7],8); # compose "X[-14]" in "X[0]" 342 &palignr(@Tx[0],@X[-2&7],8); # compose "X[-6]"
|
/external/openssl/crypto/aes/asm/ |
vpaes-x86.pl | 327 &palignr("xmm5","xmm5",12); 445 &palignr("xmm0","xmm6",8); 577 &palignr("xmm1","xmm2",15); 578 &palignr("xmm2","xmm2",15); 583 &palignr("xmm0","xmm0",1);
|
vpaes-x86_64.pl | 236 palignr \$12, %xmm5, %xmm5 362 palignr \$8,%xmm6,%xmm0 500 palignr \$15, %xmm8, %xmm1 501 palignr \$15, %xmm8, %xmm8 506 palignr \$1, %xmm0, %xmm0
|
aesni-sha1-x86_64.pl | 262 &palignr(@X[0],@X[-4&7],8); # compose "X[-14]" in "X[0]" 342 &palignr(@Tx[0],@X[-2&7],8); # compose "X[-6]"
|
/external/chromium_org/third_party/openssl/openssl/crypto/perlasm/ |
x86asm.pl | 110 sub ::palignr 115 { &::generic("palignr",@_); }
|
/external/openssl/crypto/perlasm/ |
x86asm.pl | 110 sub ::palignr 115 { &::generic("palignr",@_); }
|
/external/qemu/target-i386/ |
ops_sse_header.h | 280 DEF_HELPER_3(glue(palignr, SUFFIX), void, Reg, Reg, s32)
|
/external/chromium_org/third_party/openssl/openssl/crypto/sha/asm/ |
sha1-586.pl | 533 &palignr(@X[0],@X[-4&7],8); # compose "X[-14]" in "X[0]" 618 &palignr(@X[2],@X[-2&7],8); # compose "X[-6]" [all...] |
sha1-x86_64.pl | 380 &palignr(@X[0],@X[-4&7],8); # compose "X[-14]" in "X[0]" 460 &palignr(@Tx[0],@X[-2&7],8); # compose "X[-6]"
|
/external/openssl/crypto/sha/asm/ |
sha1-586.pl | 533 &palignr(@X[0],@X[-4&7],8); # compose "X[-14]" in "X[0]" 618 &palignr(@X[2],@X[-2&7],8); # compose "X[-6]" [all...] |
sha1-x86_64.pl | 380 &palignr(@X[0],@X[-4&7],8); # compose "X[-14]" in "X[0]" 460 &palignr(@Tx[0],@X[-2&7],8); # compose "X[-6]"
|