/external/libyuv/files/source/ |
row_posix.cc | 258 "palignr $0x8,%%xmm1,%%xmm2 \n" 261 "palignr $0xc,%%xmm0,%%xmm1 \n" 268 "palignr $0x4,%%xmm3,%%xmm3 \n" 299 "palignr $0x8,%%xmm1,%%xmm2 \n" 302 "palignr $0xc,%%xmm0,%%xmm1 \n" 309 "palignr $0x4,%%xmm3,%%xmm3 \n" [all...] |
scale.cc | 571 palignr xmm1, xmm0, 8 [all...] |
/external/chromium_org/third_party/libvpx/source/libvpx/third_party/libyuv/source/ |
row_posix.cc | 316 "palignr $0x8,%%xmm1,%%xmm2 \n" 319 "palignr $0xc,%%xmm0,%%xmm1 \n" 326 "palignr $0x4,%%xmm3,%%xmm3 \n" 357 "palignr $0x8,%%xmm1,%%xmm2 \n" 360 "palignr $0xc,%%xmm0,%%xmm1 \n" 367 "palignr $0x4,%%xmm3,%%xmm3 \n" [all...] |
row_win.cc | 384 palignr xmm2, xmm1, 8 // xmm2 = { xmm3[0:3] xmm1[8:15]} 387 palignr xmm1, xmm0, 12 // xmm1 = { xmm3[0:7] xmm0[12:15]} 394 palignr xmm3, xmm3, 4 // xmm3 = { xmm3[4:15]} 424 palignr xmm2, xmm1, 8 // xmm2 = { xmm3[0:3] xmm1[8:15]} 427 palignr xmm1, xmm0, 12 // xmm1 = { xmm3[0:7] xmm0[12:15]} 434 palignr xmm3, xmm3, 4 // xmm3 = { xmm3[4:15]} [all...] |
scale_win.cc | 436 palignr xmm1, xmm0, 8 [all...] |
x86inc.asm | 1000 AVX_INSTR palignr, 0, 1, 0
|
/external/chromium_org/third_party/libyuv/source/ |
row_posix.cc | 316 "palignr $0x8,%%xmm1,%%xmm2 \n" 319 "palignr $0xc,%%xmm0,%%xmm1 \n" 326 "palignr $0x4,%%xmm3,%%xmm3 \n" 357 "palignr $0x8,%%xmm1,%%xmm2 \n" 360 "palignr $0xc,%%xmm0,%%xmm1 \n" 367 "palignr $0x4,%%xmm3,%%xmm3 \n" [all...] |
row_win.cc | 385 palignr xmm2, xmm1, 8 // xmm2 = { xmm3[0:3] xmm1[8:15]} 388 palignr xmm1, xmm0, 12 // xmm1 = { xmm3[0:7] xmm0[12:15]} 395 palignr xmm3, xmm3, 4 // xmm3 = { xmm3[4:15]} 425 palignr xmm2, xmm1, 8 // xmm2 = { xmm3[0:3] xmm1[8:15]} 428 palignr xmm1, xmm0, 12 // xmm1 = { xmm3[0:7] xmm0[12:15]} 435 palignr xmm3, xmm3, 4 // xmm3 = { xmm3[4:15]} [all...] |
scale_win.cc | 436 palignr xmm1, xmm0, 8 [all...] |
x86inc.asm | 1000 AVX_INSTR palignr, 0, 1, 0
|
/external/elfutils/0.153/libcpu/defs/ |
i386 | [all...] |
/external/openssl/crypto/aes/asm/ |
aesni-sha1-x86_64.pl | 262 &palignr(@X[0],@X[-4&7],8); # compose "X[-14]" in "X[0]" 342 &palignr(@Tx[0],@X[-2&7],8); # compose "X[-6]"
|
/external/openssl/crypto/sha/asm/ |
sha1-586.pl | 533 &palignr(@X[0],@X[-4&7],8); # compose "X[-14]" in "X[0]" 618 &palignr(@X[2],@X[-2&7],8); # compose "X[-6]" [all...] |
sha1-x86_64.pl | 380 &palignr(@X[0],@X[-4&7],8); # compose "X[-14]" in "X[0]" 460 &palignr(@Tx[0],@X[-2&7],8); # compose "X[-6]"
|
/external/llvm/lib/Target/X86/ |
X86InstrFragmentsSIMD.td | 205 def X86PAlignr : SDNode<"X86ISD::PALIGNR", SDTShuff3OpI>;
|
X86ISelLowering.h | 323 PALIGNR, [all...] |
X86ISelLowering.cpp | [all...] |
/external/llvm/test/CodeGen/X86/ |
mmx-builtins.ll | [all...] |
/external/libvpx/libvpx/third_party/libyuv/source/ |
scale.c | 927 palignr xmm1, xmm0, 8 [all...] |
/hardware/intel/common/omx-components/videocodec/libvpx_internal/libvpx/third_party/libyuv/source/ |
scale.c | 927 palignr xmm1, xmm0, 8 [all...] |
/external/chromium_org/third_party/boringssl/src/crypto/modes/asm/ |
ghash-x86_64.pl | 528 palignr \$8,$T1,$T2 # low part is H.lo^H.hi... 546 palignr \$8,$T1,$T2 # low part is H^3.lo^H^3.hi... [all...] |
ghash-x86.pl | 968 &palignr ($T2,$T1,8); # low part is H.lo^H.hi [all...] |
/external/chromium_org/third_party/boringssl/src/crypto/sha/asm/ |
sha1-586.pl | 672 &punpcklqdq(@X[0],@X[-3&7]); # compose "X[-14]" in "X[0]", was &palignr(@X[0],@X[-4&7],8); 756 &punpcklqdq(@X[2],@X[-1&7]); # compose "X[-6]", was &palignr(@X[2],@X[-2&7],8) [all...] |
sha1-x86_64.pl | 558 &punpcklqdq(@X[0],@X[-3&7]); # compose "X[-14]" in "X[0]", was &palignr(@X[0],@X[-4&7],8); 639 &punpcklqdq(@Tx[0],@X[-1&7]); # compose "X[-6]", was &palignr(@Tx[0],@X[-2&7],8); [all...] |
/external/chromium_org/third_party/boringssl/src/crypto/perlasm/ |
x86_64-xlate.pl | 753 my $palignr = sub { [all...] |