/external/libvpx/libvpx/vpx_dsp/x86/ |
intrapred_ssse3.asm | 45 pshufb m1, m0 46 pshufb m2, m0 63 pshufb m1, m0 64 pshufb m2, m0 81 pshufb m1, m0 82 pshufb m2, m0 99 pshufb m1, m0 100 pshufb m2, m0 115 pshufb m2, m0, [GLOBAL(sh_b23456777)] 116 pshufb m1, m0, [GLOBAL(sh_b01234577) [all...] |
/hardware/intel/common/omx-components/videocodec/libvpx_internal/libvpx/vp9/common/x86/ |
vp9_intrapred_ssse3.asm | 45 pshufb m1, m0 46 pshufb m2, m0 63 pshufb m1, m0 64 pshufb m2, m0 81 pshufb m1, m0 82 pshufb m2, m0 99 pshufb m1, m0 100 pshufb m2, m0 115 pshufb m2, m0, [GLOBAL(sh_b23456777)] 116 pshufb m1, m0, [GLOBAL(sh_b01234577) [all...] |
vp9_subpixel_8t_ssse3.asm | 537 pshufb %1, [GLOBAL(shuf_t0t1)] 538 pshufb %2, [GLOBAL(shuf_t2t3)] 634 pshufb %1, [GLOBAL(shuf_t0t1)] 635 pshufb %2, [GLOBAL(shuf_t2t3)] 636 pshufb %3, [GLOBAL(shuf_t4t5)] 637 pshufb %4, [GLOBAL(shuf_t6t7)] 778 pshufb xmm0, [GLOBAL(shuf_t0t1)] 779 pshufb xmm1, [GLOBAL(shuf_t2t3)] 780 pshufb xmm2, [GLOBAL(shuf_t4t5)] 781 pshufb xmm3, [GLOBAL(shuf_t6t7) [all...] |
/external/llvm/test/CodeGen/X86/ |
avx-trunc.ll | 13 ; CHECK: pshufb 19 ; CHECK: pshufb
|
pshufb-mask-comments.ll | 3 ; Test that the pshufb mask comment is correct. 7 ; CHECK: pshufb {{.*}}# xmm0 = xmm0[1,0,0,0,0,2,0,0,0,0,3,0,0,0,0,4] 16 ; CHECK: pshufb {{.*}}# xmm0 = xmm0[15,0,0,0,0,0,0,0,0,0,1,0,0,0,0,2] 25 ; CHECK: pshufb {{.*}}# xmm0 = xmm0[1,0,0,15,0,2,0,0],zero,xmm0[0,3,0,0],zero,xmm0[0,4] 34 ; CHECK: pshufb {{.*}} 42 ; CHECK: pshufb {{.*}}
|
atom-shuf.ll | 7 ; CHECK: pshufb
|
trunc-ext-ld-st.ll | 7 ;CHECK: pshufb 48 ;CHECK: pshufb 60 ;CHECK: pshufb 72 ;CHECK: pshufb
|
extract-concat.ll | 15 ; CHECK: pshufb
|
x86-fold-pshufb.ll | 19 ; The pshufb from function @pr24562 was wrongly folded into its first operand 22 ; Check that the pshufb is correctly folded to a zero vector.
|
promote.ll | 25 ; CHECK: pshufb
|
vec_trunc_sext.ll | 8 ; There may be a better way to do this using pshufb + pmovsx,
|
vector-popcnt-128.ll | 59 ; SSSE3-NEXT: pshufb %xmm2, %xmm4 62 ; SSSE3-NEXT: pshufb %xmm0, %xmm3 75 ; SSE41-NEXT: pshufb %xmm2, %xmm4 78 ; SSE41-NEXT: pshufb %xmm0, %xmm3 161 ; SSSE3-NEXT: pshufb %xmm3, %xmm4 164 ; SSSE3-NEXT: pshufb %xmm0, %xmm1 183 ; SSE41-NEXT: pshufb %xmm3, %xmm4 186 ; SSE41-NEXT: pshufb %xmm0, %xmm1 271 ; SSSE3-NEXT: pshufb %xmm2, %xmm4 274 ; SSSE3-NEXT: pshufb %xmm0, %xmm [all...] |
/external/boringssl/src/crypto/aes/asm/ |
vpaes-x86.pl | 44 # pshufb, yet it's respectable +28%/64% improvement on Core 2 189 &pshufb ("xmm2","xmm0"); 194 &pshufb ("xmm0","xmm1"); 204 &pshufb ("xmm4","xmm2"); # 4 = sb1u 205 &pshufb ("xmm0","xmm3"); # 0 = sb1t 210 &pshufb ("xmm5","xmm2"); # 4 = sb2u 213 &pshufb ("xmm2","xmm3"); # 2 = sb2t 216 &pshufb ("xmm0","xmm1"); # 0 = B 219 &pshufb ("xmm3","xmm4"); # 3 = D 222 &pshufb ("xmm0","xmm1"); # 0 = 2B+ [all...] |
vpaes-x86_64.pl | 44 # pshufb, yet it's respectable +36%/62% improvement on Core 2 95 pshufb %xmm0, %xmm2 97 pshufb %xmm1, %xmm0 109 pshufb %xmm2, %xmm4 # 4 = sb1u 110 pshufb %xmm3, %xmm0 # 0 = sb1t 115 pshufb %xmm2, %xmm5 # 4 = sb2u 118 pshufb %xmm3, %xmm2 # 2 = sb2t 121 pshufb %xmm1, %xmm0 # 0 = B 124 pshufb %xmm4, %xmm3 # 3 = D 127 pshufb %xmm1, %xmm0 # 0 = 2B+ [all...] |
/toolchain/binutils/binutils-2.25/gas/testsuite/gas/i386/ |
ssse3.s | 37 pshufb (%ecx),%mm0 38 pshufb %mm1,%mm0 39 pshufb (%ecx),%xmm0 40 pshufb %xmm1,%xmm0
|
x86-64-ssse3.s | 37 pshufb (%rcx),%mm0 38 pshufb %mm1,%mm0 39 pshufb (%rcx),%xmm0 40 pshufb %xmm1,%xmm0
|
ssse3.d | 41 90: 0f 38 00 01[ ]+pshufb \(%ecx\),%mm0 42 94: 0f 38 00 c1[ ]+pshufb %mm1,%mm0 43 98: 66 0f 38 00 01[ ]+pshufb \(%ecx\),%xmm0 44 9d: 66 0f 38 00 c1[ ]+pshufb %xmm1,%xmm0
|
x86-64-ssse3.d | 41 90: 0f 38 00 01[ ]+pshufb \(%rcx\),%mm0 42 94: 0f 38 00 c1[ ]+pshufb %mm1,%mm0 43 98: 66 0f 38 00 01[ ]+pshufb \(%rcx\),%xmm0 44 9d: 66 0f 38 00 c1[ ]+pshufb %xmm1,%xmm0
|
sse-noavx.s | 53 pshufb %mm1,%mm0
|
x86-64-sse-noavx.s | 54 pshufb %mm1,%mm0
|
/toolchain/binutils/binutils-2.25/gas/testsuite/gas/i386/ilp32/ |
x86-64-ssse3.d | 42 90: 0f 38 00 01[ ]+pshufb \(%rcx\),%mm0 43 94: 0f 38 00 c1[ ]+pshufb %mm1,%mm0 44 98: 66 0f 38 00 01[ ]+pshufb \(%rcx\),%xmm0 45 9d: 66 0f 38 00 c1[ ]+pshufb %xmm1,%xmm0
|
/hardware/intel/common/omx-components/videocodec/libvpx_internal/libvpx/vp8/common/x86/ |
recon_sse2.asm | 389 pshufb xmm3, xmm2 408 pshufb xmm3, xmm2 409 pshufb xmm5, xmm2 534 pshufb xmm0, xmm2 535 pshufb xmm1, xmm2 548 pshufb xmm0, xmm2 549 pshufb xmm1, xmm2 647 ; FIXME use pshufb for ssse3 version 709 ; FIXME use pshufb for ssse3 version 802 ; FIXME use pshufb for ssse3 versio [all...] |
/external/libvpx/libvpx/third_party/libyuv/source/ |
scale_gcc.cc | 289 "pshufb %%xmm3,%%xmm0 \n" 290 "pshufb %%xmm4,%%xmm1 \n" 291 "pshufb %%xmm5,%%xmm2 \n" 332 "pshufb %%xmm2,%%xmm6 \n" 341 "pshufb %%xmm3,%%xmm6 \n" 351 "pshufb %%xmm4,%%xmm6 \n" 399 "pshufb %%xmm2,%%xmm6 \n" 409 "pshufb %%xmm3,%%xmm6 \n" 420 "pshufb %%xmm4,%%xmm6 \n" 450 "pshufb %%xmm4,%%xmm0 \n [all...] |
scale_win.cc | 512 pshufb xmm0, xmm3 513 pshufb xmm1, xmm4 514 pshufb xmm2, xmm5 562 pshufb xmm0, xmm2 571 pshufb xmm0, xmm3 581 pshufb xmm0, xmm4 620 pshufb xmm0, xmm2 630 pshufb xmm0, xmm3 641 pshufb xmm0, xmm4 675 pshufb xmm0, xmm [all...] |
/external/libyuv/files/source/ |
scale_win.cc | 515 pshufb xmm0, xmm3 516 pshufb xmm1, xmm4 517 pshufb xmm2, xmm5 565 pshufb xmm0, xmm2 574 pshufb xmm0, xmm3 584 pshufb xmm0, xmm4 623 pshufb xmm0, xmm2 633 pshufb xmm0, xmm3 644 pshufb xmm0, xmm4 678 pshufb xmm0, xmm [all...] |