/external/libvpx/libvpx/vp9/common/x86/ |
vp9_intrapred_ssse3.asm | 45 pshufb m1, m0 46 pshufb m2, m0 63 pshufb m1, m0 64 pshufb m2, m0 81 pshufb m1, m0 82 pshufb m2, m0 99 pshufb m1, m0 100 pshufb m2, m0 115 pshufb m2, m0, [GLOBAL(sh_b23456777)] 116 pshufb m1, m0, [GLOBAL(sh_b01234577) [all...] |
vp9_subpixel_8t_ssse3.asm | 538 pshufb %1, [GLOBAL(shuf_t0t1)] 539 pshufb %2, [GLOBAL(shuf_t2t3)] 635 pshufb %1, [GLOBAL(shuf_t0t1)] 636 pshufb %2, [GLOBAL(shuf_t2t3)] 637 pshufb %3, [GLOBAL(shuf_t4t5)] 638 pshufb %4, [GLOBAL(shuf_t6t7)] 779 pshufb xmm0, [GLOBAL(shuf_t0t1)] 780 pshufb xmm1, [GLOBAL(shuf_t2t3)] 781 pshufb xmm2, [GLOBAL(shuf_t4t5)] 782 pshufb xmm3, [GLOBAL(shuf_t6t7) [all...] |
/hardware/intel/common/omx-components/videocodec/libvpx_internal/libvpx/vp9/common/x86/ |
vp9_intrapred_ssse3.asm | 45 pshufb m1, m0 46 pshufb m2, m0 63 pshufb m1, m0 64 pshufb m2, m0 81 pshufb m1, m0 82 pshufb m2, m0 99 pshufb m1, m0 100 pshufb m2, m0 115 pshufb m2, m0, [GLOBAL(sh_b23456777)] 116 pshufb m1, m0, [GLOBAL(sh_b01234577) [all...] |
vp9_subpixel_8t_ssse3.asm | 537 pshufb %1, [GLOBAL(shuf_t0t1)] 538 pshufb %2, [GLOBAL(shuf_t2t3)] 634 pshufb %1, [GLOBAL(shuf_t0t1)] 635 pshufb %2, [GLOBAL(shuf_t2t3)] 636 pshufb %3, [GLOBAL(shuf_t4t5)] 637 pshufb %4, [GLOBAL(shuf_t6t7)] 778 pshufb xmm0, [GLOBAL(shuf_t0t1)] 779 pshufb xmm1, [GLOBAL(shuf_t2t3)] 780 pshufb xmm2, [GLOBAL(shuf_t4t5)] 781 pshufb xmm3, [GLOBAL(shuf_t6t7) [all...] |
/external/llvm/test/CodeGen/X86/ |
avx-trunc.ll | 13 ; CHECK: pshufb 19 ; CHECK: pshufb
|
pshufb-mask-comments.ll | 3 ; Test that the pshufb mask comment is correct. 7 ; CHECK: pshufb {{.*}}# xmm0 = xmm0[1,0,0,0,0,2,0,0,0,0,3,0,0,0,0,4] 16 ; CHECK: pshufb {{.*}}# xmm0 = xmm0[15,0,0,0,0,0,0,0,0,0,1,0,0,0,0,2] 25 ; CHECK: pshufb {{.*}}# xmm0 = xmm0[1,0,0,15,0,2,0,0],zero,xmm0[0,3,0,0],zero,xmm0[0,4] 34 ; CHECK: pshufb {{.*}} 42 ; CHECK: pshufb {{.*}}
|
atom-shuf.ll | 7 ; CHECK: pshufb
|
trunc-ext-ld-st.ll | 7 ;CHECK: pshufb 48 ;CHECK: pshufb 60 ;CHECK: pshufb 72 ;CHECK: pshufb
|
buildvec-insertvec.ll | 13 ; CHECK-NEXT: pshufb
|
vector-trunc.ll | 78 ; SSSE3-NEXT: pshufb %xmm2, %xmm1 79 ; SSSE3-NEXT: pshufb %xmm2, %xmm0 86 ; SSE41-NEXT: pshufb %xmm2, %xmm1 87 ; SSE41-NEXT: pshufb %xmm2, %xmm0 121 ; SSSE3-NEXT: pshufb {{.*#+}} xmm0 = xmm0[0,1,4,5,8,9,12,13,8,9,12,13,12,13,14,15] 127 ; SSE41-NEXT: pshufb {{.*#+}} xmm0 = xmm0[0,1,4,5,8,9,12,13,8,9,12,13,12,13,14,15] 158 ; SSSE3-NEXT: pshufb %xmm2, %xmm1 159 ; SSSE3-NEXT: pshufb %xmm2, %xmm0 166 ; SSE41-NEXT: pshufb %xmm2, %xmm1 167 ; SSE41-NEXT: pshufb %xmm2, %xmm [all...] |
extract-concat.ll | 15 ; CHECK: pshufb
|
vector-shuffle-128-v8.ll | 72 ; SSSE3-NEXT: pshufb {{.*#+}} xmm0 = xmm0[0,1,0,1,0,1,0,1,0,1,0,1,0,1,0,1] 77 ; SSE41-NEXT: pshufb {{.*#+}} xmm0 = xmm0[0,1,0,1,0,1,0,1,0,1,0,1,0,1,0,1] 158 ; SSSE3-NEXT: pshufb {{.*#+}} xmm0 = xmm0[8,9,8,9,8,9,8,9,0,1,0,1,0,1,0,1] 163 ; SSE41-NEXT: pshufb {{.*#+}} xmm0 = xmm0[8,9,8,9,8,9,8,9,0,1,0,1,0,1,0,1] 226 ; SSSE3-NEXT: pshufb {{.*#+}} xmm0 = xmm0[14,15,10,11,12,13,8,9,6,7,2,3,4,5,0,1] 231 ; SSE41-NEXT: pshufb {{.*#+}} xmm0 = xmm0[14,15,10,11,12,13,8,9,6,7,2,3,4,5,0,1] 252 ; SSSE3-NEXT: pshufb {{.*#+}} xmm0 = xmm0[2,3,0,1,10,11,8,9,10,11,8,9,2,3,0,1] 257 ; SSE41-NEXT: pshufb {{.*#+}} xmm0 = xmm0[2,3,0,1,10,11,8,9,10,11,8,9,2,3,0,1] 277 ; SSSE3-NEXT: pshufb {{.*#+}} xmm0 = xmm0[10,11,8,9,2,3,0,1,10,11,8,9,2,3,0,1] 282 ; SSE41-NEXT: pshufb {{.*#+}} xmm0 = xmm0[10,11,8,9,2,3,0,1,10,11,8,9,2,3,0,1 [all...] |
promote.ll | 25 ; CHECK: pshufb
|
vec_trunc_sext.ll | 8 ; There may be a better way to do this using pshufb + pmovsx,
|
vector-shuffle-128-v16.ll | 30 ; SSSE3-NEXT: pshufb %xmm1, %xmm0 36 ; SSE41-NEXT: pshufb %xmm1, %xmm0 64 ; SSSE3-NEXT: pshufb {{.*#+}} xmm0 = xmm0[0,0,0,0,0,0,0,0,1,1,1,1,1,1,1,1] 69 ; SSE41-NEXT: pshufb {{.*#+}} xmm0 = xmm0[0,0,0,0,0,0,0,0,1,1,1,1,1,1,1,1] 92 ; SSSE3-NEXT: pshufb {{.*#+}} xmm0 = xmm0[0,0,0,0,0,0,0,0,8,8,8,8,8,8,8,8] 97 ; SSE41-NEXT: pshufb {{.*#+}} xmm0 = xmm0[0,0,0,0,0,0,0,0,8,8,8,8,8,8,8,8] 153 ; SSSE3-NEXT: pshufb {{.*#+}} xmm0 = xmm0[0,0,0,0,4,4,4,4,8,8,8,8,12,12,12,12] 158 ; SSE41-NEXT: pshufb {{.*#+}} xmm0 = xmm0[0,0,0,0,4,4,4,4,8,8,8,8,12,12,12,12] 200 ; SSSE3-NEXT: pshufb {{.*#+}} xmm0 = xmm0[0,1,0,1,0,1,0,1,0,1,0,1,0,1,0,1] 205 ; SSE41-NEXT: pshufb {{.*#+}} xmm0 = xmm0[0,1,0,1,0,1,0,1,0,1,0,1,0,1,0,1 [all...] |
widen_shuffle-1.ll | 72 ; CHECK-NEXT: pshufb %xmm2, %xmm1 73 ; CHECK-NEXT: pshufb %xmm2, %xmm0 86 ; CHECK-NEXT: pshufb {{.*#+}} xmm0 = xmm0[0,2,4,6,8,10,12,14,u,u,u,u,u,u,u,u]
|
/external/boringssl/src/crypto/aes/asm/ |
vpaes-x86.pl | 44 # pshufb, yet it's respectable +28%/64% improvement on Core 2 189 &pshufb ("xmm2","xmm0"); 194 &pshufb ("xmm0","xmm1"); 204 &pshufb ("xmm4","xmm2"); # 4 = sb1u 205 &pshufb ("xmm0","xmm3"); # 0 = sb1t 210 &pshufb ("xmm5","xmm2"); # 4 = sb2u 213 &pshufb ("xmm2","xmm3"); # 2 = sb2t 216 &pshufb ("xmm0","xmm1"); # 0 = B 219 &pshufb ("xmm3","xmm4"); # 3 = D 222 &pshufb ("xmm0","xmm1"); # 0 = 2B+ [all...] |
vpaes-x86_64.pl | 44 # pshufb, yet it's respectable +36%/62% improvement on Core 2 95 pshufb %xmm0, %xmm2 97 pshufb %xmm1, %xmm0 109 pshufb %xmm2, %xmm4 # 4 = sb1u 110 pshufb %xmm3, %xmm0 # 0 = sb1t 115 pshufb %xmm2, %xmm5 # 4 = sb2u 118 pshufb %xmm3, %xmm2 # 2 = sb2t 121 pshufb %xmm1, %xmm0 # 0 = B 124 pshufb %xmm4, %xmm3 # 3 = D 127 pshufb %xmm1, %xmm0 # 0 = 2B+ [all...] |
/external/libvpx/libvpx/vp8/common/x86/ |
recon_sse2.asm | 390 pshufb xmm3, xmm2 412 pshufb xmm3, xmm2 413 pshufb xmm5, xmm2 549 pshufb xmm0, xmm2 550 pshufb xmm1, xmm2 573 pshufb xmm0, xmm2 574 pshufb xmm1, xmm2 670 ; FIXME use pshufb for ssse3 version 732 ; FIXME use pshufb for ssse3 version 825 ; FIXME use pshufb for ssse3 versio [all...] |
subpixel_ssse3.asm | 84 pshufb xmm1, [GLOBAL(shuf2bfrom1)] 86 pshufb xmm2, [GLOBAL(shuf3bfrom1)] 140 pshufb xmm0, xmm3 142 pshufb xmm2, xmm4 221 pshufb xmm1, [GLOBAL(shuf2bfrom1)] 223 pshufb xmm2, [GLOBAL(shuf3bfrom1)] 241 pshufb xmm1, [GLOBAL(shuf2bfrom1)] 242 pshufb xmm2, [GLOBAL(shuf3bfrom1)] 325 pshufb xmm0, [GLOBAL(shuf1b)] 328 pshufb xmm1, [GLOBAL(shuf2b) [all...] |
/hardware/intel/common/omx-components/videocodec/libvpx_internal/libvpx/vp8/common/x86/ |
recon_sse2.asm | 389 pshufb xmm3, xmm2 408 pshufb xmm3, xmm2 409 pshufb xmm5, xmm2 534 pshufb xmm0, xmm2 535 pshufb xmm1, xmm2 548 pshufb xmm0, xmm2 549 pshufb xmm1, xmm2 647 ; FIXME use pshufb for ssse3 version 709 ; FIXME use pshufb for ssse3 version 802 ; FIXME use pshufb for ssse3 versio [all...] |
subpixel_ssse3.asm | 84 pshufb xmm1, [GLOBAL(shuf2bfrom1)] 86 pshufb xmm2, [GLOBAL(shuf3bfrom1)] 140 pshufb xmm0, xmm3 142 pshufb xmm2, xmm4 221 pshufb xmm1, [GLOBAL(shuf2bfrom1)] 223 pshufb xmm2, [GLOBAL(shuf3bfrom1)] 241 pshufb xmm1, [GLOBAL(shuf2bfrom1)] 242 pshufb xmm2, [GLOBAL(shuf3bfrom1)] 325 pshufb xmm0, [GLOBAL(shuf1b)] 328 pshufb xmm1, [GLOBAL(shuf2b) [all...] |
/external/libvpx/libvpx/third_party/libyuv/source/ |
scale_win.cc | 437 pshufb xmm0, xmm3 438 pshufb xmm1, xmm4 439 pshufb xmm2, xmm5 489 pshufb xmm0, xmm2 498 pshufb xmm0, xmm3 508 pshufb xmm0, xmm4 549 pshufb xmm0, xmm2 559 pshufb xmm0, xmm3 570 pshufb xmm0, xmm4 605 pshufb xmm0, xmm [all...] |
/hardware/intel/common/omx-components/videocodec/libvpx_internal/libvpx/third_party/libyuv/source/ |
scale.c | 928 pshufb xmm0, xmm3 929 pshufb xmm1, xmm4 930 pshufb xmm2, xmm5 [all...] |
/hardware/intel/common/omx-components/videocodec/libvpx_internal/libvpx/vp8/encoder/x86/ |
quantize_ssse3.asm | 104 pshufb xmm2, [GLOBAL(zz_shuf)]
|