/external/llvm/test/MC/X86/ |
shuffle-comments.s | 18 vpalignr $8, %xmm0, %xmm1, %xmm2 label 20 vpalignr $8, (%rax), %xmm1, %xmm2 label 23 vpalignr $16, %xmm0, %xmm1, %xmm2 label 25 vpalignr $16, (%rax), %xmm1, %xmm2 label 28 vpalignr $0, %xmm0, %xmm1, %xmm2 label 30 vpalignr $0, (%rax), %xmm1, %xmm2 label 33 vpalignr $8, %ymm0, %ymm1, %ymm2 label 35 vpalignr $8, (%rax), %ymm1, %ymm2 label 38 vpalignr $16, %ymm0, %ymm1, %ymm2 label 40 vpalignr $16, (%rax), %ymm1, %ymm label 43 vpalignr $0, %ymm0, %ymm1, %ymm2 label 45 vpalignr $0, (%rax), %ymm1, %ymm2 label [all...] |
x86-64-avx512bw_vl.s | [all...] |
x86-64-avx512bw.s | [all...] |
/external/swiftshader/third_party/llvm-7.0/llvm/test/MC/X86/ |
shuffle-comments.s | 18 vpalignr $8, %xmm0, %xmm1, %xmm2 label 20 vpalignr $8, (%rax), %xmm1, %xmm2 label 23 vpalignr $16, %xmm0, %xmm1, %xmm2 label 25 vpalignr $16, (%rax), %xmm1, %xmm2 label 28 vpalignr $0, %xmm0, %xmm1, %xmm2 label 30 vpalignr $0, (%rax), %xmm1, %xmm2 label 33 vpalignr $8, %ymm0, %ymm1, %ymm2 label 35 vpalignr $8, (%rax), %ymm1, %ymm2 label 38 vpalignr $16, %ymm0, %ymm1, %ymm2 label 40 vpalignr $16, (%rax), %ymm1, %ymm label 43 vpalignr $0, %ymm0, %ymm1, %ymm2 label 45 vpalignr $0, (%rax), %ymm1, %ymm2 label [all...] |
x86-64-avx512bw_vl.s | [all...] |
x86-64-avx512bw.s | [all...] |
AVX2-64.s | 1133 vpalignr $0, 485498096, %ymm7, %ymm7 label 1137 vpalignr $0, 485498096, %ymm9, %ymm9 label 1141 vpalignr $0, -64(%rdx,%rax,4), %ymm7, %ymm7 label 1145 vpalignr $0, 64(%rdx,%rax,4), %ymm7, %ymm7 label 1149 vpalignr $0, -64(%rdx,%rax,4), %ymm9, %ymm9 label 1153 vpalignr $0, 64(%rdx,%rax,4), %ymm9, %ymm9 label 1157 vpalignr $0, 64(%rdx,%rax), %ymm7, %ymm7 label 1161 vpalignr $0, 64(%rdx,%rax), %ymm9, %ymm9 label 1165 vpalignr $0, 64(%rdx), %ymm7, %ymm7 label 1169 vpalignr $0, 64(%rdx), %ymm9, %ymm9 label 1173 vpalignr $0, (%rdx), %ymm7, %ymm7 label 1177 vpalignr $0, (%rdx), %ymm9, %ymm9 label 1181 vpalignr $0, %ymm7, %ymm7, %ymm7 label 1185 vpalignr $0, %ymm9, %ymm9, %ymm9 label [all...] |
AVX2-32.s | 567 // CHECK: vpalignr $0, -485498096(%edx,%eax,4), %ymm4, %ymm4 569 vpalignr $0, -485498096(%edx,%eax,4), %ymm4, %ymm4 label 571 // CHECK: vpalignr $0, 485498096(%edx,%eax,4), %ymm4, %ymm4 573 vpalignr $0, 485498096(%edx,%eax,4), %ymm4, %ymm4 label 575 // CHECK: vpalignr $0, 485498096(%edx), %ymm4, %ymm4 577 vpalignr $0, 485498096(%edx), %ymm4, %ymm4 label 579 // CHECK: vpalignr $0, 485498096, %ymm4, %ymm4 581 vpalignr $0, 485498096, %ymm4, %ymm4 label 583 // CHECK: vpalignr $0, 64(%edx,%eax), %ymm4, %ymm4 585 vpalignr $0, 64(%edx,%eax), %ymm4, %ymm4 label 589 vpalignr $0, (%edx), %ymm4, %ymm4 label 593 vpalignr $0, %ymm4, %ymm4, %ymm4 label [all...] |
/external/boringssl/linux-x86_64/crypto/cipher_extra/ |
chacha20_poly1305_x86_64.S | [all...] |
/external/boringssl/mac-x86_64/crypto/cipher_extra/ |
chacha20_poly1305_x86_64.S | [all...] |
/external/libjpeg-turbo/simd/i386/ |
jdsample-avx2.asm | 128 vpalignr ymm2, ymm1, ymm2, 15 ; ymm2=(-- 0 1 ... 28 29 30) 130 vpalignr ymm3, ymm4, ymm1, 1 ; ymm3=( 1 2 3 ... 30 31 --) 417 vpalignr ymm0, ymm0, ymm7, 2 ; ymm0=( 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 --) 424 vpalignr ymm6, ymm3, ymm6, 14 ; ymm6=(-- 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30) 430 vpalignr ymm2, ymm2, ymm3, 2 ; ymm2=(17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 --) 434 vpalignr ymm1, ymm7, ymm1, 14 ; ymm1=(-- 0 1 2 3 4 5 6 7 8 9 10 11 12 13 14) 473 vpalignr ymm7, ymm7, ymm6, 2 ; ymm7=( 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 --) 480 vpalignr ymm2, ymm4, ymm2, 14 ; ymm2=(-- 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30) 486 vpalignr ymm5, ymm5, ymm4, 2 ; ymm5=(17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 --) 490 vpalignr ymm1, ymm6, ymm1, 14 ; ymm1=(-- 0 1 2 3 4 5 6 7 8 9 10 11 12 13 14 [all...] |
/external/libjpeg-turbo/simd/x86_64/ |
jdsample-avx2.asm | 123 vpalignr ymm2, ymm1, ymm2, 15 ; ymm2=(-- 0 1 ... 28 29 30) 125 vpalignr ymm3, ymm4, ymm1, 1 ; ymm3=( 1 2 3 ... 30 31 --) 376 vpalignr ymm0, ymm0, ymm7, 2 ; ymm0=( 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 --) 383 vpalignr ymm6, ymm3, ymm6, 14 ; ymm6=(-- 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30) 389 vpalignr ymm2, ymm2, ymm3, 2 ; ymm2=(17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 --) 393 vpalignr ymm1, ymm7, ymm1, 14 ; ymm1=(-- 0 1 2 3 4 5 6 7 8 9 10 11 12 13 14) 430 vpalignr ymm7, ymm7, ymm6, 2 ; ymm7=( 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 --) 437 vpalignr ymm2, ymm4, ymm2, 14 ; ymm2=(-- 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30) 443 vpalignr ymm5, ymm5, ymm4, 2 ; ymm5=(17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 --) 447 vpalignr ymm1, ymm6, ymm1, 14 ; ymm1=(-- 0 1 2 3 4 5 6 7 8 9 10 11 12 13 14 [all...] |
/external/boringssl/linux-x86_64/crypto/fipsmodule/ |
aesni-gcm-x86_64.S | 202 vpalignr $8,%xmm4,%xmm4,%xmm0 287 vpalignr $8,%xmm4,%xmm4,%xmm8 749 vpalignr $8,%xmm8,%xmm8,%xmm14 774 vpalignr $8,%xmm8,%xmm8,%xmm14 809 vpalignr $8,%xmm8,%xmm8,%xmm2 813 vpalignr $8,%xmm8,%xmm8,%xmm2
|
sha512-x86_64.S | 1911 vpalignr $8,%xmm0,%xmm1,%xmm8 1914 vpalignr $8,%xmm4,%xmm5,%xmm11 1988 vpalignr $8,%xmm1,%xmm2,%xmm8 1991 vpalignr $8,%xmm5,%xmm6,%xmm11 2065 vpalignr $8,%xmm2,%xmm3,%xmm8 2068 vpalignr $8,%xmm6,%xmm7,%xmm11 [all...] |
sha1-x86_64.S | [all...] |
ghash-x86_64.S | 1351 vpalignr $8,%xmm3,%xmm4,%xmm5 1423 vpalignr $8,%xmm4,%xmm3,%xmm5 1599 vpalignr $8,%xmm10,%xmm10,%xmm12 1639 vpalignr $8,%xmm10,%xmm10,%xmm12 [all...] |
/external/boringssl/mac-x86_64/crypto/fipsmodule/ |
aesni-gcm-x86_64.S | 202 vpalignr $8,%xmm4,%xmm4,%xmm0 287 vpalignr $8,%xmm4,%xmm4,%xmm8 748 vpalignr $8,%xmm8,%xmm8,%xmm14 773 vpalignr $8,%xmm8,%xmm8,%xmm14 808 vpalignr $8,%xmm8,%xmm8,%xmm2 812 vpalignr $8,%xmm8,%xmm8,%xmm2
|
sha512-x86_64.S | 1910 vpalignr $8,%xmm0,%xmm1,%xmm8 1913 vpalignr $8,%xmm4,%xmm5,%xmm11 1987 vpalignr $8,%xmm1,%xmm2,%xmm8 1990 vpalignr $8,%xmm5,%xmm6,%xmm11 2064 vpalignr $8,%xmm2,%xmm3,%xmm8 2067 vpalignr $8,%xmm6,%xmm7,%xmm11 2141 vpalignr $8,%xmm3,%xmm4,%xmm8 2144 vpalignr $8,%xmm7,%xmm0,%xmm11 [all...] |
sha1-x86_64.S | [all...] |
ghash-x86_64.S | 1338 vpalignr $8,%xmm3,%xmm4,%xmm5 1410 vpalignr $8,%xmm4,%xmm3,%xmm5 1586 vpalignr $8,%xmm10,%xmm10,%xmm12 1626 vpalignr $8,%xmm10,%xmm10,%xmm12 [all...] |
/external/boringssl/src/crypto/fipsmodule/modes/asm/ |
aesni-gcm-x86_64.pl | 272 vpalignr \$8,$Z0,$Z0,$Ii # 1st phase 357 vpalignr \$8,$Z0,$Z0,$Xi # 2nd phase 876 vpalignr \$8,$Xi,$Xi,$inout5 # 1st phase 901 vpalignr \$8,$Xi,$Xi,$inout5 # 2nd phase 936 vpalignr \$8,$Xi,$Xi,$T2 # 1st phase 940 vpalignr \$8,$Xi,$Xi,$T2 # 2nd phase
|
/external/boringssl/src/crypto/cipher_extra/asm/ |
aes128gcmsiv-x86_64.pl | 333 vpalignr \$8, $T, $T, $T 346 vpalignr \$8, $T, $T, $T 380 vpalignr \$8, $T, $T, $T 384 vpalignr \$8, $T, $T, $T [all...] |
chacha20_poly1305_x86_64.pl | [all...] |
/external/boringssl/win-x86_64/crypto/fipsmodule/ |
sha512-x86_64.asm | 1938 vpalignr xmm8,xmm1,xmm0,8 1941 vpalignr xmm11,xmm5,xmm4,8 2015 vpalignr xmm8,xmm2,xmm1,8 2018 vpalignr xmm11,xmm6,xmm5,8 2092 vpalignr xmm8,xmm3,xmm2,8 2095 vpalignr xmm11,xmm7,xmm6,8 2169 vpalignr xmm8,xmm4,xmm3,8 2172 vpalignr xmm11,xmm0,xmm7,8 2246 vpalignr xmm8,xmm5,xmm4,8 2249 vpalignr xmm11,xmm1,xmm0, [all...] |
aesni-gcm-x86_64.asm | 201 vpalignr xmm0,xmm4,xmm4,8 286 vpalignr xmm8,xmm4,xmm4,8 805 vpalignr xmm14,xmm8,xmm8,8 830 vpalignr xmm14,xmm8,xmm8,8 865 vpalignr xmm2,xmm8,xmm8,8 869 vpalignr xmm2,xmm8,xmm8,8
|