HomeSort by relevance Sort by last modified time
    Searched full:vpalignr (Results 1 - 25 of 125) sorted by null

1 2 3 4 5

  /prebuilts/go/darwin-x86/src/vendor/golang_org/x/crypto/chacha20poly1305/
chacha20poly1305_amd64.s 161 // No PALIGNR in Go ASM yet (but VPALIGNR is present).
802 VPALIGNR $4, BB0, BB0, BB0; VPALIGNR $8, CC0, CC0, CC0; VPALIGNR $12, DD0, DD0, DD0
804 VPALIGNR $12, BB0, BB0, BB0; VPALIGNR $8, CC0, CC0, CC0; VPALIGNR $4, DD0, DD0, DD0
889 VPALIGNR $4, BB0, BB0, BB0; VPALIGNR $4, BB1, BB1, BB1; VPALIGNR $4, BB2, BB2, BB2; VPALIGNR $4, BB3, BB3, BB
    [all...]
  /prebuilts/go/linux-x86/src/vendor/golang_org/x/crypto/chacha20poly1305/
chacha20poly1305_amd64.s 161 // No PALIGNR in Go ASM yet (but VPALIGNR is present).
802 VPALIGNR $4, BB0, BB0, BB0; VPALIGNR $8, CC0, CC0, CC0; VPALIGNR $12, DD0, DD0, DD0
804 VPALIGNR $12, BB0, BB0, BB0; VPALIGNR $8, CC0, CC0, CC0; VPALIGNR $4, DD0, DD0, DD0
889 VPALIGNR $4, BB0, BB0, BB0; VPALIGNR $4, BB1, BB1, BB1; VPALIGNR $4, BB2, BB2, BB2; VPALIGNR $4, BB3, BB3, BB
    [all...]
  /external/boringssl/linux-x86_64/crypto/cipher_extra/
chacha20_poly1305_x86_64.S     [all...]
  /external/boringssl/mac-x86_64/crypto/cipher_extra/
chacha20_poly1305_x86_64.S     [all...]
  /external/llvm/test/MC/X86/
shuffle-comments.s 18 vpalignr $8, %xmm0, %xmm1, %xmm2 label
20 vpalignr $8, (%rax), %xmm1, %xmm2 label
23 vpalignr $16, %xmm0, %xmm1, %xmm2 label
25 vpalignr $16, (%rax), %xmm1, %xmm2 label
28 vpalignr $0, %xmm0, %xmm1, %xmm2 label
30 vpalignr $0, (%rax), %xmm1, %xmm2 label
33 vpalignr $8, %ymm0, %ymm1, %ymm2 label
35 vpalignr $8, (%rax), %ymm1, %ymm2 label
38 vpalignr $16, %ymm0, %ymm1, %ymm2 label
40 vpalignr $16, (%rax), %ymm1, %ymm label
43 vpalignr $0, %ymm0, %ymm1, %ymm2 label
45 vpalignr $0, (%rax), %ymm1, %ymm2 label
    [all...]
  /external/llvm/test/CodeGen/X86/
vector-shuffle-512-v64.ll 72 ; AVX512F-NEXT: vpalignr {{.*#+}} ymm0 = ymm2[15],ymm0[0,1,2,3,4,5,6,7,8,9,10,11,12,13,14],ymm2[31],ymm0[16,17,18,19,20,21,22,23,24,25,26,27,28,29,30]
73 ; AVX512F-NEXT: vpalignr {{.*#+}} ymm1 = ymm3[15],ymm1[0,1,2,3,4,5,6,7,8,9,10,11,12,13,14],ymm3[31],ymm1[16,17,18,19,20,21,22,23,24,25,26,27,28,29,30]
78 ; AVX512BW-NEXT: vpalignr {{.*#+}} zmm0 = zmm1[15],zmm0[0,1,2,3,4,5,6,7,8,9,10,11,12,13,14],zmm1[31],zmm0[16,17,18,19,20,21,22,23,24,25,26,27,28,29,30],zmm1[47],zmm0[32,33,34,35,36,37,38,39,40,41,42,43,44,45,46],zmm1[63],zmm0[48,49,50,51,52,53,54,55,56,57,58,59,60,61,62]
83 ; AVX512DQ-NEXT: vpalignr {{.*#+}} ymm0 = ymm2[15],ymm0[0,1,2,3,4,5,6,7,8,9,10,11,12,13,14],ymm2[31],ymm0[16,17,18,19,20,21,22,23,24,25,26,27,28,29,30]
84 ; AVX512DQ-NEXT: vpalignr {{.*#+}} ymm1 = ymm3[15],ymm1[0,1,2,3,4,5,6,7,8,9,10,11,12,13,14],ymm3[31],ymm1[16,17,18,19,20,21,22,23,24,25,26,27,28,29,30]
vector-shuffle-256-v16.ll     [all...]
  /external/boringssl/win-x86_64/crypto/fipsmodule/
sha512-x86_64.asm 1906 vpalignr xmm8,xmm1,xmm0,8
1909 vpalignr xmm11,xmm5,xmm4,8
1975 vpalignr xmm8,xmm2,xmm1,8
1978 vpalignr xmm11,xmm6,xmm5,8
2044 vpalignr xmm8,xmm3,xmm2,8
2047 vpalignr xmm11,xmm7,xmm6,8
2113 vpalignr xmm8,xmm4,xmm3,8
2116 vpalignr xmm11,xmm0,xmm7,8
2182 vpalignr xmm8,xmm5,xmm4,8
2185 vpalignr xmm11,xmm1,xmm0,
    [all...]
aesni-gcm-x86_64.asm 194 vpalignr xmm0,xmm4,xmm4,8
279 vpalignr xmm8,xmm4,xmm4,8
792 vpalignr xmm14,xmm8,xmm8,8
817 vpalignr xmm14,xmm8,xmm8,8
852 vpalignr xmm2,xmm8,xmm8,8
856 vpalignr xmm2,xmm8,xmm8,8
sha1-x86_64.asm     [all...]
  /toolchain/binutils/binutils-2.27/gas/testsuite/gas/i386/
avx512bw-wig.s 96 vpalignr $0xab, %zmm4, %zmm5, %zmm6 # AVX512BW
97 vpalignr $0xab, %zmm4, %zmm5, %zmm6{%k7} # AVX512BW
98 vpalignr $0xab, %zmm4, %zmm5, %zmm6{%k7}{z} # AVX512BW
99 vpalignr $123, %zmm4, %zmm5, %zmm6 # AVX512BW
100 vpalignr $123, (%ecx), %zmm5, %zmm6 # AVX512BW
101 vpalignr $123, -123456(%esp,%esi,8), %zmm5, %zmm6 # AVX512BW
102 vpalignr $123, 8128(%edx), %zmm5, %zmm6 # AVX512BW Disp8
103 vpalignr $123, 8192(%edx), %zmm5, %zmm6 # AVX512BW
104 vpalignr $123, -8192(%edx), %zmm5, %zmm6 # AVX512BW Disp8
105 vpalignr $123, -8256(%edx), %zmm5, %zmm6 # AVX512B
    [all...]
x86-64-avx512bw-wig.s 96 vpalignr $0xab, %zmm28, %zmm29, %zmm30 # AVX512BW
97 vpalignr $0xab, %zmm28, %zmm29, %zmm30{%k7} # AVX512BW
98 vpalignr $0xab, %zmm28, %zmm29, %zmm30{%k7}{z} # AVX512BW
99 vpalignr $123, %zmm28, %zmm29, %zmm30 # AVX512BW
100 vpalignr $123, (%rcx), %zmm29, %zmm30 # AVX512BW
101 vpalignr $123, 0x123(%rax,%r14,8), %zmm29, %zmm30 # AVX512BW
102 vpalignr $123, 8128(%rdx), %zmm29, %zmm30 # AVX512BW Disp8
103 vpalignr $123, 8192(%rdx), %zmm29, %zmm30 # AVX512BW
104 vpalignr $123, -8192(%rdx), %zmm29, %zmm30 # AVX512BW Disp8
105 vpalignr $123, -8256(%rdx), %zmm29, %zmm30 # AVX512B
    [all...]
avx512bw_vl-wig.s 166 vpalignr $0xab, %xmm4, %xmm5, %xmm6{%k7} # AVX512{BW,VL}
167 vpalignr $0xab, %xmm4, %xmm5, %xmm6{%k7}{z} # AVX512{BW,VL}
168 vpalignr $123, %xmm4, %xmm5, %xmm6{%k7} # AVX512{BW,VL}
169 vpalignr $123, (%ecx), %xmm5, %xmm6{%k7} # AVX512{BW,VL}
170 vpalignr $123, -123456(%esp,%esi,8), %xmm5, %xmm6{%k7} # AVX512{BW,VL}
171 vpalignr $123, 2032(%edx), %xmm5, %xmm6{%k7} # AVX512{BW,VL} Disp8
172 vpalignr $123, 2048(%edx), %xmm5, %xmm6{%k7} # AVX512{BW,VL}
173 vpalignr $123, -2048(%edx), %xmm5, %xmm6{%k7} # AVX512{BW,VL} Disp8
174 vpalignr $123, -2064(%edx), %xmm5, %xmm6{%k7} # AVX512{BW,VL}
175 vpalignr $0xab, %ymm4, %ymm5, %ymm6{%k7} # AVX512{BW,VL
    [all...]
x86-64-avx512bw_vl-wig.s 186 vpalignr $0xab, %xmm28, %xmm29, %xmm30 # AVX512{BW,VL}
187 vpalignr $0xab, %xmm28, %xmm29, %xmm30{%k7} # AVX512{BW,VL}
188 vpalignr $0xab, %xmm28, %xmm29, %xmm30{%k7}{z} # AVX512{BW,VL}
189 vpalignr $123, %xmm28, %xmm29, %xmm30 # AVX512{BW,VL}
190 vpalignr $123, (%rcx), %xmm29, %xmm30 # AVX512{BW,VL}
191 vpalignr $123, 0x123(%rax,%r14,8), %xmm29, %xmm30 # AVX512{BW,VL}
192 vpalignr $123, 2032(%rdx), %xmm29, %xmm30 # AVX512{BW,VL} Disp8
193 vpalignr $123, 2048(%rdx), %xmm29, %xmm30 # AVX512{BW,VL}
194 vpalignr $123, -2048(%rdx), %xmm29, %xmm30 # AVX512{BW,VL} Disp8
195 vpalignr $123, -2064(%rdx), %xmm29, %xmm30 # AVX512{BW,VL
    [all...]
avx256int.s 191 vpalignr $7,%ymm4,%ymm6,%ymm2
192 vpalignr $7,(%ecx),%ymm6,%ymm2
515 vpalignr ymm2,ymm6,ymm4,7
516 vpalignr ymm2,ymm6,YMMWORD PTR [ecx],7
517 vpalignr ymm2,ymm6,[ecx],7
avx512bw.s 124 vpalignr $0xab, %zmm4, %zmm5, %zmm6 # AVX512BW
125 vpalignr $0xab, %zmm4, %zmm5, %zmm6{%k7} # AVX512BW
126 vpalignr $0xab, %zmm4, %zmm5, %zmm6{%k7}{z} # AVX512BW
127 vpalignr $123, %zmm4, %zmm5, %zmm6 # AVX512BW
128 vpalignr $123, (%ecx), %zmm5, %zmm6 # AVX512BW
129 vpalignr $123, -123456(%esp,%esi,8), %zmm5, %zmm6 # AVX512BW
130 vpalignr $123, 8128(%edx), %zmm5, %zmm6 # AVX512BW Disp8
131 vpalignr $123, 8192(%edx), %zmm5, %zmm6 # AVX512BW
132 vpalignr $123, -8192(%edx), %zmm5, %zmm6 # AVX512BW Disp8
133 vpalignr $123, -8256(%edx), %zmm5, %zmm6 # AVX512B
    [all...]
x86-64-avx256int.s 194 vpalignr $7,%ymm4,%ymm6,%ymm2
195 vpalignr $7,(%rcx),%ymm6,%ymm2
521 vpalignr ymm2,ymm6,ymm4,7
522 vpalignr ymm2,ymm6,YMMWORD PTR [rcx],7
523 vpalignr ymm2,ymm6,[rcx],7
x86-64-avx512bw.s 124 vpalignr $0xab, %zmm28, %zmm29, %zmm30 # AVX512BW
125 vpalignr $0xab, %zmm28, %zmm29, %zmm30{%k7} # AVX512BW
126 vpalignr $0xab, %zmm28, %zmm29, %zmm30{%k7}{z} # AVX512BW
127 vpalignr $123, %zmm28, %zmm29, %zmm30 # AVX512BW
128 vpalignr $123, (%rcx), %zmm29, %zmm30 # AVX512BW
129 vpalignr $123, 0x123(%rax,%r14,8), %zmm29, %zmm30 # AVX512BW
130 vpalignr $123, 8128(%rdx), %zmm29, %zmm30 # AVX512BW Disp8
131 vpalignr $123, 8192(%rdx), %zmm29, %zmm30 # AVX512BW
132 vpalignr $123, -8192(%rdx), %zmm29, %zmm30 # AVX512BW Disp8
133 vpalignr $123, -8256(%rdx), %zmm29, %zmm30 # AVX512B
    [all...]
x86-64-avx512bw_vl.s 242 vpalignr $0xab, %xmm28, %xmm29, %xmm30 # AVX512{BW,VL}
243 vpalignr $0xab, %xmm28, %xmm29, %xmm30{%k7} # AVX512{BW,VL}
244 vpalignr $0xab, %xmm28, %xmm29, %xmm30{%k7}{z} # AVX512{BW,VL}
245 vpalignr $123, %xmm28, %xmm29, %xmm30 # AVX512{BW,VL}
246 vpalignr $123, (%rcx), %xmm29, %xmm30 # AVX512{BW,VL}
247 vpalignr $123, 0x123(%rax,%r14,8), %xmm29, %xmm30 # AVX512{BW,VL}
248 vpalignr $123, 2032(%rdx), %xmm29, %xmm30 # AVX512{BW,VL} Disp8
249 vpalignr $123, 2048(%rdx), %xmm29, %xmm30 # AVX512{BW,VL}
250 vpalignr $123, -2048(%rdx), %xmm29, %xmm30 # AVX512{BW,VL} Disp8
251 vpalignr $123, -2064(%rdx), %xmm29, %xmm30 # AVX512{BW,VL
    [all...]
  /external/boringssl/linux-x86_64/crypto/fipsmodule/
sha512-x86_64.S 1874 vpalignr $8,%xmm0,%xmm1,%xmm8
1877 vpalignr $8,%xmm4,%xmm5,%xmm11
1943 vpalignr $8,%xmm1,%xmm2,%xmm8
1946 vpalignr $8,%xmm5,%xmm6,%xmm11
2012 vpalignr $8,%xmm2,%xmm3,%xmm8
2015 vpalignr $8,%xmm6,%xmm7,%xmm11
2081 vpalignr $8,%xmm3,%xmm4,%xmm8
2084 vpalignr $8,%xmm7,%xmm0,%xmm11
    [all...]
aesni-gcm-x86_64.S 190 vpalignr $8,%xmm4,%xmm4,%xmm0
275 vpalignr $8,%xmm4,%xmm4,%xmm8
730 vpalignr $8,%xmm8,%xmm8,%xmm14
755 vpalignr $8,%xmm8,%xmm8,%xmm14
790 vpalignr $8,%xmm8,%xmm8,%xmm2
794 vpalignr $8,%xmm8,%xmm8,%xmm2
  /external/boringssl/mac-x86_64/crypto/fipsmodule/
sha512-x86_64.S 1873 vpalignr $8,%xmm0,%xmm1,%xmm8
1876 vpalignr $8,%xmm4,%xmm5,%xmm11
1942 vpalignr $8,%xmm1,%xmm2,%xmm8
1945 vpalignr $8,%xmm5,%xmm6,%xmm11
2011 vpalignr $8,%xmm2,%xmm3,%xmm8
2014 vpalignr $8,%xmm6,%xmm7,%xmm11
2080 vpalignr $8,%xmm3,%xmm4,%xmm8
2083 vpalignr $8,%xmm7,%xmm0,%xmm11
2149 vpalignr $8,%xmm4,%xmm5,%xmm8
    [all...]
aesni-gcm-x86_64.S 190 vpalignr $8,%xmm4,%xmm4,%xmm0
275 vpalignr $8,%xmm4,%xmm4,%xmm8
730 vpalignr $8,%xmm8,%xmm8,%xmm14
755 vpalignr $8,%xmm8,%xmm8,%xmm14
790 vpalignr $8,%xmm8,%xmm8,%xmm2
794 vpalignr $8,%xmm8,%xmm8,%xmm2
  /external/boringssl/src/crypto/cipher_extra/asm/
chacha20_poly1305_x86_64.pl     [all...]
  /external/boringssl/win-x86/crypto/fipsmodule/
sha1-586.asm     [all...]

Completed in 303 milliseconds

1 2 3 4 5