/toolchain/binutils/binutils-2.25/gas/testsuite/gas/i386/ |
x86-64-avx-swap.s | 10 vmovdqa %ymm8,%ymm6 27 vmovdqa %xmm8,%xmm6 42 vmovdqa ymm6,ymm8 59 vmovdqa xmm6,xmm8
|
x86-64-avx-swap-intel.d | 14 [ ]*[a-f0-9]+: c5 7d 7f c6 vmovdqa ymm6,ymm8 20 [ ]*[a-f0-9]+: c5 79 7f c6 vmovdqa xmm6,xmm8 29 [ ]*[a-f0-9]+: c5 79 7f c6 vmovdqa xmm6,xmm8 38 [ ]*[a-f0-9]+: c5 7d 7f c6 vmovdqa ymm6,ymm8 44 [ ]*[a-f0-9]+: c5 79 7f c6 vmovdqa xmm6,xmm8 53 [ ]*[a-f0-9]+: c5 79 7f c6 vmovdqa xmm6,xmm8
|
x86-64-avx-swap.d | 13 [ ]*[a-f0-9]+: c5 7d 7f c6 vmovdqa %ymm8,%ymm6 19 [ ]*[a-f0-9]+: c5 79 7f c6 vmovdqa %xmm8,%xmm6 28 [ ]*[a-f0-9]+: c5 79 7f c6 vmovdqa %xmm8,%xmm6 37 [ ]*[a-f0-9]+: c5 7d 7f c6 vmovdqa %ymm8,%ymm6 43 [ ]*[a-f0-9]+: c5 79 7f c6 vmovdqa %xmm8,%xmm6 52 [ ]*[a-f0-9]+: c5 79 7f c6 vmovdqa %xmm8,%xmm6
|
x86-64-avx.s | 244 vmovdqa %ymm4,%ymm6 245 vmovdqa (%rcx),%ymm4 278 vmovdqa %ymm4,%ymm6 279 vmovdqa %ymm4,(%rcx) 347 vmovdqa %xmm4,%xmm6 348 vmovdqa (%rcx),%xmm4 389 vmovdqa %xmm4,%xmm6 390 vmovdqa %xmm4,(%rcx) [all...] |
avx.s | 244 vmovdqa %ymm4,%ymm6 245 vmovdqa (%ecx),%ymm4 278 vmovdqa %ymm4,%ymm6 279 vmovdqa %ymm4,(%ecx) 347 vmovdqa %xmm4,%xmm6 348 vmovdqa (%ecx),%xmm4 389 vmovdqa %xmm4,%xmm6 390 vmovdqa %xmm4,(%ecx) 1223 vmovdqa 0x1234,%xmm [all...] |
opts.s | 122 vmovdqa %ymm4,%ymm6 123 vmovdqa.s %ymm4,%ymm6 154 vmovdqa %xmm4,%xmm6 155 vmovdqa.s %xmm4,%xmm6 238 vmovdqa ymm6,ymm4 239 vmovdqa.s ymm6,ymm4 270 vmovdqa xmm6,xmm4 271 vmovdqa.s xmm6,xmm4
|
x86-64-opts.s | 158 vmovdqa %ymm4,%ymm6 159 vmovdqa.s %ymm4,%ymm6 190 vmovdqa %xmm4,%xmm6 191 vmovdqa.s %xmm4,%xmm6 292 vmovdqa ymm6,ymm4 293 vmovdqa.s ymm6,ymm4 324 vmovdqa xmm6,xmm4 325 vmovdqa.s xmm6,xmm4
|
sse2avx-opts-intel.d | 124 [ ]*[a-f0-9]+: c5 fd 6f f4 vmovdqa ymm6,ymm4 125 [ ]*[a-f0-9]+: c5 fd 7f e6 vmovdqa.s ymm6,ymm4 136 [ ]*[a-f0-9]+: c5 f9 6f f4 vmovdqa xmm6,xmm4 137 [ ]*[a-f0-9]+: c5 f9 7f e6 vmovdqa.s xmm6,xmm4 154 [ ]*[a-f0-9]+: c5 f9 6f f4 vmovdqa xmm6,xmm4 155 [ ]*[a-f0-9]+: c5 f9 7f e6 vmovdqa.s xmm6,xmm4 228 [ ]*[a-f0-9]+: c5 fd 6f f4 vmovdqa ymm6,ymm4 229 [ ]*[a-f0-9]+: c5 fd 7f e6 vmovdqa.s ymm6,ymm4 240 [ ]*[a-f0-9]+: c5 f9 6f f4 vmovdqa xmm6,xmm4 241 [ ]*[a-f0-9]+: c5 f9 7f e6 vmovdqa.s xmm6,xmm [all...] |
/external/llvm/test/CodeGen/X86/ |
pr22774.ll | 8 ; CHECK: vmovdqa in(%rip), %ymm0 10 ; CHECK-NEXT: vmovdqa %xmm0, out(%rip)
|
fast-isel-vecload.ll | 10 ; AVX: vmovdqa (%rdi), %xmm0 20 ; AVX: vmovdqa (%rdi), %xmm0 30 ; AVX: vmovdqa (%rdi), %xmm0 40 ; AVX: vmovdqa (%rdi), %xmm0 130 ; AVX: vmovdqa (%rdi), %xmm0 140 ; AVX: vmovdqa (%rdi), %xmm0 150 ; AVX: vmovdqa (%rdi), %xmm0 160 ; AVX: vmovdqa (%rdi), %xmm0
|
avx-cvt-2.ll | 15 ; CHECK-NEXT: vmovdqa {{.*#+}} xmm2 = [0,1,4,5,8,9,12,13,8,9,12,13,12,13,14,15] 19 ; CHECK-NEXT: vmovdqa %xmm0, (%rdi) 32 ; CHECK-NEXT: vmovdqa {{.*#+}} xmm2 = [0,1,4,5,8,9,12,13,8,9,12,13,12,13,14,15] 36 ; CHECK-NEXT: vmovdqa %xmm0, (%rdi) 49 ; CHECK-NEXT: vmovdqa {{.*#+}} xmm2 = [0,1,4,5,8,9,12,13,8,9,12,13,12,13,14,15] 67 ; CHECK-NEXT: vmovdqa {{.*#+}} xmm2 = [0,1,4,5,8,9,12,13,8,9,12,13,12,13,14,15]
|
vector-tzcnt-256.ll | 15 ; AVX1-NEXT: vmovdqa {{.*#+}} xmm3 = [1,1] 17 ; AVX1-NEXT: vmovdqa {{.*#+}} xmm4 = [15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15] 19 ; AVX1-NEXT: vmovdqa {{.*#+}} xmm6 = [0,1,1,2,1,2,2,3,1,2,2,3,2,3,3,4] 44 ; AVX2-NEXT: vmovdqa {{.*#+}} ymm2 = [15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15] 46 ; AVX2-NEXT: vmovdqa {{.*#+}} ymm4 = [0,1,1,2,1,2,2,3,1,2,2,3,2,3,3,4,0,1,1,2,1,2,2,3,1,2,2,3,2,3,3,4] 68 ; AVX1-NEXT: vmovdqa {{.*#+}} xmm3 = [1,1] 70 ; AVX1-NEXT: vmovdqa {{.*#+}} xmm4 = [15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15] 72 ; AVX1-NEXT: vmovdqa {{.*#+}} xmm6 = [0,1,1,2,1,2,2,3,1,2,2,3,2,3,3,4] 97 ; AVX2-NEXT: vmovdqa {{.*#+}} ymm2 = [15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15] 99 ; AVX2-NEXT: vmovdqa {{.*#+}} ymm4 = [0,1,1,2,1,2,2,3,1,2,2,3,2,3,3,4,0,1,1,2,1,2,2,3,1,2,2,3,2,3,3,4 [all...] |
vector-popcnt-256.ll | 11 ; AVX1-NEXT: vmovdqa {{.*#+}} xmm4 = [0,1,1,2,1,2,2,3,1,2,2,3,2,3,3,4] 31 ; AVX2-NEXT: vmovdqa {{.*#+}} ymm1 = [15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15] 33 ; AVX2-NEXT: vmovdqa {{.*#+}} ymm3 = [0,1,1,2,1,2,2,3,1,2,2,3,2,3,3,4,0,1,1,2,1,2,2,3,1,2,2,3,2,3,3,4] 52 ; AVX1-NEXT: vmovdqa {{.*#+}} xmm4 = [0,1,1,2,1,2,2,3,1,2,2,3,2,3,3,4] 80 ; AVX2-NEXT: vmovdqa {{.*#+}} ymm1 = [15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15] 82 ; AVX2-NEXT: vmovdqa {{.*#+}} ymm3 = [0,1,1,2,1,2,2,3,1,2,2,3,2,3,3,4,0,1,1,2,1,2,2,3,1,2,2,3,2,3,3,4] 102 ; AVX1-NEXT: vmovdqa {{.*#+}} xmm1 = [15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15] 104 ; AVX1-NEXT: vmovdqa {{.*#+}} xmm3 = [0,1,1,2,1,2,2,3,1,2,2,3,2,3,3,4] 128 ; AVX2-NEXT: vmovdqa {{.*#+}} ymm1 = [15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15] 130 ; AVX2-NEXT: vmovdqa {{.*#+}} ymm3 = [0,1,1,2,1,2,2,3,1,2,2,3,2,3,3,4,0,1,1,2,1,2,2,3,1,2,2,3,2,3,3,4 [all...] |
vector-lzcnt-512.ll | 46 ; ALL-NEXT: vmovdqa {{.*#+}} ymm2 = [16,16,16,16,16,16,16,16,16,16,16,16,16,16,16,16] 60 ; AVX512BW-NEXT: vmovdqa {{.*#+}} ymm2 = [16,16,16,16,16,16,16,16,16,16,16,16,16,16,16,16] 78 ; ALL-NEXT: vmovdqa {{.*#+}} ymm2 = [16,16,16,16,16,16,16,16,16,16,16,16,16,16,16,16] 92 ; AVX512BW-NEXT: vmovdqa {{.*#+}} ymm2 = [16,16,16,16,16,16,16,16,16,16,16,16,16,16,16,16] 111 ; ALL-NEXT: vmovdqa {{.*#+}} xmm3 = [24,24,24,24,24,24,24,24,24,24,24,24,24,24,24,24] 137 ; AVX512BW-NEXT: vmovdqa {{.*#+}} xmm3 = [24,24,24,24,24,24,24,24,24,24,24,24,24,24,24,24] 167 ; ALL-NEXT: vmovdqa {{.*#+}} xmm3 = [24,24,24,24,24,24,24,24,24,24,24,24,24,24,24,24] 193 ; AVX512BW-NEXT: vmovdqa {{.*#+}} xmm3 = [24,24,24,24,24,24,24,24,24,24,24,24,24,24,24,24]
|
setcc-lowering.ll | 14 ; CHECK-NEXT: vmovdqa {{.*#+}} xmm3 = [0,1,4,5,8,9,12,13,8,9,12,13,12,13,14,15]
|
vector-shift-lshr-512.ll | 58 ; ALL-NEXT: vmovdqa {{.*#+}} ymm5 = [15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15] 63 ; ALL-NEXT: vmovdqa {{.*#+}} ymm6 = [63,63,63,63,63,63,63,63,63,63,63,63,63,63,63,63,63,63,63,63,63,63,63,63,63,63,63,63,63,63,63,63] 68 ; ALL-NEXT: vmovdqa {{.*#+}} ymm7 = [127,127,127,127,127,127,127,127,127,127,127,127,127,127,127,127,127,127,127,127,127,127,127,127,127,127,127,127,127,127,127,127] 134 ; ALL-NEXT: vmovdqa {{.*#+}} ymm4 = [15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15] 139 ; ALL-NEXT: vmovdqa {{.*#+}} ymm5 = [63,63,63,63,63,63,63,63,63,63,63,63,63,63,63,63,63,63,63,63,63,63,63,63,63,63,63,63,63,63,63,63] 144 ; ALL-NEXT: vmovdqa {{.*#+}} ymm7 = [127,127,127,127,127,127,127,127,127,127,127,127,127,127,127,127,127,127,127,127,127,127,127,127,127,127,127,127,127,127,127,127] 189 ; ALL-NEXT: vmovdqa {{.*#+}} ymm3 = [0,1,2,3,4,5,6,7,8,9,10,11,12,13,14,15] 215 ; ALL-NEXT: vmovdqa {{.*#+}} ymm3 = [15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15] 217 ; ALL-NEXT: vmovdqa {{.*#+}} ymm4 = [0,1,2,3,4,5,6,7,7,6,5,4,3,2,1,0,0,1,2,3,4,5,6,7,7,6,5,4,3,2,1,0] 221 ; ALL-NEXT: vmovdqa {{.*#+}} ymm5 = [63,63,63,63,63,63,63,63,63,63,63,63,63,63,63,63,63,63,63,63,63,63,63,63,63,63,63, (…) [all...] |
vector-tzcnt-512.ll | 142 ; ALL-NEXT: vmovdqa {{.*#+}} ymm3 = [1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1] 144 ; ALL-NEXT: vmovdqa {{.*#+}} ymm4 = [15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15] 146 ; ALL-NEXT: vmovdqa {{.*#+}} ymm6 = [0,1,1,2,1,2,2,3,1,2,2,3,2,3,3,4,0,1,1,2,1,2,2,3,1,2,2,3,2,3,3,4] 178 ; ALL-NEXT: vmovdqa {{.*#+}} ymm3 = [1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1] 180 ; ALL-NEXT: vmovdqa {{.*#+}} ymm4 = [15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15] 182 ; ALL-NEXT: vmovdqa {{.*#+}} ymm6 = [0,1,1,2,1,2,2,3,1,2,2,3,2,3,3,4,0,1,1,2,1,2,2,3,1,2,2,3,2,3,3,4] 214 ; ALL-NEXT: vmovdqa {{.*#+}} ymm3 = [1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1] 216 ; ALL-NEXT: vmovdqa {{.*#+}} ymm4 = [15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15] 218 ; ALL-NEXT: vmovdqa {{.*#+}} ymm6 = [0,1,1,2,1,2,2,3,1,2,2,3,2,3,3,4,0,1,1,2,1,2,2,3,1,2,2,3,2,3,3,4] 244 ; ALL-NEXT: vmovdqa {{.*#+}} ymm3 = [1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1 [all...] |
/external/llvm/test/MC/X86/ |
intel-syntax-unsized-memory.s | 22 // CHECK: vmovdqa %ymm0, (%rax) 23 vmovdqa [rax], ymm0 label
|
/toolchain/binutils/binutils-2.25/gas/testsuite/gas/i386/ilp32/ |
x86-64-avx-swap-intel.d | 14 [ ]*[a-f0-9]+: c5 7d 7f c6 vmovdqa ymm6,ymm8 20 [ ]*[a-f0-9]+: c5 79 7f c6 vmovdqa xmm6,xmm8 29 [ ]*[a-f0-9]+: c5 79 7f c6 vmovdqa xmm6,xmm8 38 [ ]*[a-f0-9]+: c5 7d 7f c6 vmovdqa ymm6,ymm8 44 [ ]*[a-f0-9]+: c5 79 7f c6 vmovdqa xmm6,xmm8 53 [ ]*[a-f0-9]+: c5 79 7f c6 vmovdqa xmm6,xmm8
|
x86-64-avx-swap.d | 14 [ ]*[a-f0-9]+: c5 7d 7f c6 vmovdqa %ymm8,%ymm6 20 [ ]*[a-f0-9]+: c5 79 7f c6 vmovdqa %xmm8,%xmm6 29 [ ]*[a-f0-9]+: c5 79 7f c6 vmovdqa %xmm8,%xmm6 38 [ ]*[a-f0-9]+: c5 7d 7f c6 vmovdqa %ymm8,%ymm6 44 [ ]*[a-f0-9]+: c5 79 7f c6 vmovdqa %xmm8,%xmm6 53 [ ]*[a-f0-9]+: c5 79 7f c6 vmovdqa %xmm8,%xmm6
|
/external/boringssl/win-x86/crypto/sha/ |
sha1-586.asm | [all...] |
/external/boringssl/linux-x86/crypto/sha/ |
sha1-586.S | [all...] |
/external/boringssl/mac-x86/crypto/sha/ |
sha1-586.S | [all...] |
/external/valgrind/none/tests/amd64/ |
avx2-1.c | 77 "vmovdqa 0(%0),%%ymm7" "\n\t" \ 78 "vmovdqa 32(%0),%%ymm8" "\n\t" \ 79 "vmovdqa 64(%0),%%ymm6" "\n\t" \ 80 "vmovdqa 96(%0),%%ymm9" "\n\t" \ 83 "vmovdqa %%ymm7, 0(%0)" "\n\t" \ 84 "vmovdqa %%ymm8, 32(%0)" "\n\t" \ 85 "vmovdqa %%ymm6, 64(%0)" "\n\t" \ 86 "vmovdqa %%ymm9, 96(%0)" "\n\t" \ 98 "vmovdqa 32(%0),%%ymm8" "\n\t" \ 99 "vmovdqa 64(%0),%%ymm7" "\n\t" [all...] |
/external/boringssl/src/crypto/sha/asm/ |
sha512-x86_64.pl | 1138 vmovdqa $TABLE+`$SZ*2*$rounds`(%rip),$t3 1152 vmovdqa $t0,0x00(%rsp) 1154 vmovdqa $t1,0x10(%rsp) 1156 vmovdqa $t2,0x20(%rsp) 1158 vmovdqa $t3,0x30(%rsp) [all...] |