HomeSort by relevance Sort by last modified time
    Searched full:vmovdqa (Results 1 - 25 of 111) sorted by null

1 2 3 4 5

  /toolchain/binutils/binutils-2.25/gas/testsuite/gas/i386/
x86-64-avx-swap.s 10 vmovdqa %ymm8,%ymm6
27 vmovdqa %xmm8,%xmm6
42 vmovdqa ymm6,ymm8
59 vmovdqa xmm6,xmm8
x86-64-avx-swap-intel.d 14 [ ]*[a-f0-9]+: c5 7d 7f c6 vmovdqa ymm6,ymm8
20 [ ]*[a-f0-9]+: c5 79 7f c6 vmovdqa xmm6,xmm8
29 [ ]*[a-f0-9]+: c5 79 7f c6 vmovdqa xmm6,xmm8
38 [ ]*[a-f0-9]+: c5 7d 7f c6 vmovdqa ymm6,ymm8
44 [ ]*[a-f0-9]+: c5 79 7f c6 vmovdqa xmm6,xmm8
53 [ ]*[a-f0-9]+: c5 79 7f c6 vmovdqa xmm6,xmm8
x86-64-avx-swap.d 13 [ ]*[a-f0-9]+: c5 7d 7f c6 vmovdqa %ymm8,%ymm6
19 [ ]*[a-f0-9]+: c5 79 7f c6 vmovdqa %xmm8,%xmm6
28 [ ]*[a-f0-9]+: c5 79 7f c6 vmovdqa %xmm8,%xmm6
37 [ ]*[a-f0-9]+: c5 7d 7f c6 vmovdqa %ymm8,%ymm6
43 [ ]*[a-f0-9]+: c5 79 7f c6 vmovdqa %xmm8,%xmm6
52 [ ]*[a-f0-9]+: c5 79 7f c6 vmovdqa %xmm8,%xmm6
x86-64-avx.s 244 vmovdqa %ymm4,%ymm6
245 vmovdqa (%rcx),%ymm4
278 vmovdqa %ymm4,%ymm6
279 vmovdqa %ymm4,(%rcx)
347 vmovdqa %xmm4,%xmm6
348 vmovdqa (%rcx),%xmm4
389 vmovdqa %xmm4,%xmm6
390 vmovdqa %xmm4,(%rcx)
    [all...]
avx.s 244 vmovdqa %ymm4,%ymm6
245 vmovdqa (%ecx),%ymm4
278 vmovdqa %ymm4,%ymm6
279 vmovdqa %ymm4,(%ecx)
347 vmovdqa %xmm4,%xmm6
348 vmovdqa (%ecx),%xmm4
389 vmovdqa %xmm4,%xmm6
390 vmovdqa %xmm4,(%ecx)
1223 vmovdqa 0x1234,%xmm
    [all...]
opts.s 122 vmovdqa %ymm4,%ymm6
123 vmovdqa.s %ymm4,%ymm6
154 vmovdqa %xmm4,%xmm6
155 vmovdqa.s %xmm4,%xmm6
238 vmovdqa ymm6,ymm4
239 vmovdqa.s ymm6,ymm4
270 vmovdqa xmm6,xmm4
271 vmovdqa.s xmm6,xmm4
x86-64-opts.s 158 vmovdqa %ymm4,%ymm6
159 vmovdqa.s %ymm4,%ymm6
190 vmovdqa %xmm4,%xmm6
191 vmovdqa.s %xmm4,%xmm6
292 vmovdqa ymm6,ymm4
293 vmovdqa.s ymm6,ymm4
324 vmovdqa xmm6,xmm4
325 vmovdqa.s xmm6,xmm4
sse2avx-opts-intel.d 124 [ ]*[a-f0-9]+: c5 fd 6f f4 vmovdqa ymm6,ymm4
125 [ ]*[a-f0-9]+: c5 fd 7f e6 vmovdqa.s ymm6,ymm4
136 [ ]*[a-f0-9]+: c5 f9 6f f4 vmovdqa xmm6,xmm4
137 [ ]*[a-f0-9]+: c5 f9 7f e6 vmovdqa.s xmm6,xmm4
154 [ ]*[a-f0-9]+: c5 f9 6f f4 vmovdqa xmm6,xmm4
155 [ ]*[a-f0-9]+: c5 f9 7f e6 vmovdqa.s xmm6,xmm4
228 [ ]*[a-f0-9]+: c5 fd 6f f4 vmovdqa ymm6,ymm4
229 [ ]*[a-f0-9]+: c5 fd 7f e6 vmovdqa.s ymm6,ymm4
240 [ ]*[a-f0-9]+: c5 f9 6f f4 vmovdqa xmm6,xmm4
241 [ ]*[a-f0-9]+: c5 f9 7f e6 vmovdqa.s xmm6,xmm
    [all...]
  /external/llvm/test/CodeGen/X86/
pr22774.ll 8 ; CHECK: vmovdqa in(%rip), %ymm0
10 ; CHECK-NEXT: vmovdqa %xmm0, out(%rip)
fast-isel-vecload.ll 10 ; AVX: vmovdqa (%rdi), %xmm0
20 ; AVX: vmovdqa (%rdi), %xmm0
30 ; AVX: vmovdqa (%rdi), %xmm0
40 ; AVX: vmovdqa (%rdi), %xmm0
130 ; AVX: vmovdqa (%rdi), %xmm0
140 ; AVX: vmovdqa (%rdi), %xmm0
150 ; AVX: vmovdqa (%rdi), %xmm0
160 ; AVX: vmovdqa (%rdi), %xmm0
avx-cvt-2.ll 15 ; CHECK-NEXT: vmovdqa {{.*#+}} xmm2 = [0,1,4,5,8,9,12,13,8,9,12,13,12,13,14,15]
19 ; CHECK-NEXT: vmovdqa %xmm0, (%rdi)
32 ; CHECK-NEXT: vmovdqa {{.*#+}} xmm2 = [0,1,4,5,8,9,12,13,8,9,12,13,12,13,14,15]
36 ; CHECK-NEXT: vmovdqa %xmm0, (%rdi)
49 ; CHECK-NEXT: vmovdqa {{.*#+}} xmm2 = [0,1,4,5,8,9,12,13,8,9,12,13,12,13,14,15]
67 ; CHECK-NEXT: vmovdqa {{.*#+}} xmm2 = [0,1,4,5,8,9,12,13,8,9,12,13,12,13,14,15]
vector-tzcnt-256.ll 15 ; AVX1-NEXT: vmovdqa {{.*#+}} xmm3 = [1,1]
17 ; AVX1-NEXT: vmovdqa {{.*#+}} xmm4 = [15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15]
19 ; AVX1-NEXT: vmovdqa {{.*#+}} xmm6 = [0,1,1,2,1,2,2,3,1,2,2,3,2,3,3,4]
44 ; AVX2-NEXT: vmovdqa {{.*#+}} ymm2 = [15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15]
46 ; AVX2-NEXT: vmovdqa {{.*#+}} ymm4 = [0,1,1,2,1,2,2,3,1,2,2,3,2,3,3,4,0,1,1,2,1,2,2,3,1,2,2,3,2,3,3,4]
68 ; AVX1-NEXT: vmovdqa {{.*#+}} xmm3 = [1,1]
70 ; AVX1-NEXT: vmovdqa {{.*#+}} xmm4 = [15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15]
72 ; AVX1-NEXT: vmovdqa {{.*#+}} xmm6 = [0,1,1,2,1,2,2,3,1,2,2,3,2,3,3,4]
97 ; AVX2-NEXT: vmovdqa {{.*#+}} ymm2 = [15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15]
99 ; AVX2-NEXT: vmovdqa {{.*#+}} ymm4 = [0,1,1,2,1,2,2,3,1,2,2,3,2,3,3,4,0,1,1,2,1,2,2,3,1,2,2,3,2,3,3,4
    [all...]
vector-popcnt-256.ll 11 ; AVX1-NEXT: vmovdqa {{.*#+}} xmm4 = [0,1,1,2,1,2,2,3,1,2,2,3,2,3,3,4]
31 ; AVX2-NEXT: vmovdqa {{.*#+}} ymm1 = [15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15]
33 ; AVX2-NEXT: vmovdqa {{.*#+}} ymm3 = [0,1,1,2,1,2,2,3,1,2,2,3,2,3,3,4,0,1,1,2,1,2,2,3,1,2,2,3,2,3,3,4]
52 ; AVX1-NEXT: vmovdqa {{.*#+}} xmm4 = [0,1,1,2,1,2,2,3,1,2,2,3,2,3,3,4]
80 ; AVX2-NEXT: vmovdqa {{.*#+}} ymm1 = [15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15]
82 ; AVX2-NEXT: vmovdqa {{.*#+}} ymm3 = [0,1,1,2,1,2,2,3,1,2,2,3,2,3,3,4,0,1,1,2,1,2,2,3,1,2,2,3,2,3,3,4]
102 ; AVX1-NEXT: vmovdqa {{.*#+}} xmm1 = [15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15]
104 ; AVX1-NEXT: vmovdqa {{.*#+}} xmm3 = [0,1,1,2,1,2,2,3,1,2,2,3,2,3,3,4]
128 ; AVX2-NEXT: vmovdqa {{.*#+}} ymm1 = [15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15]
130 ; AVX2-NEXT: vmovdqa {{.*#+}} ymm3 = [0,1,1,2,1,2,2,3,1,2,2,3,2,3,3,4,0,1,1,2,1,2,2,3,1,2,2,3,2,3,3,4
    [all...]
vector-lzcnt-512.ll 46 ; ALL-NEXT: vmovdqa {{.*#+}} ymm2 = [16,16,16,16,16,16,16,16,16,16,16,16,16,16,16,16]
60 ; AVX512BW-NEXT: vmovdqa {{.*#+}} ymm2 = [16,16,16,16,16,16,16,16,16,16,16,16,16,16,16,16]
78 ; ALL-NEXT: vmovdqa {{.*#+}} ymm2 = [16,16,16,16,16,16,16,16,16,16,16,16,16,16,16,16]
92 ; AVX512BW-NEXT: vmovdqa {{.*#+}} ymm2 = [16,16,16,16,16,16,16,16,16,16,16,16,16,16,16,16]
111 ; ALL-NEXT: vmovdqa {{.*#+}} xmm3 = [24,24,24,24,24,24,24,24,24,24,24,24,24,24,24,24]
137 ; AVX512BW-NEXT: vmovdqa {{.*#+}} xmm3 = [24,24,24,24,24,24,24,24,24,24,24,24,24,24,24,24]
167 ; ALL-NEXT: vmovdqa {{.*#+}} xmm3 = [24,24,24,24,24,24,24,24,24,24,24,24,24,24,24,24]
193 ; AVX512BW-NEXT: vmovdqa {{.*#+}} xmm3 = [24,24,24,24,24,24,24,24,24,24,24,24,24,24,24,24]
setcc-lowering.ll 14 ; CHECK-NEXT: vmovdqa {{.*#+}} xmm3 = [0,1,4,5,8,9,12,13,8,9,12,13,12,13,14,15]
vector-shift-lshr-512.ll 58 ; ALL-NEXT: vmovdqa {{.*#+}} ymm5 = [15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15]
63 ; ALL-NEXT: vmovdqa {{.*#+}} ymm6 = [63,63,63,63,63,63,63,63,63,63,63,63,63,63,63,63,63,63,63,63,63,63,63,63,63,63,63,63,63,63,63,63]
68 ; ALL-NEXT: vmovdqa {{.*#+}} ymm7 = [127,127,127,127,127,127,127,127,127,127,127,127,127,127,127,127,127,127,127,127,127,127,127,127,127,127,127,127,127,127,127,127]
134 ; ALL-NEXT: vmovdqa {{.*#+}} ymm4 = [15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15]
139 ; ALL-NEXT: vmovdqa {{.*#+}} ymm5 = [63,63,63,63,63,63,63,63,63,63,63,63,63,63,63,63,63,63,63,63,63,63,63,63,63,63,63,63,63,63,63,63]
144 ; ALL-NEXT: vmovdqa {{.*#+}} ymm7 = [127,127,127,127,127,127,127,127,127,127,127,127,127,127,127,127,127,127,127,127,127,127,127,127,127,127,127,127,127,127,127,127]
189 ; ALL-NEXT: vmovdqa {{.*#+}} ymm3 = [0,1,2,3,4,5,6,7,8,9,10,11,12,13,14,15]
215 ; ALL-NEXT: vmovdqa {{.*#+}} ymm3 = [15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15]
217 ; ALL-NEXT: vmovdqa {{.*#+}} ymm4 = [0,1,2,3,4,5,6,7,7,6,5,4,3,2,1,0,0,1,2,3,4,5,6,7,7,6,5,4,3,2,1,0]
221 ; ALL-NEXT: vmovdqa {{.*#+}} ymm5 = [63,63,63,63,63,63,63,63,63,63,63,63,63,63,63,63,63,63,63,63,63,63,63,63,63,63,63, (…)
    [all...]
vector-tzcnt-512.ll 142 ; ALL-NEXT: vmovdqa {{.*#+}} ymm3 = [1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1]
144 ; ALL-NEXT: vmovdqa {{.*#+}} ymm4 = [15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15]
146 ; ALL-NEXT: vmovdqa {{.*#+}} ymm6 = [0,1,1,2,1,2,2,3,1,2,2,3,2,3,3,4,0,1,1,2,1,2,2,3,1,2,2,3,2,3,3,4]
178 ; ALL-NEXT: vmovdqa {{.*#+}} ymm3 = [1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1]
180 ; ALL-NEXT: vmovdqa {{.*#+}} ymm4 = [15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15]
182 ; ALL-NEXT: vmovdqa {{.*#+}} ymm6 = [0,1,1,2,1,2,2,3,1,2,2,3,2,3,3,4,0,1,1,2,1,2,2,3,1,2,2,3,2,3,3,4]
214 ; ALL-NEXT: vmovdqa {{.*#+}} ymm3 = [1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1]
216 ; ALL-NEXT: vmovdqa {{.*#+}} ymm4 = [15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15]
218 ; ALL-NEXT: vmovdqa {{.*#+}} ymm6 = [0,1,1,2,1,2,2,3,1,2,2,3,2,3,3,4,0,1,1,2,1,2,2,3,1,2,2,3,2,3,3,4]
244 ; ALL-NEXT: vmovdqa {{.*#+}} ymm3 = [1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1
    [all...]
  /external/llvm/test/MC/X86/
intel-syntax-unsized-memory.s 22 // CHECK: vmovdqa %ymm0, (%rax)
23 vmovdqa [rax], ymm0 label
  /toolchain/binutils/binutils-2.25/gas/testsuite/gas/i386/ilp32/
x86-64-avx-swap-intel.d 14 [ ]*[a-f0-9]+: c5 7d 7f c6 vmovdqa ymm6,ymm8
20 [ ]*[a-f0-9]+: c5 79 7f c6 vmovdqa xmm6,xmm8
29 [ ]*[a-f0-9]+: c5 79 7f c6 vmovdqa xmm6,xmm8
38 [ ]*[a-f0-9]+: c5 7d 7f c6 vmovdqa ymm6,ymm8
44 [ ]*[a-f0-9]+: c5 79 7f c6 vmovdqa xmm6,xmm8
53 [ ]*[a-f0-9]+: c5 79 7f c6 vmovdqa xmm6,xmm8
x86-64-avx-swap.d 14 [ ]*[a-f0-9]+: c5 7d 7f c6 vmovdqa %ymm8,%ymm6
20 [ ]*[a-f0-9]+: c5 79 7f c6 vmovdqa %xmm8,%xmm6
29 [ ]*[a-f0-9]+: c5 79 7f c6 vmovdqa %xmm8,%xmm6
38 [ ]*[a-f0-9]+: c5 7d 7f c6 vmovdqa %ymm8,%ymm6
44 [ ]*[a-f0-9]+: c5 79 7f c6 vmovdqa %xmm8,%xmm6
53 [ ]*[a-f0-9]+: c5 79 7f c6 vmovdqa %xmm8,%xmm6
  /external/boringssl/win-x86/crypto/sha/
sha1-586.asm     [all...]
  /external/boringssl/linux-x86/crypto/sha/
sha1-586.S     [all...]
  /external/boringssl/mac-x86/crypto/sha/
sha1-586.S     [all...]
  /external/valgrind/none/tests/amd64/
avx2-1.c 77 "vmovdqa 0(%0),%%ymm7" "\n\t" \
78 "vmovdqa 32(%0),%%ymm8" "\n\t" \
79 "vmovdqa 64(%0),%%ymm6" "\n\t" \
80 "vmovdqa 96(%0),%%ymm9" "\n\t" \
83 "vmovdqa %%ymm7, 0(%0)" "\n\t" \
84 "vmovdqa %%ymm8, 32(%0)" "\n\t" \
85 "vmovdqa %%ymm6, 64(%0)" "\n\t" \
86 "vmovdqa %%ymm9, 96(%0)" "\n\t" \
98 "vmovdqa 32(%0),%%ymm8" "\n\t" \
99 "vmovdqa 64(%0),%%ymm7" "\n\t"
    [all...]
  /external/boringssl/src/crypto/sha/asm/
sha512-x86_64.pl 1138 vmovdqa $TABLE+`$SZ*2*$rounds`(%rip),$t3
1152 vmovdqa $t0,0x00(%rsp)
1154 vmovdqa $t1,0x10(%rsp)
1156 vmovdqa $t2,0x20(%rsp)
1158 vmovdqa $t3,0x30(%rsp)
    [all...]

Completed in 840 milliseconds

1 2 3 4 5