HomeSort by relevance Sort by last modified time
    Searched full:vmovdqa (Results 51 - 75 of 111) sorted by null

1 23 4 5

  /external/llvm/test/CodeGen/X86/
vector-popcnt-128.ll 86 ; AVX-NEXT: vmovdqa {{.*#+}} xmm1 = [15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15]
88 ; AVX-NEXT: vmovdqa {{.*#+}} xmm3 = [0,1,1,2,1,2,2,3,1,2,2,3,2,3,3,4]
200 ; AVX-NEXT: vmovdqa {{.*#+}} xmm1 = [15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15]
202 ; AVX-NEXT: vmovdqa {{.*#+}} xmm3 = [0,1,1,2,1,2,2,3,1,2,2,3,2,3,3,4]
302 ; AVX-NEXT: vmovdqa {{.*#+}} xmm1 = [15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15]
304 ; AVX-NEXT: vmovdqa {{.*#+}} xmm3 = [0,1,1,2,1,2,2,3,1,2,2,3,2,3,3,4]
389 ; AVX-NEXT: vmovdqa {{.*#+}} xmm1 = [15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15]
391 ; AVX-NEXT: vmovdqa {{.*#+}} xmm3 = [0,1,1,2,1,2,2,3,1,2,2,3,2,3,3,4]
pmul.ll 45 ; AVX2-NEXT: vmovdqa {{.*#+}} xmm2 = <0,2,4,6,8,10,12,14,u,u,u,u,u,u,u,u>
112 ; AVX2-NEXT: vmovdqa {{.*#+}} xmm1 = [117,117]
168 ; AVX2-NEXT: vmovdqa {{.*#+}} xmm2 = <0,2,4,6,8,10,12,14,u,u,u,u,u,u,u,u>
291 ; AVX2-NEXT: vmovdqa (%rsp), %xmm0 # 16-byte Reload
331 ; AVX2-NEXT: vmovdqa {{[0-9]+}}(%rsp), %xmm2 # 16-byte Reload
332 ; AVX2-NEXT: vmovdqa (%rsp), %xmm3 # 16-byte Reload
vec_cmp_uint-128.ll 275 ; AVX-NEXT: vmovdqa {{.*#+}} xmm2 = [9223372036854775808,9223372036854775808]
432 ; AVX-NEXT: vmovdqa {{.*#+}} xmm2 = [9223372036854775808,9223372036854775808]
458 ; AVX1-NEXT: vmovdqa {{.*#+}} xmm2 = [2147483648,2147483648,2147483648,2147483648]
500 ; AVX-NEXT: vmovdqa {{.*#+}} xmm2 = [32768,32768,32768,32768,32768,32768,32768,32768]
526 ; AVX-NEXT: vmovdqa {{.*#+}} xmm2 = [128,128,128,128,128,128,128,128,128,128,128,128,128,128,128,128]
592 ; AVX-NEXT: vmovdqa {{.*#+}} xmm2 = [9223372036854775808,9223372036854775808]
750 ; AVX-NEXT: vmovdqa {{.*#+}} xmm2 = [9223372036854775808,9223372036854775808]
777 ; AVX1-NEXT: vmovdqa {{.*#+}} xmm2 = [2147483648,2147483648,2147483648,2147483648]
820 ; AVX-NEXT: vmovdqa {{.*#+}} xmm2 = [32768,32768,32768,32768,32768,32768,32768,32768]
847 ; AVX-NEXT: vmovdqa {{.*#+}} xmm2 = [128,128,128,128,128,128,128,128,128,128,128,128,128,128,128,128
    [all...]
vector-shift-ashr-256.ll 15 ; AVX1-NEXT: vmovdqa {{.*#+}} xmm3 = [9223372036854775808,9223372036854775808]
322 ; AVX1-NEXT: vmovdqa {{.*#+}} xmm2 = [9223372036854775808,9223372036854775808]
567 ; AVX1-NEXT: vmovdqa {{.*#+}} xmm2 = [4294967296,2]
573 ; AVX1-NEXT: vmovdqa {{.*#+}} xmm2 = [4611686018427387904,72057594037927936]
582 ; AVX2-NEXT: vmovdqa {{.*#+}} ymm1 = [4611686018427387904,72057594037927936,4294967296,2]
601 ; XOPAVX2-NEXT: vmovdqa {{.*#+}} ymm1 = [4611686018427387904,72057594037927936,4294967296,2]
654 ; AVX1-NEXT: vmovdqa {{.*#+}} xmm3 = [32896,37008,41120,45232,49344,53456,57568,61680]
657 ; AVX1-NEXT: vmovdqa {{.*#+}} xmm3 = [256,8480,16704,24928,33152,41376,49600,57824]
660 ; AVX1-NEXT: vmovdqa {{.*#+}} xmm3 = [512,16960,33408,49856,768,17216,33664,50112]
663 ; AVX1-NEXT: vmovdqa {{.*#+}} xmm3 = [1024,33920,1280,34176,1536,34432,1792,34688
    [all...]
vector-shuffle-256-v8.ll     [all...]
avx2-arith.ll 69 ; CHECK-NEXT: vmovdqa {{.*#+}} xmm2 = <0,2,4,6,8,10,12,14,u,u,u,u,u,u,u,u>
88 ; CHECK-NEXT: vmovdqa {{.*#+}} xmm4 = <0,2,4,6,8,10,12,14,u,u,u,u,u,u,u,u>
vector-shuffle-256-v32.ll 329 ; AVX2-NEXT: vmovdqa {{.*#+}} ymm2 = [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,255,255,255,255,255,255,255,255,255,255,255,255,255,255,255,255,255]
350 ; AVX2-NEXT: vmovdqa {{.*#+}} ymm2 = <0,255,u,u,u,u,u,u,u,u,u,u,u,u,u,u,255,u,u,u,u,u,u,u,u,u,u,u,u,u,u,u>
372 ; AVX2-NEXT: vmovdqa {{.*#+}} ymm2 = <0,0,255,255,u,u,u,u,u,u,u,u,u,u,u,u,255,255,u,u,u,u,u,u,u,u,u,u,u,u,u,u>
394 ; AVX2-NEXT: vmovdqa {{.*#+}} ymm2 = <0,0,255,255,u,u,u,u,u,u,u,u,u,u,u,u,255,255,u,u,u,u,u,u,u,u,u,u,u,u,u,u>
    [all...]
vector-rotate-128.ll 56 ; AVX1-NEXT: vmovdqa {{.*#+}} xmm2 = [64,64]
71 ; AVX2-NEXT: vmovdqa {{.*#+}} xmm2 = [64,64]
178 ; AVX1-NEXT: vmovdqa {{.*#+}} xmm2 = [32,32,32,32]
386 ; AVX1-NEXT: vmovdqa {{.*#+}} xmm2 = [16,16,16,16,16,16,16,16]
421 ; AVX2-NEXT: vmovdqa {{.*#+}} xmm2 = [16,16,16,16,16,16,16,16]
426 ; AVX2-NEXT: vmovdqa {{.*#+}} ymm3 = [0,1,4,5,8,9,12,13,128,128,128,128,128,128,128,128,0,1,4,5,8,9,12,13,128,128,128,128,128,128,128,128]
627 ; AVX-NEXT: vmovdqa {{.*#+}} xmm2 = [8,8,8,8,8,8,8,8,8,8,8,8,8,8,8,8]
    [all...]
splat-for-size.ll 140 ; AVX-NEXT: vmovdqa {{.*#+}} xmm2 = [1,1,1,1,1,1,1,1]
176 ; AVX-NEXT: vmovdqa {{.*#+}} xmm2 = [1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1]
vector-shift-ashr-512.ll 227 ; ALL-NEXT: vmovdqa {{.*#+}} ymm3 = [0,1,2,3,4,5,6,7,8,9,10,11,12,13,14,15]
252 ; ALL-NEXT: vmovdqa {{.*#+}} ymm2 = [0,1,2,3,4,5,6,7,7,6,5,4,3,2,1,0,0,1,2,3,4,5,6,7,7,6,5,4,3,2,1,0]
335 ; ALL-NEXT: vmovdqa {{.*#+}} ymm2 = [31,31,31,31,31,31,31,31,31,31,31,31,31,31,31,31,31,31,31,31,31,31,31,31,31,31,31,31,31,31,31,31]
337 ; ALL-NEXT: vmovdqa {{.*#+}} ymm3 = [16,16,16,16,16,16,16,16,16,16,16,16,16,16,16,16,16,16,16,16,16,16,16,16,16,16,16,16,16,16,16,16]
  /toolchain/binutils/binutils-2.25/gas/testsuite/gas/i386/ilp32/
x86-64-opts-intel.d 159 [ ]*[a-f0-9]+: c5 fd 6f f4 vmovdqa ymm6,ymm4
160 [ ]*[a-f0-9]+: c5 fd 7f e6 vmovdqa.s ymm6,ymm4
189 [ ]*[a-f0-9]+: c5 f9 6f f4 vmovdqa xmm6,xmm4
190 [ ]*[a-f0-9]+: c5 f9 7f e6 vmovdqa.s xmm6,xmm4
281 [ ]*[a-f0-9]+: c5 fd 6f f4 vmovdqa ymm6,ymm4
282 [ ]*[a-f0-9]+: c5 fd 7f e6 vmovdqa.s ymm6,ymm4
311 [ ]*[a-f0-9]+: c5 f9 6f f4 vmovdqa xmm6,xmm4
312 [ ]*[a-f0-9]+: c5 f9 7f e6 vmovdqa.s xmm6,xmm4
x86-64-opts.d 159 [ ]*[a-f0-9]+: c5 fd 6f f4 vmovdqa %ymm4,%ymm6
160 [ ]*[a-f0-9]+: c5 fd 7f e6 vmovdqa.s %ymm4,%ymm6
189 [ ]*[a-f0-9]+: c5 f9 6f f4 vmovdqa %xmm4,%xmm6
190 [ ]*[a-f0-9]+: c5 f9 7f e6 vmovdqa.s %xmm4,%xmm6
281 [ ]*[a-f0-9]+: c5 fd 6f f4 vmovdqa %ymm4,%ymm6
282 [ ]*[a-f0-9]+: c5 fd 7f e6 vmovdqa.s %ymm4,%ymm6
311 [ ]*[a-f0-9]+: c5 f9 6f f4 vmovdqa %xmm4,%xmm6
312 [ ]*[a-f0-9]+: c5 f9 7f e6 vmovdqa.s %xmm4,%xmm6
x86-64-avx.d 234 [ ]*[a-f0-9]+: c5 fd 6f f4 vmovdqa %ymm4,%ymm6
235 [ ]*[a-f0-9]+: c5 fd 6f 21 vmovdqa \(%rcx\),%ymm4
266 [ ]*[a-f0-9]+: c5 fd 6f f4 vmovdqa %ymm4,%ymm6
267 [ ]*[a-f0-9]+: c5 fd 7f 21 vmovdqa %ymm4,\(%rcx\)
319 [ ]*[a-f0-9]+: c5 f9 6f f4 vmovdqa %xmm4,%xmm6
320 [ ]*[a-f0-9]+: c5 f9 6f 21 vmovdqa \(%rcx\),%xmm4
359 [ ]*[a-f0-9]+: c5 f9 6f f4 vmovdqa %xmm4,%xmm6
360 [ ]*[a-f0-9]+: c5 f9 7f 21 vmovdqa %xmm4,\(%rcx\)
    [all...]
  /toolchain/binutils/binutils-2.25/gas/testsuite/gas/i386/
opts.d 122 [ ]*[a-f0-9]+: c5 fd 6f f4 vmovdqa %ymm4,%ymm6
123 [ ]*[a-f0-9]+: c5 fd 7f e6 vmovdqa.s %ymm4,%ymm6
152 [ ]*[a-f0-9]+: c5 f9 6f f4 vmovdqa %xmm4,%xmm6
153 [ ]*[a-f0-9]+: c5 f9 7f e6 vmovdqa.s %xmm4,%xmm6
226 [ ]*[a-f0-9]+: c5 fd 6f f4 vmovdqa %ymm4,%ymm6
227 [ ]*[a-f0-9]+: c5 fd 7f e6 vmovdqa.s %ymm4,%ymm6
256 [ ]*[a-f0-9]+: c5 f9 6f f4 vmovdqa %xmm4,%xmm6
257 [ ]*[a-f0-9]+: c5 f9 7f e6 vmovdqa.s %xmm4,%xmm6
x86-64-opts-intel.d 159 [ ]*[a-f0-9]+: c5 fd 6f f4 vmovdqa ymm6,ymm4
160 [ ]*[a-f0-9]+: c5 fd 7f e6 vmovdqa.s ymm6,ymm4
189 [ ]*[a-f0-9]+: c5 f9 6f f4 vmovdqa xmm6,xmm4
190 [ ]*[a-f0-9]+: c5 f9 7f e6 vmovdqa.s xmm6,xmm4
281 [ ]*[a-f0-9]+: c5 fd 6f f4 vmovdqa ymm6,ymm4
282 [ ]*[a-f0-9]+: c5 fd 7f e6 vmovdqa.s ymm6,ymm4
311 [ ]*[a-f0-9]+: c5 f9 6f f4 vmovdqa xmm6,xmm4
312 [ ]*[a-f0-9]+: c5 f9 7f e6 vmovdqa.s xmm6,xmm4
x86-64-opts.d 158 [ ]*[a-f0-9]+: c5 fd 6f f4 vmovdqa %ymm4,%ymm6
159 [ ]*[a-f0-9]+: c5 fd 7f e6 vmovdqa.s %ymm4,%ymm6
188 [ ]*[a-f0-9]+: c5 f9 6f f4 vmovdqa %xmm4,%xmm6
189 [ ]*[a-f0-9]+: c5 f9 7f e6 vmovdqa.s %xmm4,%xmm6
280 [ ]*[a-f0-9]+: c5 fd 6f f4 vmovdqa %ymm4,%ymm6
281 [ ]*[a-f0-9]+: c5 fd 7f e6 vmovdqa.s %ymm4,%ymm6
310 [ ]*[a-f0-9]+: c5 f9 6f f4 vmovdqa %xmm4,%xmm6
311 [ ]*[a-f0-9]+: c5 f9 7f e6 vmovdqa.s %xmm4,%xmm6
x86-64-avx.d 233 [ ]*[a-f0-9]+: c5 fd 6f f4 vmovdqa %ymm4,%ymm6
234 [ ]*[a-f0-9]+: c5 fd 6f 21 vmovdqa \(%rcx\),%ymm4
265 [ ]*[a-f0-9]+: c5 fd 6f f4 vmovdqa %ymm4,%ymm6
266 [ ]*[a-f0-9]+: c5 fd 7f 21 vmovdqa %ymm4,\(%rcx\)
318 [ ]*[a-f0-9]+: c5 f9 6f f4 vmovdqa %xmm4,%xmm6
319 [ ]*[a-f0-9]+: c5 f9 6f 21 vmovdqa \(%rcx\),%xmm4
358 [ ]*[a-f0-9]+: c5 f9 6f f4 vmovdqa %xmm4,%xmm6
359 [ ]*[a-f0-9]+: c5 f9 7f 21 vmovdqa %xmm4,\(%rcx\)
    [all...]
avx.d 233 [ ]*[a-f0-9]+: c5 fd 6f f4 vmovdqa %ymm4,%ymm6
234 [ ]*[a-f0-9]+: c5 fd 6f 21 vmovdqa \(%ecx\),%ymm4
265 [ ]*[a-f0-9]+: c5 fd 6f f4 vmovdqa %ymm4,%ymm6
266 [ ]*[a-f0-9]+: c5 fd 7f 21 vmovdqa %ymm4,\(%ecx\)
318 [ ]*[a-f0-9]+: c5 f9 6f f4 vmovdqa %xmm4,%xmm6
319 [ ]*[a-f0-9]+: c5 f9 6f 21 vmovdqa \(%ecx\),%xmm4
358 [ ]*[a-f0-9]+: c5 f9 6f f4 vmovdqa %xmm4,%xmm6
359 [ ]*[a-f0-9]+: c5 f9 7f 21 vmovdqa %xmm4,\(%ecx\)
    [all...]
  /external/boringssl/linux-x86_64/crypto/sha/
sha1-x86_64.S     [all...]
  /external/boringssl/mac-x86_64/crypto/sha/
sha1-x86_64.S     [all...]
  /external/valgrind/none/tests/amd64/
avx-1.c 71 "vmovdqa 0(%0),%%ymm7" "\n\t" \
72 "vmovdqa 32(%0),%%ymm8" "\n\t" \
73 "vmovdqa 64(%0),%%ymm6" "\n\t" \
74 "vmovdqa 96(%0),%%ymm9" "\n\t" \
77 "vmovdqa %%ymm7, 0(%0)" "\n\t" \
78 "vmovdqa %%ymm8, 32(%0)" "\n\t" \
79 "vmovdqa %%ymm6, 64(%0)" "\n\t" \
80 "vmovdqa %%ymm9, 96(%0)" "\n\t" \
92 "vmovdqa 32(%0),%%ymm8" "\n\t" \
93 "vmovdqa 64(%0),%%ymm7" "\n\t"
    [all...]
  /external/boringssl/src/crypto/ec/asm/
p256-x86_64-asm.pl 1315 vmovdqa .LTwo(%rip), $TWO
1321 vmovdqa .LOne(%rip), $M0
1322 vmovdqa .LTwo(%rip), $M1
1330 vmovdqa 32*0($in_t), $T0a
1331 vmovdqa 32*1($in_t), $T0b
1332 vmovdqa 32*2($in_t), $T0c
1334 vmovdqa 32*3($in_t), $T1a
1335 vmovdqa 32*4($in_t), $T1b
1336 vmovdqa 32*5($in_t), $T1c
1420 vmovdqa .LThree(%rip), $THRE
    [all...]
  /external/boringssl/src/crypto/sha/asm/
sha256-586.pl 990 &vmovdqa ($t3,&QWP(256,$K256));
1009 &vmovdqa (&QWP(32+0,"esp"),$t0);
1010 &vmovdqa (&QWP(32+16,"esp"),$t1);
1011 &vmovdqa (&QWP(32+32,"esp"),$t2);
1012 &vmovdqa (&QWP(32+48,"esp"),$t3);
1071 &vmovdqa (&QWP(32+16*$j,"esp"),$t2);
1121 &vmovdqa ($t3,&QWP(64,$K256));
    [all...]
sha1-x86_64.pl 969 vmovdqa 64($K_XX_XX),@X[2] # pbswap mask
970 vmovdqa -64($K_XX_XX),$Kx # K_00_19
983 vmovdqa @X[0],0(%rsp) # X[]+K xfer to IALU
984 vmovdqa @X[1],16(%rsp)
985 vmovdqa @X[2],32(%rsp)
1020 &vmovdqa (eval(16*(($Xi-1)&3))."(%rsp)",@Tx[1]); # X[]+K xfer to IALU
1054 &vmovdqa ($Kx,eval(2*16*(($Xi)/5)-64)."($K_XX_XX)") if ($Xi%5==0); # K_XX_XX
1081 &vmovdqa ($Kx,eval(2*16*($Xi/5)-64)."($K_XX_XX)") if ($Xi%5==0);
1092 &vmovdqa (eval(16*(($Xi-1)&3))."(%rsp)",@Tx[1]); # X[]+K xfer to IALU
1136 &vmovdqa (eval(16*(($Xi-1)&3))."(%rsp)",@Tx[1]); # X[]+K xfer IAL
    [all...]
  /external/llvm/test/MC/Disassembler/X86/
x86-32.txt 316 # CHECK: vmovdqa %xmm1, %xmm0
319 # CHECK: vmovdqa %xmm0, %xmm1
322 # CHECK: vmovdqa %ymm1, %ymm0
325 # CHECK: vmovdqa %ymm0, %ymm1

Completed in 480 milliseconds

1 23 4 5