/external/llvm/test/CodeGen/X86/ |
vector-popcnt-128.ll | 86 ; AVX-NEXT: vmovdqa {{.*#+}} xmm1 = [15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15] 88 ; AVX-NEXT: vmovdqa {{.*#+}} xmm3 = [0,1,1,2,1,2,2,3,1,2,2,3,2,3,3,4] 200 ; AVX-NEXT: vmovdqa {{.*#+}} xmm1 = [15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15] 202 ; AVX-NEXT: vmovdqa {{.*#+}} xmm3 = [0,1,1,2,1,2,2,3,1,2,2,3,2,3,3,4] 302 ; AVX-NEXT: vmovdqa {{.*#+}} xmm1 = [15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15] 304 ; AVX-NEXT: vmovdqa {{.*#+}} xmm3 = [0,1,1,2,1,2,2,3,1,2,2,3,2,3,3,4] 389 ; AVX-NEXT: vmovdqa {{.*#+}} xmm1 = [15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15] 391 ; AVX-NEXT: vmovdqa {{.*#+}} xmm3 = [0,1,1,2,1,2,2,3,1,2,2,3,2,3,3,4]
|
pmul.ll | 45 ; AVX2-NEXT: vmovdqa {{.*#+}} xmm2 = <0,2,4,6,8,10,12,14,u,u,u,u,u,u,u,u> 112 ; AVX2-NEXT: vmovdqa {{.*#+}} xmm1 = [117,117] 168 ; AVX2-NEXT: vmovdqa {{.*#+}} xmm2 = <0,2,4,6,8,10,12,14,u,u,u,u,u,u,u,u> 291 ; AVX2-NEXT: vmovdqa (%rsp), %xmm0 # 16-byte Reload 331 ; AVX2-NEXT: vmovdqa {{[0-9]+}}(%rsp), %xmm2 # 16-byte Reload 332 ; AVX2-NEXT: vmovdqa (%rsp), %xmm3 # 16-byte Reload
|
vec_cmp_uint-128.ll | 275 ; AVX-NEXT: vmovdqa {{.*#+}} xmm2 = [9223372036854775808,9223372036854775808] 432 ; AVX-NEXT: vmovdqa {{.*#+}} xmm2 = [9223372036854775808,9223372036854775808] 458 ; AVX1-NEXT: vmovdqa {{.*#+}} xmm2 = [2147483648,2147483648,2147483648,2147483648] 500 ; AVX-NEXT: vmovdqa {{.*#+}} xmm2 = [32768,32768,32768,32768,32768,32768,32768,32768] 526 ; AVX-NEXT: vmovdqa {{.*#+}} xmm2 = [128,128,128,128,128,128,128,128,128,128,128,128,128,128,128,128] 592 ; AVX-NEXT: vmovdqa {{.*#+}} xmm2 = [9223372036854775808,9223372036854775808] 750 ; AVX-NEXT: vmovdqa {{.*#+}} xmm2 = [9223372036854775808,9223372036854775808] 777 ; AVX1-NEXT: vmovdqa {{.*#+}} xmm2 = [2147483648,2147483648,2147483648,2147483648] 820 ; AVX-NEXT: vmovdqa {{.*#+}} xmm2 = [32768,32768,32768,32768,32768,32768,32768,32768] 847 ; AVX-NEXT: vmovdqa {{.*#+}} xmm2 = [128,128,128,128,128,128,128,128,128,128,128,128,128,128,128,128 [all...] |
vector-shift-ashr-256.ll | 15 ; AVX1-NEXT: vmovdqa {{.*#+}} xmm3 = [9223372036854775808,9223372036854775808] 322 ; AVX1-NEXT: vmovdqa {{.*#+}} xmm2 = [9223372036854775808,9223372036854775808] 567 ; AVX1-NEXT: vmovdqa {{.*#+}} xmm2 = [4294967296,2] 573 ; AVX1-NEXT: vmovdqa {{.*#+}} xmm2 = [4611686018427387904,72057594037927936] 582 ; AVX2-NEXT: vmovdqa {{.*#+}} ymm1 = [4611686018427387904,72057594037927936,4294967296,2] 601 ; XOPAVX2-NEXT: vmovdqa {{.*#+}} ymm1 = [4611686018427387904,72057594037927936,4294967296,2] 654 ; AVX1-NEXT: vmovdqa {{.*#+}} xmm3 = [32896,37008,41120,45232,49344,53456,57568,61680] 657 ; AVX1-NEXT: vmovdqa {{.*#+}} xmm3 = [256,8480,16704,24928,33152,41376,49600,57824] 660 ; AVX1-NEXT: vmovdqa {{.*#+}} xmm3 = [512,16960,33408,49856,768,17216,33664,50112] 663 ; AVX1-NEXT: vmovdqa {{.*#+}} xmm3 = [1024,33920,1280,34176,1536,34432,1792,34688 [all...] |
vector-shuffle-256-v8.ll | [all...] |
avx2-arith.ll | 69 ; CHECK-NEXT: vmovdqa {{.*#+}} xmm2 = <0,2,4,6,8,10,12,14,u,u,u,u,u,u,u,u> 88 ; CHECK-NEXT: vmovdqa {{.*#+}} xmm4 = <0,2,4,6,8,10,12,14,u,u,u,u,u,u,u,u>
|
vector-shuffle-256-v32.ll | 329 ; AVX2-NEXT: vmovdqa {{.*#+}} ymm2 = [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,255,255,255,255,255,255,255,255,255,255,255,255,255,255,255,255,255] 350 ; AVX2-NEXT: vmovdqa {{.*#+}} ymm2 = <0,255,u,u,u,u,u,u,u,u,u,u,u,u,u,u,255,u,u,u,u,u,u,u,u,u,u,u,u,u,u,u> 372 ; AVX2-NEXT: vmovdqa {{.*#+}} ymm2 = <0,0,255,255,u,u,u,u,u,u,u,u,u,u,u,u,255,255,u,u,u,u,u,u,u,u,u,u,u,u,u,u> 394 ; AVX2-NEXT: vmovdqa {{.*#+}} ymm2 = <0,0,255,255,u,u,u,u,u,u,u,u,u,u,u,u,255,255,u,u,u,u,u,u,u,u,u,u,u,u,u,u> [all...] |
vector-rotate-128.ll | 56 ; AVX1-NEXT: vmovdqa {{.*#+}} xmm2 = [64,64] 71 ; AVX2-NEXT: vmovdqa {{.*#+}} xmm2 = [64,64] 178 ; AVX1-NEXT: vmovdqa {{.*#+}} xmm2 = [32,32,32,32] 386 ; AVX1-NEXT: vmovdqa {{.*#+}} xmm2 = [16,16,16,16,16,16,16,16] 421 ; AVX2-NEXT: vmovdqa {{.*#+}} xmm2 = [16,16,16,16,16,16,16,16] 426 ; AVX2-NEXT: vmovdqa {{.*#+}} ymm3 = [0,1,4,5,8,9,12,13,128,128,128,128,128,128,128,128,0,1,4,5,8,9,12,13,128,128,128,128,128,128,128,128] 627 ; AVX-NEXT: vmovdqa {{.*#+}} xmm2 = [8,8,8,8,8,8,8,8,8,8,8,8,8,8,8,8] [all...] |
splat-for-size.ll | 140 ; AVX-NEXT: vmovdqa {{.*#+}} xmm2 = [1,1,1,1,1,1,1,1] 176 ; AVX-NEXT: vmovdqa {{.*#+}} xmm2 = [1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1]
|
vector-shift-ashr-512.ll | 227 ; ALL-NEXT: vmovdqa {{.*#+}} ymm3 = [0,1,2,3,4,5,6,7,8,9,10,11,12,13,14,15] 252 ; ALL-NEXT: vmovdqa {{.*#+}} ymm2 = [0,1,2,3,4,5,6,7,7,6,5,4,3,2,1,0,0,1,2,3,4,5,6,7,7,6,5,4,3,2,1,0] 335 ; ALL-NEXT: vmovdqa {{.*#+}} ymm2 = [31,31,31,31,31,31,31,31,31,31,31,31,31,31,31,31,31,31,31,31,31,31,31,31,31,31,31,31,31,31,31,31] 337 ; ALL-NEXT: vmovdqa {{.*#+}} ymm3 = [16,16,16,16,16,16,16,16,16,16,16,16,16,16,16,16,16,16,16,16,16,16,16,16,16,16,16,16,16,16,16,16]
|
/toolchain/binutils/binutils-2.25/gas/testsuite/gas/i386/ilp32/ |
x86-64-opts-intel.d | 159 [ ]*[a-f0-9]+: c5 fd 6f f4 vmovdqa ymm6,ymm4 160 [ ]*[a-f0-9]+: c5 fd 7f e6 vmovdqa.s ymm6,ymm4 189 [ ]*[a-f0-9]+: c5 f9 6f f4 vmovdqa xmm6,xmm4 190 [ ]*[a-f0-9]+: c5 f9 7f e6 vmovdqa.s xmm6,xmm4 281 [ ]*[a-f0-9]+: c5 fd 6f f4 vmovdqa ymm6,ymm4 282 [ ]*[a-f0-9]+: c5 fd 7f e6 vmovdqa.s ymm6,ymm4 311 [ ]*[a-f0-9]+: c5 f9 6f f4 vmovdqa xmm6,xmm4 312 [ ]*[a-f0-9]+: c5 f9 7f e6 vmovdqa.s xmm6,xmm4
|
x86-64-opts.d | 159 [ ]*[a-f0-9]+: c5 fd 6f f4 vmovdqa %ymm4,%ymm6 160 [ ]*[a-f0-9]+: c5 fd 7f e6 vmovdqa.s %ymm4,%ymm6 189 [ ]*[a-f0-9]+: c5 f9 6f f4 vmovdqa %xmm4,%xmm6 190 [ ]*[a-f0-9]+: c5 f9 7f e6 vmovdqa.s %xmm4,%xmm6 281 [ ]*[a-f0-9]+: c5 fd 6f f4 vmovdqa %ymm4,%ymm6 282 [ ]*[a-f0-9]+: c5 fd 7f e6 vmovdqa.s %ymm4,%ymm6 311 [ ]*[a-f0-9]+: c5 f9 6f f4 vmovdqa %xmm4,%xmm6 312 [ ]*[a-f0-9]+: c5 f9 7f e6 vmovdqa.s %xmm4,%xmm6
|
x86-64-avx.d | 234 [ ]*[a-f0-9]+: c5 fd 6f f4 vmovdqa %ymm4,%ymm6 235 [ ]*[a-f0-9]+: c5 fd 6f 21 vmovdqa \(%rcx\),%ymm4 266 [ ]*[a-f0-9]+: c5 fd 6f f4 vmovdqa %ymm4,%ymm6 267 [ ]*[a-f0-9]+: c5 fd 7f 21 vmovdqa %ymm4,\(%rcx\) 319 [ ]*[a-f0-9]+: c5 f9 6f f4 vmovdqa %xmm4,%xmm6 320 [ ]*[a-f0-9]+: c5 f9 6f 21 vmovdqa \(%rcx\),%xmm4 359 [ ]*[a-f0-9]+: c5 f9 6f f4 vmovdqa %xmm4,%xmm6 360 [ ]*[a-f0-9]+: c5 f9 7f 21 vmovdqa %xmm4,\(%rcx\) [all...] |
/toolchain/binutils/binutils-2.25/gas/testsuite/gas/i386/ |
opts.d | 122 [ ]*[a-f0-9]+: c5 fd 6f f4 vmovdqa %ymm4,%ymm6 123 [ ]*[a-f0-9]+: c5 fd 7f e6 vmovdqa.s %ymm4,%ymm6 152 [ ]*[a-f0-9]+: c5 f9 6f f4 vmovdqa %xmm4,%xmm6 153 [ ]*[a-f0-9]+: c5 f9 7f e6 vmovdqa.s %xmm4,%xmm6 226 [ ]*[a-f0-9]+: c5 fd 6f f4 vmovdqa %ymm4,%ymm6 227 [ ]*[a-f0-9]+: c5 fd 7f e6 vmovdqa.s %ymm4,%ymm6 256 [ ]*[a-f0-9]+: c5 f9 6f f4 vmovdqa %xmm4,%xmm6 257 [ ]*[a-f0-9]+: c5 f9 7f e6 vmovdqa.s %xmm4,%xmm6
|
x86-64-opts-intel.d | 159 [ ]*[a-f0-9]+: c5 fd 6f f4 vmovdqa ymm6,ymm4 160 [ ]*[a-f0-9]+: c5 fd 7f e6 vmovdqa.s ymm6,ymm4 189 [ ]*[a-f0-9]+: c5 f9 6f f4 vmovdqa xmm6,xmm4 190 [ ]*[a-f0-9]+: c5 f9 7f e6 vmovdqa.s xmm6,xmm4 281 [ ]*[a-f0-9]+: c5 fd 6f f4 vmovdqa ymm6,ymm4 282 [ ]*[a-f0-9]+: c5 fd 7f e6 vmovdqa.s ymm6,ymm4 311 [ ]*[a-f0-9]+: c5 f9 6f f4 vmovdqa xmm6,xmm4 312 [ ]*[a-f0-9]+: c5 f9 7f e6 vmovdqa.s xmm6,xmm4
|
x86-64-opts.d | 158 [ ]*[a-f0-9]+: c5 fd 6f f4 vmovdqa %ymm4,%ymm6 159 [ ]*[a-f0-9]+: c5 fd 7f e6 vmovdqa.s %ymm4,%ymm6 188 [ ]*[a-f0-9]+: c5 f9 6f f4 vmovdqa %xmm4,%xmm6 189 [ ]*[a-f0-9]+: c5 f9 7f e6 vmovdqa.s %xmm4,%xmm6 280 [ ]*[a-f0-9]+: c5 fd 6f f4 vmovdqa %ymm4,%ymm6 281 [ ]*[a-f0-9]+: c5 fd 7f e6 vmovdqa.s %ymm4,%ymm6 310 [ ]*[a-f0-9]+: c5 f9 6f f4 vmovdqa %xmm4,%xmm6 311 [ ]*[a-f0-9]+: c5 f9 7f e6 vmovdqa.s %xmm4,%xmm6
|
x86-64-avx.d | 233 [ ]*[a-f0-9]+: c5 fd 6f f4 vmovdqa %ymm4,%ymm6 234 [ ]*[a-f0-9]+: c5 fd 6f 21 vmovdqa \(%rcx\),%ymm4 265 [ ]*[a-f0-9]+: c5 fd 6f f4 vmovdqa %ymm4,%ymm6 266 [ ]*[a-f0-9]+: c5 fd 7f 21 vmovdqa %ymm4,\(%rcx\) 318 [ ]*[a-f0-9]+: c5 f9 6f f4 vmovdqa %xmm4,%xmm6 319 [ ]*[a-f0-9]+: c5 f9 6f 21 vmovdqa \(%rcx\),%xmm4 358 [ ]*[a-f0-9]+: c5 f9 6f f4 vmovdqa %xmm4,%xmm6 359 [ ]*[a-f0-9]+: c5 f9 7f 21 vmovdqa %xmm4,\(%rcx\) [all...] |
avx.d | 233 [ ]*[a-f0-9]+: c5 fd 6f f4 vmovdqa %ymm4,%ymm6 234 [ ]*[a-f0-9]+: c5 fd 6f 21 vmovdqa \(%ecx\),%ymm4 265 [ ]*[a-f0-9]+: c5 fd 6f f4 vmovdqa %ymm4,%ymm6 266 [ ]*[a-f0-9]+: c5 fd 7f 21 vmovdqa %ymm4,\(%ecx\) 318 [ ]*[a-f0-9]+: c5 f9 6f f4 vmovdqa %xmm4,%xmm6 319 [ ]*[a-f0-9]+: c5 f9 6f 21 vmovdqa \(%ecx\),%xmm4 358 [ ]*[a-f0-9]+: c5 f9 6f f4 vmovdqa %xmm4,%xmm6 359 [ ]*[a-f0-9]+: c5 f9 7f 21 vmovdqa %xmm4,\(%ecx\) [all...] |
/external/boringssl/linux-x86_64/crypto/sha/ |
sha1-x86_64.S | [all...] |
/external/boringssl/mac-x86_64/crypto/sha/ |
sha1-x86_64.S | [all...] |
/external/valgrind/none/tests/amd64/ |
avx-1.c | 71 "vmovdqa 0(%0),%%ymm7" "\n\t" \ 72 "vmovdqa 32(%0),%%ymm8" "\n\t" \ 73 "vmovdqa 64(%0),%%ymm6" "\n\t" \ 74 "vmovdqa 96(%0),%%ymm9" "\n\t" \ 77 "vmovdqa %%ymm7, 0(%0)" "\n\t" \ 78 "vmovdqa %%ymm8, 32(%0)" "\n\t" \ 79 "vmovdqa %%ymm6, 64(%0)" "\n\t" \ 80 "vmovdqa %%ymm9, 96(%0)" "\n\t" \ 92 "vmovdqa 32(%0),%%ymm8" "\n\t" \ 93 "vmovdqa 64(%0),%%ymm7" "\n\t" [all...] |
/external/boringssl/src/crypto/ec/asm/ |
p256-x86_64-asm.pl | 1315 vmovdqa .LTwo(%rip), $TWO 1321 vmovdqa .LOne(%rip), $M0 1322 vmovdqa .LTwo(%rip), $M1 1330 vmovdqa 32*0($in_t), $T0a 1331 vmovdqa 32*1($in_t), $T0b 1332 vmovdqa 32*2($in_t), $T0c 1334 vmovdqa 32*3($in_t), $T1a 1335 vmovdqa 32*4($in_t), $T1b 1336 vmovdqa 32*5($in_t), $T1c 1420 vmovdqa .LThree(%rip), $THRE [all...] |
/external/boringssl/src/crypto/sha/asm/ |
sha256-586.pl | 990 &vmovdqa ($t3,&QWP(256,$K256)); 1009 &vmovdqa (&QWP(32+0,"esp"),$t0); 1010 &vmovdqa (&QWP(32+16,"esp"),$t1); 1011 &vmovdqa (&QWP(32+32,"esp"),$t2); 1012 &vmovdqa (&QWP(32+48,"esp"),$t3); 1071 &vmovdqa (&QWP(32+16*$j,"esp"),$t2); 1121 &vmovdqa ($t3,&QWP(64,$K256)); [all...] |
sha1-x86_64.pl | 969 vmovdqa 64($K_XX_XX),@X[2] # pbswap mask 970 vmovdqa -64($K_XX_XX),$Kx # K_00_19 983 vmovdqa @X[0],0(%rsp) # X[]+K xfer to IALU 984 vmovdqa @X[1],16(%rsp) 985 vmovdqa @X[2],32(%rsp) 1020 &vmovdqa (eval(16*(($Xi-1)&3))."(%rsp)",@Tx[1]); # X[]+K xfer to IALU 1054 &vmovdqa ($Kx,eval(2*16*(($Xi)/5)-64)."($K_XX_XX)") if ($Xi%5==0); # K_XX_XX 1081 &vmovdqa ($Kx,eval(2*16*($Xi/5)-64)."($K_XX_XX)") if ($Xi%5==0); 1092 &vmovdqa (eval(16*(($Xi-1)&3))."(%rsp)",@Tx[1]); # X[]+K xfer to IALU 1136 &vmovdqa (eval(16*(($Xi-1)&3))."(%rsp)",@Tx[1]); # X[]+K xfer IAL [all...] |
/external/llvm/test/MC/Disassembler/X86/ |
x86-32.txt | 316 # CHECK: vmovdqa %xmm1, %xmm0 319 # CHECK: vmovdqa %xmm0, %xmm1 322 # CHECK: vmovdqa %ymm1, %ymm0 325 # CHECK: vmovdqa %ymm0, %ymm1
|