/external/llvm/test/CodeGen/X86/ |
sandybridge-loads.ll | 4 ;CHECK: vmovaps 6 ;CHECK: vmovaps 24 ; CHECK: vmovaps 25 ; CHECK: vmovaps 27 ; CHECK: vmovaps 29 ; CHECK: vmovaps
|
avx-load-store.ll | 4 ; CHECK: vmovaps 5 ; CHECK: vmovaps 6 ; CHECK: vmovaps 7 ; CHECK: vmovaps 8 ; CHECK: vmovaps 9 ; CHECK: vmovaps 50 ; CHECK: vmovaps %ymm 65 ; CHECK: vmovaps %ymm 84 ; CHECK: vmovaps %xmm 85 ; CHECK: vmovaps %xm [all...] |
avx-select.ll | 4 ; CHECK: vmovaps 14 ; CHECK: vmovaps
|
vaargs.ll | 12 ; CHECK-NEXT: vmovaps %xmm0, 48(%rsp) 13 ; CHECK-NEXT: vmovaps %xmm1, 64(%rsp) 14 ; CHECK-NEXT: vmovaps %xmm2, 80(%rsp) 15 ; CHECK-NEXT: vmovaps %xmm3, 96(%rsp) 16 ; CHECK-NEXT: vmovaps %xmm4, 112(%rsp) 17 ; CHECK-NEXT: vmovaps %xmm5, 128(%rsp) 18 ; CHECK-NEXT: vmovaps %xmm6, 144(%rsp) 19 ; CHECK-NEXT: vmovaps %xmm7, 160(%rsp)
|
v8i1-masks.ll | 4 ;CHECK: vmovaps 9 ;CHECK: vmovaps 28 ;CHECK: vmovaps
|
2012-01-11-split-cv.ll | 5 ;CHECK: vmovaps
|
fold-vex.ll | 5 ;CHECK-NOT: vmovaps
|
long-extend.ll | 12 ; CHECK: vmovaps [[REG_result1]], 32(%rdi) 13 ; CHECK: vmovaps [[REG_result0]], (%rdi)
|
avx-intel-ocl.ll | 69 ; WIN64: vmovaps {{%ymm([6-9]|1[0-5])}}, {{.*(%rbp).*}} # 32-byte Spill 70 ; WIN64: vmovaps {{%ymm([6-9]|1[0-5])}}, {{.*(%rbp).*}} # 32-byte Spill 71 ; WIN64: vmovaps {{%ymm([6-9]|1[0-5])}}, {{.*(%rbp).*}} # 32-byte Spill 72 ; WIN64: vmovaps {{%ymm([6-9]|1[0-5])}}, {{.*(%rbp).*}} # 32-byte Spill 73 ; WIN64: vmovaps {{%ymm([6-9]|1[0-5])}}, {{.*(%rbp).*}} # 32-byte Spill 74 ; WIN64: vmovaps {{%ymm([6-9]|1[0-5])}}, {{.*(%rbp).*}} # 32-byte Spill 75 ; WIN64: vmovaps {{%ymm([6-9]|1[0-5])}}, {{.*(%rbp).*}} # 32-byte Spill 76 ; WIN64: vmovaps {{%ymm([6-9]|1[0-5])}}, {{.*(%rbp).*}} # 32-byte Spill 77 ; WIN64: vmovaps {{%ymm([6-9]|1[0-5])}}, {{.*(%rbp).*}} # 32-byte Spill 78 ; WIN64: vmovaps {{%ymm([6-9]|1[0-5])}}, {{.*(%rbp).*}} # 32-byte Spil [all...] |
fold-vector-sext-zext.ll | 19 ; CHECK: vmovaps 31 ; CHECK: vmovaps 43 ; CHECK: vmovaps 55 ; CHECK: vmovaps 69 ; CHECK: vmovaps 82 ; CHECK: vmovaps 98 ; CHECK: vmovaps 115 ; CHECK: vmovaps 131 ; CHECK: vmovaps 148 ; CHECK: vmovaps [all...] |
2012-05-19-avx2-store.ll | 5 ; CHECK: vmovaps
|
avx-varargs-x86_64.ll | 9 ; CHECK: vmovaps %ymm0, (%rsp)
|
memset-sse-stack-realignment.ll | 34 ; AVX1: vmovaps %ymm 39 ; AVX2: vmovaps %ymm 67 ; AVX1: vmovaps %xmm 72 ; AVX2: vmovaps %xmm
|
avx512-select.ll | 4 ; CHECK: vmovaps 14 ; CHECK: vmovaps
|
avx-vextractf128.ll | 23 ; CHECK-NOT: vmovaps %xmm0, (%rdi) 37 ; CHECK-NOT: vmovaps %xmm0, (%rdi) 51 ; CHECK-NOT: vmovaps %xmm0, (%rdi) 65 ; CHECK: vmovaps %xmm0, (%rdi) 75 ; CHECK: vmovaps %xmm0, (%rdi) 85 ; CHECK: vmovaps %xmm0, (%rdi)
|
anyregcc.ll | 418 ;AVX: vmovaps %ymm15 419 ;AVX-NEXT: vmovaps %ymm14 420 ;AVX-NEXT: vmovaps %ymm13 421 ;AVX-NEXT: vmovaps %ymm12 422 ;AVX-NEXT: vmovaps %ymm11 423 ;AVX-NEXT: vmovaps %ymm10 424 ;AVX-NEXT: vmovaps %ymm9 425 ;AVX-NEXT: vmovaps %ymm8 426 ;AVX-NEXT: vmovaps %ymm7 427 ;AVX-NEXT: vmovaps %ymm [all...] |
avx-win64.ll | 13 ; CHECK: vmovaps %ymm{{.*}}(%r 14 ; CHECK: vmovaps %ymm{{.*}}(%r 17 ; CHECK: vmovaps {{.*\(%r.*}}, %ymm 18 ; CHECK: vmovaps {{.*\(%r.*}}, %ymm
|
vec_shuffle-30.ll | 8 ; CHECK-NOT: vmovaps
|
widen_load-1.ll | 12 ; AVX: vmovaps compl+128(%rip), %xmm0 13 ; AVX: vmovaps %xmm0, (%rsp)
|
avx-vinsertf128.ll | 69 ; CHECK: vmovaps %ymm1, %ymm0 85 ; CHECK: vmovaps %ymm1, %ymm0 101 ; CHECK: vmovaps %ymm1, %ymm0 110 ; CHECK-NOT: vmovaps
|
avx2-gather.ll | 18 ; CHECK: vmovaps
|
avx-basic.ll | 10 ; CHECK: vmovaps 18 ; CHECK: vmovaps 19 ; CHECK: vmovaps
|
memset.ll | 32 ; YMM: vmovaps [[Z]],
|
/external/valgrind/main/none/tests/amd64/ |
fma.c | 103 __asm __volatile__ ("vmovaps (%1), %%xmm9; vmovaps (%2), %%xmm7; vmovaps (%3), %%xmm8;" 105 "vmovaps %%xmm9, (%0)" : : "r" (&ft.res[i]), "r" (&ft.x[i]), 109 __asm __volatile__ ("vmovaps (%1), %%xmm9; vmovaps (%3), %%xmm8;" 111 "vmovaps %%xmm9, (%0)" : : "r" (&ft.res[i]), "r" (&ft.x[i]), 115 __asm __volatile__ ("vmovaps (%1), %%xmm9; vmovaps (%3), %%xmm7; vmovaps (%2), %%xmm8; [all...] |
/external/chromium_org/third_party/boringssl/src/crypto/bn/asm/ |
rsaz-avx2.pl | 171 vmovaps %xmm6,-0xd8(%rax) 172 vmovaps %xmm7,-0xc8(%rax) 173 vmovaps %xmm8,-0xb8(%rax) 174 vmovaps %xmm9,-0xa8(%rax) 175 vmovaps %xmm10,-0x98(%rax) 176 vmovaps %xmm11,-0x88(%rax) 177 vmovaps %xmm12,-0x78(%rax) 178 vmovaps %xmm13,-0x68(%rax) 179 vmovaps %xmm14,-0x58(%rax) 180 vmovaps %xmm15,-0x48(%rax [all...] |