/external/llvm/test/CodeGen/X86/ |
avx512-vec-cmp.ll | 104 ; CHECK-NEXT: ## kill: YMM0<def> YMM0<kill> ZMM0<def> 107 ; CHECK-NEXT: ## kill: YMM0<def> YMM0<kill> ZMM0<kill> 118 ; CHECK-NEXT: ## kill: YMM0<def> YMM0<kill> ZMM0<def> 121 ; CHECK-NEXT: ## kill: YMM0<def> YMM0<kill> ZMM0<kill> 131 ; CHECK-NEXT: vpmaxud %ymm1, %ymm0, %ymm0 [all...] |
avx-vinsertf128.ll | 63 ; CHECK: vmovaps %ymm1, %ymm0 72 ; CHECK: vmovaps %ymm1, %ymm0 81 ; CHECK: vmovaps %ymm1, %ymm0
|
bswap-vector.ll | 149 ; CHECK-AVX2-NEXT: vpshufb {{.*#+}} ymm0 = ymm0[1,0,3,2,5,4,7,6,9,8,11,10,13,12,15,14,17,16,19,18,21,20,23,22,25,24,27,26,29,28,31,30] 154 ; CHECK-WIDE-AVX2-NEXT: vpshufb {{.*#+}} ymm0 = ymm0[1,0,3,2,5,4,7,6,9,8,11,10,13,12,15,14,17,16,19,18,21,20,23,22,25,24,27,26,29,28,31,30] 192 ; CHECK-AVX2-NEXT: vpshufb {{.*#+}} ymm0 = ymm0[3,2,1,0,7,6,5,4,11,10,9,8,15,14,13,12,19,18,17,16,23,22,21,20,27,26,25,24,31,30,29,28] 197 ; CHECK-WIDE-AVX2-NEXT: vpshufb {{.*#+}} ymm0 = ymm0[3,2,1,0,7,6,5,4,11,10,9,8,15,14,13,12,19,18,17,16,23,22,21,20,27,26,25,24,31,30,29,28] 239 ; CHECK-AVX2-NEXT: vpshufb {{.*#+}} ymm0 = ymm0[7,6,5,4,3,2,1,0,15,14,13,12,11,10,9,8,23,22,21,20,19,18,17,16,31,30,29,28,27,26,25,24 [all...] |
stackmap-liveness.ll | 56 %a4 = call <4 x double> asm sideeffect "", "={ymm0}"() nounwind 86 ; LiveOut Entry 3: %YMM0 (32 bytes) 101 call void asm sideeffect "", "{r8},{ah},{ymm0},{ymm1}"(i64 %a2, i8 %a3, <4 x double> %a4, <4 x double> %a5) nounwind
|
combine-avx2-intrinsics.ll | 13 ; CHECK: vpsrad $8, %ymm0, %ymm0 23 ; CHECK: vpsraw $8, %ymm0, %ymm0
|
vector-sext.ll | 51 ; AVX1-NEXT: vinsertf128 $1, %xmm0, %ymm1, %ymm0 56 ; AVX2-NEXT: vpmovsxwd %xmm0, %ymm0 147 ; AVX1-NEXT: vinsertf128 $1, %xmm0, %ymm1, %ymm0 152 ; AVX2-NEXT: vpmovsxdq %xmm0, %ymm0 489 ; AVX1-NEXT: vinsertf128 $1, %xmm0, %ymm1, %ymm0 496 ; AVX2-NEXT: vpmovsxdq %xmm0, %ymm0 554 ; AVX1-NEXT: vinsertf128 $1, %xmm1, %ymm0, %ymm0 559 ; AVX2-NEXT: vpmovsxbw (%rdi), %ymm0 655 ; AVX1-NEXT: vinsertf128 $1, %xmm0, %ymm1, %ymm0 [all...] |
fp-load-trunc.ll | 84 ; AVX-NEXT: vinsertf128 $1, %xmm1, %ymm0, %ymm0
|
musttail-fastcall.ll | 64 ; AVX-DAG: vmovups %ymm0, {{.*}} 87 ; AVX-DAG: vmovups {{.*}}, %ymm0
|
vec_uint_to_fp.ll | 146 ; AVX2-NEXT: vpblendw $170, [[LOWCST]], %ymm0, [[LOWVEC:%ymm[0-9]+]] 147 ; AVX2-NEXT: vpsrld $16, %ymm0, [[SHIFTVEC:%ymm[0-9]+]] 152 ; AVX2-NEXT: vaddps [[TMP]], [[LOWVEC]], %ymm0
|
vector-shuffle-combining.ll | [all...] |
preserve_allcc64.ll | 67 ;AVX-NEXT: vmovups %ymm0
|
avx512-cvt.ll | 78 ; CHECK: vcvtps2pd %ymm0, %zmm0
|
dynamic-allocas-VLAs.ll | 125 ; CHECK: vmovups (%rsp), %ymm0
|
/external/clang/test/CodeGen/ |
asm.c | 221 __asm__ volatile("vmovaps %0, %%ymm0" :: "m" (*(__m256i*)p) : "ymm0");
|
/external/llvm/lib/Target/X86/ |
X86CallingConv.td | 48 // 256-bit vectors are returned in YMM0 and XMM1, when they fit. YMM2 and YMM3 52 CCAssignToReg<[YMM0,YMM1,YMM2,YMM3]>>, 108 CCAssignToReg<[YMM0,YMM1,YMM2,YMM3]>>, 135 CCAssignToReg<[YMM0,YMM1,YMM2,YMM3]>>, 273 CCAssignToReg<[YMM0, YMM1, YMM2, YMM3, 360 CCAssignToReg<[YMM0, YMM1, YMM2, YMM3, YMM4, YMM5]>>, 462 CCAssignToReg<[YMM0, YMM1, YMM2, YMM3]>>>>, 520 CCAssignToReg<[YMM0, YMM1, YMM2, YMM3, YMM4, YMM5]>>, 631 CCAssignToReg<[YMM0, YMM1, YMM2, YMM3]>>,
|
X86VZeroUpper.cpp | 105 return (Reg >= X86::YMM0 && Reg <= X86::YMM15); 118 for (unsigned reg = X86::YMM0; reg <= X86::YMM15; ++reg) { 148 for (unsigned reg = X86::YMM0; reg <= X86::YMM15; ++reg) {
|
/external/lldb/test/functionalities/register/ |
TestRegisters.py | 192 self.vector_write_and_read(currentFrame, "ymm0", new_value) 194 self.expect("expr $ymm0", substrs = ['vector_type']) 196 self.runCmd("register read ymm0")
|
/external/llvm/test/MC/Disassembler/X86/ |
intel-syntax.txt | 111 # CHECK: vpgatherqq ymm2, qword ptr [rdi + 2*ymm1], ymm0
|
avx-512.txt | 46 # CHECK: vgatherdpd (%rsi,%ymm0,4), %zmm1 {%k2}
|
/external/libvpx/libvpx/third_party/libyuv/source/ |
row_posix.cc | [all...] |
/external/boringssl/src/crypto/bn/asm/ |
rsaz-avx2.pl | 120 my $ACC0="%ymm0"; 858 my $ACC0="%ymm0"; [all...] |
/external/llvm/test/MC/X86/ |
x86_64-xop-encoding.s | 616 // CHECK: vpermil2pd $3, 8(%rax), %ymm0, %ymm4, %ymm6 618 vpermil2pd $3, 8(%rax), %ymm0, %ymm4, %ymm6 645 // CHECK: vpermil2pd $1, %ymm1, 4(%rax), %ymm1, %ymm0 647 vpermil2pd $1, %ymm1, 4(%rax), %ymm1, %ymm0
|
x86_64-avx-encoding.s | [all...] |
/external/llvm/lib/Target/X86/AsmParser/ |
X86Operand.h | 240 getMemIndexReg() >= X86::YMM0 && getMemIndexReg() <= X86::YMM15; 248 getMemIndexReg() >= X86::YMM0 && getMemIndexReg() <= X86::YMM15;
|
/external/llvm/lib/Target/X86/Disassembler/ |
X86DisassemblerDecoder.h | 251 ENTRY(YMM0) \
|