/external/libvpx/libvpx/third_party/libyuv/source/ |
row_gcc.cc | 755 "vmovdqu " MEMACCESS2(0x40,0) ",%%ymm2 \n" 759 "vpmaddubsw %%ymm4,%%ymm2,%%ymm2 \n" 763 "vphaddw %%ymm3,%%ymm2,%%ymm2 \n" 765 "vpsrlw $0x7,%%ymm2,%%ymm2 \n" 766 "vpackuswb %%ymm2,%%ymm0,%%ymm0 \n" // mutates. 796 "vmovdqu " MEMACCESS2(0x40,0) ",%%ymm2 \n" [all...] |
scale_win.cc | 454 vmovdqu ymm2, [eax + esi * 2] local 456 vpavgb ymm2, ymm2, [eax + edi] local 459 vpavgb ymm0, ymm0, ymm2 462 vpand ymm2, ymm0, ymm7 // average columns (64 to 32 pixels) local 466 vpavgw ymm0, ymm0, ymm2 471 vpand ymm2, ymm0, ymm7 // average columns (32 to 16 pixels) local 473 vpavgw ymm0, ymm0, ymm2 844 vpunpcklbw ymm2, ymm3, ymm5 local 846 vpaddusw ymm0, ymm2, [edx] // sum 16 word local [all...] |
/external/libyuv/files/source/ |
scale_win.cc | 283 vmovdqu ymm2, [eax + esi] local 288 vpmaddubsw ymm2, ymm2, ymm4 local 290 vpaddw ymm0, ymm0, ymm2 // vertical add 453 vmovdqu ymm2, [eax + esi] local 457 vpmaddubsw ymm2, ymm2, ymm4 local 459 vpaddw ymm0, ymm0, ymm2 // vertical add rows 0, 1 461 vmovdqu ymm2, [eax + esi * 2] local 463 vpmaddubsw ymm2, ymm2, ymm local 467 vmovdqu ymm2, [eax + edi] local 470 vpmaddubsw ymm2, ymm2, ymm4 local 847 vpunpcklbw ymm2, ymm3, ymm5 local 849 vpaddusw ymm0, ymm2, [edx] \/\/ sum 16 words local [all...] |
scale_gcc.cc | 259 MEMOPREG(vmovdqu,0x00,0,3,1,ymm2) // vmovdqu (%0,%3,1),%%ymm2 264 "vpmaddubsw %%ymm4,%%ymm2,%%ymm2 \n" 266 "vpaddw %%ymm2,%%ymm0,%%ymm0 \n" 415 MEMOPREG(vmovdqu,0x00,0,3,1,ymm2) // vmovdqu (%0,%3,1),%%ymm2 419 "vpmaddubsw %%ymm4,%%ymm2,%%ymm2 \n" 421 "vpaddw %%ymm2,%%ymm0,%%ymm0 \n [all...] |
/external/llvm/test/CodeGen/X86/ |
vector-idiv.ll | 115 ; AVX-NEXT: vpshufd {{.*#+}} ymm2 = ymm1[1,1,3,3,5,5,7,7] 117 ; AVX-NEXT: vpmuludq %ymm2, %ymm3, %ymm2 120 ; AVX-NEXT: vpblendd {{.*#+}} ymm1 = ymm1[0],ymm2[1],ymm1[2],ymm2[3],ymm1[4],ymm2[5],ymm1[6],ymm2[7] [all...] |
fma-intrinsics-x86.ll | 124 ; CHECK-FMA-NEXT: vfmadd213ps %ymm2, %ymm1, %ymm0 126 ; CHECK-FMA4-NEXT: vfmaddps %ymm2, %ymm1, %ymm0, %ymm0 142 ; CHECK-FMA-NEXT: vfmadd213pd %ymm2, %ymm1, %ymm0 144 ; CHECK-FMA4-NEXT: vfmaddpd %ymm2, %ymm1, %ymm0, %ymm0 269 ; CHECK-FMA-NEXT: vfmsub213ps %ymm2, %ymm1, %ymm0 271 ; CHECK-FMA4-NEXT: vfmsubps %ymm2, %ymm1, %ymm0, %ymm0 287 ; CHECK-FMA-NEXT: vfmsub213pd %ymm2, %ymm1, %ymm0 289 ; CHECK-FMA4-NEXT: vfmsubpd %ymm2, %ymm1, %ymm0, %ymm0 414 ; CHECK-FMA-NEXT: vfnmadd213ps %ymm2, %ymm1, %ymm0 416 ; CHECK-FMA4-NEXT: vfnmaddps %ymm2, %ymm1, %ymm0, %ymm [all...] |
avx2-arith.ll | 83 ; CHECK-NEXT: vpmovsxbw %xmm2, %ymm2 86 ; CHECK-NEXT: vpmullw %ymm2, %ymm3, %ymm2 87 ; CHECK-NEXT: vextracti128 $1, %ymm2, %xmm3
|
avx512-build-vector.ll | 21 ; CHECK-NEXT: vinsertf128 $1, %xmm0, %ymm2, %ymm0
|
musttail-fastcall.ll | 66 ; AVX-DAG: vmovups %ymm2, {{.*}} 89 ; AVX-DAG: vmovups {{.*}}, %ymm2
|
avx512bwvl-intrinsics.ll | [all...] |
vselect-minmax.ll | [all...] |
vector-shuffle-256-v8.ll | 280 ; AVX2-NEXT: vmovaps {{.*#+}} ymm2 = <u,0,u,1,u,2,u,3> 281 ; AVX2-NEXT: vpermps %ymm1, %ymm2, %ymm1 282 ; AVX2-NEXT: vmovaps {{.*#+}} ymm2 = <0,u,1,u,2,u,3,u> 283 ; AVX2-NEXT: vpermps %ymm0, %ymm2, %ymm0 297 ; AVX1-NEXT: vinsertf128 $1, %xmm0, %ymm2, %ymm0 302 ; AVX2-NEXT: vmovaps {{.*#+}} ymm2 = <0,u,u,u,1,u,u,u> 303 ; AVX2-NEXT: vpermps %ymm0, %ymm2, %ymm0 304 ; AVX2-NEXT: vmovaps {{.*#+}} ymm2 = <u,0,1,1,u,2,3,3> 305 ; AVX2-NEXT: vpermps %ymm1, %ymm2, %ymm1 317 ; AVX1-NEXT: vinsertf128 $1, %xmm0, %ymm2, %ymm [all...] |
vector-trunc.ll | 55 ; AVX2-NEXT: vmovdqa {{.*#+}} ymm2 = <0,2,4,6,u,u,u,u> 56 ; AVX2-NEXT: vpermd %ymm0, %ymm2, %ymm0 57 ; AVX2-NEXT: vpermd %ymm1, %ymm2, %ymm1 138 ; AVX2-NEXT: vmovdqa {{.*#+}} ymm2 = <0,2,4,6,u,u,u,u> 139 ; AVX2-NEXT: vpermd %ymm0, %ymm2, %ymm0 140 ; AVX2-NEXT: vpermd %ymm1, %ymm2, %ymm1 190 ; AVX2-NEXT: vmovdqa {{.*#+}} ymm2 = <0,2,4,6,u,u,u,u> 191 ; AVX2-NEXT: vpermd %ymm0, %ymm2, %ymm0 192 ; AVX2-NEXT: vpermd %ymm1, %ymm2, %ymm1 356 ; AVX2-NEXT: vmovdqa {{.*#+}} ymm2 = [0,1,4,5,8,9,12,13,128,128,128,128,128,128,128,128,0,1,4,5,8,9,12,13,128,128,128,128,128,128,128 (…) [all...] |
vec_minmax_sint.ll | 159 ; AVX1-NEXT: vinsertf128 $1, %xmm2, %ymm3, %ymm2 160 ; AVX1-NEXT: vblendvpd %ymm2, %ymm0, %ymm1, %ymm0 165 ; AVX2-NEXT: vpcmpgtq %ymm1, %ymm0, %ymm2 166 ; AVX2-NEXT: vblendvpd %ymm2, %ymm0, %ymm1, %ymm0 171 ; AVX512-NEXT: vpcmpgtq %ymm1, %ymm0, %ymm2 172 ; AVX512-NEXT: vblendvpd %ymm2, %ymm0, %ymm1, %ymm0 563 ; AVX1-NEXT: vinsertf128 $1, %xmm2, %ymm3, %ymm2 564 ; AVX1-NEXT: vblendvpd %ymm2, %ymm0, %ymm1, %ymm0 569 ; AVX2-NEXT: vpcmpgtq %ymm0, %ymm1, %ymm2 571 ; AVX2-NEXT: vpxor %ymm3, %ymm2, %ymm [all...] |
avx512-vbroadcast.ll | 142 ; ALL-NEXT: vpxor %ymm2, %ymm2, %ymm2 167 ; ALL-NEXT: vpxor %ymm2, %ymm2, %ymm2
|
pmul.ll | 372 ; AVX2-NEXT: vpmuludq %ymm1, %ymm0, %ymm2 376 ; AVX2-NEXT: vpaddq %ymm0, %ymm2, %ymm0 412 ; AVX2-NEXT: vpmuludq %ymm1, %ymm0, %ymm2 416 ; AVX2-NEXT: vpaddq %ymm3, %ymm2, %ymm2 420 ; AVX2-NEXT: vpaddq %ymm0, %ymm2, %ymm0
|
machine-combiner.ll | 307 ; AVX-NEXT: vaddps %ymm3, %ymm2, %ymm1 322 ; AVX-NEXT: vaddpd %ymm3, %ymm2, %ymm1 337 ; AVX-NEXT: vmulps %ymm3, %ymm2, %ymm1 352 ; AVX-NEXT: vmulpd %ymm3, %ymm2, %ymm1 559 ; AVX-NEXT: vminps %ymm3, %ymm2, %ymm1 576 ; AVX-NEXT: vmaxps %ymm3, %ymm2, %ymm1 593 ; AVX-NEXT: vminpd %ymm3, %ymm2, %ymm1 610 ; AVX-NEXT: vmaxpd %ymm3, %ymm2, %ymm1
|
/toolchain/binutils/binutils-2.25/gas/testsuite/gas/i386/ |
x86-64-avx-gather-intel.d | 14 [ ]*[a-f0-9]+: c4 e2 ed 92 4c 7d 00 vgatherdpd ymm1,QWORD PTR \[rbp\+xmm7\*2\+0x0\],ymm2 15 [ ]*[a-f0-9]+: c4 e2 ed 93 4c 7d 00 vgatherqpd ymm1,QWORD PTR \[rbp\+ymm7\*2\+0x0\],ymm2 38 [ ]*[a-f0-9]+: c4 e2 6d 92 4c 7d 00 vgatherdps ymm1,DWORD PTR \[rbp\+ymm7\*2\+0x0\],ymm2 62 [ ]*[a-f0-9]+: c4 e2 6d 90 4c 7d 00 vpgatherdd ymm1,DWORD PTR \[rbp\+ymm7\*2\+0x0\],ymm2 86 [ ]*[a-f0-9]+: c4 e2 ed 90 4c 7d 00 vpgatherdq ymm1,QWORD PTR \[rbp\+xmm7\*2\+0x0\],ymm2 87 [ ]*[a-f0-9]+: c4 e2 ed 91 4c 7d 00 vpgatherqq ymm1,QWORD PTR \[rbp\+ymm7\*2\+0x0\],ymm2 110 [ ]*[a-f0-9]+: c4 e2 ed 92 4c 7d 00 vgatherdpd ymm1,QWORD PTR \[rbp\+xmm7\*2\+0x0\],ymm2 111 [ ]*[a-f0-9]+: c4 e2 ed 93 4c 7d 00 vgatherqpd ymm1,QWORD PTR \[rbp\+ymm7\*2\+0x0\],ymm2 134 [ ]*[a-f0-9]+: c4 e2 6d 92 4c 7d 00 vgatherdps ymm1,DWORD PTR \[rbp\+ymm7\*2\+0x0\],ymm2 158 [ ]*[a-f0-9]+: c4 e2 6d 90 4c 7d 00 vpgatherdd ymm1,DWORD PTR \[rbp\+ymm7\*2\+0x0\],ymm2 [all...] |
x86-64-avx-gather.d | 13 [ ]*[a-f0-9]+: c4 e2 ed 92 4c 7d 00 vgatherdpd %ymm2,0x0\(%rbp,%xmm7,2\),%ymm1 14 [ ]*[a-f0-9]+: c4 e2 ed 93 4c 7d 00 vgatherqpd %ymm2,0x0\(%rbp,%ymm7,2\),%ymm1 37 [ ]*[a-f0-9]+: c4 e2 6d 92 4c 7d 00 vgatherdps %ymm2,0x0\(%rbp,%ymm7,2\),%ymm1 61 [ ]*[a-f0-9]+: c4 e2 6d 90 4c 7d 00 vpgatherdd %ymm2,0x0\(%rbp,%ymm7,2\),%ymm1 85 [ ]*[a-f0-9]+: c4 e2 ed 90 4c 7d 00 vpgatherdq %ymm2,0x0\(%rbp,%xmm7,2\),%ymm1 86 [ ]*[a-f0-9]+: c4 e2 ed 91 4c 7d 00 vpgatherqq %ymm2,0x0\(%rbp,%ymm7,2\),%ymm1 109 [ ]*[a-f0-9]+: c4 e2 ed 92 4c 7d 00 vgatherdpd %ymm2,0x0\(%rbp,%xmm7,2\),%ymm1 110 [ ]*[a-f0-9]+: c4 e2 ed 93 4c 7d 00 vgatherqpd %ymm2,0x0\(%rbp,%ymm7,2\),%ymm1 133 [ ]*[a-f0-9]+: c4 e2 6d 92 4c 7d 00 vgatherdps %ymm2,0x0\(%rbp,%ymm7,2\),%ymm1 157 [ ]*[a-f0-9]+: c4 e2 6d 90 4c 7d 00 vpgatherdd %ymm2,0x0\(%rbp,%ymm7,2\),%ymm [all...] |
x86-64-xop.s | 24 # Tests for op VFRCZPD ymm2/mem256, ymm1 (at&t syntax) 25 VFRCZPD %ymm2,%ymm15 31 VFRCZPD %ymm2,%ymm0 58 # Tests for op VFRCZPS ymm2/mem256, ymm1 (at&t syntax) 59 VFRCZPS %ymm2,%ymm15 65 VFRCZPS %ymm2,%ymm0 126 # Tests for op VPCMOV ymm4, ymm3/mem256, ymm2, ymm1 (at&t syntax) 128 VPCMOV %ymm2,(%r9),%ymm0,%ymm0 134 VPCMOV %ymm2,%ymm0,%ymm0,%ymm15 135 VPCMOV %ymm2,(%r9),%ymm15,%ymm1 [all...] |
xop.s | 24 # Tests for op VFRCZPD ymm2/mem256, ymm1 (at&t syntax) 58 # Tests for op VFRCZPS ymm2/mem256, ymm1 (at&t syntax) 126 # Tests for op VPCMOV ymm4, ymm3/mem256, ymm2, ymm1 (at&t syntax) 160 # Tests for op VPCMOV ymm4/mem256, ymm3, ymm2, ymm1 (at&t syntax) 331 # Testing VPERMIL2PD imm8, ymm4, ymm3/mem256, ymm2, ymm1 (at&t syntax) 332 VPERMIL2PD $0x3,%ymm6,%ymm7,%ymm1,%ymm2 335 VPERMIL2PD $0x0,%ymm5,%ymm6,%ymm0,%ymm2 337 VPERMIL2PD $0x0,%ymm7,%ymm6,%ymm2,%ymm0 340 # Testing VPERMIL2PD imm8, ymm4/mem256, ymm3, ymm2, ymm1 (at&t syntax) 344 VPERMIL2PD $0x2,(%esi,%eax,1),%ymm2,%ymm6,%ymm [all...] |
/external/llvm/test/MC/X86/ |
x86_64-xop-encoding.s | 153 // CHECK: vfrczps %ymm2, %ymm4 155 vfrczps %ymm2, %ymm4 164 // CHECK: vfrczpd (%rcx,%rax), %ymm2 166 vfrczpd (%rcx,%rax), %ymm2 595 // CHECK: vpcmov %ymm1, %ymm2, %ymm3, %ymm4 597 vpcmov %ymm1, %ymm2, %ymm3, %ymm4 598 // CHECK: vpcmov (%rax), %ymm2, %ymm3, %ymm4 600 vpcmov (%rax), %ymm2, %ymm3, %ymm4 622 // CHECK: vpermil2pd $1, %ymm1, %ymm2, %ymm3, %ymm4 624 vpermil2pd $1, %ymm1, %ymm2, %ymm3, %ymm [all...] |
/external/llvm/test/MC/Disassembler/X86/ |
simple-tests.txt | 182 # CHECK: vblendps $129, %ymm2, %ymm5, %ymm1 188 # CHECK: vblendpd $129, %ymm2, %ymm5, %ymm1 200 # CHECK: vdpps $129, %ymm2, %ymm5, %ymm1 794 # CHECK: vfmaddps %ymm1, %ymm2, %ymm0, %ymm0 797 # CHECK: vfmaddps %ymm2, %ymm1, %ymm0, %ymm0 812 # CHECK: vpermil2ps $0, %ymm4, %ymm4, %ymm2, %ymm2 821 # CHECK: vgatherdpd %ymm0, (%rdi,%xmm1,2), %ymm2 833 # CHECK: vpgatherdq %ymm0, (%rdi,%xmm1,2), %ymm2 905 # CHECK: vpcmov %ymm1, %ymm2, %ymm3, %ymm [all...] |
x86-32.txt | 199 # CHECK: vblendps $129, %ymm2, %ymm5, %ymm1 205 # CHECK: vblendpd $129, %ymm2, %ymm5, %ymm1 217 # CHECK: vdpps $129, %ymm2, %ymm5, %ymm1 739 # CHECK: vpermq $238, %ymm2, %ymm2
|
/external/llvm/lib/Target/X86/ |
X86CallingConv.td | 59 // 256-bit vectors are returned in YMM0 and XMM1, when they fit. YMM2 and YMM3 63 CCAssignToReg<[YMM0,YMM1,YMM2,YMM3]>>, 119 CCAssignToReg<[YMM0,YMM1,YMM2,YMM3]>>, 146 CCAssignToReg<[YMM0,YMM1,YMM2,YMM3]>>, 308 CCAssignToReg<[YMM0, YMM1, YMM2, YMM3, 412 CCAssignToReg<[YMM0, YMM1, YMM2, YMM3, YMM4, YMM5]>>, 504 CCAssignToReg<[YMM0, YMM1, YMM2]>>>>, 523 CCAssignToReg<[YMM0, YMM1, YMM2, YMM3]>>>>, 617 CCAssignToReg<[YMM0, YMM1, YMM2, YMM3, YMM4, YMM5]>>, 728 CCAssignToReg<[YMM0, YMM1, YMM2, YMM3]>> [all...] |