HomeSort by relevance Sort by last modified time
    Searched full:ymm2 (Results 76 - 100 of 133) sorted by null

1 2 34 5 6

  /external/libvpx/libvpx/third_party/libyuv/source/
row_gcc.cc 755 "vmovdqu " MEMACCESS2(0x40,0) ",%%ymm2 \n"
759 "vpmaddubsw %%ymm4,%%ymm2,%%ymm2 \n"
763 "vphaddw %%ymm3,%%ymm2,%%ymm2 \n"
765 "vpsrlw $0x7,%%ymm2,%%ymm2 \n"
766 "vpackuswb %%ymm2,%%ymm0,%%ymm0 \n" // mutates.
796 "vmovdqu " MEMACCESS2(0x40,0) ",%%ymm2 \n"
    [all...]
scale_win.cc 454 vmovdqu ymm2, [eax + esi * 2] local
456 vpavgb ymm2, ymm2, [eax + edi] local
459 vpavgb ymm0, ymm0, ymm2
462 vpand ymm2, ymm0, ymm7 // average columns (64 to 32 pixels) local
466 vpavgw ymm0, ymm0, ymm2
471 vpand ymm2, ymm0, ymm7 // average columns (32 to 16 pixels) local
473 vpavgw ymm0, ymm0, ymm2
844 vpunpcklbw ymm2, ymm3, ymm5 local
846 vpaddusw ymm0, ymm2, [edx] // sum 16 word local
    [all...]
  /external/libyuv/files/source/
scale_win.cc 283 vmovdqu ymm2, [eax + esi] local
288 vpmaddubsw ymm2, ymm2, ymm4 local
290 vpaddw ymm0, ymm0, ymm2 // vertical add
453 vmovdqu ymm2, [eax + esi] local
457 vpmaddubsw ymm2, ymm2, ymm4 local
459 vpaddw ymm0, ymm0, ymm2 // vertical add rows 0, 1
461 vmovdqu ymm2, [eax + esi * 2] local
463 vpmaddubsw ymm2, ymm2, ymm local
467 vmovdqu ymm2, [eax + edi] local
470 vpmaddubsw ymm2, ymm2, ymm4 local
847 vpunpcklbw ymm2, ymm3, ymm5 local
849 vpaddusw ymm0, ymm2, [edx] \/\/ sum 16 words local
    [all...]
scale_gcc.cc 259 MEMOPREG(vmovdqu,0x00,0,3,1,ymm2) // vmovdqu (%0,%3,1),%%ymm2
264 "vpmaddubsw %%ymm4,%%ymm2,%%ymm2 \n"
266 "vpaddw %%ymm2,%%ymm0,%%ymm0 \n"
415 MEMOPREG(vmovdqu,0x00,0,3,1,ymm2) // vmovdqu (%0,%3,1),%%ymm2
419 "vpmaddubsw %%ymm4,%%ymm2,%%ymm2 \n"
421 "vpaddw %%ymm2,%%ymm0,%%ymm0 \n
    [all...]
  /external/llvm/test/CodeGen/X86/
vector-idiv.ll 115 ; AVX-NEXT: vpshufd {{.*#+}} ymm2 = ymm1[1,1,3,3,5,5,7,7]
117 ; AVX-NEXT: vpmuludq %ymm2, %ymm3, %ymm2
120 ; AVX-NEXT: vpblendd {{.*#+}} ymm1 = ymm1[0],ymm2[1],ymm1[2],ymm2[3],ymm1[4],ymm2[5],ymm1[6],ymm2[7]
    [all...]
fma-intrinsics-x86.ll 124 ; CHECK-FMA-NEXT: vfmadd213ps %ymm2, %ymm1, %ymm0
126 ; CHECK-FMA4-NEXT: vfmaddps %ymm2, %ymm1, %ymm0, %ymm0
142 ; CHECK-FMA-NEXT: vfmadd213pd %ymm2, %ymm1, %ymm0
144 ; CHECK-FMA4-NEXT: vfmaddpd %ymm2, %ymm1, %ymm0, %ymm0
269 ; CHECK-FMA-NEXT: vfmsub213ps %ymm2, %ymm1, %ymm0
271 ; CHECK-FMA4-NEXT: vfmsubps %ymm2, %ymm1, %ymm0, %ymm0
287 ; CHECK-FMA-NEXT: vfmsub213pd %ymm2, %ymm1, %ymm0
289 ; CHECK-FMA4-NEXT: vfmsubpd %ymm2, %ymm1, %ymm0, %ymm0
414 ; CHECK-FMA-NEXT: vfnmadd213ps %ymm2, %ymm1, %ymm0
416 ; CHECK-FMA4-NEXT: vfnmaddps %ymm2, %ymm1, %ymm0, %ymm
    [all...]
avx2-arith.ll 83 ; CHECK-NEXT: vpmovsxbw %xmm2, %ymm2
86 ; CHECK-NEXT: vpmullw %ymm2, %ymm3, %ymm2
87 ; CHECK-NEXT: vextracti128 $1, %ymm2, %xmm3
avx512-build-vector.ll 21 ; CHECK-NEXT: vinsertf128 $1, %xmm0, %ymm2, %ymm0
musttail-fastcall.ll 66 ; AVX-DAG: vmovups %ymm2, {{.*}}
89 ; AVX-DAG: vmovups {{.*}}, %ymm2
avx512bwvl-intrinsics.ll     [all...]
vselect-minmax.ll     [all...]
vector-shuffle-256-v8.ll 280 ; AVX2-NEXT: vmovaps {{.*#+}} ymm2 = <u,0,u,1,u,2,u,3>
281 ; AVX2-NEXT: vpermps %ymm1, %ymm2, %ymm1
282 ; AVX2-NEXT: vmovaps {{.*#+}} ymm2 = <0,u,1,u,2,u,3,u>
283 ; AVX2-NEXT: vpermps %ymm0, %ymm2, %ymm0
297 ; AVX1-NEXT: vinsertf128 $1, %xmm0, %ymm2, %ymm0
302 ; AVX2-NEXT: vmovaps {{.*#+}} ymm2 = <0,u,u,u,1,u,u,u>
303 ; AVX2-NEXT: vpermps %ymm0, %ymm2, %ymm0
304 ; AVX2-NEXT: vmovaps {{.*#+}} ymm2 = <u,0,1,1,u,2,3,3>
305 ; AVX2-NEXT: vpermps %ymm1, %ymm2, %ymm1
317 ; AVX1-NEXT: vinsertf128 $1, %xmm0, %ymm2, %ymm
    [all...]
vector-trunc.ll 55 ; AVX2-NEXT: vmovdqa {{.*#+}} ymm2 = <0,2,4,6,u,u,u,u>
56 ; AVX2-NEXT: vpermd %ymm0, %ymm2, %ymm0
57 ; AVX2-NEXT: vpermd %ymm1, %ymm2, %ymm1
138 ; AVX2-NEXT: vmovdqa {{.*#+}} ymm2 = <0,2,4,6,u,u,u,u>
139 ; AVX2-NEXT: vpermd %ymm0, %ymm2, %ymm0
140 ; AVX2-NEXT: vpermd %ymm1, %ymm2, %ymm1
190 ; AVX2-NEXT: vmovdqa {{.*#+}} ymm2 = <0,2,4,6,u,u,u,u>
191 ; AVX2-NEXT: vpermd %ymm0, %ymm2, %ymm0
192 ; AVX2-NEXT: vpermd %ymm1, %ymm2, %ymm1
356 ; AVX2-NEXT: vmovdqa {{.*#+}} ymm2 = [0,1,4,5,8,9,12,13,128,128,128,128,128,128,128,128,0,1,4,5,8,9,12,13,128,128,128,128,128,128,128 (…)
    [all...]
vec_minmax_sint.ll 159 ; AVX1-NEXT: vinsertf128 $1, %xmm2, %ymm3, %ymm2
160 ; AVX1-NEXT: vblendvpd %ymm2, %ymm0, %ymm1, %ymm0
165 ; AVX2-NEXT: vpcmpgtq %ymm1, %ymm0, %ymm2
166 ; AVX2-NEXT: vblendvpd %ymm2, %ymm0, %ymm1, %ymm0
171 ; AVX512-NEXT: vpcmpgtq %ymm1, %ymm0, %ymm2
172 ; AVX512-NEXT: vblendvpd %ymm2, %ymm0, %ymm1, %ymm0
563 ; AVX1-NEXT: vinsertf128 $1, %xmm2, %ymm3, %ymm2
564 ; AVX1-NEXT: vblendvpd %ymm2, %ymm0, %ymm1, %ymm0
569 ; AVX2-NEXT: vpcmpgtq %ymm0, %ymm1, %ymm2
571 ; AVX2-NEXT: vpxor %ymm3, %ymm2, %ymm
    [all...]
avx512-vbroadcast.ll 142 ; ALL-NEXT: vpxor %ymm2, %ymm2, %ymm2
167 ; ALL-NEXT: vpxor %ymm2, %ymm2, %ymm2
pmul.ll 372 ; AVX2-NEXT: vpmuludq %ymm1, %ymm0, %ymm2
376 ; AVX2-NEXT: vpaddq %ymm0, %ymm2, %ymm0
412 ; AVX2-NEXT: vpmuludq %ymm1, %ymm0, %ymm2
416 ; AVX2-NEXT: vpaddq %ymm3, %ymm2, %ymm2
420 ; AVX2-NEXT: vpaddq %ymm0, %ymm2, %ymm0
machine-combiner.ll 307 ; AVX-NEXT: vaddps %ymm3, %ymm2, %ymm1
322 ; AVX-NEXT: vaddpd %ymm3, %ymm2, %ymm1
337 ; AVX-NEXT: vmulps %ymm3, %ymm2, %ymm1
352 ; AVX-NEXT: vmulpd %ymm3, %ymm2, %ymm1
559 ; AVX-NEXT: vminps %ymm3, %ymm2, %ymm1
576 ; AVX-NEXT: vmaxps %ymm3, %ymm2, %ymm1
593 ; AVX-NEXT: vminpd %ymm3, %ymm2, %ymm1
610 ; AVX-NEXT: vmaxpd %ymm3, %ymm2, %ymm1
  /toolchain/binutils/binutils-2.25/gas/testsuite/gas/i386/
x86-64-avx-gather-intel.d 14 [ ]*[a-f0-9]+: c4 e2 ed 92 4c 7d 00 vgatherdpd ymm1,QWORD PTR \[rbp\+xmm7\*2\+0x0\],ymm2
15 [ ]*[a-f0-9]+: c4 e2 ed 93 4c 7d 00 vgatherqpd ymm1,QWORD PTR \[rbp\+ymm7\*2\+0x0\],ymm2
38 [ ]*[a-f0-9]+: c4 e2 6d 92 4c 7d 00 vgatherdps ymm1,DWORD PTR \[rbp\+ymm7\*2\+0x0\],ymm2
62 [ ]*[a-f0-9]+: c4 e2 6d 90 4c 7d 00 vpgatherdd ymm1,DWORD PTR \[rbp\+ymm7\*2\+0x0\],ymm2
86 [ ]*[a-f0-9]+: c4 e2 ed 90 4c 7d 00 vpgatherdq ymm1,QWORD PTR \[rbp\+xmm7\*2\+0x0\],ymm2
87 [ ]*[a-f0-9]+: c4 e2 ed 91 4c 7d 00 vpgatherqq ymm1,QWORD PTR \[rbp\+ymm7\*2\+0x0\],ymm2
110 [ ]*[a-f0-9]+: c4 e2 ed 92 4c 7d 00 vgatherdpd ymm1,QWORD PTR \[rbp\+xmm7\*2\+0x0\],ymm2
111 [ ]*[a-f0-9]+: c4 e2 ed 93 4c 7d 00 vgatherqpd ymm1,QWORD PTR \[rbp\+ymm7\*2\+0x0\],ymm2
134 [ ]*[a-f0-9]+: c4 e2 6d 92 4c 7d 00 vgatherdps ymm1,DWORD PTR \[rbp\+ymm7\*2\+0x0\],ymm2
158 [ ]*[a-f0-9]+: c4 e2 6d 90 4c 7d 00 vpgatherdd ymm1,DWORD PTR \[rbp\+ymm7\*2\+0x0\],ymm2
    [all...]
x86-64-avx-gather.d 13 [ ]*[a-f0-9]+: c4 e2 ed 92 4c 7d 00 vgatherdpd %ymm2,0x0\(%rbp,%xmm7,2\),%ymm1
14 [ ]*[a-f0-9]+: c4 e2 ed 93 4c 7d 00 vgatherqpd %ymm2,0x0\(%rbp,%ymm7,2\),%ymm1
37 [ ]*[a-f0-9]+: c4 e2 6d 92 4c 7d 00 vgatherdps %ymm2,0x0\(%rbp,%ymm7,2\),%ymm1
61 [ ]*[a-f0-9]+: c4 e2 6d 90 4c 7d 00 vpgatherdd %ymm2,0x0\(%rbp,%ymm7,2\),%ymm1
85 [ ]*[a-f0-9]+: c4 e2 ed 90 4c 7d 00 vpgatherdq %ymm2,0x0\(%rbp,%xmm7,2\),%ymm1
86 [ ]*[a-f0-9]+: c4 e2 ed 91 4c 7d 00 vpgatherqq %ymm2,0x0\(%rbp,%ymm7,2\),%ymm1
109 [ ]*[a-f0-9]+: c4 e2 ed 92 4c 7d 00 vgatherdpd %ymm2,0x0\(%rbp,%xmm7,2\),%ymm1
110 [ ]*[a-f0-9]+: c4 e2 ed 93 4c 7d 00 vgatherqpd %ymm2,0x0\(%rbp,%ymm7,2\),%ymm1
133 [ ]*[a-f0-9]+: c4 e2 6d 92 4c 7d 00 vgatherdps %ymm2,0x0\(%rbp,%ymm7,2\),%ymm1
157 [ ]*[a-f0-9]+: c4 e2 6d 90 4c 7d 00 vpgatherdd %ymm2,0x0\(%rbp,%ymm7,2\),%ymm
    [all...]
x86-64-xop.s 24 # Tests for op VFRCZPD ymm2/mem256, ymm1 (at&t syntax)
25 VFRCZPD %ymm2,%ymm15
31 VFRCZPD %ymm2,%ymm0
58 # Tests for op VFRCZPS ymm2/mem256, ymm1 (at&t syntax)
59 VFRCZPS %ymm2,%ymm15
65 VFRCZPS %ymm2,%ymm0
126 # Tests for op VPCMOV ymm4, ymm3/mem256, ymm2, ymm1 (at&t syntax)
128 VPCMOV %ymm2,(%r9),%ymm0,%ymm0
134 VPCMOV %ymm2,%ymm0,%ymm0,%ymm15
135 VPCMOV %ymm2,(%r9),%ymm15,%ymm1
    [all...]
xop.s 24 # Tests for op VFRCZPD ymm2/mem256, ymm1 (at&t syntax)
58 # Tests for op VFRCZPS ymm2/mem256, ymm1 (at&t syntax)
126 # Tests for op VPCMOV ymm4, ymm3/mem256, ymm2, ymm1 (at&t syntax)
160 # Tests for op VPCMOV ymm4/mem256, ymm3, ymm2, ymm1 (at&t syntax)
331 # Testing VPERMIL2PD imm8, ymm4, ymm3/mem256, ymm2, ymm1 (at&t syntax)
332 VPERMIL2PD $0x3,%ymm6,%ymm7,%ymm1,%ymm2
335 VPERMIL2PD $0x0,%ymm5,%ymm6,%ymm0,%ymm2
337 VPERMIL2PD $0x0,%ymm7,%ymm6,%ymm2,%ymm0
340 # Testing VPERMIL2PD imm8, ymm4/mem256, ymm3, ymm2, ymm1 (at&t syntax)
344 VPERMIL2PD $0x2,(%esi,%eax,1),%ymm2,%ymm6,%ymm
    [all...]
  /external/llvm/test/MC/X86/
x86_64-xop-encoding.s 153 // CHECK: vfrczps %ymm2, %ymm4
155 vfrczps %ymm2, %ymm4
164 // CHECK: vfrczpd (%rcx,%rax), %ymm2
166 vfrczpd (%rcx,%rax), %ymm2
595 // CHECK: vpcmov %ymm1, %ymm2, %ymm3, %ymm4
597 vpcmov %ymm1, %ymm2, %ymm3, %ymm4
598 // CHECK: vpcmov (%rax), %ymm2, %ymm3, %ymm4
600 vpcmov (%rax), %ymm2, %ymm3, %ymm4
622 // CHECK: vpermil2pd $1, %ymm1, %ymm2, %ymm3, %ymm4
624 vpermil2pd $1, %ymm1, %ymm2, %ymm3, %ymm
    [all...]
  /external/llvm/test/MC/Disassembler/X86/
simple-tests.txt 182 # CHECK: vblendps $129, %ymm2, %ymm5, %ymm1
188 # CHECK: vblendpd $129, %ymm2, %ymm5, %ymm1
200 # CHECK: vdpps $129, %ymm2, %ymm5, %ymm1
794 # CHECK: vfmaddps %ymm1, %ymm2, %ymm0, %ymm0
797 # CHECK: vfmaddps %ymm2, %ymm1, %ymm0, %ymm0
812 # CHECK: vpermil2ps $0, %ymm4, %ymm4, %ymm2, %ymm2
821 # CHECK: vgatherdpd %ymm0, (%rdi,%xmm1,2), %ymm2
833 # CHECK: vpgatherdq %ymm0, (%rdi,%xmm1,2), %ymm2
905 # CHECK: vpcmov %ymm1, %ymm2, %ymm3, %ymm
    [all...]
x86-32.txt 199 # CHECK: vblendps $129, %ymm2, %ymm5, %ymm1
205 # CHECK: vblendpd $129, %ymm2, %ymm5, %ymm1
217 # CHECK: vdpps $129, %ymm2, %ymm5, %ymm1
739 # CHECK: vpermq $238, %ymm2, %ymm2
  /external/llvm/lib/Target/X86/
X86CallingConv.td 59 // 256-bit vectors are returned in YMM0 and XMM1, when they fit. YMM2 and YMM3
63 CCAssignToReg<[YMM0,YMM1,YMM2,YMM3]>>,
119 CCAssignToReg<[YMM0,YMM1,YMM2,YMM3]>>,
146 CCAssignToReg<[YMM0,YMM1,YMM2,YMM3]>>,
308 CCAssignToReg<[YMM0, YMM1, YMM2, YMM3,
412 CCAssignToReg<[YMM0, YMM1, YMM2, YMM3, YMM4, YMM5]>>,
504 CCAssignToReg<[YMM0, YMM1, YMM2]>>>>,
523 CCAssignToReg<[YMM0, YMM1, YMM2, YMM3]>>>>,
617 CCAssignToReg<[YMM0, YMM1, YMM2, YMM3, YMM4, YMM5]>>,
728 CCAssignToReg<[YMM0, YMM1, YMM2, YMM3]>>
    [all...]

Completed in 234 milliseconds

1 2 34 5 6