HomeSort by relevance Sort by last modified time
    Searched full:ymm0 (Results 26 - 50 of 95) sorted by null

12 3 4

  /external/llvm/test/CodeGen/X86/
vec_cast2.ll 13 ; CHECK-NEXT: vinsertf128 $1, %xmm1, %ymm0, %ymm0
14 ; CHECK-NEXT: vcvtdq2ps %ymm0, %ymm0
26 ; CHECK-WIDE-NEXT: vinsertf128 $1, %xmm0, %ymm1, %ymm0
27 ; CHECK-WIDE-NEXT: vcvtdq2ps %ymm0, %ymm0
57 ; CHECK-NEXT: vinsertf128 $1, %xmm0, %ymm1, %ymm0
58 ; CHECK-NEXT: vandps LCPI2_0, %ymm0, %ymm0
    [all...]
commute-fcmp.ll 107 ;AVX: vcmpeqps (%rdi), %ymm0, %ymm0
118 ;AVX: vcmpneqps (%rdi), %ymm0, %ymm0
129 ;AVX: vcmpordps (%rdi), %ymm0, %ymm0
140 ;AVX: vcmpunordps (%rdi), %ymm0, %ymm0
152 ;AVX-NEXT: vcmpltps %ymm0, %ymm1, %ymm0
    [all...]
vec_shuf-insert.ll 16 ; CHECK: vinsertf128 $1, %xmm1, %ymm2, %ymm0
27 ; CHECK: vinsertf128 $1, %xmm2, %ymm0, %ymm0
fma_patterns.ll 55 ; CHECK: vfmadd213ps %ymm2, %ymm1, %ymm0
58 ; CHECK_FMA4: vfmaddps %ymm2, %ymm1, %ymm0, %ymm0
67 ; CHECK: vfmsub213ps %ymm2, %ymm1, %ymm0
70 ; CHECK_FMA4: vfmsubps %ymm2, %ymm1, %ymm0, %ymm0
79 ; CHECK: vfnmadd213ps %ymm2, %ymm1, %ymm0
82 ; CHECK_FMA4: vfnmaddps %ymm2, %ymm1, %ymm0, %ymm0
91 ; CHECK: vfnmsub213ps %ymm2, %ymm1, %ymm0
    [all...]
avx512-build-vector.ll 7 ; CHECK-NEXT: vinserti128 $1, %xmm0, %ymm0, %ymm0
9 ; CHECK-NEXT: vpblendd {{.*#+}} ymm0 = ymm1[0,1,2,3],ymm0[4],ymm1[5,6,7]
34 ; CHECK-NEXT: vinsertf128 $1, %xmm0, %ymm2, %ymm0
2012-07-15-broadcastfold.ll 8 ;CHECK: vbroadcastss [[SPILLED]], %ymm0
avx-varargs-x86_64.ll 9 ; CHECK: vmovaps %ymm0, (%rsp)
fp-trunc.ll 52 ; AVX-NEXT: vcvtpd2psy %ymm0, %xmm0
72 ; AVX-NEXT: vcvtpd2psy %ymm0, %xmm0
74 ; AVX-NEXT: vinsertf128 $1, %xmm1, %ymm0, %ymm0
pr22774.ll 8 ; CHECK: vmovdqa in(%rip), %ymm0
avx-intrinsics-x86-upgrade.ll 7 ; CHECK: vinsertf128 $1, %xmm1, %ymm0, %ymm0
15 ; CHECK: vinsertf128 $1, %xmm1, %ymm0, %ymm0
23 ; CHECK: vinsertf128 $1, %xmm1, %ymm0, %ymm0
33 ; CHECK-NOT: vinsertf128 $1, %xmm1, %ymm0, %ymm0
43 ; CHECK: vextractf128 $1, %ymm0, %xmm0
51 ; CHECK: vextractf128 $1, %ymm0, %xmm
    [all...]
avx-vextractf128.ll 22 ; CHECK-NOT: vextractf128 $1, %ymm0, %xmm0
24 ; CHECK: vextractf128 $1, %ymm0, (%rdi)
36 ; CHECK-NOT: vextractf128 $1, %ymm0, %xmm0
38 ; CHECK: vextractf128 $1, %ymm0, (%rdi)
50 ; CHECK-NOT: vextractf128 $1, %ymm0, %xmm0
52 ; CHECK: vextractf128 $1, %ymm0, (%rdi)
psubus.ll 237 ; AVX2-NEXT: vmovdqu (%rdi), %ymm0
238 ; AVX2-NEXT: vpsubusw LCPI6_0(%rip), %ymm0, %ymm0
239 ; AVX2-NEXT: vmovdqu %ymm0, (%rdi)
257 ; AVX2-NEXT: vmovdqu (%rdi), %ymm0
258 ; AVX2-NEXT: vpsubusw LCPI7_0(%rip), %ymm0, %ymm0
259 ; AVX2-NEXT: vmovdqu %ymm0, (%rdi)
280 ; AVX2-NEXT: vpbroadcastw %xmm0, %ymm0
282 ; AVX2-NEXT: vpsubusw %ymm0, %ymm1, %ymm
    [all...]
vector-zext.ll 41 ; AVX1-NEXT: vinsertf128 $1, %xmm1, %ymm0, %ymm0
46 ; AVX2-NEXT: vpmovzxwd {{.*#+}} ymm0 = xmm0[0],zero,xmm0[1],zero,xmm0[2],zero,xmm0[3],zero,xmm0[4],zero,xmm0[5],zero,xmm0[6],zero,xmm0[7],zero
89 ; AVX1-NEXT: vinsertf128 $1, %xmm1, %ymm0, %ymm0
94 ; AVX2-NEXT: vpmovzxdq {{.*#+}} ymm0 = xmm0[0],zero,xmm0[1],zero,xmm0[2],zero,xmm0[3],zero
138 ; AVX1-NEXT: vinsertf128 $1, %xmm0, %ymm1, %ymm0
139 ; AVX1-NEXT: vandps {{.*}}(%rip), %ymm0, %ymm0
144 ; AVX2-NEXT: vpmovzxwd {{.*#+}} ymm0 = xmm0[0],zero,xmm0[1],zero,xmm0[2],zero,xmm0[3],zero,xmm0[4],zero,xmm0[5],zero,xmm0[6],zero,xmm0[ (…)
    [all...]
avx512bwvl-intrinsics.ll 7 ; CHECK: vpcmpeqb %ymm1, %ymm0, %k0 ##
14 ; CHECK: vpcmpeqb %ymm1, %ymm0, %k0 {%k1} ##
23 ; CHECK: vpcmpeqw %ymm1, %ymm0, %k0 ##
30 ; CHECK: vpcmpeqw %ymm1, %ymm0, %k0 {%k1} ##
39 ; CHECK: vpcmpgtb %ymm1, %ymm0, %k0 ##
46 ; CHECK: vpcmpgtb %ymm1, %ymm0, %k0 {%k1} ##
55 ; CHECK: vpcmpgtw %ymm1, %ymm0, %k0 ##
62 ; CHECK: vpcmpgtw %ymm1, %ymm0, %k0 {%k1} ##
71 ; CHECK: vpcmpeqb %ymm1, %ymm0, %k0 ##
74 ; CHECK: vpcmpltb %ymm1, %ymm0, %k0 #
    [all...]
sqrt-fastmath.ll 126 ; ESTIMATE-NEXT: vrsqrtps %ymm0, %ymm1
128 ; ESTIMATE-NEXT: vmulps %ymm0, %ymm2, %ymm0
129 ; ESTIMATE-NEXT: vaddps {{.*}}(%rip), %ymm0, %ymm0
131 ; ESTIMATE-NEXT: vmulps %ymm1, %ymm0, %ymm0
vector-blend.ll 317 ; AVX-NEXT: vblendps {{.*#+}} ymm0 = ymm0[0],ymm1[1,2,3],ymm0[4],ymm1[5,6,7]
349 ; AVX1-NEXT: vblendps {{.*#+}} ymm0 = ymm0[0],ymm1[1,2,3],ymm0[4],ymm1[5,6,7]
354 ; AVX2-NEXT: vpblendd {{.*#+}} ymm0 = ymm0[0],ymm1[1,2,3],ymm0[4],ymm1[5,6,7]
392 ; AVX-NEXT: vblendpd {{.*#+}} ymm0 = ymm0[0],ymm2[1,2,3
    [all...]
avx512vl-intrinsics.ll 7 ; CHECK: vpcmpeqd %ymm1, %ymm0, %k0 ##
14 ; CHECK: vpcmpeqd %ymm1, %ymm0, %k0 {%k1} ##
23 ; CHECK: vpcmpeqq %ymm1, %ymm0, %k0 ##
30 ; CHECK: vpcmpeqq %ymm1, %ymm0, %k0 {%k1} ##
39 ; CHECK: vpcmpgtd %ymm1, %ymm0, %k0 ##
46 ; CHECK: vpcmpgtd %ymm1, %ymm0, %k0 {%k1} ##
55 ; CHECK: vpcmpgtq %ymm1, %ymm0, %k0 ##
62 ; CHECK: vpcmpgtq %ymm1, %ymm0, %k0 {%k1} ##
71 ; CHECK: vpcmpeqd %ymm1, %ymm0, %k0 ##
74 ; CHECK: vpcmpltd %ymm1, %ymm0, %k0 #
    [all...]
2012-1-10-buildvector.ll 21 ; CHECK-NEXT: vmovaps %ymm0, (%eax)
avx-intrinsics-x86.ll     [all...]
vector-idiv.ll 115 ; AVX-NEXT: vpshufd {{.*#+}} ymm3 = ymm0[1,1,3,3,5,5,7,7]
117 ; AVX-NEXT: vpmuludq %ymm1, %ymm0, %ymm1
120 ; AVX-NEXT: vpsubd %ymm1, %ymm0, %ymm0
121 ; AVX-NEXT: vpsrld $1, %ymm0, %ymm0
122 ; AVX-NEXT: vpaddd %ymm1, %ymm0, %ymm0
123 ; AVX-NEXT: vpsrld $2, %ymm0, %ymm0
    [all...]
  /external/libvpx/libvpx/third_party/libyuv/source/
row_win.cc 872 vmovdqu ymm0, [eax] local
876 vpmaddubsw ymm0, ymm0, ymm4 local
881 vphaddw ymm0, ymm0, ymm1 // mutates. local
883 vpsrlw ymm0, ymm0, 7 local
885 vpackuswb ymm0, ymm0, ymm2 // mutates. local
886 vpermd ymm0, ymm6, ymm0 // For vphaddw + vpackuswb mutation local
887 vpaddb ymm0, ymm0, ymm5 local
912 vmovdqu ymm0, [eax] local
916 vpmaddubsw ymm0, ymm0, ymm4 local
921 vphaddw ymm0, ymm0, ymm1 \/\/ mutates. local
923 vpaddw ymm0, ymm0, ymm5 \/\/ Add .5 for rounding. local
925 vpsrlw ymm0, ymm0, 7 local
927 vpackuswb ymm0, ymm0, ymm2 \/\/ mutates. local
928 vpermd ymm0, ymm6, ymm0 \/\/ For vphaddw + vpackuswb mutation. local
1366 vmovdqu ymm0, [eax] local
1370 vpavgb ymm0, ymm0, [eax + esi] local
1375 vshufps ymm4, ymm0, ymm1, 0x88 local
1376 vshufps ymm0, ymm0, ymm1, 0xdd local
1377 vpavgb ymm0, ymm0, ymm4 \/\/ mutated by vshufps local
1385 vpmaddubsw ymm1, ymm0, ymm7 \/\/ U local
1387 vpmaddubsw ymm0, ymm0, ymm6 \/\/ V local
1390 vphaddw ymm0, ymm0, ymm2 local
1392 vpsraw ymm0, ymm0, 8 local
1393 vpacksswb ymm0, ymm1, ymm0 \/\/ mutates local
1394 vpermq ymm0, ymm0, 0xd8 \/\/ For vpacksswb local
1395 vpshufb ymm0, ymm0, kShufARGBToUV_AVX \/\/ For vshufps + vphaddw local
1396 vpaddb ymm0, ymm0, ymm5 \/\/ -> unsigned local
1400 vextractf128 [edx], ymm0, 0 \/\/ U local
1401 vextractf128 [edx + edi], ymm0, 1 \/\/ V local
2252 vpunpcklbw ymm0, ymm0, ymm1 \/\/ UV local
2253 vpermq ymm0, ymm0, 0xd8 local
2254 vpunpcklwd ymm0, ymm0, ymm0 \/\/ UVUV local
2255 vpmaddubsw ymm2, ymm0, kUVToB_AVX \/\/ scale B UV local
2256 vpmaddubsw ymm1, ymm0, kUVToG_AVX \/\/ scale G UV local
2257 vpmaddubsw ymm0, ymm0, kUVToR_AVX \/\/ scale R UV local
2260 vpsubw ymm0, ymm0, kUVBiasR_AVX local
2271 vpaddsw ymm0, ymm0, ymm3 \/\/ R += Y local
2274 vpsraw ymm0, ymm0, 6 local
2277 vpackuswb ymm0, ymm0, ymm0 \/\/ R local
2282 vpunpcklbw ymm0, ymm0, ymm5 \/\/ RA local
2283 vpermq ymm0, ymm0, 0xd8 local
3321 vmovdqu ymm0, [eax + ecx] local
3322 vpshufb ymm0, ymm0, ymm5 local
3323 vpermq ymm0, ymm0, 0x4e \/\/ swap high and low halfs local
3446 vpermd ymm0, ymm5, [eax + ecx * 4] \/\/ permute dword order local
3547 vmovdqu ymm0, [eax] local
3550 vpsrlw ymm2, ymm0, 8 \/\/ odd bytes local
3552 vpand ymm0, ymm0, ymm5 \/\/ even bytes local
3554 vpackuswb ymm0, ymm0, ymm1 local
3556 vpermq ymm0, ymm0, 0xd8 local
3647 vmovdqu ymm0, [eax] \/\/ read 32 U's local
3650 vpunpcklbw ymm2, ymm0, ymm1 \/\/ low 16 UV pairs. mutated qqword 0,2 local
3651 vpunpckhbw ymm0, ymm0, ymm1 \/\/ high 16 UV pairs. mutated qqword 1,3 local
3652 vperm2i128 ymm1, ymm2, ymm0, 0x20 \/\/ low 128 of ymm2 and low 128 of ymm0 local
3653 vperm2i128 ymm2, ymm2, ymm0, 0x31 \/\/ high 128 of ymm2 and high 128 of ymm0 local
3770 vpcmpeqb ymm0, ymm0, ymm0 local
3771 vpsrld ymm0, ymm0, 8 \/\/ generate mask 0x00ffffff local
3839 vpcmpeqb ymm0, ymm0, ymm0 local
3840 vpsrld ymm0, ymm0, 8 \/\/ generate mask 0x00ffffff local
3924 vmovdqu ymm0, [eax] local
3927 vpand ymm0, ymm0, ymm5 \/\/ even bytes are Y local
3929 vpackuswb ymm0, ymm0, ymm1 \/\/ mutates. local
3930 vpermq ymm0, ymm0, 0xd8 local
3957 vmovdqu ymm0, [eax] local
3959 vpavgb ymm0, ymm0, [eax + esi] local
3962 vpsrlw ymm0, ymm0, 8 \/\/ YUYV -> UVUV local
3964 vpackuswb ymm0, ymm0, ymm1 \/\/ mutates. local
3965 vpermq ymm0, ymm0, 0xd8 local
3966 vpand ymm1, ymm0, ymm5 \/\/ U local
3967 vpsrlw ymm0, ymm0, 8 \/\/ V local
3969 vpackuswb ymm0, ymm0, ymm0 \/\/ mutates. local
3971 vpermq ymm0, ymm0, 0xd8 local
3973 vextractf128 [edx + edi], ymm0, 0 \/\/ V local
4000 vmovdqu ymm0, [eax] local
4003 vpsrlw ymm0, ymm0, 8 \/\/ YUYV -> UVUV local
4005 vpackuswb ymm0, ymm0, ymm1 \/\/ mutates. local
4006 vpermq ymm0, ymm0, 0xd8 local
4007 vpand ymm1, ymm0, ymm5 \/\/ U local
4008 vpsrlw ymm0, ymm0, 8 \/\/ V local
4010 vpackuswb ymm0, ymm0, ymm0 \/\/ mutates. local
4012 vpermq ymm0, ymm0, 0xd8 local
4014 vextractf128 [edx + edi], ymm0, 0 \/\/ V local
4035 vmovdqu ymm0, [eax] local
4038 vpsrlw ymm0, ymm0, 8 \/\/ odd bytes are Y local
4040 vpackuswb ymm0, ymm0, ymm1 \/\/ mutates. local
4041 vpermq ymm0, ymm0, 0xd8 local
4068 vmovdqu ymm0, [eax] local
4070 vpavgb ymm0, ymm0, [eax + esi] local
4073 vpand ymm0, ymm0, ymm5 \/\/ UYVY -> UVUV local
4075 vpackuswb ymm0, ymm0, ymm1 \/\/ mutates. local
4076 vpermq ymm0, ymm0, 0xd8 local
4077 vpand ymm1, ymm0, ymm5 \/\/ U local
4078 vpsrlw ymm0, ymm0, 8 \/\/ V local
4080 vpackuswb ymm0, ymm0, ymm0 \/\/ mutates. local
4082 vpermq ymm0, ymm0, 0xd8 local
4084 vextractf128 [edx + edi], ymm0, 0 \/\/ V local
4111 vmovdqu ymm0, [eax] local
4114 vpand ymm0, ymm0, ymm5 \/\/ UYVY -> UVUV local
4116 vpackuswb ymm0, ymm0, ymm1 \/\/ mutates. local
4117 vpermq ymm0, ymm0, 0xd8 local
4118 vpand ymm1, ymm0, ymm5 \/\/ U local
4119 vpsrlw ymm0, ymm0, 8 \/\/ V local
4121 vpackuswb ymm0, ymm0, ymm0 \/\/ mutates. local
4123 vpermq ymm0, ymm0, 0xd8 local
4125 vextractf128 [edx + edi], ymm0, 0 \/\/ V local
4959 vpunpcklbw ymm0, ymm6, ymm6 \/\/ low 4 pixels. mutated. local
4961 vpshufb ymm2, ymm0, ymm4 \/\/ low 4 alphas local
4963 vpmulhuw ymm0, ymm0, ymm2 \/\/ rgb * a local
4966 vpsrlw ymm0, ymm0, 8 local
4968 vpackuswb ymm0, ymm0, ymm1 \/\/ unmutated. local
4969 vpor ymm0, ymm0, ymm6 \/\/ copy original alpha local
5055 vpunpcklbw ymm0, ymm6, ymm6 \/\/ low 4 pixels. mutated. local
5062 vpmulhuw ymm0, ymm0, ymm2 \/\/ rgb * ia local
5064 vpackuswb ymm0, ymm0, ymm1 \/\/ unmutated. local
5118 vpunpcklbw ymm0, ymm6, ymm6 \/\/ low 4 pixels. mutated. local
5124 vpmulhuw ymm0, ymm0, ymm2 \/\/ rgb * ia local
5126 vpackuswb ymm0, ymm0, ymm1 \/\/ unmutated. local
5542 vpunpcklbw ymm0, ymm1, ymm1 \/\/ low 4 local
5546 vpmulhuw ymm0, ymm0, ymm2 \/\/ src_argb0 * src_argb1 low 4 local
5548 vpackuswb ymm0, ymm0, ymm1 local
5575 vmovdqu ymm0, [eax] \/\/ read 8 pixels from src_argb0 local
5577 vpaddusb ymm0, ymm0, [esi] \/\/ add 8 pixels from src_argb1 local
5605 vmovdqu ymm0, [eax] \/\/ read 8 pixels from src_argb0 local
5607 vpsubusb ymm0, ymm0, [esi] \/\/ src_argb0 - src_argb1 local
6230 vpxor ymm0, ymm0, ymm0 local
6231 vpermd ymm5, ymm0, ymm5 local
6235 vmovdqu ymm0, [esi] local
6237 vpunpckhbw ymm1, ymm0, ymm2 \/\/ mutates local
6238 vpunpcklbw ymm0, ymm0, ymm2 \/\/ mutates local
6239 vpmaddubsw ymm0, ymm0, ymm5 local
6241 vpsrlw ymm0, ymm0, 7 local
6243 vpackuswb ymm0, ymm0, ymm1 \/\/ unmutates local
6253 vmovdqu ymm0, [esi] local
6254 vpavgb ymm0, ymm0, [esi + edx] local
6255 vpavgb ymm0, ymm0, [esi + edx] local
6265 vmovdqu ymm0, [esi] local
6266 vpavgb ymm0, ymm0, [esi + edx] local
6276 vmovdqu ymm0, [esi + edx] local
6277 vpavgb ymm0, ymm0, [esi] local
6278 vpavgb ymm0, ymm0, [esi] local
6775 vmovdqu ymm0, [eax] local
6776 vpavgb ymm0, ymm0, [eax + edx] local
6912 vmovdqu ymm0, [eax] local
6915 vpshufb ymm0, ymm0, ymm5 local
7209 vpmovzxbd ymm0, qword ptr [eax] \/\/ 2 BGRA pixels local
7211 vcvtdq2ps ymm0, ymm0 \/\/ X 8 floats local
7212 vmulps ymm2, ymm0, ymm0 \/\/ X * X local
7213 vmulps ymm3, ymm0, ymm7 \/\/ C3 * X local
7214 vfmadd132ps ymm0, ymm4, ymm5 \/\/ result = C0 + C1 * X local
7215 vfmadd231ps ymm0, ymm2, ymm6 \/\/ result += C2 * X * X local
7216 vfmadd231ps ymm0, ymm2, ymm3 \/\/ result += C3 * X * X * X local
7217 vcvttps2dq ymm0, ymm0 local
7218 vpackusdw ymm0, ymm0, ymm0 \/\/ b0g0r0a0_00000000_b0g0r0a0_00000000 local
7219 vpermq ymm0, ymm0, 0xd8 \/\/ b0g0r0a0_b0g0r0a0_00000000_00000000 local
    [all...]
row_x86.asm 124 vperm2i128 m1, m2, m0, 0x20 // low 128 of ymm2 and low 128 of ymm0
125 vperm2i128 m2, m2, m0, 0x31 // high 128 of ymm2 and high 128 of ymm0
  /external/llvm/test/MC/Disassembler/X86/
x86-32.txt 138 # CHECK: vaddpd %ymm5, %ymm1, %ymm0
141 # CHECK: vaddps %ymm3, %ymm1, %ymm0
144 # CHECK: vandpd %ymm5, %ymm1, %ymm0
147 # CHECK: vandps %ymm3, %ymm1, %ymm0
153 # CHECK: vcvtps2pd %xmm0, %ymm0
289 # CHECK: vmovups %ymm1, %ymm0
292 # CHECK: vmovups %ymm0, %ymm1
295 # CHECK: vmovaps %ymm1, %ymm0
298 # CHECK: vmovaps %ymm0, %ymm1
319 # CHECK: vmovdqa %ymm1, %ymm0
    [all...]
simple-tests.txt 117 # CHECK: vaddpd %ymm13, %ymm1, %ymm0
120 # CHECK: vaddps %ymm3, %ymm1, %ymm0
123 # CHECK: vandpd %ymm13, %ymm1, %ymm0
126 # CHECK: vandps %ymm3, %ymm1, %ymm0
132 # CHECK: vcvtps2pd %xmm0, %ymm0
275 # CHECK: vmovups %ymm1, %ymm0
278 # CHECK: vmovups %ymm0, %ymm1
281 # CHECK: vmovups %ymm0, %ymm1
284 # CHECK: vmovaps %ymm1, %ymm0
287 # CHECK: vmovaps %ymm0, %ymm
    [all...]
  /external/llvm/include/llvm/CodeGen/
LivePhysRegs.h 22 // %YMM0<def> = ...
23 // %XMM0<def> = ... (Kills %XMM0, all %XMM0s sub-registers, and %YMM0)
25 // %YMM0<def> = ...
26 // %XMM0<def> = ..., %YMM0<imp-use> (%YMM0 and all its sub-registers are alive)

Completed in 1319 milliseconds

12 3 4