HomeSort by relevance Sort by last modified time
    Searched full:vpshufd (Results 1 - 25 of 165) sorted by null

1 2 3 4 5 6 7

  /external/swiftshader/third_party/LLVM/test/CodeGen/X86/
avx-shuffle.ll 9 ; CHECK: vpshufd
avx-basic.ll 85 ; CHECK: vpshufd $-96
86 ; CHECK: vpshufd $-6
  /external/llvm/test/CodeGen/X86/
vector-idiv.ll 64 ; AVX1-NEXT: vpshufd {{.*#+}} xmm2 = xmm1[1,1,3,3]
65 ; AVX1-NEXT: vpshufd {{.*#+}} xmm3 = xmm0[1,1,3,3]
68 ; AVX1-NEXT: vpshufd {{.*#+}} xmm0 = xmm0[1,1,3,3]
77 ; AVX2-NEXT: vpshufd {{.*#+}} xmm2 = xmm1[1,1,3,3]
78 ; AVX2-NEXT: vpshufd {{.*#+}} xmm3 = xmm0[1,1,3,3]
81 ; AVX2-NEXT: vpshufd {{.*#+}} xmm0 = xmm0[1,1,3,3]
2012-01-12-extract-sv.ll 8 ; CHECK-NEXT: vpshufd {{.*#+}} xmm0 = xmm0[0,0,1,1]
9 ; CHECK-NEXT: vpshufd {{.*#+}} xmm1 = xmm0[0,1,0,1]
avx-trunc.ll 8 ; CHECK-NEXT: vpshufd {{.*#+}} xmm1 = xmm1[0,1,0,2]
9 ; CHECK-NEXT: vpshufd {{.*#+}} xmm0 = xmm0[0,2,2,3]
swizzle-avx2.ll 30 ; CHECK: vpshufd $78
32 ; CHECK-NOT: vpshufd
42 ; CHECK: vpshufd $78
44 ; CHECK-NOT: vpshufd
vector-idiv-sdiv-256.ll 92 ; AVX1-NEXT: vpshufd {{.*#+}} xmm3 = xmm2[1,1,3,3]
94 ; AVX1-NEXT: vpshufd {{.*#+}} xmm5 = xmm4[1,1,3,3]
97 ; AVX1-NEXT: vpshufd {{.*#+}} xmm2 = xmm2[1,1,3,3]
103 ; AVX1-NEXT: vpshufd {{.*#+}} xmm3 = xmm1[1,1,3,3]
104 ; AVX1-NEXT: vpshufd {{.*#+}} xmm4 = xmm0[1,1,3,3]
107 ; AVX1-NEXT: vpshufd {{.*#+}} xmm1 = xmm1[1,1,3,3]
119 ; AVX2-NEXT: vpshufd {{.*#+}} ymm2 = ymm1[1,1,3,3,5,5,7,7]
120 ; AVX2-NEXT: vpshufd {{.*#+}} ymm3 = ymm0[1,1,3,3,5,5,7,7]
123 ; AVX2-NEXT: vpshufd {{.*#+}} ymm1 = ymm1[1,1,3,3,5,5,7,7]
169 ; AVX1-NEXT: vpshufd {{.*#+}} xmm4 = xmm1[2,3,0,1
    [all...]
vector-idiv-udiv-256.ll 99 ; AVX1-NEXT: vpshufd {{.*#+}} xmm2 = xmm1[1,1,3,3]
100 ; AVX1-NEXT: vpshufd {{.*#+}} xmm3 = xmm0[1,1,3,3]
103 ; AVX1-NEXT: vpshufd {{.*#+}} xmm3 = xmm3[1,1,3,3]
110 ; AVX1-NEXT: vpshufd {{.*#+}} xmm3 = xmm1[1,1,3,3]
112 ; AVX1-NEXT: vpshufd {{.*#+}} xmm4 = xmm0[1,1,3,3]
115 ; AVX1-NEXT: vpshufd {{.*#+}} xmm1 = xmm1[1,1,3,3]
127 ; AVX2-NEXT: vpshufd {{.*#+}} ymm2 = ymm1[1,1,3,3,5,5,7,7]
128 ; AVX2-NEXT: vpshufd {{.*#+}} ymm3 = ymm0[1,1,3,3,5,5,7,7]
131 ; AVX2-NEXT: vpshufd {{.*#+}} ymm1 = ymm1[1,1,3,3,5,5,7,7]
180 ; AVX1-NEXT: vpshufd {{.*#+}} xmm4 = xmm1[2,3,0,1
    [all...]
reduce-trunc-shl.ll 17 ; AVX2-NEXT: vpshufd {{.*#+}} ymm0 = mem[0,2,0,2,4,6,4,6]
vector-trunc.ll 43 ; AVX1-NEXT: vpshufd {{.*#+}} xmm2 = xmm2[0,1,0,2]
44 ; AVX1-NEXT: vpshufd {{.*#+}} xmm1 = xmm1[0,2,2,3]
47 ; AVX1-NEXT: vpshufd {{.*#+}} xmm2 = xmm2[0,1,0,2]
48 ; AVX1-NEXT: vpshufd {{.*#+}} xmm0 = xmm0[0,2,2,3]
55 ; AVX2-NEXT: vpshufd {{.*#+}} ymm0 = ymm0[0,2,0,2,4,6,4,6]
57 ; AVX2-NEXT: vpshufd {{.*#+}} ymm1 = ymm1[0,2,0,2,4,6,4,6]
139 ; AVX2-NEXT: vpshufd {{.*#+}} ymm0 = ymm0[0,2,0,2,4,6,4,6]
141 ; AVX2-NEXT: vpshufd {{.*#+}} ymm1 = ymm1[0,2,0,2,4,6,4,6]
193 ; AVX2-NEXT: vpshufd {{.*#+}} ymm0 = ymm0[0,2,0,2,4,6,4,6]
195 ; AVX2-NEXT: vpshufd {{.*#+}} ymm1 = ymm1[0,2,0,2,4,6,4,6
    [all...]
viabs.ll 453 ; AVX1-NEXT: vpshufd {{.*#+}} xmm3 = xmm3[1,1,3,3]
456 ; AVX1-NEXT: vpshufd {{.*#+}} xmm4 = xmm4[1,1,3,3]
463 ; AVX1-NEXT: vpshufd {{.*#+}} xmm3 = xmm3[1,1,3,3]
466 ; AVX1-NEXT: vpshufd {{.*#+}} xmm4 = xmm4[1,1,3,3]
476 ; AVX2-NEXT: vpshufd {{.*#+}} ymm2 = ymm2[1,1,3,3,5,5,7,7]
480 ; AVX2-NEXT: vpshufd {{.*#+}} ymm2 = ymm2[1,1,3,3,5,5,7,7]
530 ; AVX1-NEXT: vpshufd {{.*#+}} xmm3 = xmm3[1,1,3,3]
533 ; AVX1-NEXT: vpshufd {{.*#+}} xmm4 = xmm4[1,1,3,3]
540 ; AVX1-NEXT: vpshufd {{.*#+}} xmm3 = xmm3[1,1,3,3]
543 ; AVX1-NEXT: vpshufd {{.*#+}} xmm4 = xmm4[1,1,3,3
    [all...]
vselect-avx.ll 47 ; CHECK-NEXT: vpshufd {{.*#+}} xmm0 = xmm0[2,3,0,1]
78 ; CHECK-NEXT: vpshufd {{.*#+}} xmm4 = xmm3[1,1,3,3]
79 ; CHECK-NEXT: vpshufd {{.*#+}} xmm5 = xmm0[1,1,3,3]
82 ; CHECK-NEXT: vpshufd {{.*#+}} xmm3 = xmm3[1,1,3,3]
vector-idiv-sdiv-128.ll 124 ; AVX1-NEXT: vpshufd {{.*#+}} xmm2 = xmm1[1,1,3,3]
125 ; AVX1-NEXT: vpshufd {{.*#+}} xmm3 = xmm0[1,1,3,3]
128 ; AVX1-NEXT: vpshufd {{.*#+}} xmm1 = xmm1[1,1,3,3]
139 ; AVX2-NEXT: vpshufd {{.*#+}} xmm2 = xmm1[1,1,3,3]
140 ; AVX2-NEXT: vpshufd {{.*#+}} xmm3 = xmm0[1,1,3,3]
143 ; AVX2-NEXT: vpshufd {{.*#+}} xmm1 = xmm1[1,1,3,3]
234 ; AVX1-NEXT: vpshufd {{.*#+}} xmm3 = xmm0[2,3,0,1]
422 ; AVX1-NEXT: vpshufd {{.*#+}} xmm2 = xmm1[1,1,3,3]
423 ; AVX1-NEXT: vpshufd {{.*#+}} xmm3 = xmm0[1,1,3,3]
426 ; AVX1-NEXT: vpshufd {{.*#+}} xmm1 = xmm1[1,1,3,3
    [all...]
vector-idiv-udiv-128.ll 119 ; AVX1-NEXT: vpshufd {{.*#+}} xmm2 = xmm1[1,1,3,3]
120 ; AVX1-NEXT: vpshufd {{.*#+}} xmm3 = xmm0[1,1,3,3]
123 ; AVX1-NEXT: vpshufd {{.*#+}} xmm1 = xmm1[1,1,3,3]
134 ; AVX2-NEXT: vpshufd {{.*#+}} xmm2 = xmm1[1,1,3,3]
135 ; AVX2-NEXT: vpshufd {{.*#+}} xmm3 = xmm0[1,1,3,3]
138 ; AVX2-NEXT: vpshufd {{.*#+}} xmm1 = xmm1[1,1,3,3]
221 ; AVX1-NEXT: vpshufd {{.*#+}} xmm3 = xmm0[2,3,0,1]
402 ; AVX1-NEXT: vpshufd {{.*#+}} xmm2 = xmm1[1,1,3,3]
403 ; AVX1-NEXT: vpshufd {{.*#+}} xmm3 = xmm0[1,1,3,3]
406 ; AVX1-NEXT: vpshufd {{.*#+}} xmm1 = xmm1[1,1,3,3
    [all...]
vector-shuffle-combining.ll 101 ; AVX1-NEXT: vpshufd {{.*#+}} xmm0 = xmm0[0,0,0,0]
179 ; AVX-NEXT: vpshufd {{.*#+}} xmm0 = xmm0[0,2,1,3]
197 ; AVX-NEXT: vpshufd {{.*#+}} xmm0 = xmm0[0,2,1,3]
215 ; AVX-NEXT: vpshufd {{.*#+}} xmm0 = xmm0[0,2,1,3]
233 ; AVX-NEXT: vpshufd {{.*#+}} xmm0 = xmm0[0,2,1,3]
251 ; AVX-NEXT: vpshufd {{.*#+}} xmm0 = xmm0[0,2,1,3]
269 ; AVX-NEXT: vpshufd {{.*#+}} xmm0 = xmm0[0,2,1,3]
547 ; AVX1-NEXT: vpshufd {{.*#+}} xmm0 = xmm0[0,2,1,3]
554 ; AVX2-NEXT: vpshufd {{.*#+}} xmm0 = xmm0[0,2,1,3]
590 ; AVX1-NEXT: vpshufd {{.*#+}} xmm0 = xmm0[0,2,1,3
    [all...]
vector-trunc-math.ll 28 ; AVX1-NEXT: vpshufd {{.*#+}} xmm0 = xmm0[0,1,0,2]
29 ; AVX1-NEXT: vpshufd {{.*#+}} xmm1 = xmm2[0,2,2,3]
37 ; AVX2-NEXT: vpshufd {{.*#+}} ymm0 = ymm0[0,2,0,2,4,6,4,6]
103 ; AVX2-NEXT: vpshufd {{.*#+}} ymm0 = ymm0[0,2,0,2,4,6,4,6]
105 ; AVX2-NEXT: vpshufd {{.*#+}} ymm1 = ymm1[0,2,0,2,4,6,4,6]
241 ; AVX2-NEXT: vpshufd {{.*#+}} ymm2 = ymm2[0,2,0,2,4,6,4,6]
243 ; AVX2-NEXT: vpshufd {{.*#+}} ymm3 = ymm3[0,2,0,2,4,6,4,6]
251 ; AVX2-NEXT: vpshufd {{.*#+}} ymm0 = ymm0[0,2,0,2,4,6,4,6]
253 ; AVX2-NEXT: vpshufd {{.*#+}} ymm1 = ymm1[0,2,0,2,4,6,4,6]
419 ; AVX1-NEXT: vpshufd {{.*#+}} xmm1 = xmm1[0,2,2,3
    [all...]
vector-shuffle-256-v8.ll     [all...]
avx-splat.ll 19 ; CHECK-NEXT: vpshufd {{.*#+}} xmm0 = xmm0[2,2,3,3]
31 ; CHECK-NEXT: vpshufd {{.*#+}} xmm0 = xmm0[0,1,0,1]
104 ; CHECK-NEXT: vpshufd {{.*#+}} xmm0 = xmm0[0,1,0,0]
vector-shuffle-256-v16.ll 11 ; AVX1-NEXT: vpshufd {{.*#+}} xmm0 = xmm0[0,0,1,1]
27 ; AVX1-NEXT: vpshufd {{.*#+}} xmm1 = xmm1[0,0,1,1]
46 ; AVX1-NEXT: vpshufd {{.*#+}} xmm1 = xmm1[0,0,1,1]
65 ; AVX1-NEXT: vpshufd {{.*#+}} xmm1 = xmm1[0,0,1,1]
84 ; AVX1-NEXT: vpshufd {{.*#+}} xmm1 = xmm1[0,0,1,1]
103 ; AVX1-NEXT: vpshufd {{.*#+}} xmm1 = xmm1[0,0,1,1]
122 ; AVX1-NEXT: vpshufd {{.*#+}} xmm1 = xmm1[0,0,1,1]
141 ; AVX1-NEXT: vpshufd {{.*#+}} xmm1 = xmm1[0,0,1,1]
163 ; AVX1-NEXT: vpshufd {{.*#+}} xmm0 = xmm0[0,0,1,1]
172 ; AVX2-NEXT: vpshufd {{.*#+}} ymm0 = ymm0[0,0,1,1,4,4,5,5
    [all...]
  /external/libvpx/libvpx/third_party/libyuv/source/
compare_win.cc 91 vpshufd ymm1, ymm0, 0xee // 3, 2 + 1, 0 both lanes.
93 vpshufd ymm1, ymm0, 0x01 // 1 + 0 both lanes.
202 vpshufd xmm2, xmm1, 0x0e // upper 2 dwords
204 vpshufd xmm2, xmm1, 0x01
  /external/libyuv/files/source/
compare_win.cc 91 vpshufd ymm1, ymm0, 0xee // 3, 2 + 1, 0 both lanes.
93 vpshufd ymm1, ymm0, 0x01 // 1 + 0 both lanes.
202 vpshufd xmm2, xmm1, 0x0e // upper 2 dwords
204 vpshufd xmm2, xmm1, 0x01
  /external/boringssl/src/crypto/chacha/asm/
chacha-x86_64.pl 1305 vpshufd \$0x00,$xa3,$xa0 # smash key by lanes...
1306 vpshufd \$0x55,$xa3,$xa1
1308 vpshufd \$0xaa,$xa3,$xa2
1310 vpshufd \$0xff,$xa3,$xa3
1314 vpshufd \$0x00,$xb3,$xb0
1315 vpshufd \$0x55,$xb3,$xb1
1317 vpshufd \$0xaa,$xb3,$xb2
1319 vpshufd \$0xff,$xb3,$xb3
1323 vpshufd \$0x00,$xt3,$xt0 # "xc0"
1324 vpshufd \$0x55,$xt3,$xt1 # "xc1
    [all...]
  /external/boringssl/linux-x86_64/crypto/chacha/
chacha-x86_64.S 1011 vpshufd $0x00,%ymm11,%ymm8
1012 vpshufd $0x55,%ymm11,%ymm9
1014 vpshufd $0xaa,%ymm11,%ymm10
1016 vpshufd $0xff,%ymm11,%ymm11
1020 vpshufd $0x00,%ymm3,%ymm0
1021 vpshufd $0x55,%ymm3,%ymm1
1023 vpshufd $0xaa,%ymm3,%ymm2
1025 vpshufd $0xff,%ymm3,%ymm3
1029 vpshufd $0x00,%ymm15,%ymm12
1030 vpshufd $0x55,%ymm15,%ymm1
    [all...]
  /external/boringssl/mac-x86_64/crypto/chacha/
chacha-x86_64.S 1010 vpshufd $0x00,%ymm11,%ymm8
1011 vpshufd $0x55,%ymm11,%ymm9
1013 vpshufd $0xaa,%ymm11,%ymm10
1015 vpshufd $0xff,%ymm11,%ymm11
1019 vpshufd $0x00,%ymm3,%ymm0
1020 vpshufd $0x55,%ymm3,%ymm1
1022 vpshufd $0xaa,%ymm3,%ymm2
1024 vpshufd $0xff,%ymm3,%ymm3
1028 vpshufd $0x00,%ymm15,%ymm12
1029 vpshufd $0x55,%ymm15,%ymm1
    [all...]
  /external/boringssl/win-x86_64/crypto/chacha/
chacha-x86_64.asm 1104 vpshufd ymm8,ymm11,0x00
1105 vpshufd ymm9,ymm11,0x55
1107 vpshufd ymm10,ymm11,0xaa
1109 vpshufd ymm11,ymm11,0xff
1113 vpshufd ymm0,ymm3,0x00
1114 vpshufd ymm1,ymm3,0x55
1116 vpshufd ymm2,ymm3,0xaa
1118 vpshufd ymm3,ymm3,0xff
1122 vpshufd ymm12,ymm15,0x00
1123 vpshufd ymm13,ymm15,0x5
    [all...]

Completed in 732 milliseconds

1 2 3 4 5 6 7