HomeSort by relevance Sort by last modified time
    Searched full:xmm2 (Results 126 - 150 of 435) sorted by null

1 2 3 4 56 7 8 91011>>

  /external/boringssl/mac-x86_64/crypto/aes/
bsaes-x86_64.S 18 pxor %xmm8,%xmm2
52 pxor %xmm2,%xmm1
56 pxor %xmm1,%xmm2
81 pxor %xmm2,%xmm0
85 pxor %xmm0,%xmm2
91 movdqa %xmm2,%xmm9
92 psrlq $4,%xmm2
95 pxor %xmm6,%xmm2
97 pand %xmm7,%xmm2
99 pxor %xmm2,%xmm
    [all...]
vpaes-x86_64.S 26 movdqa L$k_ipt(%rip),%xmm2
34 pxor %xmm5,%xmm2
36 pxor %xmm2,%xmm0
53 movdqa %xmm14,%xmm2
56 pxor %xmm5,%xmm2
59 pxor %xmm2,%xmm0
82 movdqa %xmm10,%xmm2
86 pxor %xmm0,%xmm2
115 movdqa L$k_dipt(%rip),%xmm2
128 pxor %xmm5,%xmm2
    [all...]
  /external/boringssl/win-x86_64/crypto/aes/
vpaes-x86_64.asm 30 movdqa xmm2,XMMWORD[$L$k_ipt]
38 pxor xmm2,xmm5
40 pxor xmm0,xmm2
57 movdqa xmm2,xmm14
60 pxor xmm2,xmm5
63 pxor xmm0,xmm2
86 movdqa xmm2,xmm10
90 pxor xmm2,xmm0
119 movdqa xmm2,XMMWORD[$L$k_dipt]
132 pxor xmm2,xmm
    [all...]
  /external/llvm/test/CodeGen/X86/
sqrt-fastmath.ll 36 ; ESTIMATE-NEXT: vmulss {{.*}}(%rip), %xmm1, %xmm2
40 ; ESTIMATE-NEXT: vmulss %xmm2, %xmm1, %xmm1
42 ; ESTIMATE-NEXT: vxorps %xmm2, %xmm2, %xmm2
43 ; ESTIMATE-NEXT: vcmpeqss %xmm2, %xmm0, %xmm0
80 ; ESTIMATE-NEXT: vmulss {{.*}}(%rip), %xmm1, %xmm2
84 ; ESTIMATE-NEXT: vmulss %xmm2, %xmm0, %xmm0
102 ; ESTIMATE-NEXT: vmulps %xmm1, %xmm1, %xmm2
103 ; ESTIMATE-NEXT: vmulps %xmm0, %xmm2, %xmm
    [all...]
vector-sext.ll 80 ; SSE2-NEXT: movd %rax, %xmm2
85 ; SSE2-NEXT: punpcklqdq {{.*#+}} xmm2 = xmm2[0],xmm1[0]
95 ; SSE2-NEXT: movdqa %xmm2, %xmm0
103 ; SSSE3-NEXT: movd %rax, %xmm2
108 ; SSSE3-NEXT: punpcklqdq {{.*#+}} xmm2 = xmm2[0],xmm1[0]
118 ; SSSE3-NEXT: movdqa %xmm2, %xmm0
129 ; SSE41-NEXT: movd %rax, %xmm2
130 ; SSE41-NEXT: punpcklqdq {{.*#+}} xmm2 = xmm2[0],xmm3[0
    [all...]
vector-trunc.ll 77 ; SSSE3-NEXT: movdqa {{.*#+}} xmm2 = [0,1,4,5,8,9,12,13,8,9,12,13,12,13,14,15]
78 ; SSSE3-NEXT: pshufb %xmm2, %xmm1
79 ; SSSE3-NEXT: pshufb %xmm2, %xmm0
85 ; SSE41-NEXT: movdqa {{.*#+}} xmm2 = [0,1,4,5,8,9,12,13,8,9,12,13,12,13,14,15]
86 ; SSE41-NEXT: pshufb %xmm2, %xmm1
87 ; SSE41-NEXT: pshufb %xmm2, %xmm0
93 ; AVX-NEXT: vmovdqa {{.*#+}} xmm2 = [0,1,4,5,8,9,12,13,8,9,12,13,12,13,14,15]
94 ; AVX-NEXT: vpshufb %xmm2, %xmm1, %xmm1
95 ; AVX-NEXT: vpshufb %xmm2, %xmm0, %xmm0
149 ; SSE2-NEXT: movdqa {{.*#+}} xmm2 = [255,255,255,255,255,255,255,255
    [all...]
vector-shuffle-128-v16.ll 253 ; SSE2-NEXT: movdqa {{.*#+}} xmm2 = [255,0,255,0,255,0,255,0,255,0,255,0,255,0,255,0]
257 ; SSE2-NEXT: pand %xmm2, %xmm1
258 ; SSE2-NEXT: pandn %xmm0, %xmm2
259 ; SSE2-NEXT: por %xmm1, %xmm2
260 ; SSE2-NEXT: movdqa %xmm2, %xmm0
299 ; SSE2-NEXT: movdqa %xmm0, %xmm2
300 ; SSE2-NEXT: punpckhbw {{.*#+}} xmm2 = xmm2[8],xmm1[8],xmm2[9],xmm1[9],xmm2[10],xmm1[10],xmm2[11],xmm1[11],xmm2[12],xmm1[12],xmm2[13],xmm1[13],xmm2[14],xmm1[14],xmm2[15],xmm1 (…)
    [all...]
avx512-build-vector.ll 30 ; CHECK-NEXT: vxorps %xmm2, %xmm2, %xmm2
31 ; CHECK-NEXT: vmovss %xmm0, %xmm2, %xmm0
32 ; CHECK-NEXT: vmovss %xmm1, %xmm2, %xmm1
stack-folding-int-avx1.ll 14 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"()
23 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"()
32 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"()
41 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"()
50 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm1},~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"()
59 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm1},~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"()
88 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm1},~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"()
108 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"()
117 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm1},~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"()
126 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm1},~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm (…)
    [all...]
  /external/libunwind/src/x86/
regname.c 17 "xmm0", "xmm1", "xmm2", "xmm3", "xmm4", "xmm5", "xmm6", "xmm7",
  /external/libvpx/libvpx/third_party/libyuv/source/
rotate.cc 101 movq xmm2, qword ptr [eax] local
106 punpcklbw xmm2, xmm3 local
107 movdqa xmm3, xmm2
122 punpcklwd xmm0, xmm2
124 movdqa xmm2, xmm0 local
126 palignr xmm2, xmm2, 8 local
142 punpckldq xmm2, xmm6 local
143 movdqa xmm6, xmm2
145 movq qword ptr [edx], xmm2
204 movdqa xmm2, [eax] local
208 punpcklbw xmm2, xmm3 local
233 movdqa xmm2, xmm5 local
262 punpckldq xmm2, xmm6 local
    [all...]
  /external/libyuv/files/source/
rotate.cc 85 movq xmm2, qword ptr [eax] local
90 punpcklbw xmm2, xmm3 local
91 movdqa xmm3, xmm2
106 punpcklwd xmm0, xmm2
108 movdqa xmm2, xmm0 local
110 palignr xmm2, xmm2, 8 local
126 punpckldq xmm2, xmm6 local
127 movdqa xmm6, xmm2
129 movq qword ptr [edx], xmm2
188 movdqa xmm2, [eax] local
192 punpcklbw xmm2, xmm3 local
217 movdqa xmm2, xmm5 local
246 punpckldq xmm2, xmm6 local
    [all...]
scale.cc 246 movdqa xmm2, [eax + esi] local
249 pavgb xmm0, xmm2 // average rows
252 movdqa xmm2, xmm0 // average columns (32 to 16 pixels) local
256 pand xmm2, xmm5 local
258 pavgw xmm0, xmm2
321 movdqu xmm2, [eax + esi] local
324 pavgb xmm0, xmm2 // average rows
327 movdqa xmm2, xmm0 // average columns (32 to 16 pixels) local
331 pand xmm2, xmm5 local
333 pavgw xmm0, xmm2
399 movdqa xmm2, [eax + esi] local
403 movdqa xmm2, [eax + esi * 2] local
408 pavgb xmm2, xmm4 local
413 movdqa xmm2, xmm0 \/\/ average columns (32 to 16 pixels) local
417 pand xmm2, xmm7 local
423 movdqa xmm2, xmm0 \/\/ average columns (16 to 8 pixels) local
425 pand xmm2, xmm7 local
493 movdqa xmm2, [eax + esi] local
497 movdqa xmm2, [eax + esi * 2] local
503 pavgb xmm2, xmm4 local
508 movdqa xmm2, [ebp] local
512 pavgb xmm2, xmm4 local
520 pavgb xmm2, xmm4 local
570 movdqa xmm2, xmm1 local
574 pshufb xmm2, xmm5 local
612 movdqa xmm2, kShuf01 local
671 movdqa xmm2, kShuf01 local
767 movdqa xmm2, kShufAc local
833 movdqa xmm2, kShufAb0 local
903 movdqa xmm2, [eax] \/\/ read 16 pixels local
906 punpcklbw xmm2, xmm4 local
962 movdqa xmm2, [esi + edx] \/\/ row1 local
965 punpcklbw xmm2, xmm4 local
969 psubw xmm2, xmm0 \/\/ row1 - row0 local
971 pmulhw xmm2, xmm5 \/\/ scale diff local
1055 movdqa xmm2, [esi + edx] local
    [all...]
  /external/libvpx/libvpx/vp9/common/x86/
vp9_subpixel_8t_sse2.asm 24 pshuflw xmm2, xmm7, 10101010b ;k2
33 punpcklqdq xmm2, xmm3
38 movdqa k2k3, xmm2
53 punpckldq xmm2, xmm3
58 punpcklbw xmm2, zero
63 pmullw xmm2, k2k3
70 paddsw xmm0, xmm2
71 psrldq xmm2, 8
74 paddsw xmm0, xmm2
97 pshuflw xmm2, xmm7, 10101010b ;k
    [all...]
  /hardware/intel/common/omx-components/videocodec/libvpx_internal/libvpx/vp9/common/x86/
vp9_subpixel_8t_sse2.asm 24 pshuflw xmm2, xmm7, 10101010b ;k2
33 punpcklqdq xmm2, xmm3
38 movdqa k2k3, xmm2
53 punpckldq xmm2, xmm3
58 punpcklbw xmm2, zero
63 pmullw xmm2, k2k3
70 paddsw xmm0, xmm2
71 psrldq xmm2, 8
74 paddsw xmm0, xmm2
97 pshuflw xmm2, xmm7, 10101010b ;k
    [all...]
  /bionic/libc/arch-x86/silvermont/string/
sse2-memmove-slm.S 136 movdqu -16(%eax, %ecx), %xmm2
140 movdqu %xmm2, -16(%edx, %ecx)
151 movdqu 32(%eax), %xmm2
159 movdqu %xmm2, 32(%edx)
174 movdqu 32(%eax), %xmm2
188 movdqu %xmm2, 32(%edx)
211 movdqu 32(%eax, %edi), %xmm2
215 movaps %xmm2, 32(%edi)
255 movdqu -32(%esi, %ecx), %xmm2
259 movdqu %xmm2, -32(%edi, %ecx
    [all...]
sse2-memcpy-slm.S 128 movdqu -48(%eax, %ecx), %xmm2
133 movdqu %xmm2, -48(%edx, %ecx)
164 movdqu 32(%ebx, %eax), %xmm2
168 movdqa %xmm2, 32(%ebx)
177 movdqu 32(%ebx, %eax), %xmm2
185 movdqa %xmm2, 32(%ebx)
196 movdqu 32(%ebx, %eax), %xmm2
200 movdqa %xmm2, 32(%ebx)
207 movdqu 32(%eax), %xmm2
215 movdqu %xmm2, 32(%edx
    [all...]
  /bionic/libc/arch-x86_64/string/
sse2-memmove-slm.S 126 movdqu -16(%rsi, %rdx), %xmm2
130 movdqu %xmm2, -16(%rdi, %rdx)
141 movdqu 32(%rsi), %xmm2
149 movdqu %xmm2, 32(%rdi)
162 movdqu 32(%rsi), %xmm2
176 movdqu %xmm2, 32(%rdi)
199 movdqu 32(%r8, %rsi), %xmm2
203 movaps %xmm2, 32(%r8)
243 movdqu -32(%r9, %rdx), %xmm2
247 movdqu %xmm2, -32(%r8, %rdx
    [all...]
sse2-memcpy-slm.S 119 movdqu -48(%rsi, %rdx), %xmm2
124 movdqu %xmm2, -48(%rdi, %rdx)
155 movdqu 32(%r8, %rsi), %xmm2
159 movdqa %xmm2, 32(%r8)
168 movdqu 32(%r8, %rsi), %xmm2
176 movdqa %xmm2, 32(%r8)
187 movdqu 32(%r8, %rsi), %xmm2
191 movdqa %xmm2, 32(%r8)
198 movdqu 32(%rsi), %xmm2
206 movdqu %xmm2, 32(%rdi
    [all...]
  /external/boringssl/src/crypto/aes/asm/
vpaes-x86_64.pl 90 movdqa .Lk_ipt(%rip), %xmm2 # iptlo
95 pshufb %xmm0, %xmm2
98 pxor %xmm5, %xmm2
100 pxor %xmm2, %xmm0
109 pshufb %xmm2, %xmm4 # 4 = sb1u
115 pshufb %xmm2, %xmm5 # 4 = sb2u
117 movdqa %xmm14, %xmm2 # 2 : sb2t
118 pshufb %xmm3, %xmm2 # 2 = sb2t
120 pxor %xmm5, %xmm2 # 2 = 2A
123 pxor %xmm2, %xmm0 # 0 = 2A+
    [all...]
  /hardware/intel/common/omx-components/videocodec/libvpx_internal/libvpx/vp9/encoder/x86/
vp9_subpel_variance_impl_sse2.asm 53 movdqu xmm2, XMMWORD PTR [rsi+1] ;
54 pavgb xmm1, xmm2 ; xmm1 = avg(xmm1,xmm3) horizontal line i+1
170 movq xmm2, QWORD PTR [rdi]
171 punpcklbw xmm2, xmm0
172 psubw xmm5, xmm2
173 movq xmm2, QWORD PTR [rdi+8]
174 punpcklbw xmm2, xmm0
175 psubw xmm4, xmm2
278 movq xmm2, QWORD PTR [rdi+8]
279 punpcklbw xmm2, xmm
    [all...]
  /bionic/libc/arch-x86/atom/string/
sse2-strchr-atom.S 96 pxor %xmm2, %xmm2
108 pcmpeqb %xmm0, %xmm2
111 pmovmskb %xmm2, %edx
129 pxor %xmm2, %xmm2
136 pcmpeqb %xmm0, %xmm2
138 pmovmskb %xmm2, %edx
147 pcmpeqb %xmm0, %xmm2
149 pmovmskb %xmm2, %ed
    [all...]
sse2-wcsrchr-atom.S 97 pxor %xmm2, %xmm2
107 pcmpeqd %xmm0, %xmm2
110 pmovmskb %xmm2, %ecx
190 pcmpeqd %xmm0, %xmm2
193 pmovmskb %xmm2, %ecx
199 pcmpeqd %xmm3, %xmm2
202 pmovmskb %xmm2, %ecx
208 pcmpeqd %xmm4, %xmm2
211 pmovmskb %xmm2, %ec
    [all...]
  /external/llvm/test/MC/Disassembler/X86/
x86-64.txt 164 # CHECK: sha1rnds4 $1, %xmm1, %xmm2
167 # CHECK: sha1rnds4 $1, (%rax), %xmm2
170 # CHECK: sha1nexte %xmm1, %xmm2
173 # CHECK: sha1nexte (%rax), %xmm2
176 # CHECK: sha1msg1 %xmm1, %xmm2
179 # CHECK: sha1msg1 (%rax), %xmm2
182 # CHECK: sha1msg2 %xmm1, %xmm2
185 # CHECK: sha1msg2 (%rax), %xmm2
188 # CHECK: sha256rnds2 (%rax), %xmm2
191 # CHECK: sha256rnds2 %xmm1, %xmm2
    [all...]
  /external/llvm/test/MC/X86/
x86_64-fma4-encoding.s 12 // CHECK: vfmaddss %xmm2, %xmm1, %xmm0, %xmm0
14 vfmaddss %xmm2, %xmm1, %xmm0, %xmm0
24 // CHECK: vfmaddsd %xmm2, %xmm1, %xmm0, %xmm0
26 vfmaddsd %xmm2, %xmm1, %xmm0, %xmm0
40 // CHECK: vfmaddps %xmm2, %xmm1, %xmm0, %xmm0
42 vfmaddps %xmm2, %xmm1, %xmm0, %xmm0
52 // CHECK: vfmaddpd %xmm2, %xmm1, %xmm0, %xmm0
54 vfmaddpd %xmm2, %xmm1, %xmm0, %xmm0
150 // CHECK: vfmsubss %xmm2, %xmm1, %xmm0, %xmm0
152 vfmsubss %xmm2, %xmm1, %xmm0, %xmm
    [all...]

Completed in 331 milliseconds

1 2 3 4 56 7 8 91011>>