HomeSort by relevance Sort by last modified time
    Searched full:xmm3 (Results 101 - 125 of 367) sorted by null

1 2 3 45 6 7 8 91011>>

  /external/libvpx/libvpx/third_party/libyuv/source/
scale_posix.cc 138 "movdqa %%xmm1,%%xmm3 \n"
141 "pand %%xmm5,%%xmm3 \n"
143 "pavgw %%xmm3,%%xmm1 \n"
172 MEMOPREG(movdqa,0x10,0,3,1,xmm3) // movdqa 0x10(%0,%3,1),%%xmm3
175 "pavgb %%xmm3,%%xmm1 \n"
178 "movdqa %%xmm1,%%xmm3 \n"
181 "pand %%xmm5,%%xmm3 \n"
183 "pavgw %%xmm3,%%xmm1 \n"
198 , "xmm0", "xmm1", "xmm2", "xmm3", "xmm5
    [all...]
  /external/boringssl/src/crypto/aes/asm/
vpaes-x86_64.pl 110 pshufb %xmm3, %xmm0 # 0 = sb1t
118 pshufb %xmm3, %xmm2 # 2 = sb2t
119 movdqa %xmm0, %xmm3 # 3 = A
124 pshufb %xmm4, %xmm3 # 3 = D
126 pxor %xmm0, %xmm3 # 3 = 2A+B+D
130 pxor %xmm3, %xmm0 # 0 = 2A+3B+C+D
140 movdqa %xmm10, %xmm3 # 3 : 1/i
142 pshufb %xmm1, %xmm3 # 3 = 1/i
144 pxor %xmm5, %xmm3 # 3 = iak = 1/i + a/k
148 pshufb %xmm3, %xmm2 # 2 = 1/ia
    [all...]
  /bionic/libc/arch-x86/silvermont/string/
sse2-memmove-slm.S 137 movdqu -32(%eax, %ecx), %xmm3
141 movdqu %xmm3, -32(%edx, %ecx)
152 movdqu 48(%eax), %xmm3
160 movdqu %xmm3, 48(%edx)
175 movdqu 48(%eax), %xmm3
189 movdqu %xmm3, 48(%edx)
212 movdqu 48(%eax, %edi), %xmm3
216 movaps %xmm3, 48(%edi)
256 movdqu -16(%esi, %ecx), %xmm3
260 movdqu %xmm3, -16(%edi, %ecx
    [all...]
sse2-memcpy-slm.S 129 movdqu -64(%eax, %ecx), %xmm3
134 movdqu %xmm3, -64(%edx, %ecx)
165 movdqu 48(%ebx, %eax), %xmm3
169 movdqa %xmm3, 48(%ebx)
178 movdqu 48(%ebx, %eax), %xmm3
186 movdqa %xmm3, 48(%ebx)
197 movdqu 48(%ebx, %eax), %xmm3
201 movdqa %xmm3, 48(%ebx)
208 movdqu 48(%eax), %xmm3
216 movdqu %xmm3, 48(%edx
    [all...]
  /bionic/libc/arch-x86_64/string/
sse2-memmove-slm.S 127 movdqu -32(%rsi, %rdx), %xmm3
131 movdqu %xmm3, -32(%rdi, %rdx)
142 movdqu 48(%rsi), %xmm3
150 movdqu %xmm3, 48(%rdi)
163 movdqu 48(%rsi), %xmm3
177 movdqu %xmm3, 48(%rdi)
200 movdqu 48(%r8, %rsi), %xmm3
204 movaps %xmm3, 48(%r8)
244 movdqu -16(%r9, %rdx), %xmm3
248 movdqu %xmm3, -16(%r8, %rdx
    [all...]
sse2-memcpy-slm.S 120 movdqu -64(%rsi, %rdx), %xmm3
125 movdqu %xmm3, -64(%rdi, %rdx)
156 movdqu 48(%r8, %rsi), %xmm3
160 movdqa %xmm3, 48(%r8)
169 movdqu 48(%r8, %rsi), %xmm3
177 movdqa %xmm3, 48(%r8)
188 movdqu 48(%r8, %rsi), %xmm3
192 movdqa %xmm3, 48(%r8)
199 movdqu 48(%rsi), %xmm3
207 movdqu %xmm3, 48(%rdi
    [all...]
  /external/llvm/test/CodeGen/X86/
vector-blend.ll 296 ; SSE2-NEXT: movss {{.*#+}} xmm3 = xmm1[0],xmm3[1,2,3]
298 ; SSE2-NEXT: movaps %xmm3, %xmm1
304 ; SSSE3-NEXT: movss {{.*#+}} xmm3 = xmm1[0],xmm3[1,2,3]
306 ; SSSE3-NEXT: movaps %xmm3, %xmm1
312 ; SSE41-NEXT: blendps {{.*#+}} xmm1 = xmm1[0],xmm3[1,2,3]
328 ; SSE2-NEXT: movss {{.*#+}} xmm3 = xmm1[0],xmm3[1,2,3]
330 ; SSE2-NEXT: movaps %xmm3, %xmm
    [all...]
stack-folding-int-avx1.ll 14 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"()
23 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"()
32 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"()
41 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"()
50 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm1},~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"()
59 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm1},~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"()
88 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm1},~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"()
108 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"()
117 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm1},~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"()
126 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm1},~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xm (…)
    [all...]
pr3154.ll 40 %asmtmp = call { i32, i32 } asm sideeffect "1: \0A\09movapd %xmm7, %xmm1 \0A\09mulpd %xmm1, %xmm1 \0A\09movapd %xmm6, %xmm0 \0A\09subpd %xmm1, %xmm0 \0A\09pshufd $$0x4e, %xmm0, %xmm1 \0A\09cvtpi2pd ($3,$0), %xmm2 \0A\09cvtpi2pd -1*4($3,$1), %xmm3 \0A\09mulpd %xmm0, %xmm2 \0A\09mulpd %xmm1, %xmm3 \0A\09movapd %xmm2, ($2,$0,2) \0A\09movupd %xmm3, -1*8($2,$1,2) \0A\09subpd %xmm5, %xmm7 \0A\09sub $$8, $1 \0A\09add $$8, $0 \0A\09jl 1b \0A\09", "=&r,=&r,r,r,0,1,~{dirflag},~{fpsr},~{flags}"(double* %16, i32* %17, i32 %12, i32 %14) nounwind ; <{ i32, i32 }> [#uses=0]
44 %asmtmp23 = call { i32, i32 } asm sideeffect "1: \0A\09movapd %xmm7, %xmm1 \0A\09mulpd %xmm1, %xmm1 \0A\09movapd %xmm6, %xmm0 \0A\09subpd %xmm1, %xmm0 \0A\09pshufd $$0x4e, %xmm0, %xmm1 \0A\09cvtpi2pd ($3,$0), %xmm2 \0A\09cvtpi2pd -2*4($3,$1), %xmm3 \0A\09mulpd %xmm0, %xmm2 \0A\09mulpd %xmm1, %xmm3 \0A\09movapd %xmm2, ($2,$0,2) \0A\09movapd %xmm3, -2*8($2,$1,2) \0A\09subpd %xmm5, %xmm7 \0A\09sub $$8, $1 \0A\09add $$8, $0 \0A\09jl 1b \0A\09", "=&r,=&r,r,r,0,1,~{dirflag},~{fpsr},~{flags}"(double* %16, i32* %17, i32 %12, i32 %14) nounwind ; <{ i32, i32 }> [#uses=0]
84 %asmtmp32 = call i32 asm sideeffect "movsd ff_pd_1, %xmm0 \0A\09movsd ff_pd_1, %xmm1 \0A\09movsd ff_pd_1, %xmm2 \0A\091: \0A\09movapd ($4,$0), %xmm3 \0A\09movupd -8($5,$0), %xmm4 \0A\09movapd ($5,$0), %xmm5 \0A\09mulpd %xmm3, %xmm4 \0A\09mulpd %xmm3, %xmm5 \0A\09mulpd -16($5,$0), %xmm3 \0A\09addpd %xmm4, %xmm1 \0A\09addpd %xmm5, %xmm0 \0A\09addp (…)
    [all...]
stack-folding-int-sse42.ll 14 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"()
23 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"()
32 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"()
41 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"()
50 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm1},~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"()
59 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm1},~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"()
115 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm1},~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"()
135 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"()
144 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm1},~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"()
153 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm1},~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xm (…)
    [all...]
  /bionic/libm/x86/
s_expm1.S 107 movapd 96(%ebx), %xmm3
122 mulpd %xmm1, %xmm3
131 subpd %xmm3, %xmm0
133 movsd 144(%ebx), %xmm3
137 mulsd %xmm0, %xmm3
147 mulsd %xmm0, %xmm3
152 addsd %xmm3, %xmm0
153 xorpd %xmm3, %xmm3
155 pinsrw $3, %eax, %xmm3
    [all...]
  /bionic/libm/x86_64/
s_expm1.S 89 movapd 32+cv(%rip), %xmm3
104 mulpd %xmm1, %xmm3
113 subpd %xmm3, %xmm0
116 movq 80+cv(%rip), %xmm3
120 mulsd %xmm0, %xmm3
130 mulsd %xmm0, %xmm3
135 addsd %xmm3, %xmm0
136 xorpd %xmm3, %xmm3
138 pinsrw $3, %eax, %xmm3
    [all...]
s_cos.S 201 movq P_1(%rip), %xmm3
202 mulsd %xmm1, %xmm3
212 subsd %xmm3, %xmm0
214 subsd %xmm3, %xmm4
217 movq %xmm4, %xmm3
223 subsd %xmm4, %xmm3
226 subsd %xmm2, %xmm3
228 subsd %xmm3, %xmm1
229 movq 24(%rax), %xmm3
230 addsd %xmm3, %xmm
    [all...]
  /external/boringssl/linux-x86/crypto/aes/
vpaes-x86.S 104 movdqa %xmm0,%xmm3
111 pxor %xmm0,%xmm3
115 pxor %xmm3,%xmm0
123 movdqa %xmm7,%xmm3
127 pxor %xmm5,%xmm3
132 movdqa %xmm7,%xmm3
136 pxor %xmm1,%xmm3
212 movdqa %xmm7,%xmm3
216 pxor %xmm2,%xmm3
221 movdqa %xmm7,%xmm3
    [all...]
  /external/boringssl/linux-x86_64/crypto/aes/
vpaes-x86_64.S 55 movdqa %xmm0,%xmm3
62 pxor %xmm0,%xmm3
66 pxor %xmm3,%xmm0
76 movdqa %xmm10,%xmm3
80 pxor %xmm5,%xmm3
85 movdqa %xmm10,%xmm3
89 pxor %xmm1,%xmm3
182 movdqa %xmm10,%xmm3
186 pxor %xmm2,%xmm3
191 movdqa %xmm10,%xmm3
    [all...]
  /external/boringssl/mac-x86/crypto/aes/
vpaes-x86.S 101 movdqa %xmm0,%xmm3
108 pxor %xmm0,%xmm3
112 pxor %xmm3,%xmm0
120 movdqa %xmm7,%xmm3
124 pxor %xmm5,%xmm3
129 movdqa %xmm7,%xmm3
133 pxor %xmm1,%xmm3
207 movdqa %xmm7,%xmm3
211 pxor %xmm2,%xmm3
216 movdqa %xmm7,%xmm3
    [all...]
  /external/boringssl/mac-x86_64/crypto/aes/
vpaes-x86_64.S 55 movdqa %xmm0,%xmm3
62 pxor %xmm0,%xmm3
66 pxor %xmm3,%xmm0
76 movdqa %xmm10,%xmm3
80 pxor %xmm5,%xmm3
85 movdqa %xmm10,%xmm3
89 pxor %xmm1,%xmm3
182 movdqa %xmm10,%xmm3
186 pxor %xmm2,%xmm3
191 movdqa %xmm10,%xmm3
    [all...]
  /external/boringssl/win-x86/crypto/aes/
vpaes-x86.asm 112 movdqa xmm3,xmm0
119 pxor xmm3,xmm0
123 pxor xmm0,xmm3
131 movdqa xmm3,xmm7
135 pxor xmm3,xmm5
140 movdqa xmm3,xmm7
144 pxor xmm3,xmm1
217 movdqa xmm3,xmm7
221 pxor xmm3,xmm2
226 movdqa xmm3,xmm
    [all...]
  /external/boringssl/win-x86_64/crypto/aes/
vpaes-x86_64.asm 59 movdqa xmm3,xmm0
66 pxor xmm3,xmm0
70 pxor xmm0,xmm3
80 movdqa xmm3,xmm10
84 pxor xmm3,xmm5
89 movdqa xmm3,xmm10
93 pxor xmm3,xmm1
186 movdqa xmm3,xmm10
190 pxor xmm3,xmm2
195 movdqa xmm3,xmm1
    [all...]
  /external/libvpx/libvpx/vp8/common/x86/
recon_sse2.asm 48 movdqu xmm3, [rsi]
56 movdqa [rdi], xmm3
77 movdqu xmm3, [rsi]
86 movdqa [rdi], xmm3
106 movdqu xmm3, [rsi+rax]
109 movdqa [rdi+rcx], xmm3
383 movd xmm3, [rsi-1]
386 punpcklbw xmm3, xmm0
387 pshuflw xmm3, xmm3, 0x
    [all...]
  /hardware/intel/common/omx-components/videocodec/libvpx_internal/libvpx/vp8/common/x86/
recon_sse2.asm 48 movdqu xmm3, [rsi]
56 movdqa [rdi], xmm3
77 movdqu xmm3, [rsi]
86 movdqa [rdi], xmm3
106 movdqu xmm3, [rsi+rax]
109 movdqa [rdi+rcx], xmm3
382 movd xmm3, [rsi-1]
385 punpcklbw xmm3, xmm0
386 pshuflw xmm3, xmm3, 0x
    [all...]
  /external/libvpx/libvpx/vp8/encoder/x86/
encodeopt.asm 31 movdqa xmm3, [rdi+16]
34 psubw xmm2, xmm3
234 movdqa xmm3, [rdi+16]
237 psubw xmm2, xmm3
347 pxor xmm3, xmm3
357 paddd xmm3, xmm1
366 movdqa xmm1, xmm3
  /hardware/intel/common/omx-components/videocodec/libvpx_internal/libvpx/vp8/encoder/x86/
encodeopt.asm 31 movdqa xmm3, [rdi+16]
34 psubw xmm2, xmm3
234 movdqa xmm3, [rdi+16]
237 psubw xmm2, xmm3
347 pxor xmm3, xmm3
357 paddd xmm3, xmm1
366 movdqa xmm1, xmm3
  /bionic/libc/arch-x86/atom/string/
sse2-strlen-atom.S 190 pxor %xmm3, %xmm3
195 pcmpeqb (%eax), %xmm3
196 pmovmskb %xmm3, %edx
224 pcmpeqb (%eax), %xmm3
225 pmovmskb %xmm3, %edx
253 pcmpeqb (%eax), %xmm3
254 pmovmskb %xmm3, %edx
282 pcmpeqb (%eax), %xmm3
283 pmovmskb %xmm3, %ed
    [all...]
  /external/libvpx/libvpx/vp9/common/x86/
vp9_subpixel_8t_sse2.asm 25 pshuflw xmm3, xmm7, 11111111b ;k3
33 punpcklqdq xmm2, xmm3
53 punpckldq xmm2, xmm3
98 pshuflw xmm3, xmm7, 11111111b ;k3
107 punpcklwd xmm3, xmm3
116 movdqa k3, xmm3
137 movq xmm3, [rsi + rax * 2 + %1] ;3
149 punpcklbw xmm3, zero
158 pmullw xmm3, k
    [all...]

Completed in 539 milliseconds

1 2 3 45 6 7 8 91011>>