/external/libvpx/libvpx/third_party/libyuv/source/ |
scale_posix.cc | 138 "movdqa %%xmm1,%%xmm3 \n" 141 "pand %%xmm5,%%xmm3 \n" 143 "pavgw %%xmm3,%%xmm1 \n" 172 MEMOPREG(movdqa,0x10,0,3,1,xmm3) // movdqa 0x10(%0,%3,1),%%xmm3 175 "pavgb %%xmm3,%%xmm1 \n" 178 "movdqa %%xmm1,%%xmm3 \n" 181 "pand %%xmm5,%%xmm3 \n" 183 "pavgw %%xmm3,%%xmm1 \n" 198 , "xmm0", "xmm1", "xmm2", "xmm3", "xmm5 [all...] |
/external/boringssl/src/crypto/aes/asm/ |
vpaes-x86_64.pl | 110 pshufb %xmm3, %xmm0 # 0 = sb1t 118 pshufb %xmm3, %xmm2 # 2 = sb2t 119 movdqa %xmm0, %xmm3 # 3 = A 124 pshufb %xmm4, %xmm3 # 3 = D 126 pxor %xmm0, %xmm3 # 3 = 2A+B+D 130 pxor %xmm3, %xmm0 # 0 = 2A+3B+C+D 140 movdqa %xmm10, %xmm3 # 3 : 1/i 142 pshufb %xmm1, %xmm3 # 3 = 1/i 144 pxor %xmm5, %xmm3 # 3 = iak = 1/i + a/k 148 pshufb %xmm3, %xmm2 # 2 = 1/ia [all...] |
/bionic/libc/arch-x86/silvermont/string/ |
sse2-memmove-slm.S | 137 movdqu -32(%eax, %ecx), %xmm3 141 movdqu %xmm3, -32(%edx, %ecx) 152 movdqu 48(%eax), %xmm3 160 movdqu %xmm3, 48(%edx) 175 movdqu 48(%eax), %xmm3 189 movdqu %xmm3, 48(%edx) 212 movdqu 48(%eax, %edi), %xmm3 216 movaps %xmm3, 48(%edi) 256 movdqu -16(%esi, %ecx), %xmm3 260 movdqu %xmm3, -16(%edi, %ecx [all...] |
sse2-memcpy-slm.S | 129 movdqu -64(%eax, %ecx), %xmm3 134 movdqu %xmm3, -64(%edx, %ecx) 165 movdqu 48(%ebx, %eax), %xmm3 169 movdqa %xmm3, 48(%ebx) 178 movdqu 48(%ebx, %eax), %xmm3 186 movdqa %xmm3, 48(%ebx) 197 movdqu 48(%ebx, %eax), %xmm3 201 movdqa %xmm3, 48(%ebx) 208 movdqu 48(%eax), %xmm3 216 movdqu %xmm3, 48(%edx [all...] |
/bionic/libc/arch-x86_64/string/ |
sse2-memmove-slm.S | 127 movdqu -32(%rsi, %rdx), %xmm3 131 movdqu %xmm3, -32(%rdi, %rdx) 142 movdqu 48(%rsi), %xmm3 150 movdqu %xmm3, 48(%rdi) 163 movdqu 48(%rsi), %xmm3 177 movdqu %xmm3, 48(%rdi) 200 movdqu 48(%r8, %rsi), %xmm3 204 movaps %xmm3, 48(%r8) 244 movdqu -16(%r9, %rdx), %xmm3 248 movdqu %xmm3, -16(%r8, %rdx [all...] |
sse2-memcpy-slm.S | 120 movdqu -64(%rsi, %rdx), %xmm3 125 movdqu %xmm3, -64(%rdi, %rdx) 156 movdqu 48(%r8, %rsi), %xmm3 160 movdqa %xmm3, 48(%r8) 169 movdqu 48(%r8, %rsi), %xmm3 177 movdqa %xmm3, 48(%r8) 188 movdqu 48(%r8, %rsi), %xmm3 192 movdqa %xmm3, 48(%r8) 199 movdqu 48(%rsi), %xmm3 207 movdqu %xmm3, 48(%rdi [all...] |
/external/llvm/test/CodeGen/X86/ |
vector-blend.ll | 296 ; SSE2-NEXT: movss {{.*#+}} xmm3 = xmm1[0],xmm3[1,2,3] 298 ; SSE2-NEXT: movaps %xmm3, %xmm1 304 ; SSSE3-NEXT: movss {{.*#+}} xmm3 = xmm1[0],xmm3[1,2,3] 306 ; SSSE3-NEXT: movaps %xmm3, %xmm1 312 ; SSE41-NEXT: blendps {{.*#+}} xmm1 = xmm1[0],xmm3[1,2,3] 328 ; SSE2-NEXT: movss {{.*#+}} xmm3 = xmm1[0],xmm3[1,2,3] 330 ; SSE2-NEXT: movaps %xmm3, %xmm [all...] |
stack-folding-int-avx1.ll | 14 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"() 23 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"() 32 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"() 41 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"() 50 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm1},~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"() 59 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm1},~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"() 88 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm1},~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"() 108 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"() 117 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm1},~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"() 126 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm1},~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xm (…) [all...] |
pr3154.ll | 40 %asmtmp = call { i32, i32 } asm sideeffect "1: \0A\09movapd %xmm7, %xmm1 \0A\09mulpd %xmm1, %xmm1 \0A\09movapd %xmm6, %xmm0 \0A\09subpd %xmm1, %xmm0 \0A\09pshufd $$0x4e, %xmm0, %xmm1 \0A\09cvtpi2pd ($3,$0), %xmm2 \0A\09cvtpi2pd -1*4($3,$1), %xmm3 \0A\09mulpd %xmm0, %xmm2 \0A\09mulpd %xmm1, %xmm3 \0A\09movapd %xmm2, ($2,$0,2) \0A\09movupd %xmm3, -1*8($2,$1,2) \0A\09subpd %xmm5, %xmm7 \0A\09sub $$8, $1 \0A\09add $$8, $0 \0A\09jl 1b \0A\09", "=&r,=&r,r,r,0,1,~{dirflag},~{fpsr},~{flags}"(double* %16, i32* %17, i32 %12, i32 %14) nounwind ; <{ i32, i32 }> [#uses=0] 44 %asmtmp23 = call { i32, i32 } asm sideeffect "1: \0A\09movapd %xmm7, %xmm1 \0A\09mulpd %xmm1, %xmm1 \0A\09movapd %xmm6, %xmm0 \0A\09subpd %xmm1, %xmm0 \0A\09pshufd $$0x4e, %xmm0, %xmm1 \0A\09cvtpi2pd ($3,$0), %xmm2 \0A\09cvtpi2pd -2*4($3,$1), %xmm3 \0A\09mulpd %xmm0, %xmm2 \0A\09mulpd %xmm1, %xmm3 \0A\09movapd %xmm2, ($2,$0,2) \0A\09movapd %xmm3, -2*8($2,$1,2) \0A\09subpd %xmm5, %xmm7 \0A\09sub $$8, $1 \0A\09add $$8, $0 \0A\09jl 1b \0A\09", "=&r,=&r,r,r,0,1,~{dirflag},~{fpsr},~{flags}"(double* %16, i32* %17, i32 %12, i32 %14) nounwind ; <{ i32, i32 }> [#uses=0] 84 %asmtmp32 = call i32 asm sideeffect "movsd ff_pd_1, %xmm0 \0A\09movsd ff_pd_1, %xmm1 \0A\09movsd ff_pd_1, %xmm2 \0A\091: \0A\09movapd ($4,$0), %xmm3 \0A\09movupd -8($5,$0), %xmm4 \0A\09movapd ($5,$0), %xmm5 \0A\09mulpd %xmm3, %xmm4 \0A\09mulpd %xmm3, %xmm5 \0A\09mulpd -16($5,$0), %xmm3 \0A\09addpd %xmm4, %xmm1 \0A\09addpd %xmm5, %xmm0 \0A\09addp (…) [all...] |
stack-folding-int-sse42.ll | 14 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"() 23 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"() 32 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"() 41 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"() 50 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm1},~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"() 59 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm1},~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"() 115 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm1},~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"() 135 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"() 144 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm1},~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"() 153 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm1},~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xm (…) [all...] |
/bionic/libm/x86/ |
s_expm1.S | 107 movapd 96(%ebx), %xmm3 122 mulpd %xmm1, %xmm3 131 subpd %xmm3, %xmm0 133 movsd 144(%ebx), %xmm3 137 mulsd %xmm0, %xmm3 147 mulsd %xmm0, %xmm3 152 addsd %xmm3, %xmm0 153 xorpd %xmm3, %xmm3 155 pinsrw $3, %eax, %xmm3 [all...] |
/bionic/libm/x86_64/ |
s_expm1.S | 89 movapd 32+cv(%rip), %xmm3 104 mulpd %xmm1, %xmm3 113 subpd %xmm3, %xmm0 116 movq 80+cv(%rip), %xmm3 120 mulsd %xmm0, %xmm3 130 mulsd %xmm0, %xmm3 135 addsd %xmm3, %xmm0 136 xorpd %xmm3, %xmm3 138 pinsrw $3, %eax, %xmm3 [all...] |
s_cos.S | 201 movq P_1(%rip), %xmm3 202 mulsd %xmm1, %xmm3 212 subsd %xmm3, %xmm0 214 subsd %xmm3, %xmm4 217 movq %xmm4, %xmm3 223 subsd %xmm4, %xmm3 226 subsd %xmm2, %xmm3 228 subsd %xmm3, %xmm1 229 movq 24(%rax), %xmm3 230 addsd %xmm3, %xmm [all...] |
/external/boringssl/linux-x86/crypto/aes/ |
vpaes-x86.S | 104 movdqa %xmm0,%xmm3 111 pxor %xmm0,%xmm3 115 pxor %xmm3,%xmm0 123 movdqa %xmm7,%xmm3 127 pxor %xmm5,%xmm3 132 movdqa %xmm7,%xmm3 136 pxor %xmm1,%xmm3 212 movdqa %xmm7,%xmm3 216 pxor %xmm2,%xmm3 221 movdqa %xmm7,%xmm3 [all...] |
/external/boringssl/linux-x86_64/crypto/aes/ |
vpaes-x86_64.S | 55 movdqa %xmm0,%xmm3 62 pxor %xmm0,%xmm3 66 pxor %xmm3,%xmm0 76 movdqa %xmm10,%xmm3 80 pxor %xmm5,%xmm3 85 movdqa %xmm10,%xmm3 89 pxor %xmm1,%xmm3 182 movdqa %xmm10,%xmm3 186 pxor %xmm2,%xmm3 191 movdqa %xmm10,%xmm3 [all...] |
/external/boringssl/mac-x86/crypto/aes/ |
vpaes-x86.S | 101 movdqa %xmm0,%xmm3 108 pxor %xmm0,%xmm3 112 pxor %xmm3,%xmm0 120 movdqa %xmm7,%xmm3 124 pxor %xmm5,%xmm3 129 movdqa %xmm7,%xmm3 133 pxor %xmm1,%xmm3 207 movdqa %xmm7,%xmm3 211 pxor %xmm2,%xmm3 216 movdqa %xmm7,%xmm3 [all...] |
/external/boringssl/mac-x86_64/crypto/aes/ |
vpaes-x86_64.S | 55 movdqa %xmm0,%xmm3 62 pxor %xmm0,%xmm3 66 pxor %xmm3,%xmm0 76 movdqa %xmm10,%xmm3 80 pxor %xmm5,%xmm3 85 movdqa %xmm10,%xmm3 89 pxor %xmm1,%xmm3 182 movdqa %xmm10,%xmm3 186 pxor %xmm2,%xmm3 191 movdqa %xmm10,%xmm3 [all...] |
/external/boringssl/win-x86/crypto/aes/ |
vpaes-x86.asm | 112 movdqa xmm3,xmm0 119 pxor xmm3,xmm0 123 pxor xmm0,xmm3 131 movdqa xmm3,xmm7 135 pxor xmm3,xmm5 140 movdqa xmm3,xmm7 144 pxor xmm3,xmm1 217 movdqa xmm3,xmm7 221 pxor xmm3,xmm2 226 movdqa xmm3,xmm [all...] |
/external/boringssl/win-x86_64/crypto/aes/ |
vpaes-x86_64.asm | 59 movdqa xmm3,xmm0 66 pxor xmm3,xmm0 70 pxor xmm0,xmm3 80 movdqa xmm3,xmm10 84 pxor xmm3,xmm5 89 movdqa xmm3,xmm10 93 pxor xmm3,xmm1 186 movdqa xmm3,xmm10 190 pxor xmm3,xmm2 195 movdqa xmm3,xmm1 [all...] |
/external/libvpx/libvpx/vp8/common/x86/ |
recon_sse2.asm | 48 movdqu xmm3, [rsi] 56 movdqa [rdi], xmm3 77 movdqu xmm3, [rsi] 86 movdqa [rdi], xmm3 106 movdqu xmm3, [rsi+rax] 109 movdqa [rdi+rcx], xmm3 383 movd xmm3, [rsi-1] 386 punpcklbw xmm3, xmm0 387 pshuflw xmm3, xmm3, 0x [all...] |
/hardware/intel/common/omx-components/videocodec/libvpx_internal/libvpx/vp8/common/x86/ |
recon_sse2.asm | 48 movdqu xmm3, [rsi] 56 movdqa [rdi], xmm3 77 movdqu xmm3, [rsi] 86 movdqa [rdi], xmm3 106 movdqu xmm3, [rsi+rax] 109 movdqa [rdi+rcx], xmm3 382 movd xmm3, [rsi-1] 385 punpcklbw xmm3, xmm0 386 pshuflw xmm3, xmm3, 0x [all...] |
/external/libvpx/libvpx/vp8/encoder/x86/ |
encodeopt.asm | 31 movdqa xmm3, [rdi+16] 34 psubw xmm2, xmm3 234 movdqa xmm3, [rdi+16] 237 psubw xmm2, xmm3 347 pxor xmm3, xmm3 357 paddd xmm3, xmm1 366 movdqa xmm1, xmm3
|
/hardware/intel/common/omx-components/videocodec/libvpx_internal/libvpx/vp8/encoder/x86/ |
encodeopt.asm | 31 movdqa xmm3, [rdi+16] 34 psubw xmm2, xmm3 234 movdqa xmm3, [rdi+16] 237 psubw xmm2, xmm3 347 pxor xmm3, xmm3 357 paddd xmm3, xmm1 366 movdqa xmm1, xmm3
|
/bionic/libc/arch-x86/atom/string/ |
sse2-strlen-atom.S | 190 pxor %xmm3, %xmm3 195 pcmpeqb (%eax), %xmm3 196 pmovmskb %xmm3, %edx 224 pcmpeqb (%eax), %xmm3 225 pmovmskb %xmm3, %edx 253 pcmpeqb (%eax), %xmm3 254 pmovmskb %xmm3, %edx 282 pcmpeqb (%eax), %xmm3 283 pmovmskb %xmm3, %ed [all...] |
/external/libvpx/libvpx/vp9/common/x86/ |
vp9_subpixel_8t_sse2.asm | 25 pshuflw xmm3, xmm7, 11111111b ;k3 33 punpcklqdq xmm2, xmm3 53 punpckldq xmm2, xmm3 98 pshuflw xmm3, xmm7, 11111111b ;k3 107 punpcklwd xmm3, xmm3 116 movdqa k3, xmm3 137 movq xmm3, [rsi + rax * 2 + %1] ;3 149 punpcklbw xmm3, zero 158 pmullw xmm3, k [all...] |