/external/boringssl/mac-x86_64/crypto/aes/ |
bsaes-x86_64.S | 18 pxor %xmm8,%xmm2 52 pxor %xmm2,%xmm1 56 pxor %xmm1,%xmm2 81 pxor %xmm2,%xmm0 85 pxor %xmm0,%xmm2 91 movdqa %xmm2,%xmm9 92 psrlq $4,%xmm2 95 pxor %xmm6,%xmm2 97 pand %xmm7,%xmm2 99 pxor %xmm2,%xmm [all...] |
vpaes-x86_64.S | 26 movdqa L$k_ipt(%rip),%xmm2 34 pxor %xmm5,%xmm2 36 pxor %xmm2,%xmm0 53 movdqa %xmm14,%xmm2 56 pxor %xmm5,%xmm2 59 pxor %xmm2,%xmm0 82 movdqa %xmm10,%xmm2 86 pxor %xmm0,%xmm2 115 movdqa L$k_dipt(%rip),%xmm2 128 pxor %xmm5,%xmm2 [all...] |
/external/boringssl/win-x86_64/crypto/aes/ |
vpaes-x86_64.asm | 30 movdqa xmm2,XMMWORD[$L$k_ipt] 38 pxor xmm2,xmm5 40 pxor xmm0,xmm2 57 movdqa xmm2,xmm14 60 pxor xmm2,xmm5 63 pxor xmm0,xmm2 86 movdqa xmm2,xmm10 90 pxor xmm2,xmm0 119 movdqa xmm2,XMMWORD[$L$k_dipt] 132 pxor xmm2,xmm [all...] |
/external/llvm/test/CodeGen/X86/ |
sqrt-fastmath.ll | 36 ; ESTIMATE-NEXT: vmulss {{.*}}(%rip), %xmm1, %xmm2 40 ; ESTIMATE-NEXT: vmulss %xmm2, %xmm1, %xmm1 42 ; ESTIMATE-NEXT: vxorps %xmm2, %xmm2, %xmm2 43 ; ESTIMATE-NEXT: vcmpeqss %xmm2, %xmm0, %xmm0 80 ; ESTIMATE-NEXT: vmulss {{.*}}(%rip), %xmm1, %xmm2 84 ; ESTIMATE-NEXT: vmulss %xmm2, %xmm0, %xmm0 102 ; ESTIMATE-NEXT: vmulps %xmm1, %xmm1, %xmm2 103 ; ESTIMATE-NEXT: vmulps %xmm0, %xmm2, %xmm [all...] |
vector-sext.ll | 80 ; SSE2-NEXT: movd %rax, %xmm2 85 ; SSE2-NEXT: punpcklqdq {{.*#+}} xmm2 = xmm2[0],xmm1[0] 95 ; SSE2-NEXT: movdqa %xmm2, %xmm0 103 ; SSSE3-NEXT: movd %rax, %xmm2 108 ; SSSE3-NEXT: punpcklqdq {{.*#+}} xmm2 = xmm2[0],xmm1[0] 118 ; SSSE3-NEXT: movdqa %xmm2, %xmm0 129 ; SSE41-NEXT: movd %rax, %xmm2 130 ; SSE41-NEXT: punpcklqdq {{.*#+}} xmm2 = xmm2[0],xmm3[0 [all...] |
vector-trunc.ll | 77 ; SSSE3-NEXT: movdqa {{.*#+}} xmm2 = [0,1,4,5,8,9,12,13,8,9,12,13,12,13,14,15] 78 ; SSSE3-NEXT: pshufb %xmm2, %xmm1 79 ; SSSE3-NEXT: pshufb %xmm2, %xmm0 85 ; SSE41-NEXT: movdqa {{.*#+}} xmm2 = [0,1,4,5,8,9,12,13,8,9,12,13,12,13,14,15] 86 ; SSE41-NEXT: pshufb %xmm2, %xmm1 87 ; SSE41-NEXT: pshufb %xmm2, %xmm0 93 ; AVX-NEXT: vmovdqa {{.*#+}} xmm2 = [0,1,4,5,8,9,12,13,8,9,12,13,12,13,14,15] 94 ; AVX-NEXT: vpshufb %xmm2, %xmm1, %xmm1 95 ; AVX-NEXT: vpshufb %xmm2, %xmm0, %xmm0 149 ; SSE2-NEXT: movdqa {{.*#+}} xmm2 = [255,255,255,255,255,255,255,255 [all...] |
vector-shuffle-128-v16.ll | 253 ; SSE2-NEXT: movdqa {{.*#+}} xmm2 = [255,0,255,0,255,0,255,0,255,0,255,0,255,0,255,0] 257 ; SSE2-NEXT: pand %xmm2, %xmm1 258 ; SSE2-NEXT: pandn %xmm0, %xmm2 259 ; SSE2-NEXT: por %xmm1, %xmm2 260 ; SSE2-NEXT: movdqa %xmm2, %xmm0 299 ; SSE2-NEXT: movdqa %xmm0, %xmm2 300 ; SSE2-NEXT: punpckhbw {{.*#+}} xmm2 = xmm2[8],xmm1[8],xmm2[9],xmm1[9],xmm2[10],xmm1[10],xmm2[11],xmm1[11],xmm2[12],xmm1[12],xmm2[13],xmm1[13],xmm2[14],xmm1[14],xmm2[15],xmm1 (…) [all...] |
avx512-build-vector.ll | 30 ; CHECK-NEXT: vxorps %xmm2, %xmm2, %xmm2 31 ; CHECK-NEXT: vmovss %xmm0, %xmm2, %xmm0 32 ; CHECK-NEXT: vmovss %xmm1, %xmm2, %xmm1
|
stack-folding-int-avx1.ll | 14 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"() 23 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"() 32 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"() 41 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"() 50 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm1},~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"() 59 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm1},~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"() 88 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm1},~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"() 108 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"() 117 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm1},~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"() 126 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm1},~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm (…) [all...] |
/external/libunwind/src/x86/ |
regname.c | 17 "xmm0", "xmm1", "xmm2", "xmm3", "xmm4", "xmm5", "xmm6", "xmm7",
|
/external/libvpx/libvpx/third_party/libyuv/source/ |
rotate.cc | 101 movq xmm2, qword ptr [eax] local 106 punpcklbw xmm2, xmm3 local 107 movdqa xmm3, xmm2 122 punpcklwd xmm0, xmm2 124 movdqa xmm2, xmm0 local 126 palignr xmm2, xmm2, 8 local 142 punpckldq xmm2, xmm6 local 143 movdqa xmm6, xmm2 145 movq qword ptr [edx], xmm2 204 movdqa xmm2, [eax] local 208 punpcklbw xmm2, xmm3 local 233 movdqa xmm2, xmm5 local 262 punpckldq xmm2, xmm6 local [all...] |
/external/libyuv/files/source/ |
rotate.cc | 85 movq xmm2, qword ptr [eax] local 90 punpcklbw xmm2, xmm3 local 91 movdqa xmm3, xmm2 106 punpcklwd xmm0, xmm2 108 movdqa xmm2, xmm0 local 110 palignr xmm2, xmm2, 8 local 126 punpckldq xmm2, xmm6 local 127 movdqa xmm6, xmm2 129 movq qword ptr [edx], xmm2 188 movdqa xmm2, [eax] local 192 punpcklbw xmm2, xmm3 local 217 movdqa xmm2, xmm5 local 246 punpckldq xmm2, xmm6 local [all...] |
scale.cc | 246 movdqa xmm2, [eax + esi] local 249 pavgb xmm0, xmm2 // average rows 252 movdqa xmm2, xmm0 // average columns (32 to 16 pixels) local 256 pand xmm2, xmm5 local 258 pavgw xmm0, xmm2 321 movdqu xmm2, [eax + esi] local 324 pavgb xmm0, xmm2 // average rows 327 movdqa xmm2, xmm0 // average columns (32 to 16 pixels) local 331 pand xmm2, xmm5 local 333 pavgw xmm0, xmm2 399 movdqa xmm2, [eax + esi] local 403 movdqa xmm2, [eax + esi * 2] local 408 pavgb xmm2, xmm4 local 413 movdqa xmm2, xmm0 \/\/ average columns (32 to 16 pixels) local 417 pand xmm2, xmm7 local 423 movdqa xmm2, xmm0 \/\/ average columns (16 to 8 pixels) local 425 pand xmm2, xmm7 local 493 movdqa xmm2, [eax + esi] local 497 movdqa xmm2, [eax + esi * 2] local 503 pavgb xmm2, xmm4 local 508 movdqa xmm2, [ebp] local 512 pavgb xmm2, xmm4 local 520 pavgb xmm2, xmm4 local 570 movdqa xmm2, xmm1 local 574 pshufb xmm2, xmm5 local 612 movdqa xmm2, kShuf01 local 671 movdqa xmm2, kShuf01 local 767 movdqa xmm2, kShufAc local 833 movdqa xmm2, kShufAb0 local 903 movdqa xmm2, [eax] \/\/ read 16 pixels local 906 punpcklbw xmm2, xmm4 local 962 movdqa xmm2, [esi + edx] \/\/ row1 local 965 punpcklbw xmm2, xmm4 local 969 psubw xmm2, xmm0 \/\/ row1 - row0 local 971 pmulhw xmm2, xmm5 \/\/ scale diff local 1055 movdqa xmm2, [esi + edx] local [all...] |
/external/libvpx/libvpx/vp9/common/x86/ |
vp9_subpixel_8t_sse2.asm | 24 pshuflw xmm2, xmm7, 10101010b ;k2 33 punpcklqdq xmm2, xmm3 38 movdqa k2k3, xmm2 53 punpckldq xmm2, xmm3 58 punpcklbw xmm2, zero 63 pmullw xmm2, k2k3 70 paddsw xmm0, xmm2 71 psrldq xmm2, 8 74 paddsw xmm0, xmm2 97 pshuflw xmm2, xmm7, 10101010b ;k [all...] |
/hardware/intel/common/omx-components/videocodec/libvpx_internal/libvpx/vp9/common/x86/ |
vp9_subpixel_8t_sse2.asm | 24 pshuflw xmm2, xmm7, 10101010b ;k2 33 punpcklqdq xmm2, xmm3 38 movdqa k2k3, xmm2 53 punpckldq xmm2, xmm3 58 punpcklbw xmm2, zero 63 pmullw xmm2, k2k3 70 paddsw xmm0, xmm2 71 psrldq xmm2, 8 74 paddsw xmm0, xmm2 97 pshuflw xmm2, xmm7, 10101010b ;k [all...] |
/bionic/libc/arch-x86/silvermont/string/ |
sse2-memmove-slm.S | 136 movdqu -16(%eax, %ecx), %xmm2 140 movdqu %xmm2, -16(%edx, %ecx) 151 movdqu 32(%eax), %xmm2 159 movdqu %xmm2, 32(%edx) 174 movdqu 32(%eax), %xmm2 188 movdqu %xmm2, 32(%edx) 211 movdqu 32(%eax, %edi), %xmm2 215 movaps %xmm2, 32(%edi) 255 movdqu -32(%esi, %ecx), %xmm2 259 movdqu %xmm2, -32(%edi, %ecx [all...] |
sse2-memcpy-slm.S | 128 movdqu -48(%eax, %ecx), %xmm2 133 movdqu %xmm2, -48(%edx, %ecx) 164 movdqu 32(%ebx, %eax), %xmm2 168 movdqa %xmm2, 32(%ebx) 177 movdqu 32(%ebx, %eax), %xmm2 185 movdqa %xmm2, 32(%ebx) 196 movdqu 32(%ebx, %eax), %xmm2 200 movdqa %xmm2, 32(%ebx) 207 movdqu 32(%eax), %xmm2 215 movdqu %xmm2, 32(%edx [all...] |
/bionic/libc/arch-x86_64/string/ |
sse2-memmove-slm.S | 126 movdqu -16(%rsi, %rdx), %xmm2 130 movdqu %xmm2, -16(%rdi, %rdx) 141 movdqu 32(%rsi), %xmm2 149 movdqu %xmm2, 32(%rdi) 162 movdqu 32(%rsi), %xmm2 176 movdqu %xmm2, 32(%rdi) 199 movdqu 32(%r8, %rsi), %xmm2 203 movaps %xmm2, 32(%r8) 243 movdqu -32(%r9, %rdx), %xmm2 247 movdqu %xmm2, -32(%r8, %rdx [all...] |
sse2-memcpy-slm.S | 119 movdqu -48(%rsi, %rdx), %xmm2 124 movdqu %xmm2, -48(%rdi, %rdx) 155 movdqu 32(%r8, %rsi), %xmm2 159 movdqa %xmm2, 32(%r8) 168 movdqu 32(%r8, %rsi), %xmm2 176 movdqa %xmm2, 32(%r8) 187 movdqu 32(%r8, %rsi), %xmm2 191 movdqa %xmm2, 32(%r8) 198 movdqu 32(%rsi), %xmm2 206 movdqu %xmm2, 32(%rdi [all...] |
/external/boringssl/src/crypto/aes/asm/ |
vpaes-x86_64.pl | 90 movdqa .Lk_ipt(%rip), %xmm2 # iptlo 95 pshufb %xmm0, %xmm2 98 pxor %xmm5, %xmm2 100 pxor %xmm2, %xmm0 109 pshufb %xmm2, %xmm4 # 4 = sb1u 115 pshufb %xmm2, %xmm5 # 4 = sb2u 117 movdqa %xmm14, %xmm2 # 2 : sb2t 118 pshufb %xmm3, %xmm2 # 2 = sb2t 120 pxor %xmm5, %xmm2 # 2 = 2A 123 pxor %xmm2, %xmm0 # 0 = 2A+ [all...] |
/hardware/intel/common/omx-components/videocodec/libvpx_internal/libvpx/vp9/encoder/x86/ |
vp9_subpel_variance_impl_sse2.asm | 53 movdqu xmm2, XMMWORD PTR [rsi+1] ; 54 pavgb xmm1, xmm2 ; xmm1 = avg(xmm1,xmm3) horizontal line i+1 170 movq xmm2, QWORD PTR [rdi] 171 punpcklbw xmm2, xmm0 172 psubw xmm5, xmm2 173 movq xmm2, QWORD PTR [rdi+8] 174 punpcklbw xmm2, xmm0 175 psubw xmm4, xmm2 278 movq xmm2, QWORD PTR [rdi+8] 279 punpcklbw xmm2, xmm [all...] |
/bionic/libc/arch-x86/atom/string/ |
sse2-strchr-atom.S | 96 pxor %xmm2, %xmm2 108 pcmpeqb %xmm0, %xmm2 111 pmovmskb %xmm2, %edx 129 pxor %xmm2, %xmm2 136 pcmpeqb %xmm0, %xmm2 138 pmovmskb %xmm2, %edx 147 pcmpeqb %xmm0, %xmm2 149 pmovmskb %xmm2, %ed [all...] |
sse2-wcsrchr-atom.S | 97 pxor %xmm2, %xmm2 107 pcmpeqd %xmm0, %xmm2 110 pmovmskb %xmm2, %ecx 190 pcmpeqd %xmm0, %xmm2 193 pmovmskb %xmm2, %ecx 199 pcmpeqd %xmm3, %xmm2 202 pmovmskb %xmm2, %ecx 208 pcmpeqd %xmm4, %xmm2 211 pmovmskb %xmm2, %ec [all...] |
/external/llvm/test/MC/Disassembler/X86/ |
x86-64.txt | 164 # CHECK: sha1rnds4 $1, %xmm1, %xmm2 167 # CHECK: sha1rnds4 $1, (%rax), %xmm2 170 # CHECK: sha1nexte %xmm1, %xmm2 173 # CHECK: sha1nexte (%rax), %xmm2 176 # CHECK: sha1msg1 %xmm1, %xmm2 179 # CHECK: sha1msg1 (%rax), %xmm2 182 # CHECK: sha1msg2 %xmm1, %xmm2 185 # CHECK: sha1msg2 (%rax), %xmm2 188 # CHECK: sha256rnds2 (%rax), %xmm2 191 # CHECK: sha256rnds2 %xmm1, %xmm2 [all...] |
/external/llvm/test/MC/X86/ |
x86_64-fma4-encoding.s | 12 // CHECK: vfmaddss %xmm2, %xmm1, %xmm0, %xmm0 14 vfmaddss %xmm2, %xmm1, %xmm0, %xmm0 24 // CHECK: vfmaddsd %xmm2, %xmm1, %xmm0, %xmm0 26 vfmaddsd %xmm2, %xmm1, %xmm0, %xmm0 40 // CHECK: vfmaddps %xmm2, %xmm1, %xmm0, %xmm0 42 vfmaddps %xmm2, %xmm1, %xmm0, %xmm0 52 // CHECK: vfmaddpd %xmm2, %xmm1, %xmm0, %xmm0 54 vfmaddpd %xmm2, %xmm1, %xmm0, %xmm0 150 // CHECK: vfmsubss %xmm2, %xmm1, %xmm0, %xmm0 152 vfmsubss %xmm2, %xmm1, %xmm0, %xmm [all...] |