/external/boringssl/mac-x86/crypto/fipsmodule/ |
ghash-x86.S | 718 movdqa %xmm2,%xmm3 727 movdqa %xmm2,%xmm0 728 movdqa %xmm0,%xmm1 738 movdqa %xmm3,%xmm4 743 movdqa %xmm0,%xmm4 744 movdqa %xmm0,%xmm3 750 movdqa %xmm0,%xmm3 755 movdqa %xmm0,%xmm4 784 movdqa (%ecx),%xmm5 788 movdqa %xmm0,%xmm [all...] |
sha1-586.S | 1408 movdqa (%ebp),%xmm7 1409 movdqa 16(%ebp),%xmm0 1410 movdqa 32(%ebp),%xmm1 1411 movdqa 48(%ebp),%xmm2 1412 movdqa 64(%ebp),%xmm6 1419 movdqa %xmm0,112(%esp) 1420 movdqa %xmm1,128(%esp) 1421 movdqa %xmm2,144(%esp) 1423 movdqa %xmm7,160(%esp) 1425 movdqa %xmm6,176(%esp [all...] |
/external/boringssl/linux-x86/crypto/fipsmodule/ |
sha1-586.S | 1411 movdqa (%ebp),%xmm7 1412 movdqa 16(%ebp),%xmm0 1413 movdqa 32(%ebp),%xmm1 1414 movdqa 48(%ebp),%xmm2 1415 movdqa 64(%ebp),%xmm6 1422 movdqa %xmm0,112(%esp) 1423 movdqa %xmm1,128(%esp) 1424 movdqa %xmm2,144(%esp) 1426 movdqa %xmm7,160(%esp) 1428 movdqa %xmm6,176(%esp [all...] |
/external/llvm/test/CodeGen/X86/ |
vec_uint_to_fp.ll | 43 ; SSE: movdqa [[MASKCSTADDR]](%rip), [[MASK:%xmm[0-9]+]] 56 ; SSE41: movdqa [[LOWCSTADDR]](%rip), [[LOWVEC:%xmm[0-9]+]] 102 ; SSE: movdqa {{.*#+}} [[MASK:xmm[0-9]+]] = [65535,65535,65535,65535] 103 ; SSE-NEXT: movdqa %xmm0, [[VECLOW:%xmm[0-9]+]] 105 ; SSE-NEXT: movdqa {{.*#+}} [[LOWCST:xmm[0-9]+]] = [1258291200,1258291200,1258291200,1258291200] 108 ; SSE-NEXT: movdqa {{.*#+}} [[HIGHCST:xmm[0-9]+]] = [1392508928,1392508928,1392508928,1392508928] 122 ; SSE41: movdqa {{.*#+}} [[LOWCST:xmm[0-9]+]] = [1258291200,1258291200,1258291200,1258291200] 123 ; SSE41-NEXT: movdqa %xmm0, [[VECLOW:%xmm[0-9]+]] 126 ; SSE41-NEXT: movdqa {{.*#+}} [[HIGHCST:xmm[0-9]+]] = [1392508928,1392508928,1392508928,1392508928]
|
sse-domains.ll | 21 ; CHECK: movdqa 26 ; CHECK: movdqa 27 ; CHECK: movdqa
|
viabs.ll | 11 ; SSE2-NEXT: movdqa %xmm0, %xmm1 35 ; SSE2-NEXT: movdqa %xmm0, %xmm1 59 ; SSE2-NEXT: movdqa %xmm0, %xmm1 107 ; SSE2-NEXT: movdqa %xmm0, %xmm1 131 ; SSE2-NEXT: movdqa %xmm0, %xmm2 135 ; SSE2-NEXT: movdqa %xmm1, %xmm2 177 ; SSE2-NEXT: movdqa %xmm0, %xmm2 181 ; SSE2-NEXT: movdqa %xmm1, %xmm2 223 ; SSE2-NEXT: movdqa %xmm0, %xmm2 227 ; SSE2-NEXT: movdqa %xmm1, %xmm [all...] |
palignr.ll | 23 ; CHECK-NEXT: movdqa %xmm1, %xmm0 39 ; CHECK-NEXT: movdqa %xmm1, %xmm0 87 ; CHECK-NEXT: movdqa %xmm1, %xmm0 104 ; CHECK-NEXT: movdqa %xmm1, %xmm0 121 ; CHECK-NEXT: movdqa %xmm1, %xmm0 142 ; CHECK-NEXT: movdqa %xmm1, %xmm0 147 ; CHECK-YONAH-NEXT: movdqa %xmm1, %xmm0 151 ; CHECK-YONAH-NEXT: movdqa %xmm1, %xmm0
|
pmul.ll | 11 ; SSE2-NEXT: movdqa {{.*#+}} xmm1 = [117,117,117,117,117,117,117,117,117,117,117,117,117,117,117,117] 13 ; SSE2-NEXT: movdqa %xmm0, %xmm2 17 ; SSE2-NEXT: movdqa {{.*#+}} xmm3 = [255,255,255,255,255,255,255,255] 31 ; SSE41-NEXT: movdqa {{.*#+}} xmm3 = [255,255,255,255,255,255,255,255] 38 ; SSE41-NEXT: movdqa %xmm1, %xmm0 94 ; SSE2-NEXT: movdqa {{.*#+}} xmm1 = [117,117,117,117] 121 ; SSE-NEXT: movdqa {{.*#+}} xmm1 = [117,117] 122 ; SSE-NEXT: movdqa %xmm0, %xmm2 147 ; SSE2-NEXT: movdqa %xmm1, %xmm2 150 ; SSE2-NEXT: movdqa %xmm0, %xmm [all...] |
pr14161.ll | 8 ; CHECK-NEXT: movdqa (%rdi), %xmm0 27 ; CHECK-NEXT: movdqa (%rdi), %xmm0
|
vec_cmp_uint-128.ll | 231 ; SSE2-NEXT: movdqa {{.*#+}} xmm2 = [2147483648,2147483648,2147483648,2147483648] 234 ; SSE2-NEXT: movdqa %xmm1, %xmm2 248 ; SSE41-NEXT: movdqa {{.*#+}} xmm2 = [2147483648,2147483648,2147483648,2147483648] 251 ; SSE41-NEXT: movdqa %xmm1, %xmm2 265 ; SSE42-NEXT: movdqa {{.*#+}} xmm2 = [9223372036854775808,9223372036854775808] 295 ; SSE2-NEXT: movdqa {{.*#+}} xmm2 = [2147483648,2147483648,2147483648,2147483648] 394 ; SSE2-NEXT: movdqa {{.*#+}} xmm2 = [2147483648,2147483648,2147483648,2147483648] 397 ; SSE2-NEXT: movdqa %xmm0, %xmm2 409 ; SSE41-NEXT: movdqa {{.*#+}} xmm2 = [2147483648,2147483648,2147483648,2147483648] 412 ; SSE41-NEXT: movdqa %xmm0, %xmm [all...] |
vector-zext.ll | 41 ; SSE2-NEXT: movdqa %xmm0, %xmm1 49 ; SSSE3-NEXT: movdqa %xmm0, %xmm1 57 ; SSE41-NEXT: movdqa %xmm0, %xmm1 118 ; SSE2-NEXT: movdqa %xmm0, %xmm1 121 ; SSE2-NEXT: movdqa %xmm1, %xmm0 128 ; SSSE3-NEXT: movdqa %xmm0, %xmm1 131 ; SSSE3-NEXT: movdqa %xmm1, %xmm0 141 ; SSE41-NEXT: movdqa %xmm2, %xmm0 199 ; SSE2-NEXT: movdqa %xmm0, %xmm1 203 ; SSE2-NEXT: movdqa %xmm1, %xmm [all...] |
dagcombine-buildvector.ll | 17 ; CHECK: movdqa
|
vec_ins_extract-1.ll | 43 ; X32-NEXT: movdqa %xmm0, (%esp) 53 ; X64-NEXT: movdqa %xmm0, -{{[0-9]+}}(%rsp) 69 ; X32-NEXT: movdqa %xmm0, (%esp) 77 ; X64-NEXT: movdqa %xmm0, -{{[0-9]+}}(%rsp)
|
widen_conv-1.ll | 37 ; X86-NEXT: movdqa (%ecx), %xmm0 49 ; X64-NEXT: movdqa (%rsi), %xmm0 76 ; X86-NEXT: movdqa (%ecx), %xmm0 88 ; X64-NEXT: movdqa (%rsi), %xmm0
|
/bionic/libc/arch-x86/atom/string/ |
sse2-wcschr-atom.S | 119 movdqa (%edi), %xmm0 160 movdqa (%ecx), %xmm0 169 movdqa (%ecx), %xmm0 178 movdqa (%ecx), %xmm0 187 movdqa (%ecx), %xmm0
|
sse2-memrchr-atom.S | 125 movdqa 48(%ecx), %xmm0 131 movdqa 32(%ecx), %xmm2 137 movdqa 16(%ecx), %xmm3 143 movdqa (%ecx), %xmm4 153 movdqa 48(%ecx), %xmm0 159 movdqa 32(%ecx), %xmm2 165 movdqa 16(%ecx), %xmm3 171 movdqa (%ecx), %xmm3 193 movdqa (%ecx), %xmm0 194 movdqa 16(%ecx), %xmm [all...] |
/external/boringssl/win-x86_64/crypto/fipsmodule/ |
sha1-x86_64.asm | 1300 movdqa xmm6,XMMWORD[64+r14] 1301 movdqa xmm9,XMMWORD[((-64))+r14] 1314 movdqa XMMWORD[rsp],xmm0 1316 movdqa XMMWORD[16+rsp],xmm1 1318 movdqa XMMWORD[32+rsp],xmm2 1326 movdqa xmm8,xmm3 1347 movdqa XMMWORD[48+rsp],xmm9 1350 movdqa xmm10,xmm4 1354 movdqa xmm8,xmm4 1364 movdqa xmm9,xmm1 [all...] |
/bionic/libc/arch-x86_64/string/ |
sse4-memcmp-slm.S | 371 movdqa %xmm2, %xmm1 405 movdqa %xmm2, %xmm1 441 movdqa (%rdi), %xmm2 446 movdqa 16(%rdi), %xmm2 451 movdqa 32(%rdi), %xmm2 456 movdqa 48(%rdi), %xmm2 463 movdqa 64(%rdi), %xmm2 468 movdqa 80(%rdi), %xmm2 491 movdqa (%rdi), %xmm2 496 movdqa 16(%rdi), %xmm [all...] |
/external/boringssl/linux-x86_64/crypto/fipsmodule/ |
ghash-x86_64.S | 678 movdqa %xmm2,%xmm3 692 movdqa %xmm2,%xmm0 694 movdqa %xmm0,%xmm1 703 movdqa %xmm3,%xmm4 709 movdqa %xmm0,%xmm4 710 movdqa %xmm0,%xmm3 716 movdqa %xmm0,%xmm3 723 movdqa %xmm0,%xmm4 739 movdqa %xmm0,%xmm1 748 movdqa %xmm3,%xmm [all...] |
/external/boringssl/mac-x86_64/crypto/fipsmodule/ |
ghash-x86_64.S | 677 movdqa %xmm2,%xmm3 691 movdqa %xmm2,%xmm0 693 movdqa %xmm0,%xmm1 702 movdqa %xmm3,%xmm4 708 movdqa %xmm0,%xmm4 709 movdqa %xmm0,%xmm3 715 movdqa %xmm0,%xmm3 722 movdqa %xmm0,%xmm4 738 movdqa %xmm0,%xmm1 747 movdqa %xmm3,%xmm [all...] |
/external/libvpx/libvpx/third_party/libyuv/source/ |
scale_gcc.cc | 323 "movdqa %%xmm4,%%xmm5 \n" 461 "movdqa %0,%%xmm3 \n" 462 "movdqa %1,%%xmm4 \n" 463 "movdqa %2,%%xmm5 \n" 475 "movdqa %%xmm2,%%xmm1 \n" 497 "movdqa %0,%%xmm2 \n" // kShuf01 498 "movdqa %1,%%xmm3 \n" // kShuf11 499 "movdqa %2,%%xmm4 \n" // kShuf21 506 "movdqa %0,%%xmm5 \n" // kMadd01 507 "movdqa %1,%%xmm0 \n" // kMadd1 [all...] |
row_gcc.cc | 198 "movdqa %%xmm0,%%xmm1 \n" 221 "movdqa %3,%%xmm4 \n" 228 "movdqa %%xmm3,%%xmm2 \n" 259 "movdqa %3,%%xmm4 \n" 266 "movdqa %%xmm3,%%xmm2 \n" 295 "movdqa %3,%%xmm3 \n" 296 "movdqa %4,%%xmm4 \n" 297 "movdqa %5,%%xmm5 \n" 343 "movdqa %%xmm0,%%xmm1 \n" 344 "movdqa %%xmm0,%%xmm2 \n [all...] |
/external/libyuv/files/source/ |
row_gcc.cc | 165 "movdqa %%xmm0,%%xmm1 \n" 188 "movdqa %3,%%xmm4 \n" 195 "movdqa %%xmm3,%%xmm2 \n" 226 "movdqa %3,%%xmm4 \n" 233 "movdqa %%xmm3,%%xmm2 \n" 262 "movdqa %3,%%xmm3 \n" 263 "movdqa %4,%%xmm4 \n" 264 "movdqa %5,%%xmm5 \n" 310 "movdqa %%xmm0,%%xmm1 \n" 311 "movdqa %%xmm0,%%xmm2 \n [all...] |
scale_gcc.cc | 338 "movdqa %%xmm4,%%xmm5 \n" 483 "movdqa %0,%%xmm3 \n" 484 "movdqa %1,%%xmm4 \n" 485 "movdqa %2,%%xmm5 \n" 497 "movdqa %%xmm2,%%xmm1 \n" 520 "movdqa %0,%%xmm2 \n" // kShuf01 521 "movdqa %1,%%xmm3 \n" // kShuf11 522 "movdqa %2,%%xmm4 \n" // kShuf21 529 "movdqa %0,%%xmm5 \n" // kMadd01 530 "movdqa %1,%%xmm0 \n" // kMadd1 [all...] |
/external/libjpeg-turbo/simd/ |
jchuff-sse2.asm | 140 movdqa XMMWORD [esp + t1 + %1 * SIZEOF_WORD], %34 ; _mm_storeu_si128((__m128i *)(t1 + ko), x1); 141 movdqa XMMWORD [esp + t1 + (%1 + 8) * SIZEOF_WORD], %35 ; _mm_storeu_si128((__m128i *)(t1 + ko + 8), x1); 142 movdqa XMMWORD [esp + t1 + (%1 + 16) * SIZEOF_WORD], %36 ; _mm_storeu_si128((__m128i *)(t1 + ko + 16), x1); 143 movdqa XMMWORD [esp + t1 + (%1 + 24) * SIZEOF_WORD], %37 ; _mm_storeu_si128((__m128i *)(t1 + ko + 24), x1); 144 movdqa XMMWORD [esp + t2 + %1 * SIZEOF_WORD], xmm4 ; _mm_storeu_si128((__m128i *)(t2 + ko), neg); 145 movdqa XMMWORD [esp + t2 + (%1 + 8) * SIZEOF_WORD], xmm5 ; _mm_storeu_si128((__m128i *)(t2 + ko + 8), neg); 146 movdqa XMMWORD [esp + t2 + (%1 + 16) * SIZEOF_WORD], xmm6 ; _mm_storeu_si128((__m128i *)(t2 + ko + 16), neg); 147 movdqa XMMWORD [esp + t2 + (%1 + 24) * SIZEOF_WORD], xmm7 ; _mm_storeu_si128((__m128i *)(t2 + ko + 24), neg); 270 movdqa xmm0, XMMWORD [esp + t1 + 0 * SIZEOF_WORD] ; __m128i tmp0 = _mm_loadu_si128((__m128i *)(t1 + 0)); 271 movdqa xmm1, XMMWORD [esp + t1 + 8 * SIZEOF_WORD] ; __m128i tmp1 = _mm_loadu_si128((__m128i *)(t1 + 8)) [all...] |