/external/swiftshader/third_party/LLVM/test/CodeGen/X86/ |
2009-02-05-CoalescerBug.ll | 4 ; RUN: grep movdqa %t | count 2
|
dagcombine-buildvector.ll | 17 ; CHECK: movdqa
|
/external/boringssl/src/crypto/fipsmodule/bn/asm/ |
x86_64-mont5.pl | 146 movdqa 0(%r10),%xmm0 # 00000001000000010000000000000000 147 movdqa 16(%r10),%xmm1 # 00000002000000020000000200000002 152 movdqa %xmm1,%xmm4 153 movdqa %xmm1,%xmm2 162 movdqa %xmm4,%xmm3 168 movdqa %xmm0,`16*($k+0)+112`(%r10) 169 movdqa %xmm4,%xmm0 173 movdqa %xmm1,`16*($k+1)+112`(%r10) 174 movdqa %xmm4,%xmm1 178 movdqa %xmm2,`16*($k+2)+112`(%r10 [all...] |
/external/llvm/test/CodeGen/X86/ |
vector-lzcnt-128.ll | 30 ; SSE2-NEXT: movdqa %xmm1, %xmm0 48 ; SSE3-NEXT: movdqa %xmm1, %xmm0 66 ; SSSE3-NEXT: movdqa %xmm1, %xmm0 160 ; SSE2-NEXT: movdqa %xmm1, %xmm0 175 ; SSE3-NEXT: movdqa %xmm1, %xmm0 190 ; SSSE3-NEXT: movdqa %xmm1, %xmm0 290 ; SSE2-NEXT: movdqa %xmm1, %xmm0 322 ; SSE3-NEXT: movdqa %xmm1, %xmm0 354 ; SSSE3-NEXT: movdqa %xmm1, %xmm0 380 ; SSE41-NEXT: movdqa %xmm1, %xmm [all...] |
vec_cmp_sint-128.ll | 231 ; SSE2-NEXT: movdqa {{.*#+}} xmm2 = [2147483648,0,2147483648,0] 234 ; SSE2-NEXT: movdqa %xmm1, %xmm2 248 ; SSE41-NEXT: movdqa {{.*#+}} xmm2 = [2147483648,0,2147483648,0] 251 ; SSE41-NEXT: movdqa %xmm1, %xmm2 365 ; SSE2-NEXT: movdqa {{.*#+}} xmm2 = [2147483648,0,2147483648,0] 368 ; SSE2-NEXT: movdqa %xmm0, %xmm2 380 ; SSE41-NEXT: movdqa {{.*#+}} xmm2 = [2147483648,0,2147483648,0] 383 ; SSE41-NEXT: movdqa %xmm0, %xmm2 479 ; SSE2-NEXT: movdqa {{.*#+}} xmm2 = [2147483648,0,2147483648,0] 482 ; SSE2-NEXT: movdqa %xmm0, %xmm [all...] |
vector-idiv-udiv-128.ll | 33 ; SSE2-NEXT: movdqa %xmm1, %xmm0 86 ; SSE2-NEXT: movdqa {{.*#+}} xmm1 = [613566757,613566757,613566757,613566757] 87 ; SSE2-NEXT: movdqa %xmm0, %xmm2 103 ; SSE41-NEXT: movdqa {{.*#+}} xmm1 = [613566757,613566757,613566757,613566757] 152 ; SSE-NEXT: movdqa {{.*#+}} xmm1 = [9363,9363,9363,9363,9363,9363,9363,9363] 175 ; SSE2-NEXT: movdqa {{.*#+}} xmm1 = [37,37,37,37,37,37,37,37,37,37,37,37,37,37,37,37] 177 ; SSE2-NEXT: movdqa %xmm0, %xmm2 182 ; SSE2-NEXT: movdqa %xmm0, %xmm3 288 ; SSE2-NEXT: movdqa %xmm1, %xmm0 357 ; SSE2-NEXT: movdqa {{.*#+}} xmm1 = [613566757,613566757,613566757,613566757 [all...] |
shrink_vmul.ll | 96 ; CHECK-NEXT: movdqa %xmm1, %xmm0 132 ; CHECK-NEXT: movdqa %xmm0, %xmm3 134 ; CHECK-NEXT: movdqa %xmm1, %xmm4 137 ; CHECK-NEXT: movdqa %xmm4, %xmm3 143 ; CHECK-NEXT: movdqa %xmm1, %xmm0 180 ; CHECK-NEXT: movdqa %xmm1, %xmm2 215 ; CHECK-NEXT: movdqa %xmm1, %xmm2 250 ; CHECK-NEXT: movdqa %xmm1, %xmm2 253 ; CHECK-NEXT: movdqa %xmm1, %xmm0 290 ; CHECK-NEXT: movdqa %xmm2, %xmm [all...] |
unaligned-spill-folding.ll | 42 ; ALIGNED: movdqa {{.*}} # 16-byte Spill 47 ; FORCEALIGNED: movdqa {{.*}} # 16-byte Spill
|
reduce-trunc-shl.ll | 12 ; SSE2-NEXT: movdqa %xmm1, (%rdi)
|
/bionic/libc/arch-x86/silvermont/string/ |
sse2-memmove-slm.S | 184 movdqa %xmm4, (%edi) 207 movdqa %xmm0, (%edi) 414 movdqa %xmm4, -16(%edi) 415 movdqa %xmm5, -32(%edi) 416 movdqa %xmm6, -48(%edi) 417 movdqa %xmm7, -64(%edi) 438 movdqa %xmm0, -64(%edi) 439 movdqa %xmm1, -48(%edi) 440 movdqa %xmm2, -32(%edi) 441 movdqa %xmm3, -16(%edi [all...] |
/external/boringssl/src/crypto/fipsmodule/modes/asm/ |
ghash-x86.pl | 617 &movdqa ($Xhi,$Xi); # 630 &movdqa ($T2,$T1); # 645 &movdqa ($T1,$Xi); # 646 &movdqa ($Xhi,$Xi); 657 &movdqa ($T3,$T2); # 678 &movdqa ($T2,$Xi); # 679 &movdqa ($T1,$Xi); 685 &movdqa ($T1,$Xi); # 692 &movdqa ($T2,$Xi); 716 &movdqa ($T1,$Hkey) [all...] |
/external/boringssl/src/crypto/fipsmodule/aes/asm/ |
aesni-x86_64.pl | 988 movdqa .Lincrement64(%rip),$increment 989 movdqa .Lbswap_mask(%rip),$bswap_mask 995 movdqa $iv,$inout0 1030 movdqa $iv,$inout0 [all...] |
/external/boringssl/src/crypto/fipsmodule/sha/asm/ |
sha512-586.pl | 476 &movdqa (@X[1],&QWP(80*8,$K512)); # byte swap mask 480 &movdqa (&QWP(16*(($j-1)%4),$frame),@X[3]) if ($j>4); # off-load 481 &movdqa (@X[3],&QWP(16*($j%8),$K512)); 482 &movdqa (@X[2],@X[1]) if ($j<7); # perpetuate byte swap mask 484 &movdqa (@X[1],&QWP(16*(($j+1)%4),$frame)) if ($j==7);# restore @X[0] 487 &movdqa (&QWP(16*($j%8)-128,$frame),@X[3]); # xfer X[i]+K[i] 495 &movdqa (@X[2],&QWP(16*(($j+1)%4),$frame)); # pre-restore @X[1] 496 &movdqa (&QWP(16*(($j-1)%4),$frame),@X[3]); # off-load @X[3] 577 &movdqa ($t2,@X[5]); 578 &movdqa (@X[1],$t0); # restore @X[1 [all...] |
sha256-586.pl | 538 &movdqa ($TMP,&QWP(0x100-0x80,$K256)); # byte swap mask 553 &movdqa (&QWP(16,"esp"),$CDGH); # offload 555 &movdqa ($Wi,&QWP(0*16-0x80,$K256)); 561 &movdqa (&QWP(0,"esp"),$ABEF); # offload 564 &movdqa ($Wi,&QWP(1*16-0x80,$K256)); 573 &movdqa ($Wi,&QWP(2*16-0x80,$K256)); 578 &movdqa ($TMP,@MSG[3]); 585 &movdqa ($Wi,&QWP(3*16-0x80,$K256)); 590 &movdqa ($TMP,@MSG[0]); 598 &movdqa ($Wi,&QWP($i*16-0x80,$K256)) [all...] |
/external/boringssl/linux-x86_64/crypto/fipsmodule/ |
sha1-x86_64.S | 1271 movdqa 64(%r14),%xmm6 1272 movdqa -64(%r14),%xmm9 1285 movdqa %xmm0,0(%rsp) 1287 movdqa %xmm1,16(%rsp) 1289 movdqa %xmm2,32(%rsp) 1297 movdqa %xmm3,%xmm8 1318 movdqa %xmm9,48(%rsp) 1321 movdqa %xmm4,%xmm10 1325 movdqa %xmm4,%xmm8 1335 movdqa %xmm10,%xmm [all...] |
/external/boringssl/mac-x86_64/crypto/fipsmodule/ |
sha1-x86_64.S | 1270 movdqa 64(%r14),%xmm6 1271 movdqa -64(%r14),%xmm9 1284 movdqa %xmm0,0(%rsp) 1286 movdqa %xmm1,16(%rsp) 1288 movdqa %xmm2,32(%rsp) 1296 movdqa %xmm3,%xmm8 1317 movdqa %xmm9,48(%rsp) 1320 movdqa %xmm4,%xmm10 1324 movdqa %xmm4,%xmm8 1334 movdqa %xmm10,%xmm [all...] |
/device/linaro/bootloader/edk2/MdePkg/Library/BaseMemoryLibOptDxe/X64/ |
CopyMem.asm | 58 movdqa [rsp + 18h], xmm0 ; save xmm0 on stack
66 movdqa xmm0, [rsp + 18h] ; restore xmm0
|
CopyMem.nasm | 62 movdqa [rsp + 0x18], xmm0 ; save xmm0 on stack
70 movdqa xmm0, [rsp + 0x18] ; restore xmm0
|
/device/linaro/bootloader/edk2/MdePkg/Library/BaseMemoryLibSse2/X64/ |
CopyMem.asm | 58 movdqa [rsp + 18h], xmm0 ; save xmm0 on stack
66 movdqa xmm0, [rsp + 18h] ; restore xmm0
|
CopyMem.nasm | 62 movdqa [rsp + 0x18], xmm0 ; save xmm0 on stack
70 movdqa xmm0, [rsp + 0x18] ; restore xmm0
|
SetMem.S | 57 movdqa %xmm0, 0x10(%rsp) # save xmm0
66 movdqa 0x10(%rsp), %xmm0 # restore xmm0
|
SetMem.asm | 52 movdqa [rsp + 10h], xmm0 ; save xmm0
61 movdqa xmm0, [rsp + 10h] ; restore xmm0
|
SetMem.nasm | 55 movdqa [rsp + 0x10], xmm0 ; save xmm0
64 movdqa xmm0, [rsp + 0x10] ; restore xmm0
|
/external/libjpeg-turbo/simd/ |
jchuff-sse2-64.asm | 153 movdqa XMMWORD [t1 + %1 * SIZEOF_WORD], %34 ; _mm_storeu_si128((__m128i *)(t1 + ko), x1); 154 movdqa XMMWORD [t1 + (%1 + 8) * SIZEOF_WORD], %35 ; _mm_storeu_si128((__m128i *)(t1 + ko + 8), x1); 155 movdqa XMMWORD [t1 + (%1 + 16) * SIZEOF_WORD], %36 ; _mm_storeu_si128((__m128i *)(t1 + ko + 16), x1); 156 movdqa XMMWORD [t1 + (%1 + 24) * SIZEOF_WORD], %37 ; _mm_storeu_si128((__m128i *)(t1 + ko + 24), x1); 157 movdqa XMMWORD [t2 + %1 * SIZEOF_WORD], xmm8 ; _mm_storeu_si128((__m128i *)(t2 + ko), neg); 158 movdqa XMMWORD [t2 + (%1 + 8) * SIZEOF_WORD], xmm9 ; _mm_storeu_si128((__m128i *)(t2 + ko + 8), neg); 159 movdqa XMMWORD [t2 + (%1 + 16) * SIZEOF_WORD], xmm10 ; _mm_storeu_si128((__m128i *)(t2 + ko + 16), neg); 160 movdqa XMMWORD [t2 + (%1 + 24) * SIZEOF_WORD], xmm11 ; _mm_storeu_si128((__m128i *)(t2 + ko + 24), neg);
|
/external/boringssl/src/util/fipstools/testdata/x86_64-Basic/ |
in.s | 23 movdqa %xmm3,%xmm10
|