/prebuilts/go/darwin-x86/src/crypto/aes/ |
asm_amd64.s | 24 MOVUPS 0(AX), X1 25 MOVUPS 0(BX), X0 32 MOVUPS 0(AX), X1 34 MOVUPS 16(AX), X1 38 MOVUPS 0(AX), X1 40 MOVUPS 16(AX), X1 44 MOVUPS 0(AX), X1 46 MOVUPS 16(AX), X1 48 MOVUPS 32(AX), X1 50 MOVUPS 48(AX), X [all...] |
/prebuilts/go/linux-x86/src/crypto/aes/ |
asm_amd64.s | 24 MOVUPS 0(AX), X1 25 MOVUPS 0(BX), X0 32 MOVUPS 0(AX), X1 34 MOVUPS 16(AX), X1 38 MOVUPS 0(AX), X1 40 MOVUPS 16(AX), X1 44 MOVUPS 0(AX), X1 46 MOVUPS 16(AX), X1 48 MOVUPS 32(AX), X1 50 MOVUPS 48(AX), X [all...] |
/external/boringssl/win-x86/crypto/aes/ |
aesni-x86.asm | 24 movups xmm2,[eax] 27 movups xmm0,[edx] 28 movups xmm1,[16+edx] 34 movups xmm1,[edx] 40 movups [eax],xmm2 49 movups xmm2,[eax] 52 movups xmm0,[edx] 53 movups xmm1,[16+edx] 59 movups xmm1,[edx] 65 movups [eax],xmm [all...] |
/external/valgrind/memcheck/tests/amd64/ |
fxsave-amd64.c | 57 asm __volatile__("movups " VG_SYM(vecZ) ", %xmm0"); 58 asm __volatile__("movups " VG_SYM(vecZ) ", %xmm1"); 59 asm __volatile__("movups " VG_SYM(vecZ) ", %xmm2"); 60 asm __volatile__("movups " VG_SYM(vecZ) ", %xmm3"); 61 asm __volatile__("movups " VG_SYM(vecZ) ", %xmm4"); 62 asm __volatile__("movups " VG_SYM(vecZ) ", %xmm5"); 63 asm __volatile__("movups " VG_SYM(vecZ) ", %xmm6"); 64 asm __volatile__("movups " VG_SYM(vecZ) ", %xmm7"); 65 asm __volatile__("movups " VG_SYM(vecZ) ", %xmm8"); 66 asm __volatile__("movups " VG_SYM(vecZ) ", %xmm9") [all...] |
sh-mem-vec128.c | 12 "movups (%1), %%xmm7 ; movups %%xmm7, (%0)"
|
xor-undef-amd64.c | 66 "movups 0(%0), %%xmm0\n\t" 67 "movups 16(%0), %%xmm8\n\t" 69 "movups %%xmm0, 32(%0)\n\t" 78 "movups 0(%0), %%xmm0\n\t" 79 "movups 16(%0), %%xmm8\n\t" 81 "movups %%xmm0, 32(%0)\n\t" 93 "movups 0(%0), %%xmm0\n\t" 94 "movups 16(%0), %%xmm8\n\t" 96 "movups %%xmm0, 32(%0)\n\t" 105 "movups 0(%0), %%xmm0\n\t [all...] |
/external/valgrind/memcheck/tests/amd64-solaris/ |
context_sse.c | 59 "movups %[y0], %%xmm0\n" 60 "movups %[d0], %%xmm1\n" 61 "movups %[d0], %%xmm2\n" 62 "movups %[y0], %%xmm3\n" 63 "movups %[y0], %%xmm4\n" 64 "movups %[d0], %%xmm5\n" 65 "movups %[d0], %%xmm6\n" 66 "movups %[y0], %%xmm7\n" 70 "movups %%xmm0, 0x00 + %[out]\n" 71 "movups %%xmm1, 0x10 + %[out]\n [all...] |
/external/valgrind/memcheck/tests/x86-solaris/ |
context_sse.c | 57 "movups %[y0], %%xmm0\n" 58 "movups %[d0], %%xmm1\n" 59 "movups %[d0], %%xmm2\n" 60 "movups %[y0], %%xmm3\n" 61 "movups %[y0], %%xmm4\n" 62 "movups %[d0], %%xmm5\n" 63 "movups %[d0], %%xmm6\n" 64 "movups %[y0], %%xmm7\n" 75 "movups %%xmm0, 0x00 + %[out]\n" 76 "movups %%xmm1, 0x10 + %[out]\n [all...] |
/external/valgrind/memcheck/tests/x86/ |
sh-mem-vec128.c | 12 "movups (%1), %%xmm7 ; movups %%xmm7, (%0)"
|
xor-undef-x86.c | 68 "movups 0(%0), %%xmm0\n\t" 69 "movups 16(%0), %%xmm7\n\t" 71 "movups %%xmm0, 32(%0)\n\t" 82 "movups 0(%0), %%xmm0\n\t" 83 "movups 16(%0), %%xmm7\n\t" 85 "movups %%xmm0, 32(%0)\n\t" 99 "movups 0(%0), %%xmm0\n\t" 100 "movups 16(%0), %%xmm7\n\t" 102 "movups %%xmm0, 32(%0)\n\t" 113 "movups 0(%0), %%xmm0\n\t [all...] |
fxsave.c | 38 asm __volatile__("movups " VG_SYM(vecZ) ", %xmm0"); 39 asm __volatile__("movups " VG_SYM(vecZ) ", %xmm1"); 40 asm __volatile__("movups " VG_SYM(vecZ) ", %xmm2"); 41 asm __volatile__("movups " VG_SYM(vecZ) ", %xmm3"); 42 asm __volatile__("movups " VG_SYM(vecZ) ", %xmm4"); 43 asm __volatile__("movups " VG_SYM(vecZ) ", %xmm5"); 44 asm __volatile__("movups " VG_SYM(vecZ) ", %xmm6"); 45 asm __volatile__("movups " VG_SYM(vecZ) ", %xmm7"); 62 asm __volatile__("movups " VG_SYM(vec0) ", %xmm0"); 63 asm __volatile__("movups " VG_SYM(vec1) ", %xmm1") [all...] |
/external/boringssl/linux-x86/crypto/aes/ |
aesni-x86.S | 12 movups (%eax),%xmm2 15 movups (%edx),%xmm0 16 movups 16(%edx),%xmm1 22 movups (%edx),%xmm1 28 movups %xmm2,(%eax) 40 movups (%eax),%xmm2 43 movups (%edx),%xmm0 44 movups 16(%edx),%xmm1 50 movups (%edx),%xmm1 56 movups %xmm2,(%eax [all...] |
/external/boringssl/mac-x86/crypto/aes/ |
aesni-x86.S | 11 movups (%eax),%xmm2 14 movups (%edx),%xmm0 15 movups 16(%edx),%xmm1 21 movups (%edx),%xmm1 27 movups %xmm2,(%eax) 37 movups (%eax),%xmm2 40 movups (%edx),%xmm0 41 movups 16(%edx),%xmm1 47 movups (%edx),%xmm1 53 movups %xmm2,(%eax [all...] |
/external/boringssl/linux-x86_64/crypto/aes/ |
aesni-x86_64.S | 10 movups (%rdi),%xmm2 12 movups (%rdx),%xmm0 13 movups 16(%rdx),%xmm1 19 movups (%rdx),%xmm1 25 movups %xmm2,(%rsi) 35 movups (%rdi),%xmm2 37 movups (%rdx),%xmm0 38 movups 16(%rdx),%xmm1 44 movups (%rdx),%xmm1 50 movups %xmm2,(%rsi [all...] |
/external/boringssl/mac-x86_64/crypto/aes/ |
aesni-x86_64.S | 9 movups (%rdi),%xmm2 11 movups (%rdx),%xmm0 12 movups 16(%rdx),%xmm1 18 movups (%rdx),%xmm1 24 movups %xmm2,(%rsi) 34 movups (%rdi),%xmm2 36 movups (%rdx),%xmm0 37 movups 16(%rdx),%xmm1 43 movups (%rdx),%xmm1 49 movups %xmm2,(%rsi [all...] |
/external/boringssl/win-x86_64/crypto/aes/ |
aesni-x86_64.asm | 12 movups xmm2,XMMWORD[rcx] 14 movups xmm0,XMMWORD[r8] 15 movups xmm1,XMMWORD[16+r8] 21 movups xmm1,XMMWORD[r8] 27 movups XMMWORD[rdx],xmm2 36 movups xmm2,XMMWORD[rcx] 38 movups xmm0,XMMWORD[r8] 39 movups xmm1,XMMWORD[16+r8] 45 movups xmm1,XMMWORD[r8] 51 movups XMMWORD[rdx],xmm [all...] |
/external/llvm/test/CodeGen/X86/ |
2008-05-22-FoldUnalignedLoad.ll | 12 ; CHECK: movups 13 ; CHECK: movups 14 ; CHECK-NOT: movups
|
small-byval-memcpy.ll | 18 ; NEHALEM: movups 19 ; NEHALEM-NEXT: movups 22 ; BTVER2: movups 23 ; BTVER2-NEXT: movups
|
sse-align-10.ll | 1 ; RUN: llc < %s -march=x86-64 | grep movups | count 1
|
sse-align-8.ll | 1 ; RUN: llc < %s -march=x86-64 | grep movups | count 1
|
2009-11-16-UnfoldMemOpBug.ll | 9 ; CHECK: movups L_str+12(%rip), %xmm0 10 ; CHECK: movups L_str(%rip), %xmm1 17 ; CHECK: movups %xmm0, 12(%rsp)
|
musttail-fastcall.ll | 57 ; SSE2-DAG: movups %xmm0, {{.*}} 58 ; SSE2-DAG: movups %xmm1, {{.*}} 59 ; SSE2-DAG: movups %xmm2, {{.*}} 60 ; SSE2-DAG: movups %xmm3, {{.*}} 61 ; SSE2-DAG: movups %xmm4, {{.*}} 62 ; SSE2-DAG: movups %xmm5, {{.*}} 80 ; SSE2-DAG: movups {{.*}}, %xmm0 81 ; SSE2-DAG: movups {{.*}}, %xmm1 82 ; SSE2-DAG: movups {{.*}}, %xmm2 83 ; SSE2-DAG: movups {{.*}}, %xmm [all...] |
/external/valgrind/none/tests/x86-solaris/ |
coredump_single_thread_sse.c | 19 "movups (%%esp), %%xmm0\n" 24 "movups (%%esp), %%xmm1\n" 29 "movups (%%esp), %%xmm2\n" 34 "movups (%%esp), %%xmm3\n" 39 "movups (%%esp), %%xmm4\n" 44 "movups (%%esp), %%xmm5\n" 49 "movups (%%esp), %%xmm6\n" 54 "movups (%%esp), %%xmm7\n"
|
/external/valgrind/VEX/test/ |
fxsave.c | 37 asm __volatile__("movups vecZ, %xmm0"); 38 asm __volatile__("movups vecZ, %xmm1"); 39 asm __volatile__("movups vecZ, %xmm2"); 40 asm __volatile__("movups vecZ, %xmm3"); 41 asm __volatile__("movups vecZ, %xmm4"); 42 asm __volatile__("movups vecZ, %xmm5"); 43 asm __volatile__("movups vecZ, %xmm6"); 44 asm __volatile__("movups vecZ, %xmm7"); 61 asm __volatile__("movups vec0, %xmm0"); 62 asm __volatile__("movups vec1, %xmm1") [all...] |
/external/boringssl/src/crypto/aes/asm/ |
aesni-x86.pl | 71 if ($PREFIX eq "aesni") { $movekey=\&movups; } 72 else { $movekey=\&movups; } 134 &movups ($rndkey0,&QWP(0,$key)); 182 &movups ($inout0,&QWP(0,"eax")); 191 &movups (&QWP(0,"eax"),$inout0); 201 &movups ($inout0,&QWP(0,"eax")); 210 &movups (&QWP(0,"eax"),$inout0); 445 &movups (&QWP(0,$out),$inout0); 447 &movups (&QWP(0x10,$out),$inout1); 449 &movups (&QWP(0x20,$out),$inout2) [all...] |