/art/runtime/arch/x86_64/ |
jni_entrypoints_x86_64.S | 31 subq LITERAL(72 + 4 * 8), %rsp 34 movq %xmm0, 0(%rsp) 35 movq %xmm1, 8(%rsp) 36 movq %xmm2, 16(%rsp) 37 movq %xmm3, 24(%rsp) 38 movq %xmm4, 32(%rsp) 39 movq %xmm5, 40(%rsp) 40 movq %xmm6, 48(%rsp) 41 movq %xmm7, 56(%rsp) 42 movq %xmm12, 64(%rsp) [all...] |
quick_entrypoints_x86_64.S | 21 subq MACRO_LITERAL(4 * 8), %rsp 23 movq %xmm12, 0(%rsp) 24 movq %xmm13, 8(%rsp) 25 movq %xmm14, 16(%rsp) 26 movq %xmm15, 24(%rsp) 31 movq 0(%rsp), %xmm12 32 movq 8(%rsp), %xmm13 33 movq 16(%rsp), %xmm14 34 movq 24(%rsp), %xmm15 35 addq MACRO_LITERAL(4 * 8), %rsp [all...] |
/external/llvm/test/MC/ELF/ |
cfi-adjust-cfa-offset.s | 5 subq $8, %rsp 9 addq $8, %rsp
|
cfi-def-cfa-offset.s | 5 subq $8, %rsp 8 addq $8, %rsp
|
debug-line.s | 40 subq $8, %rsp
|
/external/llvm/test/Instrumentation/AddressSanitizer/X86/ |
asm_mov.s | 9 # CHECK: leaq -128(%rsp), %rsp 11 # CHECK: leaq 128(%rsp), %rsp 15 # CHECK-NEXT: leaq -128(%rsp), %rsp 17 # CHECK: leaq 128(%rsp), %rsp 38 # CHECK: leaq -128(%rsp), %rsp [all...] |
asm_swap_intel.s | 9 # CHECK: leaq -128(%rsp), %rsp 11 # CHECK: leaq 128(%rsp), %rsp 15 # CHECK-NEXT: leaq -128(%rsp), %rsp 17 # CHECK: leaq 128(%rsp), %rsp 21 # CHECK-NEXT: leaq -128(%rsp), %rsp [all...] |
/external/llvm/test/MC/X86/ |
intel-syntax-2.s | 5 // CHECK: movl $257, -4(%rsp) 6 mov DWORD PTR [RSP - 4], 257 8 // CHECK: movl $257, -4(%rsp) 9 movl $257, -4(%rsp)
|
/external/llvm/test/tools/llvm-objdump/Inputs/ |
win64-unwind.exe.coff-x86_64.asm | 7 subq $24, %rsp 9 movq %rsi, 16(%rsp) 11 movups %xmm8, (%rsp) 15 mov %rsp, %rbx 25 lea (%rbx), %rsp 27 addq $24, %rsp 45 subq $65520, %rsp 47 sub $8454128, %rsp 50 add $8454128, %rsp 51 addq $65520, %rsp [all...] |
/external/openssl/crypto/bn/asm/ |
x86_64-gf2m.S | 6 subq $128+8,%rsp 36 movq $0,0(%rsp) 38 movq %r9,8(%rsp) 40 movq %r10,16(%rsp) 42 movq %r13,24(%rsp) 45 movq %r11,32(%rsp) 47 movq %r9,40(%rsp) 49 movq %r10,48(%rsp) 51 movq %r13,56(%rsp) 54 movq %r12,64(%rsp) [all...] |
x86_64-gf2m.pl | 50 sub \$128+8,%rsp 80 movq \$0,0(%rsp) # tab[0]=0 82 mov $a1,8(%rsp) # tab[1]=a1 84 mov $a2,16(%rsp) # tab[2]=a2 86 mov $a12,24(%rsp) # tab[3]=a1^a2 89 mov $a4,32(%rsp) # tab[4]=a4 91 mov $a1,40(%rsp) # tab[5]=a1^a4 93 mov $a2,48(%rsp) # tab[6]=a2^a4 95 mov $a12,56(%rsp) # tab[7]=a1^a2^a4 98 mov $a8,64(%rsp) # tab[8]=a [all...] |
/external/llvm/test/MC/COFF/ |
seh.s | 129 subq $24, %rsp 131 movq %rsi, 16(%rsp) 133 movups %xmm8, (%rsp) 137 mov %rsp, %rbx 147 lea (%rbx), %rsp 149 addq $24, %rsp
|
seh-align2.s | 68 subq $24, %rsp 76 addq $24, %rsp
|
seh-section.s | 28 subq $8, %rsp 34 addq $8, %rsp
|
/bionic/linker/arch/x86_64/ |
begin.S | 32 mov %rsp, %rdi
|
/external/chromium_org/third_party/boringssl/win-x86_64/crypto/aes/ |
vpaes-x86_64.asm | 612 mov QWORD PTR[8+rsp],rdi ;WIN64 prologue 613 mov QWORD PTR[16+rsp],rsi 614 mov rax,rsp 621 lea rsp,QWORD PTR[((-184))+rsp] 622 movaps XMMWORD PTR[16+rsp],xmm6 623 movaps XMMWORD PTR[32+rsp],xmm7 624 movaps XMMWORD PTR[48+rsp],xmm8 625 movaps XMMWORD PTR[64+rsp],xmm9 626 movaps XMMWORD PTR[80+rsp],xmm1 [all...] |
/external/chromium_org/third_party/libvpx/source/libvpx/vp9/common/x86/ |
vp9_subpixel_8t_sse2.asm | 191 mov rbp, rsp 200 sub rsp, 16 * 6 201 %define k0k1 [rsp + 16 * 0] 202 %define k2k3 [rsp + 16 * 1] 203 %define k5k4 [rsp + 16 * 2] 204 %define k6k7 [rsp + 16 * 3] 205 %define krd [rsp + 16 * 4] 206 %define zero [rsp + 16 * 5] 235 add rsp, 16 * 6 236 pop rsp [all...] |
vp9_high_subpixel_8t_sse2.asm | 212 mov rbp, rsp 221 sub rsp, 16 * 7 222 %define k0k6 [rsp + 16 * 0] 223 %define k2k5 [rsp + 16 * 1] 224 %define k3k4 [rsp + 16 * 2] 225 %define k1k7 [rsp + 16 * 3] 226 %define krd [rsp + 16 * 4] 227 %define max [rsp + 16 * 5] 228 %define min [rsp + 16 * 6] 259 add rsp, 16 * [all...] |
/external/libvpx/libvpx/vp9/common/x86/ |
vp9_subpixel_8t_sse2.asm | 191 mov rbp, rsp 200 sub rsp, 16 * 6 201 %define k0k1 [rsp + 16 * 0] 202 %define k2k3 [rsp + 16 * 1] 203 %define k5k4 [rsp + 16 * 2] 204 %define k6k7 [rsp + 16 * 3] 205 %define krd [rsp + 16 * 4] 206 %define zero [rsp + 16 * 5] 235 add rsp, 16 * 6 236 pop rsp [all...] |
/external/openssl/crypto/sha/asm/ |
sha1-x86_64.S | 20 movq %rsp,%r11 22 subq $72,%rsp 24 andq $-64,%rsp 26 movq %r11,64(%rsp) 40 movl %edx,0(%rsp) 49 movl %ebp,4(%rsp) 62 movl %edx,8(%rsp) 75 movl %ebp,12(%rsp) 88 movl %edx,16(%rsp) 101 movl %ebp,20(%rsp) [all...] |
/external/llvm/lib/Target/X86/ |
X86CompilationCallback_Win64.asm | 20 ; Save RSP. 21 mov rbp, rsp 31 and rsp, -16 34 sub rsp, 96 35 movaps [rsp +32], xmm0 36 movaps [rsp+16+32], xmm1 37 movaps [rsp+32+32], xmm2 38 movaps [rsp+48+32], xmm3 48 movaps xmm3, [rsp+48+32] 49 movaps xmm2, [rsp+32+32 [all...] |
/external/chromium_org/third_party/libvpx/source/libvpx/vpx_ports/ |
emms.asm | 24 sub rsp, 8 25 mov [rsp], rcx ; win x64 specific 26 fldcw [rsp] 27 add rsp, 8 33 sub rsp, 8 34 fstcw [rsp] 35 mov rax, [rsp] 36 add rsp, 8
|
/external/libvpx/libvpx/vpx_ports/ |
emms.asm | 24 sub rsp, 8 25 mov [rsp], rcx ; win x64 specific 26 fldcw [rsp] 27 add rsp, 8 33 sub rsp, 8 34 fstcw [rsp] 35 mov rax, [rsp] 36 add rsp, 8
|
/external/compiler-rt/lib/builtins/x86_64/ |
floatundixf.S | 27 movq %rdi, -8(%rsp) 28 fildq -8(%rsp) 68 movq %rdi, -8(%rsp) 69 movq %rsi, -16(%rsp) 71 faddl -8(%rsp) // hi - 2^52 (as double extended, no rounding occurs) 72 faddl -16(%rsp) // hi + lo (as double extended)
|
/external/chromium_org/third_party/boringssl/linux-x86_64/crypto/sha/ |
sha1-x86_64.S | 19 movq %rsp,%rax 26 subq $72,%rsp 28 andq $-64,%rsp 30 movq %rax,64(%rsp) 46 movl %edx,0(%rsp) 59 movl %ebp,4(%rsp) 72 movl %r14d,8(%rsp) 85 movl %edx,12(%rsp) 98 movl %ebp,16(%rsp) 111 movl %r14d,20(%rsp) [all...] |