/external/valgrind/none/tests/amd64/ |
loopnel.c | 6 long rcx = 0x200000005UL; local 8 asm volatile ("1: addq $1, %0; loopnel 1b" : "+a" (rax), "+c" (rcx) : : "cc"); 9 printf ("%ld %ld\n", rax, rcx);
|
/toolchain/binutils/binutils-2.27/gas/testsuite/gas/i386/ |
rexw.s | 5 extractps $0x0,%xmm0,%rcx 6 movmskpd %xmm2,%rcx 7 movmskps %xmm2,%rcx 8 pextrb $0x0,%xmm0,%rcx 9 pextrw $0x0,%mm0,%rcx 10 pextrw $0x0,%xmm0,%rcx 11 pinsrb $0x0,%rcx,%xmm0 12 pinsrw $0x0,%rcx,%xmm0 13 pinsrw $0x0,%rcx,%mm0 18 vextractps $0x0,%xmm0,%rcx [all...] |
x86-64-inval-ept.s | 4 invept (%rcx), %bx 5 invept (%rcx), %ebx 6 invept %rbx, (%rcx) 7 invept %rbx, %rcx 8 invvpid (%rcx), %bx 9 invvpid (%rcx), %ebx 10 invvpid %rbx, (%rcx) 11 invvpid %rbx, %rcx 14 invept bx, oword ptr [rcx] 15 invept ebx, oword ptr [rcx] [all...] |
x86-64-ept.s | 4 invept (%rcx), %rbx 5 invept (%rcx), %r11 6 invvpid (%rcx), %rbx 7 invvpid (%rcx), %r11 10 invept rbx, oword ptr [rcx] 11 invept r11, oword ptr [rcx] 12 invvpid rbx, oword ptr [rcx] 13 invvpid r11, oword ptr [rcx]
|
x86-64-inval-invpcid.s | 4 invvpid (%rcx), %bx 5 invvpid (%rcx), %ebx 6 invvpid %rbx, (%rcx) 7 invvpid %rbx, %rcx 10 invvpid bx, [rcx] 11 invvpid ebx, [rcx] 12 invvpid [rcx], rbx 13 invvpid rcx, rbx
|
x86-64-inval-avx.s | 4 vcvtpd2dq (%rcx),%xmm2 5 vcvtpd2ps (%rcx),%xmm2 6 vcvttpd2dq (%rcx),%xmm2 9 vcvtpd2dq xmm2,[rcx] 10 vcvtpd2ps xmm2,[rcx] 11 vcvttpd2dq xmm2,[rcx]
|
x86-64-amdfam10.s | 5 lzcnt (%rcx),%rbx 6 lzcnt (%rcx),%ebx 7 lzcnt (%rcx),%bx 8 lzcnt %rcx,%rbx 11 popcnt (%rcx),%rbx 12 popcnt (%rcx),%ebx 13 popcnt (%rcx),%bx 14 popcnt %rcx,%rbx 21 movntsd %xmm1,(%rcx) 22 movntss %xmm1,(%rcx) [all...] |
x86-64-hle.s | 11 xacquire lock adcb $100,(%rcx) 12 lock xacquire adcb $100,(%rcx) 13 xrelease lock adcb $100,(%rcx) 14 lock xrelease adcb $100,(%rcx) 15 .byte 0xf0; .byte 0xf2; adcb $100,(%rcx) 16 .byte 0xf0; .byte 0xf3; adcb $100,(%rcx) 17 xacquire lock addb $100,(%rcx) 18 lock xacquire addb $100,(%rcx) 19 xrelease lock addb $100,(%rcx) 20 lock xrelease addb $100,(%rcx) [all...] |
x86-64-inval-avx.l | 14 [ ]*4[ ]+vcvtpd2dq \(%rcx\),%xmm2 15 [ ]*5[ ]+vcvtpd2ps \(%rcx\),%xmm2 16 [ ]*6[ ]+vcvttpd2dq \(%rcx\),%xmm2 19 [ ]*9[ ]+vcvtpd2dq xmm2,\[rcx\] 20 [ ]*10[ ]+vcvtpd2ps xmm2,\[rcx\] 21 [ ]*11[ ]+vcvttpd2dq xmm2,\[rcx\]
|
x86-64-inval-ept.l | 24 [ ]*4[ ]+invept \(%rcx\), %bx 25 [ ]*5[ ]+invept \(%rcx\), %ebx 26 [ ]*6[ ]+invept %rbx, \(%rcx\) 27 [ ]*7[ ]+invept %rbx, %rcx 28 [ ]*8[ ]+invvpid \(%rcx\), %bx 29 [ ]*9[ ]+invvpid \(%rcx\), %ebx 30 [ ]*10[ ]+invvpid %rbx, \(%rcx\) 31 [ ]*11[ ]+invvpid %rbx, %rcx 34 [ ]*14[ ]+invept bx, oword ptr \[rcx\] 35 [ ]*15[ ]+invept ebx, oword ptr \[rcx\] [all...] |
x86-64-inval-movbe.s | 4 movbe (%rcx),%bl 6 movbe %bx,%rcx 7 movbe %rbx,%rcx 8 movbe %bl,(%rcx) 11 movbe bl, byte ptr [rcx] 13 movbe rcx, bx 14 movbe rcx, rbx 15 movbe byte ptr [rcx], bl
|
x86-64-ssse3.s | 5 phaddw (%rcx),%mm0 7 phaddw (%rcx),%xmm0 9 phaddd (%rcx),%mm0 11 phaddd (%rcx),%xmm0 13 phaddsw (%rcx),%mm0 15 phaddsw (%rcx),%xmm0 17 phsubw (%rcx),%mm0 19 phsubw (%rcx),%xmm0 21 phsubd (%rcx),%mm0 23 phsubd (%rcx),%xmm [all...] |
x86-64-bmi.s | 9 tzcnt (%rcx),%bx 10 tzcnt (%rcx),%r15w 14 andn (%rcx),%ebx,%esi 16 andn (%rcx),%r15d,%r10d 20 bextr %ebx,(%rcx),%esi 22 bextr %r9d,(%rcx),%r10d 26 tzcnt (%rcx),%ebx 27 tzcnt (%rcx),%r15d 29 blsi (%rcx),%ebx 30 blsi (%rcx),%r15 [all...] |
x86-64-ept.d | 9 [ ]*[a-f0-9]+: 66 0f 38 80 19 invept \(%rcx\),%rbx 10 [ ]*[a-f0-9]+: 66 44 0f 38 80 19 invept \(%rcx\),%r11 11 [ ]*[a-f0-9]+: 66 0f 38 81 19 invvpid \(%rcx\),%rbx 12 [ ]*[a-f0-9]+: 66 44 0f 38 81 19 invvpid \(%rcx\),%r11 13 [ ]*[a-f0-9]+: 66 0f 38 80 19 invept \(%rcx\),%rbx 14 [ ]*[a-f0-9]+: 66 44 0f 38 80 19 invept \(%rcx\),%r11 15 [ ]*[a-f0-9]+: 66 0f 38 81 19 invvpid \(%rcx\),%rbx 16 [ ]*[a-f0-9]+: 66 44 0f 38 81 19 invvpid \(%rcx\),%r11
|
x86-64-inval-invpcid.l | 16 [ ]*4[ ]+invvpid \(%rcx\), %bx 17 [ ]*5[ ]+invvpid \(%rcx\), %ebx 18 [ ]*6[ ]+invvpid %rbx, \(%rcx\) 19 [ ]*7[ ]+invvpid %rbx, %rcx 22 [ ]*10[ ]+invvpid bx, \[rcx\] 23 [ ]*11[ ]+invvpid ebx, \[rcx\] 24 [ ]*12[ ]+invvpid \[rcx\], rbx 25 [ ]*13[ ]+invvpid rcx, rbx
|
x86-64-inval-movbe.l | 18 [ ]*4[ ]+movbe \(%rcx\),%bl 20 [ ]*6[ ]+movbe %bx,%rcx 21 [ ]*7[ ]+movbe %rbx,%rcx 22 [ ]*8[ ]+movbe %bl,\(%rcx\) 25 [ ]*11[ ]+movbe bl, byte ptr \[rcx\] 27 [ ]*13[ ]+movbe rcx, bx 28 [ ]*14[ ]+movbe rcx, rbx 29 [ ]*15[ ]+movbe byte ptr \[rcx\], bl
|
x86-64-sse4_1.s | 5 blendpd $0x0,(%rcx),%xmm0 7 blendps $0x0,(%rcx),%xmm0 9 blendvpd %xmm0,(%rcx),%xmm0 11 blendvpd (%rcx),%xmm0 13 blendvps %xmm0,(%rcx),%xmm0 15 blendvps (%rcx),%xmm0 17 dppd $0x0,(%rcx),%xmm0 19 dpps $0x0,(%rcx),%xmm0 21 extractps $0x0,%xmm0,%rcx 23 extractps $0x0,%xmm0,(%rcx) [all...] |
/art/runtime/interpreter/mterp/x86_64/ |
op_iget_object_quick.S | 4 movzbq rINSTbl, %rcx # rcx <- BA 6 GET_VREG OUT_32_ARG0, %rcx # vB (object we're operating on) 10 movq rSELF, %rcx 11 cmpq $$0, THREAD_EXCEPTION_OFFSET(%rcx)
|
op_iput_quick.S | 4 movzbq rINSTbl, %rcx # rcx <- BA 6 GET_VREG %ecx, %rcx # vB (object we're operating on) 12 ${store} ${reg}, (%rcx,%rax,1)
|
op_iput_wide_quick.S | 2 movzbq rINSTbl, %rcx # rcx<- BA 4 GET_VREG %ecx, %rcx # vB (object we're operating on) 8 leaq (%rcx,%rax,1), %rcx # ecx<- Address of 64-bit target 11 movq %rax, (%rcx) # obj.field<- r0/r1
|
op_move_exception.S | 2 movq rSELF, %rcx 3 movl THREAD_EXCEPTION_OFFSET(%rcx), %eax 5 movl $$0, THREAD_EXCEPTION_OFFSET(%rcx)
|
op_iget_quick.S | 4 movl rINST, %ecx # rcx <- BA 6 GET_VREG %ecx, %rcx # vB (object we're operating on) 12 movq (%rcx,%rax,1), %rax 15 ${load} (%rcx,%rax,1), %eax
|
/external/compiler-rt/lib/builtins/x86_64/ |
chkstk2.S | 16 mov %rcx,%rax // x64 _alloca is a normal function with parameter in rcx 19 push %rcx 21 lea 16(%rsp),%rcx // rsp before calling this routine -> rcx 24 sub $0x1000,%rcx 25 test %rcx,(%rcx) 30 sub %rax,%rcx 31 test %rcx,(%rcx [all...] |
chkstk.S | 20 push %rcx 23 lea 24(%rsp),%rcx 26 sub $0x1000,%rcx 27 test %rcx,(%rcx) 32 sub %rax,%rcx 33 test %rcx,(%rcx) 35 pop %rcx
|
/toolchain/binutils/binutils-2.27/ld/testsuite/ld-x86-64/ |
nogot2.s | 4 movq _GLOBAL_OFFSET_TABLE_(%rip), %rcx
|