/external/valgrind/main/none/tests/amd64/ |
sbbmisc.c | 12 VG_SYM(sbb_ib_al) ":\n" 14 "\tmovb " VG_SYM(in_b) ", %al\n" 17 "\tmovb %al, " VG_SYM(out_b1) "\n" 19 "\tmovb " VG_SYM(in_b) ", %al\n" 22 "\tmovb %al, " VG_SYM(out_b2) "\n" 30 VG_SYM(sbb_iw_ax) ":\n" 32 "\tmovw " VG_SYM(in_w) ", %ax\n" 35 "\tmovw %ax, " VG_SYM(out_w1) "\n" 37 "\tmovw " VG_SYM(in_w) ", %ax\n" 40 "\tmovw %ax, " VG_SYM(out_w2) "\n [all...] |
cmpxchg.c | 36 "\tmov " VG_SYM(rax) ",%rax\n" 37 "\tmov " VG_SYM(rbx) ",%rbx\n" 38 "\tmov " VG_SYM(rcx) ",%rcx\n" 39 "\tmov " VG_SYM(rdx) ",%rdx\n" 41 "\tmov %rax," VG_SYM(rax_out) "\n" 42 "\tmov %rbx," VG_SYM(rbx_out) "\n" 43 "\tmov %rcx," VG_SYM(rcx_out) "\n" 67 "\tmov " VG_SYM(rax) ",%rax\n" 68 "\tmov " VG_SYM(rbx) ",%rbx\n" 69 "\tmov " VG_SYM(rcx) ",%rcx\n [all...] |
/external/valgrind/main/none/tests/x86/ |
sbbmisc.c | 12 VG_SYM(sbb_ib_al) ":\n" 14 "\tmovb " VG_SYM(in_b) ", %al\n" 17 "\tmovb %al, " VG_SYM(out_b1) "\n" 19 "\tmovb " VG_SYM(in_b) ", %al\n" 22 "\tmovb %al, " VG_SYM(out_b2) "\n" 30 VG_SYM(sbb_iw_ax) ":\n" 32 "\tmovw " VG_SYM(in_w) ", %ax\n" 35 "\tmovw %ax, " VG_SYM(out_w1) "\n" 37 "\tmovw " VG_SYM(in_w) ", %ax\n" 40 "\tmovw %ax, " VG_SYM(out_w2) "\n [all...] |
looper.c | 10 VG_SYM(loop_plain) ":\n" 13 "\tmovl " VG_SYM(arg) ", %ecx\n" 21 "\tmovl %eax, " VG_SYM(res) "\n" 29 VG_SYM(loop_ne) ":\n" 32 "\tmovl " VG_SYM(arg) ", %ecx\n" 40 "\tmovl %eax, " VG_SYM(res) "\n" 48 VG_SYM(loop_e) ":\n" 51 "\tmovl " VG_SYM(arg) ", %ecx\n" 63 "\tmovl %eax, " VG_SYM(res) "\n"
|
lahf.c | 6 VG_SYM(foo_1) ":\n" 18 VG_SYM(foo_0) ":\n"
|
incdec_alt.c | 11 VG_SYM(foo) ":\n" 14 "\tmovl " VG_SYM(a1) ",%eax\n" 15 "\tmovl " VG_SYM(a2) ",%edx\n" 21 "\tmovl %eax," VG_SYM(r1) "\n" 23 "\tpopl " VG_SYM(r2) "\n" 30 "\tmovl %edx," VG_SYM(r3) "\n" 32 "\tpopl " VG_SYM(r4) "\n" 38 "\tmovl %edx," VG_SYM(r5) "\n" 40 "\tpopl " VG_SYM(r6) "\n" 47 "\tmovl %eax," VG_SYM(r7) "\n [all...] |
cmpxchg8b.c | 17 VG_SYM(foo) ":\n" 25 "\tmovl " VG_SYM(eax) ",%eax\n" 26 "\tmovl " VG_SYM(ebx) ",%ebx\n" 27 "\tmovl " VG_SYM(ecx) ",%ecx\n" 28 "\tmovl " VG_SYM(edx) ",%edx\n" 29 "\tcmpxchg8b " VG_SYM(m64) "\n" 30 "\tmovl %eax," VG_SYM(eax) "\n" 31 "\tmovl %ebx," VG_SYM(ebx) "\n" 32 "\tmovl %ecx," VG_SYM(ecx) "\n" 33 "\tmovl %edx," VG_SYM(edx) "\n [all...] |
fcmovnu.c | 18 "fstl " VG_SYM(zzz) "\n\t"
|
fxtract.c | 9 VG_SYM(do_fxtract) ":\n" 11 "\tfldl " VG_SYM(arg) "\n" 13 "\tfstpl " VG_SYM(res1) "\n" 14 "\tfstpl " VG_SYM(res2) "\n"
|
bug126147-x86.c | 23 VG_SYM(mystrncpy) ":\n" 66 VG_SYM(__ltostr) ":\n"
|
faultstatus.c | 80 asm volatile(VG_SYM(test1_ill) ": ud2");
|
/external/valgrind/main/tests/ |
asm.h | 5 // general, any symbol named in asm code should be wrapped by VG_SYM. 9 #define VG_SYM(x) "_"#x 11 #define VG_SYM(x) #x
|
/external/valgrind/main/memcheck/tests/x86/ |
fxsave.c | 38 asm __volatile__("movups " VG_SYM(vecZ) ", %xmm0"); 39 asm __volatile__("movups " VG_SYM(vecZ) ", %xmm1"); 40 asm __volatile__("movups " VG_SYM(vecZ) ", %xmm2"); 41 asm __volatile__("movups " VG_SYM(vecZ) ", %xmm3"); 42 asm __volatile__("movups " VG_SYM(vecZ) ", %xmm4"); 43 asm __volatile__("movups " VG_SYM(vecZ) ", %xmm5"); 44 asm __volatile__("movups " VG_SYM(vecZ) ", %xmm6"); 45 asm __volatile__("movups " VG_SYM(vecZ) ", %xmm7"); 62 asm __volatile__("movups " VG_SYM(vec0) ", %xmm0"); 63 asm __volatile__("movups " VG_SYM(vec1) ", %xmm1") [all...] |