/external/lzma/Asm/x86/ |
7zCrcOpt.asm | 9 rN equ r7
21 SRCDAT equ rN + rD + 4 *
42 dec rN
49 mov rN, num_VAR
51 test rN, rN
59 cmp rN, 16
61 add rN, rD
62 mov num_VAR, rN
63 sub rN, 8 [all...] |
AesOpt.asm | 19 rN equ r0
31 mov rN, num
107 sub rN, ways
110 add rN, ways
124 sub rN, 1
161 sub rN, 1
215 sub rN, ways
218 add rN, ways
231 sub rN, 1
|
/external/valgrind/main/VEX/priv/ |
host_arm64_isel.c | [all...] |
host_arm64_defs.c | 415 //ZZ ARMAModeN *mkARMAModeN_RR ( HReg rN, HReg rM ) { 418 //ZZ am->ARMamN.RR.rN = rN; 423 //ZZ ARMAModeN *mkARMAModeN_R ( HReg rN ) { 426 //ZZ am->ARMamN.R.rN = rN; 432 //ZZ addHRegUse(u, HRmRead, am->ARMamN.R.rN); 434 //ZZ addHRegUse(u, HRmRead, am->ARMamN.RR.rN); 441 //ZZ am->ARMamN.R.rN = lookupHRegRemap(m, am->ARMamN.R.rN); [all...] |
host_arm64_defs.h | 784 HReg rN; 791 HReg rN; 798 HReg rN; // address 1061 extern ARM64Instr* ARM64Instr_VLdStS ( Bool isLoad, HReg sD, HReg rN, 1063 extern ARM64Instr* ARM64Instr_VLdStD ( Bool isLoad, HReg dD, HReg rN, 1065 extern ARM64Instr* ARM64Instr_VLdStQ ( Bool isLoad, HReg rQ, HReg rN ); [all...] |
host_arm_defs.c | 379 ARMAModeN *mkARMAModeN_RR ( HReg rN, HReg rM ) { 382 am->ARMamN.RR.rN = rN; 387 ARMAModeN *mkARMAModeN_R ( HReg rN ) { 390 am->ARMamN.R.rN = rN; 396 addHRegUse(u, HRmRead, am->ARMamN.R.rN); 398 addHRegUse(u, HRmRead, am->ARMamN.RR.rN); 405 am->ARMamN.R.rN = lookupHRegRemap(m, am->ARMamN.R.rN); [all...] |
host_arm_defs.h | 212 HReg rN; 216 HReg rN; 939 /* Note: rD != rN */ 941 HReg rN; 1014 extern ARMInstr* ARMInstr_Add32 ( HReg rD, HReg rN, UInt imm32 );
|
guest_arm64_toIR.c | [all...] |
guest_arm_toIR.c | [all...] |