/art/compiler/dex/quick/mips/ |
utility_mips.cc | 526 RegStorage r_tmp = AllocTemp(); local 527 res = OpRegRegImm(kOpAdd, r_tmp, r_base, displacement); 528 load = NewLIR3(opcode, r_dest.GetLowReg(), LOWORD_OFFSET, r_tmp.GetReg()); 529 load2 = NewLIR3(opcode, r_dest.GetHighReg(), HIWORD_OFFSET, r_tmp.GetReg()); 530 FreeTemp(r_tmp); 532 RegStorage r_tmp = (r_base == r_dest) ? AllocTemp() : r_dest; local 533 res = OpRegRegImm(kOpAdd, r_tmp, r_base, displacement); 534 load = NewLIR3(opcode, r_dest.GetReg(), 0, r_tmp.GetReg()); 535 if (r_tmp != r_dest) 536 FreeTemp(r_tmp); [all...] |
/art/compiler/dex/quick/arm/ |
int_arm.cc | 895 RegStorage r_tmp = AllocTemp(); local 904 NewLIR3(kThumb2Ldrexd, r_tmp.GetReg(), r_tmp_high.GetReg(), r_ptr.GetReg()); 905 OpRegReg(kOpSub, r_tmp, rl_expected.reg.GetLow()); 911 if (r_tmp.Low8() && r_tmp_high.Low8()) { 912 NewLIR2(kThumbOrr, r_tmp.GetReg(), r_tmp_high.GetReg()); 914 NewLIR4(kThumb2OrrRRRs, r_tmp.GetReg(), r_tmp.GetReg(), r_tmp_high.GetReg(), 0); 920 NewLIR4(kThumb2Strexd /* eq */, r_tmp.GetReg(), rl_new_value.reg.GetLowReg(), rl_new_value.reg.GetHighReg(), r_ptr.GetReg()); [all...] |
utility_arm.cc | 573 RegStorage r_tmp = AllocTemp(); local 574 res = LoadConstant(r_tmp, value); 575 OpRegReg(kOpCmp, r_src1, r_tmp); 576 FreeTemp(r_tmp); [all...] |
/art/compiler/dex/quick/arm64/ |
utility_arm64.cc | 966 RegStorage r_tmp; local 969 r_tmp = AllocTempWide(); 970 res = LoadConstantWide(r_tmp, value); 972 r_tmp = AllocTemp(); 973 res = LoadConstant(r_tmp, value); 975 OpRegReg(op, r_dest_src1, r_tmp); 976 FreeTemp(r_tmp); [all...] |
fp_arm64.cc | 428 RegStorage r_tmp = (is_double) ? AllocTempDouble() : AllocTempSingle(); local 431 NewLIR3(kA64Fadd3fff | wide, r_tmp.GetReg(), rl_src.reg.GetReg(), r_imm_point5.GetReg()); 432 NewLIR2((is_double) ? kA64Fcvtms2xS : kA64Fcvtms2ws, rl_result.reg.GetReg(), r_tmp.GetReg());
|
int_arm64.cc | 749 RegStorage r_tmp; local 754 r_tmp_stored = r_tmp = AllocTempWide(); 758 // This means r_tmp_stored != r_tmp. 759 r_tmp = AllocTempRef(); 760 r_tmp_stored = As32BitReg(r_tmp); 763 r_tmp_stored = r_tmp = AllocTemp(); 766 RegStorage r_tmp32 = (r_tmp.Is32Bit()) ? r_tmp : As32BitReg(r_tmp); 769 OpRegReg(kOpCmp, r_tmp, rl_expected.reg) [all...] |
/art/compiler/dex/quick/ |
gen_invoke.cc | 147 RegStorage r_tmp = TargetReg(cu_->instruction_set == kMips ? kArg2 : kArg1, kWide); local 148 LoadValueDirectWideFixed(arg1, r_tmp); 196 RegStorage r_tmp = TargetReg(kArg0, arg0.GetWideKind()); local 197 if (r_tmp.NotExactlyEquals(arg0)) { 198 OpRegCopy(r_tmp, arg0); 209 RegStorage r_tmp = TargetReg(kArg0, arg0.GetWideKind()); local 210 if (r_tmp.NotExactlyEquals(arg0)) { 211 OpRegCopy(r_tmp, arg0); 742 RegStorage r_tmp(RegStorage::k64BitPair, arg_regs[next_reg], arg_regs[next_reg + 1]); 743 LoadValueDirectWideFixed(rl_arg, r_tmp); [all...] |
gen_common.cc | 573 RegStorage r_tmp = TargetReg(kArg2, kNotWide); local 574 LockTemp(r_tmp); 575 LIR* uninit_branch = OpCmpMemImmBranch(kCondLt, r_tmp, r_base, 583 FreeTemp(r_tmp); 658 RegStorage r_tmp = TargetReg(kArg2, kNotWide); local 659 LockTemp(r_tmp); 660 LIR* uninit_branch = OpCmpMemImmBranch(kCondLt, r_tmp, r_base, 668 FreeTemp(r_tmp); [all...] |
/art/compiler/dex/quick/x86/ |
int_x86.cc | 163 RegStorage r_tmp = AllocTempDouble(); local 164 NewLIR2(kX86MovdxrRR, r_tmp.GetReg(), r_src.GetHighReg()); 165 NewLIR2(kX86PunpckldqRR, r_dest.GetReg(), r_tmp.GetReg()); 166 FreeTemp(r_tmp); [all...] |