/art/compiler/dex/quick/x86/ |
call_x86.cc | 87 if (cu_->target64) { 125 int ex_offset = cu_->target64 ? 129 NewLIR2(cu_->target64 ? kX86Mov64RT : kX86Mov32RT, rl_result.reg.GetReg(), ex_offset); 130 NewLIR2(cu_->target64 ? kX86Mov64TI : kX86Mov32TI, ex_offset, 0); 135 DCHECK_EQ(tgt_addr_reg.Is64Bit(), cu_->target64); 138 int ct_offset = cu_->target64 ? 141 NewLIR2(cu_->target64 ? kX86Mov64RT : kX86Mov32RT, reg_card_base.GetReg(), ct_offset); 170 const InstructionSet isa = cu_->target64 ? kX86_64 : kX86; 172 const RegStorage rs_rSP = cu_->target64 ? rs_rX86_SP_64 : rs_rX86_SP_32; 177 cu_->compiler_driver->GetCompilerOptions().GetImplicitStackOverflowChecks()) [all...] |
quick_assemble_x86_test.cc | 76 cu_.reset(new CompilationUnit(pool_.get(), isa_, compiler_driver_.get(), nullptr)); 78 cu_->arena.Alloc(sizeof(DexFile::CodeItem), kArenaAllocMisc)); 80 cu_->mir_graph.reset(new MIRGraph(cu_.get(), &cu_->arena)); 81 cu_->mir_graph->current_code_item_ = code_item; 82 cu_->cg.reset(QuickCompiler::GetCodeGenerator(cu_.get(), nullptr)); 96 X86Mir2Lir* m2l = static_cast<X86Mir2Lir*>(cu_->cg.get()); 102 cu_.reset() 123 std::unique_ptr<CompilationUnit> cu_; member in class:art::QuickAssembleX86TestBase [all...] |
target_x86.cc | 152 return cu_->target64 ? x86_64_loc_c_return_ref : x86_loc_c_return_ref; 156 return cu_->target64 ? x86_64_loc_c_return_wide : x86_loc_c_return_wide; 253 return cu_->target64 ? RegStorage32FromSpecialTargetRegister_Target64[reg] 279 DCHECK(cu_->instruction_set == kX86 || cu_->instruction_set == kX86_64); 472 if (!cu_->target64) { 483 return cu_->target64 || reg.GetRegNum() < rs_rX86_SP_32.GetRegNum(); 488 if (cu_->target64) { 551 if (cu_->target64) { 572 if (cu_->target64) [all...] |
utility_x86.cc | 392 DCHECK_EQ(r_base, cu_->target64 ? rs_rX86_SP_64 : rs_rX86_SP_32); 505 if (op == kOpMul && !cu_->target64) { 508 } else if (op == kOpAnd && !cu_->target64) { 531 DCHECK_EQ(kX86, cu_->instruction_set); 544 DCHECK_EQ(kX86_64, cu_->instruction_set); 578 } else if (pc_rel_base_reg_.Valid() || cu_->target64) { 590 if (cu_->target64) { 665 if (cu_->target64) { 716 DCHECK_EQ(r_base, cu_->target64 ? rs_rX86_SP_64 : rs_rX86_SP_32); 818 if (cu_->target64) [all...] |
int_x86.cc | 39 if (cu_->target64) { 138 if (!(cu_->disable_opt & (1 << kSafeOptimizations)) && r_dest == r_src) { 405 if (cu_->target64) { 459 if (cu_->target64) { 838 DCHECK(cu_->instruction_set == kX86 || cu_->instruction_set == kX86_64); 840 if (is_long && !cu_->target64) { [all...] |
/art/compiler/dex/ |
mir_graph_test.cc | 61 cu_.mir_graph->block_id_map_.clear(); 62 cu_.mir_graph->block_list_.clear(); 69 BasicBlock* bb = cu_.mir_graph->CreateNewBB(def->type); 81 static_cast<SuccessorBlockInfo*>(cu_.arena.Alloc(sizeof(SuccessorBlockInfo), 91 cu_.arena.Alloc(sizeof(BasicBlockDataFlow), kArenaAllocDFInfo)); 94 ASSERT_EQ(count, cu_.mir_graph->block_list_.size()); 95 cu_.mir_graph->entry_block_ = cu_.mir_graph->block_list_[1]; 96 ASSERT_EQ(kEntryBlock, cu_.mir_graph->entry_block_->block_type); 97 cu_.mir_graph->exit_block_ = cu_.mir_graph->block_list_[2] 158 CompilationUnit cu_; member in class:art::TopologicalSortOrderTest [all...] |
type_inference_test.cc | 249 cu_.dex_file = dex_file_.get(); 250 cu_.method_idx = dex_file_builder_.GetMethodIdx(kClassName, test_method_signature, kMethodName); 251 cu_.access_flags = is_static ? kAccStatic : 0u; 252 cu_.mir_graph->m_units_.push_back(new (cu_.mir_graph->arena_) DexCompilationUnit( 253 &cu_, cu_.class_loader, cu_.class_linker, *cu_.dex_file, nullptr /* code_item not used */, 255 cu_.access_flags, nullptr /* verified_method not used */)) 611 CompilationUnit cu_; member in class:art::TypeInferenceTest [all...] |
mir_optimization_test.cc | 100 cu_.mir_graph->block_id_map_.clear(); 101 cu_.mir_graph->block_list_.clear(); 108 BasicBlock* bb = cu_.mir_graph->CreateNewBB(def->type); 120 static_cast<SuccessorBlockInfo*>(cu_.arena.Alloc(sizeof(SuccessorBlockInfo), 130 cu_.arena.Alloc(sizeof(BasicBlockDataFlow), kArenaAllocDFInfo)); 133 ASSERT_EQ(count, cu_.mir_graph->block_list_.size()); 134 cu_.mir_graph->entry_block_ = cu_.mir_graph->block_list_[1]; 135 ASSERT_EQ(kEntryBlock, cu_.mir_graph->entry_block_->block_type); 136 cu_.mir_graph->exit_block_ = cu_.mir_graph->block_list_[2] 339 CompilationUnit cu_; member in class:art::MirOptimizationTest [all...] |
mir_optimization.cc | 252 return InstructionSetPointerSize(cu_->instruction_set); 283 bool verbose = cu_->verbose; 440 if (cu_->verbose) { 473 if (cu_->instruction_set == kArm64 || cu_->instruction_set == kThumb2) { 476 bool use_lvn = bb->use_lvn && (cu_->disable_opt & (1u << kLocalValueNumbering)) == 0u; 481 allocator.reset(ScopedArenaAllocator::Create(&cu_->arena_stack)); 482 global_valnum.reset(new (allocator.get()) GlobalValueNumbering(cu_, allocator.get(), 543 if ((cu_->disable_opt & (1 << kBranchFusing)) != 0) { 599 if ((cu_->instruction_set == kArm64 || cu_->instruction_set == kThumb2 | [all...] |
local_value_numbering_test.cc | 95 cu_.mir_graph->ifield_lowering_infos_.clear(); 96 cu_.mir_graph->ifield_lowering_infos_.reserve(count); 105 cu_.mir_graph->ifield_lowering_infos_.push_back(field_info); 115 cu_.mir_graph->sfield_lowering_infos_.clear(); 116 cu_.mir_graph->sfield_lowering_infos_.reserve(count); 128 cu_.mir_graph->sfield_lowering_infos_.push_back(field_info); 139 mirs_ = cu_.arena.AllocArray<MIR>(count, kArenaAllocMIR); 148 ASSERT_LT(def->field_info, cu_.mir_graph->ifield_lowering_infos_.size()); 150 ASSERT_EQ(cu_.mir_graph->ifield_lowering_infos_[def->field_info].MemAccessType(), 153 ASSERT_LT(def->field_info, cu_.mir_graph->sfield_lowering_infos_.size()) 232 CompilationUnit cu_; member in class:art::LocalValueNumberingTest [all...] |
gvn_dead_code_elimination_test.cc | 144 cu_.mir_graph->ifield_lowering_infos_.clear(); 145 cu_.mir_graph->ifield_lowering_infos_.reserve(count); 156 cu_.mir_graph->ifield_lowering_infos_.push_back(field_info); 166 cu_.mir_graph->sfield_lowering_infos_.clear(); 167 cu_.mir_graph->sfield_lowering_infos_.reserve(count); 181 cu_.mir_graph->sfield_lowering_infos_.push_back(field_info); 191 cu_.mir_graph->block_id_map_.clear(); 192 cu_.mir_graph->block_list_.clear(); 199 BasicBlock* bb = cu_.mir_graph->CreateNewBB(def->type); 211 static_cast<SuccessorBlockInfo*>(cu_.arena.Alloc(sizeof(SuccessorBlockInfo) 502 CompilationUnit cu_; member in class:art::GvnDeadCodeEliminationTest [all...] |
mir_analysis.cc | 1051 if (cu_->enable_debug & (1 << kDebugShowFilterStats)) { 1059 << PrettyMethod(cu_->method_idx, *cu_->dex_file); 1068 if (cu_->compiler_driver->GetCompilerOptions().IsSmallMethod(GetNumDalvikInsns()) && 1094 if (cu_->compiler_driver->GetCompilerOptions().IsSmallMethod(GetNumDalvikInsns()) && 1108 const CompilerOptions& compiler_options = cu_->compiler_driver->GetCompilerOptions(); 1172 if (((cu_->access_flags & kAccConstructor) != 0) && ((cu_->access_flags & kAccStatic) != 0)) { 1178 if (cu_->compiler_driver->GetMethodInlinerMap() != nullptr && 1179 cu_->compiler_driver->GetMethodInlinerMap()->GetMethodInliner(cu_->dex_file [all...] |
global_value_numbering_test.cc | 141 cu_.mir_graph->ifield_lowering_infos_.clear(); 142 cu_.mir_graph->ifield_lowering_infos_.reserve(count); 151 cu_.mir_graph->ifield_lowering_infos_.push_back(field_info); 161 cu_.mir_graph->sfield_lowering_infos_.clear(); 162 cu_.mir_graph->sfield_lowering_infos_.reserve(count); 174 cu_.mir_graph->sfield_lowering_infos_.push_back(field_info); 184 cu_.mir_graph->block_id_map_.clear(); 185 cu_.mir_graph->block_list_.clear(); 192 BasicBlock* bb = cu_.mir_graph->CreateNewBB(def->type); 204 static_cast<SuccessorBlockInfo*>(cu_.arena.Alloc(sizeof(SuccessorBlockInfo) 392 CompilationUnit cu_; member in class:art::GlobalValueNumberingTest [all...] |
/art/compiler/dex/quick/ |
mir_to_lir-inl.h | 74 << PrettyMethod(cu_->method_idx, *cu_->dex_file) << " " 84 << PrettyMethod(cu_->method_idx, *cu_->dex_file) << " " 94 << PrettyMethod(cu_->method_idx, *cu_->dex_file) << " " 104 << PrettyMethod(cu_->method_idx, *cu_->dex_file) << " " 114 << PrettyMethod(cu_->method_idx, *cu_->dex_file) << " [all...] |
gen_common.cc | 100 OpPcRelDexCacheArrayLoad(cu_->dex_file, offset, r_base, false); 291 if (!cu_->compiler_driver->GetCompilerOptions().GetImplicitNullChecks()) { 303 if (!(cu_->disable_opt & (1 << kNullCheckElimination)) && (opt_flags & MIR_IGNORE_NULL_CHECK)) { 310 if (cu_->compiler_driver->GetCompilerOptions().GetImplicitNullChecks()) { 311 if (!(cu_->disable_opt & (1 << kNullCheckElimination)) && (opt_flags & MIR_IGNORE_NULL_CHECK)) { 320 if (cu_->compiler_driver->GetCompilerOptions().GetImplicitNullChecks()) { 321 if (!(cu_->disable_opt & (1 << kNullCheckElimination)) && (opt_flags & MIR_IGNORE_NULL_CHECK)) { 329 if (cu_->compiler_driver->GetCompilerOptions().GetImplicitStackOverflowChecks()) { 335 if (cu_->compiler_driver->GetCompilerOptions().GetImplicitNullChecks()) { 336 if (!(cu_->disable_opt & (1 << kNullCheckElimination)) && (opt_flags & MIR_IGNORE_NULL_CHECK)) [all...] |
gen_invoke.cc | 82 if (cu_->instruction_set == kX86 || cu_->instruction_set == kX86_64) { 151 RegStorage r_tmp = TargetReg(cu_->instruction_set == kMips ? kArg2 : kArg1, kWide); 227 if (cu_->instruction_set == kArm64 || cu_->instruction_set == kMips64 || 228 cu_->instruction_set == kX86_64) { 250 DCHECK(!cu_->target64); 255 if (cu_->instruction_set == kMips) { 262 if (cu_->instruction_set == kMips) { 272 if (cu_->instruction_set == kMips) [all...] |
codegen_util.cc | 136 if (!cu_->verbose) { 202 const bool dump_nop = (cu_->enable_debug & (1 << kDebugShowNops)); 334 DCHECK_EQ(val_reg.Is64Bit(), cu_->target64); 348 << PrettyMethod(cu_->method_idx, *cu_->dex_file); 374 cu_->dex_file->GetMethodId(cu_->method_idx); 375 const Signature signature = cu_->dex_file->GetMethodSignature(method_id); 376 const char* name = cu_->dex_file->GetMethodName(method_id); 377 const char* descriptor(cu_->dex_file->GetMethodDeclaringClassDescriptor(method_id)) [all...] |
mir_to_lir.cc | 117 int offset = StackVisitor::GetOutVROffset(in_position, cu_->instruction_set); 119 if (cu_->instruction_set == kX86) { 127 if (cu_->instruction_set == kX86_64) { 168 int offset = StackVisitor::GetOutVROffset(in_position, cu_->instruction_set); 169 if (cu_->instruction_set == kX86) { 177 if (cu_->instruction_set == kX86_64) { 211 int offset = frame_size_ + StackVisitor::GetOutVROffset(in_position, cu_->instruction_set); 223 int offset = frame_size_ + StackVisitor::GetOutVROffset(in_position, cu_->instruction_set); 287 RegisterClass ret_reg_class = ShortyToRegClass(cu_->shorty[0]); 385 RegisterClass reg_class = ShortyToRegClass(cu_->shorty[0]) [all...] |
ralloc_util.cc | 291 DCHECK_NE(cu_->instruction_set, kThumb2); 310 DCHECK_NE(cu_->instruction_set, kThumb2); 481 if (cu_->target64) { 489 if (wide && !reg.IsFloat() && !cu_->target64) { 698 if (IsTemp(rl.reg) && !(cu_->disable_opt & (1 << kSuppressLoads))) { 708 if (IsTemp(rs) && !(cu_->disable_opt & (1 << kSuppressLoads))) { [all...] |
/art/compiler/driver/ |
dex_compilation_unit.cc | 34 : cu_(cu),
|
dex_compilation_unit.h | 45 return cu_; 113 CompilationUnit* const cu_; member in class:art::DexCompilationUnit
|
/art/compiler/dex/quick/mips/ |
target_mips.cc | 156 return cu_->target64 ? mips64_loc_c_return_ref : mips_loc_c_return; 160 return cu_->target64 ? mips64_loc_c_return_wide : mips_loc_c_return_wide; 168 if (cu_->target64) { 197 if (!cu_->target64 && wide_kind == kWide) { 207 } else if (cu_->target64 && (wide_kind == kWide || wide_kind == kRef)) { 227 case kArg4: res_reg = cu_->target64 ? rs_rA4 : RegStorage::InvalidReg(); break; 228 case kArg5: res_reg = cu_->target64 ? rs_rA5 : RegStorage::InvalidReg(); break; 229 case kArg6: res_reg = cu_->target64 ? rs_rA6 : RegStorage::InvalidReg(); break; 230 case kArg7: res_reg = cu_->target64 ? rs_rA7 : RegStorage::InvalidReg(); break; 235 case kFArg4: res_reg = cu_->target64 ? rs_rF16 : RegStorage::InvalidReg(); break [all...] |
utility_mips.cc | 34 if (cu_->target64) { 90 if (cu_->target64) { 95 if (!(cu_->disable_opt & (1 << kSafeOptimizations)) && r_dest == r_src) { 292 return NewLIR2(opcode, cu_->target64 ? rRAd : rRA, r_dest_src.GetReg()); 306 bool is64bit = cu_->target64 && (r_dest.Is64Bit() || r_src1.Is64Bit() || r_src2.Is64Bit()); 350 bool is64bit = cu_->target64 && (r_dest.Is64Bit() || r_src1.Is64Bit()); 476 if (cu_->target64 && r_dest_src1.Is64Bit()) { 489 if (cu_->target64) { 492 if (cu_->compiler_driver->GetInstructionSetFeatures()->AsMipsInstructionSetFeatures() 502 if (cu_->target64) [all...] |
call_mips.cc | 215 int ex_offset = cu_->target64 ? Thread::ExceptionOffset<8>().Int32Value() : 229 if (cu_->target64) { 269 if (cu_->target64) { 277 InstructionSet target = (cu_->target64) ? kMips64 : kMips; 278 int ptr_size = cu_->target64 ? 8 : 4; 292 !cu_->compiler_driver->GetCompilerOptions().GetImplicitStackOverflowChecks(); 297 if (cu_->target64) { 358 if (cu_->target64) { 399 StoreWordDisp(rs_sp, frame_size_ - (cu_->target64 ? 8 : 4), TargetPtrReg(kLr)); 400 cfi_.RelOffset(DwarfCoreReg(rRA), frame_size_ - (cu_->target64 ? 8 : 4)) [all...] |
int_mips.cc | 56 if (cu_->target64) { 185 if (!cu_->target64) { 199 if (cu_->target64) { 210 if (!(cu_->disable_opt & (1 << kSafeOptimizations)) && r_dest == r_src) { 224 if (cu_->target64) { 372 if (!cu_->target64) { 377 if (cu_->target64) { 395 if (!cu_->target64) { 400 if (cu_->target64) { 442 if (cu_->target64) [all...] |