HomeSort by relevance Sort by last modified time
    Searched refs:cu_ (Results 1 - 25 of 52) sorted by null

1 2 3

  /art/compiler/dex/quick/x86/
call_x86.cc 87 if (cu_->target64) {
125 int ex_offset = cu_->target64 ?
129 NewLIR2(cu_->target64 ? kX86Mov64RT : kX86Mov32RT, rl_result.reg.GetReg(), ex_offset);
130 NewLIR2(cu_->target64 ? kX86Mov64TI : kX86Mov32TI, ex_offset, 0);
135 DCHECK_EQ(tgt_addr_reg.Is64Bit(), cu_->target64);
138 int ct_offset = cu_->target64 ?
141 NewLIR2(cu_->target64 ? kX86Mov64RT : kX86Mov32RT, reg_card_base.GetReg(), ct_offset);
170 const InstructionSet isa = cu_->target64 ? kX86_64 : kX86;
172 const RegStorage rs_rSP = cu_->target64 ? rs_rX86_SP_64 : rs_rX86_SP_32;
177 cu_->compiler_driver->GetCompilerOptions().GetImplicitStackOverflowChecks())
    [all...]
quick_assemble_x86_test.cc 76 cu_.reset(new CompilationUnit(pool_.get(), isa_, compiler_driver_.get(), nullptr));
78 cu_->arena.Alloc(sizeof(DexFile::CodeItem), kArenaAllocMisc));
80 cu_->mir_graph.reset(new MIRGraph(cu_.get(), &cu_->arena));
81 cu_->mir_graph->current_code_item_ = code_item;
82 cu_->cg.reset(QuickCompiler::GetCodeGenerator(cu_.get(), nullptr));
96 X86Mir2Lir* m2l = static_cast<X86Mir2Lir*>(cu_->cg.get());
102 cu_.reset()
123 std::unique_ptr<CompilationUnit> cu_; member in class:art::QuickAssembleX86TestBase
    [all...]
target_x86.cc 152 return cu_->target64 ? x86_64_loc_c_return_ref : x86_loc_c_return_ref;
156 return cu_->target64 ? x86_64_loc_c_return_wide : x86_loc_c_return_wide;
253 return cu_->target64 ? RegStorage32FromSpecialTargetRegister_Target64[reg]
279 DCHECK(cu_->instruction_set == kX86 || cu_->instruction_set == kX86_64);
472 if (!cu_->target64) {
483 return cu_->target64 || reg.GetRegNum() < rs_rX86_SP_32.GetRegNum();
488 if (cu_->target64) {
551 if (cu_->target64) {
572 if (cu_->target64)
    [all...]
utility_x86.cc 392 DCHECK_EQ(r_base, cu_->target64 ? rs_rX86_SP_64 : rs_rX86_SP_32);
505 if (op == kOpMul && !cu_->target64) {
508 } else if (op == kOpAnd && !cu_->target64) {
531 DCHECK_EQ(kX86, cu_->instruction_set);
544 DCHECK_EQ(kX86_64, cu_->instruction_set);
578 } else if (pc_rel_base_reg_.Valid() || cu_->target64) {
590 if (cu_->target64) {
665 if (cu_->target64) {
716 DCHECK_EQ(r_base, cu_->target64 ? rs_rX86_SP_64 : rs_rX86_SP_32);
818 if (cu_->target64)
    [all...]
int_x86.cc 39 if (cu_->target64) {
138 if (!(cu_->disable_opt & (1 << kSafeOptimizations)) && r_dest == r_src) {
405 if (cu_->target64) {
459 if (cu_->target64) {
838 DCHECK(cu_->instruction_set == kX86 || cu_->instruction_set == kX86_64);
840 if (is_long && !cu_->target64) {
    [all...]
  /art/compiler/dex/
mir_graph_test.cc 61 cu_.mir_graph->block_id_map_.clear();
62 cu_.mir_graph->block_list_.clear();
69 BasicBlock* bb = cu_.mir_graph->CreateNewBB(def->type);
81 static_cast<SuccessorBlockInfo*>(cu_.arena.Alloc(sizeof(SuccessorBlockInfo),
91 cu_.arena.Alloc(sizeof(BasicBlockDataFlow), kArenaAllocDFInfo));
94 ASSERT_EQ(count, cu_.mir_graph->block_list_.size());
95 cu_.mir_graph->entry_block_ = cu_.mir_graph->block_list_[1];
96 ASSERT_EQ(kEntryBlock, cu_.mir_graph->entry_block_->block_type);
97 cu_.mir_graph->exit_block_ = cu_.mir_graph->block_list_[2]
158 CompilationUnit cu_; member in class:art::TopologicalSortOrderTest
    [all...]
type_inference_test.cc 249 cu_.dex_file = dex_file_.get();
250 cu_.method_idx = dex_file_builder_.GetMethodIdx(kClassName, test_method_signature, kMethodName);
251 cu_.access_flags = is_static ? kAccStatic : 0u;
252 cu_.mir_graph->m_units_.push_back(new (cu_.mir_graph->arena_) DexCompilationUnit(
253 &cu_, cu_.class_loader, cu_.class_linker, *cu_.dex_file, nullptr /* code_item not used */,
255 cu_.access_flags, nullptr /* verified_method not used */))
611 CompilationUnit cu_; member in class:art::TypeInferenceTest
    [all...]
mir_optimization_test.cc 100 cu_.mir_graph->block_id_map_.clear();
101 cu_.mir_graph->block_list_.clear();
108 BasicBlock* bb = cu_.mir_graph->CreateNewBB(def->type);
120 static_cast<SuccessorBlockInfo*>(cu_.arena.Alloc(sizeof(SuccessorBlockInfo),
130 cu_.arena.Alloc(sizeof(BasicBlockDataFlow), kArenaAllocDFInfo));
133 ASSERT_EQ(count, cu_.mir_graph->block_list_.size());
134 cu_.mir_graph->entry_block_ = cu_.mir_graph->block_list_[1];
135 ASSERT_EQ(kEntryBlock, cu_.mir_graph->entry_block_->block_type);
136 cu_.mir_graph->exit_block_ = cu_.mir_graph->block_list_[2]
339 CompilationUnit cu_; member in class:art::MirOptimizationTest
    [all...]
mir_optimization.cc 252 return InstructionSetPointerSize(cu_->instruction_set);
283 bool verbose = cu_->verbose;
440 if (cu_->verbose) {
473 if (cu_->instruction_set == kArm64 || cu_->instruction_set == kThumb2) {
476 bool use_lvn = bb->use_lvn && (cu_->disable_opt & (1u << kLocalValueNumbering)) == 0u;
481 allocator.reset(ScopedArenaAllocator::Create(&cu_->arena_stack));
482 global_valnum.reset(new (allocator.get()) GlobalValueNumbering(cu_, allocator.get(),
543 if ((cu_->disable_opt & (1 << kBranchFusing)) != 0) {
599 if ((cu_->instruction_set == kArm64 || cu_->instruction_set == kThumb2 |
    [all...]
local_value_numbering_test.cc 95 cu_.mir_graph->ifield_lowering_infos_.clear();
96 cu_.mir_graph->ifield_lowering_infos_.reserve(count);
105 cu_.mir_graph->ifield_lowering_infos_.push_back(field_info);
115 cu_.mir_graph->sfield_lowering_infos_.clear();
116 cu_.mir_graph->sfield_lowering_infos_.reserve(count);
128 cu_.mir_graph->sfield_lowering_infos_.push_back(field_info);
139 mirs_ = cu_.arena.AllocArray<MIR>(count, kArenaAllocMIR);
148 ASSERT_LT(def->field_info, cu_.mir_graph->ifield_lowering_infos_.size());
150 ASSERT_EQ(cu_.mir_graph->ifield_lowering_infos_[def->field_info].MemAccessType(),
153 ASSERT_LT(def->field_info, cu_.mir_graph->sfield_lowering_infos_.size())
232 CompilationUnit cu_; member in class:art::LocalValueNumberingTest
    [all...]
gvn_dead_code_elimination_test.cc 144 cu_.mir_graph->ifield_lowering_infos_.clear();
145 cu_.mir_graph->ifield_lowering_infos_.reserve(count);
156 cu_.mir_graph->ifield_lowering_infos_.push_back(field_info);
166 cu_.mir_graph->sfield_lowering_infos_.clear();
167 cu_.mir_graph->sfield_lowering_infos_.reserve(count);
181 cu_.mir_graph->sfield_lowering_infos_.push_back(field_info);
191 cu_.mir_graph->block_id_map_.clear();
192 cu_.mir_graph->block_list_.clear();
199 BasicBlock* bb = cu_.mir_graph->CreateNewBB(def->type);
211 static_cast<SuccessorBlockInfo*>(cu_.arena.Alloc(sizeof(SuccessorBlockInfo)
502 CompilationUnit cu_; member in class:art::GvnDeadCodeEliminationTest
    [all...]
mir_analysis.cc 1051 if (cu_->enable_debug & (1 << kDebugShowFilterStats)) {
1059 << PrettyMethod(cu_->method_idx, *cu_->dex_file);
1068 if (cu_->compiler_driver->GetCompilerOptions().IsSmallMethod(GetNumDalvikInsns()) &&
1094 if (cu_->compiler_driver->GetCompilerOptions().IsSmallMethod(GetNumDalvikInsns()) &&
1108 const CompilerOptions& compiler_options = cu_->compiler_driver->GetCompilerOptions();
1172 if (((cu_->access_flags & kAccConstructor) != 0) && ((cu_->access_flags & kAccStatic) != 0)) {
1178 if (cu_->compiler_driver->GetMethodInlinerMap() != nullptr &&
1179 cu_->compiler_driver->GetMethodInlinerMap()->GetMethodInliner(cu_->dex_file
    [all...]
global_value_numbering_test.cc 141 cu_.mir_graph->ifield_lowering_infos_.clear();
142 cu_.mir_graph->ifield_lowering_infos_.reserve(count);
151 cu_.mir_graph->ifield_lowering_infos_.push_back(field_info);
161 cu_.mir_graph->sfield_lowering_infos_.clear();
162 cu_.mir_graph->sfield_lowering_infos_.reserve(count);
174 cu_.mir_graph->sfield_lowering_infos_.push_back(field_info);
184 cu_.mir_graph->block_id_map_.clear();
185 cu_.mir_graph->block_list_.clear();
192 BasicBlock* bb = cu_.mir_graph->CreateNewBB(def->type);
204 static_cast<SuccessorBlockInfo*>(cu_.arena.Alloc(sizeof(SuccessorBlockInfo)
392 CompilationUnit cu_; member in class:art::GlobalValueNumberingTest
    [all...]
  /art/compiler/dex/quick/
mir_to_lir-inl.h 74 << PrettyMethod(cu_->method_idx, *cu_->dex_file) << " "
84 << PrettyMethod(cu_->method_idx, *cu_->dex_file) << " "
94 << PrettyMethod(cu_->method_idx, *cu_->dex_file) << " "
104 << PrettyMethod(cu_->method_idx, *cu_->dex_file) << " "
114 << PrettyMethod(cu_->method_idx, *cu_->dex_file) << "
    [all...]
gen_common.cc 100 OpPcRelDexCacheArrayLoad(cu_->dex_file, offset, r_base, false);
291 if (!cu_->compiler_driver->GetCompilerOptions().GetImplicitNullChecks()) {
303 if (!(cu_->disable_opt & (1 << kNullCheckElimination)) && (opt_flags & MIR_IGNORE_NULL_CHECK)) {
310 if (cu_->compiler_driver->GetCompilerOptions().GetImplicitNullChecks()) {
311 if (!(cu_->disable_opt & (1 << kNullCheckElimination)) && (opt_flags & MIR_IGNORE_NULL_CHECK)) {
320 if (cu_->compiler_driver->GetCompilerOptions().GetImplicitNullChecks()) {
321 if (!(cu_->disable_opt & (1 << kNullCheckElimination)) && (opt_flags & MIR_IGNORE_NULL_CHECK)) {
329 if (cu_->compiler_driver->GetCompilerOptions().GetImplicitStackOverflowChecks()) {
335 if (cu_->compiler_driver->GetCompilerOptions().GetImplicitNullChecks()) {
336 if (!(cu_->disable_opt & (1 << kNullCheckElimination)) && (opt_flags & MIR_IGNORE_NULL_CHECK))
    [all...]
gen_invoke.cc 82 if (cu_->instruction_set == kX86 || cu_->instruction_set == kX86_64) {
151 RegStorage r_tmp = TargetReg(cu_->instruction_set == kMips ? kArg2 : kArg1, kWide);
227 if (cu_->instruction_set == kArm64 || cu_->instruction_set == kMips64 ||
228 cu_->instruction_set == kX86_64) {
250 DCHECK(!cu_->target64);
255 if (cu_->instruction_set == kMips) {
262 if (cu_->instruction_set == kMips) {
272 if (cu_->instruction_set == kMips)
    [all...]
codegen_util.cc 136 if (!cu_->verbose) {
202 const bool dump_nop = (cu_->enable_debug & (1 << kDebugShowNops));
334 DCHECK_EQ(val_reg.Is64Bit(), cu_->target64);
348 << PrettyMethod(cu_->method_idx, *cu_->dex_file);
374 cu_->dex_file->GetMethodId(cu_->method_idx);
375 const Signature signature = cu_->dex_file->GetMethodSignature(method_id);
376 const char* name = cu_->dex_file->GetMethodName(method_id);
377 const char* descriptor(cu_->dex_file->GetMethodDeclaringClassDescriptor(method_id))
    [all...]
mir_to_lir.cc 117 int offset = StackVisitor::GetOutVROffset(in_position, cu_->instruction_set);
119 if (cu_->instruction_set == kX86) {
127 if (cu_->instruction_set == kX86_64) {
168 int offset = StackVisitor::GetOutVROffset(in_position, cu_->instruction_set);
169 if (cu_->instruction_set == kX86) {
177 if (cu_->instruction_set == kX86_64) {
211 int offset = frame_size_ + StackVisitor::GetOutVROffset(in_position, cu_->instruction_set);
223 int offset = frame_size_ + StackVisitor::GetOutVROffset(in_position, cu_->instruction_set);
287 RegisterClass ret_reg_class = ShortyToRegClass(cu_->shorty[0]);
385 RegisterClass reg_class = ShortyToRegClass(cu_->shorty[0])
    [all...]
ralloc_util.cc 291 DCHECK_NE(cu_->instruction_set, kThumb2);
310 DCHECK_NE(cu_->instruction_set, kThumb2);
481 if (cu_->target64) {
489 if (wide && !reg.IsFloat() && !cu_->target64) {
698 if (IsTemp(rl.reg) && !(cu_->disable_opt & (1 << kSuppressLoads))) {
708 if (IsTemp(rs) && !(cu_->disable_opt & (1 << kSuppressLoads))) {
    [all...]
  /art/compiler/driver/
dex_compilation_unit.cc 34 : cu_(cu),
dex_compilation_unit.h 45 return cu_;
113 CompilationUnit* const cu_; member in class:art::DexCompilationUnit
  /art/compiler/dex/quick/mips/
target_mips.cc 156 return cu_->target64 ? mips64_loc_c_return_ref : mips_loc_c_return;
160 return cu_->target64 ? mips64_loc_c_return_wide : mips_loc_c_return_wide;
168 if (cu_->target64) {
197 if (!cu_->target64 && wide_kind == kWide) {
207 } else if (cu_->target64 && (wide_kind == kWide || wide_kind == kRef)) {
227 case kArg4: res_reg = cu_->target64 ? rs_rA4 : RegStorage::InvalidReg(); break;
228 case kArg5: res_reg = cu_->target64 ? rs_rA5 : RegStorage::InvalidReg(); break;
229 case kArg6: res_reg = cu_->target64 ? rs_rA6 : RegStorage::InvalidReg(); break;
230 case kArg7: res_reg = cu_->target64 ? rs_rA7 : RegStorage::InvalidReg(); break;
235 case kFArg4: res_reg = cu_->target64 ? rs_rF16 : RegStorage::InvalidReg(); break
    [all...]
utility_mips.cc 34 if (cu_->target64) {
90 if (cu_->target64) {
95 if (!(cu_->disable_opt & (1 << kSafeOptimizations)) && r_dest == r_src) {
292 return NewLIR2(opcode, cu_->target64 ? rRAd : rRA, r_dest_src.GetReg());
306 bool is64bit = cu_->target64 && (r_dest.Is64Bit() || r_src1.Is64Bit() || r_src2.Is64Bit());
350 bool is64bit = cu_->target64 && (r_dest.Is64Bit() || r_src1.Is64Bit());
476 if (cu_->target64 && r_dest_src1.Is64Bit()) {
489 if (cu_->target64) {
492 if (cu_->compiler_driver->GetInstructionSetFeatures()->AsMipsInstructionSetFeatures()
502 if (cu_->target64)
    [all...]
call_mips.cc 215 int ex_offset = cu_->target64 ? Thread::ExceptionOffset<8>().Int32Value() :
229 if (cu_->target64) {
269 if (cu_->target64) {
277 InstructionSet target = (cu_->target64) ? kMips64 : kMips;
278 int ptr_size = cu_->target64 ? 8 : 4;
292 !cu_->compiler_driver->GetCompilerOptions().GetImplicitStackOverflowChecks();
297 if (cu_->target64) {
358 if (cu_->target64) {
399 StoreWordDisp(rs_sp, frame_size_ - (cu_->target64 ? 8 : 4), TargetPtrReg(kLr));
400 cfi_.RelOffset(DwarfCoreReg(rRA), frame_size_ - (cu_->target64 ? 8 : 4))
    [all...]
int_mips.cc 56 if (cu_->target64) {
185 if (!cu_->target64) {
199 if (cu_->target64) {
210 if (!(cu_->disable_opt & (1 << kSafeOptimizations)) && r_dest == r_src) {
224 if (cu_->target64) {
372 if (!cu_->target64) {
377 if (cu_->target64) {
395 if (!cu_->target64) {
400 if (cu_->target64) {
442 if (cu_->target64)
    [all...]

Completed in 1244 milliseconds

1 2 3