/art/compiler/dex/quick/arm/ |
backend_arm.h | 24 class MIRGraph; 27 Mir2Lir* ArmCodeGenerator(CompilationUnit* const cu, MIRGraph* const mir_graph,
|
target_arm.cc | 576 ArmMir2Lir::ArmMir2Lir(CompilationUnit* cu, MIRGraph* mir_graph, ArenaAllocator* arena) 591 Mir2Lir* ArmCodeGenerator(CompilationUnit* const cu, MIRGraph* const mir_graph, [all...] |
/art/compiler/dex/quick/arm64/ |
backend_arm64.h | 24 class MIRGraph; 27 Mir2Lir* Arm64CodeGenerator(CompilationUnit* const cu, MIRGraph* const mir_graph,
|
target_arm64.cc | 606 Arm64Mir2Lir::Arm64Mir2Lir(CompilationUnit* cu, MIRGraph* mir_graph, ArenaAllocator* arena) 619 Mir2Lir* Arm64CodeGenerator(CompilationUnit* const cu, MIRGraph* const mir_graph,
|
/art/compiler/dex/quick/mips/ |
backend_mips.h | 24 class MIRGraph; 27 Mir2Lir* MipsCodeGenerator(CompilationUnit* const cu, MIRGraph* const mir_graph,
|
/art/compiler/dex/quick/x86/ |
backend_x86.h | 24 class MIRGraph; 27 Mir2Lir* X86CodeGenerator(CompilationUnit* const cu, MIRGraph* const mir_graph,
|
quick_assemble_x86_test.cc | 80 cu_->mir_graph.reset(new MIRGraph(cu_.get(), &cu_->arena));
|
/art/compiler/dex/ |
dataflow_iterator.h | 67 * @param mir_graph the MIRGraph we are interested in. 71 DataflowIterator(MIRGraph* mir_graph, int32_t start_idx, int32_t end_idx) 104 MIRGraph* const mir_graph_; /**< @brief the MIRGraph */ 115 * @brief Used to perform a Pre-order Depth-First-Search Iteration of a MIRGraph. 120 * @brief The constructor, using all of the reachable blocks of the MIRGraph. 121 * @param mir_graph The MIRGraph considered. 123 explicit PreOrderDfsIterator(MIRGraph* mir_graph) 145 * @brief Used to perform a Repeating Pre-order Depth-First-Search Iteration of a MIRGraph. 151 * @brief The constructor, using all of the reachable blocks of the MIRGraph [all...] |
vreg_analysis.cc | 28 void MIRGraph::DumpRegLocTable(RegLocation* table, int count) { 45 void MIRGraph::InitRegLocations() { 68 void MIRGraph::RemapRegLocations() {
|
ssa_transformation.cc | 27 void MIRGraph::ClearAllVisitedFlags() { 34 BasicBlock* MIRGraph::NeedsVisit(BasicBlock* bb) { 43 BasicBlock* MIRGraph::NextUnvisitedSuccessor(BasicBlock* bb) { 61 void MIRGraph::MarkPreOrder(BasicBlock* block) { 69 void MIRGraph::RecordDFSOrders(BasicBlock* block) { 92 void MIRGraph::ComputeDFSOrders() { 123 bool MIRGraph::FillDefBlockMatrix(BasicBlock* bb) { 135 void MIRGraph::ComputeDefBlockMatrix() { 171 void MIRGraph::ComputeDomPostOrderTraversal(BasicBlock* bb) { 205 void MIRGraph::CheckForDominanceFrontier(BasicBlock* dom_bb [all...] |
type_inference.h | 34 class MIRGraph; 69 TypeInference(MIRGraph* mir_graph, ScopedArenaAllocator* alloc); 372 CheckCastData(MIRGraph* mir_graph, ScopedArenaAllocator* alloc); 395 MIRGraph* const mir_graph_; 410 static Type* PrepareIFieldTypes(const DexFile* dex_file, MIRGraph* mir_graph, 412 static Type* PrepareSFieldTypes(const DexFile* dex_file, MIRGraph* mir_graph, 416 static MethodSignature* PrepareSignatures(const DexFile* dex_file, MIRGraph* mir_graph, 418 static CheckCastData* InitializeCheckCastData(MIRGraph* mir_graph, ScopedArenaAllocator* alloc); 427 MIRGraph* const mir_graph_;
|
gvn_dead_code_elimination.h | 30 class MIRGraph; 115 MIRGraph* mir_graph) const; 152 MIRGraph* const mir_graph_;
|
mir_optimization.cc | 43 void MIRGraph::SetConstant(int32_t ssa_reg, int32_t value) { 49 void MIRGraph::SetConstantWide(int32_t ssa_reg, int64_t value) { 58 void MIRGraph::DoConstantPropagation(BasicBlock* bb) { 120 MIR* MIRGraph::AdvanceMIR(BasicBlock** p_bb, MIR* mir) { 145 MIR* MIRGraph::FindMoveResult(BasicBlock* bb, MIR* mir) { 164 BasicBlock* MIRGraph::NextDominatedBlock(BasicBlock* bb) { 244 int MIRGraph::GetSSAUseCount(int s_reg) { 249 size_t MIRGraph::GetNumBytesForSpecialTemps() const { 255 size_t MIRGraph::GetNumAvailableVRTemps() { 277 CompilerTemp* MIRGraph::GetNewCompilerTemp(CompilerTempType ct_type, bool wide) [all...] |
mir_graph.cc | 44 const char* MIRGraph::extended_mir_op_names_[kMirOpLast - kMirOpFirst] = { 83 MIRGraph::MIRGraph(CompilationUnit* cu, ArenaAllocator* arena) 161 MIRGraph::~MIRGraph() { 169 int MIRGraph::ParseInsn(const uint16_t* code_ptr, MIR::DecodedInstruction* decoded_instruction) { 184 BasicBlock* MIRGraph::SplitBlock(DexOffset code_offset, 291 BasicBlock* MIRGraph::FindBlock(DexOffset code_offset, bool create, 356 void MIRGraph::ProcessTryCatchBlocks(ScopedArenaVector<uint16_t>* dex_pc_to_block_map) { 387 bool MIRGraph::IsBadMonitorExitCatch(NarrowDexOffset monitor_exit_offset [all...] |
compiler_ir.h | 37 class MIRGraph; 193 std::unique_ptr<MIRGraph> mir_graph; // MIR container.
|
post_opt_passes.cc | 52 // First get the MIRGraph here to factorize a bit the code. 53 MIRGraph *mir_graph = c_unit->mir_graph.get();
|
mir_dataflow.cc | 31 const uint64_t MIRGraph::oat_data_flow_attributes_[kMirOpLast] = { 911 void MIRGraph::HandleLiveInUse(ArenaBitVector* use_v, ArenaBitVector* def_v, 920 void MIRGraph::HandleDef(ArenaBitVector* def_v, int dalvik_reg_id) { 924 void MIRGraph::HandleExtended(ArenaBitVector* use_v, ArenaBitVector* def_v, 985 bool MIRGraph::FindLocalLiveIn(BasicBlock* bb) { 1045 int MIRGraph::AddNewSReg(int v_reg) { 1062 void MIRGraph::HandleSSAUse(int* uses, int dalvik_reg, int reg_index) { 1068 void MIRGraph::HandleSSADef(int* defs, int dalvik_reg, int reg_index) { 1075 void MIRGraph::AllocateSSAUseData(MIR *mir, int num_uses) { 1083 void MIRGraph::AllocateSSADefData(MIR *mir, int num_defs) [all...] |
mir_graph.h | 46 class MIRGraph; 181 // this on automatically disables the SCE in MIRGraph::EliminateSuspendChecksGate(). 345 // IGET/IPUT lowering info index, points to MIRGraph::ifield_lowering_infos_. Due to limit on 348 // SGET/SPUT lowering info index, points to MIRGraph::sfield_lowering_infos_. Due to limit on 351 // INVOKE data index, points to MIRGraph::method_lowering_infos_. Also used for inlined 366 MIR* Copy(MIRGraph* mir_Graph); 425 BasicBlock* Copy(MIRGraph* mir_graph); 437 void Kill(MIRGraph* mir_graph); 483 * @param mir_graph the MIRGraph. 487 MIR* GetNextUnconditionalMir(MIRGraph* mir_graph, MIR* current) [all...] |
global_value_numbering.h | 48 // Instance and static field id map is held by MIRGraph to avoid multiple recalculations 61 MIRGraph* GetMirGraph() const { 213 MIRGraph* const mir_graph_;
|
mir_analysis.cc | 960 void MIRGraph::AnalyzeBlock(BasicBlock* bb, MethodStats* stats) { 1042 bool MIRGraph::ComputeSkipCompilation(MethodStats* stats, bool skip_default, 1107 bool MIRGraph::SkipCompilation(std::string* skip_message) { 1206 void MIRGraph::DoCacheFieldLoweringInfo() { 1303 void MIRGraph::DoCacheMethodLoweringInfo() { [all...] |
type_inference.cc | 153 TypeInference::CheckCastData::CheckCastData(MIRGraph* mir_graph, ScopedArenaAllocator* alloc) 393 TypeInference::TypeInference(MIRGraph* mir_graph, ScopedArenaAllocator* alloc) 459 uint64_t attrs = MIRGraph::GetDataFlowAttributes(mir); 521 DCHECK_NE(MIRGraph::GetDataFlowAttributes(mir) & DF_SAME_TYPE_AB, 0u); 600 MIRGraph* mir_graph, 617 MIRGraph* mir_graph, 651 MIRGraph* mir_graph, 665 TypeInference::CheckCastData* TypeInference::InitializeCheckCastData(MIRGraph* mir_graph, 727 uint64_t attrs = MIRGraph::GetDataFlowAttributes(mir); [all...] |
post_opt_passes.h | 31 * @brief Convenience class for passes that check MIRGraph::MirSsaRepUpToDate(). 221 MIRGraph *mir_graph = c_unit->mir_graph.get();
|
/art/compiler/dex/quick/ |
dex_file_method_inliner.h | 37 class MIRGraph; 95 bool GenInline(MIRGraph* mir_graph, BasicBlock* bb, MIR* invoke, uint32_t method_idx) 360 static bool GenInlineConst(MIRGraph* mir_graph, BasicBlock* bb, MIR* invoke, 362 static bool GenInlineReturnArg(MIRGraph* mir_graph, BasicBlock* bb, MIR* invoke, 364 static bool GenInlineIGet(MIRGraph* mir_graph, BasicBlock* bb, MIR* invoke, 366 static bool GenInlineIPut(MIRGraph* mir_graph, BasicBlock* bb, MIR* invoke,
|
quick_cfi_test.cc | 84 cu.mir_graph.reset(new MIRGraph(&cu, &arena));
|
dex_file_method_inliner.cc | 113 MIR* AllocReplacementMIR(MIRGraph* mir_graph, MIR* invoke) { 642 bool DexFileMethodInliner::GenInline(MIRGraph* mir_graph, BasicBlock* bb, MIR* invoke, [all...] |