/art/compiler/optimizing/ |
register_allocation_resolver.cc | 281 size_t fp_spills = local 284 core_register_spill_size * core_spills + fp_register_spill_size * fp_spills;
|
code_generator.cc | 1323 const uint32_t fp_spills = codegen->GetSlowPathSpills(locations, \/* core_registers *\/ false); local 1342 const uint32_t fp_spills = codegen->GetSlowPathSpills(locations, \/* core_registers *\/ false); local [all...] |
code_generator_arm64.cc | 172 const uint32_t fp_spills = codegen->GetSlowPathSpills(locations, /* core_registers */ false); local 175 fp_spills, 180 CPURegList fp_list = CPURegList(CPURegister::kVRegister, v_reg_size, fp_spills); 227 const uint32_t fp_spills = codegen->GetSlowPathSpills(locations, /* core_registers */ false); local 228 for (uint32_t i : LowToHighBits(fp_spills)) { [all...] |
code_generator_arm.cc | 193 uint32_t fp_spills = codegen->GetSlowPathSpills(locations, /* core_registers */ false); local 195 for (uint32_t i : LowToHighBits(fp_spills)) { 202 while (fp_spills != 0u) { 203 uint32_t begin = CTZ(fp_spills); 204 uint32_t tmp = fp_spills + (1u << begin); 205 fp_spills &= tmp; // Clear the contiguous range of 1s. 235 uint32_t fp_spills = codegen->GetSlowPathSpills(locations, /* core_registers */ false); local 236 while (fp_spills != 0u) { 237 uint32_t begin = CTZ(fp_spills); 238 uint32_t tmp = fp_spills + (1u << begin) [all...] |
code_generator_arm_vixl.cc | 227 uint32_t fp_spills = codegen->GetSlowPathSpills(locations, /* core_registers */ false); local 229 for (uint32_t i : LowToHighBits(fp_spills)) { 236 while (fp_spills != 0u) { 237 uint32_t begin = CTZ(fp_spills); 238 uint32_t tmp = fp_spills + (1u << begin); 239 fp_spills &= tmp; // Clear the contiguous range of 1s. 261 uint32_t fp_spills = codegen->GetSlowPathSpills(locations, /* core_registers */ false); local 262 while (fp_spills != 0u) { 263 uint32_t begin = CTZ(fp_spills); 264 uint32_t tmp = fp_spills + (1u << begin) [all...] |