/art/compiler/optimizing/ |
code_generator_mips64.cc | 176 : SlowPathCodeMIPS64(at), cls_(cls), at_(at), dex_pc_(dex_pc), do_clinit_(do_clinit) { 189 int32_t entry_point_offset = do_clinit_ ? QUICK_ENTRY_POINT(pInitializeStaticStorage) 192 if (do_clinit_) { 224 const bool do_clinit_; member in class:art::mips64::LoadClassSlowPathMIPS64 [all...] |
code_generator_mips.cc | 217 : SlowPathCodeMIPS(at), cls_(cls), at_(at), dex_pc_(dex_pc), do_clinit_(do_clinit) { 231 int32_t entry_point_offset = do_clinit_ ? QUICK_ENTRY_POINT(pInitializeStaticStorage) 233 bool direct = do_clinit_ ? IsDirectEntrypoint(kQuickInitializeStaticStorage) 237 if (do_clinit_) { 269 const bool do_clinit_; member in class:art::mips::LoadClassSlowPathMIPS [all...] |
code_generator_arm64.cc | 282 : SlowPathCodeARM64(at), cls_(cls), at_(at), dex_pc_(dex_pc), do_clinit_(do_clinit) { 295 int32_t entry_point_offset = do_clinit_ ? QUICK_ENTRY_POINT(pInitializeStaticStorage) 298 if (do_clinit_) { 330 const bool do_clinit_; member in class:art::arm64::LoadClassSlowPathARM64 [all...] |
code_generator_arm.cc | 197 : SlowPathCode(at), cls_(cls), at_(at), dex_pc_(dex_pc), do_clinit_(do_clinit) { 210 int32_t entry_point_offset = do_clinit_ 214 if (do_clinit_) { 244 const bool do_clinit_; member in class:art::arm::LoadClassSlowPathARM [all...] |
code_generator_x86.cc | 243 : SlowPathCode(at), cls_(cls), at_(at), dex_pc_(dex_pc), do_clinit_(do_clinit) { 255 x86_codegen->InvokeRuntime(do_clinit_ ? QUICK_ENTRY_POINT(pInitializeStaticStorage) 258 if (do_clinit_) { 289 const bool do_clinit_; member in class:art::x86::LoadClassSlowPathX86 [all...] |
code_generator_x86_64.cc | 227 : SlowPathCode(at), cls_(cls), at_(at), dex_pc_(dex_pc), do_clinit_(do_clinit) { 240 x86_64_codegen->InvokeRuntime(do_clinit_ ? 246 if (do_clinit_) { 277 const bool do_clinit_; member in class:art::x86_64::LoadClassSlowPathX86_64 [all...] |