/art/compiler/utils/ |
jni_macro_assembler_test.h | 36 Ass* GetAssembler() {
|
assembler_test.h | 51 Ass* GetAssembler() { [all...] |
/art/compiler/optimizing/ |
code_generator_mips64.h | 146 Mips64Assembler* GetAssembler() const; 221 Mips64Assembler* GetAssembler() const { return assembler_; } 356 Mips64Assembler* GetAssembler() OVERRIDE { return &assembler_; } 357 const Mips64Assembler& GetAssembler() const OVERRIDE { return assembler_; }
|
code_generator_x86.h | 138 X86Assembler* GetAssembler() const; 201 X86Assembler* GetAssembler() const { return assembler_; } 364 X86Assembler* GetAssembler() OVERRIDE { 368 const X86Assembler& GetAssembler() const OVERRIDE {
|
code_generator_x86_64.h | 137 X86_64Assembler* GetAssembler() const; 202 X86_64Assembler* GetAssembler() const { return assembler_; } 342 X86_64Assembler* GetAssembler() OVERRIDE { 346 const X86_64Assembler& GetAssembler() const OVERRIDE {
|
code_generator_arm64.h | 261 Arm64Assembler* GetAssembler() const { return assembler_; } 262 vixl::aarch64::MacroAssembler* GetVIXLAssembler() { return GetAssembler()->GetVIXLAssembler(); } 392 Arm64Assembler* GetAssembler() const; 394 return GetAssembler()->GetVIXLAssembler(); 442 Arm64Assembler* GetAssembler() OVERRIDE { return &assembler_; } 443 const Arm64Assembler& GetAssembler() const OVERRIDE { return assembler_; } 444 vixl::aarch64::MacroAssembler* GetVIXLAssembler() { return GetAssembler()->GetVIXLAssembler(); } [all...] |
code_generator_arm_vixl.h | 247 ArmVIXLAssembler* GetAssembler() const; 316 ArmVIXLAssembler* GetAssembler() const { return assembler_; } 317 ArmVIXLMacroAssembler* GetVIXLAssembler() { return GetAssembler()->GetVIXLAssembler(); } 455 ArmVIXLAssembler* GetAssembler() OVERRIDE { return &assembler_; } 457 const ArmVIXLAssembler& GetAssembler() const OVERRIDE { return assembler_; } 459 ArmVIXLMacroAssembler* GetVIXLAssembler() { return GetAssembler()->GetVIXLAssembler(); } [all...] |
code_generator_mips.h | 149 MipsAssembler* GetAssembler() const; 224 MipsAssembler* GetAssembler() const { return assembler_; } 393 MipsAssembler* GetAssembler() OVERRIDE { return &assembler_; } 394 const MipsAssembler& GetAssembler() const OVERRIDE { return assembler_; }
|
code_generator_mips64.cc | 102 #define __ down_cast<CodeGeneratorMIPS64*>(codegen)->GetAssembler()-> // NOLINT [all...] |
intrinsics_mips.cc | 39 MipsAssembler* IntrinsicCodeGeneratorMIPS::GetAssembler() { 40 return reinterpret_cast<MipsAssembler*>(codegen_->GetAssembler()); 59 #define __ codegen->GetAssembler()-> 183 MoveFPToInt(invoke->GetLocations(), /* is64bit */ true, GetAssembler()); 192 MoveFPToInt(invoke->GetLocations(), /* is64bit */ false, GetAssembler()); 225 MoveIntToFP(invoke->GetLocations(), /* is64bit */ true, GetAssembler()); 234 MoveIntToFP(invoke->GetLocations(), /* is64bit */ false, GetAssembler()); 412 GetAssembler()); 426 GetAssembler()); 440 GetAssembler()); [all...] |
intrinsics_mips64.cc | 39 Mips64Assembler* IntrinsicCodeGeneratorMIPS64::GetAssembler() { 40 return reinterpret_cast<Mips64Assembler*>(codegen_->GetAssembler()); 47 #define __ codegen->GetAssembler()-> 167 MoveFPToInt(invoke->GetLocations(), /* is64bit */ true, GetAssembler()); 176 MoveFPToInt(invoke->GetLocations(), /* is64bit */ false, GetAssembler()); 204 MoveIntToFP(invoke->GetLocations(), /* is64bit */ true, GetAssembler()); 213 MoveIntToFP(invoke->GetLocations(), /* is64bit */ false, GetAssembler()); 255 GenReverseBytes(invoke->GetLocations(), Primitive::kPrimInt, GetAssembler()); 264 GenReverseBytes(invoke->GetLocations(), Primitive::kPrimLong, GetAssembler()); 273 GenReverseBytes(invoke->GetLocations(), Primitive::kPrimShort, GetAssembler()); [all...] |
intrinsics_arm_vixl.cc | 62 ArmVIXLAssembler* IntrinsicCodeGeneratorARMVIXL::GetAssembler() { 63 return codegen_->GetAssembler(); 92 ArmVIXLAssembler* assembler = down_cast<ArmVIXLAssembler*>(codegen->GetAssembler()); 181 ArmVIXLAssembler* assembler = arm_codegen->GetAssembler(); 250 assembler_(codegen->GetAssembler()), 306 MoveFPToInt(invoke->GetLocations(), /* is64bit */ true, GetAssembler()); 309 MoveIntToFP(invoke->GetLocations(), /* is64bit */ true, GetAssembler()); 320 MoveFPToInt(invoke->GetLocations(), /* is64bit */ false, GetAssembler()); 323 MoveIntToFP(invoke->GetLocations(), /* is64bit */ false, GetAssembler()); 345 ArmVIXLAssembler* assembler = codegen->GetAssembler(); [all...] |
intrinsics_x86.cc | 53 X86Assembler* IntrinsicCodeGeneratorX86::GetAssembler() { 54 return down_cast<X86Assembler*>(codegen_->GetAssembler()); 78 #define __ down_cast<X86Assembler*>(codegen->GetAssembler())-> // NOLINT 239 MoveFPToInt(invoke->GetLocations(), /* is64bit */ true, GetAssembler()); 242 MoveIntToFP(invoke->GetLocations(), /* is64bit */ true, GetAssembler()); 253 MoveFPToInt(invoke->GetLocations(), /* is64bit */ false, GetAssembler()); 256 MoveIntToFP(invoke->GetLocations(), /* is64bit */ false, GetAssembler()); 308 GenReverseBytes(invoke->GetLocations(), Primitive::kPrimInt, GetAssembler()); 324 X86Assembler* assembler = GetAssembler(); 337 GenReverseBytes(invoke->GetLocations(), Primitive::kPrimShort, GetAssembler()); [all...] |
intrinsics_x86_64.cc | 46 X86_64Assembler* IntrinsicCodeGeneratorX86_64::GetAssembler() { 47 return down_cast<X86_64Assembler*>(codegen_->GetAssembler()); 71 #define __ down_cast<X86_64Assembler*>(codegen->GetAssembler())-> // NOLINT 166 MoveFPToInt(invoke->GetLocations(), /* is64bit */ true, GetAssembler()); 169 MoveIntToFP(invoke->GetLocations(), /* is64bit */ true, GetAssembler()); 180 MoveFPToInt(invoke->GetLocations(), /* is64bit */ false, GetAssembler()); 183 MoveIntToFP(invoke->GetLocations(), /* is64bit */ false, GetAssembler()); 222 GenReverseBytes(invoke->GetLocations(), Primitive::kPrimInt, GetAssembler()); 230 GenReverseBytes(invoke->GetLocations(), Primitive::kPrimLong, GetAssembler()); 238 GenReverseBytes(invoke->GetLocations(), Primitive::kPrimShort, GetAssembler()); [all...] |
code_generator_x86_64.cc | 56 #define __ down_cast<X86_64Assembler*>(codegen->GetAssembler())-> // NOLINT [all...] |
code_generator_arm_vixl.cc | 296 arm_codegen->GetAssembler()->StoreRegisterList(core_spills, orig_offset); 330 arm_codegen->GetAssembler()->LoadRegisterList(core_spills, orig_offset); [all...] |
code_generator_mips.cc | 147 #define __ down_cast<CodeGeneratorMIPS*>(codegen)->GetAssembler()-> // NOLINT [all...] |
code_generator_x86.cc | 52 #define __ down_cast<X86Assembler*>(codegen->GetAssembler())-> // NOLINT [all...] |