/art/compiler/optimizing/ |
intrinsics_mips64.cc | 388 static void MathAbsFP(LocationSummary* locations, bool is64bit, Mips64Assembler* assembler) { 405 MathAbsFP(invoke->GetLocations(), /* is64bit */ true, GetAssembler()); 414 MathAbsFP(invoke->GetLocations(), /* is64bit */ false, GetAssembler()); [all...] |
intrinsics_arm.cc | 243 static void MathAbsFP(LocationSummary* locations, bool is64bit, ArmAssembler* assembler) { 260 MathAbsFP(invoke->GetLocations(), /* is64bit */ true, GetAssembler()); 268 MathAbsFP(invoke->GetLocations(), /* is64bit */ false, GetAssembler()); [all...] |
intrinsics_arm64.cc | 412 static void MathAbsFP(LocationSummary* locations, bool is64bit, vixl::MacroAssembler* masm) { 427 MathAbsFP(invoke->GetLocations(), /* is64bit */ true, GetVIXLAssembler()); 435 MathAbsFP(invoke->GetLocations(), /* is64bit */ false, GetVIXLAssembler()); [all...] |
intrinsics_mips.cc | 748 static void MathAbsFP(LocationSummary* locations, bool is64bit, MipsAssembler* assembler) { 765 MathAbsFP(invoke->GetLocations(), /* is64bit */ true, GetAssembler()); 774 MathAbsFP(invoke->GetLocations(), /* is64bit */ false, GetAssembler()); [all...] |
intrinsics_x86.cc | 273 static void MathAbsFP(LocationSummary* locations, 313 MathAbsFP(invoke->GetLocations(), /* is64bit */ true, GetAssembler(), codegen_); 321 MathAbsFP(invoke->GetLocations(), /* is64bit */ false, GetAssembler(), codegen_); [all...] |
intrinsics_x86_64.cc | 207 static void MathAbsFP(LocationSummary* locations, 233 MathAbsFP(invoke->GetLocations(), /* is64bit */ true, GetAssembler(), codegen_); 241 MathAbsFP(invoke->GetLocations(), /* is64bit */ false, GetAssembler(), codegen_); [all...] |