/art/compiler/optimizing/ |
intrinsics_arm64.cc | 254 static void MoveFPToInt(LocationSummary* locations, bool is64bit, MacroAssembler* masm) { 276 MoveFPToInt(invoke->GetLocations(), /* is64bit */ true, GetVIXLAssembler()); 290 MoveFPToInt(invoke->GetLocations(), /* is64bit */ false, GetVIXLAssembler()); [all...] |
intrinsics_arm.cc | 209 static void MoveFPToInt(LocationSummary* locations, bool is64bit, ArmAssembler* assembler) { 241 MoveFPToInt(invoke->GetLocations(), /* is64bit */ true, GetAssembler()); 255 MoveFPToInt(invoke->GetLocations(), /* is64bit */ false, GetAssembler()); [all...] |
intrinsics_arm_vixl.cc | 279 static void MoveFPToInt(LocationSummary* locations, bool is64bit, ArmVIXLAssembler* assembler) { 307 MoveFPToInt(invoke->GetLocations(), /* is64bit */ true, GetAssembler()); 321 MoveFPToInt(invoke->GetLocations(), /* is64bit */ false, GetAssembler()); [all...] |
intrinsics_mips.cc | 160 static void MoveFPToInt(LocationSummary* locations, bool is64bit, MipsAssembler* assembler) { 182 MoveFPToInt(invoke->GetLocations(), /* is64bit */ true, GetAssembler()); 191 MoveFPToInt(invoke->GetLocations(), /* is64bit */ false, GetAssembler()); [all...] |
intrinsics_mips64.cc | 149 static void MoveFPToInt(LocationSummary* locations, bool is64bit, Mips64Assembler* assembler) { 166 MoveFPToInt(invoke->GetLocations(), /* is64bit */ true, GetAssembler()); 175 MoveFPToInt(invoke->GetLocations(), /* is64bit */ false, GetAssembler()); [all...] |
intrinsics_x86.cc | 201 static void MoveFPToInt(LocationSummary* locations, bool is64bit, X86Assembler* assembler) { 240 MoveFPToInt(invoke->GetLocations(), /* is64bit */ true, GetAssembler()); 254 MoveFPToInt(invoke->GetLocations(), /* is64bit */ false, GetAssembler()); [all...] |
intrinsics_x86_64.cc | 147 static void MoveFPToInt(LocationSummary* locations, bool is64bit, X86_64Assembler* assembler) { 167 MoveFPToInt(invoke->GetLocations(), /* is64bit */ true, GetAssembler()); 181 MoveFPToInt(invoke->GetLocations(), /* is64bit */ false, GetAssembler()); [all...] |