/art/compiler/optimizing/ |
intrinsics_utils.h | 42 explicit IntrinsicSlowPath(HInvoke* invoke) : SlowPathCode(invoke), invoke_(invoke) { } 46 IntrinsicVisitor::MoveArguments(invoke_, codegen, &calling_convention_visitor); 54 SaveLiveRegisters(codegen, invoke_->GetLocations()); 58 if (invoke_->IsInvokeStaticOrDirect()) { 59 codegen->GenerateStaticOrDirectCall(invoke_->AsInvokeStaticOrDirect(), method_loc, this); 61 codegen->GenerateVirtualCall(invoke_->AsInvokeVirtual(), method_loc, this); 65 Location out = invoke_->GetLocations()->Out(); 68 DCHECK(!invoke_->GetLocations()->GetLiveRegisters()->ContainsCoreRegister(out.reg())); 69 codegen->MoveFromReturnRegister(out, invoke_->GetType()); 72 RestoreLiveRegisters(codegen, invoke_->GetLocations()) 80 HInvoke* const invoke_; member in class:art::IntrinsicSlowPath [all...] |
intrinsics_arm64.cc | 113 : SlowPathCodeARM64(invoke), invoke_(invoke) { } 119 SaveLiveRegisters(codegen, invoke_->GetLocations()); 121 MoveArguments(invoke_, codegen); 127 if (invoke_->IsInvokeStaticOrDirect()) { 129 invoke_->AsInvokeStaticOrDirect(), LocationFrom(kArtMethodRegister), this); 132 invoke_->AsInvokeVirtual(), LocationFrom(kArtMethodRegister), this); 137 Location out = invoke_->GetLocations()->Out(); 140 DCHECK(!invoke_->GetLocations()->GetLiveRegisters()->ContainsCoreRegister(out.reg())); 141 MoveFromReturnRegister(out, invoke_->GetType(), codegen); 144 RestoreLiveRegisters(codegen, invoke_->GetLocations()) 152 HInvoke* const invoke_; member in class:art::arm64::IntrinsicSlowPathARM64 [all...] |
intrinsics_arm_vixl.cc | 84 : SlowPathCodeARMVIXL(invoke), invoke_(invoke) {} 88 IntrinsicVisitor::MoveArguments(invoke_, codegen, &calling_convention_visitor); 96 SaveLiveRegisters(codegen, invoke_->GetLocations()); 100 if (invoke_->IsInvokeStaticOrDirect()) { 101 codegen->GenerateStaticOrDirectCall(invoke_->AsInvokeStaticOrDirect(), method_loc, this); 103 codegen->GenerateVirtualCall(invoke_->AsInvokeVirtual(), method_loc, this); 107 Location out = invoke_->GetLocations()->Out(); 110 DCHECK(!invoke_->GetLocations()->GetLiveRegisters()->ContainsCoreRegister(out.reg())); 111 codegen->MoveFromReturnRegister(out, invoke_->GetType()); 114 RestoreLiveRegisters(codegen, invoke_->GetLocations()) 122 HInvoke* const invoke_; member in class:art::arm::IntrinsicSlowPathARMVIXL [all...] |
intrinsics_mips.cc | 105 explicit IntrinsicSlowPathMIPS(HInvoke* invoke) : SlowPathCodeMIPS(invoke), invoke_(invoke) { } 112 SaveLiveRegisters(codegen, invoke_->GetLocations()); 114 MoveArguments(invoke_, codegen); 116 if (invoke_->IsInvokeStaticOrDirect()) { 118 invoke_->AsInvokeStaticOrDirect(), Location::RegisterLocation(A0), this); 121 invoke_->AsInvokeVirtual(), Location::RegisterLocation(A0), this); 125 Location out = invoke_->GetLocations()->Out(); 128 DCHECK(!invoke_->GetLocations()->GetLiveRegisters()->ContainsCoreRegister(out.reg())); 129 MoveFromReturnRegister(out, invoke_->GetType(), codegen); 132 RestoreLiveRegisters(codegen, invoke_->GetLocations()) 140 HInvoke* const invoke_; member in class:art::mips::IntrinsicSlowPathMIPS [all...] |
intrinsics_mips64.cc | 94 : SlowPathCodeMIPS64(invoke), invoke_(invoke) { } 101 SaveLiveRegisters(codegen, invoke_->GetLocations()); 103 MoveArguments(invoke_, codegen); 105 if (invoke_->IsInvokeStaticOrDirect()) { 107 invoke_->AsInvokeStaticOrDirect(), Location::RegisterLocation(A0), this); 110 invoke_->AsInvokeVirtual(), Location::RegisterLocation(A0), this); 114 Location out = invoke_->GetLocations()->Out(); 117 DCHECK(!invoke_->GetLocations()->GetLiveRegisters()->ContainsCoreRegister(out.reg())); 118 MoveFromReturnRegister(out, invoke_->GetType(), codegen); 121 RestoreLiveRegisters(codegen, invoke_->GetLocations()) 129 HInvoke* const invoke_; member in class:art::mips64::IntrinsicSlowPathMIPS64 [all...] |