/art/compiler/optimizing/ |
intrinsics_utils.h | 42 explicit IntrinsicSlowPath(HInvoke* invoke) : SlowPathCode(invoke), invoke_(invoke) { } 46 IntrinsicVisitor::MoveArguments(invoke_, codegen, &calling_convention_visitor); 54 SaveLiveRegisters(codegen, invoke_->GetLocations()); 58 if (invoke_->IsInvokeStaticOrDirect()) { 59 codegen->GenerateStaticOrDirectCall(invoke_->AsInvokeStaticOrDirect(), method_loc, this); 61 codegen->GenerateVirtualCall(invoke_->AsInvokeVirtual(), method_loc, this); 65 Location out = invoke_->GetLocations()->Out(); 68 DCHECK(!invoke_->GetLocations()->GetLiveRegisters()->ContainsCoreRegister(out.reg())); 69 codegen->MoveFromReturnRegister(out, invoke_->GetType()); 72 RestoreLiveRegisters(codegen, invoke_->GetLocations()) 80 HInvoke* const invoke_; member in class:art::IntrinsicSlowPath [all...] |
intrinsics_arm64.cc | 112 : SlowPathCodeARM64(invoke), invoke_(invoke) { } 118 SaveLiveRegisters(codegen, invoke_->GetLocations()); 120 MoveArguments(invoke_, codegen); 126 if (invoke_->IsInvokeStaticOrDirect()) { 128 invoke_->AsInvokeStaticOrDirect(), LocationFrom(kArtMethodRegister), this); 131 invoke_->AsInvokeVirtual(), LocationFrom(kArtMethodRegister), this); 136 Location out = invoke_->GetLocations()->Out(); 139 DCHECK(!invoke_->GetLocations()->GetLiveRegisters()->ContainsCoreRegister(out.reg())); 140 MoveFromReturnRegister(out, invoke_->GetType(), codegen); 143 RestoreLiveRegisters(codegen, invoke_->GetLocations()) 151 HInvoke* const invoke_; member in class:art::arm64::IntrinsicSlowPathARM64 [all...] |
intrinsics_arm_vixl.cc | 83 : SlowPathCodeARMVIXL(invoke), invoke_(invoke) {} 87 IntrinsicVisitor::MoveArguments(invoke_, codegen, &calling_convention_visitor); 95 SaveLiveRegisters(codegen, invoke_->GetLocations()); 99 if (invoke_->IsInvokeStaticOrDirect()) { 100 codegen->GenerateStaticOrDirectCall(invoke_->AsInvokeStaticOrDirect(), method_loc, this); 102 codegen->GenerateVirtualCall(invoke_->AsInvokeVirtual(), method_loc, this); 106 Location out = invoke_->GetLocations()->Out(); 109 DCHECK(!invoke_->GetLocations()->GetLiveRegisters()->ContainsCoreRegister(out.reg())); 110 codegen->MoveFromReturnRegister(out, invoke_->GetType()); 113 RestoreLiveRegisters(codegen, invoke_->GetLocations()) 121 HInvoke* const invoke_; member in class:art::arm::IntrinsicSlowPathARMVIXL [all...] |
intrinsics_mips.cc | 103 explicit IntrinsicSlowPathMIPS(HInvoke* invoke) : SlowPathCodeMIPS(invoke), invoke_(invoke) { } 110 SaveLiveRegisters(codegen, invoke_->GetLocations()); 112 MoveArguments(invoke_, codegen); 114 if (invoke_->IsInvokeStaticOrDirect()) { 116 invoke_->AsInvokeStaticOrDirect(), Location::RegisterLocation(A0), this); 119 invoke_->AsInvokeVirtual(), Location::RegisterLocation(A0), this); 123 Location out = invoke_->GetLocations()->Out(); 126 DCHECK(!invoke_->GetLocations()->GetLiveRegisters()->ContainsCoreRegister(out.reg())); 127 MoveFromReturnRegister(out, invoke_->GetType(), codegen); 130 RestoreLiveRegisters(codegen, invoke_->GetLocations()) 138 HInvoke* const invoke_; member in class:art::mips::IntrinsicSlowPathMIPS [all...] |
intrinsics_mips64.cc | 92 : SlowPathCodeMIPS64(invoke), invoke_(invoke) { } 99 SaveLiveRegisters(codegen, invoke_->GetLocations()); 101 MoveArguments(invoke_, codegen); 103 if (invoke_->IsInvokeStaticOrDirect()) { 105 invoke_->AsInvokeStaticOrDirect(), Location::RegisterLocation(A0), this); 108 invoke_->AsInvokeVirtual(), Location::RegisterLocation(A0), this); 112 Location out = invoke_->GetLocations()->Out(); 115 DCHECK(!invoke_->GetLocations()->GetLiveRegisters()->ContainsCoreRegister(out.reg())); 116 MoveFromReturnRegister(out, invoke_->GetType(), codegen); 119 RestoreLiveRegisters(codegen, invoke_->GetLocations()) 127 HInvoke* const invoke_; member in class:art::mips64::IntrinsicSlowPathMIPS64 [all...] |