1 //===-- LiveIntervalAnalysis.cpp - Live Interval Analysis -----------------===// 2 // 3 // The LLVM Compiler Infrastructure 4 // 5 // This file is distributed under the University of Illinois Open Source 6 // License. See LICENSE.TXT for details. 7 // 8 //===----------------------------------------------------------------------===// 9 // 10 // This file implements the LiveInterval analysis pass which is used 11 // by the Linear Scan Register allocator. This pass linearizes the 12 // basic blocks of the function in DFS order and uses the 13 // LiveVariables pass to conservatively compute live intervals for 14 // each virtual and physical register. 15 // 16 //===----------------------------------------------------------------------===// 17 18 #define DEBUG_TYPE "liveintervals" 19 #include "llvm/CodeGen/LiveIntervalAnalysis.h" 20 #include "VirtRegMap.h" 21 #include "llvm/Value.h" 22 #include "llvm/Analysis/AliasAnalysis.h" 23 #include "llvm/CodeGen/CalcSpillWeights.h" 24 #include "llvm/CodeGen/LiveVariables.h" 25 #include "llvm/CodeGen/MachineFrameInfo.h" 26 #include "llvm/CodeGen/MachineInstr.h" 27 #include "llvm/CodeGen/MachineInstrBuilder.h" 28 #include "llvm/CodeGen/MachineLoopInfo.h" 29 #include "llvm/CodeGen/MachineMemOperand.h" 30 #include "llvm/CodeGen/MachineRegisterInfo.h" 31 #include "llvm/CodeGen/Passes.h" 32 #include "llvm/CodeGen/ProcessImplicitDefs.h" 33 #include "llvm/Target/TargetRegisterInfo.h" 34 #include "llvm/Target/TargetInstrInfo.h" 35 #include "llvm/Target/TargetMachine.h" 36 #include "llvm/Target/TargetOptions.h" 37 #include "llvm/Support/CommandLine.h" 38 #include "llvm/Support/Debug.h" 39 #include "llvm/Support/ErrorHandling.h" 40 #include "llvm/Support/raw_ostream.h" 41 #include "llvm/ADT/DepthFirstIterator.h" 42 #include "llvm/ADT/SmallSet.h" 43 #include "llvm/ADT/Statistic.h" 44 #include "llvm/ADT/STLExtras.h" 45 #include <algorithm> 46 #include <limits> 47 #include <cmath> 48 using namespace llvm; 49 50 // Hidden options for help debugging. 51 static cl::opt<bool> DisableReMat("disable-rematerialization", 52 cl::init(false), cl::Hidden); 53 54 STATISTIC(numIntervals , "Number of original intervals"); 55 STATISTIC(numFolds , "Number of loads/stores folded into instructions"); 56 STATISTIC(numSplits , "Number of intervals split"); 57 58 char LiveIntervals::ID = 0; 59 INITIALIZE_PASS_BEGIN(LiveIntervals, "liveintervals", 60 "Live Interval Analysis", false, false) 61 INITIALIZE_PASS_DEPENDENCY(LiveVariables) 62 INITIALIZE_PASS_DEPENDENCY(MachineLoopInfo) 63 INITIALIZE_PASS_DEPENDENCY(PHIElimination) 64 INITIALIZE_PASS_DEPENDENCY(TwoAddressInstructionPass) 65 INITIALIZE_PASS_DEPENDENCY(ProcessImplicitDefs) 66 INITIALIZE_PASS_DEPENDENCY(SlotIndexes) 67 INITIALIZE_AG_DEPENDENCY(AliasAnalysis) 68 INITIALIZE_PASS_END(LiveIntervals, "liveintervals", 69 "Live Interval Analysis", false, false) 70 71 void LiveIntervals::getAnalysisUsage(AnalysisUsage &AU) const { 72 AU.setPreservesCFG(); 73 AU.addRequired<AliasAnalysis>(); 74 AU.addPreserved<AliasAnalysis>(); 75 AU.addRequired<LiveVariables>(); 76 AU.addPreserved<LiveVariables>(); 77 AU.addRequired<MachineLoopInfo>(); 78 AU.addPreserved<MachineLoopInfo>(); 79 AU.addPreservedID(MachineDominatorsID); 80 81 if (!StrongPHIElim) { 82 AU.addPreservedID(PHIEliminationID); 83 AU.addRequiredID(PHIEliminationID); 84 } 85 86 AU.addRequiredID(TwoAddressInstructionPassID); 87 AU.addPreserved<ProcessImplicitDefs>(); 88 AU.addRequired<ProcessImplicitDefs>(); 89 AU.addPreserved<SlotIndexes>(); 90 AU.addRequiredTransitive<SlotIndexes>(); 91 MachineFunctionPass::getAnalysisUsage(AU); 92 } 93 94 void LiveIntervals::releaseMemory() { 95 // Free the live intervals themselves. 96 for (DenseMap<unsigned, LiveInterval*>::iterator I = r2iMap_.begin(), 97 E = r2iMap_.end(); I != E; ++I) 98 delete I->second; 99 100 r2iMap_.clear(); 101 102 // Release VNInfo memory regions, VNInfo objects don't need to be dtor'd. 103 VNInfoAllocator.Reset(); 104 while (!CloneMIs.empty()) { 105 MachineInstr *MI = CloneMIs.back(); 106 CloneMIs.pop_back(); 107 mf_->DeleteMachineInstr(MI); 108 } 109 } 110 111 /// runOnMachineFunction - Register allocate the whole function 112 /// 113 bool LiveIntervals::runOnMachineFunction(MachineFunction &fn) { 114 mf_ = &fn; 115 mri_ = &mf_->getRegInfo(); 116 tm_ = &fn.getTarget(); 117 tri_ = tm_->getRegisterInfo(); 118 tii_ = tm_->getInstrInfo(); 119 aa_ = &getAnalysis<AliasAnalysis>(); 120 lv_ = &getAnalysis<LiveVariables>(); 121 indexes_ = &getAnalysis<SlotIndexes>(); 122 allocatableRegs_ = tri_->getAllocatableSet(fn); 123 124 computeIntervals(); 125 126 numIntervals += getNumIntervals(); 127 128 DEBUG(dump()); 129 return true; 130 } 131 132 /// print - Implement the dump method. 133 void LiveIntervals::print(raw_ostream &OS, const Module* ) const { 134 OS << "********** INTERVALS **********\n"; 135 for (const_iterator I = begin(), E = end(); I != E; ++I) { 136 I->second->print(OS, tri_); 137 OS << "\n"; 138 } 139 140 printInstrs(OS); 141 } 142 143 void LiveIntervals::printInstrs(raw_ostream &OS) const { 144 OS << "********** MACHINEINSTRS **********\n"; 145 mf_->print(OS, indexes_); 146 } 147 148 void LiveIntervals::dumpInstrs() const { 149 printInstrs(dbgs()); 150 } 151 152 bool LiveIntervals::conflictsWithPhysReg(const LiveInterval &li, 153 VirtRegMap &vrm, unsigned reg) { 154 // We don't handle fancy stuff crossing basic block boundaries 155 if (li.ranges.size() != 1) 156 return true; 157 const LiveRange &range = li.ranges.front(); 158 SlotIndex idx = range.start.getBaseIndex(); 159 SlotIndex end = range.end.getPrevSlot().getBaseIndex().getNextIndex(); 160 161 // Skip deleted instructions 162 MachineInstr *firstMI = getInstructionFromIndex(idx); 163 while (!firstMI && idx != end) { 164 idx = idx.getNextIndex(); 165 firstMI = getInstructionFromIndex(idx); 166 } 167 if (!firstMI) 168 return false; 169 170 // Find last instruction in range 171 SlotIndex lastIdx = end.getPrevIndex(); 172 MachineInstr *lastMI = getInstructionFromIndex(lastIdx); 173 while (!lastMI && lastIdx != idx) { 174 lastIdx = lastIdx.getPrevIndex(); 175 lastMI = getInstructionFromIndex(lastIdx); 176 } 177 if (!lastMI) 178 return false; 179 180 // Range cannot cross basic block boundaries or terminators 181 MachineBasicBlock *MBB = firstMI->getParent(); 182 if (MBB != lastMI->getParent() || lastMI->getDesc().isTerminator()) 183 return true; 184 185 MachineBasicBlock::const_iterator E = lastMI; 186 ++E; 187 for (MachineBasicBlock::const_iterator I = firstMI; I != E; ++I) { 188 const MachineInstr &MI = *I; 189 190 // Allow copies to and from li.reg 191 if (MI.isCopy()) 192 if (MI.getOperand(0).getReg() == li.reg || 193 MI.getOperand(1).getReg() == li.reg) 194 continue; 195 196 // Check for operands using reg 197 for (unsigned i = 0, e = MI.getNumOperands(); i != e; ++i) { 198 const MachineOperand& mop = MI.getOperand(i); 199 if (!mop.isReg()) 200 continue; 201 unsigned PhysReg = mop.getReg(); 202 if (PhysReg == 0 || PhysReg == li.reg) 203 continue; 204 if (TargetRegisterInfo::isVirtualRegister(PhysReg)) { 205 if (!vrm.hasPhys(PhysReg)) 206 continue; 207 PhysReg = vrm.getPhys(PhysReg); 208 } 209 if (PhysReg && tri_->regsOverlap(PhysReg, reg)) 210 return true; 211 } 212 } 213 214 // No conflicts found. 215 return false; 216 } 217 218 bool LiveIntervals::conflictsWithAliasRef(LiveInterval &li, unsigned Reg, 219 SmallPtrSet<MachineInstr*,32> &JoinedCopies) { 220 for (LiveInterval::Ranges::const_iterator 221 I = li.ranges.begin(), E = li.ranges.end(); I != E; ++I) { 222 for (SlotIndex index = I->start.getBaseIndex(), 223 end = I->end.getPrevSlot().getBaseIndex().getNextIndex(); 224 index != end; 225 index = index.getNextIndex()) { 226 MachineInstr *MI = getInstructionFromIndex(index); 227 if (!MI) 228 continue; // skip deleted instructions 229 230 if (JoinedCopies.count(MI)) 231 continue; 232 for (unsigned i = 0, e = MI->getNumOperands(); i != e; ++i) { 233 MachineOperand& MO = MI->getOperand(i); 234 if (!MO.isReg()) 235 continue; 236 unsigned PhysReg = MO.getReg(); 237 if (PhysReg == 0 || PhysReg == Reg || 238 TargetRegisterInfo::isVirtualRegister(PhysReg)) 239 continue; 240 if (tri_->regsOverlap(Reg, PhysReg)) 241 return true; 242 } 243 } 244 } 245 246 return false; 247 } 248 249 static 250 bool MultipleDefsBySameMI(const MachineInstr &MI, unsigned MOIdx) { 251 unsigned Reg = MI.getOperand(MOIdx).getReg(); 252 for (unsigned i = MOIdx+1, e = MI.getNumOperands(); i < e; ++i) { 253 const MachineOperand &MO = MI.getOperand(i); 254 if (!MO.isReg()) 255 continue; 256 if (MO.getReg() == Reg && MO.isDef()) { 257 assert(MI.getOperand(MOIdx).getSubReg() != MO.getSubReg() && 258 MI.getOperand(MOIdx).getSubReg() && 259 (MO.getSubReg() || MO.isImplicit())); 260 return true; 261 } 262 } 263 return false; 264 } 265 266 /// isPartialRedef - Return true if the specified def at the specific index is 267 /// partially re-defining the specified live interval. A common case of this is 268 /// a definition of the sub-register. 269 bool LiveIntervals::isPartialRedef(SlotIndex MIIdx, MachineOperand &MO, 270 LiveInterval &interval) { 271 if (!MO.getSubReg() || MO.isEarlyClobber()) 272 return false; 273 274 SlotIndex RedefIndex = MIIdx.getDefIndex(); 275 const LiveRange *OldLR = 276 interval.getLiveRangeContaining(RedefIndex.getUseIndex()); 277 MachineInstr *DefMI = getInstructionFromIndex(OldLR->valno->def); 278 if (DefMI != 0) { 279 return DefMI->findRegisterDefOperandIdx(interval.reg) != -1; 280 } 281 return false; 282 } 283 284 void LiveIntervals::handleVirtualRegisterDef(MachineBasicBlock *mbb, 285 MachineBasicBlock::iterator mi, 286 SlotIndex MIIdx, 287 MachineOperand& MO, 288 unsigned MOIdx, 289 LiveInterval &interval) { 290 DEBUG(dbgs() << "\t\tregister: " << PrintReg(interval.reg, tri_)); 291 292 // Virtual registers may be defined multiple times (due to phi 293 // elimination and 2-addr elimination). Much of what we do only has to be 294 // done once for the vreg. We use an empty interval to detect the first 295 // time we see a vreg. 296 LiveVariables::VarInfo& vi = lv_->getVarInfo(interval.reg); 297 if (interval.empty()) { 298 // Get the Idx of the defining instructions. 299 SlotIndex defIndex = MIIdx.getDefIndex(); 300 // Earlyclobbers move back one, so that they overlap the live range 301 // of inputs. 302 if (MO.isEarlyClobber()) 303 defIndex = MIIdx.getUseIndex(); 304 305 // Make sure the first definition is not a partial redefinition. Add an 306 // <imp-def> of the full register. 307 if (MO.getSubReg()) 308 mi->addRegisterDefined(interval.reg); 309 310 MachineInstr *CopyMI = NULL; 311 if (mi->isCopyLike()) { 312 CopyMI = mi; 313 } 314 315 VNInfo *ValNo = interval.getNextValue(defIndex, CopyMI, VNInfoAllocator); 316 assert(ValNo->id == 0 && "First value in interval is not 0?"); 317 318 // Loop over all of the blocks that the vreg is defined in. There are 319 // two cases we have to handle here. The most common case is a vreg 320 // whose lifetime is contained within a basic block. In this case there 321 // will be a single kill, in MBB, which comes after the definition. 322 if (vi.Kills.size() == 1 && vi.Kills[0]->getParent() == mbb) { 323 // FIXME: what about dead vars? 324 SlotIndex killIdx; 325 if (vi.Kills[0] != mi) 326 killIdx = getInstructionIndex(vi.Kills[0]).getDefIndex(); 327 else 328 killIdx = defIndex.getStoreIndex(); 329 330 // If the kill happens after the definition, we have an intra-block 331 // live range. 332 if (killIdx > defIndex) { 333 assert(vi.AliveBlocks.empty() && 334 "Shouldn't be alive across any blocks!"); 335 LiveRange LR(defIndex, killIdx, ValNo); 336 interval.addRange(LR); 337 DEBUG(dbgs() << " +" << LR << "\n"); 338 return; 339 } 340 } 341 342 // The other case we handle is when a virtual register lives to the end 343 // of the defining block, potentially live across some blocks, then is 344 // live into some number of blocks, but gets killed. Start by adding a 345 // range that goes from this definition to the end of the defining block. 346 LiveRange NewLR(defIndex, getMBBEndIdx(mbb), ValNo); 347 DEBUG(dbgs() << " +" << NewLR); 348 interval.addRange(NewLR); 349 350 bool PHIJoin = lv_->isPHIJoin(interval.reg); 351 352 if (PHIJoin) { 353 // A phi join register is killed at the end of the MBB and revived as a new 354 // valno in the killing blocks. 355 assert(vi.AliveBlocks.empty() && "Phi join can't pass through blocks"); 356 DEBUG(dbgs() << " phi-join"); 357 ValNo->setHasPHIKill(true); 358 } else { 359 // Iterate over all of the blocks that the variable is completely 360 // live in, adding [insrtIndex(begin), instrIndex(end)+4) to the 361 // live interval. 362 for (SparseBitVector<>::iterator I = vi.AliveBlocks.begin(), 363 E = vi.AliveBlocks.end(); I != E; ++I) { 364 MachineBasicBlock *aliveBlock = mf_->getBlockNumbered(*I); 365 LiveRange LR(getMBBStartIdx(aliveBlock), getMBBEndIdx(aliveBlock), ValNo); 366 interval.addRange(LR); 367 DEBUG(dbgs() << " +" << LR); 368 } 369 } 370 371 // Finally, this virtual register is live from the start of any killing 372 // block to the 'use' slot of the killing instruction. 373 for (unsigned i = 0, e = vi.Kills.size(); i != e; ++i) { 374 MachineInstr *Kill = vi.Kills[i]; 375 SlotIndex Start = getMBBStartIdx(Kill->getParent()); 376 SlotIndex killIdx = getInstructionIndex(Kill).getDefIndex(); 377 378 // Create interval with one of a NEW value number. Note that this value 379 // number isn't actually defined by an instruction, weird huh? :) 380 if (PHIJoin) { 381 assert(getInstructionFromIndex(Start) == 0 && 382 "PHI def index points at actual instruction."); 383 ValNo = interval.getNextValue(Start, 0, VNInfoAllocator); 384 ValNo->setIsPHIDef(true); 385 } 386 LiveRange LR(Start, killIdx, ValNo); 387 interval.addRange(LR); 388 DEBUG(dbgs() << " +" << LR); 389 } 390 391 } else { 392 if (MultipleDefsBySameMI(*mi, MOIdx)) 393 // Multiple defs of the same virtual register by the same instruction. 394 // e.g. %reg1031:5<def>, %reg1031:6<def> = VLD1q16 %reg1024<kill>, ... 395 // This is likely due to elimination of REG_SEQUENCE instructions. Return 396 // here since there is nothing to do. 397 return; 398 399 // If this is the second time we see a virtual register definition, it 400 // must be due to phi elimination or two addr elimination. If this is 401 // the result of two address elimination, then the vreg is one of the 402 // def-and-use register operand. 403 404 // It may also be partial redef like this: 405 // 80 %reg1041:6<def> = VSHRNv4i16 %reg1034<kill>, 12, pred:14, pred:%reg0 406 // 120 %reg1041:5<def> = VSHRNv4i16 %reg1039<kill>, 12, pred:14, pred:%reg0 407 bool PartReDef = isPartialRedef(MIIdx, MO, interval); 408 if (PartReDef || mi->isRegTiedToUseOperand(MOIdx)) { 409 // If this is a two-address definition, then we have already processed 410 // the live range. The only problem is that we didn't realize there 411 // are actually two values in the live interval. Because of this we 412 // need to take the LiveRegion that defines this register and split it 413 // into two values. 414 SlotIndex RedefIndex = MIIdx.getDefIndex(); 415 if (MO.isEarlyClobber()) 416 RedefIndex = MIIdx.getUseIndex(); 417 418 const LiveRange *OldLR = 419 interval.getLiveRangeContaining(RedefIndex.getUseIndex()); 420 VNInfo *OldValNo = OldLR->valno; 421 SlotIndex DefIndex = OldValNo->def.getDefIndex(); 422 423 // Delete the previous value, which should be short and continuous, 424 // because the 2-addr copy must be in the same MBB as the redef. 425 interval.removeRange(DefIndex, RedefIndex); 426 427 // The new value number (#1) is defined by the instruction we claimed 428 // defined value #0. 429 VNInfo *ValNo = interval.createValueCopy(OldValNo, VNInfoAllocator); 430 431 // Value#0 is now defined by the 2-addr instruction. 432 OldValNo->def = RedefIndex; 433 OldValNo->setCopy(0); 434 435 // A re-def may be a copy. e.g. %reg1030:6<def> = VMOVD %reg1026, ... 436 if (PartReDef && mi->isCopyLike()) 437 OldValNo->setCopy(&*mi); 438 439 // Add the new live interval which replaces the range for the input copy. 440 LiveRange LR(DefIndex, RedefIndex, ValNo); 441 DEBUG(dbgs() << " replace range with " << LR); 442 interval.addRange(LR); 443 444 // If this redefinition is dead, we need to add a dummy unit live 445 // range covering the def slot. 446 if (MO.isDead()) 447 interval.addRange(LiveRange(RedefIndex, RedefIndex.getStoreIndex(), 448 OldValNo)); 449 450 DEBUG({ 451 dbgs() << " RESULT: "; 452 interval.print(dbgs(), tri_); 453 }); 454 } else if (lv_->isPHIJoin(interval.reg)) { 455 // In the case of PHI elimination, each variable definition is only 456 // live until the end of the block. We've already taken care of the 457 // rest of the live range. 458 459 SlotIndex defIndex = MIIdx.getDefIndex(); 460 if (MO.isEarlyClobber()) 461 defIndex = MIIdx.getUseIndex(); 462 463 VNInfo *ValNo; 464 MachineInstr *CopyMI = NULL; 465 if (mi->isCopyLike()) 466 CopyMI = mi; 467 ValNo = interval.getNextValue(defIndex, CopyMI, VNInfoAllocator); 468 469 SlotIndex killIndex = getMBBEndIdx(mbb); 470 LiveRange LR(defIndex, killIndex, ValNo); 471 interval.addRange(LR); 472 ValNo->setHasPHIKill(true); 473 DEBUG(dbgs() << " phi-join +" << LR); 474 } else { 475 llvm_unreachable("Multiply defined register"); 476 } 477 } 478 479 DEBUG(dbgs() << '\n'); 480 } 481 482 void LiveIntervals::handlePhysicalRegisterDef(MachineBasicBlock *MBB, 483 MachineBasicBlock::iterator mi, 484 SlotIndex MIIdx, 485 MachineOperand& MO, 486 LiveInterval &interval, 487 MachineInstr *CopyMI) { 488 // A physical register cannot be live across basic block, so its 489 // lifetime must end somewhere in its defining basic block. 490 DEBUG(dbgs() << "\t\tregister: " << PrintReg(interval.reg, tri_)); 491 492 SlotIndex baseIndex = MIIdx; 493 SlotIndex start = baseIndex.getDefIndex(); 494 // Earlyclobbers move back one. 495 if (MO.isEarlyClobber()) 496 start = MIIdx.getUseIndex(); 497 SlotIndex end = start; 498 499 // If it is not used after definition, it is considered dead at 500 // the instruction defining it. Hence its interval is: 501 // [defSlot(def), defSlot(def)+1) 502 // For earlyclobbers, the defSlot was pushed back one; the extra 503 // advance below compensates. 504 if (MO.isDead()) { 505 DEBUG(dbgs() << " dead"); 506 end = start.getStoreIndex(); 507 goto exit; 508 } 509 510 // If it is not dead on definition, it must be killed by a 511 // subsequent instruction. Hence its interval is: 512 // [defSlot(def), useSlot(kill)+1) 513 baseIndex = baseIndex.getNextIndex(); 514 while (++mi != MBB->end()) { 515 516 if (mi->isDebugValue()) 517 continue; 518 if (getInstructionFromIndex(baseIndex) == 0) 519 baseIndex = indexes_->getNextNonNullIndex(baseIndex); 520 521 if (mi->killsRegister(interval.reg, tri_)) { 522 DEBUG(dbgs() << " killed"); 523 end = baseIndex.getDefIndex(); 524 goto exit; 525 } else { 526 int DefIdx = mi->findRegisterDefOperandIdx(interval.reg,false,false,tri_); 527 if (DefIdx != -1) { 528 if (mi->isRegTiedToUseOperand(DefIdx)) { 529 // Two-address instruction. 530 end = baseIndex.getDefIndex(); 531 } else { 532 // Another instruction redefines the register before it is ever read. 533 // Then the register is essentially dead at the instruction that 534 // defines it. Hence its interval is: 535 // [defSlot(def), defSlot(def)+1) 536 DEBUG(dbgs() << " dead"); 537 end = start.getStoreIndex(); 538 } 539 goto exit; 540 } 541 } 542 543 baseIndex = baseIndex.getNextIndex(); 544 } 545 546 // The only case we should have a dead physreg here without a killing or 547 // instruction where we know it's dead is if it is live-in to the function 548 // and never used. Another possible case is the implicit use of the 549 // physical register has been deleted by two-address pass. 550 end = start.getStoreIndex(); 551 552 exit: 553 assert(start < end && "did not find end of interval?"); 554 555 // Already exists? Extend old live interval. 556 VNInfo *ValNo = interval.getVNInfoAt(start); 557 bool Extend = ValNo != 0; 558 if (!Extend) 559 ValNo = interval.getNextValue(start, CopyMI, VNInfoAllocator); 560 if (Extend && MO.isEarlyClobber()) 561 ValNo->setHasRedefByEC(true); 562 LiveRange LR(start, end, ValNo); 563 interval.addRange(LR); 564 DEBUG(dbgs() << " +" << LR << '\n'); 565 } 566 567 void LiveIntervals::handleRegisterDef(MachineBasicBlock *MBB, 568 MachineBasicBlock::iterator MI, 569 SlotIndex MIIdx, 570 MachineOperand& MO, 571 unsigned MOIdx) { 572 if (TargetRegisterInfo::isVirtualRegister(MO.getReg())) 573 handleVirtualRegisterDef(MBB, MI, MIIdx, MO, MOIdx, 574 getOrCreateInterval(MO.getReg())); 575 else { 576 MachineInstr *CopyMI = NULL; 577 if (MI->isCopyLike()) 578 CopyMI = MI; 579 handlePhysicalRegisterDef(MBB, MI, MIIdx, MO, 580 getOrCreateInterval(MO.getReg()), CopyMI); 581 } 582 } 583 584 void LiveIntervals::handleLiveInRegister(MachineBasicBlock *MBB, 585 SlotIndex MIIdx, 586 LiveInterval &interval, bool isAlias) { 587 DEBUG(dbgs() << "\t\tlivein register: " << PrintReg(interval.reg, tri_)); 588 589 // Look for kills, if it reaches a def before it's killed, then it shouldn't 590 // be considered a livein. 591 MachineBasicBlock::iterator mi = MBB->begin(); 592 MachineBasicBlock::iterator E = MBB->end(); 593 // Skip over DBG_VALUE at the start of the MBB. 594 if (mi != E && mi->isDebugValue()) { 595 while (++mi != E && mi->isDebugValue()) 596 ; 597 if (mi == E) 598 // MBB is empty except for DBG_VALUE's. 599 return; 600 } 601 602 SlotIndex baseIndex = MIIdx; 603 SlotIndex start = baseIndex; 604 if (getInstructionFromIndex(baseIndex) == 0) 605 baseIndex = indexes_->getNextNonNullIndex(baseIndex); 606 607 SlotIndex end = baseIndex; 608 bool SeenDefUse = false; 609 610 while (mi != E) { 611 if (mi->killsRegister(interval.reg, tri_)) { 612 DEBUG(dbgs() << " killed"); 613 end = baseIndex.getDefIndex(); 614 SeenDefUse = true; 615 break; 616 } else if (mi->definesRegister(interval.reg, tri_)) { 617 // Another instruction redefines the register before it is ever read. 618 // Then the register is essentially dead at the instruction that defines 619 // it. Hence its interval is: 620 // [defSlot(def), defSlot(def)+1) 621 DEBUG(dbgs() << " dead"); 622 end = start.getStoreIndex(); 623 SeenDefUse = true; 624 break; 625 } 626 627 while (++mi != E && mi->isDebugValue()) 628 // Skip over DBG_VALUE. 629 ; 630 if (mi != E) 631 baseIndex = indexes_->getNextNonNullIndex(baseIndex); 632 } 633 634 // Live-in register might not be used at all. 635 if (!SeenDefUse) { 636 if (isAlias) { 637 DEBUG(dbgs() << " dead"); 638 end = MIIdx.getStoreIndex(); 639 } else { 640 DEBUG(dbgs() << " live through"); 641 end = getMBBEndIdx(MBB); 642 } 643 } 644 645 SlotIndex defIdx = getMBBStartIdx(MBB); 646 assert(getInstructionFromIndex(defIdx) == 0 && 647 "PHI def index points at actual instruction."); 648 VNInfo *vni = 649 interval.getNextValue(defIdx, 0, VNInfoAllocator); 650 vni->setIsPHIDef(true); 651 LiveRange LR(start, end, vni); 652 653 interval.addRange(LR); 654 DEBUG(dbgs() << " +" << LR << '\n'); 655 } 656 657 /// computeIntervals - computes the live intervals for virtual 658 /// registers. for some ordering of the machine instructions [1,N] a 659 /// live interval is an interval [i, j) where 1 <= i <= j < N for 660 /// which a variable is live 661 void LiveIntervals::computeIntervals() { 662 DEBUG(dbgs() << "********** COMPUTING LIVE INTERVALS **********\n" 663 << "********** Function: " 664 << ((Value*)mf_->getFunction())->getName() << '\n'); 665 666 SmallVector<unsigned, 8> UndefUses; 667 for (MachineFunction::iterator MBBI = mf_->begin(), E = mf_->end(); 668 MBBI != E; ++MBBI) { 669 MachineBasicBlock *MBB = MBBI; 670 if (MBB->empty()) 671 continue; 672 673 // Track the index of the current machine instr. 674 SlotIndex MIIndex = getMBBStartIdx(MBB); 675 DEBUG(dbgs() << "BB#" << MBB->getNumber() 676 << ":\t\t# derived from " << MBB->getName() << "\n"); 677 678 // Create intervals for live-ins to this BB first. 679 for (MachineBasicBlock::livein_iterator LI = MBB->livein_begin(), 680 LE = MBB->livein_end(); LI != LE; ++LI) { 681 handleLiveInRegister(MBB, MIIndex, getOrCreateInterval(*LI)); 682 // Multiple live-ins can alias the same register. 683 for (const unsigned* AS = tri_->getSubRegisters(*LI); *AS; ++AS) 684 if (!hasInterval(*AS)) 685 handleLiveInRegister(MBB, MIIndex, getOrCreateInterval(*AS), 686 true); 687 } 688 689 // Skip over empty initial indices. 690 if (getInstructionFromIndex(MIIndex) == 0) 691 MIIndex = indexes_->getNextNonNullIndex(MIIndex); 692 693 for (MachineBasicBlock::iterator MI = MBB->begin(), miEnd = MBB->end(); 694 MI != miEnd; ++MI) { 695 DEBUG(dbgs() << MIIndex << "\t" << *MI); 696 if (MI->isDebugValue()) 697 continue; 698 699 // Handle defs. 700 for (int i = MI->getNumOperands() - 1; i >= 0; --i) { 701 MachineOperand &MO = MI->getOperand(i); 702 if (!MO.isReg() || !MO.getReg()) 703 continue; 704 705 // handle register defs - build intervals 706 if (MO.isDef()) 707 handleRegisterDef(MBB, MI, MIIndex, MO, i); 708 else if (MO.isUndef()) 709 UndefUses.push_back(MO.getReg()); 710 } 711 712 // Move to the next instr slot. 713 MIIndex = indexes_->getNextNonNullIndex(MIIndex); 714 } 715 } 716 717 // Create empty intervals for registers defined by implicit_def's (except 718 // for those implicit_def that define values which are liveout of their 719 // blocks. 720 for (unsigned i = 0, e = UndefUses.size(); i != e; ++i) { 721 unsigned UndefReg = UndefUses[i]; 722 (void)getOrCreateInterval(UndefReg); 723 } 724 } 725 726 LiveInterval* LiveIntervals::createInterval(unsigned reg) { 727 float Weight = TargetRegisterInfo::isPhysicalRegister(reg) ? HUGE_VALF : 0.0F; 728 return new LiveInterval(reg, Weight); 729 } 730 731 /// dupInterval - Duplicate a live interval. The caller is responsible for 732 /// managing the allocated memory. 733 LiveInterval* LiveIntervals::dupInterval(LiveInterval *li) { 734 LiveInterval *NewLI = createInterval(li->reg); 735 NewLI->Copy(*li, mri_, getVNInfoAllocator()); 736 return NewLI; 737 } 738 739 /// shrinkToUses - After removing some uses of a register, shrink its live 740 /// range to just the remaining uses. This method does not compute reaching 741 /// defs for new uses, and it doesn't remove dead defs. 742 bool LiveIntervals::shrinkToUses(LiveInterval *li, 743 SmallVectorImpl<MachineInstr*> *dead) { 744 DEBUG(dbgs() << "Shrink: " << *li << '\n'); 745 assert(TargetRegisterInfo::isVirtualRegister(li->reg) 746 && "Can't only shrink physical registers"); 747 // Find all the values used, including PHI kills. 748 SmallVector<std::pair<SlotIndex, VNInfo*>, 16> WorkList; 749 750 // Visit all instructions reading li->reg. 751 for (MachineRegisterInfo::reg_iterator I = mri_->reg_begin(li->reg); 752 MachineInstr *UseMI = I.skipInstruction();) { 753 if (UseMI->isDebugValue() || !UseMI->readsVirtualRegister(li->reg)) 754 continue; 755 SlotIndex Idx = getInstructionIndex(UseMI).getUseIndex(); 756 VNInfo *VNI = li->getVNInfoAt(Idx); 757 if (!VNI) { 758 // This shouldn't happen: readsVirtualRegister returns true, but there is 759 // no live value. It is likely caused by a target getting <undef> flags 760 // wrong. 761 DEBUG(dbgs() << Idx << '\t' << *UseMI 762 << "Warning: Instr claims to read non-existent value in " 763 << *li << '\n'); 764 continue; 765 } 766 if (VNI->def == Idx) { 767 // Special case: An early-clobber tied operand reads and writes the 768 // register one slot early. 769 Idx = Idx.getPrevSlot(); 770 VNI = li->getVNInfoAt(Idx); 771 assert(VNI && "Early-clobber tied value not available"); 772 } 773 WorkList.push_back(std::make_pair(Idx, VNI)); 774 } 775 776 // Create a new live interval with only minimal live segments per def. 777 LiveInterval NewLI(li->reg, 0); 778 for (LiveInterval::vni_iterator I = li->vni_begin(), E = li->vni_end(); 779 I != E; ++I) { 780 VNInfo *VNI = *I; 781 if (VNI->isUnused()) 782 continue; 783 // We may eliminate PHI values, so recompute PHIKill flags. 784 VNI->setHasPHIKill(false); 785 NewLI.addRange(LiveRange(VNI->def, VNI->def.getNextSlot(), VNI)); 786 787 // A use tied to an early-clobber def ends at the load slot and isn't caught 788 // above. Catch it here instead. This probably only ever happens for inline 789 // assembly. 790 if (VNI->def.isUse()) 791 if (VNInfo *UVNI = li->getVNInfoAt(VNI->def.getLoadIndex())) 792 WorkList.push_back(std::make_pair(VNI->def.getLoadIndex(), UVNI)); 793 } 794 795 // Keep track of the PHIs that are in use. 796 SmallPtrSet<VNInfo*, 8> UsedPHIs; 797 798 // Extend intervals to reach all uses in WorkList. 799 while (!WorkList.empty()) { 800 SlotIndex Idx = WorkList.back().first; 801 VNInfo *VNI = WorkList.back().second; 802 WorkList.pop_back(); 803 const MachineBasicBlock *MBB = getMBBFromIndex(Idx); 804 SlotIndex BlockStart = getMBBStartIdx(MBB); 805 806 // Extend the live range for VNI to be live at Idx. 807 if (VNInfo *ExtVNI = NewLI.extendInBlock(BlockStart, Idx)) { 808 (void)ExtVNI; 809 assert(ExtVNI == VNI && "Unexpected existing value number"); 810 // Is this a PHIDef we haven't seen before? 811 if (!VNI->isPHIDef() || VNI->def != BlockStart || !UsedPHIs.insert(VNI)) 812 continue; 813 // The PHI is live, make sure the predecessors are live-out. 814 for (MachineBasicBlock::const_pred_iterator PI = MBB->pred_begin(), 815 PE = MBB->pred_end(); PI != PE; ++PI) { 816 SlotIndex Stop = getMBBEndIdx(*PI).getPrevSlot(); 817 VNInfo *PVNI = li->getVNInfoAt(Stop); 818 // A predecessor is not required to have a live-out value for a PHI. 819 if (PVNI) { 820 PVNI->setHasPHIKill(true); 821 WorkList.push_back(std::make_pair(Stop, PVNI)); 822 } 823 } 824 continue; 825 } 826 827 // VNI is live-in to MBB. 828 DEBUG(dbgs() << " live-in at " << BlockStart << '\n'); 829 NewLI.addRange(LiveRange(BlockStart, Idx.getNextSlot(), VNI)); 830 831 // Make sure VNI is live-out from the predecessors. 832 for (MachineBasicBlock::const_pred_iterator PI = MBB->pred_begin(), 833 PE = MBB->pred_end(); PI != PE; ++PI) { 834 SlotIndex Stop = getMBBEndIdx(*PI).getPrevSlot(); 835 assert(li->getVNInfoAt(Stop) == VNI && "Wrong value out of predecessor"); 836 WorkList.push_back(std::make_pair(Stop, VNI)); 837 } 838 } 839 840 // Handle dead values. 841 bool CanSeparate = false; 842 for (LiveInterval::vni_iterator I = li->vni_begin(), E = li->vni_end(); 843 I != E; ++I) { 844 VNInfo *VNI = *I; 845 if (VNI->isUnused()) 846 continue; 847 LiveInterval::iterator LII = NewLI.FindLiveRangeContaining(VNI->def); 848 assert(LII != NewLI.end() && "Missing live range for PHI"); 849 if (LII->end != VNI->def.getNextSlot()) 850 continue; 851 if (VNI->isPHIDef()) { 852 // This is a dead PHI. Remove it. 853 VNI->setIsUnused(true); 854 NewLI.removeRange(*LII); 855 DEBUG(dbgs() << "Dead PHI at " << VNI->def << " may separate interval\n"); 856 CanSeparate = true; 857 } else { 858 // This is a dead def. Make sure the instruction knows. 859 MachineInstr *MI = getInstructionFromIndex(VNI->def); 860 assert(MI && "No instruction defining live value"); 861 MI->addRegisterDead(li->reg, tri_); 862 if (dead && MI->allDefsAreDead()) { 863 DEBUG(dbgs() << "All defs dead: " << VNI->def << '\t' << *MI); 864 dead->push_back(MI); 865 } 866 } 867 } 868 869 // Move the trimmed ranges back. 870 li->ranges.swap(NewLI.ranges); 871 DEBUG(dbgs() << "Shrunk: " << *li << '\n'); 872 return CanSeparate; 873 } 874 875 876 //===----------------------------------------------------------------------===// 877 // Register allocator hooks. 878 // 879 880 MachineBasicBlock::iterator 881 LiveIntervals::getLastSplitPoint(const LiveInterval &li, 882 MachineBasicBlock *mbb) const { 883 const MachineBasicBlock *lpad = mbb->getLandingPadSuccessor(); 884 885 // If li is not live into a landing pad, we can insert spill code before the 886 // first terminator. 887 if (!lpad || !isLiveInToMBB(li, lpad)) 888 return mbb->getFirstTerminator(); 889 890 // When there is a landing pad, spill code must go before the call instruction 891 // that can throw. 892 MachineBasicBlock::iterator I = mbb->end(), B = mbb->begin(); 893 while (I != B) { 894 --I; 895 if (I->getDesc().isCall()) 896 return I; 897 } 898 // The block contains no calls that can throw, so use the first terminator. 899 return mbb->getFirstTerminator(); 900 } 901 902 void LiveIntervals::addKillFlags() { 903 for (iterator I = begin(), E = end(); I != E; ++I) { 904 unsigned Reg = I->first; 905 if (TargetRegisterInfo::isPhysicalRegister(Reg)) 906 continue; 907 if (mri_->reg_nodbg_empty(Reg)) 908 continue; 909 LiveInterval *LI = I->second; 910 911 // Every instruction that kills Reg corresponds to a live range end point. 912 for (LiveInterval::iterator RI = LI->begin(), RE = LI->end(); RI != RE; 913 ++RI) { 914 // A LOAD index indicates an MBB edge. 915 if (RI->end.isLoad()) 916 continue; 917 MachineInstr *MI = getInstructionFromIndex(RI->end); 918 if (!MI) 919 continue; 920 MI->addRegisterKilled(Reg, NULL); 921 } 922 } 923 } 924 925 /// getReMatImplicitUse - If the remat definition MI has one (for now, we only 926 /// allow one) virtual register operand, then its uses are implicitly using 927 /// the register. Returns the virtual register. 928 unsigned LiveIntervals::getReMatImplicitUse(const LiveInterval &li, 929 MachineInstr *MI) const { 930 unsigned RegOp = 0; 931 for (unsigned i = 0, e = MI->getNumOperands(); i != e; ++i) { 932 MachineOperand &MO = MI->getOperand(i); 933 if (!MO.isReg() || !MO.isUse()) 934 continue; 935 unsigned Reg = MO.getReg(); 936 if (Reg == 0 || Reg == li.reg) 937 continue; 938 939 if (TargetRegisterInfo::isPhysicalRegister(Reg) && 940 !allocatableRegs_[Reg]) 941 continue; 942 // FIXME: For now, only remat MI with at most one register operand. 943 assert(!RegOp && 944 "Can't rematerialize instruction with multiple register operand!"); 945 RegOp = MO.getReg(); 946 #ifndef NDEBUG 947 break; 948 #endif 949 } 950 return RegOp; 951 } 952 953 /// isValNoAvailableAt - Return true if the val# of the specified interval 954 /// which reaches the given instruction also reaches the specified use index. 955 bool LiveIntervals::isValNoAvailableAt(const LiveInterval &li, MachineInstr *MI, 956 SlotIndex UseIdx) const { 957 VNInfo *UValNo = li.getVNInfoAt(UseIdx); 958 return UValNo && UValNo == li.getVNInfoAt(getInstructionIndex(MI)); 959 } 960 961 /// isReMaterializable - Returns true if the definition MI of the specified 962 /// val# of the specified interval is re-materializable. 963 bool 964 LiveIntervals::isReMaterializable(const LiveInterval &li, 965 const VNInfo *ValNo, MachineInstr *MI, 966 const SmallVectorImpl<LiveInterval*> *SpillIs, 967 bool &isLoad) { 968 if (DisableReMat) 969 return false; 970 971 if (!tii_->isTriviallyReMaterializable(MI, aa_)) 972 return false; 973 974 // Target-specific code can mark an instruction as being rematerializable 975 // if it has one virtual reg use, though it had better be something like 976 // a PIC base register which is likely to be live everywhere. 977 unsigned ImpUse = getReMatImplicitUse(li, MI); 978 if (ImpUse) { 979 const LiveInterval &ImpLi = getInterval(ImpUse); 980 for (MachineRegisterInfo::use_nodbg_iterator 981 ri = mri_->use_nodbg_begin(li.reg), re = mri_->use_nodbg_end(); 982 ri != re; ++ri) { 983 MachineInstr *UseMI = &*ri; 984 SlotIndex UseIdx = getInstructionIndex(UseMI); 985 if (li.getVNInfoAt(UseIdx) != ValNo) 986 continue; 987 if (!isValNoAvailableAt(ImpLi, MI, UseIdx)) 988 return false; 989 } 990 991 // If a register operand of the re-materialized instruction is going to 992 // be spilled next, then it's not legal to re-materialize this instruction. 993 if (SpillIs) 994 for (unsigned i = 0, e = SpillIs->size(); i != e; ++i) 995 if (ImpUse == (*SpillIs)[i]->reg) 996 return false; 997 } 998 return true; 999 } 1000 1001 /// isReMaterializable - Returns true if the definition MI of the specified 1002 /// val# of the specified interval is re-materializable. 1003 bool LiveIntervals::isReMaterializable(const LiveInterval &li, 1004 const VNInfo *ValNo, MachineInstr *MI) { 1005 bool Dummy2; 1006 return isReMaterializable(li, ValNo, MI, 0, Dummy2); 1007 } 1008 1009 /// isReMaterializable - Returns true if every definition of MI of every 1010 /// val# of the specified interval is re-materializable. 1011 bool 1012 LiveIntervals::isReMaterializable(const LiveInterval &li, 1013 const SmallVectorImpl<LiveInterval*> *SpillIs, 1014 bool &isLoad) { 1015 isLoad = false; 1016 for (LiveInterval::const_vni_iterator i = li.vni_begin(), e = li.vni_end(); 1017 i != e; ++i) { 1018 const VNInfo *VNI = *i; 1019 if (VNI->isUnused()) 1020 continue; // Dead val#. 1021 // Is the def for the val# rematerializable? 1022 MachineInstr *ReMatDefMI = getInstructionFromIndex(VNI->def); 1023 if (!ReMatDefMI) 1024 return false; 1025 bool DefIsLoad = false; 1026 if (!ReMatDefMI || 1027 !isReMaterializable(li, VNI, ReMatDefMI, SpillIs, DefIsLoad)) 1028 return false; 1029 isLoad |= DefIsLoad; 1030 } 1031 return true; 1032 } 1033 1034 /// FilterFoldedOps - Filter out two-address use operands. Return 1035 /// true if it finds any issue with the operands that ought to prevent 1036 /// folding. 1037 static bool FilterFoldedOps(MachineInstr *MI, 1038 SmallVector<unsigned, 2> &Ops, 1039 unsigned &MRInfo, 1040 SmallVector<unsigned, 2> &FoldOps) { 1041 MRInfo = 0; 1042 for (unsigned i = 0, e = Ops.size(); i != e; ++i) { 1043 unsigned OpIdx = Ops[i]; 1044 MachineOperand &MO = MI->getOperand(OpIdx); 1045 // FIXME: fold subreg use. 1046 if (MO.getSubReg()) 1047 return true; 1048 if (MO.isDef()) 1049 MRInfo |= (unsigned)VirtRegMap::isMod; 1050 else { 1051 // Filter out two-address use operand(s). 1052 if (MI->isRegTiedToDefOperand(OpIdx)) { 1053 MRInfo = VirtRegMap::isModRef; 1054 continue; 1055 } 1056 MRInfo |= (unsigned)VirtRegMap::isRef; 1057 } 1058 FoldOps.push_back(OpIdx); 1059 } 1060 return false; 1061 } 1062 1063 1064 /// tryFoldMemoryOperand - Attempts to fold either a spill / restore from 1065 /// slot / to reg or any rematerialized load into ith operand of specified 1066 /// MI. If it is successul, MI is updated with the newly created MI and 1067 /// returns true. 1068 bool LiveIntervals::tryFoldMemoryOperand(MachineInstr* &MI, 1069 VirtRegMap &vrm, MachineInstr *DefMI, 1070 SlotIndex InstrIdx, 1071 SmallVector<unsigned, 2> &Ops, 1072 bool isSS, int Slot, unsigned Reg) { 1073 // If it is an implicit def instruction, just delete it. 1074 if (MI->isImplicitDef()) { 1075 RemoveMachineInstrFromMaps(MI); 1076 vrm.RemoveMachineInstrFromMaps(MI); 1077 MI->eraseFromParent(); 1078 ++numFolds; 1079 return true; 1080 } 1081 1082 // Filter the list of operand indexes that are to be folded. Abort if 1083 // any operand will prevent folding. 1084 unsigned MRInfo = 0; 1085 SmallVector<unsigned, 2> FoldOps; 1086 if (FilterFoldedOps(MI, Ops, MRInfo, FoldOps)) 1087 return false; 1088 1089 // The only time it's safe to fold into a two address instruction is when 1090 // it's folding reload and spill from / into a spill stack slot. 1091 if (DefMI && (MRInfo & VirtRegMap::isMod)) 1092 return false; 1093 1094 MachineInstr *fmi = isSS ? tii_->foldMemoryOperand(MI, FoldOps, Slot) 1095 : tii_->foldMemoryOperand(MI, FoldOps, DefMI); 1096 if (fmi) { 1097 // Remember this instruction uses the spill slot. 1098 if (isSS) vrm.addSpillSlotUse(Slot, fmi); 1099 1100 // Attempt to fold the memory reference into the instruction. If 1101 // we can do this, we don't need to insert spill code. 1102 if (isSS && !mf_->getFrameInfo()->isImmutableObjectIndex(Slot)) 1103 vrm.virtFolded(Reg, MI, fmi, (VirtRegMap::ModRef)MRInfo); 1104 vrm.transferSpillPts(MI, fmi); 1105 vrm.transferRestorePts(MI, fmi); 1106 vrm.transferEmergencySpills(MI, fmi); 1107 ReplaceMachineInstrInMaps(MI, fmi); 1108 MI->eraseFromParent(); 1109 MI = fmi; 1110 ++numFolds; 1111 return true; 1112 } 1113 return false; 1114 } 1115 1116 /// canFoldMemoryOperand - Returns true if the specified load / store 1117 /// folding is possible. 1118 bool LiveIntervals::canFoldMemoryOperand(MachineInstr *MI, 1119 SmallVector<unsigned, 2> &Ops, 1120 bool ReMat) const { 1121 // Filter the list of operand indexes that are to be folded. Abort if 1122 // any operand will prevent folding. 1123 unsigned MRInfo = 0; 1124 SmallVector<unsigned, 2> FoldOps; 1125 if (FilterFoldedOps(MI, Ops, MRInfo, FoldOps)) 1126 return false; 1127 1128 // It's only legal to remat for a use, not a def. 1129 if (ReMat && (MRInfo & VirtRegMap::isMod)) 1130 return false; 1131 1132 return tii_->canFoldMemoryOperand(MI, FoldOps); 1133 } 1134 1135 bool LiveIntervals::intervalIsInOneMBB(const LiveInterval &li) const { 1136 LiveInterval::Ranges::const_iterator itr = li.ranges.begin(); 1137 1138 MachineBasicBlock *mbb = indexes_->getMBBCoveringRange(itr->start, itr->end); 1139 1140 if (mbb == 0) 1141 return false; 1142 1143 for (++itr; itr != li.ranges.end(); ++itr) { 1144 MachineBasicBlock *mbb2 = 1145 indexes_->getMBBCoveringRange(itr->start, itr->end); 1146 1147 if (mbb2 != mbb) 1148 return false; 1149 } 1150 1151 return true; 1152 } 1153 1154 /// rewriteImplicitOps - Rewrite implicit use operands of MI (i.e. uses of 1155 /// interval on to-be re-materialized operands of MI) with new register. 1156 void LiveIntervals::rewriteImplicitOps(const LiveInterval &li, 1157 MachineInstr *MI, unsigned NewVReg, 1158 VirtRegMap &vrm) { 1159 // There is an implicit use. That means one of the other operand is 1160 // being remat'ed and the remat'ed instruction has li.reg as an 1161 // use operand. Make sure we rewrite that as well. 1162 for (unsigned i = 0, e = MI->getNumOperands(); i != e; ++i) { 1163 MachineOperand &MO = MI->getOperand(i); 1164 if (!MO.isReg()) 1165 continue; 1166 unsigned Reg = MO.getReg(); 1167 if (!TargetRegisterInfo::isVirtualRegister(Reg)) 1168 continue; 1169 if (!vrm.isReMaterialized(Reg)) 1170 continue; 1171 MachineInstr *ReMatMI = vrm.getReMaterializedMI(Reg); 1172 MachineOperand *UseMO = ReMatMI->findRegisterUseOperand(li.reg); 1173 if (UseMO) 1174 UseMO->setReg(NewVReg); 1175 } 1176 } 1177 1178 /// rewriteInstructionForSpills, rewriteInstructionsForSpills - Helper functions 1179 /// for addIntervalsForSpills to rewrite uses / defs for the given live range. 1180 bool LiveIntervals:: 1181 rewriteInstructionForSpills(const LiveInterval &li, const VNInfo *VNI, 1182 bool TrySplit, SlotIndex index, SlotIndex end, 1183 MachineInstr *MI, 1184 MachineInstr *ReMatOrigDefMI, MachineInstr *ReMatDefMI, 1185 unsigned Slot, int LdSlot, 1186 bool isLoad, bool isLoadSS, bool DefIsReMat, bool CanDelete, 1187 VirtRegMap &vrm, 1188 const TargetRegisterClass* rc, 1189 SmallVector<int, 4> &ReMatIds, 1190 const MachineLoopInfo *loopInfo, 1191 unsigned &NewVReg, unsigned ImpUse, bool &HasDef, bool &HasUse, 1192 DenseMap<unsigned,unsigned> &MBBVRegsMap, 1193 std::vector<LiveInterval*> &NewLIs) { 1194 bool CanFold = false; 1195 RestartInstruction: 1196 for (unsigned i = 0; i != MI->getNumOperands(); ++i) { 1197 MachineOperand& mop = MI->getOperand(i); 1198 if (!mop.isReg()) 1199 continue; 1200 unsigned Reg = mop.getReg(); 1201 if (!TargetRegisterInfo::isVirtualRegister(Reg)) 1202 continue; 1203 if (Reg != li.reg) 1204 continue; 1205 1206 bool TryFold = !DefIsReMat; 1207 bool FoldSS = true; // Default behavior unless it's a remat. 1208 int FoldSlot = Slot; 1209 if (DefIsReMat) { 1210 // If this is the rematerializable definition MI itself and 1211 // all of its uses are rematerialized, simply delete it. 1212 if (MI == ReMatOrigDefMI && CanDelete) { 1213 DEBUG(dbgs() << "\t\t\t\tErasing re-materializable def: " 1214 << *MI << '\n'); 1215 RemoveMachineInstrFromMaps(MI); 1216 vrm.RemoveMachineInstrFromMaps(MI); 1217 MI->eraseFromParent(); 1218 break; 1219 } 1220 1221 // If def for this use can't be rematerialized, then try folding. 1222 // If def is rematerializable and it's a load, also try folding. 1223 TryFold = !ReMatDefMI || (ReMatDefMI && (MI == ReMatOrigDefMI || isLoad)); 1224 if (isLoad) { 1225 // Try fold loads (from stack slot, constant pool, etc.) into uses. 1226 FoldSS = isLoadSS; 1227 FoldSlot = LdSlot; 1228 } 1229 } 1230 1231 // Scan all of the operands of this instruction rewriting operands 1232 // to use NewVReg instead of li.reg as appropriate. We do this for 1233 // two reasons: 1234 // 1235 // 1. If the instr reads the same spilled vreg multiple times, we 1236 // want to reuse the NewVReg. 1237 // 2. If the instr is a two-addr instruction, we are required to 1238 // keep the src/dst regs pinned. 1239 // 1240 // Keep track of whether we replace a use and/or def so that we can 1241 // create the spill interval with the appropriate range. 1242 SmallVector<unsigned, 2> Ops; 1243 tie(HasUse, HasDef) = MI->readsWritesVirtualRegister(Reg, &Ops); 1244 1245 // Create a new virtual register for the spill interval. 1246 // Create the new register now so we can map the fold instruction 1247 // to the new register so when it is unfolded we get the correct 1248 // answer. 1249 bool CreatedNewVReg = false; 1250 if (NewVReg == 0) { 1251 NewVReg = mri_->createVirtualRegister(rc); 1252 vrm.grow(); 1253 CreatedNewVReg = true; 1254 1255 // The new virtual register should get the same allocation hints as the 1256 // old one. 1257 std::pair<unsigned, unsigned> Hint = mri_->getRegAllocationHint(Reg); 1258 if (Hint.first || Hint.second) 1259 mri_->setRegAllocationHint(NewVReg, Hint.first, Hint.second); 1260 } 1261 1262 if (!TryFold) 1263 CanFold = false; 1264 else { 1265 // Do not fold load / store here if we are splitting. We'll find an 1266 // optimal point to insert a load / store later. 1267 if (!TrySplit) { 1268 if (tryFoldMemoryOperand(MI, vrm, ReMatDefMI, index, 1269 Ops, FoldSS, FoldSlot, NewVReg)) { 1270 // Folding the load/store can completely change the instruction in 1271 // unpredictable ways, rescan it from the beginning. 1272 1273 if (FoldSS) { 1274 // We need to give the new vreg the same stack slot as the 1275 // spilled interval. 1276 vrm.assignVirt2StackSlot(NewVReg, FoldSlot); 1277 } 1278 1279 HasUse = false; 1280 HasDef = false; 1281 CanFold = false; 1282 if (isNotInMIMap(MI)) 1283 break; 1284 goto RestartInstruction; 1285 } 1286 } else { 1287 // We'll try to fold it later if it's profitable. 1288 CanFold = canFoldMemoryOperand(MI, Ops, DefIsReMat); 1289 } 1290 } 1291 1292 mop.setReg(NewVReg); 1293 if (mop.isImplicit()) 1294 rewriteImplicitOps(li, MI, NewVReg, vrm); 1295 1296 // Reuse NewVReg for other reads. 1297 bool HasEarlyClobber = false; 1298 for (unsigned j = 0, e = Ops.size(); j != e; ++j) { 1299 MachineOperand &mopj = MI->getOperand(Ops[j]); 1300 mopj.setReg(NewVReg); 1301 if (mopj.isImplicit()) 1302 rewriteImplicitOps(li, MI, NewVReg, vrm); 1303 if (mopj.isEarlyClobber()) 1304 HasEarlyClobber = true; 1305 } 1306 1307 if (CreatedNewVReg) { 1308 if (DefIsReMat) { 1309 vrm.setVirtIsReMaterialized(NewVReg, ReMatDefMI); 1310 if (ReMatIds[VNI->id] == VirtRegMap::MAX_STACK_SLOT) { 1311 // Each valnum may have its own remat id. 1312 ReMatIds[VNI->id] = vrm.assignVirtReMatId(NewVReg); 1313 } else { 1314 vrm.assignVirtReMatId(NewVReg, ReMatIds[VNI->id]); 1315 } 1316 if (!CanDelete || (HasUse && HasDef)) { 1317 // If this is a two-addr instruction then its use operands are 1318 // rematerializable but its def is not. It should be assigned a 1319 // stack slot. 1320 vrm.assignVirt2StackSlot(NewVReg, Slot); 1321 } 1322 } else { 1323 vrm.assignVirt2StackSlot(NewVReg, Slot); 1324 } 1325 } else if (HasUse && HasDef && 1326 vrm.getStackSlot(NewVReg) == VirtRegMap::NO_STACK_SLOT) { 1327 // If this interval hasn't been assigned a stack slot (because earlier 1328 // def is a deleted remat def), do it now. 1329 assert(Slot != VirtRegMap::NO_STACK_SLOT); 1330 vrm.assignVirt2StackSlot(NewVReg, Slot); 1331 } 1332 1333 // Re-matting an instruction with virtual register use. Add the 1334 // register as an implicit use on the use MI. 1335 if (DefIsReMat && ImpUse) 1336 MI->addOperand(MachineOperand::CreateReg(ImpUse, false, true)); 1337 1338 // Create a new register interval for this spill / remat. 1339 LiveInterval &nI = getOrCreateInterval(NewVReg); 1340 if (CreatedNewVReg) { 1341 NewLIs.push_back(&nI); 1342 MBBVRegsMap.insert(std::make_pair(MI->getParent()->getNumber(), NewVReg)); 1343 if (TrySplit) 1344 vrm.setIsSplitFromReg(NewVReg, li.reg); 1345 } 1346 1347 if (HasUse) { 1348 if (CreatedNewVReg) { 1349 LiveRange LR(index.getLoadIndex(), index.getDefIndex(), 1350 nI.getNextValue(SlotIndex(), 0, VNInfoAllocator)); 1351 DEBUG(dbgs() << " +" << LR); 1352 nI.addRange(LR); 1353 } else { 1354 // Extend the split live interval to this def / use. 1355 SlotIndex End = index.getDefIndex(); 1356 LiveRange LR(nI.ranges[nI.ranges.size()-1].end, End, 1357 nI.getValNumInfo(nI.getNumValNums()-1)); 1358 DEBUG(dbgs() << " +" << LR); 1359 nI.addRange(LR); 1360 } 1361 } 1362 if (HasDef) { 1363 // An early clobber starts at the use slot, except for an early clobber 1364 // tied to a use operand (yes, that is a thing). 1365 LiveRange LR(HasEarlyClobber && !HasUse ? 1366 index.getUseIndex() : index.getDefIndex(), 1367 index.getStoreIndex(), 1368 nI.getNextValue(SlotIndex(), 0, VNInfoAllocator)); 1369 DEBUG(dbgs() << " +" << LR); 1370 nI.addRange(LR); 1371 } 1372 1373 DEBUG({ 1374 dbgs() << "\t\t\t\tAdded new interval: "; 1375 nI.print(dbgs(), tri_); 1376 dbgs() << '\n'; 1377 }); 1378 } 1379 return CanFold; 1380 } 1381 bool LiveIntervals::anyKillInMBBAfterIdx(const LiveInterval &li, 1382 const VNInfo *VNI, 1383 MachineBasicBlock *MBB, 1384 SlotIndex Idx) const { 1385 return li.killedInRange(Idx.getNextSlot(), getMBBEndIdx(MBB)); 1386 } 1387 1388 /// RewriteInfo - Keep track of machine instrs that will be rewritten 1389 /// during spilling. 1390 namespace { 1391 struct RewriteInfo { 1392 SlotIndex Index; 1393 MachineInstr *MI; 1394 RewriteInfo(SlotIndex i, MachineInstr *mi) : Index(i), MI(mi) {} 1395 }; 1396 1397 struct RewriteInfoCompare { 1398 bool operator()(const RewriteInfo &LHS, const RewriteInfo &RHS) const { 1399 return LHS.Index < RHS.Index; 1400 } 1401 }; 1402 } 1403 1404 void LiveIntervals:: 1405 rewriteInstructionsForSpills(const LiveInterval &li, bool TrySplit, 1406 LiveInterval::Ranges::const_iterator &I, 1407 MachineInstr *ReMatOrigDefMI, MachineInstr *ReMatDefMI, 1408 unsigned Slot, int LdSlot, 1409 bool isLoad, bool isLoadSS, bool DefIsReMat, bool CanDelete, 1410 VirtRegMap &vrm, 1411 const TargetRegisterClass* rc, 1412 SmallVector<int, 4> &ReMatIds, 1413 const MachineLoopInfo *loopInfo, 1414 BitVector &SpillMBBs, 1415 DenseMap<unsigned, std::vector<SRInfo> > &SpillIdxes, 1416 BitVector &RestoreMBBs, 1417 DenseMap<unsigned, std::vector<SRInfo> > &RestoreIdxes, 1418 DenseMap<unsigned,unsigned> &MBBVRegsMap, 1419 std::vector<LiveInterval*> &NewLIs) { 1420 bool AllCanFold = true; 1421 unsigned NewVReg = 0; 1422 SlotIndex start = I->start.getBaseIndex(); 1423 SlotIndex end = I->end.getPrevSlot().getBaseIndex().getNextIndex(); 1424 1425 // First collect all the def / use in this live range that will be rewritten. 1426 // Make sure they are sorted according to instruction index. 1427 std::vector<RewriteInfo> RewriteMIs; 1428 for (MachineRegisterInfo::reg_iterator ri = mri_->reg_begin(li.reg), 1429 re = mri_->reg_end(); ri != re; ) { 1430 MachineInstr *MI = &*ri; 1431 MachineOperand &O = ri.getOperand(); 1432 ++ri; 1433 if (MI->isDebugValue()) { 1434 // Modify DBG_VALUE now that the value is in a spill slot. 1435 if (Slot != VirtRegMap::MAX_STACK_SLOT || isLoadSS) { 1436 uint64_t Offset = MI->getOperand(1).getImm(); 1437 const MDNode *MDPtr = MI->getOperand(2).getMetadata(); 1438 DebugLoc DL = MI->getDebugLoc(); 1439 int FI = isLoadSS ? LdSlot : (int)Slot; 1440 if (MachineInstr *NewDV = tii_->emitFrameIndexDebugValue(*mf_, FI, 1441 Offset, MDPtr, DL)) { 1442 DEBUG(dbgs() << "Modifying debug info due to spill:" << "\t" << *MI); 1443 ReplaceMachineInstrInMaps(MI, NewDV); 1444 MachineBasicBlock *MBB = MI->getParent(); 1445 MBB->insert(MBB->erase(MI), NewDV); 1446 continue; 1447 } 1448 } 1449 1450 DEBUG(dbgs() << "Removing debug info due to spill:" << "\t" << *MI); 1451 RemoveMachineInstrFromMaps(MI); 1452 vrm.RemoveMachineInstrFromMaps(MI); 1453 MI->eraseFromParent(); 1454 continue; 1455 } 1456 assert(!(O.isImplicit() && O.isUse()) && 1457 "Spilling register that's used as implicit use?"); 1458 SlotIndex index = getInstructionIndex(MI); 1459 if (index < start || index >= end) 1460 continue; 1461 1462 if (O.isUndef()) 1463 // Must be defined by an implicit def. It should not be spilled. Note, 1464 // this is for correctness reason. e.g. 1465 // 8 %reg1024<def> = IMPLICIT_DEF 1466 // 12 %reg1024<def> = INSERT_SUBREG %reg1024<kill>, %reg1025, 2 1467 // The live range [12, 14) are not part of the r1024 live interval since 1468 // it's defined by an implicit def. It will not conflicts with live 1469 // interval of r1025. Now suppose both registers are spilled, you can 1470 // easily see a situation where both registers are reloaded before 1471 // the INSERT_SUBREG and both target registers that would overlap. 1472 continue; 1473 RewriteMIs.push_back(RewriteInfo(index, MI)); 1474 } 1475 std::sort(RewriteMIs.begin(), RewriteMIs.end(), RewriteInfoCompare()); 1476 1477 unsigned ImpUse = DefIsReMat ? getReMatImplicitUse(li, ReMatDefMI) : 0; 1478 // Now rewrite the defs and uses. 1479 for (unsigned i = 0, e = RewriteMIs.size(); i != e; ) { 1480 RewriteInfo &rwi = RewriteMIs[i]; 1481 ++i; 1482 SlotIndex index = rwi.Index; 1483 MachineInstr *MI = rwi.MI; 1484 // If MI def and/or use the same register multiple times, then there 1485 // are multiple entries. 1486 while (i != e && RewriteMIs[i].MI == MI) { 1487 assert(RewriteMIs[i].Index == index); 1488 ++i; 1489 } 1490 MachineBasicBlock *MBB = MI->getParent(); 1491 1492 if (ImpUse && MI != ReMatDefMI) { 1493 // Re-matting an instruction with virtual register use. Prevent interval 1494 // from being spilled. 1495 getInterval(ImpUse).markNotSpillable(); 1496 } 1497 1498 unsigned MBBId = MBB->getNumber(); 1499 unsigned ThisVReg = 0; 1500 if (TrySplit) { 1501 DenseMap<unsigned,unsigned>::iterator NVI = MBBVRegsMap.find(MBBId); 1502 if (NVI != MBBVRegsMap.end()) { 1503 ThisVReg = NVI->second; 1504 // One common case: 1505 // x = use 1506 // ... 1507 // ... 1508 // def = ... 1509 // = use 1510 // It's better to start a new interval to avoid artificially 1511 // extend the new interval. 1512 if (MI->readsWritesVirtualRegister(li.reg) == 1513 std::make_pair(false,true)) { 1514 MBBVRegsMap.erase(MBB->getNumber()); 1515 ThisVReg = 0; 1516 } 1517 } 1518 } 1519 1520 bool IsNew = ThisVReg == 0; 1521 if (IsNew) { 1522 // This ends the previous live interval. If all of its def / use 1523 // can be folded, give it a low spill weight. 1524 if (NewVReg && TrySplit && AllCanFold) { 1525 LiveInterval &nI = getOrCreateInterval(NewVReg); 1526 nI.weight /= 10.0F; 1527 } 1528 AllCanFold = true; 1529 } 1530 NewVReg = ThisVReg; 1531 1532 bool HasDef = false; 1533 bool HasUse = false; 1534 bool CanFold = rewriteInstructionForSpills(li, I->valno, TrySplit, 1535 index, end, MI, ReMatOrigDefMI, ReMatDefMI, 1536 Slot, LdSlot, isLoad, isLoadSS, DefIsReMat, 1537 CanDelete, vrm, rc, ReMatIds, loopInfo, NewVReg, 1538 ImpUse, HasDef, HasUse, MBBVRegsMap, NewLIs); 1539 if (!HasDef && !HasUse) 1540 continue; 1541 1542 AllCanFold &= CanFold; 1543 1544 // Update weight of spill interval. 1545 LiveInterval &nI = getOrCreateInterval(NewVReg); 1546 if (!TrySplit) { 1547 // The spill weight is now infinity as it cannot be spilled again. 1548 nI.markNotSpillable(); 1549 continue; 1550 } 1551 1552 // Keep track of the last def and first use in each MBB. 1553 if (HasDef) { 1554 if (MI != ReMatOrigDefMI || !CanDelete) { 1555 bool HasKill = false; 1556 if (!HasUse) 1557 HasKill = anyKillInMBBAfterIdx(li, I->valno, MBB, index.getDefIndex()); 1558 else { 1559 // If this is a two-address code, then this index starts a new VNInfo. 1560 const VNInfo *VNI = li.findDefinedVNInfoForRegInt(index.getDefIndex()); 1561 if (VNI) 1562 HasKill = anyKillInMBBAfterIdx(li, VNI, MBB, index.getDefIndex()); 1563 } 1564 DenseMap<unsigned, std::vector<SRInfo> >::iterator SII = 1565 SpillIdxes.find(MBBId); 1566 if (!HasKill) { 1567 if (SII == SpillIdxes.end()) { 1568 std::vector<SRInfo> S; 1569 S.push_back(SRInfo(index, NewVReg, true)); 1570 SpillIdxes.insert(std::make_pair(MBBId, S)); 1571 } else if (SII->second.back().vreg != NewVReg) { 1572 SII->second.push_back(SRInfo(index, NewVReg, true)); 1573 } else if (index > SII->second.back().index) { 1574 // If there is an earlier def and this is a two-address 1575 // instruction, then it's not possible to fold the store (which 1576 // would also fold the load). 1577 SRInfo &Info = SII->second.back(); 1578 Info.index = index; 1579 Info.canFold = !HasUse; 1580 } 1581 SpillMBBs.set(MBBId); 1582 } else if (SII != SpillIdxes.end() && 1583 SII->second.back().vreg == NewVReg && 1584 index > SII->second.back().index) { 1585 // There is an earlier def that's not killed (must be two-address). 1586 // The spill is no longer needed. 1587 SII->second.pop_back(); 1588 if (SII->second.empty()) { 1589 SpillIdxes.erase(MBBId); 1590 SpillMBBs.reset(MBBId); 1591 } 1592 } 1593 } 1594 } 1595 1596 if (HasUse) { 1597 DenseMap<unsigned, std::vector<SRInfo> >::iterator SII = 1598 SpillIdxes.find(MBBId); 1599 if (SII != SpillIdxes.end() && 1600 SII->second.back().vreg == NewVReg && 1601 index > SII->second.back().index) 1602 // Use(s) following the last def, it's not safe to fold the spill. 1603 SII->second.back().canFold = false; 1604 DenseMap<unsigned, std::vector<SRInfo> >::iterator RII = 1605 RestoreIdxes.find(MBBId); 1606 if (RII != RestoreIdxes.end() && RII->second.back().vreg == NewVReg) 1607 // If we are splitting live intervals, only fold if it's the first 1608 // use and there isn't another use later in the MBB. 1609 RII->second.back().canFold = false; 1610 else if (IsNew) { 1611 // Only need a reload if there isn't an earlier def / use. 1612 if (RII == RestoreIdxes.end()) { 1613 std::vector<SRInfo> Infos; 1614 Infos.push_back(SRInfo(index, NewVReg, true)); 1615 RestoreIdxes.insert(std::make_pair(MBBId, Infos)); 1616 } else { 1617 RII->second.push_back(SRInfo(index, NewVReg, true)); 1618 } 1619 RestoreMBBs.set(MBBId); 1620 } 1621 } 1622 1623 // Update spill weight. 1624 unsigned loopDepth = loopInfo->getLoopDepth(MBB); 1625 nI.weight += getSpillWeight(HasDef, HasUse, loopDepth); 1626 } 1627 1628 if (NewVReg && TrySplit && AllCanFold) { 1629 // If all of its def / use can be folded, give it a low spill weight. 1630 LiveInterval &nI = getOrCreateInterval(NewVReg); 1631 nI.weight /= 10.0F; 1632 } 1633 } 1634 1635 bool LiveIntervals::alsoFoldARestore(int Id, SlotIndex index, 1636 unsigned vr, BitVector &RestoreMBBs, 1637 DenseMap<unsigned,std::vector<SRInfo> > &RestoreIdxes) { 1638 if (!RestoreMBBs[Id]) 1639 return false; 1640 std::vector<SRInfo> &Restores = RestoreIdxes[Id]; 1641 for (unsigned i = 0, e = Restores.size(); i != e; ++i) 1642 if (Restores[i].index == index && 1643 Restores[i].vreg == vr && 1644 Restores[i].canFold) 1645 return true; 1646 return false; 1647 } 1648 1649 void LiveIntervals::eraseRestoreInfo(int Id, SlotIndex index, 1650 unsigned vr, BitVector &RestoreMBBs, 1651 DenseMap<unsigned,std::vector<SRInfo> > &RestoreIdxes) { 1652 if (!RestoreMBBs[Id]) 1653 return; 1654 std::vector<SRInfo> &Restores = RestoreIdxes[Id]; 1655 for (unsigned i = 0, e = Restores.size(); i != e; ++i) 1656 if (Restores[i].index == index && Restores[i].vreg) 1657 Restores[i].index = SlotIndex(); 1658 } 1659 1660 /// handleSpilledImpDefs - Remove IMPLICIT_DEF instructions which are being 1661 /// spilled and create empty intervals for their uses. 1662 void 1663 LiveIntervals::handleSpilledImpDefs(const LiveInterval &li, VirtRegMap &vrm, 1664 const TargetRegisterClass* rc, 1665 std::vector<LiveInterval*> &NewLIs) { 1666 for (MachineRegisterInfo::reg_iterator ri = mri_->reg_begin(li.reg), 1667 re = mri_->reg_end(); ri != re; ) { 1668 MachineOperand &O = ri.getOperand(); 1669 MachineInstr *MI = &*ri; 1670 ++ri; 1671 if (MI->isDebugValue()) { 1672 // Remove debug info for now. 1673 O.setReg(0U); 1674 DEBUG(dbgs() << "Removing debug info due to spill:" << "\t" << *MI); 1675 continue; 1676 } 1677 if (O.isDef()) { 1678 assert(MI->isImplicitDef() && 1679 "Register def was not rewritten?"); 1680 RemoveMachineInstrFromMaps(MI); 1681 vrm.RemoveMachineInstrFromMaps(MI); 1682 MI->eraseFromParent(); 1683 } else { 1684 // This must be an use of an implicit_def so it's not part of the live 1685 // interval. Create a new empty live interval for it. 1686 // FIXME: Can we simply erase some of the instructions? e.g. Stores? 1687 unsigned NewVReg = mri_->createVirtualRegister(rc); 1688 vrm.grow(); 1689 vrm.setIsImplicitlyDefined(NewVReg); 1690 NewLIs.push_back(&getOrCreateInterval(NewVReg)); 1691 for (unsigned i = 0, e = MI->getNumOperands(); i != e; ++i) { 1692 MachineOperand &MO = MI->getOperand(i); 1693 if (MO.isReg() && MO.getReg() == li.reg) { 1694 MO.setReg(NewVReg); 1695 MO.setIsUndef(); 1696 } 1697 } 1698 } 1699 } 1700 } 1701 1702 float 1703 LiveIntervals::getSpillWeight(bool isDef, bool isUse, unsigned loopDepth) { 1704 // Limit the loop depth ridiculousness. 1705 if (loopDepth > 200) 1706 loopDepth = 200; 1707 1708 // The loop depth is used to roughly estimate the number of times the 1709 // instruction is executed. Something like 10^d is simple, but will quickly 1710 // overflow a float. This expression behaves like 10^d for small d, but is 1711 // more tempered for large d. At d=200 we get 6.7e33 which leaves a bit of 1712 // headroom before overflow. 1713 // By the way, powf() might be unavailable here. For consistency, 1714 // We may take pow(double,double). 1715 float lc = std::pow(1 + (100.0 / (loopDepth + 10)), (double)loopDepth); 1716 1717 return (isDef + isUse) * lc; 1718 } 1719 1720 static void normalizeSpillWeights(std::vector<LiveInterval*> &NewLIs) { 1721 for (unsigned i = 0, e = NewLIs.size(); i != e; ++i) 1722 NewLIs[i]->weight = 1723 normalizeSpillWeight(NewLIs[i]->weight, NewLIs[i]->getSize()); 1724 } 1725 1726 std::vector<LiveInterval*> LiveIntervals:: 1727 addIntervalsForSpills(const LiveInterval &li, 1728 const SmallVectorImpl<LiveInterval*> *SpillIs, 1729 const MachineLoopInfo *loopInfo, VirtRegMap &vrm) { 1730 assert(li.isSpillable() && "attempt to spill already spilled interval!"); 1731 1732 DEBUG({ 1733 dbgs() << "\t\t\t\tadding intervals for spills for interval: "; 1734 li.print(dbgs(), tri_); 1735 dbgs() << '\n'; 1736 }); 1737 1738 // Each bit specify whether a spill is required in the MBB. 1739 BitVector SpillMBBs(mf_->getNumBlockIDs()); 1740 DenseMap<unsigned, std::vector<SRInfo> > SpillIdxes; 1741 BitVector RestoreMBBs(mf_->getNumBlockIDs()); 1742 DenseMap<unsigned, std::vector<SRInfo> > RestoreIdxes; 1743 DenseMap<unsigned,unsigned> MBBVRegsMap; 1744 std::vector<LiveInterval*> NewLIs; 1745 const TargetRegisterClass* rc = mri_->getRegClass(li.reg); 1746 1747 unsigned NumValNums = li.getNumValNums(); 1748 SmallVector<MachineInstr*, 4> ReMatDefs; 1749 ReMatDefs.resize(NumValNums, NULL); 1750 SmallVector<MachineInstr*, 4> ReMatOrigDefs; 1751 ReMatOrigDefs.resize(NumValNums, NULL); 1752 SmallVector<int, 4> ReMatIds; 1753 ReMatIds.resize(NumValNums, VirtRegMap::MAX_STACK_SLOT); 1754 BitVector ReMatDelete(NumValNums); 1755 unsigned Slot = VirtRegMap::MAX_STACK_SLOT; 1756 1757 // Spilling a split live interval. It cannot be split any further. Also, 1758 // it's also guaranteed to be a single val# / range interval. 1759 if (vrm.getPreSplitReg(li.reg)) { 1760 vrm.setIsSplitFromReg(li.reg, 0); 1761 // Unset the split kill marker on the last use. 1762 SlotIndex KillIdx = vrm.getKillPoint(li.reg); 1763 if (KillIdx != SlotIndex()) { 1764 MachineInstr *KillMI = getInstructionFromIndex(KillIdx); 1765 assert(KillMI && "Last use disappeared?"); 1766 int KillOp = KillMI->findRegisterUseOperandIdx(li.reg, true); 1767 assert(KillOp != -1 && "Last use disappeared?"); 1768 KillMI->getOperand(KillOp).setIsKill(false); 1769 } 1770 vrm.removeKillPoint(li.reg); 1771 bool DefIsReMat = vrm.isReMaterialized(li.reg); 1772 Slot = vrm.getStackSlot(li.reg); 1773 assert(Slot != VirtRegMap::MAX_STACK_SLOT); 1774 MachineInstr *ReMatDefMI = DefIsReMat ? 1775 vrm.getReMaterializedMI(li.reg) : NULL; 1776 int LdSlot = 0; 1777 bool isLoadSS = DefIsReMat && tii_->isLoadFromStackSlot(ReMatDefMI, LdSlot); 1778 bool isLoad = isLoadSS || 1779 (DefIsReMat && (ReMatDefMI->getDesc().canFoldAsLoad())); 1780 bool IsFirstRange = true; 1781 for (LiveInterval::Ranges::const_iterator 1782 I = li.ranges.begin(), E = li.ranges.end(); I != E; ++I) { 1783 // If this is a split live interval with multiple ranges, it means there 1784 // are two-address instructions that re-defined the value. Only the 1785 // first def can be rematerialized! 1786 if (IsFirstRange) { 1787 // Note ReMatOrigDefMI has already been deleted. 1788 rewriteInstructionsForSpills(li, false, I, NULL, ReMatDefMI, 1789 Slot, LdSlot, isLoad, isLoadSS, DefIsReMat, 1790 false, vrm, rc, ReMatIds, loopInfo, 1791 SpillMBBs, SpillIdxes, RestoreMBBs, RestoreIdxes, 1792 MBBVRegsMap, NewLIs); 1793 } else { 1794 rewriteInstructionsForSpills(li, false, I, NULL, 0, 1795 Slot, 0, false, false, false, 1796 false, vrm, rc, ReMatIds, loopInfo, 1797 SpillMBBs, SpillIdxes, RestoreMBBs, RestoreIdxes, 1798 MBBVRegsMap, NewLIs); 1799 } 1800 IsFirstRange = false; 1801 } 1802 1803 handleSpilledImpDefs(li, vrm, rc, NewLIs); 1804 normalizeSpillWeights(NewLIs); 1805 return NewLIs; 1806 } 1807 1808 bool TrySplit = !intervalIsInOneMBB(li); 1809 if (TrySplit) 1810 ++numSplits; 1811 bool NeedStackSlot = false; 1812 for (LiveInterval::const_vni_iterator i = li.vni_begin(), e = li.vni_end(); 1813 i != e; ++i) { 1814 const VNInfo *VNI = *i; 1815 unsigned VN = VNI->id; 1816 if (VNI->isUnused()) 1817 continue; // Dead val#. 1818 // Is the def for the val# rematerializable? 1819 MachineInstr *ReMatDefMI = getInstructionFromIndex(VNI->def); 1820 bool dummy; 1821 if (ReMatDefMI && isReMaterializable(li, VNI, ReMatDefMI, SpillIs, dummy)) { 1822 // Remember how to remat the def of this val#. 1823 ReMatOrigDefs[VN] = ReMatDefMI; 1824 // Original def may be modified so we have to make a copy here. 1825 MachineInstr *Clone = mf_->CloneMachineInstr(ReMatDefMI); 1826 CloneMIs.push_back(Clone); 1827 ReMatDefs[VN] = Clone; 1828 1829 bool CanDelete = true; 1830 if (VNI->hasPHIKill()) { 1831 // A kill is a phi node, not all of its uses can be rematerialized. 1832 // It must not be deleted. 1833 CanDelete = false; 1834 // Need a stack slot if there is any live range where uses cannot be 1835 // rematerialized. 1836 NeedStackSlot = true; 1837 } 1838 if (CanDelete) 1839 ReMatDelete.set(VN); 1840 } else { 1841 // Need a stack slot if there is any live range where uses cannot be 1842 // rematerialized. 1843 NeedStackSlot = true; 1844 } 1845 } 1846 1847 // One stack slot per live interval. 1848 if (NeedStackSlot && vrm.getPreSplitReg(li.reg) == 0) { 1849 if (vrm.getStackSlot(li.reg) == VirtRegMap::NO_STACK_SLOT) 1850 Slot = vrm.assignVirt2StackSlot(li.reg); 1851 1852 // This case only occurs when the prealloc splitter has already assigned 1853 // a stack slot to this vreg. 1854 else 1855 Slot = vrm.getStackSlot(li.reg); 1856 } 1857 1858 // Create new intervals and rewrite defs and uses. 1859 for (LiveInterval::Ranges::const_iterator 1860 I = li.ranges.begin(), E = li.ranges.end(); I != E; ++I) { 1861 MachineInstr *ReMatDefMI = ReMatDefs[I->valno->id]; 1862 MachineInstr *ReMatOrigDefMI = ReMatOrigDefs[I->valno->id]; 1863 bool DefIsReMat = ReMatDefMI != NULL; 1864 bool CanDelete = ReMatDelete[I->valno->id]; 1865 int LdSlot = 0; 1866 bool isLoadSS = DefIsReMat && tii_->isLoadFromStackSlot(ReMatDefMI, LdSlot); 1867 bool isLoad = isLoadSS || 1868 (DefIsReMat && ReMatDefMI->getDesc().canFoldAsLoad()); 1869 rewriteInstructionsForSpills(li, TrySplit, I, ReMatOrigDefMI, ReMatDefMI, 1870 Slot, LdSlot, isLoad, isLoadSS, DefIsReMat, 1871 CanDelete, vrm, rc, ReMatIds, loopInfo, 1872 SpillMBBs, SpillIdxes, RestoreMBBs, RestoreIdxes, 1873 MBBVRegsMap, NewLIs); 1874 } 1875 1876 // Insert spills / restores if we are splitting. 1877 if (!TrySplit) { 1878 handleSpilledImpDefs(li, vrm, rc, NewLIs); 1879 normalizeSpillWeights(NewLIs); 1880 return NewLIs; 1881 } 1882 1883 SmallPtrSet<LiveInterval*, 4> AddedKill; 1884 SmallVector<unsigned, 2> Ops; 1885 if (NeedStackSlot) { 1886 int Id = SpillMBBs.find_first(); 1887 while (Id != -1) { 1888 std::vector<SRInfo> &spills = SpillIdxes[Id]; 1889 for (unsigned i = 0, e = spills.size(); i != e; ++i) { 1890 SlotIndex index = spills[i].index; 1891 unsigned VReg = spills[i].vreg; 1892 LiveInterval &nI = getOrCreateInterval(VReg); 1893 bool isReMat = vrm.isReMaterialized(VReg); 1894 MachineInstr *MI = getInstructionFromIndex(index); 1895 bool CanFold = false; 1896 bool FoundUse = false; 1897 Ops.clear(); 1898 if (spills[i].canFold) { 1899 CanFold = true; 1900 for (unsigned j = 0, ee = MI->getNumOperands(); j != ee; ++j) { 1901 MachineOperand &MO = MI->getOperand(j); 1902 if (!MO.isReg() || MO.getReg() != VReg) 1903 continue; 1904 1905 Ops.push_back(j); 1906 if (MO.isDef()) 1907 continue; 1908 if (isReMat || 1909 (!FoundUse && !alsoFoldARestore(Id, index, VReg, 1910 RestoreMBBs, RestoreIdxes))) { 1911 // MI has two-address uses of the same register. If the use 1912 // isn't the first and only use in the BB, then we can't fold 1913 // it. FIXME: Move this to rewriteInstructionsForSpills. 1914 CanFold = false; 1915 break; 1916 } 1917 FoundUse = true; 1918 } 1919 } 1920 // Fold the store into the def if possible. 1921 bool Folded = false; 1922 if (CanFold && !Ops.empty()) { 1923 if (tryFoldMemoryOperand(MI, vrm, NULL, index, Ops, true, Slot,VReg)){ 1924 Folded = true; 1925 if (FoundUse) { 1926 // Also folded uses, do not issue a load. 1927 eraseRestoreInfo(Id, index, VReg, RestoreMBBs, RestoreIdxes); 1928 nI.removeRange(index.getLoadIndex(), index.getDefIndex()); 1929 } 1930 nI.removeRange(index.getDefIndex(), index.getStoreIndex()); 1931 } 1932 } 1933 1934 // Otherwise tell the spiller to issue a spill. 1935 if (!Folded) { 1936 LiveRange *LR = &nI.ranges[nI.ranges.size()-1]; 1937 bool isKill = LR->end == index.getStoreIndex(); 1938 if (!MI->registerDefIsDead(nI.reg)) 1939 // No need to spill a dead def. 1940 vrm.addSpillPoint(VReg, isKill, MI); 1941 if (isKill) 1942 AddedKill.insert(&nI); 1943 } 1944 } 1945 Id = SpillMBBs.find_next(Id); 1946 } 1947 } 1948 1949 int Id = RestoreMBBs.find_first(); 1950 while (Id != -1) { 1951 std::vector<SRInfo> &restores = RestoreIdxes[Id]; 1952 for (unsigned i = 0, e = restores.size(); i != e; ++i) { 1953 SlotIndex index = restores[i].index; 1954 if (index == SlotIndex()) 1955 continue; 1956 unsigned VReg = restores[i].vreg; 1957 LiveInterval &nI = getOrCreateInterval(VReg); 1958 bool isReMat = vrm.isReMaterialized(VReg); 1959 MachineInstr *MI = getInstructionFromIndex(index); 1960 bool CanFold = false; 1961 Ops.clear(); 1962 if (restores[i].canFold) { 1963 CanFold = true; 1964 for (unsigned j = 0, ee = MI->getNumOperands(); j != ee; ++j) { 1965 MachineOperand &MO = MI->getOperand(j); 1966 if (!MO.isReg() || MO.getReg() != VReg) 1967 continue; 1968 1969 if (MO.isDef()) { 1970 // If this restore were to be folded, it would have been folded 1971 // already. 1972 CanFold = false; 1973 break; 1974 } 1975 Ops.push_back(j); 1976 } 1977 } 1978 1979 // Fold the load into the use if possible. 1980 bool Folded = false; 1981 if (CanFold && !Ops.empty()) { 1982 if (!isReMat) 1983 Folded = tryFoldMemoryOperand(MI, vrm, NULL,index,Ops,true,Slot,VReg); 1984 else { 1985 MachineInstr *ReMatDefMI = vrm.getReMaterializedMI(VReg); 1986 int LdSlot = 0; 1987 bool isLoadSS = tii_->isLoadFromStackSlot(ReMatDefMI, LdSlot); 1988 // If the rematerializable def is a load, also try to fold it. 1989 if (isLoadSS || ReMatDefMI->getDesc().canFoldAsLoad()) 1990 Folded = tryFoldMemoryOperand(MI, vrm, ReMatDefMI, index, 1991 Ops, isLoadSS, LdSlot, VReg); 1992 if (!Folded) { 1993 unsigned ImpUse = getReMatImplicitUse(li, ReMatDefMI); 1994 if (ImpUse) { 1995 // Re-matting an instruction with virtual register use. Add the 1996 // register as an implicit use on the use MI and mark the register 1997 // interval as unspillable. 1998 LiveInterval &ImpLi = getInterval(ImpUse); 1999 ImpLi.markNotSpillable(); 2000 MI->addOperand(MachineOperand::CreateReg(ImpUse, false, true)); 2001 } 2002 } 2003 } 2004 } 2005 // If folding is not possible / failed, then tell the spiller to issue a 2006 // load / rematerialization for us. 2007 if (Folded) 2008 nI.removeRange(index.getLoadIndex(), index.getDefIndex()); 2009 else 2010 vrm.addRestorePoint(VReg, MI); 2011 } 2012 Id = RestoreMBBs.find_next(Id); 2013 } 2014 2015 // Finalize intervals: add kills, finalize spill weights, and filter out 2016 // dead intervals. 2017 std::vector<LiveInterval*> RetNewLIs; 2018 for (unsigned i = 0, e = NewLIs.size(); i != e; ++i) { 2019 LiveInterval *LI = NewLIs[i]; 2020 if (!LI->empty()) { 2021 if (!AddedKill.count(LI)) { 2022 LiveRange *LR = &LI->ranges[LI->ranges.size()-1]; 2023 SlotIndex LastUseIdx = LR->end.getBaseIndex(); 2024 MachineInstr *LastUse = getInstructionFromIndex(LastUseIdx); 2025 int UseIdx = LastUse->findRegisterUseOperandIdx(LI->reg, false); 2026 assert(UseIdx != -1); 2027 if (!LastUse->isRegTiedToDefOperand(UseIdx)) { 2028 LastUse->getOperand(UseIdx).setIsKill(); 2029 vrm.addKillPoint(LI->reg, LastUseIdx); 2030 } 2031 } 2032 RetNewLIs.push_back(LI); 2033 } 2034 } 2035 2036 handleSpilledImpDefs(li, vrm, rc, RetNewLIs); 2037 normalizeSpillWeights(RetNewLIs); 2038 return RetNewLIs; 2039 } 2040 2041 /// hasAllocatableSuperReg - Return true if the specified physical register has 2042 /// any super register that's allocatable. 2043 bool LiveIntervals::hasAllocatableSuperReg(unsigned Reg) const { 2044 for (const unsigned* AS = tri_->getSuperRegisters(Reg); *AS; ++AS) 2045 if (allocatableRegs_[*AS] && hasInterval(*AS)) 2046 return true; 2047 return false; 2048 } 2049 2050 /// getRepresentativeReg - Find the largest super register of the specified 2051 /// physical register. 2052 unsigned LiveIntervals::getRepresentativeReg(unsigned Reg) const { 2053 // Find the largest super-register that is allocatable. 2054 unsigned BestReg = Reg; 2055 for (const unsigned* AS = tri_->getSuperRegisters(Reg); *AS; ++AS) { 2056 unsigned SuperReg = *AS; 2057 if (!hasAllocatableSuperReg(SuperReg) && hasInterval(SuperReg)) { 2058 BestReg = SuperReg; 2059 break; 2060 } 2061 } 2062 return BestReg; 2063 } 2064 2065 /// getNumConflictsWithPhysReg - Return the number of uses and defs of the 2066 /// specified interval that conflicts with the specified physical register. 2067 unsigned LiveIntervals::getNumConflictsWithPhysReg(const LiveInterval &li, 2068 unsigned PhysReg) const { 2069 unsigned NumConflicts = 0; 2070 const LiveInterval &pli = getInterval(getRepresentativeReg(PhysReg)); 2071 for (MachineRegisterInfo::reg_iterator I = mri_->reg_begin(li.reg), 2072 E = mri_->reg_end(); I != E; ++I) { 2073 MachineOperand &O = I.getOperand(); 2074 MachineInstr *MI = O.getParent(); 2075 if (MI->isDebugValue()) 2076 continue; 2077 SlotIndex Index = getInstructionIndex(MI); 2078 if (pli.liveAt(Index)) 2079 ++NumConflicts; 2080 } 2081 return NumConflicts; 2082 } 2083 2084 /// spillPhysRegAroundRegDefsUses - Spill the specified physical register 2085 /// around all defs and uses of the specified interval. Return true if it 2086 /// was able to cut its interval. 2087 bool LiveIntervals::spillPhysRegAroundRegDefsUses(const LiveInterval &li, 2088 unsigned PhysReg, VirtRegMap &vrm) { 2089 unsigned SpillReg = getRepresentativeReg(PhysReg); 2090 2091 DEBUG(dbgs() << "spillPhysRegAroundRegDefsUses " << tri_->getName(PhysReg) 2092 << " represented by " << tri_->getName(SpillReg) << '\n'); 2093 2094 for (const unsigned *AS = tri_->getAliasSet(PhysReg); *AS; ++AS) 2095 // If there are registers which alias PhysReg, but which are not a 2096 // sub-register of the chosen representative super register. Assert 2097 // since we can't handle it yet. 2098 assert(*AS == SpillReg || !allocatableRegs_[*AS] || !hasInterval(*AS) || 2099 tri_->isSuperRegister(*AS, SpillReg)); 2100 2101 bool Cut = false; 2102 SmallVector<unsigned, 4> PRegs; 2103 if (hasInterval(SpillReg)) 2104 PRegs.push_back(SpillReg); 2105 for (const unsigned *SR = tri_->getSubRegisters(SpillReg); *SR; ++SR) 2106 if (hasInterval(*SR)) 2107 PRegs.push_back(*SR); 2108 2109 DEBUG({ 2110 dbgs() << "Trying to spill:"; 2111 for (unsigned i = 0, e = PRegs.size(); i != e; ++i) 2112 dbgs() << ' ' << tri_->getName(PRegs[i]); 2113 dbgs() << '\n'; 2114 }); 2115 2116 SmallPtrSet<MachineInstr*, 8> SeenMIs; 2117 for (MachineRegisterInfo::reg_iterator I = mri_->reg_begin(li.reg), 2118 E = mri_->reg_end(); I != E; ++I) { 2119 MachineOperand &O = I.getOperand(); 2120 MachineInstr *MI = O.getParent(); 2121 if (MI->isDebugValue() || SeenMIs.count(MI)) 2122 continue; 2123 SeenMIs.insert(MI); 2124 SlotIndex Index = getInstructionIndex(MI); 2125 bool LiveReg = false; 2126 for (unsigned i = 0, e = PRegs.size(); i != e; ++i) { 2127 unsigned PReg = PRegs[i]; 2128 LiveInterval &pli = getInterval(PReg); 2129 if (!pli.liveAt(Index)) 2130 continue; 2131 LiveReg = true; 2132 SlotIndex StartIdx = Index.getLoadIndex(); 2133 SlotIndex EndIdx = Index.getNextIndex().getBaseIndex(); 2134 if (!pli.isInOneLiveRange(StartIdx, EndIdx)) { 2135 std::string msg; 2136 raw_string_ostream Msg(msg); 2137 Msg << "Ran out of registers during register allocation!"; 2138 if (MI->isInlineAsm()) { 2139 Msg << "\nPlease check your inline asm statement for invalid " 2140 << "constraints:\n"; 2141 MI->print(Msg, tm_); 2142 } 2143 report_fatal_error(Msg.str()); 2144 } 2145 pli.removeRange(StartIdx, EndIdx); 2146 LiveReg = true; 2147 } 2148 if (!LiveReg) 2149 continue; 2150 DEBUG(dbgs() << "Emergency spill around " << Index << '\t' << *MI); 2151 vrm.addEmergencySpill(SpillReg, MI); 2152 Cut = true; 2153 } 2154 return Cut; 2155 } 2156 2157 LiveRange LiveIntervals::addLiveRangeToEndOfBlock(unsigned reg, 2158 MachineInstr* startInst) { 2159 LiveInterval& Interval = getOrCreateInterval(reg); 2160 VNInfo* VN = Interval.getNextValue( 2161 SlotIndex(getInstructionIndex(startInst).getDefIndex()), 2162 startInst, getVNInfoAllocator()); 2163 VN->setHasPHIKill(true); 2164 LiveRange LR( 2165 SlotIndex(getInstructionIndex(startInst).getDefIndex()), 2166 getMBBEndIdx(startInst->getParent()), VN); 2167 Interval.addRange(LR); 2168 2169 return LR; 2170 } 2171 2172