• Home
  • Line#
  • Scopes#
  • Navigate#
  • Raw
  • Download
1 //===-- AMDILCFGStructurizer.cpp - CFG Structurizer -----------------------===//
2 //
3 //                     The LLVM Compiler Infrastructure
4 //
5 // This file is distributed under the University of Illinois Open Source
6 // License. See LICENSE.TXT for details.
7 //
8 /// \file
9 //==-----------------------------------------------------------------------===//
10 
11 #include "AMDGPU.h"
12 #include "AMDGPUInstrInfo.h"
13 #include "AMDGPUSubtarget.h"
14 #include "R600InstrInfo.h"
15 #include "llvm/ADT/DepthFirstIterator.h"
16 #include "llvm/ADT/SCCIterator.h"
17 #include "llvm/ADT/SmallVector.h"
18 #include "llvm/ADT/Statistic.h"
19 #include "llvm/CodeGen/MachineDominators.h"
20 #include "llvm/CodeGen/MachineFunction.h"
21 #include "llvm/CodeGen/MachineFunctionAnalysis.h"
22 #include "llvm/CodeGen/MachineFunctionPass.h"
23 #include "llvm/CodeGen/MachineInstrBuilder.h"
24 #include "llvm/CodeGen/MachineJumpTableInfo.h"
25 #include "llvm/CodeGen/MachineLoopInfo.h"
26 #include "llvm/CodeGen/MachinePostDominators.h"
27 #include "llvm/CodeGen/MachineRegisterInfo.h"
28 #include "llvm/IR/Dominators.h"
29 #include "llvm/Support/Debug.h"
30 #include "llvm/Support/raw_ostream.h"
31 #include "llvm/Target/TargetInstrInfo.h"
32 #include "llvm/Target/TargetMachine.h"
33 #include <deque>
34 
35 using namespace llvm;
36 
37 #define DEBUG_TYPE "structcfg"
38 
39 #define DEFAULT_VEC_SLOTS 8
40 
41 // TODO: move-begin.
42 
43 //===----------------------------------------------------------------------===//
44 //
45 // Statistics for CFGStructurizer.
46 //
47 //===----------------------------------------------------------------------===//
48 
49 STATISTIC(numSerialPatternMatch,    "CFGStructurizer number of serial pattern "
50     "matched");
51 STATISTIC(numIfPatternMatch,        "CFGStructurizer number of if pattern "
52     "matched");
53 STATISTIC(numClonedBlock,           "CFGStructurizer cloned blocks");
54 STATISTIC(numClonedInstr,           "CFGStructurizer cloned instructions");
55 
56 namespace llvm {
57   void initializeAMDGPUCFGStructurizerPass(PassRegistry&);
58 }
59 
60 //===----------------------------------------------------------------------===//
61 //
62 // Miscellaneous utility for CFGStructurizer.
63 //
64 //===----------------------------------------------------------------------===//
65 namespace {
66 #define SHOWNEWINSTR(i) \
67   DEBUG(dbgs() << "New instr: " << *i << "\n");
68 
69 #define SHOWNEWBLK(b, msg) \
70 DEBUG( \
71   dbgs() << msg << "BB" << b->getNumber() << "size " << b->size(); \
72   dbgs() << "\n"; \
73 );
74 
75 #define SHOWBLK_DETAIL(b, msg) \
76 DEBUG( \
77   if (b) { \
78   dbgs() << msg << "BB" << b->getNumber() << "size " << b->size(); \
79   b->print(dbgs()); \
80   dbgs() << "\n"; \
81   } \
82 );
83 
84 #define INVALIDSCCNUM -1
85 
86 template<class NodeT>
ReverseVector(SmallVectorImpl<NodeT * > & Src)87 void ReverseVector(SmallVectorImpl<NodeT *> &Src) {
88   size_t sz = Src.size();
89   for (size_t i = 0; i < sz/2; ++i) {
90     NodeT *t = Src[i];
91     Src[i] = Src[sz - i - 1];
92     Src[sz - i - 1] = t;
93   }
94 }
95 
96 } // end anonymous namespace
97 
98 //===----------------------------------------------------------------------===//
99 //
100 // supporting data structure for CFGStructurizer
101 //
102 //===----------------------------------------------------------------------===//
103 
104 
105 namespace {
106 
107 class BlockInformation {
108 public:
109   bool IsRetired;
110   int  SccNum;
BlockInformation()111   BlockInformation() : IsRetired(false), SccNum(INVALIDSCCNUM) {}
112 };
113 
114 } // end anonymous namespace
115 
116 //===----------------------------------------------------------------------===//
117 //
118 // CFGStructurizer
119 //
120 //===----------------------------------------------------------------------===//
121 
122 namespace {
123 class AMDGPUCFGStructurizer : public MachineFunctionPass {
124 public:
125   typedef SmallVector<MachineBasicBlock *, 32> MBBVector;
126   typedef std::map<MachineBasicBlock *, BlockInformation *> MBBInfoMap;
127   typedef std::map<MachineLoop *, MachineBasicBlock *> LoopLandInfoMap;
128 
129   enum PathToKind {
130     Not_SinglePath = 0,
131     SinglePath_InPath = 1,
132     SinglePath_NotInPath = 2
133   };
134 
135   static char ID;
136 
AMDGPUCFGStructurizer()137   AMDGPUCFGStructurizer() :
138       MachineFunctionPass(ID), TII(nullptr), TRI(nullptr) {
139     initializeAMDGPUCFGStructurizerPass(*PassRegistry::getPassRegistry());
140   }
141 
getPassName() const142    const char *getPassName() const override {
143     return "AMDGPU Control Flow Graph structurizer Pass";
144   }
145 
getAnalysisUsage(AnalysisUsage & AU) const146   void getAnalysisUsage(AnalysisUsage &AU) const override {
147     AU.addPreserved<MachineFunctionAnalysis>();
148     AU.addRequired<MachineFunctionAnalysis>();
149     AU.addRequired<MachineDominatorTree>();
150     AU.addRequired<MachinePostDominatorTree>();
151     AU.addRequired<MachineLoopInfo>();
152   }
153 
154   /// Perform the CFG structurization
155   bool run();
156 
157   /// Perform the CFG preparation
158   /// This step will remove every unconditionnal/dead jump instructions and make
159   /// sure all loops have an exit block
160   bool prepare();
161 
runOnMachineFunction(MachineFunction & MF)162   bool runOnMachineFunction(MachineFunction &MF) override {
163     TII = MF.getSubtarget<R600Subtarget>().getInstrInfo();
164     TRI = &TII->getRegisterInfo();
165     DEBUG(MF.dump(););
166     OrderedBlks.clear();
167     Visited.clear();
168     FuncRep = &MF;
169     MLI = &getAnalysis<MachineLoopInfo>();
170     DEBUG(dbgs() << "LoopInfo:\n"; PrintLoopinfo(*MLI););
171     MDT = &getAnalysis<MachineDominatorTree>();
172     DEBUG(MDT->print(dbgs(), (const llvm::Module*)nullptr););
173     PDT = &getAnalysis<MachinePostDominatorTree>();
174     DEBUG(PDT->print(dbgs()););
175     prepare();
176     run();
177     DEBUG(MF.dump(););
178     return true;
179   }
180 
181 protected:
182   MachineDominatorTree *MDT;
183   MachinePostDominatorTree *PDT;
184   MachineLoopInfo *MLI;
185   const R600InstrInfo *TII;
186   const R600RegisterInfo *TRI;
187 
188   // PRINT FUNCTIONS
189   /// Print the ordered Blocks.
printOrderedBlocks() const190   void printOrderedBlocks() const {
191     size_t i = 0;
192     for (MBBVector::const_iterator iterBlk = OrderedBlks.begin(),
193         iterBlkEnd = OrderedBlks.end(); iterBlk != iterBlkEnd; ++iterBlk, ++i) {
194       dbgs() << "BB" << (*iterBlk)->getNumber();
195       dbgs() << "(" << getSCCNum(*iterBlk) << "," << (*iterBlk)->size() << ")";
196       if (i != 0 && i % 10 == 0) {
197         dbgs() << "\n";
198       } else {
199         dbgs() << " ";
200       }
201     }
202   }
PrintLoopinfo(const MachineLoopInfo & LoopInfo)203   static void PrintLoopinfo(const MachineLoopInfo &LoopInfo) {
204     for (MachineLoop::iterator iter = LoopInfo.begin(),
205          iterEnd = LoopInfo.end(); iter != iterEnd; ++iter) {
206       (*iter)->print(dbgs(), 0);
207     }
208   }
209 
210   // UTILITY FUNCTIONS
211   int getSCCNum(MachineBasicBlock *MBB) const;
212   MachineBasicBlock *getLoopLandInfo(MachineLoop *LoopRep) const;
213   bool hasBackEdge(MachineBasicBlock *MBB) const;
214   bool isRetiredBlock(MachineBasicBlock *MBB) const;
215   bool isActiveLoophead(MachineBasicBlock *MBB) const;
216   PathToKind singlePathTo(MachineBasicBlock *SrcMBB, MachineBasicBlock *DstMBB,
217       bool AllowSideEntry = true) const;
218   int countActiveBlock(MBBVector::const_iterator It,
219       MBBVector::const_iterator E) const;
220   bool needMigrateBlock(MachineBasicBlock *MBB) const;
221 
222   // Utility Functions
223   void reversePredicateSetter(MachineBasicBlock::iterator I);
224   /// Compute the reversed DFS post order of Blocks
225   void orderBlocks(MachineFunction *MF);
226 
227   // Function originally from CFGStructTraits
228   void insertInstrEnd(MachineBasicBlock *MBB, int NewOpcode,
229                       const DebugLoc &DL = DebugLoc());
230   MachineInstr *insertInstrBefore(MachineBasicBlock *MBB, int NewOpcode,
231                                   const DebugLoc &DL = DebugLoc());
232   MachineInstr *insertInstrBefore(MachineBasicBlock::iterator I, int NewOpcode);
233   void insertCondBranchBefore(MachineBasicBlock::iterator I, int NewOpcode,
234                               const DebugLoc &DL);
235   void insertCondBranchBefore(MachineBasicBlock *MBB,
236                               MachineBasicBlock::iterator I, int NewOpcode,
237                               int RegNum, const DebugLoc &DL);
238   static int getBranchNzeroOpcode(int OldOpcode);
239   static int getBranchZeroOpcode(int OldOpcode);
240   static int getContinueNzeroOpcode(int OldOpcode);
241   static int getContinueZeroOpcode(int OldOpcode);
242   static MachineBasicBlock *getTrueBranch(MachineInstr *MI);
243   static void setTrueBranch(MachineInstr *MI, MachineBasicBlock *MBB);
244   static MachineBasicBlock *getFalseBranch(MachineBasicBlock *MBB,
245       MachineInstr *MI);
246   static bool isCondBranch(MachineInstr *MI);
247   static bool isUncondBranch(MachineInstr *MI);
248   static DebugLoc getLastDebugLocInBB(MachineBasicBlock *MBB);
249   static MachineInstr *getNormalBlockBranchInstr(MachineBasicBlock *MBB);
250   /// The correct naming for this is getPossibleLoopendBlockBranchInstr.
251   ///
252   /// BB with backward-edge could have move instructions after the branch
253   /// instruction.  Such move instruction "belong to" the loop backward-edge.
254   MachineInstr *getLoopendBlockBranchInstr(MachineBasicBlock *MBB);
255   static MachineInstr *getReturnInstr(MachineBasicBlock *MBB);
256   static bool isReturnBlock(MachineBasicBlock *MBB);
257   static void cloneSuccessorList(MachineBasicBlock *DstMBB,
258       MachineBasicBlock *SrcMBB) ;
259   static MachineBasicBlock *clone(MachineBasicBlock *MBB);
260   /// MachineBasicBlock::ReplaceUsesOfBlockWith doesn't serve the purpose
261   /// because the AMDGPU instruction is not recognized as terminator fix this
262   /// and retire this routine
263   void replaceInstrUseOfBlockWith(MachineBasicBlock *SrcMBB,
264       MachineBasicBlock *OldMBB, MachineBasicBlock *NewBlk);
265   static void wrapup(MachineBasicBlock *MBB);
266 
267 
268   int patternMatch(MachineBasicBlock *MBB);
269   int patternMatchGroup(MachineBasicBlock *MBB);
270   int serialPatternMatch(MachineBasicBlock *MBB);
271   int ifPatternMatch(MachineBasicBlock *MBB);
272   int loopendPatternMatch();
273   int mergeLoop(MachineLoop *LoopRep);
274 
275   /// return true iff src1Blk->succ_size() == 0 && src1Blk and src2Blk are in
276   /// the same loop with LoopLandInfo without explicitly keeping track of
277   /// loopContBlks and loopBreakBlks, this is a method to get the information.
278   bool isSameloopDetachedContbreak(MachineBasicBlock *Src1MBB,
279       MachineBasicBlock *Src2MBB);
280   int handleJumpintoIf(MachineBasicBlock *HeadMBB,
281       MachineBasicBlock *TrueMBB, MachineBasicBlock *FalseMBB);
282   int handleJumpintoIfImp(MachineBasicBlock *HeadMBB,
283       MachineBasicBlock *TrueMBB, MachineBasicBlock *FalseMBB);
284   int improveSimpleJumpintoIf(MachineBasicBlock *HeadMBB,
285       MachineBasicBlock *TrueMBB, MachineBasicBlock *FalseMBB,
286       MachineBasicBlock **LandMBBPtr);
287   void showImproveSimpleJumpintoIf(MachineBasicBlock *HeadMBB,
288       MachineBasicBlock *TrueMBB, MachineBasicBlock *FalseMBB,
289       MachineBasicBlock *LandMBB, bool Detail = false);
290   int cloneOnSideEntryTo(MachineBasicBlock *PreMBB,
291       MachineBasicBlock *SrcMBB, MachineBasicBlock *DstMBB);
292   void mergeSerialBlock(MachineBasicBlock *DstMBB,
293       MachineBasicBlock *SrcMBB);
294 
295   void mergeIfthenelseBlock(MachineInstr *BranchMI,
296       MachineBasicBlock *MBB, MachineBasicBlock *TrueMBB,
297       MachineBasicBlock *FalseMBB, MachineBasicBlock *LandMBB);
298   void mergeLooplandBlock(MachineBasicBlock *DstMBB,
299       MachineBasicBlock *LandMBB);
300   void mergeLoopbreakBlock(MachineBasicBlock *ExitingMBB,
301       MachineBasicBlock *LandMBB);
302   void settleLoopcontBlock(MachineBasicBlock *ContingMBB,
303       MachineBasicBlock *ContMBB);
304   /// normalizeInfiniteLoopExit change
305   ///   B1:
306   ///        uncond_br LoopHeader
307   ///
308   /// to
309   ///   B1:
310   ///        cond_br 1 LoopHeader dummyExit
311   /// and return the newly added dummy exit block
312   MachineBasicBlock *normalizeInfiniteLoopExit(MachineLoop *LoopRep);
313   void removeUnconditionalBranch(MachineBasicBlock *MBB);
314   /// Remove duplicate branches instructions in a block.
315   /// For instance
316   /// B0:
317   ///    cond_br X B1 B2
318   ///    cond_br X B1 B2
319   /// is transformed to
320   /// B0:
321   ///    cond_br X B1 B2
322   void removeRedundantConditionalBranch(MachineBasicBlock *MBB);
323   void addDummyExitBlock(SmallVectorImpl<MachineBasicBlock *> &RetMBB);
324   void removeSuccessor(MachineBasicBlock *MBB);
325   MachineBasicBlock *cloneBlockForPredecessor(MachineBasicBlock *MBB,
326       MachineBasicBlock *PredMBB);
327   void migrateInstruction(MachineBasicBlock *SrcMBB,
328       MachineBasicBlock *DstMBB, MachineBasicBlock::iterator I);
329   void recordSccnum(MachineBasicBlock *MBB, int SCCNum);
330   void retireBlock(MachineBasicBlock *MBB);
331 
332 
333 private:
334   MBBInfoMap BlockInfoMap;
335   LoopLandInfoMap LLInfoMap;
336   std::map<MachineLoop *, bool> Visited;
337   MachineFunction *FuncRep;
338   SmallVector<MachineBasicBlock *, DEFAULT_VEC_SLOTS> OrderedBlks;
339 };
340 
getSCCNum(MachineBasicBlock * MBB) const341 int AMDGPUCFGStructurizer::getSCCNum(MachineBasicBlock *MBB) const {
342   MBBInfoMap::const_iterator It = BlockInfoMap.find(MBB);
343   if (It == BlockInfoMap.end())
344     return INVALIDSCCNUM;
345   return (*It).second->SccNum;
346 }
347 
getLoopLandInfo(MachineLoop * LoopRep) const348 MachineBasicBlock *AMDGPUCFGStructurizer::getLoopLandInfo(MachineLoop *LoopRep)
349     const {
350   LoopLandInfoMap::const_iterator It = LLInfoMap.find(LoopRep);
351   if (It == LLInfoMap.end())
352     return nullptr;
353   return (*It).second;
354 }
355 
hasBackEdge(MachineBasicBlock * MBB) const356 bool AMDGPUCFGStructurizer::hasBackEdge(MachineBasicBlock *MBB) const {
357   MachineLoop *LoopRep = MLI->getLoopFor(MBB);
358   if (!LoopRep)
359     return false;
360   MachineBasicBlock *LoopHeader = LoopRep->getHeader();
361   return MBB->isSuccessor(LoopHeader);
362 }
363 
isRetiredBlock(MachineBasicBlock * MBB) const364 bool AMDGPUCFGStructurizer::isRetiredBlock(MachineBasicBlock *MBB) const {
365   MBBInfoMap::const_iterator It = BlockInfoMap.find(MBB);
366   if (It == BlockInfoMap.end())
367     return false;
368   return (*It).second->IsRetired;
369 }
370 
isActiveLoophead(MachineBasicBlock * MBB) const371 bool AMDGPUCFGStructurizer::isActiveLoophead(MachineBasicBlock *MBB) const {
372   MachineLoop *LoopRep = MLI->getLoopFor(MBB);
373   while (LoopRep && LoopRep->getHeader() == MBB) {
374     MachineBasicBlock *LoopLand = getLoopLandInfo(LoopRep);
375     if(!LoopLand)
376       return true;
377     if (!isRetiredBlock(LoopLand))
378       return true;
379     LoopRep = LoopRep->getParentLoop();
380   }
381   return false;
382 }
singlePathTo(MachineBasicBlock * SrcMBB,MachineBasicBlock * DstMBB,bool AllowSideEntry) const383 AMDGPUCFGStructurizer::PathToKind AMDGPUCFGStructurizer::singlePathTo(
384     MachineBasicBlock *SrcMBB, MachineBasicBlock *DstMBB,
385     bool AllowSideEntry) const {
386   assert(DstMBB);
387   if (SrcMBB == DstMBB)
388     return SinglePath_InPath;
389   while (SrcMBB && SrcMBB->succ_size() == 1) {
390     SrcMBB = *SrcMBB->succ_begin();
391     if (SrcMBB == DstMBB)
392       return SinglePath_InPath;
393     if (!AllowSideEntry && SrcMBB->pred_size() > 1)
394       return Not_SinglePath;
395   }
396   if (SrcMBB && SrcMBB->succ_size()==0)
397     return SinglePath_NotInPath;
398   return Not_SinglePath;
399 }
400 
countActiveBlock(MBBVector::const_iterator It,MBBVector::const_iterator E) const401 int AMDGPUCFGStructurizer::countActiveBlock(MBBVector::const_iterator It,
402     MBBVector::const_iterator E) const {
403   int Count = 0;
404   while (It != E) {
405     if (!isRetiredBlock(*It))
406       ++Count;
407     ++It;
408   }
409   return Count;
410 }
411 
needMigrateBlock(MachineBasicBlock * MBB) const412 bool AMDGPUCFGStructurizer::needMigrateBlock(MachineBasicBlock *MBB) const {
413   unsigned BlockSizeThreshold = 30;
414   unsigned CloneInstrThreshold = 100;
415   bool MultiplePreds = MBB && (MBB->pred_size() > 1);
416 
417   if(!MultiplePreds)
418     return false;
419   unsigned BlkSize = MBB->size();
420   return ((BlkSize > BlockSizeThreshold) &&
421       (BlkSize * (MBB->pred_size() - 1) > CloneInstrThreshold));
422 }
423 
reversePredicateSetter(MachineBasicBlock::iterator I)424 void AMDGPUCFGStructurizer::reversePredicateSetter(
425     MachineBasicBlock::iterator I) {
426   assert(static_cast<MachineInstr *>(I) && "Expected valid iterator");
427   for (;; --I) {
428     if (I->getOpcode() == AMDGPU::PRED_X) {
429       switch (static_cast<MachineInstr *>(I)->getOperand(2).getImm()) {
430       case OPCODE_IS_ZERO_INT:
431         static_cast<MachineInstr *>(I)->getOperand(2)
432             .setImm(OPCODE_IS_NOT_ZERO_INT);
433         return;
434       case OPCODE_IS_NOT_ZERO_INT:
435         static_cast<MachineInstr *>(I)->getOperand(2)
436             .setImm(OPCODE_IS_ZERO_INT);
437         return;
438       case OPCODE_IS_ZERO:
439         static_cast<MachineInstr *>(I)->getOperand(2)
440             .setImm(OPCODE_IS_NOT_ZERO);
441         return;
442       case OPCODE_IS_NOT_ZERO:
443         static_cast<MachineInstr *>(I)->getOperand(2)
444             .setImm(OPCODE_IS_ZERO);
445         return;
446       default:
447         llvm_unreachable("PRED_X Opcode invalid!");
448       }
449     }
450   }
451 }
452 
insertInstrEnd(MachineBasicBlock * MBB,int NewOpcode,const DebugLoc & DL)453 void AMDGPUCFGStructurizer::insertInstrEnd(MachineBasicBlock *MBB,
454                                            int NewOpcode, const DebugLoc &DL) {
455   MachineInstr *MI =
456       MBB->getParent()->CreateMachineInstr(TII->get(NewOpcode), DL);
457   MBB->push_back(MI);
458   //assume the instruction doesn't take any reg operand ...
459   SHOWNEWINSTR(MI);
460 }
461 
insertInstrBefore(MachineBasicBlock * MBB,int NewOpcode,const DebugLoc & DL)462 MachineInstr *AMDGPUCFGStructurizer::insertInstrBefore(MachineBasicBlock *MBB,
463                                                        int NewOpcode,
464                                                        const DebugLoc &DL) {
465   MachineInstr *MI =
466       MBB->getParent()->CreateMachineInstr(TII->get(NewOpcode), DL);
467   if (MBB->begin() != MBB->end())
468     MBB->insert(MBB->begin(), MI);
469   else
470     MBB->push_back(MI);
471   SHOWNEWINSTR(MI);
472   return MI;
473 }
474 
insertInstrBefore(MachineBasicBlock::iterator I,int NewOpcode)475 MachineInstr *AMDGPUCFGStructurizer::insertInstrBefore(
476     MachineBasicBlock::iterator I, int NewOpcode) {
477   MachineInstr *OldMI = &(*I);
478   MachineBasicBlock *MBB = OldMI->getParent();
479   MachineInstr *NewMBB =
480       MBB->getParent()->CreateMachineInstr(TII->get(NewOpcode), DebugLoc());
481   MBB->insert(I, NewMBB);
482   //assume the instruction doesn't take any reg operand ...
483   SHOWNEWINSTR(NewMBB);
484   return NewMBB;
485 }
486 
insertCondBranchBefore(MachineBasicBlock::iterator I,int NewOpcode,const DebugLoc & DL)487 void AMDGPUCFGStructurizer::insertCondBranchBefore(
488     MachineBasicBlock::iterator I, int NewOpcode, const DebugLoc &DL) {
489   MachineInstr *OldMI = &(*I);
490   MachineBasicBlock *MBB = OldMI->getParent();
491   MachineFunction *MF = MBB->getParent();
492   MachineInstr *NewMI = MF->CreateMachineInstr(TII->get(NewOpcode), DL);
493   MBB->insert(I, NewMI);
494   MachineInstrBuilder MIB(*MF, NewMI);
495   MIB.addReg(OldMI->getOperand(1).getReg(), false);
496   SHOWNEWINSTR(NewMI);
497   //erase later oldInstr->eraseFromParent();
498 }
499 
insertCondBranchBefore(MachineBasicBlock * blk,MachineBasicBlock::iterator I,int NewOpcode,int RegNum,const DebugLoc & DL)500 void AMDGPUCFGStructurizer::insertCondBranchBefore(
501     MachineBasicBlock *blk, MachineBasicBlock::iterator I, int NewOpcode,
502     int RegNum, const DebugLoc &DL) {
503   MachineFunction *MF = blk->getParent();
504   MachineInstr *NewInstr = MF->CreateMachineInstr(TII->get(NewOpcode), DL);
505   //insert before
506   blk->insert(I, NewInstr);
507   MachineInstrBuilder(*MF, NewInstr).addReg(RegNum, false);
508   SHOWNEWINSTR(NewInstr);
509 }
510 
getBranchNzeroOpcode(int OldOpcode)511 int AMDGPUCFGStructurizer::getBranchNzeroOpcode(int OldOpcode) {
512   switch(OldOpcode) {
513   case AMDGPU::JUMP_COND:
514   case AMDGPU::JUMP: return AMDGPU::IF_PREDICATE_SET;
515   case AMDGPU::BRANCH_COND_i32:
516   case AMDGPU::BRANCH_COND_f32: return AMDGPU::IF_LOGICALNZ_f32;
517   default: llvm_unreachable("internal error");
518   }
519   return -1;
520 }
521 
getBranchZeroOpcode(int OldOpcode)522 int AMDGPUCFGStructurizer::getBranchZeroOpcode(int OldOpcode) {
523   switch(OldOpcode) {
524   case AMDGPU::JUMP_COND:
525   case AMDGPU::JUMP: return AMDGPU::IF_PREDICATE_SET;
526   case AMDGPU::BRANCH_COND_i32:
527   case AMDGPU::BRANCH_COND_f32: return AMDGPU::IF_LOGICALZ_f32;
528   default: llvm_unreachable("internal error");
529   }
530   return -1;
531 }
532 
getContinueNzeroOpcode(int OldOpcode)533 int AMDGPUCFGStructurizer::getContinueNzeroOpcode(int OldOpcode) {
534   switch(OldOpcode) {
535   case AMDGPU::JUMP_COND:
536   case AMDGPU::JUMP: return AMDGPU::CONTINUE_LOGICALNZ_i32;
537   default: llvm_unreachable("internal error");
538   };
539   return -1;
540 }
541 
getContinueZeroOpcode(int OldOpcode)542 int AMDGPUCFGStructurizer::getContinueZeroOpcode(int OldOpcode) {
543   switch(OldOpcode) {
544   case AMDGPU::JUMP_COND:
545   case AMDGPU::JUMP: return AMDGPU::CONTINUE_LOGICALZ_i32;
546   default: llvm_unreachable("internal error");
547   }
548   return -1;
549 }
550 
getTrueBranch(MachineInstr * MI)551 MachineBasicBlock *AMDGPUCFGStructurizer::getTrueBranch(MachineInstr *MI) {
552   return MI->getOperand(0).getMBB();
553 }
554 
setTrueBranch(MachineInstr * MI,MachineBasicBlock * MBB)555 void AMDGPUCFGStructurizer::setTrueBranch(MachineInstr *MI,
556     MachineBasicBlock *MBB) {
557   MI->getOperand(0).setMBB(MBB);
558 }
559 
560 MachineBasicBlock *
getFalseBranch(MachineBasicBlock * MBB,MachineInstr * MI)561 AMDGPUCFGStructurizer::getFalseBranch(MachineBasicBlock *MBB,
562     MachineInstr *MI) {
563   assert(MBB->succ_size() == 2);
564   MachineBasicBlock *TrueBranch = getTrueBranch(MI);
565   MachineBasicBlock::succ_iterator It = MBB->succ_begin();
566   MachineBasicBlock::succ_iterator Next = It;
567   ++Next;
568   return (*It == TrueBranch) ? *Next : *It;
569 }
570 
isCondBranch(MachineInstr * MI)571 bool AMDGPUCFGStructurizer::isCondBranch(MachineInstr *MI) {
572   switch (MI->getOpcode()) {
573     case AMDGPU::JUMP_COND:
574     case AMDGPU::BRANCH_COND_i32:
575     case AMDGPU::BRANCH_COND_f32: return true;
576   default:
577     return false;
578   }
579   return false;
580 }
581 
isUncondBranch(MachineInstr * MI)582 bool AMDGPUCFGStructurizer::isUncondBranch(MachineInstr *MI) {
583   switch (MI->getOpcode()) {
584   case AMDGPU::JUMP:
585   case AMDGPU::BRANCH:
586     return true;
587   default:
588     return false;
589   }
590   return false;
591 }
592 
getLastDebugLocInBB(MachineBasicBlock * MBB)593 DebugLoc AMDGPUCFGStructurizer::getLastDebugLocInBB(MachineBasicBlock *MBB) {
594   //get DebugLoc from the first MachineBasicBlock instruction with debug info
595   DebugLoc DL;
596   for (MachineBasicBlock::iterator It = MBB->begin(); It != MBB->end();
597       ++It) {
598     MachineInstr *instr = &(*It);
599     if (instr->getDebugLoc())
600       DL = instr->getDebugLoc();
601   }
602   return DL;
603 }
604 
getNormalBlockBranchInstr(MachineBasicBlock * MBB)605 MachineInstr *AMDGPUCFGStructurizer::getNormalBlockBranchInstr(
606     MachineBasicBlock *MBB) {
607   MachineBasicBlock::reverse_iterator It = MBB->rbegin();
608   MachineInstr *MI = &*It;
609   if (MI && (isCondBranch(MI) || isUncondBranch(MI)))
610     return MI;
611   return nullptr;
612 }
613 
getLoopendBlockBranchInstr(MachineBasicBlock * MBB)614 MachineInstr *AMDGPUCFGStructurizer::getLoopendBlockBranchInstr(
615     MachineBasicBlock *MBB) {
616   for (MachineBasicBlock::reverse_iterator It = MBB->rbegin(), E = MBB->rend();
617       It != E; ++It) {
618     // FIXME: Simplify
619     MachineInstr *MI = &*It;
620     if (MI) {
621       if (isCondBranch(MI) || isUncondBranch(MI))
622         return MI;
623       else if (!TII->isMov(MI->getOpcode()))
624         break;
625     }
626   }
627   return nullptr;
628 }
629 
getReturnInstr(MachineBasicBlock * MBB)630 MachineInstr *AMDGPUCFGStructurizer::getReturnInstr(MachineBasicBlock *MBB) {
631   MachineBasicBlock::reverse_iterator It = MBB->rbegin();
632   if (It != MBB->rend()) {
633     MachineInstr *instr = &(*It);
634     if (instr->getOpcode() == AMDGPU::RETURN)
635       return instr;
636   }
637   return nullptr;
638 }
639 
isReturnBlock(MachineBasicBlock * MBB)640 bool AMDGPUCFGStructurizer::isReturnBlock(MachineBasicBlock *MBB) {
641   MachineInstr *MI = getReturnInstr(MBB);
642   bool IsReturn = (MBB->succ_size() == 0);
643   if (MI)
644     assert(IsReturn);
645   else if (IsReturn)
646     DEBUG(
647       dbgs() << "BB" << MBB->getNumber()
648              <<" is return block without RETURN instr\n";);
649   return  IsReturn;
650 }
651 
cloneSuccessorList(MachineBasicBlock * DstMBB,MachineBasicBlock * SrcMBB)652 void AMDGPUCFGStructurizer::cloneSuccessorList(MachineBasicBlock *DstMBB,
653     MachineBasicBlock *SrcMBB) {
654   for (MachineBasicBlock::succ_iterator It = SrcMBB->succ_begin(),
655        iterEnd = SrcMBB->succ_end(); It != iterEnd; ++It)
656     DstMBB->addSuccessor(*It);  // *iter's predecessor is also taken care of
657 }
658 
clone(MachineBasicBlock * MBB)659 MachineBasicBlock *AMDGPUCFGStructurizer::clone(MachineBasicBlock *MBB) {
660   MachineFunction *Func = MBB->getParent();
661   MachineBasicBlock *NewMBB = Func->CreateMachineBasicBlock();
662   Func->push_back(NewMBB);  //insert to function
663   for (const MachineInstr &It : *MBB)
664     NewMBB->push_back(Func->CloneMachineInstr(&It));
665   return NewMBB;
666 }
667 
replaceInstrUseOfBlockWith(MachineBasicBlock * SrcMBB,MachineBasicBlock * OldMBB,MachineBasicBlock * NewBlk)668 void AMDGPUCFGStructurizer::replaceInstrUseOfBlockWith(
669     MachineBasicBlock *SrcMBB, MachineBasicBlock *OldMBB,
670     MachineBasicBlock *NewBlk) {
671   MachineInstr *BranchMI = getLoopendBlockBranchInstr(SrcMBB);
672   if (BranchMI && isCondBranch(BranchMI) &&
673       getTrueBranch(BranchMI) == OldMBB)
674     setTrueBranch(BranchMI, NewBlk);
675 }
676 
wrapup(MachineBasicBlock * MBB)677 void AMDGPUCFGStructurizer::wrapup(MachineBasicBlock *MBB) {
678   assert((!MBB->getParent()->getJumpTableInfo()
679           || MBB->getParent()->getJumpTableInfo()->isEmpty())
680          && "found a jump table");
681 
682    //collect continue right before endloop
683    SmallVector<MachineInstr *, DEFAULT_VEC_SLOTS> ContInstr;
684    MachineBasicBlock::iterator Pre = MBB->begin();
685    MachineBasicBlock::iterator E = MBB->end();
686    MachineBasicBlock::iterator It = Pre;
687    while (It != E) {
688      if (Pre->getOpcode() == AMDGPU::CONTINUE
689          && It->getOpcode() == AMDGPU::ENDLOOP)
690        ContInstr.push_back(&*Pre);
691      Pre = It;
692      ++It;
693    }
694 
695    //delete continue right before endloop
696    for (unsigned i = 0; i < ContInstr.size(); ++i)
697       ContInstr[i]->eraseFromParent();
698 
699    // TODO to fix up jump table so later phase won't be confused.  if
700    // (jumpTableInfo->isEmpty() == false) { need to clean the jump table, but
701    // there isn't such an interface yet.  alternatively, replace all the other
702    // blocks in the jump table with the entryBlk //}
703 
704 }
705 
706 
prepare()707 bool AMDGPUCFGStructurizer::prepare() {
708   bool Changed = false;
709 
710   //FIXME: if not reducible flow graph, make it so ???
711 
712   DEBUG(dbgs() << "AMDGPUCFGStructurizer::prepare\n";);
713 
714   orderBlocks(FuncRep);
715 
716   SmallVector<MachineBasicBlock *, DEFAULT_VEC_SLOTS> RetBlks;
717 
718   // Add an ExitBlk to loop that don't have one
719   for (MachineLoopInfo::iterator It = MLI->begin(),
720        E = MLI->end(); It != E; ++It) {
721     MachineLoop *LoopRep = (*It);
722     MBBVector ExitingMBBs;
723     LoopRep->getExitingBlocks(ExitingMBBs);
724 
725     if (ExitingMBBs.size() == 0) {
726       MachineBasicBlock* DummyExitBlk = normalizeInfiniteLoopExit(LoopRep);
727       if (DummyExitBlk)
728         RetBlks.push_back(DummyExitBlk);
729     }
730   }
731 
732   // Remove unconditional branch instr.
733   // Add dummy exit block iff there are multiple returns.
734   for (SmallVectorImpl<MachineBasicBlock *>::const_iterator
735        It = OrderedBlks.begin(), E = OrderedBlks.end(); It != E; ++It) {
736     MachineBasicBlock *MBB = *It;
737     removeUnconditionalBranch(MBB);
738     removeRedundantConditionalBranch(MBB);
739     if (isReturnBlock(MBB)) {
740       RetBlks.push_back(MBB);
741     }
742     assert(MBB->succ_size() <= 2);
743   }
744 
745   if (RetBlks.size() >= 2) {
746     addDummyExitBlock(RetBlks);
747     Changed = true;
748   }
749 
750   return Changed;
751 }
752 
run()753 bool AMDGPUCFGStructurizer::run() {
754 
755   //Assume reducible CFG...
756   DEBUG(dbgs() << "AMDGPUCFGStructurizer::run\n");
757 
758 #ifdef STRESSTEST
759   //Use the worse block ordering to test the algorithm.
760   ReverseVector(orderedBlks);
761 #endif
762 
763   DEBUG(dbgs() << "Ordered blocks:\n"; printOrderedBlocks(););
764   int NumIter = 0;
765   bool Finish = false;
766   MachineBasicBlock *MBB;
767   bool MakeProgress = false;
768   int NumRemainedBlk = countActiveBlock(OrderedBlks.begin(),
769                                         OrderedBlks.end());
770 
771   do {
772     ++NumIter;
773     DEBUG(
774       dbgs() << "numIter = " << NumIter
775              << ", numRemaintedBlk = " << NumRemainedBlk << "\n";
776     );
777 
778     SmallVectorImpl<MachineBasicBlock *>::const_iterator It =
779         OrderedBlks.begin();
780     SmallVectorImpl<MachineBasicBlock *>::const_iterator E =
781         OrderedBlks.end();
782 
783     SmallVectorImpl<MachineBasicBlock *>::const_iterator SccBeginIter =
784         It;
785     MachineBasicBlock *SccBeginMBB = nullptr;
786     int SccNumBlk = 0;  // The number of active blocks, init to a
787                         // maximum possible number.
788     int SccNumIter;     // Number of iteration in this SCC.
789 
790     while (It != E) {
791       MBB = *It;
792 
793       if (!SccBeginMBB) {
794         SccBeginIter = It;
795         SccBeginMBB = MBB;
796         SccNumIter = 0;
797         SccNumBlk = NumRemainedBlk; // Init to maximum possible number.
798         DEBUG(
799               dbgs() << "start processing SCC" << getSCCNum(SccBeginMBB);
800               dbgs() << "\n";
801         );
802       }
803 
804       if (!isRetiredBlock(MBB))
805         patternMatch(MBB);
806 
807       ++It;
808 
809       bool ContNextScc = true;
810       if (It == E
811           || getSCCNum(SccBeginMBB) != getSCCNum(*It)) {
812         // Just finish one scc.
813         ++SccNumIter;
814         int sccRemainedNumBlk = countActiveBlock(SccBeginIter, It);
815         if (sccRemainedNumBlk != 1 && sccRemainedNumBlk >= SccNumBlk) {
816           DEBUG(
817             dbgs() << "Can't reduce SCC " << getSCCNum(MBB)
818                    << ", sccNumIter = " << SccNumIter;
819             dbgs() << "doesn't make any progress\n";
820           );
821           ContNextScc = true;
822         } else if (sccRemainedNumBlk != 1 && sccRemainedNumBlk < SccNumBlk) {
823           SccNumBlk = sccRemainedNumBlk;
824           It = SccBeginIter;
825           ContNextScc = false;
826           DEBUG(
827             dbgs() << "repeat processing SCC" << getSCCNum(MBB)
828                    << "sccNumIter = " << SccNumIter << '\n';
829           );
830         } else {
831           // Finish the current scc.
832           ContNextScc = true;
833         }
834       } else {
835         // Continue on next component in the current scc.
836         ContNextScc = false;
837       }
838 
839       if (ContNextScc)
840         SccBeginMBB = nullptr;
841     } //while, "one iteration" over the function.
842 
843     MachineBasicBlock *EntryMBB =
844         &*GraphTraits<MachineFunction *>::nodes_begin(FuncRep);
845     if (EntryMBB->succ_size() == 0) {
846       Finish = true;
847       DEBUG(
848         dbgs() << "Reduce to one block\n";
849       );
850     } else {
851       int NewnumRemainedBlk
852         = countActiveBlock(OrderedBlks.begin(), OrderedBlks.end());
853       // consider cloned blocks ??
854       if (NewnumRemainedBlk == 1 || NewnumRemainedBlk < NumRemainedBlk) {
855         MakeProgress = true;
856         NumRemainedBlk = NewnumRemainedBlk;
857       } else {
858         MakeProgress = false;
859         DEBUG(
860           dbgs() << "No progress\n";
861         );
862       }
863     }
864   } while (!Finish && MakeProgress);
865 
866   // Misc wrap up to maintain the consistency of the Function representation.
867   wrapup(&*GraphTraits<MachineFunction *>::nodes_begin(FuncRep));
868 
869   // Detach retired Block, release memory.
870   for (MBBInfoMap::iterator It = BlockInfoMap.begin(), E = BlockInfoMap.end();
871       It != E; ++It) {
872     if ((*It).second && (*It).second->IsRetired) {
873       assert(((*It).first)->getNumber() != -1);
874       DEBUG(
875         dbgs() << "Erase BB" << ((*It).first)->getNumber() << "\n";
876       );
877       (*It).first->eraseFromParent();  //Remove from the parent Function.
878     }
879     delete (*It).second;
880   }
881   BlockInfoMap.clear();
882   LLInfoMap.clear();
883 
884   if (!Finish) {
885     DEBUG(FuncRep->viewCFG());
886     report_fatal_error("IRREDUCIBLE_CFG");
887   }
888 
889   return true;
890 }
891 
892 
893 
orderBlocks(MachineFunction * MF)894 void AMDGPUCFGStructurizer::orderBlocks(MachineFunction *MF) {
895   int SccNum = 0;
896   MachineBasicBlock *MBB;
897   for (scc_iterator<MachineFunction *> It = scc_begin(MF); !It.isAtEnd();
898        ++It, ++SccNum) {
899     const std::vector<MachineBasicBlock *> &SccNext = *It;
900     for (std::vector<MachineBasicBlock *>::const_iterator
901          blockIter = SccNext.begin(), blockEnd = SccNext.end();
902          blockIter != blockEnd; ++blockIter) {
903       MBB = *blockIter;
904       OrderedBlks.push_back(MBB);
905       recordSccnum(MBB, SccNum);
906     }
907   }
908 
909   //walk through all the block in func to check for unreachable
910   typedef GraphTraits<MachineFunction *> GTM;
911   MachineFunction::iterator It = GTM::nodes_begin(MF), E = GTM::nodes_end(MF);
912   for (; It != E; ++It) {
913     MachineBasicBlock *MBB = &(*It);
914     SccNum = getSCCNum(MBB);
915     if (SccNum == INVALIDSCCNUM)
916       dbgs() << "unreachable block BB" << MBB->getNumber() << "\n";
917   }
918 }
919 
patternMatch(MachineBasicBlock * MBB)920 int AMDGPUCFGStructurizer::patternMatch(MachineBasicBlock *MBB) {
921   int NumMatch = 0;
922   int CurMatch;
923 
924   DEBUG(
925         dbgs() << "Begin patternMatch BB" << MBB->getNumber() << "\n";
926   );
927 
928   while ((CurMatch = patternMatchGroup(MBB)) > 0)
929     NumMatch += CurMatch;
930 
931   DEBUG(
932         dbgs() << "End patternMatch BB" << MBB->getNumber()
933       << ", numMatch = " << NumMatch << "\n";
934   );
935 
936   return NumMatch;
937 }
938 
patternMatchGroup(MachineBasicBlock * MBB)939 int AMDGPUCFGStructurizer::patternMatchGroup(MachineBasicBlock *MBB) {
940   int NumMatch = 0;
941   NumMatch += loopendPatternMatch();
942   NumMatch += serialPatternMatch(MBB);
943   NumMatch += ifPatternMatch(MBB);
944   return NumMatch;
945 }
946 
947 
serialPatternMatch(MachineBasicBlock * MBB)948 int AMDGPUCFGStructurizer::serialPatternMatch(MachineBasicBlock *MBB) {
949   if (MBB->succ_size() != 1)
950     return 0;
951 
952   MachineBasicBlock *childBlk = *MBB->succ_begin();
953   if (childBlk->pred_size() != 1 || isActiveLoophead(childBlk))
954     return 0;
955 
956   mergeSerialBlock(MBB, childBlk);
957   ++numSerialPatternMatch;
958   return 1;
959 }
960 
ifPatternMatch(MachineBasicBlock * MBB)961 int AMDGPUCFGStructurizer::ifPatternMatch(MachineBasicBlock *MBB) {
962   //two edges
963   if (MBB->succ_size() != 2)
964     return 0;
965   if (hasBackEdge(MBB))
966     return 0;
967   MachineInstr *BranchMI = getNormalBlockBranchInstr(MBB);
968   if (!BranchMI)
969     return 0;
970 
971   assert(isCondBranch(BranchMI));
972   int NumMatch = 0;
973 
974   MachineBasicBlock *TrueMBB = getTrueBranch(BranchMI);
975   NumMatch += serialPatternMatch(TrueMBB);
976   NumMatch += ifPatternMatch(TrueMBB);
977   MachineBasicBlock *FalseMBB = getFalseBranch(MBB, BranchMI);
978   NumMatch += serialPatternMatch(FalseMBB);
979   NumMatch += ifPatternMatch(FalseMBB);
980   MachineBasicBlock *LandBlk;
981   int Cloned = 0;
982 
983   assert (!TrueMBB->succ_empty() || !FalseMBB->succ_empty());
984   // TODO: Simplify
985   if (TrueMBB->succ_size() == 1 && FalseMBB->succ_size() == 1
986     && *TrueMBB->succ_begin() == *FalseMBB->succ_begin()) {
987     // Diamond pattern
988     LandBlk = *TrueMBB->succ_begin();
989   } else if (TrueMBB->succ_size() == 1 && *TrueMBB->succ_begin() == FalseMBB) {
990     // Triangle pattern, false is empty
991     LandBlk = FalseMBB;
992     FalseMBB = nullptr;
993   } else if (FalseMBB->succ_size() == 1
994              && *FalseMBB->succ_begin() == TrueMBB) {
995     // Triangle pattern, true is empty
996     // We reverse the predicate to make a triangle, empty false pattern;
997     std::swap(TrueMBB, FalseMBB);
998     reversePredicateSetter(MBB->end());
999     LandBlk = FalseMBB;
1000     FalseMBB = nullptr;
1001   } else if (FalseMBB->succ_size() == 1
1002              && isSameloopDetachedContbreak(TrueMBB, FalseMBB)) {
1003     LandBlk = *FalseMBB->succ_begin();
1004   } else if (TrueMBB->succ_size() == 1
1005     && isSameloopDetachedContbreak(FalseMBB, TrueMBB)) {
1006     LandBlk = *TrueMBB->succ_begin();
1007   } else {
1008     return NumMatch + handleJumpintoIf(MBB, TrueMBB, FalseMBB);
1009   }
1010 
1011   // improveSimpleJumpinfoIf can handle the case where landBlk == NULL but the
1012   // new BB created for landBlk==NULL may introduce new challenge to the
1013   // reduction process.
1014   if (LandBlk &&
1015       ((TrueMBB && TrueMBB->pred_size() > 1)
1016       || (FalseMBB && FalseMBB->pred_size() > 1))) {
1017      Cloned += improveSimpleJumpintoIf(MBB, TrueMBB, FalseMBB, &LandBlk);
1018   }
1019 
1020   if (TrueMBB && TrueMBB->pred_size() > 1) {
1021     TrueMBB = cloneBlockForPredecessor(TrueMBB, MBB);
1022     ++Cloned;
1023   }
1024 
1025   if (FalseMBB && FalseMBB->pred_size() > 1) {
1026     FalseMBB = cloneBlockForPredecessor(FalseMBB, MBB);
1027     ++Cloned;
1028   }
1029 
1030   mergeIfthenelseBlock(BranchMI, MBB, TrueMBB, FalseMBB, LandBlk);
1031 
1032   ++numIfPatternMatch;
1033 
1034   numClonedBlock += Cloned;
1035 
1036   return 1 + Cloned + NumMatch;
1037 }
1038 
loopendPatternMatch()1039 int AMDGPUCFGStructurizer::loopendPatternMatch() {
1040   std::deque<MachineLoop *> NestedLoops;
1041   for (auto &It: *MLI)
1042     for (MachineLoop *ML : depth_first(It))
1043       NestedLoops.push_front(ML);
1044 
1045   if (NestedLoops.size() == 0)
1046     return 0;
1047 
1048   // Process nested loop outside->inside (we did push_front),
1049   // so "continue" to a outside loop won't be mistaken as "break"
1050   // of the current loop.
1051   int Num = 0;
1052   for (MachineLoop *ExaminedLoop : NestedLoops) {
1053     if (ExaminedLoop->getNumBlocks() == 0 || Visited[ExaminedLoop])
1054       continue;
1055     DEBUG(dbgs() << "Processing:\n"; ExaminedLoop->dump(););
1056     int NumBreak = mergeLoop(ExaminedLoop);
1057     if (NumBreak == -1)
1058       break;
1059     Num += NumBreak;
1060   }
1061   return Num;
1062 }
1063 
mergeLoop(MachineLoop * LoopRep)1064 int AMDGPUCFGStructurizer::mergeLoop(MachineLoop *LoopRep) {
1065   MachineBasicBlock *LoopHeader = LoopRep->getHeader();
1066   MBBVector ExitingMBBs;
1067   LoopRep->getExitingBlocks(ExitingMBBs);
1068   assert(!ExitingMBBs.empty() && "Infinite Loop not supported");
1069   DEBUG(dbgs() << "Loop has " << ExitingMBBs.size() << " exiting blocks\n";);
1070   // We assume a single ExitBlk
1071   MBBVector ExitBlks;
1072   LoopRep->getExitBlocks(ExitBlks);
1073   SmallPtrSet<MachineBasicBlock *, 2> ExitBlkSet;
1074   for (unsigned i = 0, e = ExitBlks.size(); i < e; ++i)
1075     ExitBlkSet.insert(ExitBlks[i]);
1076   assert(ExitBlkSet.size() == 1);
1077   MachineBasicBlock *ExitBlk = *ExitBlks.begin();
1078   assert(ExitBlk && "Loop has several exit block");
1079   MBBVector LatchBlks;
1080   typedef GraphTraits<Inverse<MachineBasicBlock*> > InvMBBTraits;
1081   InvMBBTraits::ChildIteratorType PI = InvMBBTraits::child_begin(LoopHeader),
1082       PE = InvMBBTraits::child_end(LoopHeader);
1083   for (; PI != PE; PI++) {
1084     if (LoopRep->contains(*PI))
1085       LatchBlks.push_back(*PI);
1086   }
1087 
1088   for (unsigned i = 0, e = ExitingMBBs.size(); i < e; ++i)
1089     mergeLoopbreakBlock(ExitingMBBs[i], ExitBlk);
1090   for (unsigned i = 0, e = LatchBlks.size(); i < e; ++i)
1091     settleLoopcontBlock(LatchBlks[i], LoopHeader);
1092   int Match = 0;
1093   do {
1094     Match = 0;
1095     Match += serialPatternMatch(LoopHeader);
1096     Match += ifPatternMatch(LoopHeader);
1097   } while (Match > 0);
1098   mergeLooplandBlock(LoopHeader, ExitBlk);
1099   MachineLoop *ParentLoop = LoopRep->getParentLoop();
1100   if (ParentLoop)
1101     MLI->changeLoopFor(LoopHeader, ParentLoop);
1102   else
1103     MLI->removeBlock(LoopHeader);
1104   Visited[LoopRep] = true;
1105   return 1;
1106 }
1107 
isSameloopDetachedContbreak(MachineBasicBlock * Src1MBB,MachineBasicBlock * Src2MBB)1108 bool AMDGPUCFGStructurizer::isSameloopDetachedContbreak(
1109     MachineBasicBlock *Src1MBB, MachineBasicBlock *Src2MBB) {
1110   if (Src1MBB->succ_size() == 0) {
1111     MachineLoop *LoopRep = MLI->getLoopFor(Src1MBB);
1112     if (LoopRep&& LoopRep == MLI->getLoopFor(Src2MBB)) {
1113       MachineBasicBlock *&TheEntry = LLInfoMap[LoopRep];
1114       if (TheEntry) {
1115         DEBUG(
1116           dbgs() << "isLoopContBreakBlock yes src1 = BB"
1117                  << Src1MBB->getNumber()
1118                  << " src2 = BB" << Src2MBB->getNumber() << "\n";
1119         );
1120         return true;
1121       }
1122     }
1123   }
1124   return false;
1125 }
1126 
handleJumpintoIf(MachineBasicBlock * HeadMBB,MachineBasicBlock * TrueMBB,MachineBasicBlock * FalseMBB)1127 int AMDGPUCFGStructurizer::handleJumpintoIf(MachineBasicBlock *HeadMBB,
1128     MachineBasicBlock *TrueMBB, MachineBasicBlock *FalseMBB) {
1129   int Num = handleJumpintoIfImp(HeadMBB, TrueMBB, FalseMBB);
1130   if (Num == 0) {
1131     DEBUG(
1132       dbgs() << "handleJumpintoIf swap trueBlk and FalseBlk" << "\n";
1133     );
1134     Num = handleJumpintoIfImp(HeadMBB, FalseMBB, TrueMBB);
1135   }
1136   return Num;
1137 }
1138 
handleJumpintoIfImp(MachineBasicBlock * HeadMBB,MachineBasicBlock * TrueMBB,MachineBasicBlock * FalseMBB)1139 int AMDGPUCFGStructurizer::handleJumpintoIfImp(MachineBasicBlock *HeadMBB,
1140     MachineBasicBlock *TrueMBB, MachineBasicBlock *FalseMBB) {
1141   int Num = 0;
1142   MachineBasicBlock *DownBlk;
1143 
1144   //trueBlk could be the common post dominator
1145   DownBlk = TrueMBB;
1146 
1147   DEBUG(
1148     dbgs() << "handleJumpintoIfImp head = BB" << HeadMBB->getNumber()
1149            << " true = BB" << TrueMBB->getNumber()
1150            << ", numSucc=" << TrueMBB->succ_size()
1151            << " false = BB" << FalseMBB->getNumber() << "\n";
1152   );
1153 
1154   while (DownBlk) {
1155     DEBUG(
1156       dbgs() << "check down = BB" << DownBlk->getNumber();
1157     );
1158 
1159     if (singlePathTo(FalseMBB, DownBlk) == SinglePath_InPath) {
1160       DEBUG(
1161         dbgs() << " working\n";
1162       );
1163 
1164       Num += cloneOnSideEntryTo(HeadMBB, TrueMBB, DownBlk);
1165       Num += cloneOnSideEntryTo(HeadMBB, FalseMBB, DownBlk);
1166 
1167       numClonedBlock += Num;
1168       Num += serialPatternMatch(*HeadMBB->succ_begin());
1169       Num += serialPatternMatch(*std::next(HeadMBB->succ_begin()));
1170       Num += ifPatternMatch(HeadMBB);
1171       assert(Num > 0);
1172 
1173       break;
1174     }
1175     DEBUG(
1176       dbgs() << " not working\n";
1177     );
1178     DownBlk = (DownBlk->succ_size() == 1) ? (*DownBlk->succ_begin()) : nullptr;
1179   } // walk down the postDomTree
1180 
1181   return Num;
1182 }
1183 
showImproveSimpleJumpintoIf(MachineBasicBlock * HeadMBB,MachineBasicBlock * TrueMBB,MachineBasicBlock * FalseMBB,MachineBasicBlock * LandMBB,bool Detail)1184 void AMDGPUCFGStructurizer::showImproveSimpleJumpintoIf(
1185     MachineBasicBlock *HeadMBB, MachineBasicBlock *TrueMBB,
1186     MachineBasicBlock *FalseMBB, MachineBasicBlock *LandMBB, bool Detail) {
1187   dbgs() << "head = BB" << HeadMBB->getNumber()
1188          << " size = " << HeadMBB->size();
1189   if (Detail) {
1190     dbgs() << "\n";
1191     HeadMBB->print(dbgs());
1192     dbgs() << "\n";
1193   }
1194 
1195   if (TrueMBB) {
1196     dbgs() << ", true = BB" << TrueMBB->getNumber() << " size = "
1197            << TrueMBB->size() << " numPred = " << TrueMBB->pred_size();
1198     if (Detail) {
1199       dbgs() << "\n";
1200       TrueMBB->print(dbgs());
1201       dbgs() << "\n";
1202     }
1203   }
1204   if (FalseMBB) {
1205     dbgs() << ", false = BB" << FalseMBB->getNumber() << " size = "
1206            << FalseMBB->size() << " numPred = " << FalseMBB->pred_size();
1207     if (Detail) {
1208       dbgs() << "\n";
1209       FalseMBB->print(dbgs());
1210       dbgs() << "\n";
1211     }
1212   }
1213   if (LandMBB) {
1214     dbgs() << ", land = BB" << LandMBB->getNumber() << " size = "
1215            << LandMBB->size() << " numPred = " << LandMBB->pred_size();
1216     if (Detail) {
1217       dbgs() << "\n";
1218       LandMBB->print(dbgs());
1219       dbgs() << "\n";
1220     }
1221   }
1222 
1223     dbgs() << "\n";
1224 }
1225 
improveSimpleJumpintoIf(MachineBasicBlock * HeadMBB,MachineBasicBlock * TrueMBB,MachineBasicBlock * FalseMBB,MachineBasicBlock ** LandMBBPtr)1226 int AMDGPUCFGStructurizer::improveSimpleJumpintoIf(MachineBasicBlock *HeadMBB,
1227     MachineBasicBlock *TrueMBB, MachineBasicBlock *FalseMBB,
1228     MachineBasicBlock **LandMBBPtr) {
1229   bool MigrateTrue = false;
1230   bool MigrateFalse = false;
1231 
1232   MachineBasicBlock *LandBlk = *LandMBBPtr;
1233 
1234   assert((!TrueMBB || TrueMBB->succ_size() <= 1)
1235          && (!FalseMBB || FalseMBB->succ_size() <= 1));
1236 
1237   if (TrueMBB == FalseMBB)
1238     return 0;
1239 
1240   MigrateTrue = needMigrateBlock(TrueMBB);
1241   MigrateFalse = needMigrateBlock(FalseMBB);
1242 
1243   if (!MigrateTrue && !MigrateFalse)
1244     return 0;
1245 
1246   // If we need to migrate either trueBlk and falseBlk, migrate the rest that
1247   // have more than one predecessors.  without doing this, its predecessor
1248   // rather than headBlk will have undefined value in initReg.
1249   if (!MigrateTrue && TrueMBB && TrueMBB->pred_size() > 1)
1250     MigrateTrue = true;
1251   if (!MigrateFalse && FalseMBB && FalseMBB->pred_size() > 1)
1252     MigrateFalse = true;
1253 
1254   DEBUG(
1255     dbgs() << "before improveSimpleJumpintoIf: ";
1256     showImproveSimpleJumpintoIf(HeadMBB, TrueMBB, FalseMBB, LandBlk, 0);
1257   );
1258 
1259   // org: headBlk => if () {trueBlk} else {falseBlk} => landBlk
1260   //
1261   // new: headBlk => if () {initReg = 1; org trueBlk branch} else
1262   //      {initReg = 0; org falseBlk branch }
1263   //      => landBlk => if (initReg) {org trueBlk} else {org falseBlk}
1264   //      => org landBlk
1265   //      if landBlk->pred_size() > 2, put the about if-else inside
1266   //      if (initReg !=2) {...}
1267   //
1268   // add initReg = initVal to headBlk
1269 
1270   const TargetRegisterClass * I32RC = TRI->getCFGStructurizerRegClass(MVT::i32);
1271   if (!MigrateTrue || !MigrateFalse) {
1272     // XXX: We have an opportunity here to optimize the "branch into if" case
1273     // here.  Branch into if looks like this:
1274     //                        entry
1275     //                       /     |
1276     //           diamond_head       branch_from
1277     //             /      \           |
1278     // diamond_false        diamond_true
1279     //             \      /
1280     //               done
1281     //
1282     // The diamond_head block begins the "if" and the diamond_true block
1283     // is the block being "branched into".
1284     //
1285     // If MigrateTrue is true, then TrueBB is the block being "branched into"
1286     // and if MigrateFalse is true, then FalseBB is the block being
1287     // "branched into"
1288     //
1289     // Here is the pseudo code for how I think the optimization should work:
1290     // 1. Insert MOV GPR0, 0 before the branch instruction in diamond_head.
1291     // 2. Insert MOV GPR0, 1 before the branch instruction in branch_from.
1292     // 3. Move the branch instruction from diamond_head into its own basic
1293     //    block (new_block).
1294     // 4. Add an unconditional branch from diamond_head to new_block
1295     // 5. Replace the branch instruction in branch_from with an unconditional
1296     //    branch to new_block.  If branch_from has multiple predecessors, then
1297     //    we need to replace the True/False block in the branch
1298     //    instruction instead of replacing it.
1299     // 6. Change the condition of the branch instruction in new_block from
1300     //    COND to (COND || GPR0)
1301     //
1302     // In order insert these MOV instruction, we will need to use the
1303     // RegisterScavenger.  Usually liveness stops being tracked during
1304     // the late machine optimization passes, however if we implement
1305     // bool TargetRegisterInfo::requiresRegisterScavenging(
1306     //                                                const MachineFunction &MF)
1307     // and have it return true, liveness will be tracked correctly
1308     // by generic optimization passes.  We will also need to make sure that
1309     // all of our target-specific passes that run after regalloc and before
1310     // the CFGStructurizer track liveness and we will need to modify this pass
1311     // to correctly track liveness.
1312     //
1313     // After the above changes, the new CFG should look like this:
1314     //                        entry
1315     //                       /     |
1316     //           diamond_head       branch_from
1317     //                       \     /
1318     //                      new_block
1319     //                      /      |
1320     //         diamond_false        diamond_true
1321     //                      \      /
1322     //                        done
1323     //
1324     // Without this optimization, we are forced to duplicate the diamond_true
1325     // block and we will end up with a CFG like this:
1326     //
1327     //                        entry
1328     //                       /     |
1329     //           diamond_head       branch_from
1330     //             /      \                   |
1331     // diamond_false        diamond_true      diamond_true (duplicate)
1332     //             \      /                   |
1333     //               done --------------------|
1334     //
1335     // Duplicating diamond_true can be very costly especially if it has a
1336     // lot of instructions.
1337     return 0;
1338   }
1339 
1340   int NumNewBlk = 0;
1341 
1342   bool LandBlkHasOtherPred = (LandBlk->pred_size() > 2);
1343 
1344   //insert AMDGPU::ENDIF to avoid special case "input landBlk == NULL"
1345   MachineBasicBlock::iterator I = insertInstrBefore(LandBlk, AMDGPU::ENDIF);
1346 
1347   if (LandBlkHasOtherPred) {
1348     report_fatal_error("Extra register needed to handle CFG");
1349     unsigned CmpResReg =
1350       HeadMBB->getParent()->getRegInfo().createVirtualRegister(I32RC);
1351     report_fatal_error("Extra compare instruction needed to handle CFG");
1352     insertCondBranchBefore(LandBlk, I, AMDGPU::IF_PREDICATE_SET,
1353         CmpResReg, DebugLoc());
1354   }
1355 
1356   // XXX: We are running this after RA, so creating virtual registers will
1357   // cause an assertion failure in the PostRA scheduling pass.
1358   unsigned InitReg =
1359     HeadMBB->getParent()->getRegInfo().createVirtualRegister(I32RC);
1360   insertCondBranchBefore(LandBlk, I, AMDGPU::IF_PREDICATE_SET, InitReg,
1361       DebugLoc());
1362 
1363   if (MigrateTrue) {
1364     migrateInstruction(TrueMBB, LandBlk, I);
1365     // need to uncondionally insert the assignment to ensure a path from its
1366     // predecessor rather than headBlk has valid value in initReg if
1367     // (initVal != 1).
1368     report_fatal_error("Extra register needed to handle CFG");
1369   }
1370   insertInstrBefore(I, AMDGPU::ELSE);
1371 
1372   if (MigrateFalse) {
1373     migrateInstruction(FalseMBB, LandBlk, I);
1374     // need to uncondionally insert the assignment to ensure a path from its
1375     // predecessor rather than headBlk has valid value in initReg if
1376     // (initVal != 0)
1377     report_fatal_error("Extra register needed to handle CFG");
1378   }
1379 
1380   if (LandBlkHasOtherPred) {
1381     // add endif
1382     insertInstrBefore(I, AMDGPU::ENDIF);
1383 
1384     // put initReg = 2 to other predecessors of landBlk
1385     for (MachineBasicBlock::pred_iterator PI = LandBlk->pred_begin(),
1386          PE = LandBlk->pred_end(); PI != PE; ++PI) {
1387       MachineBasicBlock *MBB = *PI;
1388       if (MBB != TrueMBB && MBB != FalseMBB)
1389         report_fatal_error("Extra register needed to handle CFG");
1390     }
1391   }
1392   DEBUG(
1393     dbgs() << "result from improveSimpleJumpintoIf: ";
1394     showImproveSimpleJumpintoIf(HeadMBB, TrueMBB, FalseMBB, LandBlk, 0);
1395   );
1396 
1397   // update landBlk
1398   *LandMBBPtr = LandBlk;
1399 
1400   return NumNewBlk;
1401 }
1402 
mergeSerialBlock(MachineBasicBlock * DstMBB,MachineBasicBlock * SrcMBB)1403 void AMDGPUCFGStructurizer::mergeSerialBlock(MachineBasicBlock *DstMBB,
1404     MachineBasicBlock *SrcMBB) {
1405   DEBUG(
1406     dbgs() << "serialPattern BB" << DstMBB->getNumber()
1407            << " <= BB" << SrcMBB->getNumber() << "\n";
1408   );
1409   DstMBB->splice(DstMBB->end(), SrcMBB, SrcMBB->begin(), SrcMBB->end());
1410 
1411   DstMBB->removeSuccessor(SrcMBB, true);
1412   cloneSuccessorList(DstMBB, SrcMBB);
1413 
1414   removeSuccessor(SrcMBB);
1415   MLI->removeBlock(SrcMBB);
1416   retireBlock(SrcMBB);
1417 }
1418 
mergeIfthenelseBlock(MachineInstr * BranchMI,MachineBasicBlock * MBB,MachineBasicBlock * TrueMBB,MachineBasicBlock * FalseMBB,MachineBasicBlock * LandMBB)1419 void AMDGPUCFGStructurizer::mergeIfthenelseBlock(MachineInstr *BranchMI,
1420     MachineBasicBlock *MBB, MachineBasicBlock *TrueMBB,
1421     MachineBasicBlock *FalseMBB, MachineBasicBlock *LandMBB) {
1422   assert (TrueMBB);
1423   DEBUG(
1424     dbgs() << "ifPattern BB" << MBB->getNumber();
1425     dbgs() << "{  ";
1426     if (TrueMBB) {
1427       dbgs() << "BB" << TrueMBB->getNumber();
1428     }
1429     dbgs() << "  } else ";
1430     dbgs() << "{  ";
1431     if (FalseMBB) {
1432       dbgs() << "BB" << FalseMBB->getNumber();
1433     }
1434     dbgs() << "  }\n ";
1435     dbgs() << "landBlock: ";
1436     if (!LandMBB) {
1437       dbgs() << "NULL";
1438     } else {
1439       dbgs() << "BB" << LandMBB->getNumber();
1440     }
1441     dbgs() << "\n";
1442   );
1443 
1444   int OldOpcode = BranchMI->getOpcode();
1445   DebugLoc BranchDL = BranchMI->getDebugLoc();
1446 
1447 //    transform to
1448 //    if cond
1449 //       trueBlk
1450 //    else
1451 //       falseBlk
1452 //    endif
1453 //    landBlk
1454 
1455   MachineBasicBlock::iterator I = BranchMI;
1456   insertCondBranchBefore(I, getBranchNzeroOpcode(OldOpcode),
1457       BranchDL);
1458 
1459   if (TrueMBB) {
1460     MBB->splice(I, TrueMBB, TrueMBB->begin(), TrueMBB->end());
1461     MBB->removeSuccessor(TrueMBB, true);
1462     if (LandMBB && TrueMBB->succ_size()!=0)
1463       TrueMBB->removeSuccessor(LandMBB, true);
1464     retireBlock(TrueMBB);
1465     MLI->removeBlock(TrueMBB);
1466   }
1467 
1468   if (FalseMBB) {
1469     insertInstrBefore(I, AMDGPU::ELSE);
1470     MBB->splice(I, FalseMBB, FalseMBB->begin(),
1471                    FalseMBB->end());
1472     MBB->removeSuccessor(FalseMBB, true);
1473     if (LandMBB && FalseMBB->succ_size() != 0)
1474       FalseMBB->removeSuccessor(LandMBB, true);
1475     retireBlock(FalseMBB);
1476     MLI->removeBlock(FalseMBB);
1477   }
1478   insertInstrBefore(I, AMDGPU::ENDIF);
1479 
1480   BranchMI->eraseFromParent();
1481 
1482   if (LandMBB && TrueMBB && FalseMBB)
1483     MBB->addSuccessor(LandMBB);
1484 
1485 }
1486 
mergeLooplandBlock(MachineBasicBlock * DstBlk,MachineBasicBlock * LandMBB)1487 void AMDGPUCFGStructurizer::mergeLooplandBlock(MachineBasicBlock *DstBlk,
1488     MachineBasicBlock *LandMBB) {
1489   DEBUG(dbgs() << "loopPattern header = BB" << DstBlk->getNumber()
1490                << " land = BB" << LandMBB->getNumber() << "\n";);
1491 
1492   insertInstrBefore(DstBlk, AMDGPU::WHILELOOP, DebugLoc());
1493   insertInstrEnd(DstBlk, AMDGPU::ENDLOOP, DebugLoc());
1494   DstBlk->replaceSuccessor(DstBlk, LandMBB);
1495 }
1496 
1497 
mergeLoopbreakBlock(MachineBasicBlock * ExitingMBB,MachineBasicBlock * LandMBB)1498 void AMDGPUCFGStructurizer::mergeLoopbreakBlock(MachineBasicBlock *ExitingMBB,
1499     MachineBasicBlock *LandMBB) {
1500   DEBUG(dbgs() << "loopbreakPattern exiting = BB" << ExitingMBB->getNumber()
1501                << " land = BB" << LandMBB->getNumber() << "\n";);
1502   MachineInstr *BranchMI = getLoopendBlockBranchInstr(ExitingMBB);
1503   assert(BranchMI && isCondBranch(BranchMI));
1504   DebugLoc DL = BranchMI->getDebugLoc();
1505   MachineBasicBlock *TrueBranch = getTrueBranch(BranchMI);
1506   MachineBasicBlock::iterator I = BranchMI;
1507   if (TrueBranch != LandMBB)
1508     reversePredicateSetter(I);
1509   insertCondBranchBefore(ExitingMBB, I, AMDGPU::IF_PREDICATE_SET, AMDGPU::PREDICATE_BIT, DL);
1510   insertInstrBefore(I, AMDGPU::BREAK);
1511   insertInstrBefore(I, AMDGPU::ENDIF);
1512   //now branchInst can be erase safely
1513   BranchMI->eraseFromParent();
1514   //now take care of successors, retire blocks
1515   ExitingMBB->removeSuccessor(LandMBB, true);
1516 }
1517 
settleLoopcontBlock(MachineBasicBlock * ContingMBB,MachineBasicBlock * ContMBB)1518 void AMDGPUCFGStructurizer::settleLoopcontBlock(MachineBasicBlock *ContingMBB,
1519     MachineBasicBlock *ContMBB) {
1520   DEBUG(dbgs() << "settleLoopcontBlock conting = BB"
1521                << ContingMBB->getNumber()
1522                << ", cont = BB" << ContMBB->getNumber() << "\n";);
1523 
1524   MachineInstr *MI = getLoopendBlockBranchInstr(ContingMBB);
1525   if (MI) {
1526     assert(isCondBranch(MI));
1527     MachineBasicBlock::iterator I = MI;
1528     MachineBasicBlock *TrueBranch = getTrueBranch(MI);
1529     int OldOpcode = MI->getOpcode();
1530     DebugLoc DL = MI->getDebugLoc();
1531 
1532     bool UseContinueLogical = ((&*ContingMBB->rbegin()) == MI);
1533 
1534     if (!UseContinueLogical) {
1535       int BranchOpcode =
1536           TrueBranch == ContMBB ? getBranchNzeroOpcode(OldOpcode) :
1537           getBranchZeroOpcode(OldOpcode);
1538       insertCondBranchBefore(I, BranchOpcode, DL);
1539       // insertEnd to ensure phi-moves, if exist, go before the continue-instr.
1540       insertInstrEnd(ContingMBB, AMDGPU::CONTINUE, DL);
1541       insertInstrEnd(ContingMBB, AMDGPU::ENDIF, DL);
1542     } else {
1543       int BranchOpcode =
1544           TrueBranch == ContMBB ? getContinueNzeroOpcode(OldOpcode) :
1545           getContinueZeroOpcode(OldOpcode);
1546       insertCondBranchBefore(I, BranchOpcode, DL);
1547     }
1548 
1549     MI->eraseFromParent();
1550   } else {
1551     // if we've arrived here then we've already erased the branch instruction
1552     // travel back up the basic block to see the last reference of our debug
1553     // location we've just inserted that reference here so it should be
1554     // representative insertEnd to ensure phi-moves, if exist, go before the
1555     // continue-instr.
1556     insertInstrEnd(ContingMBB, AMDGPU::CONTINUE,
1557         getLastDebugLocInBB(ContingMBB));
1558   }
1559 }
1560 
cloneOnSideEntryTo(MachineBasicBlock * PreMBB,MachineBasicBlock * SrcMBB,MachineBasicBlock * DstMBB)1561 int AMDGPUCFGStructurizer::cloneOnSideEntryTo(MachineBasicBlock *PreMBB,
1562     MachineBasicBlock *SrcMBB, MachineBasicBlock *DstMBB) {
1563   int Cloned = 0;
1564   assert(PreMBB->isSuccessor(SrcMBB));
1565   while (SrcMBB && SrcMBB != DstMBB) {
1566     assert(SrcMBB->succ_size() == 1);
1567     if (SrcMBB->pred_size() > 1) {
1568       SrcMBB = cloneBlockForPredecessor(SrcMBB, PreMBB);
1569       ++Cloned;
1570     }
1571 
1572     PreMBB = SrcMBB;
1573     SrcMBB = *SrcMBB->succ_begin();
1574   }
1575 
1576   return Cloned;
1577 }
1578 
1579 MachineBasicBlock *
cloneBlockForPredecessor(MachineBasicBlock * MBB,MachineBasicBlock * PredMBB)1580 AMDGPUCFGStructurizer::cloneBlockForPredecessor(MachineBasicBlock *MBB,
1581     MachineBasicBlock *PredMBB) {
1582   assert(PredMBB->isSuccessor(MBB) &&
1583          "succBlk is not a prececessor of curBlk");
1584 
1585   MachineBasicBlock *CloneMBB = clone(MBB);  //clone instructions
1586   replaceInstrUseOfBlockWith(PredMBB, MBB, CloneMBB);
1587   //srcBlk, oldBlk, newBlk
1588 
1589   PredMBB->replaceSuccessor(MBB, CloneMBB);
1590 
1591   // add all successor to cloneBlk
1592   cloneSuccessorList(CloneMBB, MBB);
1593 
1594   numClonedInstr += MBB->size();
1595 
1596   DEBUG(
1597     dbgs() << "Cloned block: " << "BB"
1598            << MBB->getNumber() << "size " << MBB->size() << "\n";
1599   );
1600 
1601   SHOWNEWBLK(CloneMBB, "result of Cloned block: ");
1602 
1603   return CloneMBB;
1604 }
1605 
migrateInstruction(MachineBasicBlock * SrcMBB,MachineBasicBlock * DstMBB,MachineBasicBlock::iterator I)1606 void AMDGPUCFGStructurizer::migrateInstruction(MachineBasicBlock *SrcMBB,
1607     MachineBasicBlock *DstMBB, MachineBasicBlock::iterator I) {
1608   MachineBasicBlock::iterator SpliceEnd;
1609   //look for the input branchinstr, not the AMDGPU branchinstr
1610   MachineInstr *BranchMI = getNormalBlockBranchInstr(SrcMBB);
1611   if (!BranchMI) {
1612     DEBUG(
1613       dbgs() << "migrateInstruction don't see branch instr\n" ;
1614     );
1615     SpliceEnd = SrcMBB->end();
1616   } else {
1617     DEBUG(dbgs() << "migrateInstruction see branch instr: " << *BranchMI);
1618     SpliceEnd = BranchMI;
1619   }
1620   DEBUG(
1621     dbgs() << "migrateInstruction before splice dstSize = " << DstMBB->size()
1622       << "srcSize = " << SrcMBB->size() << "\n";
1623   );
1624 
1625   //splice insert before insertPos
1626   DstMBB->splice(I, SrcMBB, SrcMBB->begin(), SpliceEnd);
1627 
1628   DEBUG(
1629     dbgs() << "migrateInstruction after splice dstSize = " << DstMBB->size()
1630       << "srcSize = " << SrcMBB->size() << '\n';
1631   );
1632 }
1633 
1634 MachineBasicBlock *
normalizeInfiniteLoopExit(MachineLoop * LoopRep)1635 AMDGPUCFGStructurizer::normalizeInfiniteLoopExit(MachineLoop* LoopRep) {
1636   MachineBasicBlock *LoopHeader = LoopRep->getHeader();
1637   MachineBasicBlock *LoopLatch = LoopRep->getLoopLatch();
1638 
1639   if (!LoopHeader || !LoopLatch)
1640     return nullptr;
1641   MachineInstr *BranchMI = getLoopendBlockBranchInstr(LoopLatch);
1642   // Is LoopRep an infinite loop ?
1643   if (!BranchMI || !isUncondBranch(BranchMI))
1644     return nullptr;
1645 
1646   MachineBasicBlock *DummyExitBlk = FuncRep->CreateMachineBasicBlock();
1647   FuncRep->push_back(DummyExitBlk);  //insert to function
1648   SHOWNEWBLK(DummyExitBlk, "DummyExitBlock to normalize infiniteLoop: ");
1649   DEBUG(dbgs() << "Old branch instr: " << *BranchMI << "\n";);
1650   LLVMContext &Ctx = LoopHeader->getParent()->getFunction()->getContext();
1651   Ctx.emitError("Extra register needed to handle CFG");
1652   return nullptr;
1653 }
1654 
removeUnconditionalBranch(MachineBasicBlock * MBB)1655 void AMDGPUCFGStructurizer::removeUnconditionalBranch(MachineBasicBlock *MBB) {
1656   MachineInstr *BranchMI;
1657 
1658   // I saw two unconditional branch in one basic block in example
1659   // test_fc_do_while_or.c need to fix the upstream on this to remove the loop.
1660   while ((BranchMI = getLoopendBlockBranchInstr(MBB))
1661           && isUncondBranch(BranchMI)) {
1662     DEBUG(dbgs() << "Removing uncond branch instr: " << *BranchMI);
1663     BranchMI->eraseFromParent();
1664   }
1665 }
1666 
removeRedundantConditionalBranch(MachineBasicBlock * MBB)1667 void AMDGPUCFGStructurizer::removeRedundantConditionalBranch(
1668     MachineBasicBlock *MBB) {
1669   if (MBB->succ_size() != 2)
1670     return;
1671   MachineBasicBlock *MBB1 = *MBB->succ_begin();
1672   MachineBasicBlock *MBB2 = *std::next(MBB->succ_begin());
1673   if (MBB1 != MBB2)
1674     return;
1675 
1676   MachineInstr *BranchMI = getNormalBlockBranchInstr(MBB);
1677   assert(BranchMI && isCondBranch(BranchMI));
1678   DEBUG(dbgs() << "Removing unneeded cond branch instr: " << *BranchMI);
1679   BranchMI->eraseFromParent();
1680   SHOWNEWBLK(MBB1, "Removing redundant successor");
1681   MBB->removeSuccessor(MBB1, true);
1682 }
1683 
addDummyExitBlock(SmallVectorImpl<MachineBasicBlock * > & RetMBB)1684 void AMDGPUCFGStructurizer::addDummyExitBlock(
1685     SmallVectorImpl<MachineBasicBlock*> &RetMBB) {
1686   MachineBasicBlock *DummyExitBlk = FuncRep->CreateMachineBasicBlock();
1687   FuncRep->push_back(DummyExitBlk);  //insert to function
1688   insertInstrEnd(DummyExitBlk, AMDGPU::RETURN);
1689 
1690   for (SmallVectorImpl<MachineBasicBlock *>::iterator It = RetMBB.begin(),
1691        E = RetMBB.end(); It != E; ++It) {
1692     MachineBasicBlock *MBB = *It;
1693     MachineInstr *MI = getReturnInstr(MBB);
1694     if (MI)
1695       MI->eraseFromParent();
1696     MBB->addSuccessor(DummyExitBlk);
1697     DEBUG(
1698       dbgs() << "Add dummyExitBlock to BB" << MBB->getNumber()
1699              << " successors\n";
1700     );
1701   }
1702   SHOWNEWBLK(DummyExitBlk, "DummyExitBlock: ");
1703 }
1704 
removeSuccessor(MachineBasicBlock * MBB)1705 void AMDGPUCFGStructurizer::removeSuccessor(MachineBasicBlock *MBB) {
1706   while (MBB->succ_size())
1707     MBB->removeSuccessor(*MBB->succ_begin());
1708 }
1709 
recordSccnum(MachineBasicBlock * MBB,int SccNum)1710 void AMDGPUCFGStructurizer::recordSccnum(MachineBasicBlock *MBB,
1711     int SccNum) {
1712   BlockInformation *&srcBlkInfo = BlockInfoMap[MBB];
1713   if (!srcBlkInfo)
1714     srcBlkInfo = new BlockInformation();
1715   srcBlkInfo->SccNum = SccNum;
1716 }
1717 
retireBlock(MachineBasicBlock * MBB)1718 void AMDGPUCFGStructurizer::retireBlock(MachineBasicBlock *MBB) {
1719   DEBUG(
1720         dbgs() << "Retiring BB" << MBB->getNumber() << "\n";
1721   );
1722 
1723   BlockInformation *&SrcBlkInfo = BlockInfoMap[MBB];
1724 
1725   if (!SrcBlkInfo)
1726     SrcBlkInfo = new BlockInformation();
1727 
1728   SrcBlkInfo->IsRetired = true;
1729   assert(MBB->succ_size() == 0 && MBB->pred_size() == 0
1730          && "can't retire block yet");
1731 }
1732 
1733 char AMDGPUCFGStructurizer::ID = 0;
1734 
1735 } // end anonymous namespace
1736 
1737 
1738 INITIALIZE_PASS_BEGIN(AMDGPUCFGStructurizer, "amdgpustructurizer",
1739                       "AMDGPU CFG Structurizer", false, false)
INITIALIZE_PASS_DEPENDENCY(MachineDominatorTree)1740 INITIALIZE_PASS_DEPENDENCY(MachineDominatorTree)
1741 INITIALIZE_PASS_DEPENDENCY(MachinePostDominatorTree)
1742 INITIALIZE_PASS_DEPENDENCY(MachineLoopInfo)
1743 INITIALIZE_PASS_END(AMDGPUCFGStructurizer, "amdgpustructurizer",
1744                       "AMDGPU CFG Structurizer", false, false)
1745 
1746 FunctionPass *llvm::createAMDGPUCFGStructurizerPass() {
1747   return new AMDGPUCFGStructurizer();
1748 }
1749