Lines Matching refs:ScratchReg
2052 unsigned ScratchReg = GetScratchRegister(Is64Bit, IsLP64, MF, true); in adjustForSegmentedStacks() local
2053 assert(!MF.getRegInfo().isLiveIn(ScratchReg) && in adjustForSegmentedStacks()
2121 ScratchReg = IsLP64 ? X86::RSP : X86::ESP; in adjustForSegmentedStacks()
2123 … BuildMI(checkMBB, DL, TII.get(IsLP64 ? X86::LEA64r : X86::LEA64_32r), ScratchReg).addReg(X86::RSP) in adjustForSegmentedStacks()
2126 BuildMI(checkMBB, DL, TII.get(IsLP64 ? X86::CMP64rm : X86::CMP32rm)).addReg(ScratchReg) in adjustForSegmentedStacks()
2148 ScratchReg = X86::ESP; in adjustForSegmentedStacks()
2150 BuildMI(checkMBB, DL, TII.get(X86::LEA32r), ScratchReg).addReg(X86::ESP) in adjustForSegmentedStacks()
2155 BuildMI(checkMBB, DL, TII.get(X86::CMP32rm)).addReg(ScratchReg) in adjustForSegmentedStacks()
2186 .addReg(ScratchReg) in adjustForSegmentedStacks()
2364 unsigned ScratchReg, SPReg, PReg, SPLimitOffset; in adjustForHiPEPrologue() local
2382 ScratchReg = GetScratchRegister(Is64Bit, IsLP64, MF, true); in adjustForHiPEPrologue()
2383 assert(!MF.getRegInfo().isLiveIn(ScratchReg) && in adjustForHiPEPrologue()
2387 addRegOffset(BuildMI(stackCheckMBB, DL, TII.get(LEAop), ScratchReg), in adjustForHiPEPrologue()
2391 .addReg(ScratchReg), PReg, false, SPLimitOffset); in adjustForHiPEPrologue()
2397 addRegOffset(BuildMI(incStackMBB, DL, TII.get(LEAop), ScratchReg), in adjustForHiPEPrologue()
2400 .addReg(ScratchReg), PReg, false, SPLimitOffset); in adjustForHiPEPrologue()