49#define DEBUG_TYPE "gcn-dpp-combine"
51STATISTIC(NumDPPMovsCombined,
"Number of DPP moves combined.");
67 bool IsShrinkable)
const;
71 bool IsShrinkable)
const;
74 int64_t Mask = -1)
const;
78 int getDPPOp(
unsigned Op,
bool IsShrinkable)
const;
89 GCNDPPCombineLegacy() : MachineFunctionPass(ID) {}
91 bool runOnMachineFunction(MachineFunction &MF)
override;
93 StringRef getPassName()
const override {
return "GCN DPP Combine"; }
95 void getAnalysisUsage(AnalysisUsage &AU)
const override {
100 MachineFunctionProperties getRequiredProperties()
const override {
101 return MachineFunctionProperties().setIsSSA();
110char GCNDPPCombineLegacy::
ID = 0;
115 return new GCNDPPCombineLegacy();
119 unsigned Op =
MI.getOpcode();
120 if (!
TII->isVOP3(
Op)) {
123 if (!
TII->hasVALU32BitEncoding(
Op)) {
131 if (
const auto *SDst =
TII->getNamedOperand(
MI, AMDGPU::OpName::sdst)) {
136 if (!
MRI->use_nodbg_empty(SDst->getReg()))
141 if (!hasNoImmOrEqual(
MI, AMDGPU::OpName::src0_modifiers, 0, Mask) ||
142 !hasNoImmOrEqual(
MI, AMDGPU::OpName::src1_modifiers, 0, Mask) ||
143 !hasNoImmOrEqual(
MI, AMDGPU::OpName::clamp, 0) ||
144 !hasNoImmOrEqual(
MI, AMDGPU::OpName::omod, 0) ||
145 !hasNoImmOrEqual(
MI, AMDGPU::OpName::byte_sel, 0)) {
152int GCNDPPCombine::getDPPOp(
unsigned Op,
bool IsShrinkable)
const {
157 DPP32 = (E32 == -1) ? -1 : AMDGPU::
getDPPOp32(E32);
159 if (DPP32 != -1 &&
TII->pseudoToMCOpcode(DPP32) != -1)
164 if (DPP64 != -1 &&
TII->pseudoToMCOpcode(DPP64) != -1)
173MachineOperand *GCNDPPCombine::getOldOpndValue(MachineOperand &OldOpnd)
const {
178 switch(
Def->getOpcode()) {
180 case AMDGPU::IMPLICIT_DEF:
183 case AMDGPU::V_MOV_B32_e32:
184 case AMDGPU::V_MOV_B64_PSEUDO:
185 case AMDGPU::V_MOV_B64_e32:
186 case AMDGPU::V_MOV_B64_e64: {
187 auto &Op1 =
Def->getOperand(1);
196MachineInstr *GCNDPPCombine::createDPPInst(MachineInstr &OrigMI,
200 bool IsShrinkable)
const {
202 MovMI.
getOpcode() == AMDGPU::V_MOV_B64_dpp ||
203 MovMI.
getOpcode() == AMDGPU::V_MOV_B64_DPP_PSEUDO);
209 dbgs() <<
" failed: Did not expect any 16-bit uses of dpp values\n");
212 auto DPPOp = getDPPOp(OrigOp, IsShrinkable);
219 auto *RowMaskOpnd =
TII->getNamedOperand(MovMI, AMDGPU::OpName::row_mask);
220 assert(RowMaskOpnd && RowMaskOpnd->isImm());
221 auto *BankMaskOpnd =
TII->getNamedOperand(MovMI, AMDGPU::OpName::bank_mask);
222 assert(BankMaskOpnd && BankMaskOpnd->isImm());
223 const bool MaskAllLanes =
224 RowMaskOpnd->getImm() == 0xF && BankMaskOpnd->getImm() == 0xF;
227 !(
TII->isVOPC(DPPOp) || (
TII->isVOP3(DPPOp) && OrigOpE32 != -1 &&
228 TII->isVOPC(OrigOpE32)))) &&
229 "VOPC cannot form DPP unless mask is full");
238 if (
auto *Dst =
TII->getNamedOperand(OrigMI, AMDGPU::OpName::vdst)) {
242 if (
auto *SDst =
TII->getNamedOperand(OrigMI, AMDGPU::OpName::sdst)) {
250 const int OldIdx = AMDGPU::getNamedOperandIdx(DPPOp, AMDGPU::OpName::old);
252 assert(OldIdx == NumOperands);
256 TII->getNamedOperand(MovMI, AMDGPU::OpName::vdst)->getReg()),
262 }
else if (
TII->isVOPC(DPPOp) || (
TII->isVOP3(DPPOp) && OrigOpE32 != -1 &&
263 TII->isVOPC(OrigOpE32))) {
268 LLVM_DEBUG(
dbgs() <<
" failed: no old operand in DPP instruction,"
274 auto *Mod0 =
TII->getNamedOperand(OrigMI, AMDGPU::OpName::src0_modifiers);
276 assert(NumOperands == AMDGPU::getNamedOperandIdx(DPPOp,
277 AMDGPU::OpName::src0_modifiers));
280 DPPInst.addImm(Mod0->getImm());
286 auto *Src0 =
TII->getNamedOperand(MovMI, AMDGPU::OpName::src0);
288 [[maybe_unused]]
int Src0Idx = NumOperands;
291 DPPInst->getOperand(NumOperands).setIsKill(
false);
294 auto *Mod1 =
TII->getNamedOperand(OrigMI, AMDGPU::OpName::src1_modifiers);
296 assert(NumOperands == AMDGPU::getNamedOperandIdx(DPPOp,
297 AMDGPU::OpName::src1_modifiers));
300 DPPInst.addImm(Mod1->getImm());
306 auto *Src1 =
TII->getNamedOperand(OrigMI, AMDGPU::OpName::src1);
309 "dpp version of instruction missing src1");
314 assert(
TII->getOpSize(*DPPInst, Src0Idx) ==
315 TII->getOpSize(*DPPInst, NumOperands) &&
316 "Src0 and Src1 operands should have the same size");
323 auto *Mod2 =
TII->getNamedOperand(OrigMI, AMDGPU::OpName::src2_modifiers);
326 AMDGPU::getNamedOperandIdx(DPPOp, AMDGPU::OpName::src2_modifiers));
329 DPPInst.addImm(Mod2->getImm());
332 auto *Src2 =
TII->getNamedOperand(OrigMI, AMDGPU::OpName::src2);
344 auto *ClampOpr =
TII->getNamedOperand(OrigMI, AMDGPU::OpName::clamp);
346 DPPInst.addImm(ClampOpr->getImm());
348 auto *VdstInOpr =
TII->getNamedOperand(OrigMI, AMDGPU::OpName::vdst_in);
351 DPPInst.add(*VdstInOpr);
353 auto *OmodOpr =
TII->getNamedOperand(OrigMI, AMDGPU::OpName::omod);
355 DPPInst.addImm(OmodOpr->getImm());
359 if (
TII->getNamedOperand(OrigMI, AMDGPU::OpName::op_sel)) {
364 if (Mod0 &&
TII->isVOP3(OrigMI) && !
TII->isVOP3P(OrigMI))
373 DPPInst.addImm(OpSel);
375 if (
TII->getNamedOperand(OrigMI, AMDGPU::OpName::op_sel_hi)) {
383 assert(Src2 &&
"Expected vop3p with 3 operands");
385 LLVM_DEBUG(
dbgs() <<
" failed: op_sel_hi must be all set to one\n");
390 DPPInst.addImm(OpSelHi);
392 auto *NegOpr =
TII->getNamedOperand(OrigMI, AMDGPU::OpName::neg_lo);
394 DPPInst.addImm(NegOpr->getImm());
396 auto *NegHiOpr =
TII->getNamedOperand(OrigMI, AMDGPU::OpName::neg_hi);
398 DPPInst.addImm(NegHiOpr->getImm());
400 auto *ByteSelOpr =
TII->getNamedOperand(OrigMI, AMDGPU::OpName::byte_sel);
403 DPPInst.addImm(ByteSelOpr->getImm());
405 if (MachineOperand *BitOp3 =
406 TII->getNamedOperand(OrigMI, AMDGPU::OpName::bitop3)) {
408 DPPInst.add(*BitOp3);
411 DPPInst.add(*
TII->getNamedOperand(MovMI, AMDGPU::OpName::dpp_ctrl));
412 DPPInst.add(*
TII->getNamedOperand(MovMI, AMDGPU::OpName::row_mask));
413 DPPInst.add(*
TII->getNamedOperand(MovMI, AMDGPU::OpName::bank_mask));
414 DPPInst.addImm(CombBCZ ? 1 : 0);
416 constexpr AMDGPU::OpName Srcs[] = {
417 AMDGPU::OpName::src0, AMDGPU::OpName::src1, AMDGPU::OpName::src2};
423 int OpIdx = AMDGPU::getNamedOperandIdx(DPPOp, OpName);
427 if (!
TII->isOperandLegal(*DPPInst,
OpIdx)) {
436 DPPInst.getInstr()->eraseFromParent();
440 return DPPInst.getInstr();
447 case AMDGPU::V_ADD_U32_e32:
448 case AMDGPU::V_ADD_U32_e64:
449 case AMDGPU::V_ADD_CO_U32_e32:
450 case AMDGPU::V_ADD_CO_U32_e64:
451 case AMDGPU::V_OR_B32_e32:
452 case AMDGPU::V_OR_B32_e64:
453 case AMDGPU::V_SUBREV_U32_e32:
454 case AMDGPU::V_SUBREV_U32_e64:
455 case AMDGPU::V_SUBREV_CO_U32_e32:
456 case AMDGPU::V_SUBREV_CO_U32_e64:
457 case AMDGPU::V_MAX_U32_e32:
458 case AMDGPU::V_MAX_U32_e64:
459 case AMDGPU::V_XOR_B32_e32:
460 case AMDGPU::V_XOR_B32_e64:
461 if (OldOpnd->
getImm() == 0)
464 case AMDGPU::V_AND_B32_e32:
465 case AMDGPU::V_AND_B32_e64:
466 case AMDGPU::V_MIN_U32_e32:
467 case AMDGPU::V_MIN_U32_e64:
469 std::numeric_limits<uint32_t>::max())
472 case AMDGPU::V_MIN_I32_e32:
473 case AMDGPU::V_MIN_I32_e64:
474 if (
static_cast<int32_t
>(OldOpnd->
getImm()) ==
475 std::numeric_limits<int32_t>::max())
478 case AMDGPU::V_MAX_I32_e32:
479 case AMDGPU::V_MAX_I32_e64:
480 if (
static_cast<int32_t
>(OldOpnd->
getImm()) ==
481 std::numeric_limits<int32_t>::min())
484 case AMDGPU::V_MUL_I32_I24_e32:
485 case AMDGPU::V_MUL_I32_I24_e64:
486 case AMDGPU::V_MUL_U32_U24_e32:
487 case AMDGPU::V_MUL_U32_U24_e64:
488 if (OldOpnd->
getImm() == 1)
495MachineInstr *GCNDPPCombine::createDPPInst(
496 MachineInstr &OrigMI, MachineInstr &MovMI,
RegSubRegPair CombOldVGPR,
497 MachineOperand *OldOpndValue,
bool CombBCZ,
bool IsShrinkable)
const {
499 if (!CombBCZ && OldOpndValue && OldOpndValue->
isImm()) {
500 auto *Src1 =
TII->getNamedOperand(OrigMI, AMDGPU::OpName::src1);
501 if (!Src1 || !Src1->isReg()) {
502 LLVM_DEBUG(
dbgs() <<
" failed: no src1 or it isn't a register\n");
506 LLVM_DEBUG(
dbgs() <<
" failed: old immediate isn't an identity\n");
510 auto *MovDst =
TII->getNamedOperand(MovMI, AMDGPU::OpName::vdst);
511 const TargetRegisterClass *RC =
MRI->getRegClass(MovDst->getReg());
517 return createDPPInst(OrigMI, MovMI, CombOldVGPR, CombBCZ, IsShrinkable);
522bool GCNDPPCombine::hasNoImmOrEqual(MachineInstr &
MI, AMDGPU::OpName OpndName,
523 int64_t
Value, int64_t Mask)
const {
524 auto *
Imm =
TII->getNamedOperand(
MI, OpndName);
529 return (
Imm->getImm() & Mask) ==
Value;
532bool GCNDPPCombine::combineDPPMov(MachineInstr &MovMI)
const {
534 MovMI.
getOpcode() == AMDGPU::V_MOV_B64_dpp ||
535 MovMI.
getOpcode() == AMDGPU::V_MOV_B64_DPP_PSEUDO);
538 auto *DstOpnd =
TII->getNamedOperand(MovMI, AMDGPU::OpName::vdst);
539 assert(DstOpnd && DstOpnd->isReg());
540 auto DPPMovReg = DstOpnd->getReg();
541 if (DPPMovReg.isPhysical()) {
551 auto *
DppCtrl =
TII->getNamedOperand(MovMI, AMDGPU::OpName::dpp_ctrl);
553 unsigned DppCtrlVal =
DppCtrl->getImm();
554 if ((MovMI.
getOpcode() == AMDGPU::V_MOV_B64_DPP_PSEUDO ||
555 MovMI.
getOpcode() == AMDGPU::V_MOV_B64_dpp)) {
556 if (!ST->hasFeature(AMDGPU::FeatureDPALU_DPP)) {
557 LLVM_DEBUG(
dbgs() <<
" failed: 64 bit dpp move is unsupported\n");
569 auto *RowMaskOpnd =
TII->getNamedOperand(MovMI, AMDGPU::OpName::row_mask);
570 assert(RowMaskOpnd && RowMaskOpnd->isImm());
571 auto *BankMaskOpnd =
TII->getNamedOperand(MovMI, AMDGPU::OpName::bank_mask);
572 assert(BankMaskOpnd && BankMaskOpnd->isImm());
573 const bool MaskAllLanes = RowMaskOpnd->getImm() == 0xF &&
574 BankMaskOpnd->getImm() == 0xF;
576 auto *BCZOpnd =
TII->getNamedOperand(MovMI, AMDGPU::OpName::bound_ctrl);
577 assert(BCZOpnd && BCZOpnd->isImm());
578 bool BoundCtrlZero = BCZOpnd->getImm();
580 auto *OldOpnd =
TII->getNamedOperand(MovMI, AMDGPU::OpName::old);
581 auto *SrcOpnd =
TII->getNamedOperand(MovMI, AMDGPU::OpName::src0);
583 assert(SrcOpnd && SrcOpnd->isReg());
589 auto *
const OldOpndValue = getOldOpndValue(*OldOpnd);
594 assert(!OldOpndValue || OldOpndValue->
isImm() || OldOpndValue == OldOpnd);
596 bool CombBCZ =
false;
598 if (MaskAllLanes && BoundCtrlZero) {
601 if (!OldOpndValue || !OldOpndValue->
isImm()) {
606 if (OldOpndValue->
getImm() == 0) {
611 }
else if (BoundCtrlZero) {
614 " failed: old!=0 and bctrl:0 and not all lanes isn't combinable\n");
623 dbgs() << *OldOpndValue;
624 dbgs() <<
", bound_ctrl=" << CombBCZ <<
'\n');
626 SmallVector<MachineInstr*, 4> OrigMIs, DPPMIs;
627 DenseMap<MachineInstr*, SmallVector<unsigned, 4>> RegSeqWithOpNos;
630 if (CombBCZ && OldOpndValue) {
631 const TargetRegisterClass *RC =
MRI->getRegClass(DPPMovReg);
633 MRI->createVirtualRegister(RC));
635 TII->get(AMDGPU::IMPLICIT_DEF), CombOldVGPR.
Reg);
636 DPPMIs.push_back(UndefInst.getInstr());
639 OrigMIs.push_back(&MovMI);
640 bool Rollback =
true;
644 while (!
Uses.empty()) {
645 MachineOperand *
Use =
Uses.pop_back_val();
648 auto &OrigMI = *
Use->getParent();
653 "There should not be e32 True16 instructions pre-RA");
654 if (OrigOp == AMDGPU::REG_SEQUENCE) {
656 unsigned FwdSubReg = 0;
665 for (OpNo = 1; OpNo <
E; OpNo += 2) {
675 for (
auto &
Op :
MRI->use_nodbg_operands(FwdReg)) {
676 if (
Op.getSubReg() == FwdSubReg)
679 RegSeqWithOpNos[&OrigMI].push_back(OpNo);
683 bool IsShrinkable = isShrinkable(OrigMI);
684 if (!(IsShrinkable ||
685 ((
TII->isVOP3P(OrigOp) ||
TII->isVOPC(OrigOp) ||
686 TII->isVOP3(OrigOp)) &&
688 TII->isVOP1(OrigOp) ||
TII->isVOP2(OrigOp))) {
697 auto *Src0 =
TII->getNamedOperand(OrigMI, AMDGPU::OpName::src0);
698 auto *Src1 =
TII->getNamedOperand(OrigMI, AMDGPU::OpName::src1);
699 if (Use != Src0 && !(Use == Src1 && OrigMI.
isCommutable())) {
704 auto *Src2 =
TII->getNamedOperand(OrigMI, AMDGPU::OpName::src2);
705 assert(Src0 &&
"Src1 without Src0?");
706 if ((Use == Src0 && ((Src1 && Src1->isIdenticalTo(*Src0)) ||
707 (Src2 && Src2->isIdenticalTo(*Src0)))) ||
708 (Use == Src1 && (Src1->isIdenticalTo(*Src0) ||
709 (Src2 && Src2->isIdenticalTo(*Src1))))) {
713 <<
" failed: DPP register is used more than once per instruction\n");
717 if (!ST->hasFeature(AMDGPU::FeatureDPALU_DPP) &&
720 <<
" failed: DPP ALU DPP is not supported\n");
727 <<
" failed: not valid 64-bit DPP control value\n");
733 if (
auto *DPPInst = createDPPInst(OrigMI, MovMI, CombOldVGPR,
734 OldOpndValue, CombBCZ, IsShrinkable)) {
735 DPPMIs.push_back(DPPInst);
741 auto *NewMI = BB->
getParent()->CloneMachineInstr(&OrigMI);
742 BB->
insert(OrigMI, NewMI);
743 if (
TII->commuteInstruction(*NewMI)) {
746 createDPPInst(*NewMI, MovMI, CombOldVGPR, OldOpndValue, CombBCZ,
748 DPPMIs.push_back(DPPInst);
753 NewMI->eraseFromParent();
757 OrigMIs.push_back(&OrigMI);
760 Rollback |= !
Uses.empty();
762 for (
auto *
MI : *(Rollback? &DPPMIs : &OrigMIs))
763 MI->eraseFromParent();
766 for (
auto &S : RegSeqWithOpNos) {
767 if (
MRI->use_nodbg_empty(S.first->getOperand(0).getReg())) {
768 S.first->eraseFromParent();
771 while (!S.second.empty())
772 S.first->getOperand(S.second.pop_back_val()).setIsUndef();
779bool GCNDPPCombineLegacy::runOnMachineFunction(MachineFunction &MF) {
783 return GCNDPPCombine().run(MF);
786bool GCNDPPCombine::run(MachineFunction &MF) {
795 for (
auto &
MBB : MF) {
797 if (
MI.getOpcode() == AMDGPU::V_MOV_B32_dpp && combineDPPMov(
MI)) {
799 ++NumDPPMovsCombined;
800 }
else if (
MI.getOpcode() == AMDGPU::V_MOV_B64_DPP_PSEUDO ||
801 MI.getOpcode() == AMDGPU::V_MOV_B64_dpp) {
804 ++NumDPPMovsCombined;
808 if (M && combineDPPMov(*M))
809 ++NumDPPMovsCombined;
826 bool Changed = GCNDPPCombine().run(MF);
unsigned const MachineRegisterInfo * MRI
assert(UImm &&(UImm !=~static_cast< T >(0)) &&"Invalid immediate!")
Provides AMDGPU specific target descriptions.
static GCRegistry::Add< CoreCLRGC > E("coreclr", "CoreCLR-compatible GC")
static bool isIdentityValue(unsigned OrigMIOp, MachineOperand *OldOpnd)
AMD GCN specific subclass of TargetSubtarget.
const HexagonInstrInfo * TII
TargetInstrInfo::RegSubRegPair RegSubRegPair
Promote Memory to Register
MachineInstr unsigned OpIdx
#define INITIALIZE_PASS(passName, arg, name, cfg, analysis)
Remove Loads Into Fake Uses
This file defines the 'Statistic' class, which is designed to be an easy way to expose various metric...
#define STATISTIC(VARNAME, DESC)
bool useRealTrue16Insts() const
Return true if real (non-fake) variants of True16 instructions using 16-bit registers should be code-...
LLVM_ABI void setPreservesCFG()
This function should be called by the pass, iff they do not:
Represents analyses that only rely on functions' control flow.
FunctionPass class - This class is used to implement most global optimizations.
bool hasOptNone() const
Do not optimize this function (-O0).
PreservedAnalyses run(MachineFunction &MF, MachineFunctionAnalysisManager &MAM)
const SIInstrInfo * getInstrInfo() const override
const SIRegisterInfo * getRegisterInfo() const override
bool hasDPALU_DPP() const
bool hasDPPSrc1SGPR() const
unsigned getSize(const MachineInstr &MI) const
An RAII based helper class to modify MachineFunctionProperties when running pass.
const MachineFunction * getParent() const
Return the MachineFunction containing this basic block.
MachineFunctionPass - This class adapts the FunctionPass interface to allow convenient creation of pa...
void getAnalysisUsage(AnalysisUsage &AU) const override
getAnalysisUsage - Subclasses that override getAnalysisUsage must call this.
const TargetSubtargetInfo & getSubtarget() const
getSubtarget - Return the subtarget for which this machine code is being compiled.
MachineRegisterInfo & getRegInfo()
getRegInfo - Return information about the registers currently in use.
Function & getFunction()
Return the LLVM function that this machine code represents.
void insert(iterator MBBI, MachineBasicBlock *MBB)
const MachineInstrBuilder & setMIFlags(unsigned Flags) const
Representation of each machine instruction.
unsigned getOpcode() const
Returns the opcode of this MachineInstr.
const MachineBasicBlock * getParent() const
unsigned getNumOperands() const
Retuns the total number of operands.
bool modifiesRegister(Register Reg, const TargetRegisterInfo *TRI) const
Return true if the MachineInstr modifies (fully define or partially define) the specified register.
bool isCommutable(QueryType Type=IgnoreBundle) const
Return true if this may be a 2- or 3-address instruction (of the form "X = op Y, Z,...
const DebugLoc & getDebugLoc() const
Returns the debug location id of this MachineInstr.
const MachineOperand & getOperand(unsigned i) const
uint32_t getFlags() const
Return the MI flags bitvector.
MachineOperand class - Representation of each machine instruction operand.
bool isReg() const
isReg - Tests if this is a MO_Register operand.
bool isImm() const
isImm - Tests if this is a MO_Immediate operand.
Register getReg() const
getReg - Returns the register number.
MachineRegisterInfo - Keep track of information for virtual and physical registers,...
A set of analyses that are preserved following a run of a transformation pass.
static PreservedAnalyses all()
Construct a special preserved set that preserves all passes.
constexpr bool isPhysical() const
Return true if the specified register number is in the physical register namespace.
LLVM Value Representation.
LLVM_READONLY int getVOPe32(uint16_t Opcode)
LLVM_READONLY int getDPPOp32(uint16_t Opcode)
LLVM_READNONE bool isLegalDPALU_DPPControl(const MCSubtargetInfo &ST, unsigned DC)
LLVM_READONLY bool hasNamedOperand(uint64_t Opcode, OpName NamedIdx)
bool isDPALU_DPP32BitOpc(unsigned Opc)
bool isTrue16Inst(unsigned Opc)
bool isDPALU_DPP(const MCInstrDesc &OpDesc, const MCSubtargetInfo &ST)
LLVM_READONLY int getDPPOp64(uint16_t Opcode)
constexpr std::underlying_type_t< E > Mask()
Get a bitmask with 1s in all places up to the high-order bit of E's largest value.
unsigned ID
LLVM IR allows to use arbitrary numbers as calling convention identifiers.
@ Undef
Value of the register doesn't matter.
NodeAddr< DefNode * > Def
NodeAddr< UseNode * > Use
This is an optimization pass for GlobalISel generic memory operations.
FunctionAddr VTableAddr Value
TargetInstrInfo::RegSubRegPair getRegSubRegPair(const MachineOperand &O)
Create RegSubRegPair from a register MachineOperand.
MachineInstrBuilder BuildMI(MachineFunction &MF, const MIMetadata &MIMD, const MCInstrDesc &MCID)
Builder interface. Specify how to create the initial instruction itself.
auto enumerate(FirstRange &&First, RestRanges &&...Rest)
Given two or more input ranges, returns a new range whose values are tuples (A, B,...
iterator_range< early_inc_iterator_impl< detail::IterOfRange< RangeT > > > make_early_inc_range(RangeT &&Range)
Make a range that does early increment to allow mutation of the underlying range without disrupting i...
AnalysisManager< MachineFunction > MachineFunctionAnalysisManager
LLVM_ABI PreservedAnalyses getMachineFunctionPassPreservedAnalyses()
Returns the minimum set of Analyses that all machine function passes must preserve.
auto reverse(ContainerTy &&C)
char & GCNDPPCombineLegacyID
LLVM_ABI raw_ostream & dbgs()
dbgs() - This returns a reference to a raw_ostream for debugging messages.
MachineInstr * getVRegSubRegDef(const TargetInstrInfo::RegSubRegPair &P, MachineRegisterInfo &MRI)
Return the defining instruction for a given reg:subreg pair skipping copy like instructions and subre...
class LLVM_GSL_OWNER SmallVector
Forward declaration of SmallVector so that calculateSmallVectorDefaultInlinedElements can reference s...
DWARFExpression::Operation Op
iterator_range< pointer_iterator< WrappedIteratorT > > make_pointer_range(RangeT &&Range)
bool isOfRegClass(const TargetInstrInfo::RegSubRegPair &P, const TargetRegisterClass &TRC, MachineRegisterInfo &MRI)
Returns true if a reg:subreg pair P has a TRC class.
FunctionPass * createGCNDPPCombinePass()
bool execMayBeModifiedBeforeAnyUse(const MachineRegisterInfo &MRI, Register VReg, const MachineInstr &DefMI)
Return false if EXEC is not changed between the def of VReg at DefMI and all its uses.
A pair composed of a register and a sub-register index.