32 switch (
MI.getOpcode()) {
33 case R600::INTERP_PAIR_XY:
34 case R600::INTERP_PAIR_ZW:
35 case R600::INTERP_VEC_LOAD:
46 if (
TII->isLDSRetInstr(
MI.getOpcode()))
49 if (
TII->isVector(
MI) ||
TII->isCubeOp(
MI.getOpcode()) ||
50 TII->isReductionOp(
MI.getOpcode()))
53 unsigned NumLiteral = 0;
55 E =
MI.operands_end();
58 if (MO.
isReg() && MO.
getReg() == R600::ALU_LITERAL_X)
61 return 1 + NumLiteral;
65 if (
TII->isALUInstr(
MI.getOpcode()))
67 if (
TII->isVector(
MI) ||
TII->isCubeOp(
MI.getOpcode()))
69 switch (
MI.getOpcode()) {
71 case R600::INTERP_PAIR_XY:
72 case R600::INTERP_PAIR_ZW:
73 case R600::INTERP_VEC_LOAD:
83 switch (
MI.getOpcode()) {
86 case R600::IMPLICIT_DEF:
93 std::pair<unsigned, unsigned> getAccessedBankLine(
unsigned Sel)
const {
97 return std::pair<unsigned, unsigned>(
98 ((Sel >> 2) - 512) >> 12,
104 ((((Sel >> 2) - 512) & 4095) >> 5) << 1);
109 std::vector<std::pair<unsigned, unsigned>> &CachedConsts,
110 bool UpdateInstr =
true)
const {
111 std::vector<std::pair<unsigned, unsigned>> UsedKCache;
113 if (!
TII->isALUInstr(
MI.getOpcode()) &&
MI.getOpcode() != R600::DOT_4)
119 (
TII->isALUInstr(
MI.getOpcode()) ||
MI.getOpcode() == R600::DOT_4) &&
120 "Can't assign Const");
121 for (
auto &[
Op, Sel] : Consts) {
122 if (
Op->getReg() != R600::ALU_CONST)
124 unsigned Chan = Sel & 3,
Index = ((Sel >> 2) - 512) & 31;
125 unsigned KCacheIndex =
Index * 4 + Chan;
126 const std::pair<unsigned, unsigned> &BankLine = getAccessedBankLine(Sel);
127 if (CachedConsts.empty()) {
128 CachedConsts.push_back(BankLine);
129 UsedKCache.emplace_back(0, KCacheIndex);
132 if (CachedConsts[0] == BankLine) {
133 UsedKCache.emplace_back(0, KCacheIndex);
136 if (CachedConsts.size() == 1) {
137 CachedConsts.push_back(BankLine);
138 UsedKCache.emplace_back(1, KCacheIndex);
141 if (CachedConsts[1] == BankLine) {
142 UsedKCache.emplace_back(1, KCacheIndex);
152 for (
auto &[
Op, Sel] : Consts) {
153 if (
Op->getReg() != R600::ALU_CONST)
155 switch (UsedKCache[j].first) {
157 Op->setReg(R600::R600_KC0RegClass.getRegister(UsedKCache[j].second));
160 Op->setReg(R600::R600_KC1RegClass.getRegister(UsedKCache[j].second));
170 bool canClauseLocalKillFitInClause(
171 unsigned AluInstCount,
172 std::vector<std::pair<unsigned, unsigned>> KCacheBanks,
178 if (
TRI.isPhysRegLiveAcrossClauses(MO.getReg()))
183 unsigned LastUseCount = 0;
185 AluInstCount += OccupiedDwords(*UseI);
187 if (!SubstituteKCacheBank(*UseI, KCacheBanks,
false))
193 if (AluInstCount >=
TII->getMaxAlusPerClause())
202 if (UseI->readsRegister(MO.getReg(), &
TRI))
203 LastUseCount = AluInstCount;
206 if (UseI != Def && UseI->killsRegister(MO.getReg(), &
TRI))
210 return LastUseCount <=
TII->getMaxAlusPerClause();
219 std::vector<std::pair<unsigned, unsigned>> KCacheBanks;
220 bool PushBeforeModifier =
false;
221 unsigned AluInstCount = 0;
223 if (IsTrivialInst(*
I))
227 if (AluInstCount >
TII->getMaxAlusPerClause())
229 if (
I->getOpcode() == R600::PRED_X) {
236 if (AluInstCount > 0)
239 PushBeforeModifier =
true;
250 if (
TII->mustBeLastInClause(
I->getOpcode())) {
257 if (!canClauseLocalKillFitInClause(AluInstCount, KCacheBanks,
I, E))
260 if (!SubstituteKCacheBank(*
I, KCacheBanks))
262 AluInstCount += OccupiedDwords(*
I);
264 unsigned Opcode = PushBeforeModifier ?
265 R600::CF_ALU_PUSH_BEFORE : R600::CF_ALU;
272 .
addImm(KCacheBanks.empty()?0:KCacheBanks[0].first)
273 .
addImm((KCacheBanks.size() < 2)?0:KCacheBanks[1].first)
274 .
addImm(KCacheBanks.empty()?0:2)
275 .
addImm((KCacheBanks.size() < 2)?0:2)
276 .
addImm(KCacheBanks.empty()?0:KCacheBanks[0].second)
277 .
addImm((KCacheBanks.size() < 2)?0:KCacheBanks[1].second)
290 TII =
ST.getInstrInfo();
294 if (
I !=
MBB.
end() &&
I->getOpcode() == R600::CF_ALU)
298 auto next = MakeALUClause(
MBB,
I);
309 return "R600 Emit Clause Markers Pass";
313char R600EmitClauseMarkers::ID = 0;
318 "R600 Emit Clause Markers",
false,
false)
323 return new R600EmitClauseMarkers();
assert(UImm &&(UImm !=~static_cast< T >(0)) &&"Invalid immediate!")
const HexagonInstrInfo * TII
Register const TargetRegisterInfo * TRI
#define INITIALIZE_PASS_END(passName, arg, name, cfg, analysis)
#define INITIALIZE_PASS_BEGIN(passName, arg, name, cfg, analysis)
Provides R600 specific target descriptions.
AMDGPU R600 specific subclass of TargetSubtarget.
This class represents an Operation in the Expression.
FunctionPass class - This class is used to implement most global optimizations.
LLVM_ABI DebugLoc findDebugLoc(instr_iterator MBBI)
Find the next valid DebugLoc starting at MBBI, skipping any debug instructions.
MachineFunctionPass - This class adapts the FunctionPass interface to allow convenient creation of pa...
virtual bool runOnMachineFunction(MachineFunction &MF)=0
runOnMachineFunction - This method must be overloaded to perform the desired machine code transformat...
const TargetSubtargetInfo & getSubtarget() const
getSubtarget - Return the subtarget for which this machine code is being compiled.
const MachineInstrBuilder & addImm(int64_t Val) const
Add a new immediate operand.
Representation of each machine instruction.
MachineOperand class - Representation of each machine instruction operand.
bool isReg() const
isReg - Tests if this is a MO_Register operand.
Register getReg() const
getReg - Returns the register number.
virtual StringRef getPassName() const
getPassName - Return a nice clean name for a pass.
This class consists of common code factored out of the SmallVector class to reduce code duplication b...
StringRef - Represent a constant reference to a string, i.e.
#define llvm_unreachable(msg)
Marks that the current location is not supposed to be reachable.
unsigned ID
LLVM IR allows to use arbitrary numbers as calling convention identifiers.
NodeAddr< DefNode * > Def
This is an optimization pass for GlobalISel generic memory operations.
MachineInstrBuilder BuildMI(MachineFunction &MF, const MIMetadata &MIMD, const MCInstrDesc &MCID)
Builder interface. Specify how to create the initial instruction itself.
FunctionPass * createR600EmitClauseMarkers()