40#define DEBUG_TYPE "loadstore-opt"
44using namespace MIPatternMatch;
46STATISTIC(NumStoresMerged,
"Number of stores merged");
65 AA = &getAnalysis<AAResultsWrapperPass>().getAAResults();
90 Info.setBase(BaseReg);
93 Info.setOffset(RHSCst->Value.getSExtValue());
97 Info.setIndex(PtrAddRHS);
105 auto *LdSt1 = dyn_cast<GLoadStore>(&MI1);
106 auto *LdSt2 = dyn_cast<GLoadStore>(&MI2);
107 if (!LdSt1 || !LdSt2)
132 IsAlias = !((int64_t)Size1.
getValue() <= PtrDiff);
139 IsAlias = !((PtrDiff + (int64_t)Size2.
getValue()) <= 0);
151 if (!Base0Def || !Base1Def)
155 if (Base0Def->getOpcode() != Base1Def->getOpcode())
158 if (Base0Def->getOpcode() == TargetOpcode::G_FRAME_INDEX) {
162 if (Base0Def != Base1Def &&
172 if (Base0Def->getOpcode() == TargetOpcode::G_GLOBAL_VALUE) {
173 auto GV0 = Base0Def->getOperand(1).getGlobal();
174 auto GV1 = Base1Def->getOperand(1).getGlobal();
189 struct MemUseCharacteristics {
198 auto getCharacteristics =
200 if (
const auto *LS = dyn_cast<GLoadStore>(
MI)) {
206 BaseReg = LS->getPointerReg();
211 return {LS->isVolatile(), LS->isAtomic(), BaseReg,
223 MemUseCharacteristics MUC0 = getCharacteristics(&
MI),
224 MUC1 = getCharacteristics(&
Other);
227 if (MUC0.BasePtr.isValid() && MUC0.BasePtr == MUC1.BasePtr &&
228 MUC0.Offset == MUC1.Offset)
232 if (MUC0.IsVolatile && MUC1.IsVolatile)
237 if (MUC0.IsAtomic && MUC1.IsAtomic)
242 if (MUC0.MMO && MUC1.MMO) {
243 if ((MUC0.MMO->isInvariant() && MUC1.MMO->isStore()) ||
244 (MUC1.MMO->isInvariant() && MUC0.MMO->isStore()))
250 if ((MUC0.NumBytes.isScalable() && MUC0.Offset != 0) ||
251 (MUC1.NumBytes.isScalable() && MUC1.Offset != 0))
254 const bool BothNotScalable =
255 !MUC0.NumBytes.isScalable() && !MUC1.NumBytes.isScalable();
260 if (BothNotScalable &&
265 if (!MUC0.MMO || !MUC1.MMO)
269 int64_t SrcValOffset0 = MUC0.MMO->getOffset();
270 int64_t SrcValOffset1 = MUC1.MMO->getOffset();
273 if (AA && MUC0.MMO->getValue() && MUC1.MMO->getValue() && Size0.
hasValue() &&
276 int64_t MinOffset = std::min(SrcValOffset0, SrcValOffset1);
287 MemoryLocation(MUC0.MMO->getValue(), Loc0, MUC0.MMO->getAAInfo()),
288 MemoryLocation(MUC1.MMO->getValue(), Loc1, MUC1.MMO->getAAInfo())))
299 return MI.hasUnmodeledSideEffects() ||
MI.hasOrderedMemoryRef();
305 assert(StoresToMerge.
size() > 1 &&
"Expected multiple stores to merge");
306 LLT OrigTy = MRI->
getType(StoresToMerge[0]->getValueReg());
307 LLT PtrTy = MRI->
getType(StoresToMerge[0]->getPointerReg());
310 initializeStoreMergeTargetInfo(AS);
311 const auto &LegalSizes = LegalStoreSizes[AS];
314 for (
auto *StoreMI : StoresToMerge)
318 bool AnyMerged =
false;
323 unsigned MergeSizeBits;
324 for (MergeSizeBits = MaxSizeBits; MergeSizeBits > 1; MergeSizeBits /= 2) {
328 if (LegalSizes.size() > MergeSizeBits && LegalSizes[MergeSizeBits] &&
336 unsigned NumStoresToMerge = MergeSizeBits / OrigTy.
getSizeInBits();
339 StoresToMerge.begin(), StoresToMerge.begin() + NumStoresToMerge);
340 AnyMerged |= doSingleStoreMerge(SingleMergeStores);
341 StoresToMerge.erase(StoresToMerge.begin(),
342 StoresToMerge.begin() + NumStoresToMerge);
343 }
while (StoresToMerge.size() > 1);
347bool LoadStoreOpt::isLegalOrBeforeLegalizer(
const LegalityQuery &Query,
353 return IsPreLegalizer || Action == LegalizeAction::Legal;
363 GStore *FirstStore = Stores[0];
364 const unsigned NumStores = Stores.
size();
381 for (
auto *Store : Stores) {
385 ConstantVals.
clear();
394 if (ConstantVals.
empty()) {
403 if (!isLegalOrBeforeLegalizer({TargetOpcode::G_CONSTANT, {WideValueTy}}, *MF))
416 <<
" stores into merged store: " << *NewStore);
418 NumStoresMerged += Stores.size();
425 R <<
"Merged " <<
NV(
"NumMerged", Stores.size()) <<
" stores of "
427 <<
" bytes into a single store of "
432 InstsToErase.insert_range(Stores);
436bool LoadStoreOpt::processMergeCandidate(StoreMergeCandidate &
C) {
437 if (
C.Stores.size() < 2) {
442 LLVM_DEBUG(
dbgs() <<
"Checking store merge candidate with " <<
C.Stores.size()
443 <<
" stores, starting with " << *
C.Stores[0]);
457 auto DoesStoreAliasWithPotential = [&](
unsigned Idx,
GStore &CheckStore) {
458 for (
auto AliasInfo :
reverse(
C.PotentialAliases)) {
460 unsigned PreCheckedIdx = AliasInfo.second;
461 if (
Idx < PreCheckedIdx) {
479 for (
int StoreIdx =
C.Stores.size() - 1; StoreIdx >= 0; --StoreIdx) {
480 auto *CheckStore =
C.Stores[StoreIdx];
481 if (DoesStoreAliasWithPotential(StoreIdx, *CheckStore))
487 <<
" stores remaining after alias checks. Merging...\n");
491 if (StoresToMerge.
size() < 2)
493 return mergeStores(StoresToMerge);
497 StoreMergeCandidate &
C) {
498 if (
C.Stores.empty())
501 return instMayAlias(MI, *OtherMI, *MRI, AA);
505void LoadStoreOpt::StoreMergeCandidate::addPotentialAlias(
MachineInstr &
MI) {
506 PotentialAliases.emplace_back(std::make_pair(&
MI, Stores.size() - 1));
509bool LoadStoreOpt::addStoreToCandidate(
GStore &StoreMI,
510 StoreMergeCandidate &
C) {
533 if (
C.Stores.empty()) {
534 C.BasePtr = StoreBase;
535 if (!BIO.hasValidOffset()) {
536 C.CurrentLowestOffset = 0;
538 C.CurrentLowestOffset = BIO.getOffset();
543 if (BIO.hasValidOffset() &&
546 C.Stores.emplace_back(&StoreMI);
547 LLVM_DEBUG(
dbgs() <<
"Starting a new merge candidate group with: "
563 if (
C.BasePtr != StoreBase)
567 if (!BIO.hasValidOffset())
569 if ((
C.CurrentLowestOffset -
570 static_cast<int64_t
>(ValueTy.
getSizeInBytes())) != BIO.getOffset())
574 C.Stores.emplace_back(&StoreMI);
581 bool Changed =
false;
583 StoreMergeCandidate Candidate;
585 if (InstsToErase.contains(&
MI))
588 if (
auto *StoreMI = dyn_cast<GStore>(&
MI)) {
591 if (!addStoreToCandidate(*StoreMI, Candidate)) {
594 if (operationAliasesWithCandidate(*StoreMI, Candidate)) {
595 Changed |= processMergeCandidate(Candidate);
598 Candidate.addPotentialAlias(*StoreMI);
604 if (Candidate.Stores.empty())
609 Changed |= processMergeCandidate(Candidate);
610 Candidate.Stores.clear();
614 if (!
MI.mayLoadOrStore())
617 if (operationAliasesWithCandidate(
MI, Candidate)) {
620 Changed |= processMergeCandidate(Candidate);
626 Candidate.addPotentialAlias(
MI);
630 Changed |= processMergeCandidate(Candidate);
633 for (
auto *
MI : InstsToErase)
634 MI->eraseFromParent();
635 InstsToErase.clear();
645static std::optional<int64_t>
663 if (!SrcVal.
isValid() || TruncVal == SrcVal) {
671 unsigned NarrowBits = Store.getMMO().getMemoryType().getScalarSizeInBits();
672 if (ShiftAmt % NarrowBits != 0)
674 const unsigned Offset = ShiftAmt / NarrowBits;
676 if (SrcVal.
isValid() && FoundSrcVal != SrcVal)
680 SrcVal = FoundSrcVal;
681 else if (
MRI.getType(SrcVal) !=
MRI.getType(FoundSrcVal))
708bool LoadStoreOpt::mergeTruncStore(
GStore &StoreMI,
736 auto &LastStore = StoreMI;
741 if (!
mi_match(LastStore.getPointerReg(), *MRI,
743 BaseReg = LastStore.getPointerReg();
747 GStore *LowestIdxStore = &LastStore;
748 int64_t LowestIdxOffset = LastOffset;
760 const unsigned NumStoresRequired =
764 OffsetMap[*LowestShiftAmt] = LastOffset;
767 const int MaxInstsToCheck = 10;
768 int NumInstsChecked = 0;
769 for (
auto II = ++LastStore.getReverseIterator();
770 II != LastStore.getParent()->rend() && NumInstsChecked < MaxInstsToCheck;
774 if ((NewStore = dyn_cast<GStore>(&*
II))) {
777 }
else if (
II->isLoadFoldBarrier() ||
II->mayLoad()) {
791 if (BaseReg != NewBaseReg)
795 if (!ShiftByteOffset)
797 if (MemOffset < LowestIdxOffset) {
798 LowestIdxOffset = MemOffset;
799 LowestIdxStore = NewStore;
804 if (*ShiftByteOffset < 0 || *ShiftByteOffset >= NumStoresRequired ||
805 OffsetMap[*ShiftByteOffset] !=
INT64_MAX)
807 OffsetMap[*ShiftByteOffset] = MemOffset;
812 if (FoundStores.
size() == NumStoresRequired)
816 if (FoundStores.
size() != NumStoresRequired) {
817 if (FoundStores.
size() == 1)
825 unsigned NumStoresFound = FoundStores.
size();
827 const auto &
DL = LastStore.getMF()->getDataLayout();
828 auto &
C = LastStore.getMF()->getFunction().getContext();
833 if (!Allowed || !
Fast)
839 auto checkOffsets = [&](
bool MatchLittleEndian) {
840 if (MatchLittleEndian) {
841 for (
unsigned i = 0; i != NumStoresFound; ++i)
842 if (OffsetMap[i] != i * (NarrowBits / 8) + LowestIdxOffset)
845 for (
unsigned i = 0, j = NumStoresFound - 1; i != NumStoresFound;
847 if (OffsetMap[j] != i * (NarrowBits / 8) + LowestIdxOffset)
854 bool NeedBswap =
false;
855 bool NeedRotate =
false;
856 if (!checkOffsets(
DL.isLittleEndian())) {
858 if (NarrowBits == 8 && checkOffsets(
DL.isBigEndian()))
860 else if (NumStoresFound == 2 && checkOffsets(
DL.isBigEndian()))
867 !isLegalOrBeforeLegalizer({TargetOpcode::G_BSWAP, {WideStoreTy}}, *MF))
870 !isLegalOrBeforeLegalizer(
871 {TargetOpcode::G_ROTR, {WideStoreTy, WideStoreTy}}, *MF))
876 if (WideStoreTy != MRI->
getType(WideSrcVal))
881 }
else if (NeedRotate) {
883 "Unexpected type for rotate");
895 for (
auto *ST : FoundStores) {
896 ST->eraseFromParent();
903 bool Changed =
false;
908 if (
auto *StoreMI = dyn_cast<GStore>(&
MI))
911 for (
auto *StoreMI : Stores) {
912 if (DeletedStores.
count(StoreMI))
914 if (mergeTruncStore(*StoreMI, DeletedStores))
921 bool Changed =
false;
923 Changed |= mergeBlockStores(BB);
924 Changed |= mergeTruncStoresBlock(BB);
929 for (
auto &BB : MF) {
940void LoadStoreOpt::initializeStoreMergeTargetInfo(
unsigned AddrSpace) {
945 if (LegalStoreSizes.count(AddrSpace)) {
946 assert(LegalStoreSizes[AddrSpace].
any());
952 const auto &LI = *MF->getSubtarget().getLegalizerInfo();
953 const auto &
DL = MF->getFunction().getDataLayout();
966 LegalSizes.set(
Size);
968 assert(LegalSizes.any() &&
"Expected some store sizes to be legal!");
969 LegalStoreSizes[AddrSpace] = LegalSizes;
981 bool Changed =
false;
982 Changed |= mergeFunctionStores(MF);
984 LegalStoreSizes.clear();
unsigned const MachineRegisterInfo * MRI
assert(UImm &&(UImm !=~static_cast< T >(0)) &&"Invalid immediate!")
MachineBasicBlock MachineBasicBlock::iterator DebugLoc DL
Atomic ordering constants.
Analysis containing CSE Info
Returns the sub type a function will return at a given Idx Should correspond to the result type of an ExtractValue instruction executed with just that one unsigned Idx
Performs the initial survey of the specified function
Declares convenience wrapper classes for interpreting MachineInstr instances as specific generic oper...
Interface for Targets to specify which operations they can successfully select and how the others sho...
Generic memory optimizations
const unsigned MaxStoreSizeToForm
static std::optional< int64_t > getTruncStoreByteOffset(GStore &Store, Register &SrcVal, MachineRegisterInfo &MRI)
Check if the store Store is a truncstore that can be merged.
static bool isInstHardMergeHazard(MachineInstr &MI)
Returns true if the instruction creates an unavoidable hazard that forces a boundary between store me...
Implement a low-level type suitable for MachineInstr level instruction selection.
Contains matchers for matching SSA Machine Instructions.
This file provides utility analysis objects describing memory locations.
uint64_t IntrinsicInst * II
#define INITIALIZE_PASS_END(passName, arg, name, cfg, analysis)
#define INITIALIZE_PASS_BEGIN(passName, arg, name, cfg, analysis)
This file defines the SmallPtrSet class.
This file defines the 'Statistic' class, which is designed to be an easy way to expose various metric...
#define STATISTIC(VARNAME, DESC)
This file describes how to lower LLVM code to machine code.
A wrapper pass to provide the legacy pass manager access to a suitably prepared AAResults object.
A private abstract base class describing the concept of an individual alias analysis implementation.
bool isNoAlias(const MemoryLocation &LocA, const MemoryLocation &LocB)
A trivial helper function to check to see if the specified pointers are no-alias.
Class for arbitrary precision integers.
Represent the analysis usage information of a pass.
AnalysisUsage & addRequired()
void setPreservesAll()
Set by analyses that do not transform their input at all.
static LLVM_ABI DebugLoc getMergedLocation(DebugLoc LocA, DebugLoc LocB)
When two instructions are combined into a single instruction we also need to combine the original loc...
LLVMContext & getContext() const
getContext - Return a reference to the LLVMContext associated with this function.
Helper struct to store a base, index and offset that forms an address.
int64_t getOffset() const
bool hasValidOffset() const
Register getPointerReg() const
Get the source register of the pointer value.
MachineMemOperand & getMMO() const
Get the MachineMemOperand on this instruction.
LocationSize getMemSizeInBits() const
Returns the size in bits of the memory access.
bool isSimple() const
Returns true if the memory operation is neither atomic or volatile.
Register getValueReg() const
Get the stored value register.
constexpr unsigned getScalarSizeInBits() const
constexpr bool isScalar() const
static constexpr LLT scalar(unsigned SizeInBits)
Get a low-level scalar or aggregate "bag of bits".
constexpr TypeSize getSizeInBits() const
Returns the total size of the type. Must only be called on sized types.
constexpr unsigned getAddressSpace() const
constexpr TypeSize getSizeInBytes() const
Returns the total size of the type in bytes, i.e.
LegalizeActionStep getAction(const LegalityQuery &Query) const
Determine what action should be taken to legalize the described instruction.
void getAnalysisUsage(AnalysisUsage &AU) const override
getAnalysisUsage - This function should be overriden by passes that need analysis information to do t...
bool runOnMachineFunction(MachineFunction &MF) override
runOnMachineFunction - This method must be overloaded to perform the desired machine code transformat...
static LocationSize precise(uint64_t Value)
static constexpr LocationSize beforeOrAfterPointer()
Any location before or after the base pointer (but still within the underlying object).
TypeSize getValue() const
The MachineFrameInfo class represents an abstract stack frame until prolog/epilog code is inserted.
bool isFixedObjectIndex(int ObjectIdx) const
Returns true if the specified index corresponds to a fixed stack object.
MachineFunctionPass - This class adapts the FunctionPass interface to allow convenient creation of pa...
void getAnalysisUsage(AnalysisUsage &AU) const override
getAnalysisUsage - Subclasses that override getAnalysisUsage must call this.
const TargetSubtargetInfo & getSubtarget() const
getSubtarget - Return the subtarget for which this machine code is being compiled.
StringRef getName() const
getName - Return the name of the corresponding LLVM function.
MachineMemOperand * getMachineMemOperand(MachinePointerInfo PtrInfo, MachineMemOperand::Flags f, LLT MemTy, Align base_alignment, const AAMDNodes &AAInfo=AAMDNodes(), const MDNode *Ranges=nullptr, SyncScope::ID SSID=SyncScope::System, AtomicOrdering Ordering=AtomicOrdering::NotAtomic, AtomicOrdering FailureOrdering=AtomicOrdering::NotAtomic)
getMachineMemOperand - Allocate a new MachineMemOperand.
MachineRegisterInfo & getRegInfo()
getRegInfo - Return information about the registers currently in use.
Function & getFunction()
Return the LLVM function that this machine code represents.
const MachineFunctionProperties & getProperties() const
Get the function properties.
MachineInstrBuilder buildRotateRight(const DstOp &Dst, const SrcOp &Src, const SrcOp &Amt)
Build and insert Dst = G_ROTR Src, Amt.
void setInstr(MachineInstr &MI)
Set the insertion point to before MI.
MachineInstrBuilder buildBSwap(const DstOp &Dst, const SrcOp &Src0)
Build and insert Dst = G_BSWAP Src0.
MachineInstrBuilder buildStore(const SrcOp &Val, const SrcOp &Addr, MachineMemOperand &MMO)
Build and insert G_STORE Val, Addr, MMO.
void setInstrAndDebugLoc(MachineInstr &MI)
Set the insertion point to before MI, and set the debug loc to MI's loc.
MachineInstrBuilder buildTrunc(const DstOp &Res, const SrcOp &Op, std::optional< unsigned > Flags=std::nullopt)
Build and insert Res = G_TRUNC Op.
void setDebugLoc(const DebugLoc &DL)
Set the debug location to DL for all the next build instructions.
void setMF(MachineFunction &MF)
virtual MachineInstrBuilder buildConstant(const DstOp &Res, const ConstantInt &Val)
Build and insert Res = G_CONSTANT Val.
Register getReg(unsigned Idx) const
Get the register for the operand index.
Representation of each machine instruction.
const MachineBasicBlock * getParent() const
const DebugLoc & getDebugLoc() const
Returns the debug location id of this MachineInstr.
A description of a memory reference used in the backend.
LLT getMemoryType() const
Return the memory type of the memory reference.
const MachinePointerInfo & getPointerInfo() const
LLVM_ABI Align getAlign() const
Return the minimum known alignment in bytes of the actual memory reference.
MachineRegisterInfo - Keep track of information for virtual and physical registers,...
LLT getType(Register Reg) const
Get the low-level type of Reg or LLT{} if Reg is not a generic (target independent) virtual register.
Representation for a specific memory location.
static LLVM_ABI PointerType * get(Type *ElementType, unsigned AddressSpace)
This constructs a pointer to an object of the specified type in a numbered address space.
Wrapper class representing virtual and physical registers.
constexpr bool isValid() const
A templated base class for SmallPtrSet which provides the typesafe interface that is common across al...
size_type count(ConstPtrType Ptr) const
count - Return 1 if the specified pointer is in the set, 0 otherwise.
std::pair< iterator, bool > insert(PtrType Ptr)
Inserts Ptr if and only if there is no element in the container equal to Ptr.
SmallPtrSet - This class implements a set which is optimized for holding SmallSize or less elements.
This class consists of common code factored out of the SmallVector class to reduce code duplication b...
reference emplace_back(ArgTypes &&... Args)
This is a 'vector' (really, a variable-sized array), optimized for the case when the array is small.
bool isTypeLegal(EVT VT) const
Return true if the target has native support for the specified value type.
virtual bool allowsMemoryAccess(LLVMContext &Context, const DataLayout &DL, EVT VT, unsigned AddrSpace=0, Align Alignment=Align(1), MachineMemOperand::Flags Flags=MachineMemOperand::MONone, unsigned *Fast=nullptr) const
Return true if the target supports a memory access of this type for the given address space and align...
virtual bool canMergeStoresTo(unsigned AS, EVT MemVT, const MachineFunction &MF) const
Returns if it's reasonable to merge stores to MemVT size.
virtual const LegalizerInfo * getLegalizerInfo() const
virtual const TargetLowering * getTargetLowering() const
The instances of the Type class are immutable: once they are created, they are never changed.
constexpr ScalarTy getFixedValue() const
constexpr ScalarTy getKnownMinValue() const
Returns the minimum value this quantity can represent.
constexpr bool any(E Val)
@ Fast
Attempts to make calls as fast as possible (e.g.
@ C
The default llvm calling convention, compatible with C.
LLVM_ABI bool aliasIsKnownForLoadStore(const MachineInstr &MI1, const MachineInstr &MI2, bool &IsAlias, MachineRegisterInfo &MRI)
Compute whether or not a memory access at MI1 aliases with an access at MI2.
LLVM_ABI BaseIndexOffset getPointerInfo(Register Ptr, MachineRegisterInfo &MRI)
Returns a BaseIndexOffset which describes the pointer in Ptr.
LLVM_ABI bool instMayAlias(const MachineInstr &MI, const MachineInstr &Other, MachineRegisterInfo &MRI, AliasAnalysis *AA)
Returns true if the instruction MI may alias Other.
@ Legal
The operation is expected to be selectable directly by the target, and no transformation is necessary...
@ Unsupported
This operation is completely unsupported on the target.
operand_type_match m_Reg()
ConstantMatch< APInt > m_ICst(APInt &Cst)
BinaryOp_match< LHS, RHS, TargetOpcode::G_ASHR, false > m_GAShr(const LHS &L, const RHS &R)
bool mi_match(Reg R, const MachineRegisterInfo &MRI, Pattern &&P)
BinaryOp_match< LHS, RHS, TargetOpcode::G_PTR_ADD, false > m_GPtrAdd(const LHS &L, const RHS &R)
Or< Preds... > m_any_of(Preds &&... preds)
BinaryOp_match< LHS, RHS, TargetOpcode::G_LSHR, false > m_GLShr(const LHS &L, const RHS &R)
UnaryOp_match< SrcTy, TargetOpcode::G_TRUNC > m_GTrunc(const SrcTy &Src)
DiagnosticInfoOptimizationBase::Argument NV
BaseReg
Stack frame base register. Bit 0 of FREInfo.Info.
This is an optimization pass for GlobalISel generic memory operations.
auto drop_begin(T &&RangeOrContainer, size_t N=1)
Return a range covering RangeOrContainer with the first N elements excluded.
iterator_range< early_inc_iterator_impl< detail::IterOfRange< RangeT > > > make_early_inc_range(RangeT &&Range)
Make a range that does early increment to allow mutation of the underlying range without disrupting i...
LLVM_ABI MachineInstr * getDefIgnoringCopies(Register Reg, const MachineRegisterInfo &MRI)
Find the def instruction for Reg, folding away any trivial copies.
bool any_of(R &&range, UnaryPredicate P)
Provide wrappers to std::any_of which take ranges instead of having to pass begin/end explicitly.
auto reverse(ContainerTy &&C)
LLVM_ABI raw_ostream & dbgs()
dbgs() - This returns a reference to a raw_ostream for debugging messages.
LLVM_ABI EVT getApproximateEVTForLLT(LLT Ty, LLVMContext &Ctx)
LLVM_ABI void getSelectionDAGFallbackAnalysisUsage(AnalysisUsage &AU)
Modify analysis usage so it preserves passes required for the SelectionDAG fallback.
LLVM_ABI std::optional< ValueAndVReg > getIConstantVRegValWithLookThrough(Register VReg, const MachineRegisterInfo &MRI, bool LookThroughInstrs=true)
If VReg is defined by a statically evaluable chain of instructions rooted on a G_CONSTANT returns its...
T bit_floor(T Value)
Returns the largest integral power of two no greater than Value if Value is nonzero.
LLVM_ABI LLT getLLTForType(Type &Ty, const DataLayout &DL)
Construct a low-level type based on an LLVM type.
LLVM_ABI bool isTriviallyDead(const MachineInstr &MI, const MachineRegisterInfo &MRI)
Check whether an instruction MI is dead: it only defines dead virtual registers, and doesn't have oth...
Implement std::hash so that hash_code can be used in STL containers.
The LegalityQuery object bundles together all the information that's needed to decide whether a given...
LegalizeAction Action
The action to take or the final answer.