85#ifndef LLVM_ANALYSIS_MEMORYSSA_H
86#define LLVM_ANALYSIS_MEMORYSSA_H
111template <
class GraphType>
struct GraphTraits;
116class MemorySSAWalker;
120namespace MSSAHelpers {
133template <
class T>
class memoryaccess_def_iterator_base;
142 public ilist_node<MemoryAccess, ilist_tag<MSSAHelpers::AllAccessTag>>,
143 public ilist_node<MemoryAccess, ilist_tag<MSSAHelpers::DefsOnlyTag>> {
153 void *
operator new(size_t) =
delete;
158 unsigned ID = V->getValueID();
159 return ID == MemoryUseVal ||
ID == MemoryPhiVal ||
ID == MemoryDefVal;
219 inline unsigned getID()
const;
252 void *
operator new(size_t) =
delete;
286 MemoryInstruction(
MI) {
320 void *
operator new(
size_t S) {
return User::operator
new(S, AllocMarker); }
321 void operator delete(
void *
Ptr) { User::operator
delete(
Ptr); }
330 OptimizedID = DMA->
getID();
385 void *
operator new(
size_t S) {
return User::operator
new(S, AllocMarker); }
386 void operator delete(
void *
Ptr) { User::operator
delete(
Ptr); }
394 OptimizedID = MA->
getID();
398 return cast_or_null<MemoryAccess>(getOperand(1));
402 return getOptimized() && OptimizedID == getOptimized()->getID();
407 setOperand(1,
nullptr);
428 if (
auto *MU = dyn_cast<MemoryUse>(MUD))
434 if (
auto *MU = dyn_cast<MemoryUse>(MUD))
440 if (
const auto *MU = dyn_cast<MemoryUse>(MUD))
483 void *
operator new(
size_t S) {
return User::operator
new(S, AllocMarker); }
486 void operator delete(
void *
Ptr) { User::operator
delete(
Ptr); }
493 ReservedSpace(NumPreds) {
494 allocHungoffUses(ReservedSpace);
503 return reinterpret_cast<block_iterator>(op_begin() + ReservedSpace);
513 return block_begin() + getNumOperands();
517 return make_range(block_begin(), block_end());
521 return make_range(block_begin(), block_end());
534 assert(V &&
"PHI node got a null value!");
547 assert(
this == U.getUser() &&
"Iterator doesn't point to PHI's Uses?");
548 return getIncomingBlock(
unsigned(&U - op_begin()));
554 return getIncomingBlock(
I.getUse());
558 assert(BB &&
"PHI node got a null basic block!");
559 block_begin()[
I] = BB;
564 if (getNumOperands() == ReservedSpace)
567 setNumHungOffUseOperands(getNumOperands() + 1);
568 setIncomingValue(getNumOperands() - 1, V);
569 setIncomingBlock(getNumOperands() - 1, BB);
575 for (
unsigned I = 0,
E = getNumOperands();
I !=
E; ++
I)
576 if (block_begin()[
I] == BB)
582 int Idx = getBasicBlockIndex(BB);
583 assert(
Idx >= 0 &&
"Invalid basic block argument!");
584 return getIncomingValue(
Idx);
589 unsigned E = getNumOperands();
590 assert(
I <
E &&
"Cannot remove out of bounds Phi entry.");
593 assert(
E >= 2 &&
"Cannot only remove incoming values in MemoryPhis with "
594 "at least 2 values.");
595 setIncomingValue(
I, getIncomingValue(
E - 1));
596 setIncomingBlock(
I, block_begin()[
E - 1]);
597 setOperand(
E - 1,
nullptr);
598 block_begin()[
E - 1] =
nullptr;
599 setNumHungOffUseOperands(getNumOperands() - 1);
605 for (
unsigned I = 0,
E = getNumOperands();
I !=
E; ++
I)
606 if (Pred(getIncomingValue(
I), getIncomingBlock(
I))) {
607 unorderedDeleteIncoming(
I);
608 E = getNumOperands();
611 assert(getNumOperands() >= 1 &&
612 "Cannot remove all incoming blocks in a MemoryPhi.");
617 unorderedDeleteIncomingIf(
624 unorderedDeleteIncomingIf(
629 return V->getValueID() == MemoryPhiVal;
649 unsigned ReservedSpace;
653 void growOperands() {
654 unsigned E = getNumOperands();
656 ReservedSpace = std::max(
E +
E / 2, 2u);
657 growHungoffUses(ReservedSpace,
true);
660 static void deleteMe(DerivedUser *Self);
664 assert((isa<MemoryDef>(
this) || isa<MemoryPhi>(
this)) &&
665 "only memory defs and phis have ids");
666 if (
const auto *MD = dyn_cast<MemoryDef>(
this))
668 return cast<MemoryPhi>(
this)->getID();
672 if (
const auto *MD = dyn_cast<MemoryDef>(
this))
673 return MD->isOptimized();
674 return cast<MemoryUse>(
this)->isOptimized();
678 if (
const auto *MD = dyn_cast<MemoryDef>(
this))
679 return MD->getOptimized();
680 return cast<MemoryUse>(
this)->getOptimized();
684 if (
auto *MD = dyn_cast<MemoryDef>(
this))
685 MD->setOptimized(MA);
687 cast<MemoryUse>(
this)->setOptimized(MA);
691 if (
auto *MD = dyn_cast<MemoryDef>(
this))
692 MD->resetOptimized();
694 cast<MemoryUse>(
this)->resetOptimized();
721 return cast_or_null<MemoryUseOrDef>(ValueToMemoryAccess.lookup(
I));
725 return cast_or_null<MemoryPhi>(ValueToMemoryAccess.lookup(cast<Value>(BB)));
741 return MA == LiveOnEntryDef.get();
745 return LiveOnEntryDef.get();
761 return getWritableBlockAccesses(BB);
769 return getWritableBlockDefs(BB);
789 verifyMemorySSA(VerificationLevel = VerificationLevel::Fast)
const;
800 LLVM_ABI void ensureOptimizedUses();
808 template <
typename IterT>
809 void verifyOrderingDominationAndDefUses(
811 template <
typename IterT>
void verifyDominationNumbers(IterT
Blocks)
const;
812 template <
typename IterT>
void verifyPrevDefInPhis(IterT
Blocks)
const;
816 auto It = PerBlockAccesses.find(BB);
817 return It == PerBlockAccesses.end() ? nullptr : It->second.get();
822 auto It = PerBlockDefs.find(BB);
823 return It == PerBlockDefs.
end() ? nullptr : It->second.get();
831 AccessList::iterator Where);
833 InsertionPlace Point);
838 renamePass(DT->
getNode(BB), IncomingVal, Visited,
true,
true);
846 AccessList::iterator);
850 bool CreationMustSucceed =
true);
853 class ClobberWalkerBase;
855 class SkipSelfWalker;
858 CachingWalker *getWalkerImpl();
859 template <
typename IterT>
868 void markUnreachableAsLiveOnEntry(
BasicBlock *BB);
870 template <
typename AliasAnalysisType>
878 bool SkipVisited =
false,
879 bool RenameAllUses =
false);
880 AccessList *getOrCreateAccessList(
const BasicBlock *);
881 DefsList *getOrCreateDefsList(
const BasicBlock *);
897 AccessMap PerBlockAccesses;
898 DefsMap PerBlockDefs;
899 std::unique_ptr<MemoryAccess, ValueDeleter> LiveOnEntryDef;
908 std::unique_ptr<ClobberWalkerBase> WalkerBase;
909 std::unique_ptr<CachingWalker> Walker;
910 std::unique_ptr<SkipSelfWalker> SkipWalker;
912 bool IsOptimized =
false;
950 std::unique_ptr<MemorySSA>
MSSA;
962 bool EnsureOptimizedUses;
966 :
OS(
OS), EnsureOptimizedUses(EnsureOptimizedUses) {}
1000 void releaseMemory()
override;
1006 void verifyAnalysis()
const override;
1010 std::unique_ptr<MemorySSA> MSSA;
1056 assert(MA &&
"Handed an instruction that MemorySSA doesn't recognize?");
1115 using MemorySSAWalker::getClobberingMemoryAccess;
1132 std::forward_iterator_tag, T, ptrdiff_t, T *,
1134 using BaseT =
typename memoryaccess_def_iterator_base::iterator_facade_base;
1141 return Access ==
Other.Access && (!Access || ArgNo ==
Other.ArgNo);
1151 MemoryPhi *MP = dyn_cast<MemoryPhi>(Access);
1152 assert(MP &&
"Tried to get phi arg block when not iterating over a PHI");
1156 typename std::iterator_traits<BaseT>::pointer
operator*()
const {
1157 assert(Access &&
"Tried to access past the end of our iterator");
1160 if (
const MemoryPhi *MP = dyn_cast<MemoryPhi>(Access))
1161 return MP->getIncomingValue(ArgNo);
1162 return cast<MemoryUseOrDef>(Access)->getDefiningAccess();
1165 using BaseT::operator++;
1167 assert(Access &&
"Hit end of iterator");
1168 if (
const MemoryPhi *MP = dyn_cast<MemoryPhi>(Access)) {
1169 if (++ArgNo >= MP->getNumIncomingValues()) {
1180 T *Access =
nullptr;
1230 std::forward_iterator_tag,
1231 const MemoryAccessPair> {
1232 using BaseT = upward_defs_iterator::iterator_facade_base;
1236 : DefIterator(
Info.first), Location(
Info.second),
1237 OriginalAccess(
Info.first), DT(DT) {
1238 CurrentPair.first =
nullptr;
1240 WalkingPhi =
Info.first && isa<MemoryPhi>(
Info.first);
1241 fillInCurrentPair();
1247 return DefIterator ==
Other.DefIterator;
1250 typename std::iterator_traits<BaseT>::reference
operator*()
const {
1252 "Tried to access past the end of our iterator");
1256 using BaseT::operator++;
1259 "Tried to access past the end of the iterator");
1261 if (DefIterator != OriginalAccess->
defs_end())
1262 fillInCurrentPair();
1274 void fillInCurrentPair() {
1275 CurrentPair.first = *DefIterator;
1276 CurrentPair.second = Location;
1277 if (WalkingPhi && Location.
Ptr) {
1283 Translator.translateValue(OriginalAccess->
getBlock(),
1285 if (
Addr != CurrentPair.second.Ptr)
1286 CurrentPair.second = CurrentPair.second.getWithNewPtr(
Addr);
1293 if (!IsGuaranteedLoopInvariant(CurrentPair.second.Ptr))
1294 CurrentPair.second = CurrentPair.second.getWithNewSize(
1301 MemoryLocation Location;
1302 MemoryAccess *OriginalAccess =
nullptr;
1303 DominatorTree *DT =
nullptr;
1304 bool WalkingPhi =
false;
1307inline upward_defs_iterator
1314inline iterator_range<upward_defs_iterator>
1332template <
class T,
bool UseOptimizedChain = false>
1335 std::forward_iterator_tag, MemoryAccess *> {
1343 if (
auto *MUD = dyn_cast<MemoryUseOrDef>(MA)) {
1344 if (UseOptimizedChain && MUD->isOptimized())
1345 MA = MUD->getOptimized();
1347 MA = MUD->getDefiningAccess();
1362inline iterator_range<def_chain_iterator<T>>
1364#ifdef EXPENSIVE_CHECKS
1366 "UpTo isn't in the def chain!");
assert(UImm &&(UImm !=~static_cast< T >(0)) &&"Invalid immediate!")
static void print(raw_ostream &Out, object::Archive::Kind Kind, T Val)
static GCRegistry::Add< OcamlGC > B("ocaml", "ocaml 3.10-compatible GC")
static GCRegistry::Add< ErlangGC > A("erlang", "erlang-compatible garbage collector")
static GCRegistry::Add< CoreCLRGC > E("coreclr", "CoreCLR-compatible GC")
Analysis containing CSE Info
Returns the sub type a function will return at a given Idx Should correspond to the result type of an ExtractValue instruction executed with just that one unsigned Idx
This file defines the DenseMap class.
DenseMap< Block *, BlockRelaxAux > Blocks
static bool runOnFunction(Function &F, bool PostInlining)
static MemoryAccess * getClobberingMemoryAccess(MemorySSA &MSSA, BatchAAResults &BAA, SinkAndHoistLICMFlags &Flags, MemoryUseOrDef *MA)
Machine Check Debug Module
This file provides utility analysis objects describing memory locations.
#define DEFINE_TRANSPARENT_OPERAND_ACCESSORS(CLASS, VALUECLASS)
Macro for generating out-of-class operand accessor definitions.
static bool dominates(InstrPosIndexes &PosIndexes, const MachineInstr &A, const MachineInstr &B)
This file defines the SmallPtrSet class.
This file defines the SmallVector class.
A private abstract base class describing the concept of an individual alias analysis implementation.
API to communicate dependencies between analyses during invalidation.
A container for analyses that lazily runs them and caches their results.
Represent the analysis usage information of a pass.
LLVM Basic Block Representation.
LLVM_ABI const DataLayout & getDataLayout() const
Get the data layout of the module this basic block belongs to.
This class is a wrapper over an AAResults, and it is intended to be used only when there are no IR ch...
Extension point for the Value hierarchy.
void(*)(DerivedUser *) DeleteValueTy
A MemorySSAWalker that does no alias queries, or anything else.
DomTreeNodeBase< NodeT > * getNode(const NodeT *BB) const
getNode - return the (Post)DominatorTree node for the specified basic block.
Concrete subclass of DominatorTreeBase that is used to compute a normal dominator tree.
FunctionPass class - This class is used to implement most global optimizations.
This is an important class for using LLVM in a threaded context.
static constexpr LocationSize beforeOrAfterPointer()
Any location before or after the base pointer (but still within the underlying object).
Represents a single loop in the control flow graph.
LLVM_ABI void dump() const
AllAccessType::reverse_self_iterator getReverseIterator()
AllAccessType::const_self_iterator getIterator() const
MemoryAccess(const MemoryAccess &)=delete
static bool classof(const Value *V)
DefsOnlyType::const_self_iterator getDefsIterator() const
DefsOnlyType::self_iterator getDefsIterator()
DefsOnlyType::reverse_self_iterator getReverseDefsIterator()
DefsOnlyType::const_reverse_self_iterator getReverseDefsIterator() const
memoryaccess_def_iterator defs_end()
MemoryAccess(LLVMContext &C, unsigned Vty, DeleteValueTy DeleteValue, BasicBlock *BB, AllocInfo AllocInfo)
BasicBlock * getBlock() const
user_iterator iterator
The user iterators for a memory access.
AllAccessType::const_reverse_self_iterator getReverseIterator() const
LLVM_ABI void print(raw_ostream &OS) const
unsigned getID() const
Used for debugging and tracking things about MemoryAccesses.
MemoryAccess & operator=(const MemoryAccess &)=delete
void setBlock(BasicBlock *BB)
Used by MemorySSA to change the block of a MemoryAccess when it is moved.
const_user_iterator const_iterator
memoryaccess_def_iterator defs_begin()
This iterator walks over all of the defs in a given MemoryAccess.
AllAccessType::self_iterator getIterator()
Get the iterators for the all access list and the defs only list We default to the all access list.
Represents a read-write access to memory, whether it is a must-alias, or a may-alias.
static bool classof(const Value *MA)
MemoryAccess * getOptimized() const
DECLARE_TRANSPARENT_OPERAND_ACCESSORS(MemoryAccess)
MemoryDef(LLVMContext &C, MemoryAccess *DMA, Instruction *MI, BasicBlock *BB, unsigned Ver)
void setOptimized(MemoryAccess *MA)
Representation for a specific memory location.
const Value * Ptr
The address of the start of the location.
Represents phi nodes for memory accesses.
void setIncomingBlock(unsigned I, BasicBlock *BB)
void allocHungoffUses(unsigned N)
this is more complicated than the generic User::allocHungoffUses, because we have to allocate Uses fo...
void setIncomingValue(unsigned I, MemoryAccess *V)
static bool classof(const Value *V)
void unorderedDeleteIncomingValue(const MemoryAccess *MA)
const_block_iterator block_end() const
BasicBlock * getIncomingBlock(const Use &U) const
Return incoming basic block corresponding to an operand of the PHI.
DECLARE_TRANSPARENT_OPERAND_ACCESSORS(MemoryAccess)
Provide fast operand accessors.
unsigned getNumIncomingValues() const
Return the number of incoming edges.
MemoryAccess * getIncomingValueForBlock(const BasicBlock *BB) const
block_iterator block_end()
const_block_iterator block_begin() const
iterator_range< block_iterator > blocks()
void unorderedDeleteIncomingIf(Fn &&Pred)
void unorderedDeleteIncoming(unsigned I)
BasicBlock * getIncomingBlock(unsigned I) const
Return incoming basic block number i.
const_op_range incoming_values() const
BasicBlock * getIncomingBlock(MemoryAccess::const_user_iterator I) const
Return incoming basic block corresponding to value use iterator.
static unsigned getIncomingValueNumForOperand(unsigned I)
void addIncoming(MemoryAccess *V, BasicBlock *BB)
Add an incoming value to the end of the PHI list.
op_range incoming_values()
void unorderedDeleteIncomingBlock(const BasicBlock *BB)
MemoryPhi(LLVMContext &C, BasicBlock *BB, unsigned Ver, unsigned NumPreds=0)
MemoryAccess * getIncomingValue(unsigned I) const
Return incoming value number x.
static unsigned getOperandNumForIncomingValue(unsigned I)
int getBasicBlockIndex(const BasicBlock *BB) const
Return the first index of the specified basic block in the value list for this PHI.
iterator_range< const_block_iterator > blocks() const
BasicBlock *const * const_block_iterator
block_iterator block_begin()
An analysis that produces MemorySSA for a function.
LLVM_ABI Result run(Function &F, FunctionAnalysisManager &AM)
Printer pass for MemorySSA.
LLVM_ABI PreservedAnalyses run(Function &F, FunctionAnalysisManager &AM)
MemorySSAPrinterPass(raw_ostream &OS, bool EnsureOptimizedUses)
static LLVM_ABI bool defClobbersUseOrDef(MemoryDef *MD, const MemoryUseOrDef *MU, AliasAnalysis &AA)
Printer pass for MemorySSA via the walker.
MemorySSAWalkerPrinterPass(raw_ostream &OS)
LLVM_ABI PreservedAnalyses run(Function &F, FunctionAnalysisManager &AM)
This is the generic walker interface for walkers of MemorySSA.
MemoryAccess * getClobberingMemoryAccess(const Instruction *I, BatchAAResults &AA)
Given a memory Mod/Ref/ModRef'ing instruction, calling this will give you the nearest dominating Memo...
virtual ~MemorySSAWalker()=default
MemoryAccess * getClobberingMemoryAccess(MemoryAccess *MA, const MemoryLocation &Loc)
virtual void invalidateInfo(MemoryAccess *)
Given a memory access, invalidate anything this walker knows about that access.
virtual MemoryAccess * getClobberingMemoryAccess(MemoryAccess *, const MemoryLocation &, BatchAAResults &AA)=0
Given a potentially clobbering memory access and a new location, calling this will give you the neare...
virtual MemoryAccess * getClobberingMemoryAccess(MemoryAccess *, BatchAAResults &AA)=0
Does the same thing as getClobberingMemoryAccess(const Instruction *I), but takes a MemoryAccess inst...
MemoryAccess * getClobberingMemoryAccess(const Instruction *I)
MemoryAccess * getClobberingMemoryAccess(MemoryAccess *MA)
Legacy analysis pass which computes MemorySSA.
const MemorySSA & getMSSA() const
Encapsulates MemorySSA, including all data associated with memory accesses.
const AccessList * getBlockAccesses(const BasicBlock *BB) const
Return the list of MemoryAccess's for a given basic block.
void renamePass(BasicBlock *BB, MemoryAccess *IncomingVal, SmallPtrSetImpl< BasicBlock * > &Visited)
AccessList * getWritableBlockAccesses(const BasicBlock *BB) const
InsertionPlace
Used in various insertion functions to specify whether we are talking about the beginning or end of a...
DefsList * getWritableBlockDefs(const BasicBlock *BB) const
MemorySSA(MemorySSA &&)=delete
DominatorTree & getDomTree() const
MemoryUseOrDef * getMemoryAccess(const Instruction *I) const
Given a memory Mod/Ref'ing instruction, get the MemorySSA access associated with it.
MemoryPhi * getMemoryAccess(const BasicBlock *BB) const
MemoryAccess * getLiveOnEntryDef() const
const DefsList * getBlockDefs(const BasicBlock *BB) const
Return the list of MemoryDef's and MemoryPhi's for a given basic block.
bool isLiveOnEntryDef(const MemoryAccess *MA) const
Return true if MA represents the live on entry value.
Class that has the common methods + fields of memory uses/defs.
MemoryUseOrDef(LLVMContext &C, MemoryAccess *DMA, unsigned Vty, DeleteValueTy DeleteValue, Instruction *MI, BasicBlock *BB, AllocInfo AllocInfo)
~MemoryUseOrDef()=default
MemoryAccess * getDefiningAccess() const
Get the access that produces the memory state used by this Use.
void resetOptimized()
Reset the ID of what this MemoryUse was optimized to, causing it to be rewalked by the walker if nece...
MemoryAccess * getOptimized() const
Return the MemoryAccess associated with the optimized use, or nullptr.
void setDefiningAccess(MemoryAccess *DMA, bool Optimized=false)
void setOptimized(MemoryAccess *)
Sets the optimized use for a MemoryDef.
DECLARE_TRANSPARENT_OPERAND_ACCESSORS(MemoryAccess)
Instruction * getMemoryInst() const
Get the instruction that this MemoryUse represents.
static bool classof(const Value *MA)
bool isOptimized() const
Do we have an optimized use?
Represents read-only accesses to memory.
MemoryAccess * getOptimized() const
DECLARE_TRANSPARENT_OPERAND_ACCESSORS(MemoryAccess)
MemoryUse(LLVMContext &C, MemoryAccess *DMA, Instruction *MI, BasicBlock *BB)
LLVM_ABI void print(raw_ostream &OS) const
bool isOptimized() const
Whether the MemoryUse is optimized.
static bool classof(const Value *MA)
void setOptimized(MemoryAccess *DMA)
A Module instance is used to store all the information related to an LLVM module.
PHITransAddr - An address value which tracks and handles phi translation.
A set of analyses that are preserved following a run of a transformation pass.
A templated base class for SmallPtrSet which provides the typesafe interface that is common across al...
SmallPtrSet - This class implements a set which is optimized for holding SmallSize or less elements.
This is a 'vector' (really, a variable-sized array), optimized for the case when the array is small.
The instances of the Type class are immutable: once they are created, they are never changed.
A Use represents the edge between a Value definition and its users.
LLVM_ABI void allocHungoffUses(unsigned N, bool IsPhi=false)
Allocate the array of Uses, followed by a pointer (with bottom bit set) to the User.
void setOperand(unsigned i, Value *Val)
Value * getOperand(unsigned i) const
LLVM Value Representation.
user_iterator_impl< const User > const_user_iterator
unsigned getValueID() const
Return an ID for the concrete type of this object.
LLVM_ABI void deleteValue()
Delete a pointer to a generic Value.
user_iterator_impl< User > user_iterator
typename ilist_select_iterator_type< OptionsT::has_iterator_bits, OptionsT, true, false >::type reverse_self_iterator
typename ilist_select_iterator_type< OptionsT::has_iterator_bits, OptionsT, false, true >::type const_self_iterator
typename ilist_select_iterator_type< OptionsT::has_iterator_bits, OptionsT, false, false >::type self_iterator
typename ilist_select_iterator_type< OptionsT::has_iterator_bits, OptionsT, true, true >::type const_reverse_self_iterator
reverse_self_iterator getReverseIterator()
self_iterator getIterator()
An intrusive list with ownership and callbacks specified/controlled by ilist_traits,...
CRTP base class which implements the entire standard iterator facade in terms of a minimal subset of ...
A range adaptor for a pair of iterators.
memoryaccess_def_iterator_base()=default
BasicBlock * getPhiArgBlock() const
std::iterator_traits< BaseT >::pointer operator*() const
bool operator==(const memoryaccess_def_iterator_base &Other) const
memoryaccess_def_iterator_base(T *Start)
memoryaccess_def_iterator_base & operator++()
This class implements an extremely fast bulk output stream that can only output to a stream.
A simple intrusive list implementation.
Provide an iterator that walks defs, giving both the memory access, and the current pointer location,...
upward_defs_iterator(const MemoryAccessPair &Info, DominatorTree *DT)
std::iterator_traits< BaseT >::reference operator*() const
BasicBlock * getPhiArgBlock() const
upward_defs_iterator & operator++()
bool operator==(const upward_defs_iterator &Other) const
This file defines the ilist_node class template, which is a convenient base class for creating classe...
This provides a very simple, boring adaptor for a begin and end iterator into a range type.
unsigned ID
LLVM IR allows to use arbitrary numbers as calling convention identifiers.
@ Fast
Attempts to make calls as fast as possible (e.g.
@ C
The default llvm calling convention, compatible with C.
This is an optimization pass for GlobalISel generic memory operations.
@ INVALID_MEMORYACCESS_ID
void dump(const SparseBitVector< ElementSize > &LHS, raw_ostream &out)
auto find(R &&Range, const T &Val)
Provide wrappers to std::find which take ranges instead of having to pass begin/end explicitly.
Printable print(const GCNRegPressure &RP, const GCNSubtarget *ST=nullptr, unsigned DynamicVGPRBlockSize=0)
iterator_range< T > make_range(T x, T y)
Convenience function for iterating over sub-ranges.
upward_defs_iterator upward_defs_begin(const MemoryAccessPair &Pair, DominatorTree &DT)
std::pair< const MemoryAccess *, MemoryLocation > ConstMemoryAccessPair
iterator_range< def_chain_iterator< T > > def_chain(T MA, MemoryAccess *UpTo=nullptr)
memoryaccess_def_iterator_base< MemoryAccess > memoryaccess_def_iterator
memoryaccess_def_iterator_base< const MemoryAccess > const_memoryaccess_def_iterator
LLVM_ABI bool VerifyMemorySSA
Enables verification of MemorySSA.
std::pair< MemoryAccess *, MemoryLocation > MemoryAccessPair
upward_defs_iterator upward_defs_end()
raw_ostream & operator<<(raw_ostream &OS, const APFixedPoint &FX)
OutputIt move(R &&Range, OutputIt Out)
Provide wrappers to std::move which take ranges instead of having to pass begin/end explicitly.
iterator_range< upward_defs_iterator > upward_defs(const MemoryAccessPair &Pair, DominatorTree &DT)
iterator_range< def_chain_iterator< T, true > > optimized_def_chain(T MA)
Implement std::hash so that hash_code can be used in STL containers.
A CRTP mix-in that provides informational APIs needed for analysis passes.
A special type used by analysis passes to provide an address that identifies that particular analysis...
FixedNumOperandTraits - determine the allocation regime of the Use array when it is a prefix to the U...
static ChildIteratorType child_begin(NodeRef N)
MemoryAccess::iterator ChildIteratorType
static ChildIteratorType child_end(NodeRef N)
static NodeRef getEntryNode(NodeRef N)
static ChildIteratorType child_begin(NodeRef N)
static ChildIteratorType child_end(NodeRef N)
static NodeRef getEntryNode(NodeRef N)
HungoffOperandTraits - determine the allocation regime of the Use array when it is not a prefix to th...
Result(std::unique_ptr< MemorySSA > &&MSSA)
LLVM_ABI bool invalidate(Function &F, const PreservedAnalyses &PA, FunctionAnalysisManager::Invalidator &Inv)
std::unique_ptr< MemorySSA > MSSA
Verifier pass for MemorySSA.
LLVM_ABI PreservedAnalyses run(Function &F, FunctionAnalysisManager &AM)
static unsigned operands(const MemoryUseOrDef *MUD)
static Use * op_end(MemoryUseOrDef *MUD)
static Use * op_begin(MemoryUseOrDef *MUD)
Compile-time customization of User operands.
A CRTP mix-in to automatically provide informational APIs needed for passes.
Information about how a User object was allocated, to be passed into the User constructor.
Indicates this User has operands "hung off" in another allocation.
Indicates this User has operands co-allocated.
Walks the defining accesses of MemoryDefs.
bool operator==(const def_chain_iterator &O) const
def_chain_iterator & operator++()
static void deleteNode(MemoryAccess *MA)
Use delete by default for iplist and ilist.