LLVM 22.0.0git
InferAlignment.cpp
Go to the documentation of this file.
1//===- InferAlignment.cpp -------------------------------------------------===//
2//
3// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions.
4// See https://llvm.org/LICENSE.txt for license information.
5// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception
6//
7//===----------------------------------------------------------------------===//
8//
9// Infer alignment for load, stores and other memory operations based on
10// trailing zero known bits information.
11//
12//===----------------------------------------------------------------------===//
13
21
22using namespace llvm;
23
25 const DataLayout &DL, Instruction *I,
26 function_ref<Align(Value *PtrOp, Align OldAlign, Align PrefAlign)> Fn) {
27
28 if (auto *PtrOp = getLoadStorePointerOperand(I)) {
29 Align OldAlign = getLoadStoreAlignment(I);
30 Align PrefAlign = DL.getPrefTypeAlign(getLoadStoreType(I));
31
32 Align NewAlign = Fn(PtrOp, OldAlign, PrefAlign);
33 if (NewAlign > OldAlign) {
34 setLoadStoreAlignment(I, NewAlign);
35 return true;
36 }
37 }
38 // TODO: Also handle memory intrinsics.
39 return false;
40}
41
43 const DataLayout &DL = F.getDataLayout();
44 bool Changed = false;
45
46 // Enforce preferred type alignment if possible. We do this as a separate
47 // pass first, because it may improve the alignments we infer below.
48 for (BasicBlock &BB : F) {
49 for (Instruction &I : BB) {
50 Changed |= tryToImproveAlign(
51 DL, &I, [&](Value *PtrOp, Align OldAlign, Align PrefAlign) {
52 if (PrefAlign > OldAlign)
53 return std::max(OldAlign,
54 tryEnforceAlignment(PtrOp, PrefAlign, DL));
55 return OldAlign;
56 });
57 }
58 }
59
60 // Compute alignment from known bits.
61 auto InferFromKnownBits = [&](Instruction &I, Value *PtrOp) {
62 KnownBits Known = computeKnownBits(PtrOp, DL, &AC, &I, &DT);
63 unsigned TrailZ =
65 return Align(1ull << std::min(Known.getBitWidth() - 1, TrailZ));
66 };
67
68 // Propagate alignment between loads and stores that originate from the
69 // same base pointer.
70 DenseMap<Value *, Align> BestBasePointerAligns;
71 auto InferFromBasePointer = [&](Value *PtrOp, Align LoadStoreAlign) {
72 APInt OffsetFromBase(DL.getIndexTypeSizeInBits(PtrOp->getType()), 0);
73 PtrOp = PtrOp->stripAndAccumulateConstantOffsets(DL, OffsetFromBase, true);
74 // Derive the base pointer alignment from the load/store alignment
75 // and the offset from the base pointer.
76 Align BasePointerAlign =
77 commonAlignment(LoadStoreAlign, OffsetFromBase.getLimitedValue());
78
79 auto [It, Inserted] =
80 BestBasePointerAligns.try_emplace(PtrOp, BasePointerAlign);
81 if (!Inserted) {
82 // If the stored base pointer alignment is better than the
83 // base pointer alignment we derived, we may be able to use it
84 // to improve the load/store alignment. If not, store the
85 // improved base pointer alignment for future iterations.
86 if (It->second > BasePointerAlign) {
87 Align BetterLoadStoreAlign =
88 commonAlignment(It->second, OffsetFromBase.getLimitedValue());
89 return BetterLoadStoreAlign;
90 }
91 It->second = BasePointerAlign;
92 }
93 return LoadStoreAlign;
94 };
95
96 for (BasicBlock &BB : F) {
97 // We need to reset the map for each block because alignment information
98 // can only be propagated from instruction A to B if A dominates B.
99 // This is because control flow (and exception throwing) could be dependent
100 // on the address (and its alignment) at runtime. Some sort of dominator
101 // tree approach could be better, but doing a simple forward pass through a
102 // single basic block is correct too.
103 BestBasePointerAligns.clear();
104
105 for (Instruction &I : BB) {
106 Changed |= tryToImproveAlign(
107 DL, &I, [&](Value *PtrOp, Align OldAlign, Align PrefAlign) {
108 return std::max(InferFromKnownBits(I, PtrOp),
109 InferFromBasePointer(PtrOp, OldAlign));
110 });
111 }
112 }
113
114 return Changed;
115}
116
121 inferAlignment(F, AC, DT);
122 // Changes to alignment shouldn't invalidated analyses.
123 return PreservedAnalyses::all();
124}
MachineBasicBlock MachineBasicBlock::iterator DebugLoc DL
static bool tryToImproveAlign(const DataLayout &DL, Instruction *I, function_ref< Align(Value *PtrOp, Align OldAlign, Align PrefAlign)> Fn)
bool inferAlignment(Function &F, AssumptionCache &AC, DominatorTree &DT)
#define F(x, y, z)
Definition: MD5.cpp:55
#define I(x, y, z)
Definition: MD5.cpp:58
Class for arbitrary precision integers.
Definition: APInt.h:78
uint64_t getLimitedValue(uint64_t Limit=UINT64_MAX) const
If this value is smaller than the specified limit, return it, otherwise return the limit value.
Definition: APInt.h:475
A container for analyses that lazily runs them and caches their results.
Definition: PassManager.h:255
PassT::Result & getResult(IRUnitT &IR, ExtraArgTs... ExtraArgs)
Get the result of an analysis pass for a given IR unit.
Definition: PassManager.h:412
A function analysis which provides an AssumptionCache.
A cache of @llvm.assume calls within a function.
LLVM Basic Block Representation.
Definition: BasicBlock.h:62
A parsed version of the target data layout string in and methods for querying it.
Definition: DataLayout.h:63
std::pair< iterator, bool > try_emplace(KeyT &&Key, Ts &&...Args)
Definition: DenseMap.h:245
Analysis pass which computes a DominatorTree.
Definition: Dominators.h:284
Concrete subclass of DominatorTreeBase that is used to compute a normal dominator tree.
Definition: Dominators.h:165
A set of analyses that are preserved following a run of a transformation pass.
Definition: Analysis.h:112
static PreservedAnalyses all()
Construct a special preserved set that preserves all passes.
Definition: Analysis.h:118
LLVM Value Representation.
Definition: Value.h:75
Type * getType() const
All values are typed, get the type of this value.
Definition: Value.h:256
LLVM_ABI const Value * stripAndAccumulateConstantOffsets(const DataLayout &DL, APInt &Offset, bool AllowNonInbounds, bool AllowInvariantGroup=false, function_ref< bool(Value &Value, APInt &Offset)> ExternalAnalysis=nullptr, bool LookThroughIntToPtr=false) const
Accumulate the constant offset this value has compared to a base pointer.
static constexpr unsigned MaxAlignmentExponent
The maximum alignment for instructions.
Definition: Value.h:829
An efficient, type-erasing, non-owning reference to a callable.
This is an optimization pass for GlobalISel generic memory operations.
Definition: AddressRanges.h:18
const Value * getLoadStorePointerOperand(const Value *V)
A helper function that returns the pointer operand of a load or store instruction.
Align getLoadStoreAlignment(const Value *I)
A helper function that returns the alignment of load or store instruction.
LLVM_ABI void computeKnownBits(const Value *V, KnownBits &Known, const DataLayout &DL, AssumptionCache *AC=nullptr, const Instruction *CxtI=nullptr, const DominatorTree *DT=nullptr, bool UseInstrInfo=true, unsigned Depth=0)
Determine which bits of V are known to be either zero or one and return them in the KnownZero/KnownOn...
LLVM_ABI Align tryEnforceAlignment(Value *V, Align PrefAlign, const DataLayout &DL)
If the specified pointer points to an object that we control, try to modify the object's alignment to...
Definition: Local.cpp:1517
Align commonAlignment(Align A, uint64_t Offset)
Returns the alignment that satisfies both alignments.
Definition: Alignment.h:212
Type * getLoadStoreType(const Value *I)
A helper function that returns the type of a load or store instruction.
void setLoadStoreAlignment(Value *I, Align NewAlign)
A helper function that set the alignment of load or store instruction.
This struct is a compact representation of a valid (non-zero power of two) alignment.
Definition: Alignment.h:39
PreservedAnalyses run(Function &F, FunctionAnalysisManager &AM)
unsigned countMinTrailingZeros() const
Returns the minimum number of trailing zero bits.
Definition: KnownBits.h:235
unsigned getBitWidth() const
Get the bit width of this value.
Definition: KnownBits.h:44