LLVM 22.0.0git
InstCombineAndOrXor.cpp
Go to the documentation of this file.
1//===- InstCombineAndOrXor.cpp --------------------------------------------===//
2//
3// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions.
4// See https://llvm.org/LICENSE.txt for license information.
5// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception
6//
7//===----------------------------------------------------------------------===//
8//
9// This file implements the visitAnd, visitOr, and visitXor functions.
10//
11//===----------------------------------------------------------------------===//
12
13#include "InstCombineInternal.h"
21#include "llvm/IR/Intrinsics.h"
25
26using namespace llvm;
27using namespace PatternMatch;
28
29#define DEBUG_TYPE "instcombine"
30
31/// This is the complement of getICmpCode, which turns an opcode and two
32/// operands into either a constant true or false, or a brand new ICmp
33/// instruction. The sign is passed in to determine which kind of predicate to
34/// use in the new icmp instruction.
35static Value *getNewICmpValue(unsigned Code, bool Sign, Value *LHS, Value *RHS,
36 InstCombiner::BuilderTy &Builder) {
37 ICmpInst::Predicate NewPred;
38 if (Constant *TorF = getPredForICmpCode(Code, Sign, LHS->getType(), NewPred))
39 return TorF;
40 return Builder.CreateICmp(NewPred, LHS, RHS);
41}
42
43/// This is the complement of getFCmpCode, which turns an opcode and two
44/// operands into either a FCmp instruction, or a true/false constant.
45static Value *getFCmpValue(unsigned Code, Value *LHS, Value *RHS,
46 InstCombiner::BuilderTy &Builder, FMFSource FMF) {
47 FCmpInst::Predicate NewPred;
48 if (Constant *TorF = getPredForFCmpCode(Code, LHS->getType(), NewPred))
49 return TorF;
50 return Builder.CreateFCmpFMF(NewPred, LHS, RHS, FMF);
51}
52
53/// Emit a computation of: (V >= Lo && V < Hi) if Inside is true, otherwise
54/// (V < Lo || V >= Hi). This method expects that Lo < Hi. IsSigned indicates
55/// whether to treat V, Lo, and Hi as signed or not.
57 const APInt &Hi, bool isSigned,
58 bool Inside) {
59 assert((isSigned ? Lo.slt(Hi) : Lo.ult(Hi)) &&
60 "Lo is not < Hi in range emission code!");
61
62 Type *Ty = V->getType();
63
64 // V >= Min && V < Hi --> V < Hi
65 // V < Min || V >= Hi --> V >= Hi
67 if (isSigned ? Lo.isMinSignedValue() : Lo.isMinValue()) {
68 Pred = isSigned ? ICmpInst::getSignedPredicate(Pred) : Pred;
69 return Builder.CreateICmp(Pred, V, ConstantInt::get(Ty, Hi));
70 }
71
72 // V >= Lo && V < Hi --> V - Lo u< Hi - Lo
73 // V < Lo || V >= Hi --> V - Lo u>= Hi - Lo
74 Value *VMinusLo =
75 Builder.CreateSub(V, ConstantInt::get(Ty, Lo), V->getName() + ".off");
76 Constant *HiMinusLo = ConstantInt::get(Ty, Hi - Lo);
77 return Builder.CreateICmp(Pred, VMinusLo, HiMinusLo);
78}
79
80/// Classify (icmp eq (A & B), C) and (icmp ne (A & B), C) as matching patterns
81/// that can be simplified.
82/// One of A and B is considered the mask. The other is the value. This is
83/// described as the "AMask" or "BMask" part of the enum. If the enum contains
84/// only "Mask", then both A and B can be considered masks. If A is the mask,
85/// then it was proven that (A & C) == C. This is trivial if C == A or C == 0.
86/// If both A and C are constants, this proof is also easy.
87/// For the following explanations, we assume that A is the mask.
88///
89/// "AllOnes" declares that the comparison is true only if (A & B) == A or all
90/// bits of A are set in B.
91/// Example: (icmp eq (A & 3), 3) -> AMask_AllOnes
92///
93/// "AllZeros" declares that the comparison is true only if (A & B) == 0 or all
94/// bits of A are cleared in B.
95/// Example: (icmp eq (A & 3), 0) -> Mask_AllZeroes
96///
97/// "Mixed" declares that (A & B) == C and C might or might not contain any
98/// number of one bits and zero bits.
99/// Example: (icmp eq (A & 3), 1) -> AMask_Mixed
100///
101/// "Not" means that in above descriptions "==" should be replaced by "!=".
102/// Example: (icmp ne (A & 3), 3) -> AMask_NotAllOnes
103///
104/// If the mask A contains a single bit, then the following is equivalent:
105/// (icmp eq (A & B), A) equals (icmp ne (A & B), 0)
106/// (icmp ne (A & B), A) equals (icmp eq (A & B), 0)
119
120/// Return the set of patterns (from MaskedICmpType) that (icmp SCC (A & B), C)
121/// satisfies.
122static unsigned getMaskedICmpType(Value *A, Value *B, Value *C,
123 ICmpInst::Predicate Pred) {
124 const APInt *ConstA = nullptr, *ConstB = nullptr, *ConstC = nullptr;
125 match(A, m_APInt(ConstA));
126 match(B, m_APInt(ConstB));
127 match(C, m_APInt(ConstC));
128 bool IsEq = (Pred == ICmpInst::ICMP_EQ);
129 bool IsAPow2 = ConstA && ConstA->isPowerOf2();
130 bool IsBPow2 = ConstB && ConstB->isPowerOf2();
131 unsigned MaskVal = 0;
132 if (ConstC && ConstC->isZero()) {
133 // if C is zero, then both A and B qualify as mask
134 MaskVal |= (IsEq ? (Mask_AllZeros | AMask_Mixed | BMask_Mixed)
136 if (IsAPow2)
137 MaskVal |= (IsEq ? (AMask_NotAllOnes | AMask_NotMixed)
139 if (IsBPow2)
140 MaskVal |= (IsEq ? (BMask_NotAllOnes | BMask_NotMixed)
142 return MaskVal;
143 }
144
145 if (A == C) {
146 MaskVal |= (IsEq ? (AMask_AllOnes | AMask_Mixed)
148 if (IsAPow2)
149 MaskVal |= (IsEq ? (Mask_NotAllZeros | AMask_NotMixed)
151 } else if (ConstA && ConstC && ConstC->isSubsetOf(*ConstA)) {
152 MaskVal |= (IsEq ? AMask_Mixed : AMask_NotMixed);
153 }
154
155 if (B == C) {
156 MaskVal |= (IsEq ? (BMask_AllOnes | BMask_Mixed)
158 if (IsBPow2)
159 MaskVal |= (IsEq ? (Mask_NotAllZeros | BMask_NotMixed)
161 } else if (ConstB && ConstC && ConstC->isSubsetOf(*ConstB)) {
162 MaskVal |= (IsEq ? BMask_Mixed : BMask_NotMixed);
163 }
164
165 return MaskVal;
166}
167
168/// Convert an analysis of a masked ICmp into its equivalent if all boolean
169/// operations had the opposite sense. Since each "NotXXX" flag (recording !=)
170/// is adjacent to the corresponding normal flag (recording ==), this just
171/// involves swapping those bits over.
172static unsigned conjugateICmpMask(unsigned Mask) {
173 unsigned NewMask;
174 NewMask = (Mask & (AMask_AllOnes | BMask_AllOnes | Mask_AllZeros |
176 << 1;
177
178 NewMask |= (Mask & (AMask_NotAllOnes | BMask_NotAllOnes | Mask_NotAllZeros |
180 >> 1;
181
182 return NewMask;
183}
184
185// Adapts the external decomposeBitTestICmp for local use.
187 Value *&X, Value *&Y, Value *&Z) {
188 auto Res = llvm::decomposeBitTest(Cond, /*LookThroughTrunc=*/true,
189 /*AllowNonZeroC=*/true);
190 if (!Res)
191 return false;
192
193 Pred = Res->Pred;
194 X = Res->X;
195 Y = ConstantInt::get(X->getType(), Res->Mask);
196 Z = ConstantInt::get(X->getType(), Res->C);
197 return true;
198}
199
200/// Handle (icmp(A & B) ==/!= C) &/| (icmp(A & D) ==/!= E).
201/// Return the pattern classes (from MaskedICmpType) for the left hand side and
202/// the right hand side as a pair.
203/// LHS and RHS are the left hand side and the right hand side ICmps and PredL
204/// and PredR are their predicates, respectively.
205static std::optional<std::pair<unsigned, unsigned>>
208 ICmpInst::Predicate &PredR) {
209
210 // Here comes the tricky part:
211 // LHS might be of the form L11 & L12 == X, X == L21 & L22,
212 // and L11 & L12 == L21 & L22. The same goes for RHS.
213 // Now we must find those components L** and R**, that are equal, so
214 // that we can extract the parameters A, B, C, D, and E for the canonical
215 // above.
216
217 // Check whether the icmp can be decomposed into a bit test.
218 Value *L1, *L11, *L12, *L2, *L21, *L22;
219 if (decomposeBitTestICmp(LHS, PredL, L11, L12, L2)) {
220 L21 = L22 = L1 = nullptr;
221 } else {
222 auto *LHSCMP = dyn_cast<ICmpInst>(LHS);
223 if (!LHSCMP)
224 return std::nullopt;
225
226 // Don't allow pointers. Splat vectors are fine.
227 if (!LHSCMP->getOperand(0)->getType()->isIntOrIntVectorTy())
228 return std::nullopt;
229
230 PredL = LHSCMP->getPredicate();
231 L1 = LHSCMP->getOperand(0);
232 L2 = LHSCMP->getOperand(1);
233 // Look for ANDs in the LHS icmp.
234 if (!match(L1, m_And(m_Value(L11), m_Value(L12)))) {
235 // Any icmp can be viewed as being trivially masked; if it allows us to
236 // remove one, it's worth it.
237 L11 = L1;
239 }
240
241 if (!match(L2, m_And(m_Value(L21), m_Value(L22)))) {
242 L21 = L2;
244 }
245 }
246
247 // Bail if LHS was a icmp that can't be decomposed into an equality.
248 if (!ICmpInst::isEquality(PredL))
249 return std::nullopt;
250
251 Value *R11, *R12, *R2;
252 if (decomposeBitTestICmp(RHS, PredR, R11, R12, R2)) {
253 if (R11 == L11 || R11 == L12 || R11 == L21 || R11 == L22) {
254 A = R11;
255 D = R12;
256 } else if (R12 == L11 || R12 == L12 || R12 == L21 || R12 == L22) {
257 A = R12;
258 D = R11;
259 } else {
260 return std::nullopt;
261 }
262 E = R2;
263 } else {
264 auto *RHSCMP = dyn_cast<ICmpInst>(RHS);
265 if (!RHSCMP)
266 return std::nullopt;
267 // Don't allow pointers. Splat vectors are fine.
268 if (!RHSCMP->getOperand(0)->getType()->isIntOrIntVectorTy())
269 return std::nullopt;
270
271 PredR = RHSCMP->getPredicate();
272
273 Value *R1 = RHSCMP->getOperand(0);
274 R2 = RHSCMP->getOperand(1);
275 bool Ok = false;
276 if (!match(R1, m_And(m_Value(R11), m_Value(R12)))) {
277 // As before, model no mask as a trivial mask if it'll let us do an
278 // optimization.
279 R11 = R1;
281 }
282
283 if (R11 == L11 || R11 == L12 || R11 == L21 || R11 == L22) {
284 A = R11;
285 D = R12;
286 E = R2;
287 Ok = true;
288 } else if (R12 == L11 || R12 == L12 || R12 == L21 || R12 == L22) {
289 A = R12;
290 D = R11;
291 E = R2;
292 Ok = true;
293 }
294
295 // Avoid matching against the -1 value we created for unmasked operand.
296 if (Ok && match(A, m_AllOnes()))
297 Ok = false;
298
299 // Look for ANDs on the right side of the RHS icmp.
300 if (!Ok) {
301 if (!match(R2, m_And(m_Value(R11), m_Value(R12)))) {
302 R11 = R2;
303 R12 = Constant::getAllOnesValue(R2->getType());
304 }
305
306 if (R11 == L11 || R11 == L12 || R11 == L21 || R11 == L22) {
307 A = R11;
308 D = R12;
309 E = R1;
310 } else if (R12 == L11 || R12 == L12 || R12 == L21 || R12 == L22) {
311 A = R12;
312 D = R11;
313 E = R1;
314 } else {
315 return std::nullopt;
316 }
317 }
318 }
319
320 // Bail if RHS was a icmp that can't be decomposed into an equality.
321 if (!ICmpInst::isEquality(PredR))
322 return std::nullopt;
323
324 if (L11 == A) {
325 B = L12;
326 C = L2;
327 } else if (L12 == A) {
328 B = L11;
329 C = L2;
330 } else if (L21 == A) {
331 B = L22;
332 C = L1;
333 } else if (L22 == A) {
334 B = L21;
335 C = L1;
336 }
337
338 unsigned LeftType = getMaskedICmpType(A, B, C, PredL);
339 unsigned RightType = getMaskedICmpType(A, D, E, PredR);
340 return std::optional<std::pair<unsigned, unsigned>>(
341 std::make_pair(LeftType, RightType));
342}
343
344/// Try to fold (icmp(A & B) ==/!= C) &/| (icmp(A & D) ==/!= E) into a single
345/// (icmp(A & X) ==/!= Y), where the left-hand side is of type Mask_NotAllZeros
346/// and the right hand side is of type BMask_Mixed. For example,
347/// (icmp (A & 12) != 0) & (icmp (A & 15) == 8) -> (icmp (A & 15) == 8).
348/// Also used for logical and/or, must be poison safe.
350 Value *LHS, Value *RHS, bool IsAnd, Value *A, Value *B, Value *D, Value *E,
352 InstCombiner::BuilderTy &Builder) {
353 // We are given the canonical form:
354 // (icmp ne (A & B), 0) & (icmp eq (A & D), E).
355 // where D & E == E.
356 //
357 // If IsAnd is false, we get it in negated form:
358 // (icmp eq (A & B), 0) | (icmp ne (A & D), E) ->
359 // !((icmp ne (A & B), 0) & (icmp eq (A & D), E)).
360 //
361 // We currently handle the case of B, C, D, E are constant.
362 //
363 const APInt *BCst, *DCst, *OrigECst;
364 if (!match(B, m_APInt(BCst)) || !match(D, m_APInt(DCst)) ||
365 !match(E, m_APInt(OrigECst)))
366 return nullptr;
367
369
370 // Update E to the canonical form when D is a power of two and RHS is
371 // canonicalized as,
372 // (icmp ne (A & D), 0) -> (icmp eq (A & D), D) or
373 // (icmp ne (A & D), D) -> (icmp eq (A & D), 0).
374 APInt ECst = *OrigECst;
375 if (PredR != NewCC)
376 ECst ^= *DCst;
377
378 // If B or D is zero, skip because if LHS or RHS can be trivially folded by
379 // other folding rules and this pattern won't apply any more.
380 if (*BCst == 0 || *DCst == 0)
381 return nullptr;
382
383 // If B and D don't intersect, ie. (B & D) == 0, try to fold isNaN idiom:
384 // (icmp ne (A & FractionBits), 0) & (icmp eq (A & ExpBits), ExpBits)
385 // -> isNaN(A)
386 // Otherwise, we cannot deduce anything from it.
387 if (!BCst->intersects(*DCst)) {
388 Value *Src;
389 if (*DCst == ECst && match(A, m_ElementWiseBitCast(m_Value(Src))) &&
390 !Builder.GetInsertBlock()->getParent()->hasFnAttribute(
391 Attribute::StrictFP)) {
392 Type *Ty = Src->getType()->getScalarType();
393 if (!Ty->isIEEELikeFPTy())
394 return nullptr;
395
396 APInt ExpBits = APFloat::getInf(Ty->getFltSemantics()).bitcastToAPInt();
397 if (ECst != ExpBits)
398 return nullptr;
399 APInt FractionBits = ~ExpBits;
400 FractionBits.clearSignBit();
401 if (*BCst != FractionBits)
402 return nullptr;
403
404 return Builder.CreateFCmp(IsAnd ? FCmpInst::FCMP_UNO : FCmpInst::FCMP_ORD,
405 Src, ConstantFP::getZero(Src->getType()));
406 }
407 return nullptr;
408 }
409
410 // If the following two conditions are met:
411 //
412 // 1. mask B covers only a single bit that's not covered by mask D, that is,
413 // (B & (B ^ D)) is a power of 2 (in other words, B minus the intersection of
414 // B and D has only one bit set) and,
415 //
416 // 2. RHS (and E) indicates that the rest of B's bits are zero (in other
417 // words, the intersection of B and D is zero), that is, ((B & D) & E) == 0
418 //
419 // then that single bit in B must be one and thus the whole expression can be
420 // folded to
421 // (A & (B | D)) == (B & (B ^ D)) | E.
422 //
423 // For example,
424 // (icmp ne (A & 12), 0) & (icmp eq (A & 7), 1) -> (icmp eq (A & 15), 9)
425 // (icmp ne (A & 15), 0) & (icmp eq (A & 7), 0) -> (icmp eq (A & 15), 8)
426 if ((((*BCst & *DCst) & ECst) == 0) &&
427 (*BCst & (*BCst ^ *DCst)).isPowerOf2()) {
428 APInt BorD = *BCst | *DCst;
429 APInt BandBxorDorE = (*BCst & (*BCst ^ *DCst)) | ECst;
430 Value *NewMask = ConstantInt::get(A->getType(), BorD);
431 Value *NewMaskedValue = ConstantInt::get(A->getType(), BandBxorDorE);
432 Value *NewAnd = Builder.CreateAnd(A, NewMask);
433 return Builder.CreateICmp(NewCC, NewAnd, NewMaskedValue);
434 }
435
436 auto IsSubSetOrEqual = [](const APInt *C1, const APInt *C2) {
437 return (*C1 & *C2) == *C1;
438 };
439 auto IsSuperSetOrEqual = [](const APInt *C1, const APInt *C2) {
440 return (*C1 & *C2) == *C2;
441 };
442
443 // In the following, we consider only the cases where B is a superset of D, B
444 // is a subset of D, or B == D because otherwise there's at least one bit
445 // covered by B but not D, in which case we can't deduce much from it, so
446 // no folding (aside from the single must-be-one bit case right above.)
447 // For example,
448 // (icmp ne (A & 14), 0) & (icmp eq (A & 3), 1) -> no folding.
449 if (!IsSubSetOrEqual(BCst, DCst) && !IsSuperSetOrEqual(BCst, DCst))
450 return nullptr;
451
452 // At this point, either B is a superset of D, B is a subset of D or B == D.
453
454 // If E is zero, if B is a subset of (or equal to) D, LHS and RHS contradict
455 // and the whole expression becomes false (or true if negated), otherwise, no
456 // folding.
457 // For example,
458 // (icmp ne (A & 3), 0) & (icmp eq (A & 7), 0) -> false.
459 // (icmp ne (A & 15), 0) & (icmp eq (A & 3), 0) -> no folding.
460 if (ECst.isZero()) {
461 if (IsSubSetOrEqual(BCst, DCst))
462 return ConstantInt::get(LHS->getType(), !IsAnd);
463 return nullptr;
464 }
465
466 // At this point, B, D, E aren't zero and (B & D) == B, (B & D) == D or B ==
467 // D. If B is a superset of (or equal to) D, since E is not zero, LHS is
468 // subsumed by RHS (RHS implies LHS.) So the whole expression becomes
469 // RHS. For example,
470 // (icmp ne (A & 255), 0) & (icmp eq (A & 15), 8) -> (icmp eq (A & 15), 8).
471 // (icmp ne (A & 15), 0) & (icmp eq (A & 15), 8) -> (icmp eq (A & 15), 8).
472 if (IsSuperSetOrEqual(BCst, DCst)) {
473 // We can't guarantee that samesign hold after this fold.
474 if (auto *ICmp = dyn_cast<ICmpInst>(RHS))
475 ICmp->setSameSign(false);
476 return RHS;
477 }
478 // Otherwise, B is a subset of D. If B and E have a common bit set,
479 // ie. (B & E) != 0, then LHS is subsumed by RHS. For example.
480 // (icmp ne (A & 12), 0) & (icmp eq (A & 15), 8) -> (icmp eq (A & 15), 8).
481 assert(IsSubSetOrEqual(BCst, DCst) && "Precondition due to above code");
482 if ((*BCst & ECst) != 0) {
483 // We can't guarantee that samesign hold after this fold.
484 if (auto *ICmp = dyn_cast<ICmpInst>(RHS))
485 ICmp->setSameSign(false);
486 return RHS;
487 }
488 // Otherwise, LHS and RHS contradict and the whole expression becomes false
489 // (or true if negated.) For example,
490 // (icmp ne (A & 7), 0) & (icmp eq (A & 15), 8) -> false.
491 // (icmp ne (A & 6), 0) & (icmp eq (A & 15), 8) -> false.
492 return ConstantInt::get(LHS->getType(), !IsAnd);
493}
494
495/// Try to fold (icmp(A & B) ==/!= 0) &/| (icmp(A & D) ==/!= E) into a single
496/// (icmp(A & X) ==/!= Y), where the left-hand side and the right hand side
497/// aren't of the common mask pattern type.
498/// Also used for logical and/or, must be poison safe.
500 Value *LHS, Value *RHS, bool IsAnd, Value *A, Value *B, Value *C, Value *D,
502 unsigned LHSMask, unsigned RHSMask, InstCombiner::BuilderTy &Builder) {
504 "Expected equality predicates for masked type of icmps.");
505 // Handle Mask_NotAllZeros-BMask_Mixed cases.
506 // (icmp ne/eq (A & B), C) &/| (icmp eq/ne (A & D), E), or
507 // (icmp eq/ne (A & B), C) &/| (icmp ne/eq (A & D), E)
508 // which gets swapped to
509 // (icmp ne/eq (A & D), E) &/| (icmp eq/ne (A & B), C).
510 if (!IsAnd) {
511 LHSMask = conjugateICmpMask(LHSMask);
512 RHSMask = conjugateICmpMask(RHSMask);
513 }
514 if ((LHSMask & Mask_NotAllZeros) && (RHSMask & BMask_Mixed)) {
516 LHS, RHS, IsAnd, A, B, D, E, PredL, PredR, Builder)) {
517 return V;
518 }
519 } else if ((LHSMask & BMask_Mixed) && (RHSMask & Mask_NotAllZeros)) {
521 RHS, LHS, IsAnd, A, D, B, C, PredR, PredL, Builder)) {
522 return V;
523 }
524 }
525 return nullptr;
526}
527
528/// Try to fold (icmp(A & B) ==/!= C) &/| (icmp(A & D) ==/!= E)
529/// into a single (icmp(A & X) ==/!= Y).
531 bool IsLogical,
533 const SimplifyQuery &Q) {
534 Value *A = nullptr, *B = nullptr, *C = nullptr, *D = nullptr, *E = nullptr;
535 ICmpInst::Predicate PredL, PredR;
536 std::optional<std::pair<unsigned, unsigned>> MaskPair =
537 getMaskedTypeForICmpPair(A, B, C, D, E, LHS, RHS, PredL, PredR);
538 if (!MaskPair)
539 return nullptr;
541 "Expected equality predicates for masked type of icmps.");
542 unsigned LHSMask = MaskPair->first;
543 unsigned RHSMask = MaskPair->second;
544 unsigned Mask = LHSMask & RHSMask;
545 if (Mask == 0) {
546 // Even if the two sides don't share a common pattern, check if folding can
547 // still happen.
549 LHS, RHS, IsAnd, A, B, C, D, E, PredL, PredR, LHSMask, RHSMask,
550 Builder))
551 return V;
552 return nullptr;
553 }
554
555 // In full generality:
556 // (icmp (A & B) Op C) | (icmp (A & D) Op E)
557 // == ![ (icmp (A & B) !Op C) & (icmp (A & D) !Op E) ]
558 //
559 // If the latter can be converted into (icmp (A & X) Op Y) then the former is
560 // equivalent to (icmp (A & X) !Op Y).
561 //
562 // Therefore, we can pretend for the rest of this function that we're dealing
563 // with the conjunction, provided we flip the sense of any comparisons (both
564 // input and output).
565
566 // In most cases we're going to produce an EQ for the "&&" case.
568 if (!IsAnd) {
569 // Convert the masking analysis into its equivalent with negated
570 // comparisons.
571 Mask = conjugateICmpMask(Mask);
572 }
573
574 if (Mask & Mask_AllZeros) {
575 // (icmp eq (A & B), 0) & (icmp eq (A & D), 0)
576 // -> (icmp eq (A & (B|D)), 0)
577 if (IsLogical && !isGuaranteedNotToBeUndefOrPoison(D))
578 return nullptr; // TODO: Use freeze?
579 Value *NewOr = Builder.CreateOr(B, D);
580 Value *NewAnd = Builder.CreateAnd(A, NewOr);
581 // We can't use C as zero because we might actually handle
582 // (icmp ne (A & B), B) & (icmp ne (A & D), D)
583 // with B and D, having a single bit set.
584 Value *Zero = Constant::getNullValue(A->getType());
585 return Builder.CreateICmp(NewCC, NewAnd, Zero);
586 }
587 if (Mask & BMask_AllOnes) {
588 // (icmp eq (A & B), B) & (icmp eq (A & D), D)
589 // -> (icmp eq (A & (B|D)), (B|D))
590 if (IsLogical && !isGuaranteedNotToBeUndefOrPoison(D))
591 return nullptr; // TODO: Use freeze?
592 Value *NewOr = Builder.CreateOr(B, D);
593 Value *NewAnd = Builder.CreateAnd(A, NewOr);
594 return Builder.CreateICmp(NewCC, NewAnd, NewOr);
595 }
596 if (Mask & AMask_AllOnes) {
597 // (icmp eq (A & B), A) & (icmp eq (A & D), A)
598 // -> (icmp eq (A & (B&D)), A)
599 if (IsLogical && !isGuaranteedNotToBeUndefOrPoison(D))
600 return nullptr; // TODO: Use freeze?
601 Value *NewAnd1 = Builder.CreateAnd(B, D);
602 Value *NewAnd2 = Builder.CreateAnd(A, NewAnd1);
603 return Builder.CreateICmp(NewCC, NewAnd2, A);
604 }
605
606 const APInt *ConstB, *ConstD;
607 if (match(B, m_APInt(ConstB)) && match(D, m_APInt(ConstD))) {
608 if (Mask & (Mask_NotAllZeros | BMask_NotAllOnes)) {
609 // (icmp ne (A & B), 0) & (icmp ne (A & D), 0) and
610 // (icmp ne (A & B), B) & (icmp ne (A & D), D)
611 // -> (icmp ne (A & B), 0) or (icmp ne (A & D), 0)
612 // Only valid if one of the masks is a superset of the other (check "B&D"
613 // is the same as either B or D).
614 APInt NewMask = *ConstB & *ConstD;
615 if (NewMask == *ConstB)
616 return LHS;
617 if (NewMask == *ConstD) {
618 if (IsLogical) {
619 if (auto *RHSI = dyn_cast<Instruction>(RHS))
620 RHSI->dropPoisonGeneratingFlags();
621 }
622 return RHS;
623 }
624 }
625
626 if (Mask & AMask_NotAllOnes) {
627 // (icmp ne (A & B), B) & (icmp ne (A & D), D)
628 // -> (icmp ne (A & B), A) or (icmp ne (A & D), A)
629 // Only valid if one of the masks is a superset of the other (check "B|D"
630 // is the same as either B or D).
631 APInt NewMask = *ConstB | *ConstD;
632 if (NewMask == *ConstB)
633 return LHS;
634 if (NewMask == *ConstD)
635 return RHS;
636 }
637
638 if (Mask & (BMask_Mixed | BMask_NotMixed)) {
639 // Mixed:
640 // (icmp eq (A & B), C) & (icmp eq (A & D), E)
641 // We already know that B & C == C && D & E == E.
642 // If we can prove that (B & D) & (C ^ E) == 0, that is, the bits of
643 // C and E, which are shared by both the mask B and the mask D, don't
644 // contradict, then we can transform to
645 // -> (icmp eq (A & (B|D)), (C|E))
646 // Currently, we only handle the case of B, C, D, and E being constant.
647 // We can't simply use C and E because we might actually handle
648 // (icmp ne (A & B), B) & (icmp eq (A & D), D)
649 // with B and D, having a single bit set.
650
651 // NotMixed:
652 // (icmp ne (A & B), C) & (icmp ne (A & D), E)
653 // -> (icmp ne (A & (B & D)), (C & E))
654 // Check the intersection (B & D) for inequality.
655 // Assume that (B & D) == B || (B & D) == D, i.e B/D is a subset of D/B
656 // and (B & D) & (C ^ E) == 0, bits of C and E, which are shared by both
657 // the B and the D, don't contradict. Note that we can assume (~B & C) ==
658 // 0 && (~D & E) == 0, previous operation should delete these icmps if it
659 // hadn't been met.
660
661 const APInt *OldConstC, *OldConstE;
662 if (!match(C, m_APInt(OldConstC)) || !match(E, m_APInt(OldConstE)))
663 return nullptr;
664
665 auto FoldBMixed = [&](ICmpInst::Predicate CC, bool IsNot) -> Value * {
666 CC = IsNot ? CmpInst::getInversePredicate(CC) : CC;
667 const APInt ConstC = PredL != CC ? *ConstB ^ *OldConstC : *OldConstC;
668 const APInt ConstE = PredR != CC ? *ConstD ^ *OldConstE : *OldConstE;
669
670 if (((*ConstB & *ConstD) & (ConstC ^ ConstE)).getBoolValue())
671 return IsNot ? nullptr : ConstantInt::get(LHS->getType(), !IsAnd);
672
673 if (IsNot && !ConstB->isSubsetOf(*ConstD) &&
674 !ConstD->isSubsetOf(*ConstB))
675 return nullptr;
676
677 APInt BD, CE;
678 if (IsNot) {
679 BD = *ConstB & *ConstD;
680 CE = ConstC & ConstE;
681 } else {
682 BD = *ConstB | *ConstD;
683 CE = ConstC | ConstE;
684 }
685 Value *NewAnd = Builder.CreateAnd(A, BD);
686 Value *CEVal = ConstantInt::get(A->getType(), CE);
687 return Builder.CreateICmp(CC, NewAnd, CEVal);
688 };
689
690 if (Mask & BMask_Mixed)
691 return FoldBMixed(NewCC, false);
692 if (Mask & BMask_NotMixed) // can be else also
693 return FoldBMixed(NewCC, true);
694 }
695 }
696
697 // (icmp eq (A & B), 0) | (icmp eq (A & D), 0)
698 // -> (icmp ne (A & (B|D)), (B|D))
699 // (icmp ne (A & B), 0) & (icmp ne (A & D), 0)
700 // -> (icmp eq (A & (B|D)), (B|D))
701 // iff B and D is known to be a power of two
702 if (Mask & Mask_NotAllZeros &&
703 isKnownToBeAPowerOfTwo(B, /*OrZero=*/false, Q) &&
704 isKnownToBeAPowerOfTwo(D, /*OrZero=*/false, Q)) {
705 // If this is a logical and/or, then we must prevent propagation of a
706 // poison value from the RHS by inserting freeze.
707 if (IsLogical)
708 D = Builder.CreateFreeze(D);
709 Value *Mask = Builder.CreateOr(B, D);
710 Value *Masked = Builder.CreateAnd(A, Mask);
711 return Builder.CreateICmp(NewCC, Masked, Mask);
712 }
713 return nullptr;
714}
715
716/// Try to fold a signed range checked with lower bound 0 to an unsigned icmp.
717/// Example: (icmp sge x, 0) & (icmp slt x, n) --> icmp ult x, n
718/// If \p Inverted is true then the check is for the inverted range, e.g.
719/// (icmp slt x, 0) | (icmp sgt x, n) --> icmp ugt x, n
721 bool Inverted) {
722 // Check the lower range comparison, e.g. x >= 0
723 // InstCombine already ensured that if there is a constant it's on the RHS.
724 ConstantInt *RangeStart = dyn_cast<ConstantInt>(Cmp0->getOperand(1));
725 if (!RangeStart)
726 return nullptr;
727
728 ICmpInst::Predicate Pred0 = (Inverted ? Cmp0->getInversePredicate() :
729 Cmp0->getPredicate());
730
731 // Accept x > -1 or x >= 0 (after potentially inverting the predicate).
732 if (!((Pred0 == ICmpInst::ICMP_SGT && RangeStart->isMinusOne()) ||
733 (Pred0 == ICmpInst::ICMP_SGE && RangeStart->isZero())))
734 return nullptr;
735
736 ICmpInst::Predicate Pred1 = (Inverted ? Cmp1->getInversePredicate() :
737 Cmp1->getPredicate());
738
739 Value *Input = Cmp0->getOperand(0);
740 Value *Cmp1Op0 = Cmp1->getOperand(0);
741 Value *Cmp1Op1 = Cmp1->getOperand(1);
742 Value *RangeEnd;
743 if (match(Cmp1Op0, m_SExtOrSelf(m_Specific(Input)))) {
744 // For the upper range compare we have: icmp x, n
745 Input = Cmp1Op0;
746 RangeEnd = Cmp1Op1;
747 } else if (match(Cmp1Op1, m_SExtOrSelf(m_Specific(Input)))) {
748 // For the upper range compare we have: icmp n, x
749 Input = Cmp1Op1;
750 RangeEnd = Cmp1Op0;
751 Pred1 = ICmpInst::getSwappedPredicate(Pred1);
752 } else {
753 return nullptr;
754 }
755
756 // Check the upper range comparison, e.g. x < n
757 ICmpInst::Predicate NewPred;
758 switch (Pred1) {
759 case ICmpInst::ICMP_SLT: NewPred = ICmpInst::ICMP_ULT; break;
760 case ICmpInst::ICMP_SLE: NewPred = ICmpInst::ICMP_ULE; break;
761 default: return nullptr;
762 }
763
764 // This simplification is only valid if the upper range is not negative.
765 KnownBits Known = computeKnownBits(RangeEnd, Cmp1);
766 if (!Known.isNonNegative())
767 return nullptr;
768
769 if (Inverted)
770 NewPred = ICmpInst::getInversePredicate(NewPred);
771
772 return Builder.CreateICmp(NewPred, Input, RangeEnd);
773}
774
775// (or (icmp eq X, 0), (icmp eq X, Pow2OrZero))
776// -> (icmp eq (and X, Pow2OrZero), X)
777// (and (icmp ne X, 0), (icmp ne X, Pow2OrZero))
778// -> (icmp ne (and X, Pow2OrZero), X)
779static Value *
781 ICmpInst *LHS, ICmpInst *RHS, bool IsAnd,
782 const SimplifyQuery &Q) {
784 // Make sure we have right compares for our op.
785 if (LHS->getPredicate() != Pred || RHS->getPredicate() != Pred)
786 return nullptr;
787
788 // Make it so we can match LHS against the (icmp eq/ne X, 0) just for
789 // simplicity.
790 if (match(RHS->getOperand(1), m_Zero()))
791 std::swap(LHS, RHS);
792
793 Value *Pow2, *Op;
794 // Match the desired pattern:
795 // LHS: (icmp eq/ne X, 0)
796 // RHS: (icmp eq/ne X, Pow2OrZero)
797 // Skip if Pow2OrZero is 1. Either way it gets folded to (icmp ugt X, 1) but
798 // this form ends up slightly less canonical.
799 // We could potentially be more sophisticated than requiring LHS/RHS
800 // be one-use. We don't create additional instructions if only one
801 // of them is one-use. So cases where one is one-use and the other
802 // is two-use might be profitable.
803 if (!match(LHS, m_OneUse(m_ICmp(Pred, m_Value(Op), m_Zero()))) ||
804 !match(RHS, m_OneUse(m_c_ICmp(Pred, m_Specific(Op), m_Value(Pow2)))) ||
805 match(Pow2, m_One()) ||
806 !isKnownToBeAPowerOfTwo(Pow2, Q.DL, /*OrZero=*/true, Q.AC, Q.CxtI, Q.DT))
807 return nullptr;
808
809 Value *And = Builder.CreateAnd(Op, Pow2);
810 return Builder.CreateICmp(Pred, And, Op);
811}
812
813/// General pattern:
814/// X & Y
815///
816/// Where Y is checking that all the high bits (covered by a mask 4294967168)
817/// are uniform, i.e. %arg & 4294967168 can be either 4294967168 or 0
818/// Pattern can be one of:
819/// %t = add i32 %arg, 128
820/// %r = icmp ult i32 %t, 256
821/// Or
822/// %t0 = shl i32 %arg, 24
823/// %t1 = ashr i32 %t0, 24
824/// %r = icmp eq i32 %t1, %arg
825/// Or
826/// %t0 = trunc i32 %arg to i8
827/// %t1 = sext i8 %t0 to i32
828/// %r = icmp eq i32 %t1, %arg
829/// This pattern is a signed truncation check.
830///
831/// And X is checking that some bit in that same mask is zero.
832/// I.e. can be one of:
833/// %r = icmp sgt i32 %arg, -1
834/// Or
835/// %t = and i32 %arg, 2147483648
836/// %r = icmp eq i32 %t, 0
837///
838/// Since we are checking that all the bits in that mask are the same,
839/// and a particular bit is zero, what we are really checking is that all the
840/// masked bits are zero.
841/// So this should be transformed to:
842/// %r = icmp ult i32 %arg, 128
844 Instruction &CxtI,
845 InstCombiner::BuilderTy &Builder) {
846 assert(CxtI.getOpcode() == Instruction::And);
847
848 // Match icmp ult (add %arg, C01), C1 (C1 == C01 << 1; powers of two)
849 auto tryToMatchSignedTruncationCheck = [](ICmpInst *ICmp, Value *&X,
850 APInt &SignBitMask) -> bool {
851 const APInt *I01, *I1; // powers of two; I1 == I01 << 1
853 m_Add(m_Value(X), m_Power2(I01)),
854 m_Power2(I1))) &&
855 I1->ugt(*I01) && I01->shl(1) == *I1))
856 return false;
857 // Which bit is the new sign bit as per the 'signed truncation' pattern?
858 SignBitMask = *I01;
859 return true;
860 };
861
862 // One icmp needs to be 'signed truncation check'.
863 // We need to match this first, else we will mismatch commutative cases.
864 Value *X1;
865 APInt HighestBit;
866 ICmpInst *OtherICmp;
867 if (tryToMatchSignedTruncationCheck(ICmp1, X1, HighestBit))
868 OtherICmp = ICmp0;
869 else if (tryToMatchSignedTruncationCheck(ICmp0, X1, HighestBit))
870 OtherICmp = ICmp1;
871 else
872 return nullptr;
873
874 assert(HighestBit.isPowerOf2() && "expected to be power of two (non-zero)");
875
876 // Try to match/decompose into: icmp eq (X & Mask), 0
877 auto tryToDecompose = [](ICmpInst *ICmp, Value *&X,
878 APInt &UnsetBitsMask) -> bool {
879 CmpPredicate Pred = ICmp->getPredicate();
880 // Can it be decomposed into icmp eq (X & Mask), 0 ?
882 ICmp->getOperand(0), ICmp->getOperand(1), Pred,
883 /*LookThroughTrunc=*/false, /*AllowNonZeroC=*/false,
884 /*DecomposeAnd=*/true);
885 if (Res && Res->Pred == ICmpInst::ICMP_EQ) {
886 X = Res->X;
887 UnsetBitsMask = Res->Mask;
888 return true;
889 }
890
891 return false;
892 };
893
894 // And the other icmp needs to be decomposable into a bit test.
895 Value *X0;
896 APInt UnsetBitsMask;
897 if (!tryToDecompose(OtherICmp, X0, UnsetBitsMask))
898 return nullptr;
899
900 assert(!UnsetBitsMask.isZero() && "empty mask makes no sense.");
901
902 // Are they working on the same value?
903 Value *X;
904 if (X1 == X0) {
905 // Ok as is.
906 X = X1;
907 } else if (match(X0, m_Trunc(m_Specific(X1)))) {
908 UnsetBitsMask = UnsetBitsMask.zext(X1->getType()->getScalarSizeInBits());
909 X = X1;
910 } else
911 return nullptr;
912
913 // So which bits should be uniform as per the 'signed truncation check'?
914 // (all the bits starting with (i.e. including) HighestBit)
915 APInt SignBitsMask = ~(HighestBit - 1U);
916
917 // UnsetBitsMask must have some common bits with SignBitsMask,
918 if (!UnsetBitsMask.intersects(SignBitsMask))
919 return nullptr;
920
921 // Does UnsetBitsMask contain any bits outside of SignBitsMask?
922 if (!UnsetBitsMask.isSubsetOf(SignBitsMask)) {
923 APInt OtherHighestBit = (~UnsetBitsMask) + 1U;
924 if (!OtherHighestBit.isPowerOf2())
925 return nullptr;
926 HighestBit = APIntOps::umin(HighestBit, OtherHighestBit);
927 }
928 // Else, if it does not, then all is ok as-is.
929
930 // %r = icmp ult %X, SignBit
931 return Builder.CreateICmpULT(X, ConstantInt::get(X->getType(), HighestBit),
932 CxtI.getName() + ".simplified");
933}
934
935/// Fold (icmp eq ctpop(X) 1) | (icmp eq X 0) into (icmp ult ctpop(X) 2) and
936/// fold (icmp ne ctpop(X) 1) & (icmp ne X 0) into (icmp ugt ctpop(X) 1).
937/// Also used for logical and/or, must be poison safe if range attributes are
938/// dropped.
939static Value *foldIsPowerOf2OrZero(ICmpInst *Cmp0, ICmpInst *Cmp1, bool IsAnd,
941 InstCombinerImpl &IC) {
942 CmpPredicate Pred0, Pred1;
943 Value *X;
945 m_SpecificInt(1))) ||
946 !match(Cmp1, m_ICmp(Pred1, m_Specific(X), m_ZeroInt())))
947 return nullptr;
948
949 auto *CtPop = cast<Instruction>(Cmp0->getOperand(0));
950 if (IsAnd && Pred0 == ICmpInst::ICMP_NE && Pred1 == ICmpInst::ICMP_NE) {
951 // Drop range attributes and re-infer them in the next iteration.
952 CtPop->dropPoisonGeneratingAnnotations();
953 IC.addToWorklist(CtPop);
954 return Builder.CreateICmpUGT(CtPop, ConstantInt::get(CtPop->getType(), 1));
955 }
956 if (!IsAnd && Pred0 == ICmpInst::ICMP_EQ && Pred1 == ICmpInst::ICMP_EQ) {
957 // Drop range attributes and re-infer them in the next iteration.
958 CtPop->dropPoisonGeneratingAnnotations();
959 IC.addToWorklist(CtPop);
960 return Builder.CreateICmpULT(CtPop, ConstantInt::get(CtPop->getType(), 2));
961 }
962
963 return nullptr;
964}
965
966/// Reduce a pair of compares that check if a value has exactly 1 bit set.
967/// Also used for logical and/or, must be poison safe if range attributes are
968/// dropped.
969static Value *foldIsPowerOf2(ICmpInst *Cmp0, ICmpInst *Cmp1, bool JoinedByAnd,
971 InstCombinerImpl &IC) {
972 // Handle 'and' / 'or' commutation: make the equality check the first operand.
973 if (JoinedByAnd && Cmp1->getPredicate() == ICmpInst::ICMP_NE)
974 std::swap(Cmp0, Cmp1);
975 else if (!JoinedByAnd && Cmp1->getPredicate() == ICmpInst::ICMP_EQ)
976 std::swap(Cmp0, Cmp1);
977
978 // (X != 0) && (ctpop(X) u< 2) --> ctpop(X) == 1
979 Value *X;
980 if (JoinedByAnd &&
984 m_SpecificInt(2)))) {
985 auto *CtPop = cast<Instruction>(Cmp1->getOperand(0));
986 // Drop range attributes and re-infer them in the next iteration.
987 CtPop->dropPoisonGeneratingAnnotations();
988 IC.addToWorklist(CtPop);
989 return Builder.CreateICmpEQ(CtPop, ConstantInt::get(CtPop->getType(), 1));
990 }
991 // (X == 0) || (ctpop(X) u> 1) --> ctpop(X) != 1
992 if (!JoinedByAnd &&
996 m_SpecificInt(1)))) {
997 auto *CtPop = cast<Instruction>(Cmp1->getOperand(0));
998 // Drop range attributes and re-infer them in the next iteration.
999 CtPop->dropPoisonGeneratingAnnotations();
1000 IC.addToWorklist(CtPop);
1001 return Builder.CreateICmpNE(CtPop, ConstantInt::get(CtPop->getType(), 1));
1002 }
1003 return nullptr;
1004}
1005
1006/// Try to fold (icmp(A & B) == 0) & (icmp(A & D) != E) into (icmp A u< D) iff
1007/// B is a contiguous set of ones starting from the most significant bit
1008/// (negative power of 2), D and E are equal, and D is a contiguous set of ones
1009/// starting at the most significant zero bit in B. Parameter B supports masking
1010/// using undef/poison in either scalar or vector values.
1012 Value *A, Value *B, Value *D, Value *E, ICmpInst::Predicate PredL,
1015 "Expected equality predicates for masked type of icmps.");
1016 if (PredL != ICmpInst::ICMP_EQ || PredR != ICmpInst::ICMP_NE)
1017 return nullptr;
1018
1019 if (!match(B, m_NegatedPower2()) || !match(D, m_ShiftedMask()) ||
1020 !match(E, m_ShiftedMask()))
1021 return nullptr;
1022
1023 // Test scalar arguments for conversion. B has been validated earlier to be a
1024 // negative power of two and thus is guaranteed to have one or more contiguous
1025 // ones starting from the MSB followed by zero or more contiguous zeros. D has
1026 // been validated earlier to be a shifted set of one or more contiguous ones.
1027 // In order to match, B leading ones and D leading zeros should be equal. The
1028 // predicate that B be a negative power of 2 prevents the condition of there
1029 // ever being zero leading ones. Thus 0 == 0 cannot occur. The predicate that
1030 // D always be a shifted mask prevents the condition of D equaling 0. This
1031 // prevents matching the condition where B contains the maximum number of
1032 // leading one bits (-1) and D contains the maximum number of leading zero
1033 // bits (0).
1034 auto isReducible = [](const Value *B, const Value *D, const Value *E) {
1035 const APInt *BCst, *DCst, *ECst;
1036 return match(B, m_APIntAllowPoison(BCst)) && match(D, m_APInt(DCst)) &&
1037 match(E, m_APInt(ECst)) && *DCst == *ECst &&
1038 (isa<PoisonValue>(B) ||
1039 (BCst->countLeadingOnes() == DCst->countLeadingZeros()));
1040 };
1041
1042 // Test vector type arguments for conversion.
1043 if (const auto *BVTy = dyn_cast<VectorType>(B->getType())) {
1044 const auto *BFVTy = dyn_cast<FixedVectorType>(BVTy);
1045 const auto *BConst = dyn_cast<Constant>(B);
1046 const auto *DConst = dyn_cast<Constant>(D);
1047 const auto *EConst = dyn_cast<Constant>(E);
1048
1049 if (!BFVTy || !BConst || !DConst || !EConst)
1050 return nullptr;
1051
1052 for (unsigned I = 0; I != BFVTy->getNumElements(); ++I) {
1053 const auto *BElt = BConst->getAggregateElement(I);
1054 const auto *DElt = DConst->getAggregateElement(I);
1055 const auto *EElt = EConst->getAggregateElement(I);
1056
1057 if (!BElt || !DElt || !EElt)
1058 return nullptr;
1059 if (!isReducible(BElt, DElt, EElt))
1060 return nullptr;
1061 }
1062 } else {
1063 // Test scalar type arguments for conversion.
1064 if (!isReducible(B, D, E))
1065 return nullptr;
1066 }
1067 return Builder.CreateICmp(ICmpInst::ICMP_ULT, A, D);
1068}
1069
1070/// Try to fold ((icmp X u< P) & (icmp(X & M) != M)) or ((icmp X s> -1) &
1071/// (icmp(X & M) != M)) into (icmp X u< M). Where P is a power of 2, M < P, and
1072/// M is a contiguous shifted mask starting at the right most significant zero
1073/// bit in P. SGT is supported as when P is the largest representable power of
1074/// 2, an earlier optimization converts the expression into (icmp X s> -1).
1075/// Parameter P supports masking using undef/poison in either scalar or vector
1076/// values.
1078 bool JoinedByAnd,
1079 InstCombiner::BuilderTy &Builder) {
1080 if (!JoinedByAnd)
1081 return nullptr;
1082 Value *A = nullptr, *B = nullptr, *C = nullptr, *D = nullptr, *E = nullptr;
1083 ICmpInst::Predicate CmpPred0, CmpPred1;
1084 // Assuming P is a 2^n, getMaskedTypeForICmpPair will normalize (icmp X u<
1085 // 2^n) into (icmp (X & ~(2^n-1)) == 0) and (icmp X s> -1) into (icmp (X &
1086 // SignMask) == 0).
1087 std::optional<std::pair<unsigned, unsigned>> MaskPair =
1088 getMaskedTypeForICmpPair(A, B, C, D, E, Cmp0, Cmp1, CmpPred0, CmpPred1);
1089 if (!MaskPair)
1090 return nullptr;
1091
1092 const auto compareBMask = BMask_NotMixed | BMask_NotAllOnes;
1093 unsigned CmpMask0 = MaskPair->first;
1094 unsigned CmpMask1 = MaskPair->second;
1095 if ((CmpMask0 & Mask_AllZeros) && (CmpMask1 == compareBMask)) {
1096 if (Value *V = foldNegativePower2AndShiftedMask(A, B, D, E, CmpPred0,
1097 CmpPred1, Builder))
1098 return V;
1099 } else if ((CmpMask0 == compareBMask) && (CmpMask1 & Mask_AllZeros)) {
1100 if (Value *V = foldNegativePower2AndShiftedMask(A, D, B, C, CmpPred1,
1101 CmpPred0, Builder))
1102 return V;
1103 }
1104 return nullptr;
1105}
1106
1107/// Commuted variants are assumed to be handled by calling this function again
1108/// with the parameters swapped.
1110 ICmpInst *UnsignedICmp, bool IsAnd,
1111 const SimplifyQuery &Q,
1112 InstCombiner::BuilderTy &Builder) {
1113 Value *ZeroCmpOp;
1114 CmpPredicate EqPred;
1115 if (!match(ZeroICmp, m_ICmp(EqPred, m_Value(ZeroCmpOp), m_Zero())) ||
1116 !ICmpInst::isEquality(EqPred))
1117 return nullptr;
1118
1119 CmpPredicate UnsignedPred;
1120
1121 Value *A, *B;
1122 if (match(UnsignedICmp,
1123 m_c_ICmp(UnsignedPred, m_Specific(ZeroCmpOp), m_Value(A))) &&
1124 match(ZeroCmpOp, m_c_Add(m_Specific(A), m_Value(B))) &&
1125 (ZeroICmp->hasOneUse() || UnsignedICmp->hasOneUse())) {
1126 auto GetKnownNonZeroAndOther = [&](Value *&NonZero, Value *&Other) {
1127 if (!isKnownNonZero(NonZero, Q))
1128 std::swap(NonZero, Other);
1129 return isKnownNonZero(NonZero, Q);
1130 };
1131
1132 // Given ZeroCmpOp = (A + B)
1133 // ZeroCmpOp < A && ZeroCmpOp != 0 --> (0-X) < Y iff
1134 // ZeroCmpOp >= A || ZeroCmpOp == 0 --> (0-X) >= Y iff
1135 // with X being the value (A/B) that is known to be non-zero,
1136 // and Y being remaining value.
1137 if (UnsignedPred == ICmpInst::ICMP_ULT && EqPred == ICmpInst::ICMP_NE &&
1138 IsAnd && GetKnownNonZeroAndOther(B, A))
1139 return Builder.CreateICmpULT(Builder.CreateNeg(B), A);
1140 if (UnsignedPred == ICmpInst::ICMP_UGE && EqPred == ICmpInst::ICMP_EQ &&
1141 !IsAnd && GetKnownNonZeroAndOther(B, A))
1142 return Builder.CreateICmpUGE(Builder.CreateNeg(B), A);
1143 }
1144
1145 return nullptr;
1146}
1147
1148struct IntPart {
1150 unsigned StartBit;
1151 unsigned NumBits;
1152};
1153
1154/// Match an extraction of bits from an integer.
1155static std::optional<IntPart> matchIntPart(Value *V) {
1156 Value *X;
1157 if (!match(V, m_OneUse(m_Trunc(m_Value(X)))))
1158 return std::nullopt;
1159
1160 unsigned NumOriginalBits = X->getType()->getScalarSizeInBits();
1161 unsigned NumExtractedBits = V->getType()->getScalarSizeInBits();
1162 Value *Y;
1163 const APInt *Shift;
1164 // For a trunc(lshr Y, Shift) pattern, make sure we're only extracting bits
1165 // from Y, not any shifted-in zeroes.
1166 if (match(X, m_OneUse(m_LShr(m_Value(Y), m_APInt(Shift)))) &&
1167 Shift->ule(NumOriginalBits - NumExtractedBits))
1168 return {{Y, (unsigned)Shift->getZExtValue(), NumExtractedBits}};
1169 return {{X, 0, NumExtractedBits}};
1170}
1171
1172/// Materialize an extraction of bits from an integer in IR.
1173static Value *extractIntPart(const IntPart &P, IRBuilderBase &Builder) {
1174 Value *V = P.From;
1175 if (P.StartBit)
1176 V = Builder.CreateLShr(V, P.StartBit);
1177 Type *TruncTy = V->getType()->getWithNewBitWidth(P.NumBits);
1178 if (TruncTy != V->getType())
1179 V = Builder.CreateTrunc(V, TruncTy);
1180 return V;
1181}
1182
1183/// (icmp eq X0, Y0) & (icmp eq X1, Y1) -> icmp eq X01, Y01
1184/// (icmp ne X0, Y0) | (icmp ne X1, Y1) -> icmp ne X01, Y01
1185/// where X0, X1 and Y0, Y1 are adjacent parts extracted from an integer.
1186Value *InstCombinerImpl::foldEqOfParts(Value *Cmp0, Value *Cmp1, bool IsAnd) {
1187 if (!Cmp0->hasOneUse() || !Cmp1->hasOneUse())
1188 return nullptr;
1189
1191 auto GetMatchPart = [&](Value *CmpV,
1192 unsigned OpNo) -> std::optional<IntPart> {
1193 assert(CmpV->getType()->isIntOrIntVectorTy(1) && "Must be bool");
1194
1195 Value *X, *Y;
1196 // icmp ne (and x, 1), (and y, 1) <=> trunc (xor x, y) to i1
1197 // icmp eq (and x, 1), (and y, 1) <=> not (trunc (xor x, y) to i1)
1198 if (Pred == CmpInst::ICMP_NE
1199 ? match(CmpV, m_Trunc(m_Xor(m_Value(X), m_Value(Y))))
1200 : match(CmpV, m_Not(m_Trunc(m_Xor(m_Value(X), m_Value(Y))))))
1201 return {{OpNo == 0 ? X : Y, 0, 1}};
1202
1203 auto *Cmp = dyn_cast<ICmpInst>(CmpV);
1204 if (!Cmp)
1205 return std::nullopt;
1206
1207 if (Pred == Cmp->getPredicate())
1208 return matchIntPart(Cmp->getOperand(OpNo));
1209
1210 const APInt *C;
1211 // (icmp eq (lshr x, C), (lshr y, C)) gets optimized to:
1212 // (icmp ult (xor x, y), 1 << C) so also look for that.
1213 if (Pred == CmpInst::ICMP_EQ && Cmp->getPredicate() == CmpInst::ICMP_ULT) {
1214 if (!match(Cmp->getOperand(1), m_Power2(C)) ||
1215 !match(Cmp->getOperand(0), m_Xor(m_Value(), m_Value())))
1216 return std::nullopt;
1217 }
1218
1219 // (icmp ne (lshr x, C), (lshr y, C)) gets optimized to:
1220 // (icmp ugt (xor x, y), (1 << C) - 1) so also look for that.
1221 else if (Pred == CmpInst::ICMP_NE &&
1222 Cmp->getPredicate() == CmpInst::ICMP_UGT) {
1223 if (!match(Cmp->getOperand(1), m_LowBitMask(C)) ||
1224 !match(Cmp->getOperand(0), m_Xor(m_Value(), m_Value())))
1225 return std::nullopt;
1226 } else {
1227 return std::nullopt;
1228 }
1229
1230 unsigned From = Pred == CmpInst::ICMP_NE ? C->popcount() : C->countr_zero();
1231 Instruction *I = cast<Instruction>(Cmp->getOperand(0));
1232 return {{I->getOperand(OpNo), From, C->getBitWidth() - From}};
1233 };
1234
1235 std::optional<IntPart> L0 = GetMatchPart(Cmp0, 0);
1236 std::optional<IntPart> R0 = GetMatchPart(Cmp0, 1);
1237 std::optional<IntPart> L1 = GetMatchPart(Cmp1, 0);
1238 std::optional<IntPart> R1 = GetMatchPart(Cmp1, 1);
1239 if (!L0 || !R0 || !L1 || !R1)
1240 return nullptr;
1241
1242 // Make sure the LHS/RHS compare a part of the same value, possibly after
1243 // an operand swap.
1244 if (L0->From != L1->From || R0->From != R1->From) {
1245 if (L0->From != R1->From || R0->From != L1->From)
1246 return nullptr;
1247 std::swap(L1, R1);
1248 }
1249
1250 // Make sure the extracted parts are adjacent, canonicalizing to L0/R0 being
1251 // the low part and L1/R1 being the high part.
1252 if (L0->StartBit + L0->NumBits != L1->StartBit ||
1253 R0->StartBit + R0->NumBits != R1->StartBit) {
1254 if (L1->StartBit + L1->NumBits != L0->StartBit ||
1255 R1->StartBit + R1->NumBits != R0->StartBit)
1256 return nullptr;
1257 std::swap(L0, L1);
1258 std::swap(R0, R1);
1259 }
1260
1261 // We can simplify to a comparison of these larger parts of the integers.
1262 IntPart L = {L0->From, L0->StartBit, L0->NumBits + L1->NumBits};
1263 IntPart R = {R0->From, R0->StartBit, R0->NumBits + R1->NumBits};
1266 return Builder.CreateICmp(Pred, LValue, RValue);
1267}
1268
1269/// Reduce logic-of-compares with equality to a constant by substituting a
1270/// common operand with the constant. Callers are expected to call this with
1271/// Cmp0/Cmp1 switched to handle logic op commutativity.
1273 bool IsAnd, bool IsLogical,
1274 InstCombiner::BuilderTy &Builder,
1275 const SimplifyQuery &Q) {
1276 // Match an equality compare with a non-poison constant as Cmp0.
1277 // Also, give up if the compare can be constant-folded to avoid looping.
1278 CmpPredicate Pred0;
1279 Value *X;
1280 Constant *C;
1281 if (!match(Cmp0, m_ICmp(Pred0, m_Value(X), m_Constant(C))) ||
1283 return nullptr;
1284 if ((IsAnd && Pred0 != ICmpInst::ICMP_EQ) ||
1285 (!IsAnd && Pred0 != ICmpInst::ICMP_NE))
1286 return nullptr;
1287
1288 // The other compare must include a common operand (X). Canonicalize the
1289 // common operand as operand 1 (Pred1 is swapped if the common operand was
1290 // operand 0).
1291 Value *Y;
1292 CmpPredicate Pred1;
1293 if (!match(Cmp1, m_c_ICmp(Pred1, m_Value(Y), m_Specific(X))))
1294 return nullptr;
1295
1296 // Replace variable with constant value equivalence to remove a variable use:
1297 // (X == C) && (Y Pred1 X) --> (X == C) && (Y Pred1 C)
1298 // (X != C) || (Y Pred1 X) --> (X != C) || (Y Pred1 C)
1299 // Can think of the 'or' substitution with the 'and' bool equivalent:
1300 // A || B --> A || (!A && B)
1301 Value *SubstituteCmp = simplifyICmpInst(Pred1, Y, C, Q);
1302 if (!SubstituteCmp) {
1303 // If we need to create a new instruction, require that the old compare can
1304 // be removed.
1305 if (!Cmp1->hasOneUse())
1306 return nullptr;
1307 SubstituteCmp = Builder.CreateICmp(Pred1, Y, C);
1308 }
1309 if (IsLogical)
1310 return IsAnd ? Builder.CreateLogicalAnd(Cmp0, SubstituteCmp)
1311 : Builder.CreateLogicalOr(Cmp0, SubstituteCmp);
1312 return Builder.CreateBinOp(IsAnd ? Instruction::And : Instruction::Or, Cmp0,
1313 SubstituteCmp);
1314}
1315
1316/// Fold (icmp Pred1 V1, C1) & (icmp Pred2 V2, C2)
1317/// or (icmp Pred1 V1, C1) | (icmp Pred2 V2, C2)
1318/// into a single comparison using range-based reasoning.
1319/// NOTE: This is also used for logical and/or, must be poison-safe!
1320Value *InstCombinerImpl::foldAndOrOfICmpsUsingRanges(ICmpInst *ICmp1,
1321 ICmpInst *ICmp2,
1322 bool IsAnd) {
1323 CmpPredicate Pred1, Pred2;
1324 Value *V1, *V2;
1325 const APInt *C1, *C2;
1326 if (!match(ICmp1, m_ICmp(Pred1, m_Value(V1), m_APInt(C1))) ||
1327 !match(ICmp2, m_ICmp(Pred2, m_Value(V2), m_APInt(C2))))
1328 return nullptr;
1329
1330 // Look through add of a constant offset on V1, V2, or both operands. This
1331 // allows us to interpret the V + C' < C'' range idiom into a proper range.
1332 const APInt *Offset1 = nullptr, *Offset2 = nullptr;
1333 if (V1 != V2) {
1334 Value *X;
1335 if (match(V1, m_Add(m_Value(X), m_APInt(Offset1))))
1336 V1 = X;
1337 if (match(V2, m_Add(m_Value(X), m_APInt(Offset2))))
1338 V2 = X;
1339 }
1340
1341 // Look through and with a negative power of 2 mask on V1 or V2. This
1342 // detects idioms of the form `(x == A) || ((x & Mask) == A + 1)` where A + 1
1343 // is aligned to the mask and A + 1 >= |Mask|. This pattern corresponds to a
1344 // contiguous range check, which can be folded into an addition and compare.
1345 // The same applies for `(x != A) && ((x & Mask) != A + 1)`.
1346 auto AreContiguousRangePredicates = [](CmpPredicate Pred1, CmpPredicate Pred2,
1347 bool IsAnd) {
1348 if (IsAnd)
1349 return Pred1 == ICmpInst::ICMP_NE && Pred2 == ICmpInst::ICMP_NE;
1350 return Pred1 == ICmpInst::ICMP_EQ && Pred2 == ICmpInst::ICMP_EQ;
1351 };
1352 const APInt *Mask1 = nullptr, *Mask2 = nullptr;
1353 bool MatchedAnd1 = false, MatchedAnd2 = false;
1354 if (V1 != V2 && AreContiguousRangePredicates(Pred1, Pred2, IsAnd)) {
1355 Value *X;
1356 if (match(V1, m_OneUse(m_And(m_Value(X), m_NegatedPower2(Mask1)))) &&
1357 C1->getBitWidth() == C2->getBitWidth() && *C1 == *C2 + 1 &&
1358 C1->uge(Mask1->abs()) && C1->isPowerOf2()) {
1359 MatchedAnd1 = true;
1360 V1 = X;
1361 }
1362 if (match(V2, m_OneUse(m_And(m_Value(X), m_NegatedPower2(Mask2)))) &&
1363 C1->getBitWidth() == C2->getBitWidth() && *C2 == *C1 + 1 &&
1364 C2->uge(Mask2->abs()) && C2->isPowerOf2()) {
1365 MatchedAnd2 = true;
1366 V2 = X;
1367 }
1368 }
1369
1370 if (V1 != V2)
1371 return nullptr;
1372
1373 ConstantRange CR1 =
1374 MatchedAnd1
1375 ? ConstantRange(*C1, *C1 - *Mask1)
1376 : ConstantRange::makeExactICmpRegion(
1377 IsAnd ? ICmpInst::getInverseCmpPredicate(Pred1) : Pred1, *C1);
1378 if (Offset1)
1379 CR1 = CR1.subtract(*Offset1);
1380
1381 ConstantRange CR2 =
1382 MatchedAnd2
1383 ? ConstantRange(*C2, *C2 - *Mask2)
1385 IsAnd ? ICmpInst::getInverseCmpPredicate(Pred2) : Pred2, *C2);
1386 if (Offset2)
1387 CR2 = CR2.subtract(*Offset2);
1388
1389 Type *Ty = V1->getType();
1390 Value *NewV = V1;
1391 std::optional<ConstantRange> CR = CR1.exactUnionWith(CR2);
1392 if (!CR) {
1393 if (!(ICmp1->hasOneUse() && ICmp2->hasOneUse()) || CR1.isWrappedSet() ||
1394 CR2.isWrappedSet())
1395 return nullptr;
1396
1397 // Check whether we have equal-size ranges that only differ by one bit.
1398 // In that case we can apply a mask to map one range onto the other.
1399 APInt LowerDiff = CR1.getLower() ^ CR2.getLower();
1400 APInt UpperDiff = (CR1.getUpper() - 1) ^ (CR2.getUpper() - 1);
1401 APInt CR1Size = CR1.getUpper() - CR1.getLower();
1402 if (!LowerDiff.isPowerOf2() || LowerDiff != UpperDiff ||
1403 CR1Size != CR2.getUpper() - CR2.getLower())
1404 return nullptr;
1405
1406 CR = CR1.getLower().ult(CR2.getLower()) ? CR1 : CR2;
1407 NewV = Builder.CreateAnd(NewV, ConstantInt::get(Ty, ~LowerDiff));
1408 }
1409
1410 if (IsAnd)
1411 CR = CR->inverse();
1412
1413 CmpInst::Predicate NewPred;
1414 APInt NewC, Offset;
1415 CR->getEquivalentICmp(NewPred, NewC, Offset);
1416
1417 if (Offset != 0)
1418 NewV = Builder.CreateAdd(NewV, ConstantInt::get(Ty, Offset));
1419 return Builder.CreateICmp(NewPred, NewV, ConstantInt::get(Ty, NewC));
1420}
1421
1422/// Ignore all operations which only change the sign of a value, returning the
1423/// underlying magnitude value.
1425 match(Val, m_FNeg(m_Value(Val)));
1426 match(Val, m_FAbs(m_Value(Val)));
1427 match(Val, m_CopySign(m_Value(Val), m_Value()));
1428 return Val;
1429}
1430
1431/// Matches canonical form of isnan, fcmp ord x, 0
1435
1436/// Matches fcmp u__ x, +/-inf
1441
1442/// and (fcmp ord x, 0), (fcmp u* x, inf) -> fcmp o* x, inf
1443///
1444/// Clang emits this pattern for doing an isfinite check in __builtin_isnormal.
1446 FCmpInst *RHS) {
1447 Value *LHS0 = LHS->getOperand(0), *LHS1 = LHS->getOperand(1);
1448 Value *RHS0 = RHS->getOperand(0), *RHS1 = RHS->getOperand(1);
1449 FCmpInst::Predicate PredL = LHS->getPredicate(), PredR = RHS->getPredicate();
1450
1451 if (!matchIsNotNaN(PredL, LHS0, LHS1) ||
1452 !matchUnorderedInfCompare(PredR, RHS0, RHS1))
1453 return nullptr;
1454
1455 return Builder.CreateFCmpFMF(FCmpInst::getOrderedPredicate(PredR), RHS0, RHS1,
1457}
1458
1459Value *InstCombinerImpl::foldLogicOfFCmps(FCmpInst *LHS, FCmpInst *RHS,
1460 bool IsAnd, bool IsLogicalSelect) {
1461 Value *LHS0 = LHS->getOperand(0), *LHS1 = LHS->getOperand(1);
1462 Value *RHS0 = RHS->getOperand(0), *RHS1 = RHS->getOperand(1);
1463 FCmpInst::Predicate PredL = LHS->getPredicate(), PredR = RHS->getPredicate();
1464
1465 if (LHS0 == RHS1 && RHS0 == LHS1) {
1466 // Swap RHS operands to match LHS.
1467 PredR = FCmpInst::getSwappedPredicate(PredR);
1468 std::swap(RHS0, RHS1);
1469 }
1470
1471 // Simplify (fcmp cc0 x, y) & (fcmp cc1 x, y).
1472 // Suppose the relation between x and y is R, where R is one of
1473 // U(1000), L(0100), G(0010) or E(0001), and CC0 and CC1 are the bitmasks for
1474 // testing the desired relations.
1475 //
1476 // Since (R & CC0) and (R & CC1) are either R or 0, we actually have this:
1477 // bool(R & CC0) && bool(R & CC1)
1478 // = bool((R & CC0) & (R & CC1))
1479 // = bool(R & (CC0 & CC1)) <= by re-association, commutation, and idempotency
1480 //
1481 // Since (R & CC0) and (R & CC1) are either R or 0, we actually have this:
1482 // bool(R & CC0) || bool(R & CC1)
1483 // = bool((R & CC0) | (R & CC1))
1484 // = bool(R & (CC0 | CC1)) <= by reversed distribution (contribution? ;)
1485 if (LHS0 == RHS0 && LHS1 == RHS1) {
1486 unsigned FCmpCodeL = getFCmpCode(PredL);
1487 unsigned FCmpCodeR = getFCmpCode(PredR);
1488 unsigned NewPred = IsAnd ? FCmpCodeL & FCmpCodeR : FCmpCodeL | FCmpCodeR;
1489
1490 // Intersect the fast math flags.
1491 // TODO: We can union the fast math flags unless this is a logical select.
1492 return getFCmpValue(NewPred, LHS0, LHS1, Builder,
1494 }
1495
1496 // This transform is not valid for a logical select.
1497 if (!IsLogicalSelect &&
1498 ((PredL == FCmpInst::FCMP_ORD && PredR == FCmpInst::FCMP_ORD && IsAnd) ||
1499 (PredL == FCmpInst::FCMP_UNO && PredR == FCmpInst::FCMP_UNO &&
1500 !IsAnd))) {
1501 if (LHS0->getType() != RHS0->getType())
1502 return nullptr;
1503
1504 // FCmp canonicalization ensures that (fcmp ord/uno X, X) and
1505 // (fcmp ord/uno X, C) will be transformed to (fcmp X, +0.0).
1506 if (match(LHS1, m_PosZeroFP()) && match(RHS1, m_PosZeroFP())) {
1507 // Ignore the constants because they are obviously not NANs:
1508 // (fcmp ord x, 0.0) & (fcmp ord y, 0.0) -> (fcmp ord x, y)
1509 // (fcmp uno x, 0.0) | (fcmp uno y, 0.0) -> (fcmp uno x, y)
1510 return Builder.CreateFCmpFMF(PredL, LHS0, RHS0,
1512 }
1513 }
1514
1515 // This transform is not valid for a logical select.
1516 if (!IsLogicalSelect && IsAnd &&
1517 stripSignOnlyFPOps(LHS0) == stripSignOnlyFPOps(RHS0)) {
1518 // and (fcmp ord x, 0), (fcmp u* x, inf) -> fcmp o* x, inf
1519 // and (fcmp ord x, 0), (fcmp u* fabs(x), inf) -> fcmp o* x, inf
1521 return Left;
1523 return Right;
1524 }
1525
1526 // Turn at least two fcmps with constants into llvm.is.fpclass.
1527 //
1528 // If we can represent a combined value test with one class call, we can
1529 // potentially eliminate 4-6 instructions. If we can represent a test with a
1530 // single fcmp with fneg and fabs, that's likely a better canonical form.
1531 if (LHS->hasOneUse() && RHS->hasOneUse()) {
1532 auto [ClassValRHS, ClassMaskRHS] =
1533 fcmpToClassTest(PredR, *RHS->getFunction(), RHS0, RHS1);
1534 if (ClassValRHS) {
1535 auto [ClassValLHS, ClassMaskLHS] =
1536 fcmpToClassTest(PredL, *LHS->getFunction(), LHS0, LHS1);
1537 if (ClassValLHS == ClassValRHS) {
1538 unsigned CombinedMask = IsAnd ? (ClassMaskLHS & ClassMaskRHS)
1539 : (ClassMaskLHS | ClassMaskRHS);
1540 return Builder.CreateIntrinsic(
1541 Intrinsic::is_fpclass, {ClassValLHS->getType()},
1542 {ClassValLHS, Builder.getInt32(CombinedMask)});
1543 }
1544 }
1545 }
1546
1547 // Canonicalize the range check idiom:
1548 // and (fcmp olt/ole/ult/ule x, C), (fcmp ogt/oge/ugt/uge x, -C)
1549 // --> fabs(x) olt/ole/ult/ule C
1550 // or (fcmp ogt/oge/ugt/uge x, C), (fcmp olt/ole/ult/ule x, -C)
1551 // --> fabs(x) ogt/oge/ugt/uge C
1552 // TODO: Generalize to handle a negated variable operand?
1553 const APFloat *LHSC, *RHSC;
1554 if (LHS0 == RHS0 && LHS->hasOneUse() && RHS->hasOneUse() &&
1555 FCmpInst::getSwappedPredicate(PredL) == PredR &&
1556 match(LHS1, m_APFloatAllowPoison(LHSC)) &&
1557 match(RHS1, m_APFloatAllowPoison(RHSC)) &&
1558 LHSC->bitwiseIsEqual(neg(*RHSC))) {
1559 auto IsLessThanOrLessEqual = [](FCmpInst::Predicate Pred) {
1560 switch (Pred) {
1561 case FCmpInst::FCMP_OLT:
1562 case FCmpInst::FCMP_OLE:
1563 case FCmpInst::FCMP_ULT:
1564 case FCmpInst::FCMP_ULE:
1565 return true;
1566 default:
1567 return false;
1568 }
1569 };
1570 if (IsLessThanOrLessEqual(IsAnd ? PredR : PredL)) {
1571 std::swap(LHSC, RHSC);
1572 std::swap(PredL, PredR);
1573 }
1574 if (IsLessThanOrLessEqual(IsAnd ? PredL : PredR)) {
1575 FastMathFlags NewFlag = LHS->getFastMathFlags();
1576 if (!IsLogicalSelect)
1577 NewFlag |= RHS->getFastMathFlags();
1578
1579 Value *FAbs =
1580 Builder.CreateUnaryIntrinsic(Intrinsic::fabs, LHS0, NewFlag);
1581 return Builder.CreateFCmpFMF(
1582 PredL, FAbs, ConstantFP::get(LHS0->getType(), *LHSC), NewFlag);
1583 }
1584 }
1585
1586 return nullptr;
1587}
1588
1589/// Match an fcmp against a special value that performs a test possible by
1590/// llvm.is.fpclass.
1591static bool matchIsFPClassLikeFCmp(Value *Op, Value *&ClassVal,
1592 uint64_t &ClassMask) {
1593 auto *FCmp = dyn_cast<FCmpInst>(Op);
1594 if (!FCmp || !FCmp->hasOneUse())
1595 return false;
1596
1597 std::tie(ClassVal, ClassMask) =
1598 fcmpToClassTest(FCmp->getPredicate(), *FCmp->getParent()->getParent(),
1599 FCmp->getOperand(0), FCmp->getOperand(1));
1600 return ClassVal != nullptr;
1601}
1602
1603/// or (is_fpclass x, mask0), (is_fpclass x, mask1)
1604/// -> is_fpclass x, (mask0 | mask1)
1605/// and (is_fpclass x, mask0), (is_fpclass x, mask1)
1606/// -> is_fpclass x, (mask0 & mask1)
1607/// xor (is_fpclass x, mask0), (is_fpclass x, mask1)
1608/// -> is_fpclass x, (mask0 ^ mask1)
1609Instruction *InstCombinerImpl::foldLogicOfIsFPClass(BinaryOperator &BO,
1610 Value *Op0, Value *Op1) {
1611 Value *ClassVal0 = nullptr;
1612 Value *ClassVal1 = nullptr;
1613 uint64_t ClassMask0, ClassMask1;
1614
1615 // Restrict to folding one fcmp into one is.fpclass for now, don't introduce a
1616 // new class.
1617 //
1618 // TODO: Support forming is.fpclass out of 2 separate fcmps when codegen is
1619 // better.
1620
1621 bool IsLHSClass =
1623 m_Value(ClassVal0), m_ConstantInt(ClassMask0))));
1624 bool IsRHSClass =
1626 m_Value(ClassVal1), m_ConstantInt(ClassMask1))));
1627 if ((((IsLHSClass || matchIsFPClassLikeFCmp(Op0, ClassVal0, ClassMask0)) &&
1628 (IsRHSClass || matchIsFPClassLikeFCmp(Op1, ClassVal1, ClassMask1)))) &&
1629 ClassVal0 == ClassVal1) {
1630 unsigned NewClassMask;
1631 switch (BO.getOpcode()) {
1632 case Instruction::And:
1633 NewClassMask = ClassMask0 & ClassMask1;
1634 break;
1635 case Instruction::Or:
1636 NewClassMask = ClassMask0 | ClassMask1;
1637 break;
1638 case Instruction::Xor:
1639 NewClassMask = ClassMask0 ^ ClassMask1;
1640 break;
1641 default:
1642 llvm_unreachable("not a binary logic operator");
1643 }
1644
1645 if (IsLHSClass) {
1646 auto *II = cast<IntrinsicInst>(Op0);
1647 II->setArgOperand(
1648 1, ConstantInt::get(II->getArgOperand(1)->getType(), NewClassMask));
1649 return replaceInstUsesWith(BO, II);
1650 }
1651
1652 if (IsRHSClass) {
1653 auto *II = cast<IntrinsicInst>(Op1);
1654 II->setArgOperand(
1655 1, ConstantInt::get(II->getArgOperand(1)->getType(), NewClassMask));
1656 return replaceInstUsesWith(BO, II);
1657 }
1658
1659 CallInst *NewClass =
1660 Builder.CreateIntrinsic(Intrinsic::is_fpclass, {ClassVal0->getType()},
1661 {ClassVal0, Builder.getInt32(NewClassMask)});
1662 return replaceInstUsesWith(BO, NewClass);
1663 }
1664
1665 return nullptr;
1666}
1667
1668/// Look for the pattern that conditionally negates a value via math operations:
1669/// cond.splat = sext i1 cond
1670/// sub = add cond.splat, x
1671/// xor = xor sub, cond.splat
1672/// and rewrite it to do the same, but via logical operations:
1673/// value.neg = sub 0, value
1674/// cond = select i1 neg, value.neg, value
1675Instruction *InstCombinerImpl::canonicalizeConditionalNegationViaMathToSelect(
1676 BinaryOperator &I) {
1677 assert(I.getOpcode() == BinaryOperator::Xor && "Only for xor!");
1678 Value *Cond, *X;
1679 // As per complexity ordering, `xor` is not commutative here.
1680 if (!match(&I, m_c_BinOp(m_OneUse(m_Value()), m_Value())) ||
1681 !match(I.getOperand(1), m_SExt(m_Value(Cond))) ||
1682 !Cond->getType()->isIntOrIntVectorTy(1) ||
1683 !match(I.getOperand(0), m_c_Add(m_SExt(m_Specific(Cond)), m_Value(X))))
1684 return nullptr;
1685 return SelectInst::Create(Cond, Builder.CreateNeg(X, X->getName() + ".neg"),
1686 X);
1687}
1688
1689/// This a limited reassociation for a special case (see above) where we are
1690/// checking if two values are either both NAN (unordered) or not-NAN (ordered).
1691/// This could be handled more generally in '-reassociation', but it seems like
1692/// an unlikely pattern for a large number of logic ops and fcmps.
1694 InstCombiner::BuilderTy &Builder) {
1695 Instruction::BinaryOps Opcode = BO.getOpcode();
1696 assert((Opcode == Instruction::And || Opcode == Instruction::Or) &&
1697 "Expecting and/or op for fcmp transform");
1698
1699 // There are 4 commuted variants of the pattern. Canonicalize operands of this
1700 // logic op so an fcmp is operand 0 and a matching logic op is operand 1.
1701 Value *Op0 = BO.getOperand(0), *Op1 = BO.getOperand(1), *X;
1702 if (match(Op1, m_FCmp(m_Value(), m_AnyZeroFP())))
1703 std::swap(Op0, Op1);
1704
1705 // Match inner binop and the predicate for combining 2 NAN checks into 1.
1706 Value *BO10, *BO11;
1707 FCmpInst::Predicate NanPred = Opcode == Instruction::And ? FCmpInst::FCMP_ORD
1709 if (!match(Op0, m_SpecificFCmp(NanPred, m_Value(X), m_AnyZeroFP())) ||
1710 !match(Op1, m_BinOp(Opcode, m_Value(BO10), m_Value(BO11))))
1711 return nullptr;
1712
1713 // The inner logic op must have a matching fcmp operand.
1714 Value *Y;
1715 if (!match(BO10, m_SpecificFCmp(NanPred, m_Value(Y), m_AnyZeroFP())) ||
1716 X->getType() != Y->getType())
1717 std::swap(BO10, BO11);
1718
1719 if (!match(BO10, m_SpecificFCmp(NanPred, m_Value(Y), m_AnyZeroFP())) ||
1720 X->getType() != Y->getType())
1721 return nullptr;
1722
1723 // and (fcmp ord X, 0), (and (fcmp ord Y, 0), Z) --> and (fcmp ord X, Y), Z
1724 // or (fcmp uno X, 0), (or (fcmp uno Y, 0), Z) --> or (fcmp uno X, Y), Z
1725 // Intersect FMF from the 2 source fcmps.
1726 Value *NewFCmp =
1727 Builder.CreateFCmpFMF(NanPred, X, Y, FMFSource::intersect(Op0, BO10));
1728 return BinaryOperator::Create(Opcode, NewFCmp, BO11);
1729}
1730
1731/// Match variations of De Morgan's Laws:
1732/// (~A & ~B) == (~(A | B))
1733/// (~A | ~B) == (~(A & B))
1735 InstCombiner &IC) {
1736 const Instruction::BinaryOps Opcode = I.getOpcode();
1737 assert((Opcode == Instruction::And || Opcode == Instruction::Or) &&
1738 "Trying to match De Morgan's Laws with something other than and/or");
1739
1740 // Flip the logic operation.
1741 const Instruction::BinaryOps FlippedOpcode =
1742 (Opcode == Instruction::And) ? Instruction::Or : Instruction::And;
1743
1744 Value *Op0 = I.getOperand(0), *Op1 = I.getOperand(1);
1745 Value *A, *B;
1746 if (match(Op0, m_OneUse(m_Not(m_Value(A)))) &&
1747 match(Op1, m_OneUse(m_Not(m_Value(B)))) &&
1748 !IC.isFreeToInvert(A, A->hasOneUse()) &&
1749 !IC.isFreeToInvert(B, B->hasOneUse())) {
1750 Value *AndOr =
1751 IC.Builder.CreateBinOp(FlippedOpcode, A, B, I.getName() + ".demorgan");
1752 return BinaryOperator::CreateNot(AndOr);
1753 }
1754
1755 // The 'not' ops may require reassociation.
1756 // (A & ~B) & ~C --> A & ~(B | C)
1757 // (~B & A) & ~C --> A & ~(B | C)
1758 // (A | ~B) | ~C --> A | ~(B & C)
1759 // (~B | A) | ~C --> A | ~(B & C)
1760 Value *C;
1761 if (match(Op0, m_OneUse(m_c_BinOp(Opcode, m_Value(A), m_Not(m_Value(B))))) &&
1762 match(Op1, m_Not(m_Value(C)))) {
1763 Value *FlippedBO = IC.Builder.CreateBinOp(FlippedOpcode, B, C);
1764 return BinaryOperator::Create(Opcode, A, IC.Builder.CreateNot(FlippedBO));
1765 }
1766
1767 return nullptr;
1768}
1769
1770bool InstCombinerImpl::shouldOptimizeCast(CastInst *CI) {
1771 Value *CastSrc = CI->getOperand(0);
1772
1773 // Noop casts and casts of constants should be eliminated trivially.
1774 if (CI->getSrcTy() == CI->getDestTy() || isa<Constant>(CastSrc))
1775 return false;
1776
1777 // If this cast is paired with another cast that can be eliminated, we prefer
1778 // to have it eliminated.
1779 if (const auto *PrecedingCI = dyn_cast<CastInst>(CastSrc))
1780 if (isEliminableCastPair(PrecedingCI, CI))
1781 return false;
1782
1783 return true;
1784}
1785
1786/// Fold {and,or,xor} (cast X), C.
1788 InstCombinerImpl &IC) {
1790 if (!C)
1791 return nullptr;
1792
1793 auto LogicOpc = Logic.getOpcode();
1794 Type *DestTy = Logic.getType();
1795 Type *SrcTy = Cast->getSrcTy();
1796
1797 // Move the logic operation ahead of a zext or sext if the constant is
1798 // unchanged in the smaller source type. Performing the logic in a smaller
1799 // type may provide more information to later folds, and the smaller logic
1800 // instruction may be cheaper (particularly in the case of vectors).
1801 Value *X;
1802 auto &DL = IC.getDataLayout();
1803 if (match(Cast, m_OneUse(m_ZExt(m_Value(X))))) {
1804 PreservedCastFlags Flags;
1805 if (Constant *TruncC = getLosslessUnsignedTrunc(C, SrcTy, DL, &Flags)) {
1806 // LogicOpc (zext X), C --> zext (LogicOpc X, C)
1807 Value *NewOp = IC.Builder.CreateBinOp(LogicOpc, X, TruncC);
1808 auto *ZExt = new ZExtInst(NewOp, DestTy);
1809 ZExt->setNonNeg(Flags.NNeg);
1810 ZExt->andIRFlags(Cast);
1811 return ZExt;
1812 }
1813 }
1814
1815 if (match(Cast, m_OneUse(m_SExtLike(m_Value(X))))) {
1816 if (Constant *TruncC = getLosslessSignedTrunc(C, SrcTy, DL)) {
1817 // LogicOpc (sext X), C --> sext (LogicOpc X, C)
1818 Value *NewOp = IC.Builder.CreateBinOp(LogicOpc, X, TruncC);
1819 return new SExtInst(NewOp, DestTy);
1820 }
1821 }
1822
1823 return nullptr;
1824}
1825
1826/// Fold {and,or,xor} (cast X), Y.
1827Instruction *InstCombinerImpl::foldCastedBitwiseLogic(BinaryOperator &I) {
1828 auto LogicOpc = I.getOpcode();
1829 assert(I.isBitwiseLogicOp() && "Unexpected opcode for bitwise logic folding");
1830
1831 Value *Op0 = I.getOperand(0), *Op1 = I.getOperand(1);
1832
1833 // fold bitwise(A >> BW - 1, zext(icmp)) (BW is the scalar bits of the
1834 // type of A)
1835 // -> bitwise(zext(A < 0), zext(icmp))
1836 // -> zext(bitwise(A < 0, icmp))
1837 auto FoldBitwiseICmpZeroWithICmp = [&](Value *Op0,
1838 Value *Op1) -> Instruction * {
1839 Value *A;
1840 bool IsMatched =
1841 match(Op0,
1843 m_Value(A),
1844 m_SpecificInt(Op0->getType()->getScalarSizeInBits() - 1)))) &&
1845 match(Op1, m_OneUse(m_ZExt(m_ICmp(m_Value(), m_Value()))));
1846
1847 if (!IsMatched)
1848 return nullptr;
1849
1850 auto *ICmpL =
1851 Builder.CreateICmpSLT(A, Constant::getNullValue(A->getType()));
1852 auto *ICmpR = cast<ZExtInst>(Op1)->getOperand(0);
1853 auto *BitwiseOp = Builder.CreateBinOp(LogicOpc, ICmpL, ICmpR);
1854
1855 return new ZExtInst(BitwiseOp, Op0->getType());
1856 };
1857
1858 if (auto *Ret = FoldBitwiseICmpZeroWithICmp(Op0, Op1))
1859 return Ret;
1860
1861 if (auto *Ret = FoldBitwiseICmpZeroWithICmp(Op1, Op0))
1862 return Ret;
1863
1864 CastInst *Cast0 = dyn_cast<CastInst>(Op0);
1865 if (!Cast0)
1866 return nullptr;
1867
1868 // This must be a cast from an integer or integer vector source type to allow
1869 // transformation of the logic operation to the source type.
1870 Type *DestTy = I.getType();
1871 Type *SrcTy = Cast0->getSrcTy();
1872 if (!SrcTy->isIntOrIntVectorTy())
1873 return nullptr;
1874
1875 if (Instruction *Ret = foldLogicCastConstant(I, Cast0, *this))
1876 return Ret;
1877
1878 CastInst *Cast1 = dyn_cast<CastInst>(Op1);
1879 if (!Cast1)
1880 return nullptr;
1881
1882 // Both operands of the logic operation are casts. The casts must be the
1883 // same kind for reduction.
1884 Instruction::CastOps CastOpcode = Cast0->getOpcode();
1885 if (CastOpcode != Cast1->getOpcode())
1886 return nullptr;
1887
1888 // Can't fold it profitably if no one of casts has one use.
1889 if (!Cast0->hasOneUse() && !Cast1->hasOneUse())
1890 return nullptr;
1891
1892 Value *X, *Y;
1893 if (match(Cast0, m_ZExtOrSExt(m_Value(X))) &&
1894 match(Cast1, m_ZExtOrSExt(m_Value(Y)))) {
1895 // Cast the narrower source to the wider source type.
1896 unsigned XNumBits = X->getType()->getScalarSizeInBits();
1897 unsigned YNumBits = Y->getType()->getScalarSizeInBits();
1898 if (XNumBits != YNumBits) {
1899 // Cast the narrower source to the wider source type only if both of casts
1900 // have one use to avoid creating an extra instruction.
1901 if (!Cast0->hasOneUse() || !Cast1->hasOneUse())
1902 return nullptr;
1903
1904 // If the source types do not match, but the casts are matching extends,
1905 // we can still narrow the logic op.
1906 if (XNumBits < YNumBits) {
1907 X = Builder.CreateCast(CastOpcode, X, Y->getType());
1908 } else if (YNumBits < XNumBits) {
1909 Y = Builder.CreateCast(CastOpcode, Y, X->getType());
1910 }
1911 }
1912
1913 // Do the logic op in the intermediate width, then widen more.
1914 Value *NarrowLogic = Builder.CreateBinOp(LogicOpc, X, Y, I.getName());
1915 auto *Disjoint = dyn_cast<PossiblyDisjointInst>(&I);
1916 auto *NewDisjoint = dyn_cast<PossiblyDisjointInst>(NarrowLogic);
1917 if (Disjoint && NewDisjoint)
1918 NewDisjoint->setIsDisjoint(Disjoint->isDisjoint());
1919 return CastInst::Create(CastOpcode, NarrowLogic, DestTy);
1920 }
1921
1922 // If the src type of casts are different, give up for other cast opcodes.
1923 if (SrcTy != Cast1->getSrcTy())
1924 return nullptr;
1925
1926 Value *Cast0Src = Cast0->getOperand(0);
1927 Value *Cast1Src = Cast1->getOperand(0);
1928
1929 // fold logic(cast(A), cast(B)) -> cast(logic(A, B))
1930 if (shouldOptimizeCast(Cast0) && shouldOptimizeCast(Cast1)) {
1931 Value *NewOp = Builder.CreateBinOp(LogicOpc, Cast0Src, Cast1Src,
1932 I.getName());
1933 return CastInst::Create(CastOpcode, NewOp, DestTy);
1934 }
1935
1936 return nullptr;
1937}
1938
1940 InstCombiner::BuilderTy &Builder) {
1941 assert(I.getOpcode() == Instruction::And);
1942 Value *Op0 = I.getOperand(0);
1943 Value *Op1 = I.getOperand(1);
1944 Value *A, *B;
1945
1946 // Operand complexity canonicalization guarantees that the 'or' is Op0.
1947 // (A | B) & ~(A & B) --> A ^ B
1948 // (A | B) & ~(B & A) --> A ^ B
1949 if (match(&I, m_BinOp(m_Or(m_Value(A), m_Value(B)),
1951 return BinaryOperator::CreateXor(A, B);
1952
1953 // (A | ~B) & (~A | B) --> ~(A ^ B)
1954 // (A | ~B) & (B | ~A) --> ~(A ^ B)
1955 // (~B | A) & (~A | B) --> ~(A ^ B)
1956 // (~B | A) & (B | ~A) --> ~(A ^ B)
1957 if (Op0->hasOneUse() || Op1->hasOneUse())
1960 return BinaryOperator::CreateNot(Builder.CreateXor(A, B));
1961
1962 return nullptr;
1963}
1964
1966 InstCombiner::BuilderTy &Builder) {
1967 assert(I.getOpcode() == Instruction::Or);
1968 Value *Op0 = I.getOperand(0);
1969 Value *Op1 = I.getOperand(1);
1970 Value *A, *B;
1971
1972 // Operand complexity canonicalization guarantees that the 'and' is Op0.
1973 // (A & B) | ~(A | B) --> ~(A ^ B)
1974 // (A & B) | ~(B | A) --> ~(A ^ B)
1975 if (Op0->hasOneUse() || Op1->hasOneUse())
1976 if (match(Op0, m_And(m_Value(A), m_Value(B))) &&
1978 return BinaryOperator::CreateNot(Builder.CreateXor(A, B));
1979
1980 // Operand complexity canonicalization guarantees that the 'xor' is Op0.
1981 // (A ^ B) | ~(A | B) --> ~(A & B)
1982 // (A ^ B) | ~(B | A) --> ~(A & B)
1983 if (Op0->hasOneUse() || Op1->hasOneUse())
1984 if (match(Op0, m_Xor(m_Value(A), m_Value(B))) &&
1986 return BinaryOperator::CreateNot(Builder.CreateAnd(A, B));
1987
1988 // (A & ~B) | (~A & B) --> A ^ B
1989 // (A & ~B) | (B & ~A) --> A ^ B
1990 // (~B & A) | (~A & B) --> A ^ B
1991 // (~B & A) | (B & ~A) --> A ^ B
1992 if (match(Op0, m_c_And(m_Value(A), m_Not(m_Value(B)))) &&
1994 return BinaryOperator::CreateXor(A, B);
1995
1996 return nullptr;
1997}
1998
1999/// Return true if a constant shift amount is always less than the specified
2000/// bit-width. If not, the shift could create poison in the narrower type.
2001static bool canNarrowShiftAmt(Constant *C, unsigned BitWidth) {
2002 APInt Threshold(C->getType()->getScalarSizeInBits(), BitWidth);
2003 return match(C, m_SpecificInt_ICMP(ICmpInst::ICMP_ULT, Threshold));
2004}
2005
2006/// Try to use narrower ops (sink zext ops) for an 'and' with binop operand and
2007/// a common zext operand: and (binop (zext X), C), (zext X).
2008Instruction *InstCombinerImpl::narrowMaskedBinOp(BinaryOperator &And) {
2009 // This transform could also apply to {or, and, xor}, but there are better
2010 // folds for those cases, so we don't expect those patterns here. AShr is not
2011 // handled because it should always be transformed to LShr in this sequence.
2012 // The subtract transform is different because it has a constant on the left.
2013 // Add/mul commute the constant to RHS; sub with constant RHS becomes add.
2014 Value *Op0 = And.getOperand(0), *Op1 = And.getOperand(1);
2015 Constant *C;
2016 if (!match(Op0, m_OneUse(m_Add(m_Specific(Op1), m_Constant(C)))) &&
2017 !match(Op0, m_OneUse(m_Mul(m_Specific(Op1), m_Constant(C)))) &&
2018 !match(Op0, m_OneUse(m_LShr(m_Specific(Op1), m_Constant(C)))) &&
2019 !match(Op0, m_OneUse(m_Shl(m_Specific(Op1), m_Constant(C)))) &&
2020 !match(Op0, m_OneUse(m_Sub(m_Constant(C), m_Specific(Op1)))))
2021 return nullptr;
2022
2023 Value *X;
2024 if (!match(Op1, m_ZExt(m_Value(X))) || Op1->hasNUsesOrMore(3))
2025 return nullptr;
2026
2027 Type *Ty = And.getType();
2028 if (!isa<VectorType>(Ty) && !shouldChangeType(Ty, X->getType()))
2029 return nullptr;
2030
2031 // If we're narrowing a shift, the shift amount must be safe (less than the
2032 // width) in the narrower type. If the shift amount is greater, instsimplify
2033 // usually handles that case, but we can't guarantee/assert it.
2035 if (Opc == Instruction::LShr || Opc == Instruction::Shl)
2036 if (!canNarrowShiftAmt(C, X->getType()->getScalarSizeInBits()))
2037 return nullptr;
2038
2039 // and (sub C, (zext X)), (zext X) --> zext (and (sub C', X), X)
2040 // and (binop (zext X), C), (zext X) --> zext (and (binop X, C'), X)
2041 Value *NewC = ConstantExpr::getTrunc(C, X->getType());
2042 Value *NewBO = Opc == Instruction::Sub ? Builder.CreateBinOp(Opc, NewC, X)
2043 : Builder.CreateBinOp(Opc, X, NewC);
2044 return new ZExtInst(Builder.CreateAnd(NewBO, X), Ty);
2045}
2046
2047/// Try folding relatively complex patterns for both And and Or operations
2048/// with all And and Or swapped.
2050 InstCombiner::BuilderTy &Builder) {
2051 const Instruction::BinaryOps Opcode = I.getOpcode();
2052 assert(Opcode == Instruction::And || Opcode == Instruction::Or);
2053
2054 // Flip the logic operation.
2055 const Instruction::BinaryOps FlippedOpcode =
2056 (Opcode == Instruction::And) ? Instruction::Or : Instruction::And;
2057
2058 Value *Op0 = I.getOperand(0), *Op1 = I.getOperand(1);
2059 Value *A, *B, *C, *X, *Y, *Dummy;
2060
2061 // Match following expressions:
2062 // (~(A | B) & C)
2063 // (~(A & B) | C)
2064 // Captures X = ~(A | B) or ~(A & B)
2065 const auto matchNotOrAnd =
2066 [Opcode, FlippedOpcode](Value *Op, auto m_A, auto m_B, auto m_C,
2067 Value *&X, bool CountUses = false) -> bool {
2068 if (CountUses && !Op->hasOneUse())
2069 return false;
2070
2071 if (match(Op,
2072 m_c_BinOp(FlippedOpcode,
2073 m_Value(X, m_Not(m_c_BinOp(Opcode, m_A, m_B))), m_C)))
2074 return !CountUses || X->hasOneUse();
2075
2076 return false;
2077 };
2078
2079 // (~(A | B) & C) | ... --> ...
2080 // (~(A & B) | C) & ... --> ...
2081 // TODO: One use checks are conservative. We just need to check that a total
2082 // number of multiple used values does not exceed reduction
2083 // in operations.
2084 if (matchNotOrAnd(Op0, m_Value(A), m_Value(B), m_Value(C), X)) {
2085 // (~(A | B) & C) | (~(A | C) & B) --> (B ^ C) & ~A
2086 // (~(A & B) | C) & (~(A & C) | B) --> ~((B ^ C) & A)
2087 if (matchNotOrAnd(Op1, m_Specific(A), m_Specific(C), m_Specific(B), Dummy,
2088 true)) {
2089 Value *Xor = Builder.CreateXor(B, C);
2090 return (Opcode == Instruction::Or)
2091 ? BinaryOperator::CreateAnd(Xor, Builder.CreateNot(A))
2092 : BinaryOperator::CreateNot(Builder.CreateAnd(Xor, A));
2093 }
2094
2095 // (~(A | B) & C) | (~(B | C) & A) --> (A ^ C) & ~B
2096 // (~(A & B) | C) & (~(B & C) | A) --> ~((A ^ C) & B)
2097 if (matchNotOrAnd(Op1, m_Specific(B), m_Specific(C), m_Specific(A), Dummy,
2098 true)) {
2099 Value *Xor = Builder.CreateXor(A, C);
2100 return (Opcode == Instruction::Or)
2101 ? BinaryOperator::CreateAnd(Xor, Builder.CreateNot(B))
2102 : BinaryOperator::CreateNot(Builder.CreateAnd(Xor, B));
2103 }
2104
2105 // (~(A | B) & C) | ~(A | C) --> ~((B & C) | A)
2106 // (~(A & B) | C) & ~(A & C) --> ~((B | C) & A)
2107 if (match(Op1, m_OneUse(m_Not(m_OneUse(
2108 m_c_BinOp(Opcode, m_Specific(A), m_Specific(C)))))))
2109 return BinaryOperator::CreateNot(Builder.CreateBinOp(
2110 Opcode, Builder.CreateBinOp(FlippedOpcode, B, C), A));
2111
2112 // (~(A | B) & C) | ~(B | C) --> ~((A & C) | B)
2113 // (~(A & B) | C) & ~(B & C) --> ~((A | C) & B)
2114 if (match(Op1, m_OneUse(m_Not(m_OneUse(
2115 m_c_BinOp(Opcode, m_Specific(B), m_Specific(C)))))))
2116 return BinaryOperator::CreateNot(Builder.CreateBinOp(
2117 Opcode, Builder.CreateBinOp(FlippedOpcode, A, C), B));
2118
2119 // (~(A | B) & C) | ~(C | (A ^ B)) --> ~((A | B) & (C | (A ^ B)))
2120 // Note, the pattern with swapped and/or is not handled because the
2121 // result is more undefined than a source:
2122 // (~(A & B) | C) & ~(C & (A ^ B)) --> (A ^ B ^ C) | ~(A | C) is invalid.
2123 if (Opcode == Instruction::Or && Op0->hasOneUse() &&
2124 match(Op1,
2126 Y, m_c_BinOp(Opcode, m_Specific(C),
2127 m_c_Xor(m_Specific(A), m_Specific(B)))))))) {
2128 // X = ~(A | B)
2129 // Y = (C | (A ^ B)
2130 Value *Or = cast<BinaryOperator>(X)->getOperand(0);
2131 return BinaryOperator::CreateNot(Builder.CreateAnd(Or, Y));
2132 }
2133 }
2134
2135 // (~A & B & C) | ... --> ...
2136 // (~A | B | C) | ... --> ...
2137 // TODO: One use checks are conservative. We just need to check that a total
2138 // number of multiple used values does not exceed reduction
2139 // in operations.
2140 if (match(Op0,
2141 m_OneUse(m_c_BinOp(FlippedOpcode,
2142 m_BinOp(FlippedOpcode, m_Value(B), m_Value(C)),
2143 m_Value(X, m_Not(m_Value(A)))))) ||
2144 match(Op0, m_OneUse(m_c_BinOp(FlippedOpcode,
2145 m_c_BinOp(FlippedOpcode, m_Value(C),
2146 m_Value(X, m_Not(m_Value(A)))),
2147 m_Value(B))))) {
2148 // X = ~A
2149 // (~A & B & C) | ~(A | B | C) --> ~(A | (B ^ C))
2150 // (~A | B | C) & ~(A & B & C) --> (~A | (B ^ C))
2151 if (match(Op1, m_OneUse(m_Not(m_c_BinOp(
2152 Opcode, m_c_BinOp(Opcode, m_Specific(A), m_Specific(B)),
2153 m_Specific(C))))) ||
2155 Opcode, m_c_BinOp(Opcode, m_Specific(B), m_Specific(C)),
2156 m_Specific(A))))) ||
2158 Opcode, m_c_BinOp(Opcode, m_Specific(A), m_Specific(C)),
2159 m_Specific(B)))))) {
2160 Value *Xor = Builder.CreateXor(B, C);
2161 return (Opcode == Instruction::Or)
2162 ? BinaryOperator::CreateNot(Builder.CreateOr(Xor, A))
2163 : BinaryOperator::CreateOr(Xor, X);
2164 }
2165
2166 // (~A & B & C) | ~(A | B) --> (C | ~B) & ~A
2167 // (~A | B | C) & ~(A & B) --> (C & ~B) | ~A
2168 if (match(Op1, m_OneUse(m_Not(m_OneUse(
2169 m_c_BinOp(Opcode, m_Specific(A), m_Specific(B)))))))
2171 FlippedOpcode, Builder.CreateBinOp(Opcode, C, Builder.CreateNot(B)),
2172 X);
2173
2174 // (~A & B & C) | ~(A | C) --> (B | ~C) & ~A
2175 // (~A | B | C) & ~(A & C) --> (B & ~C) | ~A
2176 if (match(Op1, m_OneUse(m_Not(m_OneUse(
2177 m_c_BinOp(Opcode, m_Specific(A), m_Specific(C)))))))
2179 FlippedOpcode, Builder.CreateBinOp(Opcode, B, Builder.CreateNot(C)),
2180 X);
2181 }
2182
2183 return nullptr;
2184}
2185
2186/// Try to reassociate a pair of binops so that values with one use only are
2187/// part of the same instruction. This may enable folds that are limited with
2188/// multi-use restrictions and makes it more likely to match other patterns that
2189/// are looking for a common operand.
2191 InstCombinerImpl::BuilderTy &Builder) {
2192 Instruction::BinaryOps Opcode = BO.getOpcode();
2193 Value *X, *Y, *Z;
2194 if (match(&BO,
2195 m_c_BinOp(Opcode, m_OneUse(m_BinOp(Opcode, m_Value(X), m_Value(Y))),
2196 m_OneUse(m_Value(Z))))) {
2197 if (!isa<Constant>(X) && !isa<Constant>(Y) && !isa<Constant>(Z)) {
2198 // (X op Y) op Z --> (Y op Z) op X
2199 if (!X->hasOneUse()) {
2200 Value *YZ = Builder.CreateBinOp(Opcode, Y, Z);
2201 return BinaryOperator::Create(Opcode, YZ, X);
2202 }
2203 // (X op Y) op Z --> (X op Z) op Y
2204 if (!Y->hasOneUse()) {
2205 Value *XZ = Builder.CreateBinOp(Opcode, X, Z);
2206 return BinaryOperator::Create(Opcode, XZ, Y);
2207 }
2208 }
2209 }
2210
2211 return nullptr;
2212}
2213
2214// Match
2215// (X + C2) | C
2216// (X + C2) ^ C
2217// (X + C2) & C
2218// and convert to do the bitwise logic first:
2219// (X | C) + C2
2220// (X ^ C) + C2
2221// (X & C) + C2
2222// iff bits affected by logic op are lower than last bit affected by math op
2224 InstCombiner::BuilderTy &Builder) {
2225 Type *Ty = I.getType();
2226 Instruction::BinaryOps OpC = I.getOpcode();
2227 Value *Op0 = I.getOperand(0);
2228 Value *Op1 = I.getOperand(1);
2229 Value *X;
2230 const APInt *C, *C2;
2231
2232 if (!(match(Op0, m_OneUse(m_Add(m_Value(X), m_APInt(C2)))) &&
2233 match(Op1, m_APInt(C))))
2234 return nullptr;
2235
2236 unsigned Width = Ty->getScalarSizeInBits();
2237 unsigned LastOneMath = Width - C2->countr_zero();
2238
2239 switch (OpC) {
2240 case Instruction::And:
2241 if (C->countl_one() < LastOneMath)
2242 return nullptr;
2243 break;
2244 case Instruction::Xor:
2245 case Instruction::Or:
2246 if (C->countl_zero() < LastOneMath)
2247 return nullptr;
2248 break;
2249 default:
2250 llvm_unreachable("Unexpected BinaryOp!");
2251 }
2252
2253 Value *NewBinOp = Builder.CreateBinOp(OpC, X, ConstantInt::get(Ty, *C));
2254 return BinaryOperator::CreateWithCopiedFlags(Instruction::Add, NewBinOp,
2255 ConstantInt::get(Ty, *C2), Op0);
2256}
2257
2258// binop(shift(ShiftedC1, ShAmt), shift(ShiftedC2, add(ShAmt, AddC))) ->
2259// shift(binop(ShiftedC1, shift(ShiftedC2, AddC)), ShAmt)
2260// where both shifts are the same and AddC is a valid shift amount.
2261Instruction *InstCombinerImpl::foldBinOpOfDisplacedShifts(BinaryOperator &I) {
2262 assert((I.isBitwiseLogicOp() || I.getOpcode() == Instruction::Add) &&
2263 "Unexpected opcode");
2264
2265 Value *ShAmt;
2266 Constant *ShiftedC1, *ShiftedC2, *AddC;
2267 Type *Ty = I.getType();
2268 unsigned BitWidth = Ty->getScalarSizeInBits();
2269 if (!match(&I, m_c_BinOp(m_Shift(m_ImmConstant(ShiftedC1), m_Value(ShAmt)),
2270 m_Shift(m_ImmConstant(ShiftedC2),
2271 m_AddLike(m_Deferred(ShAmt),
2272 m_ImmConstant(AddC))))))
2273 return nullptr;
2274
2275 // Make sure the add constant is a valid shift amount.
2276 if (!match(AddC,
2278 return nullptr;
2279
2280 // Avoid constant expressions.
2281 auto *Op0Inst = dyn_cast<Instruction>(I.getOperand(0));
2282 auto *Op1Inst = dyn_cast<Instruction>(I.getOperand(1));
2283 if (!Op0Inst || !Op1Inst)
2284 return nullptr;
2285
2286 // Both shifts must be the same.
2287 Instruction::BinaryOps ShiftOp =
2288 static_cast<Instruction::BinaryOps>(Op0Inst->getOpcode());
2289 if (ShiftOp != Op1Inst->getOpcode())
2290 return nullptr;
2291
2292 // For adds, only left shifts are supported.
2293 if (I.getOpcode() == Instruction::Add && ShiftOp != Instruction::Shl)
2294 return nullptr;
2295
2296 Value *NewC = Builder.CreateBinOp(
2297 I.getOpcode(), ShiftedC1, Builder.CreateBinOp(ShiftOp, ShiftedC2, AddC));
2298 return BinaryOperator::Create(ShiftOp, NewC, ShAmt);
2299}
2300
2301// Fold and/or/xor with two equal intrinsic IDs:
2302// bitwise(fshl (A, B, ShAmt), fshl(C, D, ShAmt))
2303// -> fshl(bitwise(A, C), bitwise(B, D), ShAmt)
2304// bitwise(fshr (A, B, ShAmt), fshr(C, D, ShAmt))
2305// -> fshr(bitwise(A, C), bitwise(B, D), ShAmt)
2306// bitwise(bswap(A), bswap(B)) -> bswap(bitwise(A, B))
2307// bitwise(bswap(A), C) -> bswap(bitwise(A, bswap(C)))
2308// bitwise(bitreverse(A), bitreverse(B)) -> bitreverse(bitwise(A, B))
2309// bitwise(bitreverse(A), C) -> bitreverse(bitwise(A, bitreverse(C)))
2310static Instruction *
2312 InstCombiner::BuilderTy &Builder) {
2313 assert(I.isBitwiseLogicOp() && "Should and/or/xor");
2314 if (!I.getOperand(0)->hasOneUse())
2315 return nullptr;
2316 IntrinsicInst *X = dyn_cast<IntrinsicInst>(I.getOperand(0));
2317 if (!X)
2318 return nullptr;
2319
2320 IntrinsicInst *Y = dyn_cast<IntrinsicInst>(I.getOperand(1));
2321 if (Y && (!Y->hasOneUse() || X->getIntrinsicID() != Y->getIntrinsicID()))
2322 return nullptr;
2323
2324 Intrinsic::ID IID = X->getIntrinsicID();
2325 const APInt *RHSC;
2326 // Try to match constant RHS.
2327 if (!Y && (!(IID == Intrinsic::bswap || IID == Intrinsic::bitreverse) ||
2328 !match(I.getOperand(1), m_APInt(RHSC))))
2329 return nullptr;
2330
2331 switch (IID) {
2332 case Intrinsic::fshl:
2333 case Intrinsic::fshr: {
2334 if (X->getOperand(2) != Y->getOperand(2))
2335 return nullptr;
2336 Value *NewOp0 =
2337 Builder.CreateBinOp(I.getOpcode(), X->getOperand(0), Y->getOperand(0));
2338 Value *NewOp1 =
2339 Builder.CreateBinOp(I.getOpcode(), X->getOperand(1), Y->getOperand(1));
2340 Function *F =
2341 Intrinsic::getOrInsertDeclaration(I.getModule(), IID, I.getType());
2342 return CallInst::Create(F, {NewOp0, NewOp1, X->getOperand(2)});
2343 }
2344 case Intrinsic::bswap:
2345 case Intrinsic::bitreverse: {
2346 Value *NewOp0 = Builder.CreateBinOp(
2347 I.getOpcode(), X->getOperand(0),
2348 Y ? Y->getOperand(0)
2349 : ConstantInt::get(I.getType(), IID == Intrinsic::bswap
2350 ? RHSC->byteSwap()
2351 : RHSC->reverseBits()));
2352 Function *F =
2353 Intrinsic::getOrInsertDeclaration(I.getModule(), IID, I.getType());
2354 return CallInst::Create(F, {NewOp0});
2355 }
2356 default:
2357 return nullptr;
2358 }
2359}
2360
2361// Try to simplify V by replacing occurrences of Op with RepOp, but only look
2362// through bitwise operations. In particular, for X | Y we try to replace Y with
2363// 0 inside X and for X & Y we try to replace Y with -1 inside X.
2364// Return the simplified result of X if successful, and nullptr otherwise.
2365// If SimplifyOnly is true, no new instructions will be created.
2367 bool SimplifyOnly,
2368 InstCombinerImpl &IC,
2369 unsigned Depth = 0) {
2370 if (Op == RepOp)
2371 return nullptr;
2372
2373 if (V == Op)
2374 return RepOp;
2375
2376 auto *I = dyn_cast<BinaryOperator>(V);
2377 if (!I || !I->isBitwiseLogicOp() || Depth >= 3)
2378 return nullptr;
2379
2380 if (!I->hasOneUse())
2381 SimplifyOnly = true;
2382
2383 Value *NewOp0 = simplifyAndOrWithOpReplaced(I->getOperand(0), Op, RepOp,
2384 SimplifyOnly, IC, Depth + 1);
2385 Value *NewOp1 = simplifyAndOrWithOpReplaced(I->getOperand(1), Op, RepOp,
2386 SimplifyOnly, IC, Depth + 1);
2387 if (!NewOp0 && !NewOp1)
2388 return nullptr;
2389
2390 if (!NewOp0)
2391 NewOp0 = I->getOperand(0);
2392 if (!NewOp1)
2393 NewOp1 = I->getOperand(1);
2394
2395 if (Value *Res = simplifyBinOp(I->getOpcode(), NewOp0, NewOp1,
2397 return Res;
2398
2399 if (SimplifyOnly)
2400 return nullptr;
2401 return IC.Builder.CreateBinOp(I->getOpcode(), NewOp0, NewOp1);
2402}
2403
2404/// Reassociate and/or expressions to see if we can fold the inner and/or ops.
2405/// TODO: Make this recursive; it's a little tricky because an arbitrary
2406/// number of and/or instructions might have to be created.
2407Value *InstCombinerImpl::reassociateBooleanAndOr(Value *LHS, Value *X, Value *Y,
2408 Instruction &I, bool IsAnd,
2409 bool RHSIsLogical) {
2410 Instruction::BinaryOps Opcode = IsAnd ? Instruction::And : Instruction::Or;
2411 // LHS bop (X lop Y) --> (LHS bop X) lop Y
2412 // LHS bop (X bop Y) --> (LHS bop X) bop Y
2413 if (Value *Res = foldBooleanAndOr(LHS, X, I, IsAnd, /*IsLogical=*/false))
2414 return RHSIsLogical ? Builder.CreateLogicalOp(Opcode, Res, Y)
2415 : Builder.CreateBinOp(Opcode, Res, Y);
2416 // LHS bop (X bop Y) --> X bop (LHS bop Y)
2417 // LHS bop (X lop Y) --> X lop (LHS bop Y)
2418 if (Value *Res = foldBooleanAndOr(LHS, Y, I, IsAnd, /*IsLogical=*/false))
2419 return RHSIsLogical ? Builder.CreateLogicalOp(Opcode, X, Res)
2420 : Builder.CreateBinOp(Opcode, X, Res);
2421 return nullptr;
2422}
2423
2424// FIXME: We use commutative matchers (m_c_*) for some, but not all, matches
2425// here. We should standardize that construct where it is needed or choose some
2426// other way to ensure that commutated variants of patterns are not missed.
2428 Type *Ty = I.getType();
2429
2430 if (Value *V = simplifyAndInst(I.getOperand(0), I.getOperand(1),
2431 SQ.getWithInstruction(&I)))
2432 return replaceInstUsesWith(I, V);
2433
2435 return &I;
2436
2438 return X;
2439
2441 return Phi;
2442
2443 // See if we can simplify any instructions used by the instruction whose sole
2444 // purpose is to compute bits we don't care about.
2446 return &I;
2447
2448 // Do this before using distributive laws to catch simple and/or/not patterns.
2450 return Xor;
2451
2453 return X;
2454
2455 // (A|B)&(A|C) -> A|(B&C) etc
2457 return replaceInstUsesWith(I, V);
2458
2460 return R;
2461
2462 Value *Op0 = I.getOperand(0), *Op1 = I.getOperand(1);
2463
2464 Value *X, *Y;
2465 const APInt *C;
2466 if ((match(Op0, m_OneUse(m_LogicalShift(m_One(), m_Value(X)))) ||
2467 (match(Op0, m_OneUse(m_Shl(m_APInt(C), m_Value(X)))) && (*C)[0])) &&
2468 match(Op1, m_One())) {
2469 // (1 >> X) & 1 --> zext(X == 0)
2470 // (C << X) & 1 --> zext(X == 0), when C is odd
2471 Value *IsZero = Builder.CreateICmpEQ(X, ConstantInt::get(Ty, 0));
2472 return new ZExtInst(IsZero, Ty);
2473 }
2474
2475 // (-(X & 1)) & Y --> (X & 1) == 0 ? 0 : Y
2476 Value *Neg;
2477 if (match(&I,
2479 m_Value(Y)))) {
2480 Value *Cmp = Builder.CreateIsNull(Neg);
2482 }
2483
2484 // Canonicalize:
2485 // (X +/- Y) & Y --> ~X & Y when Y is a power of 2.
2488 m_Sub(m_Value(X), m_Deferred(Y)))))) &&
2489 isKnownToBeAPowerOfTwo(Y, /*OrZero*/ true, &I))
2490 return BinaryOperator::CreateAnd(Builder.CreateNot(X), Y);
2491
2492 if (match(Op1, m_APInt(C))) {
2493 const APInt *XorC;
2494 if (match(Op0, m_OneUse(m_Xor(m_Value(X), m_APInt(XorC))))) {
2495 // (X ^ C1) & C2 --> (X & C2) ^ (C1&C2)
2496 Constant *NewC = ConstantInt::get(Ty, *C & *XorC);
2497 Value *And = Builder.CreateAnd(X, Op1);
2498 And->takeName(Op0);
2499 return BinaryOperator::CreateXor(And, NewC);
2500 }
2501
2502 const APInt *OrC;
2503 if (match(Op0, m_OneUse(m_Or(m_Value(X), m_APInt(OrC))))) {
2504 // (X | C1) & C2 --> (X & C2^(C1&C2)) | (C1&C2)
2505 // NOTE: This reduces the number of bits set in the & mask, which
2506 // can expose opportunities for store narrowing for scalars.
2507 // NOTE: SimplifyDemandedBits should have already removed bits from C1
2508 // that aren't set in C2. Meaning we can replace (C1&C2) with C1 in
2509 // above, but this feels safer.
2510 APInt Together = *C & *OrC;
2511 Value *And = Builder.CreateAnd(X, ConstantInt::get(Ty, Together ^ *C));
2512 And->takeName(Op0);
2513 return BinaryOperator::CreateOr(And, ConstantInt::get(Ty, Together));
2514 }
2515
2516 unsigned Width = Ty->getScalarSizeInBits();
2517 const APInt *ShiftC;
2518 if (match(Op0, m_OneUse(m_SExt(m_AShr(m_Value(X), m_APInt(ShiftC))))) &&
2519 ShiftC->ult(Width)) {
2520 if (*C == APInt::getLowBitsSet(Width, Width - ShiftC->getZExtValue())) {
2521 // We are clearing high bits that were potentially set by sext+ashr:
2522 // and (sext (ashr X, ShiftC)), C --> lshr (sext X), ShiftC
2523 Value *Sext = Builder.CreateSExt(X, Ty);
2524 Constant *ShAmtC = ConstantInt::get(Ty, ShiftC->zext(Width));
2525 return BinaryOperator::CreateLShr(Sext, ShAmtC);
2526 }
2527 }
2528
2529 // If this 'and' clears the sign-bits added by ashr, replace with lshr:
2530 // and (ashr X, ShiftC), C --> lshr X, ShiftC
2531 if (match(Op0, m_AShr(m_Value(X), m_APInt(ShiftC))) && ShiftC->ult(Width) &&
2532 C->isMask(Width - ShiftC->getZExtValue()))
2533 return BinaryOperator::CreateLShr(X, ConstantInt::get(Ty, *ShiftC));
2534
2535 const APInt *AddC;
2536 if (match(Op0, m_Add(m_Value(X), m_APInt(AddC)))) {
2537 // If we are masking the result of the add down to exactly one bit and
2538 // the constant we are adding has no bits set below that bit, then the
2539 // add is flipping a single bit. Example:
2540 // (X + 4) & 4 --> (X & 4) ^ 4
2541 if (Op0->hasOneUse() && C->isPowerOf2() && (*AddC & (*C - 1)) == 0) {
2542 assert((*C & *AddC) != 0 && "Expected common bit");
2543 Value *NewAnd = Builder.CreateAnd(X, Op1);
2544 return BinaryOperator::CreateXor(NewAnd, Op1);
2545 }
2546 }
2547
2548 // ((C1 OP zext(X)) & C2) -> zext((C1 OP X) & C2) if C2 fits in the
2549 // bitwidth of X and OP behaves well when given trunc(C1) and X.
2550 auto isNarrowableBinOpcode = [](BinaryOperator *B) {
2551 switch (B->getOpcode()) {
2552 case Instruction::Xor:
2553 case Instruction::Or:
2554 case Instruction::Mul:
2555 case Instruction::Add:
2556 case Instruction::Sub:
2557 return true;
2558 default:
2559 return false;
2560 }
2561 };
2562 BinaryOperator *BO;
2563 if (match(Op0, m_OneUse(m_BinOp(BO))) && isNarrowableBinOpcode(BO)) {
2564 Instruction::BinaryOps BOpcode = BO->getOpcode();
2565 Value *X;
2566 const APInt *C1;
2567 // TODO: The one-use restrictions could be relaxed a little if the AND
2568 // is going to be removed.
2569 // Try to narrow the 'and' and a binop with constant operand:
2570 // and (bo (zext X), C1), C --> zext (and (bo X, TruncC1), TruncC)
2571 if (match(BO, m_c_BinOp(m_OneUse(m_ZExt(m_Value(X))), m_APInt(C1))) &&
2572 C->isIntN(X->getType()->getScalarSizeInBits())) {
2573 unsigned XWidth = X->getType()->getScalarSizeInBits();
2574 Constant *TruncC1 = ConstantInt::get(X->getType(), C1->trunc(XWidth));
2575 Value *BinOp = isa<ZExtInst>(BO->getOperand(0))
2576 ? Builder.CreateBinOp(BOpcode, X, TruncC1)
2577 : Builder.CreateBinOp(BOpcode, TruncC1, X);
2578 Constant *TruncC = ConstantInt::get(X->getType(), C->trunc(XWidth));
2579 Value *And = Builder.CreateAnd(BinOp, TruncC);
2580 return new ZExtInst(And, Ty);
2581 }
2582
2583 // Similar to above: if the mask matches the zext input width, then the
2584 // 'and' can be eliminated, so we can truncate the other variable op:
2585 // and (bo (zext X), Y), C --> zext (bo X, (trunc Y))
2586 if (isa<Instruction>(BO->getOperand(0)) &&
2587 match(BO->getOperand(0), m_OneUse(m_ZExt(m_Value(X)))) &&
2588 C->isMask(X->getType()->getScalarSizeInBits())) {
2589 Y = BO->getOperand(1);
2590 Value *TrY = Builder.CreateTrunc(Y, X->getType(), Y->getName() + ".tr");
2591 Value *NewBO =
2592 Builder.CreateBinOp(BOpcode, X, TrY, BO->getName() + ".narrow");
2593 return new ZExtInst(NewBO, Ty);
2594 }
2595 // and (bo Y, (zext X)), C --> zext (bo (trunc Y), X)
2596 if (isa<Instruction>(BO->getOperand(1)) &&
2597 match(BO->getOperand(1), m_OneUse(m_ZExt(m_Value(X)))) &&
2598 C->isMask(X->getType()->getScalarSizeInBits())) {
2599 Y = BO->getOperand(0);
2600 Value *TrY = Builder.CreateTrunc(Y, X->getType(), Y->getName() + ".tr");
2601 Value *NewBO =
2602 Builder.CreateBinOp(BOpcode, TrY, X, BO->getName() + ".narrow");
2603 return new ZExtInst(NewBO, Ty);
2604 }
2605 }
2606
2607 // This is intentionally placed after the narrowing transforms for
2608 // efficiency (transform directly to the narrow logic op if possible).
2609 // If the mask is only needed on one incoming arm, push the 'and' op up.
2610 if (match(Op0, m_OneUse(m_Xor(m_Value(X), m_Value(Y)))) ||
2611 match(Op0, m_OneUse(m_Or(m_Value(X), m_Value(Y))))) {
2612 APInt NotAndMask(~(*C));
2613 BinaryOperator::BinaryOps BinOp = cast<BinaryOperator>(Op0)->getOpcode();
2614 if (MaskedValueIsZero(X, NotAndMask, &I)) {
2615 // Not masking anything out for the LHS, move mask to RHS.
2616 // and ({x}or X, Y), C --> {x}or X, (and Y, C)
2617 Value *NewRHS = Builder.CreateAnd(Y, Op1, Y->getName() + ".masked");
2618 return BinaryOperator::Create(BinOp, X, NewRHS);
2619 }
2620 if (!isa<Constant>(Y) && MaskedValueIsZero(Y, NotAndMask, &I)) {
2621 // Not masking anything out for the RHS, move mask to LHS.
2622 // and ({x}or X, Y), C --> {x}or (and X, C), Y
2623 Value *NewLHS = Builder.CreateAnd(X, Op1, X->getName() + ".masked");
2624 return BinaryOperator::Create(BinOp, NewLHS, Y);
2625 }
2626 }
2627
2628 // When the mask is a power-of-2 constant and op0 is a shifted-power-of-2
2629 // constant, test if the shift amount equals the offset bit index:
2630 // (ShiftC << X) & C --> X == (log2(C) - log2(ShiftC)) ? C : 0
2631 // (ShiftC >> X) & C --> X == (log2(ShiftC) - log2(C)) ? C : 0
2632 if (C->isPowerOf2() &&
2633 match(Op0, m_OneUse(m_LogicalShift(m_Power2(ShiftC), m_Value(X))))) {
2634 int Log2ShiftC = ShiftC->exactLogBase2();
2635 int Log2C = C->exactLogBase2();
2636 bool IsShiftLeft =
2637 cast<BinaryOperator>(Op0)->getOpcode() == Instruction::Shl;
2638 int BitNum = IsShiftLeft ? Log2C - Log2ShiftC : Log2ShiftC - Log2C;
2639 assert(BitNum >= 0 && "Expected demanded bits to handle impossible mask");
2640 Value *Cmp = Builder.CreateICmpEQ(X, ConstantInt::get(Ty, BitNum));
2641 return SelectInst::Create(Cmp, ConstantInt::get(Ty, *C),
2643 }
2644
2645 Constant *C1, *C2;
2646 const APInt *C3 = C;
2647 Value *X;
2648 if (C3->isPowerOf2()) {
2649 Constant *Log2C3 = ConstantInt::get(Ty, C3->countr_zero());
2651 m_ImmConstant(C2)))) &&
2652 match(C1, m_Power2())) {
2654 Constant *LshrC = ConstantExpr::getAdd(C2, Log2C3);
2655 KnownBits KnownLShrc = computeKnownBits(LshrC, nullptr);
2656 if (KnownLShrc.getMaxValue().ult(Width)) {
2657 // iff C1,C3 is pow2 and C2 + cttz(C3) < BitWidth:
2658 // ((C1 << X) >> C2) & C3 -> X == (cttz(C3)+C2-cttz(C1)) ? C3 : 0
2659 Constant *CmpC = ConstantExpr::getSub(LshrC, Log2C1);
2660 Value *Cmp = Builder.CreateICmpEQ(X, CmpC);
2661 return SelectInst::Create(Cmp, ConstantInt::get(Ty, *C3),
2663 }
2664 }
2665
2667 m_ImmConstant(C2)))) &&
2668 match(C1, m_Power2())) {
2670 Constant *Cmp =
2672 if (Cmp && Cmp->isZeroValue()) {
2673 // iff C1,C3 is pow2 and Log2(C3) >= C2:
2674 // ((C1 >> X) << C2) & C3 -> X == (cttz(C1)+C2-cttz(C3)) ? C3 : 0
2675 Constant *ShlC = ConstantExpr::getAdd(C2, Log2C1);
2676 Constant *CmpC = ConstantExpr::getSub(ShlC, Log2C3);
2677 Value *Cmp = Builder.CreateICmpEQ(X, CmpC);
2678 return SelectInst::Create(Cmp, ConstantInt::get(Ty, *C3),
2680 }
2681 }
2682 }
2683 }
2684
2685 // If we are clearing the sign bit of a floating-point value, convert this to
2686 // fabs, then cast back to integer.
2687 //
2688 // This is a generous interpretation for noimplicitfloat, this is not a true
2689 // floating-point operation.
2690 //
2691 // Assumes any IEEE-represented type has the sign bit in the high bit.
2692 // TODO: Unify with APInt matcher. This version allows undef unlike m_APInt
2693 Value *CastOp;
2694 if (match(Op0, m_ElementWiseBitCast(m_Value(CastOp))) &&
2695 match(Op1, m_MaxSignedValue()) &&
2696 !Builder.GetInsertBlock()->getParent()->hasFnAttribute(
2697 Attribute::NoImplicitFloat)) {
2698 Type *EltTy = CastOp->getType()->getScalarType();
2699 if (EltTy->isFloatingPointTy() &&
2701 Value *FAbs = Builder.CreateUnaryIntrinsic(Intrinsic::fabs, CastOp);
2702 return new BitCastInst(FAbs, I.getType());
2703 }
2704 }
2705
2706 // and(shl(zext(X), Y), SignMask) -> and(sext(X), SignMask)
2707 // where Y is a valid shift amount.
2709 m_SignMask())) &&
2712 APInt(Ty->getScalarSizeInBits(),
2713 Ty->getScalarSizeInBits() -
2714 X->getType()->getScalarSizeInBits())))) {
2715 auto *SExt = Builder.CreateSExt(X, Ty, X->getName() + ".signext");
2716 return BinaryOperator::CreateAnd(SExt, Op1);
2717 }
2718
2719 if (Instruction *Z = narrowMaskedBinOp(I))
2720 return Z;
2721
2722 if (I.getType()->isIntOrIntVectorTy(1)) {
2723 if (auto *SI0 = dyn_cast<SelectInst>(Op0)) {
2724 if (auto *R =
2725 foldAndOrOfSelectUsingImpliedCond(Op1, *SI0, /* IsAnd */ true))
2726 return R;
2727 }
2728 if (auto *SI1 = dyn_cast<SelectInst>(Op1)) {
2729 if (auto *R =
2730 foldAndOrOfSelectUsingImpliedCond(Op0, *SI1, /* IsAnd */ true))
2731 return R;
2732 }
2733 }
2734
2735 if (Instruction *FoldedLogic = foldBinOpIntoSelectOrPhi(I))
2736 return FoldedLogic;
2737
2738 if (Instruction *DeMorgan = matchDeMorgansLaws(I, *this))
2739 return DeMorgan;
2740
2741 {
2742 Value *A, *B, *C;
2743 // A & ~(A ^ B) --> A & B
2744 if (match(Op1, m_Not(m_c_Xor(m_Specific(Op0), m_Value(B)))))
2745 return BinaryOperator::CreateAnd(Op0, B);
2746 // ~(A ^ B) & A --> A & B
2747 if (match(Op0, m_Not(m_c_Xor(m_Specific(Op1), m_Value(B)))))
2748 return BinaryOperator::CreateAnd(Op1, B);
2749
2750 // (A ^ B) & ((B ^ C) ^ A) -> (A ^ B) & ~C
2751 if (match(Op0, m_Xor(m_Value(A), m_Value(B))) &&
2752 match(Op1, m_Xor(m_Xor(m_Specific(B), m_Value(C)), m_Specific(A)))) {
2753 Value *NotC = Op1->hasOneUse()
2754 ? Builder.CreateNot(C)
2755 : getFreelyInverted(C, C->hasOneUse(), &Builder);
2756 if (NotC != nullptr)
2757 return BinaryOperator::CreateAnd(Op0, NotC);
2758 }
2759
2760 // ((A ^ C) ^ B) & (B ^ A) -> (B ^ A) & ~C
2761 if (match(Op0, m_Xor(m_Xor(m_Value(A), m_Value(C)), m_Value(B))) &&
2762 match(Op1, m_Xor(m_Specific(B), m_Specific(A)))) {
2763 Value *NotC = Op0->hasOneUse()
2764 ? Builder.CreateNot(C)
2765 : getFreelyInverted(C, C->hasOneUse(), &Builder);
2766 if (NotC != nullptr)
2767 return BinaryOperator::CreateAnd(Op1, Builder.CreateNot(C));
2768 }
2769
2770 // (A | B) & (~A ^ B) -> A & B
2771 // (A | B) & (B ^ ~A) -> A & B
2772 // (B | A) & (~A ^ B) -> A & B
2773 // (B | A) & (B ^ ~A) -> A & B
2774 if (match(Op1, m_c_Xor(m_Not(m_Value(A)), m_Value(B))) &&
2775 match(Op0, m_c_Or(m_Specific(A), m_Specific(B))))
2776 return BinaryOperator::CreateAnd(A, B);
2777
2778 // (~A ^ B) & (A | B) -> A & B
2779 // (~A ^ B) & (B | A) -> A & B
2780 // (B ^ ~A) & (A | B) -> A & B
2781 // (B ^ ~A) & (B | A) -> A & B
2782 if (match(Op0, m_c_Xor(m_Not(m_Value(A)), m_Value(B))) &&
2783 match(Op1, m_c_Or(m_Specific(A), m_Specific(B))))
2784 return BinaryOperator::CreateAnd(A, B);
2785
2786 // (~A | B) & (A ^ B) -> ~A & B
2787 // (~A | B) & (B ^ A) -> ~A & B
2788 // (B | ~A) & (A ^ B) -> ~A & B
2789 // (B | ~A) & (B ^ A) -> ~A & B
2790 if (match(Op0, m_c_Or(m_Not(m_Value(A)), m_Value(B))) &&
2792 return BinaryOperator::CreateAnd(Builder.CreateNot(A), B);
2793
2794 // (A ^ B) & (~A | B) -> ~A & B
2795 // (B ^ A) & (~A | B) -> ~A & B
2796 // (A ^ B) & (B | ~A) -> ~A & B
2797 // (B ^ A) & (B | ~A) -> ~A & B
2798 if (match(Op1, m_c_Or(m_Not(m_Value(A)), m_Value(B))) &&
2800 return BinaryOperator::CreateAnd(Builder.CreateNot(A), B);
2801 }
2802
2803 if (Value *Res =
2804 foldBooleanAndOr(Op0, Op1, I, /*IsAnd=*/true, /*IsLogical=*/false))
2805 return replaceInstUsesWith(I, Res);
2806
2807 if (match(Op1, m_OneUse(m_LogicalAnd(m_Value(X), m_Value(Y))))) {
2808 bool IsLogical = isa<SelectInst>(Op1);
2809 if (auto *V = reassociateBooleanAndOr(Op0, X, Y, I, /*IsAnd=*/true,
2810 /*RHSIsLogical=*/IsLogical))
2811 return replaceInstUsesWith(I, V);
2812 }
2813 if (match(Op0, m_OneUse(m_LogicalAnd(m_Value(X), m_Value(Y))))) {
2814 bool IsLogical = isa<SelectInst>(Op0);
2815 if (auto *V = reassociateBooleanAndOr(Op1, X, Y, I, /*IsAnd=*/true,
2816 /*RHSIsLogical=*/IsLogical))
2817 return replaceInstUsesWith(I, V);
2818 }
2819
2820 if (Instruction *FoldedFCmps = reassociateFCmps(I, Builder))
2821 return FoldedFCmps;
2822
2823 if (Instruction *CastedAnd = foldCastedBitwiseLogic(I))
2824 return CastedAnd;
2825
2826 if (Instruction *Sel = foldBinopOfSextBoolToSelect(I))
2827 return Sel;
2828
2829 // and(sext(A), B) / and(B, sext(A)) --> A ? B : 0, where A is i1 or <N x i1>.
2830 // TODO: Move this into foldBinopOfSextBoolToSelect as a more generalized fold
2831 // with binop identity constant. But creating a select with non-constant
2832 // arm may not be reversible due to poison semantics. Is that a good
2833 // canonicalization?
2834 Value *A, *B;
2835 if (match(&I, m_c_And(m_SExt(m_Value(A)), m_Value(B))) &&
2836 A->getType()->isIntOrIntVectorTy(1))
2838
2839 // Similarly, a 'not' of the bool translates to a swap of the select arms:
2840 // ~sext(A) & B / B & ~sext(A) --> A ? 0 : B
2841 if (match(&I, m_c_And(m_Not(m_SExt(m_Value(A))), m_Value(B))) &&
2842 A->getType()->isIntOrIntVectorTy(1))
2844
2845 // and(zext(A), B) -> A ? (B & 1) : 0
2846 if (match(&I, m_c_And(m_OneUse(m_ZExt(m_Value(A))), m_Value(B))) &&
2847 A->getType()->isIntOrIntVectorTy(1))
2848 return SelectInst::Create(A, Builder.CreateAnd(B, ConstantInt::get(Ty, 1)),
2850
2851 // (-1 + A) & B --> A ? 0 : B where A is 0/1.
2853 m_Value(B)))) {
2854 if (A->getType()->isIntOrIntVectorTy(1))
2856 if (computeKnownBits(A, &I).countMaxActiveBits() <= 1) {
2857 return SelectInst::Create(
2858 Builder.CreateICmpEQ(A, Constant::getNullValue(A->getType())), B,
2860 }
2861 }
2862
2863 // (iN X s>> (N-1)) & Y --> (X s< 0) ? Y : 0 -- with optional sext
2866 m_Value(Y))) &&
2867 *C == X->getType()->getScalarSizeInBits() - 1) {
2868 Value *IsNeg = Builder.CreateIsNeg(X, "isneg");
2870 }
2871 // If there's a 'not' of the shifted value, swap the select operands:
2872 // ~(iN X s>> (N-1)) & Y --> (X s< 0) ? 0 : Y -- with optional sext
2875 m_Value(Y))) &&
2876 *C == X->getType()->getScalarSizeInBits() - 1) {
2877 Value *IsNeg = Builder.CreateIsNeg(X, "isneg");
2879 }
2880
2881 // (~x) & y --> ~(x | (~y)) iff that gets rid of inversions
2883 return &I;
2884
2885 // An and recurrence w/loop invariant step is equivelent to (and start, step)
2886 PHINode *PN = nullptr;
2887 Value *Start = nullptr, *Step = nullptr;
2888 if (matchSimpleRecurrence(&I, PN, Start, Step) && DT.dominates(Step, PN))
2889 return replaceInstUsesWith(I, Builder.CreateAnd(Start, Step));
2890
2892 return R;
2893
2894 if (Instruction *Canonicalized = canonicalizeLogicFirst(I, Builder))
2895 return Canonicalized;
2896
2897 if (Instruction *Folded = foldLogicOfIsFPClass(I, Op0, Op1))
2898 return Folded;
2899
2900 if (Instruction *Res = foldBinOpOfDisplacedShifts(I))
2901 return Res;
2902
2904 return Res;
2905
2906 if (Value *V =
2908 /*SimplifyOnly*/ false, *this))
2909 return BinaryOperator::CreateAnd(V, Op1);
2910 if (Value *V =
2912 /*SimplifyOnly*/ false, *this))
2913 return BinaryOperator::CreateAnd(Op0, V);
2914
2915 return nullptr;
2916}
2917
2919 bool MatchBSwaps,
2920 bool MatchBitReversals) {
2922 if (!recognizeBSwapOrBitReverseIdiom(&I, MatchBSwaps, MatchBitReversals,
2923 Insts))
2924 return nullptr;
2925 Instruction *LastInst = Insts.pop_back_val();
2926 LastInst->removeFromParent();
2927
2928 for (auto *Inst : Insts) {
2929 Inst->setDebugLoc(I.getDebugLoc());
2930 Worklist.push(Inst);
2931 }
2932 return LastInst;
2933}
2934
2935std::optional<std::pair<Intrinsic::ID, SmallVector<Value *, 3>>>
2937 // TODO: Can we reduce the code duplication between this and the related
2938 // rotate matching code under visitSelect and visitTrunc?
2939 assert(Or.getOpcode() == BinaryOperator::Or && "Expecting or instruction");
2940
2941 unsigned Width = Or.getType()->getScalarSizeInBits();
2942
2943 Instruction *Or0, *Or1;
2944 if (!match(Or.getOperand(0), m_Instruction(Or0)) ||
2945 !match(Or.getOperand(1), m_Instruction(Or1)))
2946 return std::nullopt;
2947
2948 bool IsFshl = true; // Sub on LSHR.
2949 SmallVector<Value *, 3> FShiftArgs;
2950
2951 // First, find an or'd pair of opposite shifts:
2952 // or (lshr ShVal0, ShAmt0), (shl ShVal1, ShAmt1)
2953 if (isa<BinaryOperator>(Or0) && isa<BinaryOperator>(Or1)) {
2954 Value *ShVal0, *ShVal1, *ShAmt0, *ShAmt1;
2955 if (!match(Or0,
2956 m_OneUse(m_LogicalShift(m_Value(ShVal0), m_Value(ShAmt0)))) ||
2957 !match(Or1,
2958 m_OneUse(m_LogicalShift(m_Value(ShVal1), m_Value(ShAmt1)))) ||
2959 Or0->getOpcode() == Or1->getOpcode())
2960 return std::nullopt;
2961
2962 // Canonicalize to or(shl(ShVal0, ShAmt0), lshr(ShVal1, ShAmt1)).
2963 if (Or0->getOpcode() == BinaryOperator::LShr) {
2964 std::swap(Or0, Or1);
2965 std::swap(ShVal0, ShVal1);
2966 std::swap(ShAmt0, ShAmt1);
2967 }
2968 assert(Or0->getOpcode() == BinaryOperator::Shl &&
2969 Or1->getOpcode() == BinaryOperator::LShr &&
2970 "Illegal or(shift,shift) pair");
2971
2972 // Match the shift amount operands for a funnel shift pattern. This always
2973 // matches a subtraction on the R operand.
2974 auto matchShiftAmount = [&](Value *L, Value *R, unsigned Width) -> Value * {
2975 // Check for constant shift amounts that sum to the bitwidth.
2976 const APInt *LI, *RI;
2977 if (match(L, m_APIntAllowPoison(LI)) && match(R, m_APIntAllowPoison(RI)))
2978 if (LI->ult(Width) && RI->ult(Width) && (*LI + *RI) == Width)
2979 return ConstantInt::get(L->getType(), *LI);
2980
2981 Constant *LC, *RC;
2982 if (match(L, m_Constant(LC)) && match(R, m_Constant(RC)) &&
2983 match(L,
2984 m_SpecificInt_ICMP(ICmpInst::ICMP_ULT, APInt(Width, Width))) &&
2985 match(R,
2986 m_SpecificInt_ICMP(ICmpInst::ICMP_ULT, APInt(Width, Width))) &&
2988 return ConstantExpr::mergeUndefsWith(LC, RC);
2989
2990 // (shl ShVal, X) | (lshr ShVal, (Width - x)) iff X < Width.
2991 // We limit this to X < Width in case the backend re-expands the
2992 // intrinsic, and has to reintroduce a shift modulo operation (InstCombine
2993 // might remove it after this fold). This still doesn't guarantee that the
2994 // final codegen will match this original pattern.
2995 if (match(R, m_OneUse(m_Sub(m_SpecificInt(Width), m_Specific(L))))) {
2996 KnownBits KnownL = computeKnownBits(L, &Or);
2997 return KnownL.getMaxValue().ult(Width) ? L : nullptr;
2998 }
2999
3000 // For non-constant cases, the following patterns currently only work for
3001 // rotation patterns.
3002 // TODO: Add general funnel-shift compatible patterns.
3003 if (ShVal0 != ShVal1)
3004 return nullptr;
3005
3006 // For non-constant cases we don't support non-pow2 shift masks.
3007 // TODO: Is it worth matching urem as well?
3008 if (!isPowerOf2_32(Width))
3009 return nullptr;
3010
3011 // The shift amount may be masked with negation:
3012 // (shl ShVal, (X & (Width - 1))) | (lshr ShVal, ((-X) & (Width - 1)))
3013 Value *X;
3014 unsigned Mask = Width - 1;
3015 if (match(L, m_And(m_Value(X), m_SpecificInt(Mask))) &&
3016 match(R, m_And(m_Neg(m_Specific(X)), m_SpecificInt(Mask))))
3017 return X;
3018
3019 // (shl ShVal, X) | (lshr ShVal, ((-X) & (Width - 1)))
3020 if (match(R, m_And(m_Neg(m_Specific(L)), m_SpecificInt(Mask))))
3021 return L;
3022
3023 // Similar to above, but the shift amount may be extended after masking,
3024 // so return the extended value as the parameter for the intrinsic.
3025 if (match(L, m_ZExt(m_And(m_Value(X), m_SpecificInt(Mask)))) &&
3026 match(R,
3028 m_SpecificInt(Mask))))
3029 return L;
3030
3031 if (match(L, m_ZExt(m_And(m_Value(X), m_SpecificInt(Mask)))) &&
3033 return L;
3034
3035 return nullptr;
3036 };
3037
3038 Value *ShAmt = matchShiftAmount(ShAmt0, ShAmt1, Width);
3039 if (!ShAmt) {
3040 ShAmt = matchShiftAmount(ShAmt1, ShAmt0, Width);
3041 IsFshl = false; // Sub on SHL.
3042 }
3043 if (!ShAmt)
3044 return std::nullopt;
3045
3046 FShiftArgs = {ShVal0, ShVal1, ShAmt};
3047 } else if (isa<ZExtInst>(Or0) || isa<ZExtInst>(Or1)) {
3048 // If there are two 'or' instructions concat variables in opposite order:
3049 //
3050 // Slot1 and Slot2 are all zero bits.
3051 // | Slot1 | Low | Slot2 | High |
3052 // LowHigh = or (shl (zext Low), ZextLowShlAmt), (zext High)
3053 // | Slot2 | High | Slot1 | Low |
3054 // HighLow = or (shl (zext High), ZextHighShlAmt), (zext Low)
3055 //
3056 // the latter 'or' can be safely convert to
3057 // -> HighLow = fshl LowHigh, LowHigh, ZextHighShlAmt
3058 // if ZextLowShlAmt + ZextHighShlAmt == Width.
3059 if (!isa<ZExtInst>(Or1))
3060 std::swap(Or0, Or1);
3061
3062 Value *High, *ZextHigh, *Low;
3063 const APInt *ZextHighShlAmt;
3064 if (!match(Or0,
3065 m_OneUse(m_Shl(m_Value(ZextHigh), m_APInt(ZextHighShlAmt)))))
3066 return std::nullopt;
3067
3068 if (!match(Or1, m_ZExt(m_Value(Low))) ||
3069 !match(ZextHigh, m_ZExt(m_Value(High))))
3070 return std::nullopt;
3071
3072 unsigned HighSize = High->getType()->getScalarSizeInBits();
3073 unsigned LowSize = Low->getType()->getScalarSizeInBits();
3074 // Make sure High does not overlap with Low and most significant bits of
3075 // High aren't shifted out.
3076 if (ZextHighShlAmt->ult(LowSize) || ZextHighShlAmt->ugt(Width - HighSize))
3077 return std::nullopt;
3078
3079 for (User *U : ZextHigh->users()) {
3080 Value *X, *Y;
3081 if (!match(U, m_Or(m_Value(X), m_Value(Y))))
3082 continue;
3083
3084 if (!isa<ZExtInst>(Y))
3085 std::swap(X, Y);
3086
3087 const APInt *ZextLowShlAmt;
3088 if (!match(X, m_Shl(m_Specific(Or1), m_APInt(ZextLowShlAmt))) ||
3089 !match(Y, m_Specific(ZextHigh)) || !DT.dominates(U, &Or))
3090 continue;
3091
3092 // HighLow is good concat. If sum of two shifts amount equals to Width,
3093 // LowHigh must also be a good concat.
3094 if (*ZextLowShlAmt + *ZextHighShlAmt != Width)
3095 continue;
3096
3097 // Low must not overlap with High and most significant bits of Low must
3098 // not be shifted out.
3099 assert(ZextLowShlAmt->uge(HighSize) &&
3100 ZextLowShlAmt->ule(Width - LowSize) && "Invalid concat");
3101
3102 FShiftArgs = {U, U, ConstantInt::get(Or0->getType(), *ZextHighShlAmt)};
3103 break;
3104 }
3105 }
3106
3107 if (FShiftArgs.empty())
3108 return std::nullopt;
3109
3110 Intrinsic::ID IID = IsFshl ? Intrinsic::fshl : Intrinsic::fshr;
3111 return std::make_pair(IID, FShiftArgs);
3112}
3113
3114/// Match UB-safe variants of the funnel shift intrinsic.
3116 if (auto Opt = IC.convertOrOfShiftsToFunnelShift(Or)) {
3117 auto [IID, FShiftArgs] = *Opt;
3118 Function *F =
3119 Intrinsic::getOrInsertDeclaration(Or.getModule(), IID, Or.getType());
3120 return CallInst::Create(F, FShiftArgs);
3121 }
3122
3123 return nullptr;
3124}
3125
3126/// Attempt to combine or(zext(x),shl(zext(y),bw/2) concat packing patterns.
3128 assert(Or.getOpcode() == Instruction::Or && "bswap requires an 'or'");
3129 Value *Op0 = Or.getOperand(0), *Op1 = Or.getOperand(1);
3130 Type *Ty = Or.getType();
3131
3132 unsigned Width = Ty->getScalarSizeInBits();
3133 if ((Width & 1) != 0)
3134 return nullptr;
3135 unsigned HalfWidth = Width / 2;
3136
3137 // Canonicalize zext (lower half) to LHS.
3138 if (!isa<ZExtInst>(Op0))
3139 std::swap(Op0, Op1);
3140
3141 // Find lower/upper half.
3142 Value *LowerSrc, *ShlVal, *UpperSrc;
3143 const APInt *C;
3144 if (!match(Op0, m_OneUse(m_ZExt(m_Value(LowerSrc)))) ||
3145 !match(Op1, m_OneUse(m_Shl(m_Value(ShlVal), m_APInt(C)))) ||
3146 !match(ShlVal, m_OneUse(m_ZExt(m_Value(UpperSrc)))))
3147 return nullptr;
3148 if (*C != HalfWidth || LowerSrc->getType() != UpperSrc->getType() ||
3149 LowerSrc->getType()->getScalarSizeInBits() != HalfWidth)
3150 return nullptr;
3151
3152 auto ConcatIntrinsicCalls = [&](Intrinsic::ID id, Value *Lo, Value *Hi) {
3153 Value *NewLower = Builder.CreateZExt(Lo, Ty);
3154 Value *NewUpper = Builder.CreateZExt(Hi, Ty);
3155 NewUpper = Builder.CreateShl(NewUpper, HalfWidth);
3156 Value *BinOp = Builder.CreateOr(NewLower, NewUpper);
3157 return Builder.CreateIntrinsic(id, Ty, BinOp);
3158 };
3159
3160 // BSWAP: Push the concat down, swapping the lower/upper sources.
3161 // concat(bswap(x),bswap(y)) -> bswap(concat(x,y))
3162 Value *LowerBSwap, *UpperBSwap;
3163 if (match(LowerSrc, m_BSwap(m_Value(LowerBSwap))) &&
3164 match(UpperSrc, m_BSwap(m_Value(UpperBSwap))))
3165 return ConcatIntrinsicCalls(Intrinsic::bswap, UpperBSwap, LowerBSwap);
3166
3167 // BITREVERSE: Push the concat down, swapping the lower/upper sources.
3168 // concat(bitreverse(x),bitreverse(y)) -> bitreverse(concat(x,y))
3169 Value *LowerBRev, *UpperBRev;
3170 if (match(LowerSrc, m_BitReverse(m_Value(LowerBRev))) &&
3171 match(UpperSrc, m_BitReverse(m_Value(UpperBRev))))
3172 return ConcatIntrinsicCalls(Intrinsic::bitreverse, UpperBRev, LowerBRev);
3173
3174 // iX ext split: extending or(zext(x),shl(zext(y),bw/2) pattern
3175 // to consume sext/ashr:
3176 // or(zext(sext(x)),shl(zext(sext(ashr(x,xbw-1))),bw/2)
3177 // or(zext(x),shl(zext(ashr(x,xbw-1)),bw/2)
3178 Value *X;
3179 if (match(LowerSrc, m_SExtOrSelf(m_Value(X))) &&
3180 match(UpperSrc,
3182 m_Specific(X),
3183 m_SpecificInt(X->getType()->getScalarSizeInBits() - 1)))))
3184 return Builder.CreateSExt(X, Ty);
3185
3186 return nullptr;
3187}
3188
3189/// If all elements of two constant vectors are 0/-1 and inverses, return true.
3191 unsigned NumElts = cast<FixedVectorType>(C1->getType())->getNumElements();
3192 for (unsigned i = 0; i != NumElts; ++i) {
3193 Constant *EltC1 = C1->getAggregateElement(i);
3194 Constant *EltC2 = C2->getAggregateElement(i);
3195 if (!EltC1 || !EltC2)
3196 return false;
3197
3198 // One element must be all ones, and the other must be all zeros.
3199 if (!((match(EltC1, m_Zero()) && match(EltC2, m_AllOnes())) ||
3200 (match(EltC2, m_Zero()) && match(EltC1, m_AllOnes()))))
3201 return false;
3202 }
3203 return true;
3204}
3205
3206/// We have an expression of the form (A & C) | (B & D). If A is a scalar or
3207/// vector composed of all-zeros or all-ones values and is the bitwise 'not' of
3208/// B, it can be used as the condition operand of a select instruction.
3209/// We will detect (A & C) | ~(B | D) when the flag ABIsTheSame enabled.
3210Value *InstCombinerImpl::getSelectCondition(Value *A, Value *B,
3211 bool ABIsTheSame) {
3212 // We may have peeked through bitcasts in the caller.
3213 // Exit immediately if we don't have (vector) integer types.
3214 Type *Ty = A->getType();
3215 if (!Ty->isIntOrIntVectorTy() || !B->getType()->isIntOrIntVectorTy())
3216 return nullptr;
3217
3218 // If A is the 'not' operand of B and has enough signbits, we have our answer.
3219 if (ABIsTheSame ? (A == B) : match(B, m_Not(m_Specific(A)))) {
3220 // If these are scalars or vectors of i1, A can be used directly.
3221 if (Ty->isIntOrIntVectorTy(1))
3222 return A;
3223
3224 // If we look through a vector bitcast, the caller will bitcast the operands
3225 // to match the condition's number of bits (N x i1).
3226 // To make this poison-safe, disallow bitcast from wide element to narrow
3227 // element. That could allow poison in lanes where it was not present in the
3228 // original code.
3230 if (A->getType()->isIntOrIntVectorTy()) {
3231 unsigned NumSignBits = ComputeNumSignBits(A);
3232 if (NumSignBits == A->getType()->getScalarSizeInBits() &&
3233 NumSignBits <= Ty->getScalarSizeInBits())
3234 return Builder.CreateTrunc(A, CmpInst::makeCmpResultType(A->getType()));
3235 }
3236 return nullptr;
3237 }
3238
3239 // TODO: add support for sext and constant case
3240 if (ABIsTheSame)
3241 return nullptr;
3242
3243 // If both operands are constants, see if the constants are inverse bitmasks.
3244 Constant *AConst, *BConst;
3245 if (match(A, m_Constant(AConst)) && match(B, m_Constant(BConst)))
3246 if (AConst == ConstantExpr::getNot(BConst) &&
3248 return Builder.CreateZExtOrTrunc(A, CmpInst::makeCmpResultType(Ty));
3249
3250 // Look for more complex patterns. The 'not' op may be hidden behind various
3251 // casts. Look through sexts and bitcasts to find the booleans.
3252 Value *Cond;
3253 Value *NotB;
3254 if (match(A, m_SExt(m_Value(Cond))) &&
3255 Cond->getType()->isIntOrIntVectorTy(1)) {
3256 // A = sext i1 Cond; B = sext (not (i1 Cond))
3257 if (match(B, m_SExt(m_Not(m_Specific(Cond)))))
3258 return Cond;
3259
3260 // A = sext i1 Cond; B = not ({bitcast} (sext (i1 Cond)))
3261 // TODO: The one-use checks are unnecessary or misplaced. If the caller
3262 // checked for uses on logic ops/casts, that should be enough to
3263 // make this transform worthwhile.
3264 if (match(B, m_OneUse(m_Not(m_Value(NotB))))) {
3265 NotB = peekThroughBitcast(NotB, true);
3266 if (match(NotB, m_SExt(m_Specific(Cond))))
3267 return Cond;
3268 }
3269 }
3270
3271 // All scalar (and most vector) possibilities should be handled now.
3272 // Try more matches that only apply to non-splat constant vectors.
3273 if (!Ty->isVectorTy())
3274 return nullptr;
3275
3276 // If both operands are xor'd with constants using the same sexted boolean
3277 // operand, see if the constants are inverse bitmasks.
3278 // TODO: Use ConstantExpr::getNot()?
3279 if (match(A, (m_Xor(m_SExt(m_Value(Cond)), m_Constant(AConst)))) &&
3280 match(B, (m_Xor(m_SExt(m_Specific(Cond)), m_Constant(BConst)))) &&
3281 Cond->getType()->isIntOrIntVectorTy(1) &&
3282 areInverseVectorBitmasks(AConst, BConst)) {
3284 return Builder.CreateXor(Cond, AConst);
3285 }
3286 return nullptr;
3287}
3288
3289/// We have an expression of the form (A & B) | (C & D). Try to simplify this
3290/// to "A' ? B : D", where A' is a boolean or vector of booleans.
3291/// When InvertFalseVal is set to true, we try to match the pattern
3292/// where we have peeked through a 'not' op and A and C are the same:
3293/// (A & B) | ~(A | D) --> (A & B) | (~A & ~D) --> A' ? B : ~D
3294Value *InstCombinerImpl::matchSelectFromAndOr(Value *A, Value *B, Value *C,
3295 Value *D, bool InvertFalseVal) {
3296 // The potential condition of the select may be bitcasted. In that case, look
3297 // through its bitcast and the corresponding bitcast of the 'not' condition.
3298 Type *OrigType = A->getType();
3299 A = peekThroughBitcast(A, true);
3300 C = peekThroughBitcast(C, true);
3301 if (Value *Cond = getSelectCondition(A, C, InvertFalseVal)) {
3302 // ((bc Cond) & B) | ((bc ~Cond) & D) --> bc (select Cond, (bc B), (bc D))
3303 // If this is a vector, we may need to cast to match the condition's length.
3304 // The bitcasts will either all exist or all not exist. The builder will
3305 // not create unnecessary casts if the types already match.
3306 Type *SelTy = A->getType();
3307 if (auto *VecTy = dyn_cast<VectorType>(Cond->getType())) {
3308 // For a fixed or scalable vector get N from <{vscale x} N x iM>
3309 unsigned Elts = VecTy->getElementCount().getKnownMinValue();
3310 // For a fixed or scalable vector, get the size in bits of N x iM; for a
3311 // scalar this is just M.
3312 unsigned SelEltSize = SelTy->getPrimitiveSizeInBits().getKnownMinValue();
3313 Type *EltTy = Builder.getIntNTy(SelEltSize / Elts);
3314 SelTy = VectorType::get(EltTy, VecTy->getElementCount());
3315 }
3316 Value *BitcastB = Builder.CreateBitCast(B, SelTy);
3317 if (InvertFalseVal)
3318 D = Builder.CreateNot(D);
3319 Value *BitcastD = Builder.CreateBitCast(D, SelTy);
3320 Value *Select = Builder.CreateSelect(Cond, BitcastB, BitcastD);
3321 return Builder.CreateBitCast(Select, OrigType);
3322 }
3323
3324 return nullptr;
3325}
3326
3327// (icmp eq X, C) | (icmp ult Other, (X - C)) -> (icmp ule Other, (X - (C + 1)))
3328// (icmp ne X, C) & (icmp uge Other, (X - C)) -> (icmp ugt Other, (X - (C + 1)))
3330 bool IsAnd, bool IsLogical,
3331 IRBuilderBase &Builder) {
3332 Value *LHS0 = LHS->getOperand(0);
3333 Value *RHS0 = RHS->getOperand(0);
3334 Value *RHS1 = RHS->getOperand(1);
3335
3336 ICmpInst::Predicate LPred =
3337 IsAnd ? LHS->getInversePredicate() : LHS->getPredicate();
3338 ICmpInst::Predicate RPred =
3339 IsAnd ? RHS->getInversePredicate() : RHS->getPredicate();
3340
3341 const APInt *CInt;
3342 if (LPred != ICmpInst::ICMP_EQ ||
3343 !match(LHS->getOperand(1), m_APIntAllowPoison(CInt)) ||
3344 !LHS0->getType()->isIntOrIntVectorTy() ||
3345 !(LHS->hasOneUse() || RHS->hasOneUse()))
3346 return nullptr;
3347
3348 auto MatchRHSOp = [LHS0, CInt](const Value *RHSOp) {
3349 return match(RHSOp,
3350 m_Add(m_Specific(LHS0), m_SpecificIntAllowPoison(-*CInt))) ||
3351 (CInt->isZero() && RHSOp == LHS0);
3352 };
3353
3354 Value *Other;
3355 if (RPred == ICmpInst::ICMP_ULT && MatchRHSOp(RHS1))
3356 Other = RHS0;
3357 else if (RPred == ICmpInst::ICMP_UGT && MatchRHSOp(RHS0))
3358 Other = RHS1;
3359 else
3360 return nullptr;
3361
3362 if (IsLogical)
3363 Other = Builder.CreateFreeze(Other);
3364
3365 return Builder.CreateICmp(
3367 Builder.CreateSub(LHS0, ConstantInt::get(LHS0->getType(), *CInt + 1)),
3368 Other);
3369}
3370
3371/// Fold (icmp)&(icmp) or (icmp)|(icmp) if possible.
3372/// If IsLogical is true, then the and/or is in select form and the transform
3373/// must be poison-safe.
3374Value *InstCombinerImpl::foldAndOrOfICmps(ICmpInst *LHS, ICmpInst *RHS,
3375 Instruction &I, bool IsAnd,
3376 bool IsLogical) {
3377 const SimplifyQuery Q = SQ.getWithInstruction(&I);
3378
3379 ICmpInst::Predicate PredL = LHS->getPredicate(), PredR = RHS->getPredicate();
3380 Value *LHS0 = LHS->getOperand(0), *RHS0 = RHS->getOperand(0);
3381 Value *LHS1 = LHS->getOperand(1), *RHS1 = RHS->getOperand(1);
3382
3383 const APInt *LHSC = nullptr, *RHSC = nullptr;
3384 match(LHS1, m_APInt(LHSC));
3385 match(RHS1, m_APInt(RHSC));
3386
3387 // (icmp1 A, B) | (icmp2 A, B) --> (icmp3 A, B)
3388 // (icmp1 A, B) & (icmp2 A, B) --> (icmp3 A, B)
3389 if (predicatesFoldable(PredL, PredR)) {
3390 if (LHS0 == RHS1 && LHS1 == RHS0) {
3391 PredL = ICmpInst::getSwappedPredicate(PredL);
3392 std::swap(LHS0, LHS1);
3393 }
3394 if (LHS0 == RHS0 && LHS1 == RHS1) {
3395 unsigned Code = IsAnd ? getICmpCode(PredL) & getICmpCode(PredR)
3396 : getICmpCode(PredL) | getICmpCode(PredR);
3397 bool IsSigned = LHS->isSigned() || RHS->isSigned();
3398 return getNewICmpValue(Code, IsSigned, LHS0, LHS1, Builder);
3399 }
3400 }
3401
3402 if (Value *V =
3403 foldAndOrOfICmpEqConstantAndICmp(LHS, RHS, IsAnd, IsLogical, Builder))
3404 return V;
3405 // We can treat logical like bitwise here, because both operands are used on
3406 // the LHS, and as such poison from both will propagate.
3408 /*IsLogical*/ false, Builder))
3409 return V;
3410
3411 if (Value *V =
3412 foldAndOrOfICmpsWithConstEq(LHS, RHS, IsAnd, IsLogical, Builder, Q))
3413 return V;
3414 // We can convert this case to bitwise and, because both operands are used
3415 // on the LHS, and as such poison from both will propagate.
3416 if (Value *V = foldAndOrOfICmpsWithConstEq(RHS, LHS, IsAnd,
3417 /*IsLogical=*/false, Builder, Q)) {
3418 // If RHS is still used, we should drop samesign flag.
3419 if (IsLogical && RHS->hasSameSign() && !RHS->use_empty()) {
3420 RHS->setSameSign(false);
3422 }
3423 return V;
3424 }
3425
3426 if (Value *V = foldIsPowerOf2OrZero(LHS, RHS, IsAnd, Builder, *this))
3427 return V;
3428 if (Value *V = foldIsPowerOf2OrZero(RHS, LHS, IsAnd, Builder, *this))
3429 return V;
3430
3431 // TODO: One of these directions is fine with logical and/or, the other could
3432 // be supported by inserting freeze.
3433 if (!IsLogical) {
3434 // E.g. (icmp slt x, 0) | (icmp sgt x, n) --> icmp ugt x, n
3435 // E.g. (icmp sge x, 0) & (icmp slt x, n) --> icmp ult x, n
3436 if (Value *V = simplifyRangeCheck(LHS, RHS, /*Inverted=*/!IsAnd))
3437 return V;
3438
3439 // E.g. (icmp sgt x, n) | (icmp slt x, 0) --> icmp ugt x, n
3440 // E.g. (icmp slt x, n) & (icmp sge x, 0) --> icmp ult x, n
3441 if (Value *V = simplifyRangeCheck(RHS, LHS, /*Inverted=*/!IsAnd))
3442 return V;
3443 }
3444
3445 // TODO: Add conjugated or fold, check whether it is safe for logical and/or.
3446 if (IsAnd && !IsLogical)
3448 return V;
3449
3450 if (Value *V = foldIsPowerOf2(LHS, RHS, IsAnd, Builder, *this))
3451 return V;
3452
3453 if (Value *V = foldPowerOf2AndShiftedMask(LHS, RHS, IsAnd, Builder))
3454 return V;
3455
3456 // TODO: Verify whether this is safe for logical and/or.
3457 if (!IsLogical) {
3458 if (Value *X = foldUnsignedUnderflowCheck(LHS, RHS, IsAnd, Q, Builder))
3459 return X;
3460 if (Value *X = foldUnsignedUnderflowCheck(RHS, LHS, IsAnd, Q, Builder))
3461 return X;
3462 }
3463
3464 // (icmp ne A, 0) | (icmp ne B, 0) --> (icmp ne (A|B), 0)
3465 // (icmp eq A, 0) & (icmp eq B, 0) --> (icmp eq (A|B), 0)
3466 // TODO: Remove this and below when foldLogOpOfMaskedICmps can handle undefs.
3467 if (PredL == (IsAnd ? ICmpInst::ICMP_EQ : ICmpInst::ICMP_NE) &&
3468 PredL == PredR && match(LHS1, m_ZeroInt()) && match(RHS1, m_ZeroInt()) &&
3469 LHS0->getType() == RHS0->getType() &&
3470 (!IsLogical || isGuaranteedNotToBePoison(RHS0))) {
3471 Value *NewOr = Builder.CreateOr(LHS0, RHS0);
3472 return Builder.CreateICmp(PredL, NewOr,
3474 }
3475
3476 // (icmp ne A, -1) | (icmp ne B, -1) --> (icmp ne (A&B), -1)
3477 // (icmp eq A, -1) & (icmp eq B, -1) --> (icmp eq (A&B), -1)
3478 if (PredL == (IsAnd ? ICmpInst::ICMP_EQ : ICmpInst::ICMP_NE) &&
3479 PredL == PredR && match(LHS1, m_AllOnes()) && match(RHS1, m_AllOnes()) &&
3480 LHS0->getType() == RHS0->getType() &&
3481 (!IsLogical || isGuaranteedNotToBePoison(RHS0))) {
3482 Value *NewAnd = Builder.CreateAnd(LHS0, RHS0);
3483 return Builder.CreateICmp(PredL, NewAnd,
3485 }
3486
3487 if (!IsLogical)
3488 if (Value *V =
3490 return V;
3491
3492 // This only handles icmp of constants: (icmp1 A, C1) | (icmp2 B, C2).
3493 if (!LHSC || !RHSC)
3494 return nullptr;
3495
3496 // (trunc x) == C1 & (and x, CA) == C2 -> (and x, CA|CMAX) == C1|C2
3497 // (trunc x) != C1 | (and x, CA) != C2 -> (and x, CA|CMAX) != C1|C2
3498 // where CMAX is the all ones value for the truncated type,
3499 // iff the lower bits of C2 and CA are zero.
3500 if (PredL == (IsAnd ? ICmpInst::ICMP_EQ : ICmpInst::ICMP_NE) &&
3501 PredL == PredR && LHS->hasOneUse() && RHS->hasOneUse()) {
3502 Value *V;
3503 const APInt *AndC, *SmallC = nullptr, *BigC = nullptr;
3504
3505 // (trunc x) == C1 & (and x, CA) == C2
3506 // (and x, CA) == C2 & (trunc x) == C1
3507 if (match(RHS0, m_Trunc(m_Value(V))) &&
3508 match(LHS0, m_And(m_Specific(V), m_APInt(AndC)))) {
3509 SmallC = RHSC;
3510 BigC = LHSC;
3511 } else if (match(LHS0, m_Trunc(m_Value(V))) &&
3512 match(RHS0, m_And(m_Specific(V), m_APInt(AndC)))) {
3513 SmallC = LHSC;
3514 BigC = RHSC;
3515 }
3516
3517 if (SmallC && BigC) {
3518 unsigned BigBitSize = BigC->getBitWidth();
3519 unsigned SmallBitSize = SmallC->getBitWidth();
3520
3521 // Check that the low bits are zero.
3522 APInt Low = APInt::getLowBitsSet(BigBitSize, SmallBitSize);
3523 if ((Low & *AndC).isZero() && (Low & *BigC).isZero()) {
3524 Value *NewAnd = Builder.CreateAnd(V, Low | *AndC);
3525 APInt N = SmallC->zext(BigBitSize) | *BigC;
3526 Value *NewVal = ConstantInt::get(NewAnd->getType(), N);
3527 return Builder.CreateICmp(PredL, NewAnd, NewVal);
3528 }
3529 }
3530 }
3531
3532 // Match naive pattern (and its inverted form) for checking if two values
3533 // share same sign. An example of the pattern:
3534 // (icmp slt (X & Y), 0) | (icmp sgt (X | Y), -1) -> (icmp sgt (X ^ Y), -1)
3535 // Inverted form (example):
3536 // (icmp slt (X | Y), 0) & (icmp sgt (X & Y), -1) -> (icmp slt (X ^ Y), 0)
3537 bool TrueIfSignedL, TrueIfSignedR;
3538 if (isSignBitCheck(PredL, *LHSC, TrueIfSignedL) &&
3539 isSignBitCheck(PredR, *RHSC, TrueIfSignedR) &&
3540 (RHS->hasOneUse() || LHS->hasOneUse())) {
3541 Value *X, *Y;
3542 if (IsAnd) {
3543 if ((TrueIfSignedL && !TrueIfSignedR &&
3544 match(LHS0, m_Or(m_Value(X), m_Value(Y))) &&
3545 match(RHS0, m_c_And(m_Specific(X), m_Specific(Y)))) ||
3546 (!TrueIfSignedL && TrueIfSignedR &&
3547 match(LHS0, m_And(m_Value(X), m_Value(Y))) &&
3548 match(RHS0, m_c_Or(m_Specific(X), m_Specific(Y))))) {
3549 Value *NewXor = Builder.CreateXor(X, Y);
3550 return Builder.CreateIsNeg(NewXor);
3551 }
3552 } else {
3553 if ((TrueIfSignedL && !TrueIfSignedR &&
3554 match(LHS0, m_And(m_Value(X), m_Value(Y))) &&
3555 match(RHS0, m_c_Or(m_Specific(X), m_Specific(Y)))) ||
3556 (!TrueIfSignedL && TrueIfSignedR &&
3557 match(LHS0, m_Or(m_Value(X), m_Value(Y))) &&
3558 match(RHS0, m_c_And(m_Specific(X), m_Specific(Y))))) {
3559 Value *NewXor = Builder.CreateXor(X, Y);
3560 return Builder.CreateIsNotNeg(NewXor);
3561 }
3562 }
3563 }
3564
3565 // (X & ExpMask) != 0 && (X & ExpMask) != ExpMask -> isnormal(X)
3566 // (X & ExpMask) == 0 || (X & ExpMask) == ExpMask -> !isnormal(X)
3567 Value *X;
3568 const APInt *MaskC;
3569 if (LHS0 == RHS0 && PredL == PredR &&
3570 PredL == (IsAnd ? ICmpInst::ICMP_NE : ICmpInst::ICMP_EQ) &&
3571 !I.getFunction()->hasFnAttribute(Attribute::NoImplicitFloat) &&
3572 LHS->hasOneUse() && RHS->hasOneUse() &&
3573 match(LHS0, m_And(m_ElementWiseBitCast(m_Value(X)), m_APInt(MaskC))) &&
3574 X->getType()->getScalarType()->isIEEELikeFPTy() &&
3575 APFloat(X->getType()->getScalarType()->getFltSemantics(), *MaskC)
3576 .isPosInfinity() &&
3577 ((LHSC->isZero() && *RHSC == *MaskC) ||
3578 (RHSC->isZero() && *LHSC == *MaskC)))
3579 return Builder.createIsFPClass(X, IsAnd ? FPClassTest::fcNormal
3581
3582 return foldAndOrOfICmpsUsingRanges(LHS, RHS, IsAnd);
3583}
3584
3585/// If IsLogical is true, then the and/or is in select form and the transform
3586/// must be poison-safe.
3587Value *InstCombinerImpl::foldBooleanAndOr(Value *LHS, Value *RHS,
3588 Instruction &I, bool IsAnd,
3589 bool IsLogical) {
3590 if (!LHS->getType()->isIntOrIntVectorTy(1))
3591 return nullptr;
3592
3593 // handle (roughly):
3594 // (icmp ne (A & B), C) | (icmp ne (A & D), E)
3595 // (icmp eq (A & B), C) & (icmp eq (A & D), E)
3596 if (Value *V = foldLogOpOfMaskedICmps(LHS, RHS, IsAnd, IsLogical, Builder,
3597 SQ.getWithInstruction(&I)))
3598 return V;
3599
3600 if (auto *LHSCmp = dyn_cast<ICmpInst>(LHS))
3601 if (auto *RHSCmp = dyn_cast<ICmpInst>(RHS))
3602 if (Value *Res = foldAndOrOfICmps(LHSCmp, RHSCmp, I, IsAnd, IsLogical))
3603 return Res;
3604
3605 if (auto *LHSCmp = dyn_cast<FCmpInst>(LHS))
3606 if (auto *RHSCmp = dyn_cast<FCmpInst>(RHS))
3607 if (Value *Res = foldLogicOfFCmps(LHSCmp, RHSCmp, IsAnd, IsLogical))
3608 return Res;
3609
3610 if (Value *Res = foldEqOfParts(LHS, RHS, IsAnd))
3611 return Res;
3612
3613 return nullptr;
3614}
3615
3617 InstCombiner::BuilderTy &Builder) {
3618 assert(I.getOpcode() == Instruction::Or &&
3619 "Simplification only supports or at the moment.");
3620
3621 Value *Cmp1, *Cmp2, *Cmp3, *Cmp4;
3622 if (!match(I.getOperand(0), m_And(m_Value(Cmp1), m_Value(Cmp2))) ||
3623 !match(I.getOperand(1), m_And(m_Value(Cmp3), m_Value(Cmp4))))
3624 return nullptr;
3625
3626 // Check if any two pairs of the and operations are inversions of each other.
3627 if (isKnownInversion(Cmp1, Cmp3) && isKnownInversion(Cmp2, Cmp4))
3628 return Builder.CreateXor(Cmp1, Cmp4);
3629 if (isKnownInversion(Cmp1, Cmp4) && isKnownInversion(Cmp2, Cmp3))
3630 return Builder.CreateXor(Cmp1, Cmp3);
3631
3632 return nullptr;
3633}
3634
3635/// Match \p V as "shufflevector -> bitcast" or "extractelement -> zext -> shl"
3636/// patterns, which extract vector elements and pack them in the same relative
3637/// positions.
3638///
3639/// \p Vec is the underlying vector being extracted from.
3640/// \p Mask is a bitmask identifying which packed elements are obtained from the
3641/// vector.
3642/// \p VecOffset is the vector element corresponding to index 0 of the
3643/// mask.
3645 int64_t &VecOffset,
3646 SmallBitVector &Mask,
3647 const DataLayout &DL) {
3648 // First try to match extractelement -> zext -> shl
3649 uint64_t VecIdx, ShlAmt;
3651 m_ConstantInt(VecIdx))),
3652 ShlAmt))) {
3653 auto *VecTy = dyn_cast<FixedVectorType>(Vec->getType());
3654 if (!VecTy)
3655 return false;
3656 auto *EltTy = dyn_cast<IntegerType>(VecTy->getElementType());
3657 if (!EltTy)
3658 return false;
3659
3660 const unsigned EltBitWidth = EltTy->getBitWidth();
3661 const unsigned TargetBitWidth = V->getType()->getIntegerBitWidth();
3662 if (TargetBitWidth % EltBitWidth != 0 || ShlAmt % EltBitWidth != 0)
3663 return false;
3664 const unsigned TargetEltWidth = TargetBitWidth / EltBitWidth;
3665 const unsigned ShlEltAmt = ShlAmt / EltBitWidth;
3666
3667 const unsigned MaskIdx =
3668 DL.isLittleEndian() ? ShlEltAmt : TargetEltWidth - ShlEltAmt - 1;
3669
3670 VecOffset = static_cast<int64_t>(VecIdx) - static_cast<int64_t>(MaskIdx);
3671 Mask.resize(TargetEltWidth);
3672 Mask.set(MaskIdx);
3673 return true;
3674 }
3675
3676 // Now try to match a bitcasted subvector.
3677 Instruction *SrcVecI;
3678 if (!match(V, m_BitCast(m_Instruction(SrcVecI))))
3679 return false;
3680
3681 auto *SrcTy = dyn_cast<FixedVectorType>(SrcVecI->getType());
3682 if (!SrcTy)
3683 return false;
3684
3685 Mask.resize(SrcTy->getNumElements());
3686
3687 // First check for a subvector obtained from a shufflevector.
3688 if (isa<ShuffleVectorInst>(SrcVecI)) {
3689 Constant *ConstVec;
3690 ArrayRef<int> ShuffleMask;
3691 if (!match(SrcVecI, m_Shuffle(m_Value(Vec), m_Constant(ConstVec),
3692 m_Mask(ShuffleMask))))
3693 return false;
3694
3695 auto *VecTy = dyn_cast<FixedVectorType>(Vec->getType());
3696 if (!VecTy)
3697 return false;
3698
3699 const unsigned NumVecElts = VecTy->getNumElements();
3700 bool FoundVecOffset = false;
3701 for (unsigned Idx = 0; Idx < ShuffleMask.size(); ++Idx) {
3702 if (ShuffleMask[Idx] == PoisonMaskElem)
3703 return false;
3704 const unsigned ShuffleIdx = ShuffleMask[Idx];
3705 if (ShuffleIdx >= NumVecElts) {
3706 const unsigned ConstIdx = ShuffleIdx - NumVecElts;
3707 auto *ConstElt =
3708 dyn_cast<ConstantInt>(ConstVec->getAggregateElement(ConstIdx));
3709 if (!ConstElt || !ConstElt->isNullValue())
3710 return false;
3711 continue;
3712 }
3713
3714 if (FoundVecOffset) {
3715 if (VecOffset + Idx != ShuffleIdx)
3716 return false;
3717 } else {
3718 if (ShuffleIdx < Idx)
3719 return false;
3720 VecOffset = ShuffleIdx - Idx;
3721 FoundVecOffset = true;
3722 }
3723 Mask.set(Idx);
3724 }
3725 return FoundVecOffset;
3726 }
3727
3728 // Check for a subvector obtained as an (insertelement V, 0, idx)
3729 uint64_t InsertIdx;
3730 if (!match(SrcVecI,
3731 m_InsertElt(m_Value(Vec), m_Zero(), m_ConstantInt(InsertIdx))))
3732 return false;
3733
3734 auto *VecTy = dyn_cast<FixedVectorType>(Vec->getType());
3735 if (!VecTy)
3736 return false;
3737 VecOffset = 0;
3738 bool AlreadyInsertedMaskedElt = Mask.test(InsertIdx);
3739 Mask.set();
3740 if (!AlreadyInsertedMaskedElt)
3741 Mask.reset(InsertIdx);
3742 return true;
3743}
3744
3745/// Try to fold the join of two scalar integers whose contents are packed
3746/// elements of the same vector.
3748 InstCombiner::BuilderTy &Builder,
3749 const DataLayout &DL) {
3750 assert(I.getOpcode() == Instruction::Or);
3751 Value *LhsVec, *RhsVec;
3752 int64_t LhsVecOffset, RhsVecOffset;
3753 SmallBitVector Mask;
3754 if (!matchSubIntegerPackFromVector(I.getOperand(0), LhsVec, LhsVecOffset,
3755 Mask, DL))
3756 return nullptr;
3757 if (!matchSubIntegerPackFromVector(I.getOperand(1), RhsVec, RhsVecOffset,
3758 Mask, DL))
3759 return nullptr;
3760 if (LhsVec != RhsVec || LhsVecOffset != RhsVecOffset)
3761 return nullptr;
3762
3763 // Convert into shufflevector -> bitcast;
3764 const unsigned ZeroVecIdx =
3765 cast<FixedVectorType>(LhsVec->getType())->getNumElements();
3766 SmallVector<int> ShuffleMask(Mask.size(), ZeroVecIdx);
3767 for (unsigned Idx : Mask.set_bits()) {
3768 assert(LhsVecOffset + Idx >= 0);
3769 ShuffleMask[Idx] = LhsVecOffset + Idx;
3770 }
3771
3772 Value *MaskedVec = Builder.CreateShuffleVector(
3773 LhsVec, Constant::getNullValue(LhsVec->getType()), ShuffleMask,
3774 I.getName() + ".v");
3775 return CastInst::Create(Instruction::BitCast, MaskedVec, I.getType());
3776}
3777
3778/// Match \p V as "lshr -> mask -> zext -> shl".
3779///
3780/// \p Int is the underlying integer being extracted from.
3781/// \p Mask is a bitmask identifying which bits of the integer are being
3782/// extracted. \p Offset identifies which bit of the result \p V corresponds to
3783/// the least significant bit of \p Int
3784static bool matchZExtedSubInteger(Value *V, Value *&Int, APInt &Mask,
3785 uint64_t &Offset, bool &IsShlNUW,
3786 bool &IsShlNSW) {
3787 Value *ShlOp0;
3788 uint64_t ShlAmt = 0;
3789 if (!match(V, m_OneUse(m_Shl(m_Value(ShlOp0), m_ConstantInt(ShlAmt)))))
3790 return false;
3791
3792 IsShlNUW = cast<BinaryOperator>(V)->hasNoUnsignedWrap();
3793 IsShlNSW = cast<BinaryOperator>(V)->hasNoSignedWrap();
3794
3795 Value *ZExtOp0;
3796 if (!match(ShlOp0, m_OneUse(m_ZExt(m_Value(ZExtOp0)))))
3797 return false;
3798
3799 Value *MaskedOp0;
3800 const APInt *ShiftedMaskConst = nullptr;
3801 if (!match(ZExtOp0, m_CombineOr(m_OneUse(m_And(m_Value(MaskedOp0),
3802 m_APInt(ShiftedMaskConst))),
3803 m_Value(MaskedOp0))))
3804 return false;
3805
3806 uint64_t LShrAmt = 0;
3807 if (!match(MaskedOp0,
3809 m_Value(Int))))
3810 return false;
3811
3812 if (LShrAmt > ShlAmt)
3813 return false;
3814 Offset = ShlAmt - LShrAmt;
3815
3816 Mask = ShiftedMaskConst ? ShiftedMaskConst->shl(LShrAmt)
3818 Int->getType()->getScalarSizeInBits(), LShrAmt);
3819
3820 return true;
3821}
3822
3823/// Try to fold the join of two scalar integers whose bits are unpacked and
3824/// zexted from the same source integer.
3826 InstCombiner::BuilderTy &Builder) {
3827
3828 Value *LhsInt, *RhsInt;
3829 APInt LhsMask, RhsMask;
3830 uint64_t LhsOffset, RhsOffset;
3831 bool IsLhsShlNUW, IsLhsShlNSW, IsRhsShlNUW, IsRhsShlNSW;
3832 if (!matchZExtedSubInteger(Lhs, LhsInt, LhsMask, LhsOffset, IsLhsShlNUW,
3833 IsLhsShlNSW))
3834 return nullptr;
3835 if (!matchZExtedSubInteger(Rhs, RhsInt, RhsMask, RhsOffset, IsRhsShlNUW,
3836 IsRhsShlNSW))
3837 return nullptr;
3838 if (LhsInt != RhsInt || LhsOffset != RhsOffset)
3839 return nullptr;
3840
3841 APInt Mask = LhsMask | RhsMask;
3842
3843 Type *DestTy = Lhs->getType();
3844 Value *Res = Builder.CreateShl(
3845 Builder.CreateZExt(
3846 Builder.CreateAnd(LhsInt, Mask, LhsInt->getName() + ".mask"), DestTy,
3847 LhsInt->getName() + ".zext"),
3848 ConstantInt::get(DestTy, LhsOffset), "", IsLhsShlNUW && IsRhsShlNUW,
3849 IsLhsShlNSW && IsRhsShlNSW);
3850 Res->takeName(Lhs);
3851 return Res;
3852}
3853
3854// A decomposition of ((X & Mask) * Factor). The NUW / NSW bools
3855// track these properities for preservation. Note that we can decompose
3856// equivalent select form of this expression (e.g. (!(X & Mask) ? 0 : Mask *
3857// Factor))
3862 bool NUW;
3863 bool NSW;
3864
3866 return X == Other.X && !Mask.intersects(Other.Mask) &&
3867 Factor == Other.Factor;
3868 }
3869};
3870
3871static std::optional<DecomposedBitMaskMul> matchBitmaskMul(Value *V) {
3873 if (!Op)
3874 return std::nullopt;
3875
3876 // Decompose (A & N) * C) into BitMaskMul
3877 Value *Original = nullptr;
3878 const APInt *Mask = nullptr;
3879 const APInt *MulConst = nullptr;
3880 if (match(Op, m_Mul(m_And(m_Value(Original), m_APInt(Mask)),
3881 m_APInt(MulConst)))) {
3882 if (MulConst->isZero() || Mask->isZero())
3883 return std::nullopt;
3884
3885 return std::optional<DecomposedBitMaskMul>(
3886 {Original, *MulConst, *Mask,
3887 cast<BinaryOperator>(Op)->hasNoUnsignedWrap(),
3888 cast<BinaryOperator>(Op)->hasNoSignedWrap()});
3889 }
3890
3891 Value *Cond = nullptr;
3892 const APInt *EqZero = nullptr, *NeZero = nullptr;
3893
3894 // Decompose ((A & N) ? 0 : N * C) into BitMaskMul
3895 if (match(Op, m_Select(m_Value(Cond), m_APInt(EqZero), m_APInt(NeZero)))) {
3896 auto ICmpDecompose =
3897 decomposeBitTest(Cond, /*LookThruTrunc=*/true,
3898 /*AllowNonZeroC=*/false, /*DecomposeBitMask=*/true);
3899 if (!ICmpDecompose.has_value())
3900 return std::nullopt;
3901
3902 assert(ICmpInst::isEquality(ICmpDecompose->Pred) &&
3903 ICmpDecompose->C.isZero());
3904
3905 if (ICmpDecompose->Pred == ICmpInst::ICMP_NE)
3906 std::swap(EqZero, NeZero);
3907
3908 if (!EqZero->isZero() || NeZero->isZero())
3909 return std::nullopt;
3910
3911 if (!ICmpDecompose->Mask.isPowerOf2() || ICmpDecompose->Mask.isZero() ||
3912 NeZero->getBitWidth() != ICmpDecompose->Mask.getBitWidth())
3913 return std::nullopt;
3914
3915 if (!NeZero->urem(ICmpDecompose->Mask).isZero())
3916 return std::nullopt;
3917
3918 return std::optional<DecomposedBitMaskMul>(
3919 {ICmpDecompose->X, NeZero->udiv(ICmpDecompose->Mask),
3920 ICmpDecompose->Mask, /*NUW=*/false, /*NSW=*/false});
3921 }
3922
3923 return std::nullopt;
3924}
3925
3926/// (A & N) * C + (A & M) * C -> (A & (N + M)) & C
3927/// This also accepts the equivalent select form of (A & N) * C
3928/// expressions i.e. !(A & N) ? 0 : N * C)
3929static Value *foldBitmaskMul(Value *Op0, Value *Op1,
3930 InstCombiner::BuilderTy &Builder) {
3931 auto Decomp1 = matchBitmaskMul(Op1);
3932 if (!Decomp1)
3933 return nullptr;
3934
3935 auto Decomp0 = matchBitmaskMul(Op0);
3936 if (!Decomp0)
3937 return nullptr;
3938
3939 if (Decomp0->isCombineableWith(*Decomp1)) {
3940 Value *NewAnd = Builder.CreateAnd(
3941 Decomp0->X,
3942 ConstantInt::get(Decomp0->X->getType(), Decomp0->Mask + Decomp1->Mask));
3943
3944 return Builder.CreateMul(
3945 NewAnd, ConstantInt::get(NewAnd->getType(), Decomp1->Factor), "",
3946 Decomp0->NUW && Decomp1->NUW, Decomp0->NSW && Decomp1->NSW);
3947 }
3948
3949 return nullptr;
3950}
3951
3952Value *InstCombinerImpl::foldDisjointOr(Value *LHS, Value *RHS) {
3953 if (Value *Res = foldBitmaskMul(LHS, RHS, Builder))
3954 return Res;
3956 return Res;
3957
3958 return nullptr;
3959}
3960
3961Value *InstCombinerImpl::reassociateDisjointOr(Value *LHS, Value *RHS) {
3962
3963 Value *X, *Y;
3965 if (Value *Res = foldDisjointOr(LHS, X))
3966 return Builder.CreateOr(Res, Y, "", /*IsDisjoint=*/true);
3967 if (Value *Res = foldDisjointOr(LHS, Y))
3968 return Builder.CreateOr(Res, X, "", /*IsDisjoint=*/true);
3969 }
3970
3972 if (Value *Res = foldDisjointOr(X, RHS))
3973 return Builder.CreateOr(Res, Y, "", /*IsDisjoint=*/true);
3974 if (Value *Res = foldDisjointOr(Y, RHS))
3975 return Builder.CreateOr(Res, X, "", /*IsDisjoint=*/true);
3976 }
3977
3978 return nullptr;
3979}
3980
3981/// Fold Res, Overflow = (umul.with.overflow x c1); (or Overflow (ugt Res c2))
3982/// --> (ugt x (c2/c1)). This code checks whether a multiplication of two
3983/// unsigned numbers (one is a constant) is mathematically greater than a
3984/// second constant.
3986 InstCombiner::BuilderTy &Builder,
3987 const DataLayout &DL) {
3988 Value *WOV, *X;
3989 const APInt *C1, *C2;
3990 if (match(&I,
3993 m_Value(X), m_APInt(C1)))),
3996 m_APInt(C2))))) &&
3997 !C1->isZero()) {
3998 Constant *NewC = ConstantInt::get(X->getType(), C2->udiv(*C1));
3999 return Builder.CreateICmp(ICmpInst::ICMP_UGT, X, NewC);
4000 }
4001 return nullptr;
4002}
4003
4004// FIXME: We use commutative matchers (m_c_*) for some, but not all, matches
4005// here. We should standardize that construct where it is needed or choose some
4006// other way to ensure that commutated variants of patterns are not missed.
4008 if (Value *V = simplifyOrInst(I.getOperand(0), I.getOperand(1),
4009 SQ.getWithInstruction(&I)))
4010 return replaceInstUsesWith(I, V);
4011
4013 return &I;
4014
4016 return X;
4017
4019 return Phi;
4020
4021 // See if we can simplify any instructions used by the instruction whose sole
4022 // purpose is to compute bits we don't care about.
4024 return &I;
4025
4026 // Do this before using distributive laws to catch simple and/or/not patterns.
4028 return Xor;
4029
4031 return X;
4032
4034 return X;
4035
4036 // (A & B) | (C & D) -> A ^ D where A == ~C && B == ~D
4037 // (A & B) | (C & D) -> A ^ C where A == ~D && B == ~C
4038 if (Value *V = foldOrOfInversions(I, Builder))
4039 return replaceInstUsesWith(I, V);
4040
4041 // (A&B)|(A&C) -> A&(B|C) etc
4043 return replaceInstUsesWith(I, V);
4044
4045 Value *Op0 = I.getOperand(0), *Op1 = I.getOperand(1);
4046 Type *Ty = I.getType();
4047 if (Ty->isIntOrIntVectorTy(1)) {
4048 if (auto *SI0 = dyn_cast<SelectInst>(Op0)) {
4049 if (auto *R =
4050 foldAndOrOfSelectUsingImpliedCond(Op1, *SI0, /* IsAnd */ false))
4051 return R;
4052 }
4053 if (auto *SI1 = dyn_cast<SelectInst>(Op1)) {
4054 if (auto *R =
4055 foldAndOrOfSelectUsingImpliedCond(Op0, *SI1, /* IsAnd */ false))
4056 return R;
4057 }
4058 }
4059
4060 if (Instruction *FoldedLogic = foldBinOpIntoSelectOrPhi(I))
4061 return FoldedLogic;
4062
4063 if (Instruction *BitOp = matchBSwapOrBitReverse(I, /*MatchBSwaps*/ true,
4064 /*MatchBitReversals*/ true))
4065 return BitOp;
4066
4067 if (Instruction *Funnel = matchFunnelShift(I, *this))
4068 return Funnel;
4069
4071 return replaceInstUsesWith(I, Concat);
4072
4074 return R;
4075
4077 return R;
4078
4079 if (cast<PossiblyDisjointInst>(I).isDisjoint()) {
4080 if (Instruction *R =
4081 foldAddLikeCommutative(I.getOperand(0), I.getOperand(1),
4082 /*NSW=*/true, /*NUW=*/true))
4083 return R;
4084 if (Instruction *R =
4085 foldAddLikeCommutative(I.getOperand(1), I.getOperand(0),
4086 /*NSW=*/true, /*NUW=*/true))
4087 return R;
4088
4089 if (Value *Res = foldDisjointOr(I.getOperand(0), I.getOperand(1)))
4090 return replaceInstUsesWith(I, Res);
4091
4092 if (Value *Res = reassociateDisjointOr(I.getOperand(0), I.getOperand(1)))
4093 return replaceInstUsesWith(I, Res);
4094 }
4095
4096 Value *X, *Y;
4097 const APInt *CV;
4098 if (match(&I, m_c_Or(m_OneUse(m_Xor(m_Value(X), m_APInt(CV))), m_Value(Y))) &&
4099 !CV->isAllOnes() && MaskedValueIsZero(Y, *CV, &I)) {
4100 // (X ^ C) | Y -> (X | Y) ^ C iff Y & C == 0
4101 // The check for a 'not' op is for efficiency (if Y is known zero --> ~X).
4102 Value *Or = Builder.CreateOr(X, Y);
4103 return BinaryOperator::CreateXor(Or, ConstantInt::get(Ty, *CV));
4104 }
4105
4106 // If the operands have no common bits set:
4107 // or (mul X, Y), X --> add (mul X, Y), X --> mul X, (Y + 1)
4109 m_Deferred(X)))) {
4110 Value *IncrementY = Builder.CreateAdd(Y, ConstantInt::get(Ty, 1));
4111 return BinaryOperator::CreateMul(X, IncrementY);
4112 }
4113
4114 // (A & C) | (B & D)
4115 Value *A, *B, *C, *D;
4116 if (match(Op0, m_And(m_Value(A), m_Value(C))) &&
4117 match(Op1, m_And(m_Value(B), m_Value(D)))) {
4118
4119 // (A & C0) | (B & C1)
4120 const APInt *C0, *C1;
4121 if (match(C, m_APInt(C0)) && match(D, m_APInt(C1))) {
4122 Value *X;
4123 if (*C0 == ~*C1) {
4124 // ((X | B) & MaskC) | (B & ~MaskC) -> (X & MaskC) | B
4125 if (match(A, m_c_Or(m_Value(X), m_Specific(B))))
4126 return BinaryOperator::CreateOr(Builder.CreateAnd(X, *C0), B);
4127 // (A & MaskC) | ((X | A) & ~MaskC) -> (X & ~MaskC) | A
4128 if (match(B, m_c_Or(m_Specific(A), m_Value(X))))
4129 return BinaryOperator::CreateOr(Builder.CreateAnd(X, *C1), A);
4130
4131 // ((X ^ B) & MaskC) | (B & ~MaskC) -> (X & MaskC) ^ B
4132 if (match(A, m_c_Xor(m_Value(X), m_Specific(B))))
4133 return BinaryOperator::CreateXor(Builder.CreateAnd(X, *C0), B);
4134 // (A & MaskC) | ((X ^ A) & ~MaskC) -> (X & ~MaskC) ^ A
4135 if (match(B, m_c_Xor(m_Specific(A), m_Value(X))))
4136 return BinaryOperator::CreateXor(Builder.CreateAnd(X, *C1), A);
4137 }
4138
4139 if ((*C0 & *C1).isZero()) {
4140 // ((X | B) & C0) | (B & C1) --> (X | B) & (C0 | C1)
4141 // iff (C0 & C1) == 0 and (X & ~C0) == 0
4142 if (match(A, m_c_Or(m_Value(X), m_Specific(B))) &&
4143 MaskedValueIsZero(X, ~*C0, &I)) {
4144 Constant *C01 = ConstantInt::get(Ty, *C0 | *C1);
4145 return BinaryOperator::CreateAnd(A, C01);
4146 }
4147 // (A & C0) | ((X | A) & C1) --> (X | A) & (C0 | C1)
4148 // iff (C0 & C1) == 0 and (X & ~C1) == 0
4149 if (match(B, m_c_Or(m_Value(X), m_Specific(A))) &&
4150 MaskedValueIsZero(X, ~*C1, &I)) {
4151 Constant *C01 = ConstantInt::get(Ty, *C0 | *C1);
4152 return BinaryOperator::CreateAnd(B, C01);
4153 }
4154 // ((X | C2) & C0) | ((X | C3) & C1) --> (X | C2 | C3) & (C0 | C1)
4155 // iff (C0 & C1) == 0 and (C2 & ~C0) == 0 and (C3 & ~C1) == 0.
4156 const APInt *C2, *C3;
4157 if (match(A, m_Or(m_Value(X), m_APInt(C2))) &&
4158 match(B, m_Or(m_Specific(X), m_APInt(C3))) &&
4159 (*C2 & ~*C0).isZero() && (*C3 & ~*C1).isZero()) {
4160 Value *Or = Builder.CreateOr(X, *C2 | *C3, "bitfield");
4161 Constant *C01 = ConstantInt::get(Ty, *C0 | *C1);
4162 return BinaryOperator::CreateAnd(Or, C01);
4163 }
4164 }
4165 }
4166
4167 // Don't try to form a select if it's unlikely that we'll get rid of at
4168 // least one of the operands. A select is generally more expensive than the
4169 // 'or' that it is replacing.
4170 if (Op0->hasOneUse() || Op1->hasOneUse()) {
4171 // (Cond & C) | (~Cond & D) -> Cond ? C : D, and commuted variants.
4172 if (Value *V = matchSelectFromAndOr(A, C, B, D))
4173 return replaceInstUsesWith(I, V);
4174 if (Value *V = matchSelectFromAndOr(A, C, D, B))
4175 return replaceInstUsesWith(I, V);
4176 if (Value *V = matchSelectFromAndOr(C, A, B, D))
4177 return replaceInstUsesWith(I, V);
4178 if (Value *V = matchSelectFromAndOr(C, A, D, B))
4179 return replaceInstUsesWith(I, V);
4180 if (Value *V = matchSelectFromAndOr(B, D, A, C))
4181 return replaceInstUsesWith(I, V);
4182 if (Value *V = matchSelectFromAndOr(B, D, C, A))
4183 return replaceInstUsesWith(I, V);
4184 if (Value *V = matchSelectFromAndOr(D, B, A, C))
4185 return replaceInstUsesWith(I, V);
4186 if (Value *V = matchSelectFromAndOr(D, B, C, A))
4187 return replaceInstUsesWith(I, V);
4188 }
4189 }
4190
4191 if (match(Op0, m_And(m_Value(A), m_Value(C))) &&
4192 match(Op1, m_Not(m_Or(m_Value(B), m_Value(D)))) &&
4193 (Op0->hasOneUse() || Op1->hasOneUse())) {
4194 // (Cond & C) | ~(Cond | D) -> Cond ? C : ~D
4195 if (Value *V = matchSelectFromAndOr(A, C, B, D, true))
4196 return replaceInstUsesWith(I, V);
4197 if (Value *V = matchSelectFromAndOr(A, C, D, B, true))
4198 return replaceInstUsesWith(I, V);
4199 if (Value *V = matchSelectFromAndOr(C, A, B, D, true))
4200 return replaceInstUsesWith(I, V);
4201 if (Value *V = matchSelectFromAndOr(C, A, D, B, true))
4202 return replaceInstUsesWith(I, V);
4203 }
4204
4205 // (A ^ B) | ((B ^ C) ^ A) -> (A ^ B) | C
4206 if (match(Op0, m_Xor(m_Value(A), m_Value(B))))
4207 if (match(Op1,
4210 return BinaryOperator::CreateOr(Op0, C);
4211
4212 // ((B ^ C) ^ A) | (A ^ B) -> (A ^ B) | C
4213 if (match(Op1, m_Xor(m_Value(A), m_Value(B))))
4214 if (match(Op0,
4217 return BinaryOperator::CreateOr(Op1, C);
4218
4219 if (Instruction *DeMorgan = matchDeMorgansLaws(I, *this))
4220 return DeMorgan;
4221
4222 // Canonicalize xor to the RHS.
4223 bool SwappedForXor = false;
4224 if (match(Op0, m_Xor(m_Value(), m_Value()))) {
4225 std::swap(Op0, Op1);
4226 SwappedForXor = true;
4227 }
4228
4229 if (match(Op1, m_Xor(m_Value(A), m_Value(B)))) {
4230 // (A | ?) | (A ^ B) --> (A | ?) | B
4231 // (B | ?) | (A ^ B) --> (B | ?) | A
4232 if (match(Op0, m_c_Or(m_Specific(A), m_Value())))
4233 return BinaryOperator::CreateOr(Op0, B);
4234 if (match(Op0, m_c_Or(m_Specific(B), m_Value())))
4235 return BinaryOperator::CreateOr(Op0, A);
4236
4237 // (A & B) | (A ^ B) --> A | B
4238 // (B & A) | (A ^ B) --> A | B
4239 if (match(Op0, m_c_And(m_Specific(A), m_Specific(B))))
4240 return BinaryOperator::CreateOr(A, B);
4241
4242 // ~A | (A ^ B) --> ~(A & B)
4243 // ~B | (A ^ B) --> ~(A & B)
4244 // The swap above should always make Op0 the 'not'.
4245 if ((Op0->hasOneUse() || Op1->hasOneUse()) &&
4246 (match(Op0, m_Not(m_Specific(A))) || match(Op0, m_Not(m_Specific(B)))))
4247 return BinaryOperator::CreateNot(Builder.CreateAnd(A, B));
4248
4249 // Same as above, but peek through an 'and' to the common operand:
4250 // ~(A & ?) | (A ^ B) --> ~((A & ?) & B)
4251 // ~(B & ?) | (A ^ B) --> ~((B & ?) & A)
4253 if ((Op0->hasOneUse() || Op1->hasOneUse()) &&
4254 match(Op0,
4256 return BinaryOperator::CreateNot(Builder.CreateAnd(And, B));
4257 if ((Op0->hasOneUse() || Op1->hasOneUse()) &&
4258 match(Op0,
4260 return BinaryOperator::CreateNot(Builder.CreateAnd(And, A));
4261
4262 // (~A | C) | (A ^ B) --> ~(A & B) | C
4263 // (~B | C) | (A ^ B) --> ~(A & B) | C
4264 if (Op0->hasOneUse() && Op1->hasOneUse() &&
4265 (match(Op0, m_c_Or(m_Not(m_Specific(A)), m_Value(C))) ||
4266 match(Op0, m_c_Or(m_Not(m_Specific(B)), m_Value(C))))) {
4267 Value *Nand = Builder.CreateNot(Builder.CreateAnd(A, B), "nand");
4268 return BinaryOperator::CreateOr(Nand, C);
4269 }
4270 }
4271
4272 if (SwappedForXor)
4273 std::swap(Op0, Op1);
4274
4275 if (Value *Res =
4276 foldBooleanAndOr(Op0, Op1, I, /*IsAnd=*/false, /*IsLogical=*/false))
4277 return replaceInstUsesWith(I, Res);
4278
4279 if (match(Op1, m_OneUse(m_LogicalOr(m_Value(X), m_Value(Y))))) {
4280 bool IsLogical = isa<SelectInst>(Op1);
4281 if (auto *V = reassociateBooleanAndOr(Op0, X, Y, I, /*IsAnd=*/false,
4282 /*RHSIsLogical=*/IsLogical))
4283 return replaceInstUsesWith(I, V);
4284 }
4285 if (match(Op0, m_OneUse(m_LogicalOr(m_Value(X), m_Value(Y))))) {
4286 bool IsLogical = isa<SelectInst>(Op0);
4287 if (auto *V = reassociateBooleanAndOr(Op1, X, Y, I, /*IsAnd=*/false,
4288 /*RHSIsLogical=*/IsLogical))
4289 return replaceInstUsesWith(I, V);
4290 }
4291
4292 if (Instruction *FoldedFCmps = reassociateFCmps(I, Builder))
4293 return FoldedFCmps;
4294
4295 if (Instruction *CastedOr = foldCastedBitwiseLogic(I))
4296 return CastedOr;
4297
4298 if (Instruction *Sel = foldBinopOfSextBoolToSelect(I))
4299 return Sel;
4300
4301 // or(sext(A), B) / or(B, sext(A)) --> A ? -1 : B, where A is i1 or <N x i1>.
4302 // TODO: Move this into foldBinopOfSextBoolToSelect as a more generalized fold
4303 // with binop identity constant. But creating a select with non-constant
4304 // arm may not be reversible due to poison semantics. Is that a good
4305 // canonicalization?
4306 if (match(&I, m_c_Or(m_OneUse(m_SExt(m_Value(A))), m_Value(B))) &&
4307 A->getType()->isIntOrIntVectorTy(1))
4309
4310 // Note: If we've gotten to the point of visiting the outer OR, then the
4311 // inner one couldn't be simplified. If it was a constant, then it won't
4312 // be simplified by a later pass either, so we try swapping the inner/outer
4313 // ORs in the hopes that we'll be able to simplify it this way.
4314 // (X|C) | V --> (X|V) | C
4315 // Pass the disjoint flag in the following two patterns:
4316 // 1. or-disjoint (or-disjoint X, C), V -->
4317 // or-disjoint (or-disjoint X, V), C
4318 //
4319 // 2. or-disjoint (or X, C), V -->
4320 // or (or-disjoint X, V), C
4321 ConstantInt *CI;
4322 if (Op0->hasOneUse() && !match(Op1, m_ConstantInt()) &&
4323 match(Op0, m_Or(m_Value(A), m_ConstantInt(CI)))) {
4324 bool IsDisjointOuter = cast<PossiblyDisjointInst>(I).isDisjoint();
4325 bool IsDisjointInner = cast<PossiblyDisjointInst>(Op0)->isDisjoint();
4326 Value *Inner = Builder.CreateOr(A, Op1);
4327 cast<PossiblyDisjointInst>(Inner)->setIsDisjoint(IsDisjointOuter);
4328 Inner->takeName(Op0);
4329 return IsDisjointOuter && IsDisjointInner
4330 ? BinaryOperator::CreateDisjointOr(Inner, CI)
4331 : BinaryOperator::CreateOr(Inner, CI);
4332 }
4333
4334 // Change (or (bool?A:B),(bool?C:D)) --> (bool?(or A,C):(or B,D))
4335 // Since this OR statement hasn't been optimized further yet, we hope
4336 // that this transformation will allow the new ORs to be optimized.
4337 {
4338 Value *X = nullptr, *Y = nullptr;
4339 if (Op0->hasOneUse() && Op1->hasOneUse() &&
4340 match(Op0, m_Select(m_Value(X), m_Value(A), m_Value(B))) &&
4341 match(Op1, m_Select(m_Value(Y), m_Value(C), m_Value(D))) && X == Y) {
4342 Value *orTrue = Builder.CreateOr(A, C);
4343 Value *orFalse = Builder.CreateOr(B, D);
4344 return SelectInst::Create(X, orTrue, orFalse);
4345 }
4346 }
4347
4348 // or(ashr(subNSW(Y, X), ScalarSizeInBits(Y) - 1), X) --> X s> Y ? -1 : X.
4349 {
4350 Value *X, *Y;
4353 m_SpecificInt(Ty->getScalarSizeInBits() - 1))),
4354 m_Deferred(X)))) {
4355 Value *NewICmpInst = Builder.CreateICmpSGT(X, Y);
4357 return SelectInst::Create(NewICmpInst, AllOnes, X);
4358 }
4359 }
4360
4361 {
4362 // ((A & B) ^ A) | ((A & B) ^ B) -> A ^ B
4363 // (A ^ (A & B)) | (B ^ (A & B)) -> A ^ B
4364 // ((A & B) ^ B) | ((A & B) ^ A) -> A ^ B
4365 // (B ^ (A & B)) | (A ^ (A & B)) -> A ^ B
4366 const auto TryXorOpt = [&](Value *Lhs, Value *Rhs) -> Instruction * {
4367 if (match(Lhs, m_c_Xor(m_And(m_Value(A), m_Value(B)), m_Deferred(A))) &&
4368 match(Rhs,
4370 return BinaryOperator::CreateXor(A, B);
4371 }
4372 return nullptr;
4373 };
4374
4375 if (Instruction *Result = TryXorOpt(Op0, Op1))
4376 return Result;
4377 if (Instruction *Result = TryXorOpt(Op1, Op0))
4378 return Result;
4379 }
4380
4381 if (Instruction *V =
4383 return V;
4384
4385 CmpPredicate Pred;
4386 Value *Mul, *Ov, *MulIsNotZero, *UMulWithOv;
4387 // Check if the OR weakens the overflow condition for umul.with.overflow by
4388 // treating any non-zero result as overflow. In that case, we overflow if both
4389 // umul.with.overflow operands are != 0, as in that case the result can only
4390 // be 0, iff the multiplication overflows.
4391 if (match(&I, m_c_Or(m_Value(Ov, m_ExtractValue<1>(m_Value(UMulWithOv))),
4392 m_Value(MulIsNotZero,
4396 m_Deferred(UMulWithOv))),
4397 m_ZeroInt())))) &&
4398 (Ov->hasOneUse() || (MulIsNotZero->hasOneUse() && Mul->hasOneUse()))) {
4399 Value *A, *B;
4401 m_Value(A), m_Value(B)))) {
4402 Value *NotNullA = Builder.CreateIsNotNull(A);
4403 Value *NotNullB = Builder.CreateIsNotNull(B);
4404 return BinaryOperator::CreateAnd(NotNullA, NotNullB);
4405 }
4406 }
4407
4408 /// Res, Overflow = xxx_with_overflow X, C1
4409 /// Try to canonicalize the pattern "Overflow | icmp pred Res, C2" into
4410 /// "Overflow | icmp pred X, C2 +/- C1".
4411 const WithOverflowInst *WO;
4412 const Value *WOV;
4413 const APInt *C1, *C2;
4415 m_Value(WOV, m_WithOverflowInst(WO)))),
4417 m_APInt(C2))))) &&
4418 (WO->getBinaryOp() == Instruction::Add ||
4419 WO->getBinaryOp() == Instruction::Sub) &&
4420 (ICmpInst::isEquality(Pred) ||
4421 WO->isSigned() == ICmpInst::isSigned(Pred)) &&
4422 match(WO->getRHS(), m_APInt(C1))) {
4423 bool Overflow;
4424 APInt NewC = WO->getBinaryOp() == Instruction::Add
4425 ? (ICmpInst::isSigned(Pred) ? C2->ssub_ov(*C1, Overflow)
4426 : C2->usub_ov(*C1, Overflow))
4427 : (ICmpInst::isSigned(Pred) ? C2->sadd_ov(*C1, Overflow)
4428 : C2->uadd_ov(*C1, Overflow));
4429 if (!Overflow || ICmpInst::isEquality(Pred)) {
4430 Value *NewCmp = Builder.CreateICmp(
4431 Pred, WO->getLHS(), ConstantInt::get(WO->getLHS()->getType(), NewC));
4432 return BinaryOperator::CreateOr(Ov, NewCmp);
4433 }
4434 }
4435
4436 // Try to fold the pattern "Overflow | icmp pred Res, C2" into a single
4437 // comparison instruction for umul.with.overflow.
4439 return replaceInstUsesWith(I, R);
4440
4441 // (~x) | y --> ~(x & (~y)) iff that gets rid of inversions
4443 return &I;
4444
4445 // Improve "get low bit mask up to and including bit X" pattern:
4446 // (1 << X) | ((1 << X) + -1) --> -1 l>> (bitwidth(x) - 1 - X)
4447 if (match(&I, m_c_Or(m_Add(m_Shl(m_One(), m_Value(X)), m_AllOnes()),
4448 m_Shl(m_One(), m_Deferred(X)))) &&
4449 match(&I, m_c_Or(m_OneUse(m_Value()), m_Value()))) {
4450 Value *Sub = Builder.CreateSub(
4451 ConstantInt::get(Ty, Ty->getScalarSizeInBits() - 1), X);
4452 return BinaryOperator::CreateLShr(Constant::getAllOnesValue(Ty), Sub);
4453 }
4454
4455 // An or recurrence w/loop invariant step is equivelent to (or start, step)
4456 PHINode *PN = nullptr;
4457 Value *Start = nullptr, *Step = nullptr;
4458 if (matchSimpleRecurrence(&I, PN, Start, Step) && DT.dominates(Step, PN))
4459 return replaceInstUsesWith(I, Builder.CreateOr(Start, Step));
4460
4461 // (A & B) | (C | D) or (C | D) | (A & B)
4462 // Can be combined if C or D is of type (A/B & X)
4464 m_OneUse(m_Or(m_Value(C), m_Value(D)))))) {
4465 // (A & B) | (C | ?) -> C | (? | (A & B))
4466 // (A & B) | (C | ?) -> C | (? | (A & B))
4467 // (A & B) | (C | ?) -> C | (? | (A & B))
4468 // (A & B) | (C | ?) -> C | (? | (A & B))
4469 // (C | ?) | (A & B) -> C | (? | (A & B))
4470 // (C | ?) | (A & B) -> C | (? | (A & B))
4471 // (C | ?) | (A & B) -> C | (? | (A & B))
4472 // (C | ?) | (A & B) -> C | (? | (A & B))
4473 if (match(D, m_OneUse(m_c_And(m_Specific(A), m_Value()))) ||
4475 return BinaryOperator::CreateOr(
4476 C, Builder.CreateOr(D, Builder.CreateAnd(A, B)));
4477 // (A & B) | (? | D) -> (? | (A & B)) | D
4478 // (A & B) | (? | D) -> (? | (A & B)) | D
4479 // (A & B) | (? | D) -> (? | (A & B)) | D
4480 // (A & B) | (? | D) -> (? | (A & B)) | D
4481 // (? | D) | (A & B) -> (? | (A & B)) | D
4482 // (? | D) | (A & B) -> (? | (A & B)) | D
4483 // (? | D) | (A & B) -> (? | (A & B)) | D
4484 // (? | D) | (A & B) -> (? | (A & B)) | D
4485 if (match(C, m_OneUse(m_c_And(m_Specific(A), m_Value()))) ||
4487 return BinaryOperator::CreateOr(
4488 Builder.CreateOr(C, Builder.CreateAnd(A, B)), D);
4489 }
4490
4492 return R;
4493
4494 if (Instruction *Canonicalized = canonicalizeLogicFirst(I, Builder))
4495 return Canonicalized;
4496
4497 if (Instruction *Folded = foldLogicOfIsFPClass(I, Op0, Op1))
4498 return Folded;
4499
4500 if (Instruction *Res = foldBinOpOfDisplacedShifts(I))
4501 return Res;
4502
4503 // If we are setting the sign bit of a floating-point value, convert
4504 // this to fneg(fabs), then cast back to integer.
4505 //
4506 // If the result isn't immediately cast back to a float, this will increase
4507 // the number of instructions. This is still probably a better canonical form
4508 // as it enables FP value tracking.
4509 //
4510 // Assumes any IEEE-represented type has the sign bit in the high bit.
4511 //
4512 // This is generous interpretation of noimplicitfloat, this is not a true
4513 // floating-point operation.
4514 Value *CastOp;
4515 if (match(Op0, m_ElementWiseBitCast(m_Value(CastOp))) &&
4516 match(Op1, m_SignMask()) &&
4517 !Builder.GetInsertBlock()->getParent()->hasFnAttribute(
4518 Attribute::NoImplicitFloat)) {
4519 Type *EltTy = CastOp->getType()->getScalarType();
4520 if (EltTy->isFloatingPointTy() &&
4522 Value *FAbs = Builder.CreateUnaryIntrinsic(Intrinsic::fabs, CastOp);
4523 Value *FNegFAbs = Builder.CreateFNeg(FAbs);
4524 return new BitCastInst(FNegFAbs, I.getType());
4525 }
4526 }
4527
4528 // (X & C1) | C2 -> X & (C1 | C2) iff (X & C2) == C2
4529 if (match(Op0, m_OneUse(m_And(m_Value(X), m_APInt(C1)))) &&
4530 match(Op1, m_APInt(C2))) {
4531 KnownBits KnownX = computeKnownBits(X, &I);
4532 if ((KnownX.One & *C2) == *C2)
4533 return BinaryOperator::CreateAnd(X, ConstantInt::get(Ty, *C1 | *C2));
4534 }
4535
4537 return Res;
4538
4539 if (Value *V =
4541 /*SimplifyOnly*/ false, *this))
4542 return BinaryOperator::CreateOr(V, Op1);
4543 if (Value *V =
4545 /*SimplifyOnly*/ false, *this))
4546 return BinaryOperator::CreateOr(Op0, V);
4547
4548 if (cast<PossiblyDisjointInst>(I).isDisjoint())
4550 return replaceInstUsesWith(I, V);
4551
4552 return nullptr;
4553}
4554
4555/// A ^ B can be specified using other logic ops in a variety of patterns. We
4556/// can fold these early and efficiently by morphing an existing instruction.
4558 InstCombiner::BuilderTy &Builder) {
4559 assert(I.getOpcode() == Instruction::Xor);
4560 Value *Op0 = I.getOperand(0);
4561 Value *Op1 = I.getOperand(1);
4562 Value *A, *B;
4563
4564 // There are 4 commuted variants for each of the basic patterns.
4565
4566 // (A & B) ^ (A | B) -> A ^ B
4567 // (A & B) ^ (B | A) -> A ^ B
4568 // (A | B) ^ (A & B) -> A ^ B
4569 // (A | B) ^ (B & A) -> A ^ B
4570 if (match(&I, m_c_Xor(m_And(m_Value(A), m_Value(B)),
4572 return BinaryOperator::CreateXor(A, B);
4573
4574 // (A | ~B) ^ (~A | B) -> A ^ B
4575 // (~B | A) ^ (~A | B) -> A ^ B
4576 // (~A | B) ^ (A | ~B) -> A ^ B
4577 // (B | ~A) ^ (A | ~B) -> A ^ B
4578 if (match(&I, m_Xor(m_c_Or(m_Value(A), m_Not(m_Value(B))),
4580 return BinaryOperator::CreateXor(A, B);
4581
4582 // (A & ~B) ^ (~A & B) -> A ^ B
4583 // (~B & A) ^ (~A & B) -> A ^ B
4584 // (~A & B) ^ (A & ~B) -> A ^ B
4585 // (B & ~A) ^ (A & ~B) -> A ^ B
4586 if (match(&I, m_Xor(m_c_And(m_Value(A), m_Not(m_Value(B))),
4588 return BinaryOperator::CreateXor(A, B);
4589
4590 // For the remaining cases we need to get rid of one of the operands.
4591 if (!Op0->hasOneUse() && !Op1->hasOneUse())
4592 return nullptr;
4593
4594 // (A | B) ^ ~(A & B) -> ~(A ^ B)
4595 // (A | B) ^ ~(B & A) -> ~(A ^ B)
4596 // (A & B) ^ ~(A | B) -> ~(A ^ B)
4597 // (A & B) ^ ~(B | A) -> ~(A ^ B)
4598 // Complexity sorting ensures the not will be on the right side.
4599 if ((match(Op0, m_Or(m_Value(A), m_Value(B))) &&
4600 match(Op1, m_Not(m_c_And(m_Specific(A), m_Specific(B))))) ||
4601 (match(Op0, m_And(m_Value(A), m_Value(B))) &&
4603 return BinaryOperator::CreateNot(Builder.CreateXor(A, B));
4604
4605 return nullptr;
4606}
4607
4608Value *InstCombinerImpl::foldXorOfICmps(ICmpInst *LHS, ICmpInst *RHS,
4609 BinaryOperator &I) {
4610 assert(I.getOpcode() == Instruction::Xor && I.getOperand(0) == LHS &&
4611 I.getOperand(1) == RHS && "Should be 'xor' with these operands");
4612
4613 ICmpInst::Predicate PredL = LHS->getPredicate(), PredR = RHS->getPredicate();
4614 Value *LHS0 = LHS->getOperand(0), *LHS1 = LHS->getOperand(1);
4615 Value *RHS0 = RHS->getOperand(0), *RHS1 = RHS->getOperand(1);
4616
4617 if (predicatesFoldable(PredL, PredR)) {
4618 if (LHS0 == RHS1 && LHS1 == RHS0) {
4619 std::swap(LHS0, LHS1);
4620 PredL = ICmpInst::getSwappedPredicate(PredL);
4621 }
4622 if (LHS0 == RHS0 && LHS1 == RHS1) {
4623 // (icmp1 A, B) ^ (icmp2 A, B) --> (icmp3 A, B)
4624 unsigned Code = getICmpCode(PredL) ^ getICmpCode(PredR);
4625 bool IsSigned = LHS->isSigned() || RHS->isSigned();
4626 return getNewICmpValue(Code, IsSigned, LHS0, LHS1, Builder);
4627 }
4628 }
4629
4630 const APInt *LC, *RC;
4631 if (match(LHS1, m_APInt(LC)) && match(RHS1, m_APInt(RC)) &&
4632 LHS0->getType() == RHS0->getType() &&
4633 LHS0->getType()->isIntOrIntVectorTy()) {
4634 // Convert xor of signbit tests to signbit test of xor'd values:
4635 // (X > -1) ^ (Y > -1) --> (X ^ Y) < 0
4636 // (X < 0) ^ (Y < 0) --> (X ^ Y) < 0
4637 // (X > -1) ^ (Y < 0) --> (X ^ Y) > -1
4638 // (X < 0) ^ (Y > -1) --> (X ^ Y) > -1
4639 bool TrueIfSignedL, TrueIfSignedR;
4640 if ((LHS->hasOneUse() || RHS->hasOneUse()) &&
4641 isSignBitCheck(PredL, *LC, TrueIfSignedL) &&
4642 isSignBitCheck(PredR, *RC, TrueIfSignedR)) {
4643 Value *XorLR = Builder.CreateXor(LHS0, RHS0);
4644 return TrueIfSignedL == TrueIfSignedR ? Builder.CreateIsNeg(XorLR) :
4645 Builder.CreateIsNotNeg(XorLR);
4646 }
4647
4648 // Fold (icmp pred1 X, C1) ^ (icmp pred2 X, C2)
4649 // into a single comparison using range-based reasoning.
4650 if (LHS0 == RHS0) {
4651 ConstantRange CR1 = ConstantRange::makeExactICmpRegion(PredL, *LC);
4652 ConstantRange CR2 = ConstantRange::makeExactICmpRegion(PredR, *RC);
4653 auto CRUnion = CR1.exactUnionWith(CR2);
4654 auto CRIntersect = CR1.exactIntersectWith(CR2);
4655 if (CRUnion && CRIntersect)
4656 if (auto CR = CRUnion->exactIntersectWith(CRIntersect->inverse())) {
4657 if (CR->isFullSet())
4658 return ConstantInt::getTrue(I.getType());
4659 if (CR->isEmptySet())
4660 return ConstantInt::getFalse(I.getType());
4661
4662 CmpInst::Predicate NewPred;
4663 APInt NewC, Offset;
4664 CR->getEquivalentICmp(NewPred, NewC, Offset);
4665
4666 if ((Offset.isZero() && (LHS->hasOneUse() || RHS->hasOneUse())) ||
4667 (LHS->hasOneUse() && RHS->hasOneUse())) {
4668 Value *NewV = LHS0;
4669 Type *Ty = LHS0->getType();
4670 if (!Offset.isZero())
4671 NewV = Builder.CreateAdd(NewV, ConstantInt::get(Ty, Offset));
4672 return Builder.CreateICmp(NewPred, NewV,
4673 ConstantInt::get(Ty, NewC));
4674 }
4675 }
4676 }
4677
4678 // Fold (icmp eq/ne (X & Pow2), 0) ^ (icmp eq/ne (Y & Pow2), 0) into
4679 // (icmp eq/ne ((X ^ Y) & Pow2), 0)
4680 Value *X, *Y, *Pow2;
4681 if (ICmpInst::isEquality(PredL) && ICmpInst::isEquality(PredR) &&
4682 LC->isZero() && RC->isZero() && LHS->hasOneUse() && RHS->hasOneUse() &&
4683 match(LHS0, m_And(m_Value(X), m_Value(Pow2))) &&
4684 match(RHS0, m_And(m_Value(Y), m_Specific(Pow2))) &&
4685 isKnownToBeAPowerOfTwo(Pow2, /*OrZero=*/true, &I)) {
4686 Value *Xor = Builder.CreateXor(X, Y);
4687 Value *And = Builder.CreateAnd(Xor, Pow2);
4688 return Builder.CreateICmp(PredL == PredR ? ICmpInst::ICMP_NE
4690 And, ConstantInt::getNullValue(Xor->getType()));
4691 }
4692 }
4693
4694 // Instead of trying to imitate the folds for and/or, decompose this 'xor'
4695 // into those logic ops. That is, try to turn this into an and-of-icmps
4696 // because we have many folds for that pattern.
4697 //
4698 // This is based on a truth table definition of xor:
4699 // X ^ Y --> (X | Y) & !(X & Y)
4700 if (Value *OrICmp = simplifyBinOp(Instruction::Or, LHS, RHS, SQ)) {
4701 // TODO: If OrICmp is true, then the definition of xor simplifies to !(X&Y).
4702 // TODO: If OrICmp is false, the whole thing is false (InstSimplify?).
4703 if (Value *AndICmp = simplifyBinOp(Instruction::And, LHS, RHS, SQ)) {
4704 // TODO: Independently handle cases where the 'and' side is a constant.
4705 ICmpInst *X = nullptr, *Y = nullptr;
4706 if (OrICmp == LHS && AndICmp == RHS) {
4707 // (LHS | RHS) & !(LHS & RHS) --> LHS & !RHS --> X & !Y
4708 X = LHS;
4709 Y = RHS;
4710 }
4711 if (OrICmp == RHS && AndICmp == LHS) {
4712 // !(LHS & RHS) & (LHS | RHS) --> !LHS & RHS --> !Y & X
4713 X = RHS;
4714 Y = LHS;
4715 }
4716 if (X && Y && (Y->hasOneUse() || canFreelyInvertAllUsersOf(Y, &I))) {
4717 // Invert the predicate of 'Y', thus inverting its output.
4718 Y->setPredicate(Y->getInversePredicate());
4719 // So, are there other uses of Y?
4720 if (!Y->hasOneUse()) {
4721 // We need to adapt other uses of Y though. Get a value that matches
4722 // the original value of Y before inversion. While this increases
4723 // immediate instruction count, we have just ensured that all the
4724 // users are freely-invertible, so that 'not' *will* get folded away.
4726 // Set insertion point to right after the Y.
4727 Builder.SetInsertPoint(Y->getParent(), ++(Y->getIterator()));
4728 Value *NotY = Builder.CreateNot(Y, Y->getName() + ".not");
4729 // Replace all uses of Y (excluding the one in NotY!) with NotY.
4730 Worklist.pushUsersToWorkList(*Y);
4731 Y->replaceUsesWithIf(NotY,
4732 [NotY](Use &U) { return U.getUser() != NotY; });
4733 }
4734 // All done.
4735 return Builder.CreateAnd(LHS, RHS);
4736 }
4737 }
4738 }
4739
4740 return nullptr;
4741}
4742
4743/// If we have a masked merge, in the canonical form of:
4744/// (assuming that A only has one use.)
4745/// | A | |B|
4746/// ((x ^ y) & M) ^ y
4747/// | D |
4748/// * If M is inverted:
4749/// | D |
4750/// ((x ^ y) & ~M) ^ y
4751/// We can canonicalize by swapping the final xor operand
4752/// to eliminate the 'not' of the mask.
4753/// ((x ^ y) & M) ^ x
4754/// * If M is a constant, and D has one use, we transform to 'and' / 'or' ops
4755/// because that shortens the dependency chain and improves analysis:
4756/// (x & M) | (y & ~M)
4758 InstCombiner::BuilderTy &Builder) {
4759 Value *B, *X, *D;
4760 Value *M;
4761 if (!match(&I, m_c_Xor(m_Value(B),
4764 m_Value(M))))))
4765 return nullptr;
4766
4767 Value *NotM;
4768 if (match(M, m_Not(m_Value(NotM)))) {
4769 // De-invert the mask and swap the value in B part.
4770 Value *NewA = Builder.CreateAnd(D, NotM);
4771 return BinaryOperator::CreateXor(NewA, X);
4772 }
4773
4774 Constant *C;
4775 if (D->hasOneUse() && match(M, m_Constant(C))) {
4776 // Propagating undef is unsafe. Clamp undef elements to -1.
4777 Type *EltTy = C->getType()->getScalarType();
4779 // Unfold.
4780 Value *LHS = Builder.CreateAnd(X, C);
4781 Value *NotC = Builder.CreateNot(C);
4782 Value *RHS = Builder.CreateAnd(B, NotC);
4783 return BinaryOperator::CreateOr(LHS, RHS);
4784 }
4785
4786 return nullptr;
4787}
4788
4790 InstCombiner::BuilderTy &Builder) {
4791 Value *X, *Y;
4792 // FIXME: one-use check is not needed in general, but currently we are unable
4793 // to fold 'not' into 'icmp', if that 'icmp' has multiple uses. (D35182)
4794 if (!match(&I, m_Not(m_OneUse(m_Xor(m_Value(X), m_Value(Y))))))
4795 return nullptr;
4796
4797 auto hasCommonOperand = [](Value *A, Value *B, Value *C, Value *D) {
4798 return A == C || A == D || B == C || B == D;
4799 };
4800
4801 Value *A, *B, *C, *D;
4802 // Canonicalize ~((A & B) ^ (A | ?)) -> (A & B) | ~(A | ?)
4803 // 4 commuted variants
4804 if (match(X, m_And(m_Value(A), m_Value(B))) &&
4805 match(Y, m_Or(m_Value(C), m_Value(D))) && hasCommonOperand(A, B, C, D)) {
4806 Value *NotY = Builder.CreateNot(Y);
4807 return BinaryOperator::CreateOr(X, NotY);
4808 };
4809
4810 // Canonicalize ~((A | ?) ^ (A & B)) -> (A & B) | ~(A | ?)
4811 // 4 commuted variants
4812 if (match(Y, m_And(m_Value(A), m_Value(B))) &&
4813 match(X, m_Or(m_Value(C), m_Value(D))) && hasCommonOperand(A, B, C, D)) {
4814 Value *NotX = Builder.CreateNot(X);
4815 return BinaryOperator::CreateOr(Y, NotX);
4816 };
4817
4818 return nullptr;
4819}
4820
4821/// Canonicalize a shifty way to code absolute value to the more common pattern
4822/// that uses negation and select.
4824 InstCombiner::BuilderTy &Builder) {
4825 assert(Xor.getOpcode() == Instruction::Xor && "Expected an xor instruction.");
4826
4827 // There are 4 potential commuted variants. Move the 'ashr' candidate to Op1.
4828 // We're relying on the fact that we only do this transform when the shift has
4829 // exactly 2 uses and the add has exactly 1 use (otherwise, we might increase
4830 // instructions).
4831 Value *Op0 = Xor.getOperand(0), *Op1 = Xor.getOperand(1);
4832 if (Op0->hasNUses(2))
4833 std::swap(Op0, Op1);
4834
4835 Type *Ty = Xor.getType();
4836 Value *A;
4837 const APInt *ShAmt;
4838 if (match(Op1, m_AShr(m_Value(A), m_APInt(ShAmt))) &&
4839 Op1->hasNUses(2) && *ShAmt == Ty->getScalarSizeInBits() - 1 &&
4840 match(Op0, m_OneUse(m_c_Add(m_Specific(A), m_Specific(Op1))))) {
4841 // Op1 = ashr i32 A, 31 ; smear the sign bit
4842 // xor (add A, Op1), Op1 ; add -1 and flip bits if negative
4843 // --> (A < 0) ? -A : A
4844 Value *IsNeg = Builder.CreateIsNeg(A);
4845 // Copy the nsw flags from the add to the negate.
4846 auto *Add = cast<BinaryOperator>(Op0);
4847 Value *NegA = Add->hasNoUnsignedWrap()
4848 ? Constant::getNullValue(A->getType())
4849 : Builder.CreateNeg(A, "", Add->hasNoSignedWrap());
4850 return SelectInst::Create(IsNeg, NegA, A);
4851 }
4852 return nullptr;
4853}
4854
4856 Instruction *IgnoredUser) {
4857 auto *I = dyn_cast<Instruction>(Op);
4858 return I && IC.isFreeToInvert(I, /*WillInvertAllUses=*/true) &&
4859 IC.canFreelyInvertAllUsersOf(I, IgnoredUser);
4860}
4861
4863 Instruction *IgnoredUser) {
4864 auto *I = cast<Instruction>(Op);
4865 IC.Builder.SetInsertPoint(*I->getInsertionPointAfterDef());
4866 Value *NotOp = IC.Builder.CreateNot(Op, Op->getName() + ".not");
4867 Op->replaceUsesWithIf(NotOp,
4868 [NotOp](Use &U) { return U.getUser() != NotOp; });
4869 IC.freelyInvertAllUsersOf(NotOp, IgnoredUser);
4870 return NotOp;
4871}
4872
4873// Transform
4874// z = ~(x &/| y)
4875// into:
4876// z = ((~x) |/& (~y))
4877// iff both x and y are free to invert and all uses of z can be freely updated.
4879 Value *Op0, *Op1;
4880 if (!match(&I, m_LogicalOp(m_Value(Op0), m_Value(Op1))))
4881 return false;
4882
4883 // If this logic op has not been simplified yet, just bail out and let that
4884 // happen first. Otherwise, the code below may wrongly invert.
4885 if (Op0 == Op1)
4886 return false;
4887
4888 // If one of the operands is a user of the other,
4889 // freelyInvert->freelyInvertAllUsersOf will change the operands of I, which
4890 // may cause miscompilation.
4891 if (match(Op0, m_Not(m_Specific(Op1))) || match(Op1, m_Not(m_Specific(Op0))))
4892 return false;
4893
4894 Instruction::BinaryOps NewOpc =
4895 match(&I, m_LogicalAnd()) ? Instruction::Or : Instruction::And;
4896 bool IsBinaryOp = isa<BinaryOperator>(I);
4897
4898 // Can our users be adapted?
4899 if (!InstCombiner::canFreelyInvertAllUsersOf(&I, /*IgnoredUser=*/nullptr))
4900 return false;
4901
4902 // And can the operands be adapted?
4903 if (!canFreelyInvert(*this, Op0, &I) || !canFreelyInvert(*this, Op1, &I))
4904 return false;
4905
4906 Op0 = freelyInvert(*this, Op0, &I);
4907 Op1 = freelyInvert(*this, Op1, &I);
4908
4909 Builder.SetInsertPoint(*I.getInsertionPointAfterDef());
4910 Value *NewLogicOp;
4911 if (IsBinaryOp)
4912 NewLogicOp = Builder.CreateBinOp(NewOpc, Op0, Op1, I.getName() + ".not");
4913 else
4914 NewLogicOp =
4915 Builder.CreateLogicalOp(NewOpc, Op0, Op1, I.getName() + ".not");
4916
4917 replaceInstUsesWith(I, NewLogicOp);
4918 // We can not just create an outer `not`, it will most likely be immediately
4919 // folded back, reconstructing our initial pattern, and causing an
4920 // infinite combine loop, so immediately manually fold it away.
4921 freelyInvertAllUsersOf(NewLogicOp);
4922 return true;
4923}
4924
4925// Transform
4926// z = (~x) &/| y
4927// into:
4928// z = ~(x |/& (~y))
4929// iff y is free to invert and all uses of z can be freely updated.
4931 Value *Op0, *Op1;
4932 if (!match(&I, m_LogicalOp(m_Value(Op0), m_Value(Op1))))
4933 return false;
4934 Instruction::BinaryOps NewOpc =
4935 match(&I, m_LogicalAnd()) ? Instruction::Or : Instruction::And;
4936 bool IsBinaryOp = isa<BinaryOperator>(I);
4937
4938 Value *NotOp0 = nullptr;
4939 Value *NotOp1 = nullptr;
4940 Value **OpToInvert = nullptr;
4941 if (match(Op0, m_Not(m_Value(NotOp0))) && canFreelyInvert(*this, Op1, &I)) {
4942 Op0 = NotOp0;
4943 OpToInvert = &Op1;
4944 } else if (match(Op1, m_Not(m_Value(NotOp1))) &&
4945 canFreelyInvert(*this, Op0, &I)) {
4946 Op1 = NotOp1;
4947 OpToInvert = &Op0;
4948 } else
4949 return false;
4950
4951 // And can our users be adapted?
4952 if (!InstCombiner::canFreelyInvertAllUsersOf(&I, /*IgnoredUser=*/nullptr))
4953 return false;
4954
4955 *OpToInvert = freelyInvert(*this, *OpToInvert, &I);
4956
4957 Builder.SetInsertPoint(*I.getInsertionPointAfterDef());
4958 Value *NewBinOp;
4959 if (IsBinaryOp)
4960 NewBinOp = Builder.CreateBinOp(NewOpc, Op0, Op1, I.getName() + ".not");
4961 else
4962 NewBinOp = Builder.CreateLogicalOp(NewOpc, Op0, Op1, I.getName() + ".not");
4963 replaceInstUsesWith(I, NewBinOp);
4964 // We can not just create an outer `not`, it will most likely be immediately
4965 // folded back, reconstructing our initial pattern, and causing an
4966 // infinite combine loop, so immediately manually fold it away.
4967 freelyInvertAllUsersOf(NewBinOp);
4968 return true;
4969}
4970
4971Instruction *InstCombinerImpl::foldNot(BinaryOperator &I) {
4972 Value *NotOp;
4973 if (!match(&I, m_Not(m_Value(NotOp))))
4974 return nullptr;
4975
4976 // Apply DeMorgan's Law for 'nand' / 'nor' logic with an inverted operand.
4977 // We must eliminate the and/or (one-use) for these transforms to not increase
4978 // the instruction count.
4979 //
4980 // ~(~X & Y) --> (X | ~Y)
4981 // ~(Y & ~X) --> (X | ~Y)
4982 //
4983 // Note: The logical matches do not check for the commuted patterns because
4984 // those are handled via SimplifySelectsFeedingBinaryOp().
4985 Type *Ty = I.getType();
4986 Value *X, *Y;
4987 if (match(NotOp, m_OneUse(m_c_And(m_Not(m_Value(X)), m_Value(Y))))) {
4988 Value *NotY = Builder.CreateNot(Y, Y->getName() + ".not");
4989 return BinaryOperator::CreateOr(X, NotY);
4990 }
4991 if (match(NotOp, m_OneUse(m_LogicalAnd(m_Not(m_Value(X)), m_Value(Y))))) {
4992 Value *NotY = Builder.CreateNot(Y, Y->getName() + ".not");
4993 return SelectInst::Create(X, ConstantInt::getTrue(Ty), NotY);
4994 }
4995
4996 // ~(~X | Y) --> (X & ~Y)
4997 // ~(Y | ~X) --> (X & ~Y)
4998 if (match(NotOp, m_OneUse(m_c_Or(m_Not(m_Value(X)), m_Value(Y))))) {
4999 Value *NotY = Builder.CreateNot(Y, Y->getName() + ".not");
5000 return BinaryOperator::CreateAnd(X, NotY);
5001 }
5002 if (match(NotOp, m_OneUse(m_LogicalOr(m_Not(m_Value(X)), m_Value(Y))))) {
5003 Value *NotY = Builder.CreateNot(Y, Y->getName() + ".not");
5004 return SelectInst::Create(X, NotY, ConstantInt::getFalse(Ty));
5005 }
5006
5007 // Is this a 'not' (~) fed by a binary operator?
5008 BinaryOperator *NotVal;
5009 if (match(NotOp, m_BinOp(NotVal))) {
5010 // ~((-X) | Y) --> (X - 1) & (~Y)
5011 if (match(NotVal,
5013 Value *DecX = Builder.CreateAdd(X, ConstantInt::getAllOnesValue(Ty));
5014 Value *NotY = Builder.CreateNot(Y);
5015 return BinaryOperator::CreateAnd(DecX, NotY);
5016 }
5017
5018 // ~(~X >>s Y) --> (X >>s Y)
5019 if (match(NotVal, m_AShr(m_Not(m_Value(X)), m_Value(Y))))
5020 return BinaryOperator::CreateAShr(X, Y);
5021
5022 // Treat lshr with non-negative operand as ashr.
5023 // ~(~X >>u Y) --> (X >>s Y) iff X is known negative
5024 if (match(NotVal, m_LShr(m_Not(m_Value(X)), m_Value(Y))) &&
5025 isKnownNegative(X, SQ.getWithInstruction(NotVal)))
5026 return BinaryOperator::CreateAShr(X, Y);
5027
5028 // Bit-hack form of a signbit test for iN type:
5029 // ~(X >>s (N - 1)) --> sext i1 (X > -1) to iN
5030 unsigned FullShift = Ty->getScalarSizeInBits() - 1;
5031 if (match(NotVal, m_OneUse(m_AShr(m_Value(X), m_SpecificInt(FullShift))))) {
5032 Value *IsNotNeg = Builder.CreateIsNotNeg(X, "isnotneg");
5033 return new SExtInst(IsNotNeg, Ty);
5034 }
5035
5036 // If we are inverting a right-shifted constant, we may be able to eliminate
5037 // the 'not' by inverting the constant and using the opposite shift type.
5038 // Canonicalization rules ensure that only a negative constant uses 'ashr',
5039 // but we must check that in case that transform has not fired yet.
5040
5041 // ~(C >>s Y) --> ~C >>u Y (when inverting the replicated sign bits)
5042 Constant *C;
5043 if (match(NotVal, m_AShr(m_Constant(C), m_Value(Y))) &&
5044 match(C, m_Negative()))
5045 return BinaryOperator::CreateLShr(ConstantExpr::getNot(C), Y);
5046
5047 // ~(C >>u Y) --> ~C >>s Y (when inverting the replicated sign bits)
5048 if (match(NotVal, m_LShr(m_Constant(C), m_Value(Y))) &&
5049 match(C, m_NonNegative()))
5050 return BinaryOperator::CreateAShr(ConstantExpr::getNot(C), Y);
5051
5052 // ~(X + C) --> ~C - X
5053 if (match(NotVal, m_Add(m_Value(X), m_ImmConstant(C))))
5054 return BinaryOperator::CreateSub(ConstantExpr::getNot(C), X);
5055
5056 // ~(X - Y) --> ~X + Y
5057 // FIXME: is it really beneficial to sink the `not` here?
5058 if (match(NotVal, m_Sub(m_Value(X), m_Value(Y))))
5059 if (isa<Constant>(X) || NotVal->hasOneUse())
5060 return BinaryOperator::CreateAdd(Builder.CreateNot(X), Y);
5061
5062 // ~(~X + Y) --> X - Y
5063 if (match(NotVal, m_c_Add(m_Not(m_Value(X)), m_Value(Y))))
5064 return BinaryOperator::CreateWithCopiedFlags(Instruction::Sub, X, Y,
5065 NotVal);
5066 }
5067
5068 // not (cmp A, B) = !cmp A, B
5069 CmpPredicate Pred;
5070 if (match(NotOp, m_Cmp(Pred, m_Value(), m_Value())) &&
5071 (NotOp->hasOneUse() ||
5073 /*IgnoredUser=*/nullptr))) {
5074 cast<CmpInst>(NotOp)->setPredicate(CmpInst::getInversePredicate(Pred));
5076 return &I;
5077 }
5078
5079 // Move a 'not' ahead of casts of a bool to enable logic reduction:
5080 // not (bitcast (sext i1 X)) --> bitcast (sext (not i1 X))
5081 if (match(NotOp, m_OneUse(m_BitCast(m_OneUse(m_SExt(m_Value(X)))))) && X->getType()->isIntOrIntVectorTy(1)) {
5082 Type *SextTy = cast<BitCastOperator>(NotOp)->getSrcTy();
5083 Value *NotX = Builder.CreateNot(X);
5084 Value *Sext = Builder.CreateSExt(NotX, SextTy);
5085 return new BitCastInst(Sext, Ty);
5086 }
5087
5088 if (auto *NotOpI = dyn_cast<Instruction>(NotOp))
5089 if (sinkNotIntoLogicalOp(*NotOpI))
5090 return &I;
5091
5092 // Eliminate a bitwise 'not' op of 'not' min/max by inverting the min/max:
5093 // ~min(~X, ~Y) --> max(X, Y)
5094 // ~max(~X, Y) --> min(X, ~Y)
5095 auto *II = dyn_cast<IntrinsicInst>(NotOp);
5096 if (II && II->hasOneUse()) {
5097 if (match(NotOp, m_c_MaxOrMin(m_Not(m_Value(X)), m_Value(Y)))) {
5098 Intrinsic::ID InvID = getInverseMinMaxIntrinsic(II->getIntrinsicID());
5099 Value *NotY = Builder.CreateNot(Y);
5100 Value *InvMaxMin = Builder.CreateBinaryIntrinsic(InvID, X, NotY);
5101 return replaceInstUsesWith(I, InvMaxMin);
5102 }
5103
5104 if (II->getIntrinsicID() == Intrinsic::is_fpclass) {
5105 ConstantInt *ClassMask = cast<ConstantInt>(II->getArgOperand(1));
5106 II->setArgOperand(
5107 1, ConstantInt::get(ClassMask->getType(),
5108 ~ClassMask->getZExtValue() & fcAllFlags));
5109 return replaceInstUsesWith(I, II);
5110 }
5111 }
5112
5113 if (NotOp->hasOneUse()) {
5114 // Pull 'not' into operands of select if both operands are one-use compares
5115 // or one is one-use compare and the other one is a constant.
5116 // Inverting the predicates eliminates the 'not' operation.
5117 // Example:
5118 // not (select ?, (cmp TPred, ?, ?), (cmp FPred, ?, ?) -->
5119 // select ?, (cmp InvTPred, ?, ?), (cmp InvFPred, ?, ?)
5120 // not (select ?, (cmp TPred, ?, ?), true -->
5121 // select ?, (cmp InvTPred, ?, ?), false
5122 if (auto *Sel = dyn_cast<SelectInst>(NotOp)) {
5123 Value *TV = Sel->getTrueValue();
5124 Value *FV = Sel->getFalseValue();
5125 auto *CmpT = dyn_cast<CmpInst>(TV);
5126 auto *CmpF = dyn_cast<CmpInst>(FV);
5127 bool InvertibleT = (CmpT && CmpT->hasOneUse()) || isa<Constant>(TV);
5128 bool InvertibleF = (CmpF && CmpF->hasOneUse()) || isa<Constant>(FV);
5129 if (InvertibleT && InvertibleF) {
5130 if (CmpT)
5131 CmpT->setPredicate(CmpT->getInversePredicate());
5132 else
5133 Sel->setTrueValue(ConstantExpr::getNot(cast<Constant>(TV)));
5134 if (CmpF)
5135 CmpF->setPredicate(CmpF->getInversePredicate());
5136 else
5137 Sel->setFalseValue(ConstantExpr::getNot(cast<Constant>(FV)));
5138 return replaceInstUsesWith(I, Sel);
5139 }
5140 }
5141 }
5142
5143 if (Instruction *NewXor = foldNotXor(I, Builder))
5144 return NewXor;
5145
5146 // TODO: Could handle multi-use better by checking if all uses of NotOp (other
5147 // than I) can be inverted.
5148 if (Value *R = getFreelyInverted(NotOp, NotOp->hasOneUse(), &Builder))
5149 return replaceInstUsesWith(I, R);
5150
5151 return nullptr;
5152}
5153
5154// FIXME: We use commutative matchers (m_c_*) for some, but not all, matches
5155// here. We should standardize that construct where it is needed or choose some
5156// other way to ensure that commutated variants of patterns are not missed.
5158 if (Value *V = simplifyXorInst(I.getOperand(0), I.getOperand(1),
5159 SQ.getWithInstruction(&I)))
5160 return replaceInstUsesWith(I, V);
5161
5163 return &I;
5164
5166 return X;
5167
5169 return Phi;
5170
5171 if (Instruction *NewXor = foldXorToXor(I, Builder))
5172 return NewXor;
5173
5174 // (A&B)^(A&C) -> A&(B^C) etc
5176 return replaceInstUsesWith(I, V);
5177
5178 // See if we can simplify any instructions used by the instruction whose sole
5179 // purpose is to compute bits we don't care about.
5181 return &I;
5182
5183 if (Instruction *R = foldNot(I))
5184 return R;
5185
5187 return R;
5188
5189 Value *Op0 = I.getOperand(0), *Op1 = I.getOperand(1);
5190 Value *X, *Y, *M;
5191
5192 // (X | Y) ^ M -> (X ^ M) ^ Y
5193 // (X | Y) ^ M -> (Y ^ M) ^ X
5195 m_Value(M)))) {
5196 if (Value *XorAC = simplifyXorInst(X, M, SQ.getWithInstruction(&I)))
5197 return BinaryOperator::CreateXor(XorAC, Y);
5198
5199 if (Value *XorBC = simplifyXorInst(Y, M, SQ.getWithInstruction(&I)))
5200 return BinaryOperator::CreateXor(XorBC, X);
5201 }
5202
5203 // Fold (X & M) ^ (Y & ~M) -> (X & M) | (Y & ~M)
5204 // This it a special case in haveNoCommonBitsSet, but the computeKnownBits
5205 // calls in there are unnecessary as SimplifyDemandedInstructionBits should
5206 // have already taken care of those cases.
5207 if (match(&I, m_c_Xor(m_c_And(m_Not(m_Value(M)), m_Value()),
5208 m_c_And(m_Deferred(M), m_Value())))) {
5210 return BinaryOperator::CreateDisjointOr(Op0, Op1);
5211 else
5212 return BinaryOperator::CreateOr(Op0, Op1);
5213 }
5214
5216 return Xor;
5217
5218 Constant *C1;
5219 if (match(Op1, m_Constant(C1))) {
5220 Constant *C2;
5221
5222 if (match(Op0, m_OneUse(m_Or(m_Value(X), m_ImmConstant(C2)))) &&
5223 match(C1, m_ImmConstant())) {
5224 // (X | C2) ^ C1 --> (X & ~C2) ^ (C1^C2)
5227 Value *And = Builder.CreateAnd(
5229 return BinaryOperator::CreateXor(
5231 }
5232
5233 // Use DeMorgan and reassociation to eliminate a 'not' op.
5234 if (match(Op0, m_OneUse(m_Or(m_Not(m_Value(X)), m_Constant(C2))))) {
5235 // (~X | C2) ^ C1 --> ((X & ~C2) ^ -1) ^ C1 --> (X & ~C2) ^ ~C1
5236 Value *And = Builder.CreateAnd(X, ConstantExpr::getNot(C2));
5237 return BinaryOperator::CreateXor(And, ConstantExpr::getNot(C1));
5238 }
5239 if (match(Op0, m_OneUse(m_And(m_Not(m_Value(X)), m_Constant(C2))))) {
5240 // (~X & C2) ^ C1 --> ((X | ~C2) ^ -1) ^ C1 --> (X | ~C2) ^ ~C1
5241 Value *Or = Builder.CreateOr(X, ConstantExpr::getNot(C2));
5242 return BinaryOperator::CreateXor(Or, ConstantExpr::getNot(C1));
5243 }
5244
5245 // Convert xor ([trunc] (ashr X, BW-1)), C =>
5246 // select(X >s -1, C, ~C)
5247 // The ashr creates "AllZeroOrAllOne's", which then optionally inverses the
5248 // constant depending on whether this input is less than 0.
5249 const APInt *CA;
5250 if (match(Op0, m_OneUse(m_TruncOrSelf(
5251 m_AShr(m_Value(X), m_APIntAllowPoison(CA))))) &&
5252 *CA == X->getType()->getScalarSizeInBits() - 1 &&
5253 !match(C1, m_AllOnes())) {
5254 assert(!C1->isZeroValue() && "Unexpected xor with 0");
5255 Value *IsNotNeg = Builder.CreateIsNotNeg(X);
5256 return SelectInst::Create(IsNotNeg, Op1, Builder.CreateNot(Op1));
5257 }
5258 }
5259
5260 Type *Ty = I.getType();
5261 {
5262 const APInt *RHSC;
5263 if (match(Op1, m_APInt(RHSC))) {
5264 Value *X;
5265 const APInt *C;
5266 // (C - X) ^ signmaskC --> (C + signmaskC) - X
5267 if (RHSC->isSignMask() && match(Op0, m_Sub(m_APInt(C), m_Value(X))))
5268 return BinaryOperator::CreateSub(ConstantInt::get(Ty, *C + *RHSC), X);
5269
5270 // (X + C) ^ signmaskC --> X + (C + signmaskC)
5271 if (RHSC->isSignMask() && match(Op0, m_Add(m_Value(X), m_APInt(C))))
5272 return BinaryOperator::CreateAdd(X, ConstantInt::get(Ty, *C + *RHSC));
5273
5274 // (X | C) ^ RHSC --> X ^ (C ^ RHSC) iff X & C == 0
5275 if (match(Op0, m_Or(m_Value(X), m_APInt(C))) &&
5276 MaskedValueIsZero(X, *C, &I))
5277 return BinaryOperator::CreateXor(X, ConstantInt::get(Ty, *C ^ *RHSC));
5278
5279 // When X is a power-of-two or zero and zero input is poison:
5280 // ctlz(i32 X) ^ 31 --> cttz(X)
5281 // cttz(i32 X) ^ 31 --> ctlz(X)
5282 auto *II = dyn_cast<IntrinsicInst>(Op0);
5283 if (II && II->hasOneUse() && *RHSC == Ty->getScalarSizeInBits() - 1) {
5284 Intrinsic::ID IID = II->getIntrinsicID();
5285 if ((IID == Intrinsic::ctlz || IID == Intrinsic::cttz) &&
5286 match(II->getArgOperand(1), m_One()) &&
5287 isKnownToBeAPowerOfTwo(II->getArgOperand(0), /*OrZero */ true)) {
5288 IID = (IID == Intrinsic::ctlz) ? Intrinsic::cttz : Intrinsic::ctlz;
5289 Function *F =
5290 Intrinsic::getOrInsertDeclaration(II->getModule(), IID, Ty);
5291 return CallInst::Create(F, {II->getArgOperand(0), Builder.getTrue()});
5292 }
5293 }
5294
5295 // If RHSC is inverting the remaining bits of shifted X,
5296 // canonicalize to a 'not' before the shift to help SCEV and codegen:
5297 // (X << C) ^ RHSC --> ~X << C
5298 if (match(Op0, m_OneUse(m_Shl(m_Value(X), m_APInt(C)))) &&
5299 *RHSC == APInt::getAllOnes(Ty->getScalarSizeInBits()).shl(*C)) {
5300 Value *NotX = Builder.CreateNot(X);
5301 return BinaryOperator::CreateShl(NotX, ConstantInt::get(Ty, *C));
5302 }
5303 // (X >>u C) ^ RHSC --> ~X >>u C
5304 if (match(Op0, m_OneUse(m_LShr(m_Value(X), m_APInt(C)))) &&
5305 *RHSC == APInt::getAllOnes(Ty->getScalarSizeInBits()).lshr(*C)) {
5306 Value *NotX = Builder.CreateNot(X);
5307 return BinaryOperator::CreateLShr(NotX, ConstantInt::get(Ty, *C));
5308 }
5309 // TODO: We could handle 'ashr' here as well. That would be matching
5310 // a 'not' op and moving it before the shift. Doing that requires
5311 // preventing the inverse fold in canShiftBinOpWithConstantRHS().
5312 }
5313
5314 // If we are XORing the sign bit of a floating-point value, convert
5315 // this to fneg, then cast back to integer.
5316 //
5317 // This is generous interpretation of noimplicitfloat, this is not a true
5318 // floating-point operation.
5319 //
5320 // Assumes any IEEE-represented type has the sign bit in the high bit.
5321 // TODO: Unify with APInt matcher. This version allows undef unlike m_APInt
5322 Value *CastOp;
5323 if (match(Op0, m_ElementWiseBitCast(m_Value(CastOp))) &&
5324 match(Op1, m_SignMask()) &&
5325 !Builder.GetInsertBlock()->getParent()->hasFnAttribute(
5326 Attribute::NoImplicitFloat)) {
5327 Type *EltTy = CastOp->getType()->getScalarType();
5328 if (EltTy->isFloatingPointTy() &&
5330 Value *FNeg = Builder.CreateFNeg(CastOp);
5331 return new BitCastInst(FNeg, I.getType());
5332 }
5333 }
5334 }
5335
5336 // FIXME: This should not be limited to scalar (pull into APInt match above).
5337 {
5338 Value *X;
5339 ConstantInt *C1, *C2, *C3;
5340 // ((X^C1) >> C2) ^ C3 -> (X>>C2) ^ ((C1>>C2)^C3)
5341 if (match(Op1, m_ConstantInt(C3)) &&
5343 m_ConstantInt(C2))) &&
5344 Op0->hasOneUse()) {
5345 // fold (C1 >> C2) ^ C3
5346 APInt FoldConst = C1->getValue().lshr(C2->getValue());
5347 FoldConst ^= C3->getValue();
5348 // Prepare the two operands.
5349 auto *Opnd0 = Builder.CreateLShr(X, C2);
5350 Opnd0->takeName(Op0);
5351 return BinaryOperator::CreateXor(Opnd0, ConstantInt::get(Ty, FoldConst));
5352 }
5353 }
5354
5355 if (Instruction *FoldedLogic = foldBinOpIntoSelectOrPhi(I))
5356 return FoldedLogic;
5357
5358 // Y ^ (X | Y) --> X & ~Y
5359 // Y ^ (Y | X) --> X & ~Y
5360 if (match(Op1, m_OneUse(m_c_Or(m_Value(X), m_Specific(Op0)))))
5361 return BinaryOperator::CreateAnd(X, Builder.CreateNot(Op0));
5362 // (X | Y) ^ Y --> X & ~Y
5363 // (Y | X) ^ Y --> X & ~Y
5364 if (match(Op0, m_OneUse(m_c_Or(m_Value(X), m_Specific(Op1)))))
5365 return BinaryOperator::CreateAnd(X, Builder.CreateNot(Op1));
5366
5367 // Y ^ (X & Y) --> ~X & Y
5368 // Y ^ (Y & X) --> ~X & Y
5369 if (match(Op1, m_OneUse(m_c_And(m_Value(X), m_Specific(Op0)))))
5370 return BinaryOperator::CreateAnd(Op0, Builder.CreateNot(X));
5371 // (X & Y) ^ Y --> ~X & Y
5372 // (Y & X) ^ Y --> ~X & Y
5373 // Canonical form is (X & C) ^ C; don't touch that.
5374 // TODO: A 'not' op is better for analysis and codegen, but demanded bits must
5375 // be fixed to prefer that (otherwise we get infinite looping).
5376 if (!match(Op1, m_Constant()) &&
5377 match(Op0, m_OneUse(m_c_And(m_Value(X), m_Specific(Op1)))))
5378 return BinaryOperator::CreateAnd(Op1, Builder.CreateNot(X));
5379
5380 Value *A, *B, *C;
5381 // (A ^ B) ^ (A | C) --> (~A & C) ^ B -- There are 4 commuted variants.
5384 return BinaryOperator::CreateXor(
5385 Builder.CreateAnd(Builder.CreateNot(A), C), B);
5386
5387 // (A ^ B) ^ (B | C) --> (~B & C) ^ A -- There are 4 commuted variants.
5390 return BinaryOperator::CreateXor(
5391 Builder.CreateAnd(Builder.CreateNot(B), C), A);
5392
5393 // (A & B) ^ (A ^ B) -> (A | B)
5394 if (match(Op0, m_And(m_Value(A), m_Value(B))) &&
5396 return BinaryOperator::CreateOr(A, B);
5397 // (A ^ B) ^ (A & B) -> (A | B)
5398 if (match(Op0, m_Xor(m_Value(A), m_Value(B))) &&
5400 return BinaryOperator::CreateOr(A, B);
5401
5402 // (A & ~B) ^ ~A -> ~(A & B)
5403 // (~B & A) ^ ~A -> ~(A & B)
5404 if (match(Op0, m_c_And(m_Value(A), m_Not(m_Value(B)))) &&
5405 match(Op1, m_Not(m_Specific(A))))
5406 return BinaryOperator::CreateNot(Builder.CreateAnd(A, B));
5407
5408 // (~A & B) ^ A --> A | B -- There are 4 commuted variants.
5410 return BinaryOperator::CreateOr(A, B);
5411
5412 // (~A | B) ^ A --> ~(A & B)
5413 if (match(Op0, m_OneUse(m_c_Or(m_Not(m_Specific(Op1)), m_Value(B)))))
5414 return BinaryOperator::CreateNot(Builder.CreateAnd(Op1, B));
5415
5416 // A ^ (~A | B) --> ~(A & B)
5417 if (match(Op1, m_OneUse(m_c_Or(m_Not(m_Specific(Op0)), m_Value(B)))))
5418 return BinaryOperator::CreateNot(Builder.CreateAnd(Op0, B));
5419
5420 // (A | B) ^ (A | C) --> (B ^ C) & ~A -- There are 4 commuted variants.
5421 // TODO: Loosen one-use restriction if common operand is a constant.
5422 Value *D;
5423 if (match(Op0, m_OneUse(m_Or(m_Value(A), m_Value(B)))) &&
5424 match(Op1, m_OneUse(m_Or(m_Value(C), m_Value(D))))) {
5425 if (B == C || B == D)
5426 std::swap(A, B);
5427 if (A == C)
5428 std::swap(C, D);
5429 if (A == D) {
5430 Value *NotA = Builder.CreateNot(A);
5431 return BinaryOperator::CreateAnd(Builder.CreateXor(B, C), NotA);
5432 }
5433 }
5434
5435 // (A & B) ^ (A | C) --> A ? ~B : C -- There are 4 commuted variants.
5436 if (I.getType()->isIntOrIntVectorTy(1) &&
5439 bool NeedFreeze = isa<SelectInst>(Op0) && isa<SelectInst>(Op1) && B == D;
5440 if (B == C || B == D)
5441 std::swap(A, B);
5442 if (A == C)
5443 std::swap(C, D);
5444 if (A == D) {
5445 if (NeedFreeze)
5446 A = Builder.CreateFreeze(A);
5447 Value *NotB = Builder.CreateNot(B);
5448 return SelectInst::Create(A, NotB, C);
5449 }
5450 }
5451
5452 if (auto *LHS = dyn_cast<ICmpInst>(I.getOperand(0)))
5453 if (auto *RHS = dyn_cast<ICmpInst>(I.getOperand(1)))
5454 if (Value *V = foldXorOfICmps(LHS, RHS, I))
5455 return replaceInstUsesWith(I, V);
5456
5457 if (Instruction *CastedXor = foldCastedBitwiseLogic(I))
5458 return CastedXor;
5459
5460 if (Instruction *Abs = canonicalizeAbs(I, Builder))
5461 return Abs;
5462
5463 // Otherwise, if all else failed, try to hoist the xor-by-constant:
5464 // (X ^ C) ^ Y --> (X ^ Y) ^ C
5465 // Just like we do in other places, we completely avoid the fold
5466 // for constantexprs, at least to avoid endless combine loop.
5468 m_ImmConstant(C1))),
5469 m_Value(Y))))
5470 return BinaryOperator::CreateXor(Builder.CreateXor(X, Y), C1);
5471
5473 return R;
5474
5475 if (Instruction *Canonicalized = canonicalizeLogicFirst(I, Builder))
5476 return Canonicalized;
5477
5478 if (Instruction *Folded = foldLogicOfIsFPClass(I, Op0, Op1))
5479 return Folded;
5480
5481 if (Instruction *Folded = canonicalizeConditionalNegationViaMathToSelect(I))
5482 return Folded;
5483
5484 if (Instruction *Res = foldBinOpOfDisplacedShifts(I))
5485 return Res;
5486
5488 return Res;
5489
5490 return nullptr;
5491}
assert(UImm &&(UImm !=~static_cast< T >(0)) &&"Invalid immediate!")
AMDGPU Register Bank Select
MachineBasicBlock MachineBasicBlock::iterator DebugLoc DL
static GCRegistry::Add< ErlangGC > A("erlang", "erlang-compatible garbage collector")
static GCRegistry::Add< StatepointGC > D("statepoint-example", "an example strategy for statepoint")
static GCRegistry::Add< CoreCLRGC > E("coreclr", "CoreCLR-compatible GC")
static GCRegistry::Add< OcamlGC > B("ocaml", "ocaml 3.10-compatible GC")
static bool isSigned(unsigned int Opcode)
static Value * foldIsPowerOf2OrZero(ICmpInst *Cmp0, ICmpInst *Cmp1, bool IsAnd, InstCombiner::BuilderTy &Builder, InstCombinerImpl &IC)
Fold (icmp eq ctpop(X) 1) | (icmp eq X 0) into (icmp ult ctpop(X) 2) and fold (icmp ne ctpop(X) 1) & ...
static Value * foldBitmaskMul(Value *Op0, Value *Op1, InstCombiner::BuilderTy &Builder)
(A & N) * C + (A & M) * C -> (A & (N + M)) & C This also accepts the equivalent select form of (A & N...
static unsigned conjugateICmpMask(unsigned Mask)
Convert an analysis of a masked ICmp into its equivalent if all boolean operations had the opposite s...
static Instruction * foldNotXor(BinaryOperator &I, InstCombiner::BuilderTy &Builder)
static Value * foldLogOpOfMaskedICmps(Value *LHS, Value *RHS, bool IsAnd, bool IsLogical, InstCombiner::BuilderTy &Builder, const SimplifyQuery &Q)
Try to fold (icmp(A & B) ==/!= C) &/| (icmp(A & D) ==/!= E) into a single (icmp(A & X) ==/!...
static Value * getFCmpValue(unsigned Code, Value *LHS, Value *RHS, InstCombiner::BuilderTy &Builder, FMFSource FMF)
This is the complement of getFCmpCode, which turns an opcode and two operands into either a FCmp inst...
static bool matchIsFPClassLikeFCmp(Value *Op, Value *&ClassVal, uint64_t &ClassMask)
Match an fcmp against a special value that performs a test possible by llvm.is.fpclass.
static Value * foldSignedTruncationCheck(ICmpInst *ICmp0, ICmpInst *ICmp1, Instruction &CxtI, InstCombiner::BuilderTy &Builder)
General pattern: X & Y.
static Instruction * visitMaskedMerge(BinaryOperator &I, InstCombiner::BuilderTy &Builder)
If we have a masked merge, in the canonical form of: (assuming that A only has one use....
static Instruction * canonicalizeAbs(BinaryOperator &Xor, InstCombiner::BuilderTy &Builder)
Canonicalize a shifty way to code absolute value to the more common pattern that uses negation and se...
static Value * foldIsPowerOf2(ICmpInst *Cmp0, ICmpInst *Cmp1, bool JoinedByAnd, InstCombiner::BuilderTy &Builder, InstCombinerImpl &IC)
Reduce a pair of compares that check if a value has exactly 1 bit set.
static Value * foldUnsignedUnderflowCheck(ICmpInst *ZeroICmp, ICmpInst *UnsignedICmp, bool IsAnd, const SimplifyQuery &Q, InstCombiner::BuilderTy &Builder)
Commuted variants are assumed to be handled by calling this function again with the parameters swappe...
static Instruction * foldOrToXor(BinaryOperator &I, InstCombiner::BuilderTy &Builder)
static Value * simplifyAndOrWithOpReplaced(Value *V, Value *Op, Value *RepOp, bool SimplifyOnly, InstCombinerImpl &IC, unsigned Depth=0)
static Instruction * matchDeMorgansLaws(BinaryOperator &I, InstCombiner &IC)
Match variations of De Morgan's Laws: (~A & ~B) == (~(A | B)) (~A | ~B) == (~(A & B))
static Value * foldLogOpOfMaskedICmpsAsymmetric(Value *LHS, Value *RHS, bool IsAnd, Value *A, Value *B, Value *C, Value *D, Value *E, ICmpInst::Predicate PredL, ICmpInst::Predicate PredR, unsigned LHSMask, unsigned RHSMask, InstCombiner::BuilderTy &Builder)
Try to fold (icmp(A & B) ==/!= 0) &/| (icmp(A & D) ==/!= E) into a single (icmp(A & X) ==/!...
static Instruction * foldAndToXor(BinaryOperator &I, InstCombiner::BuilderTy &Builder)
static unsigned getMaskedICmpType(Value *A, Value *B, Value *C, ICmpInst::Predicate Pred)
Return the set of patterns (from MaskedICmpType) that (icmp SCC (A & B), C) satisfies.
static Instruction * foldXorToXor(BinaryOperator &I, InstCombiner::BuilderTy &Builder)
A ^ B can be specified using other logic ops in a variety of patterns.
static bool canNarrowShiftAmt(Constant *C, unsigned BitWidth)
Return true if a constant shift amount is always less than the specified bit-width.
static Instruction * foldLogicCastConstant(BinaryOperator &Logic, CastInst *Cast, InstCombinerImpl &IC)
Fold {and,or,xor} (cast X), C.
static Value * foldAndOrOfICmpEqConstantAndICmp(ICmpInst *LHS, ICmpInst *RHS, bool IsAnd, bool IsLogical, IRBuilderBase &Builder)
static bool canFreelyInvert(InstCombiner &IC, Value *Op, Instruction *IgnoredUser)
static Value * foldNegativePower2AndShiftedMask(Value *A, Value *B, Value *D, Value *E, ICmpInst::Predicate PredL, ICmpInst::Predicate PredR, InstCombiner::BuilderTy &Builder)
Try to fold (icmp(A & B) == 0) & (icmp(A & D) != E) into (icmp A u< D) iff B is a contiguous set of o...
static Value * matchIsFiniteTest(InstCombiner::BuilderTy &Builder, FCmpInst *LHS, FCmpInst *RHS)
and (fcmp ord x, 0), (fcmp u* x, inf) -> fcmp o* x, inf
static Value * foldPowerOf2AndShiftedMask(ICmpInst *Cmp0, ICmpInst *Cmp1, bool JoinedByAnd, InstCombiner::BuilderTy &Builder)
Try to fold ((icmp X u< P) & (icmp(X & M) != M)) or ((icmp X s> -1) & (icmp(X & M) !...
static Value * stripSignOnlyFPOps(Value *Val)
Ignore all operations which only change the sign of a value, returning the underlying magnitude value...
static Value * foldOrUnsignedUMulOverflowICmp(BinaryOperator &I, InstCombiner::BuilderTy &Builder, const DataLayout &DL)
Fold Res, Overflow = (umul.with.overflow x c1); (or Overflow (ugt Res c2)) --> (ugt x (c2/c1)).
static Value * freelyInvert(InstCombinerImpl &IC, Value *Op, Instruction *IgnoredUser)
static Value * foldLogOpOfMaskedICmps_NotAllZeros_BMask_Mixed(Value *LHS, Value *RHS, bool IsAnd, Value *A, Value *B, Value *D, Value *E, ICmpInst::Predicate PredL, ICmpInst::Predicate PredR, InstCombiner::BuilderTy &Builder)
Try to fold (icmp(A & B) ==/!= C) &/| (icmp(A & D) ==/!= E) into a single (icmp(A & X) ==/!...
static std::optional< IntPart > matchIntPart(Value *V)
Match an extraction of bits from an integer.
static Instruction * canonicalizeLogicFirst(BinaryOperator &I, InstCombiner::BuilderTy &Builder)
static Instruction * reassociateFCmps(BinaryOperator &BO, InstCombiner::BuilderTy &Builder)
This a limited reassociation for a special case (see above) where we are checking if two values are e...
static Value * getNewICmpValue(unsigned Code, bool Sign, Value *LHS, Value *RHS, InstCombiner::BuilderTy &Builder)
This is the complement of getICmpCode, which turns an opcode and two operands into either a constant ...
static Value * extractIntPart(const IntPart &P, IRBuilderBase &Builder)
Materialize an extraction of bits from an integer in IR.
static bool matchUnorderedInfCompare(FCmpInst::Predicate P, Value *LHS, Value *RHS)
Matches fcmp u__ x, +/-inf.
static bool matchIsNotNaN(FCmpInst::Predicate P, Value *LHS, Value *RHS)
Matches canonical form of isnan, fcmp ord x, 0.
static bool areInverseVectorBitmasks(Constant *C1, Constant *C2)
If all elements of two constant vectors are 0/-1 and inverses, return true.
MaskedICmpType
Classify (icmp eq (A & B), C) and (icmp ne (A & B), C) as matching patterns that can be simplified.
@ BMask_NotAllOnes
@ AMask_NotAllOnes
@ Mask_NotAllZeros
static Instruction * foldComplexAndOrPatterns(BinaryOperator &I, InstCombiner::BuilderTy &Builder)
Try folding relatively complex patterns for both And and Or operations with all And and Or swapped.
static bool matchZExtedSubInteger(Value *V, Value *&Int, APInt &Mask, uint64_t &Offset, bool &IsShlNUW, bool &IsShlNSW)
Match V as "lshr -> mask -> zext -> shl".
static std::optional< DecomposedBitMaskMul > matchBitmaskMul(Value *V)
static Value * foldOrOfInversions(BinaryOperator &I, InstCombiner::BuilderTy &Builder)
static bool matchSubIntegerPackFromVector(Value *V, Value *&Vec, int64_t &VecOffset, SmallBitVector &Mask, const DataLayout &DL)
Match V as "shufflevector -> bitcast" or "extractelement -> zext -> shl" patterns,...
static Instruction * matchFunnelShift(Instruction &Or, InstCombinerImpl &IC)
Match UB-safe variants of the funnel shift intrinsic.
static Instruction * reassociateForUses(BinaryOperator &BO, InstCombinerImpl::BuilderTy &Builder)
Try to reassociate a pair of binops so that values with one use only are part of the same instruction...
static Value * matchOrConcat(Instruction &Or, InstCombiner::BuilderTy &Builder)
Attempt to combine or(zext(x),shl(zext(y),bw/2) concat packing patterns.
static Value * foldAndOrOfICmpsWithPow2AndWithZero(InstCombiner::BuilderTy &Builder, ICmpInst *LHS, ICmpInst *RHS, bool IsAnd, const SimplifyQuery &Q)
static Instruction * foldBitwiseLogicWithIntrinsics(BinaryOperator &I, InstCombiner::BuilderTy &Builder)
static std::optional< std::pair< unsigned, unsigned > > getMaskedTypeForICmpPair(Value *&A, Value *&B, Value *&C, Value *&D, Value *&E, Value *LHS, Value *RHS, ICmpInst::Predicate &PredL, ICmpInst::Predicate &PredR)
Handle (icmp(A & B) ==/!= C) &/| (icmp(A & D) ==/!= E).
static Instruction * foldIntegerPackFromVector(Instruction &I, InstCombiner::BuilderTy &Builder, const DataLayout &DL)
Try to fold the join of two scalar integers whose contents are packed elements of the same vector.
static Value * foldIntegerRepackThroughZExt(Value *Lhs, Value *Rhs, InstCombiner::BuilderTy &Builder)
Try to fold the join of two scalar integers whose bits are unpacked and zexted from the same source i...
static Value * foldAndOrOfICmpsWithConstEq(ICmpInst *Cmp0, ICmpInst *Cmp1, bool IsAnd, bool IsLogical, InstCombiner::BuilderTy &Builder, const SimplifyQuery &Q)
Reduce logic-of-compares with equality to a constant by substituting a common operand with the consta...
This file provides internal interfaces used to implement the InstCombine.
This file provides the interface for the instcombine pass implementation.
static bool isZero(Value *V, const DataLayout &DL, DominatorTree *DT, AssumptionCache *AC)
Definition Lint.cpp:539
#define F(x, y, z)
Definition MD5.cpp:55
#define I(x, y, z)
Definition MD5.cpp:58
#define R2(n)
uint64_t High
uint64_t IntrinsicInst * II
#define P(N)
if(PassOpts->AAPipeline)
const SmallVectorImpl< MachineOperand > & Cond
This file implements the SmallBitVector class.
static unsigned getScalarSizeInBits(Type *Ty)
static TableGen::Emitter::Opt Y("gen-skeleton-entry", EmitSkeleton, "Generate example skeleton entry")
static TableGen::Emitter::OptClass< SkeletonEmitter > X("gen-skeleton-class", "Generate example skeleton class")
static constexpr int Concat[]
Value * RHS
Value * LHS
The Input class is used to parse a yaml document into in-memory structs and vectors.
bool bitwiseIsEqual(const APFloat &RHS) const
Definition APFloat.h:1414
bool isZero() const
Definition APFloat.h:1445
APInt bitcastToAPInt() const
Definition APFloat.h:1353
static APFloat getInf(const fltSemantics &Sem, bool Negative=false)
Factory for Positive and Negative Infinity.
Definition APFloat.h:1098
Class for arbitrary precision integers.
Definition APInt.h:78
LLVM_ABI APInt udiv(const APInt &RHS) const
Unsigned division operation.
Definition APInt.cpp:1573
static APInt getAllOnes(unsigned numBits)
Return an APInt of a specified width with all bits set.
Definition APInt.h:234
LLVM_ABI APInt zext(unsigned width) const
Zero extend to a new width.
Definition APInt.cpp:1012
uint64_t getZExtValue() const
Get zero extended value.
Definition APInt.h:1540
LLVM_ABI APInt trunc(unsigned width) const
Truncate to new width.
Definition APInt.cpp:936
APInt abs() const
Get the absolute value.
Definition APInt.h:1795
unsigned countLeadingOnes() const
Definition APInt.h:1624
bool isAllOnes() const
Determine if all bits are set. This is true for zero-width values.
Definition APInt.h:371
LLVM_ABI APInt usub_ov(const APInt &RHS, bool &Overflow) const
Definition APInt.cpp:1948
bool ugt(const APInt &RHS) const
Unsigned greater than comparison.
Definition APInt.h:1182
bool isZero() const
Determine if this value is zero, i.e. all bits are clear.
Definition APInt.h:380
bool isSignMask() const
Check if the APInt's value is returned by getSignMask.
Definition APInt.h:466
unsigned getBitWidth() const
Return the number of bits in the APInt.
Definition APInt.h:1488
bool ult(const APInt &RHS) const
Unsigned less than comparison.
Definition APInt.h:1111
LLVM_ABI APInt sadd_ov(const APInt &RHS, bool &Overflow) const
Definition APInt.cpp:1928
bool intersects(const APInt &RHS) const
This operation tests if there are any pairs of corresponding bits between this APInt and RHS that are...
Definition APInt.h:1249
int32_t exactLogBase2() const
Definition APInt.h:1783
LLVM_ABI APInt reverseBits() const
Definition APInt.cpp:768
LLVM_ABI APInt uadd_ov(const APInt &RHS, bool &Overflow) const
Definition APInt.cpp:1935
unsigned countr_zero() const
Count the number of trailing zero bits.
Definition APInt.h:1639
unsigned countLeadingZeros() const
Definition APInt.h:1606
bool ule(const APInt &RHS) const
Unsigned less or equal comparison.
Definition APInt.h:1150
APInt shl(unsigned shiftAmt) const
Left-shift function.
Definition APInt.h:873
LLVM_ABI APInt byteSwap() const
Definition APInt.cpp:746
bool isSubsetOf(const APInt &RHS) const
This operation checks that all bits set in this APInt are also set in RHS.
Definition APInt.h:1257
bool isPowerOf2() const
Check if this APInt's value is a power of two greater than zero.
Definition APInt.h:440
static APInt getLowBitsSet(unsigned numBits, unsigned loBitsSet)
Constructs an APInt value that has the bottom loBitsSet bits set.
Definition APInt.h:306
LLVM_ABI APInt ssub_ov(const APInt &RHS, bool &Overflow) const
Definition APInt.cpp:1941
static APInt getBitsSetFrom(unsigned numBits, unsigned loBit)
Constructs an APInt value that has a contiguous range of bits set.
Definition APInt.h:286
APInt lshr(unsigned shiftAmt) const
Logical right-shift function.
Definition APInt.h:851
bool uge(const APInt &RHS) const
Unsigned greater or equal comparison.
Definition APInt.h:1221
void clearSignBit()
Set the sign bit to 0.
Definition APInt.h:1449
ArrayRef - Represent a constant reference to an array (0 or more elements consecutively in memory),...
Definition ArrayRef.h:41
LLVM_ABI bool isSigned() const
Whether the intrinsic is signed or unsigned.
LLVM_ABI Instruction::BinaryOps getBinaryOp() const
Returns the binary operation underlying the intrinsic.
BinaryOps getOpcode() const
Definition InstrTypes.h:374
static LLVM_ABI BinaryOperator * CreateNot(Value *Op, const Twine &Name="", InsertPosition InsertBefore=nullptr)
static LLVM_ABI BinaryOperator * Create(BinaryOps Op, Value *S1, Value *S2, const Twine &Name=Twine(), InsertPosition InsertBefore=nullptr)
Construct a binary instruction, given the opcode and the two operands.
static BinaryOperator * CreateWithCopiedFlags(BinaryOps Opc, Value *V1, Value *V2, Value *CopyO, const Twine &Name="", InsertPosition InsertBefore=nullptr)
Definition InstrTypes.h:219
This class represents a no-op cast from one type to another.
static CallInst * Create(FunctionType *Ty, Value *F, const Twine &NameStr="", InsertPosition InsertBefore=nullptr)
This is the base class for all instructions that perform data casts.
Definition InstrTypes.h:448
Type * getSrcTy() const
Return the source type, as a convenience.
Definition InstrTypes.h:617
Instruction::CastOps getOpcode() const
Return the opcode of this CastInst.
Definition InstrTypes.h:612
static LLVM_ABI CastInst * Create(Instruction::CastOps, Value *S, Type *Ty, const Twine &Name="", InsertPosition InsertBefore=nullptr)
Provides a way to construct any of the CastInst subclasses using an opcode instead of the subclass's ...
Type * getDestTy() const
Return the destination type, as a convenience.
Definition InstrTypes.h:619
static Type * makeCmpResultType(Type *opnd_type)
Create a result type for fcmp/icmp.
Definition InstrTypes.h:984
Predicate
This enumeration lists the possible predicates for CmpInst subclasses.
Definition InstrTypes.h:678
@ ICMP_SLT
signed less than
Definition InstrTypes.h:707
@ ICMP_SLE
signed less or equal
Definition InstrTypes.h:708
@ FCMP_OLT
0 1 0 0 True if ordered and less than
Definition InstrTypes.h:684
@ FCMP_ULE
1 1 0 1 True if unordered, less than, or equal
Definition InstrTypes.h:693
@ ICMP_UGE
unsigned greater or equal
Definition InstrTypes.h:702
@ ICMP_UGT
unsigned greater than
Definition InstrTypes.h:701
@ ICMP_SGT
signed greater than
Definition InstrTypes.h:705
@ FCMP_ULT
1 1 0 0 True if unordered or less than
Definition InstrTypes.h:692
@ ICMP_ULT
unsigned less than
Definition InstrTypes.h:703
@ FCMP_OLE
0 1 0 1 True if ordered and less than or equal
Definition InstrTypes.h:685
@ FCMP_ORD
0 1 1 1 True if ordered (no nans)
Definition InstrTypes.h:687
@ ICMP_NE
not equal
Definition InstrTypes.h:700
@ ICMP_SGE
signed greater or equal
Definition InstrTypes.h:706
@ ICMP_ULE
unsigned less or equal
Definition InstrTypes.h:704
@ FCMP_UNO
1 0 0 0 True if unordered: isnan(X) | isnan(Y)
Definition InstrTypes.h:688
bool isSigned() const
Definition InstrTypes.h:932
Predicate getSwappedPredicate() const
For example, EQ->EQ, SLE->SGE, ULT->UGT, OEQ->OEQ, ULE->UGE, OLT->OGT, etc.
Definition InstrTypes.h:829
Predicate getInversePredicate() const
For example, EQ -> NE, UGT -> ULE, SLT -> SGE, OEQ -> UNE, UGT -> OLE, OLT -> UGE,...
Definition InstrTypes.h:791
Predicate getPredicate() const
Return the predicate for this instruction.
Definition InstrTypes.h:767
static LLVM_ABI bool isUnordered(Predicate predicate)
Determine if the predicate is an unordered operation.
static Predicate getOrderedPredicate(Predicate Pred)
Returns the ordered variant of a floating point compare.
Definition InstrTypes.h:798
An abstraction over a floating-point predicate, and a pack of an integer predicate with samesign info...
static LLVM_ABI Constant * getSub(Constant *C1, Constant *C2, bool HasNUW=false, bool HasNSW=false)
static LLVM_ABI Constant * getNot(Constant *C)
static LLVM_ABI Constant * getXor(Constant *C1, Constant *C2)
static LLVM_ABI Constant * getAdd(Constant *C1, Constant *C2, bool HasNUW=false, bool HasNSW=false)
static LLVM_ABI Constant * getTrunc(Constant *C, Type *Ty, bool OnlyIfReduced=false)
static LLVM_ABI Constant * getExactLogBase2(Constant *C)
If C is a scalar/fixed width vector of known powers of 2, then this function returns a new scalar/fix...
static LLVM_ABI Constant * getZero(Type *Ty, bool Negative=false)
This is the shared class of boolean and integer constants.
Definition Constants.h:87
bool isMinusOne() const
This function will return true iff every bit in this constant is set to true.
Definition Constants.h:226
static LLVM_ABI ConstantInt * getTrue(LLVMContext &Context)
bool isZero() const
This is just a convenience method to make client code smaller for a common code.
Definition Constants.h:214
static LLVM_ABI ConstantInt * getFalse(LLVMContext &Context)
uint64_t getZExtValue() const
Return the constant as a 64-bit unsigned integer value after it has been zero extended as appropriate...
Definition Constants.h:163
const APInt & getValue() const
Return the constant as an APInt value reference.
Definition Constants.h:154
LLVM_ABI std::optional< ConstantRange > exactUnionWith(const ConstantRange &CR) const
Union the two ranges and return the result if it can be represented exactly, otherwise return std::nu...
LLVM_ABI ConstantRange subtract(const APInt &CI) const
Subtract the specified constant from the endpoints of this constant range.
const APInt & getLower() const
Return the lower value for this range.
LLVM_ABI bool isWrappedSet() const
Return true if this set wraps around the unsigned domain.
const APInt & getUpper() const
Return the upper value for this range.
static LLVM_ABI ConstantRange makeExactICmpRegion(CmpInst::Predicate Pred, const APInt &Other)
Produce the exact range such that all values in the returned range satisfy the given predicate with a...
LLVM_ABI std::optional< ConstantRange > exactIntersectWith(const ConstantRange &CR) const
Intersect the two ranges and return the result if it can be represented exactly, otherwise return std...
This is an important base class in LLVM.
Definition Constant.h:43
static LLVM_ABI Constant * replaceUndefsWith(Constant *C, Constant *Replacement)
Try to replace undefined constant C or undefined elements in C with Replacement.
static LLVM_ABI Constant * mergeUndefsWith(Constant *C, Constant *Other)
Merges undefs of a Constant with another Constant, along with the undefs already present.
static LLVM_ABI Constant * getAllOnesValue(Type *Ty)
static LLVM_ABI Constant * getNullValue(Type *Ty)
Constructor to create a '0' constant of arbitrary type.
LLVM_ABI Constant * getAggregateElement(unsigned Elt) const
For aggregates (struct/array/vector) return the constant that corresponds to the specified element if...
LLVM_ABI bool isZeroValue() const
Return true if the value is negative zero or null value.
Definition Constants.cpp:76
A parsed version of the target data layout string in and methods for querying it.
Definition DataLayout.h:63
This instruction compares its operands according to the predicate given to the constructor.
This provides a helper for copying FMF from an instruction or setting specified flags.
Definition IRBuilder.h:93
static FMFSource intersect(Value *A, Value *B)
Intersect the FMF from two instructions.
Definition IRBuilder.h:107
This instruction compares its operands according to the predicate given to the constructor.
CmpPredicate getInverseCmpPredicate() const
Predicate getSignedPredicate() const
For example, EQ->EQ, SLE->SLE, UGT->SGT, etc.
bool isEquality() const
Return true if this predicate is either EQ or NE.
static bool isEquality(Predicate P)
Return true if this predicate is either EQ or NE.
Common base class shared among various IRBuilders.
Definition IRBuilder.h:114
Value * CreateNot(Value *V, const Twine &Name="")
Definition IRBuilder.h:1805
Value * CreateBinOp(Instruction::BinaryOps Opc, Value *LHS, Value *RHS, const Twine &Name="", MDNode *FPMathTag=nullptr)
Definition IRBuilder.h:1708
void SetInsertPoint(BasicBlock *TheBB)
This specifies that created instructions should be appended to the end of the specified block.
Definition IRBuilder.h:207
Instruction * canonicalizeCondSignextOfHighBitExtractToSignextHighBitExtract(BinaryOperator &I)
Instruction * foldBinOpIntoSelectOrPhi(BinaryOperator &I)
This is a convenience wrapper function for the above two functions.
Instruction * visitOr(BinaryOperator &I)
bool SimplifyAssociativeOrCommutative(BinaryOperator &I)
Performs a few simplifications for operators which are associative or commutative.
Value * foldUsingDistributiveLaws(BinaryOperator &I)
Tries to simplify binary operations which some other binary operation distributes over.
Instruction * foldBinOpShiftWithShift(BinaryOperator &I)
Value * insertRangeTest(Value *V, const APInt &Lo, const APInt &Hi, bool isSigned, bool Inside)
Emit a computation of: (V >= Lo && V < Hi) if Inside is true, otherwise (V < Lo || V >= Hi).
bool sinkNotIntoLogicalOp(Instruction &I)
std::optional< std::pair< Intrinsic::ID, SmallVector< Value *, 3 > > > convertOrOfShiftsToFunnelShift(Instruction &Or)
Instruction * visitAnd(BinaryOperator &I)
bool sinkNotIntoOtherHandOfLogicalOp(Instruction &I)
Instruction * foldBinopWithPhiOperands(BinaryOperator &BO)
For a binary operator with 2 phi operands, try to hoist the binary operation before the phi.
Instruction * foldAddLikeCommutative(Value *LHS, Value *RHS, bool NSW, bool NUW)
Common transforms for add / disjoint or.
Value * simplifyRangeCheck(ICmpInst *Cmp0, ICmpInst *Cmp1, bool Inverted)
Try to fold a signed range checked with lower bound 0 to an unsigned icmp.
Instruction * tryFoldInstWithCtpopWithNot(Instruction *I)
Value * SimplifyAddWithRemainder(BinaryOperator &I)
Tries to simplify add operations using the definition of remainder.
Instruction * visitXor(BinaryOperator &I)
bool SimplifyDemandedInstructionBits(Instruction &Inst)
Tries to simplify operands to an integer instruction based on its demanded bits.
Instruction * foldVectorBinop(BinaryOperator &Inst)
Canonicalize the position of binops relative to shufflevector.
Instruction * matchBSwapOrBitReverse(Instruction &I, bool MatchBSwaps, bool MatchBitReversals)
Given an initial instruction, check to see if it is the root of a bswap/bitreverse idiom.
void freelyInvertAllUsersOf(Value *V, Value *IgnoredUser=nullptr)
Freely adapt every user of V as-if V was changed to !V.
The core instruction combiner logic.
SimplifyQuery SQ
const DataLayout & getDataLayout() const
IRBuilder< TargetFolder, IRBuilderCallbackInserter > BuilderTy
An IRBuilder that automatically inserts new instructions into the worklist.
bool isFreeToInvert(Value *V, bool WillInvertAllUses, bool &DoesConsume)
Return true if the specified value is free to invert (apply ~ to).
unsigned ComputeNumSignBits(const Value *Op, const Instruction *CxtI=nullptr, unsigned Depth=0) const
Instruction * replaceInstUsesWith(Instruction &I, Value *V)
A combiner-aware RAUW-like routine.
InstructionWorklist & Worklist
A worklist of the instructions that need to be simplified.
const DataLayout & DL
void computeKnownBits(const Value *V, KnownBits &Known, const Instruction *CxtI, unsigned Depth=0) const
static Value * peekThroughBitcast(Value *V, bool OneUseOnly=false)
Return the source operand of a potentially bitcasted value while optionally checking if it has one us...
bool canFreelyInvertAllUsersOf(Instruction *V, Value *IgnoredUser)
Given i1 V, can every user of V be freely adapted if V is changed to !V ?
void addToWorklist(Instruction *I)
bool MaskedValueIsZero(const Value *V, const APInt &Mask, const Instruction *CxtI=nullptr, unsigned Depth=0) const
DominatorTree & DT
BuilderTy & Builder
Value * getFreelyInverted(Value *V, bool WillInvertAllUses, BuilderTy *Builder, bool &DoesConsume)
const SimplifyQuery & getSimplifyQuery() const
bool isKnownToBeAPowerOfTwo(const Value *V, bool OrZero=false, const Instruction *CxtI=nullptr, unsigned Depth=0)
LLVM_ABI void removeFromParent()
This method unlinks 'this' from the containing basic block, but does not delete it.
unsigned getOpcode() const
Returns a member of one of the enums like Instruction::Add.
A wrapper class for inspecting calls to intrinsic functions.
This class represents a sign extension of integer types.
static SelectInst * Create(Value *C, Value *S1, Value *S2, const Twine &NameStr="", InsertPosition InsertBefore=nullptr, Instruction *MDFrom=nullptr)
This is a 'bitvector' (really, a variable-sized bit array), optimized for the case when the array is ...
This is a 'vector' (really, a variable-sized array), optimized for the case when the array is small.
The instances of the Type class are immutable: once they are created, they are never changed.
Definition Type.h:45
LLVM_ABI unsigned getIntegerBitWidth() const
bool isVectorTy() const
True if this is an instance of VectorType.
Definition Type.h:273
bool isIntOrIntVectorTy() const
Return true if this is an integer type or a vector of integer types.
Definition Type.h:246
Type * getScalarType() const
If this is a vector type, return the element type, otherwise return 'this'.
Definition Type.h:352
LLVM_ABI TypeSize getPrimitiveSizeInBits() const LLVM_READONLY
Return the basic size of this type if it is a primitive type.
Definition Type.cpp:198
LLVM_ABI Type * getWithNewBitWidth(unsigned NewBitWidth) const
Given an integer or vector type, change the lane bitwidth to NewBitwidth, whilst keeping the old numb...
LLVM_ABI unsigned getScalarSizeInBits() const LLVM_READONLY
If this is a vector type, return the getPrimitiveSizeInBits value for the element type.
Definition Type.cpp:231
bool isFloatingPointTy() const
Return true if this is one of the floating-point types.
Definition Type.h:184
LLVM_ABI const fltSemantics & getFltSemantics() const
Definition Type.cpp:107
A Use represents the edge between a Value definition and its users.
Definition Use.h:35
Value * getOperand(unsigned i) const
Definition User.h:232
LLVM Value Representation.
Definition Value.h:75
Type * getType() const
All values are typed, get the type of this value.
Definition Value.h:256
bool hasOneUse() const
Return true if there is exactly one use of this value.
Definition Value.h:439
iterator_range< user_iterator > users()
Definition Value.h:426
LLVM_ABI bool hasNUsesOrMore(unsigned N) const
Return true if this value has N uses or more.
Definition Value.cpp:158
LLVM_ABI bool hasNUses(unsigned N) const
Return true if this Value has exactly N uses.
Definition Value.cpp:150
bool use_empty() const
Definition Value.h:346
LLVM_ABI StringRef getName() const
Return a constant reference to the value's name.
Definition Value.cpp:322
LLVM_ABI void takeName(Value *V)
Transfer the name from V to this value.
Definition Value.cpp:396
static LLVM_ABI VectorType * get(Type *ElementType, ElementCount EC)
This static method is the primary way to construct an VectorType.
Represents an op.with.overflow intrinsic.
This class represents zero extension of integer types.
constexpr ScalarTy getKnownMinValue() const
Returns the minimum value this quantity can represent.
Definition TypeSize.h:166
#define llvm_unreachable(msg)
Marks that the current location is not supposed to be reachable.
const APInt & umin(const APInt &A, const APInt &B)
Determine the smaller of two APInts considered to be unsigned.
Definition APInt.h:2258
@ C
The default llvm calling convention, compatible with C.
Definition CallingConv.h:34
LLVM_ABI Function * getOrInsertDeclaration(Module *M, ID id, ArrayRef< Type * > Tys={})
Look up the Function declaration of the intrinsic id in the Module M.
SpecificConstantMatch m_ZeroInt()
Convenience matchers for specific integer values.
BinaryOp_match< SpecificConstantMatch, SrcTy, TargetOpcode::G_SUB > m_Neg(const SrcTy &&Src)
Matches a register negated by a G_SUB.
BinaryOp_match< SrcTy, SpecificConstantMatch, TargetOpcode::G_XOR, true > m_Not(const SrcTy &&Src)
Matches a register not-ed by a G_XOR.
OneUse_match< SubPat > m_OneUse(const SubPat &SP)
cst_pred_ty< is_all_ones > m_AllOnes()
Match an integer or vector with all bits set.
cst_pred_ty< is_lowbit_mask > m_LowBitMask()
Match an integer or vector with only the low bit(s) set.
BinaryOp_match< LHS, RHS, Instruction::And > m_And(const LHS &L, const RHS &R)
cst_pred_ty< is_negative > m_Negative()
Match an integer or vector of negative values.
BinaryOp_match< LHS, RHS, Instruction::Add > m_Add(const LHS &L, const RHS &R)
class_match< BinaryOperator > m_BinOp()
Match an arbitrary binary operation and ignore it.
CmpClass_match< LHS, RHS, FCmpInst > m_FCmp(CmpPredicate &Pred, const LHS &L, const RHS &R)
cst_pred_ty< is_sign_mask > m_SignMask()
Match an integer or vector with only the sign bit(s) set.
BinaryOp_match< LHS, RHS, Instruction::AShr > m_AShr(const LHS &L, const RHS &R)
cstfp_pred_ty< is_inf > m_Inf()
Match a positive or negative infinity FP constant.
m_Intrinsic_Ty< Opnd0 >::Ty m_BitReverse(const Opnd0 &Op0)
cst_pred_ty< is_power2 > m_Power2()
Match an integer or vector power-of-2.
match_combine_or< CastInst_match< OpTy, TruncInst >, OpTy > m_TruncOrSelf(const OpTy &Op)
auto m_LogicalOp()
Matches either L && R or L || R where L and R are arbitrary values.
class_match< Constant > m_Constant()
Match an arbitrary Constant and ignore it.
BinaryOp_match< LHS, RHS, Instruction::And, true > m_c_And(const LHS &L, const RHS &R)
Matches an And with LHS and RHS in either order.
CastInst_match< OpTy, TruncInst > m_Trunc(const OpTy &Op)
Matches Trunc.
BinaryOp_match< LHS, RHS, Instruction::Xor > m_Xor(const LHS &L, const RHS &R)
OverflowingBinaryOp_match< LHS, RHS, Instruction::Sub, OverflowingBinaryOperator::NoSignedWrap > m_NSWSub(const LHS &L, const RHS &R)
specific_intval< false > m_SpecificInt(const APInt &V)
Match a specific integer value or vector with all elements equal to the value.
match_combine_or< CastInst_match< OpTy, ZExtInst >, OpTy > m_ZExtOrSelf(const OpTy &Op)
bool match(Val *V, const Pattern &P)
cst_pred_ty< is_shifted_mask > m_ShiftedMask()
bind_ty< Instruction > m_Instruction(Instruction *&I)
Match an instruction, capturing it if we match.
cstfp_pred_ty< is_any_zero_fp > m_AnyZeroFP()
Match a floating-point negative zero or positive zero.
specificval_ty m_Specific(const Value *V)
Match if we have a specific specified value.
DisjointOr_match< LHS, RHS > m_DisjointOr(const LHS &L, const RHS &R)
constantexpr_match m_ConstantExpr()
Match a constant expression or a constant that contains a constant expression.
specific_intval< true > m_SpecificIntAllowPoison(const APInt &V)
CmpClass_match< LHS, RHS, ICmpInst, true > m_c_ICmp(CmpPredicate &Pred, const LHS &L, const RHS &R)
Matches an ICmp with a predicate over LHS and RHS in either order.
TwoOps_match< Val_t, Idx_t, Instruction::ExtractElement > m_ExtractElt(const Val_t &Val, const Idx_t &Idx)
Matches ExtractElementInst.
cst_pred_ty< is_nonnegative > m_NonNegative()
Match an integer or vector of non-negative values.
class_match< ConstantInt > m_ConstantInt()
Match an arbitrary ConstantInt and ignore it.
cst_pred_ty< is_one > m_One()
Match an integer 1 or a vector with all elements equal to 1.
IntrinsicID_match m_Intrinsic()
Match intrinsic calls like this: m_Intrinsic<Intrinsic::fabs>(m_Value(X))
ThreeOps_match< Cond, LHS, RHS, Instruction::Select > m_Select(const Cond &C, const LHS &L, const RHS &R)
Matches SelectInst.
match_combine_or< CastInst_match< OpTy, SExtInst >, OpTy > m_SExtOrSelf(const OpTy &Op)
ExtractValue_match< Ind, Val_t > m_ExtractValue(const Val_t &V)
Match a single index ExtractValue instruction.
BinOpPred_match< LHS, RHS, is_logical_shift_op > m_LogicalShift(const LHS &L, const RHS &R)
Matches logical shift operations.
ShiftLike_match< LHS, Instruction::Shl > m_ShlOrSelf(const LHS &L, uint64_t &R)
Matches shl L, ConstShAmt or L itself (R will be set to zero in this case).
bind_ty< WithOverflowInst > m_WithOverflowInst(WithOverflowInst *&I)
Match a with overflow intrinsic, capturing it if we match.
BinaryOp_match< LHS, RHS, Instruction::Xor, true > m_c_Xor(const LHS &L, const RHS &R)
Matches an Xor with LHS and RHS in either order.
SpecificCmpClass_match< LHS, RHS, CmpInst > m_SpecificCmp(CmpPredicate MatchPred, const LHS &L, const RHS &R)
BinaryOp_match< LHS, RHS, Instruction::Mul > m_Mul(const LHS &L, const RHS &R)
deferredval_ty< Value > m_Deferred(Value *const &V)
Like m_Specific(), but works if the specific value to match is determined as part of the same match()...
apint_match m_APIntAllowPoison(const APInt *&Res)
Match APInt while allowing poison in splat vector constants.
auto m_LogicalOr()
Matches L || R where L and R are arbitrary values.
TwoOps_match< V1_t, V2_t, Instruction::ShuffleVector > m_Shuffle(const V1_t &v1, const V2_t &v2)
Matches ShuffleVectorInst independently of mask value.
SpecificCmpClass_match< LHS, RHS, ICmpInst > m_SpecificICmp(CmpPredicate MatchPred, const LHS &L, const RHS &R)
CastInst_match< OpTy, ZExtInst > m_ZExt(const OpTy &Op)
Matches ZExt.
class_match< CmpInst > m_Cmp()
Matches any compare instruction and ignore it.
cst_pred_ty< is_negated_power2 > m_NegatedPower2()
Match a integer or vector negated power-of-2.
match_immconstant_ty m_ImmConstant()
Match an arbitrary immediate Constant and ignore it.
DisjointOr_match< LHS, RHS, true > m_c_DisjointOr(const LHS &L, const RHS &R)
BinaryOp_match< LHS, RHS, Instruction::Add, true > m_c_Add(const LHS &L, const RHS &R)
Matches a Add with LHS and RHS in either order.
SpecificCmpClass_match< LHS, RHS, FCmpInst > m_SpecificFCmp(CmpPredicate MatchPred, const LHS &L, const RHS &R)
apfloat_match m_APFloatAllowPoison(const APFloat *&Res)
Match APFloat while allowing poison in splat vector constants.
match_combine_or< BinaryOp_match< LHS, RHS, Instruction::Add >, DisjointOr_match< LHS, RHS > > m_AddLike(const LHS &L, const RHS &R)
Match either "add" or "or disjoint".
CastOperator_match< OpTy, Instruction::BitCast > m_BitCast(const OpTy &Op)
Matches BitCast.
match_combine_or< match_combine_or< MaxMin_match< ICmpInst, LHS, RHS, smax_pred_ty, true >, MaxMin_match< ICmpInst, LHS, RHS, smin_pred_ty, true > >, match_combine_or< MaxMin_match< ICmpInst, LHS, RHS, umax_pred_ty, true >, MaxMin_match< ICmpInst, LHS, RHS, umin_pred_ty, true > > > m_c_MaxOrMin(const LHS &L, const RHS &R)
match_combine_or< CastInst_match< OpTy, SExtInst >, NNegZExt_match< OpTy > > m_SExtLike(const OpTy &Op)
Match either "sext" or "zext nneg".
apint_match m_APInt(const APInt *&Res)
Match a ConstantInt or splatted ConstantVector, binding the specified pointer to the contained APInt.
cst_pred_ty< is_maxsignedvalue > m_MaxSignedValue()
Match an integer or vector with values having all bits except for the high bit set (0x7f....
class_match< Value > m_Value()
Match an arbitrary value and ignore it.
AnyBinaryOp_match< LHS, RHS, true > m_c_BinOp(const LHS &L, const RHS &R)
Matches a BinaryOperator with LHS and RHS in either order.
BinaryOp_match< LHS, RHS, Instruction::LShr > m_LShr(const LHS &L, const RHS &R)
CmpClass_match< LHS, RHS, ICmpInst > m_ICmp(CmpPredicate &Pred, const LHS &L, const RHS &R)
match_combine_or< CastInst_match< OpTy, ZExtInst >, CastInst_match< OpTy, SExtInst > > m_ZExtOrSExt(const OpTy &Op)
FNeg_match< OpTy > m_FNeg(const OpTy &X)
Match 'fneg X' as 'fsub -0.0, X'.
BinOpPred_match< LHS, RHS, is_shift_op > m_Shift(const LHS &L, const RHS &R)
Matches shift operations.
cstfp_pred_ty< is_pos_zero_fp > m_PosZeroFP()
Match a floating-point positive zero.
BinaryOp_match< LHS, RHS, Instruction::Shl > m_Shl(const LHS &L, const RHS &R)
auto m_LogicalAnd()
Matches L && R where L and R are arbitrary values.
BinaryOp_match< LHS, RHS, Instruction::Or > m_Or(const LHS &L, const RHS &R)
m_Intrinsic_Ty< Opnd0 >::Ty m_BSwap(const Opnd0 &Op0)
CastInst_match< OpTy, SExtInst > m_SExt(const OpTy &Op)
Matches SExt.
is_zero m_Zero()
Match any null constant or a vector with all elements equal to 0.
BinaryOp_match< LHS, RHS, Instruction::Or, true > m_c_Or(const LHS &L, const RHS &R)
Matches an Or with LHS and RHS in either order.
ThreeOps_match< Val_t, Elt_t, Idx_t, Instruction::InsertElement > m_InsertElt(const Val_t &Val, const Elt_t &Elt, const Idx_t &Idx)
Matches InsertElementInst.
ElementWiseBitCast_match< OpTy > m_ElementWiseBitCast(const OpTy &Op)
m_Intrinsic_Ty< Opnd0 >::Ty m_FAbs(const Opnd0 &Op0)
m_Intrinsic_Ty< Opnd0, Opnd1 >::Ty m_CopySign(const Opnd0 &Op0, const Opnd1 &Op1)
BinaryOp_match< LHS, RHS, Instruction::Sub > m_Sub(const LHS &L, const RHS &R)
match_unless< Ty > m_Unless(const Ty &M)
Match if the inner matcher does NOT match.
match_combine_or< LTy, RTy > m_CombineOr(const LTy &L, const RTy &R)
Combine two pattern matchers matching L || R.
cst_pred_ty< icmp_pred_with_threshold > m_SpecificInt_ICMP(ICmpInst::Predicate Predicate, const APInt &Threshold)
Match an integer or vector with every element comparing 'pred' (eg/ne/...) to Threshold.
NodeAddr< CodeNode * > Code
Definition RDFGraph.h:388
friend class Instruction
Iterator for Instructions in a `BasicBlock.
Definition BasicBlock.h:73
This is an optimization pass for GlobalISel generic memory operations.
LLVM_ABI Intrinsic::ID getInverseMinMaxIntrinsic(Intrinsic::ID MinMaxID)
@ Low
Lower the current thread's priority such that it does not affect foreground tasks significantly.
Definition Threading.h:262
@ Offset
Definition DWP.cpp:477
FunctionAddr VTableAddr Value
Definition InstrProf.h:137
Constant * getPredForFCmpCode(unsigned Code, Type *OpTy, CmpInst::Predicate &Pred)
This is the complement of getFCmpCode.
LLVM_ABI bool isSignBitCheck(ICmpInst::Predicate Pred, const APInt &RHS, bool &TrueIfSigned)
Given an exploded icmp instruction, return true if the comparison only checks the sign bit.
decltype(auto) dyn_cast(const From &Val)
dyn_cast<X> - Return the argument parameter cast to the specified type.
Definition Casting.h:649
bool predicatesFoldable(CmpInst::Predicate P1, CmpInst::Predicate P2)
Return true if both predicates match sign or if at least one of them is an equality comparison (which...
LLVM_ABI Constant * ConstantFoldCompareInstOperands(unsigned Predicate, Constant *LHS, Constant *RHS, const DataLayout &DL, const TargetLibraryInfo *TLI=nullptr, const Instruction *I=nullptr)
Attempt to constant fold a compare instruction (icmp/fcmp) with the specified operands.
LLVM_ABI Value * simplifyOrInst(Value *LHS, Value *RHS, const SimplifyQuery &Q)
Given operands for an Or, fold the result or return null.
LLVM_ABI Value * simplifyXorInst(Value *LHS, Value *RHS, const SimplifyQuery &Q)
Given operands for an Xor, fold the result or return null.
LLVM_ABI bool isGuaranteedNotToBeUndef(const Value *V, AssumptionCache *AC=nullptr, const Instruction *CtxI=nullptr, const DominatorTree *DT=nullptr, unsigned Depth=0)
Returns true if V cannot be undef, but may be poison.
LLVM_ABI bool matchSimpleRecurrence(const PHINode *P, BinaryOperator *&BO, Value *&Start, Value *&Step)
Attempt to match a simple first order recurrence cycle of the form: iv = phi Ty [Start,...
LLVM_ABI bool isKnownNegative(const Value *V, const SimplifyQuery &SQ, unsigned Depth=0)
Returns true if the given value is known be negative (i.e.
LLVM_ABI Constant * getLosslessUnsignedTrunc(Constant *C, Type *DestTy, const DataLayout &DL, PreservedCastFlags *Flags=nullptr)
LLVM_ABI bool recognizeBSwapOrBitReverseIdiom(Instruction *I, bool MatchBSwaps, bool MatchBitReversals, SmallVectorImpl< Instruction * > &InsertedInsts)
Try to match a bswap or bitreverse idiom.
Definition Local.cpp:3725
constexpr bool isPowerOf2_32(uint32_t Value)
Return true if the argument is a power of two > 0.
Definition MathExtras.h:288
LLVM_ABI Value * simplifyICmpInst(CmpPredicate Pred, Value *LHS, Value *RHS, const SimplifyQuery &Q)
Given operands for an ICmpInst, fold the result or return null.
LLVM_ABI Constant * getLosslessSignedTrunc(Constant *C, Type *DestTy, const DataLayout &DL, PreservedCastFlags *Flags=nullptr)
LLVM_ABI Value * simplifyAndInst(Value *LHS, Value *RHS, const SimplifyQuery &Q)
Given operands for an And, fold the result or return null.
LLVM_ABI bool isKnownInversion(const Value *X, const Value *Y)
Return true iff:
bool isa(const From &Val)
isa<X> - Return true if the parameter to the template is an instance of one of the template type argu...
Definition Casting.h:548
LLVM_ABI bool isKnownNonZero(const Value *V, const SimplifyQuery &Q, unsigned Depth=0)
Return true if the given value is known to be non-zero when defined.
constexpr int PoisonMaskElem
@ Other
Any other memory.
Definition ModRef.h:68
LLVM_ABI Value * simplifyBinOp(unsigned Opcode, Value *LHS, Value *RHS, const SimplifyQuery &Q)
Given operands for a BinaryOperator, fold the result or return null.
std::optional< DecomposedBitTest > decomposeBitTest(Value *Cond, bool LookThroughTrunc=true, bool AllowNonZeroC=false, bool DecomposeAnd=false)
Decompose an icmp into the form ((X & Mask) pred C) if possible.
@ Mul
Product of integers.
@ Xor
Bitwise or logical XOR of integers.
@ And
Bitwise or logical AND of integers.
@ Sub
Subtraction of integers.
@ Add
Sum of integers.
DWARFExpression::Operation Op
LLVM_ABI bool isGuaranteedNotToBeUndefOrPoison(const Value *V, AssumptionCache *AC=nullptr, const Instruction *CtxI=nullptr, const DominatorTree *DT=nullptr, unsigned Depth=0)
Return true if this function can prove that V does not have undef bits and is never poison.
constexpr unsigned BitWidth
decltype(auto) cast(const From &Val)
cast<X> - Return the argument parameter cast to the specified type.
Definition Casting.h:565
APFloat neg(APFloat X)
Returns the negated value of the argument.
Definition APFloat.h:1569
unsigned getICmpCode(CmpInst::Predicate Pred)
Encode a icmp predicate into a three bit mask.
LLVM_ABI bool isKnownToBeAPowerOfTwo(const Value *V, const DataLayout &DL, bool OrZero=false, AssumptionCache *AC=nullptr, const Instruction *CxtI=nullptr, const DominatorTree *DT=nullptr, bool UseInstrInfo=true, unsigned Depth=0)
Return true if the given value is known to have exactly one bit set when defined.
LLVM_ABI bool isGuaranteedNotToBePoison(const Value *V, AssumptionCache *AC=nullptr, const Instruction *CtxI=nullptr, const DominatorTree *DT=nullptr, unsigned Depth=0)
Returns true if V cannot be poison, but may be undef.
std::pair< Value *, FPClassTest > fcmpToClassTest(FCmpInst::Predicate Pred, const Function &F, Value *LHS, Value *RHS, bool LookThroughSrc=true)
Returns a pair of values, which if passed to llvm.is.fpclass, returns the same result as an fcmp with...
unsigned getFCmpCode(CmpInst::Predicate CC)
Similar to getICmpCode but for FCmpInst.
std::optional< DecomposedBitTest > decomposeBitTestICmp(Value *LHS, Value *RHS, CmpInst::Predicate Pred, bool LookThroughTrunc=true, bool AllowNonZeroC=false, bool DecomposeAnd=false)
Decompose an icmp into the form ((X & Mask) pred C) if possible.
Constant * getPredForICmpCode(unsigned Code, bool Sign, Type *OpTy, CmpInst::Predicate &Pred)
This is the complement of getICmpCode.
void swap(llvm::BitVector &LHS, llvm::BitVector &RHS)
Implement std::swap in terms of BitVector swap.
Definition BitVector.h:853
#define N
bool isCombineableWith(const DecomposedBitMaskMul Other)
static LLVM_ABI bool hasSignBitInMSB(const fltSemantics &)
Definition APFloat.cpp:370
bool isNonNegative() const
Returns true if this value is known to be non-negative.
Definition KnownBits.h:101
APInt getMaxValue() const
Return the maximal unsigned value possible given these KnownBits.
Definition KnownBits.h:138
Matching combinators.
const DataLayout & DL
const Instruction * CxtI
const DominatorTree * DT
SimplifyQuery getWithInstruction(const Instruction *I) const
AssumptionCache * AC