LLVM 23.0.0git
InstCombineAndOrXor.cpp
Go to the documentation of this file.
1//===- InstCombineAndOrXor.cpp --------------------------------------------===//
2//
3// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions.
4// See https://llvm.org/LICENSE.txt for license information.
5// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception
6//
7//===----------------------------------------------------------------------===//
8//
9// This file implements the visitAnd, visitOr, and visitXor functions.
10//
11//===----------------------------------------------------------------------===//
12
13#include "InstCombineInternal.h"
21#include "llvm/IR/Intrinsics.h"
26
27using namespace llvm;
28using namespace PatternMatch;
29
30#define DEBUG_TYPE "instcombine"
31
32namespace llvm {
34}
35
36/// This is the complement of getICmpCode, which turns an opcode and two
37/// operands into either a constant true or false, or a brand new ICmp
38/// instruction. The sign is passed in to determine which kind of predicate to
39/// use in the new icmp instruction.
40static Value *getNewICmpValue(unsigned Code, bool Sign, Value *LHS, Value *RHS,
41 InstCombiner::BuilderTy &Builder) {
42 ICmpInst::Predicate NewPred;
43 if (Constant *TorF = getPredForICmpCode(Code, Sign, LHS->getType(), NewPred))
44 return TorF;
45 return Builder.CreateICmp(NewPred, LHS, RHS);
46}
47
48/// This is the complement of getFCmpCode, which turns an opcode and two
49/// operands into either a FCmp instruction, or a true/false constant.
50static Value *getFCmpValue(unsigned Code, Value *LHS, Value *RHS,
51 InstCombiner::BuilderTy &Builder, FMFSource FMF) {
52 FCmpInst::Predicate NewPred;
53 if (Constant *TorF = getPredForFCmpCode(Code, LHS->getType(), NewPred))
54 return TorF;
55 return Builder.CreateFCmpFMF(NewPred, LHS, RHS, FMF);
56}
57
58/// Emit a computation of: (V >= Lo && V < Hi) if Inside is true, otherwise
59/// (V < Lo || V >= Hi). This method expects that Lo < Hi. IsSigned indicates
60/// whether to treat V, Lo, and Hi as signed or not.
62 const APInt &Hi, bool isSigned,
63 bool Inside) {
64 assert((isSigned ? Lo.slt(Hi) : Lo.ult(Hi)) &&
65 "Lo is not < Hi in range emission code!");
66
67 Type *Ty = V->getType();
68
69 // V >= Min && V < Hi --> V < Hi
70 // V < Min || V >= Hi --> V >= Hi
72 if (isSigned ? Lo.isMinSignedValue() : Lo.isMinValue()) {
73 Pred = isSigned ? ICmpInst::getSignedPredicate(Pred) : Pred;
74 return Builder.CreateICmp(Pred, V, ConstantInt::get(Ty, Hi));
75 }
76
77 // V >= Lo && V < Hi --> V - Lo u< Hi - Lo
78 // V < Lo || V >= Hi --> V - Lo u>= Hi - Lo
79 Value *VMinusLo =
80 Builder.CreateSub(V, ConstantInt::get(Ty, Lo), V->getName() + ".off");
81 Constant *HiMinusLo = ConstantInt::get(Ty, Hi - Lo);
82 return Builder.CreateICmp(Pred, VMinusLo, HiMinusLo);
83}
84
85/// Classify (icmp eq (A & B), C) and (icmp ne (A & B), C) as matching patterns
86/// that can be simplified.
87/// One of A and B is considered the mask. The other is the value. This is
88/// described as the "AMask" or "BMask" part of the enum. If the enum contains
89/// only "Mask", then both A and B can be considered masks. If A is the mask,
90/// then it was proven that (A & C) == C. This is trivial if C == A or C == 0.
91/// If both A and C are constants, this proof is also easy.
92/// For the following explanations, we assume that A is the mask.
93///
94/// "AllOnes" declares that the comparison is true only if (A & B) == A or all
95/// bits of A are set in B.
96/// Example: (icmp eq (A & 3), 3) -> AMask_AllOnes
97///
98/// "AllZeros" declares that the comparison is true only if (A & B) == 0 or all
99/// bits of A are cleared in B.
100/// Example: (icmp eq (A & 3), 0) -> Mask_AllZeroes
101///
102/// "Mixed" declares that (A & B) == C and C might or might not contain any
103/// number of one bits and zero bits.
104/// Example: (icmp eq (A & 3), 1) -> AMask_Mixed
105///
106/// "Not" means that in above descriptions "==" should be replaced by "!=".
107/// Example: (icmp ne (A & 3), 3) -> AMask_NotAllOnes
108///
109/// If the mask A contains a single bit, then the following is equivalent:
110/// (icmp eq (A & B), A) equals (icmp ne (A & B), 0)
111/// (icmp ne (A & B), A) equals (icmp eq (A & B), 0)
124
125/// Return the set of patterns (from MaskedICmpType) that (icmp SCC (A & B), C)
126/// satisfies.
127static unsigned getMaskedICmpType(Value *A, Value *B, Value *C,
128 ICmpInst::Predicate Pred) {
129 const APInt *ConstA = nullptr, *ConstB = nullptr, *ConstC = nullptr;
130 match(A, m_APInt(ConstA));
131 match(B, m_APInt(ConstB));
132 match(C, m_APInt(ConstC));
133 bool IsEq = (Pred == ICmpInst::ICMP_EQ);
134 bool IsAPow2 = ConstA && ConstA->isPowerOf2();
135 bool IsBPow2 = ConstB && ConstB->isPowerOf2();
136 unsigned MaskVal = 0;
137 if (ConstC && ConstC->isZero()) {
138 // if C is zero, then both A and B qualify as mask
139 MaskVal |= (IsEq ? (Mask_AllZeros | AMask_Mixed | BMask_Mixed)
141 if (IsAPow2)
142 MaskVal |= (IsEq ? (AMask_NotAllOnes | AMask_NotMixed)
144 if (IsBPow2)
145 MaskVal |= (IsEq ? (BMask_NotAllOnes | BMask_NotMixed)
147 return MaskVal;
148 }
149
150 if (A == C) {
151 MaskVal |= (IsEq ? (AMask_AllOnes | AMask_Mixed)
153 if (IsAPow2)
154 MaskVal |= (IsEq ? (Mask_NotAllZeros | AMask_NotMixed)
156 } else if (ConstA && ConstC && ConstC->isSubsetOf(*ConstA)) {
157 MaskVal |= (IsEq ? AMask_Mixed : AMask_NotMixed);
158 }
159
160 if (B == C) {
161 MaskVal |= (IsEq ? (BMask_AllOnes | BMask_Mixed)
163 if (IsBPow2)
164 MaskVal |= (IsEq ? (Mask_NotAllZeros | BMask_NotMixed)
166 } else if (ConstB && ConstC && ConstC->isSubsetOf(*ConstB)) {
167 MaskVal |= (IsEq ? BMask_Mixed : BMask_NotMixed);
168 }
169
170 return MaskVal;
171}
172
173/// Convert an analysis of a masked ICmp into its equivalent if all boolean
174/// operations had the opposite sense. Since each "NotXXX" flag (recording !=)
175/// is adjacent to the corresponding normal flag (recording ==), this just
176/// involves swapping those bits over.
177static unsigned conjugateICmpMask(unsigned Mask) {
178 unsigned NewMask;
179 NewMask = (Mask & (AMask_AllOnes | BMask_AllOnes | Mask_AllZeros |
181 << 1;
182
183 NewMask |= (Mask & (AMask_NotAllOnes | BMask_NotAllOnes | Mask_NotAllZeros |
185 >> 1;
186
187 return NewMask;
188}
189
190// Adapts the external decomposeBitTest for local use.
192 Value *&Y, Value *&Z) {
193 auto Res =
194 llvm::decomposeBitTest(Cond, /*LookThroughTrunc=*/true,
195 /*AllowNonZeroC=*/true, /*DecomposeAnd=*/true);
196 if (!Res)
197 return false;
198
199 Pred = Res->Pred;
200 X = Res->X;
201 Y = ConstantInt::get(X->getType(), Res->Mask);
202 Z = ConstantInt::get(X->getType(), Res->C);
203 return true;
204}
205
206/// Handle (icmp(A & B) ==/!= C) &/| (icmp(A & D) ==/!= E).
207/// Return the pattern classes (from MaskedICmpType) for the left hand side and
208/// the right hand side as a pair.
209/// LHS and RHS are the left hand side and the right hand side ICmps and PredL
210/// and PredR are their predicates, respectively.
211static std::optional<std::pair<unsigned, unsigned>>
214 ICmpInst::Predicate &PredR) {
215
216 // Here comes the tricky part:
217 // LHS might be of the form L11 & L12 == X, X == L21 & L22,
218 // and L11 & L12 == L21 & L22. The same goes for RHS.
219 // Now we must find those components L** and R**, that are equal, so
220 // that we can extract the parameters A, B, C, D, and E for the canonical
221 // above.
222
223 // Check whether the icmp can be decomposed into a bit test.
224 Value *L1, *L11, *L12, *L2, *L21, *L22;
225 if (decomposeBitTest(LHS, PredL, L11, L12, L2)) {
226 L21 = L22 = L1 = nullptr;
227 } else {
228 auto *LHSCMP = dyn_cast<ICmpInst>(LHS);
229 if (!LHSCMP)
230 return std::nullopt;
231
232 // Don't allow pointers. Splat vectors are fine.
233 if (!LHSCMP->getOperand(0)->getType()->isIntOrIntVectorTy())
234 return std::nullopt;
235
236 PredL = LHSCMP->getPredicate();
237 L1 = LHSCMP->getOperand(0);
238 L2 = LHSCMP->getOperand(1);
239 // Look for ANDs in the LHS icmp.
240 if (!match(L1, m_And(m_Value(L11), m_Value(L12)))) {
241 // Any icmp can be viewed as being trivially masked; if it allows us to
242 // remove one, it's worth it.
243 L11 = L1;
245 }
246
247 if (!match(L2, m_And(m_Value(L21), m_Value(L22)))) {
248 L21 = L2;
250 }
251 }
252
253 // Bail if LHS was a icmp that can't be decomposed into an equality.
254 if (!ICmpInst::isEquality(PredL))
255 return std::nullopt;
256
257 Value *R11, *R12, *R2;
258 if (decomposeBitTest(RHS, PredR, R11, R12, R2)) {
259 if (R11 == L11 || R11 == L12 || R11 == L21 || R11 == L22) {
260 A = R11;
261 D = R12;
262 } else if (R12 == L11 || R12 == L12 || R12 == L21 || R12 == L22) {
263 A = R12;
264 D = R11;
265 } else {
266 return std::nullopt;
267 }
268 E = R2;
269 } else {
270 auto *RHSCMP = dyn_cast<ICmpInst>(RHS);
271 if (!RHSCMP)
272 return std::nullopt;
273 // Don't allow pointers. Splat vectors are fine.
274 if (!RHSCMP->getOperand(0)->getType()->isIntOrIntVectorTy())
275 return std::nullopt;
276
277 PredR = RHSCMP->getPredicate();
278
279 Value *R1 = RHSCMP->getOperand(0);
280 R2 = RHSCMP->getOperand(1);
281 bool Ok = false;
282 if (!match(R1, m_And(m_Value(R11), m_Value(R12)))) {
283 // As before, model no mask as a trivial mask if it'll let us do an
284 // optimization.
285 R11 = R1;
287 }
288
289 if (R11 == L11 || R11 == L12 || R11 == L21 || R11 == L22) {
290 A = R11;
291 D = R12;
292 E = R2;
293 Ok = true;
294 } else if (R12 == L11 || R12 == L12 || R12 == L21 || R12 == L22) {
295 A = R12;
296 D = R11;
297 E = R2;
298 Ok = true;
299 }
300
301 // Avoid matching against the -1 value we created for unmasked operand.
302 if (Ok && match(A, m_AllOnes()))
303 Ok = false;
304
305 // Look for ANDs on the right side of the RHS icmp.
306 if (!Ok) {
307 if (!match(R2, m_And(m_Value(R11), m_Value(R12)))) {
308 R11 = R2;
309 R12 = Constant::getAllOnesValue(R2->getType());
310 }
311
312 if (R11 == L11 || R11 == L12 || R11 == L21 || R11 == L22) {
313 A = R11;
314 D = R12;
315 E = R1;
316 } else if (R12 == L11 || R12 == L12 || R12 == L21 || R12 == L22) {
317 A = R12;
318 D = R11;
319 E = R1;
320 } else {
321 return std::nullopt;
322 }
323 }
324 }
325
326 // Bail if RHS was a icmp that can't be decomposed into an equality.
327 if (!ICmpInst::isEquality(PredR))
328 return std::nullopt;
329
330 if (L11 == A) {
331 B = L12;
332 C = L2;
333 } else if (L12 == A) {
334 B = L11;
335 C = L2;
336 } else if (L21 == A) {
337 B = L22;
338 C = L1;
339 } else if (L22 == A) {
340 B = L21;
341 C = L1;
342 }
343
344 unsigned LeftType = getMaskedICmpType(A, B, C, PredL);
345 unsigned RightType = getMaskedICmpType(A, D, E, PredR);
346 return std::optional<std::pair<unsigned, unsigned>>(
347 std::make_pair(LeftType, RightType));
348}
349
350/// Try to fold (icmp(A & B) ==/!= C) &/| (icmp(A & D) ==/!= E) into a single
351/// (icmp(A & X) ==/!= Y), where the left-hand side is of type Mask_NotAllZeros
352/// and the right hand side is of type BMask_Mixed. For example,
353/// (icmp (A & 12) != 0) & (icmp (A & 15) == 8) -> (icmp (A & 15) == 8).
354/// Also used for logical and/or, must be poison safe.
356 Value *LHS, Value *RHS, bool IsAnd, Value *A, Value *B, Value *D, Value *E,
358 InstCombiner::BuilderTy &Builder) {
359 // We are given the canonical form:
360 // (icmp ne (A & B), 0) & (icmp eq (A & D), E).
361 // where D & E == E.
362 //
363 // If IsAnd is false, we get it in negated form:
364 // (icmp eq (A & B), 0) | (icmp ne (A & D), E) ->
365 // !((icmp ne (A & B), 0) & (icmp eq (A & D), E)).
366 //
367 // We currently handle the case of B, C, D, E are constant.
368 //
369 const APInt *BCst, *DCst, *OrigECst;
370 if (!match(B, m_APInt(BCst)) || !match(D, m_APInt(DCst)) ||
371 !match(E, m_APInt(OrigECst)))
372 return nullptr;
373
375
376 // Update E to the canonical form when D is a power of two and RHS is
377 // canonicalized as,
378 // (icmp ne (A & D), 0) -> (icmp eq (A & D), D) or
379 // (icmp ne (A & D), D) -> (icmp eq (A & D), 0).
380 APInt ECst = *OrigECst;
381 if (PredR != NewCC)
382 ECst ^= *DCst;
383
384 // If B or D is zero, skip because if LHS or RHS can be trivially folded by
385 // other folding rules and this pattern won't apply any more.
386 if (*BCst == 0 || *DCst == 0)
387 return nullptr;
388
389 // If B and D don't intersect, ie. (B & D) == 0, try to fold isNaN idiom:
390 // (icmp ne (A & FractionBits), 0) & (icmp eq (A & ExpBits), ExpBits)
391 // -> isNaN(A)
392 // Otherwise, we cannot deduce anything from it.
393 if (!BCst->intersects(*DCst)) {
394 Value *Src;
395 if (*DCst == ECst && match(A, m_ElementWiseBitCast(m_Value(Src))) &&
396 !Builder.GetInsertBlock()->getParent()->hasFnAttribute(
397 Attribute::StrictFP)) {
398 Type *Ty = Src->getType()->getScalarType();
399 if (!Ty->isIEEELikeFPTy())
400 return nullptr;
401
402 APInt ExpBits = APFloat::getInf(Ty->getFltSemantics()).bitcastToAPInt();
403 if (ECst != ExpBits)
404 return nullptr;
405 APInt FractionBits = ~ExpBits;
406 FractionBits.clearSignBit();
407 if (*BCst != FractionBits)
408 return nullptr;
409
410 return Builder.CreateFCmp(IsAnd ? FCmpInst::FCMP_UNO : FCmpInst::FCMP_ORD,
411 Src, ConstantFP::getZero(Src->getType()));
412 }
413 return nullptr;
414 }
415
416 // If the following two conditions are met:
417 //
418 // 1. mask B covers only a single bit that's not covered by mask D, that is,
419 // (B & (B ^ D)) is a power of 2 (in other words, B minus the intersection of
420 // B and D has only one bit set) and,
421 //
422 // 2. RHS (and E) indicates that the rest of B's bits are zero (in other
423 // words, the intersection of B and D is zero), that is, ((B & D) & E) == 0
424 //
425 // then that single bit in B must be one and thus the whole expression can be
426 // folded to
427 // (A & (B | D)) == (B & (B ^ D)) | E.
428 //
429 // For example,
430 // (icmp ne (A & 12), 0) & (icmp eq (A & 7), 1) -> (icmp eq (A & 15), 9)
431 // (icmp ne (A & 15), 0) & (icmp eq (A & 7), 0) -> (icmp eq (A & 15), 8)
432 if ((((*BCst & *DCst) & ECst) == 0) &&
433 (*BCst & (*BCst ^ *DCst)).isPowerOf2()) {
434 APInt BorD = *BCst | *DCst;
435 APInt BandBxorDorE = (*BCst & (*BCst ^ *DCst)) | ECst;
436 Value *NewMask = ConstantInt::get(A->getType(), BorD);
437 Value *NewMaskedValue = ConstantInt::get(A->getType(), BandBxorDorE);
438 Value *NewAnd = Builder.CreateAnd(A, NewMask);
439 return Builder.CreateICmp(NewCC, NewAnd, NewMaskedValue);
440 }
441
442 auto IsSubSetOrEqual = [](const APInt *C1, const APInt *C2) {
443 return (*C1 & *C2) == *C1;
444 };
445 auto IsSuperSetOrEqual = [](const APInt *C1, const APInt *C2) {
446 return (*C1 & *C2) == *C2;
447 };
448
449 // In the following, we consider only the cases where B is a superset of D, B
450 // is a subset of D, or B == D because otherwise there's at least one bit
451 // covered by B but not D, in which case we can't deduce much from it, so
452 // no folding (aside from the single must-be-one bit case right above.)
453 // For example,
454 // (icmp ne (A & 14), 0) & (icmp eq (A & 3), 1) -> no folding.
455 if (!IsSubSetOrEqual(BCst, DCst) && !IsSuperSetOrEqual(BCst, DCst))
456 return nullptr;
457
458 // At this point, either B is a superset of D, B is a subset of D or B == D.
459
460 // If E is zero, if B is a subset of (or equal to) D, LHS and RHS contradict
461 // and the whole expression becomes false (or true if negated), otherwise, no
462 // folding.
463 // For example,
464 // (icmp ne (A & 3), 0) & (icmp eq (A & 7), 0) -> false.
465 // (icmp ne (A & 15), 0) & (icmp eq (A & 3), 0) -> no folding.
466 if (ECst.isZero()) {
467 if (IsSubSetOrEqual(BCst, DCst))
468 return ConstantInt::get(LHS->getType(), !IsAnd);
469 return nullptr;
470 }
471
472 // At this point, B, D, E aren't zero and (B & D) == B, (B & D) == D or B ==
473 // D. If B is a superset of (or equal to) D, since E is not zero, LHS is
474 // subsumed by RHS (RHS implies LHS.) So the whole expression becomes
475 // RHS. For example,
476 // (icmp ne (A & 255), 0) & (icmp eq (A & 15), 8) -> (icmp eq (A & 15), 8).
477 // (icmp ne (A & 15), 0) & (icmp eq (A & 15), 8) -> (icmp eq (A & 15), 8).
478 if (IsSuperSetOrEqual(BCst, DCst)) {
479 // We can't guarantee that samesign hold after this fold.
480 if (auto *ICmp = dyn_cast<ICmpInst>(RHS))
481 ICmp->setSameSign(false);
482 return RHS;
483 }
484 // Otherwise, B is a subset of D. If B and E have a common bit set,
485 // ie. (B & E) != 0, then LHS is subsumed by RHS. For example.
486 // (icmp ne (A & 12), 0) & (icmp eq (A & 15), 8) -> (icmp eq (A & 15), 8).
487 assert(IsSubSetOrEqual(BCst, DCst) && "Precondition due to above code");
488 if ((*BCst & ECst) != 0) {
489 // We can't guarantee that samesign hold after this fold.
490 if (auto *ICmp = dyn_cast<ICmpInst>(RHS))
491 ICmp->setSameSign(false);
492 return RHS;
493 }
494 // Otherwise, LHS and RHS contradict and the whole expression becomes false
495 // (or true if negated.) For example,
496 // (icmp ne (A & 7), 0) & (icmp eq (A & 15), 8) -> false.
497 // (icmp ne (A & 6), 0) & (icmp eq (A & 15), 8) -> false.
498 return ConstantInt::get(LHS->getType(), !IsAnd);
499}
500
501/// Try to fold (icmp(A & B) ==/!= 0) &/| (icmp(A & D) ==/!= E) into a single
502/// (icmp(A & X) ==/!= Y), where the left-hand side and the right hand side
503/// aren't of the common mask pattern type.
504/// Also used for logical and/or, must be poison safe.
506 Value *LHS, Value *RHS, bool IsAnd, Value *A, Value *B, Value *C, Value *D,
508 unsigned LHSMask, unsigned RHSMask, InstCombiner::BuilderTy &Builder) {
510 "Expected equality predicates for masked type of icmps.");
511 // Handle Mask_NotAllZeros-BMask_Mixed cases.
512 // (icmp ne/eq (A & B), C) &/| (icmp eq/ne (A & D), E), or
513 // (icmp eq/ne (A & B), C) &/| (icmp ne/eq (A & D), E)
514 // which gets swapped to
515 // (icmp ne/eq (A & D), E) &/| (icmp eq/ne (A & B), C).
516 if (!IsAnd) {
517 LHSMask = conjugateICmpMask(LHSMask);
518 RHSMask = conjugateICmpMask(RHSMask);
519 }
520 if ((LHSMask & Mask_NotAllZeros) && (RHSMask & BMask_Mixed)) {
522 LHS, RHS, IsAnd, A, B, D, E, PredL, PredR, Builder)) {
523 return V;
524 }
525 } else if ((LHSMask & BMask_Mixed) && (RHSMask & Mask_NotAllZeros)) {
527 RHS, LHS, IsAnd, A, D, B, C, PredR, PredL, Builder)) {
528 return V;
529 }
530 }
531 return nullptr;
532}
533
534/// Try to fold (icmp(A & B) ==/!= C) &/| (icmp(A & D) ==/!= E)
535/// into a single (icmp(A & X) ==/!= Y).
537 bool IsLogical,
539 const SimplifyQuery &Q) {
540 Value *A = nullptr, *B = nullptr, *C = nullptr, *D = nullptr, *E = nullptr;
541 ICmpInst::Predicate PredL, PredR;
542 std::optional<std::pair<unsigned, unsigned>> MaskPair =
543 getMaskedTypeForICmpPair(A, B, C, D, E, LHS, RHS, PredL, PredR);
544 if (!MaskPair)
545 return nullptr;
547 "Expected equality predicates for masked type of icmps.");
548 unsigned LHSMask = MaskPair->first;
549 unsigned RHSMask = MaskPair->second;
550 unsigned Mask = LHSMask & RHSMask;
551 if (Mask == 0) {
552 // Even if the two sides don't share a common pattern, check if folding can
553 // still happen.
555 LHS, RHS, IsAnd, A, B, C, D, E, PredL, PredR, LHSMask, RHSMask,
556 Builder))
557 return V;
558 return nullptr;
559 }
560
561 // In full generality:
562 // (icmp (A & B) Op C) | (icmp (A & D) Op E)
563 // == ![ (icmp (A & B) !Op C) & (icmp (A & D) !Op E) ]
564 //
565 // If the latter can be converted into (icmp (A & X) Op Y) then the former is
566 // equivalent to (icmp (A & X) !Op Y).
567 //
568 // Therefore, we can pretend for the rest of this function that we're dealing
569 // with the conjunction, provided we flip the sense of any comparisons (both
570 // input and output).
571
572 // In most cases we're going to produce an EQ for the "&&" case.
574 if (!IsAnd) {
575 // Convert the masking analysis into its equivalent with negated
576 // comparisons.
577 Mask = conjugateICmpMask(Mask);
578 }
579
580 if (Mask & Mask_AllZeros) {
581 // (icmp eq (A & B), 0) & (icmp eq (A & D), 0)
582 // -> (icmp eq (A & (B|D)), 0)
583 if (IsLogical && !isGuaranteedNotToBeUndefOrPoison(D))
584 return nullptr; // TODO: Use freeze?
585 Value *NewOr = Builder.CreateOr(B, D);
586 Value *NewAnd = Builder.CreateAnd(A, NewOr);
587 // We can't use C as zero because we might actually handle
588 // (icmp ne (A & B), B) & (icmp ne (A & D), D)
589 // with B and D, having a single bit set.
590 Value *Zero = Constant::getNullValue(A->getType());
591 return Builder.CreateICmp(NewCC, NewAnd, Zero);
592 }
593 if (Mask & BMask_AllOnes) {
594 // (icmp eq (A & B), B) & (icmp eq (A & D), D)
595 // -> (icmp eq (A & (B|D)), (B|D))
596 if (IsLogical && !isGuaranteedNotToBeUndefOrPoison(D))
597 return nullptr; // TODO: Use freeze?
598 Value *NewOr = Builder.CreateOr(B, D);
599 Value *NewAnd = Builder.CreateAnd(A, NewOr);
600 return Builder.CreateICmp(NewCC, NewAnd, NewOr);
601 }
602 if (Mask & AMask_AllOnes) {
603 // (icmp eq (A & B), A) & (icmp eq (A & D), A)
604 // -> (icmp eq (A & (B&D)), A)
605 if (IsLogical && !isGuaranteedNotToBeUndefOrPoison(D))
606 return nullptr; // TODO: Use freeze?
607 Value *NewAnd1 = Builder.CreateAnd(B, D);
608 Value *NewAnd2 = Builder.CreateAnd(A, NewAnd1);
609 return Builder.CreateICmp(NewCC, NewAnd2, A);
610 }
611
612 const APInt *ConstB, *ConstD;
613 if (match(B, m_APInt(ConstB)) && match(D, m_APInt(ConstD))) {
614 if (Mask & (Mask_NotAllZeros | BMask_NotAllOnes)) {
615 // (icmp ne (A & B), 0) & (icmp ne (A & D), 0) and
616 // (icmp ne (A & B), B) & (icmp ne (A & D), D)
617 // -> (icmp ne (A & B), 0) or (icmp ne (A & D), 0)
618 // Only valid if one of the masks is a superset of the other (check "B&D"
619 // is the same as either B or D).
620 APInt NewMask = *ConstB & *ConstD;
621 if (NewMask == *ConstB)
622 return LHS;
623 if (NewMask == *ConstD) {
624 if (IsLogical) {
625 if (auto *RHSI = dyn_cast<Instruction>(RHS))
626 RHSI->dropPoisonGeneratingFlags();
627 }
628 return RHS;
629 }
630 }
631
632 if (Mask & AMask_NotAllOnes) {
633 // (icmp ne (A & B), B) & (icmp ne (A & D), D)
634 // -> (icmp ne (A & B), A) or (icmp ne (A & D), A)
635 // Only valid if one of the masks is a superset of the other (check "B|D"
636 // is the same as either B or D).
637 APInt NewMask = *ConstB | *ConstD;
638 if (NewMask == *ConstB)
639 return LHS;
640 if (NewMask == *ConstD)
641 return RHS;
642 }
643
644 if (Mask & (BMask_Mixed | BMask_NotMixed)) {
645 // Mixed:
646 // (icmp eq (A & B), C) & (icmp eq (A & D), E)
647 // We already know that B & C == C && D & E == E.
648 // If we can prove that (B & D) & (C ^ E) == 0, that is, the bits of
649 // C and E, which are shared by both the mask B and the mask D, don't
650 // contradict, then we can transform to
651 // -> (icmp eq (A & (B|D)), (C|E))
652 // Currently, we only handle the case of B, C, D, and E being constant.
653 // We can't simply use C and E because we might actually handle
654 // (icmp ne (A & B), B) & (icmp eq (A & D), D)
655 // with B and D, having a single bit set.
656
657 // NotMixed:
658 // (icmp ne (A & B), C) & (icmp ne (A & D), E)
659 // -> (icmp ne (A & (B & D)), (C & E))
660 // Check the intersection (B & D) for inequality.
661 // Assume that (B & D) == B || (B & D) == D, i.e B/D is a subset of D/B
662 // and (B & D) & (C ^ E) == 0, bits of C and E, which are shared by both
663 // the B and the D, don't contradict. Note that we can assume (~B & C) ==
664 // 0 && (~D & E) == 0, previous operation should delete these icmps if it
665 // hadn't been met.
666
667 const APInt *OldConstC, *OldConstE;
668 if (!match(C, m_APInt(OldConstC)) || !match(E, m_APInt(OldConstE)))
669 return nullptr;
670
671 auto FoldBMixed = [&](ICmpInst::Predicate CC, bool IsNot) -> Value * {
672 CC = IsNot ? CmpInst::getInversePredicate(CC) : CC;
673 const APInt ConstC = PredL != CC ? *ConstB ^ *OldConstC : *OldConstC;
674 const APInt ConstE = PredR != CC ? *ConstD ^ *OldConstE : *OldConstE;
675
676 if (((*ConstB & *ConstD) & (ConstC ^ ConstE)).getBoolValue())
677 return IsNot ? nullptr : ConstantInt::get(LHS->getType(), !IsAnd);
678
679 if (IsNot && !ConstB->isSubsetOf(*ConstD) &&
680 !ConstD->isSubsetOf(*ConstB))
681 return nullptr;
682
683 APInt BD, CE;
684 if (IsNot) {
685 BD = *ConstB & *ConstD;
686 CE = ConstC & ConstE;
687 } else {
688 BD = *ConstB | *ConstD;
689 CE = ConstC | ConstE;
690 }
691 Value *NewAnd = Builder.CreateAnd(A, BD);
692 Value *CEVal = ConstantInt::get(A->getType(), CE);
693 return Builder.CreateICmp(CC, NewAnd, CEVal);
694 };
695
696 if (Mask & BMask_Mixed)
697 return FoldBMixed(NewCC, false);
698 if (Mask & BMask_NotMixed) // can be else also
699 return FoldBMixed(NewCC, true);
700 }
701 }
702
703 // (icmp eq (A & B), 0) | (icmp eq (A & D), 0)
704 // -> (icmp ne (A & (B|D)), (B|D))
705 // (icmp ne (A & B), 0) & (icmp ne (A & D), 0)
706 // -> (icmp eq (A & (B|D)), (B|D))
707 // iff B and D is known to be a power of two
708 if (Mask & Mask_NotAllZeros &&
709 isKnownToBeAPowerOfTwo(B, /*OrZero=*/false, Q) &&
710 isKnownToBeAPowerOfTwo(D, /*OrZero=*/false, Q)) {
711 // If this is a logical and/or, then we must prevent propagation of a
712 // poison value from the RHS by inserting freeze.
713 if (IsLogical)
714 D = Builder.CreateFreeze(D);
715 Value *Mask = Builder.CreateOr(B, D);
716 Value *Masked = Builder.CreateAnd(A, Mask);
717 return Builder.CreateICmp(NewCC, Masked, Mask);
718 }
719 return nullptr;
720}
721
722/// Try to fold a signed range checked with lower bound 0 to an unsigned icmp.
723/// Example: (icmp sge x, 0) & (icmp slt x, n) --> icmp ult x, n
724/// If \p Inverted is true then the check is for the inverted range, e.g.
725/// (icmp slt x, 0) | (icmp sgt x, n) --> icmp ugt x, n
727 bool Inverted) {
728 // Check the lower range comparison, e.g. x >= 0
729 // InstCombine already ensured that if there is a constant it's on the RHS.
730 ConstantInt *RangeStart = dyn_cast<ConstantInt>(Cmp0->getOperand(1));
731 if (!RangeStart)
732 return nullptr;
733
734 ICmpInst::Predicate Pred0 = (Inverted ? Cmp0->getInversePredicate() :
735 Cmp0->getPredicate());
736
737 // Accept x > -1 or x >= 0 (after potentially inverting the predicate).
738 if (!((Pred0 == ICmpInst::ICMP_SGT && RangeStart->isMinusOne()) ||
739 (Pred0 == ICmpInst::ICMP_SGE && RangeStart->isZero())))
740 return nullptr;
741
742 ICmpInst::Predicate Pred1 = (Inverted ? Cmp1->getInversePredicate() :
743 Cmp1->getPredicate());
744
745 Value *Input = Cmp0->getOperand(0);
746 Value *Cmp1Op0 = Cmp1->getOperand(0);
747 Value *Cmp1Op1 = Cmp1->getOperand(1);
748 Value *RangeEnd;
749 if (match(Cmp1Op0, m_SExtOrSelf(m_Specific(Input)))) {
750 // For the upper range compare we have: icmp x, n
751 Input = Cmp1Op0;
752 RangeEnd = Cmp1Op1;
753 } else if (match(Cmp1Op1, m_SExtOrSelf(m_Specific(Input)))) {
754 // For the upper range compare we have: icmp n, x
755 Input = Cmp1Op1;
756 RangeEnd = Cmp1Op0;
757 Pred1 = ICmpInst::getSwappedPredicate(Pred1);
758 } else {
759 return nullptr;
760 }
761
762 // Check the upper range comparison, e.g. x < n
763 ICmpInst::Predicate NewPred;
764 switch (Pred1) {
765 case ICmpInst::ICMP_SLT: NewPred = ICmpInst::ICMP_ULT; break;
766 case ICmpInst::ICMP_SLE: NewPred = ICmpInst::ICMP_ULE; break;
767 default: return nullptr;
768 }
769
770 // This simplification is only valid if the upper range is not negative.
771 KnownBits Known = computeKnownBits(RangeEnd, Cmp1);
772 if (!Known.isNonNegative())
773 return nullptr;
774
775 if (Inverted)
776 NewPred = ICmpInst::getInversePredicate(NewPred);
777
778 return Builder.CreateICmp(NewPred, Input, RangeEnd);
779}
780
781// (or (icmp eq X, 0), (icmp eq X, Pow2OrZero))
782// -> (icmp eq (and X, Pow2OrZero), X)
783// (and (icmp ne X, 0), (icmp ne X, Pow2OrZero))
784// -> (icmp ne (and X, Pow2OrZero), X)
785static Value *
787 ICmpInst *LHS, ICmpInst *RHS, bool IsAnd,
788 const SimplifyQuery &Q) {
790 // Make sure we have right compares for our op.
791 if (LHS->getPredicate() != Pred || RHS->getPredicate() != Pred)
792 return nullptr;
793
794 // Make it so we can match LHS against the (icmp eq/ne X, 0) just for
795 // simplicity.
796 if (match(RHS->getOperand(1), m_Zero()))
797 std::swap(LHS, RHS);
798
799 Value *Pow2, *Op;
800 // Match the desired pattern:
801 // LHS: (icmp eq/ne X, 0)
802 // RHS: (icmp eq/ne X, Pow2OrZero)
803 // Skip if Pow2OrZero is 1. Either way it gets folded to (icmp ugt X, 1) but
804 // this form ends up slightly less canonical.
805 // We could potentially be more sophisticated than requiring LHS/RHS
806 // be one-use. We don't create additional instructions if only one
807 // of them is one-use. So cases where one is one-use and the other
808 // is two-use might be profitable.
809 if (!match(LHS, m_OneUse(m_ICmp(Pred, m_Value(Op), m_Zero()))) ||
810 !match(RHS, m_OneUse(m_c_ICmp(Pred, m_Specific(Op), m_Value(Pow2)))) ||
811 match(Pow2, m_One()) ||
812 !isKnownToBeAPowerOfTwo(Pow2, Q.DL, /*OrZero=*/true, Q.AC, Q.CxtI, Q.DT))
813 return nullptr;
814
815 Value *And = Builder.CreateAnd(Op, Pow2);
816 return Builder.CreateICmp(Pred, And, Op);
817}
818
819/// General pattern:
820/// X & Y
821///
822/// Where Y is checking that all the high bits (covered by a mask 4294967168)
823/// are uniform, i.e. %arg & 4294967168 can be either 4294967168 or 0
824/// Pattern can be one of:
825/// %t = add i32 %arg, 128
826/// %r = icmp ult i32 %t, 256
827/// Or
828/// %t0 = shl i32 %arg, 24
829/// %t1 = ashr i32 %t0, 24
830/// %r = icmp eq i32 %t1, %arg
831/// Or
832/// %t0 = trunc i32 %arg to i8
833/// %t1 = sext i8 %t0 to i32
834/// %r = icmp eq i32 %t1, %arg
835/// This pattern is a signed truncation check.
836///
837/// And X is checking that some bit in that same mask is zero.
838/// I.e. can be one of:
839/// %r = icmp sgt i32 %arg, -1
840/// Or
841/// %t = and i32 %arg, 2147483648
842/// %r = icmp eq i32 %t, 0
843///
844/// Since we are checking that all the bits in that mask are the same,
845/// and a particular bit is zero, what we are really checking is that all the
846/// masked bits are zero.
847/// So this should be transformed to:
848/// %r = icmp ult i32 %arg, 128
850 Instruction &CxtI,
851 InstCombiner::BuilderTy &Builder) {
852 assert(CxtI.getOpcode() == Instruction::And);
853
854 // Match icmp ult (add %arg, C01), C1 (C1 == C01 << 1; powers of two)
855 auto tryToMatchSignedTruncationCheck = [](ICmpInst *ICmp, Value *&X,
856 APInt &SignBitMask) -> bool {
857 const APInt *I01, *I1; // powers of two; I1 == I01 << 1
859 m_Add(m_Value(X), m_Power2(I01)),
860 m_Power2(I1))) &&
861 I1->ugt(*I01) && I01->shl(1) == *I1))
862 return false;
863 // Which bit is the new sign bit as per the 'signed truncation' pattern?
864 SignBitMask = *I01;
865 return true;
866 };
867
868 // One icmp needs to be 'signed truncation check'.
869 // We need to match this first, else we will mismatch commutative cases.
870 Value *X1;
871 APInt HighestBit;
872 ICmpInst *OtherICmp;
873 if (tryToMatchSignedTruncationCheck(ICmp1, X1, HighestBit))
874 OtherICmp = ICmp0;
875 else if (tryToMatchSignedTruncationCheck(ICmp0, X1, HighestBit))
876 OtherICmp = ICmp1;
877 else
878 return nullptr;
879
880 assert(HighestBit.isPowerOf2() && "expected to be power of two (non-zero)");
881
882 // Try to match/decompose into: icmp eq (X & Mask), 0
883 auto tryToDecompose = [](ICmpInst *ICmp, Value *&X,
884 APInt &UnsetBitsMask) -> bool {
885 CmpPredicate Pred = ICmp->getPredicate();
886 // Can it be decomposed into icmp eq (X & Mask), 0 ?
888 ICmp->getOperand(0), ICmp->getOperand(1), Pred,
889 /*LookThroughTrunc=*/false, /*AllowNonZeroC=*/false,
890 /*DecomposeAnd=*/true);
891 if (Res && Res->Pred == ICmpInst::ICMP_EQ) {
892 X = Res->X;
893 UnsetBitsMask = Res->Mask;
894 return true;
895 }
896
897 return false;
898 };
899
900 // And the other icmp needs to be decomposable into a bit test.
901 Value *X0;
902 APInt UnsetBitsMask;
903 if (!tryToDecompose(OtherICmp, X0, UnsetBitsMask))
904 return nullptr;
905
906 assert(!UnsetBitsMask.isZero() && "empty mask makes no sense.");
907
908 // Are they working on the same value?
909 Value *X;
910 if (X1 == X0) {
911 // Ok as is.
912 X = X1;
913 } else if (match(X0, m_Trunc(m_Specific(X1)))) {
914 UnsetBitsMask = UnsetBitsMask.zext(X1->getType()->getScalarSizeInBits());
915 X = X1;
916 } else
917 return nullptr;
918
919 // So which bits should be uniform as per the 'signed truncation check'?
920 // (all the bits starting with (i.e. including) HighestBit)
921 APInt SignBitsMask = ~(HighestBit - 1U);
922
923 // UnsetBitsMask must have some common bits with SignBitsMask,
924 if (!UnsetBitsMask.intersects(SignBitsMask))
925 return nullptr;
926
927 // Does UnsetBitsMask contain any bits outside of SignBitsMask?
928 if (!UnsetBitsMask.isSubsetOf(SignBitsMask)) {
929 APInt OtherHighestBit = (~UnsetBitsMask) + 1U;
930 if (!OtherHighestBit.isPowerOf2())
931 return nullptr;
932 HighestBit = APIntOps::umin(HighestBit, OtherHighestBit);
933 }
934 // Else, if it does not, then all is ok as-is.
935
936 // %r = icmp ult %X, SignBit
937 return Builder.CreateICmpULT(X, ConstantInt::get(X->getType(), HighestBit),
938 CxtI.getName() + ".simplified");
939}
940
941/// Fold (icmp eq ctpop(X) 1) | (icmp eq X 0) into (icmp ult ctpop(X) 2) and
942/// fold (icmp ne ctpop(X) 1) & (icmp ne X 0) into (icmp ugt ctpop(X) 1).
943/// Also used for logical and/or, must be poison safe if range attributes are
944/// dropped.
945static Value *foldIsPowerOf2OrZero(ICmpInst *Cmp0, ICmpInst *Cmp1, bool IsAnd,
947 InstCombinerImpl &IC) {
948 CmpPredicate Pred0, Pred1;
949 Value *X;
951 m_SpecificInt(1))) ||
952 !match(Cmp1, m_ICmp(Pred1, m_Specific(X), m_ZeroInt())))
953 return nullptr;
954
955 auto *CtPop = cast<Instruction>(Cmp0->getOperand(0));
956 if (IsAnd && Pred0 == ICmpInst::ICMP_NE && Pred1 == ICmpInst::ICMP_NE) {
957 // Drop range attributes and re-infer them in the next iteration.
958 CtPop->dropPoisonGeneratingAnnotations();
959 IC.addToWorklist(CtPop);
960 return Builder.CreateICmpUGT(CtPop, ConstantInt::get(CtPop->getType(), 1));
961 }
962 if (!IsAnd && Pred0 == ICmpInst::ICMP_EQ && Pred1 == ICmpInst::ICMP_EQ) {
963 // Drop range attributes and re-infer them in the next iteration.
964 CtPop->dropPoisonGeneratingAnnotations();
965 IC.addToWorklist(CtPop);
966 return Builder.CreateICmpULT(CtPop, ConstantInt::get(CtPop->getType(), 2));
967 }
968
969 return nullptr;
970}
971
972/// Reduce a pair of compares that check if a value has exactly 1 bit set.
973/// Also used for logical and/or, must be poison safe if range attributes are
974/// dropped.
975static Value *foldIsPowerOf2(ICmpInst *Cmp0, ICmpInst *Cmp1, bool JoinedByAnd,
977 InstCombinerImpl &IC) {
978 // Handle 'and' / 'or' commutation: make the equality check the first operand.
979 if (JoinedByAnd && Cmp1->getPredicate() == ICmpInst::ICMP_NE)
980 std::swap(Cmp0, Cmp1);
981 else if (!JoinedByAnd && Cmp1->getPredicate() == ICmpInst::ICMP_EQ)
982 std::swap(Cmp0, Cmp1);
983
984 // (X != 0) && (ctpop(X) u< 2) --> ctpop(X) == 1
985 Value *X;
986 if (JoinedByAnd &&
990 m_SpecificInt(2)))) {
991 auto *CtPop = cast<Instruction>(Cmp1->getOperand(0));
992 // Drop range attributes and re-infer them in the next iteration.
993 CtPop->dropPoisonGeneratingAnnotations();
994 IC.addToWorklist(CtPop);
995 return Builder.CreateICmpEQ(CtPop, ConstantInt::get(CtPop->getType(), 1));
996 }
997 // (X == 0) || (ctpop(X) u> 1) --> ctpop(X) != 1
998 if (!JoinedByAnd &&
1002 m_SpecificInt(1)))) {
1003 auto *CtPop = cast<Instruction>(Cmp1->getOperand(0));
1004 // Drop range attributes and re-infer them in the next iteration.
1005 CtPop->dropPoisonGeneratingAnnotations();
1006 IC.addToWorklist(CtPop);
1007 return Builder.CreateICmpNE(CtPop, ConstantInt::get(CtPop->getType(), 1));
1008 }
1009 return nullptr;
1010}
1011
1012/// Try to fold (icmp(A & B) == 0) & (icmp(A & D) != E) into (icmp A u< D) iff
1013/// B is a contiguous set of ones starting from the most significant bit
1014/// (negative power of 2), D and E are equal, and D is a contiguous set of ones
1015/// starting at the most significant zero bit in B. Parameter B supports masking
1016/// using undef/poison in either scalar or vector values.
1018 Value *A, Value *B, Value *D, Value *E, ICmpInst::Predicate PredL,
1021 "Expected equality predicates for masked type of icmps.");
1022 if (PredL != ICmpInst::ICMP_EQ || PredR != ICmpInst::ICMP_NE)
1023 return nullptr;
1024
1025 if (!match(B, m_NegatedPower2()) || !match(D, m_ShiftedMask()) ||
1026 !match(E, m_ShiftedMask()))
1027 return nullptr;
1028
1029 // Test scalar arguments for conversion. B has been validated earlier to be a
1030 // negative power of two and thus is guaranteed to have one or more contiguous
1031 // ones starting from the MSB followed by zero or more contiguous zeros. D has
1032 // been validated earlier to be a shifted set of one or more contiguous ones.
1033 // In order to match, B leading ones and D leading zeros should be equal. The
1034 // predicate that B be a negative power of 2 prevents the condition of there
1035 // ever being zero leading ones. Thus 0 == 0 cannot occur. The predicate that
1036 // D always be a shifted mask prevents the condition of D equaling 0. This
1037 // prevents matching the condition where B contains the maximum number of
1038 // leading one bits (-1) and D contains the maximum number of leading zero
1039 // bits (0).
1040 auto isReducible = [](const Value *B, const Value *D, const Value *E) {
1041 const APInt *BCst, *DCst, *ECst;
1042 return match(B, m_APIntAllowPoison(BCst)) && match(D, m_APInt(DCst)) &&
1043 match(E, m_APInt(ECst)) && *DCst == *ECst &&
1044 (isa<PoisonValue>(B) ||
1045 (BCst->countLeadingOnes() == DCst->countLeadingZeros()));
1046 };
1047
1048 // Test vector type arguments for conversion.
1049 if (const auto *BVTy = dyn_cast<VectorType>(B->getType())) {
1050 const auto *BFVTy = dyn_cast<FixedVectorType>(BVTy);
1051 const auto *BConst = dyn_cast<Constant>(B);
1052 const auto *DConst = dyn_cast<Constant>(D);
1053 const auto *EConst = dyn_cast<Constant>(E);
1054
1055 if (!BFVTy || !BConst || !DConst || !EConst)
1056 return nullptr;
1057
1058 for (unsigned I = 0; I != BFVTy->getNumElements(); ++I) {
1059 const auto *BElt = BConst->getAggregateElement(I);
1060 const auto *DElt = DConst->getAggregateElement(I);
1061 const auto *EElt = EConst->getAggregateElement(I);
1062
1063 if (!BElt || !DElt || !EElt)
1064 return nullptr;
1065 if (!isReducible(BElt, DElt, EElt))
1066 return nullptr;
1067 }
1068 } else {
1069 // Test scalar type arguments for conversion.
1070 if (!isReducible(B, D, E))
1071 return nullptr;
1072 }
1073 return Builder.CreateICmp(ICmpInst::ICMP_ULT, A, D);
1074}
1075
1076/// Try to fold ((icmp X u< P) & (icmp(X & M) != M)) or ((icmp X s> -1) &
1077/// (icmp(X & M) != M)) into (icmp X u< M). Where P is a power of 2, M < P, and
1078/// M is a contiguous shifted mask starting at the right most significant zero
1079/// bit in P. SGT is supported as when P is the largest representable power of
1080/// 2, an earlier optimization converts the expression into (icmp X s> -1).
1081/// Parameter P supports masking using undef/poison in either scalar or vector
1082/// values.
1084 bool JoinedByAnd,
1085 InstCombiner::BuilderTy &Builder) {
1086 if (!JoinedByAnd)
1087 return nullptr;
1088 Value *A = nullptr, *B = nullptr, *C = nullptr, *D = nullptr, *E = nullptr;
1089 ICmpInst::Predicate CmpPred0, CmpPred1;
1090 // Assuming P is a 2^n, getMaskedTypeForICmpPair will normalize (icmp X u<
1091 // 2^n) into (icmp (X & ~(2^n-1)) == 0) and (icmp X s> -1) into (icmp (X &
1092 // SignMask) == 0).
1093 std::optional<std::pair<unsigned, unsigned>> MaskPair =
1094 getMaskedTypeForICmpPair(A, B, C, D, E, Cmp0, Cmp1, CmpPred0, CmpPred1);
1095 if (!MaskPair)
1096 return nullptr;
1097
1098 const auto compareBMask = BMask_NotMixed | BMask_NotAllOnes;
1099 unsigned CmpMask0 = MaskPair->first;
1100 unsigned CmpMask1 = MaskPair->second;
1101 if ((CmpMask0 & Mask_AllZeros) && (CmpMask1 == compareBMask)) {
1102 if (Value *V = foldNegativePower2AndShiftedMask(A, B, D, E, CmpPred0,
1103 CmpPred1, Builder))
1104 return V;
1105 } else if ((CmpMask0 == compareBMask) && (CmpMask1 & Mask_AllZeros)) {
1106 if (Value *V = foldNegativePower2AndShiftedMask(A, D, B, C, CmpPred1,
1107 CmpPred0, Builder))
1108 return V;
1109 }
1110 return nullptr;
1111}
1112
1113/// Commuted variants are assumed to be handled by calling this function again
1114/// with the parameters swapped.
1116 ICmpInst *UnsignedICmp, bool IsAnd,
1117 const SimplifyQuery &Q,
1118 InstCombiner::BuilderTy &Builder) {
1119 Value *ZeroCmpOp;
1120 CmpPredicate EqPred;
1121 if (!match(ZeroICmp, m_ICmp(EqPred, m_Value(ZeroCmpOp), m_Zero())) ||
1122 !ICmpInst::isEquality(EqPred))
1123 return nullptr;
1124
1125 CmpPredicate UnsignedPred;
1126
1127 Value *A, *B;
1128 if (match(UnsignedICmp,
1129 m_c_ICmp(UnsignedPred, m_Specific(ZeroCmpOp), m_Value(A))) &&
1130 match(ZeroCmpOp, m_c_Add(m_Specific(A), m_Value(B))) &&
1131 (ZeroICmp->hasOneUse() || UnsignedICmp->hasOneUse())) {
1132 auto GetKnownNonZeroAndOther = [&](Value *&NonZero, Value *&Other) {
1133 if (!isKnownNonZero(NonZero, Q))
1134 std::swap(NonZero, Other);
1135 return isKnownNonZero(NonZero, Q);
1136 };
1137
1138 // Given ZeroCmpOp = (A + B)
1139 // ZeroCmpOp < A && ZeroCmpOp != 0 --> (0-X) < Y iff
1140 // ZeroCmpOp >= A || ZeroCmpOp == 0 --> (0-X) >= Y iff
1141 // with X being the value (A/B) that is known to be non-zero,
1142 // and Y being remaining value.
1143 if (UnsignedPred == ICmpInst::ICMP_ULT && EqPred == ICmpInst::ICMP_NE &&
1144 IsAnd && GetKnownNonZeroAndOther(B, A))
1145 return Builder.CreateICmpULT(Builder.CreateNeg(B), A);
1146 if (UnsignedPred == ICmpInst::ICMP_UGE && EqPred == ICmpInst::ICMP_EQ &&
1147 !IsAnd && GetKnownNonZeroAndOther(B, A))
1148 return Builder.CreateICmpUGE(Builder.CreateNeg(B), A);
1149 }
1150
1151 return nullptr;
1152}
1153
1154struct IntPart {
1156 unsigned StartBit;
1157 unsigned NumBits;
1158};
1159
1160/// Match an extraction of bits from an integer.
1161static std::optional<IntPart> matchIntPart(Value *V) {
1162 Value *X;
1163 if (!match(V, m_OneUse(m_Trunc(m_Value(X)))))
1164 return std::nullopt;
1165
1166 unsigned NumOriginalBits = X->getType()->getScalarSizeInBits();
1167 unsigned NumExtractedBits = V->getType()->getScalarSizeInBits();
1168 Value *Y;
1169 const APInt *Shift;
1170 // For a trunc(lshr Y, Shift) pattern, make sure we're only extracting bits
1171 // from Y, not any shifted-in zeroes.
1172 if (match(X, m_OneUse(m_LShr(m_Value(Y), m_APInt(Shift)))) &&
1173 Shift->ule(NumOriginalBits - NumExtractedBits))
1174 return {{Y, (unsigned)Shift->getZExtValue(), NumExtractedBits}};
1175 return {{X, 0, NumExtractedBits}};
1176}
1177
1178/// Materialize an extraction of bits from an integer in IR.
1179static Value *extractIntPart(const IntPart &P, IRBuilderBase &Builder) {
1180 Value *V = P.From;
1181 if (P.StartBit)
1182 V = Builder.CreateLShr(V, P.StartBit);
1183 Type *TruncTy = V->getType()->getWithNewBitWidth(P.NumBits);
1184 if (TruncTy != V->getType())
1185 V = Builder.CreateTrunc(V, TruncTy);
1186 return V;
1187}
1188
1189/// (icmp eq X0, Y0) & (icmp eq X1, Y1) -> icmp eq X01, Y01
1190/// (icmp ne X0, Y0) | (icmp ne X1, Y1) -> icmp ne X01, Y01
1191/// where X0, X1 and Y0, Y1 are adjacent parts extracted from an integer.
1192Value *InstCombinerImpl::foldEqOfParts(Value *Cmp0, Value *Cmp1, bool IsAnd) {
1193 if (!Cmp0->hasOneUse() || !Cmp1->hasOneUse())
1194 return nullptr;
1195
1197 auto GetMatchPart = [&](Value *CmpV,
1198 unsigned OpNo) -> std::optional<IntPart> {
1199 assert(CmpV->getType()->isIntOrIntVectorTy(1) && "Must be bool");
1200
1201 Value *X, *Y;
1202 // icmp ne (and x, 1), (and y, 1) <=> trunc (xor x, y) to i1
1203 // icmp eq (and x, 1), (and y, 1) <=> not (trunc (xor x, y) to i1)
1204 if (Pred == CmpInst::ICMP_NE
1205 ? match(CmpV, m_Trunc(m_Xor(m_Value(X), m_Value(Y))))
1206 : match(CmpV, m_Not(m_Trunc(m_Xor(m_Value(X), m_Value(Y))))))
1207 return {{OpNo == 0 ? X : Y, 0, 1}};
1208
1209 auto *Cmp = dyn_cast<ICmpInst>(CmpV);
1210 if (!Cmp)
1211 return std::nullopt;
1212
1213 if (Pred == Cmp->getPredicate())
1214 return matchIntPart(Cmp->getOperand(OpNo));
1215
1216 const APInt *C;
1217 // (icmp eq (lshr x, C), (lshr y, C)) gets optimized to:
1218 // (icmp ult (xor x, y), 1 << C) so also look for that.
1219 if (Pred == CmpInst::ICMP_EQ && Cmp->getPredicate() == CmpInst::ICMP_ULT) {
1220 if (!match(Cmp->getOperand(1), m_Power2(C)) ||
1221 !match(Cmp->getOperand(0), m_Xor(m_Value(), m_Value())))
1222 return std::nullopt;
1223 }
1224
1225 // (icmp ne (lshr x, C), (lshr y, C)) gets optimized to:
1226 // (icmp ugt (xor x, y), (1 << C) - 1) so also look for that.
1227 else if (Pred == CmpInst::ICMP_NE &&
1228 Cmp->getPredicate() == CmpInst::ICMP_UGT) {
1229 if (!match(Cmp->getOperand(1), m_LowBitMask(C)) ||
1230 !match(Cmp->getOperand(0), m_Xor(m_Value(), m_Value())))
1231 return std::nullopt;
1232 } else {
1233 return std::nullopt;
1234 }
1235
1236 unsigned From = Pred == CmpInst::ICMP_NE ? C->popcount() : C->countr_zero();
1237 Instruction *I = cast<Instruction>(Cmp->getOperand(0));
1238 return {{I->getOperand(OpNo), From, C->getBitWidth() - From}};
1239 };
1240
1241 std::optional<IntPart> L0 = GetMatchPart(Cmp0, 0);
1242 std::optional<IntPart> R0 = GetMatchPart(Cmp0, 1);
1243 std::optional<IntPart> L1 = GetMatchPart(Cmp1, 0);
1244 std::optional<IntPart> R1 = GetMatchPart(Cmp1, 1);
1245 if (!L0 || !R0 || !L1 || !R1)
1246 return nullptr;
1247
1248 // Make sure the LHS/RHS compare a part of the same value, possibly after
1249 // an operand swap.
1250 if (L0->From != L1->From || R0->From != R1->From) {
1251 if (L0->From != R1->From || R0->From != L1->From)
1252 return nullptr;
1253 std::swap(L1, R1);
1254 }
1255
1256 // Make sure the extracted parts are adjacent, canonicalizing to L0/R0 being
1257 // the low part and L1/R1 being the high part.
1258 if (L0->StartBit + L0->NumBits != L1->StartBit ||
1259 R0->StartBit + R0->NumBits != R1->StartBit) {
1260 if (L1->StartBit + L1->NumBits != L0->StartBit ||
1261 R1->StartBit + R1->NumBits != R0->StartBit)
1262 return nullptr;
1263 std::swap(L0, L1);
1264 std::swap(R0, R1);
1265 }
1266
1267 // We can simplify to a comparison of these larger parts of the integers.
1268 IntPart L = {L0->From, L0->StartBit, L0->NumBits + L1->NumBits};
1269 IntPart R = {R0->From, R0->StartBit, R0->NumBits + R1->NumBits};
1272 return Builder.CreateICmp(Pred, LValue, RValue);
1273}
1274
1275/// Reduce logic-of-compares with equality to a constant by substituting a
1276/// common operand with the constant. Callers are expected to call this with
1277/// Cmp0/Cmp1 switched to handle logic op commutativity.
1279 bool IsAnd, bool IsLogical,
1280 InstCombiner::BuilderTy &Builder,
1281 const SimplifyQuery &Q,
1282 Instruction &I) {
1283 // Match an equality compare with a non-poison constant as Cmp0.
1284 // Also, give up if the compare can be constant-folded to avoid looping.
1285 CmpPredicate Pred0;
1286 Value *X;
1287 Constant *C;
1288 if (!match(Cmp0, m_ICmp(Pred0, m_Value(X), m_Constant(C))) ||
1290 return nullptr;
1291 if ((IsAnd && Pred0 != ICmpInst::ICMP_EQ) ||
1292 (!IsAnd && Pred0 != ICmpInst::ICMP_NE))
1293 return nullptr;
1294
1295 // The other compare must include a common operand (X). Canonicalize the
1296 // common operand as operand 1 (Pred1 is swapped if the common operand was
1297 // operand 0).
1298 Value *Y;
1299 CmpPredicate Pred1;
1300 if (!match(Cmp1, m_c_ICmp(Pred1, m_Value(Y), m_Specific(X))))
1301 return nullptr;
1302
1303 // Replace variable with constant value equivalence to remove a variable use:
1304 // (X == C) && (Y Pred1 X) --> (X == C) && (Y Pred1 C)
1305 // (X != C) || (Y Pred1 X) --> (X != C) || (Y Pred1 C)
1306 // Can think of the 'or' substitution with the 'and' bool equivalent:
1307 // A || B --> A || (!A && B)
1308 Value *SubstituteCmp = simplifyICmpInst(Pred1, Y, C, Q);
1309 if (!SubstituteCmp) {
1310 // If we need to create a new instruction, require that the old compare can
1311 // be removed.
1312 if (!Cmp1->hasOneUse())
1313 return nullptr;
1314 SubstituteCmp = Builder.CreateICmp(Pred1, Y, C);
1315 }
1316 if (IsLogical) {
1317 Instruction *MDFrom =
1319 return IsAnd ? Builder.CreateLogicalAnd(Cmp0, SubstituteCmp, "", MDFrom)
1320 : Builder.CreateLogicalOr(Cmp0, SubstituteCmp, "", MDFrom);
1321 }
1322 return Builder.CreateBinOp(IsAnd ? Instruction::And : Instruction::Or, Cmp0,
1323 SubstituteCmp);
1324}
1325
1326/// Fold (icmp Pred1 V1, C1) & (icmp Pred2 V2, C2)
1327/// or (icmp Pred1 V1, C1) | (icmp Pred2 V2, C2)
1328/// into a single comparison using range-based reasoning.
1329/// NOTE: This is also used for logical and/or, must be poison-safe!
1330Value *InstCombinerImpl::foldAndOrOfICmpsUsingRanges(ICmpInst *ICmp1,
1331 ICmpInst *ICmp2,
1332 bool IsAnd) {
1333 // Return (V, CR) for a range check idiom V in CR.
1334 auto MatchExactRangeCheck =
1335 [](ICmpInst *ICmp) -> std::optional<std::pair<Value *, ConstantRange>> {
1336 const APInt *C;
1337 if (!match(ICmp->getOperand(1), m_APInt(C)))
1338 return std::nullopt;
1339 Value *LHS = ICmp->getOperand(0);
1340 CmpPredicate Pred = ICmp->getPredicate();
1341 Value *X;
1342 // Match (x & NegPow2) ==/!= C
1343 const APInt *Mask;
1344 if (ICmpInst::isEquality(Pred) &&
1346 C->countr_zero() >= Mask->countr_zero()) {
1347 ConstantRange CR(*C, *C - *Mask);
1348 if (Pred == ICmpInst::ICMP_NE)
1349 CR = CR.inverse();
1350 return std::make_pair(X, CR);
1351 }
1352 ConstantRange CR = ConstantRange::makeExactICmpRegion(Pred, *C);
1353 // Match (add X, C1) pred C
1354 // TODO: investigate whether we should apply the one-use check on m_AddLike.
1355 const APInt *C1;
1356 if (match(LHS, m_AddLike(m_Value(X), m_APInt(C1))))
1357 return std::make_pair(X, CR.subtract(*C1));
1358 return std::make_pair(LHS, CR);
1359 };
1360
1361 auto RC1 = MatchExactRangeCheck(ICmp1);
1362 if (!RC1)
1363 return nullptr;
1364
1365 auto RC2 = MatchExactRangeCheck(ICmp2);
1366 if (!RC2)
1367 return nullptr;
1368
1369 auto &[V1, CR1] = *RC1;
1370 auto &[V2, CR2] = *RC2;
1371 if (V1 != V2)
1372 return nullptr;
1373
1374 // For 'and', we use the De Morgan's Laws to simplify the implementation.
1375 if (IsAnd) {
1376 CR1 = CR1.inverse();
1377 CR2 = CR2.inverse();
1378 }
1379
1380 Type *Ty = V1->getType();
1381 Value *NewV = V1;
1382 std::optional<ConstantRange> CR = CR1.exactUnionWith(CR2);
1383 if (!CR) {
1384 if (!(ICmp1->hasOneUse() && ICmp2->hasOneUse()) || CR1.isWrappedSet() ||
1385 CR2.isWrappedSet())
1386 return nullptr;
1387
1388 // Check whether we have equal-size ranges that only differ by one bit.
1389 // In that case we can apply a mask to map one range onto the other.
1390 APInt LowerDiff = CR1.getLower() ^ CR2.getLower();
1391 APInt UpperDiff = (CR1.getUpper() - 1) ^ (CR2.getUpper() - 1);
1392 APInt CR1Size = CR1.getUpper() - CR1.getLower();
1393 if (!LowerDiff.isPowerOf2() || LowerDiff != UpperDiff ||
1394 CR1Size != CR2.getUpper() - CR2.getLower())
1395 return nullptr;
1396
1397 CR = CR1.getLower().ult(CR2.getLower()) ? CR1 : CR2;
1398 NewV = Builder.CreateAnd(NewV, ConstantInt::get(Ty, ~LowerDiff));
1399 }
1400
1401 if (IsAnd)
1402 CR = CR->inverse();
1403
1404 CmpInst::Predicate NewPred;
1405 APInt NewC, Offset;
1406 CR->getEquivalentICmp(NewPred, NewC, Offset);
1407
1408 if (Offset != 0)
1409 NewV = Builder.CreateAdd(NewV, ConstantInt::get(Ty, Offset));
1410 return Builder.CreateICmp(NewPred, NewV, ConstantInt::get(Ty, NewC));
1411}
1412
1413/// Matches canonical form of isnan, fcmp ord x, 0
1417
1418/// Matches fcmp u__ x, +/-inf
1423
1424/// and (fcmp ord x, 0), (fcmp u* x, inf) -> fcmp o* x, inf
1425///
1426/// Clang emits this pattern for doing an isfinite check in __builtin_isnormal.
1428 FCmpInst *RHS) {
1429 Value *LHS0 = LHS->getOperand(0), *LHS1 = LHS->getOperand(1);
1430 Value *RHS0 = RHS->getOperand(0), *RHS1 = RHS->getOperand(1);
1431 FCmpInst::Predicate PredL = LHS->getPredicate(), PredR = RHS->getPredicate();
1432
1433 if (!matchIsNotNaN(PredL, LHS0, LHS1) ||
1434 !matchUnorderedInfCompare(PredR, RHS0, RHS1))
1435 return nullptr;
1436
1437 return Builder.CreateFCmpFMF(FCmpInst::getOrderedPredicate(PredR), RHS0, RHS1,
1439}
1440
1441Value *InstCombinerImpl::foldLogicOfFCmps(FCmpInst *LHS, FCmpInst *RHS,
1442 bool IsAnd, bool IsLogicalSelect) {
1443 Value *LHS0 = LHS->getOperand(0), *LHS1 = LHS->getOperand(1);
1444 Value *RHS0 = RHS->getOperand(0), *RHS1 = RHS->getOperand(1);
1445 FCmpInst::Predicate PredL = LHS->getPredicate(), PredR = RHS->getPredicate();
1446
1447 if (LHS0 == RHS1 && RHS0 == LHS1) {
1448 // Swap RHS operands to match LHS.
1449 PredR = FCmpInst::getSwappedPredicate(PredR);
1450 std::swap(RHS0, RHS1);
1451 }
1452
1453 // Simplify (fcmp cc0 x, y) & (fcmp cc1 x, y).
1454 // Suppose the relation between x and y is R, where R is one of
1455 // U(1000), L(0100), G(0010) or E(0001), and CC0 and CC1 are the bitmasks for
1456 // testing the desired relations.
1457 //
1458 // Since (R & CC0) and (R & CC1) are either R or 0, we actually have this:
1459 // bool(R & CC0) && bool(R & CC1)
1460 // = bool((R & CC0) & (R & CC1))
1461 // = bool(R & (CC0 & CC1)) <= by re-association, commutation, and idempotency
1462 //
1463 // Since (R & CC0) and (R & CC1) are either R or 0, we actually have this:
1464 // bool(R & CC0) || bool(R & CC1)
1465 // = bool((R & CC0) | (R & CC1))
1466 // = bool(R & (CC0 | CC1)) <= by reversed distribution (contribution? ;)
1467 if (LHS0 == RHS0 && LHS1 == RHS1) {
1468 unsigned FCmpCodeL = getFCmpCode(PredL);
1469 unsigned FCmpCodeR = getFCmpCode(PredR);
1470 unsigned NewPred = IsAnd ? FCmpCodeL & FCmpCodeR : FCmpCodeL | FCmpCodeR;
1471
1472 // Intersect the fast math flags.
1473 // TODO: We can union the fast math flags unless this is a logical select.
1474 return getFCmpValue(NewPred, LHS0, LHS1, Builder,
1476 }
1477
1478 // This transform is not valid for a logical select.
1479 if (!IsLogicalSelect &&
1480 ((PredL == FCmpInst::FCMP_ORD && PredR == FCmpInst::FCMP_ORD && IsAnd) ||
1481 (PredL == FCmpInst::FCMP_UNO && PredR == FCmpInst::FCMP_UNO &&
1482 !IsAnd))) {
1483 if (LHS0->getType() != RHS0->getType())
1484 return nullptr;
1485
1486 // FCmp canonicalization ensures that (fcmp ord/uno X, X) and
1487 // (fcmp ord/uno X, C) will be transformed to (fcmp X, +0.0).
1488 if (match(LHS1, m_PosZeroFP()) && match(RHS1, m_PosZeroFP())) {
1489 // Ignore the constants because they are obviously not NANs:
1490 // (fcmp ord x, 0.0) & (fcmp ord y, 0.0) -> (fcmp ord x, y)
1491 // (fcmp uno x, 0.0) | (fcmp uno y, 0.0) -> (fcmp uno x, y)
1492 return Builder.CreateFCmpFMF(PredL, LHS0, RHS0,
1494 }
1495 }
1496
1497 // This transform is not valid for a logical select.
1498 if (!IsLogicalSelect && IsAnd &&
1499 stripSignOnlyFPOps(LHS0) == stripSignOnlyFPOps(RHS0)) {
1500 // and (fcmp ord x, 0), (fcmp u* x, inf) -> fcmp o* x, inf
1501 // and (fcmp ord x, 0), (fcmp u* fabs(x), inf) -> fcmp o* x, inf
1503 return Left;
1505 return Right;
1506 }
1507
1508 // Turn at least two fcmps with constants into llvm.is.fpclass.
1509 //
1510 // If we can represent a combined value test with one class call, we can
1511 // potentially eliminate 4-6 instructions. If we can represent a test with a
1512 // single fcmp with fneg and fabs, that's likely a better canonical form.
1513 if (LHS->hasOneUse() && RHS->hasOneUse()) {
1514 auto [ClassValRHS, ClassMaskRHS] =
1515 fcmpToClassTest(PredR, *RHS->getFunction(), RHS0, RHS1);
1516 if (ClassValRHS) {
1517 auto [ClassValLHS, ClassMaskLHS] =
1518 fcmpToClassTest(PredL, *LHS->getFunction(), LHS0, LHS1);
1519 if (ClassValLHS == ClassValRHS) {
1520 unsigned CombinedMask = IsAnd ? (ClassMaskLHS & ClassMaskRHS)
1521 : (ClassMaskLHS | ClassMaskRHS);
1522 return Builder.CreateIntrinsic(
1523 Intrinsic::is_fpclass, {ClassValLHS->getType()},
1524 {ClassValLHS, Builder.getInt32(CombinedMask)});
1525 }
1526 }
1527 }
1528
1529 // Canonicalize the range check idiom:
1530 // and (fcmp olt/ole/ult/ule x, C), (fcmp ogt/oge/ugt/uge x, -C)
1531 // --> fabs(x) olt/ole/ult/ule C
1532 // or (fcmp ogt/oge/ugt/uge x, C), (fcmp olt/ole/ult/ule x, -C)
1533 // --> fabs(x) ogt/oge/ugt/uge C
1534 // TODO: Generalize to handle a negated variable operand?
1535 const APFloat *LHSC, *RHSC;
1536 if (LHS0 == RHS0 && LHS->hasOneUse() && RHS->hasOneUse() &&
1537 FCmpInst::getSwappedPredicate(PredL) == PredR &&
1538 match(LHS1, m_APFloatAllowPoison(LHSC)) &&
1539 match(RHS1, m_APFloatAllowPoison(RHSC)) &&
1540 LHSC->bitwiseIsEqual(neg(*RHSC))) {
1541 auto IsLessThanOrLessEqual = [](FCmpInst::Predicate Pred) {
1542 switch (Pred) {
1543 case FCmpInst::FCMP_OLT:
1544 case FCmpInst::FCMP_OLE:
1545 case FCmpInst::FCMP_ULT:
1546 case FCmpInst::FCMP_ULE:
1547 return true;
1548 default:
1549 return false;
1550 }
1551 };
1552 if (IsLessThanOrLessEqual(IsAnd ? PredR : PredL)) {
1553 std::swap(LHSC, RHSC);
1554 std::swap(PredL, PredR);
1555 }
1556 if (IsLessThanOrLessEqual(IsAnd ? PredL : PredR)) {
1557 FastMathFlags NewFlag = LHS->getFastMathFlags();
1558 if (!IsLogicalSelect)
1559 NewFlag |= RHS->getFastMathFlags();
1560
1561 Value *FAbs = Builder.CreateFAbs(LHS0, NewFlag);
1562 return Builder.CreateFCmpFMF(
1563 PredL, FAbs, ConstantFP::get(LHS0->getType(), *LHSC), NewFlag);
1564 }
1565 }
1566
1567 return nullptr;
1568}
1569
1570/// Match an fcmp against a special value that performs a test possible by
1571/// llvm.is.fpclass.
1572static bool matchIsFPClassLikeFCmp(Value *Op, Value *&ClassVal,
1573 uint64_t &ClassMask) {
1574 auto *FCmp = dyn_cast<FCmpInst>(Op);
1575 if (!FCmp || !FCmp->hasOneUse())
1576 return false;
1577
1578 std::tie(ClassVal, ClassMask) =
1579 fcmpToClassTest(FCmp->getPredicate(), *FCmp->getParent()->getParent(),
1580 FCmp->getOperand(0), FCmp->getOperand(1));
1581 return ClassVal != nullptr;
1582}
1583
1584/// or (is_fpclass x, mask0), (is_fpclass x, mask1)
1585/// -> is_fpclass x, (mask0 | mask1)
1586/// and (is_fpclass x, mask0), (is_fpclass x, mask1)
1587/// -> is_fpclass x, (mask0 & mask1)
1588/// xor (is_fpclass x, mask0), (is_fpclass x, mask1)
1589/// -> is_fpclass x, (mask0 ^ mask1)
1590Instruction *InstCombinerImpl::foldLogicOfIsFPClass(BinaryOperator &BO,
1591 Value *Op0, Value *Op1) {
1592 Value *ClassVal0 = nullptr;
1593 Value *ClassVal1 = nullptr;
1594 uint64_t ClassMask0, ClassMask1;
1595
1596 // Restrict to folding one fcmp into one is.fpclass for now, don't introduce a
1597 // new class.
1598 //
1599 // TODO: Support forming is.fpclass out of 2 separate fcmps when codegen is
1600 // better.
1601
1602 bool IsLHSClass =
1604 m_Value(ClassVal0), m_ConstantInt(ClassMask0))));
1605 bool IsRHSClass =
1607 m_Value(ClassVal1), m_ConstantInt(ClassMask1))));
1608 if ((((IsLHSClass || matchIsFPClassLikeFCmp(Op0, ClassVal0, ClassMask0)) &&
1609 (IsRHSClass || matchIsFPClassLikeFCmp(Op1, ClassVal1, ClassMask1)))) &&
1610 ClassVal0 == ClassVal1) {
1611 unsigned NewClassMask;
1612 switch (BO.getOpcode()) {
1613 case Instruction::And:
1614 NewClassMask = ClassMask0 & ClassMask1;
1615 break;
1616 case Instruction::Or:
1617 NewClassMask = ClassMask0 | ClassMask1;
1618 break;
1619 case Instruction::Xor:
1620 NewClassMask = ClassMask0 ^ ClassMask1;
1621 break;
1622 default:
1623 llvm_unreachable("not a binary logic operator");
1624 }
1625
1626 if (IsLHSClass) {
1627 auto *II = cast<IntrinsicInst>(Op0);
1628 II->setArgOperand(
1629 1, ConstantInt::get(II->getArgOperand(1)->getType(), NewClassMask));
1630 return replaceInstUsesWith(BO, II);
1631 }
1632
1633 if (IsRHSClass) {
1634 auto *II = cast<IntrinsicInst>(Op1);
1635 II->setArgOperand(
1636 1, ConstantInt::get(II->getArgOperand(1)->getType(), NewClassMask));
1637 return replaceInstUsesWith(BO, II);
1638 }
1639
1640 CallInst *NewClass =
1641 Builder.CreateIntrinsic(Intrinsic::is_fpclass, {ClassVal0->getType()},
1642 {ClassVal0, Builder.getInt32(NewClassMask)});
1643 return replaceInstUsesWith(BO, NewClass);
1644 }
1645
1646 return nullptr;
1647}
1648
1649/// Look for the pattern that conditionally negates a value via math operations:
1650/// cond.splat = sext i1 cond
1651/// sub = add cond.splat, x
1652/// xor = xor sub, cond.splat
1653/// and rewrite it to do the same, but via logical operations:
1654/// value.neg = sub 0, value
1655/// cond = select i1 neg, value.neg, value
1656Instruction *InstCombinerImpl::canonicalizeConditionalNegationViaMathToSelect(
1657 BinaryOperator &I) {
1658 assert(I.getOpcode() == BinaryOperator::Xor && "Only for xor!");
1659 Value *Cond, *X;
1660 // As per complexity ordering, `xor` is not commutative here.
1661 if (!match(&I, m_c_BinOp(m_OneUse(m_Value()), m_Value())) ||
1662 !match(I.getOperand(1), m_SExt(m_Value(Cond))) ||
1663 !Cond->getType()->isIntOrIntVectorTy(1) ||
1664 !match(I.getOperand(0), m_c_Add(m_SExt(m_Specific(Cond)), m_Value(X))))
1665 return nullptr;
1666 return createSelectInstWithUnknownProfile(
1667 Cond, Builder.CreateNeg(X, X->getName() + ".neg"), X);
1668}
1669
1670/// This a limited reassociation for a special case (see above) where we are
1671/// checking if two values are either both NAN (unordered) or not-NAN (ordered).
1672/// This could be handled more generally in '-reassociation', but it seems like
1673/// an unlikely pattern for a large number of logic ops and fcmps.
1675 InstCombiner::BuilderTy &Builder) {
1676 Instruction::BinaryOps Opcode = BO.getOpcode();
1677 assert((Opcode == Instruction::And || Opcode == Instruction::Or) &&
1678 "Expecting and/or op for fcmp transform");
1679
1680 // There are 4 commuted variants of the pattern. Canonicalize operands of this
1681 // logic op so an fcmp is operand 0 and a matching logic op is operand 1.
1682 Value *Op0 = BO.getOperand(0), *Op1 = BO.getOperand(1), *X;
1683 if (match(Op1, m_FCmp(m_Value(), m_AnyZeroFP())))
1684 std::swap(Op0, Op1);
1685
1686 // Match inner binop and the predicate for combining 2 NAN checks into 1.
1687 Value *BO10, *BO11;
1688 FCmpInst::Predicate NanPred = Opcode == Instruction::And ? FCmpInst::FCMP_ORD
1690 if (!match(Op0, m_SpecificFCmp(NanPred, m_Value(X), m_AnyZeroFP())) ||
1691 !match(Op1, m_BinOp(Opcode, m_Value(BO10), m_Value(BO11))))
1692 return nullptr;
1693
1694 // The inner logic op must have a matching fcmp operand.
1695 Value *Y;
1696 if (!match(BO10, m_SpecificFCmp(NanPred, m_Value(Y), m_AnyZeroFP())) ||
1697 X->getType() != Y->getType())
1698 std::swap(BO10, BO11);
1699
1700 if (!match(BO10, m_SpecificFCmp(NanPred, m_Value(Y), m_AnyZeroFP())) ||
1701 X->getType() != Y->getType())
1702 return nullptr;
1703
1704 // and (fcmp ord X, 0), (and (fcmp ord Y, 0), Z) --> and (fcmp ord X, Y), Z
1705 // or (fcmp uno X, 0), (or (fcmp uno Y, 0), Z) --> or (fcmp uno X, Y), Z
1706 // Intersect FMF from the 2 source fcmps.
1707 Value *NewFCmp =
1708 Builder.CreateFCmpFMF(NanPred, X, Y, FMFSource::intersect(Op0, BO10));
1709 return BinaryOperator::Create(Opcode, NewFCmp, BO11);
1710}
1711
1712/// Match variations of De Morgan's Laws:
1713/// (~A & ~B) == (~(A | B))
1714/// (~A | ~B) == (~(A & B))
1716 InstCombiner &IC) {
1717 const Instruction::BinaryOps Opcode = I.getOpcode();
1718 assert((Opcode == Instruction::And || Opcode == Instruction::Or) &&
1719 "Trying to match De Morgan's Laws with something other than and/or");
1720
1721 // Flip the logic operation.
1722 const Instruction::BinaryOps FlippedOpcode =
1723 (Opcode == Instruction::And) ? Instruction::Or : Instruction::And;
1724
1725 Value *Op0 = I.getOperand(0), *Op1 = I.getOperand(1);
1726 Value *A, *B;
1727 if (match(Op0, m_OneUse(m_Not(m_Value(A)))) &&
1728 match(Op1, m_OneUse(m_Not(m_Value(B)))) &&
1729 !IC.isFreeToInvert(A, A->hasOneUse()) &&
1730 !IC.isFreeToInvert(B, B->hasOneUse())) {
1731 Value *AndOr =
1732 IC.Builder.CreateBinOp(FlippedOpcode, A, B, I.getName() + ".demorgan");
1733 return BinaryOperator::CreateNot(AndOr);
1734 }
1735
1736 // The 'not' ops may require reassociation.
1737 // (A & ~B) & ~C --> A & ~(B | C)
1738 // (~B & A) & ~C --> A & ~(B | C)
1739 // (A | ~B) | ~C --> A | ~(B & C)
1740 // (~B | A) | ~C --> A | ~(B & C)
1741 Value *C;
1742 if (match(Op0, m_OneUse(m_c_BinOp(Opcode, m_Value(A), m_Not(m_Value(B))))) &&
1743 match(Op1, m_Not(m_Value(C)))) {
1744 Value *FlippedBO = IC.Builder.CreateBinOp(FlippedOpcode, B, C);
1745 return BinaryOperator::Create(Opcode, A, IC.Builder.CreateNot(FlippedBO));
1746 }
1747
1748 return nullptr;
1749}
1750
1751bool InstCombinerImpl::shouldOptimizeCast(CastInst *CI) {
1752 Value *CastSrc = CI->getOperand(0);
1753
1754 // Noop casts and casts of constants should be eliminated trivially.
1755 if (CI->getSrcTy() == CI->getDestTy() || isa<Constant>(CastSrc))
1756 return false;
1757
1758 // If this cast is paired with another cast that can be eliminated, we prefer
1759 // to have it eliminated.
1760 if (const auto *PrecedingCI = dyn_cast<CastInst>(CastSrc))
1761 if (isEliminableCastPair(PrecedingCI, CI))
1762 return false;
1763
1764 return true;
1765}
1766
1767/// Fold {and,or,xor} (cast X), C.
1769 InstCombinerImpl &IC) {
1771 if (!C)
1772 return nullptr;
1773
1774 auto LogicOpc = Logic.getOpcode();
1775 Type *DestTy = Logic.getType();
1776 Type *SrcTy = Cast->getSrcTy();
1777
1778 // Move the logic operation ahead of a zext or sext if the constant is
1779 // unchanged in the smaller source type. Performing the logic in a smaller
1780 // type may provide more information to later folds, and the smaller logic
1781 // instruction may be cheaper (particularly in the case of vectors).
1782 Value *X;
1783 auto &DL = IC.getDataLayout();
1784 if (match(Cast, m_OneUse(m_ZExt(m_Value(X))))) {
1785 PreservedCastFlags Flags;
1786 if (Constant *TruncC = getLosslessUnsignedTrunc(C, SrcTy, DL, &Flags)) {
1787 // LogicOpc (zext X), C --> zext (LogicOpc X, C)
1788 Value *NewOp = IC.Builder.CreateBinOp(LogicOpc, X, TruncC);
1789 auto *ZExt = new ZExtInst(NewOp, DestTy);
1790 ZExt->setNonNeg(Flags.NNeg);
1791 ZExt->andIRFlags(Cast);
1792 return ZExt;
1793 }
1794 }
1795
1796 if (match(Cast, m_OneUse(m_SExtLike(m_Value(X))))) {
1797 if (Constant *TruncC = getLosslessSignedTrunc(C, SrcTy, DL)) {
1798 // LogicOpc (sext X), C --> sext (LogicOpc X, C)
1799 Value *NewOp = IC.Builder.CreateBinOp(LogicOpc, X, TruncC);
1800 return new SExtInst(NewOp, DestTy);
1801 }
1802 }
1803
1804 return nullptr;
1805}
1806
1807/// Fold {and,or,xor} (cast X), Y.
1808Instruction *InstCombinerImpl::foldCastedBitwiseLogic(BinaryOperator &I) {
1809 auto LogicOpc = I.getOpcode();
1810 assert(I.isBitwiseLogicOp() && "Unexpected opcode for bitwise logic folding");
1811
1812 Value *Op0 = I.getOperand(0), *Op1 = I.getOperand(1);
1813
1814 // fold bitwise(A >> BW - 1, zext(icmp)) (BW is the scalar bits of the
1815 // type of A)
1816 // -> bitwise(zext(A < 0), zext(icmp))
1817 // -> zext(bitwise(A < 0, icmp))
1818 auto FoldBitwiseICmpZeroWithICmp = [&](Value *Op0,
1819 Value *Op1) -> Instruction * {
1820 Value *A;
1821 bool IsMatched =
1822 match(Op0,
1824 m_Value(A),
1825 m_SpecificInt(Op0->getType()->getScalarSizeInBits() - 1)))) &&
1826 match(Op1, m_OneUse(m_ZExt(m_ICmp(m_Value(), m_Value()))));
1827
1828 if (!IsMatched)
1829 return nullptr;
1830
1831 auto *ICmpL =
1832 Builder.CreateICmpSLT(A, Constant::getNullValue(A->getType()));
1833 auto *ICmpR = cast<ZExtInst>(Op1)->getOperand(0);
1834 auto *BitwiseOp = Builder.CreateBinOp(LogicOpc, ICmpL, ICmpR);
1835
1836 return new ZExtInst(BitwiseOp, Op0->getType());
1837 };
1838
1839 if (auto *Ret = FoldBitwiseICmpZeroWithICmp(Op0, Op1))
1840 return Ret;
1841
1842 if (auto *Ret = FoldBitwiseICmpZeroWithICmp(Op1, Op0))
1843 return Ret;
1844
1845 CastInst *Cast0 = dyn_cast<CastInst>(Op0);
1846 if (!Cast0)
1847 return nullptr;
1848
1849 // This must be a cast from an integer or integer vector source type to allow
1850 // transformation of the logic operation to the source type.
1851 Type *DestTy = I.getType();
1852 Type *SrcTy = Cast0->getSrcTy();
1853 if (!SrcTy->isIntOrIntVectorTy())
1854 return nullptr;
1855
1856 if (Instruction *Ret = foldLogicCastConstant(I, Cast0, *this))
1857 return Ret;
1858
1859 CastInst *Cast1 = dyn_cast<CastInst>(Op1);
1860 if (!Cast1)
1861 return nullptr;
1862
1863 // Both operands of the logic operation are casts. The casts must be the
1864 // same kind for reduction.
1865 Instruction::CastOps CastOpcode = Cast0->getOpcode();
1866 if (CastOpcode != Cast1->getOpcode())
1867 return nullptr;
1868
1869 // Can't fold it profitably if no one of casts has one use.
1870 if (!Cast0->hasOneUse() && !Cast1->hasOneUse())
1871 return nullptr;
1872
1873 Value *X, *Y;
1874 if (match(Cast0, m_ZExtOrSExt(m_Value(X))) &&
1875 match(Cast1, m_ZExtOrSExt(m_Value(Y)))) {
1876 // Cast the narrower source to the wider source type.
1877 unsigned XNumBits = X->getType()->getScalarSizeInBits();
1878 unsigned YNumBits = Y->getType()->getScalarSizeInBits();
1879 if (XNumBits != YNumBits) {
1880 // Cast the narrower source to the wider source type only if both of casts
1881 // have one use to avoid creating an extra instruction.
1882 if (!Cast0->hasOneUse() || !Cast1->hasOneUse())
1883 return nullptr;
1884
1885 // If the source types do not match, but the casts are matching extends,
1886 // we can still narrow the logic op.
1887 if (XNumBits < YNumBits) {
1888 X = Builder.CreateCast(CastOpcode, X, Y->getType());
1889 } else if (YNumBits < XNumBits) {
1890 Y = Builder.CreateCast(CastOpcode, Y, X->getType());
1891 }
1892 }
1893
1894 // Do the logic op in the intermediate width, then widen more.
1895 Value *NarrowLogic = Builder.CreateBinOp(LogicOpc, X, Y, I.getName());
1896 auto *Disjoint = dyn_cast<PossiblyDisjointInst>(&I);
1897 auto *NewDisjoint = dyn_cast<PossiblyDisjointInst>(NarrowLogic);
1898 if (Disjoint && NewDisjoint)
1899 NewDisjoint->setIsDisjoint(Disjoint->isDisjoint());
1900 return CastInst::Create(CastOpcode, NarrowLogic, DestTy);
1901 }
1902
1903 // If the src type of casts are different, give up for other cast opcodes.
1904 if (SrcTy != Cast1->getSrcTy())
1905 return nullptr;
1906
1907 Value *Cast0Src = Cast0->getOperand(0);
1908 Value *Cast1Src = Cast1->getOperand(0);
1909
1910 // fold logic(cast(A), cast(B)) -> cast(logic(A, B))
1911 if (shouldOptimizeCast(Cast0) && shouldOptimizeCast(Cast1)) {
1912 Value *NewOp = Builder.CreateBinOp(LogicOpc, Cast0Src, Cast1Src,
1913 I.getName());
1914 return CastInst::Create(CastOpcode, NewOp, DestTy);
1915 }
1916
1917 return nullptr;
1918}
1919
1921 InstCombiner::BuilderTy &Builder) {
1922 assert(I.getOpcode() == Instruction::And);
1923 Value *Op0 = I.getOperand(0);
1924 Value *Op1 = I.getOperand(1);
1925 Value *A, *B;
1926
1927 // Operand complexity canonicalization guarantees that the 'or' is Op0.
1928 // (A | B) & ~(A & B) --> A ^ B
1929 // (A | B) & ~(B & A) --> A ^ B
1930 if (match(&I, m_BinOp(m_Or(m_Value(A), m_Value(B)),
1932 return BinaryOperator::CreateXor(A, B);
1933
1934 // (A | ~B) & (~A | B) --> ~(A ^ B)
1935 // (A | ~B) & (B | ~A) --> ~(A ^ B)
1936 // (~B | A) & (~A | B) --> ~(A ^ B)
1937 // (~B | A) & (B | ~A) --> ~(A ^ B)
1938 if (Op0->hasOneUse() || Op1->hasOneUse())
1941 return BinaryOperator::CreateNot(Builder.CreateXor(A, B));
1942
1943 return nullptr;
1944}
1945
1947 InstCombiner::BuilderTy &Builder) {
1948 assert(I.getOpcode() == Instruction::Or);
1949 Value *Op0 = I.getOperand(0);
1950 Value *Op1 = I.getOperand(1);
1951 Value *A, *B;
1952
1953 // Operand complexity canonicalization guarantees that the 'and' is Op0.
1954 // (A & B) | ~(A | B) --> ~(A ^ B)
1955 // (A & B) | ~(B | A) --> ~(A ^ B)
1956 if (Op0->hasOneUse() || Op1->hasOneUse())
1957 if (match(Op0, m_And(m_Value(A), m_Value(B))) &&
1959 return BinaryOperator::CreateNot(Builder.CreateXor(A, B));
1960
1961 // Operand complexity canonicalization guarantees that the 'xor' is Op0.
1962 // (A ^ B) | ~(A | B) --> ~(A & B)
1963 // (A ^ B) | ~(B | A) --> ~(A & B)
1964 if (Op0->hasOneUse() || Op1->hasOneUse())
1965 if (match(Op0, m_Xor(m_Value(A), m_Value(B))) &&
1967 return BinaryOperator::CreateNot(Builder.CreateAnd(A, B));
1968
1969 // (A & ~B) | (~A & B) --> A ^ B
1970 // (A & ~B) | (B & ~A) --> A ^ B
1971 // (~B & A) | (~A & B) --> A ^ B
1972 // (~B & A) | (B & ~A) --> A ^ B
1973 if (match(Op0, m_c_And(m_Value(A), m_Not(m_Value(B)))) &&
1975 return BinaryOperator::CreateXor(A, B);
1976
1977 return nullptr;
1978}
1979
1980/// Return true if a constant shift amount is always less than the specified
1981/// bit-width. If not, the shift could create poison in the narrower type.
1982static bool canNarrowShiftAmt(Constant *C, unsigned BitWidth) {
1983 APInt Threshold(C->getType()->getScalarSizeInBits(), BitWidth);
1984 return match(C, m_SpecificInt_ICMP(ICmpInst::ICMP_ULT, Threshold));
1985}
1986
1987/// Try to use narrower ops (sink zext ops) for an 'and' with binop operand and
1988/// a common zext operand: and (binop (zext X), C), (zext X).
1989Instruction *InstCombinerImpl::narrowMaskedBinOp(BinaryOperator &And) {
1990 // This transform could also apply to {or, and, xor}, but there are better
1991 // folds for those cases, so we don't expect those patterns here. AShr is not
1992 // handled because it should always be transformed to LShr in this sequence.
1993 // The subtract transform is different because it has a constant on the left.
1994 // Add/mul commute the constant to RHS; sub with constant RHS becomes add.
1995 Value *Op0 = And.getOperand(0), *Op1 = And.getOperand(1);
1996 Constant *C;
1997 if (!match(Op0, m_OneUse(m_Add(m_Specific(Op1), m_Constant(C)))) &&
1998 !match(Op0, m_OneUse(m_Mul(m_Specific(Op1), m_Constant(C)))) &&
1999 !match(Op0, m_OneUse(m_LShr(m_Specific(Op1), m_Constant(C)))) &&
2000 !match(Op0, m_OneUse(m_Shl(m_Specific(Op1), m_Constant(C)))) &&
2001 !match(Op0, m_OneUse(m_Sub(m_Constant(C), m_Specific(Op1)))))
2002 return nullptr;
2003
2004 Value *X;
2005 if (!match(Op1, m_ZExt(m_Value(X))) || Op1->hasNUsesOrMore(3))
2006 return nullptr;
2007
2008 Type *Ty = And.getType();
2009 if (!isa<VectorType>(Ty) && !shouldChangeType(Ty, X->getType()))
2010 return nullptr;
2011
2012 // If we're narrowing a shift, the shift amount must be safe (less than the
2013 // width) in the narrower type. If the shift amount is greater, instsimplify
2014 // usually handles that case, but we can't guarantee/assert it.
2016 if (Opc == Instruction::LShr || Opc == Instruction::Shl)
2017 if (!canNarrowShiftAmt(C, X->getType()->getScalarSizeInBits()))
2018 return nullptr;
2019
2020 // and (sub C, (zext X)), (zext X) --> zext (and (sub C', X), X)
2021 // and (binop (zext X), C), (zext X) --> zext (and (binop X, C'), X)
2022 Value *NewC = ConstantExpr::getTrunc(C, X->getType());
2023 Value *NewBO = Opc == Instruction::Sub ? Builder.CreateBinOp(Opc, NewC, X)
2024 : Builder.CreateBinOp(Opc, X, NewC);
2025 return new ZExtInst(Builder.CreateAnd(NewBO, X), Ty);
2026}
2027
2028/// Try folding relatively complex patterns for both And and Or operations
2029/// with all And and Or swapped.
2031 InstCombiner::BuilderTy &Builder) {
2032 const Instruction::BinaryOps Opcode = I.getOpcode();
2033 assert(Opcode == Instruction::And || Opcode == Instruction::Or);
2034
2035 // Flip the logic operation.
2036 const Instruction::BinaryOps FlippedOpcode =
2037 (Opcode == Instruction::And) ? Instruction::Or : Instruction::And;
2038
2039 Value *Op0 = I.getOperand(0), *Op1 = I.getOperand(1);
2040 Value *A, *B, *C, *X, *Y, *Dummy;
2041
2042 // Match following expressions:
2043 // (~(A | B) & C)
2044 // (~(A & B) | C)
2045 // Captures X = ~(A | B) or ~(A & B)
2046 const auto matchNotOrAnd =
2047 [Opcode, FlippedOpcode](Value *Op, auto m_A, auto m_B, auto m_C,
2048 Value *&X, bool CountUses = false) -> bool {
2049 if (CountUses && !Op->hasOneUse())
2050 return false;
2051
2052 if (match(Op,
2053 m_c_BinOp(FlippedOpcode,
2054 m_Value(X, m_Not(m_c_BinOp(Opcode, m_A, m_B))), m_C)))
2055 return !CountUses || X->hasOneUse();
2056
2057 return false;
2058 };
2059
2060 // (~(A | B) & C) | ... --> ...
2061 // (~(A & B) | C) & ... --> ...
2062 // TODO: One use checks are conservative. We just need to check that a total
2063 // number of multiple used values does not exceed reduction
2064 // in operations.
2065 if (matchNotOrAnd(Op0, m_Value(A), m_Value(B), m_Value(C), X)) {
2066 // (~(A | B) & C) | (~(A | C) & B) --> (B ^ C) & ~A
2067 // (~(A & B) | C) & (~(A & C) | B) --> ~((B ^ C) & A)
2068 if (matchNotOrAnd(Op1, m_Specific(A), m_Specific(C), m_Specific(B), Dummy,
2069 true)) {
2070 Value *Xor = Builder.CreateXor(B, C);
2071 return (Opcode == Instruction::Or)
2072 ? BinaryOperator::CreateAnd(Xor, Builder.CreateNot(A))
2073 : BinaryOperator::CreateNot(Builder.CreateAnd(Xor, A));
2074 }
2075
2076 // (~(A | B) & C) | (~(B | C) & A) --> (A ^ C) & ~B
2077 // (~(A & B) | C) & (~(B & C) | A) --> ~((A ^ C) & B)
2078 if (matchNotOrAnd(Op1, m_Specific(B), m_Specific(C), m_Specific(A), Dummy,
2079 true)) {
2080 Value *Xor = Builder.CreateXor(A, C);
2081 return (Opcode == Instruction::Or)
2082 ? BinaryOperator::CreateAnd(Xor, Builder.CreateNot(B))
2083 : BinaryOperator::CreateNot(Builder.CreateAnd(Xor, B));
2084 }
2085
2086 // (~(A | B) & C) | ~(A | C) --> ~((B & C) | A)
2087 // (~(A & B) | C) & ~(A & C) --> ~((B | C) & A)
2088 if (match(Op1, m_OneUse(m_Not(m_OneUse(
2089 m_c_BinOp(Opcode, m_Specific(A), m_Specific(C)))))))
2090 return BinaryOperator::CreateNot(Builder.CreateBinOp(
2091 Opcode, Builder.CreateBinOp(FlippedOpcode, B, C), A));
2092
2093 // (~(A | B) & C) | ~(B | C) --> ~((A & C) | B)
2094 // (~(A & B) | C) & ~(B & C) --> ~((A | C) & B)
2095 if (match(Op1, m_OneUse(m_Not(m_OneUse(
2096 m_c_BinOp(Opcode, m_Specific(B), m_Specific(C)))))))
2097 return BinaryOperator::CreateNot(Builder.CreateBinOp(
2098 Opcode, Builder.CreateBinOp(FlippedOpcode, A, C), B));
2099
2100 // (~(A | B) & C) | ~(C | (A ^ B)) --> ~((A | B) & (C | (A ^ B)))
2101 // Note, the pattern with swapped and/or is not handled because the
2102 // result is more undefined than a source:
2103 // (~(A & B) | C) & ~(C & (A ^ B)) --> (A ^ B ^ C) | ~(A | C) is invalid.
2104 if (Opcode == Instruction::Or && Op0->hasOneUse() &&
2105 match(Op1,
2107 Y, m_c_BinOp(Opcode, m_Specific(C),
2108 m_c_Xor(m_Specific(A), m_Specific(B)))))))) {
2109 // X = ~(A | B)
2110 // Y = (C | (A ^ B)
2111 Value *Or = cast<BinaryOperator>(X)->getOperand(0);
2112 return BinaryOperator::CreateNot(Builder.CreateAnd(Or, Y));
2113 }
2114 }
2115
2116 // (~A & B & C) | ... --> ...
2117 // (~A | B | C) | ... --> ...
2118 // TODO: One use checks are conservative. We just need to check that a total
2119 // number of multiple used values does not exceed reduction
2120 // in operations.
2121 if (match(Op0,
2122 m_OneUse(m_c_BinOp(FlippedOpcode,
2123 m_BinOp(FlippedOpcode, m_Value(B), m_Value(C)),
2124 m_Value(X, m_Not(m_Value(A)))))) ||
2125 match(Op0, m_OneUse(m_c_BinOp(FlippedOpcode,
2126 m_c_BinOp(FlippedOpcode, m_Value(C),
2127 m_Value(X, m_Not(m_Value(A)))),
2128 m_Value(B))))) {
2129 // X = ~A
2130 // (~A & B & C) | ~(A | B | C) --> ~(A | (B ^ C))
2131 // (~A | B | C) & ~(A & B & C) --> (~A | (B ^ C))
2132 if (match(Op1, m_OneUse(m_Not(m_c_BinOp(
2133 Opcode, m_c_BinOp(Opcode, m_Specific(A), m_Specific(B)),
2134 m_Specific(C))))) ||
2136 Opcode, m_c_BinOp(Opcode, m_Specific(B), m_Specific(C)),
2137 m_Specific(A))))) ||
2139 Opcode, m_c_BinOp(Opcode, m_Specific(A), m_Specific(C)),
2140 m_Specific(B)))))) {
2141 Value *Xor = Builder.CreateXor(B, C);
2142 return (Opcode == Instruction::Or)
2143 ? BinaryOperator::CreateNot(Builder.CreateOr(Xor, A))
2144 : BinaryOperator::CreateOr(Xor, X);
2145 }
2146
2147 // (~A & B & C) | ~(A | B) --> (C | ~B) & ~A
2148 // (~A | B | C) & ~(A & B) --> (C & ~B) | ~A
2149 if (match(Op1, m_OneUse(m_Not(m_OneUse(
2150 m_c_BinOp(Opcode, m_Specific(A), m_Specific(B)))))))
2152 FlippedOpcode, Builder.CreateBinOp(Opcode, C, Builder.CreateNot(B)),
2153 X);
2154
2155 // (~A & B & C) | ~(A | C) --> (B | ~C) & ~A
2156 // (~A | B | C) & ~(A & C) --> (B & ~C) | ~A
2157 if (match(Op1, m_OneUse(m_Not(m_OneUse(
2158 m_c_BinOp(Opcode, m_Specific(A), m_Specific(C)))))))
2160 FlippedOpcode, Builder.CreateBinOp(Opcode, B, Builder.CreateNot(C)),
2161 X);
2162 }
2163
2164 return nullptr;
2165}
2166
2167/// Try to reassociate a pair of binops so that values with one use only are
2168/// part of the same instruction. This may enable folds that are limited with
2169/// multi-use restrictions and makes it more likely to match other patterns that
2170/// are looking for a common operand.
2172 InstCombinerImpl::BuilderTy &Builder) {
2173 Instruction::BinaryOps Opcode = BO.getOpcode();
2174 Value *X, *Y, *Z;
2175 if (match(&BO,
2176 m_c_BinOp(Opcode, m_OneUse(m_BinOp(Opcode, m_Value(X), m_Value(Y))),
2177 m_OneUse(m_Value(Z))))) {
2178 if (!isa<Constant>(X) && !isa<Constant>(Y) && !isa<Constant>(Z)) {
2179 // (X op Y) op Z --> (Y op Z) op X
2180 if (!X->hasOneUse()) {
2181 Value *YZ = Builder.CreateBinOp(Opcode, Y, Z);
2182 return BinaryOperator::Create(Opcode, YZ, X);
2183 }
2184 // (X op Y) op Z --> (X op Z) op Y
2185 if (!Y->hasOneUse()) {
2186 Value *XZ = Builder.CreateBinOp(Opcode, X, Z);
2187 return BinaryOperator::Create(Opcode, XZ, Y);
2188 }
2189 }
2190 }
2191
2192 return nullptr;
2193}
2194
2195// Match
2196// (X + C2) | C
2197// (X + C2) ^ C
2198// (X + C2) & C
2199// and convert to do the bitwise logic first:
2200// (X | C) + C2
2201// (X ^ C) + C2
2202// (X & C) + C2
2203// iff bits affected by logic op are lower than last bit affected by math op
2205 InstCombiner::BuilderTy &Builder) {
2206 Type *Ty = I.getType();
2207 Instruction::BinaryOps OpC = I.getOpcode();
2208 Value *Op0 = I.getOperand(0);
2209 Value *Op1 = I.getOperand(1);
2210 Value *X;
2211 const APInt *C, *C2;
2212
2213 if (!(match(Op0, m_OneUse(m_Add(m_Value(X), m_APInt(C2)))) &&
2214 match(Op1, m_APInt(C))))
2215 return nullptr;
2216
2217 unsigned Width = Ty->getScalarSizeInBits();
2218 unsigned LastOneMath = Width - C2->countr_zero();
2219
2220 switch (OpC) {
2221 case Instruction::And:
2222 if (C->countl_one() < LastOneMath)
2223 return nullptr;
2224 break;
2225 case Instruction::Xor:
2226 case Instruction::Or:
2227 if (C->countl_zero() < LastOneMath)
2228 return nullptr;
2229 break;
2230 default:
2231 llvm_unreachable("Unexpected BinaryOp!");
2232 }
2233
2234 Value *NewBinOp = Builder.CreateBinOp(OpC, X, ConstantInt::get(Ty, *C));
2235 return BinaryOperator::CreateWithCopiedFlags(Instruction::Add, NewBinOp,
2236 ConstantInt::get(Ty, *C2), Op0);
2237}
2238
2239// binop(shift(ShiftedC1, ShAmt), shift(ShiftedC2, add(ShAmt, AddC))) ->
2240// shift(binop(ShiftedC1, shift(ShiftedC2, AddC)), ShAmt)
2241// where both shifts are the same and AddC is a valid shift amount.
2242Instruction *InstCombinerImpl::foldBinOpOfDisplacedShifts(BinaryOperator &I) {
2243 assert((I.isBitwiseLogicOp() || I.getOpcode() == Instruction::Add) &&
2244 "Unexpected opcode");
2245
2246 Value *ShAmt;
2247 Constant *ShiftedC1, *ShiftedC2, *AddC;
2248 Type *Ty = I.getType();
2249 unsigned BitWidth = Ty->getScalarSizeInBits();
2250 if (!match(&I, m_c_BinOp(m_Shift(m_ImmConstant(ShiftedC1), m_Value(ShAmt)),
2251 m_Shift(m_ImmConstant(ShiftedC2),
2252 m_AddLike(m_Deferred(ShAmt),
2253 m_ImmConstant(AddC))))))
2254 return nullptr;
2255
2256 // Make sure the add constant is a valid shift amount.
2257 if (!match(AddC,
2259 return nullptr;
2260
2261 // Avoid constant expressions.
2262 auto *Op0Inst = dyn_cast<Instruction>(I.getOperand(0));
2263 auto *Op1Inst = dyn_cast<Instruction>(I.getOperand(1));
2264 if (!Op0Inst || !Op1Inst)
2265 return nullptr;
2266
2267 // Both shifts must be the same.
2268 Instruction::BinaryOps ShiftOp =
2269 static_cast<Instruction::BinaryOps>(Op0Inst->getOpcode());
2270 if (ShiftOp != Op1Inst->getOpcode())
2271 return nullptr;
2272
2273 // For adds, only left shifts are supported.
2274 if (I.getOpcode() == Instruction::Add && ShiftOp != Instruction::Shl)
2275 return nullptr;
2276
2277 Value *NewC = Builder.CreateBinOp(
2278 I.getOpcode(), ShiftedC1, Builder.CreateBinOp(ShiftOp, ShiftedC2, AddC));
2279 return BinaryOperator::Create(ShiftOp, NewC, ShAmt);
2280}
2281
2282// Fold and/or/xor with two equal intrinsic IDs:
2283// bitwise(fshl (A, B, ShAmt), fshl(C, D, ShAmt))
2284// -> fshl(bitwise(A, C), bitwise(B, D), ShAmt)
2285// bitwise(fshr (A, B, ShAmt), fshr(C, D, ShAmt))
2286// -> fshr(bitwise(A, C), bitwise(B, D), ShAmt)
2287// bitwise(bswap(A), bswap(B)) -> bswap(bitwise(A, B))
2288// bitwise(bswap(A), C) -> bswap(bitwise(A, bswap(C)))
2289// bitwise(bitreverse(A), bitreverse(B)) -> bitreverse(bitwise(A, B))
2290// bitwise(bitreverse(A), C) -> bitreverse(bitwise(A, bitreverse(C)))
2291static Instruction *
2293 InstCombiner::BuilderTy &Builder) {
2294 assert(I.isBitwiseLogicOp() && "Should and/or/xor");
2295 if (!I.getOperand(0)->hasOneUse())
2296 return nullptr;
2297 IntrinsicInst *X = dyn_cast<IntrinsicInst>(I.getOperand(0));
2298 if (!X)
2299 return nullptr;
2300
2301 IntrinsicInst *Y = dyn_cast<IntrinsicInst>(I.getOperand(1));
2302 if (Y && (!Y->hasOneUse() || X->getIntrinsicID() != Y->getIntrinsicID()))
2303 return nullptr;
2304
2305 Intrinsic::ID IID = X->getIntrinsicID();
2306 const APInt *RHSC;
2307 // Try to match constant RHS.
2308 if (!Y && (!(IID == Intrinsic::bswap || IID == Intrinsic::bitreverse) ||
2309 !match(I.getOperand(1), m_APInt(RHSC))))
2310 return nullptr;
2311
2312 switch (IID) {
2313 case Intrinsic::fshl:
2314 case Intrinsic::fshr: {
2315 if (X->getOperand(2) != Y->getOperand(2))
2316 return nullptr;
2317 Value *NewOp0 =
2318 Builder.CreateBinOp(I.getOpcode(), X->getOperand(0), Y->getOperand(0));
2319 Value *NewOp1 =
2320 Builder.CreateBinOp(I.getOpcode(), X->getOperand(1), Y->getOperand(1));
2321 Function *F =
2322 Intrinsic::getOrInsertDeclaration(I.getModule(), IID, I.getType());
2323 return CallInst::Create(F, {NewOp0, NewOp1, X->getOperand(2)});
2324 }
2325 case Intrinsic::bswap:
2326 case Intrinsic::bitreverse: {
2327 Value *NewOp0 = Builder.CreateBinOp(
2328 I.getOpcode(), X->getOperand(0),
2329 Y ? Y->getOperand(0)
2330 : ConstantInt::get(I.getType(), IID == Intrinsic::bswap
2331 ? RHSC->byteSwap()
2332 : RHSC->reverseBits()));
2333 Function *F =
2334 Intrinsic::getOrInsertDeclaration(I.getModule(), IID, I.getType());
2335 return CallInst::Create(F, {NewOp0});
2336 }
2337 default:
2338 return nullptr;
2339 }
2340}
2341
2342// Try to simplify V by replacing occurrences of Op with RepOp, but only look
2343// through bitwise operations. In particular, for X | Y we try to replace Y with
2344// 0 inside X and for X & Y we try to replace Y with -1 inside X.
2345// Return the simplified result of X if successful, and nullptr otherwise.
2346// If SimplifyOnly is true, no new instructions will be created.
2348 bool SimplifyOnly,
2349 InstCombinerImpl &IC,
2350 unsigned Depth = 0) {
2351 if (Op == RepOp)
2352 return nullptr;
2353
2354 if (V == Op)
2355 return RepOp;
2356
2357 auto *I = dyn_cast<BinaryOperator>(V);
2358 if (!I || !I->isBitwiseLogicOp() || Depth >= 3)
2359 return nullptr;
2360
2361 if (!I->hasOneUse())
2362 SimplifyOnly = true;
2363
2364 Value *NewOp0 = simplifyAndOrWithOpReplaced(I->getOperand(0), Op, RepOp,
2365 SimplifyOnly, IC, Depth + 1);
2366 Value *NewOp1 = simplifyAndOrWithOpReplaced(I->getOperand(1), Op, RepOp,
2367 SimplifyOnly, IC, Depth + 1);
2368 if (!NewOp0 && !NewOp1)
2369 return nullptr;
2370
2371 if (!NewOp0)
2372 NewOp0 = I->getOperand(0);
2373 if (!NewOp1)
2374 NewOp1 = I->getOperand(1);
2375
2376 if (Value *Res = simplifyBinOp(I->getOpcode(), NewOp0, NewOp1,
2378 return Res;
2379
2380 if (SimplifyOnly)
2381 return nullptr;
2382 return IC.Builder.CreateBinOp(I->getOpcode(), NewOp0, NewOp1);
2383}
2384
2385/// Reassociate and/or expressions to see if we can fold the inner and/or ops.
2386/// TODO: Make this recursive; it's a little tricky because an arbitrary
2387/// number of and/or instructions might have to be created.
2388Value *InstCombinerImpl::reassociateBooleanAndOr(Value *LHS, Value *X, Value *Y,
2389 Instruction &I, bool IsAnd,
2390 bool RHSIsLogical) {
2391 Instruction::BinaryOps Opcode = IsAnd ? Instruction::And : Instruction::Or;
2392 Value *Folded = nullptr;
2393 // LHS bop (X lop Y) --> (LHS bop X) lop Y
2394 // LHS bop (X bop Y) --> (LHS bop X) bop Y
2395 if (Value *Res = foldBooleanAndOr(LHS, X, I, IsAnd, /*IsLogical=*/false))
2396 Folded = RHSIsLogical ? Builder.CreateLogicalOp(Opcode, Res, Y)
2397 : Builder.CreateBinOp(Opcode, Res, Y);
2398 // LHS bop (X bop Y) --> X bop (LHS bop Y)
2399 // LHS bop (X lop Y) --> X lop (LHS bop Y)
2400 else if (Value *Res = foldBooleanAndOr(LHS, Y, I, IsAnd, /*IsLogical=*/false))
2401 Folded = RHSIsLogical ? Builder.CreateLogicalOp(Opcode, X, Res)
2402 : Builder.CreateBinOp(Opcode, X, Res);
2403 if (SelectInst *SI = dyn_cast_or_null<SelectInst>(Folded);
2404 SI != nullptr && !ProfcheckDisableMetadataFixes)
2405 // If the bop I was originally a lop, we could recover branch weight
2406 // information using that lop's weights. However, InstCombine usually
2407 // replaces the lop with a bop by the time we get here, deleting the branch
2408 // weight information. Therefore, we can only assume unknown branch weights.
2409 // TODO: see if it's possible to recover branch weight information from the
2410 // original lop (https://github.com/llvm/llvm-project/issues/183864).
2412 I.getFunction());
2413 return Folded;
2414}
2415
2416// FIXME: We use commutative matchers (m_c_*) for some, but not all, matches
2417// here. We should standardize that construct where it is needed or choose some
2418// other way to ensure that commutated variants of patterns are not missed.
2420 Type *Ty = I.getType();
2421
2422 if (Value *V = simplifyAndInst(I.getOperand(0), I.getOperand(1),
2423 SQ.getWithInstruction(&I)))
2424 return replaceInstUsesWith(I, V);
2425
2427 return &I;
2428
2430 return X;
2431
2433 return Phi;
2434
2435 // See if we can simplify any instructions used by the instruction whose sole
2436 // purpose is to compute bits we don't care about.
2438 return &I;
2439
2440 // Do this before using distributive laws to catch simple and/or/not patterns.
2442 return Xor;
2443
2445 return X;
2446
2447 // (A|B)&(A|C) -> A|(B&C) etc
2449 return replaceInstUsesWith(I, V);
2450
2452 return R;
2453
2454 Value *Op0 = I.getOperand(0), *Op1 = I.getOperand(1);
2455
2456 Value *X, *Y;
2457 const APInt *C;
2458 if ((match(Op0, m_OneUse(m_LogicalShift(m_One(), m_Value(X)))) ||
2459 (match(Op0, m_OneUse(m_Shl(m_APInt(C), m_Value(X)))) && (*C)[0])) &&
2460 match(Op1, m_One())) {
2461 // (1 >> X) & 1 --> zext(X == 0)
2462 // (C << X) & 1 --> zext(X == 0), when C is odd
2463 Value *IsZero = Builder.CreateICmpEQ(X, ConstantInt::get(Ty, 0));
2464 return new ZExtInst(IsZero, Ty);
2465 }
2466
2467 // (-(X & 1)) & Y --> (X & 1) == 0 ? 0 : Y
2468 Value *Neg;
2469 if (match(&I,
2471 m_Value(Y)))) {
2472 Value *Cmp = Builder.CreateIsNull(Neg);
2473 return createSelectInstWithUnknownProfile(Cmp,
2475 }
2476
2477 // Canonicalize:
2478 // (X +/- Y) & Y --> ~X & Y when Y is a power of 2.
2481 m_Sub(m_Value(X), m_Deferred(Y)))))) &&
2482 isKnownToBeAPowerOfTwo(Y, /*OrZero*/ true, &I))
2483 return BinaryOperator::CreateAnd(Builder.CreateNot(X), Y);
2484
2485 if (match(Op1, m_APInt(C))) {
2486 const APInt *XorC;
2487 if (match(Op0, m_OneUse(m_Xor(m_Value(X), m_APInt(XorC))))) {
2488 // (X ^ C1) & C2 --> (X & C2) ^ (C1&C2)
2489 Constant *NewC = ConstantInt::get(Ty, *C & *XorC);
2490 Value *And = Builder.CreateAnd(X, Op1);
2491 And->takeName(Op0);
2492 return BinaryOperator::CreateXor(And, NewC);
2493 }
2494
2495 const APInt *OrC;
2496 if (match(Op0, m_OneUse(m_Or(m_Value(X), m_APInt(OrC))))) {
2497 // (X | C1) & C2 --> (X & C2^(C1&C2)) | (C1&C2)
2498 // NOTE: This reduces the number of bits set in the & mask, which
2499 // can expose opportunities for store narrowing for scalars.
2500 // NOTE: SimplifyDemandedBits should have already removed bits from C1
2501 // that aren't set in C2. Meaning we can replace (C1&C2) with C1 in
2502 // above, but this feels safer.
2503 APInt Together = *C & *OrC;
2504 Value *And = Builder.CreateAnd(X, ConstantInt::get(Ty, Together ^ *C));
2505 And->takeName(Op0);
2506 return BinaryOperator::CreateOr(And, ConstantInt::get(Ty, Together));
2507 }
2508
2509 unsigned Width = Ty->getScalarSizeInBits();
2510 const APInt *ShiftC;
2511 if (match(Op0, m_OneUse(m_SExt(m_AShr(m_Value(X), m_APInt(ShiftC))))) &&
2512 ShiftC->ult(Width)) {
2513 if (*C == APInt::getLowBitsSet(Width, Width - ShiftC->getZExtValue())) {
2514 // We are clearing high bits that were potentially set by sext+ashr:
2515 // and (sext (ashr X, ShiftC)), C --> lshr (sext X), ShiftC
2516 Value *Sext = Builder.CreateSExt(X, Ty);
2517 Constant *ShAmtC = ConstantInt::get(Ty, ShiftC->zext(Width));
2518 return BinaryOperator::CreateLShr(Sext, ShAmtC);
2519 }
2520 }
2521
2522 // If this 'and' clears the sign-bits added by ashr, replace with lshr:
2523 // and (ashr X, ShiftC), C --> lshr X, ShiftC
2524 if (match(Op0, m_AShr(m_Value(X), m_APInt(ShiftC))) && ShiftC->ult(Width) &&
2525 C->isMask(Width - ShiftC->getZExtValue()))
2526 return BinaryOperator::CreateLShr(X, ConstantInt::get(Ty, *ShiftC));
2527
2528 const APInt *AddC;
2529 if (match(Op0, m_Add(m_Value(X), m_APInt(AddC)))) {
2530 // If we are masking the result of the add down to exactly one bit and
2531 // the constant we are adding has no bits set below that bit, then the
2532 // add is flipping a single bit. Example:
2533 // (X + 4) & 4 --> (X & 4) ^ 4
2534 if (Op0->hasOneUse() && C->isPowerOf2() && (*AddC & (*C - 1)) == 0) {
2535 assert((*C & *AddC) != 0 && "Expected common bit");
2536 Value *NewAnd = Builder.CreateAnd(X, Op1);
2537 return BinaryOperator::CreateXor(NewAnd, Op1);
2538 }
2539 }
2540
2541 // ((C1 OP zext(X)) & C2) -> zext((C1 OP X) & C2) if C2 fits in the
2542 // bitwidth of X and OP behaves well when given trunc(C1) and X.
2543 auto isNarrowableBinOpcode = [](BinaryOperator *B) {
2544 switch (B->getOpcode()) {
2545 case Instruction::Xor:
2546 case Instruction::Or:
2547 case Instruction::Mul:
2548 case Instruction::Add:
2549 case Instruction::Sub:
2550 return true;
2551 default:
2552 return false;
2553 }
2554 };
2555 BinaryOperator *BO;
2556 if (match(Op0, m_OneUse(m_BinOp(BO))) && isNarrowableBinOpcode(BO)) {
2557 Instruction::BinaryOps BOpcode = BO->getOpcode();
2558 Value *X;
2559 const APInt *C1;
2560 // TODO: The one-use restrictions could be relaxed a little if the AND
2561 // is going to be removed.
2562 // Try to narrow the 'and' and a binop with constant operand:
2563 // and (bo (zext X), C1), C --> zext (and (bo X, TruncC1), TruncC)
2564 if (match(BO, m_c_BinOp(m_OneUse(m_ZExt(m_Value(X))), m_APInt(C1))) &&
2565 C->isIntN(X->getType()->getScalarSizeInBits())) {
2566 unsigned XWidth = X->getType()->getScalarSizeInBits();
2567 Constant *TruncC1 = ConstantInt::get(X->getType(), C1->trunc(XWidth));
2568 Value *BinOp = isa<ZExtInst>(BO->getOperand(0))
2569 ? Builder.CreateBinOp(BOpcode, X, TruncC1)
2570 : Builder.CreateBinOp(BOpcode, TruncC1, X);
2571 Constant *TruncC = ConstantInt::get(X->getType(), C->trunc(XWidth));
2572 Value *And = Builder.CreateAnd(BinOp, TruncC);
2573 return new ZExtInst(And, Ty);
2574 }
2575
2576 // Similar to above: if the mask matches the zext input width, then the
2577 // 'and' can be eliminated, so we can truncate the other variable op:
2578 // and (bo (zext X), Y), C --> zext (bo X, (trunc Y))
2579 if (isa<Instruction>(BO->getOperand(0)) &&
2580 match(BO->getOperand(0), m_OneUse(m_ZExt(m_Value(X)))) &&
2581 C->isMask(X->getType()->getScalarSizeInBits())) {
2582 Y = BO->getOperand(1);
2583 Value *TrY = Builder.CreateTrunc(Y, X->getType(), Y->getName() + ".tr");
2584 Value *NewBO =
2585 Builder.CreateBinOp(BOpcode, X, TrY, BO->getName() + ".narrow");
2586 return new ZExtInst(NewBO, Ty);
2587 }
2588 // and (bo Y, (zext X)), C --> zext (bo (trunc Y), X)
2589 if (isa<Instruction>(BO->getOperand(1)) &&
2590 match(BO->getOperand(1), m_OneUse(m_ZExt(m_Value(X)))) &&
2591 C->isMask(X->getType()->getScalarSizeInBits())) {
2592 Y = BO->getOperand(0);
2593 Value *TrY = Builder.CreateTrunc(Y, X->getType(), Y->getName() + ".tr");
2594 Value *NewBO =
2595 Builder.CreateBinOp(BOpcode, TrY, X, BO->getName() + ".narrow");
2596 return new ZExtInst(NewBO, Ty);
2597 }
2598 }
2599
2600 // This is intentionally placed after the narrowing transforms for
2601 // efficiency (transform directly to the narrow logic op if possible).
2602 // If the mask is only needed on one incoming arm, push the 'and' op up.
2603 if (match(Op0, m_OneUse(m_Xor(m_Value(X), m_Value(Y)))) ||
2604 match(Op0, m_OneUse(m_Or(m_Value(X), m_Value(Y))))) {
2605 APInt NotAndMask(~(*C));
2606 BinaryOperator::BinaryOps BinOp = cast<BinaryOperator>(Op0)->getOpcode();
2607 if (MaskedValueIsZero(X, NotAndMask, &I)) {
2608 // Not masking anything out for the LHS, move mask to RHS.
2609 // and ({x}or X, Y), C --> {x}or X, (and Y, C)
2610 Value *NewRHS = Builder.CreateAnd(Y, Op1, Y->getName() + ".masked");
2611 return BinaryOperator::Create(BinOp, X, NewRHS);
2612 }
2613 if (!isa<Constant>(Y) && MaskedValueIsZero(Y, NotAndMask, &I)) {
2614 // Not masking anything out for the RHS, move mask to LHS.
2615 // and ({x}or X, Y), C --> {x}or (and X, C), Y
2616 Value *NewLHS = Builder.CreateAnd(X, Op1, X->getName() + ".masked");
2617 return BinaryOperator::Create(BinOp, NewLHS, Y);
2618 }
2619 }
2620
2621 // When the mask is a power-of-2 constant and op0 is a shifted-power-of-2
2622 // constant, test if the shift amount equals the offset bit index:
2623 // (ShiftC << X) & C --> X == (log2(C) - log2(ShiftC)) ? C : 0
2624 // (ShiftC >> X) & C --> X == (log2(ShiftC) - log2(C)) ? C : 0
2625 if (C->isPowerOf2() &&
2626 match(Op0, m_OneUse(m_LogicalShift(m_Power2(ShiftC), m_Value(X))))) {
2627 int Log2ShiftC = ShiftC->exactLogBase2();
2628 int Log2C = C->exactLogBase2();
2629 bool IsShiftLeft =
2630 cast<BinaryOperator>(Op0)->getOpcode() == Instruction::Shl;
2631 int BitNum = IsShiftLeft ? Log2C - Log2ShiftC : Log2ShiftC - Log2C;
2632 assert(BitNum >= 0 && "Expected demanded bits to handle impossible mask");
2633 Value *Cmp = Builder.CreateICmpEQ(X, ConstantInt::get(Ty, BitNum));
2634 return createSelectInstWithUnknownProfile(Cmp, ConstantInt::get(Ty, *C),
2636 }
2637
2638 Constant *C1, *C2;
2639 const APInt *C3 = C;
2640 Value *X;
2641 if (C3->isPowerOf2()) {
2642 Constant *Log2C3 = ConstantInt::get(Ty, C3->countr_zero());
2644 m_ImmConstant(C2)))) &&
2645 match(C1, m_Power2())) {
2647 Constant *LshrC = ConstantExpr::getAdd(C2, Log2C3);
2648 KnownBits KnownLShrc = computeKnownBits(LshrC, nullptr);
2649 if (KnownLShrc.getMaxValue().ult(Width)) {
2650 // iff C1,C3 is pow2 and C2 + cttz(C3) < BitWidth:
2651 // ((C1 << X) >> C2) & C3 -> X == (cttz(C3)+C2-cttz(C1)) ? C3 : 0
2652 Constant *CmpC = ConstantExpr::getSub(LshrC, Log2C1);
2653 Value *Cmp = Builder.CreateICmpEQ(X, CmpC);
2654 return createSelectInstWithUnknownProfile(
2655 Cmp, ConstantInt::get(Ty, *C3), ConstantInt::getNullValue(Ty));
2656 }
2657 }
2658
2660 m_ImmConstant(C2)))) &&
2661 match(C1, m_Power2())) {
2663 Constant *Cmp =
2665 if (Cmp && Cmp->isNullValue()) {
2666 // iff C1,C3 is pow2 and Log2(C3) >= C2:
2667 // ((C1 >> X) << C2) & C3 -> X == (cttz(C1)+C2-cttz(C3)) ? C3 : 0
2668 Constant *ShlC = ConstantExpr::getAdd(C2, Log2C1);
2669 Constant *CmpC = ConstantExpr::getSub(ShlC, Log2C3);
2670 Value *Cmp = Builder.CreateICmpEQ(X, CmpC);
2671 return createSelectInstWithUnknownProfile(
2672 Cmp, ConstantInt::get(Ty, *C3), ConstantInt::getNullValue(Ty));
2673 }
2674 }
2675 }
2676 }
2677
2678 // If we are clearing the sign bit of a floating-point value, convert this to
2679 // fabs, then cast back to integer.
2680 //
2681 // This is a generous interpretation for noimplicitfloat, this is not a true
2682 // floating-point operation.
2683 //
2684 // Assumes any IEEE-represented type has the sign bit in the high bit.
2685 // TODO: Unify with APInt matcher. This version allows undef unlike m_APInt
2686 Value *CastOp;
2687 if (match(Op0, m_ElementWiseBitCast(m_Value(CastOp))) &&
2688 match(Op1, m_MaxSignedValue()) &&
2689 !Builder.GetInsertBlock()->getParent()->hasFnAttribute(
2690 Attribute::NoImplicitFloat)) {
2691 Type *EltTy = CastOp->getType()->getScalarType();
2692 if (EltTy->isFloatingPointTy() &&
2694 Value *FAbs = Builder.CreateFAbs(CastOp);
2695 return new BitCastInst(FAbs, I.getType());
2696 }
2697 }
2698
2699 // and(shl(zext(X), Y), SignMask) -> and(sext(X), SignMask)
2700 // where Y is a valid shift amount.
2702 m_SignMask())) &&
2705 APInt(Ty->getScalarSizeInBits(),
2706 Ty->getScalarSizeInBits() -
2707 X->getType()->getScalarSizeInBits())))) {
2708 auto *SExt = Builder.CreateSExt(X, Ty, X->getName() + ".signext");
2709 return BinaryOperator::CreateAnd(SExt, Op1);
2710 }
2711
2712 if (Instruction *Z = narrowMaskedBinOp(I))
2713 return Z;
2714
2715 if (I.getType()->isIntOrIntVectorTy(1)) {
2716 if (auto *SI0 = dyn_cast<SelectInst>(Op0)) {
2717 if (auto *R =
2718 foldAndOrOfSelectUsingImpliedCond(Op1, *SI0, /* IsAnd */ true))
2719 return R;
2720 }
2721 if (auto *SI1 = dyn_cast<SelectInst>(Op1)) {
2722 if (auto *R =
2723 foldAndOrOfSelectUsingImpliedCond(Op0, *SI1, /* IsAnd */ true))
2724 return R;
2725 }
2726 }
2727
2728 if (Instruction *FoldedLogic = foldBinOpIntoSelectOrPhi(I))
2729 return FoldedLogic;
2730
2731 if (Instruction *DeMorgan = matchDeMorgansLaws(I, *this))
2732 return DeMorgan;
2733
2734 {
2735 Value *A, *B, *C;
2736 // A & ~(A ^ B) --> A & B
2737 if (match(Op1, m_Not(m_c_Xor(m_Specific(Op0), m_Value(B)))))
2738 return BinaryOperator::CreateAnd(Op0, B);
2739 // ~(A ^ B) & A --> A & B
2740 if (match(Op0, m_Not(m_c_Xor(m_Specific(Op1), m_Value(B)))))
2741 return BinaryOperator::CreateAnd(Op1, B);
2742
2743 // (A ^ B) & ((B ^ C) ^ A) -> (A ^ B) & ~C
2744 if (match(Op0, m_Xor(m_Value(A), m_Value(B))) &&
2745 match(Op1, m_Xor(m_Xor(m_Specific(B), m_Value(C)), m_Specific(A)))) {
2746 Value *NotC = Op1->hasOneUse()
2747 ? Builder.CreateNot(C)
2748 : getFreelyInverted(C, C->hasOneUse(), &Builder);
2749 if (NotC != nullptr)
2750 return BinaryOperator::CreateAnd(Op0, NotC);
2751 }
2752
2753 // ((A ^ C) ^ B) & (B ^ A) -> (B ^ A) & ~C
2754 if (match(Op0, m_Xor(m_Xor(m_Value(A), m_Value(C)), m_Value(B))) &&
2755 match(Op1, m_Xor(m_Specific(B), m_Specific(A)))) {
2756 Value *NotC = Op0->hasOneUse()
2757 ? Builder.CreateNot(C)
2758 : getFreelyInverted(C, C->hasOneUse(), &Builder);
2759 if (NotC != nullptr)
2760 return BinaryOperator::CreateAnd(Op1, Builder.CreateNot(C));
2761 }
2762
2763 // (A | B) & (~A ^ B) -> A & B
2764 // (A | B) & (B ^ ~A) -> A & B
2765 // (B | A) & (~A ^ B) -> A & B
2766 // (B | A) & (B ^ ~A) -> A & B
2767 if (match(Op1, m_c_Xor(m_Not(m_Value(A)), m_Value(B))) &&
2768 match(Op0, m_c_Or(m_Specific(A), m_Specific(B))))
2769 return BinaryOperator::CreateAnd(A, B);
2770
2771 // (~A ^ B) & (A | B) -> A & B
2772 // (~A ^ B) & (B | A) -> A & B
2773 // (B ^ ~A) & (A | B) -> A & B
2774 // (B ^ ~A) & (B | A) -> A & B
2775 if (match(Op0, m_c_Xor(m_Not(m_Value(A)), m_Value(B))) &&
2776 match(Op1, m_c_Or(m_Specific(A), m_Specific(B))))
2777 return BinaryOperator::CreateAnd(A, B);
2778
2779 // (~A | B) & (A ^ B) -> ~A & B
2780 // (~A | B) & (B ^ A) -> ~A & B
2781 // (B | ~A) & (A ^ B) -> ~A & B
2782 // (B | ~A) & (B ^ A) -> ~A & B
2783 if (match(Op0, m_c_Or(m_Not(m_Value(A)), m_Value(B))) &&
2785 return BinaryOperator::CreateAnd(Builder.CreateNot(A), B);
2786
2787 // (A ^ B) & (~A | B) -> ~A & B
2788 // (B ^ A) & (~A | B) -> ~A & B
2789 // (A ^ B) & (B | ~A) -> ~A & B
2790 // (B ^ A) & (B | ~A) -> ~A & B
2791 if (match(Op1, m_c_Or(m_Not(m_Value(A)), m_Value(B))) &&
2793 return BinaryOperator::CreateAnd(Builder.CreateNot(A), B);
2794 }
2795
2796 if (Value *Res =
2797 foldBooleanAndOr(Op0, Op1, I, /*IsAnd=*/true, /*IsLogical=*/false))
2798 return replaceInstUsesWith(I, Res);
2799
2800 if (match(Op1, m_OneUse(m_LogicalAnd(m_Value(X), m_Value(Y))))) {
2801 bool IsLogical = isa<SelectInst>(Op1);
2802 if (auto *V = reassociateBooleanAndOr(Op0, X, Y, I, /*IsAnd=*/true,
2803 /*RHSIsLogical=*/IsLogical))
2804 return replaceInstUsesWith(I, V);
2805 }
2806 if (match(Op0, m_OneUse(m_LogicalAnd(m_Value(X), m_Value(Y))))) {
2807 bool IsLogical = isa<SelectInst>(Op0);
2808 if (auto *V = reassociateBooleanAndOr(Op1, X, Y, I, /*IsAnd=*/true,
2809 /*RHSIsLogical=*/IsLogical))
2810 return replaceInstUsesWith(I, V);
2811 }
2812
2813 if (Instruction *FoldedFCmps = reassociateFCmps(I, Builder))
2814 return FoldedFCmps;
2815
2816 if (Instruction *CastedAnd = foldCastedBitwiseLogic(I))
2817 return CastedAnd;
2818
2819 if (Instruction *Sel = foldBinopOfSextBoolToSelect(I))
2820 return Sel;
2821
2822 // and(sext(A), B) / and(B, sext(A)) --> A ? B : 0, where A is i1 or <N x i1>.
2823 // TODO: Move this into foldBinopOfSextBoolToSelect as a more generalized fold
2824 // with binop identity constant. But creating a select with non-constant
2825 // arm may not be reversible due to poison semantics. Is that a good
2826 // canonicalization?
2827 Value *A, *B;
2828 if (match(&I, m_c_And(m_SExt(m_Value(A)), m_Value(B))) &&
2829 A->getType()->isIntOrIntVectorTy(1))
2830 return createSelectInstWithUnknownProfile(A, B, Constant::getNullValue(Ty));
2831
2832 // Similarly, a 'not' of the bool translates to a swap of the select arms:
2833 // ~sext(A) & B / B & ~sext(A) --> A ? 0 : B
2834 if (match(&I, m_c_And(m_Not(m_SExt(m_Value(A))), m_Value(B))) &&
2835 A->getType()->isIntOrIntVectorTy(1))
2836 return createSelectInstWithUnknownProfile(A, Constant::getNullValue(Ty), B);
2837
2838 // and(zext(A), B) -> A ? (B & 1) : 0
2839 if (match(&I, m_c_And(m_OneUse(m_ZExt(m_Value(A))), m_Value(B))) &&
2840 A->getType()->isIntOrIntVectorTy(1))
2841 return createSelectInstWithUnknownProfile(
2842 A, Builder.CreateAnd(B, ConstantInt::get(Ty, 1)),
2844
2845 // (-1 + A) & B --> A ? 0 : B where A is 0/1.
2847 m_Value(B)))) {
2848 if (A->getType()->isIntOrIntVectorTy(1))
2849 return createSelectInstWithUnknownProfile(A, Constant::getNullValue(Ty),
2850 B);
2851 if (computeKnownBits(A, &I).countMaxActiveBits() <= 1) {
2852 return createSelectInstWithUnknownProfile(
2853 Builder.CreateICmpEQ(A, Constant::getNullValue(A->getType())), B,
2855 }
2856 }
2857
2858 // (iN X s>> (N-1)) & Y --> (X s< 0) ? Y : 0 -- with optional sext
2861 m_Value(Y))) &&
2862 *C == X->getType()->getScalarSizeInBits() - 1) {
2863 Value *IsNeg = Builder.CreateIsNeg(X, "isneg");
2864 return createSelectInstWithUnknownProfile(IsNeg, Y,
2866 }
2867 // If there's a 'not' of the shifted value, swap the select operands:
2868 // ~(iN X s>> (N-1)) & Y --> (X s< 0) ? 0 : Y -- with optional sext
2871 m_Value(Y))) &&
2872 *C == X->getType()->getScalarSizeInBits() - 1) {
2873 Value *IsNeg = Builder.CreateIsNeg(X, "isneg");
2874 return createSelectInstWithUnknownProfile(IsNeg,
2876 }
2877
2878 // (~x) & y --> ~(x | (~y)) iff that gets rid of inversions
2880 return &I;
2881
2882 // An and recurrence w/loop invariant step is equivelent to (and start, step)
2883 PHINode *PN = nullptr;
2884 Value *Start = nullptr, *Step = nullptr;
2885 if (matchSimpleRecurrence(&I, PN, Start, Step) && DT.dominates(Step, PN))
2886 return replaceInstUsesWith(I, Builder.CreateAnd(Start, Step));
2887
2889 return R;
2890
2891 if (Instruction *Canonicalized = canonicalizeLogicFirst(I, Builder))
2892 return Canonicalized;
2893
2894 if (Instruction *Folded = foldLogicOfIsFPClass(I, Op0, Op1))
2895 return Folded;
2896
2897 if (Instruction *Res = foldBinOpOfDisplacedShifts(I))
2898 return Res;
2899
2901 return Res;
2902
2903 if (Value *V =
2905 /*SimplifyOnly*/ false, *this))
2906 return BinaryOperator::CreateAnd(V, Op1);
2907 if (Value *V =
2909 /*SimplifyOnly*/ false, *this))
2910 return BinaryOperator::CreateAnd(Op0, V);
2911
2912 return nullptr;
2913}
2914
2916 bool MatchBSwaps,
2917 bool MatchBitReversals) {
2919 if (!recognizeBSwapOrBitReverseIdiom(&I, MatchBSwaps, MatchBitReversals,
2920 Insts))
2921 return nullptr;
2922 Instruction *LastInst = Insts.pop_back_val();
2923 LastInst->removeFromParent();
2924
2925 for (auto *Inst : Insts) {
2926 Inst->setDebugLoc(I.getDebugLoc());
2927 Worklist.push(Inst);
2928 }
2929 return LastInst;
2930}
2931
2932std::optional<std::pair<Intrinsic::ID, SmallVector<Value *, 3>>>
2934 // TODO: Can we reduce the code duplication between this and the related
2935 // rotate matching code under visitSelect and visitTrunc?
2936 assert(Or.getOpcode() == BinaryOperator::Or && "Expecting or instruction");
2937
2938 unsigned Width = Or.getType()->getScalarSizeInBits();
2939
2940 Instruction *Or0, *Or1;
2941 if (!match(Or.getOperand(0), m_Instruction(Or0)) ||
2942 !match(Or.getOperand(1), m_Instruction(Or1)))
2943 return std::nullopt;
2944
2945 bool IsFshl = true; // Sub on LSHR.
2946 SmallVector<Value *, 3> FShiftArgs;
2947
2948 // First, find an or'd pair of opposite shifts:
2949 // or (lshr ShVal0, ShAmt0), (shl ShVal1, ShAmt1)
2950 if (isa<BinaryOperator>(Or0) && isa<BinaryOperator>(Or1)) {
2951 Value *ShVal0, *ShVal1, *ShAmt0, *ShAmt1;
2952 if (!match(Or0,
2953 m_OneUse(m_LogicalShift(m_Value(ShVal0), m_Value(ShAmt0)))) ||
2954 !match(Or1,
2955 m_OneUse(m_LogicalShift(m_Value(ShVal1), m_Value(ShAmt1)))) ||
2956 Or0->getOpcode() == Or1->getOpcode())
2957 return std::nullopt;
2958
2959 // Canonicalize to or(shl(ShVal0, ShAmt0), lshr(ShVal1, ShAmt1)).
2960 if (Or0->getOpcode() == BinaryOperator::LShr) {
2961 std::swap(Or0, Or1);
2962 std::swap(ShVal0, ShVal1);
2963 std::swap(ShAmt0, ShAmt1);
2964 }
2965 assert(Or0->getOpcode() == BinaryOperator::Shl &&
2966 Or1->getOpcode() == BinaryOperator::LShr &&
2967 "Illegal or(shift,shift) pair");
2968
2969 // Match the shift amount operands for a funnel shift pattern. This always
2970 // matches a subtraction on the R operand.
2971 auto matchShiftAmount = [&](Value *L, Value *R, unsigned Width) -> Value * {
2972 // Check for constant shift amounts that sum to the bitwidth.
2973 const APInt *LI, *RI;
2974 if (match(L, m_APIntAllowPoison(LI)) && match(R, m_APIntAllowPoison(RI)))
2975 if (LI->ult(Width) && RI->ult(Width) && (*LI + *RI) == Width)
2976 return ConstantInt::get(L->getType(), *LI);
2977
2978 Constant *LC, *RC;
2979 if (match(L, m_Constant(LC)) && match(R, m_Constant(RC)) &&
2980 match(L,
2981 m_SpecificInt_ICMP(ICmpInst::ICMP_ULT, APInt(Width, Width))) &&
2982 match(R,
2983 m_SpecificInt_ICMP(ICmpInst::ICMP_ULT, APInt(Width, Width))) &&
2985 return ConstantExpr::mergeUndefsWith(LC, RC);
2986
2987 // (shl ShVal, X) | (lshr ShVal, (Width - x)) iff X < Width.
2988 // We limit this to X < Width in case the backend re-expands the
2989 // intrinsic, and has to reintroduce a shift modulo operation (InstCombine
2990 // might remove it after this fold). This still doesn't guarantee that the
2991 // final codegen will match this original pattern.
2992 if (match(R, m_OneUse(m_Sub(m_SpecificInt(Width), m_Specific(L))))) {
2993 KnownBits KnownL = computeKnownBits(L, &Or);
2994 return KnownL.getMaxValue().ult(Width) ? L : nullptr;
2995 }
2996
2997 // For non-constant cases, the following patterns currently only work for
2998 // rotation patterns.
2999 // TODO: Add general funnel-shift compatible patterns.
3000 if (ShVal0 != ShVal1)
3001 return nullptr;
3002
3003 // For non-constant cases we don't support non-pow2 shift masks.
3004 // TODO: Is it worth matching urem as well?
3005 if (!isPowerOf2_32(Width))
3006 return nullptr;
3007
3008 // The shift amount may be masked with negation:
3009 // (shl ShVal, (X & (Width - 1))) | (lshr ShVal, ((-X) & (Width - 1)))
3010 Value *X;
3011 unsigned Mask = Width - 1;
3012 if (match(L, m_And(m_Value(X), m_SpecificInt(Mask))) &&
3013 match(R, m_And(m_Neg(m_Specific(X)), m_SpecificInt(Mask))))
3014 return X;
3015
3016 // (shl ShVal, X) | (lshr ShVal, ((-X) & (Width - 1)))
3017 if (match(R, m_And(m_Neg(m_Specific(L)), m_SpecificInt(Mask))))
3018 return L;
3019
3020 // Similar to above, but the shift amount may be extended after masking,
3021 // so return the extended value as the parameter for the intrinsic.
3022 if (match(L, m_ZExt(m_And(m_Value(X), m_SpecificInt(Mask)))) &&
3023 match(R,
3025 m_SpecificInt(Mask))))
3026 return L;
3027
3028 if (match(L, m_ZExt(m_And(m_Value(X), m_SpecificInt(Mask)))) &&
3030 return L;
3031
3032 return nullptr;
3033 };
3034
3035 Value *ShAmt = matchShiftAmount(ShAmt0, ShAmt1, Width);
3036 if (!ShAmt) {
3037 ShAmt = matchShiftAmount(ShAmt1, ShAmt0, Width);
3038 IsFshl = false; // Sub on SHL.
3039 }
3040 if (!ShAmt)
3041 return std::nullopt;
3042
3043 FShiftArgs = {ShVal0, ShVal1, ShAmt};
3044 } else if (isa<ZExtInst>(Or0) || isa<ZExtInst>(Or1)) {
3045 // If there are two 'or' instructions concat variables in opposite order:
3046 //
3047 // Slot1 and Slot2 are all zero bits.
3048 // | Slot1 | Low | Slot2 | High |
3049 // LowHigh = or (shl (zext Low), ZextLowShlAmt), (zext High)
3050 // | Slot2 | High | Slot1 | Low |
3051 // HighLow = or (shl (zext High), ZextHighShlAmt), (zext Low)
3052 //
3053 // the latter 'or' can be safely convert to
3054 // -> HighLow = fshl LowHigh, LowHigh, ZextHighShlAmt
3055 // if ZextLowShlAmt + ZextHighShlAmt == Width.
3056 if (!isa<ZExtInst>(Or1))
3057 std::swap(Or0, Or1);
3058
3059 Value *High, *ZextHigh, *Low;
3060 const APInt *ZextHighShlAmt;
3061 if (!match(Or0,
3062 m_OneUse(m_Shl(m_Value(ZextHigh), m_APInt(ZextHighShlAmt)))))
3063 return std::nullopt;
3064
3065 if (!match(Or1, m_ZExt(m_Value(Low))) ||
3066 !match(ZextHigh, m_ZExt(m_Value(High))))
3067 return std::nullopt;
3068
3069 unsigned HighSize = High->getType()->getScalarSizeInBits();
3070 unsigned LowSize = Low->getType()->getScalarSizeInBits();
3071 // Make sure High does not overlap with Low and most significant bits of
3072 // High aren't shifted out.
3073 if (ZextHighShlAmt->ult(LowSize) || ZextHighShlAmt->ugt(Width - HighSize))
3074 return std::nullopt;
3075
3076 for (User *U : ZextHigh->users()) {
3077 Value *X, *Y;
3078 if (!match(U, m_Or(m_Value(X), m_Value(Y))))
3079 continue;
3080
3081 if (!isa<ZExtInst>(Y))
3082 std::swap(X, Y);
3083
3084 const APInt *ZextLowShlAmt;
3085 if (!match(X, m_Shl(m_Specific(Or1), m_APInt(ZextLowShlAmt))) ||
3086 !match(Y, m_Specific(ZextHigh)) || !DT.dominates(U, &Or))
3087 continue;
3088
3089 // HighLow is good concat. If sum of two shifts amount equals to Width,
3090 // LowHigh must also be a good concat.
3091 if (*ZextLowShlAmt + *ZextHighShlAmt != Width)
3092 continue;
3093
3094 // Low must not overlap with High and most significant bits of Low must
3095 // not be shifted out.
3096 assert(ZextLowShlAmt->uge(HighSize) &&
3097 ZextLowShlAmt->ule(Width - LowSize) && "Invalid concat");
3098
3099 // We cannot reuse the result if it may produce poison.
3100 // Drop poison generating flags in the expression tree.
3101 // Or
3102 cast<Instruction>(U)->dropPoisonGeneratingFlags();
3103 // Shl
3104 cast<Instruction>(X)->dropPoisonGeneratingFlags();
3105
3106 FShiftArgs = {U, U, ConstantInt::get(Or0->getType(), *ZextHighShlAmt)};
3107 break;
3108 }
3109 }
3110
3111 if (FShiftArgs.empty())
3112 return std::nullopt;
3113
3114 Intrinsic::ID IID = IsFshl ? Intrinsic::fshl : Intrinsic::fshr;
3115 return std::make_pair(IID, FShiftArgs);
3116}
3117
3118/// Match UB-safe variants of the funnel shift intrinsic.
3120 if (auto Opt = IC.convertOrOfShiftsToFunnelShift(Or)) {
3121 auto [IID, FShiftArgs] = *Opt;
3122 Function *F =
3123 Intrinsic::getOrInsertDeclaration(Or.getModule(), IID, Or.getType());
3124 return CallInst::Create(F, FShiftArgs);
3125 }
3126
3127 return nullptr;
3128}
3129
3130/// Attempt to combine or(zext(x),shl(zext(y),bw/2) concat packing patterns.
3132 assert(Or.getOpcode() == Instruction::Or && "bswap requires an 'or'");
3133 Value *Op0 = Or.getOperand(0), *Op1 = Or.getOperand(1);
3134 Type *Ty = Or.getType();
3135
3136 unsigned Width = Ty->getScalarSizeInBits();
3137 if ((Width & 1) != 0)
3138 return nullptr;
3139 unsigned HalfWidth = Width / 2;
3140
3141 // Canonicalize zext (lower half) to LHS.
3142 if (!isa<ZExtInst>(Op0))
3143 std::swap(Op0, Op1);
3144
3145 // Find lower/upper half.
3146 Value *LowerSrc, *ShlVal, *UpperSrc;
3147 const APInt *C;
3148 if (!match(Op0, m_OneUse(m_ZExt(m_Value(LowerSrc)))) ||
3149 !match(Op1, m_OneUse(m_Shl(m_Value(ShlVal), m_APInt(C)))) ||
3150 !match(ShlVal, m_OneUse(m_ZExt(m_Value(UpperSrc)))))
3151 return nullptr;
3152 if (*C != HalfWidth || LowerSrc->getType() != UpperSrc->getType() ||
3153 LowerSrc->getType()->getScalarSizeInBits() != HalfWidth)
3154 return nullptr;
3155
3156 auto ConcatIntrinsicCalls = [&](Intrinsic::ID id, Value *Lo, Value *Hi) {
3157 Value *NewLower = Builder.CreateZExt(Lo, Ty);
3158 Value *NewUpper = Builder.CreateZExt(Hi, Ty);
3159 NewUpper = Builder.CreateShl(NewUpper, HalfWidth);
3160 Value *BinOp = Builder.CreateOr(NewLower, NewUpper);
3161 return Builder.CreateIntrinsic(id, Ty, BinOp);
3162 };
3163
3164 // BSWAP: Push the concat down, swapping the lower/upper sources.
3165 // concat(bswap(x),bswap(y)) -> bswap(concat(x,y))
3166 Value *LowerBSwap, *UpperBSwap;
3167 if (match(LowerSrc, m_BSwap(m_Value(LowerBSwap))) &&
3168 match(UpperSrc, m_BSwap(m_Value(UpperBSwap))))
3169 return ConcatIntrinsicCalls(Intrinsic::bswap, UpperBSwap, LowerBSwap);
3170
3171 // BITREVERSE: Push the concat down, swapping the lower/upper sources.
3172 // concat(bitreverse(x),bitreverse(y)) -> bitreverse(concat(x,y))
3173 Value *LowerBRev, *UpperBRev;
3174 if (match(LowerSrc, m_BitReverse(m_Value(LowerBRev))) &&
3175 match(UpperSrc, m_BitReverse(m_Value(UpperBRev))))
3176 return ConcatIntrinsicCalls(Intrinsic::bitreverse, UpperBRev, LowerBRev);
3177
3178 // iX ext split: extending or(zext(x),shl(zext(y),bw/2) pattern
3179 // to consume sext/ashr:
3180 // or(zext(sext(x)),shl(zext(sext(ashr(x,xbw-1))),bw/2)
3181 // or(zext(x),shl(zext(ashr(x,xbw-1)),bw/2)
3182 Value *X;
3183 if (match(LowerSrc, m_SExtOrSelf(m_Value(X))) &&
3184 match(UpperSrc,
3186 m_Specific(X),
3187 m_SpecificInt(X->getType()->getScalarSizeInBits() - 1)))))
3188 return Builder.CreateSExt(X, Ty);
3189
3190 return nullptr;
3191}
3192
3193/// If all elements of two constant vectors are 0/-1 and inverses, return true.
3195 unsigned NumElts = cast<FixedVectorType>(C1->getType())->getNumElements();
3196 for (unsigned i = 0; i != NumElts; ++i) {
3197 Constant *EltC1 = C1->getAggregateElement(i);
3198 Constant *EltC2 = C2->getAggregateElement(i);
3199 if (!EltC1 || !EltC2)
3200 return false;
3201
3202 // One element must be all ones, and the other must be all zeros.
3203 if (!((match(EltC1, m_Zero()) && match(EltC2, m_AllOnes())) ||
3204 (match(EltC2, m_Zero()) && match(EltC1, m_AllOnes()))))
3205 return false;
3206 }
3207 return true;
3208}
3209
3210/// We have an expression of the form (A & C) | (B & D). If A is a scalar or
3211/// vector composed of all-zeros or all-ones values and is the bitwise 'not' of
3212/// B, it can be used as the condition operand of a select instruction.
3213/// We will detect (A & C) | ~(B | D) when the flag ABIsTheSame enabled.
3214Value *InstCombinerImpl::getSelectCondition(Value *A, Value *B,
3215 bool ABIsTheSame) {
3216 // We may have peeked through bitcasts in the caller.
3217 // Exit immediately if we don't have (vector) integer types.
3218 Type *Ty = A->getType();
3219 if (!Ty->isIntOrIntVectorTy() || !B->getType()->isIntOrIntVectorTy())
3220 return nullptr;
3221
3222 // If A is the 'not' operand of B and has enough signbits, we have our answer.
3223 if (ABIsTheSame ? (A == B) : match(B, m_Not(m_Specific(A)))) {
3224 // If these are scalars or vectors of i1, A can be used directly.
3225 if (Ty->isIntOrIntVectorTy(1))
3226 return A;
3227
3228 // If we look through a vector bitcast, the caller will bitcast the operands
3229 // to match the condition's number of bits (N x i1).
3230 // To make this poison-safe, disallow bitcast from wide element to narrow
3231 // element. That could allow poison in lanes where it was not present in the
3232 // original code.
3234 if (A->getType()->isIntOrIntVectorTy()) {
3235 unsigned NumSignBits = ComputeNumSignBits(A);
3236 if (NumSignBits == A->getType()->getScalarSizeInBits() &&
3237 NumSignBits <= Ty->getScalarSizeInBits())
3238 return Builder.CreateTrunc(A, CmpInst::makeCmpResultType(A->getType()));
3239 }
3240 return nullptr;
3241 }
3242
3243 // TODO: add support for sext and constant case
3244 if (ABIsTheSame)
3245 return nullptr;
3246
3247 // If both operands are constants, see if the constants are inverse bitmasks.
3248 Constant *AConst, *BConst;
3249 if (match(A, m_Constant(AConst)) && match(B, m_Constant(BConst)))
3250 if (AConst == ConstantExpr::getNot(BConst) &&
3252 return Builder.CreateZExtOrTrunc(A, CmpInst::makeCmpResultType(Ty));
3253
3254 // Look for more complex patterns. The 'not' op may be hidden behind various
3255 // casts. Look through sexts and bitcasts to find the booleans.
3256 Value *Cond;
3257 Value *NotB;
3258 if (match(A, m_SExt(m_Value(Cond))) &&
3259 Cond->getType()->isIntOrIntVectorTy(1)) {
3260 // A = sext i1 Cond; B = sext (not (i1 Cond))
3261 if (match(B, m_SExt(m_Not(m_Specific(Cond)))))
3262 return Cond;
3263
3264 // A = sext i1 Cond; B = not ({bitcast} (sext (i1 Cond)))
3265 // TODO: The one-use checks are unnecessary or misplaced. If the caller
3266 // checked for uses on logic ops/casts, that should be enough to
3267 // make this transform worthwhile.
3268 if (match(B, m_OneUse(m_Not(m_Value(NotB))))) {
3269 NotB = peekThroughBitcast(NotB, true);
3270 if (match(NotB, m_SExt(m_Specific(Cond))))
3271 return Cond;
3272 }
3273 }
3274
3275 // All scalar (and most vector) possibilities should be handled now.
3276 // Try more matches that only apply to non-splat constant vectors.
3277 if (!Ty->isVectorTy())
3278 return nullptr;
3279
3280 // If both operands are xor'd with constants using the same sexted boolean
3281 // operand, see if the constants are inverse bitmasks.
3282 // TODO: Use ConstantExpr::getNot()?
3283 if (match(A, (m_Xor(m_SExt(m_Value(Cond)), m_Constant(AConst)))) &&
3284 match(B, (m_Xor(m_SExt(m_Specific(Cond)), m_Constant(BConst)))) &&
3285 Cond->getType()->isIntOrIntVectorTy(1) &&
3286 areInverseVectorBitmasks(AConst, BConst)) {
3288 return Builder.CreateXor(Cond, AConst);
3289 }
3290 return nullptr;
3291}
3292
3293/// We have an expression of the form (A & B) | (C & D). Try to simplify this
3294/// to "A' ? B : D", where A' is a boolean or vector of booleans.
3295/// When InvertFalseVal is set to true, we try to match the pattern
3296/// where we have peeked through a 'not' op and A and C are the same:
3297/// (A & B) | ~(A | D) --> (A & B) | (~A & ~D) --> A' ? B : ~D
3298Value *InstCombinerImpl::matchSelectFromAndOr(Value *A, Value *B, Value *C,
3299 Value *D, bool InvertFalseVal) {
3300 // The potential condition of the select may be bitcasted. In that case, look
3301 // through its bitcast and the corresponding bitcast of the 'not' condition.
3302 Type *OrigType = A->getType();
3303 A = peekThroughBitcast(A, true);
3304 C = peekThroughBitcast(C, true);
3305 if (Value *Cond = getSelectCondition(A, C, InvertFalseVal)) {
3306 // ((bc Cond) & B) | ((bc ~Cond) & D) --> bc (select Cond, (bc B), (bc D))
3307 // If this is a vector, we may need to cast to match the condition's length.
3308 // The bitcasts will either all exist or all not exist. The builder will
3309 // not create unnecessary casts if the types already match.
3310 Type *SelTy = A->getType();
3311 if (auto *VecTy = dyn_cast<VectorType>(Cond->getType())) {
3312 // For a fixed or scalable vector get N from <{vscale x} N x iM>
3313 unsigned Elts = VecTy->getElementCount().getKnownMinValue();
3314 // For a fixed or scalable vector, get the size in bits of N x iM; for a
3315 // scalar this is just M.
3316 unsigned SelEltSize = SelTy->getPrimitiveSizeInBits().getKnownMinValue();
3317 Type *EltTy = Builder.getIntNTy(SelEltSize / Elts);
3318 SelTy = VectorType::get(EltTy, VecTy->getElementCount());
3319 }
3320 Value *BitcastB = Builder.CreateBitCast(B, SelTy);
3321 if (InvertFalseVal)
3322 D = Builder.CreateNot(D);
3323 Value *BitcastD = Builder.CreateBitCast(D, SelTy);
3324 Value *Select = Builder.CreateSelect(Cond, BitcastB, BitcastD);
3325 return Builder.CreateBitCast(Select, OrigType);
3326 }
3327
3328 return nullptr;
3329}
3330
3331// (icmp eq X, C) | (icmp ult Other, (X - C)) -> (icmp ule Other, (X - (C + 1)))
3332// (icmp ne X, C) & (icmp uge Other, (X - C)) -> (icmp ugt Other, (X - (C + 1)))
3334 bool IsAnd, bool IsLogical,
3335 IRBuilderBase &Builder) {
3336 Value *LHS0 = LHS->getOperand(0);
3337 Value *RHS0 = RHS->getOperand(0);
3338 Value *RHS1 = RHS->getOperand(1);
3339
3340 ICmpInst::Predicate LPred =
3341 IsAnd ? LHS->getInversePredicate() : LHS->getPredicate();
3342 ICmpInst::Predicate RPred =
3343 IsAnd ? RHS->getInversePredicate() : RHS->getPredicate();
3344
3345 const APInt *CInt;
3346 if (LPred != ICmpInst::ICMP_EQ ||
3347 !match(LHS->getOperand(1), m_APIntAllowPoison(CInt)) ||
3348 !LHS0->getType()->isIntOrIntVectorTy() ||
3349 !(LHS->hasOneUse() || RHS->hasOneUse()))
3350 return nullptr;
3351
3352 auto MatchRHSOp = [LHS0, CInt](const Value *RHSOp) {
3353 return match(RHSOp,
3354 m_Add(m_Specific(LHS0), m_SpecificIntAllowPoison(-*CInt))) ||
3355 (CInt->isZero() && RHSOp == LHS0);
3356 };
3357
3358 Value *Other;
3359 if (RPred == ICmpInst::ICMP_ULT && MatchRHSOp(RHS1))
3360 Other = RHS0;
3361 else if (RPred == ICmpInst::ICMP_UGT && MatchRHSOp(RHS0))
3362 Other = RHS1;
3363 else
3364 return nullptr;
3365
3366 if (IsLogical)
3367 Other = Builder.CreateFreeze(Other);
3368
3369 return Builder.CreateICmp(
3371 Builder.CreateSub(LHS0, ConstantInt::get(LHS0->getType(), *CInt + 1)),
3372 Other);
3373}
3374
3375/// Fold (icmp)&(icmp) or (icmp)|(icmp) if possible.
3376/// If IsLogical is true, then the and/or is in select form and the transform
3377/// must be poison-safe.
3378Value *InstCombinerImpl::foldAndOrOfICmps(ICmpInst *LHS, ICmpInst *RHS,
3379 Instruction &I, bool IsAnd,
3380 bool IsLogical) {
3381 const SimplifyQuery Q = SQ.getWithInstruction(&I);
3382
3383 ICmpInst::Predicate PredL = LHS->getPredicate(), PredR = RHS->getPredicate();
3384 Value *LHS0 = LHS->getOperand(0), *RHS0 = RHS->getOperand(0);
3385 Value *LHS1 = LHS->getOperand(1), *RHS1 = RHS->getOperand(1);
3386
3387 const APInt *LHSC = nullptr, *RHSC = nullptr;
3388 match(LHS1, m_APInt(LHSC));
3389 match(RHS1, m_APInt(RHSC));
3390
3391 // (icmp1 A, B) | (icmp2 A, B) --> (icmp3 A, B)
3392 // (icmp1 A, B) & (icmp2 A, B) --> (icmp3 A, B)
3393 if (predicatesFoldable(PredL, PredR)) {
3394 if (LHS0 == RHS1 && LHS1 == RHS0) {
3395 PredL = ICmpInst::getSwappedPredicate(PredL);
3396 std::swap(LHS0, LHS1);
3397 }
3398 if (LHS0 == RHS0 && LHS1 == RHS1) {
3399 unsigned Code = IsAnd ? getICmpCode(PredL) & getICmpCode(PredR)
3400 : getICmpCode(PredL) | getICmpCode(PredR);
3401 bool IsSigned = LHS->isSigned() || RHS->isSigned();
3402 return getNewICmpValue(Code, IsSigned, LHS0, LHS1, Builder);
3403 }
3404 }
3405
3406 if (Value *V =
3407 foldAndOrOfICmpEqConstantAndICmp(LHS, RHS, IsAnd, IsLogical, Builder))
3408 return V;
3409 // We can treat logical like bitwise here, because both operands are used on
3410 // the LHS, and as such poison from both will propagate.
3412 /*IsLogical*/ false, Builder))
3413 return V;
3414
3415 if (Value *V = foldAndOrOfICmpsWithConstEq(LHS, RHS, IsAnd, IsLogical,
3416 Builder, Q, I))
3417 return V;
3418 // We can convert this case to bitwise and, because both operands are used
3419 // on the LHS, and as such poison from both will propagate.
3421 RHS, LHS, IsAnd, /*IsLogical=*/false, Builder, Q, I)) {
3422 // If RHS is still used, we should drop samesign flag.
3423 if (IsLogical && RHS->hasSameSign() && !RHS->use_empty()) {
3424 RHS->setSameSign(false);
3426 }
3427 return V;
3428 }
3429
3430 if (Value *V = foldIsPowerOf2OrZero(LHS, RHS, IsAnd, Builder, *this))
3431 return V;
3432 if (Value *V = foldIsPowerOf2OrZero(RHS, LHS, IsAnd, Builder, *this))
3433 return V;
3434
3435 // TODO: One of these directions is fine with logical and/or, the other could
3436 // be supported by inserting freeze.
3437 if (!IsLogical) {
3438 // E.g. (icmp slt x, 0) | (icmp sgt x, n) --> icmp ugt x, n
3439 // E.g. (icmp sge x, 0) & (icmp slt x, n) --> icmp ult x, n
3440 if (Value *V = simplifyRangeCheck(LHS, RHS, /*Inverted=*/!IsAnd))
3441 return V;
3442
3443 // E.g. (icmp sgt x, n) | (icmp slt x, 0) --> icmp ugt x, n
3444 // E.g. (icmp slt x, n) & (icmp sge x, 0) --> icmp ult x, n
3445 if (Value *V = simplifyRangeCheck(RHS, LHS, /*Inverted=*/!IsAnd))
3446 return V;
3447 }
3448
3449 // TODO: Add conjugated or fold, check whether it is safe for logical and/or.
3450 if (IsAnd && !IsLogical)
3452 return V;
3453
3454 if (Value *V = foldIsPowerOf2(LHS, RHS, IsAnd, Builder, *this))
3455 return V;
3456
3457 if (Value *V = foldPowerOf2AndShiftedMask(LHS, RHS, IsAnd, Builder))
3458 return V;
3459
3460 // TODO: Verify whether this is safe for logical and/or.
3461 if (!IsLogical) {
3462 if (Value *X = foldUnsignedUnderflowCheck(LHS, RHS, IsAnd, Q, Builder))
3463 return X;
3464 if (Value *X = foldUnsignedUnderflowCheck(RHS, LHS, IsAnd, Q, Builder))
3465 return X;
3466 }
3467
3468 // (icmp ne A, 0) | (icmp ne B, 0) --> (icmp ne (A|B), 0)
3469 // (icmp eq A, 0) & (icmp eq B, 0) --> (icmp eq (A|B), 0)
3470 // TODO: Remove this and below when foldLogOpOfMaskedICmps can handle undefs.
3471 if (PredL == (IsAnd ? ICmpInst::ICMP_EQ : ICmpInst::ICMP_NE) &&
3472 PredL == PredR && match(LHS1, m_ZeroInt()) && match(RHS1, m_ZeroInt()) &&
3473 LHS0->getType() == RHS0->getType() &&
3474 (!IsLogical || isGuaranteedNotToBePoison(RHS0))) {
3475 Value *NewOr = Builder.CreateOr(LHS0, RHS0);
3476 return Builder.CreateICmp(PredL, NewOr,
3478 }
3479
3480 // (icmp ne A, -1) | (icmp ne B, -1) --> (icmp ne (A&B), -1)
3481 // (icmp eq A, -1) & (icmp eq B, -1) --> (icmp eq (A&B), -1)
3482 if (PredL == (IsAnd ? ICmpInst::ICMP_EQ : ICmpInst::ICMP_NE) &&
3483 PredL == PredR && match(LHS1, m_AllOnes()) && match(RHS1, m_AllOnes()) &&
3484 LHS0->getType() == RHS0->getType() &&
3485 (!IsLogical || isGuaranteedNotToBePoison(RHS0))) {
3486 Value *NewAnd = Builder.CreateAnd(LHS0, RHS0);
3487 return Builder.CreateICmp(PredL, NewAnd,
3489 }
3490
3491 if (!IsLogical)
3492 if (Value *V =
3494 return V;
3495
3496 // This only handles icmp of constants: (icmp1 A, C1) | (icmp2 B, C2).
3497 if (!LHSC || !RHSC)
3498 return nullptr;
3499
3500 // (trunc x) == C1 & (and x, CA) == C2 -> (and x, CA|CMAX) == C1|C2
3501 // (trunc x) != C1 | (and x, CA) != C2 -> (and x, CA|CMAX) != C1|C2
3502 // where CMAX is the all ones value for the truncated type,
3503 // iff the lower bits of C2 and CA are zero.
3504 if (PredL == (IsAnd ? ICmpInst::ICMP_EQ : ICmpInst::ICMP_NE) &&
3505 PredL == PredR && LHS->hasOneUse() && RHS->hasOneUse()) {
3506 Value *V;
3507 const APInt *AndC, *SmallC = nullptr, *BigC = nullptr;
3508
3509 // (trunc x) == C1 & (and x, CA) == C2
3510 // (and x, CA) == C2 & (trunc x) == C1
3511 if (match(RHS0, m_Trunc(m_Value(V))) &&
3512 match(LHS0, m_And(m_Specific(V), m_APInt(AndC)))) {
3513 SmallC = RHSC;
3514 BigC = LHSC;
3515 } else if (match(LHS0, m_Trunc(m_Value(V))) &&
3516 match(RHS0, m_And(m_Specific(V), m_APInt(AndC)))) {
3517 SmallC = LHSC;
3518 BigC = RHSC;
3519 }
3520
3521 if (SmallC && BigC) {
3522 unsigned BigBitSize = BigC->getBitWidth();
3523 unsigned SmallBitSize = SmallC->getBitWidth();
3524
3525 // Check that the low bits are zero.
3526 APInt Low = APInt::getLowBitsSet(BigBitSize, SmallBitSize);
3527 if ((Low & *AndC).isZero() && (Low & *BigC).isZero()) {
3528 Value *NewAnd = Builder.CreateAnd(V, Low | *AndC);
3529 APInt N = SmallC->zext(BigBitSize) | *BigC;
3530 Value *NewVal = ConstantInt::get(NewAnd->getType(), N);
3531 return Builder.CreateICmp(PredL, NewAnd, NewVal);
3532 }
3533 }
3534 }
3535
3536 // Match naive pattern (and its inverted form) for checking if two values
3537 // share same sign. An example of the pattern:
3538 // (icmp slt (X & Y), 0) | (icmp sgt (X | Y), -1) -> (icmp sgt (X ^ Y), -1)
3539 // Inverted form (example):
3540 // (icmp slt (X | Y), 0) & (icmp sgt (X & Y), -1) -> (icmp slt (X ^ Y), 0)
3541 bool TrueIfSignedL, TrueIfSignedR;
3542 if (isSignBitCheck(PredL, *LHSC, TrueIfSignedL) &&
3543 isSignBitCheck(PredR, *RHSC, TrueIfSignedR) &&
3544 (RHS->hasOneUse() || LHS->hasOneUse())) {
3545 Value *X, *Y;
3546 if (IsAnd) {
3547 if ((TrueIfSignedL && !TrueIfSignedR &&
3548 match(LHS0, m_Or(m_Value(X), m_Value(Y))) &&
3549 match(RHS0, m_c_And(m_Specific(X), m_Specific(Y)))) ||
3550 (!TrueIfSignedL && TrueIfSignedR &&
3551 match(LHS0, m_And(m_Value(X), m_Value(Y))) &&
3552 match(RHS0, m_c_Or(m_Specific(X), m_Specific(Y))))) {
3553 Value *NewXor = Builder.CreateXor(X, Y);
3554 return Builder.CreateIsNeg(NewXor);
3555 }
3556 } else {
3557 if ((TrueIfSignedL && !TrueIfSignedR &&
3558 match(LHS0, m_And(m_Value(X), m_Value(Y))) &&
3559 match(RHS0, m_c_Or(m_Specific(X), m_Specific(Y)))) ||
3560 (!TrueIfSignedL && TrueIfSignedR &&
3561 match(LHS0, m_Or(m_Value(X), m_Value(Y))) &&
3562 match(RHS0, m_c_And(m_Specific(X), m_Specific(Y))))) {
3563 Value *NewXor = Builder.CreateXor(X, Y);
3564 return Builder.CreateIsNotNeg(NewXor);
3565 }
3566 }
3567 }
3568
3569 // (X & ExpMask) != 0 && (X & ExpMask) != ExpMask -> isnormal(X)
3570 // (X & ExpMask) == 0 || (X & ExpMask) == ExpMask -> !isnormal(X)
3571 Value *X;
3572 const APInt *MaskC;
3573 if (LHS0 == RHS0 && PredL == PredR &&
3574 PredL == (IsAnd ? ICmpInst::ICMP_NE : ICmpInst::ICMP_EQ) &&
3575 !I.getFunction()->hasFnAttribute(Attribute::NoImplicitFloat) &&
3576 LHS->hasOneUse() && RHS->hasOneUse() &&
3577 match(LHS0, m_And(m_ElementWiseBitCast(m_Value(X)), m_APInt(MaskC))) &&
3578 X->getType()->getScalarType()->isIEEELikeFPTy() &&
3579 APFloat(X->getType()->getScalarType()->getFltSemantics(), *MaskC)
3580 .isPosInfinity() &&
3581 ((LHSC->isZero() && *RHSC == *MaskC) ||
3582 (RHSC->isZero() && *LHSC == *MaskC)))
3583 return Builder.createIsFPClass(X, IsAnd ? FPClassTest::fcNormal
3585
3586 return foldAndOrOfICmpsUsingRanges(LHS, RHS, IsAnd);
3587}
3588
3589/// If IsLogical is true, then the and/or is in select form and the transform
3590/// must be poison-safe.
3591Value *InstCombinerImpl::foldBooleanAndOr(Value *LHS, Value *RHS,
3592 Instruction &I, bool IsAnd,
3593 bool IsLogical) {
3594 if (!LHS->getType()->isIntOrIntVectorTy(1))
3595 return nullptr;
3596
3597 // handle (roughly):
3598 // (icmp ne (A & B), C) | (icmp ne (A & D), E)
3599 // (icmp eq (A & B), C) & (icmp eq (A & D), E)
3600 if (Value *V = foldLogOpOfMaskedICmps(LHS, RHS, IsAnd, IsLogical, Builder,
3601 SQ.getWithInstruction(&I)))
3602 return V;
3603
3604 if (auto *LHSCmp = dyn_cast<ICmpInst>(LHS))
3605 if (auto *RHSCmp = dyn_cast<ICmpInst>(RHS))
3606 if (Value *Res = foldAndOrOfICmps(LHSCmp, RHSCmp, I, IsAnd, IsLogical))
3607 return Res;
3608
3609 if (auto *LHSCmp = dyn_cast<FCmpInst>(LHS))
3610 if (auto *RHSCmp = dyn_cast<FCmpInst>(RHS))
3611 if (Value *Res = foldLogicOfFCmps(LHSCmp, RHSCmp, IsAnd, IsLogical))
3612 return Res;
3613
3614 if (Value *Res = foldEqOfParts(LHS, RHS, IsAnd))
3615 return Res;
3616
3617 return nullptr;
3618}
3619
3621 InstCombiner::BuilderTy &Builder) {
3622 assert(I.getOpcode() == Instruction::Or &&
3623 "Simplification only supports or at the moment.");
3624
3625 Value *Cmp1, *Cmp2, *Cmp3, *Cmp4;
3626 if (!match(I.getOperand(0), m_And(m_Value(Cmp1), m_Value(Cmp2))) ||
3627 !match(I.getOperand(1), m_And(m_Value(Cmp3), m_Value(Cmp4))))
3628 return nullptr;
3629
3630 // Check if any two pairs of the and operations are inversions of each other.
3631 if (isKnownInversion(Cmp1, Cmp3) && isKnownInversion(Cmp2, Cmp4))
3632 return Builder.CreateXor(Cmp1, Cmp4);
3633 if (isKnownInversion(Cmp1, Cmp4) && isKnownInversion(Cmp2, Cmp3))
3634 return Builder.CreateXor(Cmp1, Cmp3);
3635
3636 return nullptr;
3637}
3638
3639/// Match \p V as "shufflevector -> bitcast" or "extractelement -> zext -> shl"
3640/// patterns, which extract vector elements and pack them in the same relative
3641/// positions.
3642///
3643/// \p Vec is the underlying vector being extracted from.
3644/// \p Mask is a bitmask identifying which packed elements are obtained from the
3645/// vector.
3646/// \p VecOffset is the vector element corresponding to index 0 of the
3647/// mask.
3649 int64_t &VecOffset,
3650 SmallBitVector &Mask,
3651 const DataLayout &DL) {
3652 // First try to match extractelement -> zext -> shl
3653 uint64_t VecIdx, ShlAmt;
3655 m_ConstantInt(VecIdx))),
3656 ShlAmt))) {
3657 auto *VecTy = dyn_cast<FixedVectorType>(Vec->getType());
3658 if (!VecTy)
3659 return false;
3660 auto *EltTy = dyn_cast<IntegerType>(VecTy->getElementType());
3661 if (!EltTy)
3662 return false;
3663
3664 const unsigned EltBitWidth = EltTy->getBitWidth();
3665 const unsigned TargetBitWidth = V->getType()->getIntegerBitWidth();
3666 if (TargetBitWidth % EltBitWidth != 0 || ShlAmt % EltBitWidth != 0)
3667 return false;
3668 const unsigned TargetEltWidth = TargetBitWidth / EltBitWidth;
3669 const unsigned ShlEltAmt = ShlAmt / EltBitWidth;
3670
3671 const unsigned MaskIdx =
3672 DL.isLittleEndian() ? ShlEltAmt : TargetEltWidth - ShlEltAmt - 1;
3673
3674 VecOffset = static_cast<int64_t>(VecIdx) - static_cast<int64_t>(MaskIdx);
3675 Mask.resize(TargetEltWidth);
3676 Mask.set(MaskIdx);
3677 return true;
3678 }
3679
3680 // Now try to match a bitcasted subvector.
3681 Instruction *SrcVecI;
3682 if (!match(V, m_BitCast(m_Instruction(SrcVecI))))
3683 return false;
3684
3685 auto *SrcTy = dyn_cast<FixedVectorType>(SrcVecI->getType());
3686 if (!SrcTy)
3687 return false;
3688
3689 Mask.resize(SrcTy->getNumElements());
3690
3691 // First check for a subvector obtained from a shufflevector.
3692 if (isa<ShuffleVectorInst>(SrcVecI)) {
3693 Constant *ConstVec;
3694 ArrayRef<int> ShuffleMask;
3695 if (!match(SrcVecI, m_Shuffle(m_Value(Vec), m_Constant(ConstVec),
3696 m_Mask(ShuffleMask))))
3697 return false;
3698
3699 auto *VecTy = dyn_cast<FixedVectorType>(Vec->getType());
3700 if (!VecTy)
3701 return false;
3702
3703 const unsigned NumVecElts = VecTy->getNumElements();
3704 bool FoundVecOffset = false;
3705 for (unsigned Idx = 0; Idx < ShuffleMask.size(); ++Idx) {
3706 if (ShuffleMask[Idx] == PoisonMaskElem)
3707 return false;
3708 const unsigned ShuffleIdx = ShuffleMask[Idx];
3709 if (ShuffleIdx >= NumVecElts) {
3710 const unsigned ConstIdx = ShuffleIdx - NumVecElts;
3711 auto *ConstElt =
3712 dyn_cast<ConstantInt>(ConstVec->getAggregateElement(ConstIdx));
3713 if (!ConstElt || !ConstElt->isNullValue())
3714 return false;
3715 continue;
3716 }
3717
3718 if (FoundVecOffset) {
3719 if (VecOffset + Idx != ShuffleIdx)
3720 return false;
3721 } else {
3722 if (ShuffleIdx < Idx)
3723 return false;
3724 VecOffset = ShuffleIdx - Idx;
3725 FoundVecOffset = true;
3726 }
3727 Mask.set(Idx);
3728 }
3729 return FoundVecOffset;
3730 }
3731
3732 // Check for a subvector obtained as an (insertelement V, 0, idx)
3733 uint64_t InsertIdx;
3734 if (!match(SrcVecI,
3735 m_InsertElt(m_Value(Vec), m_Zero(), m_ConstantInt(InsertIdx))))
3736 return false;
3737
3738 auto *VecTy = dyn_cast<FixedVectorType>(Vec->getType());
3739 if (!VecTy)
3740 return false;
3741 VecOffset = 0;
3742 bool AlreadyInsertedMaskedElt = Mask.test(InsertIdx);
3743 Mask.set();
3744 if (!AlreadyInsertedMaskedElt)
3745 Mask.reset(InsertIdx);
3746 return true;
3747}
3748
3749/// Try to fold the join of two scalar integers whose contents are packed
3750/// elements of the same vector.
3752 InstCombiner::BuilderTy &Builder,
3753 const DataLayout &DL) {
3754 assert(I.getOpcode() == Instruction::Or);
3755 Value *LhsVec, *RhsVec;
3756 int64_t LhsVecOffset, RhsVecOffset;
3757 SmallBitVector Mask;
3758 if (!matchSubIntegerPackFromVector(I.getOperand(0), LhsVec, LhsVecOffset,
3759 Mask, DL))
3760 return nullptr;
3761 if (!matchSubIntegerPackFromVector(I.getOperand(1), RhsVec, RhsVecOffset,
3762 Mask, DL))
3763 return nullptr;
3764 if (LhsVec != RhsVec || LhsVecOffset != RhsVecOffset)
3765 return nullptr;
3766
3767 // Convert into shufflevector -> bitcast;
3768 const unsigned ZeroVecIdx =
3769 cast<FixedVectorType>(LhsVec->getType())->getNumElements();
3770 SmallVector<int> ShuffleMask(Mask.size(), ZeroVecIdx);
3771 for (unsigned Idx : Mask.set_bits()) {
3772 assert(LhsVecOffset + Idx >= 0);
3773 ShuffleMask[Idx] = LhsVecOffset + Idx;
3774 }
3775
3776 Value *MaskedVec = Builder.CreateShuffleVector(
3777 LhsVec, Constant::getNullValue(LhsVec->getType()), ShuffleMask,
3778 I.getName() + ".v");
3779 return CastInst::Create(Instruction::BitCast, MaskedVec, I.getType());
3780}
3781
3782/// Match \p V as "lshr -> mask -> zext -> shl".
3783///
3784/// \p Int is the underlying integer being extracted from.
3785/// \p Mask is a bitmask identifying which bits of the integer are being
3786/// extracted. \p Offset identifies which bit of the result \p V corresponds to
3787/// the least significant bit of \p Int
3788static bool matchZExtedSubInteger(Value *V, Value *&Int, APInt &Mask,
3789 uint64_t &Offset, bool &IsShlNUW,
3790 bool &IsShlNSW) {
3791 Value *ShlOp0;
3792 uint64_t ShlAmt = 0;
3793 if (!match(V, m_OneUse(m_Shl(m_Value(ShlOp0), m_ConstantInt(ShlAmt)))))
3794 return false;
3795
3796 IsShlNUW = cast<BinaryOperator>(V)->hasNoUnsignedWrap();
3797 IsShlNSW = cast<BinaryOperator>(V)->hasNoSignedWrap();
3798
3799 Value *ZExtOp0;
3800 if (!match(ShlOp0, m_OneUse(m_ZExt(m_Value(ZExtOp0)))))
3801 return false;
3802
3803 Value *MaskedOp0;
3804 const APInt *ShiftedMaskConst = nullptr;
3805 if (!match(ZExtOp0, m_CombineOr(m_OneUse(m_And(m_Value(MaskedOp0),
3806 m_APInt(ShiftedMaskConst))),
3807 m_Value(MaskedOp0))))
3808 return false;
3809
3810 uint64_t LShrAmt = 0;
3811 if (!match(MaskedOp0,
3813 m_Value(Int))))
3814 return false;
3815
3816 if (LShrAmt > ShlAmt)
3817 return false;
3818 Offset = ShlAmt - LShrAmt;
3819
3820 Mask = ShiftedMaskConst ? ShiftedMaskConst->shl(LShrAmt)
3822 Int->getType()->getScalarSizeInBits(), LShrAmt);
3823
3824 return true;
3825}
3826
3827/// Try to fold the join of two scalar integers whose bits are unpacked and
3828/// zexted from the same source integer.
3830 InstCombiner::BuilderTy &Builder) {
3831
3832 Value *LhsInt, *RhsInt;
3833 APInt LhsMask, RhsMask;
3834 uint64_t LhsOffset, RhsOffset;
3835 bool IsLhsShlNUW, IsLhsShlNSW, IsRhsShlNUW, IsRhsShlNSW;
3836 if (!matchZExtedSubInteger(Lhs, LhsInt, LhsMask, LhsOffset, IsLhsShlNUW,
3837 IsLhsShlNSW))
3838 return nullptr;
3839 if (!matchZExtedSubInteger(Rhs, RhsInt, RhsMask, RhsOffset, IsRhsShlNUW,
3840 IsRhsShlNSW))
3841 return nullptr;
3842 if (LhsInt != RhsInt || LhsOffset != RhsOffset)
3843 return nullptr;
3844
3845 APInt Mask = LhsMask | RhsMask;
3846
3847 Type *DestTy = Lhs->getType();
3848 Value *Res = Builder.CreateShl(
3849 Builder.CreateZExt(
3850 Builder.CreateAnd(LhsInt, Mask, LhsInt->getName() + ".mask"), DestTy,
3851 LhsInt->getName() + ".zext"),
3852 ConstantInt::get(DestTy, LhsOffset), "", IsLhsShlNUW && IsRhsShlNUW,
3853 IsLhsShlNSW && IsRhsShlNSW);
3854 Res->takeName(Lhs);
3855 return Res;
3856}
3857
3858// A decomposition of ((X & Mask) * Factor). The NUW / NSW bools
3859// track these properities for preservation. Note that we can decompose
3860// equivalent select form of this expression (e.g. (!(X & Mask) ? 0 : Mask *
3861// Factor))
3866 bool NUW;
3867 bool NSW;
3868
3870 return X == Other.X && !Mask.intersects(Other.Mask) &&
3871 Factor == Other.Factor;
3872 }
3873};
3874
3875static std::optional<DecomposedBitMaskMul> matchBitmaskMul(Value *V) {
3877 if (!Op)
3878 return std::nullopt;
3879
3880 // Decompose (A & N) * C) into BitMaskMul
3881 Value *Original = nullptr;
3882 const APInt *Mask = nullptr;
3883 const APInt *MulConst = nullptr;
3884 if (match(Op, m_Mul(m_And(m_Value(Original), m_APInt(Mask)),
3885 m_APInt(MulConst)))) {
3886 if (MulConst->isZero() || Mask->isZero())
3887 return std::nullopt;
3888
3889 return std::optional<DecomposedBitMaskMul>(
3890 {Original, *MulConst, *Mask,
3891 cast<BinaryOperator>(Op)->hasNoUnsignedWrap(),
3892 cast<BinaryOperator>(Op)->hasNoSignedWrap()});
3893 }
3894
3895 Value *Cond = nullptr;
3896 const APInt *EqZero = nullptr, *NeZero = nullptr;
3897
3898 // Decompose ((A & N) ? 0 : N * C) into BitMaskMul
3899 if (match(Op, m_Select(m_Value(Cond), m_APInt(EqZero), m_APInt(NeZero)))) {
3900 auto ICmpDecompose =
3901 decomposeBitTest(Cond, /*LookThroughTrunc=*/true,
3902 /*AllowNonZeroC=*/false, /*DecomposeBitMask=*/true);
3903 if (!ICmpDecompose.has_value())
3904 return std::nullopt;
3905
3906 assert(ICmpInst::isEquality(ICmpDecompose->Pred) &&
3907 ICmpDecompose->C.isZero());
3908
3909 if (ICmpDecompose->Pred == ICmpInst::ICMP_NE)
3910 std::swap(EqZero, NeZero);
3911
3912 if (!EqZero->isZero() || NeZero->isZero())
3913 return std::nullopt;
3914
3915 if (!ICmpDecompose->Mask.isPowerOf2() || ICmpDecompose->Mask.isZero() ||
3916 NeZero->getBitWidth() != ICmpDecompose->Mask.getBitWidth())
3917 return std::nullopt;
3918
3919 if (!NeZero->urem(ICmpDecompose->Mask).isZero())
3920 return std::nullopt;
3921
3922 return std::optional<DecomposedBitMaskMul>(
3923 {ICmpDecompose->X, NeZero->udiv(ICmpDecompose->Mask),
3924 ICmpDecompose->Mask, /*NUW=*/false, /*NSW=*/false});
3925 }
3926
3927 return std::nullopt;
3928}
3929
3930/// (A & N) * C + (A & M) * C -> (A & (N + M)) & C
3931/// This also accepts the equivalent select form of (A & N) * C
3932/// expressions i.e. !(A & N) ? 0 : N * C)
3933static Value *foldBitmaskMul(Value *Op0, Value *Op1,
3934 InstCombiner::BuilderTy &Builder) {
3935 auto Decomp1 = matchBitmaskMul(Op1);
3936 if (!Decomp1)
3937 return nullptr;
3938
3939 auto Decomp0 = matchBitmaskMul(Op0);
3940 if (!Decomp0)
3941 return nullptr;
3942
3943 if (Decomp0->isCombineableWith(*Decomp1)) {
3944 Value *NewAnd = Builder.CreateAnd(
3945 Decomp0->X,
3946 ConstantInt::get(Decomp0->X->getType(), Decomp0->Mask + Decomp1->Mask));
3947
3948 return Builder.CreateMul(
3949 NewAnd, ConstantInt::get(NewAnd->getType(), Decomp1->Factor), "",
3950 Decomp0->NUW && Decomp1->NUW, Decomp0->NSW && Decomp1->NSW);
3951 }
3952
3953 return nullptr;
3954}
3955
3956Value *InstCombinerImpl::foldDisjointOr(Value *LHS, Value *RHS) {
3957 if (Value *Res = foldBitmaskMul(LHS, RHS, Builder))
3958 return Res;
3960 return Res;
3961
3962 return nullptr;
3963}
3964
3965Value *InstCombinerImpl::reassociateDisjointOr(Value *LHS, Value *RHS) {
3966
3967 Value *X, *Y;
3969 if (Value *Res = foldDisjointOr(LHS, X))
3970 return Builder.CreateOr(Res, Y, "", /*IsDisjoint=*/true);
3971 if (Value *Res = foldDisjointOr(LHS, Y))
3972 return Builder.CreateOr(Res, X, "", /*IsDisjoint=*/true);
3973 }
3974
3976 if (Value *Res = foldDisjointOr(X, RHS))
3977 return Builder.CreateOr(Res, Y, "", /*IsDisjoint=*/true);
3978 if (Value *Res = foldDisjointOr(Y, RHS))
3979 return Builder.CreateOr(Res, X, "", /*IsDisjoint=*/true);
3980 }
3981
3982 return nullptr;
3983}
3984
3985/// Fold Res, Overflow = (umul.with.overflow x c1); (or Overflow (ugt Res c2))
3986/// --> (ugt x (c2/c1)). This code checks whether a multiplication of two
3987/// unsigned numbers (one is a constant) is mathematically greater than a
3988/// second constant.
3990 InstCombiner::BuilderTy &Builder,
3991 const DataLayout &DL) {
3992 Value *WOV, *X;
3993 const APInt *C1, *C2;
3994 if (match(&I,
3997 m_Value(X), m_APInt(C1)))),
4000 m_APInt(C2))))) &&
4001 !C1->isZero()) {
4002 Constant *NewC = ConstantInt::get(X->getType(), C2->udiv(*C1));
4003 return Builder.CreateICmp(ICmpInst::ICMP_UGT, X, NewC);
4004 }
4005 return nullptr;
4006}
4007
4008/// Fold select(X >s 0, 0, -X) | smax(X, 0) --> abs(X)
4009/// select(X <s 0, -X, 0) | smax(X, 0) --> abs(X)
4011 InstCombiner::BuilderTy &Builder) {
4012 Value *X;
4013 Value *Sel;
4014 if (match(&I,
4016 auto NegX = m_Neg(m_Specific(X));
4018 m_ZeroInt()),
4019 m_ZeroInt(), NegX)) ||
4021 m_ZeroInt()),
4022 NegX, m_ZeroInt())))
4023 return Builder.CreateBinaryIntrinsic(Intrinsic::abs, X,
4024 Builder.getFalse());
4025 }
4026 return nullptr;
4027}
4028
4030 Value *C, *A, *B;
4031 // (C && A) || (!C && B)
4032 // (C && A) || (B && !C)
4033 // (A && C) || (!C && B)
4034 // (A && C) || (B && !C) (may require freeze)
4035 //
4036 // => select C, A, B
4037 if (match(Op1, m_c_LogicalAnd(m_Not(m_Value(C)), m_Value(B))) &&
4039 auto *SelOp0 = dyn_cast<SelectInst>(Op0);
4040 auto *SelOp1 = dyn_cast<SelectInst>(Op1);
4041
4042 bool MayNeedFreeze = SelOp0 && SelOp1 &&
4043 match(SelOp1->getTrueValue(),
4044 m_Not(m_Specific(SelOp0->getTrueValue())));
4045 if (MayNeedFreeze)
4046 C = Builder.CreateFreeze(C);
4048 Value *C2 = nullptr, *A2 = nullptr, *B2 = nullptr;
4049 if (match(Op0, m_LogicalAnd(m_Specific(C), m_Value(A2))) && SelOp0) {
4050 return SelectInst::Create(C, A, B, "", nullptr, SelOp0);
4051 } else if (match(Op1, m_LogicalAnd(m_Not(m_Value(C2)), m_Value(B2))) &&
4052 SelOp1) {
4053 SelectInst *NewSI = SelectInst::Create(C, A, B, "", nullptr, SelOp1);
4054 NewSI->swapProfMetadata();
4055 return NewSI;
4056 } else {
4057 return createSelectInstWithUnknownProfile(C, A, B);
4058 }
4059 }
4060 return SelectInst::Create(C, A, B);
4061 }
4062
4063 // (!C && A) || (C && B)
4064 // (A && !C) || (C && B)
4065 // (!C && A) || (B && C)
4066 // (A && !C) || (B && C) (may require freeze)
4067 //
4068 // => select C, B, A
4069 if (match(Op0, m_c_LogicalAnd(m_Not(m_Value(C)), m_Value(A))) &&
4071 auto *SelOp0 = dyn_cast<SelectInst>(Op0);
4072 auto *SelOp1 = dyn_cast<SelectInst>(Op1);
4073 bool MayNeedFreeze = SelOp0 && SelOp1 &&
4074 match(SelOp0->getTrueValue(),
4075 m_Not(m_Specific(SelOp1->getTrueValue())));
4076 if (MayNeedFreeze)
4077 C = Builder.CreateFreeze(C);
4079 Value *C2 = nullptr, *A2 = nullptr, *B2 = nullptr;
4080 if (match(Op0, m_LogicalAnd(m_Not(m_Value(C2)), m_Value(A2))) && SelOp0) {
4081 SelectInst *NewSI = SelectInst::Create(C, B, A, "", nullptr, SelOp0);
4082 NewSI->swapProfMetadata();
4083 return NewSI;
4084 } else if (match(Op1, m_LogicalAnd(m_Specific(C), m_Value(B2))) &&
4085 SelOp1) {
4086 return SelectInst::Create(C, B, A, "", nullptr, SelOp1);
4087 } else {
4088 return createSelectInstWithUnknownProfile(C, B, A);
4089 }
4090 }
4091 return SelectInst::Create(C, B, A);
4092 }
4093
4094 return nullptr;
4095}
4096
4097// FIXME: We use commutative matchers (m_c_*) for some, but not all, matches
4098// here. We should standardize that construct where it is needed or choose some
4099// other way to ensure that commutated variants of patterns are not missed.
4101 if (Value *V = simplifyOrInst(I.getOperand(0), I.getOperand(1),
4102 SQ.getWithInstruction(&I)))
4103 return replaceInstUsesWith(I, V);
4104
4106 return &I;
4107
4109 return X;
4110
4112 return Phi;
4113
4114 // See if we can simplify any instructions used by the instruction whose sole
4115 // purpose is to compute bits we don't care about.
4117 return &I;
4118
4119 // Do this before using distributive laws to catch simple and/or/not patterns.
4121 return Xor;
4122
4124 return X;
4125
4127 return X;
4128
4129 // (A & B) | (C & D) -> A ^ D where A == ~C && B == ~D
4130 // (A & B) | (C & D) -> A ^ C where A == ~D && B == ~C
4131 if (Value *V = foldOrOfInversions(I, Builder))
4132 return replaceInstUsesWith(I, V);
4133
4134 // (A&B)|(A&C) -> A&(B|C) etc
4136 return replaceInstUsesWith(I, V);
4137
4138 Value *Op0 = I.getOperand(0), *Op1 = I.getOperand(1);
4139 Type *Ty = I.getType();
4140 if (Ty->isIntOrIntVectorTy(1)) {
4141 if (auto *SI0 = dyn_cast<SelectInst>(Op0)) {
4142 if (auto *R =
4143 foldAndOrOfSelectUsingImpliedCond(Op1, *SI0, /* IsAnd */ false))
4144 return R;
4145 }
4146 if (auto *SI1 = dyn_cast<SelectInst>(Op1)) {
4147 if (auto *R =
4148 foldAndOrOfSelectUsingImpliedCond(Op0, *SI1, /* IsAnd */ false))
4149 return R;
4150 }
4151 }
4152
4153 if (Instruction *FoldedLogic = foldBinOpIntoSelectOrPhi(I))
4154 return FoldedLogic;
4155
4156 if (Instruction *FoldedLogic = foldBinOpSelectBinOp(I))
4157 return FoldedLogic;
4158
4159 if (Instruction *BitOp = matchBSwapOrBitReverse(I, /*MatchBSwaps*/ true,
4160 /*MatchBitReversals*/ true))
4161 return BitOp;
4162
4163 if (Instruction *Funnel = matchFunnelShift(I, *this))
4164 return Funnel;
4165
4167 return replaceInstUsesWith(I, Concat);
4168
4170 return R;
4171
4173 return R;
4174
4175 if (cast<PossiblyDisjointInst>(I).isDisjoint()) {
4176 if (Instruction *R =
4177 foldAddLikeCommutative(I.getOperand(0), I.getOperand(1),
4178 /*NSW=*/true, /*NUW=*/true))
4179 return R;
4180 if (Instruction *R =
4181 foldAddLikeCommutative(I.getOperand(1), I.getOperand(0),
4182 /*NSW=*/true, /*NUW=*/true))
4183 return R;
4184
4185 if (Value *Res = foldDisjointOr(I.getOperand(0), I.getOperand(1)))
4186 return replaceInstUsesWith(I, Res);
4187
4188 if (Value *Res = reassociateDisjointOr(I.getOperand(0), I.getOperand(1)))
4189 return replaceInstUsesWith(I, Res);
4190 }
4191
4192 Value *X, *Y;
4193 const APInt *CV;
4194 if (match(&I, m_c_Or(m_OneUse(m_Xor(m_Value(X), m_APInt(CV))), m_Value(Y))) &&
4195 !CV->isAllOnes() && MaskedValueIsZero(Y, *CV, &I)) {
4196 // (X ^ C) | Y -> (X | Y) ^ C iff Y & C == 0
4197 // The check for a 'not' op is for efficiency (if Y is known zero --> ~X).
4198 Value *Or = Builder.CreateOr(X, Y);
4199 return BinaryOperator::CreateXor(Or, ConstantInt::get(Ty, *CV));
4200 }
4201
4202 // If the operands have no common bits set:
4203 // or (mul X, Y), X --> add (mul X, Y), X --> mul X, (Y + 1)
4205 m_Deferred(X)))) {
4206 Value *IncrementY = Builder.CreateAdd(Y, ConstantInt::get(Ty, 1));
4207 return BinaryOperator::CreateMul(X, IncrementY);
4208 }
4209
4210 // Canonicalization to achieve lowering to Bit Manipulation Instructions (BMI)
4211 // ~X | (X-1) => ~(X & -X)
4212 Value *Op;
4215 Value *NegX = Builder.CreateNeg(Op);
4216 Value *And = Builder.CreateAnd(Op, NegX);
4218 }
4219
4220 // (C && A) || (C && B) => select C, A, B (and similar cases)
4221 //
4222 // Note: This is the same transformation used in `foldSelectOfBools`,
4223 // except that it's an `or` instead of `select`.
4224 if (I.getType()->isIntOrIntVectorTy(1) &&
4225 (Op0->hasOneUse() || Op1->hasOneUse())) {
4226 if (Instruction *V = FoldOrOfLogicalAnds(Op0, Op1)) {
4227 return V;
4228 }
4229 }
4230
4231 // (A & C) | (B & D)
4232 Value *A, *B, *C, *D;
4233 if (match(Op0, m_And(m_Value(A), m_Value(C))) &&
4234 match(Op1, m_And(m_Value(B), m_Value(D)))) {
4235
4236 // (A & C0) | (B & C1)
4237 const APInt *C0, *C1;
4238 if (match(C, m_APInt(C0)) && match(D, m_APInt(C1))) {
4239 Value *X;
4240 if (*C0 == ~*C1) {
4241 // ((X | B) & MaskC) | (B & ~MaskC) -> (X & MaskC) | B
4242 if (match(A, m_c_Or(m_Value(X), m_Specific(B))))
4243 return BinaryOperator::CreateOr(Builder.CreateAnd(X, *C0), B);
4244 // (A & MaskC) | ((X | A) & ~MaskC) -> (X & ~MaskC) | A
4245 if (match(B, m_c_Or(m_Specific(A), m_Value(X))))
4246 return BinaryOperator::CreateOr(Builder.CreateAnd(X, *C1), A);
4247
4248 // ((X ^ B) & MaskC) | (B & ~MaskC) -> (X & MaskC) ^ B
4249 if (match(A, m_c_Xor(m_Value(X), m_Specific(B))))
4250 return BinaryOperator::CreateXor(Builder.CreateAnd(X, *C0), B);
4251 // (A & MaskC) | ((X ^ A) & ~MaskC) -> (X & ~MaskC) ^ A
4252 if (match(B, m_c_Xor(m_Specific(A), m_Value(X))))
4253 return BinaryOperator::CreateXor(Builder.CreateAnd(X, *C1), A);
4254 }
4255
4256 if ((*C0 & *C1).isZero()) {
4257 // ((X | B) & C0) | (B & C1) --> (X | B) & (C0 | C1)
4258 // iff (C0 & C1) == 0 and (X & ~C0) == 0
4259 if (match(A, m_c_Or(m_Value(X), m_Specific(B))) &&
4260 MaskedValueIsZero(X, ~*C0, &I)) {
4261 Constant *C01 = ConstantInt::get(Ty, *C0 | *C1);
4262 return BinaryOperator::CreateAnd(A, C01);
4263 }
4264 // (A & C0) | ((X | A) & C1) --> (X | A) & (C0 | C1)
4265 // iff (C0 & C1) == 0 and (X & ~C1) == 0
4266 if (match(B, m_c_Or(m_Value(X), m_Specific(A))) &&
4267 MaskedValueIsZero(X, ~*C1, &I)) {
4268 Constant *C01 = ConstantInt::get(Ty, *C0 | *C1);
4269 return BinaryOperator::CreateAnd(B, C01);
4270 }
4271 // ((X | C2) & C0) | ((X | C3) & C1) --> (X | C2 | C3) & (C0 | C1)
4272 // iff (C0 & C1) == 0 and (C2 & ~C0) == 0 and (C3 & ~C1) == 0.
4273 const APInt *C2, *C3;
4274 if (match(A, m_Or(m_Value(X), m_APInt(C2))) &&
4275 match(B, m_Or(m_Specific(X), m_APInt(C3))) &&
4276 (*C2 & ~*C0).isZero() && (*C3 & ~*C1).isZero()) {
4277 Value *Or = Builder.CreateOr(X, *C2 | *C3, "bitfield");
4278 Constant *C01 = ConstantInt::get(Ty, *C0 | *C1);
4279 return BinaryOperator::CreateAnd(Or, C01);
4280 }
4281 }
4282 }
4283
4284 // Don't try to form a select if it's unlikely that we'll get rid of at
4285 // least one of the operands. A select is generally more expensive than the
4286 // 'or' that it is replacing.
4287 if (Op0->hasOneUse() || Op1->hasOneUse()) {
4288 // (Cond & C) | (~Cond & D) -> Cond ? C : D, and commuted variants.
4289 if (Value *V = matchSelectFromAndOr(A, C, B, D))
4290 return replaceInstUsesWith(I, V);
4291 if (Value *V = matchSelectFromAndOr(A, C, D, B))
4292 return replaceInstUsesWith(I, V);
4293 if (Value *V = matchSelectFromAndOr(C, A, B, D))
4294 return replaceInstUsesWith(I, V);
4295 if (Value *V = matchSelectFromAndOr(C, A, D, B))
4296 return replaceInstUsesWith(I, V);
4297 if (Value *V = matchSelectFromAndOr(B, D, A, C))
4298 return replaceInstUsesWith(I, V);
4299 if (Value *V = matchSelectFromAndOr(B, D, C, A))
4300 return replaceInstUsesWith(I, V);
4301 if (Value *V = matchSelectFromAndOr(D, B, A, C))
4302 return replaceInstUsesWith(I, V);
4303 if (Value *V = matchSelectFromAndOr(D, B, C, A))
4304 return replaceInstUsesWith(I, V);
4305 }
4306 }
4307
4308 if (match(Op0, m_And(m_Value(A), m_Value(C))) &&
4309 match(Op1, m_Not(m_Or(m_Value(B), m_Value(D)))) &&
4310 (Op0->hasOneUse() || Op1->hasOneUse())) {
4311 // (Cond & C) | ~(Cond | D) -> Cond ? C : ~D
4312 if (Value *V = matchSelectFromAndOr(A, C, B, D, true))
4313 return replaceInstUsesWith(I, V);
4314 if (Value *V = matchSelectFromAndOr(A, C, D, B, true))
4315 return replaceInstUsesWith(I, V);
4316 if (Value *V = matchSelectFromAndOr(C, A, B, D, true))
4317 return replaceInstUsesWith(I, V);
4318 if (Value *V = matchSelectFromAndOr(C, A, D, B, true))
4319 return replaceInstUsesWith(I, V);
4320 }
4321
4322 // (A ^ B) | ((B ^ C) ^ A) -> (A ^ B) | C
4323 if (match(Op0, m_Xor(m_Value(A), m_Value(B))))
4324 if (match(Op1,
4327 return BinaryOperator::CreateOr(Op0, C);
4328
4329 // ((B ^ C) ^ A) | (A ^ B) -> (A ^ B) | C
4330 if (match(Op1, m_Xor(m_Value(A), m_Value(B))))
4331 if (match(Op0,
4334 return BinaryOperator::CreateOr(Op1, C);
4335
4336 if (Instruction *DeMorgan = matchDeMorgansLaws(I, *this))
4337 return DeMorgan;
4338
4339 // Canonicalize xor to the RHS.
4340 bool SwappedForXor = false;
4341 if (match(Op0, m_Xor(m_Value(), m_Value()))) {
4342 std::swap(Op0, Op1);
4343 SwappedForXor = true;
4344 }
4345
4346 if (match(Op1, m_Xor(m_Value(A), m_Value(B)))) {
4347 // (A | ?) | (A ^ B) --> (A | ?) | B
4348 // (B | ?) | (A ^ B) --> (B | ?) | A
4349 if (match(Op0, m_c_Or(m_Specific(A), m_Value())))
4350 return BinaryOperator::CreateOr(Op0, B);
4351 if (match(Op0, m_c_Or(m_Specific(B), m_Value())))
4352 return BinaryOperator::CreateOr(Op0, A);
4353
4354 // (A & B) | (A ^ B) --> A | B
4355 // (B & A) | (A ^ B) --> A | B
4356 if (match(Op0, m_c_And(m_Specific(A), m_Specific(B))))
4357 return BinaryOperator::CreateOr(A, B);
4358
4359 // ~A | (A ^ B) --> ~(A & B)
4360 // ~B | (A ^ B) --> ~(A & B)
4361 // The swap above should always make Op0 the 'not'.
4362 if ((Op0->hasOneUse() || Op1->hasOneUse()) &&
4363 (match(Op0, m_Not(m_Specific(A))) || match(Op0, m_Not(m_Specific(B)))))
4364 return BinaryOperator::CreateNot(Builder.CreateAnd(A, B));
4365
4366 // Same as above, but peek through an 'and' to the common operand:
4367 // ~(A & ?) | (A ^ B) --> ~((A & ?) & B)
4368 // ~(B & ?) | (A ^ B) --> ~((B & ?) & A)
4370 if ((Op0->hasOneUse() || Op1->hasOneUse()) &&
4371 match(Op0,
4373 return BinaryOperator::CreateNot(Builder.CreateAnd(And, B));
4374 if ((Op0->hasOneUse() || Op1->hasOneUse()) &&
4375 match(Op0,
4377 return BinaryOperator::CreateNot(Builder.CreateAnd(And, A));
4378
4379 // (~A | C) | (A ^ B) --> ~(A & B) | C
4380 // (~B | C) | (A ^ B) --> ~(A & B) | C
4381 if (Op0->hasOneUse() && Op1->hasOneUse() &&
4382 (match(Op0, m_c_Or(m_Not(m_Specific(A)), m_Value(C))) ||
4383 match(Op0, m_c_Or(m_Not(m_Specific(B)), m_Value(C))))) {
4384 Value *Nand = Builder.CreateNot(Builder.CreateAnd(A, B), "nand");
4385 return BinaryOperator::CreateOr(Nand, C);
4386 }
4387 }
4388
4389 if (SwappedForXor)
4390 std::swap(Op0, Op1);
4391
4392 if (Value *Res =
4393 foldBooleanAndOr(Op0, Op1, I, /*IsAnd=*/false, /*IsLogical=*/false))
4394 return replaceInstUsesWith(I, Res);
4395
4396 if (match(Op1, m_OneUse(m_LogicalOr(m_Value(X), m_Value(Y))))) {
4397 bool IsLogical = isa<SelectInst>(Op1);
4398 if (auto *V = reassociateBooleanAndOr(Op0, X, Y, I, /*IsAnd=*/false,
4399 /*RHSIsLogical=*/IsLogical))
4400 return replaceInstUsesWith(I, V);
4401 }
4402 if (match(Op0, m_OneUse(m_LogicalOr(m_Value(X), m_Value(Y))))) {
4403 bool IsLogical = isa<SelectInst>(Op0);
4404 if (auto *V = reassociateBooleanAndOr(Op1, X, Y, I, /*IsAnd=*/false,
4405 /*RHSIsLogical=*/IsLogical))
4406 return replaceInstUsesWith(I, V);
4407 }
4408
4409 if (Instruction *FoldedFCmps = reassociateFCmps(I, Builder))
4410 return FoldedFCmps;
4411
4412 if (Instruction *CastedOr = foldCastedBitwiseLogic(I))
4413 return CastedOr;
4414
4415 if (Instruction *Sel = foldBinopOfSextBoolToSelect(I))
4416 return Sel;
4417
4418 // or(sext(A), B) / or(B, sext(A)) --> A ? -1 : B, where A is i1 or <N x i1>.
4419 // TODO: Move this into foldBinopOfSextBoolToSelect as a more generalized fold
4420 // with binop identity constant. But creating a select with non-constant
4421 // arm may not be reversible due to poison semantics. Is that a good
4422 // canonicalization?
4423 if (match(&I, m_c_Or(m_OneUse(m_SExt(m_Value(A))), m_Value(B))) &&
4424 A->getType()->isIntOrIntVectorTy(1))
4425 return createSelectInstWithUnknownProfile(
4427
4428 // Note: If we've gotten to the point of visiting the outer OR, then the
4429 // inner one couldn't be simplified. If it was a constant, then it won't
4430 // be simplified by a later pass either, so we try swapping the inner/outer
4431 // ORs in the hopes that we'll be able to simplify it this way.
4432 // (X|C) | V --> (X|V) | C
4433 // Pass the disjoint flag in the following two patterns:
4434 // 1. or-disjoint (or-disjoint X, C), V -->
4435 // or-disjoint (or-disjoint X, V), C
4436 //
4437 // 2. or-disjoint (or X, C), V -->
4438 // or (or-disjoint X, V), C
4439 ConstantInt *CI;
4440 if (Op0->hasOneUse() && !match(Op1, m_ConstantInt()) &&
4441 match(Op0, m_Or(m_Value(A), m_ConstantInt(CI)))) {
4442 bool IsDisjointOuter = cast<PossiblyDisjointInst>(I).isDisjoint();
4443 bool IsDisjointInner = cast<PossiblyDisjointInst>(Op0)->isDisjoint();
4444 Value *Inner = Builder.CreateOr(A, Op1);
4445 cast<PossiblyDisjointInst>(Inner)->setIsDisjoint(IsDisjointOuter);
4446 Inner->takeName(Op0);
4447 return IsDisjointOuter && IsDisjointInner
4448 ? BinaryOperator::CreateDisjointOr(Inner, CI)
4449 : BinaryOperator::CreateOr(Inner, CI);
4450 }
4451
4452 // Change (or (bool?A:B),(bool?C:D)) --> (bool?(or A,C):(or B,D))
4453 // Since this OR statement hasn't been optimized further yet, we hope
4454 // that this transformation will allow the new ORs to be optimized.
4455 {
4456 Value *X = nullptr, *Y = nullptr;
4457 if (Op0->hasOneUse() && Op1->hasOneUse() &&
4458 match(Op0, m_Select(m_Value(X), m_Value(A), m_Value(B))) &&
4459 match(Op1, m_Select(m_Value(Y), m_Value(C), m_Value(D))) && X == Y) {
4460 Value *orTrue = Builder.CreateOr(A, C);
4461 Value *orFalse = Builder.CreateOr(B, D);
4462 return SelectInst::Create(X, orTrue, orFalse);
4463 }
4464 }
4465
4466 // or(ashr(subNSW(Y, X), ScalarSizeInBits(Y) - 1), X) --> X s> Y ? -1 : X.
4467 {
4468 Value *X, *Y;
4471 m_SpecificInt(Ty->getScalarSizeInBits() - 1))),
4472 m_Deferred(X)))) {
4473 Value *NewICmpInst = Builder.CreateICmpSGT(X, Y);
4475 return createSelectInstWithUnknownProfile(NewICmpInst, AllOnes, X);
4476 }
4477 }
4478
4479 {
4480 // ((A & B) ^ A) | ((A & B) ^ B) -> A ^ B
4481 // (A ^ (A & B)) | (B ^ (A & B)) -> A ^ B
4482 // ((A & B) ^ B) | ((A & B) ^ A) -> A ^ B
4483 // (B ^ (A & B)) | (A ^ (A & B)) -> A ^ B
4484 const auto TryXorOpt = [&](Value *Lhs, Value *Rhs) -> Instruction * {
4485 if (match(Lhs, m_c_Xor(m_And(m_Value(A), m_Value(B)), m_Deferred(A))) &&
4486 match(Rhs,
4488 return BinaryOperator::CreateXor(A, B);
4489 }
4490 return nullptr;
4491 };
4492
4493 if (Instruction *Result = TryXorOpt(Op0, Op1))
4494 return Result;
4495 if (Instruction *Result = TryXorOpt(Op1, Op0))
4496 return Result;
4497 }
4498
4499 if (Instruction *V =
4501 return V;
4502
4503 CmpPredicate Pred;
4504 Value *Mul, *Ov, *MulIsNotZero, *UMulWithOv;
4505 // Check if the OR weakens the overflow condition for umul.with.overflow by
4506 // treating any non-zero result as overflow. In that case, we overflow if both
4507 // umul.with.overflow operands are != 0, as in that case the result can only
4508 // be 0, iff the multiplication overflows.
4509 if (match(&I, m_c_Or(m_Value(Ov, m_ExtractValue<1>(m_Value(UMulWithOv))),
4510 m_Value(MulIsNotZero,
4514 m_Deferred(UMulWithOv))),
4515 m_ZeroInt())))) &&
4516 (Ov->hasOneUse() || (MulIsNotZero->hasOneUse() && Mul->hasOneUse()))) {
4517 Value *A, *B;
4519 m_Value(A), m_Value(B)))) {
4520 Value *NotNullA = Builder.CreateIsNotNull(A);
4521 Value *NotNullB = Builder.CreateIsNotNull(B);
4522 return BinaryOperator::CreateAnd(NotNullA, NotNullB);
4523 }
4524 }
4525
4526 /// Res, Overflow = xxx_with_overflow X, C1
4527 /// Try to canonicalize the pattern "Overflow | icmp pred Res, C2" into
4528 /// "Overflow | icmp pred X, C2 +/- C1".
4529 const WithOverflowInst *WO;
4530 const Value *WOV;
4531 const APInt *C1, *C2;
4533 m_Value(WOV, m_WithOverflowInst(WO)))),
4535 m_APInt(C2))))) &&
4536 (WO->getBinaryOp() == Instruction::Add ||
4537 WO->getBinaryOp() == Instruction::Sub) &&
4538 (ICmpInst::isEquality(Pred) ||
4539 WO->isSigned() == ICmpInst::isSigned(Pred)) &&
4540 match(WO->getRHS(), m_APInt(C1))) {
4541 bool Overflow;
4542 APInt NewC = WO->getBinaryOp() == Instruction::Add
4543 ? (ICmpInst::isSigned(Pred) ? C2->ssub_ov(*C1, Overflow)
4544 : C2->usub_ov(*C1, Overflow))
4545 : (ICmpInst::isSigned(Pred) ? C2->sadd_ov(*C1, Overflow)
4546 : C2->uadd_ov(*C1, Overflow));
4547 if (!Overflow || ICmpInst::isEquality(Pred)) {
4548 Value *NewCmp = Builder.CreateICmp(
4549 Pred, WO->getLHS(), ConstantInt::get(WO->getLHS()->getType(), NewC));
4550 return BinaryOperator::CreateOr(Ov, NewCmp);
4551 }
4552 }
4553
4554 // Try to fold the pattern "Overflow | icmp pred Res, C2" into a single
4555 // comparison instruction for umul.with.overflow.
4557 return replaceInstUsesWith(I, R);
4558
4559 // (~x) | y --> ~(x & (~y)) iff that gets rid of inversions
4561 return &I;
4562
4563 // Improve "get low bit mask up to and including bit X" pattern:
4564 // (1 << X) | ((1 << X) + -1) --> -1 l>> (bitwidth(x) - 1 - X)
4565 if (match(&I, m_c_Or(m_Add(m_Shl(m_One(), m_Value(X)), m_AllOnes()),
4566 m_Shl(m_One(), m_Deferred(X)))) &&
4567 match(&I, m_c_Or(m_OneUse(m_Value()), m_Value()))) {
4568 Value *Sub = Builder.CreateSub(
4569 ConstantInt::get(Ty, Ty->getScalarSizeInBits() - 1), X);
4570 return BinaryOperator::CreateLShr(Constant::getAllOnesValue(Ty), Sub);
4571 }
4572
4573 // An or recurrence w/loop invariant step is equivelent to (or start, step)
4574 PHINode *PN = nullptr;
4575 Value *Start = nullptr, *Step = nullptr;
4576 if (matchSimpleRecurrence(&I, PN, Start, Step) && DT.dominates(Step, PN))
4577 return replaceInstUsesWith(I, Builder.CreateOr(Start, Step));
4578
4579 // (A & B) | (C | D) or (C | D) | (A & B)
4580 // Can be combined if C or D is of type (A/B & X)
4582 m_OneUse(m_Or(m_Value(C), m_Value(D)))))) {
4583 // (A & B) | (C | ?) -> C | (? | (A & B))
4584 // (A & B) | (C | ?) -> C | (? | (A & B))
4585 // (A & B) | (C | ?) -> C | (? | (A & B))
4586 // (A & B) | (C | ?) -> C | (? | (A & B))
4587 // (C | ?) | (A & B) -> C | (? | (A & B))
4588 // (C | ?) | (A & B) -> C | (? | (A & B))
4589 // (C | ?) | (A & B) -> C | (? | (A & B))
4590 // (C | ?) | (A & B) -> C | (? | (A & B))
4591 if (match(D, m_OneUse(m_c_And(m_Specific(A), m_Value()))) ||
4593 return BinaryOperator::CreateOr(
4594 C, Builder.CreateOr(D, Builder.CreateAnd(A, B)));
4595 // (A & B) | (? | D) -> (? | (A & B)) | D
4596 // (A & B) | (? | D) -> (? | (A & B)) | D
4597 // (A & B) | (? | D) -> (? | (A & B)) | D
4598 // (A & B) | (? | D) -> (? | (A & B)) | D
4599 // (? | D) | (A & B) -> (? | (A & B)) | D
4600 // (? | D) | (A & B) -> (? | (A & B)) | D
4601 // (? | D) | (A & B) -> (? | (A & B)) | D
4602 // (? | D) | (A & B) -> (? | (A & B)) | D
4603 if (match(C, m_OneUse(m_c_And(m_Specific(A), m_Value()))) ||
4605 return BinaryOperator::CreateOr(
4606 Builder.CreateOr(C, Builder.CreateAnd(A, B)), D);
4607 }
4608
4610 return R;
4611
4612 if (Instruction *Canonicalized = canonicalizeLogicFirst(I, Builder))
4613 return Canonicalized;
4614
4615 if (Instruction *Folded = foldLogicOfIsFPClass(I, Op0, Op1))
4616 return Folded;
4617
4618 if (Instruction *Res = foldBinOpOfDisplacedShifts(I))
4619 return Res;
4620
4621 // If we are setting the sign bit of a floating-point value, convert
4622 // this to fneg(fabs), then cast back to integer.
4623 //
4624 // If the result isn't immediately cast back to a float, this will increase
4625 // the number of instructions. This is still probably a better canonical form
4626 // as it enables FP value tracking.
4627 //
4628 // Assumes any IEEE-represented type has the sign bit in the high bit.
4629 //
4630 // This is generous interpretation of noimplicitfloat, this is not a true
4631 // floating-point operation.
4632 Value *CastOp;
4633 if (match(Op0, m_ElementWiseBitCast(m_Value(CastOp))) &&
4634 match(Op1, m_SignMask()) &&
4635 !Builder.GetInsertBlock()->getParent()->hasFnAttribute(
4636 Attribute::NoImplicitFloat)) {
4637 Type *EltTy = CastOp->getType()->getScalarType();
4638 if (EltTy->isFloatingPointTy() &&
4640 Value *FAbs = Builder.CreateFAbs(CastOp);
4641 Value *FNegFAbs = Builder.CreateFNeg(FAbs);
4642 return new BitCastInst(FNegFAbs, I.getType());
4643 }
4644 }
4645
4646 // (X & C1) | C2 -> X & (C1 | C2) iff (X & C2) == C2
4647 if (match(Op0, m_OneUse(m_And(m_Value(X), m_APInt(C1)))) &&
4648 match(Op1, m_APInt(C2))) {
4649 KnownBits KnownX = computeKnownBits(X, &I);
4650 if ((KnownX.One & *C2) == *C2)
4651 return BinaryOperator::CreateAnd(X, ConstantInt::get(Ty, *C1 | *C2));
4652 }
4653
4655 return Res;
4656
4657 if (Value *V =
4659 /*SimplifyOnly*/ false, *this))
4660 return BinaryOperator::CreateOr(V, Op1);
4661 if (Value *V =
4663 /*SimplifyOnly*/ false, *this))
4664 return BinaryOperator::CreateOr(Op0, V);
4665
4666 if (cast<PossiblyDisjointInst>(I).isDisjoint())
4668 return replaceInstUsesWith(I, V);
4669
4671 return replaceInstUsesWith(I, Res);
4672
4673 return nullptr;
4674}
4675
4676/// A ^ B can be specified using other logic ops in a variety of patterns. We
4677/// can fold these early and efficiently by morphing an existing instruction.
4679 InstCombiner::BuilderTy &Builder) {
4680 assert(I.getOpcode() == Instruction::Xor);
4681 Value *Op0 = I.getOperand(0);
4682 Value *Op1 = I.getOperand(1);
4683 Value *A, *B;
4684
4685 // There are 4 commuted variants for each of the basic patterns.
4686
4687 // (A & B) ^ (A | B) -> A ^ B
4688 // (A & B) ^ (B | A) -> A ^ B
4689 // (A | B) ^ (A & B) -> A ^ B
4690 // (A | B) ^ (B & A) -> A ^ B
4691 if (match(&I, m_c_Xor(m_And(m_Value(A), m_Value(B)),
4693 return BinaryOperator::CreateXor(A, B);
4694
4695 // (A | ~B) ^ (~A | B) -> A ^ B
4696 // (~B | A) ^ (~A | B) -> A ^ B
4697 // (~A | B) ^ (A | ~B) -> A ^ B
4698 // (B | ~A) ^ (A | ~B) -> A ^ B
4699 if (match(&I, m_Xor(m_c_Or(m_Value(A), m_Not(m_Value(B))),
4701 return BinaryOperator::CreateXor(A, B);
4702
4703 // (A & ~B) ^ (~A & B) -> A ^ B
4704 // (~B & A) ^ (~A & B) -> A ^ B
4705 // (~A & B) ^ (A & ~B) -> A ^ B
4706 // (B & ~A) ^ (A & ~B) -> A ^ B
4707 if (match(&I, m_Xor(m_c_And(m_Value(A), m_Not(m_Value(B))),
4709 return BinaryOperator::CreateXor(A, B);
4710
4711 // For the remaining cases we need to get rid of one of the operands.
4712 if (!Op0->hasOneUse() && !Op1->hasOneUse())
4713 return nullptr;
4714
4715 // (A | B) ^ ~(A & B) -> ~(A ^ B)
4716 // (A | B) ^ ~(B & A) -> ~(A ^ B)
4717 // (A & B) ^ ~(A | B) -> ~(A ^ B)
4718 // (A & B) ^ ~(B | A) -> ~(A ^ B)
4719 // Complexity sorting ensures the not will be on the right side.
4720 if ((match(Op0, m_Or(m_Value(A), m_Value(B))) &&
4721 match(Op1, m_Not(m_c_And(m_Specific(A), m_Specific(B))))) ||
4722 (match(Op0, m_And(m_Value(A), m_Value(B))) &&
4724 return BinaryOperator::CreateNot(Builder.CreateXor(A, B));
4725
4726 return nullptr;
4727}
4728
4729Value *InstCombinerImpl::foldXorOfICmps(ICmpInst *LHS, ICmpInst *RHS,
4730 BinaryOperator &I) {
4731 assert(I.getOpcode() == Instruction::Xor && I.getOperand(0) == LHS &&
4732 I.getOperand(1) == RHS && "Should be 'xor' with these operands");
4733
4734 ICmpInst::Predicate PredL = LHS->getPredicate(), PredR = RHS->getPredicate();
4735 Value *LHS0 = LHS->getOperand(0), *LHS1 = LHS->getOperand(1);
4736 Value *RHS0 = RHS->getOperand(0), *RHS1 = RHS->getOperand(1);
4737
4738 if (predicatesFoldable(PredL, PredR)) {
4739 if (LHS0 == RHS1 && LHS1 == RHS0) {
4740 std::swap(LHS0, LHS1);
4741 PredL = ICmpInst::getSwappedPredicate(PredL);
4742 }
4743 if (LHS0 == RHS0 && LHS1 == RHS1) {
4744 // (icmp1 A, B) ^ (icmp2 A, B) --> (icmp3 A, B)
4745 unsigned Code = getICmpCode(PredL) ^ getICmpCode(PredR);
4746 bool IsSigned = LHS->isSigned() || RHS->isSigned();
4747 return getNewICmpValue(Code, IsSigned, LHS0, LHS1, Builder);
4748 }
4749 }
4750
4751 const APInt *LC, *RC;
4752 if (match(LHS1, m_APInt(LC)) && match(RHS1, m_APInt(RC)) &&
4753 LHS0->getType() == RHS0->getType() &&
4754 LHS0->getType()->isIntOrIntVectorTy()) {
4755 // Convert xor of signbit tests to signbit test of xor'd values:
4756 // (X > -1) ^ (Y > -1) --> (X ^ Y) < 0
4757 // (X < 0) ^ (Y < 0) --> (X ^ Y) < 0
4758 // (X > -1) ^ (Y < 0) --> (X ^ Y) > -1
4759 // (X < 0) ^ (Y > -1) --> (X ^ Y) > -1
4760 bool TrueIfSignedL, TrueIfSignedR;
4761 if ((LHS->hasOneUse() || RHS->hasOneUse()) &&
4762 isSignBitCheck(PredL, *LC, TrueIfSignedL) &&
4763 isSignBitCheck(PredR, *RC, TrueIfSignedR)) {
4764 Value *XorLR = Builder.CreateXor(LHS0, RHS0);
4765 return TrueIfSignedL == TrueIfSignedR ? Builder.CreateIsNeg(XorLR) :
4766 Builder.CreateIsNotNeg(XorLR);
4767 }
4768
4769 // Fold (icmp pred1 X, C1) ^ (icmp pred2 X, C2)
4770 // into a single comparison using range-based reasoning.
4771 if (LHS0 == RHS0) {
4772 ConstantRange CR1 = ConstantRange::makeExactICmpRegion(PredL, *LC);
4773 ConstantRange CR2 = ConstantRange::makeExactICmpRegion(PredR, *RC);
4774 auto CRUnion = CR1.exactUnionWith(CR2);
4775 auto CRIntersect = CR1.exactIntersectWith(CR2);
4776 if (CRUnion && CRIntersect)
4777 if (auto CR = CRUnion->exactIntersectWith(CRIntersect->inverse())) {
4778 if (CR->isFullSet())
4779 return ConstantInt::getTrue(I.getType());
4780 if (CR->isEmptySet())
4781 return ConstantInt::getFalse(I.getType());
4782
4783 CmpInst::Predicate NewPred;
4784 APInt NewC, Offset;
4785 CR->getEquivalentICmp(NewPred, NewC, Offset);
4786
4787 if ((Offset.isZero() && (LHS->hasOneUse() || RHS->hasOneUse())) ||
4788 (LHS->hasOneUse() && RHS->hasOneUse())) {
4789 Value *NewV = LHS0;
4790 Type *Ty = LHS0->getType();
4791 if (!Offset.isZero())
4792 NewV = Builder.CreateAdd(NewV, ConstantInt::get(Ty, Offset));
4793 return Builder.CreateICmp(NewPred, NewV,
4794 ConstantInt::get(Ty, NewC));
4795 }
4796 }
4797 }
4798
4799 // Fold (icmp eq/ne (X & Pow2), 0) ^ (icmp eq/ne (Y & Pow2), 0) into
4800 // (icmp eq/ne ((X ^ Y) & Pow2), 0)
4801 Value *X, *Y, *Pow2;
4802 if (ICmpInst::isEquality(PredL) && ICmpInst::isEquality(PredR) &&
4803 LC->isZero() && RC->isZero() && LHS->hasOneUse() && RHS->hasOneUse() &&
4804 match(LHS0, m_And(m_Value(X), m_Value(Pow2))) &&
4805 match(RHS0, m_And(m_Value(Y), m_Specific(Pow2))) &&
4806 isKnownToBeAPowerOfTwo(Pow2, /*OrZero=*/true, &I)) {
4807 Value *Xor = Builder.CreateXor(X, Y);
4808 Value *And = Builder.CreateAnd(Xor, Pow2);
4809 return Builder.CreateICmp(PredL == PredR ? ICmpInst::ICMP_NE
4811 And, ConstantInt::getNullValue(Xor->getType()));
4812 }
4813 }
4814
4815 // Instead of trying to imitate the folds for and/or, decompose this 'xor'
4816 // into those logic ops. That is, try to turn this into an and-of-icmps
4817 // because we have many folds for that pattern.
4818 //
4819 // This is based on a truth table definition of xor:
4820 // X ^ Y --> (X | Y) & !(X & Y)
4821 if (Value *OrICmp = simplifyBinOp(Instruction::Or, LHS, RHS, SQ)) {
4822 // TODO: If OrICmp is true, then the definition of xor simplifies to !(X&Y).
4823 // TODO: If OrICmp is false, the whole thing is false (InstSimplify?).
4824 if (Value *AndICmp = simplifyBinOp(Instruction::And, LHS, RHS, SQ)) {
4825 // TODO: Independently handle cases where the 'and' side is a constant.
4826 ICmpInst *X = nullptr, *Y = nullptr;
4827 if (OrICmp == LHS && AndICmp == RHS) {
4828 // (LHS | RHS) & !(LHS & RHS) --> LHS & !RHS --> X & !Y
4829 X = LHS;
4830 Y = RHS;
4831 }
4832 if (OrICmp == RHS && AndICmp == LHS) {
4833 // !(LHS & RHS) & (LHS | RHS) --> !LHS & RHS --> !Y & X
4834 X = RHS;
4835 Y = LHS;
4836 }
4837 if (X && Y && (Y->hasOneUse() || canFreelyInvertAllUsersOf(Y, &I))) {
4838 // Invert the predicate of 'Y', thus inverting its output.
4839 Y->setPredicate(Y->getInversePredicate());
4840 // So, are there other uses of Y?
4841 if (!Y->hasOneUse()) {
4842 // We need to adapt other uses of Y though. Get a value that matches
4843 // the original value of Y before inversion. While this increases
4844 // immediate instruction count, we have just ensured that all the
4845 // users are freely-invertible, so that 'not' *will* get folded away.
4847 // Set insertion point to right after the Y.
4848 Builder.SetInsertPoint(Y->getParent(), ++(Y->getIterator()));
4849 Value *NotY = Builder.CreateNot(Y, Y->getName() + ".not");
4850 // Replace all uses of Y (excluding the one in NotY!) with NotY.
4851 Worklist.pushUsersToWorkList(*Y);
4852 Y->replaceUsesWithIf(NotY,
4853 [NotY](Use &U) { return U.getUser() != NotY; });
4854 }
4855 // All done.
4856 return Builder.CreateAnd(LHS, RHS);
4857 }
4858 }
4859 }
4860
4861 return nullptr;
4862}
4863
4864/// If we have a masked merge, in the canonical form of:
4865/// (assuming that A only has one use.)
4866/// | A | |B|
4867/// ((x ^ y) & M) ^ y
4868/// | D |
4869/// * If M is inverted:
4870/// | D |
4871/// ((x ^ y) & ~M) ^ y
4872/// We can canonicalize by swapping the final xor operand
4873/// to eliminate the 'not' of the mask.
4874/// ((x ^ y) & M) ^ x
4875/// * If M is a constant, and D has one use, we transform to 'and' / 'or' ops
4876/// because that shortens the dependency chain and improves analysis:
4877/// (x & M) | (y & ~M)
4879 InstCombiner::BuilderTy &Builder) {
4880 Value *B, *X, *D;
4881 Value *M;
4882 if (!match(&I, m_c_Xor(m_Value(B),
4885 m_Value(M))))))
4886 return nullptr;
4887
4888 Value *NotM;
4889 if (match(M, m_Not(m_Value(NotM)))) {
4890 // De-invert the mask and swap the value in B part.
4891 Value *NewA = Builder.CreateAnd(D, NotM);
4892 return BinaryOperator::CreateXor(NewA, X);
4893 }
4894
4895 Constant *C;
4896 if (D->hasOneUse() && match(M, m_Constant(C))) {
4897 // Propagating undef is unsafe. Clamp undef elements to -1.
4898 Type *EltTy = C->getType()->getScalarType();
4900 // Unfold.
4901 Value *LHS = Builder.CreateAnd(X, C);
4902 Value *NotC = Builder.CreateNot(C);
4903 Value *RHS = Builder.CreateAnd(B, NotC);
4904 return BinaryOperator::CreateOr(LHS, RHS);
4905 }
4906
4907 return nullptr;
4908}
4909
4911 InstCombiner::BuilderTy &Builder) {
4912 Value *X, *Y;
4913 // FIXME: one-use check is not needed in general, but currently we are unable
4914 // to fold 'not' into 'icmp', if that 'icmp' has multiple uses. (D35182)
4915 if (!match(&I, m_Not(m_OneUse(m_Xor(m_Value(X), m_Value(Y))))))
4916 return nullptr;
4917
4918 auto hasCommonOperand = [](Value *A, Value *B, Value *C, Value *D) {
4919 return A == C || A == D || B == C || B == D;
4920 };
4921
4922 Value *A, *B, *C, *D;
4923 // Canonicalize ~((A & B) ^ (A | ?)) -> (A & B) | ~(A | ?)
4924 // 4 commuted variants
4925 if (match(X, m_And(m_Value(A), m_Value(B))) &&
4926 match(Y, m_Or(m_Value(C), m_Value(D))) && hasCommonOperand(A, B, C, D)) {
4927 Value *NotY = Builder.CreateNot(Y);
4928 return BinaryOperator::CreateOr(X, NotY);
4929 };
4930
4931 // Canonicalize ~((A | ?) ^ (A & B)) -> (A & B) | ~(A | ?)
4932 // 4 commuted variants
4933 if (match(Y, m_And(m_Value(A), m_Value(B))) &&
4934 match(X, m_Or(m_Value(C), m_Value(D))) && hasCommonOperand(A, B, C, D)) {
4935 Value *NotX = Builder.CreateNot(X);
4936 return BinaryOperator::CreateOr(Y, NotX);
4937 };
4938
4939 return nullptr;
4940}
4941
4942/// Canonicalize a shifty way to code absolute value to the more common pattern
4943/// that uses negation and select.
4945 InstCombiner::BuilderTy &Builder) {
4946 assert(Xor.getOpcode() == Instruction::Xor && "Expected an xor instruction.");
4947
4948 // There are 4 potential commuted variants. Move the 'ashr' candidate to Op1.
4949 // We're relying on the fact that we only do this transform when the shift has
4950 // exactly 2 uses and the add has exactly 1 use (otherwise, we might increase
4951 // instructions).
4952 Value *Op0 = Xor.getOperand(0), *Op1 = Xor.getOperand(1);
4953 if (Op0->hasNUses(2))
4954 std::swap(Op0, Op1);
4955
4956 Type *Ty = Xor.getType();
4957 Value *A;
4958 const APInt *ShAmt;
4959 if (match(Op1, m_AShr(m_Value(A), m_APInt(ShAmt))) &&
4960 Op1->hasNUses(2) && *ShAmt == Ty->getScalarSizeInBits() - 1 &&
4961 match(Op0, m_OneUse(m_c_Add(m_Specific(A), m_Specific(Op1))))) {
4962 // Op1 = ashr i32 A, 31 ; smear the sign bit
4963 // xor (add A, Op1), Op1 ; add -1 and flip bits if negative
4964 // --> (A < 0) ? -A : A
4965 Value *IsNeg = Builder.CreateIsNeg(A);
4966 // Copy the nsw flags from the add to the negate.
4967 auto *Add = cast<BinaryOperator>(Op0);
4968 Value *NegA = Add->hasNoUnsignedWrap()
4969 ? Constant::getNullValue(A->getType())
4970 : Builder.CreateNeg(A, "", Add->hasNoSignedWrap());
4971 return SelectInst::Create(IsNeg, NegA, A);
4972 }
4973 return nullptr;
4974}
4975
4977 Instruction *IgnoredUser) {
4978 auto *I = dyn_cast<Instruction>(Op);
4979 return I && IC.isFreeToInvert(I, /*WillInvertAllUses=*/true) &&
4980 IC.canFreelyInvertAllUsersOf(I, IgnoredUser);
4981}
4982
4984 Instruction *IgnoredUser) {
4985 auto *I = cast<Instruction>(Op);
4986 IC.Builder.SetInsertPoint(*I->getInsertionPointAfterDef());
4987 Value *NotOp = IC.Builder.CreateNot(Op, Op->getName() + ".not");
4988 Op->replaceUsesWithIf(NotOp,
4989 [NotOp](Use &U) { return U.getUser() != NotOp; });
4990 IC.freelyInvertAllUsersOf(NotOp, IgnoredUser);
4991 return NotOp;
4992}
4993
4994// Transform
4995// z = ~(x &/| y)
4996// into:
4997// z = ((~x) |/& (~y))
4998// iff both x and y are free to invert and all uses of z can be freely updated.
5000 Value *Op0, *Op1;
5001 if (!match(&I, m_LogicalOp(m_Value(Op0), m_Value(Op1))))
5002 return false;
5003
5004 // If this logic op has not been simplified yet, just bail out and let that
5005 // happen first. Otherwise, the code below may wrongly invert.
5006 if (Op0 == Op1)
5007 return false;
5008
5009 // If one of the operands is a user of the other,
5010 // freelyInvert->freelyInvertAllUsersOf will change the operands of I, which
5011 // may cause miscompilation.
5012 if (match(Op0, m_Not(m_Specific(Op1))) || match(Op1, m_Not(m_Specific(Op0))))
5013 return false;
5014
5015 Instruction::BinaryOps NewOpc =
5016 match(&I, m_LogicalAnd()) ? Instruction::Or : Instruction::And;
5017 bool IsBinaryOp = isa<BinaryOperator>(I);
5018
5019 // Can our users be adapted?
5020 if (!InstCombiner::canFreelyInvertAllUsersOf(&I, /*IgnoredUser=*/nullptr))
5021 return false;
5022
5023 // And can the operands be adapted?
5024 if (!canFreelyInvert(*this, Op0, &I) || !canFreelyInvert(*this, Op1, &I))
5025 return false;
5026
5027 Op0 = freelyInvert(*this, Op0, &I);
5028 Op1 = freelyInvert(*this, Op1, &I);
5029
5030 Builder.SetInsertPoint(*I.getInsertionPointAfterDef());
5031 Value *NewLogicOp;
5032 if (IsBinaryOp) {
5033 NewLogicOp = Builder.CreateBinOp(NewOpc, Op0, Op1, I.getName() + ".not");
5034 } else {
5035 NewLogicOp =
5036 Builder.CreateLogicalOp(NewOpc, Op0, Op1, I.getName() + ".not",
5037 ProfcheckDisableMetadataFixes ? nullptr : &I);
5038 if (SelectInst *SI = dyn_cast<SelectInst>(NewLogicOp))
5039 SI->swapProfMetadata();
5040 }
5041
5042 replaceInstUsesWith(I, NewLogicOp);
5043 // We can not just create an outer `not`, it will most likely be immediately
5044 // folded back, reconstructing our initial pattern, and causing an
5045 // infinite combine loop, so immediately manually fold it away.
5046 freelyInvertAllUsersOf(NewLogicOp);
5047 return true;
5048}
5049
5050// Transform
5051// z = (~x) &/| y
5052// into:
5053// z = ~(x |/& (~y))
5054// iff y is free to invert and all uses of z can be freely updated.
5056 Value *Op0, *Op1;
5057 if (!match(&I, m_LogicalOp(m_Value(Op0), m_Value(Op1))))
5058 return false;
5059 Instruction::BinaryOps NewOpc =
5060 match(&I, m_LogicalAnd()) ? Instruction::Or : Instruction::And;
5061 bool IsBinaryOp = isa<BinaryOperator>(I);
5062
5063 Value *NotOp0 = nullptr;
5064 Value *NotOp1 = nullptr;
5065 Value **OpToInvert = nullptr;
5066 if (match(Op0, m_Not(m_Value(NotOp0))) && canFreelyInvert(*this, Op1, &I)) {
5067 Op0 = NotOp0;
5068 OpToInvert = &Op1;
5069 } else if (match(Op1, m_Not(m_Value(NotOp1))) &&
5070 canFreelyInvert(*this, Op0, &I)) {
5071 Op1 = NotOp1;
5072 OpToInvert = &Op0;
5073 } else
5074 return false;
5075
5076 // And can our users be adapted?
5077 if (!InstCombiner::canFreelyInvertAllUsersOf(&I, /*IgnoredUser=*/nullptr))
5078 return false;
5079
5080 *OpToInvert = freelyInvert(*this, *OpToInvert, &I);
5081
5082 Builder.SetInsertPoint(*I.getInsertionPointAfterDef());
5083 Value *NewBinOp;
5084 if (IsBinaryOp)
5085 NewBinOp = Builder.CreateBinOp(NewOpc, Op0, Op1, I.getName() + ".not");
5086 else
5087 NewBinOp = Builder.CreateLogicalOp(NewOpc, Op0, Op1, I.getName() + ".not");
5088 replaceInstUsesWith(I, NewBinOp);
5089 // We can not just create an outer `not`, it will most likely be immediately
5090 // folded back, reconstructing our initial pattern, and causing an
5091 // infinite combine loop, so immediately manually fold it away.
5092 freelyInvertAllUsersOf(NewBinOp);
5093 return true;
5094}
5095
5096Instruction *InstCombinerImpl::foldNot(BinaryOperator &I) {
5097 Value *NotOp;
5098 if (!match(&I, m_Not(m_Value(NotOp))))
5099 return nullptr;
5100
5101 // Apply DeMorgan's Law for 'nand' / 'nor' logic with an inverted operand.
5102 // We must eliminate the and/or (one-use) for these transforms to not increase
5103 // the instruction count.
5104 //
5105 // ~(~X & Y) --> (X | ~Y)
5106 // ~(Y & ~X) --> (X | ~Y)
5107 //
5108 // Note: The logical matches do not check for the commuted patterns because
5109 // those are handled via SimplifySelectsFeedingBinaryOp().
5110 Type *Ty = I.getType();
5111 Value *X, *Y;
5112 if (match(NotOp, m_OneUse(m_c_And(m_Not(m_Value(X)), m_Value(Y))))) {
5113 Value *NotY = Builder.CreateNot(Y, Y->getName() + ".not");
5114 return BinaryOperator::CreateOr(X, NotY);
5115 }
5116 if (match(NotOp, m_OneUse(m_LogicalAnd(m_Not(m_Value(X)), m_Value(Y))))) {
5117 Value *NotY = Builder.CreateNot(Y, Y->getName() + ".not");
5119 X, ConstantInt::getTrue(Ty), NotY, "", nullptr,
5121 SI->swapProfMetadata();
5122 return SI;
5123 }
5124
5125 // ~(~X | Y) --> (X & ~Y)
5126 // ~(Y | ~X) --> (X & ~Y)
5127 if (match(NotOp, m_OneUse(m_c_Or(m_Not(m_Value(X)), m_Value(Y))))) {
5128 Value *NotY = Builder.CreateNot(Y, Y->getName() + ".not");
5129 return BinaryOperator::CreateAnd(X, NotY);
5130 }
5131 if (match(NotOp, m_OneUse(m_LogicalOr(m_Not(m_Value(X)), m_Value(Y))))) {
5132 Value *NotY = Builder.CreateNot(Y, Y->getName() + ".not");
5133 SelectInst *SI = SelectInst::Create(
5134 X, NotY, ConstantInt::getFalse(Ty), "", nullptr,
5136 SI->swapProfMetadata();
5137 return SI;
5138 }
5139
5140 // Is this a 'not' (~) fed by a binary operator?
5141 BinaryOperator *NotVal;
5142 if (match(NotOp, m_BinOp(NotVal))) {
5143 // ~((-X) | Y) --> (X - 1) & (~Y)
5144 if (match(NotVal,
5146 Value *DecX = Builder.CreateAdd(X, ConstantInt::getAllOnesValue(Ty));
5147 Value *NotY = Builder.CreateNot(Y);
5148 return BinaryOperator::CreateAnd(DecX, NotY);
5149 }
5150
5151 // ~(~X >>s Y) --> (X >>s Y)
5152 if (match(NotVal, m_AShr(m_Not(m_Value(X)), m_Value(Y))))
5153 return BinaryOperator::CreateAShr(X, Y);
5154
5155 // Treat lshr with non-negative operand as ashr.
5156 // ~(~X >>u Y) --> (X >>s Y) iff X is known negative
5157 if (match(NotVal, m_LShr(m_Not(m_Value(X)), m_Value(Y))) &&
5158 isKnownNegative(X, SQ.getWithInstruction(NotVal)))
5159 return BinaryOperator::CreateAShr(X, Y);
5160
5161 // Bit-hack form of a signbit test for iN type:
5162 // ~(X >>s (N - 1)) --> sext i1 (X > -1) to iN
5163 unsigned FullShift = Ty->getScalarSizeInBits() - 1;
5164 if (match(NotVal, m_OneUse(m_AShr(m_Value(X), m_SpecificInt(FullShift))))) {
5165 Value *IsNotNeg = Builder.CreateIsNotNeg(X, "isnotneg");
5166 return new SExtInst(IsNotNeg, Ty);
5167 }
5168
5169 // If we are inverting a right-shifted constant, we may be able to eliminate
5170 // the 'not' by inverting the constant and using the opposite shift type.
5171 // Canonicalization rules ensure that only a negative constant uses 'ashr',
5172 // but we must check that in case that transform has not fired yet.
5173
5174 // ~(C >>s Y) --> ~C >>u Y (when inverting the replicated sign bits)
5175 Constant *C;
5176 if (match(NotVal, m_AShr(m_Constant(C), m_Value(Y))) &&
5177 match(C, m_Negative()))
5178 return BinaryOperator::CreateLShr(ConstantExpr::getNot(C), Y);
5179
5180 // ~(C >>u Y) --> ~C >>s Y (when inverting the replicated sign bits)
5181 if (match(NotVal, m_LShr(m_Constant(C), m_Value(Y))) &&
5182 match(C, m_NonNegative()))
5183 return BinaryOperator::CreateAShr(ConstantExpr::getNot(C), Y);
5184
5185 // ~(X + C) --> ~C - X
5186 if (match(NotVal, m_Add(m_Value(X), m_ImmConstant(C))))
5187 return BinaryOperator::CreateSub(ConstantExpr::getNot(C), X);
5188
5189 // ~(X - Y) --> ~X + Y
5190 // FIXME: is it really beneficial to sink the `not` here?
5191 if (match(NotVal, m_Sub(m_Value(X), m_Value(Y))))
5192 if (isa<Constant>(X) || NotVal->hasOneUse())
5193 return BinaryOperator::CreateAdd(Builder.CreateNot(X), Y);
5194
5195 // ~(~X + Y) --> X - Y
5196 if (match(NotVal, m_c_Add(m_Not(m_Value(X)), m_Value(Y))))
5197 return BinaryOperator::CreateWithCopiedFlags(Instruction::Sub, X, Y,
5198 NotVal);
5199 }
5200
5201 // not (cmp A, B) = !cmp A, B
5202 CmpPredicate Pred;
5203 if (match(NotOp, m_Cmp(Pred, m_Value(), m_Value())) &&
5204 (NotOp->hasOneUse() ||
5206 /*IgnoredUser=*/nullptr))) {
5207 cast<CmpInst>(NotOp)->setPredicate(CmpInst::getInversePredicate(Pred));
5209 return &I;
5210 }
5211
5212 // not (bitcast (cmp A, B) --> bitcast (!cmp A, B)
5213 if (match(NotOp, m_OneUse(m_BitCast(m_Value(X)))) &&
5214 match(X, m_OneUse(m_Cmp(Pred, m_Value(), m_Value())))) {
5215 cast<CmpInst>(X)->setPredicate(CmpInst::getInversePredicate(Pred));
5216 return new BitCastInst(X, Ty);
5217 }
5218
5219 // Move a 'not' ahead of casts of a bool to enable logic reduction:
5220 // not (bitcast (sext i1 X)) --> bitcast (sext (not i1 X))
5221 if (match(NotOp, m_OneUse(m_BitCast(m_OneUse(m_SExt(m_Value(X)))))) &&
5222 X->getType()->isIntOrIntVectorTy(1)) {
5223 Type *SextTy = cast<BitCastOperator>(NotOp)->getSrcTy();
5224 Value *NotX = Builder.CreateNot(X);
5225 Value *Sext = Builder.CreateSExt(NotX, SextTy);
5226 return new BitCastInst(Sext, Ty);
5227 }
5228
5229 if (auto *NotOpI = dyn_cast<Instruction>(NotOp))
5230 if (sinkNotIntoLogicalOp(*NotOpI))
5231 return &I;
5232
5233 // Eliminate a bitwise 'not' op of 'not' min/max by inverting the min/max:
5234 // ~min(~X, ~Y) --> max(X, Y)
5235 // ~max(~X, Y) --> min(X, ~Y)
5236 auto *II = dyn_cast<IntrinsicInst>(NotOp);
5237 if (II && II->hasOneUse()) {
5238 if (match(NotOp, m_c_MaxOrMin(m_Not(m_Value(X)), m_Value(Y)))) {
5239 Intrinsic::ID InvID = getInverseMinMaxIntrinsic(II->getIntrinsicID());
5240 Value *NotY = Builder.CreateNot(Y);
5241 Value *InvMaxMin = Builder.CreateBinaryIntrinsic(InvID, X, NotY);
5242 return replaceInstUsesWith(I, InvMaxMin);
5243 }
5244
5245 if (II->getIntrinsicID() == Intrinsic::is_fpclass) {
5246 ConstantInt *ClassMask = cast<ConstantInt>(II->getArgOperand(1));
5247 II->setArgOperand(
5248 1, ConstantInt::get(ClassMask->getType(),
5249 ~ClassMask->getZExtValue() & fcAllFlags));
5250 return replaceInstUsesWith(I, II);
5251 }
5252 }
5253
5254 if (NotOp->hasOneUse()) {
5255 // Pull 'not' into operands of select if both operands are one-use compares
5256 // or one is one-use compare and the other one is a constant.
5257 // Inverting the predicates eliminates the 'not' operation.
5258 // Example:
5259 // not (select ?, (cmp TPred, ?, ?), (cmp FPred, ?, ?) -->
5260 // select ?, (cmp InvTPred, ?, ?), (cmp InvFPred, ?, ?)
5261 // not (select ?, (cmp TPred, ?, ?), true -->
5262 // select ?, (cmp InvTPred, ?, ?), false
5263 if (auto *Sel = dyn_cast<SelectInst>(NotOp)) {
5264 Value *TV = Sel->getTrueValue();
5265 Value *FV = Sel->getFalseValue();
5266 auto *CmpT = dyn_cast<CmpInst>(TV);
5267 auto *CmpF = dyn_cast<CmpInst>(FV);
5268 bool InvertibleT = (CmpT && CmpT->hasOneUse()) || isa<Constant>(TV);
5269 bool InvertibleF = (CmpF && CmpF->hasOneUse()) || isa<Constant>(FV);
5270 if (InvertibleT && InvertibleF) {
5271 if (CmpT)
5272 CmpT->setPredicate(CmpT->getInversePredicate());
5273 else
5274 Sel->setTrueValue(ConstantExpr::getNot(cast<Constant>(TV)));
5275 if (CmpF)
5276 CmpF->setPredicate(CmpF->getInversePredicate());
5277 else
5278 Sel->setFalseValue(ConstantExpr::getNot(cast<Constant>(FV)));
5279 return replaceInstUsesWith(I, Sel);
5280 }
5281 }
5282 }
5283
5284 if (Instruction *NewXor = foldNotXor(I, Builder))
5285 return NewXor;
5286
5287 // TODO: Could handle multi-use better by checking if all uses of NotOp (other
5288 // than I) can be inverted.
5289 if (Value *R = getFreelyInverted(NotOp, NotOp->hasOneUse(), &Builder))
5290 return replaceInstUsesWith(I, R);
5291
5292 return nullptr;
5293}
5294
5295// ((X + C) & M) ^ M --> (~C − X) & M
5297 InstCombiner::BuilderTy &Builder) {
5298 Value *X, *Mask;
5299 Constant *AddC;
5300 BinaryOperator *AddInst;
5301 if (match(&I,
5303 m_BinOp(AddInst),
5304 m_Add(m_Value(X), m_ImmConstant(AddC)))),
5305 m_Value(Mask))),
5306 m_Deferred(Mask)))) {
5307 Value *NotC = Builder.CreateNot(AddC);
5308 Value *NewSub = Builder.CreateSub(NotC, X, "", AddInst->hasNoUnsignedWrap(),
5309 AddInst->hasNoSignedWrap());
5310 return BinaryOperator::CreateAnd(NewSub, Mask);
5311 }
5312
5313 return nullptr;
5314}
5315
5316// FIXME: We use commutative matchers (m_c_*) for some, but not all, matches
5317// here. We should standardize that construct where it is needed or choose some
5318// other way to ensure that commutated variants of patterns are not missed.
5320 if (Value *V = simplifyXorInst(I.getOperand(0), I.getOperand(1),
5321 SQ.getWithInstruction(&I)))
5322 return replaceInstUsesWith(I, V);
5323
5325 return &I;
5326
5328 return X;
5329
5331 return Phi;
5332
5333 if (Instruction *NewXor = foldXorToXor(I, Builder))
5334 return NewXor;
5335
5336 // (A&B)^(A&C) -> A&(B^C) etc
5338 return replaceInstUsesWith(I, V);
5339
5340 // See if we can simplify any instructions used by the instruction whose sole
5341 // purpose is to compute bits we don't care about.
5343 return &I;
5344
5345 if (Instruction *R = foldNot(I))
5346 return R;
5347
5349 return R;
5350
5351 Value *Op0 = I.getOperand(0), *Op1 = I.getOperand(1);
5352 Value *X, *Y, *M;
5353
5354 // (X | Y) ^ M -> (X ^ M) ^ Y
5355 // (X | Y) ^ M -> (Y ^ M) ^ X
5357 m_Value(M)))) {
5358 if (Value *XorAC = simplifyXorInst(X, M, SQ.getWithInstruction(&I)))
5359 return BinaryOperator::CreateXor(XorAC, Y);
5360
5361 if (Value *XorBC = simplifyXorInst(Y, M, SQ.getWithInstruction(&I)))
5362 return BinaryOperator::CreateXor(XorBC, X);
5363 }
5364
5365 // Fold (X & M) ^ (Y & ~M) -> (X & M) | (Y & ~M)
5366 // This it a special case in haveNoCommonBitsSet, but the computeKnownBits
5367 // calls in there are unnecessary as SimplifyDemandedInstructionBits should
5368 // have already taken care of those cases.
5369 if (match(&I, m_c_Xor(m_c_And(m_Not(m_Value(M)), m_Value()),
5370 m_c_And(m_Deferred(M), m_Value())))) {
5372 return BinaryOperator::CreateDisjointOr(Op0, Op1);
5373 else
5374 return BinaryOperator::CreateOr(Op0, Op1);
5375 }
5376
5378 return Xor;
5379
5380 Constant *C1;
5381 if (match(Op1, m_Constant(C1))) {
5382 Constant *C2;
5383
5384 if (match(Op0, m_OneUse(m_Or(m_Value(X), m_ImmConstant(C2)))) &&
5385 match(C1, m_ImmConstant())) {
5386 // (X | C2) ^ C1 --> (X & ~C2) ^ (C1^C2)
5389 Value *And = Builder.CreateAnd(
5391 return BinaryOperator::CreateXor(
5393 }
5394
5395 // Use DeMorgan and reassociation to eliminate a 'not' op.
5396 if (match(Op0, m_OneUse(m_Or(m_Not(m_Value(X)), m_Constant(C2))))) {
5397 // (~X | C2) ^ C1 --> ((X & ~C2) ^ -1) ^ C1 --> (X & ~C2) ^ ~C1
5398 Value *And = Builder.CreateAnd(X, ConstantExpr::getNot(C2));
5399 return BinaryOperator::CreateXor(And, ConstantExpr::getNot(C1));
5400 }
5401 if (match(Op0, m_OneUse(m_And(m_Not(m_Value(X)), m_Constant(C2))))) {
5402 // (~X & C2) ^ C1 --> ((X | ~C2) ^ -1) ^ C1 --> (X | ~C2) ^ ~C1
5403 Value *Or = Builder.CreateOr(X, ConstantExpr::getNot(C2));
5404 return BinaryOperator::CreateXor(Or, ConstantExpr::getNot(C1));
5405 }
5406
5407 // Convert xor ([trunc] (ashr X, BW-1)), C =>
5408 // select(X >s -1, C, ~C)
5409 // The ashr creates "AllZeroOrAllOne's", which then optionally inverses the
5410 // constant depending on whether this input is less than 0.
5411 const APInt *CA;
5412 if (match(Op0, m_OneUse(m_TruncOrSelf(
5413 m_AShr(m_Value(X), m_APIntAllowPoison(CA))))) &&
5414 *CA == X->getType()->getScalarSizeInBits() - 1 &&
5415 !match(C1, m_AllOnes())) {
5416 assert(!C1->isNullValue() && "Unexpected xor with 0");
5417 Value *IsNotNeg = Builder.CreateIsNotNeg(X);
5418 return createSelectInstWithUnknownProfile(IsNotNeg, Op1,
5419 Builder.CreateNot(Op1));
5420 }
5421 }
5422
5423 Type *Ty = I.getType();
5424 {
5425 const APInt *RHSC;
5426 if (match(Op1, m_APInt(RHSC))) {
5427 Value *X;
5428 const APInt *C;
5429 // (C - X) ^ signmaskC --> (C + signmaskC) - X
5430 if (RHSC->isSignMask() && match(Op0, m_Sub(m_APInt(C), m_Value(X))))
5431 return BinaryOperator::CreateSub(ConstantInt::get(Ty, *C + *RHSC), X);
5432
5433 // (X + C) ^ signmaskC --> X + (C + signmaskC)
5434 if (RHSC->isSignMask() && match(Op0, m_Add(m_Value(X), m_APInt(C))))
5435 return BinaryOperator::CreateAdd(X, ConstantInt::get(Ty, *C + *RHSC));
5436
5437 // (X | C) ^ RHSC --> X ^ (C ^ RHSC) iff X & C == 0
5438 if (match(Op0, m_Or(m_Value(X), m_APInt(C))) &&
5439 MaskedValueIsZero(X, *C, &I))
5440 return BinaryOperator::CreateXor(X, ConstantInt::get(Ty, *C ^ *RHSC));
5441
5442 // When X is a power-of-two or zero and zero input is poison:
5443 // ctlz(i32 X) ^ 31 --> cttz(X)
5444 // cttz(i32 X) ^ 31 --> ctlz(X)
5445 auto *II = dyn_cast<IntrinsicInst>(Op0);
5446 if (II && II->hasOneUse() && *RHSC == Ty->getScalarSizeInBits() - 1) {
5447 Intrinsic::ID IID = II->getIntrinsicID();
5448 if ((IID == Intrinsic::ctlz || IID == Intrinsic::cttz) &&
5449 match(II->getArgOperand(1), m_One()) &&
5450 isKnownToBeAPowerOfTwo(II->getArgOperand(0), /*OrZero */ true)) {
5451 IID = (IID == Intrinsic::ctlz) ? Intrinsic::cttz : Intrinsic::ctlz;
5452 Function *F =
5453 Intrinsic::getOrInsertDeclaration(II->getModule(), IID, Ty);
5454 return CallInst::Create(F, {II->getArgOperand(0), Builder.getTrue()});
5455 }
5456 }
5457
5458 // If RHSC is inverting the remaining bits of shifted X,
5459 // canonicalize to a 'not' before the shift to help SCEV and codegen:
5460 // (X << C) ^ RHSC --> ~X << C
5461 if (match(Op0, m_OneUse(m_Shl(m_Value(X), m_APInt(C)))) &&
5462 *RHSC == APInt::getAllOnes(Ty->getScalarSizeInBits()).shl(*C)) {
5463 Value *NotX = Builder.CreateNot(X);
5464 return BinaryOperator::CreateShl(NotX, ConstantInt::get(Ty, *C));
5465 }
5466 // (X >>u C) ^ RHSC --> ~X >>u C
5467 if (match(Op0, m_OneUse(m_LShr(m_Value(X), m_APInt(C)))) &&
5468 *RHSC == APInt::getAllOnes(Ty->getScalarSizeInBits()).lshr(*C)) {
5469 Value *NotX = Builder.CreateNot(X);
5470 return BinaryOperator::CreateLShr(NotX, ConstantInt::get(Ty, *C));
5471 }
5472 // TODO: We could handle 'ashr' here as well. That would be matching
5473 // a 'not' op and moving it before the shift. Doing that requires
5474 // preventing the inverse fold in canShiftBinOpWithConstantRHS().
5475 }
5476
5477 // If we are XORing the sign bit of a floating-point value, convert
5478 // this to fneg, then cast back to integer.
5479 //
5480 // This is generous interpretation of noimplicitfloat, this is not a true
5481 // floating-point operation.
5482 //
5483 // Assumes any IEEE-represented type has the sign bit in the high bit.
5484 // TODO: Unify with APInt matcher. This version allows undef unlike m_APInt
5485 Value *CastOp;
5486 if (match(Op0, m_ElementWiseBitCast(m_Value(CastOp))) &&
5487 match(Op1, m_SignMask()) &&
5488 !Builder.GetInsertBlock()->getParent()->hasFnAttribute(
5489 Attribute::NoImplicitFloat)) {
5490 Type *EltTy = CastOp->getType()->getScalarType();
5491 if (EltTy->isFloatingPointTy() &&
5493 Value *FNeg = Builder.CreateFNeg(CastOp);
5494 return new BitCastInst(FNeg, I.getType());
5495 }
5496 }
5497 }
5498
5499 // FIXME: This should not be limited to scalar (pull into APInt match above).
5500 {
5501 Value *X;
5502 ConstantInt *C1, *C2, *C3;
5503 // ((X^C1) >> C2) ^ C3 -> (X>>C2) ^ ((C1>>C2)^C3)
5504 if (match(Op1, m_ConstantInt(C3)) &&
5506 m_ConstantInt(C2))) &&
5507 Op0->hasOneUse()) {
5508 // fold (C1 >> C2) ^ C3
5509 APInt FoldConst = C1->getValue().lshr(C2->getValue());
5510 FoldConst ^= C3->getValue();
5511 // Prepare the two operands.
5512 auto *Opnd0 = Builder.CreateLShr(X, C2);
5513 Opnd0->takeName(Op0);
5514 return BinaryOperator::CreateXor(Opnd0, ConstantInt::get(Ty, FoldConst));
5515 }
5516 }
5517
5518 if (Instruction *FoldedLogic = foldBinOpIntoSelectOrPhi(I))
5519 return FoldedLogic;
5520
5521 if (Instruction *FoldedLogic = foldBinOpSelectBinOp(I))
5522 return FoldedLogic;
5523
5524 // Y ^ (X | Y) --> X & ~Y
5525 // Y ^ (Y | X) --> X & ~Y
5526 if (match(Op1, m_OneUse(m_c_Or(m_Value(X), m_Specific(Op0)))))
5527 return BinaryOperator::CreateAnd(X, Builder.CreateNot(Op0));
5528 // (X | Y) ^ Y --> X & ~Y
5529 // (Y | X) ^ Y --> X & ~Y
5530 if (match(Op0, m_OneUse(m_c_Or(m_Value(X), m_Specific(Op1)))))
5531 return BinaryOperator::CreateAnd(X, Builder.CreateNot(Op1));
5532
5533 // Y ^ (X & Y) --> ~X & Y
5534 // Y ^ (Y & X) --> ~X & Y
5535 if (match(Op1, m_OneUse(m_c_And(m_Value(X), m_Specific(Op0)))))
5536 return BinaryOperator::CreateAnd(Op0, Builder.CreateNot(X));
5537 // (X & Y) ^ Y --> ~X & Y
5538 // (Y & X) ^ Y --> ~X & Y
5539 // Canonical form is (X & C) ^ C; don't touch that.
5540 // TODO: A 'not' op is better for analysis and codegen, but demanded bits must
5541 // be fixed to prefer that (otherwise we get infinite looping).
5542 if (!match(Op1, m_Constant()) &&
5543 match(Op0, m_OneUse(m_c_And(m_Value(X), m_Specific(Op1)))))
5544 return BinaryOperator::CreateAnd(Op1, Builder.CreateNot(X));
5545
5546 Value *A, *B, *C;
5547 // (A ^ B) ^ (A | C) --> (~A & C) ^ B -- There are 4 commuted variants.
5550 return BinaryOperator::CreateXor(
5551 Builder.CreateAnd(Builder.CreateNot(A), C), B);
5552
5553 // (A ^ B) ^ (B | C) --> (~B & C) ^ A -- There are 4 commuted variants.
5556 return BinaryOperator::CreateXor(
5557 Builder.CreateAnd(Builder.CreateNot(B), C), A);
5558
5559 // (A & B) ^ (A ^ B) -> (A | B)
5560 if (match(Op0, m_And(m_Value(A), m_Value(B))) &&
5562 return BinaryOperator::CreateOr(A, B);
5563 // (A ^ B) ^ (A & B) -> (A | B)
5564 if (match(Op0, m_Xor(m_Value(A), m_Value(B))) &&
5566 return BinaryOperator::CreateOr(A, B);
5567
5568 // (A & ~B) ^ ~A -> ~(A & B)
5569 // (~B & A) ^ ~A -> ~(A & B)
5570 if (match(Op0, m_c_And(m_Value(A), m_Not(m_Value(B)))) &&
5571 match(Op1, m_Not(m_Specific(A))))
5572 return BinaryOperator::CreateNot(Builder.CreateAnd(A, B));
5573
5574 // (~A & B) ^ A --> A | B -- There are 4 commuted variants.
5576 return BinaryOperator::CreateOr(A, B);
5577
5578 // (~A | B) ^ A --> ~(A & B)
5579 if (match(Op0, m_OneUse(m_c_Or(m_Not(m_Specific(Op1)), m_Value(B)))))
5580 return BinaryOperator::CreateNot(Builder.CreateAnd(Op1, B));
5581
5582 // A ^ (~A | B) --> ~(A & B)
5583 if (match(Op1, m_OneUse(m_c_Or(m_Not(m_Specific(Op0)), m_Value(B)))))
5584 return BinaryOperator::CreateNot(Builder.CreateAnd(Op0, B));
5585
5586 // (A | B) ^ (A | C) --> (B ^ C) & ~A -- There are 4 commuted variants.
5587 // TODO: Loosen one-use restriction if common operand is a constant.
5588 Value *D;
5589 if (match(Op0, m_OneUse(m_Or(m_Value(A), m_Value(B)))) &&
5590 match(Op1, m_OneUse(m_Or(m_Value(C), m_Value(D))))) {
5591 if (B == C || B == D)
5592 std::swap(A, B);
5593 if (A == C)
5594 std::swap(C, D);
5595 if (A == D) {
5596 Value *NotA = Builder.CreateNot(A);
5597 return BinaryOperator::CreateAnd(Builder.CreateXor(B, C), NotA);
5598 }
5599 }
5600
5601 // (A & B) ^ (A | C) --> A ? ~B : C -- There are 4 commuted variants.
5602 if (I.getType()->isIntOrIntVectorTy(1) &&
5605 bool NeedFreeze = isa<SelectInst>(Op0) && isa<SelectInst>(Op1) && B == D;
5606 Instruction *MDFrom = cast<Instruction>(Op0);
5607 if (B == C || B == D) {
5608 std::swap(A, B);
5609 MDFrom = B == C ? cast<Instruction>(Op1) : nullptr;
5610 }
5611 if (A == C)
5612 std::swap(C, D);
5613 if (A == D) {
5614 if (NeedFreeze)
5615 A = Builder.CreateFreeze(A);
5616 Value *NotB = Builder.CreateNot(B);
5617 return MDFrom == nullptr || ProfcheckDisableMetadataFixes
5618 ? createSelectInstWithUnknownProfile(A, NotB, C)
5619 : SelectInst::Create(A, NotB, C, "", nullptr, MDFrom);
5620 }
5621 }
5622
5623 if (auto *LHS = dyn_cast<ICmpInst>(I.getOperand(0)))
5624 if (auto *RHS = dyn_cast<ICmpInst>(I.getOperand(1)))
5625 if (Value *V = foldXorOfICmps(LHS, RHS, I))
5626 return replaceInstUsesWith(I, V);
5627
5628 if (Instruction *CastedXor = foldCastedBitwiseLogic(I))
5629 return CastedXor;
5630
5631 if (Instruction *Abs = canonicalizeAbs(I, Builder))
5632 return Abs;
5633
5634 // Otherwise, if all else failed, try to hoist the xor-by-constant:
5635 // (X ^ C) ^ Y --> (X ^ Y) ^ C
5636 // Just like we do in other places, we completely avoid the fold
5637 // for constantexprs, at least to avoid endless combine loop.
5639 m_ImmConstant(C1))),
5640 m_Value(Y))))
5641 return BinaryOperator::CreateXor(Builder.CreateXor(X, Y), C1);
5642
5644 return R;
5645
5646 if (Instruction *Canonicalized = canonicalizeLogicFirst(I, Builder))
5647 return Canonicalized;
5648
5649 if (Instruction *Folded = foldLogicOfIsFPClass(I, Op0, Op1))
5650 return Folded;
5651
5652 if (Instruction *Folded = canonicalizeConditionalNegationViaMathToSelect(I))
5653 return Folded;
5654
5655 if (Instruction *Res = foldBinOpOfDisplacedShifts(I))
5656 return Res;
5657
5659 return Res;
5660
5662 return Res;
5663
5664 return nullptr;
5665}
assert(UImm &&(UImm !=~static_cast< T >(0)) &&"Invalid immediate!")
AMDGPU Register Bank Select
MachineBasicBlock MachineBasicBlock::iterator DebugLoc DL
#define X(NUM, ENUM, NAME)
Definition ELF.h:851
static GCRegistry::Add< ErlangGC > A("erlang", "erlang-compatible garbage collector")
static GCRegistry::Add< StatepointGC > D("statepoint-example", "an example strategy for statepoint")
static GCRegistry::Add< CoreCLRGC > E("coreclr", "CoreCLR-compatible GC")
static GCRegistry::Add< OcamlGC > B("ocaml", "ocaml 3.10-compatible GC")
#define DEBUG_TYPE
static Value * foldAndOrOfICmpsWithConstEq(ICmpInst *Cmp0, ICmpInst *Cmp1, bool IsAnd, bool IsLogical, InstCombiner::BuilderTy &Builder, const SimplifyQuery &Q, Instruction &I)
Reduce logic-of-compares with equality to a constant by substituting a common operand with the consta...
static Value * foldIsPowerOf2OrZero(ICmpInst *Cmp0, ICmpInst *Cmp1, bool IsAnd, InstCombiner::BuilderTy &Builder, InstCombinerImpl &IC)
Fold (icmp eq ctpop(X) 1) | (icmp eq X 0) into (icmp ult ctpop(X) 2) and fold (icmp ne ctpop(X) 1) & ...
static Value * foldBitmaskMul(Value *Op0, Value *Op1, InstCombiner::BuilderTy &Builder)
(A & N) * C + (A & M) * C -> (A & (N + M)) & C This also accepts the equivalent select form of (A & N...
static unsigned conjugateICmpMask(unsigned Mask)
Convert an analysis of a masked ICmp into its equivalent if all boolean operations had the opposite s...
static Instruction * foldNotXor(BinaryOperator &I, InstCombiner::BuilderTy &Builder)
static Value * foldLogOpOfMaskedICmps(Value *LHS, Value *RHS, bool IsAnd, bool IsLogical, InstCombiner::BuilderTy &Builder, const SimplifyQuery &Q)
Try to fold (icmp(A & B) ==/!= C) &/| (icmp(A & D) ==/!= E) into a single (icmp(A & X) ==/!...
static Value * getFCmpValue(unsigned Code, Value *LHS, Value *RHS, InstCombiner::BuilderTy &Builder, FMFSource FMF)
This is the complement of getFCmpCode, which turns an opcode and two operands into either a FCmp inst...
static bool matchIsFPClassLikeFCmp(Value *Op, Value *&ClassVal, uint64_t &ClassMask)
Match an fcmp against a special value that performs a test possible by llvm.is.fpclass.
static Value * foldSignedTruncationCheck(ICmpInst *ICmp0, ICmpInst *ICmp1, Instruction &CxtI, InstCombiner::BuilderTy &Builder)
General pattern: X & Y.
static Instruction * visitMaskedMerge(BinaryOperator &I, InstCombiner::BuilderTy &Builder)
If we have a masked merge, in the canonical form of: (assuming that A only has one use....
static Instruction * canonicalizeAbs(BinaryOperator &Xor, InstCombiner::BuilderTy &Builder)
Canonicalize a shifty way to code absolute value to the more common pattern that uses negation and se...
static Value * foldIsPowerOf2(ICmpInst *Cmp0, ICmpInst *Cmp1, bool JoinedByAnd, InstCombiner::BuilderTy &Builder, InstCombinerImpl &IC)
Reduce a pair of compares that check if a value has exactly 1 bit set.
static Value * foldUnsignedUnderflowCheck(ICmpInst *ZeroICmp, ICmpInst *UnsignedICmp, bool IsAnd, const SimplifyQuery &Q, InstCombiner::BuilderTy &Builder)
Commuted variants are assumed to be handled by calling this function again with the parameters swappe...
static Instruction * foldOrToXor(BinaryOperator &I, InstCombiner::BuilderTy &Builder)
static Value * simplifyAndOrWithOpReplaced(Value *V, Value *Op, Value *RepOp, bool SimplifyOnly, InstCombinerImpl &IC, unsigned Depth=0)
static Instruction * matchDeMorgansLaws(BinaryOperator &I, InstCombiner &IC)
Match variations of De Morgan's Laws: (~A & ~B) == (~(A | B)) (~A | ~B) == (~(A & B))
static Value * foldLogOpOfMaskedICmpsAsymmetric(Value *LHS, Value *RHS, bool IsAnd, Value *A, Value *B, Value *C, Value *D, Value *E, ICmpInst::Predicate PredL, ICmpInst::Predicate PredR, unsigned LHSMask, unsigned RHSMask, InstCombiner::BuilderTy &Builder)
Try to fold (icmp(A & B) ==/!= 0) &/| (icmp(A & D) ==/!= E) into a single (icmp(A & X) ==/!...
static Value * FoldOrOfSelectSmaxToAbs(BinaryOperator &I, InstCombiner::BuilderTy &Builder)
Fold select(X >s 0, 0, -X) | smax(X, 0) --> abs(X) select(X <s 0, -X, 0) | smax(X,...
static Instruction * foldAndToXor(BinaryOperator &I, InstCombiner::BuilderTy &Builder)
static unsigned getMaskedICmpType(Value *A, Value *B, Value *C, ICmpInst::Predicate Pred)
Return the set of patterns (from MaskedICmpType) that (icmp SCC (A & B), C) satisfies.
static Instruction * foldXorToXor(BinaryOperator &I, InstCombiner::BuilderTy &Builder)
A ^ B can be specified using other logic ops in a variety of patterns.
static bool canNarrowShiftAmt(Constant *C, unsigned BitWidth)
Return true if a constant shift amount is always less than the specified bit-width.
static Instruction * foldLogicCastConstant(BinaryOperator &Logic, CastInst *Cast, InstCombinerImpl &IC)
Fold {and,or,xor} (cast X), C.
static Value * foldAndOrOfICmpEqConstantAndICmp(ICmpInst *LHS, ICmpInst *RHS, bool IsAnd, bool IsLogical, IRBuilderBase &Builder)
static bool canFreelyInvert(InstCombiner &IC, Value *Op, Instruction *IgnoredUser)
static Value * foldNegativePower2AndShiftedMask(Value *A, Value *B, Value *D, Value *E, ICmpInst::Predicate PredL, ICmpInst::Predicate PredR, InstCombiner::BuilderTy &Builder)
Try to fold (icmp(A & B) == 0) & (icmp(A & D) != E) into (icmp A u< D) iff B is a contiguous set of o...
static Value * matchIsFiniteTest(InstCombiner::BuilderTy &Builder, FCmpInst *LHS, FCmpInst *RHS)
and (fcmp ord x, 0), (fcmp u* x, inf) -> fcmp o* x, inf
static Value * foldPowerOf2AndShiftedMask(ICmpInst *Cmp0, ICmpInst *Cmp1, bool JoinedByAnd, InstCombiner::BuilderTy &Builder)
Try to fold ((icmp X u< P) & (icmp(X & M) != M)) or ((icmp X s> -1) & (icmp(X & M) !...
static Value * foldOrUnsignedUMulOverflowICmp(BinaryOperator &I, InstCombiner::BuilderTy &Builder, const DataLayout &DL)
Fold Res, Overflow = (umul.with.overflow x c1); (or Overflow (ugt Res c2)) --> (ugt x (c2/c1)).
static Value * freelyInvert(InstCombinerImpl &IC, Value *Op, Instruction *IgnoredUser)
static Value * foldLogOpOfMaskedICmps_NotAllZeros_BMask_Mixed(Value *LHS, Value *RHS, bool IsAnd, Value *A, Value *B, Value *D, Value *E, ICmpInst::Predicate PredL, ICmpInst::Predicate PredR, InstCombiner::BuilderTy &Builder)
Try to fold (icmp(A & B) ==/!= C) &/| (icmp(A & D) ==/!= E) into a single (icmp(A & X) ==/!...
static std::optional< IntPart > matchIntPart(Value *V)
Match an extraction of bits from an integer.
static Instruction * canonicalizeLogicFirst(BinaryOperator &I, InstCombiner::BuilderTy &Builder)
static Instruction * reassociateFCmps(BinaryOperator &BO, InstCombiner::BuilderTy &Builder)
This a limited reassociation for a special case (see above) where we are checking if two values are e...
static Value * getNewICmpValue(unsigned Code, bool Sign, Value *LHS, Value *RHS, InstCombiner::BuilderTy &Builder)
This is the complement of getICmpCode, which turns an opcode and two operands into either a constant ...
static Value * extractIntPart(const IntPart &P, IRBuilderBase &Builder)
Materialize an extraction of bits from an integer in IR.
static bool matchUnorderedInfCompare(FCmpInst::Predicate P, Value *LHS, Value *RHS)
Matches fcmp u__ x, +/-inf.
static bool matchIsNotNaN(FCmpInst::Predicate P, Value *LHS, Value *RHS)
Matches canonical form of isnan, fcmp ord x, 0.
static bool areInverseVectorBitmasks(Constant *C1, Constant *C2)
If all elements of two constant vectors are 0/-1 and inverses, return true.
MaskedICmpType
Classify (icmp eq (A & B), C) and (icmp ne (A & B), C) as matching patterns that can be simplified.
@ BMask_NotAllOnes
@ AMask_NotAllOnes
@ Mask_NotAllZeros
static Instruction * foldComplexAndOrPatterns(BinaryOperator &I, InstCombiner::BuilderTy &Builder)
Try folding relatively complex patterns for both And and Or operations with all And and Or swapped.
static bool matchZExtedSubInteger(Value *V, Value *&Int, APInt &Mask, uint64_t &Offset, bool &IsShlNUW, bool &IsShlNSW)
Match V as "lshr -> mask -> zext -> shl".
static std::optional< DecomposedBitMaskMul > matchBitmaskMul(Value *V)
static Value * foldOrOfInversions(BinaryOperator &I, InstCombiner::BuilderTy &Builder)
static bool matchSubIntegerPackFromVector(Value *V, Value *&Vec, int64_t &VecOffset, SmallBitVector &Mask, const DataLayout &DL)
Match V as "shufflevector -> bitcast" or "extractelement -> zext -> shl" patterns,...
static Instruction * matchFunnelShift(Instruction &Or, InstCombinerImpl &IC)
Match UB-safe variants of the funnel shift intrinsic.
static Instruction * reassociateForUses(BinaryOperator &BO, InstCombinerImpl::BuilderTy &Builder)
Try to reassociate a pair of binops so that values with one use only are part of the same instruction...
static Value * matchOrConcat(Instruction &Or, InstCombiner::BuilderTy &Builder)
Attempt to combine or(zext(x),shl(zext(y),bw/2) concat packing patterns.
static Value * foldAndOrOfICmpsWithPow2AndWithZero(InstCombiner::BuilderTy &Builder, ICmpInst *LHS, ICmpInst *RHS, bool IsAnd, const SimplifyQuery &Q)
static Instruction * foldMaskedAddXorPattern(BinaryOperator &I, InstCombiner::BuilderTy &Builder)
static Instruction * foldBitwiseLogicWithIntrinsics(BinaryOperator &I, InstCombiner::BuilderTy &Builder)
static std::optional< std::pair< unsigned, unsigned > > getMaskedTypeForICmpPair(Value *&A, Value *&B, Value *&C, Value *&D, Value *&E, Value *LHS, Value *RHS, ICmpInst::Predicate &PredL, ICmpInst::Predicate &PredR)
Handle (icmp(A & B) ==/!= C) &/| (icmp(A & D) ==/!= E).
static Instruction * foldIntegerPackFromVector(Instruction &I, InstCombiner::BuilderTy &Builder, const DataLayout &DL)
Try to fold the join of two scalar integers whose contents are packed elements of the same vector.
static Value * foldIntegerRepackThroughZExt(Value *Lhs, Value *Rhs, InstCombiner::BuilderTy &Builder)
Try to fold the join of two scalar integers whose bits are unpacked and zexted from the same source i...
This file provides internal interfaces used to implement the InstCombine.
This file provides the interface for the instcombine pass implementation.
static bool isZero(Value *V, const DataLayout &DL, DominatorTree *DT, AssumptionCache *AC)
Definition Lint.cpp:539
#define F(x, y, z)
Definition MD5.cpp:54
#define I(x, y, z)
Definition MD5.cpp:57
#define R2(n)
uint64_t High
uint64_t IntrinsicInst * II
#define P(N)
This file contains the declarations for profiling metadata utility functions.
const SmallVectorImpl< MachineOperand > & Cond
This file implements the SmallBitVector class.
static unsigned getScalarSizeInBits(Type *Ty)
static TableGen::Emitter::Opt Y("gen-skeleton-entry", EmitSkeleton, "Generate example skeleton entry")
static constexpr int Concat[]
Value * RHS
Value * LHS
The Input class is used to parse a yaml document into in-memory structs and vectors.
static LLVM_ABI bool hasSignBitInMSB(const fltSemantics &)
Definition APFloat.cpp:260
bool bitwiseIsEqual(const APFloat &RHS) const
Definition APFloat.h:1481
bool isZero() const
Definition APFloat.h:1512
APInt bitcastToAPInt() const
Definition APFloat.h:1408
static APFloat getInf(const fltSemantics &Sem, bool Negative=false)
Factory for Positive and Negative Infinity.
Definition APFloat.h:1153
Class for arbitrary precision integers.
Definition APInt.h:78
LLVM_ABI APInt udiv(const APInt &RHS) const
Unsigned division operation.
Definition APInt.cpp:1615
static APInt getAllOnes(unsigned numBits)
Return an APInt of a specified width with all bits set.
Definition APInt.h:235
LLVM_ABI APInt zext(unsigned width) const
Zero extend to a new width.
Definition APInt.cpp:1054
uint64_t getZExtValue() const
Get zero extended value.
Definition APInt.h:1563
LLVM_ABI APInt trunc(unsigned width) const
Truncate to new width.
Definition APInt.cpp:967
unsigned countLeadingOnes() const
Definition APInt.h:1647
bool isAllOnes() const
Determine if all bits are set. This is true for zero-width values.
Definition APInt.h:372
LLVM_ABI APInt usub_ov(const APInt &RHS, bool &Overflow) const
Definition APInt.cpp:1999
bool ugt(const APInt &RHS) const
Unsigned greater than comparison.
Definition APInt.h:1189
bool isZero() const
Determine if this value is zero, i.e. all bits are clear.
Definition APInt.h:381
bool isSignMask() const
Check if the APInt's value is returned by getSignMask.
Definition APInt.h:467
unsigned getBitWidth() const
Return the number of bits in the APInt.
Definition APInt.h:1511
bool ult(const APInt &RHS) const
Unsigned less than comparison.
Definition APInt.h:1118
LLVM_ABI APInt sadd_ov(const APInt &RHS, bool &Overflow) const
Definition APInt.cpp:1979
bool intersects(const APInt &RHS) const
This operation tests if there are any pairs of corresponding bits between this APInt and RHS that are...
Definition APInt.h:1256
int32_t exactLogBase2() const
Definition APInt.h:1806
LLVM_ABI APInt reverseBits() const
Definition APInt.cpp:789
LLVM_ABI APInt uadd_ov(const APInt &RHS, bool &Overflow) const
Definition APInt.cpp:1986
unsigned countr_zero() const
Count the number of trailing zero bits.
Definition APInt.h:1662
unsigned countLeadingZeros() const
Definition APInt.h:1629
bool ule(const APInt &RHS) const
Unsigned less or equal comparison.
Definition APInt.h:1157
APInt shl(unsigned shiftAmt) const
Left-shift function.
Definition APInt.h:880
LLVM_ABI APInt byteSwap() const
Definition APInt.cpp:767
bool isSubsetOf(const APInt &RHS) const
This operation checks that all bits set in this APInt are also set in RHS.
Definition APInt.h:1264
bool isPowerOf2() const
Check if this APInt's value is a power of two greater than zero.
Definition APInt.h:441
static APInt getLowBitsSet(unsigned numBits, unsigned loBitsSet)
Constructs an APInt value that has the bottom loBitsSet bits set.
Definition APInt.h:307
LLVM_ABI APInt ssub_ov(const APInt &RHS, bool &Overflow) const
Definition APInt.cpp:1992
static APInt getBitsSetFrom(unsigned numBits, unsigned loBit)
Constructs an APInt value that has a contiguous range of bits set.
Definition APInt.h:287
APInt lshr(unsigned shiftAmt) const
Logical right-shift function.
Definition APInt.h:858
bool uge(const APInt &RHS) const
Unsigned greater or equal comparison.
Definition APInt.h:1228
void clearSignBit()
Set the sign bit to 0.
Definition APInt.h:1472
ArrayRef - Represent a constant reference to an array (0 or more elements consecutively in memory),...
Definition ArrayRef.h:40
LLVM_ABI bool isSigned() const
Whether the intrinsic is signed or unsigned.
LLVM_ABI Instruction::BinaryOps getBinaryOp() const
Returns the binary operation underlying the intrinsic.
BinaryOps getOpcode() const
Definition InstrTypes.h:374
static LLVM_ABI BinaryOperator * CreateNot(Value *Op, const Twine &Name="", InsertPosition InsertBefore=nullptr)
static LLVM_ABI BinaryOperator * Create(BinaryOps Op, Value *S1, Value *S2, const Twine &Name=Twine(), InsertPosition InsertBefore=nullptr)
Construct a binary instruction, given the opcode and the two operands.
static BinaryOperator * CreateWithCopiedFlags(BinaryOps Opc, Value *V1, Value *V2, Value *CopyO, const Twine &Name="", InsertPosition InsertBefore=nullptr)
Definition InstrTypes.h:219
This class represents a no-op cast from one type to another.
static CallInst * Create(FunctionType *Ty, Value *F, const Twine &NameStr="", InsertPosition InsertBefore=nullptr)
This is the base class for all instructions that perform data casts.
Definition InstrTypes.h:448
Type * getSrcTy() const
Return the source type, as a convenience.
Definition InstrTypes.h:615
Instruction::CastOps getOpcode() const
Return the opcode of this CastInst.
Definition InstrTypes.h:610
static LLVM_ABI CastInst * Create(Instruction::CastOps, Value *S, Type *Ty, const Twine &Name="", InsertPosition InsertBefore=nullptr)
Provides a way to construct any of the CastInst subclasses using an opcode instead of the subclass's ...
Type * getDestTy() const
Return the destination type, as a convenience.
Definition InstrTypes.h:617
static Type * makeCmpResultType(Type *opnd_type)
Create a result type for fcmp/icmp.
Definition InstrTypes.h:986
Predicate
This enumeration lists the possible predicates for CmpInst subclasses.
Definition InstrTypes.h:676
@ ICMP_SLT
signed less than
Definition InstrTypes.h:705
@ ICMP_SLE
signed less or equal
Definition InstrTypes.h:706
@ FCMP_OLT
0 1 0 0 True if ordered and less than
Definition InstrTypes.h:682
@ FCMP_ULE
1 1 0 1 True if unordered, less than, or equal
Definition InstrTypes.h:691
@ ICMP_UGE
unsigned greater or equal
Definition InstrTypes.h:700
@ ICMP_UGT
unsigned greater than
Definition InstrTypes.h:699
@ ICMP_SGT
signed greater than
Definition InstrTypes.h:703
@ FCMP_ULT
1 1 0 0 True if unordered or less than
Definition InstrTypes.h:690
@ ICMP_ULT
unsigned less than
Definition InstrTypes.h:701
@ FCMP_OLE
0 1 0 1 True if ordered and less than or equal
Definition InstrTypes.h:683
@ FCMP_ORD
0 1 1 1 True if ordered (no nans)
Definition InstrTypes.h:685
@ ICMP_NE
not equal
Definition InstrTypes.h:698
@ ICMP_SGE
signed greater or equal
Definition InstrTypes.h:704
@ ICMP_ULE
unsigned less or equal
Definition InstrTypes.h:702
@ FCMP_UNO
1 0 0 0 True if unordered: isnan(X) | isnan(Y)
Definition InstrTypes.h:686
bool isSigned() const
Definition InstrTypes.h:930
Predicate getSwappedPredicate() const
For example, EQ->EQ, SLE->SGE, ULT->UGT, OEQ->OEQ, ULE->UGE, OLT->OGT, etc.
Definition InstrTypes.h:827
Predicate getInversePredicate() const
For example, EQ -> NE, UGT -> ULE, SLT -> SGE, OEQ -> UNE, UGT -> OLE, OLT -> UGE,...
Definition InstrTypes.h:789
Predicate getPredicate() const
Return the predicate for this instruction.
Definition InstrTypes.h:765
static LLVM_ABI bool isUnordered(Predicate predicate)
Determine if the predicate is an unordered operation.
static Predicate getOrderedPredicate(Predicate Pred)
Returns the ordered variant of a floating point compare.
Definition InstrTypes.h:796
An abstraction over a floating-point predicate, and a pack of an integer predicate with samesign info...
static LLVM_ABI Constant * getSub(Constant *C1, Constant *C2, bool HasNUW=false, bool HasNSW=false)
static LLVM_ABI Constant * getNot(Constant *C)
static LLVM_ABI Constant * getXor(Constant *C1, Constant *C2)
static LLVM_ABI Constant * getAdd(Constant *C1, Constant *C2, bool HasNUW=false, bool HasNSW=false)
static LLVM_ABI Constant * getTrunc(Constant *C, Type *Ty, bool OnlyIfReduced=false)
static LLVM_ABI Constant * getExactLogBase2(Constant *C)
If C is a scalar/fixed width vector of known powers of 2, then this function returns a new scalar/fix...
static LLVM_ABI Constant * getZero(Type *Ty, bool Negative=false)
This is the shared class of boolean and integer constants.
Definition Constants.h:87
bool isMinusOne() const
This function will return true iff every bit in this constant is set to true.
Definition Constants.h:231
static LLVM_ABI ConstantInt * getTrue(LLVMContext &Context)
bool isZero() const
This is just a convenience method to make client code smaller for a common code.
Definition Constants.h:219
static LLVM_ABI ConstantInt * getFalse(LLVMContext &Context)
uint64_t getZExtValue() const
Return the constant as a 64-bit unsigned integer value after it has been zero extended as appropriate...
Definition Constants.h:168
const APInt & getValue() const
Return the constant as an APInt value reference.
Definition Constants.h:159
LLVM_ABI std::optional< ConstantRange > exactUnionWith(const ConstantRange &CR) const
Union the two ranges and return the result if it can be represented exactly, otherwise return std::nu...
LLVM_ABI ConstantRange subtract(const APInt &CI) const
Subtract the specified constant from the endpoints of this constant range.
static LLVM_ABI ConstantRange makeExactICmpRegion(CmpInst::Predicate Pred, const APInt &Other)
Produce the exact range such that all values in the returned range satisfy the given predicate with a...
LLVM_ABI std::optional< ConstantRange > exactIntersectWith(const ConstantRange &CR) const
Intersect the two ranges and return the result if it can be represented exactly, otherwise return std...
This is an important base class in LLVM.
Definition Constant.h:43
static LLVM_ABI Constant * replaceUndefsWith(Constant *C, Constant *Replacement)
Try to replace undefined constant C or undefined elements in C with Replacement.
static LLVM_ABI Constant * mergeUndefsWith(Constant *C, Constant *Other)
Merges undefs of a Constant with another Constant, along with the undefs already present.
static LLVM_ABI Constant * getAllOnesValue(Type *Ty)
static LLVM_ABI Constant * getNullValue(Type *Ty)
Constructor to create a '0' constant of arbitrary type.
LLVM_ABI Constant * getAggregateElement(unsigned Elt) const
For aggregates (struct/array/vector) return the constant that corresponds to the specified element if...
LLVM_ABI bool isNullValue() const
Return true if this is the value that would be returned by getNullValue.
Definition Constants.cpp:74
A parsed version of the target data layout string in and methods for querying it.
Definition DataLayout.h:64
This instruction compares its operands according to the predicate given to the constructor.
This provides a helper for copying FMF from an instruction or setting specified flags.
Definition IRBuilder.h:93
static FMFSource intersect(Value *A, Value *B)
Intersect the FMF from two instructions.
Definition IRBuilder.h:107
This instruction compares its operands according to the predicate given to the constructor.
Predicate getSignedPredicate() const
For example, EQ->EQ, SLE->SLE, UGT->SGT, etc.
bool isEquality() const
Return true if this predicate is either EQ or NE.
static bool isEquality(Predicate P)
Return true if this predicate is either EQ or NE.
Common base class shared among various IRBuilders.
Definition IRBuilder.h:114
Value * CreateNot(Value *V, const Twine &Name="")
Definition IRBuilder.h:1849
Value * CreateBinOp(Instruction::BinaryOps Opc, Value *LHS, Value *RHS, const Twine &Name="", MDNode *FPMathTag=nullptr)
Definition IRBuilder.h:1748
void SetInsertPoint(BasicBlock *TheBB)
This specifies that created instructions should be appended to the end of the specified block.
Definition IRBuilder.h:207
Instruction * canonicalizeCondSignextOfHighBitExtractToSignextHighBitExtract(BinaryOperator &I)
Instruction * foldBinOpIntoSelectOrPhi(BinaryOperator &I)
This is a convenience wrapper function for the above two functions.
Instruction * visitOr(BinaryOperator &I)
bool SimplifyAssociativeOrCommutative(BinaryOperator &I)
Performs a few simplifications for operators which are associative or commutative.
Value * foldUsingDistributiveLaws(BinaryOperator &I)
Tries to simplify binary operations which some other binary operation distributes over.
Instruction * foldBinOpShiftWithShift(BinaryOperator &I)
Value * insertRangeTest(Value *V, const APInt &Lo, const APInt &Hi, bool isSigned, bool Inside)
Emit a computation of: (V >= Lo && V < Hi) if Inside is true, otherwise (V < Lo || V >= Hi).
Instruction * foldBinOpSelectBinOp(BinaryOperator &Op)
In some cases it is beneficial to fold a select into a binary operator.
bool sinkNotIntoLogicalOp(Instruction &I)
std::optional< std::pair< Intrinsic::ID, SmallVector< Value *, 3 > > > convertOrOfShiftsToFunnelShift(Instruction &Or)
Instruction * visitAnd(BinaryOperator &I)
bool sinkNotIntoOtherHandOfLogicalOp(Instruction &I)
Instruction * foldBinopWithPhiOperands(BinaryOperator &BO)
For a binary operator with 2 phi operands, try to hoist the binary operation before the phi.
Instruction * foldAddLikeCommutative(Value *LHS, Value *RHS, bool NSW, bool NUW)
Common transforms for add / disjoint or.
Value * simplifyRangeCheck(ICmpInst *Cmp0, ICmpInst *Cmp1, bool Inverted)
Try to fold a signed range checked with lower bound 0 to an unsigned icmp.
Instruction * tryFoldInstWithCtpopWithNot(Instruction *I)
Instruction * FoldOrOfLogicalAnds(Value *Op0, Value *Op1)
Value * SimplifyAddWithRemainder(BinaryOperator &I)
Tries to simplify add operations using the definition of remainder.
Instruction * visitXor(BinaryOperator &I)
bool SimplifyDemandedInstructionBits(Instruction &Inst)
Tries to simplify operands to an integer instruction based on its demanded bits.
Instruction * foldVectorBinop(BinaryOperator &Inst)
Canonicalize the position of binops relative to shufflevector.
Instruction * matchBSwapOrBitReverse(Instruction &I, bool MatchBSwaps, bool MatchBitReversals)
Given an initial instruction, check to see if it is the root of a bswap/bitreverse idiom.
void freelyInvertAllUsersOf(Value *V, Value *IgnoredUser=nullptr)
Freely adapt every user of V as-if V was changed to !V.
The core instruction combiner logic.
SimplifyQuery SQ
const DataLayout & getDataLayout() const
IRBuilder< TargetFolder, IRBuilderCallbackInserter > BuilderTy
An IRBuilder that automatically inserts new instructions into the worklist.
bool isFreeToInvert(Value *V, bool WillInvertAllUses, bool &DoesConsume)
Return true if the specified value is free to invert (apply ~ to).
unsigned ComputeNumSignBits(const Value *Op, const Instruction *CxtI=nullptr, unsigned Depth=0) const
Instruction * replaceInstUsesWith(Instruction &I, Value *V)
A combiner-aware RAUW-like routine.
InstructionWorklist & Worklist
A worklist of the instructions that need to be simplified.
const DataLayout & DL
void computeKnownBits(const Value *V, KnownBits &Known, const Instruction *CxtI, unsigned Depth=0) const
static Value * peekThroughBitcast(Value *V, bool OneUseOnly=false)
Return the source operand of a potentially bitcasted value while optionally checking if it has one us...
bool canFreelyInvertAllUsersOf(Instruction *V, Value *IgnoredUser)
Given i1 V, can every user of V be freely adapted if V is changed to !V ?
void addToWorklist(Instruction *I)
static Value * stripSignOnlyFPOps(Value *Val)
Ignore all operations which only change the sign of a value, returning the underlying magnitude value...
bool MaskedValueIsZero(const Value *V, const APInt &Mask, const Instruction *CxtI=nullptr, unsigned Depth=0) const
DominatorTree & DT
BuilderTy & Builder
Value * getFreelyInverted(Value *V, bool WillInvertAllUses, BuilderTy *Builder, bool &DoesConsume)
const SimplifyQuery & getSimplifyQuery() const
bool isKnownToBeAPowerOfTwo(const Value *V, bool OrZero=false, const Instruction *CxtI=nullptr, unsigned Depth=0)
LLVM_ABI void removeFromParent()
This method unlinks 'this' from the containing basic block, but does not delete it.
LLVM_ABI bool hasNoUnsignedWrap() const LLVM_READONLY
Determine whether the no unsigned wrap flag is set.
LLVM_ABI bool hasNoSignedWrap() const LLVM_READONLY
Determine whether the no signed wrap flag is set.
LLVM_ABI void swapProfMetadata()
If the instruction has "branch_weights" MD_prof metadata and the MDNode has three operands (including...
unsigned getOpcode() const
Returns a member of one of the enums like Instruction::Add.
A wrapper class for inspecting calls to intrinsic functions.
This class represents a sign extension of integer types.
This class represents the LLVM 'select' instruction.
static SelectInst * Create(Value *C, Value *S1, Value *S2, const Twine &NameStr="", InsertPosition InsertBefore=nullptr, const Instruction *MDFrom=nullptr)
This is a 'bitvector' (really, a variable-sized bit array), optimized for the case when the array is ...
This is a 'vector' (really, a variable-sized array), optimized for the case when the array is small.
The instances of the Type class are immutable: once they are created, they are never changed.
Definition Type.h:46
LLVM_ABI unsigned getIntegerBitWidth() const
bool isVectorTy() const
True if this is an instance of VectorType.
Definition Type.h:290
bool isIntOrIntVectorTy() const
Return true if this is an integer type or a vector of integer types.
Definition Type.h:263
Type * getScalarType() const
If this is a vector type, return the element type, otherwise return 'this'.
Definition Type.h:370
LLVM_ABI TypeSize getPrimitiveSizeInBits() const LLVM_READONLY
Return the basic size of this type if it is a primitive type.
Definition Type.cpp:201
LLVM_ABI Type * getWithNewBitWidth(unsigned NewBitWidth) const
Given an integer or vector type, change the lane bitwidth to NewBitwidth, whilst keeping the old numb...
LLVM_ABI unsigned getScalarSizeInBits() const LLVM_READONLY
If this is a vector type, return the getPrimitiveSizeInBits value for the element type.
Definition Type.cpp:236
bool isFloatingPointTy() const
Return true if this is one of the floating-point types.
Definition Type.h:186
LLVM_ABI const fltSemantics & getFltSemantics() const
Definition Type.cpp:110
A Use represents the edge between a Value definition and its users.
Definition Use.h:35
Value * getOperand(unsigned i) const
Definition User.h:207
LLVM Value Representation.
Definition Value.h:75
Type * getType() const
All values are typed, get the type of this value.
Definition Value.h:255
bool hasOneUse() const
Return true if there is exactly one use of this value.
Definition Value.h:439
iterator_range< user_iterator > users()
Definition Value.h:426
LLVM_ABI bool hasNUsesOrMore(unsigned N) const
Return true if this value has N uses or more.
Definition Value.cpp:154
LLVM_ABI bool hasNUses(unsigned N) const
Return true if this Value has exactly N uses.
Definition Value.cpp:146
bool use_empty() const
Definition Value.h:346
LLVM_ABI StringRef getName() const
Return a constant reference to the value's name.
Definition Value.cpp:318
LLVM_ABI void takeName(Value *V)
Transfer the name from V to this value.
Definition Value.cpp:399
static LLVM_ABI VectorType * get(Type *ElementType, ElementCount EC)
This static method is the primary way to construct an VectorType.
Represents an op.with.overflow intrinsic.
This class represents zero extension of integer types.
constexpr ScalarTy getKnownMinValue() const
Returns the minimum value this quantity can represent.
Definition TypeSize.h:165
#define llvm_unreachable(msg)
Marks that the current location is not supposed to be reachable.
const APInt & umin(const APInt &A, const APInt &B)
Determine the smaller of two APInts considered to be unsigned.
Definition APInt.h:2287
constexpr std::underlying_type_t< E > Mask()
Get a bitmask with 1s in all places up to the high-order bit of E's largest value.
@ C
The default llvm calling convention, compatible with C.
Definition CallingConv.h:34
LLVM_ABI Function * getOrInsertDeclaration(Module *M, ID id, ArrayRef< Type * > OverloadTys={})
Look up the Function declaration of the intrinsic id in the Module M.
SpecificConstantMatch m_ZeroInt()
Convenience matchers for specific integer values.
BinaryOp_match< SpecificConstantMatch, SrcTy, TargetOpcode::G_SUB > m_Neg(const SrcTy &&Src)
Matches a register negated by a G_SUB.
BinaryOp_match< SrcTy, SpecificConstantMatch, TargetOpcode::G_XOR, true > m_Not(const SrcTy &&Src)
Matches a register not-ed by a G_XOR.
OneUse_match< SubPat > m_OneUse(const SubPat &SP)
match_combine_or< Ty... > m_CombineOr(const Ty &...Ps)
Combine pattern matchers matching any of Ps patterns.
match_combine_and< Ty... > m_CombineAnd(const Ty &...Ps)
Combine pattern matchers matching all of Ps patterns.
cst_pred_ty< is_all_ones > m_AllOnes()
Match an integer or vector with all bits set.
cst_pred_ty< is_lowbit_mask > m_LowBitMask()
Match an integer or vector with only the low bit(s) set.
BinaryOp_match< LHS, RHS, Instruction::And > m_And(const LHS &L, const RHS &R)
cst_pred_ty< is_negative > m_Negative()
Match an integer or vector of negative values.
auto m_Cmp()
Matches any compare instruction and ignore it.
BinaryOp_match< LHS, RHS, Instruction::Add > m_Add(const LHS &L, const RHS &R)
CmpClass_match< LHS, RHS, FCmpInst > m_FCmp(CmpPredicate &Pred, const LHS &L, const RHS &R)
cst_pred_ty< is_sign_mask > m_SignMask()
Match an integer or vector with only the sign bit(s) set.
BinaryOp_match< LHS, RHS, Instruction::AShr > m_AShr(const LHS &L, const RHS &R)
cstfp_pred_ty< is_inf > m_Inf()
Match a positive or negative infinity FP constant.
m_Intrinsic_Ty< Opnd0 >::Ty m_BitReverse(const Opnd0 &Op0)
cst_pred_ty< is_power2 > m_Power2()
Match an integer or vector power-of-2.
match_combine_or< CastInst_match< OpTy, TruncInst >, OpTy > m_TruncOrSelf(const OpTy &Op)
auto m_LogicalOp()
Matches either L && R or L || R where L and R are arbitrary values.
ap_match< APInt > m_APInt(const APInt *&Res)
Match a ConstantInt or splatted ConstantVector, binding the specified pointer to the contained APInt.
BinaryOp_match< LHS, RHS, Instruction::And, true > m_c_And(const LHS &L, const RHS &R)
Matches an And with LHS and RHS in either order.
CastInst_match< OpTy, TruncInst > m_Trunc(const OpTy &Op)
Matches Trunc.
BinaryOp_match< LHS, RHS, Instruction::Xor > m_Xor(const LHS &L, const RHS &R)
ap_match< APInt > m_APIntAllowPoison(const APInt *&Res)
Match APInt while allowing poison in splat vector constants.
OverflowingBinaryOp_match< LHS, RHS, Instruction::Sub, OverflowingBinaryOperator::NoSignedWrap > m_NSWSub(const LHS &L, const RHS &R)
specific_intval< false > m_SpecificInt(const APInt &V)
Match a specific integer value or vector with all elements equal to the value.
match_combine_or< CastInst_match< OpTy, ZExtInst >, OpTy > m_ZExtOrSelf(const OpTy &Op)
bool match(Val *V, const Pattern &P)
match_bind< Instruction > m_Instruction(Instruction *&I)
Match an instruction, capturing it if we match.
cst_pred_ty< is_shifted_mask > m_ShiftedMask()
match_deferred< Value > m_Deferred(Value *const &V)
Like m_Specific(), but works if the specific value to match is determined as part of the same match()...
cstfp_pred_ty< is_any_zero_fp > m_AnyZeroFP()
Match a floating-point negative zero or positive zero.
specificval_ty m_Specific(const Value *V)
Match if we have a specific specified value.
DisjointOr_match< LHS, RHS > m_DisjointOr(const LHS &L, const RHS &R)
constantexpr_match m_ConstantExpr()
Match a constant expression or a constant that contains a constant expression.
specific_intval< true > m_SpecificIntAllowPoison(const APInt &V)
ap_match< APFloat > m_APFloatAllowPoison(const APFloat *&Res)
Match APFloat while allowing poison in splat vector constants.
CmpClass_match< LHS, RHS, ICmpInst, true > m_c_ICmp(CmpPredicate &Pred, const LHS &L, const RHS &R)
Matches an ICmp with a predicate over LHS and RHS in either order.
TwoOps_match< Val_t, Idx_t, Instruction::ExtractElement > m_ExtractElt(const Val_t &Val, const Idx_t &Idx)
Matches ExtractElementInst.
cst_pred_ty< is_nonnegative > m_NonNegative()
Match an integer or vector of non-negative values.
cst_pred_ty< is_one > m_One()
Match an integer 1 or a vector with all elements equal to 1.
IntrinsicID_match m_Intrinsic()
Match intrinsic calls like this: m_Intrinsic<Intrinsic::fabs>(m_Value(X))
ThreeOps_match< Cond, LHS, RHS, Instruction::Select > m_Select(const Cond &C, const LHS &L, const RHS &R)
Matches SelectInst.
auto m_BinOp()
Match an arbitrary binary operation and ignore it.
match_combine_or< CastInst_match< OpTy, SExtInst >, OpTy > m_SExtOrSelf(const OpTy &Op)
ExtractValue_match< Ind, Val_t > m_ExtractValue(const Val_t &V)
Match a single index ExtractValue instruction.
BinOpPred_match< LHS, RHS, is_logical_shift_op > m_LogicalShift(const LHS &L, const RHS &R)
Matches logical shift operations.
auto m_Value()
Match an arbitrary value and ignore it.
ShiftLike_match< LHS, Instruction::Shl > m_ShlOrSelf(const LHS &L, uint64_t &R)
Matches shl L, ConstShAmt or L itself (R will be set to zero in this case).
BinaryOp_match< LHS, RHS, Instruction::Xor, true > m_c_Xor(const LHS &L, const RHS &R)
Matches an Xor with LHS and RHS in either order.
SpecificCmpClass_match< LHS, RHS, CmpInst > m_SpecificCmp(CmpPredicate MatchPred, const LHS &L, const RHS &R)
BinaryOp_match< LHS, RHS, Instruction::Mul > m_Mul(const LHS &L, const RHS &R)
auto m_Constant()
Match an arbitrary Constant and ignore it.
auto m_LogicalOr()
Matches L || R where L and R are arbitrary values.
TwoOps_match< V1_t, V2_t, Instruction::ShuffleVector > m_Shuffle(const V1_t &v1, const V2_t &v2)
Matches ShuffleVectorInst independently of mask value.
match_bind< WithOverflowInst > m_WithOverflowInst(WithOverflowInst *&I)
Match a with overflow intrinsic, capturing it if we match.
SpecificCmpClass_match< LHS, RHS, ICmpInst > m_SpecificICmp(CmpPredicate MatchPred, const LHS &L, const RHS &R)
CastInst_match< OpTy, ZExtInst > m_ZExt(const OpTy &Op)
Matches ZExt.
cst_pred_ty< is_negated_power2 > m_NegatedPower2()
Match a integer or vector negated power-of-2.
match_immconstant_ty m_ImmConstant()
Match an arbitrary immediate Constant and ignore it.
DisjointOr_match< LHS, RHS, true > m_c_DisjointOr(const LHS &L, const RHS &R)
BinaryOp_match< LHS, RHS, Instruction::Add, true > m_c_Add(const LHS &L, const RHS &R)
Matches a Add with LHS and RHS in either order.
SpecificCmpClass_match< LHS, RHS, FCmpInst > m_SpecificFCmp(CmpPredicate MatchPred, const LHS &L, const RHS &R)
match_combine_or< BinaryOp_match< LHS, RHS, Instruction::Add >, DisjointOr_match< LHS, RHS > > m_AddLike(const LHS &L, const RHS &R)
Match either "add" or "or disjoint".
CastOperator_match< OpTy, Instruction::BitCast > m_BitCast(const OpTy &Op)
Matches BitCast.
match_combine_or< CastInst_match< OpTy, SExtInst >, NNegZExt_match< OpTy > > m_SExtLike(const OpTy &Op)
Match either "sext" or "zext nneg".
auto m_c_MaxOrMin(const LHS &L, const RHS &R)
MaxMin_match< ICmpInst, LHS, RHS, smax_pred_ty > m_SMax(const LHS &L, const RHS &R)
cst_pred_ty< is_maxsignedvalue > m_MaxSignedValue()
Match an integer or vector with values having all bits except for the high bit set (0x7f....
AnyBinaryOp_match< LHS, RHS, true > m_c_BinOp(const LHS &L, const RHS &R)
Matches a BinaryOperator with LHS and RHS in either order.
BinaryOp_match< LHS, RHS, Instruction::LShr > m_LShr(const LHS &L, const RHS &R)
CmpClass_match< LHS, RHS, ICmpInst > m_ICmp(CmpPredicate &Pred, const LHS &L, const RHS &R)
match_combine_or< CastInst_match< OpTy, ZExtInst >, CastInst_match< OpTy, SExtInst > > m_ZExtOrSExt(const OpTy &Op)
BinOpPred_match< LHS, RHS, is_shift_op > m_Shift(const LHS &L, const RHS &R)
Matches shift operations.
cstfp_pred_ty< is_pos_zero_fp > m_PosZeroFP()
Match a floating-point positive zero.
LogicalOp_match< LHS, RHS, Instruction::And, true > m_c_LogicalAnd(const LHS &L, const RHS &R)
Matches L && R with LHS and RHS in either order.
BinaryOp_match< LHS, RHS, Instruction::Shl > m_Shl(const LHS &L, const RHS &R)
auto m_LogicalAnd()
Matches L && R where L and R are arbitrary values.
BinaryOp_match< LHS, RHS, Instruction::Or > m_Or(const LHS &L, const RHS &R)
m_Intrinsic_Ty< Opnd0 >::Ty m_BSwap(const Opnd0 &Op0)
CastInst_match< OpTy, SExtInst > m_SExt(const OpTy &Op)
Matches SExt.
is_zero m_Zero()
Match any null constant or a vector with all elements equal to 0.
BinaryOp_match< LHS, RHS, Instruction::Or, true > m_c_Or(const LHS &L, const RHS &R)
Matches an Or with LHS and RHS in either order.
ThreeOps_match< Val_t, Elt_t, Idx_t, Instruction::InsertElement > m_InsertElt(const Val_t &Val, const Elt_t &Elt, const Idx_t &Idx)
Matches InsertElementInst.
ElementWiseBitCast_match< OpTy > m_ElementWiseBitCast(const OpTy &Op)
BinaryOp_match< LHS, RHS, Instruction::Sub > m_Sub(const LHS &L, const RHS &R)
match_unless< Ty > m_Unless(const Ty &M)
Match if the inner matcher does NOT match.
cst_pred_ty< icmp_pred_with_threshold > m_SpecificInt_ICMP(ICmpInst::Predicate Predicate, const APInt &Threshold)
Match an integer or vector with every element comparing 'pred' (eg/ne/...) to Threshold.
auto m_ConstantInt()
Match an arbitrary ConstantInt and ignore it.
NodeAddr< CodeNode * > Code
Definition RDFGraph.h:388
friend class Instruction
Iterator for Instructions in a `BasicBlock.
Definition BasicBlock.h:73
This is an optimization pass for GlobalISel generic memory operations.
LLVM_ABI Intrinsic::ID getInverseMinMaxIntrinsic(Intrinsic::ID MinMaxID)
@ Low
Lower the current thread's priority such that it does not affect foreground tasks significantly.
Definition Threading.h:280
@ Offset
Definition DWP.cpp:532
FunctionAddr VTableAddr Value
Definition InstrProf.h:137
Constant * getPredForFCmpCode(unsigned Code, Type *OpTy, CmpInst::Predicate &Pred)
This is the complement of getFCmpCode.
cl::opt< bool > ProfcheckDisableMetadataFixes
Definition LoopInfo.cpp:60
LLVM_ABI bool isSignBitCheck(ICmpInst::Predicate Pred, const APInt &RHS, bool &TrueIfSigned)
Given an exploded icmp instruction, return true if the comparison only checks the sign bit.
LLVM_ABI void setExplicitlyUnknownBranchWeightsIfProfiled(Instruction &I, StringRef PassName, const Function *F=nullptr)
Like setExplicitlyUnknownBranchWeights(...), but only sets unknown branch weights in the new instruct...
decltype(auto) dyn_cast(const From &Val)
dyn_cast<X> - Return the argument parameter cast to the specified type.
Definition Casting.h:643
bool predicatesFoldable(CmpInst::Predicate P1, CmpInst::Predicate P2)
Return true if both predicates match sign or if at least one of them is an equality comparison (which...
LLVM_ABI Constant * ConstantFoldCompareInstOperands(unsigned Predicate, Constant *LHS, Constant *RHS, const DataLayout &DL, const TargetLibraryInfo *TLI=nullptr, const Instruction *I=nullptr)
Attempt to constant fold a compare instruction (icmp/fcmp) with the specified operands.
LLVM_ABI Value * simplifyOrInst(Value *LHS, Value *RHS, const SimplifyQuery &Q)
Given operands for an Or, fold the result or return null.
LLVM_ABI Value * simplifyXorInst(Value *LHS, Value *RHS, const SimplifyQuery &Q)
Given operands for an Xor, fold the result or return null.
LLVM_ABI bool isGuaranteedNotToBeUndef(const Value *V, AssumptionCache *AC=nullptr, const Instruction *CtxI=nullptr, const DominatorTree *DT=nullptr, unsigned Depth=0)
Returns true if V cannot be undef, but may be poison.
LLVM_ABI bool matchSimpleRecurrence(const PHINode *P, BinaryOperator *&BO, Value *&Start, Value *&Step)
Attempt to match a simple first order recurrence cycle of the form: iv = phi Ty [Start,...
auto dyn_cast_or_null(const Y &Val)
Definition Casting.h:753
LLVM_ABI bool isKnownNegative(const Value *V, const SimplifyQuery &SQ, unsigned Depth=0)
Returns true if the given value is known be negative (i.e.
LLVM_ABI Constant * getLosslessUnsignedTrunc(Constant *C, Type *DestTy, const DataLayout &DL, PreservedCastFlags *Flags=nullptr)
LLVM_ABI bool recognizeBSwapOrBitReverseIdiom(Instruction *I, bool MatchBSwaps, bool MatchBitReversals, SmallVectorImpl< Instruction * > &InsertedInsts)
Try to match a bswap or bitreverse idiom.
Definition Local.cpp:3785
constexpr bool isPowerOf2_32(uint32_t Value)
Return true if the argument is a power of two > 0.
Definition MathExtras.h:279
LLVM_ABI Value * simplifyICmpInst(CmpPredicate Pred, Value *LHS, Value *RHS, const SimplifyQuery &Q)
Given operands for an ICmpInst, fold the result or return null.
LLVM_ABI Constant * getLosslessSignedTrunc(Constant *C, Type *DestTy, const DataLayout &DL, PreservedCastFlags *Flags=nullptr)
LLVM_ABI Value * simplifyAndInst(Value *LHS, Value *RHS, const SimplifyQuery &Q)
Given operands for an And, fold the result or return null.
LLVM_ABI bool isKnownInversion(const Value *X, const Value *Y)
Return true iff:
bool isa(const From &Val)
isa<X> - Return true if the parameter to the template is an instance of one of the template type argu...
Definition Casting.h:547
LLVM_ABI bool isKnownNonZero(const Value *V, const SimplifyQuery &Q, unsigned Depth=0)
Return true if the given value is known to be non-zero when defined.
constexpr int PoisonMaskElem
@ Other
Any other memory.
Definition ModRef.h:68
LLVM_ABI Value * simplifyBinOp(unsigned Opcode, Value *LHS, Value *RHS, const SimplifyQuery &Q)
Given operands for a BinaryOperator, fold the result or return null.
std::optional< DecomposedBitTest > decomposeBitTest(Value *Cond, bool LookThroughTrunc=true, bool AllowNonZeroC=false, bool DecomposeAnd=false)
Decompose an icmp into the form ((X & Mask) pred C) if possible.
@ Mul
Product of integers.
@ Xor
Bitwise or logical XOR of integers.
@ And
Bitwise or logical AND of integers.
@ Sub
Subtraction of integers.
@ Add
Sum of integers.
DWARFExpression::Operation Op
LLVM_ABI bool isGuaranteedNotToBeUndefOrPoison(const Value *V, AssumptionCache *AC=nullptr, const Instruction *CtxI=nullptr, const DominatorTree *DT=nullptr, unsigned Depth=0)
Return true if this function can prove that V does not have undef bits and is never poison.
constexpr unsigned BitWidth
decltype(auto) cast(const From &Val)
cast<X> - Return the argument parameter cast to the specified type.
Definition Casting.h:559
APFloat neg(APFloat X)
Returns the negated value of the argument.
Definition APFloat.h:1636
unsigned getICmpCode(CmpInst::Predicate Pred)
Encode a icmp predicate into a three bit mask.
LLVM_ABI bool isKnownToBeAPowerOfTwo(const Value *V, const DataLayout &DL, bool OrZero=false, AssumptionCache *AC=nullptr, const Instruction *CxtI=nullptr, const DominatorTree *DT=nullptr, bool UseInstrInfo=true, unsigned Depth=0)
Return true if the given value is known to have exactly one bit set when defined.
LLVM_ABI bool isGuaranteedNotToBePoison(const Value *V, AssumptionCache *AC=nullptr, const Instruction *CtxI=nullptr, const DominatorTree *DT=nullptr, unsigned Depth=0)
Returns true if V cannot be poison, but may be undef.
std::pair< Value *, FPClassTest > fcmpToClassTest(FCmpInst::Predicate Pred, const Function &F, Value *LHS, Value *RHS, bool LookThroughSrc=true)
Returns a pair of values, which if passed to llvm.is.fpclass, returns the same result as an fcmp with...
unsigned getFCmpCode(CmpInst::Predicate CC)
Similar to getICmpCode but for FCmpInst.
std::optional< DecomposedBitTest > decomposeBitTestICmp(Value *LHS, Value *RHS, CmpInst::Predicate Pred, bool LookThroughTrunc=true, bool AllowNonZeroC=false, bool DecomposeAnd=false)
Decompose an icmp into the form ((X & Mask) pred C) if possible.
Constant * getPredForICmpCode(unsigned Code, bool Sign, Type *OpTy, CmpInst::Predicate &Pred)
This is the complement of getICmpCode.
void swap(llvm::BitVector &LHS, llvm::BitVector &RHS)
Implement std::swap in terms of BitVector swap.
Definition BitVector.h:872
#define N
bool isCombineableWith(const DecomposedBitMaskMul Other)
bool isNonNegative() const
Returns true if this value is known to be non-negative.
Definition KnownBits.h:106
APInt getMaxValue() const
Return the maximal unsigned value possible given these KnownBits.
Definition KnownBits.h:146
Matching combinators.
const DataLayout & DL
const Instruction * CxtI
const DominatorTree * DT
SimplifyQuery getWithInstruction(const Instruction *I) const
AssumptionCache * AC