LLVM 23.0.0git
InstCombineCompares.cpp
Go to the documentation of this file.
1//===- InstCombineCompares.cpp --------------------------------------------===//
2//
3// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions.
4// See https://llvm.org/LICENSE.txt for license information.
5// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception
6//
7//===----------------------------------------------------------------------===//
8//
9// This file implements the visitICmp and visitFCmp functions.
10//
11//===----------------------------------------------------------------------===//
12
13#include "InstCombineInternal.h"
14#include "llvm/ADT/APFloat.h"
15#include "llvm/ADT/APSInt.h"
16#include "llvm/ADT/SetVector.h"
17#include "llvm/ADT/Statistic.h"
22#include "llvm/Analysis/Loads.h"
26#include "llvm/IR/Constants.h"
27#include "llvm/IR/DataLayout.h"
28#include "llvm/IR/InstrTypes.h"
29#include "llvm/IR/Instruction.h"
35#include <bitset>
36
37using namespace llvm;
38using namespace PatternMatch;
39
40#define DEBUG_TYPE "instcombine"
41
42// How many times is a select replaced by one of its operands?
43STATISTIC(NumSel, "Number of select opts");
44
45namespace llvm {
47}
48
49/// Compute Result = In1+In2, returning true if the result overflowed for this
50/// type.
51static bool addWithOverflow(APInt &Result, const APInt &In1, const APInt &In2,
52 bool IsSigned = false) {
53 bool Overflow;
54 if (IsSigned)
55 Result = In1.sadd_ov(In2, Overflow);
56 else
57 Result = In1.uadd_ov(In2, Overflow);
58
59 return Overflow;
60}
61
62/// Compute Result = In1-In2, returning true if the result overflowed for this
63/// type.
64static bool subWithOverflow(APInt &Result, const APInt &In1, const APInt &In2,
65 bool IsSigned = false) {
66 bool Overflow;
67 if (IsSigned)
68 Result = In1.ssub_ov(In2, Overflow);
69 else
70 Result = In1.usub_ov(In2, Overflow);
71
72 return Overflow;
73}
74
75/// Given an icmp instruction, return true if any use of this comparison is a
76/// branch on sign bit comparison.
77static bool hasBranchUse(ICmpInst &I) {
78 for (auto *U : I.users())
79 if (isa<CondBrInst>(U))
80 return true;
81 return false;
82}
83
84/// Returns true if the exploded icmp can be expressed as a signed comparison
85/// to zero and updates the predicate accordingly.
86/// The signedness of the comparison is preserved.
87/// TODO: Refactor with decomposeBitTestICmp()?
88static bool isSignTest(ICmpInst::Predicate &Pred, const APInt &C) {
89 if (!ICmpInst::isSigned(Pred))
90 return false;
91
92 if (C.isZero())
93 return ICmpInst::isRelational(Pred);
94
95 if (C.isOne()) {
96 if (Pred == ICmpInst::ICMP_SLT) {
97 Pred = ICmpInst::ICMP_SLE;
98 return true;
99 }
100 } else if (C.isAllOnes()) {
101 if (Pred == ICmpInst::ICMP_SGT) {
102 Pred = ICmpInst::ICMP_SGE;
103 return true;
104 }
105 }
106
107 return false;
108}
109
110/// This is called when we see this pattern:
111/// cmp pred (load (gep GV, ...)), cmpcst
112/// where GV is a global variable with a constant initializer. Try to simplify
113/// this into some simple computation that does not need the load. For example
114/// we can optimize "icmp eq (load (gep "foo", 0, i)), 0" into "icmp eq i, 3".
115///
116/// If AndCst is non-null, then the loaded value is masked with that constant
117/// before doing the comparison. This handles cases like "A[i]&4 == 0".
119 LoadInst *LI, GetElementPtrInst *GEP, CmpInst &ICI, ConstantInt *AndCst) {
121 if (LI->isVolatile() || !GV || !GV->isConstant() ||
122 !GV->hasDefinitiveInitializer())
123 return nullptr;
124
125 Type *EltTy = LI->getType();
126 TypeSize EltSize = DL.getTypeStoreSize(EltTy);
127 if (EltSize.isScalable())
128 return nullptr;
129
131 if (!Expr.Index || Expr.BasePtr != GV || Expr.Offset.getBitWidth() > 64)
132 return nullptr;
133
134 Constant *Init = GV->getInitializer();
135 TypeSize GlobalSize = DL.getTypeAllocSize(Init->getType());
136
137 Value *Idx = Expr.Index;
138 const APInt &Stride = Expr.Scale;
139 const APInt &ConstOffset = Expr.Offset;
140
141 // Allow an additional context offset, but only within the stride.
142 if (!ConstOffset.ult(Stride))
143 return nullptr;
144
145 // Don't handle overlapping loads for now.
146 if (!Stride.uge(EltSize.getFixedValue()))
147 return nullptr;
148
149 // Don't blow up on huge arrays.
150 uint64_t ArrayElementCount =
151 divideCeil((GlobalSize.getFixedValue() - ConstOffset.getZExtValue()),
152 Stride.getZExtValue());
153 if (ArrayElementCount > MaxArraySizeForCombine)
154 return nullptr;
155
156 enum { Overdefined = -3, Undefined = -2 };
157
158 // Variables for our state machines.
159
160 // FirstTrueElement/SecondTrueElement - Used to emit a comparison of the form
161 // "i == 47 | i == 87", where 47 is the first index the condition is true for,
162 // and 87 is the second (and last) index. FirstTrueElement is -2 when
163 // undefined, otherwise set to the first true element. SecondTrueElement is
164 // -2 when undefined, -3 when overdefined and >= 0 when that index is true.
165 int FirstTrueElement = Undefined, SecondTrueElement = Undefined;
166
167 // FirstFalseElement/SecondFalseElement - Used to emit a comparison of the
168 // form "i != 47 & i != 87". Same state transitions as for true elements.
169 int FirstFalseElement = Undefined, SecondFalseElement = Undefined;
170
171 /// TrueRangeEnd/FalseRangeEnd - In conjunction with First*Element, these
172 /// define a state machine that triggers for ranges of values that the index
173 /// is true or false for. This triggers on things like "abbbbc"[i] == 'b'.
174 /// This is -2 when undefined, -3 when overdefined, and otherwise the last
175 /// index in the range (inclusive). We use -2 for undefined here because we
176 /// use relative comparisons and don't want 0-1 to match -1.
177 int TrueRangeEnd = Undefined, FalseRangeEnd = Undefined;
178
179 // MagicBitvector - This is a magic bitvector where we set a bit if the
180 // comparison is true for element 'i'. If there are 64 elements or less in
181 // the array, this will fully represent all the comparison results.
182 uint64_t MagicBitvector = 0;
183
184 // Scan the array and see if one of our patterns matches.
185 Constant *CompareRHS = cast<Constant>(ICI.getOperand(1));
186 APInt Offset = ConstOffset;
187 for (unsigned i = 0, e = ArrayElementCount; i != e; ++i, Offset += Stride) {
189 if (!Elt)
190 return nullptr;
191
192 // If the element is masked, handle it.
193 if (AndCst) {
194 Elt = ConstantFoldBinaryOpOperands(Instruction::And, Elt, AndCst, DL);
195 if (!Elt)
196 return nullptr;
197 }
198
199 // Find out if the comparison would be true or false for the i'th element.
201 CompareRHS, DL, &TLI);
202 if (!C)
203 return nullptr;
204
205 // If the result is undef for this element, ignore it.
206 if (isa<UndefValue>(C)) {
207 // Extend range state machines to cover this element in case there is an
208 // undef in the middle of the range.
209 if (TrueRangeEnd == (int)i - 1)
210 TrueRangeEnd = i;
211 if (FalseRangeEnd == (int)i - 1)
212 FalseRangeEnd = i;
213 continue;
214 }
215
216 // If we can't compute the result for any of the elements, we have to give
217 // up evaluating the entire conditional.
218 if (!isa<ConstantInt>(C))
219 return nullptr;
220
221 // Otherwise, we know if the comparison is true or false for this element,
222 // update our state machines.
223 bool IsTrueForElt = !cast<ConstantInt>(C)->isZero();
224
225 // State machine for single/double/range index comparison.
226 if (IsTrueForElt) {
227 // Update the TrueElement state machine.
228 if (FirstTrueElement == Undefined)
229 FirstTrueElement = TrueRangeEnd = i; // First true element.
230 else {
231 // Update double-compare state machine.
232 if (SecondTrueElement == Undefined)
233 SecondTrueElement = i;
234 else
235 SecondTrueElement = Overdefined;
236
237 // Update range state machine.
238 if (TrueRangeEnd == (int)i - 1)
239 TrueRangeEnd = i;
240 else
241 TrueRangeEnd = Overdefined;
242 }
243 } else {
244 // Update the FalseElement state machine.
245 if (FirstFalseElement == Undefined)
246 FirstFalseElement = FalseRangeEnd = i; // First false element.
247 else {
248 // Update double-compare state machine.
249 if (SecondFalseElement == Undefined)
250 SecondFalseElement = i;
251 else
252 SecondFalseElement = Overdefined;
253
254 // Update range state machine.
255 if (FalseRangeEnd == (int)i - 1)
256 FalseRangeEnd = i;
257 else
258 FalseRangeEnd = Overdefined;
259 }
260 }
261
262 // If this element is in range, update our magic bitvector.
263 if (i < 64 && IsTrueForElt)
264 MagicBitvector |= 1ULL << i;
265
266 // If all of our states become overdefined, bail out early. Since the
267 // predicate is expensive, only check it every 8 elements. This is only
268 // really useful for really huge arrays.
269 if ((i & 8) == 0 && i >= 64 && SecondTrueElement == Overdefined &&
270 SecondFalseElement == Overdefined && TrueRangeEnd == Overdefined &&
271 FalseRangeEnd == Overdefined)
272 return nullptr;
273 }
274
275 // Now that we've scanned the entire array, emit our new comparison(s). We
276 // order the state machines in complexity of the generated code.
277
278 // If inbounds keyword is not present, Idx * Stride can overflow.
279 // Let's assume that Stride is 2 and the wanted value is at offset 0.
280 // Then, there are two possible values for Idx to match offset 0:
281 // 0x00..00, 0x80..00.
282 // Emitting 'icmp eq Idx, 0' isn't correct in this case because the
283 // comparison is false if Idx was 0x80..00.
284 // We need to erase the highest countTrailingZeros(ElementSize) bits of Idx.
285 auto MaskIdx = [&](Value *Idx) {
286 if (!Expr.Flags.isInBounds() && Stride.countr_zero() != 0) {
288 Mask = Builder.CreateLShr(Mask, Stride.countr_zero());
289 Idx = Builder.CreateAnd(Idx, Mask);
290 }
291 return Idx;
292 };
293
294 // If the comparison is only true for one or two elements, emit direct
295 // comparisons.
296 if (SecondTrueElement != Overdefined) {
297 Idx = MaskIdx(Idx);
298 // None true -> false.
299 if (FirstTrueElement == Undefined)
300 return replaceInstUsesWith(ICI, Builder.getFalse());
301
302 Value *FirstTrueIdx = ConstantInt::get(Idx->getType(), FirstTrueElement);
303
304 // True for one element -> 'i == 47'.
305 if (SecondTrueElement == Undefined)
306 return new ICmpInst(ICmpInst::ICMP_EQ, Idx, FirstTrueIdx);
307
308 // True for two elements -> 'i == 47 | i == 72'.
309 Value *C1 = Builder.CreateICmpEQ(Idx, FirstTrueIdx);
310 Value *SecondTrueIdx = ConstantInt::get(Idx->getType(), SecondTrueElement);
311 Value *C2 = Builder.CreateICmpEQ(Idx, SecondTrueIdx);
312 return BinaryOperator::CreateOr(C1, C2);
313 }
314
315 // If the comparison is only false for one or two elements, emit direct
316 // comparisons.
317 if (SecondFalseElement != Overdefined) {
318 Idx = MaskIdx(Idx);
319 // None false -> true.
320 if (FirstFalseElement == Undefined)
321 return replaceInstUsesWith(ICI, Builder.getTrue());
322
323 Value *FirstFalseIdx = ConstantInt::get(Idx->getType(), FirstFalseElement);
324
325 // False for one element -> 'i != 47'.
326 if (SecondFalseElement == Undefined)
327 return new ICmpInst(ICmpInst::ICMP_NE, Idx, FirstFalseIdx);
328
329 // False for two elements -> 'i != 47 & i != 72'.
330 Value *C1 = Builder.CreateICmpNE(Idx, FirstFalseIdx);
331 Value *SecondFalseIdx =
332 ConstantInt::get(Idx->getType(), SecondFalseElement);
333 Value *C2 = Builder.CreateICmpNE(Idx, SecondFalseIdx);
334 return BinaryOperator::CreateAnd(C1, C2);
335 }
336
337 // If the comparison can be replaced with a range comparison for the elements
338 // where it is true, emit the range check.
339 if (TrueRangeEnd != Overdefined) {
340 assert(TrueRangeEnd != FirstTrueElement && "Should emit single compare");
341 Idx = MaskIdx(Idx);
342
343 // Generate (i-FirstTrue) <u (TrueRangeEnd-FirstTrue+1).
344 if (FirstTrueElement) {
345 Value *Offs = ConstantInt::getSigned(Idx->getType(), -FirstTrueElement);
346 Idx = Builder.CreateAdd(Idx, Offs);
347 }
348
349 Value *End =
350 ConstantInt::get(Idx->getType(), TrueRangeEnd - FirstTrueElement + 1);
351 return new ICmpInst(ICmpInst::ICMP_ULT, Idx, End);
352 }
353
354 // False range check.
355 if (FalseRangeEnd != Overdefined) {
356 assert(FalseRangeEnd != FirstFalseElement && "Should emit single compare");
357 Idx = MaskIdx(Idx);
358 // Generate (i-FirstFalse) >u (FalseRangeEnd-FirstFalse).
359 if (FirstFalseElement) {
360 Value *Offs = ConstantInt::getSigned(Idx->getType(), -FirstFalseElement);
361 Idx = Builder.CreateAdd(Idx, Offs);
362 }
363
364 Value *End =
365 ConstantInt::get(Idx->getType(), FalseRangeEnd - FirstFalseElement);
366 return new ICmpInst(ICmpInst::ICMP_UGT, Idx, End);
367 }
368
369 // If a magic bitvector captures the entire comparison state
370 // of this load, replace it with computation that does:
371 // ((magic_cst >> i) & 1) != 0
372 {
373 Type *Ty = nullptr;
374
375 // Look for an appropriate type:
376 // - The type of Idx if the magic fits
377 // - The smallest fitting legal type
378 if (ArrayElementCount <= Idx->getType()->getIntegerBitWidth())
379 Ty = Idx->getType();
380 else
381 Ty = DL.getSmallestLegalIntType(Init->getContext(), ArrayElementCount);
382
383 if (Ty) {
384 Idx = MaskIdx(Idx);
385 Value *V = Builder.CreateIntCast(Idx, Ty, false);
386 V = Builder.CreateLShr(ConstantInt::get(Ty, MagicBitvector), V);
387 V = Builder.CreateAnd(ConstantInt::get(Ty, 1), V);
388 return new ICmpInst(ICmpInst::ICMP_NE, V, ConstantInt::get(Ty, 0));
389 }
390 }
391
392 return nullptr;
393}
394
395/// Returns true if we can rewrite Start as a GEP with pointer Base
396/// and some integer offset. The nodes that need to be re-written
397/// for this transformation will be added to Explored.
399 const DataLayout &DL,
400 SetVector<Value *> &Explored) {
401 SmallVector<Value *, 16> WorkList(1, Start);
402 Explored.insert(Base);
403
404 // The following traversal gives us an order which can be used
405 // when doing the final transformation. Since in the final
406 // transformation we create the PHI replacement instructions first,
407 // we don't have to get them in any particular order.
408 //
409 // However, for other instructions we will have to traverse the
410 // operands of an instruction first, which means that we have to
411 // do a post-order traversal.
412 while (!WorkList.empty()) {
414
415 while (!WorkList.empty()) {
416 if (Explored.size() >= 100)
417 return false;
418
419 Value *V = WorkList.back();
420
421 if (Explored.contains(V)) {
422 WorkList.pop_back();
423 continue;
424 }
425
427 // We've found some value that we can't explore which is different from
428 // the base. Therefore we can't do this transformation.
429 return false;
430
431 if (auto *GEP = dyn_cast<GEPOperator>(V)) {
432 // Only allow inbounds GEPs with at most one variable offset.
433 auto IsNonConst = [](Value *V) { return !isa<ConstantInt>(V); };
434 if (!GEP->isInBounds() || count_if(GEP->indices(), IsNonConst) > 1)
435 return false;
436
437 NW = NW.intersectForOffsetAdd(GEP->getNoWrapFlags());
438 if (!Explored.contains(GEP->getOperand(0)))
439 WorkList.push_back(GEP->getOperand(0));
440 }
441
442 if (WorkList.back() == V) {
443 WorkList.pop_back();
444 // We've finished visiting this node, mark it as such.
445 Explored.insert(V);
446 }
447
448 if (auto *PN = dyn_cast<PHINode>(V)) {
449 // We cannot transform PHIs on unsplittable basic blocks.
450 if (isa<CatchSwitchInst>(PN->getParent()->getTerminator()))
451 return false;
452 Explored.insert(PN);
453 PHIs.insert(PN);
454 }
455 }
456
457 // Explore the PHI nodes further.
458 for (auto *PN : PHIs)
459 for (Value *Op : PN->incoming_values())
460 if (!Explored.contains(Op))
461 WorkList.push_back(Op);
462 }
463
464 // Make sure that we can do this. Since we can't insert GEPs in a basic
465 // block before a PHI node, we can't easily do this transformation if
466 // we have PHI node users of transformed instructions.
467 for (Value *Val : Explored) {
468 for (Value *Use : Val->uses()) {
469
470 auto *PHI = dyn_cast<PHINode>(Use);
471 auto *Inst = dyn_cast<Instruction>(Val);
472
473 if (Inst == Base || Inst == PHI || !Inst || !PHI ||
474 !Explored.contains(PHI))
475 continue;
476
477 if (PHI->getParent() == Inst->getParent())
478 return false;
479 }
480 }
481 return true;
482}
483
484// Sets the appropriate insert point on Builder where we can add
485// a replacement Instruction for V (if that is possible).
486static void setInsertionPoint(IRBuilder<> &Builder, Value *V,
487 bool Before = true) {
488 if (auto *PHI = dyn_cast<PHINode>(V)) {
489 BasicBlock *Parent = PHI->getParent();
490 Builder.SetInsertPoint(Parent, Parent->getFirstInsertionPt());
491 return;
492 }
493 if (auto *I = dyn_cast<Instruction>(V)) {
494 if (!Before)
495 I = &*std::next(I->getIterator());
496 Builder.SetInsertPoint(I);
497 return;
498 }
499 if (auto *A = dyn_cast<Argument>(V)) {
500 // Set the insertion point in the entry block.
501 BasicBlock &Entry = A->getParent()->getEntryBlock();
502 Builder.SetInsertPoint(&Entry, Entry.getFirstInsertionPt());
503 return;
504 }
505 // Otherwise, this is a constant and we don't need to set a new
506 // insertion point.
507 assert(isa<Constant>(V) && "Setting insertion point for unknown value!");
508}
509
510/// Returns a re-written value of Start as an indexed GEP using Base as a
511/// pointer.
513 const DataLayout &DL,
514 SetVector<Value *> &Explored,
515 InstCombiner &IC) {
516 // Perform all the substitutions. This is a bit tricky because we can
517 // have cycles in our use-def chains.
518 // 1. Create the PHI nodes without any incoming values.
519 // 2. Create all the other values.
520 // 3. Add the edges for the PHI nodes.
521 // 4. Emit GEPs to get the original pointers.
522 // 5. Remove the original instructions.
523 Type *IndexType = IntegerType::get(
524 Base->getContext(), DL.getIndexTypeSizeInBits(Start->getType()));
525
527 NewInsts[Base] = ConstantInt::getNullValue(IndexType);
528
529 // Create the new PHI nodes, without adding any incoming values.
530 for (Value *Val : Explored) {
531 if (Val == Base)
532 continue;
533 // Create empty phi nodes. This avoids cyclic dependencies when creating
534 // the remaining instructions.
535 if (auto *PHI = dyn_cast<PHINode>(Val))
536 NewInsts[PHI] =
537 PHINode::Create(IndexType, PHI->getNumIncomingValues(),
538 PHI->getName() + ".idx", PHI->getIterator());
539 }
540 IRBuilder<> Builder(Base->getContext());
541
542 // Create all the other instructions.
543 for (Value *Val : Explored) {
544 if (NewInsts.contains(Val))
545 continue;
546
547 if (auto *GEP = dyn_cast<GEPOperator>(Val)) {
548 setInsertionPoint(Builder, GEP);
549 Value *Op = NewInsts[GEP->getOperand(0)];
550 Value *OffsetV = emitGEPOffset(&Builder, DL, GEP);
552 NewInsts[GEP] = OffsetV;
553 else
554 NewInsts[GEP] = Builder.CreateAdd(
555 Op, OffsetV, GEP->getOperand(0)->getName() + ".add",
556 /*NUW=*/NW.hasNoUnsignedWrap(),
557 /*NSW=*/NW.hasNoUnsignedSignedWrap());
558 continue;
559 }
560 if (isa<PHINode>(Val))
561 continue;
562
563 llvm_unreachable("Unexpected instruction type");
564 }
565
566 // Add the incoming values to the PHI nodes.
567 for (Value *Val : Explored) {
568 if (Val == Base)
569 continue;
570 // All the instructions have been created, we can now add edges to the
571 // phi nodes.
572 if (auto *PHI = dyn_cast<PHINode>(Val)) {
573 PHINode *NewPhi = static_cast<PHINode *>(NewInsts[PHI]);
574 for (unsigned I = 0, E = PHI->getNumIncomingValues(); I < E; ++I) {
575 Value *NewIncoming = PHI->getIncomingValue(I);
576
577 auto It = NewInsts.find(NewIncoming);
578 if (It != NewInsts.end())
579 NewIncoming = It->second;
580
581 NewPhi->addIncoming(NewIncoming, PHI->getIncomingBlock(I));
582 }
583 }
584 }
585
586 for (Value *Val : Explored) {
587 if (Val == Base)
588 continue;
589
590 setInsertionPoint(Builder, Val, false);
591 // Create GEP for external users.
592 Value *NewVal = Builder.CreateGEP(Builder.getInt8Ty(), Base, NewInsts[Val],
593 Val->getName() + ".ptr", NW);
594 IC.replaceInstUsesWith(*cast<Instruction>(Val), NewVal);
595 // Add old instruction to worklist for DCE. We don't directly remove it
596 // here because the original compare is one of the users.
598 }
599
600 return NewInsts[Start];
601}
602
603/// Converts (CMP GEPLHS, RHS) if this change would make RHS a constant.
604/// We can look through PHIs, GEPs and casts in order to determine a common base
605/// between GEPLHS and RHS.
608 const DataLayout &DL,
609 InstCombiner &IC) {
610 // FIXME: Support vector of pointers.
611 if (GEPLHS->getType()->isVectorTy())
612 return nullptr;
613
614 if (!GEPLHS->hasAllConstantIndices())
615 return nullptr;
616
617 APInt Offset(DL.getIndexTypeSizeInBits(GEPLHS->getType()), 0);
618 Value *PtrBase =
620 /*AllowNonInbounds*/ false);
621
622 // Bail if we looked through addrspacecast.
623 if (PtrBase->getType() != GEPLHS->getType())
624 return nullptr;
625
626 // The set of nodes that will take part in this transformation.
627 SetVector<Value *> Nodes;
628 GEPNoWrapFlags NW = GEPLHS->getNoWrapFlags();
629 if (!canRewriteGEPAsOffset(RHS, PtrBase, NW, DL, Nodes))
630 return nullptr;
631
632 // We know we can re-write this as
633 // ((gep Ptr, OFFSET1) cmp (gep Ptr, OFFSET2)
634 // Since we've only looked through inbouds GEPs we know that we
635 // can't have overflow on either side. We can therefore re-write
636 // this as:
637 // OFFSET1 cmp OFFSET2
638 Value *NewRHS = rewriteGEPAsOffset(RHS, PtrBase, NW, DL, Nodes, IC);
639
640 // RewriteGEPAsOffset has replaced RHS and all of its uses with a re-written
641 // GEP having PtrBase as the pointer base, and has returned in NewRHS the
642 // offset. Since Index is the offset of LHS to the base pointer, we will now
643 // compare the offsets instead of comparing the pointers.
645 IC.Builder.getInt(Offset), NewRHS);
646}
647
648/// Fold comparisons between a GEP instruction and something else. At this point
649/// we know that the GEP is on the LHS of the comparison.
652 // Don't transform signed compares of GEPs into index compares. Even if the
653 // GEP is inbounds, the final add of the base pointer can have signed overflow
654 // and would change the result of the icmp.
655 // e.g. "&foo[0] <s &foo[1]" can't be folded to "true" because "foo" could be
656 // the maximum signed value for the pointer type.
658 return nullptr;
659
660 // Look through bitcasts and addrspacecasts. We do not however want to remove
661 // 0 GEPs.
662 if (!isa<GetElementPtrInst>(RHS))
663 RHS = RHS->stripPointerCasts();
664
665 auto CanFold = [Cond](GEPNoWrapFlags NW) {
667 return true;
668
669 // Unsigned predicates can be folded if the GEPs have *any* nowrap flags.
671 return NW != GEPNoWrapFlags::none();
672 };
673
674 auto NewICmp = [Cond](GEPNoWrapFlags NW, Value *Op1, Value *Op2) {
675 if (!NW.hasNoUnsignedWrap()) {
676 // Convert signed to unsigned comparison.
677 return new ICmpInst(ICmpInst::getSignedPredicate(Cond), Op1, Op2);
678 }
679
680 auto *I = new ICmpInst(Cond, Op1, Op2);
681 I->setSameSign(NW.hasNoUnsignedSignedWrap());
682 return I;
683 };
684
686 if (Base.Ptr == RHS && CanFold(Base.LHSNW) && !Base.isExpensive()) {
687 // ((gep Ptr, OFFSET) cmp Ptr) ---> (OFFSET cmp 0).
688 Type *IdxTy = DL.getIndexType(GEPLHS->getType());
689 Value *Offset =
690 EmitGEPOffsets(Base.LHSGEPs, Base.LHSNW, IdxTy, /*RewriteGEPs=*/true);
691 return NewICmp(Base.LHSNW, Offset,
692 Constant::getNullValue(Offset->getType()));
693 }
694
695 if (GEPLHS->isInBounds() && ICmpInst::isEquality(Cond) &&
696 isa<Constant>(RHS) && cast<Constant>(RHS)->isNullValue() &&
697 !NullPointerIsDefined(I.getFunction(),
698 RHS->getType()->getPointerAddressSpace())) {
699 // For most address spaces, an allocation can't be placed at null, but null
700 // itself is treated as a 0 size allocation in the in bounds rules. Thus,
701 // the only valid inbounds address derived from null, is null itself.
702 // Thus, we have four cases to consider:
703 // 1) Base == nullptr, Offset == 0 -> inbounds, null
704 // 2) Base == nullptr, Offset != 0 -> poison as the result is out of bounds
705 // 3) Base != nullptr, Offset == (-base) -> poison (crossing allocations)
706 // 4) Base != nullptr, Offset != (-base) -> nonnull (and possibly poison)
707 //
708 // (Note if we're indexing a type of size 0, that simply collapses into one
709 // of the buckets above.)
710 //
711 // In general, we're allowed to make values less poison (i.e. remove
712 // sources of full UB), so in this case, we just select between the two
713 // non-poison cases (1 and 4 above).
714 //
715 // For vectors, we apply the same reasoning on a per-lane basis.
716 auto *Base = GEPLHS->getPointerOperand();
717 if (GEPLHS->getType()->isVectorTy() && Base->getType()->isPointerTy()) {
718 auto EC = cast<VectorType>(GEPLHS->getType())->getElementCount();
719 Base = Builder.CreateVectorSplat(EC, Base);
720 }
721 return new ICmpInst(Cond, Base,
723 cast<Constant>(RHS), Base->getType()));
724 } else if (GEPOperator *GEPRHS = dyn_cast<GEPOperator>(RHS)) {
725 GEPNoWrapFlags NW = GEPLHS->getNoWrapFlags() & GEPRHS->getNoWrapFlags();
726
727 // If the base pointers are different, but the indices are the same, just
728 // compare the base pointer.
729 if (GEPLHS->getOperand(0) != GEPRHS->getOperand(0)) {
730 bool IndicesTheSame =
731 GEPLHS->getNumOperands() == GEPRHS->getNumOperands() &&
732 GEPLHS->getPointerOperand()->getType() ==
733 GEPRHS->getPointerOperand()->getType() &&
734 GEPLHS->getSourceElementType() == GEPRHS->getSourceElementType();
735 if (IndicesTheSame)
736 for (unsigned i = 1, e = GEPLHS->getNumOperands(); i != e; ++i)
737 if (GEPLHS->getOperand(i) != GEPRHS->getOperand(i)) {
738 IndicesTheSame = false;
739 break;
740 }
741
742 // If all indices are the same, just compare the base pointers.
743 Type *BaseType = GEPLHS->getOperand(0)->getType();
744 if (IndicesTheSame &&
745 CmpInst::makeCmpResultType(BaseType) == I.getType() && CanFold(NW))
746 return new ICmpInst(Cond, GEPLHS->getOperand(0), GEPRHS->getOperand(0));
747
748 // If we're comparing GEPs with two base pointers that only differ in type
749 // and both GEPs have only constant indices or just one use, then fold
750 // the compare with the adjusted indices.
751 // FIXME: Support vector of pointers.
752 if (GEPLHS->isInBounds() && GEPRHS->isInBounds() &&
753 (GEPLHS->hasAllConstantIndices() || GEPLHS->hasOneUse()) &&
754 (GEPRHS->hasAllConstantIndices() || GEPRHS->hasOneUse()) &&
755 GEPLHS->getOperand(0)->stripPointerCasts() ==
756 GEPRHS->getOperand(0)->stripPointerCasts() &&
757 !GEPLHS->getType()->isVectorTy()) {
758 Value *LOffset = EmitGEPOffset(GEPLHS);
759 Value *ROffset = EmitGEPOffset(GEPRHS);
760
761 // If we looked through an addrspacecast between different sized address
762 // spaces, the LHS and RHS pointers are different sized
763 // integers. Truncate to the smaller one.
764 Type *LHSIndexTy = LOffset->getType();
765 Type *RHSIndexTy = ROffset->getType();
766 if (LHSIndexTy != RHSIndexTy) {
767 if (LHSIndexTy->getPrimitiveSizeInBits().getFixedValue() <
768 RHSIndexTy->getPrimitiveSizeInBits().getFixedValue()) {
769 ROffset = Builder.CreateTrunc(ROffset, LHSIndexTy);
770 } else
771 LOffset = Builder.CreateTrunc(LOffset, RHSIndexTy);
772 }
773
775 LOffset, ROffset);
776 return replaceInstUsesWith(I, Cmp);
777 }
778 }
779
780 if (GEPLHS->getOperand(0) == GEPRHS->getOperand(0) &&
781 GEPLHS->getNumOperands() == GEPRHS->getNumOperands() &&
782 GEPLHS->getSourceElementType() == GEPRHS->getSourceElementType()) {
783 // If the GEPs only differ by one index, compare it.
784 unsigned NumDifferences = 0; // Keep track of # differences.
785 unsigned DiffOperand = 0; // The operand that differs.
786 for (unsigned i = 1, e = GEPRHS->getNumOperands(); i != e; ++i)
787 if (GEPLHS->getOperand(i) != GEPRHS->getOperand(i)) {
788 Type *LHSType = GEPLHS->getOperand(i)->getType();
789 Type *RHSType = GEPRHS->getOperand(i)->getType();
790 // FIXME: Better support for vector of pointers.
791 if (LHSType->getPrimitiveSizeInBits() !=
792 RHSType->getPrimitiveSizeInBits() ||
793 (GEPLHS->getType()->isVectorTy() &&
794 (!LHSType->isVectorTy() || !RHSType->isVectorTy()))) {
795 // Irreconcilable differences.
796 NumDifferences = 2;
797 break;
798 }
799
800 if (NumDifferences++)
801 break;
802 DiffOperand = i;
803 }
804
805 if (NumDifferences == 0) // SAME GEP?
806 return replaceInstUsesWith(
807 I, // No comparison is needed here.
808 ConstantInt::get(I.getType(), ICmpInst::isTrueWhenEqual(Cond)));
809 // If two GEPs only differ by an index, compare them.
810 // Note that nowrap flags are always needed when comparing two indices.
811 else if (NumDifferences == 1 && NW != GEPNoWrapFlags::none()) {
812 Value *LHSV = GEPLHS->getOperand(DiffOperand);
813 Value *RHSV = GEPRHS->getOperand(DiffOperand);
814 return NewICmp(NW, LHSV, RHSV);
815 }
816 }
817
818 if (Base.Ptr && CanFold(Base.LHSNW & Base.RHSNW) && !Base.isExpensive()) {
819 // ((gep Ptr, OFFSET1) cmp (gep Ptr, OFFSET2) ---> (OFFSET1 cmp OFFSET2)
820 Type *IdxTy = DL.getIndexType(GEPLHS->getType());
821 Value *L =
822 EmitGEPOffsets(Base.LHSGEPs, Base.LHSNW, IdxTy, /*RewriteGEP=*/true);
823 Value *R =
824 EmitGEPOffsets(Base.RHSGEPs, Base.RHSNW, IdxTy, /*RewriteGEP=*/true);
825 return NewICmp(Base.LHSNW & Base.RHSNW, L, R);
826 }
827 }
828
829 // Try convert this to an indexed compare by looking through PHIs/casts as a
830 // last resort.
831 return transformToIndexedCompare(GEPLHS, RHS, Cond, DL, *this);
832}
833
835 // It would be tempting to fold away comparisons between allocas and any
836 // pointer not based on that alloca (e.g. an argument). However, even
837 // though such pointers cannot alias, they can still compare equal.
838 //
839 // But LLVM doesn't specify where allocas get their memory, so if the alloca
840 // doesn't escape we can argue that it's impossible to guess its value, and we
841 // can therefore act as if any such guesses are wrong.
842 //
843 // However, we need to ensure that this folding is consistent: We can't fold
844 // one comparison to false, and then leave a different comparison against the
845 // same value alone (as it might evaluate to true at runtime, leading to a
846 // contradiction). As such, this code ensures that all comparisons are folded
847 // at the same time, and there are no other escapes.
848
849 struct CmpCaptureTracker : public CaptureTracker {
850 AllocaInst *Alloca;
851 bool Captured = false;
852 /// The value of the map is a bit mask of which icmp operands the alloca is
853 /// used in.
855
856 CmpCaptureTracker(AllocaInst *Alloca) : Alloca(Alloca) {}
857
858 void tooManyUses() override { Captured = true; }
859
860 Action captured(const Use *U, UseCaptureInfo CI) override {
861 // TODO(captures): Use UseCaptureInfo.
862 auto *ICmp = dyn_cast<ICmpInst>(U->getUser());
863 // We need to check that U is based *only* on the alloca, and doesn't
864 // have other contributions from a select/phi operand.
865 // TODO: We could check whether getUnderlyingObjects() reduces to one
866 // object, which would allow looking through phi nodes.
867 if (ICmp && ICmp->isEquality() && getUnderlyingObject(*U) == Alloca) {
868 // Collect equality icmps of the alloca, and don't treat them as
869 // captures.
870 ICmps[ICmp] |= 1u << U->getOperandNo();
871 return Continue;
872 }
873
874 Captured = true;
875 return Stop;
876 }
877 };
878
879 CmpCaptureTracker Tracker(Alloca);
880 PointerMayBeCaptured(Alloca, &Tracker);
881 if (Tracker.Captured)
882 return false;
883
884 bool Changed = false;
885 for (auto [ICmp, Operands] : Tracker.ICmps) {
886 switch (Operands) {
887 case 1:
888 case 2: {
889 // The alloca is only used in one icmp operand. Assume that the
890 // equality is false.
891 auto *Res = ConstantInt::get(ICmp->getType(),
892 ICmp->getPredicate() == ICmpInst::ICMP_NE);
893 replaceInstUsesWith(*ICmp, Res);
895 Changed = true;
896 break;
897 }
898 case 3:
899 // Both icmp operands are based on the alloca, so this is comparing
900 // pointer offsets, without leaking any information about the address
901 // of the alloca. Ignore such comparisons.
902 break;
903 default:
904 llvm_unreachable("Cannot happen");
905 }
906 }
907
908 return Changed;
909}
910
911/// Fold "icmp pred (X+C), X".
913 CmpPredicate Pred) {
914 // From this point on, we know that (X+C <= X) --> (X+C < X) because C != 0,
915 // so the values can never be equal. Similarly for all other "or equals"
916 // operators.
917 assert(!!C && "C should not be zero!");
918
919 // (X+1) <u X --> X >u (MAXUINT-1) --> X == 255
920 // (X+2) <u X --> X >u (MAXUINT-2) --> X > 253
921 // (X+MAXUINT) <u X --> X >u (MAXUINT-MAXUINT) --> X != 0
922 if (Pred == ICmpInst::ICMP_ULT || Pred == ICmpInst::ICMP_ULE) {
923 Constant *R =
924 ConstantInt::get(X->getType(), APInt::getMaxValue(C.getBitWidth()) - C);
925 return new ICmpInst(ICmpInst::ICMP_UGT, X, R);
926 }
927
928 // (X+1) >u X --> X <u (0-1) --> X != 255
929 // (X+2) >u X --> X <u (0-2) --> X <u 254
930 // (X+MAXUINT) >u X --> X <u (0-MAXUINT) --> X <u 1 --> X == 0
931 if (Pred == ICmpInst::ICMP_UGT || Pred == ICmpInst::ICMP_UGE)
932 return new ICmpInst(ICmpInst::ICMP_ULT, X,
933 ConstantInt::get(X->getType(), -C));
934
935 APInt SMax = APInt::getSignedMaxValue(C.getBitWidth());
936
937 // (X+ 1) <s X --> X >s (MAXSINT-1) --> X == 127
938 // (X+ 2) <s X --> X >s (MAXSINT-2) --> X >s 125
939 // (X+MAXSINT) <s X --> X >s (MAXSINT-MAXSINT) --> X >s 0
940 // (X+MINSINT) <s X --> X >s (MAXSINT-MINSINT) --> X >s -1
941 // (X+ -2) <s X --> X >s (MAXSINT- -2) --> X >s 126
942 // (X+ -1) <s X --> X >s (MAXSINT- -1) --> X != 127
943 if (Pred == ICmpInst::ICMP_SLT || Pred == ICmpInst::ICMP_SLE)
944 return new ICmpInst(ICmpInst::ICMP_SGT, X,
945 ConstantInt::get(X->getType(), SMax - C));
946
947 // (X+ 1) >s X --> X <s (MAXSINT-(1-1)) --> X != 127
948 // (X+ 2) >s X --> X <s (MAXSINT-(2-1)) --> X <s 126
949 // (X+MAXSINT) >s X --> X <s (MAXSINT-(MAXSINT-1)) --> X <s 1
950 // (X+MINSINT) >s X --> X <s (MAXSINT-(MINSINT-1)) --> X <s -2
951 // (X+ -2) >s X --> X <s (MAXSINT-(-2-1)) --> X <s -126
952 // (X+ -1) >s X --> X <s (MAXSINT-(-1-1)) --> X == -128
953
954 assert(Pred == ICmpInst::ICMP_SGT || Pred == ICmpInst::ICMP_SGE);
955 return new ICmpInst(ICmpInst::ICMP_SLT, X,
956 ConstantInt::get(X->getType(), SMax - (C - 1)));
957}
958
959/// Handle "(icmp eq/ne (ashr/lshr AP2, A), AP1)" ->
960/// (icmp eq/ne A, Log2(AP2/AP1)) ->
961/// (icmp eq/ne A, Log2(AP2) - Log2(AP1)).
963 const APInt &AP1,
964 const APInt &AP2) {
965 assert(I.isEquality() && "Cannot fold icmp gt/lt");
966
967 auto getICmp = [&I](CmpInst::Predicate Pred, Value *LHS, Value *RHS) {
968 if (I.getPredicate() == I.ICMP_NE)
969 Pred = CmpInst::getInversePredicate(Pred);
970 return new ICmpInst(Pred, LHS, RHS);
971 };
972
973 // Don't bother doing any work for cases which InstSimplify handles.
974 if (AP2.isZero())
975 return nullptr;
976
977 bool IsAShr = isa<AShrOperator>(I.getOperand(0));
978 if (IsAShr) {
979 if (AP2.isAllOnes())
980 return nullptr;
981 if (AP2.isNegative() != AP1.isNegative())
982 return nullptr;
983 if (AP2.sgt(AP1))
984 return nullptr;
985 }
986
987 if (!AP1)
988 // 'A' must be large enough to shift out the highest set bit.
989 return getICmp(I.ICMP_UGT, A,
990 ConstantInt::get(A->getType(), AP2.logBase2()));
991
992 if (AP1 == AP2)
993 return getICmp(I.ICMP_EQ, A, ConstantInt::getNullValue(A->getType()));
994
995 int Shift;
996 if (IsAShr && AP1.isNegative())
997 Shift = AP1.countl_one() - AP2.countl_one();
998 else
999 Shift = AP1.countl_zero() - AP2.countl_zero();
1000
1001 if (Shift > 0) {
1002 if (IsAShr && AP1 == AP2.ashr(Shift)) {
1003 // There are multiple solutions if we are comparing against -1 and the LHS
1004 // of the ashr is not a power of two.
1005 if (AP1.isAllOnes() && !AP2.isPowerOf2())
1006 return getICmp(I.ICMP_UGE, A, ConstantInt::get(A->getType(), Shift));
1007 return getICmp(I.ICMP_EQ, A, ConstantInt::get(A->getType(), Shift));
1008 } else if (AP1 == AP2.lshr(Shift)) {
1009 return getICmp(I.ICMP_EQ, A, ConstantInt::get(A->getType(), Shift));
1010 }
1011 }
1012
1013 // Shifting const2 will never be equal to const1.
1014 // FIXME: This should always be handled by InstSimplify?
1015 auto *TorF = ConstantInt::get(I.getType(), I.getPredicate() == I.ICMP_NE);
1016 return replaceInstUsesWith(I, TorF);
1017}
1018
1019/// Handle "(icmp eq/ne (shl AP2, A), AP1)" ->
1020/// (icmp eq/ne A, TrailingZeros(AP1) - TrailingZeros(AP2)).
1022 const APInt &AP1,
1023 const APInt &AP2) {
1024 assert(I.isEquality() && "Cannot fold icmp gt/lt");
1025
1026 auto getICmp = [&I](CmpInst::Predicate Pred, Value *LHS, Value *RHS) {
1027 if (I.getPredicate() == I.ICMP_NE)
1028 Pred = CmpInst::getInversePredicate(Pred);
1029 return new ICmpInst(Pred, LHS, RHS);
1030 };
1031
1032 // Don't bother doing any work for cases which InstSimplify handles.
1033 if (AP2.isZero())
1034 return nullptr;
1035
1036 unsigned AP2TrailingZeros = AP2.countr_zero();
1037
1038 if (!AP1 && AP2TrailingZeros != 0)
1039 return getICmp(
1040 I.ICMP_UGE, A,
1041 ConstantInt::get(A->getType(), AP2.getBitWidth() - AP2TrailingZeros));
1042
1043 if (AP1 == AP2)
1044 return getICmp(I.ICMP_EQ, A, ConstantInt::getNullValue(A->getType()));
1045
1046 // Get the distance between the lowest bits that are set.
1047 int Shift = AP1.countr_zero() - AP2TrailingZeros;
1048
1049 if (Shift > 0 && AP2.shl(Shift) == AP1)
1050 return getICmp(I.ICMP_EQ, A, ConstantInt::get(A->getType(), Shift));
1051
1052 // Shifting const2 will never be equal to const1.
1053 // FIXME: This should always be handled by InstSimplify?
1054 auto *TorF = ConstantInt::get(I.getType(), I.getPredicate() == I.ICMP_NE);
1055 return replaceInstUsesWith(I, TorF);
1056}
1057
1058/// The caller has matched a pattern of the form:
1059/// I = icmp ugt (add (add A, B), CI2), CI1
1060/// If this is of the form:
1061/// sum = a + b
1062/// if (sum+128 >u 255)
1063/// Then replace it with llvm.sadd.with.overflow.i8.
1064///
1066 ConstantInt *CI2, ConstantInt *CI1,
1067 InstCombinerImpl &IC) {
1068 // The transformation we're trying to do here is to transform this into an
1069 // llvm.sadd.with.overflow. To do this, we have to replace the original add
1070 // with a narrower add, and discard the add-with-constant that is part of the
1071 // range check (if we can't eliminate it, this isn't profitable).
1072
1073 // In order to eliminate the add-with-constant, the compare can be its only
1074 // use.
1075 Instruction *AddWithCst = cast<Instruction>(I.getOperand(0));
1076 if (!AddWithCst->hasOneUse())
1077 return nullptr;
1078
1079 // If CI2 is 2^7, 2^15, 2^31, then it might be an sadd.with.overflow.
1080 if (!CI2->getValue().isPowerOf2())
1081 return nullptr;
1082 unsigned NewWidth = CI2->getValue().countr_zero();
1083 if (NewWidth != 7 && NewWidth != 15 && NewWidth != 31)
1084 return nullptr;
1085
1086 // The width of the new add formed is 1 more than the bias.
1087 ++NewWidth;
1088
1089 // Check to see that CI1 is an all-ones value with NewWidth bits.
1090 if (CI1->getBitWidth() == NewWidth ||
1091 CI1->getValue() != APInt::getLowBitsSet(CI1->getBitWidth(), NewWidth))
1092 return nullptr;
1093
1094 // This is only really a signed overflow check if the inputs have been
1095 // sign-extended; check for that condition. For example, if CI2 is 2^31 and
1096 // the operands of the add are 64 bits wide, we need at least 33 sign bits.
1097 if (IC.ComputeMaxSignificantBits(A, &I) > NewWidth ||
1098 IC.ComputeMaxSignificantBits(B, &I) > NewWidth)
1099 return nullptr;
1100
1101 // In order to replace the original add with a narrower
1102 // llvm.sadd.with.overflow, the only uses allowed are the add-with-constant
1103 // and truncates that discard the high bits of the add. Verify that this is
1104 // the case.
1105 Instruction *OrigAdd = cast<Instruction>(AddWithCst->getOperand(0));
1106 for (User *U : OrigAdd->users()) {
1107 if (U == AddWithCst)
1108 continue;
1109
1110 // Only accept truncates for now. We would really like a nice recursive
1111 // predicate like SimplifyDemandedBits, but which goes downwards the use-def
1112 // chain to see which bits of a value are actually demanded. If the
1113 // original add had another add which was then immediately truncated, we
1114 // could still do the transformation.
1116 if (!TI || TI->getType()->getPrimitiveSizeInBits() > NewWidth)
1117 return nullptr;
1118 }
1119
1120 // If the pattern matches, truncate the inputs to the narrower type and
1121 // use the sadd_with_overflow intrinsic to efficiently compute both the
1122 // result and the overflow bit.
1123 Type *NewType = IntegerType::get(OrigAdd->getContext(), NewWidth);
1125 I.getModule(), Intrinsic::sadd_with_overflow, NewType);
1126
1127 InstCombiner::BuilderTy &Builder = IC.Builder;
1128
1129 // Put the new code above the original add, in case there are any uses of the
1130 // add between the add and the compare.
1131 Builder.SetInsertPoint(OrigAdd);
1132
1133 Value *TruncA = Builder.CreateTrunc(A, NewType, A->getName() + ".trunc");
1134 Value *TruncB = Builder.CreateTrunc(B, NewType, B->getName() + ".trunc");
1135 CallInst *Call = Builder.CreateCall(F, {TruncA, TruncB}, "sadd");
1136 Value *Add = Builder.CreateExtractValue(Call, 0, "sadd.result");
1137 Value *ZExt = Builder.CreateZExt(Add, OrigAdd->getType());
1138
1139 // The inner add was the result of the narrow add, zero extended to the
1140 // wider type. Replace it with the result computed by the intrinsic.
1141 IC.replaceInstUsesWith(*OrigAdd, ZExt);
1142 IC.eraseInstFromFunction(*OrigAdd);
1143
1144 // The original icmp gets replaced with the overflow value.
1145 return ExtractValueInst::Create(Call, 1, "sadd.overflow");
1146}
1147
1148/// If we have:
1149/// icmp eq/ne (urem/srem %x, %y), 0
1150/// iff %y is a power-of-two, we can replace this with a bit test:
1151/// icmp eq/ne (and %x, (add %y, -1)), 0
1153 // This fold is only valid for equality predicates.
1154 if (!I.isEquality())
1155 return nullptr;
1156 CmpPredicate Pred;
1157 Value *X, *Y, *Zero;
1158 if (!match(&I, m_ICmp(Pred, m_OneUse(m_IRem(m_Value(X), m_Value(Y))),
1159 m_CombineAnd(m_Zero(), m_Value(Zero)))))
1160 return nullptr;
1161 if (!isKnownToBeAPowerOfTwo(Y, /*OrZero*/ true, &I))
1162 return nullptr;
1163 // This may increase instruction count, we don't enforce that Y is a constant.
1164 Value *Mask = Builder.CreateAdd(Y, Constant::getAllOnesValue(Y->getType()));
1165 Value *Masked = Builder.CreateAnd(X, Mask);
1166 return ICmpInst::Create(Instruction::ICmp, Pred, Masked, Zero);
1167}
1168
1169/// Fold equality-comparison between zero and any (maybe truncated) right-shift
1170/// by one-less-than-bitwidth into a sign test on the original value.
1172 Instruction *Val;
1173 CmpPredicate Pred;
1174 if (!I.isEquality() || !match(&I, m_ICmp(Pred, m_Instruction(Val), m_Zero())))
1175 return nullptr;
1176
1177 Value *X;
1178 Type *XTy;
1179
1180 Constant *C;
1181 if (match(Val, m_TruncOrSelf(m_Shr(m_Value(X), m_Constant(C))))) {
1182 XTy = X->getType();
1183 unsigned XBitWidth = XTy->getScalarSizeInBits();
1185 APInt(XBitWidth, XBitWidth - 1))))
1186 return nullptr;
1187 } else if (isa<BinaryOperator>(Val) &&
1189 cast<BinaryOperator>(Val), SQ.getWithInstruction(Val),
1190 /*AnalyzeForSignBitExtraction=*/true))) {
1191 XTy = X->getType();
1192 } else
1193 return nullptr;
1194
1195 return ICmpInst::Create(Instruction::ICmp,
1199}
1200
1201// Handle icmp pred X, 0
1203 CmpInst::Predicate Pred = Cmp.getPredicate();
1204 if (!match(Cmp.getOperand(1), m_Zero()))
1205 return nullptr;
1206
1207 // (icmp sgt smin(PosA, B) 0) -> (icmp sgt B 0)
1208 if (Pred == ICmpInst::ICMP_SGT) {
1209 Value *A, *B;
1210 if (match(Cmp.getOperand(0), m_SMin(m_Value(A), m_Value(B)))) {
1211 if (isKnownPositive(A, SQ.getWithInstruction(&Cmp)))
1212 return new ICmpInst(Pred, B, Cmp.getOperand(1));
1213 if (isKnownPositive(B, SQ.getWithInstruction(&Cmp)))
1214 return new ICmpInst(Pred, A, Cmp.getOperand(1));
1215 }
1216 }
1217
1219 return New;
1220
1221 // Given:
1222 // icmp eq/ne (urem %x, %y), 0
1223 // Iff %x has 0 or 1 bits set, and %y has at least 2 bits set, omit 'urem':
1224 // icmp eq/ne %x, 0
1225 Value *X, *Y;
1226 if (match(Cmp.getOperand(0), m_URem(m_Value(X), m_Value(Y))) &&
1227 ICmpInst::isEquality(Pred)) {
1228 KnownBits XKnown = computeKnownBits(X, &Cmp);
1229 KnownBits YKnown = computeKnownBits(Y, &Cmp);
1230 if (XKnown.countMaxPopulation() == 1 && YKnown.countMinPopulation() >= 2)
1231 return new ICmpInst(Pred, X, Cmp.getOperand(1));
1232 }
1233
1234 // (icmp eq/ne (mul X Y)) -> (icmp eq/ne X/Y) if we know about whether X/Y are
1235 // odd/non-zero/there is no overflow.
1236 if (match(Cmp.getOperand(0), m_Mul(m_Value(X), m_Value(Y))) &&
1237 ICmpInst::isEquality(Pred)) {
1238
1239 KnownBits XKnown = computeKnownBits(X, &Cmp);
1240 // if X % 2 != 0
1241 // (icmp eq/ne Y)
1242 if (XKnown.countMaxTrailingZeros() == 0)
1243 return new ICmpInst(Pred, Y, Cmp.getOperand(1));
1244
1245 KnownBits YKnown = computeKnownBits(Y, &Cmp);
1246 // if Y % 2 != 0
1247 // (icmp eq/ne X)
1248 if (YKnown.countMaxTrailingZeros() == 0)
1249 return new ICmpInst(Pred, X, Cmp.getOperand(1));
1250
1251 auto *BO0 = cast<OverflowingBinaryOperator>(Cmp.getOperand(0));
1252 if (BO0->hasNoUnsignedWrap() || BO0->hasNoSignedWrap()) {
1253 const SimplifyQuery Q = SQ.getWithInstruction(&Cmp);
1254 // `isKnownNonZero` does more analysis than just `!KnownBits.One.isZero()`
1255 // but to avoid unnecessary work, first just if this is an obvious case.
1256
1257 // if X non-zero and NoOverflow(X * Y)
1258 // (icmp eq/ne Y)
1259 if (!XKnown.One.isZero() || isKnownNonZero(X, Q))
1260 return new ICmpInst(Pred, Y, Cmp.getOperand(1));
1261
1262 // if Y non-zero and NoOverflow(X * Y)
1263 // (icmp eq/ne X)
1264 if (!YKnown.One.isZero() || isKnownNonZero(Y, Q))
1265 return new ICmpInst(Pred, X, Cmp.getOperand(1));
1266 }
1267 // Note, we are skipping cases:
1268 // if Y % 2 != 0 AND X % 2 != 0
1269 // (false/true)
1270 // if X non-zero and Y non-zero and NoOverflow(X * Y)
1271 // (false/true)
1272 // Those can be simplified later as we would have already replaced the (icmp
1273 // eq/ne (mul X, Y)) with (icmp eq/ne X/Y) and if X/Y is known non-zero that
1274 // will fold to a constant elsewhere.
1275 }
1276
1277 // (icmp eq/ne f(X), 0) -> (icmp eq/ne X, 0)
1278 // where f(X) == 0 if and only if X == 0
1279 if (ICmpInst::isEquality(Pred))
1280 if (Value *Stripped = stripNullTest(Cmp.getOperand(0)))
1281 return new ICmpInst(Pred, Stripped,
1282 Constant::getNullValue(Stripped->getType()));
1283
1284 return nullptr;
1285}
1286
1287/// Fold icmp eq (num + mask) & ~mask, num
1288/// to
1289/// icmp eq (and num, mask), 0
1290/// Where mask is a low bit mask.
1292 Value *Num;
1293 CmpPredicate Pred;
1294 const APInt *Mask, *Neg;
1295
1296 if (!match(&Cmp,
1297 m_c_ICmp(Pred, m_Value(Num),
1299 m_LowBitMask(Mask))),
1300 m_APInt(Neg))))))
1301 return nullptr;
1302
1303 if (*Neg != ~*Mask)
1304 return nullptr;
1305
1306 if (!ICmpInst::isEquality(Pred))
1307 return nullptr;
1308
1309 // Create new icmp eq (num & mask), 0
1310 auto *NewAnd = Builder.CreateAnd(Num, *Mask);
1311 auto *Zero = Constant::getNullValue(Num->getType());
1312
1313 return new ICmpInst(Pred, NewAnd, Zero);
1314}
1315
1316/// Fold icmp Pred X, C.
1317/// TODO: This code structure does not make sense. The saturating add fold
1318/// should be moved to some other helper and extended as noted below (it is also
1319/// possible that code has been made unnecessary - do we canonicalize IR to
1320/// overflow/saturating intrinsics or not?).
1322 // Match the following pattern, which is a common idiom when writing
1323 // overflow-safe integer arithmetic functions. The source performs an addition
1324 // in wider type and explicitly checks for overflow using comparisons against
1325 // INT_MIN and INT_MAX. Simplify by using the sadd_with_overflow intrinsic.
1326 //
1327 // TODO: This could probably be generalized to handle other overflow-safe
1328 // operations if we worked out the formulas to compute the appropriate magic
1329 // constants.
1330 //
1331 // sum = a + b
1332 // if (sum+128 >u 255) ... -> llvm.sadd.with.overflow.i8
1333 CmpInst::Predicate Pred = Cmp.getPredicate();
1334 Value *Op0 = Cmp.getOperand(0), *Op1 = Cmp.getOperand(1);
1335 Value *A, *B;
1336 ConstantInt *CI, *CI2; // I = icmp ugt (add (add A, B), CI2), CI
1337 if (Pred == ICmpInst::ICMP_UGT && match(Op1, m_ConstantInt(CI)) &&
1338 match(Op0, m_Add(m_Add(m_Value(A), m_Value(B)), m_ConstantInt(CI2))))
1339 if (Instruction *Res = processUGT_ADDCST_ADD(Cmp, A, B, CI2, CI, *this))
1340 return Res;
1341
1342 // icmp(phi(C1, C2, ...), C) -> phi(icmp(C1, C), icmp(C2, C), ...).
1344 if (!C)
1345 return nullptr;
1346
1347 if (auto *Phi = dyn_cast<PHINode>(Op0))
1348 if (all_of(Phi->operands(), IsaPred<Constant>)) {
1350 for (Value *V : Phi->incoming_values()) {
1351 Constant *Res =
1353 if (!Res)
1354 return nullptr;
1355 Ops.push_back(Res);
1356 }
1357 Builder.SetInsertPoint(Phi);
1358 PHINode *NewPhi = Builder.CreatePHI(Cmp.getType(), Phi->getNumOperands());
1359 for (auto [V, Pred] : zip(Ops, Phi->blocks()))
1360 NewPhi->addIncoming(V, Pred);
1361 return replaceInstUsesWith(Cmp, NewPhi);
1362 }
1363
1365 return R;
1366
1367 return nullptr;
1368}
1369
1370/// Canonicalize icmp instructions based on dominating conditions.
1372 // We already checked simple implication in InstSimplify, only handle complex
1373 // cases here.
1374 Value *X = Cmp.getOperand(0), *Y = Cmp.getOperand(1);
1375 const APInt *C;
1376 if (!match(Y, m_APInt(C)))
1377 return nullptr;
1378
1379 CmpInst::Predicate Pred = Cmp.getPredicate();
1381
1382 auto handleDomCond = [&](ICmpInst::Predicate DomPred,
1383 const APInt *DomC) -> Instruction * {
1384 // We have 2 compares of a variable with constants. Calculate the constant
1385 // ranges of those compares to see if we can transform the 2nd compare:
1386 // DomBB:
1387 // DomCond = icmp DomPred X, DomC
1388 // br DomCond, CmpBB, FalseBB
1389 // CmpBB:
1390 // Cmp = icmp Pred X, C
1391 ConstantRange DominatingCR =
1392 ConstantRange::makeExactICmpRegion(DomPred, *DomC);
1393 ConstantRange Intersection = DominatingCR.intersectWith(CR);
1394 ConstantRange Difference = DominatingCR.difference(CR);
1395 if (Intersection.isEmptySet())
1396 return replaceInstUsesWith(Cmp, Builder.getFalse());
1397 if (Difference.isEmptySet())
1398 return replaceInstUsesWith(Cmp, Builder.getTrue());
1399
1400 // Canonicalizing a sign bit comparison that gets used in a branch,
1401 // pessimizes codegen by generating branch on zero instruction instead
1402 // of a test and branch. So we avoid canonicalizing in such situations
1403 // because test and branch instruction has better branch displacement
1404 // than compare and branch instruction.
1405 bool UnusedBit;
1406 bool IsSignBit = isSignBitCheck(Pred, *C, UnusedBit);
1407 if (Cmp.isEquality() || (IsSignBit && hasBranchUse(Cmp)))
1408 return nullptr;
1409
1410 // Avoid an infinite loop with min/max canonicalization.
1411 // TODO: This will be unnecessary if we canonicalize to min/max intrinsics.
1412 if (Cmp.hasOneUse() &&
1413 match(Cmp.user_back(), m_MaxOrMin(m_Value(), m_Value())))
1414 return nullptr;
1415
1416 if (const APInt *EqC = Intersection.getSingleElement())
1417 return new ICmpInst(ICmpInst::ICMP_EQ, X, Builder.getInt(*EqC));
1418 if (const APInt *NeC = Difference.getSingleElement())
1419 return new ICmpInst(ICmpInst::ICMP_NE, X, Builder.getInt(*NeC));
1420 return nullptr;
1421 };
1422
1423 for (CondBrInst *BI : DC.conditionsFor(X)) {
1424 CmpPredicate DomPred;
1425 const APInt *DomC;
1426 if (!match(BI->getCondition(),
1427 m_ICmp(DomPred, m_Specific(X), m_APInt(DomC))))
1428 continue;
1429
1430 BasicBlockEdge Edge0(BI->getParent(), BI->getSuccessor(0));
1431 if (DT.dominates(Edge0, Cmp.getParent())) {
1432 if (auto *V = handleDomCond(DomPred, DomC))
1433 return V;
1434 } else {
1435 BasicBlockEdge Edge1(BI->getParent(), BI->getSuccessor(1));
1436 if (DT.dominates(Edge1, Cmp.getParent()))
1437 if (auto *V =
1438 handleDomCond(CmpInst::getInversePredicate(DomPred), DomC))
1439 return V;
1440 }
1441 }
1442
1443 return nullptr;
1444}
1445
1446/// Fold icmp (trunc X), C.
1448 TruncInst *Trunc,
1449 const APInt &C) {
1450 ICmpInst::Predicate Pred = Cmp.getPredicate();
1451 Value *X = Trunc->getOperand(0);
1452 Type *SrcTy = X->getType();
1453 unsigned DstBits = Trunc->getType()->getScalarSizeInBits(),
1454 SrcBits = SrcTy->getScalarSizeInBits();
1455
1456 // Match (icmp pred (trunc nuw/nsw X), C)
1457 // Which we can convert to (icmp pred X, (sext/zext C))
1458 if (shouldChangeType(Trunc->getType(), SrcTy)) {
1459 if (Trunc->hasNoSignedWrap())
1460 return new ICmpInst(Pred, X, ConstantInt::get(SrcTy, C.sext(SrcBits)));
1461 if (!Cmp.isSigned() && Trunc->hasNoUnsignedWrap())
1462 return new ICmpInst(Pred, X, ConstantInt::get(SrcTy, C.zext(SrcBits)));
1463 }
1464
1465 if (C.isOne() && C.getBitWidth() > 1) {
1466 // icmp slt trunc(signum(V)) 1 --> icmp slt V, 1
1467 Value *V = nullptr;
1468 if (Pred == ICmpInst::ICMP_SLT && match(X, m_Signum(m_Value(V))))
1469 return new ICmpInst(ICmpInst::ICMP_SLT, V,
1470 ConstantInt::get(V->getType(), 1));
1471 }
1472
1473 // TODO: Handle non-equality predicates.
1474 Value *Y;
1475 const APInt *Pow2;
1476 if (Cmp.isEquality() && match(X, m_Shl(m_Power2(Pow2), m_Value(Y))) &&
1477 DstBits > Pow2->logBase2()) {
1478 // (trunc (Pow2 << Y) to iN) == 0 --> Y u>= N - log2(Pow2)
1479 // (trunc (Pow2 << Y) to iN) != 0 --> Y u< N - log2(Pow2)
1480 // iff N > log2(Pow2)
1481 if (C.isZero()) {
1482 auto NewPred = (Pred == Cmp.ICMP_EQ) ? Cmp.ICMP_UGE : Cmp.ICMP_ULT;
1483 return new ICmpInst(NewPred, Y,
1484 ConstantInt::get(SrcTy, DstBits - Pow2->logBase2()));
1485 }
1486 // (trunc (Pow2 << Y) to iN) == 2**C --> Y == C - log2(Pow2)
1487 // (trunc (Pow2 << Y) to iN) != 2**C --> Y != C - log2(Pow2)
1488 if (C.isPowerOf2())
1489 return new ICmpInst(
1490 Pred, Y, ConstantInt::get(SrcTy, C.logBase2() - Pow2->logBase2()));
1491 }
1492
1493 if (Cmp.isEquality() && (Trunc->hasOneUse() || Trunc->hasNoUnsignedWrap())) {
1494 // Canonicalize to a mask and wider compare if the wide type is suitable:
1495 // (trunc X to i8) == C --> (X & 0xff) == (zext C)
1496 if (!SrcTy->isVectorTy() && shouldChangeType(DstBits, SrcBits)) {
1497 Constant *Mask =
1498 ConstantInt::get(SrcTy, APInt::getLowBitsSet(SrcBits, DstBits));
1499 Value *And = Trunc->hasNoUnsignedWrap() ? X : Builder.CreateAnd(X, Mask);
1500 Constant *WideC = ConstantInt::get(SrcTy, C.zext(SrcBits));
1501 return new ICmpInst(Pred, And, WideC);
1502 }
1503
1504 // Simplify icmp eq (trunc x to i8), 42 -> icmp eq x, 42|highbits if all
1505 // of the high bits truncated out of x are known.
1506 KnownBits Known = computeKnownBits(X, &Cmp);
1507
1508 // If all the high bits are known, we can do this xform.
1509 if ((Known.Zero | Known.One).countl_one() >= SrcBits - DstBits) {
1510 // Pull in the high bits from known-ones set.
1511 APInt NewRHS = C.zext(SrcBits);
1512 NewRHS |= Known.One & APInt::getHighBitsSet(SrcBits, SrcBits - DstBits);
1513 return new ICmpInst(Pred, X, ConstantInt::get(SrcTy, NewRHS));
1514 }
1515 }
1516
1517 // Look through truncated right-shift of the sign-bit for a sign-bit check:
1518 // trunc iN (ShOp >> ShAmtC) to i[N - ShAmtC] < 0 --> ShOp < 0
1519 // trunc iN (ShOp >> ShAmtC) to i[N - ShAmtC] > -1 --> ShOp > -1
1520 Value *ShOp;
1521 uint64_t ShAmt;
1522 bool TrueIfSigned;
1523 if (isSignBitCheck(Pred, C, TrueIfSigned) &&
1524 match(X, m_Shr(m_Value(ShOp), m_ConstantInt(ShAmt))) &&
1525 DstBits == SrcBits - ShAmt) {
1526 return TrueIfSigned ? new ICmpInst(ICmpInst::ICMP_SLT, ShOp,
1528 : new ICmpInst(ICmpInst::ICMP_SGT, ShOp,
1530 }
1531
1532 return nullptr;
1533}
1534
1535/// Fold icmp (trunc nuw/nsw X), (trunc nuw/nsw Y).
1536/// Fold icmp (trunc nuw/nsw X), (zext/sext Y).
1539 const SimplifyQuery &Q) {
1540 Value *X, *Y;
1541 CmpPredicate Pred;
1542 bool YIsSExt = false;
1543 // Try to match icmp (trunc X), (trunc Y)
1544 if (match(&Cmp, m_ICmp(Pred, m_Trunc(m_Value(X)), m_Trunc(m_Value(Y))))) {
1545 unsigned NoWrapFlags = cast<TruncInst>(Cmp.getOperand(0))->getNoWrapKind() &
1546 cast<TruncInst>(Cmp.getOperand(1))->getNoWrapKind();
1547 if (Cmp.isSigned()) {
1548 // For signed comparisons, both truncs must be nsw.
1549 if (!(NoWrapFlags & TruncInst::NoSignedWrap))
1550 return nullptr;
1551 } else {
1552 // For unsigned and equality comparisons, either both must be nuw or
1553 // both must be nsw, we don't care which.
1554 if (!NoWrapFlags)
1555 return nullptr;
1556 }
1557
1558 if (X->getType() != Y->getType() &&
1559 (!Cmp.getOperand(0)->hasOneUse() || !Cmp.getOperand(1)->hasOneUse()))
1560 return nullptr;
1561 if (!isDesirableIntType(X->getType()->getScalarSizeInBits()) &&
1562 isDesirableIntType(Y->getType()->getScalarSizeInBits())) {
1563 std::swap(X, Y);
1564 Pred = Cmp.getSwappedPredicate(Pred);
1565 }
1566 YIsSExt = !(NoWrapFlags & TruncInst::NoUnsignedWrap);
1567 }
1568 // Try to match icmp (trunc nuw X), (zext Y)
1569 else if (!Cmp.isSigned() &&
1570 match(&Cmp, m_c_ICmp(Pred, m_NUWTrunc(m_Value(X)),
1571 m_OneUse(m_ZExt(m_Value(Y)))))) {
1572 // Can fold trunc nuw + zext for unsigned and equality predicates.
1573 }
1574 // Try to match icmp (trunc nsw X), (sext Y)
1575 else if (match(&Cmp, m_c_ICmp(Pred, m_NSWTrunc(m_Value(X)),
1577 // Can fold trunc nsw + zext/sext for all predicates.
1578 YIsSExt =
1579 isa<SExtInst>(Cmp.getOperand(0)) || isa<SExtInst>(Cmp.getOperand(1));
1580 } else
1581 return nullptr;
1582
1583 Type *TruncTy = Cmp.getOperand(0)->getType();
1584 unsigned TruncBits = TruncTy->getScalarSizeInBits();
1585
1586 // If this transform will end up changing from desirable types -> undesirable
1587 // types skip it.
1588 if (isDesirableIntType(TruncBits) &&
1589 !isDesirableIntType(X->getType()->getScalarSizeInBits()))
1590 return nullptr;
1591
1592 Value *NewY = Builder.CreateIntCast(Y, X->getType(), YIsSExt);
1593 return new ICmpInst(Pred, X, NewY);
1594}
1595
1596/// Fold icmp (xor X, Y), C.
1599 const APInt &C) {
1600 if (Instruction *I = foldICmpXorShiftConst(Cmp, Xor, C))
1601 return I;
1602
1603 Value *X = Xor->getOperand(0);
1604 Value *Y = Xor->getOperand(1);
1605 const APInt *XorC;
1606 if (!match(Y, m_APInt(XorC)))
1607 return nullptr;
1608
1609 // If this is a comparison that tests the signbit (X < 0) or (x > -1),
1610 // fold the xor.
1611 ICmpInst::Predicate Pred = Cmp.getPredicate();
1612 bool TrueIfSigned = false;
1613 if (isSignBitCheck(Cmp.getPredicate(), C, TrueIfSigned)) {
1614
1615 // If the sign bit of the XorCst is not set, there is no change to
1616 // the operation, just stop using the Xor.
1617 if (!XorC->isNegative())
1618 return replaceOperand(Cmp, 0, X);
1619
1620 // Emit the opposite comparison.
1621 if (TrueIfSigned)
1622 return new ICmpInst(ICmpInst::ICMP_SGT, X,
1623 ConstantInt::getAllOnesValue(X->getType()));
1624 else
1625 return new ICmpInst(ICmpInst::ICMP_SLT, X,
1626 ConstantInt::getNullValue(X->getType()));
1627 }
1628
1629 if (Xor->hasOneUse()) {
1630 // (icmp u/s (xor X SignMask), C) -> (icmp s/u X, (xor C SignMask))
1631 if (!Cmp.isEquality() && XorC->isSignMask()) {
1632 Pred = Cmp.getFlippedSignednessPredicate();
1633 return new ICmpInst(Pred, X, ConstantInt::get(X->getType(), C ^ *XorC));
1634 }
1635
1636 // (icmp u/s (xor X ~SignMask), C) -> (icmp s/u X, (xor C ~SignMask))
1637 if (!Cmp.isEquality() && XorC->isMaxSignedValue()) {
1638 Pred = Cmp.getFlippedSignednessPredicate();
1639 Pred = Cmp.getSwappedPredicate(Pred);
1640 return new ICmpInst(Pred, X, ConstantInt::get(X->getType(), C ^ *XorC));
1641 }
1642 }
1643
1644 // Mask constant magic can eliminate an 'xor' with unsigned compares.
1645 if (Pred == ICmpInst::ICMP_UGT) {
1646 // (xor X, ~C) >u C --> X <u ~C (when C+1 is a power of 2)
1647 if (*XorC == ~C && (C + 1).isPowerOf2())
1648 return new ICmpInst(ICmpInst::ICMP_ULT, X, Y);
1649 // (xor X, C) >u C --> X >u C (when C+1 is a power of 2)
1650 if (*XorC == C && (C + 1).isPowerOf2())
1651 return new ICmpInst(ICmpInst::ICMP_UGT, X, Y);
1652 }
1653 if (Pred == ICmpInst::ICMP_ULT) {
1654 // (xor X, -C) <u C --> X >u ~C (when C is a power of 2)
1655 if (*XorC == -C && C.isPowerOf2())
1656 return new ICmpInst(ICmpInst::ICMP_UGT, X,
1657 ConstantInt::get(X->getType(), ~C));
1658 // (xor X, C) <u C --> X >u ~C (when -C is a power of 2)
1659 if (*XorC == C && (-C).isPowerOf2())
1660 return new ICmpInst(ICmpInst::ICMP_UGT, X,
1661 ConstantInt::get(X->getType(), ~C));
1662 }
1663 return nullptr;
1664}
1665
1666/// For power-of-2 C:
1667/// ((X s>> ShiftC) ^ X) u< C --> (X + C) u< (C << 1)
1668/// ((X s>> ShiftC) ^ X) u> (C - 1) --> (X + C) u> ((C << 1) - 1)
1671 const APInt &C) {
1672 CmpInst::Predicate Pred = Cmp.getPredicate();
1673 APInt PowerOf2;
1674 if (Pred == ICmpInst::ICMP_ULT)
1675 PowerOf2 = C;
1676 else if (Pred == ICmpInst::ICMP_UGT && !C.isMaxValue())
1677 PowerOf2 = C + 1;
1678 else
1679 return nullptr;
1680 if (!PowerOf2.isPowerOf2())
1681 return nullptr;
1682 Value *X;
1683 const APInt *ShiftC;
1685 m_AShr(m_Deferred(X), m_APInt(ShiftC))))))
1686 return nullptr;
1687 uint64_t Shift = ShiftC->getLimitedValue();
1688 Type *XType = X->getType();
1689 if (Shift == 0 || PowerOf2.isMinSignedValue())
1690 return nullptr;
1691 Value *Add = Builder.CreateAdd(X, ConstantInt::get(XType, PowerOf2));
1692 APInt Bound =
1693 Pred == ICmpInst::ICMP_ULT ? PowerOf2 << 1 : ((PowerOf2 << 1) - 1);
1694 return new ICmpInst(Pred, Add, ConstantInt::get(XType, Bound));
1695}
1696
1697/// Fold icmp (and (sh X, Y), C2), C1.
1700 const APInt &C1,
1701 const APInt &C2) {
1702 BinaryOperator *Shift = dyn_cast<BinaryOperator>(And->getOperand(0));
1703 if (!Shift || !Shift->isShift())
1704 return nullptr;
1705
1706 // If this is: (X >> C3) & C2 != C1 (where any shift and any compare could
1707 // exist), turn it into (X & (C2 << C3)) != (C1 << C3). This happens a LOT in
1708 // code produced by the clang front-end, for bitfield access.
1709 // This seemingly simple opportunity to fold away a shift turns out to be
1710 // rather complicated. See PR17827 for details.
1711 unsigned ShiftOpcode = Shift->getOpcode();
1712 bool IsShl = ShiftOpcode == Instruction::Shl;
1713 const APInt *C3;
1714 if (match(Shift->getOperand(1), m_APInt(C3))) {
1715 APInt NewAndCst, NewCmpCst;
1716 bool AnyCmpCstBitsShiftedOut;
1717 if (ShiftOpcode == Instruction::Shl) {
1718 // For a left shift, we can fold if the comparison is not signed. We can
1719 // also fold a signed comparison if the mask value and comparison value
1720 // are not negative. These constraints may not be obvious, but we can
1721 // prove that they are correct using an SMT solver.
1722 if (Cmp.isSigned() && (C2.isNegative() || C1.isNegative()))
1723 return nullptr;
1724
1725 NewCmpCst = C1.lshr(*C3);
1726 NewAndCst = C2.lshr(*C3);
1727 AnyCmpCstBitsShiftedOut = NewCmpCst.shl(*C3) != C1;
1728 } else if (ShiftOpcode == Instruction::LShr) {
1729 // For a logical right shift, we can fold if the comparison is not signed.
1730 // We can also fold a signed comparison if the shifted mask value and the
1731 // shifted comparison value are not negative. These constraints may not be
1732 // obvious, but we can prove that they are correct using an SMT solver.
1733 NewCmpCst = C1.shl(*C3);
1734 NewAndCst = C2.shl(*C3);
1735 AnyCmpCstBitsShiftedOut = NewCmpCst.lshr(*C3) != C1;
1736 if (Cmp.isSigned() && (NewAndCst.isNegative() || NewCmpCst.isNegative()))
1737 return nullptr;
1738 } else {
1739 // For an arithmetic shift, check that both constants don't use (in a
1740 // signed sense) the top bits being shifted out.
1741 assert(ShiftOpcode == Instruction::AShr && "Unknown shift opcode");
1742 NewCmpCst = C1.shl(*C3);
1743 NewAndCst = C2.shl(*C3);
1744 AnyCmpCstBitsShiftedOut = NewCmpCst.ashr(*C3) != C1;
1745 if (NewAndCst.ashr(*C3) != C2)
1746 return nullptr;
1747 }
1748
1749 if (AnyCmpCstBitsShiftedOut) {
1750 // If we shifted bits out, the fold is not going to work out. As a
1751 // special case, check to see if this means that the result is always
1752 // true or false now.
1753 if (Cmp.getPredicate() == ICmpInst::ICMP_EQ)
1754 return replaceInstUsesWith(Cmp, ConstantInt::getFalse(Cmp.getType()));
1755 if (Cmp.getPredicate() == ICmpInst::ICMP_NE)
1756 return replaceInstUsesWith(Cmp, ConstantInt::getTrue(Cmp.getType()));
1757 } else {
1758 Value *NewAnd = Builder.CreateAnd(
1759 Shift->getOperand(0), ConstantInt::get(And->getType(), NewAndCst));
1760 return new ICmpInst(Cmp.getPredicate(), NewAnd,
1761 ConstantInt::get(And->getType(), NewCmpCst));
1762 }
1763 }
1764
1765 // Turn ((X >> Y) & C2) == 0 into (X & (C2 << Y)) == 0. The latter is
1766 // preferable because it allows the C2 << Y expression to be hoisted out of a
1767 // loop if Y is invariant and X is not.
1768 if (Shift->hasOneUse() && C1.isZero() && Cmp.isEquality() &&
1769 !Shift->isArithmeticShift() &&
1770 ((!IsShl && C2.isOne()) || !isa<Constant>(Shift->getOperand(0)))) {
1771 // Compute C2 << Y.
1772 Value *NewShift =
1773 IsShl ? Builder.CreateLShr(And->getOperand(1), Shift->getOperand(1))
1774 : Builder.CreateShl(And->getOperand(1), Shift->getOperand(1));
1775
1776 // Compute X & (C2 << Y).
1777 Value *NewAnd = Builder.CreateAnd(Shift->getOperand(0), NewShift);
1778 return new ICmpInst(Cmp.getPredicate(), NewAnd, Cmp.getOperand(1));
1779 }
1780
1781 return nullptr;
1782}
1783
1784/// Fold icmp (and X, C2), C1.
1787 const APInt &C1) {
1788 bool isICMP_NE = Cmp.getPredicate() == ICmpInst::ICMP_NE;
1789
1790 // icmp ne (and X, 1), 0 --> trunc X to i1
1791 if (isICMP_NE && C1.isZero() && match(And->getOperand(1), m_One()))
1792 return new TruncInst(And->getOperand(0), Cmp.getType());
1793
1794 const APInt *C2;
1795 Value *X;
1796 if (!match(And, m_And(m_Value(X), m_APInt(C2))))
1797 return nullptr;
1798
1799 // (and X, highmask) s> [0, ~highmask] --> X s> ~highmask
1800 if (Cmp.getPredicate() == ICmpInst::ICMP_SGT && C1.ule(~*C2) &&
1801 C2->isNegatedPowerOf2())
1802 return new ICmpInst(ICmpInst::ICMP_SGT, X,
1803 ConstantInt::get(X->getType(), ~*C2));
1804 // (and X, highmask) s< [1, -highmask] --> X s< -highmask
1805 if (Cmp.getPredicate() == ICmpInst::ICMP_SLT && !C1.isSignMask() &&
1806 (C1 - 1).ule(~*C2) && C2->isNegatedPowerOf2() && !C2->isSignMask())
1807 return new ICmpInst(ICmpInst::ICMP_SLT, X,
1808 ConstantInt::get(X->getType(), -*C2));
1809
1810 // Don't perform the following transforms if the AND has multiple uses
1811 if (!And->hasOneUse())
1812 return nullptr;
1813
1814 if (Cmp.isEquality() && C1.isZero()) {
1815 // Restrict this fold to single-use 'and' (PR10267).
1816 // Replace (and X, (1 << size(X)-1) != 0) with X s< 0
1817 if (C2->isSignMask()) {
1818 Constant *Zero = Constant::getNullValue(X->getType());
1819 auto NewPred = isICMP_NE ? ICmpInst::ICMP_SLT : ICmpInst::ICMP_SGE;
1820 return new ICmpInst(NewPred, X, Zero);
1821 }
1822
1823 APInt NewC2 = *C2;
1824 KnownBits Know = computeKnownBits(And->getOperand(0), And);
1825 // Set high zeros of C2 to allow matching negated power-of-2.
1826 NewC2 = *C2 | APInt::getHighBitsSet(C2->getBitWidth(),
1827 Know.countMinLeadingZeros());
1828
1829 // Restrict this fold only for single-use 'and' (PR10267).
1830 // ((%x & C) == 0) --> %x u< (-C) iff (-C) is power of two.
1831 if (NewC2.isNegatedPowerOf2()) {
1832 Constant *NegBOC = ConstantInt::get(And->getType(), -NewC2);
1833 auto NewPred = isICMP_NE ? ICmpInst::ICMP_UGE : ICmpInst::ICMP_ULT;
1834 return new ICmpInst(NewPred, X, NegBOC);
1835 }
1836 }
1837
1838 // If the LHS is an 'and' of a truncate and we can widen the and/compare to
1839 // the input width without changing the value produced, eliminate the cast:
1840 //
1841 // icmp (and (trunc W), C2), C1 -> icmp (and W, C2'), C1'
1842 //
1843 // We can do this transformation if the constants do not have their sign bits
1844 // set or if it is an equality comparison. Extending a relational comparison
1845 // when we're checking the sign bit would not work.
1846 Value *W;
1847 if (match(And->getOperand(0), m_OneUse(m_Trunc(m_Value(W)))) &&
1848 (Cmp.isEquality() || (!C1.isNegative() && !C2->isNegative()))) {
1849 // TODO: Is this a good transform for vectors? Wider types may reduce
1850 // throughput. Should this transform be limited (even for scalars) by using
1851 // shouldChangeType()?
1852 if (!Cmp.getType()->isVectorTy()) {
1853 Type *WideType = W->getType();
1854 unsigned WideScalarBits = WideType->getScalarSizeInBits();
1855 Constant *ZextC1 = ConstantInt::get(WideType, C1.zext(WideScalarBits));
1856 Constant *ZextC2 = ConstantInt::get(WideType, C2->zext(WideScalarBits));
1857 Value *NewAnd = Builder.CreateAnd(W, ZextC2, And->getName());
1858 return new ICmpInst(Cmp.getPredicate(), NewAnd, ZextC1);
1859 }
1860 }
1861
1862 if (Instruction *I = foldICmpAndShift(Cmp, And, C1, *C2))
1863 return I;
1864
1865 // (icmp pred (and (or (lshr A, B), A), 1), 0) -->
1866 // (icmp pred (and A, (or (shl 1, B), 1), 0))
1867 //
1868 // iff pred isn't signed
1869 if (!Cmp.isSigned() && C1.isZero() && And->getOperand(0)->hasOneUse() &&
1870 match(And->getOperand(1), m_One())) {
1871 Constant *One = cast<Constant>(And->getOperand(1));
1872 Value *Or = And->getOperand(0);
1873 Value *A, *B, *LShr;
1874 if (match(Or, m_Or(m_Value(LShr), m_Value(A))) &&
1875 match(LShr, m_LShr(m_Specific(A), m_Value(B)))) {
1876 unsigned UsesRemoved = 0;
1877 if (And->hasOneUse())
1878 ++UsesRemoved;
1879 if (Or->hasOneUse())
1880 ++UsesRemoved;
1881 if (LShr->hasOneUse())
1882 ++UsesRemoved;
1883
1884 // Compute A & ((1 << B) | 1)
1885 unsigned RequireUsesRemoved = match(B, m_ImmConstant()) ? 1 : 3;
1886 if (UsesRemoved >= RequireUsesRemoved) {
1887 Value *NewOr =
1888 Builder.CreateOr(Builder.CreateShl(One, B, LShr->getName(),
1889 /*HasNUW=*/true),
1890 One, Or->getName());
1891 Value *NewAnd = Builder.CreateAnd(A, NewOr, And->getName());
1892 return new ICmpInst(Cmp.getPredicate(), NewAnd, Cmp.getOperand(1));
1893 }
1894 }
1895 }
1896
1897 // (icmp eq (and (bitcast X to int), ExponentMask), ExponentMask) -->
1898 // llvm.is.fpclass(X, fcInf|fcNan)
1899 // (icmp ne (and (bitcast X to int), ExponentMask), ExponentMask) -->
1900 // llvm.is.fpclass(X, ~(fcInf|fcNan))
1901 // (icmp eq (and (bitcast X to int), ExponentMask), 0) -->
1902 // llvm.is.fpclass(X, fcSubnormal|fcZero)
1903 // (icmp ne (and (bitcast X to int), ExponentMask), 0) -->
1904 // llvm.is.fpclass(X, ~(fcSubnormal|fcZero))
1905 Value *V;
1906 if (!Cmp.getParent()->getParent()->hasFnAttribute(
1907 Attribute::NoImplicitFloat) &&
1908 Cmp.isEquality() &&
1910 Type *FPType = V->getType()->getScalarType();
1911 if (FPType->isIEEELikeFPTy() && (C1.isZero() || C1 == *C2)) {
1912 APInt ExponentMask =
1913 APFloat::getInf(FPType->getFltSemantics()).bitcastToAPInt();
1914 if (*C2 == ExponentMask) {
1915 unsigned Mask = C1.isZero()
1918 if (isICMP_NE)
1919 Mask = ~Mask & fcAllFlags;
1920 return replaceInstUsesWith(Cmp, Builder.createIsFPClass(V, Mask));
1921 }
1922 }
1923 }
1924
1925 return nullptr;
1926}
1927
1928/// Fold icmp (and X, Y), C.
1931 const APInt &C) {
1932 if (Instruction *I = foldICmpAndConstConst(Cmp, And, C))
1933 return I;
1934
1935 const ICmpInst::Predicate Pred = Cmp.getPredicate();
1936 bool TrueIfNeg;
1937 if (isSignBitCheck(Pred, C, TrueIfNeg)) {
1938 // ((X - 1) & ~X) < 0 --> X == 0
1939 // ((X - 1) & ~X) >= 0 --> X != 0
1940 Value *X;
1941 if (match(And->getOperand(0), m_Add(m_Value(X), m_AllOnes())) &&
1942 match(And->getOperand(1), m_Not(m_Specific(X)))) {
1943 auto NewPred = TrueIfNeg ? CmpInst::ICMP_EQ : CmpInst::ICMP_NE;
1944 return new ICmpInst(NewPred, X, ConstantInt::getNullValue(X->getType()));
1945 }
1946 // (X & -X) < 0 --> X == MinSignedC
1947 // (X & -X) > -1 --> X != MinSignedC
1948 if (match(And, m_c_And(m_Neg(m_Value(X)), m_Deferred(X)))) {
1949 Constant *MinSignedC = ConstantInt::get(
1950 X->getType(),
1951 APInt::getSignedMinValue(X->getType()->getScalarSizeInBits()));
1952 auto NewPred = TrueIfNeg ? CmpInst::ICMP_EQ : CmpInst::ICMP_NE;
1953 return new ICmpInst(NewPred, X, MinSignedC);
1954 }
1955 }
1956
1957 // TODO: These all require that Y is constant too, so refactor with the above.
1958
1959 // Try to optimize things like "A[i] & 42 == 0" to index computations.
1960 Value *X = And->getOperand(0);
1961 Value *Y = And->getOperand(1);
1962 if (auto *C2 = dyn_cast<ConstantInt>(Y))
1963 if (auto *LI = dyn_cast<LoadInst>(X))
1964 if (auto *GEP = dyn_cast<GetElementPtrInst>(LI->getOperand(0)))
1965 if (Instruction *Res = foldCmpLoadFromIndexedGlobal(LI, GEP, Cmp, C2))
1966 return Res;
1967
1968 if (!Cmp.isEquality())
1969 return nullptr;
1970
1971 // X & -C == -C -> X > u ~C
1972 // X & -C != -C -> X <= u ~C
1973 // iff C is a power of 2
1974 if (Cmp.getOperand(1) == Y && C.isNegatedPowerOf2()) {
1975 auto NewPred =
1977 return new ICmpInst(NewPred, X, SubOne(cast<Constant>(Cmp.getOperand(1))));
1978 }
1979
1980 // ((zext i1 X) & Y) == 0 --> !((trunc Y) & X)
1981 // ((zext i1 X) & Y) != 0 --> ((trunc Y) & X)
1982 // ((zext i1 X) & Y) == 1 --> ((trunc Y) & X)
1983 // ((zext i1 X) & Y) != 1 --> !((trunc Y) & X)
1985 X->getType()->isIntOrIntVectorTy(1) && (C.isZero() || C.isOne())) {
1986 Value *TruncY = Builder.CreateTrunc(Y, X->getType());
1987 if (C.isZero() ^ (Pred == CmpInst::ICMP_NE)) {
1988 Value *And = Builder.CreateAnd(TruncY, X);
1990 }
1991 return BinaryOperator::CreateAnd(TruncY, X);
1992 }
1993
1994 // (icmp eq/ne (and (shl -1, X), Y), 0)
1995 // -> (icmp eq/ne (lshr Y, X), 0)
1996 // We could technically handle any C == 0 or (C < 0 && isOdd(C)) but it seems
1997 // highly unlikely the non-zero case will ever show up in code.
1998 if (C.isZero() &&
2000 m_Value(Y))))) {
2001 Value *LShr = Builder.CreateLShr(Y, X);
2002 return new ICmpInst(Pred, LShr, Constant::getNullValue(LShr->getType()));
2003 }
2004
2005 // (icmp eq/ne (and (add A, Addend), Msk), C)
2006 // -> (icmp eq/ne (and A, Msk), (and (sub C, Addend), Msk))
2007 {
2008 Value *A;
2009 const APInt *Addend, *Msk;
2011 m_LowBitMask(Msk)))) &&
2012 C.ule(*Msk)) {
2013 APInt NewComperand = (C - *Addend) & *Msk;
2014 Value *MaskA = Builder.CreateAnd(A, ConstantInt::get(A->getType(), *Msk));
2015 return new ICmpInst(Pred, MaskA,
2016 ConstantInt::get(MaskA->getType(), NewComperand));
2017 }
2018 }
2019
2020 return nullptr;
2021}
2022
2023/// Fold icmp eq/ne (or (xor/sub (X1, X2), xor/sub (X3, X4))), 0.
2025 InstCombiner::BuilderTy &Builder) {
2026 // Are we using xors or subs to bitwise check for a pair or pairs of
2027 // (in)equalities? Convert to a shorter form that has more potential to be
2028 // folded even further.
2029 // ((X1 ^/- X2) || (X3 ^/- X4)) == 0 --> (X1 == X2) && (X3 == X4)
2030 // ((X1 ^/- X2) || (X3 ^/- X4)) != 0 --> (X1 != X2) || (X3 != X4)
2031 // ((X1 ^/- X2) || (X3 ^/- X4) || (X5 ^/- X6)) == 0 -->
2032 // (X1 == X2) && (X3 == X4) && (X5 == X6)
2033 // ((X1 ^/- X2) || (X3 ^/- X4) || (X5 ^/- X6)) != 0 -->
2034 // (X1 != X2) || (X3 != X4) || (X5 != X6)
2036 SmallVector<Value *, 16> WorkList(1, Or);
2037
2038 while (!WorkList.empty()) {
2039 auto MatchOrOperatorArgument = [&](Value *OrOperatorArgument) {
2040 Value *Lhs, *Rhs;
2041
2042 if (match(OrOperatorArgument,
2043 m_OneUse(m_Xor(m_Value(Lhs), m_Value(Rhs))))) {
2044 CmpValues.emplace_back(Lhs, Rhs);
2045 return;
2046 }
2047
2048 if (match(OrOperatorArgument,
2049 m_OneUse(m_Sub(m_Value(Lhs), m_Value(Rhs))))) {
2050 CmpValues.emplace_back(Lhs, Rhs);
2051 return;
2052 }
2053
2054 WorkList.push_back(OrOperatorArgument);
2055 };
2056
2057 Value *CurrentValue = WorkList.pop_back_val();
2058 Value *OrOperatorLhs, *OrOperatorRhs;
2059
2060 if (!match(CurrentValue,
2061 m_Or(m_Value(OrOperatorLhs), m_Value(OrOperatorRhs)))) {
2062 return nullptr;
2063 }
2064
2065 MatchOrOperatorArgument(OrOperatorRhs);
2066 MatchOrOperatorArgument(OrOperatorLhs);
2067 }
2068
2069 ICmpInst::Predicate Pred = Cmp.getPredicate();
2070 auto BOpc = Pred == CmpInst::ICMP_EQ ? Instruction::And : Instruction::Or;
2071 Value *LhsCmp = Builder.CreateICmp(Pred, CmpValues.rbegin()->first,
2072 CmpValues.rbegin()->second);
2073
2074 for (auto It = CmpValues.rbegin() + 1; It != CmpValues.rend(); ++It) {
2075 Value *RhsCmp = Builder.CreateICmp(Pred, It->first, It->second);
2076 LhsCmp = Builder.CreateBinOp(BOpc, LhsCmp, RhsCmp);
2077 }
2078
2079 return LhsCmp;
2080}
2081
2082/// Fold icmp (or X, Y), C.
2085 const APInt &C) {
2086 ICmpInst::Predicate Pred = Cmp.getPredicate();
2087 if (C.isOne()) {
2088 // icmp slt signum(V) 1 --> icmp slt V, 1
2089 Value *V = nullptr;
2090 if (Pred == ICmpInst::ICMP_SLT && match(Or, m_Signum(m_Value(V))))
2091 return new ICmpInst(ICmpInst::ICMP_SLT, V,
2092 ConstantInt::get(V->getType(), 1));
2093 }
2094
2095 Value *OrOp0 = Or->getOperand(0), *OrOp1 = Or->getOperand(1);
2096
2097 // (icmp eq/ne (or disjoint x, C0), C1)
2098 // -> (icmp eq/ne x, C0^C1)
2099 if (Cmp.isEquality() && match(OrOp1, m_ImmConstant()) &&
2100 cast<PossiblyDisjointInst>(Or)->isDisjoint()) {
2101 Value *NewC =
2102 Builder.CreateXor(OrOp1, ConstantInt::get(OrOp1->getType(), C));
2103 return new ICmpInst(Pred, OrOp0, NewC);
2104 }
2105
2106 const APInt *MaskC;
2107 if (match(OrOp1, m_APInt(MaskC)) && Cmp.isEquality()) {
2108 if (*MaskC == C && (C + 1).isPowerOf2()) {
2109 // X | C == C --> X <=u C
2110 // X | C != C --> X >u C
2111 // iff C+1 is a power of 2 (C is a bitmask of the low bits)
2113 return new ICmpInst(Pred, OrOp0, OrOp1);
2114 }
2115
2116 // More general: canonicalize 'equality with set bits mask' to
2117 // 'equality with clear bits mask'.
2118 // (X | MaskC) == C --> (X & ~MaskC) == C ^ MaskC
2119 // (X | MaskC) != C --> (X & ~MaskC) != C ^ MaskC
2120 if (Or->hasOneUse()) {
2121 Value *And = Builder.CreateAnd(OrOp0, ~(*MaskC));
2122 Constant *NewC = ConstantInt::get(Or->getType(), C ^ (*MaskC));
2123 return new ICmpInst(Pred, And, NewC);
2124 }
2125 }
2126
2127 // (X | (X-1)) s< 0 --> X s< 1
2128 // (X | (X-1)) s> -1 --> X s> 0
2129 Value *X;
2130 bool TrueIfSigned;
2131 if (isSignBitCheck(Pred, C, TrueIfSigned) &&
2133 auto NewPred = TrueIfSigned ? ICmpInst::ICMP_SLT : ICmpInst::ICMP_SGT;
2134 Constant *NewC = ConstantInt::get(X->getType(), TrueIfSigned ? 1 : 0);
2135 return new ICmpInst(NewPred, X, NewC);
2136 }
2137
2138 const APInt *OrC;
2139 // icmp(X | OrC, C) --> icmp(X, 0)
2140 if (C.isNonNegative() && match(Or, m_Or(m_Value(X), m_APInt(OrC)))) {
2141 switch (Pred) {
2142 // X | OrC s< C --> X s< 0 iff OrC s>= C s>= 0
2143 case ICmpInst::ICMP_SLT:
2144 // X | OrC s>= C --> X s>= 0 iff OrC s>= C s>= 0
2145 case ICmpInst::ICMP_SGE:
2146 if (OrC->sge(C))
2147 return new ICmpInst(Pred, X, ConstantInt::getNullValue(X->getType()));
2148 break;
2149 // X | OrC s<= C --> X s< 0 iff OrC s> C s>= 0
2150 case ICmpInst::ICMP_SLE:
2151 // X | OrC s> C --> X s>= 0 iff OrC s> C s>= 0
2152 case ICmpInst::ICMP_SGT:
2153 if (OrC->sgt(C))
2155 ConstantInt::getNullValue(X->getType()));
2156 break;
2157 default:
2158 break;
2159 }
2160 }
2161
2162 if (!Cmp.isEquality() || !C.isZero() || !Or->hasOneUse())
2163 return nullptr;
2164
2165 Value *P, *Q;
2167 // Simplify icmp eq (or (ptrtoint P), (ptrtoint Q)), 0
2168 // -> and (icmp eq P, null), (icmp eq Q, null).
2169 Value *CmpP =
2170 Builder.CreateICmp(Pred, P, ConstantInt::getNullValue(P->getType()));
2171 Value *CmpQ =
2172 Builder.CreateICmp(Pred, Q, ConstantInt::getNullValue(Q->getType()));
2173 auto BOpc = Pred == CmpInst::ICMP_EQ ? Instruction::And : Instruction::Or;
2174 return BinaryOperator::Create(BOpc, CmpP, CmpQ);
2175 }
2176
2177 if (Value *V = foldICmpOrXorSubChain(Cmp, Or, Builder))
2178 return replaceInstUsesWith(Cmp, V);
2179
2180 return nullptr;
2181}
2182
2183/// Fold icmp (mul X, Y), C.
2186 const APInt &C) {
2187 ICmpInst::Predicate Pred = Cmp.getPredicate();
2188 Type *MulTy = Mul->getType();
2189 Value *X = Mul->getOperand(0);
2190
2191 // If there's no overflow:
2192 // X * X == 0 --> X == 0
2193 // X * X != 0 --> X != 0
2194 if (Cmp.isEquality() && C.isZero() && X == Mul->getOperand(1) &&
2195 (Mul->hasNoUnsignedWrap() || Mul->hasNoSignedWrap()))
2196 return new ICmpInst(Pred, X, ConstantInt::getNullValue(MulTy));
2197
2198 const APInt *MulC;
2199 if (!match(Mul->getOperand(1), m_APInt(MulC)))
2200 return nullptr;
2201
2202 // If this is a test of the sign bit and the multiply is sign-preserving with
2203 // a constant operand, use the multiply LHS operand instead:
2204 // (X * +MulC) < 0 --> X < 0
2205 // (X * -MulC) < 0 --> X > 0
2206 if (isSignTest(Pred, C) && Mul->hasNoSignedWrap()) {
2207 if (MulC->isNegative())
2208 Pred = ICmpInst::getSwappedPredicate(Pred);
2209 return new ICmpInst(Pred, X, ConstantInt::getNullValue(MulTy));
2210 }
2211
2212 if (MulC->isZero())
2213 return nullptr;
2214
2215 // If the multiply does not wrap or the constant is odd, try to divide the
2216 // compare constant by the multiplication factor.
2217 if (Cmp.isEquality()) {
2218 // (mul nsw X, MulC) eq/ne C --> X eq/ne C /s MulC
2219 if (Mul->hasNoSignedWrap() && C.srem(*MulC).isZero()) {
2220 Constant *NewC = ConstantInt::get(MulTy, C.sdiv(*MulC));
2221 return new ICmpInst(Pred, X, NewC);
2222 }
2223
2224 // C % MulC == 0 is weaker than we could use if MulC is odd because it
2225 // correct to transform if MulC * N == C including overflow. I.e with i8
2226 // (icmp eq (mul X, 5), 101) -> (icmp eq X, 225) but since 101 % 5 != 0, we
2227 // miss that case.
2228 if (C.urem(*MulC).isZero()) {
2229 // (mul nuw X, MulC) eq/ne C --> X eq/ne C /u MulC
2230 // (mul X, OddC) eq/ne N * C --> X eq/ne N
2231 if ((*MulC & 1).isOne() || Mul->hasNoUnsignedWrap()) {
2232 Constant *NewC = ConstantInt::get(MulTy, C.udiv(*MulC));
2233 return new ICmpInst(Pred, X, NewC);
2234 }
2235 }
2236 }
2237
2238 // With a matching no-overflow guarantee, fold the constants:
2239 // (X * MulC) < C --> X < (C / MulC)
2240 // (X * MulC) > C --> X > (C / MulC)
2241 // TODO: Assert that Pred is not equal to SGE, SLE, UGE, ULE?
2242 Constant *NewC = nullptr;
2243 if (Mul->hasNoSignedWrap() && ICmpInst::isSigned(Pred)) {
2244 // MININT / -1 --> overflow.
2245 if (C.isMinSignedValue() && MulC->isAllOnes())
2246 return nullptr;
2247 if (MulC->isNegative())
2248 Pred = ICmpInst::getSwappedPredicate(Pred);
2249
2250 if (Pred == ICmpInst::ICMP_SLT || Pred == ICmpInst::ICMP_SGE) {
2251 NewC = ConstantInt::get(
2253 } else {
2254 assert((Pred == ICmpInst::ICMP_SLE || Pred == ICmpInst::ICMP_SGT) &&
2255 "Unexpected predicate");
2256 NewC = ConstantInt::get(
2258 }
2259 } else if (Mul->hasNoUnsignedWrap() && ICmpInst::isUnsigned(Pred)) {
2260 if (Pred == ICmpInst::ICMP_ULT || Pred == ICmpInst::ICMP_UGE) {
2261 NewC = ConstantInt::get(
2263 } else {
2264 assert((Pred == ICmpInst::ICMP_ULE || Pred == ICmpInst::ICMP_UGT) &&
2265 "Unexpected predicate");
2266 NewC = ConstantInt::get(
2268 }
2269 }
2270
2271 return NewC ? new ICmpInst(Pred, X, NewC) : nullptr;
2272}
2273
2274/// Fold icmp (shl nuw C2, Y), C.
2276 const APInt &C) {
2277 Value *Y;
2278 const APInt *C2;
2279 if (!match(Shl, m_NUWShl(m_APInt(C2), m_Value(Y))))
2280 return nullptr;
2281
2282 Type *ShiftType = Shl->getType();
2283 unsigned TypeBits = C.getBitWidth();
2284 ICmpInst::Predicate Pred = Cmp.getPredicate();
2285 if (Cmp.isUnsigned()) {
2286 if (C2->isZero() || C2->ugt(C))
2287 return nullptr;
2288 APInt Div, Rem;
2289 APInt::udivrem(C, *C2, Div, Rem);
2290 bool CIsPowerOf2 = Rem.isZero() && Div.isPowerOf2();
2291
2292 // (1 << Y) pred C -> Y pred Log2(C)
2293 if (!CIsPowerOf2) {
2294 // (1 << Y) < 30 -> Y <= 4
2295 // (1 << Y) <= 30 -> Y <= 4
2296 // (1 << Y) >= 30 -> Y > 4
2297 // (1 << Y) > 30 -> Y > 4
2298 if (Pred == ICmpInst::ICMP_ULT)
2299 Pred = ICmpInst::ICMP_ULE;
2300 else if (Pred == ICmpInst::ICMP_UGE)
2301 Pred = ICmpInst::ICMP_UGT;
2302 }
2303
2304 unsigned CLog2 = Div.logBase2();
2305 return new ICmpInst(Pred, Y, ConstantInt::get(ShiftType, CLog2));
2306 } else if (Cmp.isSigned() && C2->isOne()) {
2307 Constant *BitWidthMinusOne = ConstantInt::get(ShiftType, TypeBits - 1);
2308 // (1 << Y) > 0 -> Y != 31
2309 // (1 << Y) > C -> Y != 31 if C is negative.
2310 if (Pred == ICmpInst::ICMP_SGT && C.sle(0))
2311 return new ICmpInst(ICmpInst::ICMP_NE, Y, BitWidthMinusOne);
2312
2313 // (1 << Y) < 0 -> Y == 31
2314 // (1 << Y) < 1 -> Y == 31
2315 // (1 << Y) < C -> Y == 31 if C is negative and not signed min.
2316 // Exclude signed min by subtracting 1 and lower the upper bound to 0.
2317 if (Pred == ICmpInst::ICMP_SLT && (C - 1).sle(0))
2318 return new ICmpInst(ICmpInst::ICMP_EQ, Y, BitWidthMinusOne);
2319 }
2320
2321 return nullptr;
2322}
2323
2324/// Fold icmp (shl X, Y), C.
2326 BinaryOperator *Shl,
2327 const APInt &C) {
2328 const APInt *ShiftVal;
2329 if (Cmp.isEquality() && match(Shl->getOperand(0), m_APInt(ShiftVal)))
2330 return foldICmpShlConstConst(Cmp, Shl->getOperand(1), C, *ShiftVal);
2331
2332 ICmpInst::Predicate Pred = Cmp.getPredicate();
2333 // (icmp pred (shl nuw&nsw X, Y), Csle0)
2334 // -> (icmp pred X, Csle0)
2335 //
2336 // The idea is the nuw/nsw essentially freeze the sign bit for the shift op
2337 // so X's must be what is used.
2338 if (C.sle(0) && Shl->hasNoUnsignedWrap() && Shl->hasNoSignedWrap())
2339 return new ICmpInst(Pred, Shl->getOperand(0), Cmp.getOperand(1));
2340
2341 // (icmp eq/ne (shl nuw|nsw X, Y), 0)
2342 // -> (icmp eq/ne X, 0)
2343 if (ICmpInst::isEquality(Pred) && C.isZero() &&
2344 (Shl->hasNoUnsignedWrap() || Shl->hasNoSignedWrap()))
2345 return new ICmpInst(Pred, Shl->getOperand(0), Cmp.getOperand(1));
2346
2347 // (icmp slt (shl nsw X, Y), 0/1)
2348 // -> (icmp slt X, 0/1)
2349 // (icmp sgt (shl nsw X, Y), 0/-1)
2350 // -> (icmp sgt X, 0/-1)
2351 //
2352 // NB: sge/sle with a constant will canonicalize to sgt/slt.
2353 if (Shl->hasNoSignedWrap() &&
2354 (Pred == ICmpInst::ICMP_SGT || Pred == ICmpInst::ICMP_SLT))
2355 if (C.isZero() || (Pred == ICmpInst::ICMP_SGT ? C.isAllOnes() : C.isOne()))
2356 return new ICmpInst(Pred, Shl->getOperand(0), Cmp.getOperand(1));
2357
2358 const APInt *ShiftAmt;
2359 if (!match(Shl->getOperand(1), m_APInt(ShiftAmt)))
2360 return foldICmpShlLHSC(Cmp, Shl, C);
2361
2362 // Check that the shift amount is in range. If not, don't perform undefined
2363 // shifts. When the shift is visited, it will be simplified.
2364 unsigned TypeBits = C.getBitWidth();
2365 if (ShiftAmt->uge(TypeBits))
2366 return nullptr;
2367
2368 Value *X = Shl->getOperand(0);
2369 Type *ShType = Shl->getType();
2370
2371 // NSW guarantees that we are only shifting out sign bits from the high bits,
2372 // so we can ASHR the compare constant without needing a mask and eliminate
2373 // the shift.
2374 if (Shl->hasNoSignedWrap()) {
2375 if (Pred == ICmpInst::ICMP_SGT) {
2376 // icmp Pred (shl nsw X, ShiftAmt), C --> icmp Pred X, (C >>s ShiftAmt)
2377 APInt ShiftedC = C.ashr(*ShiftAmt);
2378 return new ICmpInst(Pred, X, ConstantInt::get(ShType, ShiftedC));
2379 }
2380 if ((Pred == ICmpInst::ICMP_EQ || Pred == ICmpInst::ICMP_NE) &&
2381 C.ashr(*ShiftAmt).shl(*ShiftAmt) == C) {
2382 APInt ShiftedC = C.ashr(*ShiftAmt);
2383 return new ICmpInst(Pred, X, ConstantInt::get(ShType, ShiftedC));
2384 }
2385 if (Pred == ICmpInst::ICMP_SLT) {
2386 // SLE is the same as above, but SLE is canonicalized to SLT, so convert:
2387 // (X << S) <=s C is equiv to X <=s (C >> S) for all C
2388 // (X << S) <s (C + 1) is equiv to X <s (C >> S) + 1 if C <s SMAX
2389 // (X << S) <s C is equiv to X <s ((C - 1) >> S) + 1 if C >s SMIN
2390 assert(!C.isMinSignedValue() && "Unexpected icmp slt");
2391 APInt ShiftedC = (C - 1).ashr(*ShiftAmt) + 1;
2392 return new ICmpInst(Pred, X, ConstantInt::get(ShType, ShiftedC));
2393 }
2394 }
2395
2396 // NUW guarantees that we are only shifting out zero bits from the high bits,
2397 // so we can LSHR the compare constant without needing a mask and eliminate
2398 // the shift.
2399 if (Shl->hasNoUnsignedWrap()) {
2400 if (Pred == ICmpInst::ICMP_UGT) {
2401 // icmp Pred (shl nuw X, ShiftAmt), C --> icmp Pred X, (C >>u ShiftAmt)
2402 APInt ShiftedC = C.lshr(*ShiftAmt);
2403 return new ICmpInst(Pred, X, ConstantInt::get(ShType, ShiftedC));
2404 }
2405 if ((Pred == ICmpInst::ICMP_EQ || Pred == ICmpInst::ICMP_NE) &&
2406 C.lshr(*ShiftAmt).shl(*ShiftAmt) == C) {
2407 APInt ShiftedC = C.lshr(*ShiftAmt);
2408 return new ICmpInst(Pred, X, ConstantInt::get(ShType, ShiftedC));
2409 }
2410 if (Pred == ICmpInst::ICMP_ULT) {
2411 // ULE is the same as above, but ULE is canonicalized to ULT, so convert:
2412 // (X << S) <=u C is equiv to X <=u (C >> S) for all C
2413 // (X << S) <u (C + 1) is equiv to X <u (C >> S) + 1 if C <u ~0u
2414 // (X << S) <u C is equiv to X <u ((C - 1) >> S) + 1 if C >u 0
2415 assert(C.ugt(0) && "ult 0 should have been eliminated");
2416 APInt ShiftedC = (C - 1).lshr(*ShiftAmt) + 1;
2417 return new ICmpInst(Pred, X, ConstantInt::get(ShType, ShiftedC));
2418 }
2419 }
2420
2421 if (Cmp.isEquality() && Shl->hasOneUse()) {
2422 // Strength-reduce the shift into an 'and'.
2423 Constant *Mask = ConstantInt::get(
2424 ShType,
2425 APInt::getLowBitsSet(TypeBits, TypeBits - ShiftAmt->getZExtValue()));
2426 Value *And = Builder.CreateAnd(X, Mask, Shl->getName() + ".mask");
2427 Constant *LShrC = ConstantInt::get(ShType, C.lshr(*ShiftAmt));
2428 return new ICmpInst(Pred, And, LShrC);
2429 }
2430
2431 // Otherwise, if this is a comparison of the sign bit, simplify to and/test.
2432 bool TrueIfSigned = false;
2433 if (Shl->hasOneUse() && isSignBitCheck(Pred, C, TrueIfSigned)) {
2434 // (X << 31) <s 0 --> (X & 1) != 0
2435 Constant *Mask = ConstantInt::get(
2436 ShType,
2437 APInt::getOneBitSet(TypeBits, TypeBits - ShiftAmt->getZExtValue() - 1));
2438 Value *And = Builder.CreateAnd(X, Mask, Shl->getName() + ".mask");
2439 return new ICmpInst(TrueIfSigned ? ICmpInst::ICMP_NE : ICmpInst::ICMP_EQ,
2440 And, Constant::getNullValue(ShType));
2441 }
2442
2443 // Simplify 'shl' inequality test into 'and' equality test.
2444 if (Cmp.isUnsigned() && Shl->hasOneUse()) {
2445 // (X l<< C2) u<=/u> C1 iff C1+1 is power of two -> X & (~C1 l>> C2) ==/!= 0
2446 if ((C + 1).isPowerOf2() &&
2447 (Pred == ICmpInst::ICMP_ULE || Pred == ICmpInst::ICMP_UGT)) {
2448 Value *And = Builder.CreateAnd(X, (~C).lshr(ShiftAmt->getZExtValue()));
2449 return new ICmpInst(Pred == ICmpInst::ICMP_ULE ? ICmpInst::ICMP_EQ
2451 And, Constant::getNullValue(ShType));
2452 }
2453 // (X l<< C2) u</u>= C1 iff C1 is power of two -> X & (-C1 l>> C2) ==/!= 0
2454 if (C.isPowerOf2() &&
2455 (Pred == ICmpInst::ICMP_ULT || Pred == ICmpInst::ICMP_UGE)) {
2456 Value *And =
2457 Builder.CreateAnd(X, (~(C - 1)).lshr(ShiftAmt->getZExtValue()));
2458 return new ICmpInst(Pred == ICmpInst::ICMP_ULT ? ICmpInst::ICMP_EQ
2460 And, Constant::getNullValue(ShType));
2461 }
2462 }
2463
2464 // Transform (icmp pred iM (shl iM %v, N), C)
2465 // -> (icmp pred i(M-N) (trunc %v iM to i(M-N)), (trunc (C>>N))
2466 // Transform the shl to a trunc if (trunc (C>>N)) has no loss and M-N.
2467 // This enables us to get rid of the shift in favor of a trunc that may be
2468 // free on the target. It has the additional benefit of comparing to a
2469 // smaller constant that may be more target-friendly.
2470 unsigned Amt = ShiftAmt->getLimitedValue(TypeBits - 1);
2471 if (Shl->hasOneUse() && Amt != 0 &&
2472 shouldChangeType(ShType->getScalarSizeInBits(), TypeBits - Amt)) {
2473 ICmpInst::Predicate CmpPred = Pred;
2474 APInt RHSC = C;
2475
2476 if (RHSC.countr_zero() < Amt && ICmpInst::isStrictPredicate(CmpPred)) {
2477 // Try the flipped strictness predicate.
2478 // e.g.:
2479 // icmp ult i64 (shl X, 32), 8589934593 ->
2480 // icmp ule i64 (shl X, 32), 8589934592 ->
2481 // icmp ule i32 (trunc X, i32), 2 ->
2482 // icmp ult i32 (trunc X, i32), 3
2483 if (auto FlippedStrictness = getFlippedStrictnessPredicateAndConstant(
2484 Pred, ConstantInt::get(ShType->getContext(), C))) {
2485 CmpPred = FlippedStrictness->first;
2486 RHSC = cast<ConstantInt>(FlippedStrictness->second)->getValue();
2487 }
2488 }
2489
2490 if (RHSC.countr_zero() >= Amt) {
2491 Type *TruncTy = ShType->getWithNewBitWidth(TypeBits - Amt);
2492 Constant *NewC =
2493 ConstantInt::get(TruncTy, RHSC.ashr(*ShiftAmt).trunc(TypeBits - Amt));
2494 return new ICmpInst(CmpPred,
2495 Builder.CreateTrunc(X, TruncTy, "", /*IsNUW=*/false,
2496 Shl->hasNoSignedWrap()),
2497 NewC);
2498 }
2499 }
2500
2501 return nullptr;
2502}
2503
2504/// Fold icmp ({al}shr X, Y), C.
2506 BinaryOperator *Shr,
2507 const APInt &C) {
2508 // An exact shr only shifts out zero bits, so:
2509 // icmp eq/ne (shr X, Y), 0 --> icmp eq/ne X, 0
2510 Value *X = Shr->getOperand(0);
2511 CmpInst::Predicate Pred = Cmp.getPredicate();
2512 if (Cmp.isEquality() && Shr->isExact() && C.isZero())
2513 return new ICmpInst(Pred, X, Cmp.getOperand(1));
2514
2515 bool IsAShr = Shr->getOpcode() == Instruction::AShr;
2516 const APInt *ShiftValC;
2517 if (match(X, m_APInt(ShiftValC))) {
2518 if (Cmp.isEquality())
2519 return foldICmpShrConstConst(Cmp, Shr->getOperand(1), C, *ShiftValC);
2520
2521 // (ShiftValC >> Y) >s -1 --> Y != 0 with ShiftValC < 0
2522 // (ShiftValC >> Y) <s 0 --> Y == 0 with ShiftValC < 0
2523 bool TrueIfSigned;
2524 if (!IsAShr && ShiftValC->isNegative() &&
2525 isSignBitCheck(Pred, C, TrueIfSigned))
2526 return new ICmpInst(TrueIfSigned ? CmpInst::ICMP_EQ : CmpInst::ICMP_NE,
2527 Shr->getOperand(1),
2528 ConstantInt::getNullValue(X->getType()));
2529
2530 // If the shifted constant is a power-of-2, test the shift amount directly:
2531 // (ShiftValC >> Y) >u C --> X <u (LZ(C) - LZ(ShiftValC))
2532 // (ShiftValC >> Y) <u C --> X >=u (LZ(C-1) - LZ(ShiftValC))
2533 if (!IsAShr && ShiftValC->isPowerOf2() &&
2534 (Pred == CmpInst::ICMP_UGT || Pred == CmpInst::ICMP_ULT)) {
2535 bool IsUGT = Pred == CmpInst::ICMP_UGT;
2536 assert(ShiftValC->uge(C) && "Expected simplify of compare");
2537 assert((IsUGT || !C.isZero()) && "Expected X u< 0 to simplify");
2538
2539 unsigned CmpLZ = IsUGT ? C.countl_zero() : (C - 1).countl_zero();
2540 unsigned ShiftLZ = ShiftValC->countl_zero();
2541 Constant *NewC = ConstantInt::get(Shr->getType(), CmpLZ - ShiftLZ);
2542 auto NewPred = IsUGT ? CmpInst::ICMP_ULT : CmpInst::ICMP_UGE;
2543 return new ICmpInst(NewPred, Shr->getOperand(1), NewC);
2544 }
2545 }
2546
2547 const APInt *ShiftAmtC;
2548 if (!match(Shr->getOperand(1), m_APInt(ShiftAmtC)))
2549 return nullptr;
2550
2551 // Check that the shift amount is in range. If not, don't perform undefined
2552 // shifts. When the shift is visited it will be simplified.
2553 unsigned TypeBits = C.getBitWidth();
2554 unsigned ShAmtVal = ShiftAmtC->getLimitedValue(TypeBits);
2555 if (ShAmtVal >= TypeBits || ShAmtVal == 0)
2556 return nullptr;
2557
2558 bool IsExact = Shr->isExact();
2559 Type *ShrTy = Shr->getType();
2560 // TODO: If we could guarantee that InstSimplify would handle all of the
2561 // constant-value-based preconditions in the folds below, then we could assert
2562 // those conditions rather than checking them. This is difficult because of
2563 // undef/poison (PR34838).
2564 if (IsAShr && Shr->hasOneUse()) {
2565 if (IsExact && (Pred == CmpInst::ICMP_SLT || Pred == CmpInst::ICMP_ULT) &&
2566 (C - 1).isPowerOf2() && C.countLeadingZeros() > ShAmtVal) {
2567 // When C - 1 is a power of two and the transform can be legally
2568 // performed, prefer this form so the produced constant is close to a
2569 // power of two.
2570 // icmp slt/ult (ashr exact X, ShAmtC), C
2571 // --> icmp slt/ult X, (C - 1) << ShAmtC) + 1
2572 APInt ShiftedC = (C - 1).shl(ShAmtVal) + 1;
2573 return new ICmpInst(Pred, X, ConstantInt::get(ShrTy, ShiftedC));
2574 }
2575 if (IsExact || Pred == CmpInst::ICMP_SLT || Pred == CmpInst::ICMP_ULT) {
2576 // When ShAmtC can be shifted losslessly:
2577 // icmp PRED (ashr exact X, ShAmtC), C --> icmp PRED X, (C << ShAmtC)
2578 // icmp slt/ult (ashr X, ShAmtC), C --> icmp slt/ult X, (C << ShAmtC)
2579 APInt ShiftedC = C.shl(ShAmtVal);
2580 if (ShiftedC.ashr(ShAmtVal) == C)
2581 return new ICmpInst(Pred, X, ConstantInt::get(ShrTy, ShiftedC));
2582 }
2583 if (Pred == CmpInst::ICMP_SGT) {
2584 // icmp sgt (ashr X, ShAmtC), C --> icmp sgt X, ((C + 1) << ShAmtC) - 1
2585 APInt ShiftedC = (C + 1).shl(ShAmtVal) - 1;
2586 if (!C.isMaxSignedValue() && !(C + 1).shl(ShAmtVal).isMinSignedValue() &&
2587 (ShiftedC + 1).ashr(ShAmtVal) == (C + 1))
2588 return new ICmpInst(Pred, X, ConstantInt::get(ShrTy, ShiftedC));
2589 }
2590 if (Pred == CmpInst::ICMP_UGT) {
2591 // icmp ugt (ashr X, ShAmtC), C --> icmp ugt X, ((C + 1) << ShAmtC) - 1
2592 // 'C + 1 << ShAmtC' can overflow as a signed number, so the 2nd
2593 // clause accounts for that pattern.
2594 APInt ShiftedC = (C + 1).shl(ShAmtVal) - 1;
2595 if ((ShiftedC + 1).ashr(ShAmtVal) == (C + 1) ||
2596 (C + 1).shl(ShAmtVal).isMinSignedValue())
2597 return new ICmpInst(Pred, X, ConstantInt::get(ShrTy, ShiftedC));
2598 }
2599
2600 // If the compare constant has significant bits above the lowest sign-bit,
2601 // then convert an unsigned cmp to a test of the sign-bit:
2602 // (ashr X, ShiftC) u> C --> X s< 0
2603 // (ashr X, ShiftC) u< C --> X s> -1
2604 if (C.getBitWidth() > 2 && C.getNumSignBits() <= ShAmtVal) {
2605 if (Pred == CmpInst::ICMP_UGT) {
2606 return new ICmpInst(CmpInst::ICMP_SLT, X,
2608 }
2609 if (Pred == CmpInst::ICMP_ULT) {
2610 return new ICmpInst(CmpInst::ICMP_SGT, X,
2612 }
2613 }
2614 } else if (!IsAShr) {
2615 if (Pred == CmpInst::ICMP_ULT || (Pred == CmpInst::ICMP_UGT && IsExact)) {
2616 // icmp ult (lshr X, ShAmtC), C --> icmp ult X, (C << ShAmtC)
2617 // icmp ugt (lshr exact X, ShAmtC), C --> icmp ugt X, (C << ShAmtC)
2618 APInt ShiftedC = C.shl(ShAmtVal);
2619 if (ShiftedC.lshr(ShAmtVal) == C)
2620 return new ICmpInst(Pred, X, ConstantInt::get(ShrTy, ShiftedC));
2621 }
2622 if (Pred == CmpInst::ICMP_UGT) {
2623 // icmp ugt (lshr X, ShAmtC), C --> icmp ugt X, ((C + 1) << ShAmtC) - 1
2624 APInt ShiftedC = (C + 1).shl(ShAmtVal) - 1;
2625 if ((ShiftedC + 1).lshr(ShAmtVal) == (C + 1))
2626 return new ICmpInst(Pred, X, ConstantInt::get(ShrTy, ShiftedC));
2627 }
2628 }
2629
2630 if (!Cmp.isEquality())
2631 return nullptr;
2632
2633 // Handle equality comparisons of shift-by-constant.
2634
2635 // If the comparison constant changes with the shift, the comparison cannot
2636 // succeed (bits of the comparison constant cannot match the shifted value).
2637 // This should be known by InstSimplify and already be folded to true/false.
2638 assert(((IsAShr && C.shl(ShAmtVal).ashr(ShAmtVal) == C) ||
2639 (!IsAShr && C.shl(ShAmtVal).lshr(ShAmtVal) == C)) &&
2640 "Expected icmp+shr simplify did not occur.");
2641
2642 // If the bits shifted out are known zero, compare the unshifted value:
2643 // (X & 4) >> 1 == 2 --> (X & 4) == 4.
2644 if (Shr->isExact())
2645 return new ICmpInst(Pred, X, ConstantInt::get(ShrTy, C << ShAmtVal));
2646
2647 if (Shr->hasOneUse()) {
2648 // Canonicalize the shift into an 'and':
2649 // icmp eq/ne (shr X, ShAmt), C --> icmp eq/ne (and X, HiMask), (C << ShAmt)
2650 APInt Val(APInt::getHighBitsSet(TypeBits, TypeBits - ShAmtVal));
2651 Constant *Mask = ConstantInt::get(ShrTy, Val);
2652 Value *And = Builder.CreateAnd(X, Mask, Shr->getName() + ".mask");
2653 return new ICmpInst(Pred, And, ConstantInt::get(ShrTy, C << ShAmtVal));
2654 }
2655
2656 return nullptr;
2657}
2658
2660 BinaryOperator *SRem,
2661 const APInt &C) {
2662 const ICmpInst::Predicate Pred = Cmp.getPredicate();
2663 if (Pred == ICmpInst::ICMP_UGT || Pred == ICmpInst::ICMP_ULT) {
2664 // Canonicalize unsigned predicates to signed:
2665 // (X s% DivisorC) u> C -> (X s% DivisorC) s< 0
2666 // iff (C s< 0 ? ~C : C) u>= abs(DivisorC)-1
2667 // (X s% DivisorC) u< C+1 -> (X s% DivisorC) s> -1
2668 // iff (C+1 s< 0 ? ~C : C) u>= abs(DivisorC)-1
2669
2670 const APInt *DivisorC;
2671 if (!match(SRem->getOperand(1), m_APInt(DivisorC)))
2672 return nullptr;
2673 if (DivisorC->isZero())
2674 return nullptr;
2675
2676 APInt NormalizedC = C;
2677 if (Pred == ICmpInst::ICMP_ULT) {
2678 assert(!NormalizedC.isZero() &&
2679 "ult X, 0 should have been simplified already.");
2680 --NormalizedC;
2681 }
2682 if (C.isNegative())
2683 NormalizedC.flipAllBits();
2684 if (!NormalizedC.uge(DivisorC->abs() - 1))
2685 return nullptr;
2686
2687 Type *Ty = SRem->getType();
2688 if (Pred == ICmpInst::ICMP_UGT)
2689 return new ICmpInst(ICmpInst::ICMP_SLT, SRem,
2691 return new ICmpInst(ICmpInst::ICMP_SGT, SRem,
2693 }
2694 // Match an 'is positive' or 'is negative' comparison of remainder by a
2695 // constant power-of-2 value:
2696 // (X % pow2C) sgt/slt 0
2697 if (Pred != ICmpInst::ICMP_SGT && Pred != ICmpInst::ICMP_SLT &&
2698 Pred != ICmpInst::ICMP_EQ && Pred != ICmpInst::ICMP_NE)
2699 return nullptr;
2700
2701 // TODO: The one-use check is standard because we do not typically want to
2702 // create longer instruction sequences, but this might be a special-case
2703 // because srem is not good for analysis or codegen.
2704 if (!SRem->hasOneUse())
2705 return nullptr;
2706
2707 const APInt *DivisorC;
2708 if (!match(SRem->getOperand(1), m_Power2(DivisorC)))
2709 return nullptr;
2710
2711 // For cmp_sgt/cmp_slt only zero valued C is handled.
2712 // For cmp_eq/cmp_ne only positive valued C is handled.
2713 if (((Pred == ICmpInst::ICMP_SGT || Pred == ICmpInst::ICMP_SLT) &&
2714 !C.isZero()) ||
2715 ((Pred == ICmpInst::ICMP_EQ || Pred == ICmpInst::ICMP_NE) &&
2716 !C.isStrictlyPositive()))
2717 return nullptr;
2718
2719 // Mask off the sign bit and the modulo bits (low-bits).
2720 Type *Ty = SRem->getType();
2721 APInt SignMask = APInt::getSignMask(Ty->getScalarSizeInBits());
2722 Constant *MaskC = ConstantInt::get(Ty, SignMask | (*DivisorC - 1));
2723 Value *And = Builder.CreateAnd(SRem->getOperand(0), MaskC);
2724
2725 if (Pred == ICmpInst::ICMP_EQ || Pred == ICmpInst::ICMP_NE)
2726 return new ICmpInst(Pred, And, ConstantInt::get(Ty, C));
2727
2728 // For 'is positive?' check that the sign-bit is clear and at least 1 masked
2729 // bit is set. Example:
2730 // (i8 X % 32) s> 0 --> (X & 159) s> 0
2731 if (Pred == ICmpInst::ICMP_SGT)
2733
2734 // For 'is negative?' check that the sign-bit is set and at least 1 masked
2735 // bit is set. Example:
2736 // (i16 X % 4) s< 0 --> (X & 32771) u> 32768
2737 return new ICmpInst(ICmpInst::ICMP_UGT, And, ConstantInt::get(Ty, SignMask));
2738}
2739
2740/// Fold icmp (udiv X, Y), C.
2742 BinaryOperator *UDiv,
2743 const APInt &C) {
2744 ICmpInst::Predicate Pred = Cmp.getPredicate();
2745 Value *X = UDiv->getOperand(0);
2746 Value *Y = UDiv->getOperand(1);
2747 Type *Ty = UDiv->getType();
2748
2749 const APInt *C2;
2750 if (!match(X, m_APInt(C2)))
2751 return nullptr;
2752
2753 assert(*C2 != 0 && "udiv 0, X should have been simplified already.");
2754
2755 // (icmp ugt (udiv C2, Y), C) -> (icmp ule Y, C2/(C+1))
2756 if (Pred == ICmpInst::ICMP_UGT) {
2757 assert(!C.isMaxValue() &&
2758 "icmp ugt X, UINT_MAX should have been simplified already.");
2759 return new ICmpInst(ICmpInst::ICMP_ULE, Y,
2760 ConstantInt::get(Ty, C2->udiv(C + 1)));
2761 }
2762
2763 // (icmp ult (udiv C2, Y), C) -> (icmp ugt Y, C2/C)
2764 if (Pred == ICmpInst::ICMP_ULT) {
2765 assert(C != 0 && "icmp ult X, 0 should have been simplified already.");
2766 return new ICmpInst(ICmpInst::ICMP_UGT, Y,
2767 ConstantInt::get(Ty, C2->udiv(C)));
2768 }
2769
2770 return nullptr;
2771}
2772
2773/// Fold icmp ({su}div X, Y), C.
2775 BinaryOperator *Div,
2776 const APInt &C) {
2777 ICmpInst::Predicate Pred = Cmp.getPredicate();
2778 Value *X = Div->getOperand(0);
2779 Value *Y = Div->getOperand(1);
2780 Type *Ty = Div->getType();
2781 bool DivIsSigned = Div->getOpcode() == Instruction::SDiv;
2782
2783 // If unsigned division and the compare constant is bigger than
2784 // UMAX/2 (negative), there's only one pair of values that satisfies an
2785 // equality check, so eliminate the division:
2786 // (X u/ Y) == C --> (X == C) && (Y == 1)
2787 // (X u/ Y) != C --> (X != C) || (Y != 1)
2788 // Similarly, if signed division and the compare constant is exactly SMIN:
2789 // (X s/ Y) == SMIN --> (X == SMIN) && (Y == 1)
2790 // (X s/ Y) != SMIN --> (X != SMIN) || (Y != 1)
2791 if (Cmp.isEquality() && Div->hasOneUse() && C.isSignBitSet() &&
2792 (!DivIsSigned || C.isMinSignedValue())) {
2793 Value *XBig = Builder.CreateICmp(Pred, X, ConstantInt::get(Ty, C));
2794 Value *YOne = Builder.CreateICmp(Pred, Y, ConstantInt::get(Ty, 1));
2795 auto Logic = Pred == ICmpInst::ICMP_EQ ? Instruction::And : Instruction::Or;
2796 return BinaryOperator::Create(Logic, XBig, YOne);
2797 }
2798
2799 // Fold: icmp pred ([us]div X, C2), C -> range test
2800 // Fold this div into the comparison, producing a range check.
2801 // Determine, based on the divide type, what the range is being
2802 // checked. If there is an overflow on the low or high side, remember
2803 // it, otherwise compute the range [low, hi) bounding the new value.
2804 // See: InsertRangeTest above for the kinds of replacements possible.
2805 const APInt *C2;
2806 if (!match(Y, m_APInt(C2)))
2807 return nullptr;
2808
2809 // FIXME: If the operand types don't match the type of the divide
2810 // then don't attempt this transform. The code below doesn't have the
2811 // logic to deal with a signed divide and an unsigned compare (and
2812 // vice versa). This is because (x /s C2) <s C produces different
2813 // results than (x /s C2) <u C or (x /u C2) <s C or even
2814 // (x /u C2) <u C. Simply casting the operands and result won't
2815 // work. :( The if statement below tests that condition and bails
2816 // if it finds it.
2817 // However, when the divisor is a positive constant and the dividend is
2818 // known non-negative, sdiv is equivalent to udiv, so we can lower
2819 // DivIsSigned and proceed through the unsigned path.
2820 if (!Cmp.isEquality() && DivIsSigned != Cmp.isSigned()) {
2821 if (!DivIsSigned || !C2->isStrictlyPositive() ||
2822 !isKnownNonNegative(X, SQ.getWithInstruction(&Cmp)))
2823 return nullptr;
2824 DivIsSigned = false;
2825 }
2826
2827 // The ProdOV computation fails on divide by 0 and divide by -1. Cases with
2828 // INT_MIN will also fail if the divisor is 1. Although folds of all these
2829 // division-by-constant cases should be present, we can not assert that they
2830 // have happened before we reach this icmp instruction.
2831 if (C2->isZero() || C2->isOne() || (DivIsSigned && C2->isAllOnes()))
2832 return nullptr;
2833
2834 // Compute Prod = C * C2. We are essentially solving an equation of
2835 // form X / C2 = C. We solve for X by multiplying C2 and C.
2836 // By solving for X, we can turn this into a range check instead of computing
2837 // a divide.
2838 APInt Prod = C * *C2;
2839
2840 // Determine if the product overflows by seeing if the product is not equal to
2841 // the divide. Make sure we do the same kind of divide as in the LHS
2842 // instruction that we're folding.
2843 bool ProdOV = (DivIsSigned ? Prod.sdiv(*C2) : Prod.udiv(*C2)) != C;
2844
2845 // If the division is known to be exact, then there is no remainder from the
2846 // divide, so the covered range size is unit, otherwise it is the divisor.
2847 APInt RangeSize = Div->isExact() ? APInt(C2->getBitWidth(), 1) : *C2;
2848
2849 // Figure out the interval that is being checked. For example, a comparison
2850 // like "X /u 5 == 0" is really checking that X is in the interval [0, 5).
2851 // Compute this interval based on the constants involved and the signedness of
2852 // the compare/divide. This computes a half-open interval, keeping track of
2853 // whether either value in the interval overflows. After analysis each
2854 // overflow variable is set to 0 if it's corresponding bound variable is valid
2855 // -1 if overflowed off the bottom end, or +1 if overflowed off the top end.
2856 int LoOverflow = 0, HiOverflow = 0;
2857 APInt LoBound, HiBound;
2858
2859 if (!DivIsSigned) { // udiv
2860 // e.g. X/5 op 3 --> [15, 20)
2861 LoBound = Prod;
2862 HiOverflow = LoOverflow = ProdOV;
2863 if (!HiOverflow) {
2864 // If this is not an exact divide, then many values in the range collapse
2865 // to the same result value.
2866 HiOverflow = addWithOverflow(HiBound, LoBound, RangeSize, false);
2867 }
2868 } else if (C2->isStrictlyPositive()) { // Divisor is > 0.
2869 if (C.isZero()) { // (X / pos) op 0
2870 // Can't overflow. e.g. X/2 op 0 --> [-1, 2)
2871 LoBound = -(RangeSize - 1);
2872 HiBound = RangeSize;
2873 } else if (C.isStrictlyPositive()) { // (X / pos) op pos
2874 LoBound = Prod; // e.g. X/5 op 3 --> [15, 20)
2875 HiOverflow = LoOverflow = ProdOV;
2876 if (!HiOverflow)
2877 HiOverflow = addWithOverflow(HiBound, Prod, RangeSize, true);
2878 } else { // (X / pos) op neg
2879 // e.g. X/5 op -3 --> [-15-4, -15+1) --> [-19, -14)
2880 HiBound = Prod + 1;
2881 LoOverflow = HiOverflow = ProdOV ? -1 : 0;
2882 if (!LoOverflow) {
2883 APInt DivNeg = -RangeSize;
2884 LoOverflow = addWithOverflow(LoBound, HiBound, DivNeg, true) ? -1 : 0;
2885 }
2886 }
2887 } else if (C2->isNegative()) { // Divisor is < 0.
2888 if (Div->isExact())
2889 RangeSize.negate();
2890 if (C.isZero()) { // (X / neg) op 0
2891 // e.g. X/-5 op 0 --> [-4, 5)
2892 LoBound = RangeSize + 1;
2893 HiBound = -RangeSize;
2894 if (HiBound == *C2) { // -INTMIN = INTMIN
2895 HiOverflow = 1; // [INTMIN+1, overflow)
2896 HiBound = APInt(); // e.g. X/INTMIN = 0 --> X > INTMIN
2897 }
2898 } else if (C.isStrictlyPositive()) { // (X / neg) op pos
2899 // e.g. X/-5 op 3 --> [-19, -14)
2900 HiBound = Prod + 1;
2901 HiOverflow = LoOverflow = ProdOV ? -1 : 0;
2902 if (!LoOverflow)
2903 LoOverflow =
2904 addWithOverflow(LoBound, HiBound, RangeSize, true) ? -1 : 0;
2905 } else { // (X / neg) op neg
2906 LoBound = Prod; // e.g. X/-5 op -3 --> [15, 20)
2907 LoOverflow = HiOverflow = ProdOV;
2908 if (!HiOverflow)
2909 HiOverflow = subWithOverflow(HiBound, Prod, RangeSize, true);
2910 }
2911
2912 // Dividing by a negative swaps the condition. LT <-> GT
2913 Pred = ICmpInst::getSwappedPredicate(Pred);
2914 }
2915
2916 switch (Pred) {
2917 default:
2918 llvm_unreachable("Unhandled icmp predicate!");
2919 case ICmpInst::ICMP_EQ:
2920 if (LoOverflow && HiOverflow)
2921 return replaceInstUsesWith(Cmp, Builder.getFalse());
2922 if (HiOverflow)
2923 return new ICmpInst(DivIsSigned ? ICmpInst::ICMP_SGE : ICmpInst::ICMP_UGE,
2924 X, ConstantInt::get(Ty, LoBound));
2925 if (LoOverflow)
2926 return new ICmpInst(DivIsSigned ? ICmpInst::ICMP_SLT : ICmpInst::ICMP_ULT,
2927 X, ConstantInt::get(Ty, HiBound));
2928 return replaceInstUsesWith(
2929 Cmp, insertRangeTest(X, LoBound, HiBound, DivIsSigned, true));
2930 case ICmpInst::ICMP_NE:
2931 if (LoOverflow && HiOverflow)
2932 return replaceInstUsesWith(Cmp, Builder.getTrue());
2933 if (HiOverflow)
2934 return new ICmpInst(DivIsSigned ? ICmpInst::ICMP_SLT : ICmpInst::ICMP_ULT,
2935 X, ConstantInt::get(Ty, LoBound));
2936 if (LoOverflow)
2937 return new ICmpInst(DivIsSigned ? ICmpInst::ICMP_SGE : ICmpInst::ICMP_UGE,
2938 X, ConstantInt::get(Ty, HiBound));
2939 return replaceInstUsesWith(
2940 Cmp, insertRangeTest(X, LoBound, HiBound, DivIsSigned, false));
2941 case ICmpInst::ICMP_ULT:
2942 case ICmpInst::ICMP_SLT:
2943 if (LoOverflow == +1) // Low bound is greater than input range.
2944 return replaceInstUsesWith(Cmp, Builder.getTrue());
2945 if (LoOverflow == -1) // Low bound is less than input range.
2946 return replaceInstUsesWith(Cmp, Builder.getFalse());
2947 return new ICmpInst(Pred, X, ConstantInt::get(Ty, LoBound));
2948 case ICmpInst::ICMP_UGT:
2949 case ICmpInst::ICMP_SGT:
2950 if (HiOverflow == +1) // High bound greater than input range.
2951 return replaceInstUsesWith(Cmp, Builder.getFalse());
2952 if (HiOverflow == -1) // High bound less than input range.
2953 return replaceInstUsesWith(Cmp, Builder.getTrue());
2954 if (Pred == ICmpInst::ICMP_UGT)
2955 return new ICmpInst(ICmpInst::ICMP_UGE, X, ConstantInt::get(Ty, HiBound));
2956 return new ICmpInst(ICmpInst::ICMP_SGE, X, ConstantInt::get(Ty, HiBound));
2957 }
2958
2959 return nullptr;
2960}
2961
2962/// Fold icmp (sub X, Y), C.
2965 const APInt &C) {
2966 Value *X = Sub->getOperand(0), *Y = Sub->getOperand(1);
2967 ICmpInst::Predicate Pred = Cmp.getPredicate();
2968 Type *Ty = Sub->getType();
2969
2970 // (SubC - Y) == C) --> Y == (SubC - C)
2971 // (SubC - Y) != C) --> Y != (SubC - C)
2972 Constant *SubC;
2973 if (Cmp.isEquality() && match(X, m_ImmConstant(SubC))) {
2974 return new ICmpInst(Pred, Y,
2975 ConstantExpr::getSub(SubC, ConstantInt::get(Ty, C)));
2976 }
2977
2978 // (icmp P (sub nuw|nsw C2, Y), C) -> (icmp swap(P) Y, C2-C)
2979 const APInt *C2;
2980 APInt SubResult;
2981 ICmpInst::Predicate SwappedPred = Cmp.getSwappedPredicate();
2982 bool HasNSW = Sub->hasNoSignedWrap();
2983 bool HasNUW = Sub->hasNoUnsignedWrap();
2984 if (match(X, m_APInt(C2)) &&
2985 ((Cmp.isUnsigned() && HasNUW) || (Cmp.isSigned() && HasNSW)) &&
2986 !subWithOverflow(SubResult, *C2, C, Cmp.isSigned()))
2987 return new ICmpInst(SwappedPred, Y, ConstantInt::get(Ty, SubResult));
2988
2989 // X - Y == 0 --> X == Y.
2990 // X - Y != 0 --> X != Y.
2991 // TODO: We allow this with multiple uses as long as the other uses are not
2992 // in phis. The phi use check is guarding against a codegen regression
2993 // for a loop test. If the backend could undo this (and possibly
2994 // subsequent transforms), we would not need this hack.
2995 if (Cmp.isEquality() && C.isZero() &&
2996 none_of((Sub->users()), [](const User *U) { return isa<PHINode>(U); }))
2997 return new ICmpInst(Pred, X, Y);
2998
2999 // The following transforms are only worth it if the only user of the subtract
3000 // is the icmp.
3001 // TODO: This is an artificial restriction for all of the transforms below
3002 // that only need a single replacement icmp. Can these use the phi test
3003 // like the transform above here?
3004 if (!Sub->hasOneUse())
3005 return nullptr;
3006
3007 if (Sub->hasNoSignedWrap()) {
3008 // (icmp sgt (sub nsw X, Y), -1) -> (icmp sge X, Y)
3009 if (Pred == ICmpInst::ICMP_SGT && C.isAllOnes())
3010 return new ICmpInst(ICmpInst::ICMP_SGE, X, Y);
3011
3012 // (icmp sgt (sub nsw X, Y), 0) -> (icmp sgt X, Y)
3013 if (Pred == ICmpInst::ICMP_SGT && C.isZero())
3014 return new ICmpInst(ICmpInst::ICMP_SGT, X, Y);
3015
3016 // (icmp slt (sub nsw X, Y), 0) -> (icmp slt X, Y)
3017 if (Pred == ICmpInst::ICMP_SLT && C.isZero())
3018 return new ICmpInst(ICmpInst::ICMP_SLT, X, Y);
3019
3020 // (icmp slt (sub nsw X, Y), 1) -> (icmp sle X, Y)
3021 if (Pred == ICmpInst::ICMP_SLT && C.isOne())
3022 return new ICmpInst(ICmpInst::ICMP_SLE, X, Y);
3023 }
3024
3025 if (!match(X, m_APInt(C2)))
3026 return nullptr;
3027
3028 // C2 - Y <u C -> (Y | (C - 1)) == C2
3029 // iff (C2 & (C - 1)) == C - 1 and C is a power of 2
3030 if (Pred == ICmpInst::ICMP_ULT && C.isPowerOf2() &&
3031 (*C2 & (C - 1)) == (C - 1))
3032 return new ICmpInst(ICmpInst::ICMP_EQ, Builder.CreateOr(Y, C - 1), X);
3033
3034 // C2 - Y >u C -> (Y | C) != C2
3035 // iff C2 & C == C and C + 1 is a power of 2
3036 if (Pred == ICmpInst::ICMP_UGT && (C + 1).isPowerOf2() && (*C2 & C) == C)
3037 return new ICmpInst(ICmpInst::ICMP_NE, Builder.CreateOr(Y, C), X);
3038
3039 // We have handled special cases that reduce.
3040 // Canonicalize any remaining sub to add as:
3041 // (C2 - Y) > C --> (Y + ~C2) < ~C
3042 Value *Add = Builder.CreateAdd(Y, ConstantInt::get(Ty, ~(*C2)), "notsub",
3043 HasNUW, HasNSW);
3044 return new ICmpInst(SwappedPred, Add, ConstantInt::get(Ty, ~C));
3045}
3046
3047static Value *createLogicFromTable(const std::bitset<4> &Table, Value *Op0,
3048 Value *Op1, IRBuilderBase &Builder,
3049 bool HasOneUse) {
3050 auto FoldConstant = [&](bool Val) {
3051 Constant *Res = Val ? Builder.getTrue() : Builder.getFalse();
3052 if (Op0->getType()->isVectorTy())
3054 cast<VectorType>(Op0->getType())->getElementCount(), Res);
3055 return Res;
3056 };
3057
3058 switch (Table.to_ulong()) {
3059 case 0: // 0 0 0 0
3060 return FoldConstant(false);
3061 case 1: // 0 0 0 1
3062 return HasOneUse ? Builder.CreateNot(Builder.CreateOr(Op0, Op1)) : nullptr;
3063 case 2: // 0 0 1 0
3064 return HasOneUse ? Builder.CreateAnd(Builder.CreateNot(Op0), Op1) : nullptr;
3065 case 3: // 0 0 1 1
3066 return Builder.CreateNot(Op0);
3067 case 4: // 0 1 0 0
3068 return HasOneUse ? Builder.CreateAnd(Op0, Builder.CreateNot(Op1)) : nullptr;
3069 case 5: // 0 1 0 1
3070 return Builder.CreateNot(Op1);
3071 case 6: // 0 1 1 0
3072 return Builder.CreateXor(Op0, Op1);
3073 case 7: // 0 1 1 1
3074 return HasOneUse ? Builder.CreateNot(Builder.CreateAnd(Op0, Op1)) : nullptr;
3075 case 8: // 1 0 0 0
3076 return Builder.CreateAnd(Op0, Op1);
3077 case 9: // 1 0 0 1
3078 return HasOneUse ? Builder.CreateNot(Builder.CreateXor(Op0, Op1)) : nullptr;
3079 case 10: // 1 0 1 0
3080 return Op1;
3081 case 11: // 1 0 1 1
3082 return HasOneUse ? Builder.CreateOr(Builder.CreateNot(Op0), Op1) : nullptr;
3083 case 12: // 1 1 0 0
3084 return Op0;
3085 case 13: // 1 1 0 1
3086 return HasOneUse ? Builder.CreateOr(Op0, Builder.CreateNot(Op1)) : nullptr;
3087 case 14: // 1 1 1 0
3088 return Builder.CreateOr(Op0, Op1);
3089 case 15: // 1 1 1 1
3090 return FoldConstant(true);
3091 default:
3092 llvm_unreachable("Invalid Operation");
3093 }
3094 return nullptr;
3095}
3096
3098 ICmpInst &Cmp, BinaryOperator *BO, const APInt &C) {
3099 Value *A, *B;
3100 Constant *C1, *C2, *C3, *C4;
3101 if (!match(BO->getOperand(0),
3103 !match(BO->getOperand(1),
3105 Cmp.getType() != A->getType() || Cmp.getType() != B->getType())
3106 return nullptr;
3107
3108 std::bitset<4> Table;
3109 auto ComputeTable = [&](bool First, bool Second) -> std::optional<bool> {
3110 Constant *L = First ? C1 : C2;
3111 Constant *R = Second ? C3 : C4;
3112 if (auto *Res = ConstantFoldBinaryOpOperands(BO->getOpcode(), L, R, DL)) {
3113 auto *Val = Res->getType()->isVectorTy() ? Res->getSplatValue() : Res;
3114 if (auto *CI = dyn_cast_or_null<ConstantInt>(Val))
3115 return ICmpInst::compare(CI->getValue(), C, Cmp.getPredicate());
3116 }
3117 return std::nullopt;
3118 };
3119
3120 for (unsigned I = 0; I < 4; ++I) {
3121 bool First = (I >> 1) & 1;
3122 bool Second = I & 1;
3123 if (auto Res = ComputeTable(First, Second))
3124 Table[I] = *Res;
3125 else
3126 return nullptr;
3127 }
3128
3129 // Synthesize optimal logic.
3130 if (auto *Cond = createLogicFromTable(Table, A, B, Builder, BO->hasOneUse()))
3131 return replaceInstUsesWith(Cmp, Cond);
3132 return nullptr;
3133}
3134
3135/// Fold icmp (add X, Y), C.
3138 const APInt &C) {
3139 Value *Y = Add->getOperand(1);
3140 Value *X = Add->getOperand(0);
3141 const CmpPredicate Pred = Cmp.getCmpPredicate();
3142
3143 // icmp ult (add nuw A, (lshr A, ShAmtC)), C --> icmp ult A, C
3144 // when C <= (1 << ShAmtC).
3145 const APInt *ShAmtC;
3146 Value *A;
3147 unsigned BitWidth = C.getBitWidth();
3148 if (Pred == ICmpInst::ICMP_ULT &&
3149 match(Add,
3150 m_c_NUWAdd(m_Value(A), m_LShr(m_Deferred(A), m_APInt(ShAmtC)))) &&
3151 ShAmtC->ult(BitWidth) &&
3152 C.ule(APInt::getOneBitSet(BitWidth, ShAmtC->getZExtValue())))
3153 return new ICmpInst(Pred, A, ConstantInt::get(A->getType(), C));
3154
3155 const APInt *C2;
3156 if (Cmp.isEquality() || !match(Y, m_APInt(C2)))
3157 return nullptr;
3158
3159 // Fold icmp pred (add X, C2), C.
3160 Type *Ty = Add->getType();
3161
3162 // If the add does not wrap, we can always adjust the compare by subtracting
3163 // the constants. Equality comparisons are handled elsewhere. SGE/SLE/UGE/ULE
3164 // have been canonicalized to SGT/SLT/UGT/ULT.
3165 if (Add->hasNoUnsignedWrap() &&
3166 (Pred == ICmpInst::ICMP_UGT || Pred == ICmpInst::ICMP_ULT)) {
3167 bool Overflow;
3168 APInt NewC = C.usub_ov(*C2, Overflow);
3169 // If there is overflow, the result must be true or false.
3170 if (!Overflow)
3171 // icmp Pred (add nsw X, C2), C --> icmp Pred X, (C - C2)
3172 return new ICmpInst(Pred, X, ConstantInt::get(Ty, NewC));
3173 }
3174
3175 CmpInst::Predicate ChosenPred = Pred.getPreferredSignedPredicate();
3176
3177 if (Add->hasNoSignedWrap() &&
3178 (ChosenPred == ICmpInst::ICMP_SGT || ChosenPred == ICmpInst::ICMP_SLT)) {
3179 bool Overflow;
3180 APInt NewC = C.ssub_ov(*C2, Overflow);
3181 if (!Overflow)
3182 // icmp samesign ugt/ult (add nsw X, C2), C
3183 // -> icmp sgt/slt X, (C - C2)
3184 return new ICmpInst(ChosenPred, X, ConstantInt::get(Ty, NewC));
3185 }
3186
3187 if (ICmpInst::isUnsigned(Pred) && Add->hasNoSignedWrap() &&
3188 C.isNonNegative() && (C - *C2).isNonNegative() &&
3189 computeConstantRange(X, /*ForSigned=*/true, SQ.getWithInstruction(&Cmp))
3190 .add(*C2)
3191 .isAllNonNegative())
3192 return new ICmpInst(ICmpInst::getSignedPredicate(Pred), X,
3193 ConstantInt::get(Ty, C - *C2));
3194
3195 auto CR = ConstantRange::makeExactICmpRegion(Pred, C).subtract(*C2);
3196 const APInt &Upper = CR.getUpper();
3197 const APInt &Lower = CR.getLower();
3198 if (Cmp.isSigned()) {
3199 if (Lower.isSignMask())
3200 return new ICmpInst(ICmpInst::ICMP_SLT, X, ConstantInt::get(Ty, Upper));
3201 if (Upper.isSignMask())
3202 return new ICmpInst(ICmpInst::ICMP_SGE, X, ConstantInt::get(Ty, Lower));
3203 } else {
3204 if (Lower.isMinValue())
3205 return new ICmpInst(ICmpInst::ICMP_ULT, X, ConstantInt::get(Ty, Upper));
3206 if (Upper.isMinValue())
3207 return new ICmpInst(ICmpInst::ICMP_UGE, X, ConstantInt::get(Ty, Lower));
3208 }
3209
3210 // This set of folds is intentionally placed after folds that use no-wrapping
3211 // flags because those folds are likely better for later analysis/codegen.
3212 const APInt SMax = APInt::getSignedMaxValue(Ty->getScalarSizeInBits());
3213 const APInt SMin = APInt::getSignedMinValue(Ty->getScalarSizeInBits());
3214
3215 // Fold compare with offset to opposite sign compare if it eliminates offset:
3216 // (X + C2) >u C --> X <s -C2 (if C == C2 + SMAX)
3217 if (Pred == CmpInst::ICMP_UGT && C == *C2 + SMax)
3218 return new ICmpInst(ICmpInst::ICMP_SLT, X, ConstantInt::get(Ty, -(*C2)));
3219
3220 // (X + C2) <u C --> X >s ~C2 (if C == C2 + SMIN)
3221 if (Pred == CmpInst::ICMP_ULT && C == *C2 + SMin)
3222 return new ICmpInst(ICmpInst::ICMP_SGT, X, ConstantInt::get(Ty, ~(*C2)));
3223
3224 // (X + C2) >s C --> X <u (SMAX - C) (if C == C2 - 1)
3225 if (Pred == CmpInst::ICMP_SGT && C == *C2 - 1)
3226 return new ICmpInst(ICmpInst::ICMP_ULT, X, ConstantInt::get(Ty, SMax - C));
3227
3228 // (X + C2) <s C --> X >u (C ^ SMAX) (if C == C2)
3229 if (Pred == CmpInst::ICMP_SLT && C == *C2)
3230 return new ICmpInst(ICmpInst::ICMP_UGT, X, ConstantInt::get(Ty, C ^ SMax));
3231
3232 // (X + -1) <u C --> X <=u C (if X is never null)
3233 if (Pred == CmpInst::ICMP_ULT && C2->isAllOnes()) {
3234 const SimplifyQuery Q = SQ.getWithInstruction(&Cmp);
3235 if (llvm::isKnownNonZero(X, Q))
3236 return new ICmpInst(ICmpInst::ICMP_ULE, X, ConstantInt::get(Ty, C));
3237 }
3238
3239 if (!Add->hasOneUse())
3240 return nullptr;
3241
3242 // X+C <u C2 -> (X & -C2) == C
3243 // iff C & (C2-1) == 0
3244 // C2 is a power of 2
3245 if (Pred == ICmpInst::ICMP_ULT && C.isPowerOf2() && (*C2 & (C - 1)) == 0)
3246 return new ICmpInst(ICmpInst::ICMP_EQ, Builder.CreateAnd(X, -C),
3248
3249 // X+C2 <u C -> (X & C) == 2C
3250 // iff C == -(C2)
3251 // C2 is a power of 2
3252 if (Pred == ICmpInst::ICMP_ULT && C2->isPowerOf2() && C == -*C2)
3253 return new ICmpInst(ICmpInst::ICMP_NE, Builder.CreateAnd(X, C),
3254 ConstantInt::get(Ty, C * 2));
3255
3256 // X+C >u C2 -> (X & ~C2) != C
3257 // iff C & C2 == 0
3258 // C2+1 is a power of 2
3259 if (Pred == ICmpInst::ICMP_UGT && (C + 1).isPowerOf2() && (*C2 & C) == 0)
3260 return new ICmpInst(ICmpInst::ICMP_NE, Builder.CreateAnd(X, ~C),
3262
3263 // The range test idiom can use either ult or ugt. Arbitrarily canonicalize
3264 // to the ult form.
3265 // X+C2 >u C -> X+(C2-C-1) <u ~C
3266 if (Pred == ICmpInst::ICMP_UGT)
3267 return new ICmpInst(ICmpInst::ICMP_ULT,
3268 Builder.CreateAdd(X, ConstantInt::get(Ty, *C2 - C - 1)),
3269 ConstantInt::get(Ty, ~C));
3270
3271 // zext(V) + C2 pred C -> V + C3 pred' C4
3272 Value *V;
3273 if (match(X, m_ZExt(m_Value(V)))) {
3274 Type *NewCmpTy = V->getType();
3275 unsigned NewCmpBW = NewCmpTy->getScalarSizeInBits();
3276 if (shouldChangeType(Ty, NewCmpTy)) {
3277 ConstantRange SrcCR = CR.truncate(NewCmpBW, TruncInst::NoUnsignedWrap);
3278 CmpInst::Predicate EquivPred;
3279 APInt EquivInt;
3280 APInt EquivOffset;
3281
3282 SrcCR.getEquivalentICmp(EquivPred, EquivInt, EquivOffset);
3283 return new ICmpInst(
3284 EquivPred,
3285 EquivOffset.isZero()
3286 ? V
3287 : Builder.CreateAdd(V, ConstantInt::get(NewCmpTy, EquivOffset)),
3288 ConstantInt::get(NewCmpTy, EquivInt));
3289 }
3290 }
3291
3292 return nullptr;
3293}
3294
3296 Value *&RHS, ConstantInt *&Less,
3297 ConstantInt *&Equal,
3298 ConstantInt *&Greater) {
3299 // TODO: Generalize this to work with other comparison idioms or ensure
3300 // they get canonicalized into this form.
3301
3302 // select i1 (a == b),
3303 // i32 Equal,
3304 // i32 (select i1 (a < b), i32 Less, i32 Greater)
3305 // where Equal, Less and Greater are placeholders for any three constants.
3306 CmpPredicate PredA;
3307 if (!match(SI->getCondition(), m_ICmp(PredA, m_Value(LHS), m_Value(RHS))) ||
3308 !ICmpInst::isEquality(PredA))
3309 return false;
3310 Value *EqualVal = SI->getTrueValue();
3311 Value *UnequalVal = SI->getFalseValue();
3312 // We still can get non-canonical predicate here, so canonicalize.
3313 if (PredA == ICmpInst::ICMP_NE)
3314 std::swap(EqualVal, UnequalVal);
3315 if (!match(EqualVal, m_ConstantInt(Equal)))
3316 return false;
3317 CmpPredicate PredB;
3318 Value *LHS2, *RHS2;
3319 if (!match(UnequalVal, m_Select(m_ICmp(PredB, m_Value(LHS2), m_Value(RHS2)),
3320 m_ConstantInt(Less), m_ConstantInt(Greater))))
3321 return false;
3322 // We can get predicate mismatch here, so canonicalize if possible:
3323 // First, ensure that 'LHS' match.
3324 if (LHS2 != LHS) {
3325 // x sgt y <--> y slt x
3326 std::swap(LHS2, RHS2);
3327 PredB = ICmpInst::getSwappedPredicate(PredB);
3328 }
3329 if (LHS2 != LHS)
3330 return false;
3331 // We also need to canonicalize 'RHS'.
3332 if (PredB == ICmpInst::ICMP_SGT && isa<Constant>(RHS2)) {
3333 // x sgt C-1 <--> x sge C <--> not(x slt C)
3334 auto FlippedStrictness =
3336 if (!FlippedStrictness)
3337 return false;
3338 assert(FlippedStrictness->first == ICmpInst::ICMP_SGE &&
3339 "basic correctness failure");
3340 RHS2 = FlippedStrictness->second;
3341 // And kind-of perform the result swap.
3342 std::swap(Less, Greater);
3343 PredB = ICmpInst::ICMP_SLT;
3344 }
3345 return PredB == ICmpInst::ICMP_SLT && RHS == RHS2;
3346}
3347
3350 ConstantInt *C) {
3351
3352 assert(C && "Cmp RHS should be a constant int!");
3353 // If we're testing a constant value against the result of a three way
3354 // comparison, the result can be expressed directly in terms of the
3355 // original values being compared. Note: We could possibly be more
3356 // aggressive here and remove the hasOneUse test. The original select is
3357 // really likely to simplify or sink when we remove a test of the result.
3358 Value *OrigLHS, *OrigRHS;
3359 ConstantInt *C1LessThan, *C2Equal, *C3GreaterThan;
3360 if (Cmp.hasOneUse() &&
3361 matchThreeWayIntCompare(Select, OrigLHS, OrigRHS, C1LessThan, C2Equal,
3362 C3GreaterThan)) {
3363 assert(C1LessThan && C2Equal && C3GreaterThan);
3364
3365 bool TrueWhenLessThan = ICmpInst::compare(
3366 C1LessThan->getValue(), C->getValue(), Cmp.getPredicate());
3367 bool TrueWhenEqual = ICmpInst::compare(C2Equal->getValue(), C->getValue(),
3368 Cmp.getPredicate());
3369 bool TrueWhenGreaterThan = ICmpInst::compare(
3370 C3GreaterThan->getValue(), C->getValue(), Cmp.getPredicate());
3371
3372 // This generates the new instruction that will replace the original Cmp
3373 // Instruction. Instead of enumerating the various combinations when
3374 // TrueWhenLessThan, TrueWhenEqual and TrueWhenGreaterThan are true versus
3375 // false, we rely on chaining of ORs and future passes of InstCombine to
3376 // simplify the OR further (i.e. a s< b || a == b becomes a s<= b).
3377
3378 // When none of the three constants satisfy the predicate for the RHS (C),
3379 // the entire original Cmp can be simplified to a false.
3380 Value *Cond = Builder.getFalse();
3381 if (TrueWhenLessThan)
3382 Cond = Builder.CreateOr(
3383 Cond, Builder.CreateICmp(ICmpInst::ICMP_SLT, OrigLHS, OrigRHS));
3384 if (TrueWhenEqual)
3385 Cond = Builder.CreateOr(
3386 Cond, Builder.CreateICmp(ICmpInst::ICMP_EQ, OrigLHS, OrigRHS));
3387 if (TrueWhenGreaterThan)
3388 Cond = Builder.CreateOr(
3389 Cond, Builder.CreateICmp(ICmpInst::ICMP_SGT, OrigLHS, OrigRHS));
3390
3391 return replaceInstUsesWith(Cmp, Cond);
3392 }
3393 return nullptr;
3394}
3395
3397 auto *Bitcast = dyn_cast<BitCastInst>(Cmp.getOperand(0));
3398 if (!Bitcast)
3399 return nullptr;
3400
3401 ICmpInst::Predicate Pred = Cmp.getPredicate();
3402 Value *Op1 = Cmp.getOperand(1);
3403 Value *BCSrcOp = Bitcast->getOperand(0);
3404 Type *SrcType = Bitcast->getSrcTy();
3405 Type *DstType = Bitcast->getType();
3406
3407 // Make sure the bitcast doesn't change between scalar and vector and
3408 // doesn't change the number of vector elements.
3409 if (SrcType->isVectorTy() == DstType->isVectorTy() &&
3410 SrcType->getScalarSizeInBits() == DstType->getScalarSizeInBits()) {
3411 // Zero-equality and sign-bit checks are preserved through sitofp + bitcast.
3412 Value *X;
3413 if (match(BCSrcOp, m_SIToFP(m_Value(X)))) {
3414 // icmp eq (bitcast (sitofp X)), 0 --> icmp eq X, 0
3415 // icmp ne (bitcast (sitofp X)), 0 --> icmp ne X, 0
3416 // icmp slt (bitcast (sitofp X)), 0 --> icmp slt X, 0
3417 // icmp sgt (bitcast (sitofp X)), 0 --> icmp sgt X, 0
3418 if ((Pred == ICmpInst::ICMP_EQ || Pred == ICmpInst::ICMP_SLT ||
3419 Pred == ICmpInst::ICMP_NE || Pred == ICmpInst::ICMP_SGT) &&
3420 match(Op1, m_Zero()))
3421 return new ICmpInst(Pred, X, ConstantInt::getNullValue(X->getType()));
3422
3423 // icmp slt (bitcast (sitofp X)), 1 --> icmp slt X, 1
3424 if (Pred == ICmpInst::ICMP_SLT && match(Op1, m_One()))
3425 return new ICmpInst(Pred, X, ConstantInt::get(X->getType(), 1));
3426
3427 // icmp sgt (bitcast (sitofp X)), -1 --> icmp sgt X, -1
3428 if (Pred == ICmpInst::ICMP_SGT && match(Op1, m_AllOnes()))
3429 return new ICmpInst(Pred, X,
3430 ConstantInt::getAllOnesValue(X->getType()));
3431 }
3432
3433 // Zero-equality checks are preserved through unsigned floating-point casts:
3434 // icmp eq (bitcast (uitofp X)), 0 --> icmp eq X, 0
3435 // icmp ne (bitcast (uitofp X)), 0 --> icmp ne X, 0
3436 if (match(BCSrcOp, m_UIToFP(m_Value(X))))
3437 if (Cmp.isEquality() && match(Op1, m_Zero()))
3438 return new ICmpInst(Pred, X, ConstantInt::getNullValue(X->getType()));
3439
3440 const APInt *C;
3441 bool TrueIfSigned;
3442 if (match(Op1, m_APInt(C)) && Bitcast->hasOneUse()) {
3443 // If this is a sign-bit test of a bitcast of a casted FP value, eliminate
3444 // the FP extend/truncate because that cast does not change the sign-bit.
3445 // This is true for all standard IEEE-754 types and the X86 80-bit type.
3446 // The sign-bit is always the most significant bit in those types.
3447 if (isSignBitCheck(Pred, *C, TrueIfSigned) &&
3448 (match(BCSrcOp, m_FPExt(m_Value(X))) ||
3449 match(BCSrcOp, m_FPTrunc(m_Value(X))))) {
3450 // (bitcast (fpext/fptrunc X)) to iX) < 0 --> (bitcast X to iY) < 0
3451 // (bitcast (fpext/fptrunc X)) to iX) > -1 --> (bitcast X to iY) > -1
3452 Type *XType = X->getType();
3453
3454 // We can't currently handle Power style floating point operations here.
3455 if (!(XType->isPPC_FP128Ty() || SrcType->isPPC_FP128Ty())) {
3456 Type *NewType = Builder.getIntNTy(XType->getScalarSizeInBits());
3457 if (auto *XVTy = dyn_cast<VectorType>(XType))
3458 NewType = VectorType::get(NewType, XVTy->getElementCount());
3459 Value *NewBitcast = Builder.CreateBitCast(X, NewType);
3460 if (TrueIfSigned)
3461 return new ICmpInst(ICmpInst::ICMP_SLT, NewBitcast,
3462 ConstantInt::getNullValue(NewType));
3463 else
3464 return new ICmpInst(ICmpInst::ICMP_SGT, NewBitcast,
3466 }
3467 }
3468
3469 // icmp eq/ne (bitcast X to int), special fp -> llvm.is.fpclass(X, class)
3470 Type *FPType = SrcType->getScalarType();
3471 if (!Cmp.getParent()->getParent()->hasFnAttribute(
3472 Attribute::NoImplicitFloat) &&
3473 Cmp.isEquality() && FPType->isIEEELikeFPTy()) {
3474 FPClassTest Mask = APFloat(FPType->getFltSemantics(), *C).classify();
3475 if (Mask & (fcInf | fcZero)) {
3476 if (Pred == ICmpInst::ICMP_NE)
3477 Mask = ~Mask;
3478 return replaceInstUsesWith(Cmp,
3479 Builder.createIsFPClass(BCSrcOp, Mask));
3480 }
3481 }
3482 }
3483 }
3484
3485 const APInt *C;
3486 if (!match(Cmp.getOperand(1), m_APInt(C)) || !DstType->isIntegerTy() ||
3487 !SrcType->isIntOrIntVectorTy())
3488 return nullptr;
3489
3490 // If this is checking if all elements of a vector compare are set or not,
3491 // invert the casted vector equality compare and test if all compare
3492 // elements are clear or not. Compare against zero is generally easier for
3493 // analysis and codegen.
3494 // icmp eq/ne (bitcast (not X) to iN), -1 --> icmp eq/ne (bitcast X to iN), 0
3495 // Example: are all elements equal? --> are zero elements not equal?
3496 // TODO: Try harder to reduce compare of 2 freely invertible operands?
3497 if (Cmp.isEquality() && C->isAllOnes() && Bitcast->hasOneUse()) {
3498 if (Value *NotBCSrcOp =
3499 getFreelyInverted(BCSrcOp, BCSrcOp->hasOneUse(), &Builder)) {
3500 Value *Cast = Builder.CreateBitCast(NotBCSrcOp, DstType);
3501 return new ICmpInst(Pred, Cast, ConstantInt::getNullValue(DstType));
3502 }
3503 }
3504
3505 // If this is checking if all elements of an extended vector are clear or not,
3506 // compare in a narrow type to eliminate the extend:
3507 // icmp eq/ne (bitcast (ext X) to iN), 0 --> icmp eq/ne (bitcast X to iM), 0
3508 Value *X;
3509 if (Cmp.isEquality() && C->isZero() && Bitcast->hasOneUse() &&
3510 match(BCSrcOp, m_ZExtOrSExt(m_Value(X)))) {
3511 if (auto *VecTy = dyn_cast<FixedVectorType>(X->getType())) {
3512 Type *NewType = Builder.getIntNTy(VecTy->getPrimitiveSizeInBits());
3513 Value *NewCast = Builder.CreateBitCast(X, NewType);
3514 return new ICmpInst(Pred, NewCast, ConstantInt::getNullValue(NewType));
3515 }
3516 }
3517
3518 // Folding: icmp <pred> iN X, C
3519 // where X = bitcast <M x iK> (shufflevector <M x iK> %vec, undef, SC)) to iN
3520 // and C is a splat of a K-bit pattern
3521 // and SC is a constant vector = <C', C', C', ..., C'>
3522 // Into:
3523 // %E = extractelement <M x iK> %vec, i32 C'
3524 // icmp <pred> iK %E, trunc(C)
3525 Value *Vec;
3526 ArrayRef<int> Mask;
3527 if (match(BCSrcOp, m_Shuffle(m_Value(Vec), m_Undef(), m_Mask(Mask)))) {
3528 // Check whether every element of Mask is the same constant
3529 if (all_equal(Mask)) {
3530 auto *VecTy = cast<VectorType>(SrcType);
3531 auto *EltTy = cast<IntegerType>(VecTy->getElementType());
3532 if (C->isSplat(EltTy->getBitWidth())) {
3533 // Fold the icmp based on the value of C
3534 // If C is M copies of an iK sized bit pattern,
3535 // then:
3536 // => %E = extractelement <N x iK> %vec, i32 Elem
3537 // icmp <pred> iK %SplatVal, <pattern>
3538 Value *Elem = Builder.getInt32(Mask[0]);
3539 Value *Extract = Builder.CreateExtractElement(Vec, Elem);
3540 Value *NewC = ConstantInt::get(EltTy, C->trunc(EltTy->getBitWidth()));
3541 return new ICmpInst(Pred, Extract, NewC);
3542 }
3543 }
3544 }
3545 return nullptr;
3546}
3547
3548/// Try to fold integer comparisons with a constant operand: icmp Pred X, C
3549/// where X is some kind of instruction.
3551 const APInt *C;
3552
3553 if (match(Cmp.getOperand(1), m_APInt(C))) {
3554 if (auto *BO = dyn_cast<BinaryOperator>(Cmp.getOperand(0)))
3555 if (Instruction *I = foldICmpBinOpWithConstant(Cmp, BO, *C))
3556 return I;
3557
3558 if (auto *SI = dyn_cast<SelectInst>(Cmp.getOperand(0)))
3559 // For now, we only support constant integers while folding the
3560 // ICMP(SELECT)) pattern. We can extend this to support vector of integers
3561 // similar to the cases handled by binary ops above.
3562 if (auto *ConstRHS = dyn_cast<ConstantInt>(Cmp.getOperand(1)))
3563 if (Instruction *I = foldICmpSelectConstant(Cmp, SI, ConstRHS))
3564 return I;
3565
3566 if (auto *TI = dyn_cast<TruncInst>(Cmp.getOperand(0)))
3567 if (Instruction *I = foldICmpTruncConstant(Cmp, TI, *C))
3568 return I;
3569
3570 if (auto *II = dyn_cast<IntrinsicInst>(Cmp.getOperand(0)))
3572 return I;
3573
3574 // (extractval ([s/u]subo X, Y), 0) == 0 --> X == Y
3575 // (extractval ([s/u]subo X, Y), 0) != 0 --> X != Y
3576 // TODO: This checks one-use, but that is not strictly necessary.
3577 Value *Cmp0 = Cmp.getOperand(0);
3578 Value *X, *Y;
3579 if (C->isZero() && Cmp.isEquality() && Cmp0->hasOneUse() &&
3580 (match(Cmp0,
3582 m_Value(X), m_Value(Y)))) ||
3583 match(Cmp0,
3585 m_Value(X), m_Value(Y))))))
3586 return new ICmpInst(Cmp.getPredicate(), X, Y);
3587 }
3588
3589 if (match(Cmp.getOperand(1), m_APIntAllowPoison(C)))
3591
3592 return nullptr;
3593}
3594
3595/// Fold an icmp equality instruction with binary operator LHS and constant RHS:
3596/// icmp eq/ne BO, C.
3598 ICmpInst &Cmp, BinaryOperator *BO, const APInt &C) {
3599 // TODO: Some of these folds could work with arbitrary constants, but this
3600 // function is limited to scalar and vector splat constants.
3601 if (!Cmp.isEquality())
3602 return nullptr;
3603
3604 ICmpInst::Predicate Pred = Cmp.getPredicate();
3605 bool isICMP_NE = Pred == ICmpInst::ICMP_NE;
3606 Constant *RHS = cast<Constant>(Cmp.getOperand(1));
3607 Value *BOp0 = BO->getOperand(0), *BOp1 = BO->getOperand(1);
3608
3609 switch (BO->getOpcode()) {
3610 case Instruction::SRem:
3611 // If we have a signed (X % (2^c)) == 0, turn it into an unsigned one.
3612 if (C.isZero() && BO->hasOneUse()) {
3613 const APInt *BOC;
3614 if (match(BOp1, m_APInt(BOC)) && BOC->sgt(1) && BOC->isPowerOf2()) {
3615 Value *NewRem = Builder.CreateURem(BOp0, BOp1, BO->getName());
3616 return new ICmpInst(Pred, NewRem,
3618 }
3619 }
3620 break;
3621 case Instruction::Add: {
3622 // (A + C2) == C --> A == (C - C2)
3623 // (A + C2) != C --> A != (C - C2)
3624 // TODO: Remove the one-use limitation? See discussion in D58633.
3625 if (Constant *C2 = dyn_cast<Constant>(BOp1)) {
3626 if (BO->hasOneUse())
3627 return new ICmpInst(Pred, BOp0, ConstantExpr::getSub(RHS, C2));
3628 } else if (C.isZero()) {
3629 // Replace ((add A, B) != 0) with (A != -B) if A or B is
3630 // efficiently invertible, or if the add has just this one use.
3631 if (Value *NegVal = dyn_castNegVal(BOp1))
3632 return new ICmpInst(Pred, BOp0, NegVal);
3633 if (Value *NegVal = dyn_castNegVal(BOp0))
3634 return new ICmpInst(Pred, NegVal, BOp1);
3635 if (BO->hasOneUse()) {
3636 // (add nuw A, B) != 0 -> (or A, B) != 0
3637 if (match(BO, m_NUWAdd(m_Value(), m_Value()))) {
3638 Value *Or = Builder.CreateOr(BOp0, BOp1);
3639 return new ICmpInst(Pred, Or, Constant::getNullValue(BO->getType()));
3640 }
3641 Value *Neg = Builder.CreateNeg(BOp1);
3642 Neg->takeName(BO);
3643 return new ICmpInst(Pred, BOp0, Neg);
3644 }
3645 }
3646 break;
3647 }
3648 case Instruction::Xor:
3649 if (Constant *BOC = dyn_cast<Constant>(BOp1)) {
3650 // For the xor case, we can xor two constants together, eliminating
3651 // the explicit xor.
3652 return new ICmpInst(Pred, BOp0, ConstantExpr::getXor(RHS, BOC));
3653 } else if (C.isZero()) {
3654 // Replace ((xor A, B) != 0) with (A != B)
3655 return new ICmpInst(Pred, BOp0, BOp1);
3656 }
3657 break;
3658 case Instruction::Or: {
3659 const APInt *BOC;
3660 if (match(BOp1, m_APInt(BOC)) && BO->hasOneUse() && RHS->isAllOnesValue()) {
3661 // Comparing if all bits outside of a constant mask are set?
3662 // Replace (X | C) == -1 with (X & ~C) == ~C.
3663 // This removes the -1 constant.
3665 Value *And = Builder.CreateAnd(BOp0, NotBOC);
3666 return new ICmpInst(Pred, And, NotBOC);
3667 }
3668 // (icmp eq (or (select cond, 0, NonZero), Other), 0)
3669 // -> (and cond, (icmp eq Other, 0))
3670 // (icmp ne (or (select cond, NonZero, 0), Other), 0)
3671 // -> (or cond, (icmp ne Other, 0))
3672 Value *Cond, *TV, *FV, *Other, *Sel;
3673 if (C.isZero() &&
3674 match(BO,
3677 m_Value(FV))),
3678 m_Value(Other)))) &&
3679 Cond->getType() == Cmp.getType()) {
3680 const SimplifyQuery Q = SQ.getWithInstruction(&Cmp);
3681 // Easy case is if eq/ne matches whether 0 is trueval/falseval.
3682 if (Pred == ICmpInst::ICMP_EQ
3683 ? (match(TV, m_Zero()) && isKnownNonZero(FV, Q))
3684 : (match(FV, m_Zero()) && isKnownNonZero(TV, Q))) {
3685 Value *Cmp = Builder.CreateICmp(
3686 Pred, Other, Constant::getNullValue(Other->getType()));
3688 Pred == ICmpInst::ICMP_EQ ? Instruction::And : Instruction::Or, Cmp,
3689 Cond);
3690 }
3691 // Harder case is if eq/ne matches whether 0 is falseval/trueval. In this
3692 // case we need to invert the select condition so we need to be careful to
3693 // avoid creating extra instructions.
3694 // (icmp ne (or (select cond, 0, NonZero), Other), 0)
3695 // -> (or (not cond), (icmp ne Other, 0))
3696 // (icmp eq (or (select cond, NonZero, 0), Other), 0)
3697 // -> (and (not cond), (icmp eq Other, 0))
3698 //
3699 // Only do this if the inner select has one use, in which case we are
3700 // replacing `select` with `(not cond)`. Otherwise, we will create more
3701 // uses. NB: Trying to freely invert cond doesn't make sense here, as if
3702 // cond was freely invertable, the select arms would have been inverted.
3703 if (Sel->hasOneUse() &&
3704 (Pred == ICmpInst::ICMP_EQ
3705 ? (match(FV, m_Zero()) && isKnownNonZero(TV, Q))
3706 : (match(TV, m_Zero()) && isKnownNonZero(FV, Q)))) {
3707 Value *NotCond = Builder.CreateNot(Cond);
3708 Value *Cmp = Builder.CreateICmp(
3709 Pred, Other, Constant::getNullValue(Other->getType()));
3711 Pred == ICmpInst::ICMP_EQ ? Instruction::And : Instruction::Or, Cmp,
3712 NotCond);
3713 }
3714 }
3715 break;
3716 }
3717 case Instruction::UDiv:
3718 case Instruction::SDiv:
3719 if (BO->isExact()) {
3720 // div exact X, Y eq/ne 0 -> X eq/ne 0
3721 // div exact X, Y eq/ne 1 -> X eq/ne Y
3722 // div exact X, Y eq/ne C ->
3723 // if Y * C never-overflow && OneUse:
3724 // -> Y * C eq/ne X
3725 if (C.isZero())
3726 return new ICmpInst(Pred, BOp0, Constant::getNullValue(BO->getType()));
3727 else if (C.isOne())
3728 return new ICmpInst(Pred, BOp0, BOp1);
3729 else if (BO->hasOneUse()) {
3731 Instruction::Mul, BO->getOpcode() == Instruction::SDiv, BOp1,
3732 Cmp.getOperand(1), BO);
3734 Value *YC =
3735 Builder.CreateMul(BOp1, ConstantInt::get(BO->getType(), C));
3736 return new ICmpInst(Pred, YC, BOp0);
3737 }
3738 }
3739 }
3740 if (BO->getOpcode() == Instruction::UDiv && C.isZero()) {
3741 // (icmp eq/ne (udiv A, B), 0) -> (icmp ugt/ule i32 B, A)
3742 auto NewPred = isICMP_NE ? ICmpInst::ICMP_ULE : ICmpInst::ICMP_UGT;
3743 return new ICmpInst(NewPred, BOp1, BOp0);
3744 }
3745 break;
3746 default:
3747 break;
3748 }
3749 return nullptr;
3750}
3751
3753 const APInt &CRhs,
3754 InstCombiner::BuilderTy &Builder,
3755 const SimplifyQuery &Q) {
3756 assert(CtpopLhs->getIntrinsicID() == Intrinsic::ctpop &&
3757 "Non-ctpop intrin in ctpop fold");
3758 if (!CtpopLhs->hasOneUse())
3759 return nullptr;
3760
3761 // Power of 2 test:
3762 // isPow2OrZero : ctpop(X) u< 2
3763 // isPow2 : ctpop(X) == 1
3764 // NotPow2OrZero: ctpop(X) u> 1
3765 // NotPow2 : ctpop(X) != 1
3766 // If we know any bit of X can be folded to:
3767 // IsPow2 : X & (~Bit) == 0
3768 // NotPow2 : X & (~Bit) != 0
3769 const ICmpInst::Predicate Pred = I.getPredicate();
3770 if (((I.isEquality() || Pred == ICmpInst::ICMP_UGT) && CRhs == 1) ||
3771 (Pred == ICmpInst::ICMP_ULT && CRhs == 2)) {
3772 Value *Op = CtpopLhs->getArgOperand(0);
3773 KnownBits OpKnown = computeKnownBits(Op, Q.DL, Q.AC, Q.CxtI, Q.DT);
3774 // No need to check for count > 1, that should be already constant folded.
3775 if (OpKnown.countMinPopulation() == 1) {
3776 Value *And = Builder.CreateAnd(
3777 Op, Constant::getIntegerValue(Op->getType(), ~(OpKnown.One)));
3778 return new ICmpInst(
3779 (Pred == ICmpInst::ICMP_EQ || Pred == ICmpInst::ICMP_ULT)
3782 And, Constant::getNullValue(Op->getType()));
3783 }
3784 }
3785
3786 return nullptr;
3787}
3788
3789/// Fold an equality icmp with LLVM intrinsic and constant operand.
3791 ICmpInst &Cmp, IntrinsicInst *II, const APInt &C) {
3792 Type *Ty = II->getType();
3793 unsigned BitWidth = C.getBitWidth();
3794 const ICmpInst::Predicate Pred = Cmp.getPredicate();
3795
3796 switch (II->getIntrinsicID()) {
3797 case Intrinsic::abs:
3798 // abs(A) == 0 -> A == 0
3799 // abs(A) == INT_MIN -> A == INT_MIN
3800 if (C.isZero() || C.isMinSignedValue())
3801 return new ICmpInst(Pred, II->getArgOperand(0), ConstantInt::get(Ty, C));
3802 break;
3803
3804 case Intrinsic::bswap:
3805 // bswap(A) == C -> A == bswap(C)
3806 return new ICmpInst(Pred, II->getArgOperand(0),
3807 ConstantInt::get(Ty, C.byteSwap()));
3808
3809 case Intrinsic::bitreverse:
3810 // bitreverse(A) == C -> A == bitreverse(C)
3811 return new ICmpInst(Pred, II->getArgOperand(0),
3812 ConstantInt::get(Ty, C.reverseBits()));
3813
3814 case Intrinsic::ctlz:
3815 case Intrinsic::cttz: {
3816 // ctz(A) == bitwidth(A) -> A == 0 and likewise for !=
3817 if (C == BitWidth)
3818 return new ICmpInst(Pred, II->getArgOperand(0),
3820
3821 // ctz(A) == C -> A & Mask1 == Mask2, where Mask2 only has bit C set
3822 // and Mask1 has bits 0..C+1 set. Similar for ctl, but for high bits.
3823 // Limit to one use to ensure we don't increase instruction count.
3824 unsigned Num = C.getLimitedValue(BitWidth);
3825 if (Num != BitWidth && II->hasOneUse()) {
3826 bool IsTrailing = II->getIntrinsicID() == Intrinsic::cttz;
3827 APInt Mask1 = IsTrailing ? APInt::getLowBitsSet(BitWidth, Num + 1)
3828 : APInt::getHighBitsSet(BitWidth, Num + 1);
3829 APInt Mask2 = IsTrailing
3832 return new ICmpInst(Pred, Builder.CreateAnd(II->getArgOperand(0), Mask1),
3833 ConstantInt::get(Ty, Mask2));
3834 }
3835 break;
3836 }
3837
3838 case Intrinsic::ctpop: {
3839 // popcount(A) == 0 -> A == 0 and likewise for !=
3840 // popcount(A) == bitwidth(A) -> A == -1 and likewise for !=
3841 bool IsZero = C.isZero();
3842 if (IsZero || C == BitWidth)
3843 return new ICmpInst(Pred, II->getArgOperand(0),
3844 IsZero ? Constant::getNullValue(Ty)
3846
3847 break;
3848 }
3849
3850 case Intrinsic::fshl:
3851 case Intrinsic::fshr:
3852 if (II->getArgOperand(0) == II->getArgOperand(1)) {
3853 const APInt *RotAmtC;
3854 // ror(X, RotAmtC) == C --> X == rol(C, RotAmtC)
3855 // rol(X, RotAmtC) == C --> X == ror(C, RotAmtC)
3856 if (match(II->getArgOperand(2), m_APInt(RotAmtC)))
3857 return new ICmpInst(Pred, II->getArgOperand(0),
3858 II->getIntrinsicID() == Intrinsic::fshl
3859 ? ConstantInt::get(Ty, C.rotr(*RotAmtC))
3860 : ConstantInt::get(Ty, C.rotl(*RotAmtC)));
3861 }
3862 break;
3863
3864 case Intrinsic::umax:
3865 case Intrinsic::uadd_sat: {
3866 // uadd.sat(a, b) == 0 -> (a | b) == 0
3867 // umax(a, b) == 0 -> (a | b) == 0
3868 if (C.isZero() && II->hasOneUse()) {
3869 Value *Or = Builder.CreateOr(II->getArgOperand(0), II->getArgOperand(1));
3870 return new ICmpInst(Pred, Or, Constant::getNullValue(Ty));
3871 }
3872 break;
3873 }
3874
3875 case Intrinsic::ssub_sat:
3876 // ssub.sat(a, b) == 0 -> a == b
3877 //
3878 // Note this doesn't work for ssub.sat.i1 because ssub.sat.i1 0, -1 = 0
3879 // (because 1 saturates to 0). Just skip the optimization for i1.
3880 if (C.isZero() && II->getType()->getScalarSizeInBits() > 1)
3881 return new ICmpInst(Pred, II->getArgOperand(0), II->getArgOperand(1));
3882 break;
3883 case Intrinsic::usub_sat: {
3884 // usub.sat(a, b) == 0 -> a <= b
3885 if (C.isZero()) {
3886 ICmpInst::Predicate NewPred =
3888 return new ICmpInst(NewPred, II->getArgOperand(0), II->getArgOperand(1));
3889 }
3890 break;
3891 }
3892 default:
3893 break;
3894 }
3895
3896 return nullptr;
3897}
3898
3899/// Fold an icmp with LLVM intrinsics
3900static Instruction *
3902 InstCombiner::BuilderTy &Builder) {
3903 assert(Cmp.isEquality());
3904
3905 ICmpInst::Predicate Pred = Cmp.getPredicate();
3906 Value *Op0 = Cmp.getOperand(0);
3907 Value *Op1 = Cmp.getOperand(1);
3908 const auto *IIOp0 = dyn_cast<IntrinsicInst>(Op0);
3909 const auto *IIOp1 = dyn_cast<IntrinsicInst>(Op1);
3910 if (!IIOp0 || !IIOp1 || IIOp0->getIntrinsicID() != IIOp1->getIntrinsicID())
3911 return nullptr;
3912
3913 switch (IIOp0->getIntrinsicID()) {
3914 case Intrinsic::bswap:
3915 case Intrinsic::bitreverse:
3916 // If both operands are byte-swapped or bit-reversed, just compare the
3917 // original values.
3918 return new ICmpInst(Pred, IIOp0->getOperand(0), IIOp1->getOperand(0));
3919 case Intrinsic::fshl:
3920 case Intrinsic::fshr: {
3921 // If both operands are rotated by same amount, just compare the
3922 // original values.
3923 if (IIOp0->getOperand(0) != IIOp0->getOperand(1))
3924 break;
3925 if (IIOp1->getOperand(0) != IIOp1->getOperand(1))
3926 break;
3927 if (IIOp0->getOperand(2) == IIOp1->getOperand(2))
3928 return new ICmpInst(Pred, IIOp0->getOperand(0), IIOp1->getOperand(0));
3929
3930 // rotate(X, AmtX) == rotate(Y, AmtY)
3931 // -> rotate(X, AmtX - AmtY) == Y
3932 // Do this if either both rotates have one use or if only one has one use
3933 // and AmtX/AmtY are constants.
3934 unsigned OneUses = IIOp0->hasOneUse() + IIOp1->hasOneUse();
3935 if (OneUses == 2 ||
3936 (OneUses == 1 && match(IIOp0->getOperand(2), m_ImmConstant()) &&
3937 match(IIOp1->getOperand(2), m_ImmConstant()))) {
3938 Value *SubAmt =
3939 Builder.CreateSub(IIOp0->getOperand(2), IIOp1->getOperand(2));
3940 Value *CombinedRotate = Builder.CreateIntrinsic(
3941 Op0->getType(), IIOp0->getIntrinsicID(),
3942 {IIOp0->getOperand(0), IIOp0->getOperand(0), SubAmt});
3943 return new ICmpInst(Pred, IIOp1->getOperand(0), CombinedRotate);
3944 }
3945 } break;
3946 default:
3947 break;
3948 }
3949
3950 return nullptr;
3951}
3952
3953/// Try to fold integer comparisons with a constant operand: icmp Pred X, C
3954/// where X is some kind of instruction and C is AllowPoison.
3955/// TODO: Move more folds which allow poison to this function.
3958 const APInt &C) {
3959 const ICmpInst::Predicate Pred = Cmp.getPredicate();
3960 if (auto *II = dyn_cast<IntrinsicInst>(Cmp.getOperand(0))) {
3961 switch (II->getIntrinsicID()) {
3962 default:
3963 break;
3964 case Intrinsic::fshl:
3965 case Intrinsic::fshr:
3966 if (Cmp.isEquality() && II->getArgOperand(0) == II->getArgOperand(1)) {
3967 // (rot X, ?) == 0/-1 --> X == 0/-1
3968 if (C.isZero() || C.isAllOnes())
3969 return new ICmpInst(Pred, II->getArgOperand(0), Cmp.getOperand(1));
3970 }
3971 break;
3972 }
3973 }
3974
3975 return nullptr;
3976}
3977
3978/// Fold an icmp with BinaryOp and constant operand: icmp Pred BO, C.
3980 BinaryOperator *BO,
3981 const APInt &C) {
3982 switch (BO->getOpcode()) {
3983 case Instruction::Xor:
3984 if (Instruction *I = foldICmpXorConstant(Cmp, BO, C))
3985 return I;
3986 break;
3987 case Instruction::And:
3988 if (Instruction *I = foldICmpAndConstant(Cmp, BO, C))
3989 return I;
3990 break;
3991 case Instruction::Or:
3992 if (Instruction *I = foldICmpOrConstant(Cmp, BO, C))
3993 return I;
3994 break;
3995 case Instruction::Mul:
3996 if (Instruction *I = foldICmpMulConstant(Cmp, BO, C))
3997 return I;
3998 break;
3999 case Instruction::Shl:
4000 if (Instruction *I = foldICmpShlConstant(Cmp, BO, C))
4001 return I;
4002 break;
4003 case Instruction::LShr:
4004 case Instruction::AShr:
4005 if (Instruction *I = foldICmpShrConstant(Cmp, BO, C))
4006 return I;
4007 break;
4008 case Instruction::SRem:
4009 if (Instruction *I = foldICmpSRemConstant(Cmp, BO, C))
4010 return I;
4011 break;
4012 case Instruction::UDiv:
4013 if (Instruction *I = foldICmpUDivConstant(Cmp, BO, C))
4014 return I;
4015 [[fallthrough]];
4016 case Instruction::SDiv:
4017 if (Instruction *I = foldICmpDivConstant(Cmp, BO, C))
4018 return I;
4019 break;
4020 case Instruction::Sub:
4021 if (Instruction *I = foldICmpSubConstant(Cmp, BO, C))
4022 return I;
4023 break;
4024 case Instruction::Add:
4025 if (Instruction *I = foldICmpAddConstant(Cmp, BO, C))
4026 return I;
4027 break;
4028 default:
4029 break;
4030 }
4031
4032 // TODO: These folds could be refactored to be part of the above calls.
4034 return I;
4035
4036 // Fall back to handling `icmp pred (select A ? C1 : C2) binop (select B ? C3
4037 // : C4), C5` pattern, by computing a truth table of the four constant
4038 // variants.
4040}
4041
4042static Instruction *
4044 const APInt &C,
4045 InstCombiner::BuilderTy &Builder) {
4046 // This transform may end up producing more than one instruction for the
4047 // intrinsic, so limit it to one user of the intrinsic.
4048 if (!II->hasOneUse())
4049 return nullptr;
4050
4051 // Let Y = [add/sub]_sat(X, C) pred C2
4052 // SatVal = The saturating value for the operation
4053 // WillWrap = Whether or not the operation will underflow / overflow
4054 // => Y = (WillWrap ? SatVal : (X binop C)) pred C2
4055 // => Y = WillWrap ? (SatVal pred C2) : ((X binop C) pred C2)
4056 //
4057 // When (SatVal pred C2) is true, then
4058 // Y = WillWrap ? true : ((X binop C) pred C2)
4059 // => Y = WillWrap || ((X binop C) pred C2)
4060 // else
4061 // Y = WillWrap ? false : ((X binop C) pred C2)
4062 // => Y = !WillWrap ? ((X binop C) pred C2) : false
4063 // => Y = !WillWrap && ((X binop C) pred C2)
4064 Value *Op0 = II->getOperand(0);
4065 Value *Op1 = II->getOperand(1);
4066
4067 const APInt *COp1;
4068 // This transform only works when the intrinsic has an integral constant or
4069 // splat vector as the second operand.
4070 if (!match(Op1, m_APInt(COp1)))
4071 return nullptr;
4072
4073 APInt SatVal;
4074 switch (II->getIntrinsicID()) {
4075 default:
4077 "This function only works with usub_sat and uadd_sat for now!");
4078 case Intrinsic::uadd_sat:
4079 SatVal = APInt::getAllOnes(C.getBitWidth());
4080 break;
4081 case Intrinsic::usub_sat:
4082 SatVal = APInt::getZero(C.getBitWidth());
4083 break;
4084 }
4085
4086 // Check (SatVal pred C2)
4087 bool SatValCheck = ICmpInst::compare(SatVal, C, Pred);
4088
4089 // !WillWrap.
4091 II->getBinaryOp(), *COp1, II->getNoWrapKind());
4092
4093 // WillWrap.
4094 if (SatValCheck)
4095 C1 = C1.inverse();
4096
4098 if (II->getBinaryOp() == Instruction::Add)
4099 C2 = C2.sub(*COp1);
4100 else
4101 C2 = C2.add(*COp1);
4102
4103 Instruction::BinaryOps CombiningOp =
4104 SatValCheck ? Instruction::BinaryOps::Or : Instruction::BinaryOps::And;
4105
4106 std::optional<ConstantRange> Combination;
4107 if (CombiningOp == Instruction::BinaryOps::Or)
4108 Combination = C1.exactUnionWith(C2);
4109 else /* CombiningOp == Instruction::BinaryOps::And */
4110 Combination = C1.exactIntersectWith(C2);
4111
4112 if (!Combination)
4113 return nullptr;
4114
4115 CmpInst::Predicate EquivPred;
4116 APInt EquivInt;
4117 APInt EquivOffset;
4118
4119 Combination->getEquivalentICmp(EquivPred, EquivInt, EquivOffset);
4120
4121 return new ICmpInst(
4122 EquivPred,
4123 Builder.CreateAdd(Op0, ConstantInt::get(Op1->getType(), EquivOffset)),
4124 ConstantInt::get(Op1->getType(), EquivInt));
4125}
4126
4127static Instruction *
4129 const APInt &C,
4130 InstCombiner::BuilderTy &Builder) {
4131 std::optional<ICmpInst::Predicate> NewPredicate = std::nullopt;
4132 switch (Pred) {
4133 case ICmpInst::ICMP_EQ:
4134 case ICmpInst::ICMP_NE:
4135 if (C.isZero())
4136 NewPredicate = Pred;
4137 else if (C.isOne())
4138 NewPredicate =
4140 else if (C.isAllOnes())
4141 NewPredicate =
4143 break;
4144
4145 case ICmpInst::ICMP_SGT:
4146 if (C.isAllOnes())
4147 NewPredicate = ICmpInst::ICMP_UGE;
4148 else if (C.isZero())
4149 NewPredicate = ICmpInst::ICMP_UGT;
4150 break;
4151
4152 case ICmpInst::ICMP_SLT:
4153 if (C.isZero())
4154 NewPredicate = ICmpInst::ICMP_ULT;
4155 else if (C.isOne())
4156 NewPredicate = ICmpInst::ICMP_ULE;
4157 break;
4158
4159 case ICmpInst::ICMP_ULT:
4160 if (C.ugt(1))
4161 NewPredicate = ICmpInst::ICMP_UGE;
4162 break;
4163
4164 case ICmpInst::ICMP_UGT:
4165 if (!C.isZero() && !C.isAllOnes())
4166 NewPredicate = ICmpInst::ICMP_ULT;
4167 break;
4168
4169 default:
4170 break;
4171 }
4172
4173 if (!NewPredicate)
4174 return nullptr;
4175
4176 if (I->getIntrinsicID() == Intrinsic::scmp)
4177 NewPredicate = ICmpInst::getSignedPredicate(*NewPredicate);
4178 Value *LHS = I->getOperand(0);
4179 Value *RHS = I->getOperand(1);
4180 return new ICmpInst(*NewPredicate, LHS, RHS);
4181}
4182
4183/// Fold an icmp with LLVM intrinsic and constant operand: icmp Pred II, C.
4186 const APInt &C) {
4187 ICmpInst::Predicate Pred = Cmp.getPredicate();
4188
4189 // Handle folds that apply for any kind of icmp.
4190 switch (II->getIntrinsicID()) {
4191 default:
4192 break;
4193 case Intrinsic::uadd_sat:
4194 case Intrinsic::usub_sat:
4195 if (auto *Folded = foldICmpUSubSatOrUAddSatWithConstant(
4196 Pred, cast<SaturatingInst>(II), C, Builder))
4197 return Folded;
4198 break;
4199 case Intrinsic::ctpop: {
4200 const SimplifyQuery Q = SQ.getWithInstruction(&Cmp);
4201 if (Instruction *R = foldCtpopPow2Test(Cmp, II, C, Builder, Q))
4202 return R;
4203 } break;
4204 case Intrinsic::scmp:
4205 case Intrinsic::ucmp:
4206 if (auto *Folded = foldICmpOfCmpIntrinsicWithConstant(Pred, II, C, Builder))
4207 return Folded;
4208 break;
4209 }
4210
4211 if (Cmp.isEquality())
4212 return foldICmpEqIntrinsicWithConstant(Cmp, II, C);
4213
4214 Type *Ty = II->getType();
4215 unsigned BitWidth = C.getBitWidth();
4216 switch (II->getIntrinsicID()) {
4217 case Intrinsic::ctpop: {
4218 // (ctpop X > BitWidth - 1) --> X == -1
4219 Value *X = II->getArgOperand(0);
4220 if (C == BitWidth - 1 && Pred == ICmpInst::ICMP_UGT)
4221 return CmpInst::Create(Instruction::ICmp, ICmpInst::ICMP_EQ, X,
4223 // (ctpop X < BitWidth) --> X != -1
4224 if (C == BitWidth && Pred == ICmpInst::ICMP_ULT)
4225 return CmpInst::Create(Instruction::ICmp, ICmpInst::ICMP_NE, X,
4227 break;
4228 }
4229 case Intrinsic::ctlz: {
4230 // ctlz(0bXXXXXXXX) > 3 -> 0bXXXXXXXX < 0b00010000
4231 if (Pred == ICmpInst::ICMP_UGT && C.ult(BitWidth)) {
4232 unsigned Num = C.getLimitedValue();
4233 APInt Limit = APInt::getOneBitSet(BitWidth, BitWidth - Num - 1);
4234 return CmpInst::Create(Instruction::ICmp, ICmpInst::ICMP_ULT,
4235 II->getArgOperand(0), ConstantInt::get(Ty, Limit));
4236 }
4237
4238 // ctlz(0bXXXXXXXX) < 3 -> 0bXXXXXXXX > 0b00011111
4239 if (Pred == ICmpInst::ICMP_ULT && C.uge(1) && C.ule(BitWidth)) {
4240 unsigned Num = C.getLimitedValue();
4242 return CmpInst::Create(Instruction::ICmp, ICmpInst::ICMP_UGT,
4243 II->getArgOperand(0), ConstantInt::get(Ty, Limit));
4244 }
4245 break;
4246 }
4247 case Intrinsic::cttz: {
4248 // Limit to one use to ensure we don't increase instruction count.
4249 if (!II->hasOneUse())
4250 return nullptr;
4251
4252 // cttz(0bXXXXXXXX) > 3 -> 0bXXXXXXXX & 0b00001111 == 0
4253 if (Pred == ICmpInst::ICMP_UGT && C.ult(BitWidth)) {
4254 APInt Mask = APInt::getLowBitsSet(BitWidth, C.getLimitedValue() + 1);
4255 return CmpInst::Create(Instruction::ICmp, ICmpInst::ICMP_EQ,
4256 Builder.CreateAnd(II->getArgOperand(0), Mask),
4258 }
4259
4260 // cttz(0bXXXXXXXX) < 3 -> 0bXXXXXXXX & 0b00000111 != 0
4261 if (Pred == ICmpInst::ICMP_ULT && C.uge(1) && C.ule(BitWidth)) {
4262 APInt Mask = APInt::getLowBitsSet(BitWidth, C.getLimitedValue());
4263 return CmpInst::Create(Instruction::ICmp, ICmpInst::ICMP_NE,
4264 Builder.CreateAnd(II->getArgOperand(0), Mask),
4266 }
4267 break;
4268 }
4269 case Intrinsic::ssub_sat:
4270 // ssub.sat(a, b) spred 0 -> a spred b
4271 //
4272 // Note this doesn't work for ssub.sat.i1 because ssub.sat.i1 0, -1 = 0
4273 // (because 1 saturates to 0). Just skip the optimization for i1.
4274 if (ICmpInst::isSigned(Pred) && C.getBitWidth() > 1) {
4275 if (C.isZero())
4276 return new ICmpInst(Pred, II->getArgOperand(0), II->getArgOperand(1));
4277 // X s<= 0 is cannonicalized to X s< 1
4278 if (Pred == ICmpInst::ICMP_SLT && C.isOne())
4279 return new ICmpInst(ICmpInst::ICMP_SLE, II->getArgOperand(0),
4280 II->getArgOperand(1));
4281 // X s>= 0 is cannonicalized to X s> -1
4282 if (Pred == ICmpInst::ICMP_SGT && C.isAllOnes())
4283 return new ICmpInst(ICmpInst::ICMP_SGE, II->getArgOperand(0),
4284 II->getArgOperand(1));
4285 }
4286 break;
4287 case Intrinsic::abs: {
4288 if (!II->hasOneUse())
4289 return nullptr;
4290
4291 Value *X = II->getArgOperand(0);
4292 bool IsIntMinPoison =
4293 cast<ConstantInt>(II->getArgOperand(1))->getValue().isOne();
4294
4295 // If C >= 0:
4296 // abs(X) u> C --> X + C u> 2 * C
4297 if (Pred == CmpInst::ICMP_UGT && C.isNonNegative()) {
4298 return new ICmpInst(ICmpInst::ICMP_UGT,
4299 Builder.CreateAdd(X, ConstantInt::get(Ty, C)),
4300 ConstantInt::get(Ty, 2 * C));
4301 }
4302
4303 // If abs(INT_MIN) is poison and C >= 1:
4304 // abs(X) u< C --> X + (C - 1) u<= 2 * (C - 1)
4305 if (IsIntMinPoison && Pred == CmpInst::ICMP_ULT && C.sge(1)) {
4306 return new ICmpInst(ICmpInst::ICMP_ULE,
4307 Builder.CreateAdd(X, ConstantInt::get(Ty, C - 1)),
4308 ConstantInt::get(Ty, 2 * (C - 1)));
4309 }
4310
4311 break;
4312 }
4313 default:
4314 break;
4315 }
4316
4317 return nullptr;
4318}
4319
4320/// Handle icmp with constant (but not simple integer constant) RHS.
4322 Value *Op0 = I.getOperand(0), *Op1 = I.getOperand(1);
4323 Constant *RHSC = dyn_cast<Constant>(Op1);
4325 if (!RHSC || !LHSI)
4326 return nullptr;
4327
4328 switch (LHSI->getOpcode()) {
4329 case Instruction::IntToPtr:
4330 // icmp pred inttoptr(X), null -> icmp pred X, 0
4331 if (RHSC->isNullValue() &&
4332 DL.getIntPtrType(RHSC->getType()) == LHSI->getOperand(0)->getType())
4333 return new ICmpInst(
4334 I.getPredicate(), LHSI->getOperand(0),
4336 break;
4337
4338 case Instruction::Load:
4339 // Try to optimize things like "A[i] > 4" to index computations.
4340 if (GetElementPtrInst *GEP =
4342 if (Instruction *Res =
4344 return Res;
4345 break;
4346 }
4347
4348 return nullptr;
4349}
4350
4352 Value *RHS, const ICmpInst &I) {
4353 // Try to fold the comparison into the select arms, which will cause the
4354 // select to be converted into a logical and/or.
4355 auto SimplifyOp = [&](Value *Op, bool SelectCondIsTrue) -> Value * {
4356 if (Value *Res = simplifyICmpInst(Pred, Op, RHS, SQ))
4357 return Res;
4358 if (std::optional<bool> Impl = isImpliedCondition(
4359 SI->getCondition(), Pred, Op, RHS, DL, SelectCondIsTrue))
4360 return ConstantInt::get(I.getType(), *Impl);
4361 return nullptr;
4362 };
4363
4364 ConstantInt *CI = nullptr;
4365 Value *Op1 = SimplifyOp(SI->getOperand(1), true);
4366 if (Op1)
4367 CI = dyn_cast<ConstantInt>(Op1);
4368
4369 Value *Op2 = SimplifyOp(SI->getOperand(2), false);
4370 if (Op2)
4371 CI = dyn_cast<ConstantInt>(Op2);
4372
4373 auto Simplifies = [&](Value *Op, unsigned Idx) {
4374 // A comparison of ucmp/scmp with a constant will fold into an icmp.
4375 const APInt *Dummy;
4376 return Op ||
4377 (isa<CmpIntrinsic>(SI->getOperand(Idx)) &&
4378 SI->getOperand(Idx)->hasOneUse() && match(RHS, m_APInt(Dummy)));
4379 };
4380
4381 // We only want to perform this transformation if it will not lead to
4382 // additional code. This is true if either both sides of the select
4383 // fold to a constant (in which case the icmp is replaced with a select
4384 // which will usually simplify) or this is the only user of the
4385 // select (in which case we are trading a select+icmp for a simpler
4386 // select+icmp) or all uses of the select can be replaced based on
4387 // dominance information ("Global cases").
4388 bool Transform = false;
4389 if (Op1 && Op2)
4390 Transform = true;
4391 else if (Simplifies(Op1, 1) || Simplifies(Op2, 2)) {
4392 // Local case
4393 if (SI->hasOneUse())
4394 Transform = true;
4395 // Global cases
4396 else if (CI && !CI->isZero())
4397 // When Op1 is constant try replacing select with second operand.
4398 // Otherwise Op2 is constant and try replacing select with first
4399 // operand.
4400 Transform = replacedSelectWithOperand(SI, &I, Op1 ? 2 : 1);
4401 }
4402 if (Transform) {
4403 if (!Op1)
4404 Op1 = Builder.CreateICmp(Pred, SI->getOperand(1), RHS, I.getName());
4405 if (!Op2)
4406 Op2 = Builder.CreateICmp(Pred, SI->getOperand(2), RHS, I.getName());
4407 return SelectInst::Create(SI->getOperand(0), Op1, Op2, "", nullptr,
4408 ProfcheckDisableMetadataFixes ? nullptr : SI);
4409 }
4410
4411 return nullptr;
4412}
4413
4414// Returns whether V is a Mask ((X + 1) & X == 0) or ~Mask (-Pow2OrZero)
4415static bool isMaskOrZero(const Value *V, bool Not, const SimplifyQuery &Q,
4416 unsigned Depth = 0) {
4417 if (Not ? match(V, m_NegatedPower2OrZero()) : match(V, m_LowBitMaskOrZero()))
4418 return true;
4419 if (V->getType()->getScalarSizeInBits() == 1)
4420 return true;
4422 return false;
4423 Value *X;
4425 if (!I)
4426 return false;
4427 switch (I->getOpcode()) {
4428 case Instruction::ZExt:
4429 // ZExt(Mask) is a Mask.
4430 return !Not && isMaskOrZero(I->getOperand(0), Not, Q, Depth);
4431 case Instruction::SExt:
4432 // SExt(Mask) is a Mask.
4433 // SExt(~Mask) is a ~Mask.
4434 return isMaskOrZero(I->getOperand(0), Not, Q, Depth);
4435 case Instruction::And:
4436 case Instruction::Or:
4437 // Mask0 | Mask1 is a Mask.
4438 // Mask0 & Mask1 is a Mask.
4439 // ~Mask0 | ~Mask1 is a ~Mask.
4440 // ~Mask0 & ~Mask1 is a ~Mask.
4441 return isMaskOrZero(I->getOperand(1), Not, Q, Depth) &&
4442 isMaskOrZero(I->getOperand(0), Not, Q, Depth);
4443 case Instruction::Xor:
4444 if (match(V, m_Not(m_Value(X))))
4445 return isMaskOrZero(X, !Not, Q, Depth);
4446
4447 // (X ^ -X) is a ~Mask
4448 if (Not)
4449 return match(V, m_c_Xor(m_Value(X), m_Neg(m_Deferred(X))));
4450 // (X ^ (X - 1)) is a Mask
4451 else
4452 return match(V, m_c_Xor(m_Value(X), m_Add(m_Deferred(X), m_AllOnes())));
4453 case Instruction::Select:
4454 // c ? Mask0 : Mask1 is a Mask.
4455 return isMaskOrZero(I->getOperand(1), Not, Q, Depth) &&
4456 isMaskOrZero(I->getOperand(2), Not, Q, Depth);
4457 case Instruction::Shl:
4458 // (~Mask) << X is a ~Mask.
4459 return Not && isMaskOrZero(I->getOperand(0), Not, Q, Depth);
4460 case Instruction::LShr:
4461 // Mask >> X is a Mask.
4462 return !Not && isMaskOrZero(I->getOperand(0), Not, Q, Depth);
4463 case Instruction::AShr:
4464 // Mask s>> X is a Mask.
4465 // ~Mask s>> X is a ~Mask.
4466 return isMaskOrZero(I->getOperand(0), Not, Q, Depth);
4467 case Instruction::Add:
4468 // Pow2 - 1 is a Mask.
4469 if (!Not && match(I->getOperand(1), m_AllOnes()))
4470 return isKnownToBeAPowerOfTwo(I->getOperand(0), Q.DL, /*OrZero*/ true,
4471 Q.AC, Q.CxtI, Q.DT, Depth);
4472 break;
4473 case Instruction::Sub:
4474 // -Pow2 is a ~Mask.
4475 if (Not && match(I->getOperand(0), m_Zero()))
4476 return isKnownToBeAPowerOfTwo(I->getOperand(1), Q.DL, /*OrZero*/ true,
4477 Q.AC, Q.CxtI, Q.DT, Depth);
4478 break;
4479 case Instruction::Call: {
4480 if (auto *II = dyn_cast<IntrinsicInst>(I)) {
4481 switch (II->getIntrinsicID()) {
4482 // min/max(Mask0, Mask1) is a Mask.
4483 // min/max(~Mask0, ~Mask1) is a ~Mask.
4484 case Intrinsic::umax:
4485 case Intrinsic::smax:
4486 case Intrinsic::umin:
4487 case Intrinsic::smin:
4488 return isMaskOrZero(II->getArgOperand(1), Not, Q, Depth) &&
4489 isMaskOrZero(II->getArgOperand(0), Not, Q, Depth);
4490
4491 // In the context of masks, bitreverse(Mask) == ~Mask
4492 case Intrinsic::bitreverse:
4493 return isMaskOrZero(II->getArgOperand(0), !Not, Q, Depth);
4494 default:
4495 break;
4496 }
4497 }
4498 break;
4499 }
4500 default:
4501 break;
4502 }
4503 return false;
4504}
4505
4506/// Some comparisons can be simplified.
4507/// In this case, we are looking for comparisons that look like
4508/// a check for a lossy truncation.
4509/// Folds:
4510/// icmp SrcPred (x & Mask), x to icmp DstPred x, Mask
4511/// icmp SrcPred (x & ~Mask), ~Mask to icmp DstPred x, ~Mask
4512/// icmp eq/ne (x & ~Mask), 0 to icmp DstPred x, Mask
4513/// icmp eq/ne (~x | Mask), -1 to icmp DstPred x, Mask
4514/// Where Mask is some pattern that produces all-ones in low bits:
4515/// (-1 >> y)
4516/// ((-1 << y) >> y) <- non-canonical, has extra uses
4517/// ~(-1 << y)
4518/// ((1 << y) + (-1)) <- non-canonical, has extra uses
4519/// The Mask can be a constant, too.
4520/// For some predicates, the operands are commutative.
4521/// For others, x can only be on a specific side.
4523 Value *Op1, const SimplifyQuery &Q,
4524 InstCombiner &IC) {
4525
4526 ICmpInst::Predicate DstPred;
4527 switch (Pred) {
4529 // x & Mask == x
4530 // x & ~Mask == 0
4531 // ~x | Mask == -1
4532 // -> x u<= Mask
4533 // x & ~Mask == ~Mask
4534 // -> ~Mask u<= x
4536 break;
4538 // x & Mask != x
4539 // x & ~Mask != 0
4540 // ~x | Mask != -1
4541 // -> x u> Mask
4542 // x & ~Mask != ~Mask
4543 // -> ~Mask u> x
4545 break;
4547 // x & Mask u< x
4548 // -> x u> Mask
4549 // x & ~Mask u< ~Mask
4550 // -> ~Mask u> x
4552 break;
4554 // x & Mask u>= x
4555 // -> x u<= Mask
4556 // x & ~Mask u>= ~Mask
4557 // -> ~Mask u<= x
4559 break;
4561 // x & Mask s< x [iff Mask s>= 0]
4562 // -> x s> Mask
4563 // x & ~Mask s< ~Mask [iff ~Mask != 0]
4564 // -> ~Mask s> x
4566 break;
4568 // x & Mask s>= x [iff Mask s>= 0]
4569 // -> x s<= Mask
4570 // x & ~Mask s>= ~Mask [iff ~Mask != 0]
4571 // -> ~Mask s<= x
4573 break;
4574 default:
4575 // We don't support sgt,sle
4576 // ult/ugt are simplified to true/false respectively.
4577 return nullptr;
4578 }
4579
4580 Value *X, *M;
4581 // Put search code in lambda for early positive returns.
4582 auto IsLowBitMask = [&]() {
4583 if (match(Op0, m_c_And(m_Specific(Op1), m_Value(M)))) {
4584 X = Op1;
4585 // Look for: x & Mask pred x
4586 if (isMaskOrZero(M, /*Not=*/false, Q)) {
4587 return !ICmpInst::isSigned(Pred) ||
4588 (match(M, m_NonNegative()) || isKnownNonNegative(M, Q));
4589 }
4590
4591 // Look for: x & ~Mask pred ~Mask
4592 if (isMaskOrZero(X, /*Not=*/true, Q)) {
4593 return !ICmpInst::isSigned(Pred) || isKnownNonZero(X, Q);
4594 }
4595 return false;
4596 }
4597 if (ICmpInst::isEquality(Pred) && match(Op1, m_AllOnes()) &&
4598 match(Op0, m_OneUse(m_Or(m_Value(X), m_Value(M))))) {
4599
4600 auto Check = [&]() {
4601 // Look for: ~x | Mask == -1
4602 if (isMaskOrZero(M, /*Not=*/false, Q)) {
4603 if (Value *NotX =
4604 IC.getFreelyInverted(X, X->hasOneUse(), &IC.Builder)) {
4605 X = NotX;
4606 return true;
4607 }
4608 }
4609 return false;
4610 };
4611 if (Check())
4612 return true;
4613 std::swap(X, M);
4614 return Check();
4615 }
4616 if (ICmpInst::isEquality(Pred) && match(Op1, m_Zero()) &&
4617 match(Op0, m_OneUse(m_And(m_Value(X), m_Value(M))))) {
4618 auto Check = [&]() {
4619 // Look for: x & ~Mask == 0
4620 if (isMaskOrZero(M, /*Not=*/true, Q)) {
4621 if (Value *NotM =
4622 IC.getFreelyInverted(M, M->hasOneUse(), &IC.Builder)) {
4623 M = NotM;
4624 return true;
4625 }
4626 }
4627 return false;
4628 };
4629 if (Check())
4630 return true;
4631 std::swap(X, M);
4632 return Check();
4633 }
4634 return false;
4635 };
4636
4637 if (!IsLowBitMask())
4638 return nullptr;
4639
4640 return IC.Builder.CreateICmp(DstPred, X, M);
4641}
4642
4643/// Some comparisons can be simplified.
4644/// In this case, we are looking for comparisons that look like
4645/// a check for a lossy signed truncation.
4646/// Folds: (MaskedBits is a constant.)
4647/// ((%x << MaskedBits) a>> MaskedBits) SrcPred %x
4648/// Into:
4649/// (add %x, (1 << (KeptBits-1))) DstPred (1 << KeptBits)
4650/// Where KeptBits = bitwidth(%x) - MaskedBits
4651static Value *
4653 InstCombiner::BuilderTy &Builder) {
4654 CmpPredicate SrcPred;
4655 Value *X;
4656 const APInt *C0, *C1; // FIXME: non-splats, potentially with undef.
4657 // We are ok with 'shl' having multiple uses, but 'ashr' must be one-use.
4658 if (!match(&I, m_c_ICmp(SrcPred,
4660 m_APInt(C1))),
4661 m_Deferred(X))))
4662 return nullptr;
4663
4664 // Potential handling of non-splats: for each element:
4665 // * if both are undef, replace with constant 0.
4666 // Because (1<<0) is OK and is 1, and ((1<<0)>>1) is also OK and is 0.
4667 // * if both are not undef, and are different, bailout.
4668 // * else, only one is undef, then pick the non-undef one.
4669
4670 // The shift amount must be equal.
4671 if (*C0 != *C1)
4672 return nullptr;
4673 const APInt &MaskedBits = *C0;
4674 assert(MaskedBits != 0 && "shift by zero should be folded away already.");
4675
4676 ICmpInst::Predicate DstPred;
4677 switch (SrcPred) {
4679 // ((%x << MaskedBits) a>> MaskedBits) == %x
4680 // =>
4681 // (add %x, (1 << (KeptBits-1))) u< (1 << KeptBits)
4683 break;
4685 // ((%x << MaskedBits) a>> MaskedBits) != %x
4686 // =>
4687 // (add %x, (1 << (KeptBits-1))) u>= (1 << KeptBits)
4689 break;
4690 // FIXME: are more folds possible?
4691 default:
4692 return nullptr;
4693 }
4694
4695 auto *XType = X->getType();
4696 const unsigned XBitWidth = XType->getScalarSizeInBits();
4697 const APInt BitWidth = APInt(XBitWidth, XBitWidth);
4698 assert(BitWidth.ugt(MaskedBits) && "shifts should leave some bits untouched");
4699
4700 // KeptBits = bitwidth(%x) - MaskedBits
4701 const APInt KeptBits = BitWidth - MaskedBits;
4702 assert(KeptBits.ugt(0) && KeptBits.ult(BitWidth) && "unreachable");
4703 // ICmpCst = (1 << KeptBits)
4704 const APInt ICmpCst = APInt(XBitWidth, 1).shl(KeptBits);
4705 assert(ICmpCst.isPowerOf2());
4706 // AddCst = (1 << (KeptBits-1))
4707 const APInt AddCst = ICmpCst.lshr(1);
4708 assert(AddCst.ult(ICmpCst) && AddCst.isPowerOf2());
4709
4710 // T0 = add %x, AddCst
4711 Value *T0 = Builder.CreateAdd(X, ConstantInt::get(XType, AddCst));
4712 // T1 = T0 DstPred ICmpCst
4713 Value *T1 = Builder.CreateICmp(DstPred, T0, ConstantInt::get(XType, ICmpCst));
4714
4715 return T1;
4716}
4717
4718// Given pattern:
4719// icmp eq/ne (and ((x shift Q), (y oppositeshift K))), 0
4720// we should move shifts to the same hand of 'and', i.e. rewrite as
4721// icmp eq/ne (and (x shift (Q+K)), y), 0 iff (Q+K) u< bitwidth(x)
4722// We are only interested in opposite logical shifts here.
4723// One of the shifts can be truncated.
4724// If we can, we want to end up creating 'lshr' shift.
4725static Value *
4727 InstCombiner::BuilderTy &Builder) {
4728 if (!I.isEquality() || !match(I.getOperand(1), m_Zero()) ||
4729 !I.getOperand(0)->hasOneUse())
4730 return nullptr;
4731
4732 auto m_AnyLogicalShift = m_LogicalShift(m_Value(), m_Value());
4733
4734 // Look for an 'and' of two logical shifts, one of which may be truncated.
4735 // We use m_TruncOrSelf() on the RHS to correctly handle commutative case.
4736 Instruction *XShift, *MaybeTruncation, *YShift;
4737 if (!match(
4738 I.getOperand(0),
4739 m_c_And(m_CombineAnd(m_AnyLogicalShift, m_Instruction(XShift)),
4741 m_AnyLogicalShift, m_Instruction(YShift))),
4742 m_Instruction(MaybeTruncation)))))
4743 return nullptr;
4744
4745 // We potentially looked past 'trunc', but only when matching YShift,
4746 // therefore YShift must have the widest type.
4747 Instruction *WidestShift = YShift;
4748 // Therefore XShift must have the shallowest type.
4749 // Or they both have identical types if there was no truncation.
4750 Instruction *NarrowestShift = XShift;
4751
4752 Type *WidestTy = WidestShift->getType();
4753 Type *NarrowestTy = NarrowestShift->getType();
4754 assert(NarrowestTy == I.getOperand(0)->getType() &&
4755 "We did not look past any shifts while matching XShift though.");
4756 bool HadTrunc = WidestTy != I.getOperand(0)->getType();
4757
4758 // If YShift is a 'lshr', swap the shifts around.
4759 if (match(YShift, m_LShr(m_Value(), m_Value())))
4760 std::swap(XShift, YShift);
4761
4762 // The shifts must be in opposite directions.
4763 auto XShiftOpcode = XShift->getOpcode();
4764 if (XShiftOpcode == YShift->getOpcode())
4765 return nullptr; // Do not care about same-direction shifts here.
4766
4767 Value *X, *XShAmt, *Y, *YShAmt;
4768 match(XShift, m_BinOp(m_Value(X), m_ZExtOrSelf(m_Value(XShAmt))));
4769 match(YShift, m_BinOp(m_Value(Y), m_ZExtOrSelf(m_Value(YShAmt))));
4770
4771 // If one of the values being shifted is a constant, then we will end with
4772 // and+icmp, and [zext+]shift instrs will be constant-folded. If they are not,
4773 // however, we will need to ensure that we won't increase instruction count.
4774 if (!isa<Constant>(X) && !isa<Constant>(Y)) {
4775 // At least one of the hands of the 'and' should be one-use shift.
4776 if (!match(I.getOperand(0),
4777 m_c_And(m_OneUse(m_AnyLogicalShift), m_Value())))
4778 return nullptr;
4779 if (HadTrunc) {
4780 // Due to the 'trunc', we will need to widen X. For that either the old
4781 // 'trunc' or the shift amt in the non-truncated shift should be one-use.
4782 if (!MaybeTruncation->hasOneUse() &&
4783 !NarrowestShift->getOperand(1)->hasOneUse())
4784 return nullptr;
4785 }
4786 }
4787
4788 // We have two shift amounts from two different shifts. The types of those
4789 // shift amounts may not match. If that's the case let's bailout now.
4790 if (XShAmt->getType() != YShAmt->getType())
4791 return nullptr;
4792
4793 // As input, we have the following pattern:
4794 // icmp eq/ne (and ((x shift Q), (y oppositeshift K))), 0
4795 // We want to rewrite that as:
4796 // icmp eq/ne (and (x shift (Q+K)), y), 0 iff (Q+K) u< bitwidth(x)
4797 // While we know that originally (Q+K) would not overflow
4798 // (because 2 * (N-1) u<= iN -1), we have looked past extensions of
4799 // shift amounts. so it may now overflow in smaller bitwidth.
4800 // To ensure that does not happen, we need to ensure that the total maximal
4801 // shift amount is still representable in that smaller bit width.
4802 unsigned MaximalPossibleTotalShiftAmount =
4803 (WidestTy->getScalarSizeInBits() - 1) +
4804 (NarrowestTy->getScalarSizeInBits() - 1);
4805 APInt MaximalRepresentableShiftAmount =
4807 if (MaximalRepresentableShiftAmount.ult(MaximalPossibleTotalShiftAmount))
4808 return nullptr;
4809
4810 // Can we fold (XShAmt+YShAmt) ?
4811 auto *NewShAmt = dyn_cast_or_null<Constant>(
4812 simplifyAddInst(XShAmt, YShAmt, /*isNSW=*/false,
4813 /*isNUW=*/false, SQ.getWithInstruction(&I)));
4814 if (!NewShAmt)
4815 return nullptr;
4816 if (NewShAmt->getType() != WidestTy) {
4817 NewShAmt =
4818 ConstantFoldCastOperand(Instruction::ZExt, NewShAmt, WidestTy, SQ.DL);
4819 if (!NewShAmt)
4820 return nullptr;
4821 }
4822 unsigned WidestBitWidth = WidestTy->getScalarSizeInBits();
4823
4824 // Is the new shift amount smaller than the bit width?
4825 // FIXME: could also rely on ConstantRange.
4826 if (!match(NewShAmt,
4828 APInt(WidestBitWidth, WidestBitWidth))))
4829 return nullptr;
4830
4831 // An extra legality check is needed if we had trunc-of-lshr.
4832 if (HadTrunc && match(WidestShift, m_LShr(m_Value(), m_Value()))) {
4833 auto CanFold = [NewShAmt, WidestBitWidth, NarrowestShift, SQ,
4834 WidestShift]() {
4835 // It isn't obvious whether it's worth it to analyze non-constants here.
4836 // Also, let's basically give up on non-splat cases, pessimizing vectors.
4837 // If *any* of these preconditions matches we can perform the fold.
4838 Constant *NewShAmtSplat = NewShAmt->getType()->isVectorTy()
4839 ? NewShAmt->getSplatValue()
4840 : NewShAmt;
4841 // If it's edge-case shift (by 0 or by WidestBitWidth-1) we can fold.
4842 if (NewShAmtSplat &&
4843 (NewShAmtSplat->isNullValue() ||
4844 NewShAmtSplat->getUniqueInteger() == WidestBitWidth - 1))
4845 return true;
4846 // We consider *min* leading zeros so a single outlier
4847 // blocks the transform as opposed to allowing it.
4848 if (auto *C = dyn_cast<Constant>(NarrowestShift->getOperand(0))) {
4849 KnownBits Known = computeKnownBits(C, SQ.DL);
4850 unsigned MinLeadZero = Known.countMinLeadingZeros();
4851 // If the value being shifted has at most lowest bit set we can fold.
4852 unsigned MaxActiveBits = Known.getBitWidth() - MinLeadZero;
4853 if (MaxActiveBits <= 1)
4854 return true;
4855 // Precondition: NewShAmt u<= countLeadingZeros(C)
4856 if (NewShAmtSplat && NewShAmtSplat->getUniqueInteger().ule(MinLeadZero))
4857 return true;
4858 }
4859 if (auto *C = dyn_cast<Constant>(WidestShift->getOperand(0))) {
4860 KnownBits Known = computeKnownBits(C, SQ.DL);
4861 unsigned MinLeadZero = Known.countMinLeadingZeros();
4862 // If the value being shifted has at most lowest bit set we can fold.
4863 unsigned MaxActiveBits = Known.getBitWidth() - MinLeadZero;
4864 if (MaxActiveBits <= 1)
4865 return true;
4866 // Precondition: ((WidestBitWidth-1)-NewShAmt) u<= countLeadingZeros(C)
4867 if (NewShAmtSplat) {
4868 APInt AdjNewShAmt =
4869 (WidestBitWidth - 1) - NewShAmtSplat->getUniqueInteger();
4870 if (AdjNewShAmt.ule(MinLeadZero))
4871 return true;
4872 }
4873 }
4874 return false; // Can't tell if it's ok.
4875 };
4876 if (!CanFold())
4877 return nullptr;
4878 }
4879
4880 // All good, we can do this fold.
4881 X = Builder.CreateZExt(X, WidestTy);
4882 Y = Builder.CreateZExt(Y, WidestTy);
4883 // The shift is the same that was for X.
4884 Value *T0 = XShiftOpcode == Instruction::BinaryOps::LShr
4885 ? Builder.CreateLShr(X, NewShAmt)
4886 : Builder.CreateShl(X, NewShAmt);
4887 Value *T1 = Builder.CreateAnd(T0, Y);
4888 return Builder.CreateICmp(I.getPredicate(), T1,
4889 Constant::getNullValue(WidestTy));
4890}
4891
4892/// Fold
4893/// (-1 u/ x) u< y
4894/// ((x * y) ?/ x) != y
4895/// to
4896/// @llvm.?mul.with.overflow(x, y) plus extraction of overflow bit
4897/// Note that the comparison is commutative, while inverted (u>=, ==) predicate
4898/// will mean that we are looking for the opposite answer.
4900 CmpPredicate Pred;
4901 Value *X, *Y;
4903 Instruction *Div;
4904 bool NeedNegation;
4905 // Look for: (-1 u/ x) u</u>= y
4906 if (!I.isEquality() &&
4907 match(&I, m_c_ICmp(Pred,
4909 m_Instruction(Div)),
4910 m_Value(Y)))) {
4911 Mul = nullptr;
4912
4913 // Are we checking that overflow does not happen, or does happen?
4914 switch (Pred) {
4916 NeedNegation = false;
4917 break; // OK
4919 NeedNegation = true;
4920 break; // OK
4921 default:
4922 return nullptr; // Wrong predicate.
4923 }
4924 } else // Look for: ((x * y) / x) !=/== y
4925 if (I.isEquality() &&
4926 match(&I, m_c_ICmp(Pred, m_Value(Y),
4929 m_Value(X)),
4931 m_Deferred(X))),
4932 m_Instruction(Div))))) {
4933 NeedNegation = Pred == ICmpInst::Predicate::ICMP_EQ;
4934 } else
4935 return nullptr;
4936
4938 // If the pattern included (x * y), we'll want to insert new instructions
4939 // right before that original multiplication so that we can replace it.
4940 bool MulHadOtherUses = Mul && !Mul->hasOneUse();
4941 if (MulHadOtherUses)
4942 Builder.SetInsertPoint(Mul);
4943
4944 CallInst *Call = Builder.CreateIntrinsic(
4945 Div->getOpcode() == Instruction::UDiv ? Intrinsic::umul_with_overflow
4946 : Intrinsic::smul_with_overflow,
4947 X->getType(), {X, Y}, /*FMFSource=*/nullptr, "mul");
4948
4949 // If the multiplication was used elsewhere, to ensure that we don't leave
4950 // "duplicate" instructions, replace uses of that original multiplication
4951 // with the multiplication result from the with.overflow intrinsic.
4952 if (MulHadOtherUses)
4953 replaceInstUsesWith(*Mul, Builder.CreateExtractValue(Call, 0, "mul.val"));
4954
4955 Value *Res = Builder.CreateExtractValue(Call, 1, "mul.ov");
4956 if (NeedNegation) // This technically increases instruction count.
4957 Res = Builder.CreateNot(Res, "mul.not.ov");
4958
4959 // If we replaced the mul, erase it. Do this after all uses of Builder,
4960 // as the mul is used as insertion point.
4961 if (MulHadOtherUses)
4963
4964 return Res;
4965}
4966
4968 InstCombiner::BuilderTy &Builder) {
4969 CmpPredicate Pred;
4970 Value *X;
4971 if (match(&I, m_c_ICmp(Pred, m_NSWNeg(m_Value(X)), m_Deferred(X)))) {
4972
4973 if (ICmpInst::isSigned(Pred))
4974 Pred = ICmpInst::getSwappedPredicate(Pred);
4975 else if (ICmpInst::isUnsigned(Pred))
4976 Pred = ICmpInst::getSignedPredicate(Pred);
4977 // else for equality-comparisons just keep the predicate.
4978
4979 return ICmpInst::Create(Instruction::ICmp, Pred, X,
4980 Constant::getNullValue(X->getType()), I.getName());
4981 }
4982
4983 // A value is not equal to its negation unless that value is 0 or
4984 // MinSignedValue, ie: a != -a --> (a & MaxSignedVal) != 0
4985 if (match(&I, m_c_ICmp(Pred, m_OneUse(m_Neg(m_Value(X))), m_Deferred(X))) &&
4986 ICmpInst::isEquality(Pred)) {
4987 Type *Ty = X->getType();
4988 uint32_t BitWidth = Ty->getScalarSizeInBits();
4989 Constant *MaxSignedVal =
4990 ConstantInt::get(Ty, APInt::getSignedMaxValue(BitWidth));
4991 Value *And = Builder.CreateAnd(X, MaxSignedVal);
4992 Constant *Zero = Constant::getNullValue(Ty);
4993 return CmpInst::Create(Instruction::ICmp, Pred, And, Zero);
4994 }
4995
4996 return nullptr;
4997}
4998
5000 InstCombinerImpl &IC) {
5001 Value *Op0 = I.getOperand(0), *Op1 = I.getOperand(1), *A;
5002 // Normalize and operand as operand 0.
5003 CmpInst::Predicate Pred = I.getPredicate();
5004 if (match(Op1, m_c_And(m_Specific(Op0), m_Value()))) {
5005 std::swap(Op0, Op1);
5006 Pred = ICmpInst::getSwappedPredicate(Pred);
5007 }
5008
5009 if (!match(Op0, m_c_And(m_Specific(Op1), m_Value(A))))
5010 return nullptr;
5011
5012 // (icmp (X & Y) u< X --> (X & Y) != X
5013 if (Pred == ICmpInst::ICMP_ULT)
5014 return new ICmpInst(ICmpInst::ICMP_NE, Op0, Op1);
5015
5016 // (icmp (X & Y) u>= X --> (X & Y) == X
5017 if (Pred == ICmpInst::ICMP_UGE)
5018 return new ICmpInst(ICmpInst::ICMP_EQ, Op0, Op1);
5019
5020 if (ICmpInst::isEquality(Pred) && Op0->hasOneUse()) {
5021 // icmp (X & Y) eq/ne Y --> (X | ~Y) eq/ne -1 if Y is freely invertible and
5022 // Y is non-constant. If Y is constant the `X & C == C` form is preferable
5023 // so don't do this fold.
5024 if (!match(Op1, m_ImmConstant()))
5025 if (auto *NotOp1 =
5026 IC.getFreelyInverted(Op1, !Op1->hasNUsesOrMore(3), &IC.Builder))
5027 return new ICmpInst(Pred, IC.Builder.CreateOr(A, NotOp1),
5028 Constant::getAllOnesValue(Op1->getType()));
5029 // icmp (X & Y) eq/ne Y --> (~X & Y) eq/ne 0 if X is freely invertible.
5030 if (auto *NotA = IC.getFreelyInverted(A, A->hasOneUse(), &IC.Builder))
5031 return new ICmpInst(Pred, IC.Builder.CreateAnd(Op1, NotA),
5032 Constant::getNullValue(Op1->getType()));
5033 }
5034
5035 if (!ICmpInst::isSigned(Pred))
5036 return nullptr;
5037
5038 KnownBits KnownY = IC.computeKnownBits(A, &I);
5039 // (X & NegY) spred X --> (X & NegY) upred X
5040 if (KnownY.isNegative())
5041 return new ICmpInst(ICmpInst::getUnsignedPredicate(Pred), Op0, Op1);
5042
5043 if (Pred != ICmpInst::ICMP_SLE && Pred != ICmpInst::ICMP_SGT)
5044 return nullptr;
5045
5046 if (KnownY.isNonNegative())
5047 // (X & PosY) s<= X --> X s>= 0
5048 // (X & PosY) s> X --> X s< 0
5049 return new ICmpInst(ICmpInst::getSwappedPredicate(Pred), Op1,
5050 Constant::getNullValue(Op1->getType()));
5051
5053 // (NegX & Y) s<= NegX --> Y s< 0
5054 // (NegX & Y) s> NegX --> Y s>= 0
5056 Constant::getNullValue(A->getType()));
5057
5058 return nullptr;
5059}
5060
5062 InstCombinerImpl &IC) {
5063 Value *Op0 = I.getOperand(0), *Op1 = I.getOperand(1), *A;
5064
5065 // Normalize or operand as operand 0.
5066 CmpInst::Predicate Pred = I.getPredicate();
5067 if (match(Op1, m_c_Or(m_Specific(Op0), m_Value(A)))) {
5068 std::swap(Op0, Op1);
5069 Pred = ICmpInst::getSwappedPredicate(Pred);
5070 } else if (!match(Op0, m_c_Or(m_Specific(Op1), m_Value(A)))) {
5071 return nullptr;
5072 }
5073
5074 // icmp (X | Y) u<= X --> (X | Y) == X
5075 if (Pred == ICmpInst::ICMP_ULE)
5076 return new ICmpInst(ICmpInst::ICMP_EQ, Op0, Op1);
5077
5078 // icmp (X | Y) u> X --> (X | Y) != X
5079 if (Pred == ICmpInst::ICMP_UGT)
5080 return new ICmpInst(ICmpInst::ICMP_NE, Op0, Op1);
5081
5082 if (ICmpInst::isEquality(Pred) && Op0->hasOneUse()) {
5083 // icmp (X | Y) eq/ne Y --> (X & ~Y) eq/ne 0 if Y is freely invertible
5084 if (Value *NotOp1 = IC.getFreelyInverted(
5085 Op1, !isa<Constant>(Op1) && !Op1->hasNUsesOrMore(3), &IC.Builder))
5086 return new ICmpInst(Pred, IC.Builder.CreateAnd(A, NotOp1),
5087 Constant::getNullValue(Op1->getType()));
5088 // icmp (X | Y) eq/ne Y --> (~X | Y) eq/ne -1 if X is freely invertible.
5089 if (Value *NotA = IC.getFreelyInverted(A, A->hasOneUse(), &IC.Builder))
5090 return new ICmpInst(Pred, IC.Builder.CreateOr(Op1, NotA),
5091 Constant::getAllOnesValue(Op1->getType()));
5092 }
5093 return nullptr;
5094}
5095
5097 InstCombinerImpl &IC) {
5098 Value *Op0 = I.getOperand(0), *Op1 = I.getOperand(1), *A;
5099 // Normalize xor operand as operand 0.
5100 CmpInst::Predicate Pred = I.getPredicate();
5101 if (match(Op1, m_c_Xor(m_Specific(Op0), m_Value()))) {
5102 std::swap(Op0, Op1);
5103 Pred = ICmpInst::getSwappedPredicate(Pred);
5104 }
5105 if (!match(Op0, m_c_Xor(m_Specific(Op1), m_Value(A))))
5106 return nullptr;
5107
5108 // icmp (X ^ Y_NonZero) u>= X --> icmp (X ^ Y_NonZero) u> X
5109 // icmp (X ^ Y_NonZero) u<= X --> icmp (X ^ Y_NonZero) u< X
5110 // icmp (X ^ Y_NonZero) s>= X --> icmp (X ^ Y_NonZero) s> X
5111 // icmp (X ^ Y_NonZero) s<= X --> icmp (X ^ Y_NonZero) s< X
5113 if (PredOut != Pred && isKnownNonZero(A, Q))
5114 return new ICmpInst(PredOut, Op0, Op1);
5115
5116 // These transform work when A is negative.
5117 // X s< X^A, X s<= X^A, X u> X^A, X u>= X^A --> X s< 0
5118 // X s> X^A, X s>= X^A, X u< X^A, X u<= X^A --> X s>= 0
5119 if (match(A, m_Negative())) {
5120 CmpInst::Predicate NewPred;
5121 switch (ICmpInst::getStrictPredicate(Pred)) {
5122 default:
5123 return nullptr;
5124 case ICmpInst::ICMP_SLT:
5125 case ICmpInst::ICMP_UGT:
5126 NewPred = ICmpInst::ICMP_SLT;
5127 break;
5128 case ICmpInst::ICMP_SGT:
5129 case ICmpInst::ICMP_ULT:
5130 NewPred = ICmpInst::ICMP_SGE;
5131 break;
5132 }
5133 Constant *Const = Constant::getNullValue(Op0->getType());
5134 return new ICmpInst(NewPred, Op0, Const);
5135 }
5136
5137 return nullptr;
5138}
5139
5140/// Return true if X is a multiple of C.
5141/// TODO: Handle non-power-of-2 factors.
5142static bool isMultipleOf(Value *X, const APInt &C, const SimplifyQuery &Q) {
5143 if (C.isOne())
5144 return true;
5145
5146 if (!C.isPowerOf2())
5147 return false;
5148
5149 return MaskedValueIsZero(X, C - 1, Q);
5150}
5151
5152/// Try to fold icmp (binop), X or icmp X, (binop).
5153/// TODO: A large part of this logic is duplicated in InstSimplify's
5154/// simplifyICmpWithBinOp(). We should be able to share that and avoid the code
5155/// duplication.
5157 const SimplifyQuery &SQ) {
5158 const SimplifyQuery Q = SQ.getWithInstruction(&I);
5159 Value *Op0 = I.getOperand(0), *Op1 = I.getOperand(1);
5160
5161 // Special logic for binary operators.
5164 if (!BO0 && !BO1)
5165 return nullptr;
5166
5167 if (Instruction *NewICmp = foldICmpXNegX(I, Builder))
5168 return NewICmp;
5169
5170 const CmpInst::Predicate Pred = I.getPredicate();
5171 Value *X;
5172
5173 // Convert add-with-unsigned-overflow comparisons into a 'not' with compare.
5174 // (Op1 + X) u</u>= Op1 --> ~Op1 u</u>= X
5175 if (match(Op0, m_OneUse(m_c_Add(m_Specific(Op1), m_Value(X)))) &&
5176 (Pred == ICmpInst::ICMP_ULT || Pred == ICmpInst::ICMP_UGE))
5177 return new ICmpInst(Pred, Builder.CreateNot(Op1), X);
5178 // Op0 u>/u<= (Op0 + X) --> X u>/u<= ~Op0
5179 if (match(Op1, m_OneUse(m_c_Add(m_Specific(Op0), m_Value(X)))) &&
5180 (Pred == ICmpInst::ICMP_UGT || Pred == ICmpInst::ICMP_ULE))
5181 return new ICmpInst(Pred, X, Builder.CreateNot(Op0));
5182
5183 {
5184 // (Op1 + X) + C u</u>= Op1 --> ~C - X u</u>= Op1
5185 Constant *C;
5186 if (match(Op0, m_OneUse(m_Add(m_c_Add(m_Specific(Op1), m_Value(X)),
5187 m_ImmConstant(C)))) &&
5188 (Pred == ICmpInst::ICMP_ULT || Pred == ICmpInst::ICMP_UGE)) {
5190 return new ICmpInst(Pred, Builder.CreateSub(C2, X), Op1);
5191 }
5192 // Op0 u>/u<= (Op0 + X) + C --> Op0 u>/u<= ~C - X
5193 if (match(Op1, m_OneUse(m_Add(m_c_Add(m_Specific(Op0), m_Value(X)),
5194 m_ImmConstant(C)))) &&
5195 (Pred == ICmpInst::ICMP_UGT || Pred == ICmpInst::ICMP_ULE)) {
5197 return new ICmpInst(Pred, Op0, Builder.CreateSub(C2, X));
5198 }
5199 }
5200
5201 // (icmp eq/ne (X, -P2), INT_MIN)
5202 // -> (icmp slt/sge X, INT_MIN + P2)
5203 if (ICmpInst::isEquality(Pred) && BO0 &&
5204 match(I.getOperand(1), m_SignMask()) &&
5206 // Will Constant fold.
5207 Value *NewC = Builder.CreateSub(I.getOperand(1), BO0->getOperand(1));
5208 return new ICmpInst(Pred == ICmpInst::ICMP_EQ ? ICmpInst::ICMP_SLT
5210 BO0->getOperand(0), NewC);
5211 }
5212
5213 {
5214 // Similar to above: an unsigned overflow comparison may use offset + mask:
5215 // ((Op1 + C) & C) u< Op1 --> Op1 != 0
5216 // ((Op1 + C) & C) u>= Op1 --> Op1 == 0
5217 // Op0 u> ((Op0 + C) & C) --> Op0 != 0
5218 // Op0 u<= ((Op0 + C) & C) --> Op0 == 0
5219 BinaryOperator *BO;
5220 const APInt *C;
5221 if ((Pred == ICmpInst::ICMP_ULT || Pred == ICmpInst::ICMP_UGE) &&
5222 match(Op0, m_And(m_BinOp(BO), m_LowBitMask(C))) &&
5224 CmpInst::Predicate NewPred =
5226 Constant *Zero = ConstantInt::getNullValue(Op1->getType());
5227 return new ICmpInst(NewPred, Op1, Zero);
5228 }
5229
5230 if ((Pred == ICmpInst::ICMP_UGT || Pred == ICmpInst::ICMP_ULE) &&
5231 match(Op1, m_And(m_BinOp(BO), m_LowBitMask(C))) &&
5233 CmpInst::Predicate NewPred =
5235 Constant *Zero = ConstantInt::getNullValue(Op1->getType());
5236 return new ICmpInst(NewPred, Op0, Zero);
5237 }
5238 }
5239
5240 bool NoOp0WrapProblem = false, NoOp1WrapProblem = false;
5241 bool Op0HasNUW = false, Op1HasNUW = false;
5242 bool Op0HasNSW = false, Op1HasNSW = false;
5243 // Analyze the case when either Op0 or Op1 is an add instruction.
5244 // Op0 = A + B (or A and B are null); Op1 = C + D (or C and D are null).
5245 auto hasNoWrapProblem = [](const BinaryOperator &BO, CmpInst::Predicate Pred,
5246 bool &HasNSW, bool &HasNUW) -> bool {
5248 HasNUW = BO.hasNoUnsignedWrap();
5249 HasNSW = BO.hasNoSignedWrap();
5250 return ICmpInst::isEquality(Pred) ||
5251 (CmpInst::isUnsigned(Pred) && HasNUW) ||
5252 (CmpInst::isSigned(Pred) && HasNSW);
5253 } else if (BO.getOpcode() == Instruction::Or) {
5254 HasNUW = true;
5255 HasNSW = true;
5256 return true;
5257 } else {
5258 return false;
5259 }
5260 };
5261 Value *A = nullptr, *B = nullptr, *C = nullptr, *D = nullptr;
5262
5263 if (BO0) {
5264 match(BO0, m_AddLike(m_Value(A), m_Value(B)));
5265 NoOp0WrapProblem = hasNoWrapProblem(*BO0, Pred, Op0HasNSW, Op0HasNUW);
5266 }
5267 if (BO1) {
5268 match(BO1, m_AddLike(m_Value(C), m_Value(D)));
5269 NoOp1WrapProblem = hasNoWrapProblem(*BO1, Pred, Op1HasNSW, Op1HasNUW);
5270 }
5271
5272 // icmp (A+B), A -> icmp B, 0 for equalities or if there is no overflow.
5273 // icmp (A+B), B -> icmp A, 0 for equalities or if there is no overflow.
5274 if ((A == Op1 || B == Op1) && NoOp0WrapProblem)
5275 return new ICmpInst(Pred, A == Op1 ? B : A,
5276 Constant::getNullValue(Op1->getType()));
5277
5278 // icmp C, (C+D) -> icmp 0, D for equalities or if there is no overflow.
5279 // icmp D, (C+D) -> icmp 0, C for equalities or if there is no overflow.
5280 if ((C == Op0 || D == Op0) && NoOp1WrapProblem)
5281 return new ICmpInst(Pred, Constant::getNullValue(Op0->getType()),
5282 C == Op0 ? D : C);
5283
5284 // icmp (A+B), (A+D) -> icmp B, D for equalities or if there is no overflow.
5285 if (A && C && (A == C || A == D || B == C || B == D) && NoOp0WrapProblem &&
5286 NoOp1WrapProblem) {
5287 // Determine Y and Z in the form icmp (X+Y), (X+Z).
5288 Value *Y, *Z;
5289 if (A == C) {
5290 // C + B == C + D -> B == D
5291 Y = B;
5292 Z = D;
5293 } else if (A == D) {
5294 // D + B == C + D -> B == C
5295 Y = B;
5296 Z = C;
5297 } else if (B == C) {
5298 // A + C == C + D -> A == D
5299 Y = A;
5300 Z = D;
5301 } else {
5302 assert(B == D);
5303 // A + D == C + D -> A == C
5304 Y = A;
5305 Z = C;
5306 }
5307 return new ICmpInst(Pred, Y, Z);
5308 }
5309
5310 if (ICmpInst::isRelational(Pred)) {
5311 // Return if both X and Y is divisible by Z/-Z.
5312 // TODO: Generalize to check if (X - Y) is divisible by Z/-Z.
5313 auto ShareCommonDivisor = [&Q](Value *X, Value *Y, Value *Z,
5314 bool IsNegative) -> bool {
5315 const APInt *OffsetC;
5316 if (!match(Z, m_APInt(OffsetC)))
5317 return false;
5318
5319 // Fast path for Z == 1/-1.
5320 if (IsNegative ? OffsetC->isAllOnes() : OffsetC->isOne())
5321 return true;
5322
5323 APInt C = *OffsetC;
5324 if (IsNegative)
5325 C.negate();
5326 // Note: -INT_MIN is also negative.
5327 if (!C.isStrictlyPositive())
5328 return false;
5329
5330 return isMultipleOf(X, C, Q) && isMultipleOf(Y, C, Q);
5331 };
5332
5333 // TODO: The subtraction-related identities shown below also hold, but
5334 // canonicalization from (X -nuw 1) to (X + -1) means that the combinations
5335 // wouldn't happen even if they were implemented.
5336 //
5337 // icmp ult (A - 1), Op1 -> icmp ule A, Op1
5338 // icmp uge (A - 1), Op1 -> icmp ugt A, Op1
5339 // icmp ugt Op0, (C - 1) -> icmp uge Op0, C
5340 // icmp ule Op0, (C - 1) -> icmp ult Op0, C
5341
5342 // icmp slt (A + -1), Op1 -> icmp sle A, Op1
5343 // icmp sge (A + -1), Op1 -> icmp sgt A, Op1
5344 // icmp sle (A + 1), Op1 -> icmp slt A, Op1
5345 // icmp sgt (A + 1), Op1 -> icmp sge A, Op1
5346 // icmp ule (A + 1), Op0 -> icmp ult A, Op1
5347 // icmp ugt (A + 1), Op0 -> icmp uge A, Op1
5348 if (A && NoOp0WrapProblem &&
5349 ShareCommonDivisor(A, Op1, B,
5350 ICmpInst::isLT(Pred) || ICmpInst::isGE(Pred)))
5352 Op1);
5353
5354 // icmp sgt Op0, (C + -1) -> icmp sge Op0, C
5355 // icmp sle Op0, (C + -1) -> icmp slt Op0, C
5356 // icmp sge Op0, (C + 1) -> icmp sgt Op0, C
5357 // icmp slt Op0, (C + 1) -> icmp sle Op0, C
5358 // icmp uge Op0, (C + 1) -> icmp ugt Op0, C
5359 // icmp ult Op0, (C + 1) -> icmp ule Op0, C
5360 if (C && NoOp1WrapProblem &&
5361 ShareCommonDivisor(Op0, C, D,
5362 ICmpInst::isGT(Pred) || ICmpInst::isLE(Pred)))
5364 C);
5365 }
5366
5367 // if C1 has greater magnitude than C2:
5368 // icmp (A + C1), (C + C2) -> icmp (A + C3), C
5369 // s.t. C3 = C1 - C2
5370 //
5371 // if C2 has greater magnitude than C1:
5372 // icmp (A + C1), (C + C2) -> icmp A, (C + C3)
5373 // s.t. C3 = C2 - C1
5374 if (A && C && NoOp0WrapProblem && NoOp1WrapProblem &&
5375 (BO0->hasOneUse() || BO1->hasOneUse()) && !I.isUnsigned()) {
5376 const APInt *AP1, *AP2;
5377 // TODO: Support non-uniform vectors.
5378 // TODO: Allow poison passthrough if B or D's element is poison.
5379 if (match(B, m_APIntAllowPoison(AP1)) &&
5380 match(D, m_APIntAllowPoison(AP2)) &&
5381 AP1->isNegative() == AP2->isNegative()) {
5382 APInt AP1Abs = AP1->abs();
5383 APInt AP2Abs = AP2->abs();
5384 if (AP1Abs.uge(AP2Abs)) {
5385 APInt Diff = *AP1 - *AP2;
5386 Constant *C3 = Constant::getIntegerValue(BO0->getType(), Diff);
5387 Value *NewAdd = Builder.CreateAdd(
5388 A, C3, "", Op0HasNUW && Diff.ule(*AP1), Op0HasNSW);
5389 return new ICmpInst(Pred, NewAdd, C);
5390 } else {
5391 APInt Diff = *AP2 - *AP1;
5392 Constant *C3 = Constant::getIntegerValue(BO0->getType(), Diff);
5393 Value *NewAdd = Builder.CreateAdd(
5394 C, C3, "", Op1HasNUW && Diff.ule(*AP2), Op1HasNSW);
5395 return new ICmpInst(Pred, A, NewAdd);
5396 }
5397 }
5398 Constant *Cst1, *Cst2;
5399 if (match(B, m_ImmConstant(Cst1)) && match(D, m_ImmConstant(Cst2)) &&
5400 ICmpInst::isEquality(Pred)) {
5401 Constant *Diff = ConstantExpr::getSub(Cst2, Cst1);
5402 Value *NewAdd = Builder.CreateAdd(C, Diff);
5403 return new ICmpInst(Pred, A, NewAdd);
5404 }
5405 }
5406
5407 // Analyze the case when either Op0 or Op1 is a sub instruction.
5408 // Op0 = A - B (or A and B are null); Op1 = C - D (or C and D are null).
5409 A = nullptr;
5410 B = nullptr;
5411 C = nullptr;
5412 D = nullptr;
5413 if (BO0 && BO0->getOpcode() == Instruction::Sub) {
5414 A = BO0->getOperand(0);
5415 B = BO0->getOperand(1);
5416 }
5417 if (BO1 && BO1->getOpcode() == Instruction::Sub) {
5418 C = BO1->getOperand(0);
5419 D = BO1->getOperand(1);
5420 }
5421
5422 // icmp (A-B), A -> icmp 0, B for equalities or if there is no overflow.
5423 if (A == Op1 && NoOp0WrapProblem)
5424 return new ICmpInst(Pred, Constant::getNullValue(Op1->getType()), B);
5425 // icmp C, (C-D) -> icmp D, 0 for equalities or if there is no overflow.
5426 if (C == Op0 && NoOp1WrapProblem)
5427 return new ICmpInst(Pred, D, Constant::getNullValue(Op0->getType()));
5428
5429 // Convert sub-with-unsigned-overflow comparisons into a comparison of args.
5430 // (A - B) u>/u<= A --> B u>/u<= A
5431 if (A == Op1 && (Pred == ICmpInst::ICMP_UGT || Pred == ICmpInst::ICMP_ULE))
5432 return new ICmpInst(Pred, B, A);
5433 // C u</u>= (C - D) --> C u</u>= D
5434 if (C == Op0 && (Pred == ICmpInst::ICMP_ULT || Pred == ICmpInst::ICMP_UGE))
5435 return new ICmpInst(Pred, C, D);
5436 // (A - B) u>=/u< A --> B u>/u<= A iff B != 0
5437 if (A == Op1 && (Pred == ICmpInst::ICMP_UGE || Pred == ICmpInst::ICMP_ULT) &&
5438 isKnownNonZero(B, Q))
5440 // C u<=/u> (C - D) --> C u</u>= D iff B != 0
5441 if (C == Op0 && (Pred == ICmpInst::ICMP_ULE || Pred == ICmpInst::ICMP_UGT) &&
5442 isKnownNonZero(D, Q))
5444
5445 // icmp (A-B), (C-B) -> icmp A, C for equalities or if there is no overflow.
5446 if (B && D && B == D && NoOp0WrapProblem && NoOp1WrapProblem)
5447 return new ICmpInst(Pred, A, C);
5448
5449 // icmp (A-B), (A-D) -> icmp D, B for equalities or if there is no overflow.
5450 if (A && C && A == C && NoOp0WrapProblem && NoOp1WrapProblem)
5451 return new ICmpInst(Pred, D, B);
5452
5453 // icmp (0-X) < cst --> x > -cst
5454 if (NoOp0WrapProblem && ICmpInst::isSigned(Pred)) {
5455 Value *X;
5456 if (match(BO0, m_Neg(m_Value(X))))
5457 if (Constant *RHSC = dyn_cast<Constant>(Op1))
5458 if (RHSC->isNotMinSignedValue())
5459 return new ICmpInst(I.getSwappedPredicate(), X,
5460 ConstantExpr::getNeg(RHSC));
5461 }
5462
5463 if (Instruction *R = foldICmpXorXX(I, Q, *this))
5464 return R;
5465 if (Instruction *R = foldICmpOrXX(I, Q, *this))
5466 return R;
5467
5468 {
5469 // Try to remove shared multiplier from comparison:
5470 // X * Z pred Y * Z
5471 Value *X, *Y, *Z;
5472 if ((match(Op0, m_Mul(m_Value(X), m_Value(Z))) &&
5473 match(Op1, m_c_Mul(m_Specific(Z), m_Value(Y)))) ||
5474 (match(Op0, m_Mul(m_Value(Z), m_Value(X))) &&
5475 match(Op1, m_c_Mul(m_Specific(Z), m_Value(Y))))) {
5476 if (ICmpInst::isSigned(Pred)) {
5477 if (Op0HasNSW && Op1HasNSW) {
5478 KnownBits ZKnown = computeKnownBits(Z, &I);
5479 if (ZKnown.isStrictlyPositive())
5480 return new ICmpInst(Pred, X, Y);
5481 if (ZKnown.isNegative())
5482 return new ICmpInst(ICmpInst::getSwappedPredicate(Pred), X, Y);
5484 SQ.getWithInstruction(&I));
5485 if (LessThan && match(LessThan, m_One()))
5486 return new ICmpInst(ICmpInst::getSwappedPredicate(Pred), Z,
5487 Constant::getNullValue(Z->getType()));
5488 Value *GreaterThan = simplifyICmpInst(ICmpInst::ICMP_SGT, X, Y,
5489 SQ.getWithInstruction(&I));
5490 if (GreaterThan && match(GreaterThan, m_One()))
5491 return new ICmpInst(Pred, Z, Constant::getNullValue(Z->getType()));
5492 }
5493 } else {
5494 bool NonZero;
5495 if (ICmpInst::isEquality(Pred)) {
5496 // If X != Y, fold (X *nw Z) eq/ne (Y *nw Z) -> Z eq/ne 0
5497 if (((Op0HasNSW && Op1HasNSW) || (Op0HasNUW && Op1HasNUW)) &&
5498 isKnownNonEqual(X, Y, SQ))
5499 return new ICmpInst(Pred, Z, Constant::getNullValue(Z->getType()));
5500
5501 KnownBits ZKnown = computeKnownBits(Z, &I);
5502 // if Z % 2 != 0
5503 // X * Z eq/ne Y * Z -> X eq/ne Y
5504 if (ZKnown.countMaxTrailingZeros() == 0)
5505 return new ICmpInst(Pred, X, Y);
5506 NonZero = !ZKnown.One.isZero() || isKnownNonZero(Z, Q);
5507 // if Z != 0 and nsw(X * Z) and nsw(Y * Z)
5508 // X * Z eq/ne Y * Z -> X eq/ne Y
5509 if (NonZero && BO0 && BO1 && Op0HasNSW && Op1HasNSW)
5510 return new ICmpInst(Pred, X, Y);
5511 } else
5512 NonZero = isKnownNonZero(Z, Q);
5513
5514 // If Z != 0 and nuw(X * Z) and nuw(Y * Z)
5515 // X * Z u{lt/le/gt/ge}/eq/ne Y * Z -> X u{lt/le/gt/ge}/eq/ne Y
5516 if (NonZero && BO0 && BO1 && Op0HasNUW && Op1HasNUW)
5517 return new ICmpInst(Pred, X, Y);
5518 }
5519 }
5520 }
5521
5522 BinaryOperator *SRem = nullptr;
5523 // icmp (srem X, Y), Y
5524 if (BO0 && BO0->getOpcode() == Instruction::SRem && Op1 == BO0->getOperand(1))
5525 SRem = BO0;
5526 // icmp Y, (srem X, Y)
5527 else if (BO1 && BO1->getOpcode() == Instruction::SRem &&
5528 Op0 == BO1->getOperand(1))
5529 SRem = BO1;
5530 if (SRem) {
5531 // We don't check hasOneUse to avoid increasing register pressure because
5532 // the value we use is the same value this instruction was already using.
5533 switch (SRem == BO0 ? ICmpInst::getSwappedPredicate(Pred) : Pred) {
5534 default:
5535 break;
5536 case ICmpInst::ICMP_EQ:
5537 return replaceInstUsesWith(I, ConstantInt::getFalse(I.getType()));
5538 case ICmpInst::ICMP_NE:
5539 return replaceInstUsesWith(I, ConstantInt::getTrue(I.getType()));
5540 case ICmpInst::ICMP_SGT:
5541 case ICmpInst::ICMP_SGE:
5542 return new ICmpInst(ICmpInst::ICMP_SGT, SRem->getOperand(1),
5544 case ICmpInst::ICMP_SLT:
5545 case ICmpInst::ICMP_SLE:
5546 return new ICmpInst(ICmpInst::ICMP_SLT, SRem->getOperand(1),
5548 }
5549 }
5550
5551 if (BO0 && BO1 && BO0->getOpcode() == BO1->getOpcode() &&
5552 (BO0->hasOneUse() || BO1->hasOneUse()) &&
5553 BO0->getOperand(1) == BO1->getOperand(1)) {
5554 switch (BO0->getOpcode()) {
5555 default:
5556 break;
5557 case Instruction::Add:
5558 case Instruction::Sub:
5559 case Instruction::Xor: {
5560 if (I.isEquality()) // a+x icmp eq/ne b+x --> a icmp b
5561 return new ICmpInst(Pred, BO0->getOperand(0), BO1->getOperand(0));
5562
5563 const APInt *C;
5564 if (match(BO0->getOperand(1), m_APInt(C))) {
5565 // icmp u/s (a ^ signmask), (b ^ signmask) --> icmp s/u a, b
5566 if (C->isSignMask()) {
5567 ICmpInst::Predicate NewPred = I.getFlippedSignednessPredicate();
5568 return new ICmpInst(NewPred, BO0->getOperand(0), BO1->getOperand(0));
5569 }
5570
5571 // icmp u/s (a ^ maxsignval), (b ^ maxsignval) --> icmp s/u' a, b
5572 if (BO0->getOpcode() == Instruction::Xor && C->isMaxSignedValue()) {
5573 ICmpInst::Predicate NewPred = I.getFlippedSignednessPredicate();
5574 NewPred = I.getSwappedPredicate(NewPred);
5575 return new ICmpInst(NewPred, BO0->getOperand(0), BO1->getOperand(0));
5576 }
5577 }
5578 break;
5579 }
5580 case Instruction::Mul: {
5581 if (!I.isEquality())
5582 break;
5583
5584 const APInt *C;
5585 if (match(BO0->getOperand(1), m_APInt(C)) && !C->isZero() &&
5586 !C->isOne()) {
5587 // icmp eq/ne (X * C), (Y * C) --> icmp (X & Mask), (Y & Mask)
5588 // Mask = -1 >> count-trailing-zeros(C).
5589 if (unsigned TZs = C->countr_zero()) {
5590 Constant *Mask = ConstantInt::get(
5591 BO0->getType(),
5592 APInt::getLowBitsSet(C->getBitWidth(), C->getBitWidth() - TZs));
5593 Value *And1 = Builder.CreateAnd(BO0->getOperand(0), Mask);
5594 Value *And2 = Builder.CreateAnd(BO1->getOperand(0), Mask);
5595 return new ICmpInst(Pred, And1, And2);
5596 }
5597 }
5598 break;
5599 }
5600 case Instruction::UDiv:
5601 case Instruction::LShr:
5602 if (I.isSigned() || !BO0->isExact() || !BO1->isExact())
5603 break;
5604 return new ICmpInst(Pred, BO0->getOperand(0), BO1->getOperand(0));
5605
5606 case Instruction::SDiv:
5607 if (!(I.isEquality() || match(BO0->getOperand(1), m_NonNegative())) ||
5608 !BO0->isExact() || !BO1->isExact())
5609 break;
5610 return new ICmpInst(Pred, BO0->getOperand(0), BO1->getOperand(0));
5611
5612 case Instruction::AShr:
5613 if (!BO0->isExact() || !BO1->isExact())
5614 break;
5615 return new ICmpInst(Pred, BO0->getOperand(0), BO1->getOperand(0));
5616
5617 case Instruction::Shl: {
5618 bool NUW = Op0HasNUW && Op1HasNUW;
5619 bool NSW = Op0HasNSW && Op1HasNSW;
5620 if (!NUW && !NSW)
5621 break;
5622 if (!NSW && I.isSigned())
5623 break;
5624 return new ICmpInst(Pred, BO0->getOperand(0), BO1->getOperand(0));
5625 }
5626 }
5627 }
5628
5629 if (BO0) {
5630 // Transform A & (L - 1) `ult` L --> L != 0
5631 auto LSubOne = m_Add(m_Specific(Op1), m_AllOnes());
5632 auto BitwiseAnd = m_c_And(m_Value(), LSubOne);
5633
5634 if (match(BO0, BitwiseAnd) && Pred == ICmpInst::ICMP_ULT) {
5635 auto *Zero = Constant::getNullValue(BO0->getType());
5636 return new ICmpInst(ICmpInst::ICMP_NE, Op1, Zero);
5637 }
5638 }
5639
5640 // For unsigned predicates / eq / ne:
5641 // icmp pred (x << 1), x --> icmp getSignedPredicate(pred) x, 0
5642 // icmp pred x, (x << 1) --> icmp getSignedPredicate(pred) 0, x
5643 if (!ICmpInst::isSigned(Pred)) {
5644 if (match(Op0, m_Shl(m_Specific(Op1), m_One())))
5645 return new ICmpInst(ICmpInst::getSignedPredicate(Pred), Op1,
5646 Constant::getNullValue(Op1->getType()));
5647 else if (match(Op1, m_Shl(m_Specific(Op0), m_One())))
5648 return new ICmpInst(ICmpInst::getSignedPredicate(Pred),
5649 Constant::getNullValue(Op0->getType()), Op0);
5650 }
5651
5653 return replaceInstUsesWith(I, V);
5654
5655 if (Instruction *R = foldICmpAndXX(I, Q, *this))
5656 return R;
5657
5659 return replaceInstUsesWith(I, V);
5660
5662 return replaceInstUsesWith(I, V);
5663
5664 return nullptr;
5665}
5666
5667/// Fold icmp Pred min|max(X, Y), Z.
5670 Value *Z, CmpPredicate Pred) {
5671 Value *X = MinMax->getLHS();
5672 Value *Y = MinMax->getRHS();
5673 if (ICmpInst::isSigned(Pred) && !MinMax->isSigned())
5674 return nullptr;
5675 if (ICmpInst::isUnsigned(Pred) && MinMax->isSigned()) {
5676 // Revert the transform signed pred -> unsigned pred
5677 // TODO: We can flip the signedness of predicate if both operands of icmp
5678 // are negative.
5679 if (isKnownNonNegative(Z, SQ.getWithInstruction(&I)) &&
5680 isKnownNonNegative(MinMax, SQ.getWithInstruction(&I))) {
5682 } else
5683 return nullptr;
5684 }
5685 SimplifyQuery Q = SQ.getWithInstruction(&I);
5686 auto IsCondKnownTrue = [](Value *Val) -> std::optional<bool> {
5687 if (!Val)
5688 return std::nullopt;
5689 if (match(Val, m_One()))
5690 return true;
5691 if (match(Val, m_Zero()))
5692 return false;
5693 return std::nullopt;
5694 };
5695 // Remove samesign here since it is illegal to keep it when we speculatively
5696 // execute comparisons. For example, `icmp samesign ult umax(X, -46), -32`
5697 // cannot be decomposed into `(icmp samesign ult X, -46) or (icmp samesign ult
5698 // -46, -32)`. `X` is allowed to be non-negative here.
5699 Pred = Pred.dropSameSign();
5700 auto CmpXZ = IsCondKnownTrue(simplifyICmpInst(Pred, X, Z, Q));
5701 auto CmpYZ = IsCondKnownTrue(simplifyICmpInst(Pred, Y, Z, Q));
5702 if (!CmpXZ.has_value() && !CmpYZ.has_value())
5703 return nullptr;
5704 if (!CmpXZ.has_value()) {
5705 std::swap(X, Y);
5706 std::swap(CmpXZ, CmpYZ);
5707 }
5708
5709 auto FoldIntoCmpYZ = [&]() -> Instruction * {
5710 if (CmpYZ.has_value())
5711 return replaceInstUsesWith(I, ConstantInt::getBool(I.getType(), *CmpYZ));
5712 return ICmpInst::Create(Instruction::ICmp, Pred, Y, Z);
5713 };
5714
5715 switch (Pred) {
5716 case ICmpInst::ICMP_EQ:
5717 case ICmpInst::ICMP_NE: {
5718 // If X == Z:
5719 // Expr Result
5720 // min(X, Y) == Z X <= Y
5721 // max(X, Y) == Z X >= Y
5722 // min(X, Y) != Z X > Y
5723 // max(X, Y) != Z X < Y
5724 if ((Pred == ICmpInst::ICMP_EQ) == *CmpXZ) {
5725 ICmpInst::Predicate NewPred =
5726 ICmpInst::getNonStrictPredicate(MinMax->getPredicate());
5727 if (Pred == ICmpInst::ICMP_NE)
5728 NewPred = ICmpInst::getInversePredicate(NewPred);
5729 return ICmpInst::Create(Instruction::ICmp, NewPred, X, Y);
5730 }
5731 // Otherwise (X != Z):
5732 ICmpInst::Predicate NewPred = MinMax->getPredicate();
5733 auto MinMaxCmpXZ = IsCondKnownTrue(simplifyICmpInst(NewPred, X, Z, Q));
5734 if (!MinMaxCmpXZ.has_value()) {
5735 std::swap(X, Y);
5736 std::swap(CmpXZ, CmpYZ);
5737 // Re-check pre-condition X != Z
5738 if (!CmpXZ.has_value() || (Pred == ICmpInst::ICMP_EQ) == *CmpXZ)
5739 break;
5740 MinMaxCmpXZ = IsCondKnownTrue(simplifyICmpInst(NewPred, X, Z, Q));
5741 }
5742 if (!MinMaxCmpXZ.has_value())
5743 break;
5744 if (*MinMaxCmpXZ) {
5745 // Expr Fact Result
5746 // min(X, Y) == Z X < Z false
5747 // max(X, Y) == Z X > Z false
5748 // min(X, Y) != Z X < Z true
5749 // max(X, Y) != Z X > Z true
5750 return replaceInstUsesWith(
5751 I, ConstantInt::getBool(I.getType(), Pred == ICmpInst::ICMP_NE));
5752 } else {
5753 // Expr Fact Result
5754 // min(X, Y) == Z X > Z Y == Z
5755 // max(X, Y) == Z X < Z Y == Z
5756 // min(X, Y) != Z X > Z Y != Z
5757 // max(X, Y) != Z X < Z Y != Z
5758 return FoldIntoCmpYZ();
5759 }
5760 break;
5761 }
5762 case ICmpInst::ICMP_SLT:
5763 case ICmpInst::ICMP_ULT:
5764 case ICmpInst::ICMP_SLE:
5765 case ICmpInst::ICMP_ULE:
5766 case ICmpInst::ICMP_SGT:
5767 case ICmpInst::ICMP_UGT:
5768 case ICmpInst::ICMP_SGE:
5769 case ICmpInst::ICMP_UGE: {
5770 bool IsSame = MinMax->getPredicate() == ICmpInst::getStrictPredicate(Pred);
5771 if (*CmpXZ) {
5772 if (IsSame) {
5773 // Expr Fact Result
5774 // min(X, Y) < Z X < Z true
5775 // min(X, Y) <= Z X <= Z true
5776 // max(X, Y) > Z X > Z true
5777 // max(X, Y) >= Z X >= Z true
5778 return replaceInstUsesWith(I, ConstantInt::getTrue(I.getType()));
5779 } else {
5780 // Expr Fact Result
5781 // max(X, Y) < Z X < Z Y < Z
5782 // max(X, Y) <= Z X <= Z Y <= Z
5783 // min(X, Y) > Z X > Z Y > Z
5784 // min(X, Y) >= Z X >= Z Y >= Z
5785 return FoldIntoCmpYZ();
5786 }
5787 } else {
5788 if (IsSame) {
5789 // Expr Fact Result
5790 // min(X, Y) < Z X >= Z Y < Z
5791 // min(X, Y) <= Z X > Z Y <= Z
5792 // max(X, Y) > Z X <= Z Y > Z
5793 // max(X, Y) >= Z X < Z Y >= Z
5794 return FoldIntoCmpYZ();
5795 } else {
5796 // Expr Fact Result
5797 // max(X, Y) < Z X >= Z false
5798 // max(X, Y) <= Z X > Z false
5799 // min(X, Y) > Z X <= Z false
5800 // min(X, Y) >= Z X < Z false
5801 return replaceInstUsesWith(I, ConstantInt::getFalse(I.getType()));
5802 }
5803 }
5804 break;
5805 }
5806 default:
5807 break;
5808 }
5809
5810 return nullptr;
5811}
5812
5813/// Match and fold patterns like:
5814/// icmp eq/ne X, min(max(X, Lo), Hi)
5815/// which represents a range check and can be represented as a ConstantRange.
5816///
5817/// For icmp eq, build ConstantRange [Lo, Hi + 1) and convert to:
5818/// (X - Lo) u< (Hi + 1 - Lo)
5819/// For icmp ne, build ConstantRange [Hi + 1, Lo) and convert to:
5820/// (X - (Hi + 1)) u< (Lo - (Hi + 1))
5822 MinMaxIntrinsic *Min) {
5823 if (!I.isEquality() || !Min->hasOneUse() || !Min->isMin())
5824 return nullptr;
5825
5826 const APInt *Lo = nullptr, *Hi = nullptr;
5827 if (Min->isSigned()) {
5828 if (!match(Min->getLHS(), m_OneUse(m_SMax(m_Specific(X), m_APInt(Lo)))) ||
5829 !match(Min->getRHS(), m_APInt(Hi)) || !Lo->slt(*Hi))
5830 return nullptr;
5831 } else {
5832 if (!match(Min->getLHS(), m_OneUse(m_UMax(m_Specific(X), m_APInt(Lo)))) ||
5833 !match(Min->getRHS(), m_APInt(Hi)) || !Lo->ult(*Hi))
5834 return nullptr;
5835 }
5836
5839 APInt C, Offset;
5840 if (I.getPredicate() == ICmpInst::ICMP_EQ)
5841 CR.getEquivalentICmp(Pred, C, Offset);
5842 else
5843 CR.inverse().getEquivalentICmp(Pred, C, Offset);
5844
5845 if (!Offset.isZero())
5846 X = Builder.CreateAdd(X, ConstantInt::get(X->getType(), Offset));
5847
5848 return replaceInstUsesWith(
5849 I, Builder.CreateICmp(Pred, X, ConstantInt::get(X->getType(), C)));
5850}
5851
5852// Canonicalize checking for a power-of-2-or-zero value:
5854 InstCombiner::BuilderTy &Builder) {
5855 Value *Op0 = I.getOperand(0), *Op1 = I.getOperand(1);
5856 const CmpInst::Predicate Pred = I.getPredicate();
5857 Value *A = nullptr;
5858 bool CheckIs;
5859 if (I.isEquality()) {
5860 // (A & (A-1)) == 0 --> ctpop(A) < 2 (two commuted variants)
5861 // ((A-1) & A) != 0 --> ctpop(A) > 1 (two commuted variants)
5862 if (!match(Op0, m_OneUse(m_c_And(m_Add(m_Value(A), m_AllOnes()),
5863 m_Deferred(A)))) ||
5864 !match(Op1, m_ZeroInt()))
5865 A = nullptr;
5866
5867 // (A & -A) == A --> ctpop(A) < 2 (four commuted variants)
5868 // (-A & A) != A --> ctpop(A) > 1 (four commuted variants)
5869 if (match(Op0, m_OneUse(m_c_And(m_Neg(m_Specific(Op1)), m_Specific(Op1)))))
5870 A = Op1;
5871 else if (match(Op1,
5873 A = Op0;
5874
5875 CheckIs = Pred == ICmpInst::ICMP_EQ;
5876 } else if (ICmpInst::isUnsigned(Pred)) {
5877 // (A ^ (A-1)) u>= A --> ctpop(A) < 2 (two commuted variants)
5878 // ((A-1) ^ A) u< A --> ctpop(A) > 1 (two commuted variants)
5879
5880 if ((Pred == ICmpInst::ICMP_UGE || Pred == ICmpInst::ICMP_ULT) &&
5882 m_Specific(Op1))))) {
5883 A = Op1;
5884 CheckIs = Pred == ICmpInst::ICMP_UGE;
5885 } else if ((Pred == ICmpInst::ICMP_UGT || Pred == ICmpInst::ICMP_ULE) &&
5887 m_Specific(Op0))))) {
5888 A = Op0;
5889 CheckIs = Pred == ICmpInst::ICMP_ULE;
5890 }
5891 }
5892
5893 if (A) {
5894 Type *Ty = A->getType();
5895 CallInst *CtPop = Builder.CreateUnaryIntrinsic(Intrinsic::ctpop, A);
5896 return CheckIs ? new ICmpInst(ICmpInst::ICMP_ULT, CtPop,
5897 ConstantInt::get(Ty, 2))
5898 : new ICmpInst(ICmpInst::ICMP_UGT, CtPop,
5899 ConstantInt::get(Ty, 1));
5900 }
5901
5902 return nullptr;
5903}
5904
5905/// Find all possible pairs (BinOp, RHS) that BinOp V, RHS can be simplified.
5906using OffsetOp = std::pair<Instruction::BinaryOps, Value *>;
5908 bool AllowRecursion) {
5910 if (!Inst || !Inst->hasOneUse())
5911 return;
5912
5913 switch (Inst->getOpcode()) {
5914 case Instruction::Add:
5915 Offsets.emplace_back(Instruction::Sub, Inst->getOperand(1));
5916 Offsets.emplace_back(Instruction::Sub, Inst->getOperand(0));
5917 break;
5918 case Instruction::Sub:
5919 Offsets.emplace_back(Instruction::Add, Inst->getOperand(1));
5920 break;
5921 case Instruction::Xor:
5922 Offsets.emplace_back(Instruction::Xor, Inst->getOperand(1));
5923 Offsets.emplace_back(Instruction::Xor, Inst->getOperand(0));
5924 break;
5925 case Instruction::Shl:
5926 if (Inst->hasNoSignedWrap())
5927 Offsets.emplace_back(Instruction::AShr, Inst->getOperand(1));
5928 if (Inst->hasNoUnsignedWrap())
5929 Offsets.emplace_back(Instruction::LShr, Inst->getOperand(1));
5930 break;
5931 case Instruction::Select:
5932 if (AllowRecursion) {
5933 collectOffsetOp(Inst->getOperand(1), Offsets, /*AllowRecursion=*/false);
5934 collectOffsetOp(Inst->getOperand(2), Offsets, /*AllowRecursion=*/false);
5935 }
5936 break;
5937 default:
5938 break;
5939 }
5940}
5941
5943
5948
5950 return {OffsetKind::Invalid, nullptr, nullptr, nullptr, nullptr};
5951 }
5953 return {OffsetKind::Value, V, nullptr, nullptr, nullptr};
5954 }
5955 static OffsetResult select(Value *Cond, Value *TrueV, Value *FalseV,
5957 return {OffsetKind::Select, Cond, TrueV, FalseV, MDFrom};
5958 }
5959 bool isValid() const { return Kind != OffsetKind::Invalid; }
5961 switch (Kind) {
5963 llvm_unreachable("Invalid offset result");
5964 case OffsetKind::Value:
5965 return V0;
5966 case OffsetKind::Select:
5967 return Builder.CreateSelect(
5968 V0, V1, V2, "", ProfcheckDisableMetadataFixes ? nullptr : MDFrom);
5969 }
5970 llvm_unreachable("Unknown OffsetKind enum");
5971 }
5972};
5973
5974/// Offset both sides of an equality icmp to see if we can save some
5975/// instructions: icmp eq/ne X, Y -> icmp eq/ne X op Z, Y op Z.
5976/// Note: This operation should not introduce poison.
5978 InstCombiner::BuilderTy &Builder,
5979 const SimplifyQuery &SQ) {
5980 assert(I.isEquality() && "Expected an equality icmp");
5981 Value *Op0 = I.getOperand(0), *Op1 = I.getOperand(1);
5982 if (!Op0->getType()->isIntOrIntVectorTy())
5983 return nullptr;
5984
5985 SmallVector<OffsetOp, 4> OffsetOps;
5986 collectOffsetOp(Op0, OffsetOps, /*AllowRecursion=*/true);
5987 collectOffsetOp(Op1, OffsetOps, /*AllowRecursion=*/true);
5988
5989 auto ApplyOffsetImpl = [&](Value *V, unsigned BinOpc, Value *RHS) -> Value * {
5990 switch (BinOpc) {
5991 // V = shl nsw X, RHS => X = ashr V, RHS
5992 case Instruction::AShr: {
5993 const APInt *CV, *CRHS;
5994 if (!(match(V, m_APInt(CV)) && match(RHS, m_APInt(CRHS)) &&
5995 CV->ashr(*CRHS).shl(*CRHS) == *CV) &&
5997 return nullptr;
5998 break;
5999 }
6000 // V = shl nuw X, RHS => X = lshr V, RHS
6001 case Instruction::LShr: {
6002 const APInt *CV, *CRHS;
6003 if (!(match(V, m_APInt(CV)) && match(RHS, m_APInt(CRHS)) &&
6004 CV->lshr(*CRHS).shl(*CRHS) == *CV) &&
6006 return nullptr;
6007 break;
6008 }
6009 default:
6010 break;
6011 }
6012
6013 Value *Simplified = simplifyBinOp(BinOpc, V, RHS, SQ);
6014 if (!Simplified)
6015 return nullptr;
6016 // Reject constant expressions as they don't simplify things.
6017 if (isa<Constant>(Simplified) && !match(Simplified, m_ImmConstant()))
6018 return nullptr;
6019 // Check if the transformation introduces poison.
6020 return impliesPoison(RHS, V) ? Simplified : nullptr;
6021 };
6022
6023 auto ApplyOffset = [&](Value *V, unsigned BinOpc,
6024 Value *RHS) -> OffsetResult {
6025 if (auto *Sel = dyn_cast<SelectInst>(V)) {
6026 if (!Sel->hasOneUse())
6027 return OffsetResult::invalid();
6028 Value *TrueVal = ApplyOffsetImpl(Sel->getTrueValue(), BinOpc, RHS);
6029 if (!TrueVal)
6030 return OffsetResult::invalid();
6031 Value *FalseVal = ApplyOffsetImpl(Sel->getFalseValue(), BinOpc, RHS);
6032 if (!FalseVal)
6033 return OffsetResult::invalid();
6034 return OffsetResult::select(Sel->getCondition(), TrueVal, FalseVal, Sel);
6035 }
6036 if (Value *Simplified = ApplyOffsetImpl(V, BinOpc, RHS))
6037 return OffsetResult::value(Simplified);
6038 return OffsetResult::invalid();
6039 };
6040
6041 for (auto [BinOp, RHS] : OffsetOps) {
6042 auto BinOpc = static_cast<unsigned>(BinOp);
6043
6044 auto Op0Result = ApplyOffset(Op0, BinOpc, RHS);
6045 if (!Op0Result.isValid())
6046 continue;
6047 auto Op1Result = ApplyOffset(Op1, BinOpc, RHS);
6048 if (!Op1Result.isValid())
6049 continue;
6050
6051 Value *NewLHS = Op0Result.materialize(Builder);
6052 Value *NewRHS = Op1Result.materialize(Builder);
6053 return new ICmpInst(I.getPredicate(), NewLHS, NewRHS);
6054 }
6055
6056 return nullptr;
6057}
6058
6060 if (!I.isEquality())
6061 return nullptr;
6062
6063 Value *Op0 = I.getOperand(0), *Op1 = I.getOperand(1);
6064 const CmpInst::Predicate Pred = I.getPredicate();
6065 Value *A, *B, *C, *D;
6066 if (match(Op0, m_Xor(m_Value(A), m_Value(B)))) {
6067 if (A == Op1 || B == Op1) { // (A^B) == A -> B == 0
6068 Value *OtherVal = A == Op1 ? B : A;
6069 return new ICmpInst(Pred, OtherVal, Constant::getNullValue(A->getType()));
6070 }
6071
6072 if (match(Op1, m_Xor(m_Value(C), m_Value(D)))) {
6073 // A^c1 == C^c2 --> A == C^(c1^c2)
6074 ConstantInt *C1, *C2;
6075 if (match(B, m_ConstantInt(C1)) && match(D, m_ConstantInt(C2)) &&
6076 Op1->hasOneUse()) {
6077 Constant *NC = Builder.getInt(C1->getValue() ^ C2->getValue());
6078 Value *Xor = Builder.CreateXor(C, NC);
6079 return new ICmpInst(Pred, A, Xor);
6080 }
6081
6082 // A^B == A^D -> B == D
6083 if (A == C)
6084 return new ICmpInst(Pred, B, D);
6085 if (A == D)
6086 return new ICmpInst(Pred, B, C);
6087 if (B == C)
6088 return new ICmpInst(Pred, A, D);
6089 if (B == D)
6090 return new ICmpInst(Pred, A, C);
6091 }
6092 }
6093
6094 if (match(Op1, m_Xor(m_Value(A), m_Value(B))) && (A == Op0 || B == Op0)) {
6095 // A == (A^B) -> B == 0
6096 Value *OtherVal = A == Op0 ? B : A;
6097 return new ICmpInst(Pred, OtherVal, Constant::getNullValue(A->getType()));
6098 }
6099
6100 // (X&Z) == (Y&Z) -> (X^Y) & Z == 0
6101 if (match(Op0, m_And(m_Value(A), m_Value(B))) &&
6102 match(Op1, m_And(m_Value(C), m_Value(D)))) {
6103 Value *X = nullptr, *Y = nullptr, *Z = nullptr;
6104
6105 if (A == C) {
6106 X = B;
6107 Y = D;
6108 Z = A;
6109 } else if (A == D) {
6110 X = B;
6111 Y = C;
6112 Z = A;
6113 } else if (B == C) {
6114 X = A;
6115 Y = D;
6116 Z = B;
6117 } else if (B == D) {
6118 X = A;
6119 Y = C;
6120 Z = B;
6121 }
6122
6123 if (X) {
6124 // If X^Y is a negative power of two, then `icmp eq/ne (Z & NegP2), 0`
6125 // will fold to `icmp ult/uge Z, -NegP2` incurringb no additional
6126 // instructions.
6127 const APInt *C0, *C1;
6128 bool XorIsNegP2 = match(X, m_APInt(C0)) && match(Y, m_APInt(C1)) &&
6129 (*C0 ^ *C1).isNegatedPowerOf2();
6130
6131 // If either Op0/Op1 are both one use or X^Y will constant fold and one of
6132 // Op0/Op1 are one use, proceed. In those cases we are instruction neutral
6133 // but `icmp eq/ne A, 0` is easier to analyze than `icmp eq/ne A, B`.
6134 int UseCnt =
6135 int(Op0->hasOneUse()) + int(Op1->hasOneUse()) +
6136 (int(match(X, m_ImmConstant()) && match(Y, m_ImmConstant())));
6137 if (XorIsNegP2 || UseCnt >= 2) {
6138 // Build (X^Y) & Z
6139 Op1 = Builder.CreateXor(X, Y);
6140 Op1 = Builder.CreateAnd(Op1, Z);
6141 return new ICmpInst(Pred, Op1, Constant::getNullValue(Op1->getType()));
6142 }
6143 }
6144 }
6145
6146 {
6147 // Similar to above, but specialized for constant because invert is needed:
6148 // (X | C) == (Y | C) --> (X ^ Y) & ~C == 0
6149 Value *X, *Y;
6150 Constant *C;
6151 if (match(Op0, m_OneUse(m_Or(m_Value(X), m_Constant(C)))) &&
6152 match(Op1, m_OneUse(m_Or(m_Value(Y), m_Specific(C))))) {
6153 Value *Xor = Builder.CreateXor(X, Y);
6154 Value *And = Builder.CreateAnd(Xor, ConstantExpr::getNot(C));
6155 return new ICmpInst(Pred, And, Constant::getNullValue(And->getType()));
6156 }
6157 }
6158
6159 if (match(Op1, m_ZExt(m_Value(A))) &&
6160 (Op0->hasOneUse() || Op1->hasOneUse())) {
6161 // (B & (Pow2C-1)) == zext A --> A == trunc B
6162 // (B & (Pow2C-1)) != zext A --> A != trunc B
6163 const APInt *MaskC;
6164 if (match(Op0, m_And(m_Value(B), m_LowBitMask(MaskC))) &&
6165 MaskC->countr_one() == A->getType()->getScalarSizeInBits())
6166 return new ICmpInst(Pred, A, Builder.CreateTrunc(B, A->getType()));
6167 }
6168
6169 // (A >> C) == (B >> C) --> (A^B) u< (1 << C)
6170 // For lshr and ashr pairs.
6171 const APInt *AP1, *AP2;
6172 if ((match(Op0, m_OneUse(m_LShr(m_Value(A), m_APIntAllowPoison(AP1)))) &&
6173 match(Op1, m_OneUse(m_LShr(m_Value(B), m_APIntAllowPoison(AP2))))) ||
6174 (match(Op0, m_OneUse(m_AShr(m_Value(A), m_APIntAllowPoison(AP1)))) &&
6175 match(Op1, m_OneUse(m_AShr(m_Value(B), m_APIntAllowPoison(AP2)))))) {
6176 if (*AP1 != *AP2)
6177 return nullptr;
6178 unsigned TypeBits = AP1->getBitWidth();
6179 unsigned ShAmt = AP1->getLimitedValue(TypeBits);
6180 if (ShAmt < TypeBits && ShAmt != 0) {
6181 ICmpInst::Predicate NewPred =
6183 Value *Xor = Builder.CreateXor(A, B, I.getName() + ".unshifted");
6184 APInt CmpVal = APInt::getOneBitSet(TypeBits, ShAmt);
6185 return new ICmpInst(NewPred, Xor, ConstantInt::get(A->getType(), CmpVal));
6186 }
6187 }
6188
6189 // (A << C) == (B << C) --> ((A^B) & (~0U >> C)) == 0
6190 ConstantInt *Cst1;
6191 if (match(Op0, m_OneUse(m_Shl(m_Value(A), m_ConstantInt(Cst1)))) &&
6192 match(Op1, m_OneUse(m_Shl(m_Value(B), m_Specific(Cst1))))) {
6193 unsigned TypeBits = Cst1->getBitWidth();
6194 unsigned ShAmt = (unsigned)Cst1->getLimitedValue(TypeBits);
6195 if (ShAmt < TypeBits && ShAmt != 0) {
6196 Value *Xor = Builder.CreateXor(A, B, I.getName() + ".unshifted");
6197 APInt AndVal = APInt::getLowBitsSet(TypeBits, TypeBits - ShAmt);
6198 Value *And =
6199 Builder.CreateAnd(Xor, Builder.getInt(AndVal), I.getName() + ".mask");
6200 return new ICmpInst(Pred, And, Constant::getNullValue(Cst1->getType()));
6201 }
6202 }
6203
6204 // Transform "icmp eq (trunc (lshr(X, cst1)), cst" to
6205 // "icmp (and X, mask), cst"
6206 uint64_t ShAmt = 0;
6207 if (Op0->hasOneUse() &&
6208 match(Op0, m_Trunc(m_OneUse(m_LShr(m_Value(A), m_ConstantInt(ShAmt))))) &&
6209 match(Op1, m_ConstantInt(Cst1)) &&
6210 // Only do this when A has multiple uses. This is most important to do
6211 // when it exposes other optimizations.
6212 !A->hasOneUse()) {
6213 unsigned ASize = cast<IntegerType>(A->getType())->getPrimitiveSizeInBits();
6214
6215 if (ShAmt < ASize) {
6216 APInt MaskV =
6218 MaskV <<= ShAmt;
6219
6220 APInt CmpV = Cst1->getValue().zext(ASize);
6221 CmpV <<= ShAmt;
6222
6223 Value *Mask = Builder.CreateAnd(A, Builder.getInt(MaskV));
6224 return new ICmpInst(Pred, Mask, Builder.getInt(CmpV));
6225 }
6226 }
6227
6229 return ICmp;
6230
6231 // Match icmp eq (trunc (lshr A, BW), (ashr (trunc A), BW-1)), which checks
6232 // the top BW/2 + 1 bits are all the same. Create "A >=s INT_MIN && A <=s
6233 // INT_MAX", which we generate as "icmp ult (add A, 2^(BW-1)), 2^BW" to skip a
6234 // few steps of instcombine.
6235 unsigned BitWidth = Op0->getType()->getScalarSizeInBits();
6236 if (match(Op0, m_AShr(m_Trunc(m_Value(A)), m_SpecificInt(BitWidth - 1))) &&
6238 A->getType()->getScalarSizeInBits() == BitWidth * 2 &&
6239 (I.getOperand(0)->hasOneUse() || I.getOperand(1)->hasOneUse())) {
6241 Value *Add = Builder.CreateAdd(A, ConstantInt::get(A->getType(), C));
6242 return new ICmpInst(Pred == ICmpInst::ICMP_EQ ? ICmpInst::ICMP_ULT
6244 Add, ConstantInt::get(A->getType(), C.shl(1)));
6245 }
6246
6247 // Canonicalize:
6248 // Assume B_Pow2 != 0
6249 // 1. A & B_Pow2 != B_Pow2 -> A & B_Pow2 == 0
6250 // 2. A & B_Pow2 == B_Pow2 -> A & B_Pow2 != 0
6251 if (match(Op0, m_c_And(m_Specific(Op1), m_Value())) &&
6252 isKnownToBeAPowerOfTwo(Op1, /* OrZero */ false, &I))
6253 return new ICmpInst(CmpInst::getInversePredicate(Pred), Op0,
6255
6256 if (match(Op1, m_c_And(m_Specific(Op0), m_Value())) &&
6257 isKnownToBeAPowerOfTwo(Op0, /* OrZero */ false, &I))
6258 return new ICmpInst(CmpInst::getInversePredicate(Pred), Op1,
6259 ConstantInt::getNullValue(Op1->getType()));
6260
6261 // Canonicalize:
6262 // icmp eq/ne X, OneUse(rotate-right(X))
6263 // -> icmp eq/ne X, rotate-left(X)
6264 // We generally try to convert rotate-right -> rotate-left, this just
6265 // canonicalizes another case.
6266 if (match(&I, m_c_ICmp(m_Value(A),
6268 m_Deferred(A), m_Deferred(A), m_Value(B))))))
6269 return new ICmpInst(
6270 Pred, A,
6271 Builder.CreateIntrinsic(Op0->getType(), Intrinsic::fshl, {A, A, B}));
6272
6273 // Canonicalize:
6274 // icmp eq/ne OneUse(A ^ Cst), B --> icmp eq/ne (A ^ B), Cst
6275 Constant *Cst;
6278 return new ICmpInst(Pred, Builder.CreateXor(A, B), Cst);
6279
6280 {
6281 // (icmp eq/ne (and (add/sub/xor X, P2), P2), P2)
6282 auto m_Matcher =
6285 m_Sub(m_Value(B), m_Deferred(A)));
6286 std::optional<bool> IsZero = std::nullopt;
6287 if (match(&I, m_c_ICmp(m_OneUse(m_c_And(m_Value(A), m_Matcher)),
6288 m_Deferred(A))))
6289 IsZero = false;
6290 // (icmp eq/ne (and (add/sub/xor X, P2), P2), 0)
6291 else if (match(&I,
6292 m_ICmp(m_OneUse(m_c_And(m_Value(A), m_Matcher)), m_Zero())))
6293 IsZero = true;
6294
6295 if (IsZero && isKnownToBeAPowerOfTwo(A, /* OrZero */ true, &I))
6296 // (icmp eq/ne (and (add/sub/xor X, P2), P2), P2)
6297 // -> (icmp eq/ne (and X, P2), 0)
6298 // (icmp eq/ne (and (add/sub/xor X, P2), P2), 0)
6299 // -> (icmp eq/ne (and X, P2), P2)
6300 return new ICmpInst(Pred, Builder.CreateAnd(B, A),
6301 *IsZero ? A
6302 : ConstantInt::getNullValue(A->getType()));
6303 }
6304
6305 if (auto *Res = foldICmpEqualityWithOffset(
6306 I, Builder, getSimplifyQuery().getWithInstruction(&I)))
6307 return Res;
6308
6309 return nullptr;
6310}
6311
6313 ICmpInst::Predicate Pred = ICmp.getPredicate();
6314 Value *Op0 = ICmp.getOperand(0), *Op1 = ICmp.getOperand(1);
6315
6316 // Try to canonicalize trunc + compare-to-constant into a mask + cmp.
6317 // The trunc masks high bits while the compare may effectively mask low bits.
6318 Value *X;
6319 const APInt *C;
6320 if (!match(Op0, m_OneUse(m_Trunc(m_Value(X)))) || !match(Op1, m_APInt(C)))
6321 return nullptr;
6322
6323 // This matches patterns corresponding to tests of the signbit as well as:
6324 // (trunc X) pred C2 --> (X & Mask) == C
6325 if (auto Res = decomposeBitTestICmp(Op0, Op1, Pred, /*LookThroughTrunc=*/true,
6326 /*AllowNonZeroC=*/true)) {
6327 Value *And = Builder.CreateAnd(Res->X, Res->Mask);
6328 Constant *C = ConstantInt::get(Res->X->getType(), Res->C);
6329 return new ICmpInst(Res->Pred, And, C);
6330 }
6331
6332 unsigned SrcBits = X->getType()->getScalarSizeInBits();
6333 if (auto *II = dyn_cast<IntrinsicInst>(X)) {
6334 if (II->getIntrinsicID() == Intrinsic::cttz ||
6335 II->getIntrinsicID() == Intrinsic::ctlz) {
6336 unsigned MaxRet = SrcBits;
6337 // If the "is_zero_poison" argument is set, then we know at least
6338 // one bit is set in the input, so the result is always at least one
6339 // less than the full bitwidth of that input.
6340 if (match(II->getArgOperand(1), m_One()))
6341 MaxRet--;
6342
6343 // Make sure the destination is wide enough to hold the largest output of
6344 // the intrinsic.
6345 if (llvm::Log2_32(MaxRet) + 1 <= Op0->getType()->getScalarSizeInBits())
6346 if (Instruction *I =
6347 foldICmpIntrinsicWithConstant(ICmp, II, C->zext(SrcBits)))
6348 return I;
6349 }
6350 }
6351
6352 return nullptr;
6353}
6354
6356 assert(isa<CastInst>(ICmp.getOperand(0)) && "Expected cast for operand 0");
6357 auto *CastOp0 = cast<CastInst>(ICmp.getOperand(0));
6358 Value *X;
6359 if (!match(CastOp0, m_ZExtOrSExt(m_Value(X))))
6360 return nullptr;
6361
6362 bool IsSignedExt = CastOp0->getOpcode() == Instruction::SExt;
6363 bool IsSignedCmp = ICmp.isSigned();
6364
6365 // icmp Pred (ext X), (ext Y)
6366 Value *Y;
6367 if (match(ICmp.getOperand(1), m_ZExtOrSExt(m_Value(Y)))) {
6368 bool IsZext0 = isa<ZExtInst>(ICmp.getOperand(0));
6369 bool IsZext1 = isa<ZExtInst>(ICmp.getOperand(1));
6370
6371 if (IsZext0 != IsZext1) {
6372 // If X and Y and both i1
6373 // (icmp eq/ne (zext X) (sext Y))
6374 // eq -> (icmp eq (or X, Y), 0)
6375 // ne -> (icmp ne (or X, Y), 0)
6376 if (ICmp.isEquality() && X->getType()->isIntOrIntVectorTy(1) &&
6377 Y->getType()->isIntOrIntVectorTy(1))
6378 return new ICmpInst(ICmp.getPredicate(), Builder.CreateOr(X, Y),
6379 Constant::getNullValue(X->getType()));
6380
6381 // If we have mismatched casts and zext has the nneg flag, we can
6382 // treat the "zext nneg" as "sext". Otherwise, we cannot fold and quit.
6383
6384 auto *NonNegInst0 = dyn_cast<PossiblyNonNegInst>(ICmp.getOperand(0));
6385 auto *NonNegInst1 = dyn_cast<PossiblyNonNegInst>(ICmp.getOperand(1));
6386
6387 bool IsNonNeg0 = NonNegInst0 && NonNegInst0->hasNonNeg();
6388 bool IsNonNeg1 = NonNegInst1 && NonNegInst1->hasNonNeg();
6389
6390 if ((IsZext0 && IsNonNeg0) || (IsZext1 && IsNonNeg1))
6391 IsSignedExt = true;
6392 else
6393 return nullptr;
6394 }
6395
6396 // Not an extension from the same type?
6397 Type *XTy = X->getType(), *YTy = Y->getType();
6398 if (XTy != YTy) {
6399 // One of the casts must have one use because we are creating a new cast.
6400 if (!ICmp.getOperand(0)->hasOneUse() && !ICmp.getOperand(1)->hasOneUse())
6401 return nullptr;
6402 // Extend the narrower operand to the type of the wider operand.
6403 CastInst::CastOps CastOpcode =
6404 IsSignedExt ? Instruction::SExt : Instruction::ZExt;
6405 if (XTy->getScalarSizeInBits() < YTy->getScalarSizeInBits())
6406 X = Builder.CreateCast(CastOpcode, X, YTy);
6407 else if (YTy->getScalarSizeInBits() < XTy->getScalarSizeInBits())
6408 Y = Builder.CreateCast(CastOpcode, Y, XTy);
6409 else
6410 return nullptr;
6411 }
6412
6413 // (zext X) == (zext Y) --> X == Y
6414 // (sext X) == (sext Y) --> X == Y
6415 if (ICmp.isEquality())
6416 return new ICmpInst(ICmp.getPredicate(), X, Y);
6417
6418 // A signed comparison of sign extended values simplifies into a
6419 // signed comparison.
6420 if (IsSignedCmp && IsSignedExt)
6421 return new ICmpInst(ICmp.getPredicate(), X, Y);
6422
6423 // The other three cases all fold into an unsigned comparison.
6424 return new ICmpInst(ICmp.getUnsignedPredicate(), X, Y);
6425 }
6426
6427 // Below here, we are only folding a compare with constant.
6428 auto *C = dyn_cast<Constant>(ICmp.getOperand(1));
6429 if (!C)
6430 return nullptr;
6431
6432 // If a lossless truncate is possible...
6433 Type *SrcTy = CastOp0->getSrcTy();
6434 Constant *Res = getLosslessInvCast(C, SrcTy, CastOp0->getOpcode(), DL);
6435 if (Res) {
6436 if (ICmp.isEquality())
6437 return new ICmpInst(ICmp.getPredicate(), X, Res);
6438
6439 // A signed comparison of sign extended values simplifies into a
6440 // signed comparison.
6441 if (IsSignedExt && IsSignedCmp)
6442 return new ICmpInst(ICmp.getPredicate(), X, Res);
6443
6444 // The other three cases all fold into an unsigned comparison.
6445 return new ICmpInst(ICmp.getUnsignedPredicate(), X, Res);
6446 }
6447
6448 // The re-extended constant changed, partly changed (in the case of a vector),
6449 // or could not be determined to be equal (in the case of a constant
6450 // expression), so the constant cannot be represented in the shorter type.
6451 // All the cases that fold to true or false will have already been handled
6452 // by simplifyICmpInst, so only deal with the tricky case.
6453 if (IsSignedCmp || !IsSignedExt || !isa<ConstantInt>(C))
6454 return nullptr;
6455
6456 // Is source op positive?
6457 // icmp ult (sext X), C --> icmp sgt X, -1
6458 if (ICmp.getPredicate() == ICmpInst::ICMP_ULT)
6460
6461 // Is source op negative?
6462 // icmp ugt (sext X), C --> icmp slt X, 0
6463 assert(ICmp.getPredicate() == ICmpInst::ICMP_UGT && "ICmp should be folded!");
6465}
6466
6467/// Handle icmp (cast x), (cast or constant).
6469 // If any operand of ICmp is a inttoptr roundtrip cast then remove it as
6470 // icmp compares only pointer's value.
6471 // icmp (inttoptr (ptrtoint p1)), p2 --> icmp p1, p2.
6472 Value *SimplifiedOp0 = simplifyIntToPtrRoundTripCast(ICmp.getOperand(0));
6473 Value *SimplifiedOp1 = simplifyIntToPtrRoundTripCast(ICmp.getOperand(1));
6474 if (SimplifiedOp0 || SimplifiedOp1)
6475 return new ICmpInst(ICmp.getPredicate(),
6476 SimplifiedOp0 ? SimplifiedOp0 : ICmp.getOperand(0),
6477 SimplifiedOp1 ? SimplifiedOp1 : ICmp.getOperand(1));
6478
6479 auto *CastOp0 = dyn_cast<CastInst>(ICmp.getOperand(0));
6480 if (!CastOp0)
6481 return nullptr;
6482 if (!isa<Constant>(ICmp.getOperand(1)) && !isa<CastInst>(ICmp.getOperand(1)))
6483 return nullptr;
6484
6485 Value *Op0Src = CastOp0->getOperand(0);
6486 Type *SrcTy = CastOp0->getSrcTy();
6487 Type *DestTy = CastOp0->getDestTy();
6488
6489 // Turn icmp (ptrtoint x), (ptrtoint/c) into a compare of the input if the
6490 // integer type is the same size as the pointer type.
6491 auto CompatibleSizes = [&](Type *PtrTy, Type *IntTy) {
6492 if (isa<VectorType>(PtrTy)) {
6493 PtrTy = cast<VectorType>(PtrTy)->getElementType();
6494 IntTy = cast<VectorType>(IntTy)->getElementType();
6495 }
6496 return DL.getPointerTypeSizeInBits(PtrTy) == IntTy->getIntegerBitWidth();
6497 };
6498 if (CastOp0->getOpcode() == Instruction::PtrToInt &&
6499 CompatibleSizes(SrcTy, DestTy)) {
6500 Value *NewOp1 = nullptr;
6501 if (auto *PtrToIntOp1 = dyn_cast<PtrToIntOperator>(ICmp.getOperand(1))) {
6502 Value *PtrSrc = PtrToIntOp1->getOperand(0);
6503 if (PtrSrc->getType() == Op0Src->getType())
6504 NewOp1 = PtrToIntOp1->getOperand(0);
6505 } else if (auto *RHSC = dyn_cast<Constant>(ICmp.getOperand(1))) {
6506 NewOp1 = ConstantExpr::getIntToPtr(RHSC, SrcTy);
6507 }
6508
6509 if (NewOp1)
6510 return new ICmpInst(ICmp.getPredicate(), Op0Src, NewOp1);
6511 }
6512
6513 // Do the same in the other direction for icmp (inttoptr x), (inttoptr/c).
6514 if (CastOp0->getOpcode() == Instruction::IntToPtr &&
6515 CompatibleSizes(DestTy, SrcTy)) {
6516 Value *NewOp1 = nullptr;
6517 if (auto *IntToPtrOp1 = dyn_cast<IntToPtrInst>(ICmp.getOperand(1))) {
6518 Value *IntSrc = IntToPtrOp1->getOperand(0);
6519 if (IntSrc->getType() == Op0Src->getType())
6520 NewOp1 = IntToPtrOp1->getOperand(0);
6521 } else if (auto *RHSC = dyn_cast<Constant>(ICmp.getOperand(1))) {
6522 NewOp1 = ConstantFoldConstant(ConstantExpr::getPtrToInt(RHSC, SrcTy), DL);
6523 }
6524
6525 if (NewOp1)
6526 return new ICmpInst(ICmp.getPredicate(), Op0Src, NewOp1);
6527 }
6528
6529 if (Instruction *R = foldICmpWithTrunc(ICmp))
6530 return R;
6531
6532 return foldICmpWithZextOrSext(ICmp);
6533}
6534
6536 bool IsSigned) {
6537 switch (BinaryOp) {
6538 default:
6539 llvm_unreachable("Unsupported binary op");
6540 case Instruction::Add:
6541 case Instruction::Sub:
6542 return match(RHS, m_Zero());
6543 case Instruction::Mul:
6544 return !(RHS->getType()->isIntOrIntVectorTy(1) && IsSigned) &&
6545 match(RHS, m_One());
6546 }
6547}
6548
6551 bool IsSigned, Value *LHS, Value *RHS,
6552 Instruction *CxtI) const {
6553 switch (BinaryOp) {
6554 default:
6555 llvm_unreachable("Unsupported binary op");
6556 case Instruction::Add:
6557 if (IsSigned)
6558 return computeOverflowForSignedAdd(LHS, RHS, CxtI);
6559 else
6560 return computeOverflowForUnsignedAdd(LHS, RHS, CxtI);
6561 case Instruction::Sub:
6562 if (IsSigned)
6563 return computeOverflowForSignedSub(LHS, RHS, CxtI);
6564 else
6565 return computeOverflowForUnsignedSub(LHS, RHS, CxtI);
6566 case Instruction::Mul:
6567 if (IsSigned)
6568 return computeOverflowForSignedMul(LHS, RHS, CxtI);
6569 else
6570 return computeOverflowForUnsignedMul(LHS, RHS, CxtI);
6571 }
6572}
6573
6574bool InstCombinerImpl::OptimizeOverflowCheck(Instruction::BinaryOps BinaryOp,
6575 bool IsSigned, Value *LHS,
6576 Value *RHS, Instruction &OrigI,
6577 Value *&Result,
6578 Constant *&Overflow) {
6579 if (OrigI.isCommutative() && isa<Constant>(LHS) && !isa<Constant>(RHS))
6580 std::swap(LHS, RHS);
6581
6582 // If the overflow check was an add followed by a compare, the insertion point
6583 // may be pointing to the compare. We want to insert the new instructions
6584 // before the add in case there are uses of the add between the add and the
6585 // compare.
6586 Builder.SetInsertPoint(&OrigI);
6587
6588 Type *OverflowTy = Type::getInt1Ty(LHS->getContext());
6589 if (auto *LHSTy = dyn_cast<VectorType>(LHS->getType()))
6590 OverflowTy = VectorType::get(OverflowTy, LHSTy->getElementCount());
6591
6592 if (isNeutralValue(BinaryOp, RHS, IsSigned)) {
6593 Result = LHS;
6594 Overflow = ConstantInt::getFalse(OverflowTy);
6595 return true;
6596 }
6597
6598 switch (computeOverflow(BinaryOp, IsSigned, LHS, RHS, &OrigI)) {
6600 return false;
6603 Result = Builder.CreateBinOp(BinaryOp, LHS, RHS);
6604 Result->takeName(&OrigI);
6605 Overflow = ConstantInt::getTrue(OverflowTy);
6606 return true;
6608 Result = Builder.CreateBinOp(BinaryOp, LHS, RHS);
6609 Result->takeName(&OrigI);
6610 Overflow = ConstantInt::getFalse(OverflowTy);
6611 if (auto *Inst = dyn_cast<Instruction>(Result)) {
6612 if (IsSigned)
6613 Inst->setHasNoSignedWrap();
6614 else
6615 Inst->setHasNoUnsignedWrap();
6616 }
6617 return true;
6618 }
6619
6620 llvm_unreachable("Unexpected overflow result");
6621}
6622
6623/// Recognize and process idiom involving test for multiplication
6624/// overflow.
6625///
6626/// The caller has matched a pattern of the form:
6627/// I = cmp u (mul(zext A, zext B), V
6628/// The function checks if this is a test for overflow and if so replaces
6629/// multiplication with call to 'mul.with.overflow' intrinsic.
6630///
6631/// \param I Compare instruction.
6632/// \param MulVal Result of 'mult' instruction. It is one of the arguments of
6633/// the compare instruction. Must be of integer type.
6634/// \param OtherVal The other argument of compare instruction.
6635/// \returns Instruction which must replace the compare instruction, NULL if no
6636/// replacement required.
6638 const APInt *OtherVal,
6639 InstCombinerImpl &IC) {
6640 // Don't bother doing this transformation for pointers, don't do it for
6641 // vectors.
6642 if (!isa<IntegerType>(MulVal->getType()))
6643 return nullptr;
6644
6645 auto *MulInstr = dyn_cast<Instruction>(MulVal);
6646 if (!MulInstr)
6647 return nullptr;
6648 assert(MulInstr->getOpcode() == Instruction::Mul);
6649
6650 auto *LHS = cast<ZExtInst>(MulInstr->getOperand(0)),
6651 *RHS = cast<ZExtInst>(MulInstr->getOperand(1));
6652 assert(LHS->getOpcode() == Instruction::ZExt);
6653 assert(RHS->getOpcode() == Instruction::ZExt);
6654 Value *A = LHS->getOperand(0), *B = RHS->getOperand(0);
6655
6656 // Calculate type and width of the result produced by mul.with.overflow.
6657 Type *TyA = A->getType(), *TyB = B->getType();
6658 unsigned WidthA = TyA->getPrimitiveSizeInBits(),
6659 WidthB = TyB->getPrimitiveSizeInBits();
6660 unsigned MulWidth;
6661 Type *MulType;
6662 if (WidthB > WidthA) {
6663 MulWidth = WidthB;
6664 MulType = TyB;
6665 } else {
6666 MulWidth = WidthA;
6667 MulType = TyA;
6668 }
6669
6670 // In order to replace the original mul with a narrower mul.with.overflow,
6671 // all uses must ignore upper bits of the product. The number of used low
6672 // bits must be not greater than the width of mul.with.overflow.
6673 if (MulVal->hasNUsesOrMore(2))
6674 for (User *U : MulVal->users()) {
6675 if (U == &I)
6676 continue;
6677 if (TruncInst *TI = dyn_cast<TruncInst>(U)) {
6678 // Check if truncation ignores bits above MulWidth.
6679 unsigned TruncWidth = TI->getType()->getPrimitiveSizeInBits();
6680 if (TruncWidth > MulWidth)
6681 return nullptr;
6682 } else if (BinaryOperator *BO = dyn_cast<BinaryOperator>(U)) {
6683 // Check if AND ignores bits above MulWidth.
6684 if (BO->getOpcode() != Instruction::And)
6685 return nullptr;
6686 if (ConstantInt *CI = dyn_cast<ConstantInt>(BO->getOperand(1))) {
6687 const APInt &CVal = CI->getValue();
6688 if (CVal.getBitWidth() - CVal.countl_zero() > MulWidth)
6689 return nullptr;
6690 } else {
6691 // In this case we could have the operand of the binary operation
6692 // being defined in another block, and performing the replacement
6693 // could break the dominance relation.
6694 return nullptr;
6695 }
6696 } else {
6697 // Other uses prohibit this transformation.
6698 return nullptr;
6699 }
6700 }
6701
6702 // Recognize patterns
6703 switch (I.getPredicate()) {
6704 case ICmpInst::ICMP_UGT: {
6705 // Recognize pattern:
6706 // mulval = mul(zext A, zext B)
6707 // cmp ugt mulval, max
6708 APInt MaxVal = APInt::getMaxValue(MulWidth);
6709 MaxVal = MaxVal.zext(OtherVal->getBitWidth());
6710 if (MaxVal.eq(*OtherVal))
6711 break; // Recognized
6712 return nullptr;
6713 }
6714
6715 case ICmpInst::ICMP_ULT: {
6716 // Recognize pattern:
6717 // mulval = mul(zext A, zext B)
6718 // cmp ule mulval, max + 1
6719 APInt MaxVal = APInt::getOneBitSet(OtherVal->getBitWidth(), MulWidth);
6720 if (MaxVal.eq(*OtherVal))
6721 break; // Recognized
6722 return nullptr;
6723 }
6724
6725 default:
6726 return nullptr;
6727 }
6728
6729 InstCombiner::BuilderTy &Builder = IC.Builder;
6730 Builder.SetInsertPoint(MulInstr);
6731
6732 // Replace: mul(zext A, zext B) --> mul.with.overflow(A, B)
6733 Value *MulA = A, *MulB = B;
6734 if (WidthA < MulWidth)
6735 MulA = Builder.CreateZExt(A, MulType);
6736 if (WidthB < MulWidth)
6737 MulB = Builder.CreateZExt(B, MulType);
6738 CallInst *Call =
6739 Builder.CreateIntrinsic(Intrinsic::umul_with_overflow, MulType,
6740 {MulA, MulB}, /*FMFSource=*/nullptr, "umul");
6741 IC.addToWorklist(MulInstr);
6742
6743 // If there are uses of mul result other than the comparison, we know that
6744 // they are truncation or binary AND. Change them to use result of
6745 // mul.with.overflow and adjust properly mask/size.
6746 if (MulVal->hasNUsesOrMore(2)) {
6747 Value *Mul = Builder.CreateExtractValue(Call, 0, "umul.value");
6748 for (User *U : make_early_inc_range(MulVal->users())) {
6749 if (U == &I)
6750 continue;
6751 if (TruncInst *TI = dyn_cast<TruncInst>(U)) {
6752 if (TI->getType()->getPrimitiveSizeInBits() == MulWidth)
6753 IC.replaceInstUsesWith(*TI, Mul);
6754 else
6755 TI->setOperand(0, Mul);
6756 } else if (BinaryOperator *BO = dyn_cast<BinaryOperator>(U)) {
6757 assert(BO->getOpcode() == Instruction::And);
6758 // Replace (mul & mask) --> zext (mul.with.overflow & short_mask)
6759 ConstantInt *CI = cast<ConstantInt>(BO->getOperand(1));
6760 APInt ShortMask = CI->getValue().trunc(MulWidth);
6761 Value *ShortAnd = Builder.CreateAnd(Mul, ShortMask);
6762 Value *Zext = Builder.CreateZExt(ShortAnd, BO->getType());
6763 IC.replaceInstUsesWith(*BO, Zext);
6764 } else {
6765 llvm_unreachable("Unexpected Binary operation");
6766 }
6768 }
6769 }
6770
6771 // The original icmp gets replaced with the overflow value, maybe inverted
6772 // depending on predicate.
6773 if (I.getPredicate() == ICmpInst::ICMP_ULT) {
6774 Value *Res = Builder.CreateExtractValue(Call, 1);
6775 return BinaryOperator::CreateNot(Res);
6776 }
6777
6778 return ExtractValueInst::Create(Call, 1);
6779}
6780
6781/// When performing a comparison against a constant, it is possible that not all
6782/// the bits in the LHS are demanded. This helper method computes the mask that
6783/// IS demanded.
6785 const APInt *RHS;
6786 if (!match(I.getOperand(1), m_APInt(RHS)))
6788
6789 // If this is a normal comparison, it demands all bits. If it is a sign bit
6790 // comparison, it only demands the sign bit.
6791 bool UnusedBit;
6792 if (isSignBitCheck(I.getPredicate(), *RHS, UnusedBit))
6794
6795 switch (I.getPredicate()) {
6796 // For a UGT comparison, we don't care about any bits that
6797 // correspond to the trailing ones of the comparand. The value of these
6798 // bits doesn't impact the outcome of the comparison, because any value
6799 // greater than the RHS must differ in a bit higher than these due to carry.
6800 case ICmpInst::ICMP_UGT:
6801 return APInt::getBitsSetFrom(BitWidth, RHS->countr_one());
6802
6803 // Similarly, for a ULT comparison, we don't care about the trailing zeros.
6804 // Any value less than the RHS must differ in a higher bit because of carries.
6805 case ICmpInst::ICMP_ULT:
6806 return APInt::getBitsSetFrom(BitWidth, RHS->countr_zero());
6807
6808 default:
6810 }
6811}
6812
6813/// Check that one use is in the same block as the definition and all
6814/// other uses are in blocks dominated by a given block.
6815///
6816/// \param DI Definition
6817/// \param UI Use
6818/// \param DB Block that must dominate all uses of \p DI outside
6819/// the parent block
6820/// \return true when \p UI is the only use of \p DI in the parent block
6821/// and all other uses of \p DI are in blocks dominated by \p DB.
6822///
6824 const Instruction *UI,
6825 const BasicBlock *DB) const {
6826 assert(DI && UI && "Instruction not defined\n");
6827 // Ignore incomplete definitions.
6828 if (!DI->getParent())
6829 return false;
6830 // DI and UI must be in the same block.
6831 if (DI->getParent() != UI->getParent())
6832 return false;
6833 // Protect from self-referencing blocks.
6834 if (DI->getParent() == DB)
6835 return false;
6836 for (const User *U : DI->users()) {
6837 auto *Usr = cast<Instruction>(U);
6838 if (Usr != UI && !DT.dominates(DB, Usr->getParent()))
6839 return false;
6840 }
6841 return true;
6842}
6843
6844/// Return true when the instruction sequence within a block is select-cmp-br.
6846 const BasicBlock *BB = SI->getParent();
6847 if (!BB)
6848 return false;
6850 if (!BI)
6851 return false;
6852 auto *IC = dyn_cast<ICmpInst>(BI->getCondition());
6853 if (!IC || (IC->getOperand(0) != SI && IC->getOperand(1) != SI))
6854 return false;
6855 return true;
6856}
6857
6858/// True when a select result is replaced by one of its operands
6859/// in select-icmp sequence. This will eventually result in the elimination
6860/// of the select.
6861///
6862/// \param SI Select instruction
6863/// \param Icmp Compare instruction
6864/// \param SIOpd Operand that replaces the select
6865///
6866/// Notes:
6867/// - The replacement is global and requires dominator information
6868/// - The caller is responsible for the actual replacement
6869///
6870/// Example:
6871///
6872/// entry:
6873/// %4 = select i1 %3, %C* %0, %C* null
6874/// %5 = icmp eq %C* %4, null
6875/// br i1 %5, label %9, label %7
6876/// ...
6877/// ; <label>:7 ; preds = %entry
6878/// %8 = getelementptr inbounds %C* %4, i64 0, i32 0
6879/// ...
6880///
6881/// can be transformed to
6882///
6883/// %5 = icmp eq %C* %0, null
6884/// %6 = select i1 %3, i1 %5, i1 true
6885/// br i1 %6, label %9, label %7
6886/// ...
6887/// ; <label>:7 ; preds = %entry
6888/// %8 = getelementptr inbounds %C* %0, i64 0, i32 0 // replace by %0!
6889///
6890/// Similar when the first operand of the select is a constant or/and
6891/// the compare is for not equal rather than equal.
6892///
6893/// NOTE: The function is only called when the select and compare constants
6894/// are equal, the optimization can work only for EQ predicates. This is not a
6895/// major restriction since a NE compare should be 'normalized' to an equal
6896/// compare, which usually happens in the combiner and test case
6897/// select-cmp-br.ll checks for it.
6899 const ICmpInst *Icmp,
6900 const unsigned SIOpd) {
6901 assert((SIOpd == 1 || SIOpd == 2) && "Invalid select operand!");
6903 BasicBlock *Succ = SI->getParent()->getTerminator()->getSuccessor(1);
6904 // The check for the single predecessor is not the best that can be
6905 // done. But it protects efficiently against cases like when SI's
6906 // home block has two successors, Succ and Succ1, and Succ1 predecessor
6907 // of Succ. Then SI can't be replaced by SIOpd because the use that gets
6908 // replaced can be reached on either path. So the uniqueness check
6909 // guarantees that the path all uses of SI (outside SI's parent) are on
6910 // is disjoint from all other paths out of SI. But that information
6911 // is more expensive to compute, and the trade-off here is in favor
6912 // of compile-time. It should also be noticed that we check for a single
6913 // predecessor and not only uniqueness. This to handle the situation when
6914 // Succ and Succ1 points to the same basic block.
6915 if (Succ->getSinglePredecessor() && dominatesAllUses(SI, Icmp, Succ)) {
6916 NumSel++;
6917 SI->replaceUsesOutsideBlock(SI->getOperand(SIOpd), SI->getParent());
6918 return true;
6919 }
6920 }
6921 return false;
6922}
6923
6924/// Try to fold the comparison based on range information we can get by checking
6925/// whether bits are known to be zero or one in the inputs.
6927 Value *Op0 = I.getOperand(0), *Op1 = I.getOperand(1);
6928 Type *Ty = Op0->getType();
6929 ICmpInst::Predicate Pred = I.getPredicate();
6930
6931 // Get scalar or pointer size.
6932 unsigned BitWidth = Ty->isIntOrIntVectorTy()
6933 ? Ty->getScalarSizeInBits()
6934 : DL.getPointerTypeSizeInBits(Ty->getScalarType());
6935
6936 if (!BitWidth)
6937 return nullptr;
6938
6939 KnownBits Op0Known(BitWidth);
6940 KnownBits Op1Known(BitWidth);
6941
6942 {
6943 // Don't use dominating conditions when folding icmp using known bits. This
6944 // may convert signed into unsigned predicates in ways that other passes
6945 // (especially IndVarSimplify) may not be able to reliably undo.
6946 SimplifyQuery Q = SQ.getWithoutDomCondCache().getWithInstruction(&I);
6948 Op0Known, Q))
6949 return &I;
6950
6951 if (SimplifyDemandedBits(&I, 1, APInt::getAllOnes(BitWidth), Op1Known, Q))
6952 return &I;
6953 }
6954
6955 if (!isa<Constant>(Op0) && Op0Known.isConstant())
6956 return new ICmpInst(
6957 Pred, ConstantExpr::getIntegerValue(Ty, Op0Known.getConstant()), Op1);
6958 if (!isa<Constant>(Op1) && Op1Known.isConstant())
6959 return new ICmpInst(
6960 Pred, Op0, ConstantExpr::getIntegerValue(Ty, Op1Known.getConstant()));
6961
6962 if (std::optional<bool> Res = ICmpInst::compare(Op0Known, Op1Known, Pred))
6963 return replaceInstUsesWith(I, ConstantInt::getBool(I.getType(), *Res));
6964
6965 // Given the known and unknown bits, compute a range that the LHS could be
6966 // in. Compute the Min, Max and RHS values based on the known bits. For the
6967 // EQ and NE we use unsigned values.
6968 APInt Op0Min(BitWidth, 0), Op0Max(BitWidth, 0);
6969 APInt Op1Min(BitWidth, 0), Op1Max(BitWidth, 0);
6970 if (I.isSigned()) {
6971 Op0Min = Op0Known.getSignedMinValue();
6972 Op0Max = Op0Known.getSignedMaxValue();
6973 Op1Min = Op1Known.getSignedMinValue();
6974 Op1Max = Op1Known.getSignedMaxValue();
6975 } else {
6976 Op0Min = Op0Known.getMinValue();
6977 Op0Max = Op0Known.getMaxValue();
6978 Op1Min = Op1Known.getMinValue();
6979 Op1Max = Op1Known.getMaxValue();
6980 }
6981
6982 // Don't break up a clamp pattern -- (min(max X, Y), Z) -- by replacing a
6983 // min/max canonical compare with some other compare. That could lead to
6984 // conflict with select canonicalization and infinite looping.
6985 // FIXME: This constraint may go away if min/max intrinsics are canonical.
6986 auto isMinMaxCmp = [&](Instruction &Cmp) {
6987 if (!Cmp.hasOneUse())
6988 return false;
6989 Value *A, *B;
6990 SelectPatternFlavor SPF = matchSelectPattern(Cmp.user_back(), A, B).Flavor;
6992 return false;
6993 return match(Op0, m_MaxOrMin(m_Value(), m_Value())) ||
6994 match(Op1, m_MaxOrMin(m_Value(), m_Value()));
6995 };
6996 if (!isMinMaxCmp(I)) {
6997 switch (Pred) {
6998 default:
6999 break;
7000 case ICmpInst::ICMP_ULT: {
7001 if (Op1Min == Op0Max) // A <u B -> A != B if max(A) == min(B)
7002 return new ICmpInst(ICmpInst::ICMP_NE, Op0, Op1);
7003 const APInt *CmpC;
7004 if (match(Op1, m_APInt(CmpC))) {
7005 // A <u C -> A == C-1 if min(A)+1 == C
7006 if (*CmpC == Op0Min + 1)
7007 return new ICmpInst(ICmpInst::ICMP_EQ, Op0,
7008 ConstantInt::get(Op1->getType(), *CmpC - 1));
7009 // X <u C --> X == 0, if the number of zero bits in the bottom of X
7010 // exceeds the log2 of C.
7011 if (Op0Known.countMinTrailingZeros() >= CmpC->ceilLogBase2())
7012 return new ICmpInst(ICmpInst::ICMP_EQ, Op0,
7013 Constant::getNullValue(Op1->getType()));
7014 }
7015 break;
7016 }
7017 case ICmpInst::ICMP_UGT: {
7018 if (Op1Max == Op0Min) // A >u B -> A != B if min(A) == max(B)
7019 return new ICmpInst(ICmpInst::ICMP_NE, Op0, Op1);
7020 const APInt *CmpC;
7021 if (match(Op1, m_APInt(CmpC))) {
7022 // A >u C -> A == C+1 if max(a)-1 == C
7023 if (*CmpC == Op0Max - 1)
7024 return new ICmpInst(ICmpInst::ICMP_EQ, Op0,
7025 ConstantInt::get(Op1->getType(), *CmpC + 1));
7026 // X >u C --> X != 0, if the number of zero bits in the bottom of X
7027 // exceeds the log2 of C.
7028 if (Op0Known.countMinTrailingZeros() >= CmpC->getActiveBits())
7029 return new ICmpInst(ICmpInst::ICMP_NE, Op0,
7030 Constant::getNullValue(Op1->getType()));
7031 }
7032 break;
7033 }
7034 case ICmpInst::ICMP_SLT: {
7035 if (Op1Min == Op0Max) // A <s B -> A != B if max(A) == min(B)
7036 return new ICmpInst(ICmpInst::ICMP_NE, Op0, Op1);
7037 const APInt *CmpC;
7038 if (match(Op1, m_APInt(CmpC))) {
7039 if (*CmpC == Op0Min + 1) // A <s C -> A == C-1 if min(A)+1 == C
7040 return new ICmpInst(ICmpInst::ICMP_EQ, Op0,
7041 ConstantInt::get(Op1->getType(), *CmpC - 1));
7042 }
7043 break;
7044 }
7045 case ICmpInst::ICMP_SGT: {
7046 if (Op1Max == Op0Min) // A >s B -> A != B if min(A) == max(B)
7047 return new ICmpInst(ICmpInst::ICMP_NE, Op0, Op1);
7048 const APInt *CmpC;
7049 if (match(Op1, m_APInt(CmpC))) {
7050 if (*CmpC == Op0Max - 1) // A >s C -> A == C+1 if max(A)-1 == C
7051 return new ICmpInst(ICmpInst::ICMP_EQ, Op0,
7052 ConstantInt::get(Op1->getType(), *CmpC + 1));
7053 }
7054 break;
7055 }
7056 }
7057 }
7058
7059 // Based on the range information we know about the LHS, see if we can
7060 // simplify this comparison. For example, (x&4) < 8 is always true.
7061 switch (Pred) {
7062 default:
7063 break;
7064 case ICmpInst::ICMP_EQ:
7065 case ICmpInst::ICMP_NE: {
7066 // If all bits are known zero except for one, then we know at most one bit
7067 // is set. If the comparison is against zero, then this is a check to see if
7068 // *that* bit is set.
7069 APInt Op0KnownZeroInverted = ~Op0Known.Zero;
7070 if (Op1Known.isZero()) {
7071 // If the LHS is an AND with the same constant, look through it.
7072 Value *LHS = nullptr;
7073 const APInt *LHSC;
7074 if (!match(Op0, m_And(m_Value(LHS), m_APInt(LHSC))) ||
7075 *LHSC != Op0KnownZeroInverted)
7076 LHS = Op0;
7077
7078 Value *X;
7079 const APInt *C1;
7080 if (match(LHS, m_Shl(m_Power2(C1), m_Value(X)))) {
7081 Type *XTy = X->getType();
7082 unsigned Log2C1 = C1->countr_zero();
7083 APInt C2 = Op0KnownZeroInverted;
7084 APInt C2Pow2 = (C2 & ~(*C1 - 1)) + *C1;
7085 if (C2Pow2.isPowerOf2()) {
7086 // iff (C1 is pow2) & ((C2 & ~(C1-1)) + C1) is pow2):
7087 // ((C1 << X) & C2) == 0 -> X >= (Log2(C2+C1) - Log2(C1))
7088 // ((C1 << X) & C2) != 0 -> X < (Log2(C2+C1) - Log2(C1))
7089 unsigned Log2C2 = C2Pow2.countr_zero();
7090 auto *CmpC = ConstantInt::get(XTy, Log2C2 - Log2C1);
7091 auto NewPred =
7093 return new ICmpInst(NewPred, X, CmpC);
7094 }
7095 }
7096 }
7097
7098 // Op0 eq C_Pow2 -> Op0 ne 0 if Op0 is known to be C_Pow2 or zero.
7099 if (Op1Known.isConstant() && Op1Known.getConstant().isPowerOf2() &&
7100 (Op0Known & Op1Known) == Op0Known)
7101 return new ICmpInst(CmpInst::getInversePredicate(Pred), Op0,
7102 ConstantInt::getNullValue(Op1->getType()));
7103 break;
7104 }
7105 case ICmpInst::ICMP_SGE:
7106 if (Op1Min == Op0Max) // A >=s B -> A == B if max(A) == min(B)
7107 return new ICmpInst(ICmpInst::ICMP_EQ, Op0, Op1);
7108 break;
7109 case ICmpInst::ICMP_SLE:
7110 if (Op1Max == Op0Min) // A <=s B -> A == B if min(A) == max(B)
7111 return new ICmpInst(ICmpInst::ICMP_EQ, Op0, Op1);
7112 break;
7113 case ICmpInst::ICMP_UGE:
7114 if (Op1Min == Op0Max) // A >=u B -> A == B if max(A) == min(B)
7115 return new ICmpInst(ICmpInst::ICMP_EQ, Op0, Op1);
7116 break;
7117 case ICmpInst::ICMP_ULE:
7118 if (Op1Max == Op0Min) // A <=u B -> A == B if min(A) == max(B)
7119 return new ICmpInst(ICmpInst::ICMP_EQ, Op0, Op1);
7120 break;
7121 }
7122
7123 // Turn a signed comparison into an unsigned one if both operands are known to
7124 // have the same sign. Set samesign if possible (except for equality
7125 // predicates).
7126 if ((I.isSigned() || (I.isUnsigned() && !I.hasSameSign())) &&
7127 ((Op0Known.Zero.isNegative() && Op1Known.Zero.isNegative()) ||
7128 (Op0Known.One.isNegative() && Op1Known.One.isNegative()))) {
7129 I.setPredicate(I.getUnsignedPredicate());
7130 I.setSameSign();
7131 return &I;
7132 }
7133
7134 return nullptr;
7135}
7136
7137/// If one operand of an icmp is effectively a bool (value range of {0,1}),
7138/// then try to reduce patterns based on that limit.
7140 Value *X, *Y;
7141 CmpPredicate Pred;
7142
7143 // X must be 0 and bool must be true for "ULT":
7144 // X <u (zext i1 Y) --> (X == 0) & Y
7145 if (match(&I, m_c_ICmp(Pred, m_Value(X), m_OneUse(m_ZExt(m_Value(Y))))) &&
7146 Y->getType()->isIntOrIntVectorTy(1) && Pred == ICmpInst::ICMP_ULT)
7147 return BinaryOperator::CreateAnd(Builder.CreateIsNull(X), Y);
7148
7149 // X must be 0 or bool must be true for "ULE":
7150 // X <=u (sext i1 Y) --> (X == 0) | Y
7151 if (match(&I, m_c_ICmp(Pred, m_Value(X), m_OneUse(m_SExt(m_Value(Y))))) &&
7152 Y->getType()->isIntOrIntVectorTy(1) && Pred == ICmpInst::ICMP_ULE)
7153 return BinaryOperator::CreateOr(Builder.CreateIsNull(X), Y);
7154
7155 // icmp eq/ne X, (zext/sext (icmp eq/ne X, C))
7156 CmpPredicate Pred1, Pred2;
7157 const APInt *C;
7158 Instruction *ExtI;
7159 if (match(&I, m_c_ICmp(Pred1, m_Value(X),
7162 m_APInt(C)))))) &&
7163 ICmpInst::isEquality(Pred1) && ICmpInst::isEquality(Pred2)) {
7164 bool IsSExt = ExtI->getOpcode() == Instruction::SExt;
7165 bool HasOneUse = ExtI->hasOneUse() && ExtI->getOperand(0)->hasOneUse();
7166 auto CreateRangeCheck = [&] {
7167 Value *CmpV1 =
7168 Builder.CreateICmp(Pred1, X, Constant::getNullValue(X->getType()));
7169 Value *CmpV2 = Builder.CreateICmp(
7170 Pred1, X, ConstantInt::getSigned(X->getType(), IsSExt ? -1 : 1));
7172 Pred1 == ICmpInst::ICMP_EQ ? Instruction::Or : Instruction::And,
7173 CmpV1, CmpV2);
7174 };
7175 if (C->isZero()) {
7176 if (Pred2 == ICmpInst::ICMP_EQ) {
7177 // icmp eq X, (zext/sext (icmp eq X, 0)) --> false
7178 // icmp ne X, (zext/sext (icmp eq X, 0)) --> true
7179 return replaceInstUsesWith(
7180 I, ConstantInt::getBool(I.getType(), Pred1 == ICmpInst::ICMP_NE));
7181 } else if (!IsSExt || HasOneUse) {
7182 // icmp eq X, (zext (icmp ne X, 0)) --> X == 0 || X == 1
7183 // icmp ne X, (zext (icmp ne X, 0)) --> X != 0 && X != 1
7184 // icmp eq X, (sext (icmp ne X, 0)) --> X == 0 || X == -1
7185 // icmp ne X, (sext (icmp ne X, 0)) --> X != 0 && X != -1
7186 return CreateRangeCheck();
7187 }
7188 } else if (IsSExt ? C->isAllOnes() : C->isOne()) {
7189 if (Pred2 == ICmpInst::ICMP_NE) {
7190 // icmp eq X, (zext (icmp ne X, 1)) --> false
7191 // icmp ne X, (zext (icmp ne X, 1)) --> true
7192 // icmp eq X, (sext (icmp ne X, -1)) --> false
7193 // icmp ne X, (sext (icmp ne X, -1)) --> true
7194 return replaceInstUsesWith(
7195 I, ConstantInt::getBool(I.getType(), Pred1 == ICmpInst::ICMP_NE));
7196 } else if (!IsSExt || HasOneUse) {
7197 // icmp eq X, (zext (icmp eq X, 1)) --> X == 0 || X == 1
7198 // icmp ne X, (zext (icmp eq X, 1)) --> X != 0 && X != 1
7199 // icmp eq X, (sext (icmp eq X, -1)) --> X == 0 || X == -1
7200 // icmp ne X, (sext (icmp eq X, -1)) --> X != 0 && X == -1
7201 return CreateRangeCheck();
7202 }
7203 } else {
7204 // when C != 0 && C != 1:
7205 // icmp eq X, (zext (icmp eq X, C)) --> icmp eq X, 0
7206 // icmp eq X, (zext (icmp ne X, C)) --> icmp eq X, 1
7207 // icmp ne X, (zext (icmp eq X, C)) --> icmp ne X, 0
7208 // icmp ne X, (zext (icmp ne X, C)) --> icmp ne X, 1
7209 // when C != 0 && C != -1:
7210 // icmp eq X, (sext (icmp eq X, C)) --> icmp eq X, 0
7211 // icmp eq X, (sext (icmp ne X, C)) --> icmp eq X, -1
7212 // icmp ne X, (sext (icmp eq X, C)) --> icmp ne X, 0
7213 // icmp ne X, (sext (icmp ne X, C)) --> icmp ne X, -1
7214 return ICmpInst::Create(
7215 Instruction::ICmp, Pred1, X,
7216 ConstantInt::getSigned(X->getType(), Pred2 == ICmpInst::ICMP_NE
7217 ? (IsSExt ? -1 : 1)
7218 : 0));
7219 }
7220 }
7221
7222 return nullptr;
7223}
7224
7225/// If we have an icmp le or icmp ge instruction with a constant operand, turn
7226/// it into the appropriate icmp lt or icmp gt instruction. This transform
7227/// allows them to be folded in visitICmpInst.
7229 ICmpInst::Predicate Pred = I.getPredicate();
7230 if (ICmpInst::isEquality(Pred) || !ICmpInst::isIntPredicate(Pred) ||
7232 return nullptr;
7233
7234 Value *Op0 = I.getOperand(0);
7235 Value *Op1 = I.getOperand(1);
7236 auto *Op1C = dyn_cast<Constant>(Op1);
7237 if (!Op1C)
7238 return nullptr;
7239
7240 auto FlippedStrictness = getFlippedStrictnessPredicateAndConstant(Pred, Op1C);
7241 if (!FlippedStrictness)
7242 return nullptr;
7243
7244 return new ICmpInst(FlippedStrictness->first, Op0, FlippedStrictness->second);
7245}
7246
7247/// If we have a comparison with a non-canonical predicate, if we can update
7248/// all the users, invert the predicate and adjust all the users.
7250 // Is the predicate already canonical?
7251 CmpInst::Predicate Pred = I.getPredicate();
7253 return nullptr;
7254
7255 // Can all users be adjusted to predicate inversion?
7256 if (!InstCombiner::canFreelyInvertAllUsersOf(&I, /*IgnoredUser=*/nullptr))
7257 return nullptr;
7258
7259 // Ok, we can canonicalize comparison!
7260 // Let's first invert the comparison's predicate.
7261 I.setPredicate(CmpInst::getInversePredicate(Pred));
7262 I.setName(I.getName() + ".not");
7263
7264 // And, adapt users.
7266
7267 return &I;
7268}
7269
7270/// Integer compare with boolean values can always be turned into bitwise ops.
7272 InstCombiner::BuilderTy &Builder) {
7273 Value *A = I.getOperand(0), *B = I.getOperand(1);
7274 assert(A->getType()->isIntOrIntVectorTy(1) && "Bools only");
7275
7276 // A boolean compared to true/false can be simplified to Op0/true/false in
7277 // 14 out of the 20 (10 predicates * 2 constants) possible combinations.
7278 // Cases not handled by InstSimplify are always 'not' of Op0.
7279 if (match(B, m_Zero())) {
7280 switch (I.getPredicate()) {
7281 case CmpInst::ICMP_EQ: // A == 0 -> !A
7282 case CmpInst::ICMP_ULE: // A <=u 0 -> !A
7283 case CmpInst::ICMP_SGE: // A >=s 0 -> !A
7285 default:
7286 llvm_unreachable("ICmp i1 X, C not simplified as expected.");
7287 }
7288 } else if (match(B, m_One())) {
7289 switch (I.getPredicate()) {
7290 case CmpInst::ICMP_NE: // A != 1 -> !A
7291 case CmpInst::ICMP_ULT: // A <u 1 -> !A
7292 case CmpInst::ICMP_SGT: // A >s -1 -> !A
7294 default:
7295 llvm_unreachable("ICmp i1 X, C not simplified as expected.");
7296 }
7297 }
7298
7299 switch (I.getPredicate()) {
7300 default:
7301 llvm_unreachable("Invalid icmp instruction!");
7302 case ICmpInst::ICMP_EQ:
7303 // icmp eq i1 A, B -> ~(A ^ B)
7304 return BinaryOperator::CreateNot(Builder.CreateXor(A, B));
7305
7306 case ICmpInst::ICMP_NE:
7307 // icmp ne i1 A, B -> A ^ B
7308 return BinaryOperator::CreateXor(A, B);
7309
7310 case ICmpInst::ICMP_UGT:
7311 // icmp ugt -> icmp ult
7312 std::swap(A, B);
7313 [[fallthrough]];
7314 case ICmpInst::ICMP_ULT:
7315 // icmp ult i1 A, B -> ~A & B
7316 return BinaryOperator::CreateAnd(Builder.CreateNot(A), B);
7317
7318 case ICmpInst::ICMP_SGT:
7319 // icmp sgt -> icmp slt
7320 std::swap(A, B);
7321 [[fallthrough]];
7322 case ICmpInst::ICMP_SLT:
7323 // icmp slt i1 A, B -> A & ~B
7324 return BinaryOperator::CreateAnd(Builder.CreateNot(B), A);
7325
7326 case ICmpInst::ICMP_UGE:
7327 // icmp uge -> icmp ule
7328 std::swap(A, B);
7329 [[fallthrough]];
7330 case ICmpInst::ICMP_ULE:
7331 // icmp ule i1 A, B -> ~A | B
7332 return BinaryOperator::CreateOr(Builder.CreateNot(A), B);
7333
7334 case ICmpInst::ICMP_SGE:
7335 // icmp sge -> icmp sle
7336 std::swap(A, B);
7337 [[fallthrough]];
7338 case ICmpInst::ICMP_SLE:
7339 // icmp sle i1 A, B -> A | ~B
7340 return BinaryOperator::CreateOr(Builder.CreateNot(B), A);
7341 }
7342}
7343
7344// Transform pattern like:
7345// (1 << Y) u<= X or ~(-1 << Y) u< X or ((1 << Y)+(-1)) u< X
7346// (1 << Y) u> X or ~(-1 << Y) u>= X or ((1 << Y)+(-1)) u>= X
7347// Into:
7348// (X l>> Y) != 0
7349// (X l>> Y) == 0
7351 InstCombiner::BuilderTy &Builder) {
7352 CmpPredicate Pred, NewPred;
7353 Value *X, *Y;
7354 if (match(&Cmp,
7355 m_c_ICmp(Pred, m_OneUse(m_Shl(m_One(), m_Value(Y))), m_Value(X)))) {
7356 switch (Pred) {
7357 case ICmpInst::ICMP_ULE:
7358 NewPred = ICmpInst::ICMP_NE;
7359 break;
7360 case ICmpInst::ICMP_UGT:
7361 NewPred = ICmpInst::ICMP_EQ;
7362 break;
7363 default:
7364 return nullptr;
7365 }
7366 } else if (match(&Cmp, m_c_ICmp(Pred,
7369 m_Add(m_Shl(m_One(), m_Value(Y)),
7370 m_AllOnes()))),
7371 m_Value(X)))) {
7372 // The variant with 'add' is not canonical, (the variant with 'not' is)
7373 // we only get it because it has extra uses, and can't be canonicalized,
7374
7375 switch (Pred) {
7376 case ICmpInst::ICMP_ULT:
7377 NewPred = ICmpInst::ICMP_NE;
7378 break;
7379 case ICmpInst::ICMP_UGE:
7380 NewPred = ICmpInst::ICMP_EQ;
7381 break;
7382 default:
7383 return nullptr;
7384 }
7385 } else
7386 return nullptr;
7387
7388 Value *NewX = Builder.CreateLShr(X, Y, X->getName() + ".highbits");
7389 Constant *Zero = Constant::getNullValue(NewX->getType());
7390 return CmpInst::Create(Instruction::ICmp, NewPred, NewX, Zero);
7391}
7392
7394 InstCombiner::BuilderTy &Builder) {
7395 const CmpInst::Predicate Pred = Cmp.getPredicate();
7396 Value *LHS = Cmp.getOperand(0), *RHS = Cmp.getOperand(1);
7397 Value *V1, *V2;
7398
7399 auto createCmpReverse = [&](CmpInst::Predicate Pred, Value *X, Value *Y) {
7400 Value *V = Builder.CreateCmp(Pred, X, Y, Cmp.getName());
7401 if (auto *I = dyn_cast<Instruction>(V))
7402 I->copyIRFlags(&Cmp);
7403 Module *M = Cmp.getModule();
7405 M, Intrinsic::vector_reverse, V->getType());
7406 return CallInst::Create(F, V);
7407 };
7408
7409 if (match(LHS, m_VecReverse(m_Value(V1)))) {
7410 // cmp Pred, rev(V1), rev(V2) --> rev(cmp Pred, V1, V2)
7411 if (match(RHS, m_VecReverse(m_Value(V2))) &&
7412 (LHS->hasOneUse() || RHS->hasOneUse()))
7413 return createCmpReverse(Pred, V1, V2);
7414
7415 // cmp Pred, rev(V1), RHSSplat --> rev(cmp Pred, V1, RHSSplat)
7416 if (LHS->hasOneUse() && isSplatValue(RHS))
7417 return createCmpReverse(Pred, V1, RHS);
7418 }
7419 // cmp Pred, LHSSplat, rev(V2) --> rev(cmp Pred, LHSSplat, V2)
7420 else if (isSplatValue(LHS) && match(RHS, m_OneUse(m_VecReverse(m_Value(V2)))))
7421 return createCmpReverse(Pred, LHS, V2);
7422
7423 ArrayRef<int> M;
7424 if (!match(LHS, m_Shuffle(m_Value(V1), m_Undef(), m_Mask(M))))
7425 return nullptr;
7426
7427 // If both arguments of the cmp are shuffles that use the same mask and
7428 // shuffle within a single vector, move the shuffle after the cmp:
7429 // cmp (shuffle V1, M), (shuffle V2, M) --> shuffle (cmp V1, V2), M
7430 Type *V1Ty = V1->getType();
7431 if (match(RHS, m_Shuffle(m_Value(V2), m_Undef(), m_SpecificMask(M))) &&
7432 V1Ty == V2->getType() && (LHS->hasOneUse() || RHS->hasOneUse())) {
7433 Value *NewCmp = Builder.CreateCmp(Pred, V1, V2);
7434 return new ShuffleVectorInst(NewCmp, M);
7435 }
7436
7437 // Try to canonicalize compare with splatted operand and splat constant.
7438 // TODO: We could generalize this for more than splats. See/use the code in
7439 // InstCombiner::foldVectorBinop().
7440 Constant *C;
7441 if (!LHS->hasOneUse() || !match(RHS, m_Constant(C)))
7442 return nullptr;
7443
7444 // Length-changing splats are ok, so adjust the constants as needed:
7445 // cmp (shuffle V1, M), C --> shuffle (cmp V1, C'), M
7446 Constant *ScalarC = C->getSplatValue(/* AllowPoison */ true);
7447 int MaskSplatIndex;
7448 if (ScalarC && match(M, m_SplatOrPoisonMask(MaskSplatIndex))) {
7449 // We allow poison in matching, but this transform removes it for safety.
7450 // Demanded elements analysis should be able to recover some/all of that.
7451 C = ConstantVector::getSplat(cast<VectorType>(V1Ty)->getElementCount(),
7452 ScalarC);
7453 SmallVector<int, 8> NewM(M.size(), MaskSplatIndex);
7454 Value *NewCmp = Builder.CreateCmp(Pred, V1, C);
7455 return new ShuffleVectorInst(NewCmp, NewM);
7456 }
7457
7458 return nullptr;
7459}
7460
7461// extract(uadd.with.overflow(A, B), 0) ult A
7462// -> extract(uadd.with.overflow(A, B), 1)
7464 CmpInst::Predicate Pred = I.getPredicate();
7465 Value *Op0 = I.getOperand(0), *Op1 = I.getOperand(1);
7466
7467 Value *UAddOv;
7468 Value *A, *B;
7469 auto UAddOvResultPat = m_ExtractValue<0>(
7471 if (match(Op0, UAddOvResultPat) &&
7472 ((Pred == ICmpInst::ICMP_ULT && (Op1 == A || Op1 == B)) ||
7473 (Pred == ICmpInst::ICMP_EQ && match(Op1, m_ZeroInt()) &&
7474 (match(A, m_One()) || match(B, m_One()))) ||
7475 (Pred == ICmpInst::ICMP_NE && match(Op1, m_AllOnes()) &&
7476 (match(A, m_AllOnes()) || match(B, m_AllOnes())))))
7477 // extract(uadd.with.overflow(A, B), 0) < A
7478 // extract(uadd.with.overflow(A, 1), 0) == 0
7479 // extract(uadd.with.overflow(A, -1), 0) != -1
7480 UAddOv = cast<ExtractValueInst>(Op0)->getAggregateOperand();
7481 else if (match(Op1, UAddOvResultPat) && Pred == ICmpInst::ICMP_UGT &&
7482 (Op0 == A || Op0 == B))
7483 // A > extract(uadd.with.overflow(A, B), 0)
7484 UAddOv = cast<ExtractValueInst>(Op1)->getAggregateOperand();
7485 else
7486 return nullptr;
7487
7488 return ExtractValueInst::Create(UAddOv, 1);
7489}
7490
7492 if (!I.getOperand(0)->getType()->isPointerTy() ||
7494 I.getParent()->getParent(),
7495 I.getOperand(0)->getType()->getPointerAddressSpace())) {
7496 return nullptr;
7497 }
7498 Instruction *Op;
7499 if (match(I.getOperand(0), m_Instruction(Op)) &&
7500 match(I.getOperand(1), m_Zero()) &&
7501 Op->isLaunderOrStripInvariantGroup()) {
7502 return ICmpInst::Create(Instruction::ICmp, I.getPredicate(),
7503 Op->getOperand(0), I.getOperand(1));
7504 }
7505 return nullptr;
7506}
7507
7509 IRBuilderBase &Builder) {
7510 if (!ICmpInst::isEquality(I.getPredicate()))
7511 return nullptr;
7512
7513 // The caller puts constants after non-constants.
7514 Value *Op = I.getOperand(0);
7515 Value *Const = I.getOperand(1);
7516
7517 // For Cond an equality condition, fold
7518 //
7519 // icmp (eq|ne) (vreduce_(or|and) Op), (Zero|AllOnes) ->
7520 // icmp (eq|ne) Op, (Zero|AllOnes)
7521 //
7522 // with a bitcast.
7523 Value *Vec;
7524 if ((match(Const, m_ZeroInt()) &&
7526 m_Value(Vec))))) ||
7527 (match(Const, m_AllOnes()) &&
7529 m_Value(Vec)))))) {
7530 auto *VecTy = dyn_cast<FixedVectorType>(Vec->getType());
7531 if (!VecTy)
7532 return nullptr;
7533 Type *VecEltTy = VecTy->getElementType();
7534 unsigned ScalarBW =
7535 DL.getTypeSizeInBits(VecEltTy) * VecTy->getNumElements();
7536 if (!DL.fitsInLegalInteger(ScalarBW))
7537 return nullptr;
7538 Type *ScalarTy = IntegerType::get(I.getContext(), ScalarBW);
7539 Value *NewConst = match(Const, m_ZeroInt())
7540 ? ConstantInt::get(ScalarTy, 0)
7541 : ConstantInt::getAllOnesValue(ScalarTy);
7542 return CmpInst::Create(Instruction::ICmp, I.getPredicate(),
7543 Builder.CreateBitCast(Vec, ScalarTy), NewConst);
7544 }
7545 return nullptr;
7546}
7547
7548/// This function folds patterns produced by lowering of reduce idioms, such as
7549/// llvm.vector.reduce.and which are lowered into instruction chains. This code
7550/// attempts to generate fewer number of scalar comparisons instead of vector
7551/// comparisons when possible.
7553 InstCombiner::BuilderTy &Builder,
7554 const DataLayout &DL) {
7555 if (I.getType()->isVectorTy())
7556 return nullptr;
7557 CmpPredicate OuterPred, InnerPred;
7558 Value *LHS, *RHS;
7559
7560 // Match lowering of @llvm.vector.reduce.and. Turn
7561 /// %vec_ne = icmp ne <8 x i8> %lhs, %rhs
7562 /// %scalar_ne = bitcast <8 x i1> %vec_ne to i8
7563 /// %res = icmp <pred> i8 %scalar_ne, 0
7564 ///
7565 /// into
7566 ///
7567 /// %lhs.scalar = bitcast <8 x i8> %lhs to i64
7568 /// %rhs.scalar = bitcast <8 x i8> %rhs to i64
7569 /// %res = icmp <pred> i64 %lhs.scalar, %rhs.scalar
7570 ///
7571 /// for <pred> in {ne, eq}.
7572 if (!match(&I, m_ICmp(OuterPred,
7574 m_ICmp(InnerPred, m_Value(LHS), m_Value(RHS))))),
7575 m_Zero())))
7576 return nullptr;
7577 auto *LHSTy = dyn_cast<FixedVectorType>(LHS->getType());
7578 if (!LHSTy || !LHSTy->getElementType()->isIntegerTy())
7579 return nullptr;
7580 unsigned NumBits =
7581 LHSTy->getNumElements() * LHSTy->getElementType()->getIntegerBitWidth();
7582 // TODO: Relax this to "not wider than max legal integer type"?
7583 if (!DL.isLegalInteger(NumBits))
7584 return nullptr;
7585
7586 if (ICmpInst::isEquality(OuterPred) && InnerPred == ICmpInst::ICMP_NE) {
7587 auto *ScalarTy = Builder.getIntNTy(NumBits);
7588 LHS = Builder.CreateBitCast(LHS, ScalarTy, LHS->getName() + ".scalar");
7589 RHS = Builder.CreateBitCast(RHS, ScalarTy, RHS->getName() + ".scalar");
7590 return ICmpInst::Create(Instruction::ICmp, OuterPred, LHS, RHS,
7591 I.getName());
7592 }
7593
7594 return nullptr;
7595}
7596
7597// This helper will be called with icmp operands in both orders.
7599 Value *Op0, Value *Op1,
7600 ICmpInst &CxtI) {
7601 // Try to optimize 'icmp GEP, P' or 'icmp P, GEP'.
7602 if (auto *GEP = dyn_cast<GEPOperator>(Op0))
7603 if (Instruction *NI = foldGEPICmp(GEP, Op1, Pred, CxtI))
7604 return NI;
7605
7606 if (auto *SI = dyn_cast<SelectInst>(Op0))
7607 if (Instruction *NI = foldSelectICmp(Pred, SI, Op1, CxtI))
7608 return NI;
7609
7610 if (auto *MinMax = dyn_cast<MinMaxIntrinsic>(Op0)) {
7611 if (Instruction *Res = foldICmpWithMinMax(CxtI, MinMax, Op1, Pred))
7612 return Res;
7613
7614 if (Instruction *Res = foldICmpWithClamp(CxtI, Op1, MinMax))
7615 return Res;
7616 }
7617
7618 {
7619 Value *X;
7620 const APInt *C;
7621 // icmp X+Cst, X
7622 if (match(Op0, m_Add(m_Value(X), m_APInt(C))) && Op1 == X)
7623 return foldICmpAddOpConst(X, *C, Pred);
7624 }
7625
7626 // abs(X) >= X --> true
7627 // abs(X) u<= X --> true
7628 // abs(X) < X --> false
7629 // abs(X) u> X --> false
7630 // abs(X) u>= X --> IsIntMinPosion ? `X > -1`: `X u<= INTMIN`
7631 // abs(X) <= X --> IsIntMinPosion ? `X > -1`: `X u<= INTMIN`
7632 // abs(X) == X --> IsIntMinPosion ? `X > -1`: `X u<= INTMIN`
7633 // abs(X) u< X --> IsIntMinPosion ? `X < 0` : `X > INTMIN`
7634 // abs(X) > X --> IsIntMinPosion ? `X < 0` : `X > INTMIN`
7635 // abs(X) != X --> IsIntMinPosion ? `X < 0` : `X > INTMIN`
7636 {
7637 Value *X;
7638 Constant *C;
7640 match(Op1, m_Specific(X))) {
7641 Value *NullValue = Constant::getNullValue(X->getType());
7642 Value *AllOnesValue = Constant::getAllOnesValue(X->getType());
7643 const APInt SMin =
7644 APInt::getSignedMinValue(X->getType()->getScalarSizeInBits());
7645 bool IsIntMinPosion = C->isAllOnesValue();
7646 switch (Pred) {
7647 case CmpInst::ICMP_ULE:
7648 case CmpInst::ICMP_SGE:
7649 return replaceInstUsesWith(CxtI, ConstantInt::getTrue(CxtI.getType()));
7650 case CmpInst::ICMP_UGT:
7651 case CmpInst::ICMP_SLT:
7653 case CmpInst::ICMP_UGE:
7654 case CmpInst::ICMP_SLE:
7655 case CmpInst::ICMP_EQ: {
7656 return replaceInstUsesWith(
7657 CxtI, IsIntMinPosion
7658 ? Builder.CreateICmpSGT(X, AllOnesValue)
7659 : Builder.CreateICmpULT(
7660 X, ConstantInt::get(X->getType(), SMin + 1)));
7661 }
7662 case CmpInst::ICMP_ULT:
7663 case CmpInst::ICMP_SGT:
7664 case CmpInst::ICMP_NE: {
7665 return replaceInstUsesWith(
7666 CxtI, IsIntMinPosion
7667 ? Builder.CreateICmpSLT(X, NullValue)
7668 : Builder.CreateICmpUGT(
7669 X, ConstantInt::get(X->getType(), SMin)));
7670 }
7671 default:
7672 llvm_unreachable("Invalid predicate!");
7673 }
7674 }
7675 }
7676
7677 const SimplifyQuery Q = SQ.getWithInstruction(&CxtI);
7678 if (Value *V = foldICmpWithLowBitMaskedVal(Pred, Op0, Op1, Q, *this))
7679 return replaceInstUsesWith(CxtI, V);
7680
7681 // Folding (X / Y) pred X => X swap(pred) 0 for constant Y other than 0 or 1
7682 auto CheckUGT1 = [](const APInt &Divisor) { return Divisor.ugt(1); };
7683 {
7684 if (match(Op0, m_UDiv(m_Specific(Op1), m_CheckedInt(CheckUGT1)))) {
7685 return new ICmpInst(ICmpInst::getSwappedPredicate(Pred), Op1,
7687 }
7688
7689 if (!ICmpInst::isUnsigned(Pred) &&
7690 match(Op0, m_SDiv(m_Specific(Op1), m_CheckedInt(CheckUGT1)))) {
7691 return new ICmpInst(ICmpInst::getSwappedPredicate(Pred), Op1,
7693 }
7694 }
7695
7696 // Another case of this fold is (X >> Y) pred X => X swap(pred) 0 if Y != 0
7697 auto CheckNE0 = [](const APInt &Shift) { return !Shift.isZero(); };
7698 {
7699 if (match(Op0, m_LShr(m_Specific(Op1), m_CheckedInt(CheckNE0)))) {
7700 return new ICmpInst(ICmpInst::getSwappedPredicate(Pred), Op1,
7702 }
7703
7704 if ((Pred == CmpInst::ICMP_SLT || Pred == CmpInst::ICMP_SGE) &&
7705 match(Op0, m_AShr(m_Specific(Op1), m_CheckedInt(CheckNE0)))) {
7706 return new ICmpInst(ICmpInst::getSwappedPredicate(Pred), Op1,
7708 }
7709 }
7710
7711 return nullptr;
7712}
7713
7715 bool Changed = false;
7716 const SimplifyQuery Q = SQ.getWithInstruction(&I);
7717 Value *Op0 = I.getOperand(0), *Op1 = I.getOperand(1);
7718 unsigned Op0Cplxity = getComplexity(Op0);
7719 unsigned Op1Cplxity = getComplexity(Op1);
7720
7721 /// Orders the operands of the compare so that they are listed from most
7722 /// complex to least complex. This puts constants before unary operators,
7723 /// before binary operators.
7724 if (Op0Cplxity < Op1Cplxity) {
7725 I.swapOperands();
7726 std::swap(Op0, Op1);
7727 Changed = true;
7728 }
7729
7730 if (Value *V = simplifyICmpInst(I.getCmpPredicate(), Op0, Op1, Q))
7731 return replaceInstUsesWith(I, V);
7732
7733 // Comparing -val or val with non-zero is the same as just comparing val
7734 // ie, abs(val) != 0 -> val != 0
7735 if (I.getPredicate() == ICmpInst::ICMP_NE && match(Op1, m_Zero())) {
7736 Value *Cond, *SelectTrue, *SelectFalse;
7737 if (match(Op0, m_Select(m_Value(Cond), m_Value(SelectTrue),
7738 m_Value(SelectFalse)))) {
7739 if (Value *V = dyn_castNegVal(SelectTrue)) {
7740 if (V == SelectFalse)
7741 return CmpInst::Create(Instruction::ICmp, I.getPredicate(), V, Op1);
7742 } else if (Value *V = dyn_castNegVal(SelectFalse)) {
7743 if (V == SelectTrue)
7744 return CmpInst::Create(Instruction::ICmp, I.getPredicate(), V, Op1);
7745 }
7746 }
7747 }
7748
7750 return Res;
7751
7752 if (Op0->getType()->isIntOrIntVectorTy(1))
7754 return Res;
7755
7757 return Res;
7758
7760 return Res;
7761
7763 return Res;
7764
7766 return Res;
7767
7769 return Res;
7770
7772 return Res;
7773
7775 return Res;
7776
7777 // Test if the ICmpInst instruction is used exclusively by a select as
7778 // part of a minimum or maximum operation. If so, refrain from doing
7779 // any other folding. This helps out other analyses which understand
7780 // non-obfuscated minimum and maximum idioms, such as ScalarEvolution
7781 // and CodeGen. And in this case, at least one of the comparison
7782 // operands has at least one user besides the compare (the select),
7783 // which would often largely negate the benefit of folding anyway.
7784 //
7785 // Do the same for the other patterns recognized by matchSelectPattern.
7786 if (I.hasOneUse())
7787 if (SelectInst *SI = dyn_cast<SelectInst>(I.user_back())) {
7788 Value *A, *B;
7790 if (SPR.Flavor != SPF_UNKNOWN)
7791 return nullptr;
7792 }
7793
7794 // Do this after checking for min/max to prevent infinite looping.
7795 if (Instruction *Res = foldICmpWithZero(I))
7796 return Res;
7797
7798 Value *X;
7799 const APInt *C;
7800 if (I.getPredicate() == ICmpInst::ICMP_UGT &&
7801 match(Op0, m_UMax(m_Value(X), m_APInt(C))) &&
7802 match(Op1, m_Not(m_Specific(X)))) {
7803 if (C->isNonNegative())
7804 return new ICmpInst(ICmpInst::ICMP_SLT, X,
7805 Constant::getNullValue(X->getType()));
7806 return new ICmpInst(ICmpInst::ICMP_UGT, X,
7807 ConstantInt::get(X->getType(), ~*C));
7808 }
7809
7810 if (I.getPredicate() == ICmpInst::ICMP_ULT &&
7811 match(Op0, m_UMax(m_Value(X), m_APInt(C))) &&
7812 match(Op1, m_Not(m_Specific(X)))) {
7813 if (C->isNonNegative())
7814 return new ICmpInst(ICmpInst::ICMP_SGT, X,
7815 Constant::getAllOnesValue(X->getType()));
7816 return new ICmpInst(ICmpInst::ICMP_ULT, X,
7817 ConstantInt::get(X->getType(), ~*C));
7818 }
7819
7820 // FIXME: We only do this after checking for min/max to prevent infinite
7821 // looping caused by a reverse canonicalization of these patterns for min/max.
7822 // FIXME: The organization of folds is a mess. These would naturally go into
7823 // canonicalizeCmpWithConstant(), but we can't move all of the above folds
7824 // down here after the min/max restriction.
7825 ICmpInst::Predicate Pred = I.getPredicate();
7826 if (match(Op1, m_APInt(C))) {
7827 // For i32: x >u 2147483647 -> x <s 0 -> true if sign bit set
7828 if (Pred == ICmpInst::ICMP_UGT && C->isMaxSignedValue()) {
7829 Constant *Zero = Constant::getNullValue(Op0->getType());
7830 return new ICmpInst(ICmpInst::ICMP_SLT, Op0, Zero);
7831 }
7832
7833 // For i32: x <u 2147483648 -> x >s -1 -> true if sign bit clear
7834 if (Pred == ICmpInst::ICMP_ULT && C->isMinSignedValue()) {
7836 return new ICmpInst(ICmpInst::ICMP_SGT, Op0, AllOnes);
7837 }
7838 }
7839
7840 // The folds in here may rely on wrapping flags and special constants, so
7841 // they can break up min/max idioms in some cases but not seemingly similar
7842 // patterns.
7843 // FIXME: It may be possible to enhance select folding to make this
7844 // unnecessary. It may also be moot if we canonicalize to min/max
7845 // intrinsics.
7846 if (Instruction *Res = foldICmpBinOp(I, Q))
7847 return Res;
7848
7850 return Res;
7851
7852 // Try to match comparison as a sign bit test. Intentionally do this after
7853 // foldICmpInstWithConstant() to potentially let other folds to happen first.
7854 if (Instruction *New = foldSignBitTest(I))
7855 return New;
7856
7857 if (auto *PN = dyn_cast<PHINode>(Op0))
7858 if (Instruction *NV = foldOpIntoPhi(I, PN))
7859 return NV;
7860 if (auto *PN = dyn_cast<PHINode>(Op1))
7861 if (Instruction *NV = foldOpIntoPhi(I, PN))
7862 return NV;
7863
7865 return Res;
7866
7867 if (Instruction *Res = foldICmpCommutative(I.getCmpPredicate(), Op0, Op1, I))
7868 return Res;
7869 if (Instruction *Res =
7870 foldICmpCommutative(I.getSwappedCmpPredicate(), Op1, Op0, I))
7871 return Res;
7872
7873 if (I.isCommutative()) {
7874 if (auto Pair = matchSymmetricPair(I.getOperand(0), I.getOperand(1))) {
7875 replaceOperand(I, 0, Pair->first);
7876 replaceOperand(I, 1, Pair->second);
7877 return &I;
7878 }
7879 }
7880
7881 // Fold icmp pred (select C1, TV1, FV1), (select C2, TV2, FV2)
7882 // when all select arms are constants, via truth table.
7884 return R;
7885
7886 // In case of a comparison with two select instructions having the same
7887 // condition, check whether one of the resulting branches can be simplified.
7888 // If so, just compare the other branch and select the appropriate result.
7889 // For example:
7890 // %tmp1 = select i1 %cmp, i32 %y, i32 %x
7891 // %tmp2 = select i1 %cmp, i32 %z, i32 %x
7892 // %cmp2 = icmp slt i32 %tmp2, %tmp1
7893 // The icmp will result false for the false value of selects and the result
7894 // will depend upon the comparison of true values of selects if %cmp is
7895 // true. Thus, transform this into:
7896 // %cmp = icmp slt i32 %y, %z
7897 // %sel = select i1 %cond, i1 %cmp, i1 false
7898 // This handles similar cases to transform.
7899 {
7900 Value *Cond, *A, *B, *C, *D;
7901 if (match(Op0, m_Select(m_Value(Cond), m_Value(A), m_Value(B))) &&
7903 (Op0->hasOneUse() || Op1->hasOneUse())) {
7904 // Check whether comparison of TrueValues can be simplified
7905 if (Value *Res = simplifyICmpInst(Pred, A, C, SQ)) {
7906 Value *NewICMP = Builder.CreateICmp(Pred, B, D);
7907 return SelectInst::Create(
7908 Cond, Res, NewICMP, /*NameStr=*/"", /*InsertBefore=*/nullptr,
7910 }
7911 // Check whether comparison of FalseValues can be simplified
7912 if (Value *Res = simplifyICmpInst(Pred, B, D, SQ)) {
7913 Value *NewICMP = Builder.CreateICmp(Pred, A, C);
7914 return SelectInst::Create(
7915 Cond, NewICMP, Res, /*NameStr=*/"", /*InsertBefore=*/nullptr,
7917 }
7918 }
7919 }
7920
7921 // icmp slt (sub nsw x, y), (add nsw x, y) --> icmp sgt y, 0
7922 // icmp ult (sub nuw x, y), (add nuw x, y) --> icmp ugt y, 0
7923 // icmp eq (sub nsw/nuw x, y), (add nsw/nuw x, y) --> icmp eq y, 0
7924 {
7925 Value *A, *B;
7926 CmpPredicate CmpPred;
7927 if (match(&I, m_c_ICmp(CmpPred, m_Sub(m_Value(A), m_Value(B)),
7929 auto *I0 = cast<OverflowingBinaryOperator>(Op0);
7930 auto *I1 = cast<OverflowingBinaryOperator>(Op1);
7931 bool I0NUW = I0->hasNoUnsignedWrap();
7932 bool I1NUW = I1->hasNoUnsignedWrap();
7933 bool I0NSW = I0->hasNoSignedWrap();
7934 bool I1NSW = I1->hasNoSignedWrap();
7935 if ((ICmpInst::isUnsigned(Pred) && I0NUW && I1NUW) ||
7936 (ICmpInst::isSigned(Pred) && I0NSW && I1NSW) ||
7937 (ICmpInst::isEquality(Pred) &&
7938 ((I0NUW || I0NSW) && (I1NUW || I1NSW)))) {
7939 return new ICmpInst(CmpPredicate::getSwapped(CmpPred), B,
7940 ConstantInt::get(Op0->getType(), 0));
7941 }
7942 }
7943 }
7944
7945 // Try to optimize equality comparisons against alloca-based pointers.
7946 if (Op0->getType()->isPointerTy() && I.isEquality()) {
7947 assert(Op1->getType()->isPointerTy() &&
7948 "Comparing pointer with non-pointer?");
7949 if (auto *Alloca = dyn_cast<AllocaInst>(getUnderlyingObject(Op0)))
7950 if (foldAllocaCmp(Alloca))
7951 return nullptr;
7952 if (auto *Alloca = dyn_cast<AllocaInst>(getUnderlyingObject(Op1)))
7953 if (foldAllocaCmp(Alloca))
7954 return nullptr;
7955 }
7956
7957 if (Instruction *Res = foldICmpBitCast(I))
7958 return Res;
7959
7960 // TODO: Hoist this above the min/max bailout.
7962 return R;
7963
7964 {
7965 Value *X, *Y;
7966 // Transform (X & ~Y) == 0 --> (X & Y) != 0
7967 // and (X & ~Y) != 0 --> (X & Y) == 0
7968 // if A is a power of 2.
7969 if (match(Op0, m_And(m_Value(X), m_Not(m_Value(Y)))) &&
7970 match(Op1, m_Zero()) && isKnownToBeAPowerOfTwo(X, false, &I) &&
7971 I.isEquality())
7972 return new ICmpInst(I.getInversePredicate(), Builder.CreateAnd(X, Y),
7973 Op1);
7974
7975 // Op0 pred Op1 -> ~Op1 pred ~Op0, if this allows us to drop an instruction.
7976 if (Op0->getType()->isIntOrIntVectorTy()) {
7977 bool ConsumesOp0, ConsumesOp1;
7978 if (isFreeToInvert(Op0, Op0->hasOneUse(), ConsumesOp0) &&
7979 isFreeToInvert(Op1, Op1->hasOneUse(), ConsumesOp1) &&
7980 (ConsumesOp0 || ConsumesOp1)) {
7981 Value *InvOp0 = getFreelyInverted(Op0, Op0->hasOneUse(), &Builder);
7982 Value *InvOp1 = getFreelyInverted(Op1, Op1->hasOneUse(), &Builder);
7983 assert(InvOp0 && InvOp1 &&
7984 "Mismatch between isFreeToInvert and getFreelyInverted");
7985 return new ICmpInst(I.getSwappedPredicate(), InvOp0, InvOp1);
7986 }
7987 }
7988
7989 Instruction *AddI = nullptr;
7991 m_Instruction(AddI))) &&
7992 isa<IntegerType>(X->getType())) {
7993 Value *Result;
7994 Constant *Overflow;
7995 // m_UAddWithOverflow can match patterns that do not include an explicit
7996 // "add" instruction, so check the opcode of the matched op.
7997 if (AddI->getOpcode() == Instruction::Add &&
7998 OptimizeOverflowCheck(Instruction::Add, /*Signed*/ false, X, Y, *AddI,
7999 Result, Overflow)) {
8000 replaceInstUsesWith(*AddI, Result);
8001 eraseInstFromFunction(*AddI);
8002 return replaceInstUsesWith(I, Overflow);
8003 }
8004 }
8005
8006 // (zext X) * (zext Y) --> llvm.umul.with.overflow.
8007 if (match(Op0, m_NUWMul(m_ZExt(m_Value(X)), m_ZExt(m_Value(Y)))) &&
8008 match(Op1, m_APInt(C))) {
8009 if (Instruction *R = processUMulZExtIdiom(I, Op0, C, *this))
8010 return R;
8011 }
8012
8013 // Signbit test folds
8014 // Fold (X u>> BitWidth - 1 Pred ZExt(i1)) --> X s< 0 Pred i1
8015 // Fold (X s>> BitWidth - 1 Pred SExt(i1)) --> X s< 0 Pred i1
8016 Instruction *ExtI;
8017 if ((I.isUnsigned() || I.isEquality()) &&
8018 match(Op1,
8020 Y->getType()->getScalarSizeInBits() == 1 &&
8021 (Op0->hasOneUse() || Op1->hasOneUse())) {
8022 unsigned OpWidth = Op0->getType()->getScalarSizeInBits();
8023 Instruction *ShiftI;
8024 if (match(Op0, m_CombineAnd(m_Instruction(ShiftI),
8026 OpWidth - 1))))) {
8027 unsigned ExtOpc = ExtI->getOpcode();
8028 unsigned ShiftOpc = ShiftI->getOpcode();
8029 if ((ExtOpc == Instruction::ZExt && ShiftOpc == Instruction::LShr) ||
8030 (ExtOpc == Instruction::SExt && ShiftOpc == Instruction::AShr)) {
8031 Value *SLTZero =
8032 Builder.CreateICmpSLT(X, Constant::getNullValue(X->getType()));
8033 Value *Cmp = Builder.CreateICmp(Pred, SLTZero, Y, I.getName());
8034 return replaceInstUsesWith(I, Cmp);
8035 }
8036 }
8037 }
8038 }
8039
8040 if (Instruction *Res = foldICmpEquality(I))
8041 return Res;
8042
8044 return Res;
8045
8046 if (Instruction *Res = foldICmpOfUAddOv(I))
8047 return Res;
8048
8050 return Res;
8051
8052 // The 'cmpxchg' instruction returns an aggregate containing the old value and
8053 // an i1 which indicates whether or not we successfully did the swap.
8054 //
8055 // Replace comparisons between the old value and the expected value with the
8056 // indicator that 'cmpxchg' returns.
8057 //
8058 // N.B. This transform is only valid when the 'cmpxchg' is not permitted to
8059 // spuriously fail. In those cases, the old value may equal the expected
8060 // value but it is possible for the swap to not occur.
8061 if (I.getPredicate() == ICmpInst::ICMP_EQ)
8062 if (auto *EVI = dyn_cast<ExtractValueInst>(Op0))
8063 if (auto *ACXI = dyn_cast<AtomicCmpXchgInst>(EVI->getAggregateOperand()))
8064 if (EVI->getIndices()[0] == 0 && ACXI->getCompareOperand() == Op1 &&
8065 !ACXI->isWeak())
8066 return ExtractValueInst::Create(ACXI, 1);
8067
8069 return Res;
8070
8071 if (I.getType()->isVectorTy())
8072 if (Instruction *Res = foldVectorCmp(I, Builder))
8073 return Res;
8074
8076 return Res;
8077
8079 return Res;
8080
8081 {
8082 Value *A;
8083 const APInt *C1, *C2;
8084 ICmpInst::Predicate Pred = I.getPredicate();
8085 if (ICmpInst::isEquality(Pred)) {
8086 // sext(a) & c1 == c2 --> a & c3 == trunc(c2)
8087 // sext(a) & c1 != c2 --> a & c3 != trunc(c2)
8088 if (match(Op0, m_And(m_SExt(m_Value(A)), m_APInt(C1))) &&
8089 match(Op1, m_APInt(C2))) {
8090 Type *InputTy = A->getType();
8091 unsigned InputBitWidth = InputTy->getScalarSizeInBits();
8092 // c2 must be non-negative at the bitwidth of a.
8093 if (C2->getActiveBits() < InputBitWidth) {
8094 APInt TruncC1 = C1->trunc(InputBitWidth);
8095 // Check if there are 1s in C1 high bits of size InputBitWidth.
8096 if (C1->uge(APInt::getOneBitSet(C1->getBitWidth(), InputBitWidth)))
8097 TruncC1.setBit(InputBitWidth - 1);
8098 Value *AndInst = Builder.CreateAnd(A, TruncC1);
8099 return new ICmpInst(
8100 Pred, AndInst,
8101 ConstantInt::get(InputTy, C2->trunc(InputBitWidth)));
8102 }
8103 }
8104 }
8105 }
8106
8107 return Changed ? &I : nullptr;
8108}
8109
8110/// Fold fcmp ([us]itofp x, cst) if possible.
8112 Instruction *LHSI,
8113 Constant *RHSC) {
8114 const APFloat *RHS;
8115 if (!match(RHSC, m_APFloat(RHS)))
8116 return nullptr;
8117
8118 // Get the width of the mantissa. We don't want to hack on conversions that
8119 // might lose information from the integer, e.g. "i64 -> float"
8120 int MantissaWidth = LHSI->getType()->getFPMantissaWidth();
8121 if (MantissaWidth == -1)
8122 return nullptr; // Unknown.
8123
8124 Type *IntTy = LHSI->getOperand(0)->getType();
8125 unsigned IntWidth = IntTy->getScalarSizeInBits();
8126 bool LHSUnsigned = isa<UIToFPInst>(LHSI);
8127
8128 if (I.isEquality()) {
8129 FCmpInst::Predicate P = I.getPredicate();
8130 bool IsExact = false;
8131 APSInt RHSCvt(IntWidth, LHSUnsigned);
8132 RHS->convertToInteger(RHSCvt, APFloat::rmNearestTiesToEven, &IsExact);
8133
8134 // If the floating point constant isn't an integer value, we know if we will
8135 // ever compare equal / not equal to it.
8136 if (!IsExact) {
8137 // TODO: Can never be -0.0 and other non-representable values
8138 APFloat RHSRoundInt(*RHS);
8140 if (*RHS != RHSRoundInt) {
8142 return replaceInstUsesWith(I, ConstantInt::getFalse(I.getType()));
8143
8145 return replaceInstUsesWith(I, ConstantInt::getTrue(I.getType()));
8146 }
8147 }
8148
8149 // TODO: If the constant is exactly representable, is it always OK to do
8150 // equality compares as integer?
8151 }
8152
8153 // Check to see that the input is converted from an integer type that is small
8154 // enough that preserves all bits. TODO: check here for "known" sign bits.
8155 // This would allow us to handle (fptosi (x >>s 62) to float) if x is i64 f.e.
8156
8157 // Following test does NOT adjust IntWidth downwards for signed inputs,
8158 // because the most negative value still requires all the mantissa bits
8159 // to distinguish it from one less than that value.
8160 if ((int)IntWidth > MantissaWidth) {
8161 // Conversion would lose accuracy. Check if loss can impact comparison.
8162 int Exp = ilogb(*RHS);
8163 if (Exp == APFloat::IEK_Inf) {
8164 int MaxExponent = ilogb(APFloat::getLargest(RHS->getSemantics()));
8165 if (MaxExponent < (int)IntWidth - !LHSUnsigned)
8166 // Conversion could create infinity.
8167 return nullptr;
8168 } else {
8169 // Note that if RHS is zero or NaN, then Exp is negative
8170 // and first condition is trivially false.
8171 if (MantissaWidth <= Exp && Exp <= (int)IntWidth - !LHSUnsigned)
8172 // Conversion could affect comparison.
8173 return nullptr;
8174 }
8175 }
8176
8177 // Otherwise, we can potentially simplify the comparison. We know that it
8178 // will always come through as an integer value and we know the constant is
8179 // not a NAN (it would have been previously simplified).
8180 assert(!RHS->isNaN() && "NaN comparison not already folded!");
8181
8183 switch (I.getPredicate()) {
8184 default:
8185 llvm_unreachable("Unexpected predicate!");
8186 case FCmpInst::FCMP_UEQ:
8187 case FCmpInst::FCMP_OEQ:
8188 Pred = ICmpInst::ICMP_EQ;
8189 break;
8190 case FCmpInst::FCMP_UGT:
8191 case FCmpInst::FCMP_OGT:
8192 Pred = LHSUnsigned ? ICmpInst::ICMP_UGT : ICmpInst::ICMP_SGT;
8193 break;
8194 case FCmpInst::FCMP_UGE:
8195 case FCmpInst::FCMP_OGE:
8196 Pred = LHSUnsigned ? ICmpInst::ICMP_UGE : ICmpInst::ICMP_SGE;
8197 break;
8198 case FCmpInst::FCMP_ULT:
8199 case FCmpInst::FCMP_OLT:
8200 Pred = LHSUnsigned ? ICmpInst::ICMP_ULT : ICmpInst::ICMP_SLT;
8201 break;
8202 case FCmpInst::FCMP_ULE:
8203 case FCmpInst::FCMP_OLE:
8204 Pred = LHSUnsigned ? ICmpInst::ICMP_ULE : ICmpInst::ICMP_SLE;
8205 break;
8206 case FCmpInst::FCMP_UNE:
8207 case FCmpInst::FCMP_ONE:
8208 Pred = ICmpInst::ICMP_NE;
8209 break;
8210 case FCmpInst::FCMP_ORD:
8211 return replaceInstUsesWith(I, ConstantInt::getTrue(I.getType()));
8212 case FCmpInst::FCMP_UNO:
8213 return replaceInstUsesWith(I, ConstantInt::getFalse(I.getType()));
8214 }
8215
8216 // Now we know that the APFloat is a normal number, zero or inf.
8217
8218 // See if the FP constant is too large for the integer. For example,
8219 // comparing an i8 to 300.0.
8220 if (!LHSUnsigned) {
8221 // If the RHS value is > SignedMax, fold the comparison. This handles +INF
8222 // and large values.
8223 APFloat SMax(RHS->getSemantics());
8224 SMax.convertFromAPInt(APInt::getSignedMaxValue(IntWidth), true,
8226 if (SMax < *RHS) { // smax < 13123.0
8227 if (Pred == ICmpInst::ICMP_NE || Pred == ICmpInst::ICMP_SLT ||
8228 Pred == ICmpInst::ICMP_SLE)
8229 return replaceInstUsesWith(I, ConstantInt::getTrue(I.getType()));
8230 return replaceInstUsesWith(I, ConstantInt::getFalse(I.getType()));
8231 }
8232 } else {
8233 // If the RHS value is > UnsignedMax, fold the comparison. This handles
8234 // +INF and large values.
8235 APFloat UMax(RHS->getSemantics());
8236 UMax.convertFromAPInt(APInt::getMaxValue(IntWidth), false,
8238 if (UMax < *RHS) { // umax < 13123.0
8239 if (Pred == ICmpInst::ICMP_NE || Pred == ICmpInst::ICMP_ULT ||
8240 Pred == ICmpInst::ICMP_ULE)
8241 return replaceInstUsesWith(I, ConstantInt::getTrue(I.getType()));
8242 return replaceInstUsesWith(I, ConstantInt::getFalse(I.getType()));
8243 }
8244 }
8245
8246 if (!LHSUnsigned) {
8247 // See if the RHS value is < SignedMin.
8248 APFloat SMin(RHS->getSemantics());
8249 SMin.convertFromAPInt(APInt::getSignedMinValue(IntWidth), true,
8251 if (SMin > *RHS) { // smin > 12312.0
8252 if (Pred == ICmpInst::ICMP_NE || Pred == ICmpInst::ICMP_SGT ||
8253 Pred == ICmpInst::ICMP_SGE)
8254 return replaceInstUsesWith(I, ConstantInt::getTrue(I.getType()));
8255 return replaceInstUsesWith(I, ConstantInt::getFalse(I.getType()));
8256 }
8257 } else {
8258 // See if the RHS value is < UnsignedMin.
8259 APFloat UMin(RHS->getSemantics());
8260 UMin.convertFromAPInt(APInt::getMinValue(IntWidth), false,
8262 if (UMin > *RHS) { // umin > 12312.0
8263 if (Pred == ICmpInst::ICMP_NE || Pred == ICmpInst::ICMP_UGT ||
8264 Pred == ICmpInst::ICMP_UGE)
8265 return replaceInstUsesWith(I, ConstantInt::getTrue(I.getType()));
8266 return replaceInstUsesWith(I, ConstantInt::getFalse(I.getType()));
8267 }
8268 }
8269
8270 // Okay, now we know that the FP constant fits in the range [SMIN, SMAX] or
8271 // [0, UMAX], but it may still be fractional. Check whether this is the case
8272 // using the IsExact flag.
8273 // Don't do this for zero, because -0.0 is not fractional.
8274 APSInt RHSInt(IntWidth, LHSUnsigned);
8275 bool IsExact;
8276 RHS->convertToInteger(RHSInt, APFloat::rmTowardZero, &IsExact);
8277 if (!RHS->isZero()) {
8278 if (!IsExact) {
8279 // If we had a comparison against a fractional value, we have to adjust
8280 // the compare predicate and sometimes the value. RHSC is rounded towards
8281 // zero at this point.
8282 switch (Pred) {
8283 default:
8284 llvm_unreachable("Unexpected integer comparison!");
8285 case ICmpInst::ICMP_NE: // (float)int != 4.4 --> true
8286 return replaceInstUsesWith(I, ConstantInt::getTrue(I.getType()));
8287 case ICmpInst::ICMP_EQ: // (float)int == 4.4 --> false
8288 return replaceInstUsesWith(I, ConstantInt::getFalse(I.getType()));
8289 case ICmpInst::ICMP_ULE:
8290 // (float)int <= 4.4 --> int <= 4
8291 // (float)int <= -4.4 --> false
8292 if (RHS->isNegative())
8293 return replaceInstUsesWith(I, ConstantInt::getFalse(I.getType()));
8294 break;
8295 case ICmpInst::ICMP_SLE:
8296 // (float)int <= 4.4 --> int <= 4
8297 // (float)int <= -4.4 --> int < -4
8298 if (RHS->isNegative())
8299 Pred = ICmpInst::ICMP_SLT;
8300 break;
8301 case ICmpInst::ICMP_ULT:
8302 // (float)int < -4.4 --> false
8303 // (float)int < 4.4 --> int <= 4
8304 if (RHS->isNegative())
8305 return replaceInstUsesWith(I, ConstantInt::getFalse(I.getType()));
8306 Pred = ICmpInst::ICMP_ULE;
8307 break;
8308 case ICmpInst::ICMP_SLT:
8309 // (float)int < -4.4 --> int < -4
8310 // (float)int < 4.4 --> int <= 4
8311 if (!RHS->isNegative())
8312 Pred = ICmpInst::ICMP_SLE;
8313 break;
8314 case ICmpInst::ICMP_UGT:
8315 // (float)int > 4.4 --> int > 4
8316 // (float)int > -4.4 --> true
8317 if (RHS->isNegative())
8318 return replaceInstUsesWith(I, ConstantInt::getTrue(I.getType()));
8319 break;
8320 case ICmpInst::ICMP_SGT:
8321 // (float)int > 4.4 --> int > 4
8322 // (float)int > -4.4 --> int >= -4
8323 if (RHS->isNegative())
8324 Pred = ICmpInst::ICMP_SGE;
8325 break;
8326 case ICmpInst::ICMP_UGE:
8327 // (float)int >= -4.4 --> true
8328 // (float)int >= 4.4 --> int > 4
8329 if (RHS->isNegative())
8330 return replaceInstUsesWith(I, ConstantInt::getTrue(I.getType()));
8331 Pred = ICmpInst::ICMP_UGT;
8332 break;
8333 case ICmpInst::ICMP_SGE:
8334 // (float)int >= -4.4 --> int >= -4
8335 // (float)int >= 4.4 --> int > 4
8336 if (!RHS->isNegative())
8337 Pred = ICmpInst::ICMP_SGT;
8338 break;
8339 }
8340 }
8341 }
8342
8343 // Lower this FP comparison into an appropriate integer version of the
8344 // comparison.
8345 return new ICmpInst(Pred, LHSI->getOperand(0),
8346 ConstantInt::get(LHSI->getOperand(0)->getType(), RHSInt));
8347}
8348
8349/// Fold fcmp/icmp pred (select C1, TV1, FV1), (select C2, TV2, FV2)
8350/// where all true/false values are constants that allow the compare to be
8351/// constant-folded for every combination of C1 and C2.
8352/// We compute a 4-entry truth table and use createLogicFromTable to
8353/// synthesize a boolean expression of C1 and C2.
8355 Value *Op0 = I.getOperand(0), *Op1 = I.getOperand(1);
8356 Value *C1, *C2;
8357 Constant *TV1, *FV1, *TV2, *FV2;
8358
8359 if (!match(Op0, m_Select(m_Value(C1), m_Constant(TV1), m_Constant(FV1))) ||
8360 !match(Op1, m_Select(m_Value(C2), m_Constant(TV2), m_Constant(FV2))))
8361 return nullptr;
8362
8363 if (I.getType() != C1->getType() || I.getType() != C2->getType())
8364 return nullptr;
8365
8366 unsigned Pred = I.getPredicate();
8367 const DataLayout &DL = I.getDataLayout();
8368
8369 Constant *Res00 = ConstantFoldCompareInstOperands(Pred, FV1, FV2, DL);
8370 Constant *Res01 = ConstantFoldCompareInstOperands(Pred, FV1, TV2, DL);
8371 Constant *Res10 = ConstantFoldCompareInstOperands(Pred, TV1, FV2, DL);
8372 Constant *Res11 = ConstantFoldCompareInstOperands(Pred, TV1, TV2, DL);
8373
8374 if (!Res00 || !Res01 || !Res10 || !Res11)
8375 return nullptr;
8376
8377 if ((!Res00->isNullValue() && !Res00->isAllOnesValue()) ||
8378 (!Res01->isNullValue() && !Res01->isAllOnesValue()) ||
8379 (!Res10->isNullValue() && !Res10->isAllOnesValue()) ||
8380 (!Res11->isNullValue() && !Res11->isAllOnesValue()))
8381 return nullptr;
8382
8383 std::bitset<4> Table;
8384 if (!Res00->isNullValue())
8385 Table.set(0);
8386 if (!Res01->isNullValue())
8387 Table.set(1);
8388 if (!Res10->isNullValue())
8389 Table.set(2);
8390 if (!Res11->isNullValue())
8391 Table.set(3);
8392
8393 Value *Res = createLogicFromTable(Table, C1, C2, Builder,
8394 Op0->hasOneUse() && Op1->hasOneUse());
8395 if (!Res)
8396 return nullptr;
8397 return replaceInstUsesWith(I, Res);
8398}
8399
8400/// Fold (C / X) < 0.0 --> X < 0.0 if possible. Swap predicate if necessary.
8402 Constant *RHSC) {
8403 // When C is not 0.0 and infinities are not allowed:
8404 // (C / X) < 0.0 is a sign-bit test of X
8405 // (C / X) < 0.0 --> X < 0.0 (if C is positive)
8406 // (C / X) < 0.0 --> X > 0.0 (if C is negative, swap the predicate)
8407 //
8408 // Proof:
8409 // Multiply (C / X) < 0.0 by X * X / C.
8410 // - X is non zero, if it is the flag 'ninf' is violated.
8411 // - C defines the sign of X * X * C. Thus it also defines whether to swap
8412 // the predicate. C is also non zero by definition.
8413 //
8414 // Thus X * X / C is non zero and the transformation is valid. [qed]
8415
8416 FCmpInst::Predicate Pred = I.getPredicate();
8417
8418 // Check that predicates are valid.
8419 if ((Pred != FCmpInst::FCMP_OGT) && (Pred != FCmpInst::FCMP_OLT) &&
8420 (Pred != FCmpInst::FCMP_OGE) && (Pred != FCmpInst::FCMP_OLE))
8421 return nullptr;
8422
8423 // Check that RHS operand is zero.
8424 if (!match(RHSC, m_AnyZeroFP()))
8425 return nullptr;
8426
8427 // Check fastmath flags ('ninf').
8428 if (!LHSI->hasNoInfs() || !I.hasNoInfs())
8429 return nullptr;
8430
8431 // Check the properties of the dividend. It must not be zero to avoid a
8432 // division by zero (see Proof).
8433 const APFloat *C;
8434 if (!match(LHSI->getOperand(0), m_APFloat(C)))
8435 return nullptr;
8436
8437 if (C->isZero())
8438 return nullptr;
8439
8440 // Get swapped predicate if necessary.
8441 if (C->isNegative())
8442 Pred = I.getSwappedPredicate();
8443
8444 return new FCmpInst(Pred, LHSI->getOperand(1), RHSC, "", &I);
8445}
8446
8447// Transform 'fptrunc(x) cmp C' to 'x cmp ext(C)' if possible.
8448// Patterns include:
8449// fptrunc(x) < C --> x < ext(C)
8450// fptrunc(x) <= C --> x <= ext(C)
8451// fptrunc(x) > C --> x > ext(C)
8452// fptrunc(x) >= C --> x >= ext(C)
8453// fptrunc(x) ord/uno C --> x ord/uno 0
8454// where 'ext(C)' is the extension of 'C' to the type of 'x' with a small bias
8455// due to precision loss.
8457 const Constant &C) {
8458 FCmpInst::Predicate Pred = I.getPredicate();
8459 Type *DestType = FPTrunc.getOperand(0)->getType();
8460
8461 const APFloat *CValue;
8462 // TODO: support vec
8463 if (!match(&C, m_APFloat(CValue)))
8464 return nullptr;
8465
8466 // Handle ord/uno
8467 if (Pred == FCmpInst::FCMP_ORD || Pred == FCmpInst::FCMP_UNO) {
8468 assert(!CValue->isNaN() &&
8469 "X ord/uno NaN should be folded away by simplifyFCmpInst()");
8470 return new FCmpInst(Pred, FPTrunc.getOperand(0),
8471 ConstantFP::getZero(DestType), "", &I);
8472 }
8473
8474 // Handle <, >, <=, >=
8475 bool RoundDown = false;
8476
8477 if (Pred == FCmpInst::FCMP_OGE || Pred == FCmpInst::FCMP_UGE ||
8478 Pred == FCmpInst::FCMP_OLT || Pred == FCmpInst::FCMP_ULT)
8479 RoundDown = true;
8480 else if (Pred == FCmpInst::FCMP_OGT || Pred == FCmpInst::FCMP_UGT ||
8481 Pred == FCmpInst::FCMP_OLE || Pred == FCmpInst::FCMP_ULE)
8482 RoundDown = false;
8483 else
8484 return nullptr;
8485
8486 if (CValue->isNaN() || CValue->isInfinity())
8487 return nullptr;
8488
8489 auto ConvertFltSema = [](const APFloat &Src, const fltSemantics &Sema) {
8490 bool LosesInfo;
8491 APFloat Dest = Src;
8492 Dest.convert(Sema, APFloat::rmNearestTiesToEven, &LosesInfo);
8493 return Dest;
8494 };
8495
8496 auto NextValue = [](const APFloat &Value, bool RoundDown) {
8497 APFloat NextValue = Value;
8498 NextValue.next(RoundDown);
8499 return NextValue;
8500 };
8501
8502 APFloat NextCValue = NextValue(*CValue, RoundDown);
8503
8504 const fltSemantics &DestFltSema =
8505 DestType->getScalarType()->getFltSemantics();
8506
8507 APFloat ExtCValue = ConvertFltSema(*CValue, DestFltSema);
8508 APFloat ExtNextCValue = ConvertFltSema(NextCValue, DestFltSema);
8509
8510 // When 'NextCValue' is infinity, use an imaged 'NextCValue' that equals
8511 // 'CValue + bias' to avoid the infinity after conversion. The bias is
8512 // estimated as 'CValue - PrevCValue', where 'PrevCValue' is the previous
8513 // value of 'CValue'.
8514 if (NextCValue.isInfinity()) {
8515 APFloat PrevCValue = NextValue(*CValue, !RoundDown);
8516 APFloat Bias = ConvertFltSema(*CValue - PrevCValue, DestFltSema);
8517
8518 ExtNextCValue = ExtCValue + Bias;
8519 }
8520
8521 APFloat ExtMidValue =
8522 scalbn(ExtCValue + ExtNextCValue, -1, APFloat::rmNearestTiesToEven);
8523
8524 const fltSemantics &SrcFltSema =
8525 C.getType()->getScalarType()->getFltSemantics();
8526
8527 // 'MidValue' might be rounded to 'NextCValue'. Correct it here.
8528 APFloat MidValue = ConvertFltSema(ExtMidValue, SrcFltSema);
8529 if (MidValue != *CValue)
8530 ExtMidValue.next(!RoundDown);
8531
8532 // Check whether 'ExtMidValue' is a valid result since the assumption on
8533 // imaged 'NextCValue' might not hold for new float types.
8534 // ppc_fp128 can't pass here when converting from max float because of
8535 // APFloat implementation.
8536 if (NextCValue.isInfinity()) {
8537 // ExtMidValue --- narrowed ---> Finite
8538 if (ConvertFltSema(ExtMidValue, SrcFltSema).isInfinity())
8539 return nullptr;
8540
8541 // NextExtMidValue --- narrowed ---> Infinity
8542 APFloat NextExtMidValue = NextValue(ExtMidValue, RoundDown);
8543 if (ConvertFltSema(NextExtMidValue, SrcFltSema).isFinite())
8544 return nullptr;
8545 }
8546
8547 return new FCmpInst(Pred, FPTrunc.getOperand(0),
8548 ConstantFP::get(DestType, ExtMidValue), "", &I);
8549}
8550
8551/// Optimize fabs(X) compared with zero.
8553 Value *X;
8554 if (!match(I.getOperand(0), m_FAbs(m_Value(X))))
8555 return nullptr;
8556
8557 const APFloat *C;
8558 if (!match(I.getOperand(1), m_APFloat(C)))
8559 return nullptr;
8560
8561 if (!C->isPosZero()) {
8562 if (!C->isSmallestNormalized())
8563 return nullptr;
8564
8565 const Function *F = I.getFunction();
8566 DenormalMode Mode = F->getDenormalMode(C->getSemantics());
8567 if (Mode.Input == DenormalMode::PreserveSign ||
8569
8570 auto replaceFCmp = [](FCmpInst *I, FCmpInst::Predicate P, Value *X) {
8571 Constant *Zero = ConstantFP::getZero(X->getType());
8572 return new FCmpInst(P, X, Zero, "", I);
8573 };
8574
8575 switch (I.getPredicate()) {
8576 case FCmpInst::FCMP_OLT:
8577 // fcmp olt fabs(x), smallest_normalized_number -> fcmp oeq x, 0.0
8578 return replaceFCmp(&I, FCmpInst::FCMP_OEQ, X);
8579 case FCmpInst::FCMP_UGE:
8580 // fcmp uge fabs(x), smallest_normalized_number -> fcmp une x, 0.0
8581 return replaceFCmp(&I, FCmpInst::FCMP_UNE, X);
8582 case FCmpInst::FCMP_OGE:
8583 // fcmp oge fabs(x), smallest_normalized_number -> fcmp one x, 0.0
8584 return replaceFCmp(&I, FCmpInst::FCMP_ONE, X);
8585 case FCmpInst::FCMP_ULT:
8586 // fcmp ult fabs(x), smallest_normalized_number -> fcmp ueq x, 0.0
8587 return replaceFCmp(&I, FCmpInst::FCMP_UEQ, X);
8588 default:
8589 break;
8590 }
8591 }
8592
8593 return nullptr;
8594 }
8595
8596 auto replacePredAndOp0 = [&IC](FCmpInst *I, FCmpInst::Predicate P, Value *X) {
8597 I->setPredicate(P);
8598 return IC.replaceOperand(*I, 0, X);
8599 };
8600
8601 switch (I.getPredicate()) {
8602 case FCmpInst::FCMP_UGE:
8603 case FCmpInst::FCMP_OLT:
8604 // fabs(X) >= 0.0 --> true
8605 // fabs(X) < 0.0 --> false
8606 llvm_unreachable("fcmp should have simplified");
8607
8608 case FCmpInst::FCMP_OGT:
8609 // fabs(X) > 0.0 --> X != 0.0
8610 return replacePredAndOp0(&I, FCmpInst::FCMP_ONE, X);
8611
8612 case FCmpInst::FCMP_UGT:
8613 // fabs(X) u> 0.0 --> X u!= 0.0
8614 return replacePredAndOp0(&I, FCmpInst::FCMP_UNE, X);
8615
8616 case FCmpInst::FCMP_OLE:
8617 // fabs(X) <= 0.0 --> X == 0.0
8618 return replacePredAndOp0(&I, FCmpInst::FCMP_OEQ, X);
8619
8620 case FCmpInst::FCMP_ULE:
8621 // fabs(X) u<= 0.0 --> X u== 0.0
8622 return replacePredAndOp0(&I, FCmpInst::FCMP_UEQ, X);
8623
8624 case FCmpInst::FCMP_OGE:
8625 // fabs(X) >= 0.0 --> !isnan(X)
8626 assert(!I.hasNoNaNs() && "fcmp should have simplified");
8627 return replacePredAndOp0(&I, FCmpInst::FCMP_ORD, X);
8628
8629 case FCmpInst::FCMP_ULT:
8630 // fabs(X) u< 0.0 --> isnan(X)
8631 assert(!I.hasNoNaNs() && "fcmp should have simplified");
8632 return replacePredAndOp0(&I, FCmpInst::FCMP_UNO, X);
8633
8634 case FCmpInst::FCMP_OEQ:
8635 case FCmpInst::FCMP_UEQ:
8636 case FCmpInst::FCMP_ONE:
8637 case FCmpInst::FCMP_UNE:
8638 case FCmpInst::FCMP_ORD:
8639 case FCmpInst::FCMP_UNO:
8640 // Look through the fabs() because it doesn't change anything but the sign.
8641 // fabs(X) == 0.0 --> X == 0.0,
8642 // fabs(X) != 0.0 --> X != 0.0
8643 // isnan(fabs(X)) --> isnan(X)
8644 // !isnan(fabs(X) --> !isnan(X)
8645 return replacePredAndOp0(&I, I.getPredicate(), X);
8646
8647 default:
8648 return nullptr;
8649 }
8650}
8651
8652/// Optimize sqrt(X) compared with zero.
8654 Value *X;
8655 if (!match(I.getOperand(0), m_Sqrt(m_Value(X))))
8656 return nullptr;
8657
8658 if (!match(I.getOperand(1), m_PosZeroFP()))
8659 return nullptr;
8660
8661 auto ReplacePredAndOp0 = [&](FCmpInst::Predicate P) {
8662 I.setPredicate(P);
8663 return IC.replaceOperand(I, 0, X);
8664 };
8665
8666 // Clear ninf flag if sqrt doesn't have it.
8667 if (!cast<Instruction>(I.getOperand(0))->hasNoInfs())
8668 I.setHasNoInfs(false);
8669
8670 switch (I.getPredicate()) {
8671 case FCmpInst::FCMP_OLT:
8672 case FCmpInst::FCMP_UGE:
8673 // sqrt(X) < 0.0 --> false
8674 // sqrt(X) u>= 0.0 --> true
8675 llvm_unreachable("fcmp should have simplified");
8676 case FCmpInst::FCMP_ULT:
8677 case FCmpInst::FCMP_ULE:
8678 case FCmpInst::FCMP_OGT:
8679 case FCmpInst::FCMP_OGE:
8680 case FCmpInst::FCMP_OEQ:
8681 case FCmpInst::FCMP_UNE:
8682 // sqrt(X) u< 0.0 --> X u< 0.0
8683 // sqrt(X) u<= 0.0 --> X u<= 0.0
8684 // sqrt(X) > 0.0 --> X > 0.0
8685 // sqrt(X) >= 0.0 --> X >= 0.0
8686 // sqrt(X) == 0.0 --> X == 0.0
8687 // sqrt(X) u!= 0.0 --> X u!= 0.0
8688 return IC.replaceOperand(I, 0, X);
8689
8690 case FCmpInst::FCMP_OLE:
8691 // sqrt(X) <= 0.0 --> X == 0.0
8692 return ReplacePredAndOp0(FCmpInst::FCMP_OEQ);
8693 case FCmpInst::FCMP_UGT:
8694 // sqrt(X) u> 0.0 --> X u!= 0.0
8695 return ReplacePredAndOp0(FCmpInst::FCMP_UNE);
8696 case FCmpInst::FCMP_UEQ:
8697 // sqrt(X) u== 0.0 --> X u<= 0.0
8698 return ReplacePredAndOp0(FCmpInst::FCMP_ULE);
8699 case FCmpInst::FCMP_ONE:
8700 // sqrt(X) != 0.0 --> X > 0.0
8701 return ReplacePredAndOp0(FCmpInst::FCMP_OGT);
8702 case FCmpInst::FCMP_ORD:
8703 // !isnan(sqrt(X)) --> X >= 0.0
8704 return ReplacePredAndOp0(FCmpInst::FCMP_OGE);
8705 case FCmpInst::FCMP_UNO:
8706 // isnan(sqrt(X)) --> X u< 0.0
8707 return ReplacePredAndOp0(FCmpInst::FCMP_ULT);
8708 default:
8709 llvm_unreachable("Unexpected predicate!");
8710 }
8711}
8712
8714 CmpInst::Predicate Pred = I.getPredicate();
8715 Value *Op0 = I.getOperand(0), *Op1 = I.getOperand(1);
8716
8717 // Canonicalize fneg as Op1.
8718 if (match(Op0, m_FNeg(m_Value())) && !match(Op1, m_FNeg(m_Value()))) {
8719 std::swap(Op0, Op1);
8720 Pred = I.getSwappedPredicate();
8721 }
8722
8723 if (!match(Op1, m_FNeg(m_Specific(Op0))))
8724 return nullptr;
8725
8726 // Replace the negated operand with 0.0:
8727 // fcmp Pred Op0, -Op0 --> fcmp Pred Op0, 0.0
8728 Constant *Zero = ConstantFP::getZero(Op0->getType());
8729 return new FCmpInst(Pred, Op0, Zero, "", &I);
8730}
8731
8733 Constant *RHSC, InstCombinerImpl &CI) {
8734 const CmpInst::Predicate Pred = I.getPredicate();
8735 Value *X = LHSI->getOperand(0);
8736 Value *Y = LHSI->getOperand(1);
8737 switch (Pred) {
8738 default:
8739 break;
8740 case FCmpInst::FCMP_UGT:
8741 case FCmpInst::FCMP_ULT:
8742 case FCmpInst::FCMP_UNE:
8743 case FCmpInst::FCMP_OEQ:
8744 case FCmpInst::FCMP_OGE:
8745 case FCmpInst::FCMP_OLE:
8746 // The optimization is not valid if X and Y are infinities of the same
8747 // sign, i.e. the inf - inf = nan case. If the fsub has the ninf or nnan
8748 // flag then we can assume we do not have that case. Otherwise we might be
8749 // able to prove that either X or Y is not infinity.
8750 if (!LHSI->hasNoNaNs() && !LHSI->hasNoInfs() &&
8754 break;
8755
8756 [[fallthrough]];
8757 case FCmpInst::FCMP_OGT:
8758 case FCmpInst::FCMP_OLT:
8759 case FCmpInst::FCMP_ONE:
8760 case FCmpInst::FCMP_UEQ:
8761 case FCmpInst::FCMP_UGE:
8762 case FCmpInst::FCMP_ULE:
8763 // fcmp pred (x - y), 0 --> fcmp pred x, y
8764 if (match(RHSC, m_AnyZeroFP()) &&
8765 I.getFunction()->getDenormalMode(
8766 LHSI->getType()->getScalarType()->getFltSemantics()) ==
8768 CI.replaceOperand(I, 0, X);
8769 CI.replaceOperand(I, 1, Y);
8770 I.setHasNoInfs(LHSI->hasNoInfs());
8771 if (LHSI->hasNoNaNs())
8772 I.setHasNoNaNs(true);
8773 return &I;
8774 }
8775 // fcmp `pred (C - Y), C` -> `fcmp swap(pred), Y, 0`
8776 // where C and Y can't be arbitrary floating-point values.
8777 // For example, with `C = 1.0f` and `Y = 0x1p-149`, `1.0f - Y` rounds back
8778 // to `1.0f`, so the source compare is false while the rewritten compare is
8779 // true.
8780 // We need to make sure (C - Y) never rounds back to C
8781 const APFloat *C;
8782 Value *IntSrc;
8783 if (match(RHSC, m_APFloat(C)) &&
8784 match(LHSI, m_FSub(m_Specific(RHSC), m_IToFP(m_Value(IntSrc)))) &&
8785 C->isNormal()) {
8786 // Requirements on C and Y:
8787 // 1. C is finite, nonzero, normal.
8788 // 2. C shouldn't be too large, that is, ULP(C) <= 1.
8789 // 3. Y must be the form of `[su]itofp`, so the finite nonzero result of Y
8790 // must be integer-valued with an absolute value of at least 1;
8791 // as long as the step size near C does not exceed 1,
8792 // C - Y cannot be rounded back to C when Y != 0.
8793 // 4. If Y = 0, `fcmp pred (C - 0), C` are equivalent to `fcmp swap(pred)
8794 // 0, 0` for ordered and unordered predicates as long as C is finite and
8795 // nonzero.
8796 int MantissaWidth = LHSI->getType()->getFPMantissaWidth();
8797 if (MantissaWidth != -1 && ilogb(*C) < MantissaWidth) {
8798 Constant *ZeroC = ConstantFP::getZero(LHSI->getType());
8799 I.setPredicate(I.getSwappedPredicate());
8800 CI.replaceOperand(I, 0, Y);
8801 CI.replaceOperand(I, 1, ZeroC);
8802 return &I;
8803 }
8804 }
8805 break;
8806 }
8807
8808 return nullptr;
8809}
8810
8811/// Fold: fabs(uitofp(a) - uitofp(b)) pred C --> a == b
8812/// where 'pred' is olt, ult, ogt, ugt, oge or uge and C is a positive, Non-NaN
8813/// float when the uitofp casts are exact and C is in the valid range.
8814///
8815/// Since exact uitofp means distinct integers map to distinct floats, the only
8816/// values fabs(uitofp(a) - uitofp(b)) can take are {0.0, 1.0, 2.0, ...}.
8817/// There are no values in the open interval (0, 1), so:
8818/// fabs(...) < C where 0 < C <= 1.0 --> a == b (strict lt: C=1.0 ok)
8819// fabs(..) >= C where C >= 1.0 -> a != b
8820///
8821/// The same logic applies to sitofp.
8823 Value *FAbsArg;
8824 if (!match(I.getOperand(0), m_FAbs(m_Value(FAbsArg))))
8825 return nullptr;
8826
8827 const APFloat *C;
8828 if (!match(I.getOperand(1), PatternMatch::m_FiniteNonZero(C)))
8829 return nullptr;
8830
8831 FCmpInst::Predicate Pred = I.getPredicate();
8832 bool IsStrictLt = Pred == FCmpInst::FCMP_OLT || Pred == FCmpInst::FCMP_ULT;
8833 bool IsLe = Pred == FCmpInst::FCMP_OLE || Pred == FCmpInst::FCMP_ULE;
8834 bool IsStrictGt = Pred == FCmpInst::FCMP_OGT || Pred == FCmpInst::FCMP_UGT;
8835 bool IsGe = Pred == FCmpInst::FCMP_OGE || Pred == FCmpInst::FCMP_UGE;
8836 if (!IsStrictLt && !IsStrictGt && !IsGe)
8837 return nullptr;
8838
8839 APFloat One = APFloat::getOne(C->getSemantics());
8840 APFloat::cmpResult Cmp = C->compare(One);
8841
8842 // For strict-lt (olt/ult): C must be in (0, 1.0] -- C == 1.0 is fine since
8843 // the next possible value after 0.0 is 1.0, and < 1.0 excludes it.
8844 if (IsStrictLt && Cmp == APFloat::cmpGreaterThan)
8845 return nullptr;
8846 if (IsGe && Cmp == APFloat::cmpGreaterThan)
8847 return nullptr;
8848 if (IsLe && Cmp != APFloat::cmpGreaterThan)
8849 return nullptr;
8850 if (IsStrictGt && Cmp != APFloat::cmpLessThan)
8851 return nullptr;
8852
8853 // Match: fsub(uitofp(A), uitofp(B)) where both casts are uitofp or sitofp
8854 Value *A, *B;
8855 bool IsSigned;
8856 if (match(FAbsArg, m_FSub(m_UIToFP(m_Value(A)), m_UIToFP(m_Value(B))))) {
8857 IsSigned = false;
8858 } else if (match(FAbsArg,
8860 IsSigned = true;
8861 } else {
8862 return nullptr;
8863 }
8864
8865 // A and B must have the same integer type
8866 if (A->getType() != B->getType())
8867 return nullptr;
8868
8869 Type *FPTy = FAbsArg->getType();
8870 if (!IC.canBeCastedExactlyIntToFP(A, FPTy, IsSigned, &I) ||
8871 !IC.canBeCastedExactlyIntToFP(B, FPTy, IsSigned, &I))
8872 return nullptr;
8873 ICmpInst::Predicate ResultPred =
8874 IsStrictLt || IsLe ? ICmpInst::ICMP_EQ : ICmpInst::ICMP_NE;
8875 return new ICmpInst(ResultPred, A, B);
8876}
8877
8879 InstCombinerImpl &IC) {
8880 Value *LHS = I.getOperand(0), *RHS = I.getOperand(1);
8881 Type *OpType = LHS->getType();
8882 CmpInst::Predicate Pred = I.getPredicate();
8883
8886
8887 if (!FloorX && !CeilX) {
8890 std::swap(LHS, RHS);
8891 Pred = I.getSwappedPredicate();
8892 }
8893 }
8894
8895 if ((FloorX || CeilX) && FCmpInst::isCommutative(Pred) && LHS->hasOneUse()) {
8896 // fcmp pred floor(x), x => fcmp pred trunc(x), x
8897 // fcmp pred ceil(x), x => fcmp pred trunc(x), x
8898 // where pred is oeq, one, ord, ueq, une, uno.
8899 Value *TruncX = IC.Builder.CreateUnaryIntrinsic(Intrinsic::trunc, RHS);
8900 return new FCmpInst(Pred, TruncX, RHS, "", &I);
8901 }
8902
8903 switch (Pred) {
8904 case FCmpInst::FCMP_OLE:
8905 // fcmp ole floor(x), x => fcmp ord x, 0
8906 if (FloorX)
8908 "", &I);
8909 break;
8910 case FCmpInst::FCMP_OGT:
8911 // fcmp ogt floor(x), x => false
8912 if (FloorX)
8913 return IC.replaceInstUsesWith(I, ConstantInt::getFalse(I.getType()));
8914 break;
8915 case FCmpInst::FCMP_OGE:
8916 // fcmp oge ceil(x), x => fcmp ord x, 0
8917 if (CeilX)
8919 "", &I);
8920 break;
8921 case FCmpInst::FCMP_OLT:
8922 // fcmp olt ceil(x), x => false
8923 if (CeilX)
8924 return IC.replaceInstUsesWith(I, ConstantInt::getFalse(I.getType()));
8925 break;
8926 case FCmpInst::FCMP_ULE:
8927 // fcmp ule floor(x), x => true
8928 if (FloorX)
8929 return IC.replaceInstUsesWith(I, ConstantInt::getTrue(I.getType()));
8930 break;
8931 case FCmpInst::FCMP_UGT:
8932 // fcmp ugt floor(x), x => fcmp uno x, 0
8933 if (FloorX)
8935 "", &I);
8936 break;
8937 case FCmpInst::FCMP_UGE:
8938 // fcmp uge ceil(x), x => true
8939 if (CeilX)
8940 return IC.replaceInstUsesWith(I, ConstantInt::getTrue(I.getType()));
8941 break;
8942 case FCmpInst::FCMP_ULT:
8943 // fcmp ult ceil(x), x => fcmp uno x, 0
8944 if (CeilX)
8946 "", &I);
8947 break;
8948 default:
8949 break;
8950 }
8951
8952 return nullptr;
8953}
8954
8955/// Returns true if a select that implements a min/max is redundant and
8956/// select result can be replaced with its non-constant operand, e.g.,
8957/// select ( (si/ui-to-fp A) <= C ), C, (si/ui-to-fp A)
8958/// where C is the FP constant equal to the minimum integer value
8959/// representable by A.
8961 Value *B) {
8962 const APFloat *APF;
8963 if (!match(B, m_APFloat(APF)))
8964 return false;
8965
8966 auto *I = dyn_cast<Instruction>(A);
8967 if (!I || !(I->getOpcode() == Instruction::SIToFP ||
8968 I->getOpcode() == Instruction::UIToFP))
8969 return false;
8970
8971 bool IsUnsigned = I->getOpcode() == Instruction::UIToFP;
8972 unsigned BitWidth = I->getOperand(0)->getType()->getScalarSizeInBits();
8973 APSInt IntBoundary = (Flavor == SPF_FMAXNUM)
8974 ? APSInt::getMinValue(BitWidth, IsUnsigned)
8975 : APSInt::getMaxValue(BitWidth, IsUnsigned);
8976 APSInt ConvertedInt(BitWidth, IsUnsigned);
8977 bool IsExact;
8979 APF->convertToInteger(ConvertedInt, APFloat::rmTowardZero, &IsExact);
8980 return Status == APFloat::opOK && IsExact && ConvertedInt == IntBoundary;
8981}
8982
8984 bool Changed = false;
8985
8986 /// Orders the operands of the compare so that they are listed from most
8987 /// complex to least complex. This puts constants before unary operators,
8988 /// before binary operators.
8989 if (getComplexity(I.getOperand(0)) < getComplexity(I.getOperand(1))) {
8990 I.swapOperands();
8991 Changed = true;
8992 }
8993
8994 const CmpInst::Predicate Pred = I.getPredicate();
8995 Value *Op0 = I.getOperand(0), *Op1 = I.getOperand(1);
8996 if (Value *V = simplifyFCmpInst(Pred, Op0, Op1, I.getFastMathFlags(),
8997 SQ.getWithInstruction(&I)))
8998 return replaceInstUsesWith(I, V);
8999
9000 // Simplify 'fcmp pred X, X'
9001 Type *OpType = Op0->getType();
9002 assert(OpType == Op1->getType() && "fcmp with different-typed operands?");
9003 if (Op0 == Op1) {
9004 switch (Pred) {
9005 default:
9006 break;
9007 case FCmpInst::FCMP_UNO: // True if unordered: isnan(X) | isnan(Y)
9008 case FCmpInst::FCMP_ULT: // True if unordered or less than
9009 case FCmpInst::FCMP_UGT: // True if unordered or greater than
9010 case FCmpInst::FCMP_UNE: // True if unordered or not equal
9011 // Canonicalize these to be 'fcmp uno %X, 0.0'.
9012 I.setPredicate(FCmpInst::FCMP_UNO);
9013 I.setOperand(1, Constant::getNullValue(OpType));
9014 return &I;
9015
9016 case FCmpInst::FCMP_ORD: // True if ordered (no nans)
9017 case FCmpInst::FCMP_OEQ: // True if ordered and equal
9018 case FCmpInst::FCMP_OGE: // True if ordered and greater than or equal
9019 case FCmpInst::FCMP_OLE: // True if ordered and less than or equal
9020 // Canonicalize these to be 'fcmp ord %X, 0.0'.
9021 I.setPredicate(FCmpInst::FCMP_ORD);
9022 I.setOperand(1, Constant::getNullValue(OpType));
9023 return &I;
9024 }
9025 }
9026
9027 if (I.isCommutative()) {
9028 if (auto Pair = matchSymmetricPair(I.getOperand(0), I.getOperand(1))) {
9029 replaceOperand(I, 0, Pair->first);
9030 replaceOperand(I, 1, Pair->second);
9031 return &I;
9032 }
9033 }
9034
9035 // If we're just checking for a NaN (ORD/UNO) and have a non-NaN operand,
9036 // then canonicalize the operand to 0.0.
9037 if (Pred == CmpInst::FCMP_ORD || Pred == CmpInst::FCMP_UNO) {
9038 if (!match(Op0, m_PosZeroFP()) &&
9039 isKnownNeverNaN(Op0, getSimplifyQuery().getWithInstruction(&I)))
9040 return replaceOperand(I, 0, ConstantFP::getZero(OpType));
9041
9042 if (!match(Op1, m_PosZeroFP()) &&
9043 isKnownNeverNaN(Op1, getSimplifyQuery().getWithInstruction(&I)))
9044 return replaceOperand(I, 1, ConstantFP::getZero(OpType));
9045 }
9046
9047 // fcmp pred (fneg X), (fneg Y) -> fcmp swap(pred) X, Y
9048 Value *X, *Y;
9049 if (match(Op0, m_FNeg(m_Value(X))) && match(Op1, m_FNeg(m_Value(Y))))
9050 return new FCmpInst(I.getSwappedPredicate(), X, Y, "", &I);
9051
9053 return R;
9054
9055 // Test if the FCmpInst instruction is used exclusively by a select as
9056 // part of a minimum or maximum operation. If so, refrain from doing
9057 // any other folding. This helps out other analyses which understand
9058 // non-obfuscated minimum and maximum idioms, such as ScalarEvolution
9059 // and CodeGen. And in this case, at least one of the comparison
9060 // operands has at least one user besides the compare (the select),
9061 // which would often largely negate the benefit of folding anyway.
9062 if (I.hasOneUse())
9063 if (SelectInst *SI = dyn_cast<SelectInst>(I.user_back())) {
9064 Value *A, *B;
9066 bool IsRedundantMinMaxClamp =
9067 (SPR.Flavor == SPF_FMAXNUM || SPR.Flavor == SPF_FMINNUM) &&
9069 if (SPR.Flavor != SPF_UNKNOWN && !IsRedundantMinMaxClamp)
9070 return nullptr;
9071 }
9072
9073 // The sign of 0.0 is ignored by fcmp, so canonicalize to +0.0:
9074 // fcmp Pred X, -0.0 --> fcmp Pred X, 0.0
9075 if (match(Op1, m_AnyZeroFP()) && !match(Op1, m_PosZeroFP()))
9076 return replaceOperand(I, 1, ConstantFP::getZero(OpType));
9077
9078 // Canonicalize:
9079 // fcmp olt X, +inf -> fcmp one X, +inf
9080 // fcmp ole X, +inf -> fcmp ord X, 0
9081 // fcmp ogt X, +inf -> false
9082 // fcmp oge X, +inf -> fcmp oeq X, +inf
9083 // fcmp ult X, +inf -> fcmp une X, +inf
9084 // fcmp ule X, +inf -> true
9085 // fcmp ugt X, +inf -> fcmp uno X, 0
9086 // fcmp uge X, +inf -> fcmp ueq X, +inf
9087 // fcmp olt X, -inf -> false
9088 // fcmp ole X, -inf -> fcmp oeq X, -inf
9089 // fcmp ogt X, -inf -> fcmp one X, -inf
9090 // fcmp oge X, -inf -> fcmp ord X, 0
9091 // fcmp ult X, -inf -> fcmp uno X, 0
9092 // fcmp ule X, -inf -> fcmp ueq X, -inf
9093 // fcmp ugt X, -inf -> fcmp une X, -inf
9094 // fcmp uge X, -inf -> true
9095 const APFloat *C;
9096 if (match(Op1, m_APFloat(C)) && C->isInfinity()) {
9097 switch (C->isNegative() ? FCmpInst::getSwappedPredicate(Pred) : Pred) {
9098 default:
9099 break;
9100 case FCmpInst::FCMP_ORD:
9101 case FCmpInst::FCMP_UNO:
9104 case FCmpInst::FCMP_OGT:
9105 case FCmpInst::FCMP_ULE:
9106 llvm_unreachable("Should be simplified by InstSimplify");
9107 case FCmpInst::FCMP_OLT:
9108 return new FCmpInst(FCmpInst::FCMP_ONE, Op0, Op1, "", &I);
9109 case FCmpInst::FCMP_OLE:
9110 return new FCmpInst(FCmpInst::FCMP_ORD, Op0, ConstantFP::getZero(OpType),
9111 "", &I);
9112 case FCmpInst::FCMP_OGE:
9113 return new FCmpInst(FCmpInst::FCMP_OEQ, Op0, Op1, "", &I);
9114 case FCmpInst::FCMP_ULT:
9115 return new FCmpInst(FCmpInst::FCMP_UNE, Op0, Op1, "", &I);
9116 case FCmpInst::FCMP_UGT:
9117 return new FCmpInst(FCmpInst::FCMP_UNO, Op0, ConstantFP::getZero(OpType),
9118 "", &I);
9119 case FCmpInst::FCMP_UGE:
9120 return new FCmpInst(FCmpInst::FCMP_UEQ, Op0, Op1, "", &I);
9121 }
9122 }
9123
9124 // Ignore signbit of bitcasted int when comparing equality to FP 0.0:
9125 // fcmp oeq/une (bitcast X), 0.0 --> (and X, SignMaskC) ==/!= 0
9126 if (match(Op1, m_PosZeroFP()) &&
9128 !F.getDenormalMode(Op1->getType()->getScalarType()->getFltSemantics())
9129 .inputsMayBeZero()) {
9131 if (Pred == FCmpInst::FCMP_OEQ)
9132 IntPred = ICmpInst::ICMP_EQ;
9133 else if (Pred == FCmpInst::FCMP_UNE)
9134 IntPred = ICmpInst::ICMP_NE;
9135
9136 if (IntPred != ICmpInst::BAD_ICMP_PREDICATE) {
9137 Type *IntTy = X->getType();
9138 const APInt &SignMask = ~APInt::getSignMask(IntTy->getScalarSizeInBits());
9139 Value *MaskX = Builder.CreateAnd(X, ConstantInt::get(IntTy, SignMask));
9140 return new ICmpInst(IntPred, MaskX, ConstantInt::getNullValue(IntTy));
9141 }
9142 }
9143
9144 // Handle fcmp with instruction LHS and constant RHS.
9145 Instruction *LHSI;
9146 Constant *RHSC;
9147 if (match(Op0, m_Instruction(LHSI)) && match(Op1, m_Constant(RHSC))) {
9148 switch (LHSI->getOpcode()) {
9149 case Instruction::Select:
9150 // fcmp eq (cond ? x : -x), 0 --> fcmp eq x, 0
9151 if (FCmpInst::isEquality(Pred) && match(RHSC, m_AnyZeroFP()) &&
9153 return replaceOperand(I, 0, X);
9155 return NV;
9156 break;
9157 case Instruction::FSub:
9158 if (LHSI->hasOneUse())
9159 if (Instruction *NV = foldFCmpFSubIntoFCmp(I, LHSI, RHSC, *this))
9160 return NV;
9161 break;
9162 case Instruction::PHI:
9163 if (Instruction *NV = foldOpIntoPhi(I, cast<PHINode>(LHSI)))
9164 return NV;
9165 break;
9166 case Instruction::SIToFP:
9167 case Instruction::UIToFP:
9168 if (Instruction *NV = foldFCmpIntToFPConst(I, LHSI, RHSC))
9169 return NV;
9170 break;
9171 case Instruction::FDiv:
9172 if (Instruction *NV = foldFCmpReciprocalAndZero(I, LHSI, RHSC))
9173 return NV;
9174 break;
9175 case Instruction::Load:
9176 if (auto *GEP = dyn_cast<GetElementPtrInst>(LHSI->getOperand(0)))
9177 if (Instruction *Res =
9179 return Res;
9180 break;
9181 case Instruction::FPTrunc:
9182 if (Instruction *NV = foldFCmpFpTrunc(I, *LHSI, *RHSC))
9183 return NV;
9184 break;
9185 }
9186 }
9187
9188 if (Instruction *R = foldFabsWithFcmpZero(I, *this))
9189 return R;
9190
9191 if (Instruction *R = foldFCmpFAbsFSubIntToFP(I, *this))
9192 return R;
9193
9194 if (Instruction *R = foldSqrtWithFcmpZero(I, *this))
9195 return R;
9196
9197 if (Instruction *R = foldFCmpWithFloorAndCeil(I, *this))
9198 return R;
9199
9201 return R;
9202
9203 if (match(Op0, m_FNeg(m_Value(X)))) {
9204 // fcmp pred (fneg X), C --> fcmp swap(pred) X, -C
9205 Constant *C;
9206 if (match(Op1, m_Constant(C)))
9207 if (Constant *NegC = ConstantFoldUnaryOpOperand(Instruction::FNeg, C, DL))
9208 return new FCmpInst(I.getSwappedPredicate(), X, NegC, "", &I);
9209 }
9210
9211 // fcmp (fadd X, 0.0), Y --> fcmp X, Y
9212 if (match(Op0, m_FAdd(m_Value(X), m_AnyZeroFP())))
9213 return new FCmpInst(Pred, X, Op1, "", &I);
9214
9215 // fcmp X, (fadd Y, 0.0) --> fcmp X, Y
9216 if (match(Op1, m_FAdd(m_Value(Y), m_AnyZeroFP())))
9217 return new FCmpInst(Pred, Op0, Y, "", &I);
9218
9219 // fcmp ord/uno (fptrunc X), (fptrunc Y) -> fcmp ord/uno X, Y
9220 if ((Pred == FCmpInst::FCMP_ORD || Pred == FCmpInst::FCMP_UNO) &&
9221 match(Op0, m_FPTrunc(m_Value(X))) && match(Op1, m_FPTrunc(m_Value(Y))) &&
9222 X->getType() == Y->getType())
9223 return new FCmpInst(Pred, X, Y, "", &I);
9224
9225 if (match(Op0, m_FPExt(m_Value(X)))) {
9226 // fcmp (fpext X), (fpext Y) -> fcmp X, Y
9227 if (match(Op1, m_FPExt(m_Value(Y))) && X->getType() == Y->getType())
9228 return new FCmpInst(Pred, X, Y, "", &I);
9229
9230 const APFloat *C;
9231 if (match(Op1, m_APFloat(C))) {
9232 const fltSemantics &FPSem =
9233 X->getType()->getScalarType()->getFltSemantics();
9234 bool Lossy;
9235 APFloat TruncC = *C;
9236 TruncC.convert(FPSem, APFloat::rmNearestTiesToEven, &Lossy);
9237
9238 if (Lossy) {
9239 // X can't possibly equal the higher-precision constant, so reduce any
9240 // equality comparison.
9241 // TODO: Other predicates can be handled via getFCmpCode().
9242 switch (Pred) {
9243 case FCmpInst::FCMP_OEQ:
9244 // X is ordered and equal to an impossible constant --> false
9245 return replaceInstUsesWith(I, ConstantInt::getFalse(I.getType()));
9246 case FCmpInst::FCMP_ONE:
9247 // X is ordered and not equal to an impossible constant --> ordered
9248 return new FCmpInst(FCmpInst::FCMP_ORD, X,
9249 ConstantFP::getZero(X->getType()));
9250 case FCmpInst::FCMP_UEQ:
9251 // X is unordered or equal to an impossible constant --> unordered
9252 return new FCmpInst(FCmpInst::FCMP_UNO, X,
9253 ConstantFP::getZero(X->getType()));
9254 case FCmpInst::FCMP_UNE:
9255 // X is unordered or not equal to an impossible constant --> true
9256 return replaceInstUsesWith(I, ConstantInt::getTrue(I.getType()));
9257 default:
9258 break;
9259 }
9260 }
9261
9262 // fcmp (fpext X), C -> fcmp X, (fptrunc C) if fptrunc is lossless
9263 // Avoid lossy conversions and denormals.
9264 // Zero is a special case that's OK to convert.
9265 APFloat Fabs = TruncC;
9266 Fabs.clearSign();
9267 if (!Lossy &&
9268 (Fabs.isZero() || !(Fabs < APFloat::getSmallestNormalized(FPSem)))) {
9269 Constant *NewC = ConstantFP::get(X->getType(), TruncC);
9270 return new FCmpInst(Pred, X, NewC, "", &I);
9271 }
9272 }
9273 }
9274
9275 // Convert a sign-bit test of an FP value into a cast and integer compare.
9276 // TODO: Simplify if the copysign constant is 0.0 or NaN.
9277 // TODO: Handle non-zero compare constants.
9278 // TODO: Handle other predicates.
9280 m_Value(X)))) &&
9281 match(Op1, m_AnyZeroFP()) && !C->isZero() && !C->isNaN()) {
9282 Type *IntType = Builder.getIntNTy(X->getType()->getScalarSizeInBits());
9283 if (auto *VecTy = dyn_cast<VectorType>(OpType))
9284 IntType = VectorType::get(IntType, VecTy->getElementCount());
9285
9286 // copysign(non-zero constant, X) < 0.0 --> (bitcast X) < 0
9287 if (Pred == FCmpInst::FCMP_OLT) {
9288 Value *IntX = Builder.CreateBitCast(X, IntType);
9289 return new ICmpInst(ICmpInst::ICMP_SLT, IntX,
9290 ConstantInt::getNullValue(IntType));
9291 }
9292 }
9293
9294 {
9295 Value *CanonLHS = nullptr;
9297 // (canonicalize(x) == x) => (x == x)
9298 if (CanonLHS == Op1)
9299 return new FCmpInst(Pred, Op1, Op1, "", &I);
9300
9301 Value *CanonRHS = nullptr;
9303 // (x == canonicalize(x)) => (x == x)
9304 if (CanonRHS == Op0)
9305 return new FCmpInst(Pred, Op0, Op0, "", &I);
9306
9307 // (canonicalize(x) == canonicalize(y)) => (x == y)
9308 if (CanonLHS && CanonRHS)
9309 return new FCmpInst(Pred, CanonLHS, CanonRHS, "", &I);
9310 }
9311
9312 if (I.getType()->isVectorTy())
9313 if (Instruction *Res = foldVectorCmp(I, Builder))
9314 return Res;
9315
9316 return Changed ? &I : nullptr;
9317}
assert(UImm &&(UImm !=~static_cast< T >(0)) &&"Invalid immediate!")
AMDGPU Register Bank Select
Rewrite undef for PHI
This file declares a class to represent arbitrary precision floating point values and provide a varie...
This file implements the APSInt class, which is a simple class that represents an arbitrary sized int...
MachineBasicBlock MachineBasicBlock::iterator DebugLoc DL
#define X(NUM, ENUM, NAME)
Definition ELF.h:851
static GCRegistry::Add< ErlangGC > A("erlang", "erlang-compatible garbage collector")
static GCRegistry::Add< StatepointGC > D("statepoint-example", "an example strategy for statepoint")
static GCRegistry::Add< CoreCLRGC > E("coreclr", "CoreCLR-compatible GC")
static GCRegistry::Add< OcamlGC > B("ocaml", "ocaml 3.10-compatible GC")
This file contains the declarations for the subclasses of Constant, which represent the different fla...
#define Check(C,...)
Hexagon Common GEP
static Instruction * foldFCmpReciprocalAndZero(FCmpInst &I, Instruction *LHSI, Constant *RHSC)
Fold (C / X) < 0.0 --> X < 0.0 if possible. Swap predicate if necessary.
static Instruction * foldFabsWithFcmpZero(FCmpInst &I, InstCombinerImpl &IC)
Optimize fabs(X) compared with zero.
static void collectOffsetOp(Value *V, SmallVectorImpl< OffsetOp > &Offsets, bool AllowRecursion)
static Value * rewriteGEPAsOffset(Value *Start, Value *Base, GEPNoWrapFlags NW, const DataLayout &DL, SetVector< Value * > &Explored, InstCombiner &IC)
Returns a re-written value of Start as an indexed GEP using Base as a pointer.
static bool isMinMaxCmpSelectEliminable(SelectPatternFlavor Flavor, Value *A, Value *B)
Returns true if a select that implements a min/max is redundant and select result can be replaced wit...
static Instruction * foldICmpEqualityWithOffset(ICmpInst &I, InstCombiner::BuilderTy &Builder, const SimplifyQuery &SQ)
Offset both sides of an equality icmp to see if we can save some instructions: icmp eq/ne X,...
static bool addWithOverflow(APInt &Result, const APInt &In1, const APInt &In2, bool IsSigned=false)
Compute Result = In1+In2, returning true if the result overflowed for this type.
static Instruction * foldICmpOfVectorReduce(ICmpInst &I, const DataLayout &DL, IRBuilderBase &Builder)
static Instruction * foldICmpAndXX(ICmpInst &I, const SimplifyQuery &Q, InstCombinerImpl &IC)
static Instruction * foldVectorCmp(CmpInst &Cmp, InstCombiner::BuilderTy &Builder)
static bool isMaskOrZero(const Value *V, bool Not, const SimplifyQuery &Q, unsigned Depth=0)
static Value * createLogicFromTable(const std::bitset< 4 > &Table, Value *Op0, Value *Op1, IRBuilderBase &Builder, bool HasOneUse)
static Instruction * foldICmpOfUAddOv(ICmpInst &I)
static bool isChainSelectCmpBranch(const SelectInst *SI)
Return true when the instruction sequence within a block is select-cmp-br.
static Instruction * foldICmpInvariantGroup(ICmpInst &I)
std::pair< Instruction::BinaryOps, Value * > OffsetOp
Find all possible pairs (BinOp, RHS) that BinOp V, RHS can be simplified.
static Instruction * foldReductionIdiom(ICmpInst &I, InstCombiner::BuilderTy &Builder, const DataLayout &DL)
This function folds patterns produced by lowering of reduce idioms, such as llvm.vector....
static Instruction * canonicalizeICmpBool(ICmpInst &I, InstCombiner::BuilderTy &Builder)
Integer compare with boolean values can always be turned into bitwise ops.
static Instruction * foldFCmpFSubIntoFCmp(FCmpInst &I, Instruction *LHSI, Constant *RHSC, InstCombinerImpl &CI)
static Value * foldICmpOrXorSubChain(ICmpInst &Cmp, BinaryOperator *Or, InstCombiner::BuilderTy &Builder)
Fold icmp eq/ne (or (xor/sub (X1, X2), xor/sub (X3, X4))), 0.
static bool hasBranchUse(ICmpInst &I)
Given an icmp instruction, return true if any use of this comparison is a branch on sign bit comparis...
static Value * foldICmpWithLowBitMaskedVal(CmpPredicate Pred, Value *Op0, Value *Op1, const SimplifyQuery &Q, InstCombiner &IC)
Some comparisons can be simplified.
static APInt getDemandedBitsLHSMask(ICmpInst &I, unsigned BitWidth)
When performing a comparison against a constant, it is possible that not all the bits in the LHS are ...
static Instruction * foldICmpShlLHSC(ICmpInst &Cmp, Instruction *Shl, const APInt &C)
Fold icmp (shl nuw C2, Y), C.
static Instruction * foldFCmpWithFloorAndCeil(FCmpInst &I, InstCombinerImpl &IC)
static Instruction * foldICmpXorXX(ICmpInst &I, const SimplifyQuery &Q, InstCombinerImpl &IC)
static Instruction * foldICmpOfCmpIntrinsicWithConstant(CmpPredicate Pred, IntrinsicInst *I, const APInt &C, InstCombiner::BuilderTy &Builder)
static Instruction * processUMulZExtIdiom(ICmpInst &I, Value *MulVal, const APInt *OtherVal, InstCombinerImpl &IC)
Recognize and process idiom involving test for multiplication overflow.
static Instruction * foldSqrtWithFcmpZero(FCmpInst &I, InstCombinerImpl &IC)
Optimize sqrt(X) compared with zero.
static Instruction * foldFCmpFNegCommonOp(FCmpInst &I)
static Instruction * foldICmpWithHighBitMask(ICmpInst &Cmp, InstCombiner::BuilderTy &Builder)
static ICmpInst * canonicalizeCmpWithConstant(ICmpInst &I)
If we have an icmp le or icmp ge instruction with a constant operand, turn it into the appropriate ic...
static Instruction * foldICmpIntrinsicWithIntrinsic(ICmpInst &Cmp, InstCombiner::BuilderTy &Builder)
Fold an icmp with LLVM intrinsics.
static Instruction * foldICmpUSubSatOrUAddSatWithConstant(CmpPredicate Pred, SaturatingInst *II, const APInt &C, InstCombiner::BuilderTy &Builder)
static Instruction * foldICmpPow2Test(ICmpInst &I, InstCombiner::BuilderTy &Builder)
static bool subWithOverflow(APInt &Result, const APInt &In1, const APInt &In2, bool IsSigned=false)
Compute Result = In1-In2, returning true if the result overflowed for this type.
static bool canRewriteGEPAsOffset(Value *Start, Value *Base, GEPNoWrapFlags &NW, const DataLayout &DL, SetVector< Value * > &Explored)
Returns true if we can rewrite Start as a GEP with pointer Base and some integer offset.
static Instruction * foldFCmpFpTrunc(FCmpInst &I, const Instruction &FPTrunc, const Constant &C)
static Instruction * foldICmpXNegX(ICmpInst &I, InstCombiner::BuilderTy &Builder)
static Instruction * processUGT_ADDCST_ADD(ICmpInst &I, Value *A, Value *B, ConstantInt *CI2, ConstantInt *CI1, InstCombinerImpl &IC)
The caller has matched a pattern of the form: I = icmp ugt (add (add A, B), CI2), CI1 If this is of t...
static Value * foldShiftIntoShiftInAnotherHandOfAndInICmp(ICmpInst &I, const SimplifyQuery SQ, InstCombiner::BuilderTy &Builder)
static bool isSignTest(ICmpInst::Predicate &Pred, const APInt &C)
Returns true if the exploded icmp can be expressed as a signed comparison to zero and updates the pre...
static Instruction * transformToIndexedCompare(GEPOperator *GEPLHS, Value *RHS, CmpPredicate Cond, const DataLayout &DL, InstCombiner &IC)
Converts (CMP GEPLHS, RHS) if this change would make RHS a constant.
static Instruction * foldCtpopPow2Test(ICmpInst &I, IntrinsicInst *CtpopLhs, const APInt &CRhs, InstCombiner::BuilderTy &Builder, const SimplifyQuery &Q)
static Instruction * foldFCmpFAbsFSubIntToFP(FCmpInst &I, InstCombinerImpl &IC)
Fold: fabs(uitofp(a) - uitofp(b)) pred C --> a == b where 'pred' is olt, ult, ogt,...
static void setInsertionPoint(IRBuilder<> &Builder, Value *V, bool Before=true)
static bool isNeutralValue(Instruction::BinaryOps BinaryOp, Value *RHS, bool IsSigned)
static bool isMultipleOf(Value *X, const APInt &C, const SimplifyQuery &Q)
Return true if X is a multiple of C.
static Value * foldICmpWithTruncSignExtendedVal(ICmpInst &I, InstCombiner::BuilderTy &Builder)
Some comparisons can be simplified.
static Instruction * foldICmpOrXX(ICmpInst &I, const SimplifyQuery &Q, InstCombinerImpl &IC)
This file provides internal interfaces used to implement the InstCombine.
This file provides the interface for the instcombine pass implementation.
const AbstractManglingParser< Derived, Alloc >::OperatorInfo AbstractManglingParser< Derived, Alloc >::Ops[]
static bool isZero(Value *V, const DataLayout &DL, DominatorTree *DT, AssumptionCache *AC)
Definition Lint.cpp:539
#define F(x, y, z)
Definition MD5.cpp:54
#define I(x, y, z)
Definition MD5.cpp:57
#define T1
uint64_t IntrinsicInst * II
#define P(N)
const SmallVectorImpl< MachineOperand > & Cond
static cl::opt< RegAllocEvictionAdvisorAnalysisLegacy::AdvisorMode > Mode("regalloc-enable-advisor", cl::Hidden, cl::init(RegAllocEvictionAdvisorAnalysisLegacy::AdvisorMode::Default), cl::desc("Enable regalloc advisor mode"), cl::values(clEnumValN(RegAllocEvictionAdvisorAnalysisLegacy::AdvisorMode::Default, "default", "Default"), clEnumValN(RegAllocEvictionAdvisorAnalysisLegacy::AdvisorMode::Release, "release", "precompiled"), clEnumValN(RegAllocEvictionAdvisorAnalysisLegacy::AdvisorMode::Development, "development", "for training")))
This file implements a set that has insertion order iteration characteristics.
This file defines the 'Statistic' class, which is designed to be an easy way to expose various metric...
#define STATISTIC(VARNAME, DESC)
Definition Statistic.h:171
static TableGen::Emitter::Opt Y("gen-skeleton-entry", EmitSkeleton, "Generate example skeleton entry")
static SymbolRef::Type getType(const Symbol *Sym)
Definition TapiFile.cpp:39
Value * RHS
Value * LHS
BinaryOperator * Mul
cmpResult
IEEE-754R 5.11: Floating Point Comparison Relations.
Definition APFloat.h:334
static constexpr roundingMode rmTowardZero
Definition APFloat.h:348
static constexpr roundingMode rmNearestTiesToEven
Definition APFloat.h:344
opStatus
IEEE-754R 7: Default exception handling.
Definition APFloat.h:360
LLVM_ABI opStatus convert(const fltSemantics &ToSemantics, roundingMode RM, bool *losesInfo)
Definition APFloat.cpp:5890
void clearSign()
Definition APFloat.h:1353
bool isNaN() const
Definition APFloat.h:1514
static APFloat getOne(const fltSemantics &Sem, bool Negative=false)
Factory for Positive and Negative One.
Definition APFloat.h:1143
bool isZero() const
Definition APFloat.h:1512
static APFloat getSmallestNormalized(const fltSemantics &Sem, bool Negative=false)
Returns the smallest (by magnitude) normalized finite number in the given semantics.
Definition APFloat.h:1213
APInt bitcastToAPInt() const
Definition APFloat.h:1408
static APFloat getLargest(const fltSemantics &Sem, bool Negative=false)
Returns the largest finite number in the given semantics.
Definition APFloat.h:1193
opStatus convertToInteger(MutableArrayRef< integerPart > Input, unsigned int Width, bool IsSigned, roundingMode RM, bool *IsExact) const
Definition APFloat.h:1387
opStatus next(bool nextDown)
Definition APFloat.h:1309
static APFloat getInf(const fltSemantics &Sem, bool Negative=false)
Factory for Positive and Negative Infinity.
Definition APFloat.h:1153
LLVM_ABI FPClassTest classify() const
Return the FPClassTest which will return true for the value.
Definition APFloat.cpp:5819
opStatus roundToIntegral(roundingMode RM)
Definition APFloat.h:1303
bool isInfinity() const
Definition APFloat.h:1513
Class for arbitrary precision integers.
Definition APInt.h:78
LLVM_ABI APInt udiv(const APInt &RHS) const
Unsigned division operation.
Definition APInt.cpp:1615
static APInt getAllOnes(unsigned numBits)
Return an APInt of a specified width with all bits set.
Definition APInt.h:235
static LLVM_ABI void udivrem(const APInt &LHS, const APInt &RHS, APInt &Quotient, APInt &Remainder)
Dual division/remainder interface.
Definition APInt.cpp:1809
bool isNegatedPowerOf2() const
Check if this APInt's negated value is a power of two greater than zero.
Definition APInt.h:450
LLVM_ABI APInt zext(unsigned width) const
Zero extend to a new width.
Definition APInt.cpp:1054
static APInt getSignMask(unsigned BitWidth)
Get the SignMask for a specific bit width.
Definition APInt.h:230
bool isMinSignedValue() const
Determine if this is the smallest signed value.
Definition APInt.h:424
uint64_t getZExtValue() const
Get zero extended value.
Definition APInt.h:1563
unsigned getActiveBits() const
Compute the number of active bits in the value.
Definition APInt.h:1535
LLVM_ABI APInt trunc(unsigned width) const
Truncate to new width.
Definition APInt.cpp:967
static APInt getMaxValue(unsigned numBits)
Gets maximum unsigned value of APInt for specific bit width.
Definition APInt.h:207
void setBit(unsigned BitPosition)
Set the given bit to 1 whose position is given as "bitPosition".
Definition APInt.h:1353
APInt abs() const
Get the absolute value.
Definition APInt.h:1818
unsigned ceilLogBase2() const
Definition APInt.h:1787
bool sgt(const APInt &RHS) const
Signed greater than comparison.
Definition APInt.h:1208
bool isAllOnes() const
Determine if all bits are set. This is true for zero-width values.
Definition APInt.h:372
LLVM_ABI APInt usub_ov(const APInt &RHS, bool &Overflow) const
Definition APInt.cpp:1999
bool ugt(const APInt &RHS) const
Unsigned greater than comparison.
Definition APInt.h:1189
bool isZero() const
Determine if this value is zero, i.e. all bits are clear.
Definition APInt.h:381
bool isSignMask() const
Check if the APInt's value is returned by getSignMask.
Definition APInt.h:467
unsigned getBitWidth() const
Return the number of bits in the APInt.
Definition APInt.h:1511
bool ult(const APInt &RHS) const
Unsigned less than comparison.
Definition APInt.h:1118
static APInt getSignedMaxValue(unsigned numBits)
Gets maximum signed value of APInt for a specific bit width.
Definition APInt.h:210
static APInt getMinValue(unsigned numBits)
Gets minimum unsigned value of APInt for a specific bit width.
Definition APInt.h:217
bool isNegative() const
Determine sign of this APInt.
Definition APInt.h:330
LLVM_ABI APInt sadd_ov(const APInt &RHS, bool &Overflow) const
Definition APInt.cpp:1979
bool eq(const APInt &RHS) const
Equality comparison.
Definition APInt.h:1086
LLVM_ABI APInt sdiv(const APInt &RHS) const
Signed division function for APInt.
Definition APInt.cpp:1686
LLVM_ABI APInt uadd_ov(const APInt &RHS, bool &Overflow) const
Definition APInt.cpp:1986
void negate()
Negate this APInt in place.
Definition APInt.h:1491
unsigned countr_zero() const
Count the number of trailing zero bits.
Definition APInt.h:1662
unsigned countl_zero() const
The APInt version of std::countl_zero.
Definition APInt.h:1621
static APInt getSignedMinValue(unsigned numBits)
Gets minimum signed value of APInt for a specific bit width.
Definition APInt.h:220
bool isStrictlyPositive() const
Determine if this APInt Value is positive.
Definition APInt.h:357
void flipAllBits()
Toggle every bit to its opposite value.
Definition APInt.h:1475
unsigned countl_one() const
Count the number of leading one bits.
Definition APInt.h:1638
unsigned logBase2() const
Definition APInt.h:1784
uint64_t getLimitedValue(uint64_t Limit=UINT64_MAX) const
If this value is smaller than the specified limit, return it, otherwise return the limit value.
Definition APInt.h:476
APInt ashr(unsigned ShiftAmt) const
Arithmetic right-shift function.
Definition APInt.h:834
bool isMaxSignedValue() const
Determine if this is the largest signed value.
Definition APInt.h:406
bool ule(const APInt &RHS) const
Unsigned less or equal comparison.
Definition APInt.h:1157
APInt shl(unsigned shiftAmt) const
Left-shift function.
Definition APInt.h:880
bool isPowerOf2() const
Check if this APInt's value is a power of two greater than zero.
Definition APInt.h:441
static APInt getLowBitsSet(unsigned numBits, unsigned loBitsSet)
Constructs an APInt value that has the bottom loBitsSet bits set.
Definition APInt.h:307
static APInt getHighBitsSet(unsigned numBits, unsigned hiBitsSet)
Constructs an APInt value that has the top hiBitsSet bits set.
Definition APInt.h:297
static APInt getZero(unsigned numBits)
Get the '0' value for the specified bit-width.
Definition APInt.h:201
bool sge(const APInt &RHS) const
Signed greater or equal comparison.
Definition APInt.h:1244
LLVM_ABI APInt ssub_ov(const APInt &RHS, bool &Overflow) const
Definition APInt.cpp:1992
bool isOne() const
Determine if this is a value of 1.
Definition APInt.h:390
static APInt getBitsSetFrom(unsigned numBits, unsigned loBit)
Constructs an APInt value that has a contiguous range of bits set.
Definition APInt.h:287
static APInt getOneBitSet(unsigned numBits, unsigned BitNo)
Return an APInt with exactly one bit set in the result.
Definition APInt.h:240
APInt lshr(unsigned shiftAmt) const
Logical right-shift function.
Definition APInt.h:858
unsigned countr_one() const
Count the number of trailing one bits.
Definition APInt.h:1679
bool uge(const APInt &RHS) const
Unsigned greater or equal comparison.
Definition APInt.h:1228
An arbitrary precision integer that knows its signedness.
Definition APSInt.h:24
static APSInt getMinValue(uint32_t numBits, bool Unsigned)
Return the APSInt representing the minimum integer value with the given bit width and signedness.
Definition APSInt.h:310
static APSInt getMaxValue(uint32_t numBits, bool Unsigned)
Return the APSInt representing the maximum integer value with the given bit width and signedness.
Definition APSInt.h:302
an instruction to allocate memory on the stack
ArrayRef - Represent a constant reference to an array (0 or more elements consecutively in memory),...
Definition ArrayRef.h:40
LLVM Basic Block Representation.
Definition BasicBlock.h:62
LLVM_ABI const_iterator getFirstInsertionPt() const
Returns an iterator to the first instruction in this block that is suitable for inserting a non-PHI i...
LLVM_ABI const BasicBlock * getSinglePredecessor() const
Return the predecessor of this block if it has a single predecessor block.
const Instruction * getTerminator() const LLVM_READONLY
Returns the terminator instruction; assumes that the block is well-formed.
Definition BasicBlock.h:237
BinaryOps getOpcode() const
Definition InstrTypes.h:374
static LLVM_ABI BinaryOperator * CreateNot(Value *Op, const Twine &Name="", InsertPosition InsertBefore=nullptr)
static LLVM_ABI BinaryOperator * Create(BinaryOps Op, Value *S1, Value *S2, const Twine &Name=Twine(), InsertPosition InsertBefore=nullptr)
Construct a binary instruction, given the opcode and the two operands.
Value * getArgOperand(unsigned i) const
This class represents a function call, abstracting a target machine's calling convention.
static CallInst * Create(FunctionType *Ty, Value *F, const Twine &NameStr="", InsertPosition InsertBefore=nullptr)
This class is the base class for the comparison instructions.
Definition InstrTypes.h:664
static Type * makeCmpResultType(Type *opnd_type)
Create a result type for fcmp/icmp.
Definition InstrTypes.h:986
Predicate getStrictPredicate() const
For example, SGE -> SGT, SLE -> SLT, ULE -> ULT, UGE -> UGT.
Definition InstrTypes.h:858
static LLVM_ABI Predicate getFlippedStrictnessPredicate(Predicate pred)
This is a static version that you can use without an instruction available.
Predicate
This enumeration lists the possible predicates for CmpInst subclasses.
Definition InstrTypes.h:676
@ FCMP_OEQ
0 0 0 1 True if ordered and equal
Definition InstrTypes.h:679
@ FCMP_TRUE
1 1 1 1 Always true (always folded)
Definition InstrTypes.h:693
@ ICMP_SLT
signed less than
Definition InstrTypes.h:705
@ ICMP_SLE
signed less or equal
Definition InstrTypes.h:706
@ FCMP_OLT
0 1 0 0 True if ordered and less than
Definition InstrTypes.h:682
@ FCMP_ULE
1 1 0 1 True if unordered, less than, or equal
Definition InstrTypes.h:691
@ FCMP_OGT
0 0 1 0 True if ordered and greater than
Definition InstrTypes.h:680
@ FCMP_OGE
0 0 1 1 True if ordered and greater than or equal
Definition InstrTypes.h:681
@ ICMP_UGE
unsigned greater or equal
Definition InstrTypes.h:700
@ ICMP_UGT
unsigned greater than
Definition InstrTypes.h:699
@ ICMP_SGT
signed greater than
Definition InstrTypes.h:703
@ FCMP_ULT
1 1 0 0 True if unordered or less than
Definition InstrTypes.h:690
@ FCMP_ONE
0 1 1 0 True if ordered and operands are unequal
Definition InstrTypes.h:684
@ FCMP_UEQ
1 0 0 1 True if unordered or equal
Definition InstrTypes.h:687
@ ICMP_ULT
unsigned less than
Definition InstrTypes.h:701
@ FCMP_UGT
1 0 1 0 True if unordered or greater than
Definition InstrTypes.h:688
@ FCMP_OLE
0 1 0 1 True if ordered and less than or equal
Definition InstrTypes.h:683
@ FCMP_ORD
0 1 1 1 True if ordered (no nans)
Definition InstrTypes.h:685
@ ICMP_NE
not equal
Definition InstrTypes.h:698
@ ICMP_SGE
signed greater or equal
Definition InstrTypes.h:704
@ FCMP_UNE
1 1 1 0 True if unordered or not equal
Definition InstrTypes.h:692
@ ICMP_ULE
unsigned less or equal
Definition InstrTypes.h:702
@ FCMP_UGE
1 0 1 1 True if unordered, greater than, or equal
Definition InstrTypes.h:689
@ FCMP_FALSE
0 0 0 0 Always false (always folded)
Definition InstrTypes.h:678
@ FCMP_UNO
1 0 0 0 True if unordered: isnan(X) | isnan(Y)
Definition InstrTypes.h:686
bool isSigned() const
Definition InstrTypes.h:930
Predicate getSwappedPredicate() const
For example, EQ->EQ, SLE->SGE, ULT->UGT, OEQ->OEQ, ULE->UGE, OLT->OGT, etc.
Definition InstrTypes.h:827
bool isTrueWhenEqual() const
This is just a convenience.
Definition InstrTypes.h:942
static LLVM_ABI CmpInst * Create(OtherOps Op, Predicate Pred, Value *S1, Value *S2, const Twine &Name="", InsertPosition InsertBefore=nullptr)
Construct a compare instruction, given the opcode, the predicate and the two operands.
Predicate getNonStrictPredicate() const
For example, SGT -> SGE, SLT -> SLE, ULT -> ULE, UGT -> UGE.
Definition InstrTypes.h:871
static LLVM_ABI bool isStrictPredicate(Predicate predicate)
This is a static version that you can use without an instruction available.
Predicate getInversePredicate() const
For example, EQ -> NE, UGT -> ULE, SLT -> SGE, OEQ -> UNE, UGT -> OLE, OLT -> UGE,...
Definition InstrTypes.h:789
Predicate getPredicate() const
Return the predicate for this instruction.
Definition InstrTypes.h:765
static bool isIntPredicate(Predicate P)
Definition InstrTypes.h:776
bool isUnsigned() const
Definition InstrTypes.h:936
An abstraction over a floating-point predicate, and a pack of an integer predicate with samesign info...
static LLVM_ABI CmpPredicate getSwapped(CmpPredicate P)
Get the swapped predicate of a CmpPredicate.
Conditional Branch instruction.
static LLVM_ABI Constant * getIntToPtr(Constant *C, Type *Ty, bool OnlyIfReduced=false)
static LLVM_ABI Constant * getPointerBitCastOrAddrSpaceCast(Constant *C, Type *Ty)
Create a BitCast or AddrSpaceCast for a pointer type depending on the address space.
static LLVM_ABI Constant * getSub(Constant *C1, Constant *C2, bool HasNUW=false, bool HasNSW=false)
static LLVM_ABI Constant * getNot(Constant *C)
static LLVM_ABI Constant * getPtrToInt(Constant *C, Type *Ty, bool OnlyIfReduced=false)
static LLVM_ABI Constant * getXor(Constant *C1, Constant *C2)
static LLVM_ABI Constant * getNeg(Constant *C, bool HasNSW=false)
static LLVM_ABI Constant * getZero(Type *Ty, bool Negative=false)
This is the shared class of boolean and integer constants.
Definition Constants.h:87
uint64_t getLimitedValue(uint64_t Limit=~0ULL) const
getLimitedValue - If the value is smaller than the specified limit, return it, otherwise return the l...
Definition Constants.h:269
static LLVM_ABI ConstantInt * getTrue(LLVMContext &Context)
static ConstantInt * getSigned(IntegerType *Ty, int64_t V, bool ImplicitTrunc=false)
Return a ConstantInt with the specified value for the specified type.
Definition Constants.h:135
bool isZero() const
This is just a convenience method to make client code smaller for a common code.
Definition Constants.h:219
static LLVM_ABI ConstantInt * getFalse(LLVMContext &Context)
unsigned getBitWidth() const
getBitWidth - Return the scalar bitwidth of this constant.
Definition Constants.h:162
const APInt & getValue() const
Return the constant as an APInt value reference.
Definition Constants.h:159
static LLVM_ABI ConstantInt * getBool(LLVMContext &Context, bool V)
This class represents a range of values.
LLVM_ABI ConstantRange add(const ConstantRange &Other) const
Return a new range representing the possible values resulting from an addition of a value in this ran...
LLVM_ABI std::optional< ConstantRange > exactUnionWith(const ConstantRange &CR) const
Union the two ranges and return the result if it can be represented exactly, otherwise return std::nu...
LLVM_ABI bool getEquivalentICmp(CmpInst::Predicate &Pred, APInt &RHS) const
Set up Pred and RHS such that ConstantRange::makeExactICmpRegion(Pred, RHS) == *this.
LLVM_ABI ConstantRange subtract(const APInt &CI) const
Subtract the specified constant from the endpoints of this constant range.
const APInt * getSingleElement() const
If this set contains a single element, return it, otherwise return null.
LLVM_ABI ConstantRange difference(const ConstantRange &CR) const
Subtract the specified range from this range (aka relative complement of the sets).
LLVM_ABI bool isEmptySet() const
Return true if this set contains no members.
LLVM_ABI ConstantRange truncate(uint32_t BitWidth, unsigned NoWrapKind=0) const
Return a new range in the specified integer type, which must be strictly smaller than the current typ...
static LLVM_ABI ConstantRange makeExactICmpRegion(CmpInst::Predicate Pred, const APInt &Other)
Produce the exact range such that all values in the returned range satisfy the given predicate with a...
LLVM_ABI ConstantRange inverse() const
Return a new range that is the logical not of the current set.
LLVM_ABI std::optional< ConstantRange > exactIntersectWith(const ConstantRange &CR) const
Intersect the two ranges and return the result if it can be represented exactly, otherwise return std...
LLVM_ABI ConstantRange intersectWith(const ConstantRange &CR, PreferredRangeType Type=Smallest) const
Return the range that results from the intersection of this range with another range.
static ConstantRange getNonEmpty(APInt Lower, APInt Upper)
Create non-empty constant range with the given bounds.
LLVM_ABI ConstantRange sub(const ConstantRange &Other) const
Return a new range representing the possible values resulting from a subtraction of a value in this r...
static LLVM_ABI ConstantRange makeExactNoWrapRegion(Instruction::BinaryOps BinOp, const APInt &Other, unsigned NoWrapKind)
Produce the range that contains X if and only if "X BinOp Other" does not wrap.
static LLVM_ABI Constant * getSplat(ElementCount EC, Constant *Elt)
Return a ConstantVector with the specified constant in each element.
This is an important base class in LLVM.
Definition Constant.h:43
static LLVM_ABI Constant * getIntegerValue(Type *Ty, const APInt &V)
Return the value for an integer or pointer constant, or a vector thereof, with the given scalar value...
static LLVM_ABI Constant * getAllOnesValue(Type *Ty)
LLVM_ABI bool isAllOnesValue() const
Return true if this is the value that would be returned by getAllOnesValue.
Definition Constants.cpp:95
LLVM_ABI const APInt & getUniqueInteger() const
If C is a constant integer then return its value, otherwise C must be a vector of constant integers,...
static LLVM_ABI Constant * getNullValue(Type *Ty)
Constructor to create a '0' constant of arbitrary type.
LLVM_ABI bool isNullValue() const
Return true if this is the value that would be returned by getNullValue.
Definition Constants.cpp:74
A parsed version of the target data layout string in and methods for querying it.
Definition DataLayout.h:64
iterator find(const_arg_type_t< KeyT > Val)
Definition DenseMap.h:178
iterator end()
Definition DenseMap.h:81
bool contains(const_arg_type_t< KeyT > Val) const
Return true if the specified key is in the map, false otherwise.
Definition DenseMap.h:169
static ExtractValueInst * Create(Value *Agg, ArrayRef< unsigned > Idxs, const Twine &NameStr="", InsertPosition InsertBefore=nullptr)
This instruction compares its operands according to the predicate given to the constructor.
static bool isCommutative(Predicate Pred)
static bool isEquality(Predicate Pred)
Represents flags for the getelementptr instruction/expression.
bool hasNoUnsignedSignedWrap() const
bool hasNoUnsignedWrap() const
bool isInBounds() const
GEPNoWrapFlags intersectForOffsetAdd(GEPNoWrapFlags Other) const
Given (gep (gep p, x), y), determine the nowrap flags for (gep p, x+y).
static GEPNoWrapFlags none()
bool isInBounds() const
Test whether this is an inbounds GEP, as defined by LangRef.html.
Definition Operator.h:430
LLVM_ABI Type * getSourceElementType() const
Definition Operator.cpp:82
Value * getPointerOperand()
Definition Operator.h:457
GEPNoWrapFlags getNoWrapFlags() const
Definition Operator.h:425
bool hasAllConstantIndices() const
Return true if all of the indices of this GEP are constant integers.
Definition Operator.h:504
an instruction for type-safe pointer arithmetic to access elements of arrays and structs
This instruction compares its operands according to the predicate given to the constructor.
static bool isGE(Predicate P)
Return true if the predicate is SGE or UGE.
static LLVM_ABI bool compare(const APInt &LHS, const APInt &RHS, ICmpInst::Predicate Pred)
Return result of LHS Pred RHS comparison.
static bool isLT(Predicate P)
Return true if the predicate is SLT or ULT.
static bool isGT(Predicate P)
Return true if the predicate is SGT or UGT.
Predicate getFlippedSignednessPredicate() const
For example, SLT->ULT, ULT->SLT, SLE->ULE, ULE->SLE, EQ->EQ.
Predicate getSignedPredicate() const
For example, EQ->EQ, SLE->SLE, UGT->SGT, etc.
bool isEquality() const
Return true if this predicate is either EQ or NE.
static bool isEquality(Predicate P)
Return true if this predicate is either EQ or NE.
bool isRelational() const
Return true if the predicate is relational (not EQ or NE).
Predicate getUnsignedPredicate() const
For example, EQ->EQ, SLE->ULE, UGT->UGT, etc.
static bool isLE(Predicate P)
Return true if the predicate is SLE or ULE.
Common base class shared among various IRBuilders.
Definition IRBuilder.h:114
LLVM_ABI CallInst * CreateUnaryIntrinsic(Intrinsic::ID ID, Value *V, FMFSource FMFSource={}, const Twine &Name="")
Create a call to intrinsic ID with 1 operand which is mangled on its type.
Value * CreateAnd(Value *LHS, Value *RHS, const Twine &Name="")
Definition IRBuilder.h:1591
void SetInsertPoint(BasicBlock *TheBB)
This specifies that created instructions should be appended to the end of the specified block.
Definition IRBuilder.h:207
Value * CreateICmp(CmpInst::Predicate P, Value *LHS, Value *RHS, const Twine &Name="")
Definition IRBuilder.h:2477
Value * CreateOr(Value *LHS, Value *RHS, const Twine &Name="", bool IsDisjoint=false)
Definition IRBuilder.h:1613
ConstantInt * getInt(const APInt &AI)
Get a constant integer value.
Definition IRBuilder.h:544
This provides a uniform API for creating instructions and inserting them into a basic block: either a...
Definition IRBuilder.h:2847
Instruction * foldICmpShrConstant(ICmpInst &Cmp, BinaryOperator *Shr, const APInt &C)
Fold icmp ({al}shr X, Y), C.
Instruction * foldICmpWithZextOrSext(ICmpInst &ICmp)
Instruction * foldICmpSelectConstant(ICmpInst &Cmp, SelectInst *Select, ConstantInt *C)
Instruction * foldICmpSRemConstant(ICmpInst &Cmp, BinaryOperator *UDiv, const APInt &C)
Instruction * foldICmpBinOpWithConstant(ICmpInst &Cmp, BinaryOperator *BO, const APInt &C)
Fold an icmp with BinaryOp and constant operand: icmp Pred BO, C.
Instruction * foldICmpOrConstant(ICmpInst &Cmp, BinaryOperator *Or, const APInt &C)
Fold icmp (or X, Y), C.
Instruction * foldICmpTruncWithTruncOrExt(ICmpInst &Cmp, const SimplifyQuery &Q)
Fold icmp (trunc nuw/nsw X), (trunc nuw/nsw Y).
Instruction * foldSignBitTest(ICmpInst &I)
Fold equality-comparison between zero and any (maybe truncated) right-shift by one-less-than-bitwidth...
Instruction * foldOpIntoPhi(Instruction &I, PHINode *PN, bool AllowMultipleUses=false)
Given a binary operator, cast instruction, or select which has a PHI node as operand #0,...
Value * insertRangeTest(Value *V, const APInt &Lo, const APInt &Hi, bool isSigned, bool Inside)
Emit a computation of: (V >= Lo && V < Hi) if Inside is true, otherwise (V < Lo || V >= Hi).
Instruction * foldICmpBinOp(ICmpInst &Cmp, const SimplifyQuery &SQ)
Try to fold icmp (binop), X or icmp X, (binop).
Instruction * foldCmpLoadFromIndexedGlobal(LoadInst *LI, GetElementPtrInst *GEP, CmpInst &ICI, ConstantInt *AndCst=nullptr)
This is called when we see this pattern: cmp pred (load (gep GV, ...)), cmpcst where GV is a global v...
Instruction * foldICmpSubConstant(ICmpInst &Cmp, BinaryOperator *Sub, const APInt &C)
Fold icmp (sub X, Y), C.
Instruction * foldICmpWithClamp(ICmpInst &Cmp, Value *X, MinMaxIntrinsic *Min)
Match and fold patterns like: icmp eq/ne X, min(max(X, Lo), Hi) which represents a range check and ca...
Instruction * foldICmpInstWithConstantNotInt(ICmpInst &Cmp)
Handle icmp with constant (but not simple integer constant) RHS.
bool SimplifyDemandedBits(Instruction *I, unsigned Op, const APInt &DemandedMask, KnownBits &Known, const SimplifyQuery &Q, unsigned Depth=0) override
This form of SimplifyDemandedBits simplifies the specified instruction operand if possible,...
Instruction * foldICmpShlConstConst(ICmpInst &I, Value *ShAmt, const APInt &C1, const APInt &C2)
Handle "(icmp eq/ne (shl AP2, A), AP1)" -> (icmp eq/ne A, TrailingZeros(AP1) - TrailingZeros(AP2)).
Value * reassociateShiftAmtsOfTwoSameDirectionShifts(BinaryOperator *Sh0, const SimplifyQuery &SQ, bool AnalyzeForSignBitExtraction=false)
Instruction * foldICmpEqIntrinsicWithConstant(ICmpInst &ICI, IntrinsicInst *II, const APInt &C)
Fold an equality icmp with LLVM intrinsic and constant operand.
Instruction * FoldOpIntoSelect(Instruction &Op, SelectInst *SI, bool FoldWithMultiUse=false, bool SimplifyBothArms=false)
Given an instruction with a select as one operand and a constant as the other operand,...
Value * foldMultiplicationOverflowCheck(ICmpInst &Cmp)
Fold (-1 u/ x) u< y ((x * y) ?
Instruction * foldICmpWithConstant(ICmpInst &Cmp)
Fold icmp Pred X, C.
CmpInst * canonicalizeICmpPredicate(CmpInst &I)
If we have a comparison with a non-canonical predicate, if we can update all the users,...
Instruction * eraseInstFromFunction(Instruction &I) override
Combiner aware instruction erasure.
Instruction * foldICmpWithZero(ICmpInst &Cmp)
Instruction * foldICmpCommutative(CmpPredicate Pred, Value *Op0, Value *Op1, ICmpInst &CxtI)
Instruction * foldICmpBinOpEqualityWithConstant(ICmpInst &Cmp, BinaryOperator *BO, const APInt &C)
Fold an icmp equality instruction with binary operator LHS and constant RHS: icmp eq/ne BO,...
Instruction * foldICmpUsingBoolRange(ICmpInst &I)
If one operand of an icmp is effectively a bool (value range of {0,1}), then try to reduce patterns b...
Instruction * foldICmpWithTrunc(ICmpInst &Cmp)
Instruction * foldCmpSelectOfConstants(CmpInst &I)
Fold fcmp/icmp pred (select C1, TV1, FV1), (select C2, TV2, FV2) where all true/false values are cons...
Instruction * foldICmpIntrinsicWithConstant(ICmpInst &ICI, IntrinsicInst *II, const APInt &C)
Fold an icmp with LLVM intrinsic and constant operand: icmp Pred II, C.
bool matchThreeWayIntCompare(SelectInst *SI, Value *&LHS, Value *&RHS, ConstantInt *&Less, ConstantInt *&Equal, ConstantInt *&Greater)
Match a select chain which produces one of three values based on whether the LHS is less than,...
Instruction * visitFCmpInst(FCmpInst &I)
Instruction * foldICmpUsingKnownBits(ICmpInst &Cmp)
Try to fold the comparison based on range information we can get by checking whether bits are known t...
Instruction * foldICmpDivConstant(ICmpInst &Cmp, BinaryOperator *Div, const APInt &C)
Fold icmp ({su}div X, Y), C.
Instruction * foldIRemByPowerOfTwoToBitTest(ICmpInst &I)
If we have: icmp eq/ne (urem/srem x, y), 0 iff y is a power-of-two, we can replace this with a bit te...
Instruction * foldFCmpIntToFPConst(FCmpInst &I, Instruction *LHSI, Constant *RHSC)
Fold fcmp ([us]itofp x, cst) if possible.
Instruction * foldICmpUDivConstant(ICmpInst &Cmp, BinaryOperator *UDiv, const APInt &C)
Fold icmp (udiv X, Y), C.
Instruction * foldICmpAddOpConst(Value *X, const APInt &C, CmpPredicate Pred)
Fold "icmp pred (X+C), X".
Instruction * foldICmpWithCastOp(ICmpInst &ICmp)
Handle icmp (cast x), (cast or constant).
Instruction * foldICmpTruncConstant(ICmpInst &Cmp, TruncInst *Trunc, const APInt &C)
Fold icmp (trunc X), C.
Instruction * foldICmpAddConstant(ICmpInst &Cmp, BinaryOperator *Add, const APInt &C)
Fold icmp (add X, Y), C.
Instruction * foldICmpMulConstant(ICmpInst &Cmp, BinaryOperator *Mul, const APInt &C)
Fold icmp (mul X, Y), C.
Instruction * tryFoldInstWithCtpopWithNot(Instruction *I)
Instruction * foldICmpXorConstant(ICmpInst &Cmp, BinaryOperator *Xor, const APInt &C)
Fold icmp (xor X, Y), C.
Instruction * foldSelectICmp(CmpPredicate Pred, SelectInst *SI, Value *RHS, const ICmpInst &I)
Instruction * foldICmpInstWithConstantAllowPoison(ICmpInst &Cmp, const APInt &C)
Try to fold integer comparisons with a constant operand: icmp Pred X, C where X is some kind of instr...
Instruction * foldIsMultipleOfAPowerOfTwo(ICmpInst &Cmp)
Fold icmp eq (num + mask) & ~mask, num to icmp eq (and num, mask), 0 Where mask is a low bit mask.
Instruction * foldICmpAndShift(ICmpInst &Cmp, BinaryOperator *And, const APInt &C1, const APInt &C2)
Fold icmp (and (sh X, Y), C2), C1.
Instruction * foldICmpBinOpWithConstantViaTruthTable(ICmpInst &Cmp, BinaryOperator *BO, const APInt &C)
Instruction * foldICmpInstWithConstant(ICmpInst &Cmp)
Try to fold integer comparisons with a constant operand: icmp Pred X, C where X is some kind of instr...
Instruction * foldICmpXorShiftConst(ICmpInst &Cmp, BinaryOperator *Xor, const APInt &C)
For power-of-2 C: ((X s>> ShiftC) ^ X) u< C --> (X + C) u< (C << 1) ((X s>> ShiftC) ^ X) u> (C - 1) -...
Instruction * foldICmpShlConstant(ICmpInst &Cmp, BinaryOperator *Shl, const APInt &C)
Fold icmp (shl X, Y), C.
Instruction * foldICmpAndConstant(ICmpInst &Cmp, BinaryOperator *And, const APInt &C)
Fold icmp (and X, Y), C.
Instruction * foldICmpEquality(ICmpInst &Cmp)
Instruction * foldICmpWithMinMax(Instruction &I, MinMaxIntrinsic *MinMax, Value *Z, CmpPredicate Pred)
Fold icmp Pred min|max(X, Y), Z.
bool dominatesAllUses(const Instruction *DI, const Instruction *UI, const BasicBlock *DB) const
True when DB dominates all uses of DI except UI.
bool foldAllocaCmp(AllocaInst *Alloca)
Instruction * visitICmpInst(ICmpInst &I)
OverflowResult computeOverflow(Instruction::BinaryOps BinaryOp, bool IsSigned, Value *LHS, Value *RHS, Instruction *CxtI) const
Instruction * foldICmpWithDominatingICmp(ICmpInst &Cmp)
Canonicalize icmp instructions based on dominating conditions.
bool replacedSelectWithOperand(SelectInst *SI, const ICmpInst *Icmp, const unsigned SIOpd)
Try to replace select with select operand SIOpd in SI-ICmp sequence.
Instruction * foldICmpShrConstConst(ICmpInst &I, Value *ShAmt, const APInt &C1, const APInt &C2)
Handle "(icmp eq/ne (ashr/lshr AP2, A), AP1)" -> (icmp eq/ne A, Log2(AP2/AP1)) -> (icmp eq/ne A,...
void freelyInvertAllUsersOf(Value *V, Value *IgnoredUser=nullptr)
Freely adapt every user of V as-if V was changed to !V.
Instruction * foldICmpAndConstConst(ICmpInst &Cmp, BinaryOperator *And, const APInt &C1)
Fold icmp (and X, C2), C1.
Instruction * foldICmpBitCast(ICmpInst &Cmp)
Instruction * foldGEPICmp(GEPOperator *GEPLHS, Value *RHS, CmpPredicate Cond, Instruction &I)
Fold comparisons between a GEP instruction and something else.
The core instruction combiner logic.
OverflowResult computeOverflowForSignedSub(const Value *LHS, const Value *RHS, const Instruction *CxtI) const
SimplifyQuery SQ
unsigned ComputeMaxSignificantBits(const Value *Op, const Instruction *CxtI=nullptr, unsigned Depth=0) const
IRBuilder< TargetFolder, IRBuilderCallbackInserter > BuilderTy
An IRBuilder that automatically inserts new instructions into the worklist.
bool isFreeToInvert(Value *V, bool WillInvertAllUses, bool &DoesConsume)
Return true if the specified value is free to invert (apply ~ to).
OverflowResult computeOverflowForUnsignedMul(const Value *LHS, const Value *RHS, const Instruction *CxtI, bool IsNSW=false) const
static unsigned getComplexity(Value *V)
Assign a complexity or rank value to LLVM Values.
TargetLibraryInfo & TLI
Instruction * replaceInstUsesWith(Instruction &I, Value *V)
A combiner-aware RAUW-like routine.
uint64_t MaxArraySizeForCombine
Maximum size of array considered when transforming.
bool canBeCastedExactlyIntToFP(Value *V, Type *FPTy, bool IsSigned, const Instruction *CxtI=nullptr) const
OverflowResult computeOverflowForSignedAdd(const WithCache< const Value * > &LHS, const WithCache< const Value * > &RHS, const Instruction *CxtI) const
static Constant * SubOne(Constant *C)
Subtract one from a Constant.
OverflowResult computeOverflowForUnsignedSub(const Value *LHS, const Value *RHS, const Instruction *CxtI) const
static bool isCanonicalPredicate(CmpPredicate Pred)
Predicate canonicalization reduces the number of patterns that need to be matched by other transforms...
const DataLayout & DL
DomConditionCache DC
void computeKnownBits(const Value *V, KnownBits &Known, const Instruction *CxtI, unsigned Depth=0) const
bool canFreelyInvertAllUsersOf(Instruction *V, Value *IgnoredUser)
Given i1 V, can every user of V be freely adapted if V is changed to !V ?
void addToWorklist(Instruction *I)
Instruction * replaceOperand(Instruction &I, unsigned OpNum, Value *V)
Replace operand of instruction and add old operand to the worklist.
DominatorTree & DT
OverflowResult computeOverflowForSignedMul(const Value *LHS, const Value *RHS, const Instruction *CxtI) const
BuilderTy & Builder
OverflowResult computeOverflowForUnsignedAdd(const WithCache< const Value * > &LHS, const WithCache< const Value * > &RHS, const Instruction *CxtI) const
Value * getFreelyInverted(Value *V, bool WillInvertAllUses, BuilderTy *Builder, bool &DoesConsume)
const SimplifyQuery & getSimplifyQuery() const
bool isKnownToBeAPowerOfTwo(const Value *V, bool OrZero=false, const Instruction *CxtI=nullptr, unsigned Depth=0)
LLVM_ABI bool hasNoNaNs() const LLVM_READONLY
Determine whether the no-NaNs flag is set.
LLVM_ABI bool hasNoUnsignedWrap() const LLVM_READONLY
Determine whether the no unsigned wrap flag is set.
LLVM_ABI bool hasNoInfs() const LLVM_READONLY
Determine whether the no-infs flag is set.
bool isArithmeticShift() const
Return true if this is an arithmetic shift right.
LLVM_ABI bool hasNoSignedWrap() const LLVM_READONLY
Determine whether the no signed wrap flag is set.
LLVM_ABI bool isCommutative() const LLVM_READONLY
Return true if the instruction is commutative:
LLVM_ABI bool isExact() const LLVM_READONLY
Determine whether the exact flag is set.
unsigned getOpcode() const
Returns a member of one of the enums like Instruction::Add.
bool isShift() const
static LLVM_ABI IntegerType * get(LLVMContext &C, unsigned NumBits)
This static method is the primary way of constructing an IntegerType.
Definition Type.cpp:354
A wrapper class for inspecting calls to intrinsic functions.
Intrinsic::ID getIntrinsicID() const
Return the intrinsic ID of this intrinsic.
An instruction for reading from memory.
bool isVolatile() const
Return true if this is a load from a volatile memory location.
This class represents min/max intrinsics.
Value * getLHS() const
Value * getRHS() const
static bool isMin(Intrinsic::ID ID)
Whether the intrinsic is a smin or umin.
static bool isSigned(Intrinsic::ID ID)
Whether the intrinsic is signed or unsigned.
A Module instance is used to store all the information related to an LLVM module.
Definition Module.h:67
void addIncoming(Value *V, BasicBlock *BB)
Add an incoming value to the end of the PHI list.
static PHINode * Create(Type *Ty, unsigned NumReservedValues, const Twine &NameStr="", InsertPosition InsertBefore=nullptr)
Constructors - NumReservedValues is a hint for the number of incoming edges that this phi node will h...
Represents a saturating add/sub intrinsic.
This class represents the LLVM 'select' instruction.
static SelectInst * Create(Value *C, Value *S1, Value *S2, const Twine &NameStr="", InsertPosition InsertBefore=nullptr, const Instruction *MDFrom=nullptr)
A vector that has set insertion semantics.
Definition SetVector.h:57
size_type size() const
Determine the number of elements in the SetVector.
Definition SetVector.h:103
bool contains(const_arg_type key) const
Check if the SetVector contains the given key.
Definition SetVector.h:252
bool insert(const value_type &X)
Insert a new element into the SetVector.
Definition SetVector.h:151
This instruction constructs a fixed permutation of two input vectors.
This class consists of common code factored out of the SmallVector class to reduce code duplication b...
reference emplace_back(ArgTypes &&... Args)
void push_back(const T &Elt)
This is a 'vector' (really, a variable-sized array), optimized for the case when the array is small.
This class represents a truncation of integer types.
bool hasNoSignedWrap() const
Test whether this operation is known to never undergo signed overflow, aka the nsw property.
bool hasNoUnsignedWrap() const
Test whether this operation is known to never undergo unsigned overflow, aka the nuw property.
The instances of the Type class are immutable: once they are created, they are never changed.
Definition Type.h:46
bool isVectorTy() const
True if this is an instance of VectorType.
Definition Type.h:290
bool isIntOrIntVectorTy() const
Return true if this is an integer type or a vector of integer types.
Definition Type.h:263
bool isPointerTy() const
True if this is an instance of PointerType.
Definition Type.h:284
bool isPPC_FP128Ty() const
Return true if this is powerpc long double.
Definition Type.h:167
Type * getScalarType() const
If this is a vector type, return the element type, otherwise return 'this'.
Definition Type.h:370
LLVM_ABI TypeSize getPrimitiveSizeInBits() const LLVM_READONLY
Return the basic size of this type if it is a primitive type.
Definition Type.cpp:201
LLVM_ABI Type * getWithNewBitWidth(unsigned NewBitWidth) const
Given an integer or vector type, change the lane bitwidth to NewBitwidth, whilst keeping the old numb...
LLVMContext & getContext() const
Return the LLVMContext in which this type was uniqued.
Definition Type.h:130
LLVM_ABI unsigned getScalarSizeInBits() const LLVM_READONLY
If this is a vector type, return the getPrimitiveSizeInBits value for the element type.
Definition Type.cpp:236
static LLVM_ABI IntegerType * getInt1Ty(LLVMContext &C)
Definition Type.cpp:310
LLVM_ABI int getFPMantissaWidth() const
Return the width of the mantissa of this type.
Definition Type.cpp:241
LLVM_ABI const fltSemantics & getFltSemantics() const
Definition Type.cpp:110
A Use represents the edge between a Value definition and its users.
Definition Use.h:35
void setOperand(unsigned i, Value *Val)
Definition User.h:212
Value * getOperand(unsigned i) const
Definition User.h:207
unsigned getNumOperands() const
Definition User.h:229
LLVM Value Representation.
Definition Value.h:75
Type * getType() const
All values are typed, get the type of this value.
Definition Value.h:255
bool hasOneUse() const
Return true if there is exactly one use of this value.
Definition Value.h:439
LLVMContext & getContext() const
All values hold a context through their type.
Definition Value.h:258
iterator_range< user_iterator > users()
Definition Value.h:426
LLVM_ABI bool hasNUsesOrMore(unsigned N) const
Return true if this value has N uses or more.
Definition Value.cpp:154
LLVM_ABI const Value * stripAndAccumulateConstantOffsets(const DataLayout &DL, APInt &Offset, bool AllowNonInbounds, bool AllowInvariantGroup=false, function_ref< bool(Value &Value, APInt &Offset)> ExternalAnalysis=nullptr, bool LookThroughIntToPtr=false) const
Accumulate the constant offset this value has compared to a base pointer.
LLVM_ABI const Value * stripPointerCasts() const
Strip off pointer casts, all-zero GEPs and address space casts.
Definition Value.cpp:709
iterator_range< use_iterator > uses()
Definition Value.h:380
LLVM_ABI StringRef getName() const
Return a constant reference to the value's name.
Definition Value.cpp:318
LLVM_ABI void takeName(Value *V)
Transfer the name from V to this value.
Definition Value.cpp:399
static LLVM_ABI VectorType * get(Type *ElementType, ElementCount EC)
This static method is the primary way to construct an VectorType.
constexpr ScalarTy getFixedValue() const
Definition TypeSize.h:200
constexpr bool isScalable() const
Returns whether the quantity is scaled by a runtime quantity (vscale).
Definition TypeSize.h:168
const ParentTy * getParent() const
Definition ilist_node.h:34
CallInst * Call
Changed
#define llvm_unreachable(msg)
Marks that the current location is not supposed to be reachable.
LLVM_ABI APInt RoundingUDiv(const APInt &A, const APInt &B, APInt::Rounding RM)
Return A unsign-divided by B, rounded by the given rounding mode.
Definition APInt.cpp:2814
LLVM_ABI APInt RoundingSDiv(const APInt &A, const APInt &B, APInt::Rounding RM)
Return A sign-divided by B, rounded by the given rounding mode.
Definition APInt.cpp:2832
@ C
The default llvm calling convention, compatible with C.
Definition CallingConv.h:34
LLVM_ABI Function * getOrInsertDeclaration(Module *M, ID id, ArrayRef< Type * > OverloadTys={})
Look up the Function declaration of the intrinsic id in the Module M.
SpecificConstantMatch m_ZeroInt()
Convenience matchers for specific integer values.
BinaryOp_match< SpecificConstantMatch, SrcTy, TargetOpcode::G_SUB > m_Neg(const SrcTy &&Src)
Matches a register negated by a G_SUB.
BinaryOp_match< SrcTy, SpecificConstantMatch, TargetOpcode::G_XOR, true > m_Not(const SrcTy &&Src)
Matches a register not-ed by a G_XOR.
OneUse_match< SubPat > m_OneUse(const SubPat &SP)
match_combine_or< Ty... > m_CombineOr(const Ty &...Ps)
Combine pattern matchers matching any of Ps patterns.
match_combine_and< Ty... > m_CombineAnd(const Ty &...Ps)
Combine pattern matchers matching all of Ps patterns.
cst_pred_ty< is_all_ones > m_AllOnes()
Match an integer or vector with all bits set.
cst_pred_ty< is_lowbit_mask > m_LowBitMask()
Match an integer or vector with only the low bit(s) set.
BinaryOp_match< LHS, RHS, Instruction::And > m_And(const LHS &L, const RHS &R)
cst_pred_ty< is_negative > m_Negative()
Match an integer or vector of negative values.
BinaryOp_match< LHS, RHS, Instruction::Add > m_Add(const LHS &L, const RHS &R)
cst_pred_ty< is_sign_mask > m_SignMask()
Match an integer or vector with only the sign bit(s) set.
OverflowingBinaryOp_match< LHS, RHS, Instruction::Add, OverflowingBinaryOperator::NoUnsignedWrap > m_NUWAdd(const LHS &L, const RHS &R)
BinaryOp_match< LHS, RHS, Instruction::AShr > m_AShr(const LHS &L, const RHS &R)
BinaryOp_match< LHS, RHS, Instruction::FSub > m_FSub(const LHS &L, const RHS &R)
cst_pred_ty< is_power2 > m_Power2()
Match an integer or vector power-of-2.
BinaryOp_match< LHS, RHS, Instruction::URem > m_URem(const LHS &L, const RHS &R)
match_combine_or< CastInst_match< OpTy, TruncInst >, OpTy > m_TruncOrSelf(const OpTy &Op)
ap_match< APInt > m_APInt(const APInt *&Res)
Match a ConstantInt or splatted ConstantVector, binding the specified pointer to the contained APInt.
BinaryOp_match< LHS, RHS, Instruction::And, true > m_c_And(const LHS &L, const RHS &R)
Matches an And with LHS and RHS in either order.
CastInst_match< OpTy, TruncInst > m_Trunc(const OpTy &Op)
Matches Trunc.
BinaryOp_match< LHS, RHS, Instruction::Xor > m_Xor(const LHS &L, const RHS &R)
ap_match< APInt > m_APIntAllowPoison(const APInt *&Res)
Match APInt while allowing poison in splat vector constants.
specific_intval< false > m_SpecificInt(const APInt &V)
Match a specific integer value or vector with all elements equal to the value.
match_combine_or< CastInst_match< OpTy, ZExtInst >, OpTy > m_ZExtOrSelf(const OpTy &Op)
bool match(Val *V, const Pattern &P)
BinOpPred_match< LHS, RHS, is_idiv_op > m_IDiv(const LHS &L, const RHS &R)
Matches integer division operations.
match_bind< Instruction > m_Instruction(Instruction *&I)
Match an instruction, capturing it if we match.
match_deferred< Value > m_Deferred(Value *const &V)
Like m_Specific(), but works if the specific value to match is determined as part of the same match()...
cstfp_pred_ty< is_any_zero_fp > m_AnyZeroFP()
Match a floating-point negative zero or positive zero.
specificval_ty m_Specific(const Value *V)
Match if we have a specific specified value.
BinOpPred_match< LHS, RHS, is_right_shift_op > m_Shr(const LHS &L, const RHS &R)
Matches logical shift operations.
specific_intval< true > m_SpecificIntAllowPoison(const APInt &V)
ap_match< APFloat > m_APFloat(const APFloat *&Res)
Match a ConstantFP or splatted ConstantVector, binding the specified pointer to the contained APFloat...
CmpClass_match< LHS, RHS, ICmpInst, true > m_c_ICmp(CmpPredicate &Pred, const LHS &L, const RHS &R)
Matches an ICmp with a predicate over LHS and RHS in either order.
OverflowingBinaryOp_match< LHS, RHS, Instruction::Add, OverflowingBinaryOperator::NoUnsignedWrap, true > m_c_NUWAdd(const LHS &L, const RHS &R)
OverflowingBinaryOp_match< cst_pred_ty< is_zero_int >, ValTy, Instruction::Sub, OverflowingBinaryOperator::NoSignedWrap > m_NSWNeg(const ValTy &V)
Matches a 'Neg' as 'sub nsw 0, V'.
cst_pred_ty< is_nonnegative > m_NonNegative()
Match an integer or vector of non-negative values.
cst_pred_ty< is_one > m_One()
Match an integer 1 or a vector with all elements equal to 1.
IntrinsicID_match m_Intrinsic()
Match intrinsic calls like this: m_Intrinsic<Intrinsic::fabs>(m_Value(X))
ThreeOps_match< Cond, LHS, RHS, Instruction::Select > m_Select(const Cond &C, const LHS &L, const RHS &R)
Matches SelectInst.
auto m_BinOp()
Match an arbitrary binary operation and ignore it.
ExtractValue_match< Ind, Val_t > m_ExtractValue(const Val_t &V)
Match a single index ExtractValue instruction.
BinOpPred_match< LHS, RHS, is_logical_shift_op > m_LogicalShift(const LHS &L, const RHS &R)
Matches logical shift operations.
MaxMin_match< ICmpInst, LHS, RHS, smin_pred_ty > m_SMin(const LHS &L, const RHS &R)
match_combine_or< CastInst_match< OpTy, UIToFPInst >, CastInst_match< OpTy, SIToFPInst > > m_IToFP(const OpTy &Op)
auto m_Value()
Match an arbitrary value and ignore it.
m_Intrinsic_Ty< Opnd0 >::Ty m_Sqrt(const Opnd0 &Op0)
BinaryOp_match< LHS, RHS, Instruction::Xor, true > m_c_Xor(const LHS &L, const RHS &R)
Matches an Xor with LHS and RHS in either order.
BinaryOp_match< LHS, RHS, Instruction::FAdd > m_FAdd(const LHS &L, const RHS &R)
BinaryOp_match< LHS, RHS, Instruction::Mul > m_Mul(const LHS &L, const RHS &R)
auto m_Constant()
Match an arbitrary Constant and ignore it.
NoWrapTrunc_match< OpTy, TruncInst::NoSignedWrap > m_NSWTrunc(const OpTy &Op)
Matches trunc nsw.
TwoOps_match< V1_t, V2_t, Instruction::ShuffleVector > m_Shuffle(const V1_t &v1, const V2_t &v2)
Matches ShuffleVectorInst independently of mask value.
ThreeOps_match< decltype(m_Value()), LHS, RHS, Instruction::Select, true > m_c_Select(const LHS &L, const RHS &R)
Match Select(C, LHS, RHS) or Select(C, RHS, LHS)
CastInst_match< OpTy, FPExtInst > m_FPExt(const OpTy &Op)
OverflowingBinaryOp_match< LHS, RHS, Instruction::Shl, OverflowingBinaryOperator::NoSignedWrap > m_NSWShl(const LHS &L, const RHS &R)
CastInst_match< OpTy, ZExtInst > m_ZExt(const OpTy &Op)
Matches ZExt.
OverflowingBinaryOp_match< LHS, RHS, Instruction::Shl, OverflowingBinaryOperator::NoUnsignedWrap > m_NUWShl(const LHS &L, const RHS &R)
OverflowingBinaryOp_match< LHS, RHS, Instruction::Mul, OverflowingBinaryOperator::NoUnsignedWrap > m_NUWMul(const LHS &L, const RHS &R)
BinaryOp_match< LHS, RHS, Instruction::UDiv > m_UDiv(const LHS &L, const RHS &R)
MaxMin_match< ICmpInst, LHS, RHS, umax_pred_ty > m_UMax(const LHS &L, const RHS &R)
match_immconstant_ty m_ImmConstant()
Match an arbitrary immediate Constant and ignore it.
cst_pred_ty< is_negated_power2_or_zero > m_NegatedPower2OrZero()
Match a integer or vector negated power-of-2.
NoWrapTrunc_match< OpTy, TruncInst::NoUnsignedWrap > m_NUWTrunc(const OpTy &Op)
Matches trunc nuw.
cst_pred_ty< custom_checkfn< APInt > > m_CheckedInt(function_ref< bool(const APInt &)> CheckFn)
Match an integer or vector where CheckFn(ele) for each element is true.
SelectLike_match< CondTy, LTy, RTy > m_SelectLike(const CondTy &C, const LTy &TrueC, const RTy &FalseC)
Matches a value that behaves like a boolean-controlled select, i.e.
cst_pred_ty< is_lowbit_mask_or_zero > m_LowBitMaskOrZero()
Match an integer or vector with only the low bit(s) set.
auto m_MaxOrMin(const LHS &L, const RHS &R)
BinaryOp_match< LHS, RHS, Instruction::Add, true > m_c_Add(const LHS &L, const RHS &R)
Matches a Add with LHS and RHS in either order.
match_combine_or< BinaryOp_match< LHS, RHS, Instruction::Add >, DisjointOr_match< LHS, RHS > > m_AddLike(const LHS &L, const RHS &R)
Match either "add" or "or disjoint".
CastInst_match< OpTy, UIToFPInst > m_UIToFP(const OpTy &Op)
CastOperator_match< OpTy, Instruction::BitCast > m_BitCast(const OpTy &Op)
Matches BitCast.
cstfp_pred_ty< is_finitenonzero > m_FiniteNonZero()
Match a finite non-zero FP constant.
BinaryOp_match< LHS, RHS, Instruction::SDiv > m_SDiv(const LHS &L, const RHS &R)
MaxMin_match< ICmpInst, LHS, RHS, smax_pred_ty > m_SMax(const LHS &L, const RHS &R)
Signum_match< Val_t > m_Signum(const Val_t &V)
Matches a signum pattern.
CastInst_match< OpTy, SIToFPInst > m_SIToFP(const OpTy &Op)
BinaryOp_match< LHS, RHS, Instruction::LShr > m_LShr(const LHS &L, const RHS &R)
CmpClass_match< LHS, RHS, ICmpInst > m_ICmp(CmpPredicate &Pred, const LHS &L, const RHS &R)
match_combine_or< CastInst_match< OpTy, ZExtInst >, CastInst_match< OpTy, SExtInst > > m_ZExtOrSExt(const OpTy &Op)
FNeg_match< OpTy > m_FNeg(const OpTy &X)
Match 'fneg X' as 'fsub -0.0, X'.
cstfp_pred_ty< is_pos_zero_fp > m_PosZeroFP()
Match a floating-point positive zero.
BinaryOp_match< LHS, RHS, Instruction::Shl > m_Shl(const LHS &L, const RHS &R)
UAddWithOverflow_match< LHS_t, RHS_t, Sum_t > m_UAddWithOverflow(const LHS_t &L, const RHS_t &R, const Sum_t &S)
Match an icmp instruction checking for unsigned overflow on addition.
m_Intrinsic_Ty< Opnd0 >::Ty m_VecReverse(const Opnd0 &Op0)
BinOpPred_match< LHS, RHS, is_irem_op > m_IRem(const LHS &L, const RHS &R)
Matches integer remainder operations.
CastInst_match< OpTy, FPTruncInst > m_FPTrunc(const OpTy &Op)
auto m_Undef()
Match an arbitrary undef constant.
BinaryOp_match< LHS, RHS, Instruction::Or > m_Or(const LHS &L, const RHS &R)
CastInst_match< OpTy, SExtInst > m_SExt(const OpTy &Op)
Matches SExt.
is_zero m_Zero()
Match any null constant or a vector with all elements equal to 0.
BinaryOp_match< LHS, RHS, Instruction::Or, true > m_c_Or(const LHS &L, const RHS &R)
Matches an Or with LHS and RHS in either order.
ElementWiseBitCast_match< OpTy > m_ElementWiseBitCast(const OpTy &Op)
m_Intrinsic_Ty< Opnd0 >::Ty m_FAbs(const Opnd0 &Op0)
BinaryOp_match< LHS, RHS, Instruction::Mul, true > m_c_Mul(const LHS &L, const RHS &R)
Matches a Mul with LHS and RHS in either order.
CastOperator_match< OpTy, Instruction::PtrToInt > m_PtrToInt(const OpTy &Op)
Matches PtrToInt.
BinaryOp_match< LHS, RHS, Instruction::Sub > m_Sub(const LHS &L, const RHS &R)
match_unless< Ty > m_Unless(const Ty &M)
Match if the inner matcher does NOT match.
cst_pred_ty< icmp_pred_with_threshold > m_SpecificInt_ICMP(ICmpInst::Predicate Predicate, const APInt &Threshold)
Match an integer or vector with every element comparing 'pred' (eg/ne/...) to Threshold.
auto m_ConstantInt()
Match an arbitrary ConstantInt and ignore it.
This is an optimization pass for GlobalISel generic memory operations.
@ Offset
Definition DWP.cpp:532
detail::zippy< detail::zip_shortest, T, U, Args... > zip(T &&t, U &&u, Args &&...args)
zip iterator for two or more iteratable types.
Definition STLExtras.h:831
@ NeverOverflows
Never overflows.
@ AlwaysOverflowsHigh
Always overflows in the direction of signed/unsigned max value.
@ AlwaysOverflowsLow
Always overflows in the direction of signed/unsigned min value.
@ MayOverflow
May or may not overflow.
cl::opt< bool > ProfcheckDisableMetadataFixes
Definition LoopInfo.cpp:60
bool all_of(R &&range, UnaryPredicate P)
Provide wrappers to std::all_of which take ranges instead of having to pass begin/end explicitly.
Definition STLExtras.h:1739
LLVM_ABI bool isKnownNeverInfinity(const Value *V, const SimplifyQuery &SQ, unsigned Depth=0)
Return true if the floating-point scalar value is not an infinity or if the floating-point vector val...
LLVM_ABI bool isSignBitCheck(ICmpInst::Predicate Pred, const APInt &RHS, bool &TrueIfSigned)
Given an exploded icmp instruction, return true if the comparison only checks the sign bit.
decltype(auto) dyn_cast(const From &Val)
dyn_cast<X> - Return the argument parameter cast to the specified type.
Definition Casting.h:643
LLVM_ABI Value * stripNullTest(Value *V)
Returns the inner value X if the expression has the form f(X) where f(X) == 0 if and only if X == 0,...
LLVM_ABI Constant * ConstantFoldCompareInstOperands(unsigned Predicate, Constant *LHS, Constant *RHS, const DataLayout &DL, const TargetLibraryInfo *TLI=nullptr, const Instruction *I=nullptr)
Attempt to constant fold a compare instruction (icmp/fcmp) with the specified operands.
iterator_range< early_inc_iterator_impl< detail::IterOfRange< RangeT > > > make_early_inc_range(RangeT &&Range)
Make a range that does early increment to allow mutation of the underlying range without disrupting i...
Definition STLExtras.h:634
LLVM_ABI Value * simplifyFCmpInst(CmpPredicate Predicate, Value *LHS, Value *RHS, FastMathFlags FMF, const SimplifyQuery &Q)
Given operands for an FCmpInst, fold the result or return null.
int ilogb(const APFloat &Arg)
Returns the exponent of the internal representation of the APFloat.
Definition APFloat.h:1601
LLVM_ABI bool MaskedValueIsZero(const Value *V, const APInt &Mask, const SimplifyQuery &SQ, unsigned Depth=0)
Return true if 'V & Mask' is known to be zero.
LLVM_ABI Value * simplifyAddInst(Value *LHS, Value *RHS, bool IsNSW, bool IsNUW, const SimplifyQuery &Q)
Given operands for an Add, fold the result or return null.
LLVM_ABI Constant * ConstantFoldConstant(const Constant *C, const DataLayout &DL, const TargetLibraryInfo *TLI=nullptr)
ConstantFoldConstant - Fold the constant using the specified DataLayout.
auto dyn_cast_or_null(const Y &Val)
Definition Casting.h:753
LLVM_ABI bool isSplatValue(const Value *V, int Index=-1, unsigned Depth=0)
Return true if each element of the vector value V is poisoned or equal to every other non-poisoned el...
unsigned Log2_32(uint32_t Value)
Return the floor log base 2 of the specified value, -1 if the value is zero.
Definition MathExtras.h:331
int countl_zero(T Val)
Count number of 0's from the most significant bit to the least stopping at the first 1.
Definition bit.h:263
LLVM_ABI Value * emitGEPOffset(IRBuilderBase *Builder, const DataLayout &DL, User *GEP, bool NoAssumptions=false)
Given a getelementptr instruction/constantexpr, emit the code necessary to compute the offset from th...
Definition Local.cpp:22
constexpr unsigned MaxAnalysisRecursionDepth
LLVM_ABI Constant * ConstantFoldUnaryOpOperand(unsigned Opcode, Constant *Op, const DataLayout &DL)
Attempt to constant fold a unary operation with the specified operand.
LLVM_ABI bool isKnownNegative(const Value *V, const SimplifyQuery &SQ, unsigned Depth=0)
Returns true if the given value is known be negative (i.e.
SelectPatternFlavor
Specific patterns of select instructions we can match.
@ SPF_FMAXNUM
Floating point minnum.
@ SPF_UNKNOWN
@ SPF_FMINNUM
Unsigned maximum.
LLVM_ABI bool impliesPoison(const Value *ValAssumedPoison, const Value *V)
Return true if V is poison given that ValAssumedPoison is already poison.
LLVM_ABI LinearExpression decomposeLinearExpression(const DataLayout &DL, Value *Ptr)
Decompose a pointer into a linear expression.
Definition Loads.cpp:907
LLVM_ABI bool isFinite(const Loop *L)
Return true if this loop can be assumed to run for a finite number of iterations.
FPClassTest
Floating-point class tests, supported by 'is_fpclass' intrinsic.
APFloat scalbn(APFloat X, int Exp, APFloat::roundingMode RM)
Returns: X * 2^Exp for integral exponents.
Definition APFloat.h:1610
LLVM_ABI void computeKnownBits(const Value *V, KnownBits &Known, const DataLayout &DL, AssumptionCache *AC=nullptr, const Instruction *CxtI=nullptr, const DominatorTree *DT=nullptr, bool UseInstrInfo=true, unsigned Depth=0)
Determine which bits of V are known to be either zero or one and return them in the KnownZero/KnownOn...
LLVM_ABI SelectPatternResult matchSelectPattern(Value *V, Value *&LHS, Value *&RHS, Instruction::CastOps *CastOp=nullptr, unsigned Depth=0)
Pattern match integer [SU]MIN, [SU]MAX and ABS idioms, returning the kind and providing the out param...
LLVM_ABI bool NullPointerIsDefined(const Function *F, unsigned AS=0)
Check whether null pointer dereferencing is considered undefined behavior for a given function or an ...
bool none_of(R &&Range, UnaryPredicate P)
Provide wrappers to std::none_of which take ranges instead of having to pass begin/end explicitly.
Definition STLExtras.h:1753
LLVM_ABI Value * simplifyICmpInst(CmpPredicate Pred, Value *LHS, Value *RHS, const SimplifyQuery &Q)
Given operands for an ICmpInst, fold the result or return null.
LLVM_ABI Constant * ConstantFoldCastOperand(unsigned Opcode, Constant *C, Type *DestTy, const DataLayout &DL)
Attempt to constant fold a cast with the specified operand.
LLVM_ABI Constant * ConstantFoldLoadFromConst(Constant *C, Type *Ty, const APInt &Offset, const DataLayout &DL)
Extract value of C at the given Offset reinterpreted as Ty.
bool isa(const From &Val)
isa<X> - Return true if the parameter to the template is an instance of one of the template type argu...
Definition Casting.h:547
LLVM_ABI Constant * ConstantFoldBinaryOpOperands(unsigned Opcode, Constant *LHS, Constant *RHS, const DataLayout &DL)
Attempt to constant fold a binary operation with the specified operands.
LLVM_ABI bool isKnownNonZero(const Value *V, const SimplifyQuery &Q, unsigned Depth=0)
Return true if the given value is known to be non-zero when defined.
constexpr T divideCeil(U Numerator, V Denominator)
Returns the integer ceil(Numerator / Denominator).
Definition MathExtras.h:394
@ Other
Any other memory.
Definition ModRef.h:68
@ First
Helpers to iterate all locations in the MemoryEffectsBase class.
Definition ModRef.h:74
LLVM_ABI Value * simplifyBinOp(unsigned Opcode, Value *LHS, Value *RHS, const SimplifyQuery &Q)
Given operands for a BinaryOperator, fold the result or return null.
@ UMin
Unsigned integer min implemented in terms of select(cmp()).
@ Mul
Product of integers.
@ Xor
Bitwise or logical XOR of integers.
@ SMax
Signed integer max implemented in terms of select(cmp()).
@ SMin
Signed integer min implemented in terms of select(cmp()).
@ Sub
Subtraction of integers.
@ Add
Sum of integers.
@ UMax
Unsigned integer max implemented in terms of select(cmp()).
LLVM_ABI bool isKnownNonEqual(const Value *V1, const Value *V2, const SimplifyQuery &SQ, unsigned Depth=0)
Return true if the given values are known to be non-equal when defined.
DWARFExpression::Operation Op
LLVM_ABI bool PointerMayBeCaptured(const Value *V, bool ReturnCaptures, unsigned MaxUsesToExplore=0)
PointerMayBeCaptured - Return true if this pointer value may be captured by the enclosing function (w...
constexpr unsigned BitWidth
LLVM_ABI Constant * getLosslessInvCast(Constant *C, Type *InvCastTo, unsigned CastOp, const DataLayout &DL, PreservedCastFlags *Flags=nullptr)
Try to cast C to InvC losslessly, satisfying CastOp(InvC) equals C, or CastOp(InvC) is a refined valu...
auto count_if(R &&Range, UnaryPredicate P)
Wrapper function around std::count_if to count the number of times an element satisfying a given pred...
Definition STLExtras.h:2019
decltype(auto) cast(const From &Val)
cast<X> - Return the argument parameter cast to the specified type.
Definition Casting.h:559
LLVM_ABI bool isKnownNeverNaN(const Value *V, const SimplifyQuery &SQ, unsigned Depth=0)
Return true if the floating-point scalar value is not a NaN or if the floating-point vector value has...
LLVM_ABI std::optional< std::pair< CmpPredicate, Constant * > > getFlippedStrictnessPredicateAndConstant(CmpPredicate Pred, Constant *C)
Convert an integer comparison with a constant RHS into an equivalent form with the strictness flipped...
bool all_equal(std::initializer_list< T > Values)
Returns true if all Values in the initializer lists are equal or the list.
Definition STLExtras.h:2166
LLVM_ABI bool isKnownToBeAPowerOfTwo(const Value *V, const DataLayout &DL, bool OrZero=false, AssumptionCache *AC=nullptr, const Instruction *CxtI=nullptr, const DominatorTree *DT=nullptr, bool UseInstrInfo=true, unsigned Depth=0)
Return true if the given value is known to have exactly one bit set when defined.
@ Continue
Definition DWP.h:22
LLVM_ABI const Value * getUnderlyingObject(const Value *V, unsigned MaxLookup=MaxLookupSearchDepth)
This method strips off any GEP address adjustments, pointer casts or llvm.threadlocal....
LLVM_ABI bool isKnownPositive(const Value *V, const SimplifyQuery &SQ, unsigned Depth=0)
Returns true if the given value is known be positive (i.e.
LLVM_ABI bool isKnownNonNegative(const Value *V, const SimplifyQuery &SQ, unsigned Depth=0)
Returns true if the give value is known to be non-negative.
constexpr detail::IsaCheckPredicate< Types... > IsaPred
Function object wrapper for the llvm::isa type check.
Definition Casting.h:866
LLVM_ABI std::optional< bool > isImpliedCondition(const Value *LHS, const Value *RHS, const DataLayout &DL, bool LHSIsTrue=true, unsigned Depth=0)
Return true if RHS is known to be implied true by LHS.
std::optional< DecomposedBitTest > decomposeBitTestICmp(Value *LHS, Value *RHS, CmpInst::Predicate Pred, bool LookThroughTrunc=true, bool AllowNonZeroC=false, bool DecomposeAnd=false)
Decompose an icmp into the form ((X & Mask) pred C) if possible.
LLVM_ABI ConstantRange computeConstantRange(const Value *V, bool ForSigned, const SimplifyQuery &SQ, unsigned Depth=0)
Determine the possible constant range of an integer or vector of integer value.
void swap(llvm::BitVector &LHS, llvm::BitVector &RHS)
Implement std::swap in terms of BitVector swap.
Definition BitVector.h:872
#define NC
Definition regutils.h:42
Value * materialize(InstCombiner::BuilderTy &Builder) const
static OffsetResult select(Value *Cond, Value *TrueV, Value *FalseV, Instruction *MDFrom)
static OffsetResult value(Value *V)
static OffsetResult invalid()
This callback is used in conjunction with PointerMayBeCaptured.
static CommonPointerBase compute(Value *LHS, Value *RHS)
Represent subnormal handling kind for floating point instruction inputs and outputs.
@ PreserveSign
The sign of a flushed-to-zero number is preserved in the sign of 0.
@ PositiveZero
Denormals are flushed to positive zero.
static constexpr DenormalMode getIEEE()
bool isNonNegative() const
Returns true if this value is known to be non-negative.
Definition KnownBits.h:106
bool isZero() const
Returns true if value is all zero.
Definition KnownBits.h:78
unsigned countMinTrailingZeros() const
Returns the minimum number of trailing zero bits.
Definition KnownBits.h:256
unsigned countMaxTrailingZeros() const
Returns the maximum number of trailing zero bits possible.
Definition KnownBits.h:288
APInt getSignedMaxValue() const
Return the maximal signed value possible given these KnownBits.
Definition KnownBits.h:152
unsigned countMaxPopulation() const
Returns the maximum number of bits that could be one.
Definition KnownBits.h:303
unsigned getBitWidth() const
Get the bit width of this value.
Definition KnownBits.h:44
bool isConstant() const
Returns true if we know the value of all bits.
Definition KnownBits.h:54
unsigned countMinLeadingZeros() const
Returns the minimum number of leading zero bits.
Definition KnownBits.h:262
APInt getMaxValue() const
Return the maximal unsigned value possible given these KnownBits.
Definition KnownBits.h:146
APInt getMinValue() const
Return the minimal unsigned value possible given these KnownBits.
Definition KnownBits.h:130
bool isStrictlyPositive() const
Returns true if this value is known to be positive.
Definition KnownBits.h:112
bool isNegative() const
Returns true if this value is known to be negative.
Definition KnownBits.h:103
unsigned countMinPopulation() const
Returns the number of bits known to be one.
Definition KnownBits.h:300
APInt getSignedMinValue() const
Return the minimal signed value possible given these KnownBits.
Definition KnownBits.h:136
const APInt & getConstant() const
Returns the value when all bits have a known value.
Definition KnownBits.h:58
Linear expression BasePtr + Index * Scale + Offset.
Definition Loads.h:211
GEPNoWrapFlags Flags
Definition Loads.h:216
Matching combinators.
SelectPatternFlavor Flavor
static bool isMinOrMax(SelectPatternFlavor SPF)
When implementing this min/max pattern as fcmp; select, does the fcmp have to be ordered?
const DataLayout & DL
const Instruction * CxtI
const DominatorTree * DT
SimplifyQuery getWithInstruction(const Instruction *I) const
AssumptionCache * AC
A MapVector that performs no allocations if smaller than a certain size.
Definition MapVector.h:276
Capture information for a specific Use.