18#ifndef LLVM_TRANSFORMS_INSTCOMBINE_INSTCOMBINER_H
19#define LLVM_TRANSFORMS_INSTCOMBINE_INSTCOMBINER_H
31#define DEBUG_TYPE "instcombine"
38class OptimizationRemarkEmitter;
39class ProfileSummaryInfo;
40class TargetLibraryInfo;
41class TargetTransformInfo;
87 bool MadeIRChange =
false;
102 :
TTI(
TTI), Builder(Builder), Worklist(Worklist),
103 MinimizeSize(MinimizeSize), AA(AA), AC(AC), TLI(TLI), DT(DT),
DL(
DL),
104 SQ(
DL, &TLI, &DT, &AC, nullptr,
true,
106 ORE(ORE), BFI(BFI), BPI(BPI), PSI(PSI), LI(LI) {}
114 if (
auto *BitCast = dyn_cast<BitCastInst>(V))
115 if (!OneUseOnly || BitCast->hasOneUse())
116 return BitCast->getOperand(0);
138 if (isa<Constant>(V))
139 return isa<UndefValue>(V) ? 0 : 1;
141 if (isa<CastInst>(V) ||
match(V, m_Neg(PatternMatch::m_Value())) ||
142 match(V, m_Not(PatternMatch::m_Value())) ||
156 case CmpInst::ICMP_NE:
157 case CmpInst::ICMP_ULE:
158 case CmpInst::ICMP_SLE:
159 case CmpInst::ICMP_UGE:
160 case CmpInst::ICMP_SGE:
162 case CmpInst::FCMP_ONE:
163 case CmpInst::FCMP_OLE:
164 case CmpInst::FCMP_OGE:
173 return ConstantExpr::getAdd(
C, ConstantInt::get(
C->getType(), 1));
178 return ConstantExpr::getSub(
C, ConstantInt::get(
C->getType(), 1));
181 std::optional<std::pair<
193 return match(&
SI, PatternMatch::m_LogicalAnd(PatternMatch::m_Value(),
194 PatternMatch::m_Value())) ||
195 match(&
SI, PatternMatch::m_LogicalOr(PatternMatch::m_Value(),
196 PatternMatch::m_Value()));
206 Value *getFreelyInvertedImpl(
Value *V,
bool WillInvertAllUses,
213 return getFreelyInvertedImpl(V, WillInvertAllUses, Builder, DoesConsume,
220 return getFreelyInverted(V, WillInvertAllUses, Builder, Unused);
231 return getFreelyInverted(V, WillInvertAllUses,
nullptr,
232 DoesConsume) !=
nullptr;
237 return isFreeToInvert(V, WillInvertAllUses, Unused);
247 for (
Use &U : V->uses()) {
248 if (U.getUser() == IgnoredUser)
251 auto *
I = cast<Instruction>(U.getUser());
252 switch (
I->getOpcode()) {
253 case Instruction::Select:
254 if (U.getOperandNo() != 0)
256 if (shouldAvoidAbsorbingNotIntoSelect(*cast<SelectInst>(
I)))
259 case Instruction::Br:
260 assert(U.getOperandNo() == 0 &&
"Must be branching on that value.");
262 case Instruction::Xor:
264 if (!
match(
I, m_Not(PatternMatch::m_Value())))
282 bool IsRHSConstant) {
283 auto *InVTy = cast<FixedVectorType>(In->getType());
285 Type *EltTy = InVTy->getElementType();
286 auto *SafeC = ConstantExpr::getBinOpIdentity(Opcode, EltTy, IsRHSConstant);
292 case Instruction::SRem:
293 case Instruction::URem:
294 SafeC = ConstantInt::get(EltTy, 1);
296 case Instruction::FRem:
297 SafeC = ConstantFP::get(EltTy, 1.0);
301 "Only rem opcodes have no identity constant for RHS");
305 case Instruction::Shl:
306 case Instruction::LShr:
307 case Instruction::AShr:
308 case Instruction::SDiv:
309 case Instruction::UDiv:
310 case Instruction::SRem:
311 case Instruction::URem:
312 case Instruction::Sub:
313 case Instruction::FSub:
314 case Instruction::FDiv:
315 case Instruction::FRem:
316 SafeC = Constant::getNullValue(EltTy);
323 assert(SafeC &&
"Must have safe constant for binop");
324 unsigned NumElts = InVTy->getNumElements();
326 for (
unsigned i = 0; i != NumElts; ++i) {
327 Constant *
C = In->getAggregateElement(i);
328 Out[i] = isa<UndefValue>(
C) ? SafeC :
C;
330 return ConstantVector::get(Out);
348 std::optional<Instruction *> targetInstCombineIntrinsic(
IntrinsicInst &
II);
349 std::optional<Value *>
352 bool &KnownBitsComputed);
353 std::optional<Value *> targetSimplifyDemandedVectorEltsIntrinsic(
364 assert(New && !New->getParent() &&
365 "New instruction already inserted into a basic block!");
366 New->insertBefore(Old);
373 New->setDebugLoc(Old->getDebugLoc());
374 return InsertNewInstBefore(New, Old);
386 if (
I.use_empty())
return nullptr;
393 V = PoisonValue::get(
I.getType());
396 <<
" with " << *V <<
'\n');
399 if (V->use_empty() && isa<Instruction>(V) && !V->hasName() &&
I.hasName())
402 I.replaceAllUsesWith(V);
408 Value *OldOp =
I.getOperand(OpNum);
409 I.setOperand(OpNum, V);
462 bool IsNSW =
false)
const {
508 return SimplifyDemandedBits(
I, OpNo, DemandedMask, Known,
515 bool AllowMultipleUsers =
false) = 0;
517 bool isValidAddrSpaceCast(
unsigned FromAS,
unsigned ToAS)
const;
MachineBasicBlock MachineBasicBlock::iterator DebugLoc DL
IRBuilder< TargetFolder > BuilderTy
static GCRegistry::Add< ShadowStackGC > C("shadow-stack", "Very portable GC for uncooperative code generators")
#define LLVM_LIBRARY_VISIBILITY
uint64_t IntrinsicInst * II
StandardInstrumentations SI(Mod->getContext(), Debug, VerifyEach)
assert(ImpDefSCC.getReg()==AMDGPU::SCC &&ImpDefSCC.isDef())
Class for arbitrary precision integers.
A cache of @llvm.assume calls within a function.
InstListType::iterator iterator
Instruction iterators...
BlockFrequencyInfo pass uses BlockFrequencyInfoImpl implementation to estimate IR basic block frequen...
Analysis providing branch probability information.
This class is the base class for the comparison instructions.
Predicate
This enumeration lists the possible predicates for CmpInst subclasses.
This is an important base class in LLVM.
This class represents an Operation in the Expression.
A parsed version of the target data layout string in and methods for querying it.
Concrete subclass of DominatorTreeBase that is used to compute a normal dominator tree.
The core instruction combiner logic.
OverflowResult computeOverflowForSignedSub(const Value *LHS, const Value *RHS, const Instruction *CxtI) const
const DataLayout & getDataLayout() const
static bool isCanonicalPredicate(CmpInst::Predicate Pred)
Predicate canonicalization reduces the number of patterns that need to be matched by other transforms...
bool isFreeToInvert(Value *V, bool WillInvertAllUses)
virtual Instruction * eraseInstFromFunction(Instruction &I)=0
Combiner aware instruction erasure.
bool isFreeToInvert(Value *V, bool WillInvertAllUses, bool &DoesConsume)
Return true if the specified value is free to invert (apply ~ to).
DominatorTree & getDominatorTree() const
OverflowResult computeOverflowForUnsignedMul(const Value *LHS, const Value *RHS, const Instruction *CxtI, bool IsNSW=false) const
virtual ~InstCombiner()=default
LoopInfo * getLoopInfo() const
InstCombiner(InstructionWorklist &Worklist, BuilderTy &Builder, bool MinimizeSize, AAResults *AA, AssumptionCache &AC, TargetLibraryInfo &TLI, TargetTransformInfo &TTI, DominatorTree &DT, OptimizationRemarkEmitter &ORE, BlockFrequencyInfo *BFI, BranchProbabilityInfo *BPI, ProfileSummaryInfo *PSI, const DataLayout &DL, LoopInfo *LI)
static unsigned getComplexity(Value *V)
Assign a complexity or rank value to LLVM Values.
SmallDenseMap< BasicBlock *, SmallVector< BasicBlock * >, 8 > PredOrder
Order of predecessors to canonicalize phi nodes towards.
TargetLibraryInfo & getTargetLibraryInfo() const
BlockFrequencyInfo * getBlockFrequencyInfo() const
bool isKnownToBeAPowerOfTwo(const Value *V, bool OrZero=false, unsigned Depth=0, const Instruction *CxtI=nullptr)
Instruction * InsertNewInstBefore(Instruction *New, BasicBlock::iterator Old)
Inserts an instruction New before instruction Old.
Instruction * replaceInstUsesWith(Instruction &I, Value *V)
A combiner-aware RAUW-like routine.
static bool shouldAvoidAbsorbingNotIntoSelect(const SelectInst &SI)
OverflowResult computeOverflowForSignedAdd(const WithCache< const Value * > &LHS, const WithCache< const Value * > &RHS, const Instruction *CxtI) const
static Constant * SubOne(Constant *C)
Subtract one from a Constant.
KnownBits computeKnownBits(const Value *V, unsigned Depth, const Instruction *CxtI) const
void replaceUse(Use &U, Value *NewValue)
Replace use and add the previously used value to the worklist.
OverflowResult computeOverflowForUnsignedSub(const Value *LHS, const Value *RHS, const Instruction *CxtI) const
InstructionWorklist & Worklist
A worklist of the instructions that need to be simplified.
Instruction * InsertNewInstWith(Instruction *New, BasicBlock::iterator Old)
Same as InsertNewInstBefore, but also sets the debug loc.
BranchProbabilityInfo * BPI
bool SimplifyDemandedBits(Instruction *I, unsigned OpNo, const APInt &DemandedMask, KnownBits &Known)
unsigned ComputeNumSignBits(const Value *Op, unsigned Depth=0, const Instruction *CxtI=nullptr) const
virtual bool SimplifyDemandedBits(Instruction *I, unsigned OpNo, const APInt &DemandedMask, KnownBits &Known, unsigned Depth, const SimplifyQuery &Q)=0
virtual Value * SimplifyDemandedVectorElts(Value *V, APInt DemandedElts, APInt &UndefElts, unsigned Depth=0, bool AllowMultipleUsers=false)=0
static Value * peekThroughBitcast(Value *V, bool OneUseOnly=false)
Return the source operand of a potentially bitcasted value while optionally checking if it has one us...
bool canFreelyInvertAllUsersOf(Instruction *V, Value *IgnoredUser)
Given i1 V, can every user of V be freely adapted if V is changed to !V ? InstCombine's freelyInvertA...
Value * getFreelyInverted(Value *V, bool WillInvertAllUses, BuilderTy *Builder)
void addToWorklist(Instruction *I)
Instruction * replaceOperand(Instruction &I, unsigned OpNum, Value *V)
Replace operand of instruction and add old operand to the worklist.
static Constant * getSafeVectorConstantForBinop(BinaryOperator::BinaryOps Opcode, Constant *In, bool IsRHSConstant)
Some binary operators require special handling to avoid poison and undefined behavior.
OverflowResult computeOverflowForSignedMul(const Value *LHS, const Value *RHS, const Instruction *CxtI) const
ProfileSummaryInfo * getProfileSummaryInfo() const
OptimizationRemarkEmitter & getOptimizationRemarkEmitter() const
SmallDenseSet< std::pair< BasicBlock *, BasicBlock * >, 8 > DeadEdges
Edges that are known to never be taken.
void computeKnownBits(const Value *V, KnownBits &Known, unsigned Depth, const Instruction *CxtI) const
AssumptionCache & getAssumptionCache() const
bool MaskedValueIsZero(const Value *V, const APInt &Mask, unsigned Depth=0, const Instruction *CxtI=nullptr) const
OptimizationRemarkEmitter & ORE
OverflowResult computeOverflowForUnsignedAdd(const WithCache< const Value * > &LHS, const WithCache< const Value * > &RHS, const Instruction *CxtI) const
Value * getFreelyInverted(Value *V, bool WillInvertAllUses, BuilderTy *Builder, bool &DoesConsume)
const SimplifyQuery & getSimplifyQuery() const
static Constant * AddOne(Constant *C)
Add one to a Constant.
unsigned ComputeMaxSignificantBits(const Value *Op, unsigned Depth=0, const Instruction *CxtI=nullptr) const
InstructionWorklist - This is the worklist management logic for InstCombine and other simplification ...
void pushUsersToWorkList(Instruction &I)
When an instruction is simplified, add all users of the instruction to the work lists because they mi...
void add(Instruction *I)
Add instruction to the worklist.
void push(Instruction *I)
Push the instruction onto the worklist stack.
void handleUseCountDecrement(Value *V)
Should be called after decrementing the use-count on V.
A wrapper class for inspecting calls to intrinsic functions.
Analysis providing profile information.
This class represents the LLVM 'select' instruction.
Implements a dense probed hash-table based set with some number of buckets stored inline.
This is a 'vector' (really, a variable-sized array), optimized for the case when the array is small.
Provides information about what library functions are available for the current target.
The instances of the Type class are immutable: once they are created, they are never changed.
A Use represents the edge between a Value definition and its users.
LLVM Value Representation.
#define llvm_unreachable(msg)
Marks that the current location is not supposed to be reachable.
bool match(Val *V, const Pattern &P)
FNeg_match< OpTy > m_FNeg(const OpTy &X)
Match 'fneg X' as 'fsub -0.0, X'.
This is an optimization pass for GlobalISel generic memory operations.
bool MaskedValueIsZero(const Value *V, const APInt &Mask, const SimplifyQuery &DL, unsigned Depth=0)
Return true if 'V & Mask' is known to be zero.
bool isKnownToBeAPowerOfTwo(const Value *V, const DataLayout &DL, bool OrZero=false, unsigned Depth=0, AssumptionCache *AC=nullptr, const Instruction *CxtI=nullptr, const DominatorTree *DT=nullptr, bool UseInstrInfo=true)
Return true if the given value is known to have exactly one bit set when defined.
OverflowResult computeOverflowForUnsignedMul(const Value *LHS, const Value *RHS, const SimplifyQuery &SQ, bool IsNSW=false)
OverflowResult computeOverflowForSignedSub(const Value *LHS, const Value *RHS, const SimplifyQuery &SQ)
raw_ostream & dbgs()
dbgs() - This returns a reference to a raw_ostream for debugging messages.
OverflowResult computeOverflowForSignedMul(const Value *LHS, const Value *RHS, const SimplifyQuery &SQ)
OverflowResult computeOverflowForSignedAdd(const WithCache< const Value * > &LHS, const WithCache< const Value * > &RHS, const SimplifyQuery &SQ)
void computeKnownBits(const Value *V, KnownBits &Known, const DataLayout &DL, unsigned Depth=0, AssumptionCache *AC=nullptr, const Instruction *CxtI=nullptr, const DominatorTree *DT=nullptr, bool UseInstrInfo=true)
Determine which bits of V are known to be either zero or one and return them in the KnownZero/KnownOn...
OverflowResult computeOverflowForUnsignedSub(const Value *LHS, const Value *RHS, const SimplifyQuery &SQ)
unsigned ComputeNumSignBits(const Value *Op, const DataLayout &DL, unsigned Depth=0, AssumptionCache *AC=nullptr, const Instruction *CxtI=nullptr, const DominatorTree *DT=nullptr, bool UseInstrInfo=true)
Return the number of times the sign bit of the register is replicated into the other bits.
OverflowResult computeOverflowForUnsignedAdd(const WithCache< const Value * > &LHS, const WithCache< const Value * > &RHS, const SimplifyQuery &SQ)
unsigned ComputeMaxSignificantBits(const Value *Op, const DataLayout &DL, unsigned Depth=0, AssumptionCache *AC=nullptr, const Instruction *CxtI=nullptr, const DominatorTree *DT=nullptr)
Get the upper bound on bit size for this Value Op as a signed integer.
SimplifyQuery getWithInstruction(const Instruction *I) const