36#define DEBUG_TYPE "bpf-lower"
40 cl::desc(
"Expand memcpy into load/store pairs in order"));
44 cl::desc(
"Set minimum number of entries to use a jump table on BPF"));
93 for (
auto VT : {MVT::i8, MVT::i16, MVT::i32}) {
108 for (
auto VT : {MVT::i32, MVT::i64}) {
113 for (
auto VT : { MVT::i32, MVT::i64 }) {
195 unsigned CommonMaxStores =
215 unsigned *
Fast)
const {
217 if (!AllowsMisalignedMemAccess)
235bool BPFTargetLowering::isTruncateFree(
Type *Ty1,
Type *Ty2)
const {
240 return NumBits1 > NumBits2;
243bool BPFTargetLowering::isTruncateFree(
EVT VT1,
EVT VT2)
const {
248 return NumBits1 > NumBits2;
251bool BPFTargetLowering::isZExtFree(
Type *Ty1,
Type *Ty2)
const {
256 return NumBits1 == 32 && NumBits2 == 64;
259bool BPFTargetLowering::isZExtFree(
EVT VT1,
EVT VT2)
const {
264 return NumBits1 == 32 && NumBits2 == 64;
267bool BPFTargetLowering::isZExtFree(
SDValue Val,
EVT VT2)
const {
272 if ((MT1 == MVT::i8 || MT1 == MVT::i16 || MT1 == MVT::i32) &&
273 (MT2 == MVT::i32 || MT2 == MVT::i64))
285 if (Constraint.
size() == 1) {
286 switch (Constraint[0]) {
297std::pair<unsigned, const TargetRegisterClass *>
301 if (Constraint.
size() == 1) {
303 switch (Constraint[0]) {
305 return std::make_pair(0U, &BPF::GPRRegClass);
308 return std::make_pair(0U, &BPF::GPR32RegClass);
318void BPFTargetLowering::ReplaceNodeResults(
332 Msg =
"unsupported atomic operation, please use 32/64 bit version";
334 Msg =
"unsupported atomic operation, please use 64 bit version";
348 switch (
Op.getOpcode()) {
352 return LowerBR_CC(
Op, DAG);
354 return LowerJumpTable(
Op, DAG);
356 return LowerGlobalAddress(
Op, DAG);
358 return LowerConstantPool(
Op, DAG);
360 return LowerBlockAddress(
Op, DAG);
362 return LowerSELECT_CC(
Op, DAG);
365 return LowerSDIVSREM(
Op, DAG);
369 return LowerShiftParts(
Op, DAG);
371 return LowerDYNAMIC_STACKALLOC(
Op, DAG);
374 return LowerATOMIC_LOAD_STORE(
Op, DAG);
376 return LowerTRAP(
Op, DAG);
381#include "BPFGenCallingConv.inc"
398SDValue BPFTargetLowering::LowerFormalArguments(
411 MachineRegisterInfo &RegInfo = MF.
getRegInfo();
415 CCState CCInfo(CallConv, IsVarArg, MF, ArgLocs, *DAG.
getContext());
416 CCInfo.AnalyzeFormalArguments(Ins,
getHasAlu32() ? CC_BPF32 : CC_BPF64);
418 for (
size_t I = 0;
I < ArgLocs.
size(); ++
I) {
419 auto &VA = ArgLocs[
I];
420 EVT RegVT = VA.getLocVT();
429 raw_string_ostream OS(Str);
437 SimpleTy == MVT::i64 ? &BPF::GPRRegClass : &BPF::GPR32RegClass);
450 int Off = VA.getLocMemOffset() + 8;
451 if (Off > INT16_MAX) {
452 fail(
DL, DAG,
"extra parameter stack depth exceeded limit");
458 DAG.
getVTList(MVT::i64, MVT::Other), Chain,
468 fail(
DL, DAG,
"variadic functions are not supported");
475 RegMask[SubReg / 32] &= ~(1u << (SubReg % 32));
483 memcpy(RegMask, BaseRegMask,
sizeof(RegMask[0]) * RegMaskSize);
489 SelectionDAG &DAG = CLI.
DAG;
490 auto &Outs = CLI.
Outs;
513 CCState CCInfo(CallConv, IsVarArg, MF, ArgLocs, *DAG.
getContext());
515 CCInfo.AnalyzeCallOperands(Outs,
getHasAlu32() ? CC_BPF32 : CC_BPF64);
517 unsigned NumBytes = CCInfo.getStackSize();
519 for (
auto &Arg : Outs) {
520 ISD::ArgFlagsTy
Flags = Arg.Flags;
521 if (!
Flags.isByVal())
523 fail(CLI.
DL, DAG,
"pass by value not supported", Callee);
533 for (
size_t i = 0; i < OutVals.size(); ++i) {
534 CCValAssign &VA = ArgLocs[i];
562 if (Off < INT16_MIN) {
563 fail(CLI.
DL, DAG,
"extra parameter stack depth exceeded limit");
573 Chain = DAG.
getNode(BPFISD::STORE_STACK_ARG, CLI.
DL, MVT::Other, Chain,
586 for (
auto &
Reg : RegsToPass) {
599 StringRef Sym =
E->getSymbol();
600 if (Sym !=
BPF_TRAP && Sym !=
"__multi3" && Sym !=
"__divti3" &&
601 Sym !=
"__modti3" && Sym !=
"__udivti3" && Sym !=
"__umodti3" &&
602 Sym !=
"memcpy" && Sym !=
"memset" && Sym !=
"memmove")
605 Twine(
"A call to built-in function '" + Sym +
"' is not supported."));
609 SDVTList NodeTys = DAG.
getVTList(MVT::Other, MVT::Glue);
611 Ops.push_back(Chain);
612 Ops.push_back(Callee);
616 for (
auto &
Reg : RegsToPass)
625 for (
auto const &RegPair : RegsToPass)
636 Ops.push_back(InGlue);
638 Chain = DAG.
getNode(BPFISD::CALL, CLI.
DL, NodeTys,
Ops);
649 return LowerCallResult(Chain, InGlue, CallConv, IsVarArg, Ins, CLI.
DL, DAG,
659 unsigned Opc = BPFISD::RET_GLUE;
666 CCState CCInfo(CallConv, IsVarArg, MF, RVLocs, *DAG.
getContext());
669 CCInfo.AnalyzeReturn(Outs,
getHasAlu32() ? RetCC_BPF32 : RetCC_BPF64);
675 for (
size_t i = 0; i != RVLocs.
size(); ++i) {
676 CCValAssign &VA = RVLocs[i];
692 RetOps.push_back(Glue);
697SDValue BPFTargetLowering::LowerCallResult(
705 CCState CCInfo(CallConv, IsVarArg, MF, RVLocs, *DAG.
getContext());
707 CCInfo.AnalyzeCallResult(Ins,
getHasAlu32() ? RetCC_BPF32 : RetCC_BPF64);
710 for (
auto &Val : RVLocs) {
712 Val.getValVT(), InGlue).
getValue(1);
737 "unsupported signed division, please convert to unsigned div/mod.");
751 fail(
DL, DAG,
"unsupported dynamic stack allocation");
785 return DAG.
getNode(BPFISD::SELECT_CC,
DL,
Op.getValueType(),
Ops);
790 SDNode *
N =
Op.getNode();
796 "sequentially consistent (seq_cst) "
797 "atomic load/store is not supported");
803 if (
auto *Fn = M->getFunction(
BPF_TRAP))
813 if (M->debug_compile_units().empty())
817 DITypeArray ParamTypes =
823 DINode::FlagZero, DISubprogram::SPFlagZero);
830 TargetLowering::CallLoweringInfo CLI(DAG);
832 SDNode *
N =
Op.getNode();
842 CLI.
DL = std::move(
DL);
845 return LowerCall(CLI, InVals);
850 return getAddr(
N, DAG);
856 N->getOffset(), Flags);
864template <
class NodeTy>
866 unsigned Flags)
const {
871 return DAG.
getNode(BPFISD::Wrapper,
DL, MVT::i64, GA);
877 if (
N->getOffset() != 0)
879 Twine(
N->getOffset()));
881 const GlobalValue *GVal =
N->getGlobal();
895 return getAddr(
N, DAG);
915 int RShiftOp =
isSigned ? BPF::SRA_ri : BPF::SRL_ri;
919 MachineRegisterInfo &RegInfo =
F->getRegInfo();
943BPFTargetLowering::EmitInstrWithCustomInserterMemcpy(
MachineInstr &
MI,
946 MachineFunction *MF =
MI.getParent()->getParent();
948 MachineInstrBuilder MIB(*MF,
MI);
967 MIB.addReg(ScratchReg,
976 const BPFInstrInfo *
TII = MF->
getSubtarget<BPFSubtarget>().getInstrInfo();
978 MachineRegisterInfo &RegInfo = MF->
getRegInfo();
982 DenseMap<const BasicBlock *, MachineBasicBlock *> AddressTakenBBs;
983 for (MachineBasicBlock &
MBB : *MF) {
986 AddressTakenBBs[BB] = &
MBB;
989 MachineOperand &MO =
MI.getOperand(1);
992 Register ResultReg =
MI.getOperand(0).getReg();
995 std::vector<MachineBasicBlock *> Targets;
1000 MachineBasicBlock *TgtMBB = AddressTakenBBs[BA->
getBasicBlock()];
1003 Targets.push_back(TgtMBB);
1005 ->createJumpTableIndex(Targets);
1012 MI.eraseFromParent();
1017 auto emitLDImm64 = [&](
const GlobalValue *GV =
nullptr,
unsigned JTI = -1) {
1018 auto MIB =
BuildMI(*BB,
MI,
DL,
TII->get(BPF::LD_imm64), ResultReg);
1020 MIB.addGlobalAddress(GV);
1022 MIB.addJumpTableIndex(JTI);
1023 MI.eraseFromParent();
1028 const GlobalValue *GVal = MO.
getGlobal();
1032 !GV->isConstant() || !GV->hasInitializer())
1033 return emitLDImm64(GVal);
1037 return emitLDImm64(GVal);
1039 for (
const Use &
Op : CA->operands()) {
1041 return emitLDImm64(GVal);
1043 MachineBasicBlock *TgtMBB = AddressTakenBBs[BA->
getBasicBlock()];
1045 Targets.push_back(TgtMBB);
1049 ->createJumpTableIndex(Targets);
1050 return emitLDImm64(
nullptr, JTI);
1058 unsigned Opc =
MI.getOpcode();
1059 bool isSelectRROp = (
Opc == BPF::Select ||
1060 Opc == BPF::Select_64_32 ||
1061 Opc == BPF::Select_32 ||
1062 Opc == BPF::Select_32_64);
1064 bool isMemcpyOp =
Opc == BPF::MEMCPY;
1065 bool isLDimm64Op =
Opc == BPF::LDIMM64;
1068 bool isSelectRIOp = (
Opc == BPF::Select_Ri ||
1069 Opc == BPF::Select_Ri_64_32 ||
1070 Opc == BPF::Select_Ri_32 ||
1071 Opc == BPF::Select_Ri_32_64);
1073 if (!(isSelectRROp || isSelectRIOp || isMemcpyOp || isLDimm64Op))
1078 return EmitInstrWithCustomInserterMemcpy(
MI, BB);
1081 return EmitInstrWithCustomInserterLDimm64(
MI, BB);
1083 bool is32BitCmp = (
Opc == BPF::Select_32 ||
1084 Opc == BPF::Select_32_64 ||
1085 Opc == BPF::Select_Ri_32 ||
1086 Opc == BPF::Select_Ri_32_64);
1105 F->insert(
I, Copy0MBB);
1106 F->insert(
I, Copy1MBB);
1117 int CC =
MI.getOperand(3).getImm();
1120#define SET_NEWCC(X, Y) \
1122 if (is32BitCmp && HasJmp32) \
1123 NewCC = isSelectRROp ? BPF::Y##_rr_32 : BPF::Y##_ri_32; \
1125 NewCC = isSelectRROp ? BPF::Y##_rr : BPF::Y##_ri; \
1154 if (is32BitCmp && !HasJmp32)
1155 LHS = EmitSubregExt(
MI, BB, LHS, isSignedCmp);
1160 if (is32BitCmp && !HasJmp32)
1161 RHS = EmitSubregExt(
MI, BB, RHS, isSignedCmp);
1165 int64_t imm32 =
MI.getOperand(2).getImm();
1191 MI.eraseFromParent();
1202 return (
getHasAlu32() && VT == MVT::i32) ? MVT::i32 : MVT::i64;
1205bool BPFTargetLowering::isLegalAddressingMode(
const DataLayout &
DL,
1227bool BPFTargetLowering::CanLowerReturn(
1230 const Type *RetTy)
const {
1232 CCState CCInfo(CallConv, IsVarArg, MF, RVLocs,
Context);
1233 return CCInfo.CheckReturn(Outs,
getHasAlu32() ? RetCC_BPF32 : RetCC_BPF64);
assert(UImm &&(UImm !=~static_cast< T >(0)) &&"Invalid immediate!")
MachineBasicBlock MachineBasicBlock::iterator DebugLoc DL
Function Alias Analysis Results
static uint32_t * regMaskFromTemplate(const TargetRegisterInfo *TRI, MachineFunction &MF, const uint32_t *BaseRegMask)
static Function * createBPFUnreachable(Module *M)
static SDValue getTargetNode(ConstantPoolSDNode *N, const SDLoc &DL, EVT Ty, SelectionDAG &DAG, unsigned Flags)
static cl::opt< bool > BPFExpandMemcpyInOrder("bpf-expand-memcpy-in-order", cl::Hidden, cl::init(false), cl::desc("Expand memcpy into load/store pairs in order"))
static SDValue convertLocValType(SelectionDAG &DAG, const SDLoc &DL, const CCValAssign &VA, EVT RegVT, SDValue ArgValue)
static void fail(const SDLoc &DL, SelectionDAG &DAG, const Twine &Msg, SDValue Val={})
static cl::opt< unsigned > BPFMinimumJumpTableEntries("bpf-min-jump-table-entries", cl::init(13), cl::Hidden, cl::desc("Set minimum number of entries to use a jump table on BPF"))
static void resetRegMaskBit(const TargetRegisterInfo *TRI, uint32_t *RegMask, MCRegister Reg)
static void NegateCC(SDValue &LHS, SDValue &RHS, ISD::CondCode &CC)
static GCRegistry::Add< CoreCLRGC > E("coreclr", "CoreCLR-compatible GC")
static bool isSigned(unsigned Opcode)
const HexagonInstrInfo * TII
Module.h This file contains the declarations for the Module class.
const AbstractManglingParser< Derived, Alloc >::OperatorInfo AbstractManglingParser< Derived, Alloc >::Ops[]
Register const TargetRegisterInfo * TRI
Promote Memory to Register
unsigned getCommonMaxStoresPerMemFunc() const
bool getAllowsMisalignedMemAccess() const
bool getHasJmpExt() const
const BPFSelectionDAGInfo * getSelectionDAGInfo() const override
const BPFRegisterInfo * getRegisterInfo() const override
BPFTargetLowering::ConstraintType getConstraintType(StringRef Constraint) const override
Given a constraint, return the type of constraint it is for this target.
unsigned getJumpTableEncoding() const override
Return the entry encoding for a jump table in the current function.
bool getHasJmpExt() const
bool isOffsetFoldingLegal(const GlobalAddressSDNode *GA) const override
Return true if folding a constant offset with the given GlobalAddress is legal.
bool allowsMisalignedMemoryAccesses(EVT VT, unsigned, Align, MachineMemOperand::Flags, unsigned *) const override
Determine if the target supports unaligned memory accesses.
EVT getSetCCResultType(const DataLayout &DL, LLVMContext &Context, EVT VT) const override
Return the ValueType of the result of SETCC operations.
BPFTargetLowering(const TargetMachine &TM, const BPFSubtarget &STI)
SDValue LowerOperation(SDValue Op, SelectionDAG &DAG) const override
This callback is invoked for operations that are unsupported by the target, which are registered to u...
MachineBasicBlock * EmitInstrWithCustomInserter(MachineInstr &MI, MachineBasicBlock *BB) const override
This method should be implemented by targets that mark instructions with the 'usesCustomInserter' fla...
MVT getScalarShiftAmountTy(const DataLayout &, EVT) const override
Return the type to use for a scalar shift opcode, given the shifted amount type.
std::pair< unsigned, const TargetRegisterClass * > getRegForInlineAsmConstraint(const TargetRegisterInfo *TRI, StringRef Constraint, MVT VT) const override
Given a physical register constraint (e.g.
LLVM Basic Block Representation.
BasicBlock * getBasicBlock() const
CCValAssign - Represent assignment of one arg/retval to a location.
Register getLocReg() const
LocInfo getLocInfo() const
int64_t getLocMemOffset() const
bool hasFnAttr(Attribute::AttrKind Kind) const
Determine whether this call has the given attribute.
LLVM_ABI DISubroutineType * createSubroutineType(DITypeArray ParameterTypes, DINode::DIFlags Flags=DINode::FlagZero, unsigned CC=0)
Create subroutine type.
LLVM_ABI DISubprogram * createFunction(DIScope *Scope, StringRef Name, StringRef LinkageName, DIFile *File, unsigned LineNo, DISubroutineType *Ty, unsigned ScopeLine, DINode::DIFlags Flags=DINode::FlagZero, DISubprogram::DISPFlags SPFlags=DISubprogram::SPFlagZero, DITemplateParameterArray TParams=nullptr, DISubprogram *Decl=nullptr, DITypeArray ThrownTypes=nullptr, DINodeArray Annotations=nullptr, StringRef TargetFuncName="", bool UseKeyInstructions=false)
Create a new descriptor for the specified subprogram.
LLVM_ABI DITypeArray getOrCreateTypeArray(ArrayRef< Metadata * > Elements)
Get a DITypeArray, create one if required.
Subprogram description. Uses SubclassData1.
Type array for a subprogram.
A parsed version of the target data layout string in and methods for querying it.
Diagnostic information for unsupported feature in backend.
static LLVM_ABI FunctionType * get(Type *Result, ArrayRef< Type * > Params, bool isVarArg)
This static method is the primary way of constructing a FunctionType.
void setSubprogram(DISubprogram *SP)
Set the attached subprogram.
static Function * Create(FunctionType *Ty, LinkageTypes Linkage, unsigned AddrSpace, const Twine &N="", Module *M=nullptr)
void setCallingConv(CallingConv::ID CC)
LLVM_ABI void setSection(StringRef S)
Change the section for this global.
LinkageTypes getLinkage() const
Module * getParent()
Get the module that this global value is contained inside of...
void setDSOLocal(bool Local)
@ PrivateLinkage
Like Internal, but omit from symbol table.
@ ExternalWeakLinkage
ExternalWeak linkage description.
This is an important class for using LLVM in a threaded context.
LLVM_ABI void diagnose(const DiagnosticInfo &DI)
Report a message to the currently installed diagnostic handler.
Wrapper class representing physical registers. Should be passed by value.
static auto integer_valuetypes()
LLVM_ABI void transferSuccessorsAndUpdatePHIs(MachineBasicBlock *FromMBB)
Transfers all the successors, as in transferSuccessors, and update PHI operands in the successor bloc...
const BasicBlock * getBasicBlock() const
Return the LLVM basic block that this instance corresponded to originally.
bool hasAddressTaken() const
Test whether this block is used as something other than the target of a terminator,...
LLVM_ABI void addSuccessor(MachineBasicBlock *Succ, BranchProbability Prob=BranchProbability::getUnknown())
Add Succ as a successor of this MachineBasicBlock.
const MachineFunction * getParent() const
Return the MachineFunction containing this basic block.
void splice(iterator Where, MachineBasicBlock *Other, iterator From)
Take an instruction from MBB 'Other' at the position From, and insert it into this MBB right before '...
MachineInstrBundleIterator< MachineInstr > iterator
const TargetSubtargetInfo & getSubtarget() const
getSubtarget - Return the subtarget for which this machine code is being compiled.
uint32_t * allocateRegMask()
Allocate and initialize a register mask with NumRegister bits.
MachineRegisterInfo & getRegInfo()
getRegInfo - Return information about the registers currently in use.
const DataLayout & getDataLayout() const
Return the DataLayout attached to the Module associated to this MF.
Function & getFunction()
Return the LLVM function that this machine code represents.
BasicBlockListType::iterator iterator
const MachineInstrBuilder & addReg(Register RegNo, RegState Flags={}, unsigned SubReg=0) const
Add a new virtual register operand.
const MachineInstrBuilder & addImm(int64_t Val) const
Add a new immediate operand.
const MachineInstrBuilder & addJumpTableIndex(unsigned Idx, unsigned TargetFlags=0) const
const MachineInstrBuilder & addMBB(MachineBasicBlock *MBB, unsigned TargetFlags=0) const
Representation of each machine instruction.
@ EK_BlockAddress
EK_BlockAddress - Each entry is a plain address of block, e.g.: .word LBB123.
Flags
Flags values. These may be or'd together.
const GlobalValue * getGlobal() const
const BlockAddress * getBlockAddress() const
static unsigned getRegMaskSize(unsigned NumRegs)
Returns number of elements needed for a regmask array.
bool isGlobal() const
isGlobal - Tests if this is a MO_GlobalAddress operand.
bool isBlockAddress() const
isBlockAddress - Tests if this is a MO_BlockAddress operand.
LLVM_ABI Register createVirtualRegister(const TargetRegisterClass *RegClass, StringRef Name="")
createVirtualRegister - Create and return a new virtual register in the function with the specified r...
void addLiveIn(MCRegister Reg, Register vreg=Register())
addLiveIn - Add the specified register as a live-in.
A Module instance is used to store all the information related to an LLVM module.
Wrapper class representing virtual and physical registers.
Wrapper class for IR location info (IR ordering and DebugLoc) to be passed into SDNode creation funct...
Represents one node in the SelectionDAG.
const SDValue & getOperand(unsigned Num) const
Unlike LLVM values, Selection DAG nodes may return multiple values as the result of a computation.
SDNode * getNode() const
get the SDNode which holds the desired result
SDValue getValue(unsigned R) const
EVT getValueType() const
Return the ValueType of the referenced return value.
unsigned getOpcode() const
This is used to represent a portion of an LLVM function in a low-level Data Dependence DAG representa...
SDValue getTargetGlobalAddress(const GlobalValue *GV, const SDLoc &DL, EVT VT, int64_t offset=0, unsigned TargetFlags=0)
SDValue getCopyToReg(SDValue Chain, const SDLoc &dl, Register Reg, SDValue N)
LLVM_ABI SDValue getMergeValues(ArrayRef< SDValue > Ops, const SDLoc &dl)
Create a MERGE_VALUES node from the given operands.
LLVM_ABI SDVTList getVTList(EVT VT)
Return an SDVTList that represents the list of values specified.
LLVM_ABI MachineSDNode * getMachineNode(unsigned Opcode, const SDLoc &dl, EVT VT)
These are used for target selectors to create a new node with specified return type(s),...
LLVM_ABI SDValue getRegister(Register Reg, EVT VT)
void addNoMergeSiteInfo(const SDNode *Node, bool NoMerge)
Set NoMergeSiteInfo to be associated with Node if NoMerge is true.
SDValue getTargetJumpTable(int JTI, EVT VT, unsigned TargetFlags=0)
SDValue getUNDEF(EVT VT)
Return an UNDEF node. UNDEF does not have a useful SDLoc.
SDValue getCALLSEQ_END(SDValue Chain, SDValue Op1, SDValue Op2, SDValue InGlue, const SDLoc &DL)
Return a new CALLSEQ_END node, which always must have a glue result (to ensure it's not CSE'd).
SDValue getCopyFromReg(SDValue Chain, const SDLoc &dl, Register Reg, EVT VT)
LLVM_ABI SDValue getConstant(uint64_t Val, const SDLoc &DL, EVT VT, bool isTarget=false, bool isOpaque=false)
Create a ConstantSDNode wrapping a constant value.
SDValue getCALLSEQ_START(SDValue Chain, uint64_t InSize, uint64_t OutSize, const SDLoc &DL)
Return a new CALLSEQ_START node, that starts new call frame, in which InSize bytes are set up inside ...
LLVM_ABI SDValue getValueType(EVT)
LLVM_ABI SDValue getNode(unsigned Opcode, const SDLoc &DL, EVT VT, ArrayRef< SDUse > Ops)
Gets or creates the specified node.
SDValue getTargetBlockAddress(const BlockAddress *BA, EVT VT, int64_t Offset=0, unsigned TargetFlags=0)
MachineFunction & getMachineFunction() const
LLVM_ABI SDValue getRegisterMask(const uint32_t *RegMask)
LLVMContext * getContext() const
LLVM_ABI SDValue getTargetExternalSymbol(const char *Sym, EVT VT, unsigned TargetFlags=0)
SDValue getTargetConstantPool(const Constant *C, EVT VT, MaybeAlign Align=std::nullopt, int Offset=0, unsigned TargetFlags=0)
This class consists of common code factored out of the SmallVector class to reduce code duplication b...
void push_back(const T &Elt)
Represent a constant reference to a string, i.e.
constexpr size_t size() const
Get the string size.
TargetInstrInfo - Interface to description of machine instruction set.
void setOperationAction(unsigned Op, MVT VT, LegalizeAction Action)
Indicate that the specified operation does not work with the specified type and indicate what to do a...
unsigned MaxStoresPerMemcpyOptSize
Likewise for functions with the OptSize attribute.
virtual const TargetRegisterClass * getRegClassFor(MVT VT, bool isDivergent=false) const
Return the register class that should be used for the specified value type.
unsigned MaxLoadsPerMemcmp
Specify maximum number of load instructions per memcmp call.
virtual bool isZExtFree(Type *FromTy, Type *ToTy) const
Return true if any actual instruction that defines a value of type FromTy implicitly zero-extends the...
void setMaxAtomicSizeInBitsSupported(unsigned SizeInBits)
Set the maximum atomic operation size supported by the backend.
void setMinFunctionAlignment(Align Alignment)
Set the target's minimum function alignment.
unsigned MaxStoresPerMemsetOptSize
Likewise for functions with the OptSize attribute.
void setBooleanContents(BooleanContent Ty)
Specify how the target extends the result of integer and floating point boolean values from i1 to a w...
unsigned MaxStoresPerMemmove
Specify maximum number of store instructions per memmove call.
void computeRegisterProperties(const TargetRegisterInfo *TRI)
Once all of the register classes are added, this allows us to compute derived properties we expose.
unsigned MaxStoresPerMemmoveOptSize
Likewise for functions with the OptSize attribute.
void addRegisterClass(MVT VT, const TargetRegisterClass *RC)
Add the specified register class as an available regclass for the specified value type.
virtual MVT getPointerTy(const DataLayout &DL, uint32_t AS=0) const
Return the pointer type for the given address space, defaults to the pointer type from the data layou...
void setPrefFunctionAlignment(Align Alignment)
Set the target's preferred function alignment.
unsigned MaxStoresPerMemset
Specify maximum number of store instructions per memset call.
void setMinimumJumpTableEntries(unsigned Val)
Indicate the minimum number of blocks to generate jump tables.
@ ZeroOrOneBooleanContent
unsigned MaxLoadsPerMemcmpOptSize
Likewise for functions with the OptSize attribute.
void setStackPointerRegisterToSaveRestore(Register R)
If set to a physical register, this specifies the register that llvm.savestack/llvm....
void setLoadExtAction(unsigned ExtType, MVT ValVT, MVT MemVT, LegalizeAction Action)
Indicate that the specified load with extension does not work with the specified type and indicate wh...
unsigned MaxStoresPerMemcpy
Specify maximum number of store instructions per memcpy call.
virtual ConstraintType getConstraintType(StringRef Constraint) const
Given a constraint, return the type of constraint it is for this target.
virtual std::pair< unsigned, const TargetRegisterClass * > getRegForInlineAsmConstraint(const TargetRegisterInfo *TRI, StringRef Constraint, MVT VT) const
Given a physical register constraint (e.g.
TargetLowering(const TargetLowering &)=delete
void expandShiftParts(SDNode *N, SDValue &Lo, SDValue &Hi, SelectionDAG &DAG) const
Expand shift-by-parts.
Primary interface to the complete machine description for the target machine.
TargetRegisterInfo base class - We assume that the target defines a static array of TargetRegisterDes...
virtual const TargetInstrInfo * getInstrInfo() const
virtual const TargetRegisterInfo * getRegisterInfo() const =0
Return the target's register information.
Twine - A lightweight data structure for efficiently representing the concatenation of temporary valu...
The instances of the Type class are immutable: once they are created, they are never changed.
static LLVM_ABI Type * getVoidTy(LLVMContext &C)
LLVM_ABI TypeSize getPrimitiveSizeInBits() const LLVM_READONLY
Return the basic size of this type if it is a primitive type.
bool isIntegerTy() const
True if this is an instance of IntegerType.
bool isVoidTy() const
Return true if this is 'void'.
Type * getType() const
All values are typed, get the type of this value.
self_iterator getIterator()
A raw_ostream that writes to an std::string.
unsigned ID
LLVM IR allows to use arbitrary numbers as calling convention identifiers.
@ PreserveAll
Used for runtime calls that preserves (almost) all registers.
@ Fast
Attempts to make calls as fast as possible (e.g.
@ C
The default llvm calling convention, compatible with C.
@ SETCC
SetCC operator - This evaluates to a true value iff the condition is true.
@ STACKRESTORE
STACKRESTORE has two operands, an input chain and a pointer to restore to it returns an output chain.
@ STACKSAVE
STACKSAVE - STACKSAVE has one operand, an input chain.
@ SMUL_LOHI
SMUL_LOHI/UMUL_LOHI - Multiply two integers of type iN, producing a signed/unsigned value of type i[2...
@ BSWAP
Byte Swap and Counting operators.
@ ATOMIC_STORE
OUTCHAIN = ATOMIC_STORE(INCHAIN, val, ptr) This corresponds to "store atomic" instruction.
@ LOAD
LOAD and STORE have token chains as their first operand, then the same operands as an LLVM load/store...
@ ANY_EXTEND
ANY_EXTEND - Used for integer types. The high bits are undefined.
@ ATOMIC_CMP_SWAP_WITH_SUCCESS
Val, Success, OUTCHAIN = ATOMIC_CMP_SWAP_WITH_SUCCESS(INCHAIN, ptr, cmp, swap) N.b.
@ SDIVREM
SDIVREM/UDIVREM - Divide two integers and produce both a quotient and remainder result.
@ SIGN_EXTEND
Conversion operators.
@ BR_CC
BR_CC - Conditional branch.
@ BRIND
BRIND - Indirect branch.
@ BR_JT
BR_JT - Jumptable branch.
@ SELECT
Select(COND, TRUEVAL, FALSEVAL).
@ ATOMIC_LOAD
Val, OUTCHAIN = ATOMIC_LOAD(INCHAIN, ptr) This corresponds to "load atomic" instruction.
@ MULHU
MULHU/MULHS - Multiply high - Multiply two integers of type iN, producing an unsigned/signed value of...
@ ZERO_EXTEND
ZERO_EXTEND - Used for integer types, zeroing the new bits.
@ SELECT_CC
Select with condition operator - This selects between a true value and a false value (ops #2 and #3) ...
@ DYNAMIC_STACKALLOC
DYNAMIC_STACKALLOC - Allocate some number of bytes on the stack aligned to a specified boundary.
@ SIGN_EXTEND_INREG
SIGN_EXTEND_INREG - This operator atomically performs a SHL/SRA pair to sign extend a small value in ...
@ TRAP
TRAP - Trapping instruction.
@ ATOMIC_SWAP
Val, OUTCHAIN = ATOMIC_SWAP(INCHAIN, ptr, amt) Val, OUTCHAIN = ATOMIC_LOAD_[OpName](INCHAIN,...
@ CTTZ_ZERO_POISON
Bit counting operators with a poisoned result for zero inputs.
@ TRUNCATE
TRUNCATE - Completely drop the high bits.
@ BRCOND
BRCOND - Conditional branch.
@ SHL_PARTS
SHL_PARTS/SRA_PARTS/SRL_PARTS - These operators are used for expanded integer shift operations.
@ AssertSext
AssertSext, AssertZext - These nodes record if a register contains a value that has already been zero...
LLVM_ABI CondCode getSetCCSwappedOperands(CondCode Operation)
Return the operation corresponding to (Y op X) when given the operation for (X op Y).
CondCode
ISD::CondCode enum - These are ordered carefully to make the bitfields below work out,...
initializer< Ty > init(const Ty &Val)
This is an optimization pass for GlobalISel generic memory operations.
MachineInstrBuilder BuildMI(MachineFunction &MF, const MIMetadata &MIMD, const MCInstrDesc &MCID)
Builder interface. Specify how to create the initial instruction itself.
constexpr bool isInt(int64_t x)
Checks if an integer fits into the given bit width.
@ EarlyClobber
Register definition happens before uses.
@ Define
Register definition.
decltype(auto) dyn_cast(const From &Val)
dyn_cast<X> - Return the argument parameter cast to the specified type.
detail::concat_range< ValueT, RangeTs... > concat(RangeTs &&...Ranges)
Returns a concatenated range across two or more ranges.
LLVM_ABI void report_fatal_error(Error Err, bool gen_crash_diag=true)
class LLVM_GSL_OWNER SmallVector
Forward declaration of SmallVector so that calculateSmallVectorDefaultInlinedElements can reference s...
bool isa(const From &Val)
isa<X> - Return true if the parameter to the template is an instance of one of the template type argu...
uint16_t MCPhysReg
An unsigned integer type large enough to represent all physical registers, but not necessarily virtua...
DWARFExpression::Operation Op
decltype(auto) cast(const From &Val)
cast<X> - Return the argument parameter cast to the specified type.
void swap(llvm::BitVector &LHS, llvm::BitVector &RHS)
Implement std::swap in terms of BitVector swap.
This struct is a compact representation of a valid (non-zero power of two) alignment.
bool isSimple() const
Test if the given EVT is simple (as opposed to being extended).
TypeSize getSizeInBits() const
Return the size of the specified value type in bits.
MVT getSimpleVT() const
Return the SimpleValueType held in the specified simple EVT.
void print(raw_ostream &OS) const
Implement operator<<.
bool isInteger() const
Return true if this is an integer or a vector integer type.
This structure contains all information that is necessary for lowering calls.
SmallVector< ISD::InputArg, 32 > Ins
SmallVector< ISD::OutputArg, 32 > Outs
SmallVector< SDValue, 32 > OutVals