27#define GET_REGINFO_TARGET_DESC
28#include "RISCVGenRegisterInfo.inc"
37 cl::desc(
"Disable two address hints for register "
40static_assert(RISCV::X1 == RISCV::X0 + 1,
"Register list not consecutive");
41static_assert(RISCV::X31 == RISCV::X0 + 31,
"Register list not consecutive");
42static_assert(RISCV::F1_H == RISCV::F0_H + 1,
"Register list not consecutive");
43static_assert(RISCV::F31_H == RISCV::F0_H + 31,
44 "Register list not consecutive");
45static_assert(RISCV::F1_F == RISCV::F0_F + 1,
"Register list not consecutive");
46static_assert(RISCV::F31_F == RISCV::F0_F + 31,
47 "Register list not consecutive");
48static_assert(RISCV::F1_D == RISCV::F0_D + 1,
"Register list not consecutive");
49static_assert(RISCV::F31_D == RISCV::F0_D + 31,
50 "Register list not consecutive");
51static_assert(RISCV::F1_Q == RISCV::F0_Q + 1,
"Register list not consecutive");
52static_assert(RISCV::F31_Q == RISCV::F0_Q + 31,
53 "Register list not consecutive");
54static_assert(RISCV::V1 == RISCV::V0 + 1,
"Register list not consecutive");
55static_assert(RISCV::V31 == RISCV::V0 + 31,
"Register list not consecutive");
63 return CSR_IPRA_SaveList;
70 return CSR_NoRegs_SaveList;
72 return Subtarget.hasStdExtE() ? CSR_RT_MostRegs_RVE_SaveList
73 : CSR_RT_MostRegs_SaveList;
75 if (Subtarget.hasVInstructions()) {
76 if (Subtarget.hasStdExtD())
77 return Subtarget.hasStdExtE() ? CSR_XLEN_F64_V_Interrupt_RVE_SaveList
78 : CSR_XLEN_F64_V_Interrupt_SaveList;
79 if (Subtarget.hasStdExtF())
80 return Subtarget.hasStdExtE() ? CSR_XLEN_F32_V_Interrupt_RVE_SaveList
81 : CSR_XLEN_F32_V_Interrupt_SaveList;
82 return Subtarget.hasStdExtE() ? CSR_XLEN_V_Interrupt_RVE_SaveList
83 : CSR_XLEN_V_Interrupt_SaveList;
85 if (Subtarget.hasStdExtD())
86 return Subtarget.hasStdExtE() ? CSR_XLEN_F64_Interrupt_RVE_SaveList
87 : CSR_XLEN_F64_Interrupt_SaveList;
88 if (Subtarget.hasStdExtF())
89 return Subtarget.hasStdExtE() ? CSR_XLEN_F32_Interrupt_RVE_SaveList
90 : CSR_XLEN_F32_Interrupt_SaveList;
91 return Subtarget.hasStdExtE() ? CSR_Interrupt_RVE_SaveList
92 : CSR_Interrupt_SaveList;
97 Subtarget.hasVInstructions();
99 switch (Subtarget.getTargetABI()) {
104 return CSR_ILP32E_LP64E_SaveList;
108 return CSR_ILP32_LP64_V_SaveList;
109 return CSR_ILP32_LP64_SaveList;
113 return CSR_ILP32F_LP64F_V_SaveList;
114 return CSR_ILP32F_LP64F_SaveList;
118 return CSR_ILP32D_LP64D_V_SaveList;
119 return CSR_ILP32D_LP64D_SaveList;
128 for (
size_t Reg = 0; Reg < getNumRegs(); Reg++) {
130 if (Subtarget.isRegisterReservedByUser(Reg))
134 if (isConstantPhysReg(Reg))
139 markSuperRegs(
Reserved, RISCV::X2_H);
140 markSuperRegs(
Reserved, RISCV::X3_H);
141 markSuperRegs(
Reserved, RISCV::X4_H);
143 markSuperRegs(
Reserved, RISCV::X8_H);
151 markSuperRegs(
Reserved, RISCV::DUMMY_REG_PAIR_WITH_X0);
154 if (Subtarget.hasStdExtE())
155 for (
MCPhysReg Reg = RISCV::X16_H; Reg <= RISCV::X31_H; Reg++)
160 markSuperRegs(
Reserved, RISCV::VTYPE);
161 markSuperRegs(
Reserved, RISCV::VXSAT);
162 markSuperRegs(
Reserved, RISCV::VXRM);
165 markSuperRegs(
Reserved, RISCV::FRM);
166 markSuperRegs(
Reserved, RISCV::FFLAGS);
169 markSuperRegs(
Reserved, RISCV::SF_VCIX_STATE);
172 if (Subtarget.hasStdExtE())
174 markSuperRegs(
Reserved, RISCV::X23_H);
175 markSuperRegs(
Reserved, RISCV::X27_H);
179 markSuperRegs(
Reserved, RISCV::SSP);
191 return CSR_NoRegs_RegMask;
201 if (DestReg == SrcReg && !
Offset.getFixed() && !
Offset.getScalable())
210 if (
Offset.getScalable()) {
211 if (
auto VLEN = ST.getRealVLen()) {
213 const int64_t VLENB = *VLEN / 8;
215 "Reserve the stack by the multiple of one vector size.");
216 const int64_t NumOfVReg =
Offset.getScalable() / 8;
217 const int64_t FixedOffset = NumOfVReg * VLENB;
220 "Frame size outside of the signed 32-bit range not supported");
226 bool KillSrcReg =
false;
228 if (
Offset.getScalable()) {
229 unsigned ScalableAdjOpc = RISCV::ADD;
230 int64_t ScalableValue =
Offset.getScalable();
231 if (ScalableValue < 0) {
232 ScalableValue = -ScalableValue;
233 ScalableAdjOpc = RISCV::SUB;
237 if (DestReg == SrcReg)
238 ScratchReg =
MRI.createVirtualRegister(&RISCV::GPRRegClass);
240 assert(ScalableValue > 0 &&
"There is no need to get VLEN scaled value.");
242 "Reserve the stack by the multiple of one vector size.");
244 "Expect the number of vector registers within 32-bits.");
248 bool IsPrologueOrEpilogue =
250 bool UseVsetvliRatherThanVlenb =
251 IsPrologueOrEpilogue && ST.preferVsetvliOverReadVLENB();
252 if (UseVsetvliRatherThanVlenb && (NumOfVReg == 1 || NumOfVReg == 2 ||
253 NumOfVReg == 4 || NumOfVReg == 8)) {
263 if (UseVsetvliRatherThanVlenb)
272 if (ScalableAdjOpc == RISCV::ADD && ST.hasStdExtZba() &&
273 (NumOfVReg == 2 || NumOfVReg == 4 || NumOfVReg == 8)) {
274 unsigned Opc = NumOfVReg == 2
276 : (NumOfVReg == 4 ? RISCV::SH2ADD : RISCV::SH3ADD);
282 TII->mulImm(MF,
MBB,
II,
DL, ScratchReg, NumOfVReg, Flag);
293 int64_t Val =
Offset.getFixed();
294 if (DestReg == SrcReg && Val == 0)
309 if (ST.hasVendorXqcilia() &&
isInt<26>(Val)) {
313 int Hi20 = (Val & 0xFFFFF000) >> 12;
315 ((Val & 0xFFF) == 0) && (Hi20 != 0) &&
316 (
isUInt<5>(Hi20) || (Hi20 >= 0xfffe0 && Hi20 <= 0xfffff));
317 bool IsCompressAddSub =
318 (SrcReg == DestReg) &&
319 ((Val > 0 && RISCV::GPRNoX0RegClass.
contains(SrcReg)) ||
320 (Val < 0 && RISCV::GPRCRegClass.
contains(SrcReg)));
322 if (!(IsCompressLUI && IsCompressAddSub)) {
337 assert(
Align < 2048 &&
"Required alignment too large");
338 int64_t MaxPosAdjStep = 2048 -
Align;
339 if (Val > -4096 && Val <= (2 * MaxPosAdjStep)) {
340 int64_t FirstAdj = Val < 0 ? -2048 : MaxPosAdjStep;
358 if (ST.hasStdExtZba() && (Val & 0xFFF) != 0) {
368 Register ScratchReg =
MRI.createVirtualRegister(&RISCV::GPRRegClass);
369 TII->movImm(
MBB,
II,
DL, ScratchReg, Val, Flag);
378 unsigned Opc = RISCV::ADD;
384 Register ScratchReg =
MRI.createVirtualRegister(&RISCV::GPRRegClass);
385 TII->movImm(
MBB,
II,
DL, ScratchReg, Val, Flag);
392static std::tuple<RISCVVType::VLMUL, const TargetRegisterClass &, unsigned>
394 if (NumRemaining >= 8 && RegEncoding % 8 == 0)
396 IsSpill ? RISCV::VS8R_V : RISCV::VL8RE8_V};
397 if (NumRemaining >= 4 && RegEncoding % 4 == 0)
399 IsSpill ? RISCV::VS4R_V : RISCV::VL4RE8_V};
400 if (NumRemaining >= 2 && RegEncoding % 2 == 0)
402 IsSpill ? RISCV::VS2R_V : RISCV::VL2RE8_V};
404 IsSpill ? RISCV::VS1R_V : RISCV::VL1RE8_V};
410 bool IsSpill)
const {
420 unsigned NF = ZvlssegInfo->first;
421 unsigned LMUL = ZvlssegInfo->second;
422 unsigned NumRegs = NF * LMUL;
423 assert(NumRegs <= 8 &&
"Invalid NF/LMUL combinations.");
428 bool IsBaseKill =
II->getOperand(1).isKill();
429 Register NewBase =
MRI.createVirtualRegister(&RISCV::GPRRegClass);
431 auto *OldMMO = *(
II->memoperands_begin());
437 unsigned PreHandledNum = 0;
439 while (
I != NumRegs) {
440 auto [LMulHandled, RegClass, Opcode] =
443 bool IsLast =
I + RegNumHandled == NumRegs;
448 int64_t
Offset = *VLEN / 8 * PreHandledNum;
449 Step =
MRI.createVirtualRegister(&RISCV::GPRRegClass);
453 VLENB =
MRI.createVirtualRegister(&RISCV::GPRRegClass);
457 if (ShiftAmount == 0)
460 Step =
MRI.createVirtualRegister(&RISCV::GPRRegClass);
479 VRegSize * RegNumHandled));
488 PreHandledNum = RegNumHandled;
489 RegEncoding += RegNumHandled;
492 II->eraseFromParent();
496 int SPAdj,
unsigned FIOperandNum,
498 assert(SPAdj == 0 &&
"Unexpected non-zero SPAdj value");
505 int FrameIndex =
MI.getOperand(FIOperandNum).getIndex();
508 getFrameLowering(MF)->getFrameIndexReference(MF, FrameIndex, FrameReg);
515 "Frame offsets outside of the signed 32-bit range not supported");
519 int64_t Val =
Offset.getFixed();
521 unsigned Opc =
MI.getOpcode();
529 MI.getOperand(FIOperandNum + 1).ChangeToImmediate(0);
530 }
else if ((
Opc == RISCV::PREFETCH_I ||
Opc == RISCV::PREFETCH_R ||
531 Opc == RISCV::PREFETCH_W) &&
532 (Lo12 & 0b11111) != 0) {
534 MI.getOperand(FIOperandNum + 1).ChangeToImmediate(0);
535 }
else if (
Opc == RISCV::MIPS_PREF && !
isUInt<9>(Val)) {
537 MI.getOperand(FIOperandNum + 1).ChangeToImmediate(0);
538 }
else if ((
Opc == RISCV::PseudoRV32ZdinxLD ||
539 Opc == RISCV::PseudoRV32ZdinxSD) &&
544 MI.getOperand(FIOperandNum + 1).ChangeToImmediate(0);
549 MI.getOperand(FIOperandNum + 1).ChangeToImmediate(Lo12);
557 if (
MI.getOpcode() == RISCV::ADDI)
558 DestReg =
MI.getOperand(0).getReg();
560 DestReg =
MRI.createVirtualRegister(&RISCV::GPRRegClass);
563 MI.getOperand(FIOperandNum).ChangeToRegister(DestReg,
false,
567 MI.getOperand(FIOperandNum).ChangeToRegister(FrameReg,
false,
573 if (
MI.getOpcode() == RISCV::ADDI &&
574 MI.getOperand(0).getReg() ==
MI.getOperand(1).getReg() &&
575 MI.getOperand(2).getImm() == 0) {
576 MI.eraseFromParent();
582 switch (
MI.getOpcode()) {
583 case RISCV::PseudoVSPILL2_M1:
584 case RISCV::PseudoVSPILL2_M2:
585 case RISCV::PseudoVSPILL2_M4:
586 case RISCV::PseudoVSPILL3_M1:
587 case RISCV::PseudoVSPILL3_M2:
588 case RISCV::PseudoVSPILL4_M1:
589 case RISCV::PseudoVSPILL4_M2:
590 case RISCV::PseudoVSPILL5_M1:
591 case RISCV::PseudoVSPILL6_M1:
592 case RISCV::PseudoVSPILL7_M1:
593 case RISCV::PseudoVSPILL8_M1:
596 case RISCV::PseudoVRELOAD2_M1:
597 case RISCV::PseudoVRELOAD2_M2:
598 case RISCV::PseudoVRELOAD2_M4:
599 case RISCV::PseudoVRELOAD3_M1:
600 case RISCV::PseudoVRELOAD3_M2:
601 case RISCV::PseudoVRELOAD4_M1:
602 case RISCV::PseudoVRELOAD4_M2:
603 case RISCV::PseudoVRELOAD5_M1:
604 case RISCV::PseudoVRELOAD6_M1:
605 case RISCV::PseudoVRELOAD7_M1:
606 case RISCV::PseudoVRELOAD8_M1:
625 unsigned FIOperandNum = 0;
626 for (; !
MI->getOperand(FIOperandNum).isFI(); FIOperandNum++)
627 assert(FIOperandNum < MI->getNumOperands() &&
628 "Instr doesn't have FrameIndex operand");
637 if (!
MI->mayLoad() && !
MI->mayStore())
645 if (TFI->
hasFP(MF) && !shouldRealignStack(MF)) {
649 unsigned CalleeSavedSize = 0;
652 if (Subtarget.isRegisterReservedByUser(Reg))
655 if (RISCV::GPRRegClass.
contains(Reg))
656 CalleeSavedSize += getSpillSize(RISCV::GPRRegClass);
657 else if (RISCV::FPR64RegClass.
contains(Reg))
658 CalleeSavedSize += getSpillSize(RISCV::FPR64RegClass);
659 else if (RISCV::FPR32RegClass.
contains(Reg))
660 CalleeSavedSize += getSpillSize(RISCV::FPR32RegClass);
664 int64_t MaxFPOffset =
Offset - CalleeSavedSize;
672 int64_t MaxSPOffset =
Offset + 128;
682 unsigned FIOperandNum = 0;
683 while (!
MI->getOperand(FIOperandNum).isFI()) {
685 assert(FIOperandNum < MI->getNumOperands() &&
686 "Instr does not have a FrameIndex operand!");
718 unsigned FIOperandNum = 0;
719 while (!
MI.getOperand(FIOperandNum).isFI()) {
721 assert(FIOperandNum <
MI.getNumOperands() &&
722 "Instr does not have a FrameIndex operand!");
728 MI.getOperand(FIOperandNum).ChangeToRegister(BaseReg,
false);
729 MI.getOperand(FIOperandNum + 1).ChangeToImmediate(
Offset);
738 "The MI must be I or S format.");
739 assert(
MI->getOperand(Idx).isFI() &&
"The Idx'th operand of MI is not a "
740 "FrameIndex operand");
741 return MI->getOperand(Idx + 1).getImm();
746 return TFI->
hasFP(MF) ? RISCV::X8 : RISCV::X2;
750 if (Reg == RISCV::SF_VCIX_STATE)
751 return "sf.vcix_state";
761 return CSR_NoRegs_RegMask;
765 return CSR_RT_MostRegs_RVE_RegMask;
766 return CSR_RT_MostRegs_RegMask;
773 return CSR_ILP32E_LP64E_RegMask;
777 return CSR_ILP32_LP64_V_RegMask;
778 return CSR_ILP32_LP64_RegMask;
782 return CSR_ILP32F_LP64F_V_RegMask;
783 return CSR_ILP32F_LP64F_RegMask;
787 return CSR_ILP32D_LP64D_V_RegMask;
788 return CSR_ILP32D_LP64D_RegMask;
795 if (RC == &RISCV::VMV0RegClass)
796 return &RISCV::VRRegClass;
797 if (RC == &RISCV::VRNoV0RegClass)
798 return &RISCV::VRRegClass;
799 if (RC == &RISCV::VRM2NoV0RegClass)
800 return &RISCV::VRM2RegClass;
801 if (RC == &RISCV::VRM4NoV0RegClass)
802 return &RISCV::VRM4RegClass;
803 if (RC == &RISCV::VRM8NoV0RegClass)
804 return &RISCV::VRM8RegClass;
813 assert(
Offset.getScalable() % 8 == 0 &&
"Invalid frame offset");
819 int64_t VLENBSized =
Offset.getScalable() / 8;
820 if (VLENBSized > 0) {
821 Ops.push_back(dwarf::DW_OP_constu);
822 Ops.push_back(VLENBSized);
823 Ops.append({dwarf::DW_OP_bregx, VLENB, 0ULL});
824 Ops.push_back(dwarf::DW_OP_mul);
825 Ops.push_back(dwarf::DW_OP_plus);
826 }
else if (VLENBSized < 0) {
827 Ops.push_back(dwarf::DW_OP_constu);
828 Ops.push_back(-VLENBSized);
829 Ops.append({dwarf::DW_OP_bregx, VLENB, 0ULL});
830 Ops.push_back(dwarf::DW_OP_mul);
831 Ops.push_back(dwarf::DW_OP_minus);
844 return getRegClassWeight(RC).RegWeight;
857 VirtReg, Order, Hints, MF, VRM,
Matrix);
860 return BaseImplRetVal;
866 bool NeedGPRC) ->
void {
872 if (PhysReg && (!NeedGPRC || RISCV::GPRCRegClass.
contains(PhysReg)) &&
873 !MO.getSubReg() && !VRRegMO.
getSubReg()) {
875 TwoAddrHints.
insert(PhysReg);
881 auto isCompressible = [&Subtarget](
const MachineInstr &
MI,
bool &NeedGPRC) {
883 switch (
MI.getOpcode()) {
896 if (!
MI.getOperand(2).isImm())
898 int64_t Imm =
MI.getOperand(2).getImm();
902 return Subtarget.hasStdExtZcb() && Imm == 255;
913 return MI.getOperand(2).isImm() &&
isInt<6>(
MI.getOperand(2).getImm());
917 case RISCV::ZEXT_H_RV32:
918 case RISCV::ZEXT_H_RV64:
921 return Subtarget.hasStdExtZcb();
925 return Subtarget.hasStdExtZcb() &&
MI.getOperand(2).isReg() &&
926 MI.getOperand(2).getReg() == RISCV::X0;
930 return Subtarget.hasStdExtZcb() &&
MI.getOperand(2).isImm() &&
931 MI.getOperand(2).getImm() == -1;
944 return PhysReg && RISCV::GPRCRegClass.contains(PhysReg);
947 for (
auto &MO :
MRI->reg_nodbg_operands(VirtReg)) {
949 unsigned OpIdx = MO.getOperandNo();
951 if (isCompressible(
MI, NeedGPRC)) {
952 if (
OpIdx == 0 &&
MI.getOperand(1).isReg()) {
953 if (!NeedGPRC ||
MI.getNumExplicitOperands() < 3 ||
954 MI.getOpcode() == RISCV::ADD_UW ||
955 isCompressibleOpnd(
MI.getOperand(2)))
956 tryAddHint(MO,
MI.getOperand(1), NeedGPRC);
957 if (
MI.isCommutable() &&
MI.getOperand(2).isReg() &&
958 (!NeedGPRC || isCompressibleOpnd(
MI.getOperand(1))))
959 tryAddHint(MO,
MI.getOperand(2), NeedGPRC);
960 }
else if (
OpIdx == 1 && (!NeedGPRC ||
MI.getNumExplicitOperands() < 3 ||
961 isCompressibleOpnd(
MI.getOperand(2)))) {
962 tryAddHint(MO,
MI.getOperand(0), NeedGPRC);
963 }
else if (
MI.isCommutable() &&
OpIdx == 2 &&
964 (!NeedGPRC || isCompressibleOpnd(
MI.getOperand(1)))) {
965 tryAddHint(MO,
MI.getOperand(0), NeedGPRC);
972 if ((
MI.getOpcode() == RISCV::ADDIW ||
MI.getOpcode() == RISCV::ADDI) &&
973 MI.getOperand(1).isReg()) {
977 if (
I !=
MBB.begin()) {
979 if ((
I->getOpcode() == RISCV::LUI ||
I->getOpcode() == RISCV::AUIPC) &&
980 I->getOperand(0).getReg() ==
MI.getOperand(1).getReg()) {
982 tryAddHint(MO,
MI.getOperand(1),
false);
984 tryAddHint(MO,
MI.getOperand(0),
false);
991 if (TwoAddrHints.
count(OrderReg))
994 return BaseImplRetVal;
1003 return getMatchingSuperReg(Reg, RISCV::sub_vrm1_0, &RegClass);
unsigned const MachineRegisterInfo * MRI
assert(UImm &&(UImm !=~static_cast< T >(0)) &&"Invalid immediate!")
MachineBasicBlock MachineBasicBlock::iterator DebugLoc DL
MachineBasicBlock MachineBasicBlock::iterator MBBI
This file contains constants used for implementing Dwarf debug support.
const HexagonInstrInfo * TII
const AbstractManglingParser< Derived, Alloc >::OperatorInfo AbstractManglingParser< Derived, Alloc >::Ops[]
Register const TargetRegisterInfo * TRI
Promote Memory to Register
MachineInstr unsigned OpIdx
uint64_t IntrinsicInst * II
static cl::opt< bool > DisableRegAllocHints("riscv-disable-regalloc-hints", cl::Hidden, cl::init(false), cl::desc("Disable two address hints for register " "allocation"))
static cl::opt< bool > DisableCostPerUse("riscv-disable-cost-per-use", cl::init(false), cl::Hidden)
static std::tuple< RISCVVType::VLMUL, const TargetRegisterClass &, unsigned > getSpillReloadInfo(unsigned NumRemaining, uint16_t RegEncoding, bool IsSpill)
This file declares the machine register scavenger class.
static bool contains(SmallPtrSetImpl< ConstantExpr * > &Cache, ConstantExpr *Expr, Constant *C)
This file defines the SmallSet class.
static unsigned getDwarfRegNum(MCRegister Reg, const TargetRegisterInfo *TRI)
Go up the super-register chain until we hit a valid dwarf register number.
ArrayRef - Represent a constant reference to an array (0 or more elements consecutively in memory),...
static LLVM_ABI void appendOffset(SmallVectorImpl< uint64_t > &Ops, int64_t Offset)
Append Ops with operations to apply the Offset.
CallingConv::ID getCallingConv() const
getCallingConv()/setCallingConv(CC) - These method get and set the calling convention of this functio...
bool hasFnAttribute(Attribute::AttrKind Kind) const
Return true if the function has the attribute.
TypeSize getValue() const
Wrapper class representing physical registers. Should be passed by value.
MachineInstrBundleIterator< const MachineInstr > const_iterator
MachineInstrBundleIterator< MachineInstr > iterator
The MachineFrameInfo class represents an abstract stack frame until prolog/epilog code is inserted.
int64_t getLocalFrameSize() const
Get the size of the local object blob.
const TargetSubtargetInfo & getSubtarget() const
getSubtarget - Return the subtarget for which this machine code is being compiled.
MachineMemOperand * getMachineMemOperand(MachinePointerInfo PtrInfo, MachineMemOperand::Flags f, LLT MemTy, Align base_alignment, const AAMDNodes &AAInfo=AAMDNodes(), const MDNode *Ranges=nullptr, SyncScope::ID SSID=SyncScope::System, AtomicOrdering Ordering=AtomicOrdering::NotAtomic, AtomicOrdering FailureOrdering=AtomicOrdering::NotAtomic)
getMachineMemOperand - Allocate a new MachineMemOperand.
MachineFrameInfo & getFrameInfo()
getFrameInfo - Return the frame info object for the current function.
MachineRegisterInfo & getRegInfo()
getRegInfo - Return information about the registers currently in use.
Function & getFunction()
Return the LLVM function that this machine code represents.
const MachineInstrBuilder & setMIFlag(MachineInstr::MIFlag Flag) const
const MachineInstrBuilder & addImm(int64_t Val) const
Add a new immediate operand.
const MachineInstrBuilder & addFrameIndex(int Idx) const
const MachineInstrBuilder & addReg(Register RegNo, unsigned flags=0, unsigned SubReg=0) const
Add a new virtual register operand.
const MachineInstrBuilder & addMemOperand(MachineMemOperand *MMO) const
Representation of each machine instruction.
MachineOperand class - Representation of each machine instruction operand.
unsigned getSubReg() const
MachineRegisterInfo - Keep track of information for virtual and physical registers,...
LLVM_ABI Register createVirtualRegister(const TargetRegisterClass *RegClass, StringRef Name="")
createVirtualRegister - Create and return a new virtual register in the function with the specified r...
bool hasBP(const MachineFunction &MF) const
void movImm(MachineBasicBlock &MBB, MachineBasicBlock::iterator MBBI, const DebugLoc &DL, Register DstReg, uint64_t Val, MachineInstr::MIFlag Flag=MachineInstr::NoFlags, bool DstRenamable=false, bool DstIsDead=false) const
std::optional< unsigned > getRealVLen() const
const RISCVRegisterInfo * getRegisterInfo() const override
const RISCVInstrInfo * getInstrInfo() const override
Wrapper class representing virtual and physical registers.
constexpr bool isPhysical() const
Return true if the specified register number is in the physical register namespace.
SmallSet - This maintains a set of unique values, optimizing for the case when the set is small (less...
size_type count(const T &V) const
count - Return 1 if the element is in the set, 0 otherwise.
std::pair< const_iterator, bool > insert(const T &V)
insert - Insert an element into the set if it isn't already there.
This class consists of common code factored out of the SmallVector class to reduce code duplication b...
void push_back(const T &Elt)
StackOffset holds a fixed and a scalable offset in bytes.
int64_t getFixed() const
Returns the fixed component of the stack.
static StackOffset get(int64_t Fixed, int64_t Scalable)
StringRef - Represent a constant reference to a string, i.e.
Information about stack frame layout on the target.
bool hasFP(const MachineFunction &MF) const
hasFP - Return true if the specified function should have a dedicated frame pointer register.
TargetInstrInfo - Interface to description of machine instruction set.
const uint8_t TSFlags
Configurable target specific flags.
TargetRegisterInfo base class - We assume that the target defines a static array of TargetRegisterDes...
virtual StringRef getRegAsmName(MCRegister Reg) const
Return the assembly name for Reg.
virtual bool getRegAllocationHints(Register VirtReg, ArrayRef< MCPhysReg > Order, SmallVectorImpl< MCPhysReg > &Hints, const MachineFunction &MF, const VirtRegMap *VRM=nullptr, const LiveRegMatrix *Matrix=nullptr) const
Get a list of 'hint' registers that the register allocator should try first when allocating a physica...
virtual bool isRegisterReservedByUser(Register R) const
virtual const TargetInstrInfo * getInstrInfo() const
MCRegister getPhys(Register virtReg) const
returns the physical register mapped to the specified virtual register
constexpr bool isKnownMultipleOf(ScalarTy RHS) const
This function tells the caller whether the element count is known at compile time to be a multiple of...
constexpr LeafTy divideCoefficientBy(ScalarTy RHS) const
We do not provide the '/' operator here because division for polynomial types does not work in the sa...
#define llvm_unreachable(msg)
Marks that the current location is not supposed to be reachable.
unsigned ID
LLVM IR allows to use arbitrary numbers as calling convention identifiers.
@ RISCV_VectorCall
Calling convention used for RISC-V V-extension.
@ PreserveMost
Used for runtime calls that preserves most registers.
@ GHC
Used by the Glasgow Haskell Compiler (GHC).
@ GRAAL
Used by GraalVM. Two additional registers are reserved.
static unsigned getFormat(uint64_t TSFlags)
static RISCVVType::VLMUL getLMul(uint8_t TSFlags)
LLVM_ABI std::pair< unsigned, bool > decodeVLMUL(VLMUL VLMul)
std::optional< std::pair< unsigned, unsigned > > isRVVSpillForZvlsseg(unsigned Opcode)
bool isRVVSpill(const MachineInstr &MI)
static constexpr unsigned RVVBytesPerBlock
@ Implicit
Not emitted register (e.g. carry, or temporary result).
@ Kill
The last use of a register.
initializer< Ty > init(const Ty &Val)
This is an optimization pass for GlobalISel generic memory operations.
MachineInstrBuilder BuildMI(MachineFunction &MF, const MIMetadata &MIMD, const MCInstrDesc &MCID)
Builder interface. Specify how to create the initial instruction itself.
constexpr bool isInt(int64_t x)
Checks if an integer fits into the given bit width.
unsigned Log2_32(uint32_t Value)
Return the floor log base 2 of the specified value, -1 if the value is zero.
constexpr bool isUInt(uint64_t x)
Checks if an unsigned integer fits into the given bit width.
IterT skipDebugInstructionsBackward(IterT It, IterT Begin, bool SkipPseudoOp=true)
Decrement It until it points to a non-debug instruction or to Begin and return the resulting iterator...
unsigned getDefRegState(bool B)
unsigned getKillRegState(bool B)
uint16_t MCPhysReg
An unsigned integer type large enough to represent all physical registers, but not necessarily virtua...
constexpr bool isShiftedInt(int64_t x)
Checks if a signed integer is an N bit number shifted left by S.
bool is_contained(R &&Range, const E &Element)
Returns true if Element is found in Range.
constexpr int64_t SignExtend64(uint64_t x)
Sign-extend the number in the bottom B bits of X to a 64-bit integer.
LLVM_ABI void reportFatalUsageError(Error Err)
Report a fatal error that does not indicate a bug in LLVM.
This struct is a compact representation of a valid (non-zero power of two) alignment.
uint64_t value() const
This is a hole in the type system and should not be abused.
This struct is a compact representation of a valid (power of two) or undefined (0) alignment.
Align valueOrOne() const
For convenience, returns a valid alignment or 1 if undefined.
bool needsFrameBaseReg(MachineInstr *MI, int64_t Offset) const override
bool requiresVirtualBaseRegisters(const MachineFunction &MF) const override
Register findVRegWithEncoding(const TargetRegisterClass &RegClass, uint16_t Encoding) const
const TargetRegisterClass * getLargestLegalSuperClass(const TargetRegisterClass *RC, const MachineFunction &) const override
const MCPhysReg * getCalleeSavedRegs(const MachineFunction *MF) const override
BitVector getReservedRegs(const MachineFunction &MF) const override
const uint32_t * getCallPreservedMask(const MachineFunction &MF, CallingConv::ID) const override
Register materializeFrameBaseRegister(MachineBasicBlock *MBB, int FrameIdx, int64_t Offset) const override
RISCVRegisterInfo(unsigned HwMode)
void getOffsetOpcodes(const StackOffset &Offset, SmallVectorImpl< uint64_t > &Ops) const override
bool isFrameOffsetLegal(const MachineInstr *MI, Register BaseReg, int64_t Offset) const override
Register getFrameRegister(const MachineFunction &MF) const override
const MCPhysReg * getIPRACSRegs(const MachineFunction *MF) const override
void lowerSegmentSpillReload(MachineBasicBlock::iterator II, bool IsSpill) const
void adjustReg(MachineBasicBlock &MBB, MachineBasicBlock::iterator II, const DebugLoc &DL, Register DestReg, Register SrcReg, StackOffset Offset, MachineInstr::MIFlag Flag, MaybeAlign RequiredAlign) const
bool isAsmClobberable(const MachineFunction &MF, MCRegister PhysReg) const override
const uint32_t * getNoPreservedMask() const override
float getSpillWeightScaleFactor(const TargetRegisterClass *RC) const override
void resolveFrameIndex(MachineInstr &MI, Register BaseReg, int64_t Offset) const override
bool getRegAllocationHints(Register VirtReg, ArrayRef< MCPhysReg > Order, SmallVectorImpl< MCPhysReg > &Hints, const MachineFunction &MF, const VirtRegMap *VRM, const LiveRegMatrix *Matrix) const override
int64_t getFrameIndexInstrOffset(const MachineInstr *MI, int Idx) const override
unsigned getRegisterCostTableIndex(const MachineFunction &MF) const override
StringRef getRegAsmName(MCRegister Reg) const override
bool eliminateFrameIndex(MachineBasicBlock::iterator MI, int SPAdj, unsigned FIOperandNum, RegScavenger *RS=nullptr) const override