41#define GEN_CHECK_COMPRESS_INSTR
42#include "RISCVGenCompressInstEmitter.inc"
44#define GET_INSTRINFO_CTOR_DTOR
45#include "RISCVGenInstrInfo.inc"
47#define DEBUG_TYPE "riscv-instr-info"
49 "Number of registers within vector register groups spilled");
51 "Number of registers within vector register groups reloaded");
55 cl::desc(
"Prefer whole register move for vector registers."));
58 "riscv-force-machine-combiner-strategy",
cl::Hidden,
59 cl::desc(
"Force machine combiner to use a specific strategy for machine "
60 "trace metrics evaluation."),
65 "MinInstrCount strategy.")));
69 cl::desc(
"Enable RegSave strategy in machine outliner (save X5 to a "
70 "temporary register when X5 is live across outlined calls)."));
76#define GET_RISCVVPseudosTable_IMPL
77#include "RISCVGenSearchableTables.inc"
83#define GET_RISCVMaskedPseudosTable_IMPL
84#include "RISCVGenSearchableTables.inc"
90 RISCV::ADJCALLSTACKUP),
93#define GET_INSTRINFO_HELPERS
94#include "RISCVGenInstrInfo.inc"
97 if (
STI.hasStdExtZca())
106 int &FrameIndex)
const {
116 case RISCV::VL1RE8_V:
117 case RISCV::VL1RE16_V:
118 case RISCV::VL1RE32_V:
119 case RISCV::VL1RE64_V:
122 case RISCV::VL2RE8_V:
123 case RISCV::VL2RE16_V:
124 case RISCV::VL2RE32_V:
125 case RISCV::VL2RE64_V:
128 case RISCV::VL4RE8_V:
129 case RISCV::VL4RE16_V:
130 case RISCV::VL4RE32_V:
131 case RISCV::VL4RE64_V:
134 case RISCV::VL8RE8_V:
135 case RISCV::VL8RE16_V:
136 case RISCV::VL8RE32_V:
137 case RISCV::VL8RE64_V:
145 switch (
MI.getOpcode()) {
169 case RISCV::VL1RE8_V:
170 case RISCV::VL2RE8_V:
171 case RISCV::VL4RE8_V:
172 case RISCV::VL8RE8_V:
173 if (!
MI.getOperand(1).isFI())
175 FrameIndex =
MI.getOperand(1).getIndex();
178 return MI.getOperand(0).getReg();
181 if (
MI.getOperand(1).isFI() &&
MI.getOperand(2).isImm() &&
182 MI.getOperand(2).getImm() == 0) {
183 FrameIndex =
MI.getOperand(1).getIndex();
184 return MI.getOperand(0).getReg();
191 int &FrameIndex)
const {
199 switch (
MI.getOpcode()) {
224 if (!
MI.getOperand(1).isFI())
226 FrameIndex =
MI.getOperand(1).getIndex();
229 return MI.getOperand(0).getReg();
232 if (
MI.getOperand(1).isFI() &&
MI.getOperand(2).isImm() &&
233 MI.getOperand(2).getImm() == 0) {
234 FrameIndex =
MI.getOperand(1).getIndex();
235 return MI.getOperand(0).getReg();
245 case RISCV::VFMV_V_F:
248 case RISCV::VFMV_S_F:
250 return MI.getOperand(1).isUndef();
258 return DstReg > SrcReg && (DstReg - SrcReg) < NumRegs;
269 assert(
MBBI->getOpcode() == TargetOpcode::COPY &&
270 "Unexpected COPY instruction.");
274 bool FoundDef =
false;
275 bool FirstVSetVLI =
false;
276 unsigned FirstSEW = 0;
279 if (
MBBI->isMetaInstruction())
282 if (RISCVInstrInfo::isVectorConfigInstr(*
MBBI)) {
292 unsigned FirstVType =
MBBI->getOperand(2).getImm();
297 if (FirstLMul != LMul)
302 if (!RISCVInstrInfo::isVLPreservingConfig(*
MBBI))
308 unsigned VType =
MBBI->getOperand(2).getImm();
326 }
else if (
MBBI->isInlineAsm() ||
MBBI->isCall()) {
328 }
else if (
MBBI->getNumDefs()) {
331 if (
MBBI->modifiesRegister(RISCV::VL,
nullptr))
337 if (!MO.isReg() || !MO.isDef())
339 if (!FoundDef &&
TRI->regsOverlap(MO.getReg(), SrcReg)) {
354 if (MO.getReg() != SrcReg)
395 uint16_t SrcEncoding =
TRI->getEncodingValue(SrcReg);
396 uint16_t DstEncoding =
TRI->getEncodingValue(DstReg);
398 assert(!Fractional &&
"It is impossible be fractional lmul here.");
399 unsigned NumRegs = NF * LMulVal;
405 SrcEncoding += NumRegs - 1;
406 DstEncoding += NumRegs - 1;
412 unsigned,
unsigned> {
420 uint16_t Diff = DstEncoding - SrcEncoding;
421 if (
I + 8 <= NumRegs && Diff >= 8 && SrcEncoding % 8 == 7 &&
422 DstEncoding % 8 == 7)
424 RISCV::PseudoVMV_V_V_M8, RISCV::PseudoVMV_V_I_M8};
425 if (
I + 4 <= NumRegs && Diff >= 4 && SrcEncoding % 4 == 3 &&
426 DstEncoding % 4 == 3)
428 RISCV::PseudoVMV_V_V_M4, RISCV::PseudoVMV_V_I_M4};
429 if (
I + 2 <= NumRegs && Diff >= 2 && SrcEncoding % 2 == 1 &&
430 DstEncoding % 2 == 1)
432 RISCV::PseudoVMV_V_V_M2, RISCV::PseudoVMV_V_I_M2};
435 RISCV::PseudoVMV_V_V_M1, RISCV::PseudoVMV_V_I_M1};
440 if (
I + 8 <= NumRegs && SrcEncoding % 8 == 0 && DstEncoding % 8 == 0)
442 RISCV::PseudoVMV_V_V_M8, RISCV::PseudoVMV_V_I_M8};
443 if (
I + 4 <= NumRegs && SrcEncoding % 4 == 0 && DstEncoding % 4 == 0)
445 RISCV::PseudoVMV_V_V_M4, RISCV::PseudoVMV_V_I_M4};
446 if (
I + 2 <= NumRegs && SrcEncoding % 2 == 0 && DstEncoding % 2 == 0)
448 RISCV::PseudoVMV_V_V_M2, RISCV::PseudoVMV_V_I_M2};
451 RISCV::PseudoVMV_V_V_M1, RISCV::PseudoVMV_V_I_M1};
454 while (
I != NumRegs) {
459 auto [LMulCopied, RegClass,
Opc, VVOpc, VIOpc] =
460 GetCopyInfo(SrcEncoding, DstEncoding);
464 if (LMul == LMulCopied &&
467 if (DefMBBI->getOpcode() == VIOpc)
474 RegClass, ReversedCopy ? (SrcEncoding - NumCopied + 1) : SrcEncoding);
476 RegClass, ReversedCopy ? (DstEncoding - NumCopied + 1) : DstEncoding);
484 MIB = MIB.add(DefMBBI->getOperand(2));
492 MIB.addImm(Log2SEW ? Log2SEW : 3);
504 SrcEncoding += (ReversedCopy ? -NumCopied : NumCopied);
505 DstEncoding += (ReversedCopy ? -NumCopied : NumCopied);
514 bool RenamableDest,
bool RenamableSrc)
const {
518 if (RISCV::GPRRegClass.
contains(DstReg, SrcReg)) {
525 if (RISCV::GPRF16RegClass.
contains(DstReg, SrcReg)) {
531 if (RISCV::GPRF32RegClass.
contains(DstReg, SrcReg)) {
537 if (RISCV::GPRPairRegClass.
contains(DstReg, SrcReg)) {
539 if (
STI.hasStdExtZdinx()) {
548 if (
STI.hasStdExtP()) {
557 MCRegister EvenReg =
TRI->getSubReg(SrcReg, RISCV::sub_gpr_even);
558 MCRegister OddReg =
TRI->getSubReg(SrcReg, RISCV::sub_gpr_odd);
560 if (OddReg == RISCV::DUMMY_REG_PAIR_WITH_X0)
562 assert(DstReg != RISCV::X0_Pair &&
"Cannot write to X0_Pair");
566 TRI->getSubReg(DstReg, RISCV::sub_gpr_even))
567 .
addReg(EvenReg, KillFlag)
570 TRI->getSubReg(DstReg, RISCV::sub_gpr_odd))
577 if (RISCV::VCSRRegClass.
contains(SrcReg) &&
578 RISCV::GPRRegClass.
contains(DstReg)) {
580 .
addImm(RISCVSysReg::lookupSysRegByName(
TRI->getName(SrcReg))->Encoding)
585 if (RISCV::FPR16RegClass.
contains(DstReg, SrcReg)) {
587 if (
STI.hasStdExtZfh()) {
588 Opc = RISCV::FSGNJ_H;
591 (
STI.hasStdExtZfhmin() ||
STI.hasStdExtZfbfmin()) &&
592 "Unexpected extensions");
594 DstReg =
TRI->getMatchingSuperReg(DstReg, RISCV::sub_16,
595 &RISCV::FPR32RegClass);
596 SrcReg =
TRI->getMatchingSuperReg(SrcReg, RISCV::sub_16,
597 &RISCV::FPR32RegClass);
598 Opc = RISCV::FSGNJ_S;
602 .
addReg(SrcReg, KillFlag);
606 if (RISCV::FPR32RegClass.
contains(DstReg, SrcReg)) {
609 .
addReg(SrcReg, KillFlag);
613 if (RISCV::FPR64RegClass.
contains(DstReg, SrcReg)) {
616 .
addReg(SrcReg, KillFlag);
620 if (RISCV::FPR32RegClass.
contains(DstReg) &&
621 RISCV::GPRRegClass.
contains(SrcReg)) {
623 .
addReg(SrcReg, KillFlag);
627 if (RISCV::GPRRegClass.
contains(DstReg) &&
628 RISCV::FPR32RegClass.
contains(SrcReg)) {
630 .
addReg(SrcReg, KillFlag);
634 if (RISCV::FPR64RegClass.
contains(DstReg) &&
635 RISCV::GPRRegClass.
contains(SrcReg)) {
636 assert(
STI.getXLen() == 64 &&
"Unexpected GPR size");
638 .
addReg(SrcReg, KillFlag);
642 if (RISCV::GPRRegClass.
contains(DstReg) &&
643 RISCV::FPR64RegClass.
contains(SrcReg)) {
644 assert(
STI.getXLen() == 64 &&
"Unexpected GPR size");
646 .
addReg(SrcReg, KillFlag);
652 TRI->getCommonMinimalPhysRegClass(SrcReg, DstReg);
663 Register SrcReg,
bool IsKill,
int FI,
672 if (RISCV::GPRRegClass.hasSubClassEq(RC)) {
673 Opcode = RegInfo.getRegSizeInBits(RISCV::GPRRegClass) == 32 ? RISCV::SW
675 }
else if (RISCV::GPRF16RegClass.hasSubClassEq(RC)) {
676 Opcode = RISCV::SH_INX;
677 }
else if (RISCV::GPRF32RegClass.hasSubClassEq(RC)) {
678 Opcode = RISCV::SW_INX;
679 }
else if (RISCV::GPRPairRegClass.hasSubClassEq(RC)) {
680 if (!
STI.is64Bit() &&
STI.hasStdExtZilsd() &&
681 Alignment >=
STI.getZilsdAlign()) {
682 Opcode = RISCV::SD_RV32;
684 Opcode = RISCV::PseudoRV32ZdinxSD;
686 }
else if (RISCV::FPR16RegClass.hasSubClassEq(RC)) {
688 }
else if (RISCV::FPR32RegClass.hasSubClassEq(RC)) {
690 }
else if (RISCV::FPR64RegClass.hasSubClassEq(RC)) {
692 }
else if (RISCV::VRRegClass.hasSubClassEq(RC)) {
693 Opcode = RISCV::VS1R_V;
694 }
else if (RISCV::VRM2RegClass.hasSubClassEq(RC)) {
695 Opcode = RISCV::VS2R_V;
696 }
else if (RISCV::VRM4RegClass.hasSubClassEq(RC)) {
697 Opcode = RISCV::VS4R_V;
698 }
else if (RISCV::VRM8RegClass.hasSubClassEq(RC)) {
699 Opcode = RISCV::VS8R_V;
700 }
else if (RISCV::VRN2M1RegClass.hasSubClassEq(RC))
701 Opcode = RISCV::PseudoVSPILL2_M1;
702 else if (RISCV::VRN2M2RegClass.hasSubClassEq(RC))
703 Opcode = RISCV::PseudoVSPILL2_M2;
704 else if (RISCV::VRN2M4RegClass.hasSubClassEq(RC))
705 Opcode = RISCV::PseudoVSPILL2_M4;
706 else if (RISCV::VRN3M1RegClass.hasSubClassEq(RC))
707 Opcode = RISCV::PseudoVSPILL3_M1;
708 else if (RISCV::VRN3M2RegClass.hasSubClassEq(RC))
709 Opcode = RISCV::PseudoVSPILL3_M2;
710 else if (RISCV::VRN4M1RegClass.hasSubClassEq(RC))
711 Opcode = RISCV::PseudoVSPILL4_M1;
712 else if (RISCV::VRN4M2RegClass.hasSubClassEq(RC))
713 Opcode = RISCV::PseudoVSPILL4_M2;
714 else if (RISCV::VRN5M1RegClass.hasSubClassEq(RC))
715 Opcode = RISCV::PseudoVSPILL5_M1;
716 else if (RISCV::VRN6M1RegClass.hasSubClassEq(RC))
717 Opcode = RISCV::PseudoVSPILL6_M1;
718 else if (RISCV::VRN7M1RegClass.hasSubClassEq(RC))
719 Opcode = RISCV::PseudoVSPILL7_M1;
720 else if (RISCV::VRN8M1RegClass.hasSubClassEq(RC))
721 Opcode = RISCV::PseudoVSPILL8_M1;
764 if (RISCV::GPRRegClass.hasSubClassEq(RC)) {
765 Opcode = RegInfo.getRegSizeInBits(RISCV::GPRRegClass) == 32 ? RISCV::LW
767 }
else if (RISCV::GPRF16RegClass.hasSubClassEq(RC)) {
768 Opcode = RISCV::LH_INX;
769 }
else if (RISCV::GPRF32RegClass.hasSubClassEq(RC)) {
770 Opcode = RISCV::LW_INX;
771 }
else if (RISCV::GPRPairRegClass.hasSubClassEq(RC)) {
772 if (!
STI.is64Bit() &&
STI.hasStdExtZilsd() &&
773 Alignment >=
STI.getZilsdAlign()) {
774 Opcode = RISCV::LD_RV32;
776 Opcode = RISCV::PseudoRV32ZdinxLD;
778 }
else if (RISCV::FPR16RegClass.hasSubClassEq(RC)) {
780 }
else if (RISCV::FPR32RegClass.hasSubClassEq(RC)) {
782 }
else if (RISCV::FPR64RegClass.hasSubClassEq(RC)) {
784 }
else if (RISCV::VRRegClass.hasSubClassEq(RC)) {
785 Opcode = RISCV::VL1RE8_V;
786 }
else if (RISCV::VRM2RegClass.hasSubClassEq(RC)) {
787 Opcode = RISCV::VL2RE8_V;
788 }
else if (RISCV::VRM4RegClass.hasSubClassEq(RC)) {
789 Opcode = RISCV::VL4RE8_V;
790 }
else if (RISCV::VRM8RegClass.hasSubClassEq(RC)) {
791 Opcode = RISCV::VL8RE8_V;
792 }
else if (RISCV::VRN2M1RegClass.hasSubClassEq(RC))
793 Opcode = RISCV::PseudoVRELOAD2_M1;
794 else if (RISCV::VRN2M2RegClass.hasSubClassEq(RC))
795 Opcode = RISCV::PseudoVRELOAD2_M2;
796 else if (RISCV::VRN2M4RegClass.hasSubClassEq(RC))
797 Opcode = RISCV::PseudoVRELOAD2_M4;
798 else if (RISCV::VRN3M1RegClass.hasSubClassEq(RC))
799 Opcode = RISCV::PseudoVRELOAD3_M1;
800 else if (RISCV::VRN3M2RegClass.hasSubClassEq(RC))
801 Opcode = RISCV::PseudoVRELOAD3_M2;
802 else if (RISCV::VRN4M1RegClass.hasSubClassEq(RC))
803 Opcode = RISCV::PseudoVRELOAD4_M1;
804 else if (RISCV::VRN4M2RegClass.hasSubClassEq(RC))
805 Opcode = RISCV::PseudoVRELOAD4_M2;
806 else if (RISCV::VRN5M1RegClass.hasSubClassEq(RC))
807 Opcode = RISCV::PseudoVRELOAD5_M1;
808 else if (RISCV::VRN6M1RegClass.hasSubClassEq(RC))
809 Opcode = RISCV::PseudoVRELOAD6_M1;
810 else if (RISCV::VRN7M1RegClass.hasSubClassEq(RC))
811 Opcode = RISCV::PseudoVRELOAD7_M1;
812 else if (RISCV::VRN8M1RegClass.hasSubClassEq(RC))
813 Opcode = RISCV::PseudoVRELOAD8_M1;
851 if (
Ops.size() != 1 ||
Ops[0] != 1)
854 switch (
MI.getOpcode()) {
856 if (RISCVInstrInfo::isSEXT_W(
MI))
858 if (RISCVInstrInfo::isZEXT_W(
MI))
860 if (RISCVInstrInfo::isZEXT_B(
MI))
867 case RISCV::ZEXT_H_RV32:
868 case RISCV::ZEXT_H_RV64:
875 case RISCV::VMV_X_S: {
878 if (ST.getXLen() < (1U << Log2SEW))
893 case RISCV::VFMV_F_S: {
920 return BuildMI(*
MI.getParent(), InsertPt,
MI.getDebugLoc(),
get(*LoadOpc),
929 return RISCV::PseudoCCLB;
931 return RISCV::PseudoCCLBU;
933 return RISCV::PseudoCCLH;
935 return RISCV::PseudoCCLHU;
937 return RISCV::PseudoCCLW;
939 return RISCV::PseudoCCLWU;
941 return RISCV::PseudoCCLD;
943 return RISCV::PseudoCCQC_E_LB;
944 case RISCV::QC_E_LBU:
945 return RISCV::PseudoCCQC_E_LBU;
947 return RISCV::PseudoCCQC_E_LH;
948 case RISCV::QC_E_LHU:
949 return RISCV::PseudoCCQC_E_LHU;
951 return RISCV::PseudoCCQC_E_LW;
962 if (
MI.getOpcode() != RISCV::PseudoCCMOVGPR)
967 if (!
STI.hasShortForwardBranchILoad() || !PredOpc)
971 if (
Ops.size() != 1 || (
Ops[0] != 1 &&
Ops[0] != 2))
974 bool Invert =
Ops[0] == 2;
983 MI.getDebugLoc(),
get(PredOpc), DestReg);
994 unsigned BCC =
MI.getOperand(
MI.getNumExplicitOperands() - 3).getImm();
1000 NewMI.
add({
MI.getOperand(
MI.getNumExplicitOperands() - 2),
1001 MI.getOperand(
MI.getNumExplicitOperands() - 1)});
1010 bool DstIsDead)
const {
1026 bool SrcRenamable =
false;
1030 bool LastItem = ++Num == Seq.
size();
1035 switch (Inst.getOpndKind()) {
1045 .
addReg(SrcReg, SrcRegState)
1052 .
addReg(SrcReg, SrcRegState)
1053 .
addReg(SrcReg, SrcRegState)
1059 .
addReg(SrcReg, SrcRegState)
1067 SrcRenamable = DstRenamable;
1077 case RISCV::CV_BEQIMM:
1078 case RISCV::QC_BEQI:
1079 case RISCV::QC_E_BEQI:
1080 case RISCV::NDS_BBC:
1081 case RISCV::NDS_BEQC:
1085 case RISCV::QC_BNEI:
1086 case RISCV::QC_E_BNEI:
1087 case RISCV::CV_BNEIMM:
1088 case RISCV::NDS_BBS:
1089 case RISCV::NDS_BNEC:
1092 case RISCV::QC_BLTI:
1093 case RISCV::QC_E_BLTI:
1096 case RISCV::QC_BGEI:
1097 case RISCV::QC_E_BGEI:
1100 case RISCV::QC_BLTUI:
1101 case RISCV::QC_E_BLTUI:
1104 case RISCV::QC_BGEUI:
1105 case RISCV::QC_E_BGEUI:
1137 "Unknown conditional branch");
1148 case RISCV::QC_MVEQ:
1149 return RISCV::QC_MVNE;
1150 case RISCV::QC_MVNE:
1151 return RISCV::QC_MVEQ;
1152 case RISCV::QC_MVLT:
1153 return RISCV::QC_MVGE;
1154 case RISCV::QC_MVGE:
1155 return RISCV::QC_MVLT;
1156 case RISCV::QC_MVLTU:
1157 return RISCV::QC_MVGEU;
1158 case RISCV::QC_MVGEU:
1159 return RISCV::QC_MVLTU;
1160 case RISCV::QC_MVEQI:
1161 return RISCV::QC_MVNEI;
1162 case RISCV::QC_MVNEI:
1163 return RISCV::QC_MVEQI;
1164 case RISCV::QC_MVLTI:
1165 return RISCV::QC_MVGEI;
1166 case RISCV::QC_MVGEI:
1167 return RISCV::QC_MVLTI;
1168 case RISCV::QC_MVLTUI:
1169 return RISCV::QC_MVGEUI;
1170 case RISCV::QC_MVGEUI:
1171 return RISCV::QC_MVLTUI;
1176 switch (SelectOpc) {
1195 case RISCV::Select_GPR_Using_CC_Imm5_Zibi:
1205 case RISCV::Select_GPR_Using_CC_SImm5_CV:
1210 return RISCV::CV_BEQIMM;
1212 return RISCV::CV_BNEIMM;
1215 case RISCV::Select_GPRNoX0_Using_CC_SImm5NonZero_QC:
1220 return RISCV::QC_BEQI;
1222 return RISCV::QC_BNEI;
1224 return RISCV::QC_BLTI;
1226 return RISCV::QC_BGEI;
1229 case RISCV::Select_GPRNoX0_Using_CC_UImm5NonZero_QC:
1234 return RISCV::QC_BLTUI;
1236 return RISCV::QC_BGEUI;
1239 case RISCV::Select_GPRNoX0_Using_CC_SImm16NonZero_QC:
1244 return RISCV::QC_E_BEQI;
1246 return RISCV::QC_E_BNEI;
1248 return RISCV::QC_E_BLTI;
1250 return RISCV::QC_E_BGEI;
1253 case RISCV::Select_GPRNoX0_Using_CC_UImm16NonZero_QC:
1258 return RISCV::QC_E_BLTUI;
1260 return RISCV::QC_E_BGEUI;
1263 case RISCV::Select_GPR_Using_CC_UImmLog2XLen_NDS:
1268 return RISCV::NDS_BBC;
1270 return RISCV::NDS_BBS;
1273 case RISCV::Select_GPR_Using_CC_UImm7_NDS:
1278 return RISCV::NDS_BEQC;
1280 return RISCV::NDS_BNEC;
1326 case RISCV::CV_BEQIMM:
1327 return RISCV::CV_BNEIMM;
1328 case RISCV::CV_BNEIMM:
1329 return RISCV::CV_BEQIMM;
1330 case RISCV::QC_BEQI:
1331 return RISCV::QC_BNEI;
1332 case RISCV::QC_BNEI:
1333 return RISCV::QC_BEQI;
1334 case RISCV::QC_BLTI:
1335 return RISCV::QC_BGEI;
1336 case RISCV::QC_BGEI:
1337 return RISCV::QC_BLTI;
1338 case RISCV::QC_BLTUI:
1339 return RISCV::QC_BGEUI;
1340 case RISCV::QC_BGEUI:
1341 return RISCV::QC_BLTUI;
1342 case RISCV::QC_E_BEQI:
1343 return RISCV::QC_E_BNEI;
1344 case RISCV::QC_E_BNEI:
1345 return RISCV::QC_E_BEQI;
1346 case RISCV::QC_E_BLTI:
1347 return RISCV::QC_E_BGEI;
1348 case RISCV::QC_E_BGEI:
1349 return RISCV::QC_E_BLTI;
1350 case RISCV::QC_E_BLTUI:
1351 return RISCV::QC_E_BGEUI;
1352 case RISCV::QC_E_BGEUI:
1353 return RISCV::QC_E_BLTUI;
1354 case RISCV::NDS_BBC:
1355 return RISCV::NDS_BBS;
1356 case RISCV::NDS_BBS:
1357 return RISCV::NDS_BBC;
1358 case RISCV::NDS_BEQC:
1359 return RISCV::NDS_BNEC;
1360 case RISCV::NDS_BNEC:
1361 return RISCV::NDS_BEQC;
1369 bool AllowModify)
const {
1370 TBB = FBB =
nullptr;
1375 if (
I ==
MBB.end() || !isUnpredicatedTerminator(*
I))
1381 int NumTerminators = 0;
1382 for (
auto J =
I.getReverse(); J !=
MBB.rend() && isUnpredicatedTerminator(*J);
1385 if (J->getDesc().isUnconditionalBranch() ||
1386 J->getDesc().isIndirectBranch()) {
1393 if (AllowModify && FirstUncondOrIndirectBr !=
MBB.end()) {
1394 while (std::next(FirstUncondOrIndirectBr) !=
MBB.end()) {
1395 std::next(FirstUncondOrIndirectBr)->eraseFromParent();
1398 I = FirstUncondOrIndirectBr;
1402 if (
I->getDesc().isIndirectBranch())
1406 if (
I->isPreISelOpcode())
1410 if (NumTerminators > 2)
1414 if (NumTerminators == 1 &&
I->getDesc().isUnconditionalBranch()) {
1420 if (NumTerminators == 1 &&
I->getDesc().isConditionalBranch()) {
1426 if (NumTerminators == 2 && std::prev(
I)->getDesc().isConditionalBranch() &&
1427 I->getDesc().isUnconditionalBranch()) {
1438 int *BytesRemoved)
const {
1445 if (!
I->getDesc().isUnconditionalBranch() &&
1446 !
I->getDesc().isConditionalBranch())
1452 I->eraseFromParent();
1456 if (
I ==
MBB.begin())
1459 if (!
I->getDesc().isConditionalBranch())
1465 I->eraseFromParent();
1478 assert(
TBB &&
"insertBranch must not be told to insert a fallthrough");
1480 "RISC-V branch conditions have two components!");
1514 assert(RS &&
"RegScavenger required for long branching");
1516 "new block should be inserted for expanding unconditional branch");
1519 "restore block should be inserted for restoring clobbered registers");
1528 "Branch offsets outside of the signed 32-bit range not supported");
1534 auto II =
MBB.end();
1540 RS->enterBasicBlockEnd(
MBB);
1547 RC = &RISCV::GPRX7RegClass;
1549 RS->scavengeRegisterBackwards(*RC,
MI.getIterator(),
1553 RS->setRegUsed(TmpGPR);
1558 TmpGPR =
STI.hasStdExtE() ? RISCV::X9 : RISCV::X27;
1564 if (FrameIndex == -1)
1569 TRI->eliminateFrameIndex(std::prev(
MI.getIterator()),
1572 MI.getOperand(1).setMBB(&RestoreBB);
1576 TRI->eliminateFrameIndex(RestoreBB.
back(),
1586 assert((
Cond.size() == 3) &&
"Invalid branch condition!");
1596 if (
MI->getOpcode() == RISCV::ADDI &&
MI->getOperand(1).isReg() &&
1597 MI->getOperand(1).getReg() == RISCV::X0) {
1598 Imm =
MI->getOperand(2).getImm();
1603 if (
MI->getOpcode() == RISCV::BSETI &&
MI->getOperand(1).isReg() &&
1604 MI->getOperand(1).getReg() == RISCV::X0 &&
1605 MI->getOperand(2).getImm() == 11) {
1619 if (Reg == RISCV::X0) {
1627 bool IsSigned =
false;
1628 bool IsEquality =
false;
1629 switch (
MI.getOpcode()) {
1665 MI.eraseFromParent();
1691 auto searchConst = [&](int64_t C1) ->
Register {
1693 auto DefC1 = std::find_if(++
II, E, [&](
const MachineInstr &
I) ->
bool {
1696 I.getOperand(0).getReg().isVirtual();
1699 return DefC1->getOperand(0).getReg();
1711 if (
isFromLoadImm(MRI, LHS, C0) && C0 != 0 && LHS.getReg().isVirtual() &&
1712 MRI.
hasOneUse(LHS.getReg()) && (IsSigned || C0 != -1)) {
1714 if (
Register RegZ = searchConst(C0 + 1)) {
1722 MI.eraseFromParent();
1732 if (
isFromLoadImm(MRI, RHS, C0) && C0 != 0 && RHS.getReg().isVirtual() &&
1735 if (
Register RegZ = searchConst(C0 - 1)) {
1743 MI.eraseFromParent();
1753 assert(
MI.getDesc().isBranch() &&
"Unexpected opcode!");
1755 int NumOp =
MI.getNumExplicitOperands();
1756 return MI.getOperand(NumOp - 1).getMBB();
1760 int64_t BrOffset)
const {
1761 unsigned XLen =
STI.getXLen();
1768 case RISCV::NDS_BBC:
1769 case RISCV::NDS_BBS:
1770 case RISCV::NDS_BEQC:
1771 case RISCV::NDS_BNEC:
1781 case RISCV::CV_BEQIMM:
1782 case RISCV::CV_BNEIMM:
1783 case RISCV::QC_BEQI:
1784 case RISCV::QC_BNEI:
1785 case RISCV::QC_BGEI:
1786 case RISCV::QC_BLTI:
1787 case RISCV::QC_BLTUI:
1788 case RISCV::QC_BGEUI:
1789 case RISCV::QC_E_BEQI:
1790 case RISCV::QC_E_BNEI:
1791 case RISCV::QC_E_BGEI:
1792 case RISCV::QC_E_BLTI:
1793 case RISCV::QC_E_BLTUI:
1794 case RISCV::QC_E_BGEUI:
1797 case RISCV::PseudoBR:
1799 case RISCV::PseudoJump:
1810 case RISCV::ADD:
return RISCV::PseudoCCADD;
1811 case RISCV::SUB:
return RISCV::PseudoCCSUB;
1812 case RISCV::SLL:
return RISCV::PseudoCCSLL;
1813 case RISCV::SRL:
return RISCV::PseudoCCSRL;
1814 case RISCV::SRA:
return RISCV::PseudoCCSRA;
1815 case RISCV::AND:
return RISCV::PseudoCCAND;
1816 case RISCV::OR:
return RISCV::PseudoCCOR;
1817 case RISCV::XOR:
return RISCV::PseudoCCXOR;
1818 case RISCV::MAX:
return RISCV::PseudoCCMAX;
1819 case RISCV::MAXU:
return RISCV::PseudoCCMAXU;
1820 case RISCV::MIN:
return RISCV::PseudoCCMIN;
1821 case RISCV::MINU:
return RISCV::PseudoCCMINU;
1822 case RISCV::MUL:
return RISCV::PseudoCCMUL;
1823 case RISCV::LUI:
return RISCV::PseudoCCLUI;
1824 case RISCV::QC_LI:
return RISCV::PseudoCCQC_LI;
1825 case RISCV::QC_E_LI:
return RISCV::PseudoCCQC_E_LI;
1827 case RISCV::ADDI:
return RISCV::PseudoCCADDI;
1828 case RISCV::SLLI:
return RISCV::PseudoCCSLLI;
1829 case RISCV::SRLI:
return RISCV::PseudoCCSRLI;
1830 case RISCV::SRAI:
return RISCV::PseudoCCSRAI;
1831 case RISCV::ANDI:
return RISCV::PseudoCCANDI;
1832 case RISCV::ORI:
return RISCV::PseudoCCORI;
1833 case RISCV::XORI:
return RISCV::PseudoCCXORI;
1835 case RISCV::ADDW:
return RISCV::PseudoCCADDW;
1836 case RISCV::SUBW:
return RISCV::PseudoCCSUBW;
1837 case RISCV::SLLW:
return RISCV::PseudoCCSLLW;
1838 case RISCV::SRLW:
return RISCV::PseudoCCSRLW;
1839 case RISCV::SRAW:
return RISCV::PseudoCCSRAW;
1841 case RISCV::ADDIW:
return RISCV::PseudoCCADDIW;
1842 case RISCV::SLLIW:
return RISCV::PseudoCCSLLIW;
1843 case RISCV::SRLIW:
return RISCV::PseudoCCSRLIW;
1844 case RISCV::SRAIW:
return RISCV::PseudoCCSRAIW;
1846 case RISCV::ANDN:
return RISCV::PseudoCCANDN;
1847 case RISCV::ORN:
return RISCV::PseudoCCORN;
1848 case RISCV::XNOR:
return RISCV::PseudoCCXNOR;
1850 case RISCV::NDS_BFOS:
return RISCV::PseudoCCNDS_BFOS;
1851 case RISCV::NDS_BFOZ:
return RISCV::PseudoCCNDS_BFOZ;
1855 return RISCV::INSTRUCTION_LIST_END;
1864 if (!
Reg.isVirtual())
1872 if (!STI.hasShortForwardBranchIMinMax() &&
1873 (
MI->getOpcode() == RISCV::MAX ||
MI->getOpcode() == RISCV::MIN ||
1874 MI->getOpcode() == RISCV::MINU ||
MI->getOpcode() == RISCV::MAXU))
1877 if (!STI.hasShortForwardBranchIMul() &&
MI->getOpcode() == RISCV::MUL)
1884 if (
MI->getOpcode() == RISCV::ADDI &&
MI->getOperand(1).isReg() &&
1885 MI->getOperand(1).getReg() == RISCV::X0)
1890 if (MO.isFI() || MO.isCPI() || MO.isJTI())
1903 bool DontMoveAcrossStores =
true;
1904 if (!
MI->isSafeToMove(DontMoveAcrossStores))
1912 bool PreferFalse)
const {
1913 assert(
MI.getOpcode() == RISCV::PseudoCCMOVGPR &&
1914 "Unknown select instruction");
1915 if (!
STI.hasShortForwardBranchIALU())
1921 bool Invert = !
DefMI;
1929 Register DestReg =
MI.getOperand(0).getReg();
1935 assert(PredOpc != RISCV::INSTRUCTION_LIST_END &&
"Unexpected opcode!");
1942 NewMI.
add(FalseReg);
1950 unsigned BCCOpcode =
MI.getOperand(
MI.getNumExplicitOperands() - 3).getImm();
1956 NewMI.
add(
MI.getOperand(
MI.getNumExplicitOperands() - 2));
1957 NewMI.
add(
MI.getOperand(
MI.getNumExplicitOperands() - 1));
1967 if (
DefMI->getParent() !=
MI.getParent())
1971 DefMI->eraseFromParent();
1976 if (
MI.isMetaInstruction())
1979 unsigned Opcode =
MI.getOpcode();
1981 if (Opcode == TargetOpcode::INLINEASM ||
1982 Opcode == TargetOpcode::INLINEASM_BR) {
1984 return getInlineAsmLength(
MI.getOperand(0).getSymbolName(),
1989 if (
STI.hasStdExtZca()) {
1990 if (isCompressibleInst(
MI,
STI))
1997 if (Opcode == TargetOpcode::BUNDLE)
1998 return getInstBundleLength(
MI);
2000 if (
MI.getParent() &&
MI.getParent()->getParent()) {
2001 if (isCompressibleInst(
MI,
STI))
2006 case RISCV::PseudoMV_FPR16INX:
2007 case RISCV::PseudoMV_FPR32INX:
2009 return STI.hasStdExtZca() ? 2 : 4;
2011 case RISCV::PseudoCCMOVGPRNoX0:
2012 return get(
MI.getOperand(
MI.getNumExplicitOperands() - 3).getImm())
2015 case RISCV::PseudoCCMOVGPR:
2016 case RISCV::PseudoCCADD:
2017 case RISCV::PseudoCCSUB:
2018 case RISCV::PseudoCCSLL:
2019 case RISCV::PseudoCCSRL:
2020 case RISCV::PseudoCCSRA:
2021 case RISCV::PseudoCCAND:
2022 case RISCV::PseudoCCOR:
2023 case RISCV::PseudoCCXOR:
2024 case RISCV::PseudoCCADDI:
2025 case RISCV::PseudoCCANDI:
2026 case RISCV::PseudoCCORI:
2027 case RISCV::PseudoCCXORI:
2028 case RISCV::PseudoCCLUI:
2029 case RISCV::PseudoCCSLLI:
2030 case RISCV::PseudoCCSRLI:
2031 case RISCV::PseudoCCSRAI:
2032 case RISCV::PseudoCCADDW:
2033 case RISCV::PseudoCCSUBW:
2034 case RISCV::PseudoCCSLLW:
2035 case RISCV::PseudoCCSRLW:
2036 case RISCV::PseudoCCSRAW:
2037 case RISCV::PseudoCCADDIW:
2038 case RISCV::PseudoCCSLLIW:
2039 case RISCV::PseudoCCSRLIW:
2040 case RISCV::PseudoCCSRAIW:
2041 case RISCV::PseudoCCANDN:
2042 case RISCV::PseudoCCORN:
2043 case RISCV::PseudoCCXNOR:
2044 case RISCV::PseudoCCMAX:
2045 case RISCV::PseudoCCMIN:
2046 case RISCV::PseudoCCMAXU:
2047 case RISCV::PseudoCCMINU:
2048 case RISCV::PseudoCCMUL:
2049 case RISCV::PseudoCCLB:
2050 case RISCV::PseudoCCLH:
2051 case RISCV::PseudoCCLW:
2052 case RISCV::PseudoCCLHU:
2053 case RISCV::PseudoCCLBU:
2054 case RISCV::PseudoCCLWU:
2055 case RISCV::PseudoCCLD:
2056 case RISCV::PseudoCCQC_LI:
2057 return get(
MI.getOperand(
MI.getNumExplicitOperands() - 3).getImm())
2060 case RISCV::PseudoCCQC_E_LI:
2061 case RISCV::PseudoCCQC_E_LB:
2062 case RISCV::PseudoCCQC_E_LH:
2063 case RISCV::PseudoCCQC_E_LW:
2064 case RISCV::PseudoCCQC_E_LHU:
2065 case RISCV::PseudoCCQC_E_LBU:
2066 return get(
MI.getOperand(
MI.getNumExplicitOperands() - 3).getImm())
2069 case TargetOpcode::STACKMAP:
2072 case TargetOpcode::PATCHPOINT:
2075 case TargetOpcode::STATEPOINT: {
2079 return std::max(NumBytes, 8U);
2081 case TargetOpcode::PATCHABLE_FUNCTION_ENTER:
2082 case TargetOpcode::PATCHABLE_FUNCTION_EXIT:
2083 case TargetOpcode::PATCHABLE_TAIL_CALL: {
2086 if (Opcode == TargetOpcode::PATCHABLE_FUNCTION_ENTER &&
2087 F.hasFnAttribute(
"patchable-function-entry")) {
2089 if (
F.getFnAttribute(
"patchable-function-entry")
2091 .getAsInteger(10, Num))
2092 return get(Opcode).getSize();
2095 return (
STI.hasStdExtZca() ? 2 : 4) * Num;
2099 return STI.is64Bit() ? 68 : 44;
2102 return get(Opcode).getSize();
2106unsigned RISCVInstrInfo::getInstBundleLength(
const MachineInstr &
MI)
const {
2110 while (++
I != E &&
I->isInsideBundle()) {
2111 assert(!
I->isBundle() &&
"No nested bundle!");
2118 const unsigned Opcode =
MI.getOpcode();
2122 case RISCV::FSGNJ_D:
2123 case RISCV::FSGNJ_S:
2124 case RISCV::FSGNJ_H:
2125 case RISCV::FSGNJ_D_INX:
2126 case RISCV::FSGNJ_D_IN32X:
2127 case RISCV::FSGNJ_S_INX:
2128 case RISCV::FSGNJ_H_INX:
2130 return MI.getOperand(1).isReg() &&
MI.getOperand(2).isReg() &&
2131 MI.getOperand(1).getReg() ==
MI.getOperand(2).getReg();
2135 return (
MI.getOperand(1).isReg() &&
2136 MI.getOperand(1).getReg() == RISCV::X0) ||
2137 (
MI.getOperand(2).isImm() &&
MI.getOperand(2).getImm() == 0);
2139 return MI.isAsCheapAsAMove();
2142std::optional<DestSourcePair>
2146 switch (
MI.getOpcode()) {
2152 if (
MI.getOperand(1).isReg() &&
MI.getOperand(1).getReg() == RISCV::X0 &&
2153 MI.getOperand(2).isReg())
2155 if (
MI.getOperand(2).isReg() &&
MI.getOperand(2).getReg() == RISCV::X0 &&
2156 MI.getOperand(1).isReg())
2161 if (
MI.getOperand(1).isReg() &&
MI.getOperand(2).isImm() &&
2162 MI.getOperand(2).getImm() == 0)
2166 if (
MI.getOperand(2).isReg() &&
MI.getOperand(2).getReg() == RISCV::X0 &&
2167 MI.getOperand(1).isReg())
2171 case RISCV::SH1ADD_UW:
2173 case RISCV::SH2ADD_UW:
2175 case RISCV::SH3ADD_UW:
2176 if (
MI.getOperand(1).isReg() &&
MI.getOperand(1).getReg() == RISCV::X0 &&
2177 MI.getOperand(2).isReg())
2180 case RISCV::FSGNJ_D:
2181 case RISCV::FSGNJ_S:
2182 case RISCV::FSGNJ_H:
2183 case RISCV::FSGNJ_D_INX:
2184 case RISCV::FSGNJ_D_IN32X:
2185 case RISCV::FSGNJ_S_INX:
2186 case RISCV::FSGNJ_H_INX:
2188 if (
MI.getOperand(1).isReg() &&
MI.getOperand(2).isReg() &&
2189 MI.getOperand(1).getReg() ==
MI.getOperand(2).getReg())
2193 return std::nullopt;
2201 const auto &SchedModel =
STI.getSchedModel();
2202 return (!SchedModel.hasInstrSchedModel() || SchedModel.isOutOfOrder())
2214 RISCV::getNamedOperandIdx(Root.
getOpcode(), RISCV::OpName::frm);
2218 return RISCV::getNamedOperandIdx(
MI->getOpcode(),
2219 RISCV::OpName::frm) < 0;
2221 "New instructions require FRM whereas the old one does not have it");
2228 for (
auto *NewMI : InsInstrs) {
2230 if (
static_cast<unsigned>(RISCV::getNamedOperandIdx(
2231 NewMI->getOpcode(), RISCV::OpName::frm)) != NewMI->getNumOperands())
2273bool RISCVInstrInfo::isVectorAssociativeAndCommutative(
const MachineInstr &Inst,
2274 bool Invert)
const {
2275#define OPCODE_LMUL_CASE(OPC) \
2276 case RISCV::OPC##_M1: \
2277 case RISCV::OPC##_M2: \
2278 case RISCV::OPC##_M4: \
2279 case RISCV::OPC##_M8: \
2280 case RISCV::OPC##_MF2: \
2281 case RISCV::OPC##_MF4: \
2282 case RISCV::OPC##_MF8
2284#define OPCODE_LMUL_MASK_CASE(OPC) \
2285 case RISCV::OPC##_M1_MASK: \
2286 case RISCV::OPC##_M2_MASK: \
2287 case RISCV::OPC##_M4_MASK: \
2288 case RISCV::OPC##_M8_MASK: \
2289 case RISCV::OPC##_MF2_MASK: \
2290 case RISCV::OPC##_MF4_MASK: \
2291 case RISCV::OPC##_MF8_MASK
2296 Opcode = *InvOpcode;
2313#undef OPCODE_LMUL_MASK_CASE
2314#undef OPCODE_LMUL_CASE
2317bool RISCVInstrInfo::areRVVInstsReassociable(
const MachineInstr &Root,
2328 const uint64_t TSFlags =
Desc.TSFlags;
2330 auto checkImmOperand = [&](
unsigned OpIdx) {
2334 auto checkRegOperand = [&](
unsigned OpIdx) {
2342 if (!checkRegOperand(1))
2357 bool SeenMI2 =
false;
2358 for (
auto End =
MBB->
rend(), It = It1; It != End; ++It) {
2367 if (It->modifiesRegister(RISCV::V0,
TRI)) {
2368 Register SrcReg = It->getOperand(1).getReg();
2386 if (MI1VReg != SrcReg)
2395 assert(SeenMI2 &&
"Prev is expected to appear before Root");
2435bool RISCVInstrInfo::hasReassociableVectorSibling(
const MachineInstr &Inst,
2436 bool &Commuted)
const {
2440 "Expect the present of passthrough operand.");
2446 Commuted = !areRVVInstsReassociable(Inst, *MI1) &&
2447 areRVVInstsReassociable(Inst, *MI2);
2451 return areRVVInstsReassociable(Inst, *MI1) &&
2452 (isVectorAssociativeAndCommutative(*MI1) ||
2453 isVectorAssociativeAndCommutative(*MI1,
true)) &&
2460 if (!isVectorAssociativeAndCommutative(Inst) &&
2461 !isVectorAssociativeAndCommutative(Inst,
true))
2487 for (
unsigned I = 0;
I < 5; ++
I)
2493 bool &Commuted)
const {
2494 if (isVectorAssociativeAndCommutative(Inst) ||
2495 isVectorAssociativeAndCommutative(Inst,
true))
2496 return hasReassociableVectorSibling(Inst, Commuted);
2502 unsigned OperandIdx = Commuted ? 2 : 1;
2506 int16_t InstFrmOpIdx =
2507 RISCV::getNamedOperandIdx(Inst.
getOpcode(), RISCV::OpName::frm);
2508 int16_t SiblingFrmOpIdx =
2509 RISCV::getNamedOperandIdx(Sibling.
getOpcode(), RISCV::OpName::frm);
2511 return (InstFrmOpIdx < 0 && SiblingFrmOpIdx < 0) ||
2516 bool Invert)
const {
2517 if (isVectorAssociativeAndCommutative(Inst, Invert))
2525 Opc = *InverseOpcode;
2570std::optional<unsigned>
2572#define RVV_OPC_LMUL_CASE(OPC, INV) \
2573 case RISCV::OPC##_M1: \
2574 return RISCV::INV##_M1; \
2575 case RISCV::OPC##_M2: \
2576 return RISCV::INV##_M2; \
2577 case RISCV::OPC##_M4: \
2578 return RISCV::INV##_M4; \
2579 case RISCV::OPC##_M8: \
2580 return RISCV::INV##_M8; \
2581 case RISCV::OPC##_MF2: \
2582 return RISCV::INV##_MF2; \
2583 case RISCV::OPC##_MF4: \
2584 return RISCV::INV##_MF4; \
2585 case RISCV::OPC##_MF8: \
2586 return RISCV::INV##_MF8
2588#define RVV_OPC_LMUL_MASK_CASE(OPC, INV) \
2589 case RISCV::OPC##_M1_MASK: \
2590 return RISCV::INV##_M1_MASK; \
2591 case RISCV::OPC##_M2_MASK: \
2592 return RISCV::INV##_M2_MASK; \
2593 case RISCV::OPC##_M4_MASK: \
2594 return RISCV::INV##_M4_MASK; \
2595 case RISCV::OPC##_M8_MASK: \
2596 return RISCV::INV##_M8_MASK; \
2597 case RISCV::OPC##_MF2_MASK: \
2598 return RISCV::INV##_MF2_MASK; \
2599 case RISCV::OPC##_MF4_MASK: \
2600 return RISCV::INV##_MF4_MASK; \
2601 case RISCV::OPC##_MF8_MASK: \
2602 return RISCV::INV##_MF8_MASK
2606 return std::nullopt;
2608 return RISCV::FSUB_H;
2610 return RISCV::FSUB_S;
2612 return RISCV::FSUB_D;
2614 return RISCV::FADD_H;
2616 return RISCV::FADD_S;
2618 return RISCV::FADD_D;
2635#undef RVV_OPC_LMUL_MASK_CASE
2636#undef RVV_OPC_LMUL_CASE
2641 bool DoRegPressureReduce) {
2668 bool DoRegPressureReduce) {
2675 DoRegPressureReduce)) {
2681 DoRegPressureReduce)) {
2691 bool DoRegPressureReduce) {
2699 unsigned CombineOpc) {
2706 if (!
MI ||
MI->getParent() != &
MBB ||
MI->getOpcode() != CombineOpc)
2720 unsigned OuterShiftAmt) {
2726 if (InnerShiftAmt < OuterShiftAmt || (InnerShiftAmt - OuterShiftAmt) > 3)
2753 case RISCV::SH1ADD_UW:
2755 case RISCV::SH2ADD_UW:
2757 case RISCV::SH3ADD_UW:
2803 bool DoRegPressureReduce)
const {
2812 DoRegPressureReduce);
2820 return RISCV::FMADD_H;
2822 return RISCV::FMADD_S;
2824 return RISCV::FMADD_D;
2869 bool Mul1IsKill = Mul1.
isKill();
2870 bool Mul2IsKill = Mul2.
isKill();
2871 bool AddendIsKill = Addend.
isKill();
2880 BuildMI(*MF, MergedLoc,
TII->get(FusedOpc), DstReg)
2905 assert(OuterShiftAmt != 0 &&
"Unexpected opcode");
2912 assert(InnerShiftAmt >= OuterShiftAmt &&
"Unexpected shift amount");
2915 switch (InnerShiftAmt - OuterShiftAmt) {
2919 InnerOpc = RISCV::ADD;
2922 InnerOpc = RISCV::SH1ADD;
2925 InnerOpc = RISCV::SH2ADD;
2928 InnerOpc = RISCV::SH3ADD;
2946 InstrIdxForVirtReg.
insert(std::make_pair(NewVR, 0));
2963 DelInstrs, InstrIdxForVirtReg);
2990 for (
const auto &[Index, Operand] :
enumerate(
Desc.operands())) {
2992 unsigned OpType = Operand.OperandType;
2998 ErrInfo =
"Expected an immediate operand.";
3001 int64_t Imm = MO.
getImm();
3007#define CASE_OPERAND_UIMM(NUM) \
3008 case RISCVOp::OPERAND_UIMM##NUM: \
3009 Ok = isUInt<NUM>(Imm); \
3011#define CASE_OPERAND_UIMM_LSB_ZEROS(BITS, SUFFIX) \
3012 case RISCVOp::OPERAND_UIMM##BITS##_LSB##SUFFIX: { \
3013 constexpr size_t NumZeros = sizeof(#SUFFIX) - 1; \
3014 Ok = isShiftedUInt<BITS - NumZeros, NumZeros>(Imm); \
3017#define CASE_OPERAND_SIMM(NUM) \
3018 case RISCVOp::OPERAND_SIMM##NUM: \
3019 Ok = isInt<NUM>(Imm); \
3053 Ok = Imm >= 1 && Imm <= 32;
3056 Ok = Imm >= 1 && Imm <= 64;
3077 Ok = (
isUInt<5>(Imm) && Imm != 0) || Imm == -1;
3088 Ok = Imm >= -15 && Imm <= 16;
3116 Ok = Ok && Imm != 0;
3119 Ok = (
isUInt<5>(Imm) && Imm != 0) || (Imm >= 0xfffe0 && Imm <= 0xfffff);
3122 Ok = Imm >= 0 && Imm <= 10;
3125 Ok = Imm >= 0 && Imm <= 7;
3128 Ok = Imm >= 1 && Imm <= 10;
3131 Ok = Imm >= 2 && Imm <= 14;
3140 Ok = Imm >= 0 && Imm <= 48 && Imm % 16 == 0;
3175 Ok = Imm == 1 || Imm == 2 || Imm == 4;
3179 ErrInfo =
"Invalid immediate";
3188 ErrInfo =
"Expected a non-register operand.";
3192 ErrInfo =
"Invalid immediate";
3201 ErrInfo =
"Expected a non-register operand.";
3205 ErrInfo =
"Invalid immediate";
3213 ErrInfo =
"Expected a non-register operand.";
3217 ErrInfo =
"Invalid immediate";
3223 int64_t Imm = MO.
getImm();
3226 ErrInfo =
"Invalid immediate";
3229 }
else if (!MO.
isReg()) {
3230 ErrInfo =
"Expected a register or immediate operand.";
3236 ErrInfo =
"Expected a register or immediate operand.";
3246 if (!
Op.isImm() && !
Op.isReg()) {
3247 ErrInfo =
"Invalid operand type for VL operand";
3250 if (
Op.isReg() &&
Op.getReg().isValid()) {
3253 if (!RISCV::GPRNoX0RegClass.hasSubClassEq(RC)) {
3254 ErrInfo =
"Invalid register class for VL operand";
3259 ErrInfo =
"VL operand w/o SEW operand?";
3265 if (!
MI.getOperand(
OpIdx).isImm()) {
3266 ErrInfo =
"SEW value expected to be an immediate";
3271 ErrInfo =
"Unexpected SEW value";
3274 unsigned SEW = Log2SEW ? 1 << Log2SEW : 8;
3276 ErrInfo =
"Unexpected SEW value";
3282 if (!
MI.getOperand(
OpIdx).isImm()) {
3283 ErrInfo =
"Policy operand expected to be an immediate";
3288 ErrInfo =
"Invalid Policy Value";
3292 ErrInfo =
"policy operand w/o VL operand?";
3300 if (!
MI.isRegTiedToUseOperand(0, &UseOpIdx)) {
3301 ErrInfo =
"policy operand w/o tied operand?";
3308 !
MI.readsRegister(RISCV::FRM,
nullptr)) {
3309 ErrInfo =
"dynamic rounding mode should read FRM";
3331 case RISCV::LD_RV32:
3341 case RISCV::SD_RV32:
3357 int64_t NewOffset = OldOffset + Disp;
3379 "Addressing mode not supported for folding");
3452 case RISCV::LD_RV32:
3455 case RISCV::SD_RV32:
3462 OffsetIsScalable =
false;
3478 if (BaseOps1.
front()->isIdenticalTo(*BaseOps2.
front()))
3486 if (MO1->getAddrSpace() != MO2->getAddrSpace())
3489 auto Base1 = MO1->getValue();
3490 auto Base2 = MO2->getValue();
3491 if (!Base1 || !Base2)
3499 return Base1 == Base2;
3505 int64_t Offset2,
bool OffsetIsScalable2,
unsigned ClusterSize,
3506 unsigned NumBytes)
const {
3509 if (!BaseOps1.
empty() && !BaseOps2.
empty()) {
3514 }
else if (!BaseOps1.
empty() || !BaseOps2.
empty()) {
3520 BaseOps1.
front()->getParent()->getMF()->getSubtarget().getCacheLineSize();
3526 return ClusterSize <= 4 && std::abs(Offset1 - Offset2) <
CacheLineSize;
3576 int64_t OffsetA = 0, OffsetB = 0;
3582 int LowOffset = std::min(OffsetA, OffsetB);
3583 int HighOffset = std::max(OffsetA, OffsetB);
3584 LocationSize LowWidth = (LowOffset == OffsetA) ? WidthA : WidthB;
3586 LowOffset + (
int)LowWidth.
getValue() <= HighOffset)
3593std::pair<unsigned, unsigned>
3596 return std::make_pair(TF & Mask, TF & ~Mask);
3602 static const std::pair<unsigned, const char *> TargetFlags[] = {
3603 {MO_CALL,
"riscv-call"},
3604 {MO_LO,
"riscv-lo"},
3605 {MO_HI,
"riscv-hi"},
3606 {MO_PCREL_LO,
"riscv-pcrel-lo"},
3607 {MO_PCREL_HI,
"riscv-pcrel-hi"},
3608 {MO_GOT_HI,
"riscv-got-hi"},
3609 {MO_TPREL_LO,
"riscv-tprel-lo"},
3610 {MO_TPREL_HI,
"riscv-tprel-hi"},
3611 {MO_TPREL_ADD,
"riscv-tprel-add"},
3612 {MO_TLS_GOT_HI,
"riscv-tls-got-hi"},
3613 {MO_TLS_GD_HI,
"riscv-tls-gd-hi"},
3614 {MO_TLSDESC_HI,
"riscv-tlsdesc-hi"},
3615 {MO_TLSDESC_LOAD_LO,
"riscv-tlsdesc-load-lo"},
3616 {MO_TLSDESC_ADD_LO,
"riscv-tlsdesc-add-lo"},
3617 {MO_TLSDESC_CALL,
"riscv-tlsdesc-call"}};
3625 if (!OutlineFromLinkOnceODRs &&
F.hasLinkOnceODRLinkage())
3638 unsigned &Flags)
const {
3658 return F.getFnAttribute(
"fentry-call").getValueAsBool() ||
3659 F.hasFnAttribute(
"patchable-function-entry");
3664 return MI.readsRegister(RegNo,
TRI) ||
3665 MI.getDesc().hasImplicitUseOfPhysReg(RegNo);
3670 return MI.modifiesRegister(RegNo,
TRI) ||
3671 MI.getDesc().hasImplicitDefOfPhysReg(RegNo);
3675 if (!
MBB.back().isReturn())
3714 if (
C.isAvailableAcrossAndOutOfSeq(
Reg,
TRI) &&
3715 C.isAvailableInsideSeq(
Reg,
TRI)) {
3729 if (
C.back().isReturn() &&
3730 !
C.isAvailableAcrossAndOutOfSeq(TailExpandUseReg, RegInfo)) {
3732 LLVM_DEBUG(
dbgs() <<
"Cannot be outlined between: " <<
C.front() <<
"and "
3734 LLVM_DEBUG(
dbgs() <<
"Because the tail-call register is live across "
3735 "the proposed outlined function call\n");
3741 if (
C.back().isReturn()) {
3743 "The candidate who uses return instruction must be outlined "
3750 if (!
C.isAvailableInsideSeq(RISCV::X5, RegInfo))
3754 if (
C.isAvailableAcrossAndOutOfSeq(RISCV::X5, RegInfo))
3764std::optional<std::unique_ptr<outliner::OutlinedFunction>>
3767 std::vector<outliner::Candidate> &RepeatedSequenceLocs,
3768 unsigned MinRepeats)
const {
3776 if (RepeatedSequenceLocs.size() < MinRepeats)
3777 return std::nullopt;
3781 unsigned InstrSizeCExt =
3783 unsigned CallOverhead = 0, FrameOverhead = 0;
3786 unsigned CFICount = 0;
3787 for (
auto &
I : Candidate) {
3788 if (
I.isCFIInstruction())
3799 std::vector<MCCFIInstruction> CFIInstructions =
3800 C.getMF()->getFrameInstructions();
3802 if (CFICount > 0 && CFICount != CFIInstructions.size())
3803 return std::nullopt;
3811 CallOverhead = 4 + InstrSizeCExt;
3818 FrameOverhead = InstrSizeCExt;
3824 return std::nullopt;
3828 for (
auto &
C : RepeatedSequenceLocs) {
3830 if (
C.isAvailableAcrossAndOutOfSeq(RISCV::X5, RegInfo)) {
3832 unsigned CandCallOverhead = 8;
3837 unsigned CandCallOverhead = InstrSizeCExt + 8 + InstrSizeCExt;
3842 for (
auto &
C : RepeatedSequenceLocs)
3843 C.setCallInfo(MOCI, CallOverhead);
3846 unsigned SequenceSize = 0;
3847 for (
auto &
MI : Candidate)
3850 return std::make_unique<outliner::OutlinedFunction>(
3851 RepeatedSequenceLocs, SequenceSize, FrameOverhead, MOCI);
3857 unsigned Flags)
const {
3861 MBB->getParent()->getSubtarget().getRegisterInfo();
3862 const auto &
F =
MI.getMF()->getFunction();
3867 if (
MI.isCFIInstruction())
3875 for (
const auto &MO :
MI.operands()) {
3880 (
MI.getMF()->getTarget().getFunctionSections() ||
F.hasComdat() ||
3881 F.hasSection() ||
F.getSectionPrefix()))
3898 MBB.addLiveIn(RISCV::X5);
3913 .addGlobalAddress(M.getNamedValue(MF.
getName()),
3920 assert(SaveReg &&
"Cannot find an available register to save/restore X5.");
3931 .addGlobalAddress(M.getNamedValue(MF.
getName()), 0,
3947 .addGlobalAddress(M.getNamedValue(MF.
getName()), 0,
3958 return std::nullopt;
3962 if (
MI.getOpcode() == RISCV::ADDI &&
MI.getOperand(1).isReg() &&
3963 MI.getOperand(2).isImm())
3964 return RegImmPair{
MI.getOperand(1).getReg(),
MI.getOperand(2).getImm()};
3966 return std::nullopt;
3974 std::string GenericComment =
3976 if (!GenericComment.empty())
3977 return GenericComment;
3981 return std::string();
3983 std::string Comment;
3990 switch (OpInfo.OperandType) {
3993 unsigned Imm =
Op.getImm();
3998 unsigned Imm =
Op.getImm();
4003 unsigned Imm =
Op.getImm();
4009 unsigned Log2SEW =
Op.getImm();
4010 unsigned SEW = Log2SEW ? 1 << Log2SEW : 8;
4016 unsigned Policy =
Op.getImm();
4018 "Invalid Policy Value");
4024 if (
Op.isImm() &&
Op.getImm() == -1)
4046#define CASE_RVV_OPCODE_UNMASK_LMUL(OP, LMUL) \
4047 RISCV::Pseudo##OP##_##LMUL
4049#define CASE_RVV_OPCODE_MASK_LMUL(OP, LMUL) \
4050 RISCV::Pseudo##OP##_##LMUL##_MASK
4052#define CASE_RVV_OPCODE_LMUL(OP, LMUL) \
4053 CASE_RVV_OPCODE_UNMASK_LMUL(OP, LMUL): \
4054 case CASE_RVV_OPCODE_MASK_LMUL(OP, LMUL)
4056#define CASE_RVV_OPCODE_UNMASK_WIDEN(OP) \
4057 CASE_RVV_OPCODE_UNMASK_LMUL(OP, MF8): \
4058 case CASE_RVV_OPCODE_UNMASK_LMUL(OP, MF4): \
4059 case CASE_RVV_OPCODE_UNMASK_LMUL(OP, MF2): \
4060 case CASE_RVV_OPCODE_UNMASK_LMUL(OP, M1): \
4061 case CASE_RVV_OPCODE_UNMASK_LMUL(OP, M2): \
4062 case CASE_RVV_OPCODE_UNMASK_LMUL(OP, M4)
4064#define CASE_RVV_OPCODE_UNMASK(OP) \
4065 CASE_RVV_OPCODE_UNMASK_WIDEN(OP): \
4066 case CASE_RVV_OPCODE_UNMASK_LMUL(OP, M8)
4068#define CASE_RVV_OPCODE_MASK_WIDEN(OP) \
4069 CASE_RVV_OPCODE_MASK_LMUL(OP, MF8): \
4070 case CASE_RVV_OPCODE_MASK_LMUL(OP, MF4): \
4071 case CASE_RVV_OPCODE_MASK_LMUL(OP, MF2): \
4072 case CASE_RVV_OPCODE_MASK_LMUL(OP, M1): \
4073 case CASE_RVV_OPCODE_MASK_LMUL(OP, M2): \
4074 case CASE_RVV_OPCODE_MASK_LMUL(OP, M4)
4076#define CASE_RVV_OPCODE_MASK(OP) \
4077 CASE_RVV_OPCODE_MASK_WIDEN(OP): \
4078 case CASE_RVV_OPCODE_MASK_LMUL(OP, M8)
4080#define CASE_RVV_OPCODE_WIDEN(OP) \
4081 CASE_RVV_OPCODE_UNMASK_WIDEN(OP): \
4082 case CASE_RVV_OPCODE_MASK_WIDEN(OP)
4084#define CASE_RVV_OPCODE(OP) \
4085 CASE_RVV_OPCODE_UNMASK(OP): \
4086 case CASE_RVV_OPCODE_MASK(OP)
4090#define CASE_VMA_OPCODE_COMMON(OP, TYPE, LMUL) \
4091 RISCV::PseudoV##OP##_##TYPE##_##LMUL
4093#define CASE_VMA_OPCODE_LMULS(OP, TYPE) \
4094 CASE_VMA_OPCODE_COMMON(OP, TYPE, MF8): \
4095 case CASE_VMA_OPCODE_COMMON(OP, TYPE, MF4): \
4096 case CASE_VMA_OPCODE_COMMON(OP, TYPE, MF2): \
4097 case CASE_VMA_OPCODE_COMMON(OP, TYPE, M1): \
4098 case CASE_VMA_OPCODE_COMMON(OP, TYPE, M2): \
4099 case CASE_VMA_OPCODE_COMMON(OP, TYPE, M4): \
4100 case CASE_VMA_OPCODE_COMMON(OP, TYPE, M8)
4103#define CASE_VFMA_OPCODE_COMMON(OP, TYPE, LMUL, SEW) \
4104 RISCV::PseudoV##OP##_##TYPE##_##LMUL##_##SEW
4106#define CASE_VFMA_OPCODE_LMULS_M1(OP, TYPE, SEW) \
4107 CASE_VFMA_OPCODE_COMMON(OP, TYPE, M1, SEW): \
4108 case CASE_VFMA_OPCODE_COMMON(OP, TYPE, M2, SEW): \
4109 case CASE_VFMA_OPCODE_COMMON(OP, TYPE, M4, SEW): \
4110 case CASE_VFMA_OPCODE_COMMON(OP, TYPE, M8, SEW)
4112#define CASE_VFMA_OPCODE_LMULS_MF2(OP, TYPE, SEW) \
4113 CASE_VFMA_OPCODE_COMMON(OP, TYPE, MF2, SEW): \
4114 case CASE_VFMA_OPCODE_LMULS_M1(OP, TYPE, SEW)
4116#define CASE_VFMA_OPCODE_LMULS_MF4(OP, TYPE, SEW) \
4117 CASE_VFMA_OPCODE_COMMON(OP, TYPE, MF4, SEW): \
4118 case CASE_VFMA_OPCODE_LMULS_MF2(OP, TYPE, SEW)
4120#define CASE_VFMA_OPCODE_VV(OP) \
4121 CASE_VFMA_OPCODE_LMULS_MF4(OP, VV, E16): \
4122 case CASE_VFMA_OPCODE_LMULS_MF4(OP##_ALT, VV, E16): \
4123 case CASE_VFMA_OPCODE_LMULS_MF2(OP, VV, E32): \
4124 case CASE_VFMA_OPCODE_LMULS_M1(OP, VV, E64)
4126#define CASE_VFMA_SPLATS(OP) \
4127 CASE_VFMA_OPCODE_LMULS_MF4(OP, VFPR16, E16): \
4128 case CASE_VFMA_OPCODE_LMULS_MF4(OP##_ALT, VFPR16, E16): \
4129 case CASE_VFMA_OPCODE_LMULS_MF2(OP, VFPR32, E32): \
4130 case CASE_VFMA_OPCODE_LMULS_M1(OP, VFPR64, E64)
4134 unsigned &SrcOpIdx1,
4135 unsigned &SrcOpIdx2)
const {
4137 if (!
Desc.isCommutable())
4140 switch (
MI.getOpcode()) {
4141 case RISCV::TH_MVEQZ:
4142 case RISCV::TH_MVNEZ:
4146 if (
MI.getOperand(2).getReg() == RISCV::X0)
4149 return fixCommutedOpIndices(SrcOpIdx1, SrcOpIdx2, 1, 2);
4150 case RISCV::QC_SELECTIEQ:
4151 case RISCV::QC_SELECTINE:
4152 case RISCV::QC_SELECTIIEQ:
4153 case RISCV::QC_SELECTIINE:
4154 return fixCommutedOpIndices(SrcOpIdx1, SrcOpIdx2, 1, 2);
4155 case RISCV::QC_MVEQ:
4156 case RISCV::QC_MVNE:
4157 case RISCV::QC_MVLT:
4158 case RISCV::QC_MVGE:
4159 case RISCV::QC_MVLTU:
4160 case RISCV::QC_MVGEU:
4161 case RISCV::QC_MVEQI:
4162 case RISCV::QC_MVNEI:
4163 case RISCV::QC_MVLTI:
4164 case RISCV::QC_MVGEI:
4165 case RISCV::QC_MVLTUI:
4166 case RISCV::QC_MVGEUI:
4167 return fixCommutedOpIndices(SrcOpIdx1, SrcOpIdx2, 1, 4);
4168 case RISCV::TH_MULA:
4169 case RISCV::TH_MULAW:
4170 case RISCV::TH_MULAH:
4171 case RISCV::TH_MULS:
4172 case RISCV::TH_MULSW:
4173 case RISCV::TH_MULSH:
4175 return fixCommutedOpIndices(SrcOpIdx1, SrcOpIdx2, 2, 3);
4176 case RISCV::PseudoCCMOVGPRNoX0:
4177 case RISCV::PseudoCCMOVGPR:
4179 return fixCommutedOpIndices(SrcOpIdx1, SrcOpIdx2, 1, 2);
4220 return fixCommutedOpIndices(SrcOpIdx1, SrcOpIdx2, 2, 3);
4247 unsigned CommutableOpIdx1 = 1;
4248 unsigned CommutableOpIdx2 = 3;
4249 if (!fixCommutedOpIndices(SrcOpIdx1, SrcOpIdx2, CommutableOpIdx1,
4270 if (SrcOpIdx1 != CommuteAnyOperandIndex && SrcOpIdx1 > 3)
4272 if (SrcOpIdx2 != CommuteAnyOperandIndex && SrcOpIdx2 > 3)
4276 if (SrcOpIdx1 != CommuteAnyOperandIndex &&
4277 SrcOpIdx2 != CommuteAnyOperandIndex && SrcOpIdx1 != 1 && SrcOpIdx2 != 1)
4283 if (SrcOpIdx1 == CommuteAnyOperandIndex ||
4284 SrcOpIdx2 == CommuteAnyOperandIndex) {
4287 unsigned CommutableOpIdx1 = SrcOpIdx1;
4288 if (SrcOpIdx1 == SrcOpIdx2) {
4291 CommutableOpIdx1 = 1;
4292 }
else if (SrcOpIdx1 == CommuteAnyOperandIndex) {
4294 CommutableOpIdx1 = SrcOpIdx2;
4299 unsigned CommutableOpIdx2;
4300 if (CommutableOpIdx1 != 1) {
4302 CommutableOpIdx2 = 1;
4304 Register Op1Reg =
MI.getOperand(CommutableOpIdx1).getReg();
4309 if (Op1Reg !=
MI.getOperand(2).getReg())
4310 CommutableOpIdx2 = 2;
4312 CommutableOpIdx2 = 3;
4317 if (!fixCommutedOpIndices(SrcOpIdx1, SrcOpIdx2, CommutableOpIdx1,
4330#define CASE_VMA_CHANGE_OPCODE_COMMON(OLDOP, NEWOP, TYPE, LMUL) \
4331 case RISCV::PseudoV##OLDOP##_##TYPE##_##LMUL: \
4332 Opc = RISCV::PseudoV##NEWOP##_##TYPE##_##LMUL; \
4335#define CASE_VMA_CHANGE_OPCODE_LMULS(OLDOP, NEWOP, TYPE) \
4336 CASE_VMA_CHANGE_OPCODE_COMMON(OLDOP, NEWOP, TYPE, MF8) \
4337 CASE_VMA_CHANGE_OPCODE_COMMON(OLDOP, NEWOP, TYPE, MF4) \
4338 CASE_VMA_CHANGE_OPCODE_COMMON(OLDOP, NEWOP, TYPE, MF2) \
4339 CASE_VMA_CHANGE_OPCODE_COMMON(OLDOP, NEWOP, TYPE, M1) \
4340 CASE_VMA_CHANGE_OPCODE_COMMON(OLDOP, NEWOP, TYPE, M2) \
4341 CASE_VMA_CHANGE_OPCODE_COMMON(OLDOP, NEWOP, TYPE, M4) \
4342 CASE_VMA_CHANGE_OPCODE_COMMON(OLDOP, NEWOP, TYPE, M8)
4345#define CASE_VFMA_CHANGE_OPCODE_COMMON(OLDOP, NEWOP, TYPE, LMUL, SEW) \
4346 case RISCV::PseudoV##OLDOP##_##TYPE##_##LMUL##_##SEW: \
4347 Opc = RISCV::PseudoV##NEWOP##_##TYPE##_##LMUL##_##SEW; \
4350#define CASE_VFMA_CHANGE_OPCODE_LMULS_M1(OLDOP, NEWOP, TYPE, SEW) \
4351 CASE_VFMA_CHANGE_OPCODE_COMMON(OLDOP, NEWOP, TYPE, M1, SEW) \
4352 CASE_VFMA_CHANGE_OPCODE_COMMON(OLDOP, NEWOP, TYPE, M2, SEW) \
4353 CASE_VFMA_CHANGE_OPCODE_COMMON(OLDOP, NEWOP, TYPE, M4, SEW) \
4354 CASE_VFMA_CHANGE_OPCODE_COMMON(OLDOP, NEWOP, TYPE, M8, SEW)
4356#define CASE_VFMA_CHANGE_OPCODE_LMULS_MF2(OLDOP, NEWOP, TYPE, SEW) \
4357 CASE_VFMA_CHANGE_OPCODE_COMMON(OLDOP, NEWOP, TYPE, MF2, SEW) \
4358 CASE_VFMA_CHANGE_OPCODE_LMULS_M1(OLDOP, NEWOP, TYPE, SEW)
4360#define CASE_VFMA_CHANGE_OPCODE_LMULS_MF4(OLDOP, NEWOP, TYPE, SEW) \
4361 CASE_VFMA_CHANGE_OPCODE_COMMON(OLDOP, NEWOP, TYPE, MF4, SEW) \
4362 CASE_VFMA_CHANGE_OPCODE_LMULS_MF2(OLDOP, NEWOP, TYPE, SEW)
4364#define CASE_VFMA_CHANGE_OPCODE_VV(OLDOP, NEWOP) \
4365 CASE_VFMA_CHANGE_OPCODE_LMULS_MF4(OLDOP, NEWOP, VV, E16) \
4366 CASE_VFMA_CHANGE_OPCODE_LMULS_MF4(OLDOP##_ALT, NEWOP##_ALT, VV, E16) \
4367 CASE_VFMA_CHANGE_OPCODE_LMULS_MF2(OLDOP, NEWOP, VV, E32) \
4368 CASE_VFMA_CHANGE_OPCODE_LMULS_M1(OLDOP, NEWOP, VV, E64)
4370#define CASE_VFMA_CHANGE_OPCODE_SPLATS(OLDOP, NEWOP) \
4371 CASE_VFMA_CHANGE_OPCODE_LMULS_MF4(OLDOP, NEWOP, VFPR16, E16) \
4372 CASE_VFMA_CHANGE_OPCODE_LMULS_MF4(OLDOP##_ALT, NEWOP##_ALT, VFPR16, E16) \
4373 CASE_VFMA_CHANGE_OPCODE_LMULS_MF2(OLDOP, NEWOP, VFPR32, E32) \
4374 CASE_VFMA_CHANGE_OPCODE_LMULS_M1(OLDOP, NEWOP, VFPR64, E64)
4380 unsigned OpIdx2)
const {
4383 return *
MI.getParent()->getParent()->CloneMachineInstr(&
MI);
4387 switch (
MI.getOpcode()) {
4388 case RISCV::TH_MVEQZ:
4389 case RISCV::TH_MVNEZ: {
4390 auto &WorkingMI = cloneIfNew(
MI);
4391 WorkingMI.setDesc(
get(
MI.getOpcode() == RISCV::TH_MVEQZ ? RISCV::TH_MVNEZ
4392 : RISCV::TH_MVEQZ));
4396 case RISCV::QC_SELECTIEQ:
4397 case RISCV::QC_SELECTINE:
4398 case RISCV::QC_SELECTIIEQ:
4399 case RISCV::QC_SELECTIINE:
4401 case RISCV::QC_MVEQ:
4402 case RISCV::QC_MVNE:
4403 case RISCV::QC_MVLT:
4404 case RISCV::QC_MVGE:
4405 case RISCV::QC_MVLTU:
4406 case RISCV::QC_MVGEU:
4407 case RISCV::QC_MVEQI:
4408 case RISCV::QC_MVNEI:
4409 case RISCV::QC_MVLTI:
4410 case RISCV::QC_MVGEI:
4411 case RISCV::QC_MVLTUI:
4412 case RISCV::QC_MVGEUI: {
4413 auto &WorkingMI = cloneIfNew(
MI);
4418 case RISCV::PseudoCCMOVGPRNoX0:
4419 case RISCV::PseudoCCMOVGPR: {
4421 unsigned BCC =
MI.getOperand(
MI.getNumExplicitOperands() - 3).getImm();
4423 auto &WorkingMI = cloneIfNew(
MI);
4424 WorkingMI.getOperand(
MI.getNumExplicitOperands() - 3).setImm(BCC);
4448 assert((OpIdx1 == 1 || OpIdx2 == 1) &&
"Unexpected opcode index");
4449 assert((OpIdx1 == 3 || OpIdx2 == 3) &&
"Unexpected opcode index");
4451 switch (
MI.getOpcode()) {
4474 auto &WorkingMI = cloneIfNew(
MI);
4475 WorkingMI.setDesc(
get(
Opc));
4485 assert((OpIdx1 == 1 || OpIdx2 == 1) &&
"Unexpected opcode index");
4488 if (OpIdx1 == 3 || OpIdx2 == 3) {
4490 switch (
MI.getOpcode()) {
4501 auto &WorkingMI = cloneIfNew(
MI);
4502 WorkingMI.setDesc(
get(
Opc));
4514#undef CASE_VMA_CHANGE_OPCODE_COMMON
4515#undef CASE_VMA_CHANGE_OPCODE_LMULS
4516#undef CASE_VFMA_CHANGE_OPCODE_COMMON
4517#undef CASE_VFMA_CHANGE_OPCODE_LMULS_M1
4518#undef CASE_VFMA_CHANGE_OPCODE_LMULS_MF2
4519#undef CASE_VFMA_CHANGE_OPCODE_LMULS_MF4
4520#undef CASE_VFMA_CHANGE_OPCODE_VV
4521#undef CASE_VFMA_CHANGE_OPCODE_SPLATS
4523#undef CASE_RVV_OPCODE_UNMASK_LMUL
4524#undef CASE_RVV_OPCODE_MASK_LMUL
4525#undef CASE_RVV_OPCODE_LMUL
4526#undef CASE_RVV_OPCODE_UNMASK_WIDEN
4527#undef CASE_RVV_OPCODE_UNMASK
4528#undef CASE_RVV_OPCODE_MASK_WIDEN
4529#undef CASE_RVV_OPCODE_MASK
4530#undef CASE_RVV_OPCODE_WIDEN
4531#undef CASE_RVV_OPCODE
4533#undef CASE_VMA_OPCODE_COMMON
4534#undef CASE_VMA_OPCODE_LMULS
4535#undef CASE_VFMA_OPCODE_COMMON
4536#undef CASE_VFMA_OPCODE_LMULS_M1
4537#undef CASE_VFMA_OPCODE_LMULS_MF2
4538#undef CASE_VFMA_OPCODE_LMULS_MF4
4539#undef CASE_VFMA_OPCODE_VV
4540#undef CASE_VFMA_SPLATS
4543 switch (
MI.getOpcode()) {
4551 if (
MI.getOperand(1).getReg() == RISCV::X0)
4552 commuteInstruction(
MI);
4554 if (
MI.getOperand(2).getReg() == RISCV::X0) {
4555 MI.getOperand(2).ChangeToImmediate(0);
4556 MI.setDesc(
get(RISCV::ADDI));
4560 if (
MI.getOpcode() == RISCV::XOR &&
4561 MI.getOperand(1).getReg() ==
MI.getOperand(2).getReg()) {
4562 MI.getOperand(1).setReg(RISCV::X0);
4563 MI.getOperand(2).ChangeToImmediate(0);
4564 MI.setDesc(
get(RISCV::ADDI));
4571 if (
MI.getOperand(1).getReg() == RISCV::X0) {
4572 MI.setDesc(
get(RISCV::ADDI));
4578 if (
MI.getOperand(2).getReg() == RISCV::X0) {
4579 MI.getOperand(2).ChangeToImmediate(0);
4580 MI.setDesc(
get(RISCV::ADDI));
4586 if (
MI.getOperand(2).getReg() == RISCV::X0) {
4587 MI.getOperand(2).ChangeToImmediate(0);
4588 MI.setDesc(
get(RISCV::ADDIW));
4595 if (
MI.getOperand(1).getReg() == RISCV::X0)
4596 commuteInstruction(
MI);
4598 if (
MI.getOperand(2).getReg() == RISCV::X0) {
4599 MI.getOperand(2).ChangeToImmediate(0);
4600 MI.setDesc(
get(RISCV::ADDIW));
4605 case RISCV::SH1ADD_UW:
4607 case RISCV::SH2ADD_UW:
4609 case RISCV::SH3ADD_UW:
4611 if (
MI.getOperand(1).getReg() == RISCV::X0) {
4612 MI.removeOperand(1);
4614 MI.setDesc(
get(RISCV::ADDI));
4618 if (
MI.getOperand(2).getReg() == RISCV::X0) {
4619 MI.removeOperand(2);
4620 unsigned Opc =
MI.getOpcode();
4621 if (
Opc == RISCV::SH1ADD_UW ||
Opc == RISCV::SH2ADD_UW ||
4622 Opc == RISCV::SH3ADD_UW) {
4624 MI.setDesc(
get(RISCV::SLLI_UW));
4628 MI.setDesc(
get(RISCV::SLLI));
4642 if (
MI.getOperand(1).getReg() == RISCV::X0 ||
4643 MI.getOperand(2).getReg() == RISCV::X0) {
4644 MI.getOperand(1).setReg(RISCV::X0);
4645 MI.getOperand(2).ChangeToImmediate(0);
4646 MI.setDesc(
get(RISCV::ADDI));
4652 if (
MI.getOperand(1).getReg() == RISCV::X0) {
4653 MI.getOperand(2).setImm(0);
4654 MI.setDesc(
get(RISCV::ADDI));
4662 if (
MI.getOperand(1).getReg() == RISCV::X0) {
4663 MI.getOperand(2).ChangeToImmediate(0);
4664 MI.setDesc(
get(RISCV::ADDI));
4668 if (
MI.getOperand(2).getReg() == RISCV::X0) {
4669 MI.getOperand(2).ChangeToImmediate(0);
4670 MI.setDesc(
get(RISCV::ADDI));
4678 if (
MI.getOperand(1).getReg() == RISCV::X0) {
4679 MI.getOperand(2).ChangeToImmediate(0);
4680 MI.setDesc(
get(RISCV::ADDI));
4690 case RISCV::SLLI_UW:
4692 if (
MI.getOperand(1).getReg() == RISCV::X0) {
4693 MI.getOperand(2).setImm(0);
4694 MI.setDesc(
get(RISCV::ADDI));
4702 if (
MI.getOperand(1).getReg() == RISCV::X0 &&
4703 MI.getOperand(2).getReg() == RISCV::X0) {
4704 MI.getOperand(2).ChangeToImmediate(0);
4705 MI.setDesc(
get(RISCV::ADDI));
4709 if (
MI.getOpcode() == RISCV::ADD_UW &&
4710 MI.getOperand(1).getReg() == RISCV::X0) {
4711 MI.removeOperand(1);
4713 MI.setDesc(
get(RISCV::ADDI));
4719 if (
MI.getOperand(1).getReg() == RISCV::X0) {
4720 MI.getOperand(2).setImm(
MI.getOperand(2).getImm() != 0);
4721 MI.setDesc(
get(RISCV::ADDI));
4727 case RISCV::ZEXT_H_RV32:
4728 case RISCV::ZEXT_H_RV64:
4731 if (
MI.getOperand(1).getReg() == RISCV::X0) {
4733 MI.setDesc(
get(RISCV::ADDI));
4742 if (
MI.getOperand(1).getReg() ==
MI.getOperand(2).getReg()) {
4743 MI.getOperand(2).ChangeToImmediate(0);
4744 MI.setDesc(
get(RISCV::ADDI));
4751 if (
MI.getOperand(0).getReg() == RISCV::X0) {
4753 MI.removeOperand(0);
4754 MI.insert(
MI.operands_begin() + 1, {MO0});
4759 if (
MI.getOperand(0).getReg() == RISCV::X0) {
4761 MI.removeOperand(0);
4762 MI.insert(
MI.operands_begin() + 1, {MO0});
4763 MI.setDesc(
get(RISCV::BNE));
4768 if (
MI.getOperand(0).getReg() == RISCV::X0) {
4770 MI.removeOperand(0);
4771 MI.insert(
MI.operands_begin() + 1, {MO0});
4772 MI.setDesc(
get(RISCV::BEQ));
4780#define CASE_WIDEOP_OPCODE_COMMON(OP, LMUL) \
4781 RISCV::PseudoV##OP##_##LMUL##_TIED
4783#define CASE_WIDEOP_OPCODE_LMULS(OP) \
4784 CASE_WIDEOP_OPCODE_COMMON(OP, MF8): \
4785 case CASE_WIDEOP_OPCODE_COMMON(OP, MF4): \
4786 case CASE_WIDEOP_OPCODE_COMMON(OP, MF2): \
4787 case CASE_WIDEOP_OPCODE_COMMON(OP, M1): \
4788 case CASE_WIDEOP_OPCODE_COMMON(OP, M2): \
4789 case CASE_WIDEOP_OPCODE_COMMON(OP, M4)
4791#define CASE_WIDEOP_CHANGE_OPCODE_COMMON(OP, LMUL) \
4792 case RISCV::PseudoV##OP##_##LMUL##_TIED: \
4793 NewOpc = RISCV::PseudoV##OP##_##LMUL; \
4796#define CASE_WIDEOP_CHANGE_OPCODE_LMULS(OP) \
4797 CASE_WIDEOP_CHANGE_OPCODE_COMMON(OP, MF8) \
4798 CASE_WIDEOP_CHANGE_OPCODE_COMMON(OP, MF4) \
4799 CASE_WIDEOP_CHANGE_OPCODE_COMMON(OP, MF2) \
4800 CASE_WIDEOP_CHANGE_OPCODE_COMMON(OP, M1) \
4801 CASE_WIDEOP_CHANGE_OPCODE_COMMON(OP, M2) \
4802 CASE_WIDEOP_CHANGE_OPCODE_COMMON(OP, M4)
4805#define CASE_FP_WIDEOP_OPCODE_COMMON(OP, LMUL, SEW) \
4806 RISCV::PseudoV##OP##_##LMUL##_##SEW##_TIED
4808#define CASE_FP_WIDEOP_OPCODE_LMULS(OP) \
4809 CASE_FP_WIDEOP_OPCODE_COMMON(OP, MF4, E16): \
4810 case CASE_FP_WIDEOP_OPCODE_COMMON(OP, MF2, E16): \
4811 case CASE_FP_WIDEOP_OPCODE_COMMON(OP, MF2, E32): \
4812 case CASE_FP_WIDEOP_OPCODE_COMMON(OP, M1, E16): \
4813 case CASE_FP_WIDEOP_OPCODE_COMMON(OP, M1, E32): \
4814 case CASE_FP_WIDEOP_OPCODE_COMMON(OP, M2, E16): \
4815 case CASE_FP_WIDEOP_OPCODE_COMMON(OP, M2, E32): \
4816 case CASE_FP_WIDEOP_OPCODE_COMMON(OP, M4, E16): \
4817 case CASE_FP_WIDEOP_OPCODE_COMMON(OP, M4, E32) \
4819#define CASE_FP_WIDEOP_CHANGE_OPCODE_COMMON(OP, LMUL, SEW) \
4820 case RISCV::PseudoV##OP##_##LMUL##_##SEW##_TIED: \
4821 NewOpc = RISCV::PseudoV##OP##_##LMUL##_##SEW; \
4824#define CASE_FP_WIDEOP_CHANGE_OPCODE_LMULS(OP) \
4825 CASE_FP_WIDEOP_CHANGE_OPCODE_COMMON(OP, MF4, E16) \
4826 CASE_FP_WIDEOP_CHANGE_OPCODE_COMMON(OP, MF2, E16) \
4827 CASE_FP_WIDEOP_CHANGE_OPCODE_COMMON(OP, MF2, E32) \
4828 CASE_FP_WIDEOP_CHANGE_OPCODE_COMMON(OP, M1, E16) \
4829 CASE_FP_WIDEOP_CHANGE_OPCODE_COMMON(OP, M1, E32) \
4830 CASE_FP_WIDEOP_CHANGE_OPCODE_COMMON(OP, M2, E16) \
4831 CASE_FP_WIDEOP_CHANGE_OPCODE_COMMON(OP, M2, E32) \
4832 CASE_FP_WIDEOP_CHANGE_OPCODE_COMMON(OP, M4, E16) \
4833 CASE_FP_WIDEOP_CHANGE_OPCODE_COMMON(OP, M4, E32) \
4835#define CASE_FP_WIDEOP_OPCODE_LMULS_ALT(OP) \
4836 CASE_FP_WIDEOP_OPCODE_COMMON(OP, MF4, E16): \
4837 case CASE_FP_WIDEOP_OPCODE_COMMON(OP, MF2, E16): \
4838 case CASE_FP_WIDEOP_OPCODE_COMMON(OP, M1, E16): \
4839 case CASE_FP_WIDEOP_OPCODE_COMMON(OP, M2, E16): \
4840 case CASE_FP_WIDEOP_OPCODE_COMMON(OP, M4, E16)
4842#define CASE_FP_WIDEOP_CHANGE_OPCODE_LMULS_ALT(OP) \
4843 CASE_FP_WIDEOP_CHANGE_OPCODE_COMMON(OP, MF4, E16) \
4844 CASE_FP_WIDEOP_CHANGE_OPCODE_COMMON(OP, MF2, E16) \
4845 CASE_FP_WIDEOP_CHANGE_OPCODE_COMMON(OP, M1, E16) \
4846 CASE_FP_WIDEOP_CHANGE_OPCODE_COMMON(OP, M2, E16) \
4847 CASE_FP_WIDEOP_CHANGE_OPCODE_COMMON(OP, M4, E16)
4854 switch (
MI.getOpcode()) {
4862 MI.getNumExplicitOperands() == 7 &&
4863 "Expect 7 explicit operands rd, rs2, rs1, rm, vl, sew, policy");
4870 switch (
MI.getOpcode()) {
4882 .
add(
MI.getOperand(0))
4884 .
add(
MI.getOperand(1))
4885 .
add(
MI.getOperand(2))
4886 .
add(
MI.getOperand(3))
4887 .
add(
MI.getOperand(4))
4888 .
add(
MI.getOperand(5))
4889 .
add(
MI.getOperand(6));
4898 MI.getNumExplicitOperands() == 6);
4905 switch (
MI.getOpcode()) {
4917 .
add(
MI.getOperand(0))
4919 .
add(
MI.getOperand(1))
4920 .
add(
MI.getOperand(2))
4921 .
add(
MI.getOperand(3))
4922 .
add(
MI.getOperand(4))
4923 .
add(
MI.getOperand(5));
4930 unsigned NumOps =
MI.getNumOperands();
4933 if (
Op.isReg() &&
Op.isKill())
4941 if (
MI.getOperand(0).isEarlyClobber()) {
4955#undef CASE_WIDEOP_OPCODE_COMMON
4956#undef CASE_WIDEOP_OPCODE_LMULS
4957#undef CASE_WIDEOP_CHANGE_OPCODE_COMMON
4958#undef CASE_WIDEOP_CHANGE_OPCODE_LMULS
4959#undef CASE_FP_WIDEOP_OPCODE_COMMON
4960#undef CASE_FP_WIDEOP_OPCODE_LMULS
4961#undef CASE_FP_WIDEOP_CHANGE_OPCODE_COMMON
4962#undef CASE_FP_WIDEOP_CHANGE_OPCODE_LMULS
4971 if (ShiftAmount == 0)
4977 }
else if (
int ShXAmount, ShiftAmount;
4979 (ShXAmount =
isShifted359(Amount, ShiftAmount)) != 0) {
4982 switch (ShXAmount) {
4984 Opc = RISCV::SH1ADD;
4987 Opc = RISCV::SH2ADD;
4990 Opc = RISCV::SH3ADD;
5026 }
else if (
STI.hasStdExtZmmul()) {
5036 for (
uint32_t ShiftAmount = 0; Amount >> ShiftAmount; ShiftAmount++) {
5037 if (Amount & (1U << ShiftAmount)) {
5041 .
addImm(ShiftAmount - PrevShiftAmount)
5043 if (Amount >> (ShiftAmount + 1)) {
5057 PrevShiftAmount = ShiftAmount;
5060 assert(Acc &&
"Expected valid accumulator");
5070 static const std::pair<MachineMemOperand::Flags, const char *> TargetFlags[] =
5078 ?
STI.getTailDupAggressiveThreshold()
5085 unsigned Opcode =
MI.getOpcode();
5086 if (!RISCVVPseudosTable::getPseudoInfo(Opcode) &&
5095 return MI.isCopy() &&
MI.getOperand(0).getReg().isPhysical() &&
5097 TRI->getMinimalPhysRegClass(
MI.getOperand(0).getReg()));
5100std::optional<std::pair<unsigned, unsigned>>
5104 return std::nullopt;
5105 case RISCV::PseudoVSPILL2_M1:
5106 case RISCV::PseudoVRELOAD2_M1:
5107 return std::make_pair(2u, 1u);
5108 case RISCV::PseudoVSPILL2_M2:
5109 case RISCV::PseudoVRELOAD2_M2:
5110 return std::make_pair(2u, 2u);
5111 case RISCV::PseudoVSPILL2_M4:
5112 case RISCV::PseudoVRELOAD2_M4:
5113 return std::make_pair(2u, 4u);
5114 case RISCV::PseudoVSPILL3_M1:
5115 case RISCV::PseudoVRELOAD3_M1:
5116 return std::make_pair(3u, 1u);
5117 case RISCV::PseudoVSPILL3_M2:
5118 case RISCV::PseudoVRELOAD3_M2:
5119 return std::make_pair(3u, 2u);
5120 case RISCV::PseudoVSPILL4_M1:
5121 case RISCV::PseudoVRELOAD4_M1:
5122 return std::make_pair(4u, 1u);
5123 case RISCV::PseudoVSPILL4_M2:
5124 case RISCV::PseudoVRELOAD4_M2:
5125 return std::make_pair(4u, 2u);
5126 case RISCV::PseudoVSPILL5_M1:
5127 case RISCV::PseudoVRELOAD5_M1:
5128 return std::make_pair(5u, 1u);
5129 case RISCV::PseudoVSPILL6_M1:
5130 case RISCV::PseudoVRELOAD6_M1:
5131 return std::make_pair(6u, 1u);
5132 case RISCV::PseudoVSPILL7_M1:
5133 case RISCV::PseudoVRELOAD7_M1:
5134 return std::make_pair(7u, 1u);
5135 case RISCV::PseudoVSPILL8_M1:
5136 case RISCV::PseudoVRELOAD8_M1:
5137 return std::make_pair(8u, 1u);
5142 int16_t MI1FrmOpIdx =
5143 RISCV::getNamedOperandIdx(MI1.
getOpcode(), RISCV::OpName::frm);
5144 int16_t MI2FrmOpIdx =
5145 RISCV::getNamedOperandIdx(MI2.
getOpcode(), RISCV::OpName::frm);
5146 if (MI1FrmOpIdx < 0 || MI2FrmOpIdx < 0)
5153std::optional<unsigned>
5157 return std::nullopt;
5160 case RISCV::VSLL_VX:
5161 case RISCV::VSRL_VX:
5162 case RISCV::VSRA_VX:
5164 case RISCV::VSSRL_VX:
5165 case RISCV::VSSRA_VX:
5167 case RISCV::VROL_VX:
5168 case RISCV::VROR_VX:
5173 case RISCV::VNSRL_WX:
5174 case RISCV::VNSRA_WX:
5176 case RISCV::VNCLIPU_WX:
5177 case RISCV::VNCLIP_WX:
5179 case RISCV::VWSLL_VX:
5184 case RISCV::VADD_VX:
5185 case RISCV::VSUB_VX:
5186 case RISCV::VRSUB_VX:
5188 case RISCV::VWADDU_VX:
5189 case RISCV::VWSUBU_VX:
5190 case RISCV::VWADD_VX:
5191 case RISCV::VWSUB_VX:
5192 case RISCV::VWADDU_WX:
5193 case RISCV::VWSUBU_WX:
5194 case RISCV::VWADD_WX:
5195 case RISCV::VWSUB_WX:
5197 case RISCV::VADC_VXM:
5198 case RISCV::VADC_VIM:
5199 case RISCV::VMADC_VXM:
5200 case RISCV::VMADC_VIM:
5201 case RISCV::VMADC_VX:
5202 case RISCV::VSBC_VXM:
5203 case RISCV::VMSBC_VXM:
5204 case RISCV::VMSBC_VX:
5206 case RISCV::VAND_VX:
5208 case RISCV::VXOR_VX:
5210 case RISCV::VMSEQ_VX:
5211 case RISCV::VMSNE_VX:
5212 case RISCV::VMSLTU_VX:
5213 case RISCV::VMSLT_VX:
5214 case RISCV::VMSLEU_VX:
5215 case RISCV::VMSLE_VX:
5216 case RISCV::VMSGTU_VX:
5217 case RISCV::VMSGT_VX:
5219 case RISCV::VMINU_VX:
5220 case RISCV::VMIN_VX:
5221 case RISCV::VMAXU_VX:
5222 case RISCV::VMAX_VX:
5224 case RISCV::VMUL_VX:
5225 case RISCV::VMULH_VX:
5226 case RISCV::VMULHU_VX:
5227 case RISCV::VMULHSU_VX:
5229 case RISCV::VDIVU_VX:
5230 case RISCV::VDIV_VX:
5231 case RISCV::VREMU_VX:
5232 case RISCV::VREM_VX:
5234 case RISCV::VWMUL_VX:
5235 case RISCV::VWMULU_VX:
5236 case RISCV::VWMULSU_VX:
5238 case RISCV::VMACC_VX:
5239 case RISCV::VNMSAC_VX:
5240 case RISCV::VMADD_VX:
5241 case RISCV::VNMSUB_VX:
5243 case RISCV::VWMACCU_VX:
5244 case RISCV::VWMACC_VX:
5245 case RISCV::VWMACCSU_VX:
5246 case RISCV::VWMACCUS_VX:
5248 case RISCV::VMERGE_VXM:
5250 case RISCV::VMV_V_X:
5252 case RISCV::VSADDU_VX:
5253 case RISCV::VSADD_VX:
5254 case RISCV::VSSUBU_VX:
5255 case RISCV::VSSUB_VX:
5257 case RISCV::VAADDU_VX:
5258 case RISCV::VAADD_VX:
5259 case RISCV::VASUBU_VX:
5260 case RISCV::VASUB_VX:
5262 case RISCV::VSMUL_VX:
5264 case RISCV::VMV_S_X:
5266 case RISCV::VANDN_VX:
5267 return 1U << Log2SEW;
5273 RISCVVPseudosTable::getPseudoInfo(RVVPseudoOpcode);
5276 return RVV->BaseInstr;
5286 unsigned Scaled = Log2SEW + (DestEEW - 1);
5300 return std::nullopt;
5305 assert((LHS.isImm() || LHS.getParent()->getMF()->getRegInfo().isSSA()) &&
5306 (RHS.isImm() || RHS.getParent()->getMF()->getRegInfo().isSSA()));
5307 if (LHS.isReg() && RHS.isReg() && LHS.getReg().isVirtual() &&
5308 LHS.getReg() == RHS.getReg())
5312 if (LHS.isImm() && LHS.getImm() == 0)
5318 if (!LHSImm || !RHSImm)
5320 return LHSImm <= RHSImm;
5332 : LHS(LHS), RHS(RHS),
Cond(
Cond.begin(),
Cond.end()) {}
5334 bool shouldIgnoreForPipelining(
const MachineInstr *
MI)
const override {
5344 std::optional<bool> createTripCountGreaterCondition(
5345 int TC, MachineBasicBlock &
MBB,
5346 SmallVectorImpl<MachineOperand> &CondParam)
override {
5354 void setPreheader(MachineBasicBlock *NewPreheader)
override {}
5356 void adjustTripCount(
int TripCountAdjust)
override {}
5360std::unique_ptr<TargetInstrInfo::PipelinerLoopInfo>
5368 if (
TBB == LoopBB && FBB == LoopBB)
5375 assert((
TBB == LoopBB || FBB == LoopBB) &&
5376 "The Loop must be a single-basic-block loop");
5387 if (!Reg.isVirtual())
5394 if (LHS && LHS->isPHI())
5396 if (RHS && RHS->isPHI())
5399 return std::make_unique<RISCVPipelinerLoopInfo>(LHS, RHS,
Cond);
5405 Opc = RVVMCOpcode ? RVVMCOpcode :
Opc;
5422 case RISCV::FDIV_H_INX:
5423 case RISCV::FDIV_S_INX:
5424 case RISCV::FDIV_D_INX:
5425 case RISCV::FDIV_D_IN32X:
5426 case RISCV::FSQRT_H:
5427 case RISCV::FSQRT_S:
5428 case RISCV::FSQRT_D:
5429 case RISCV::FSQRT_H_INX:
5430 case RISCV::FSQRT_S_INX:
5431 case RISCV::FSQRT_D_INX:
5432 case RISCV::FSQRT_D_IN32X:
5434 case RISCV::VDIV_VV:
5435 case RISCV::VDIV_VX:
5436 case RISCV::VDIVU_VV:
5437 case RISCV::VDIVU_VX:
5438 case RISCV::VREM_VV:
5439 case RISCV::VREM_VX:
5440 case RISCV::VREMU_VV:
5441 case RISCV::VREMU_VX:
5443 case RISCV::VFDIV_VV:
5444 case RISCV::VFDIV_VF:
5445 case RISCV::VFRDIV_VF:
5446 case RISCV::VFSQRT_V:
5447 case RISCV::VFRSQRT7_V:
5453 if (
MI->getOpcode() != TargetOpcode::COPY)
5458 Register DstReg =
MI->getOperand(0).getReg();
5461 :
TRI->getMinimalPhysRegClass(DstReg);
5471 auto [RCLMul, RCFractional] =
5473 return (!RCFractional && LMul == RCLMul) || (RCFractional && LMul == 1);
5477 if (
MI.memoperands_empty())
5492 if (MO.getReg().isPhysical())
5495 if (MO.getReg().isPhysical())
5497 bool SawStore =
false;
5500 if (
II->definesRegister(PhysReg,
nullptr))
5503 if (
II->definesRegister(PhysReg,
nullptr) ||
5504 II->readsRegister(PhysReg,
nullptr))
5506 if (
II->mayStore()) {
MachineInstrBuilder MachineInstrBuilder & DefMI
static bool forwardCopyWillClobberTuple(unsigned DestReg, unsigned SrcReg, unsigned NumRegs)
static void parseCondBranch(MachineInstr *LastInst, MachineBasicBlock *&Target, SmallVectorImpl< MachineOperand > &Cond)
@ MachineOutlinerTailCall
Emit a save, restore, call, and return.
@ MachineOutlinerRegSave
Emit a call and tail-call.
assert(UImm &&(UImm !=~static_cast< T >(0)) &&"Invalid immediate!")
SmallVector< int16_t, MAX_SRC_OPERANDS_NUM > OperandIndices
MachineBasicBlock MachineBasicBlock::iterator DebugLoc DL
MachineBasicBlock MachineBasicBlock::iterator MBBI
#define clEnumValN(ENUMVAL, FLAGNAME, DESC)
const HexagonInstrInfo * TII
Module.h This file contains the declarations for the Module class.
const size_t AbstractManglingParser< Derived, Alloc >::NumOps
const AbstractManglingParser< Derived, Alloc >::OperatorInfo AbstractManglingParser< Derived, Alloc >::Ops[]
Register const TargetRegisterInfo * TRI
Promote Memory to Register
This file provides utility analysis objects describing memory locations.
MachineInstr unsigned OpIdx
uint64_t IntrinsicInst * II
static bool cannotInsertTailCall(const MachineBasicBlock &MBB)
#define CASE_VFMA_CHANGE_OPCODE_SPLATS(OLDOP, NEWOP)
#define CASE_FP_WIDEOP_CHANGE_OPCODE_LMULS_ALT(OP)
#define CASE_FP_WIDEOP_OPCODE_LMULS(OP)
#define CASE_OPERAND_SIMM(NUM)
static std::optional< unsigned > getLMULForRVVWholeLoadStore(unsigned Opcode)
#define CASE_VFMA_CHANGE_OPCODE_VV(OLDOP, NEWOP)
static unsigned getFPFusedMultiplyOpcode(unsigned RootOpc, unsigned Pattern)
std::optional< unsigned > getFoldedOpcode(MachineFunction &MF, MachineInstr &MI, ArrayRef< unsigned > Ops, const RISCVSubtarget &ST)
#define RVV_OPC_LMUL_CASE(OPC, INV)
#define CASE_FP_WIDEOP_CHANGE_OPCODE_LMULS(OP)
static void combineFPFusedMultiply(MachineInstr &Root, MachineInstr &Prev, unsigned Pattern, SmallVectorImpl< MachineInstr * > &InsInstrs, SmallVectorImpl< MachineInstr * > &DelInstrs)
static unsigned getAddendOperandIdx(unsigned Pattern)
#define CASE_RVV_OPCODE_UNMASK(OP)
#define CASE_WIDEOP_CHANGE_OPCODE_LMULS(OP)
static cl::opt< bool > PreferWholeRegisterMove("riscv-prefer-whole-register-move", cl::init(false), cl::Hidden, cl::desc("Prefer whole register move for vector registers."))
#define CASE_VFMA_SPLATS(OP)
unsigned getPredicatedOpcode(unsigned Opcode)
#define CASE_FP_WIDEOP_OPCODE_LMULS_ALT(OP)
#define CASE_WIDEOP_OPCODE_LMULS(OP)
static bool isMIReadsReg(const MachineInstr &MI, const TargetRegisterInfo *TRI, MCRegister RegNo)
#define OPCODE_LMUL_MASK_CASE(OPC)
#define CASE_OPERAND_UIMM_LSB_ZEROS(BITS, SUFFIX)
static bool isFSUB(unsigned Opc)
#define CASE_VMA_CHANGE_OPCODE_LMULS(OLDOP, NEWOP, TYPE)
#define CASE_RVV_OPCODE(OP)
static std::optional< int64_t > getEffectiveImm(const MachineOperand &MO)
#define CASE_VFMA_OPCODE_VV(OP)
static cl::opt< bool > OutlinerEnableRegSave("riscv-outliner-regsave", cl::init(true), cl::Hidden, cl::desc("Enable RegSave strategy in machine outliner (save X5 to a " "temporary register when X5 is live across outlined calls)."))
MachineOutlinerConstructionID
#define CASE_RVV_OPCODE_WIDEN(OP)
static unsigned getLoadPredicatedOpcode(unsigned Opcode)
static unsigned getSHXADDUWShiftAmount(unsigned Opc)
#define CASE_VMA_OPCODE_LMULS(OP, TYPE)
static bool isConvertibleToVMV_V_V(const RISCVSubtarget &STI, const MachineBasicBlock &MBB, MachineBasicBlock::const_iterator MBBI, MachineBasicBlock::const_iterator &DefMBBI, RISCVVType::VLMUL LMul)
static bool isFMUL(unsigned Opc)
static unsigned getInverseXqcicmOpcode(unsigned Opcode)
static bool getFPPatterns(MachineInstr &Root, SmallVectorImpl< unsigned > &Patterns, bool DoRegPressureReduce)
#define OPCODE_LMUL_CASE(OPC)
#define CASE_OPERAND_UIMM(NUM)
static Register findRegisterToSaveX5To(outliner::Candidate &C, const TargetRegisterInfo &TRI)
static bool canCombineShiftIntoShXAdd(const MachineBasicBlock &MBB, const MachineOperand &MO, unsigned OuterShiftAmt)
Utility routine that checks if.
static bool isCandidatePatchable(const MachineBasicBlock &MBB)
static bool isFADD(unsigned Opc)
static void genShXAddAddShift(MachineInstr &Root, unsigned AddOpIdx, SmallVectorImpl< MachineInstr * > &InsInstrs, SmallVectorImpl< MachineInstr * > &DelInstrs, DenseMap< Register, unsigned > &InstrIdxForVirtReg)
static bool isLoadImm(const MachineInstr *MI, int64_t &Imm)
static bool isMIModifiesReg(const MachineInstr &MI, const TargetRegisterInfo *TRI, MCRegister RegNo)
#define CASE_RVV_OPCODE_LMUL(OP, LMUL)
static bool canCombineFPFusedMultiply(const MachineInstr &Root, const MachineOperand &MO, bool DoRegPressureReduce)
static bool getSHXADDPatterns(const MachineInstr &Root, SmallVectorImpl< unsigned > &Patterns)
static bool getFPFusedMultiplyPatterns(MachineInstr &Root, SmallVectorImpl< unsigned > &Patterns, bool DoRegPressureReduce)
static cl::opt< MachineTraceStrategy > ForceMachineCombinerStrategy("riscv-force-machine-combiner-strategy", cl::Hidden, cl::desc("Force machine combiner to use a specific strategy for machine " "trace metrics evaluation."), cl::init(MachineTraceStrategy::TS_NumStrategies), cl::values(clEnumValN(MachineTraceStrategy::TS_Local, "local", "Local strategy."), clEnumValN(MachineTraceStrategy::TS_MinInstrCount, "min-instr", "MinInstrCount strategy.")))
static unsigned getSHXADDShiftAmount(unsigned Opc)
#define CASE_RVV_OPCODE_MASK(OP)
#define RVV_OPC_LMUL_MASK_CASE(OPC, INV)
static MachineInstr * canFoldAsPredicatedOp(Register Reg, const MachineRegisterInfo &MRI, const TargetInstrInfo *TII, const RISCVSubtarget &STI)
Identify instructions that can be folded into a CCMOV instruction, and return the defining instructio...
const SmallVectorImpl< MachineOperand > MachineBasicBlock * TBB
const SmallVectorImpl< MachineOperand > & Cond
This file declares the machine register scavenger class.
static bool memOpsHaveSameBasePtr(const MachineInstr &MI1, ArrayRef< const MachineOperand * > BaseOps1, const MachineInstr &MI2, ArrayRef< const MachineOperand * > BaseOps2)
static bool contains(SmallPtrSetImpl< ConstantExpr * > &Cache, ConstantExpr *Expr, Constant *C)
This file defines the SmallVector class.
This file defines the 'Statistic' class, which is designed to be an easy way to expose various metric...
#define STATISTIC(VARNAME, DESC)
static TableGen::Emitter::Opt Y("gen-skeleton-entry", EmitSkeleton, "Generate example skeleton entry")
static bool canCombine(MachineBasicBlock &MBB, MachineOperand &MO, unsigned CombineOpc=0)
ArrayRef - Represent a constant reference to an array (0 or more elements consecutively in memory),...
const T & front() const
front - Get the first element.
bool empty() const
empty - Check if the array is empty.
static LLVM_ABI DILocation * getMergedLocation(DILocation *LocA, DILocation *LocB)
Attempts to merge LocA and LocB into a single location; see DebugLoc::getMergedLocation for more deta...
std::pair< iterator, bool > insert(const std::pair< KeyT, ValueT > &KV)
bool hasMinSize() const
Optimize this function for minimum size (-Oz).
LiveInterval - This class represents the liveness of a register, or stack slot.
LiveInterval & getInterval(Register Reg)
SlotIndex ReplaceMachineInstrInMaps(MachineInstr &MI, MachineInstr &NewMI)
const Segment * getSegmentContaining(SlotIndex Idx) const
Return the segment that contains the specified index, or null if there is none.
LLVM_ABI void replaceKillInstruction(Register Reg, MachineInstr &OldMI, MachineInstr &NewMI)
replaceKillInstruction - Update register kill info by replacing a kill instruction with a new one.
static LocationSize precise(uint64_t Value)
TypeSize getValue() const
MCInstBuilder & addReg(MCRegister Reg)
Add a new register operand.
MCInstBuilder & addImm(int64_t Val)
Add a new integer immediate operand.
Instances of this class represent a single low-level machine instruction.
Describe properties that are true of each instruction in the target description file.
unsigned getNumOperands() const
Return the number of declared MachineOperands for this MachineInstruction.
bool isConditionalBranch() const
Return true if this is a branch which may fall through to the next instruction or may transfer contro...
This holds information about one operand of a machine instruction, indicating the register class for ...
Wrapper class representing physical registers. Should be passed by value.
const FeatureBitset & getFeatureBits() const
MachineInstrBundleIterator< const MachineInstr > const_iterator
MachineInstrBundleIterator< MachineInstr, true > reverse_iterator
Instructions::const_iterator const_instr_iterator
const MachineFunction * getParent() const
Return the MachineFunction containing this basic block.
MachineInstrBundleIterator< MachineInstr > iterator
MachineInstrBundleIterator< const MachineInstr, true > const_reverse_iterator
The MachineFrameInfo class represents an abstract stack frame until prolog/epilog code is inserted.
void setStackID(int ObjectIdx, uint8_t ID)
Align getObjectAlign(int ObjectIdx) const
Return the alignment of the specified stack object.
int64_t getObjectSize(int ObjectIdx) const
Return the size of the specified object.
const TargetSubtargetInfo & getSubtarget() const
getSubtarget - Return the subtarget for which this machine code is being compiled.
StringRef getName() const
getName - Return the name of the corresponding LLVM function.
MachineMemOperand * getMachineMemOperand(MachinePointerInfo PtrInfo, MachineMemOperand::Flags f, LLT MemTy, Align base_alignment, const AAMDNodes &AAInfo=AAMDNodes(), const MDNode *Ranges=nullptr, SyncScope::ID SSID=SyncScope::System, AtomicOrdering Ordering=AtomicOrdering::NotAtomic, AtomicOrdering FailureOrdering=AtomicOrdering::NotAtomic)
getMachineMemOperand - Allocate a new MachineMemOperand.
MachineFrameInfo & getFrameInfo()
getFrameInfo - Return the frame info object for the current function.
MachineRegisterInfo & getRegInfo()
getRegInfo - Return information about the registers currently in use.
const DataLayout & getDataLayout() const
Return the DataLayout attached to the Module associated to this MF.
Function & getFunction()
Return the LLVM function that this machine code represents.
Ty * getInfo()
getInfo - Keep track of various per-function pieces of information for backends that would like to do...
const TargetMachine & getTarget() const
getTarget - Return the target machine this machine code is compiled with
const MachineInstrBuilder & setMemRefs(ArrayRef< MachineMemOperand * > MMOs) const
const MachineInstrBuilder & addUse(Register RegNo, RegState Flags={}, unsigned SubReg=0) const
Add a virtual register use operand.
const MachineInstrBuilder & addReg(Register RegNo, RegState Flags={}, unsigned SubReg=0) const
Add a new virtual register operand.
const MachineInstrBuilder & setMIFlag(MachineInstr::MIFlag Flag) const
const MachineInstrBuilder & addImm(int64_t Val) const
Add a new immediate operand.
const MachineInstrBuilder & add(const MachineOperand &MO) const
const MachineInstrBuilder & addFrameIndex(int Idx) const
const MachineInstrBuilder & addMBB(MachineBasicBlock *MBB, unsigned TargetFlags=0) const
const MachineInstrBuilder & cloneMemRefs(const MachineInstr &OtherMI) const
const MachineInstrBuilder & setMIFlags(unsigned Flags) const
const MachineInstrBuilder & copyImplicitOps(const MachineInstr &OtherMI) const
Copy all the implicit operands from OtherMI onto this one.
const MachineInstrBuilder & addMemOperand(MachineMemOperand *MMO) const
reverse_iterator getReverse() const
Get a reverse iterator to the same node.
Representation of each machine instruction.
unsigned getOpcode() const
Returns the opcode of this MachineInstr.
bool isReturn(QueryType Type=AnyInBundle) const
bool mayLoadOrStore(QueryType Type=AnyInBundle) const
Return true if this instruction could possibly read or modify memory.
const MachineBasicBlock * getParent() const
filtered_mop_range all_defs()
Returns an iterator range over all operands that are (explicit or implicit) register defs.
bool getFlag(MIFlag Flag) const
Return whether an MI flag is set.
LLVM_ABI bool isSafeToMove(bool &SawStore) const
Return true if it is safe to move this instruction.
LLVM_ABI unsigned getNumExplicitOperands() const
Returns the number of non-implicit operands.
bool modifiesRegister(Register Reg, const TargetRegisterInfo *TRI) const
Return true if the MachineInstr modifies (fully define or partially define) the specified register.
bool mayLoad(QueryType Type=AnyInBundle) const
Return true if this instruction could possibly read memory.
const MCInstrDesc & getDesc() const
Returns the target instruction descriptor of this MachineInstr.
LLVM_ABI bool hasUnmodeledSideEffects() const
Return true if this instruction has side effects that are not modeled by mayLoad / mayStore,...
bool hasOneMemOperand() const
Return true if this instruction has exactly one MachineMemOperand.
mmo_iterator memoperands_begin() const
Access to memory operands of the instruction.
LLVM_ABI bool hasOrderedMemoryRef() const
Return true if this instruction may have an ordered or volatile memory reference, or if the informati...
LLVM_ABI const MachineFunction * getMF() const
Return the function that contains the basic block that this instruction belongs to.
ArrayRef< MachineMemOperand * > memoperands() const
Access to memory operands of the instruction.
const DebugLoc & getDebugLoc() const
Returns the debug location id of this MachineInstr.
filtered_mop_range all_uses()
Returns an iterator range over all operands that are (explicit or implicit) register uses.
const MachineOperand & getOperand(unsigned i) const
uint32_t getFlags() const
Return the MI flags bitvector.
LLVM_ABI void clearKillInfo()
Clears kill flags on all operands.
A description of a memory reference used in the backend.
bool isNonTemporal() const
@ MOLoad
The memory access reads data.
@ MOStore
The memory access writes data.
This class contains meta information specific to a module.
MachineOperand class - Representation of each machine instruction operand.
bool isReg() const
isReg - Tests if this is a MO_Register operand.
MachineBasicBlock * getMBB() const
bool isImm() const
isImm - Tests if this is a MO_Immediate operand.
MachineInstr * getParent()
getParent - Return the instruction that this operand belongs to.
static MachineOperand CreateImm(int64_t Val)
MachineOperandType getType() const
getType - Returns the MachineOperandType for this operand.
Register getReg() const
getReg - Returns the register number.
bool isFI() const
isFI - Tests if this is a MO_FrameIndex operand.
LLVM_ABI bool isIdenticalTo(const MachineOperand &Other) const
Returns true if this operand is identical to the specified operand except for liveness related flags ...
@ MO_Immediate
Immediate operand.
@ MO_Register
Register operand.
MachineRegisterInfo - Keep track of information for virtual and physical registers,...
LLVM_ABI bool hasOneNonDBGUse(Register RegNo) const
hasOneNonDBGUse - Return true if there is exactly one non-Debug use of the specified register.
const TargetRegisterClass * getRegClass(Register Reg) const
Return the register class of the specified virtual register.
LLVM_ABI void clearKillFlags(Register Reg) const
clearKillFlags - Iterate over all the uses of the given register and clear the kill flag from the Mac...
LLVM_ABI MachineInstr * getVRegDef(Register Reg) const
getVRegDef - Return the machine instr that defines the specified virtual register or null if none is ...
bool isReserved(MCRegister PhysReg) const
isReserved - Returns true when PhysReg is a reserved register.
LLVM_ABI Register createVirtualRegister(const TargetRegisterClass *RegClass, StringRef Name="")
createVirtualRegister - Create and return a new virtual register in the function with the specified r...
bool hasOneUse(Register RegNo) const
hasOneUse - Return true if there is exactly one instruction using the specified register.
LLVM_ABI void clearVirtRegs()
clearVirtRegs - Remove all virtual registers (after physreg assignment).
const TargetRegisterInfo * getTargetRegisterInfo() const
LLVM_ABI bool isConstantPhysReg(MCRegister PhysReg) const
Returns true if PhysReg is unallocatable and constant throughout the function.
LLVM_ABI const TargetRegisterClass * constrainRegClass(Register Reg, const TargetRegisterClass *RC, unsigned MinNumRegs=0)
constrainRegClass - Constrain the register class of the specified virtual register to be a common sub...
LLVM_ABI void replaceRegWith(Register FromReg, Register ToReg)
replaceRegWith - Replace all instances of FromReg with ToReg in the machine function.
LLVM_ABI MachineInstr * getUniqueVRegDef(Register Reg) const
getUniqueVRegDef - Return the unique machine instr that defines the specified virtual register or nul...
A Module instance is used to store all the information related to an LLVM module.
MI-level patchpoint operands.
uint32_t getNumPatchBytes() const
Return the number of patchable bytes the given patchpoint should emit.
void storeRegToStackSlot(MachineBasicBlock &MBB, MachineBasicBlock::iterator MBBI, Register SrcReg, bool IsKill, int FrameIndex, const TargetRegisterClass *RC, Register VReg, MachineInstr::MIFlag Flags=MachineInstr::NoFlags) const override
MachineInstr * convertToThreeAddress(MachineInstr &MI, LiveVariables *LV, LiveIntervals *LIS) const override
Register isLoadFromStackSlot(const MachineInstr &MI, int &FrameIndex) const override
std::optional< std::unique_ptr< outliner::OutlinedFunction > > getOutliningCandidateInfo(const MachineModuleInfo &MMI, std::vector< outliner::Candidate > &RepeatedSequenceLocs, unsigned MinRepeats) const override
unsigned removeBranch(MachineBasicBlock &MBB, int *BytesRemoved=nullptr) const override
void genAlternativeCodeSequence(MachineInstr &Root, unsigned Pattern, SmallVectorImpl< MachineInstr * > &InsInstrs, SmallVectorImpl< MachineInstr * > &DelInstrs, DenseMap< Register, unsigned > &InstrIdxForVirtReg) const override
void movImm(MachineBasicBlock &MBB, MachineBasicBlock::iterator MBBI, const DebugLoc &DL, Register DstReg, uint64_t Val, MachineInstr::MIFlag Flag=MachineInstr::NoFlags, bool DstRenamable=false, bool DstIsDead=false) const
MachineInstr * emitLdStWithAddr(MachineInstr &MemI, const ExtAddrMode &AM) const override
void mulImm(MachineFunction &MF, MachineBasicBlock &MBB, MachineBasicBlock::iterator II, const DebugLoc &DL, Register DestReg, uint32_t Amt, MachineInstr::MIFlag Flag) const
Generate code to multiply the value in DestReg by Amt - handles all the common optimizations for this...
static bool isPairableLdStInstOpc(unsigned Opc)
Return true if pairing the given load or store may be paired with another.
RISCVInstrInfo(const RISCVSubtarget &STI)
void loadRegFromStackSlot(MachineBasicBlock &MBB, MachineBasicBlock::iterator MBBI, Register DstReg, int FrameIndex, const TargetRegisterClass *RC, Register VReg, unsigned SubReg=0, MachineInstr::MIFlag Flags=MachineInstr::NoFlags) const override
bool isFunctionSafeToOutlineFrom(MachineFunction &MF, bool OutlineFromLinkOnceODRs) const override
std::unique_ptr< TargetInstrInfo::PipelinerLoopInfo > analyzeLoopForPipelining(MachineBasicBlock *LoopBB) const override
unsigned insertBranch(MachineBasicBlock &MBB, MachineBasicBlock *TBB, MachineBasicBlock *FBB, ArrayRef< MachineOperand > Cond, const DebugLoc &dl, int *BytesAdded=nullptr) const override
bool hasReassociableSibling(const MachineInstr &Inst, bool &Commuted) const override
static bool isLdStSafeToPair(const MachineInstr &LdSt, const TargetRegisterInfo *TRI)
void copyPhysRegVector(MachineBasicBlock &MBB, MachineBasicBlock::iterator MBBI, const DebugLoc &DL, MCRegister DstReg, MCRegister SrcReg, bool KillSrc, const TargetRegisterClass *RegClass) const
bool isReMaterializableImpl(const MachineInstr &MI) const override
MachineInstr * optimizeSelect(MachineInstr &MI, SmallPtrSetImpl< MachineInstr * > &SeenMIs, bool) const override
bool isVRegCopy(const MachineInstr *MI, unsigned LMul=0) const
Return true if MI is a COPY to a vector register of a specific LMul, or any kind of vector registers ...
bool canFoldIntoAddrMode(const MachineInstr &MemI, Register Reg, const MachineInstr &AddrI, ExtAddrMode &AM) const override
void insertIndirectBranch(MachineBasicBlock &MBB, MachineBasicBlock &NewDestBB, MachineBasicBlock &RestoreBB, const DebugLoc &DL, int64_t BrOffset, RegScavenger *RS) const override
bool isAsCheapAsAMove(const MachineInstr &MI) const override
bool verifyInstruction(const MachineInstr &MI, StringRef &ErrInfo) const override
bool getMemOperandWithOffsetWidth(const MachineInstr &LdSt, const MachineOperand *&BaseOp, int64_t &Offset, LocationSize &Width, const TargetRegisterInfo *TRI) const
unsigned getTailDuplicateSize(CodeGenOptLevel OptLevel) const override
void getReassociateOperandIndices(const MachineInstr &Root, unsigned Pattern, std::array< unsigned, 5 > &OperandIndices) const override
const RISCVSubtarget & STI
Register isStoreToStackSlot(const MachineInstr &MI, int &FrameIndex) const override
static bool isSafeToMove(const MachineInstr &From, const MachineInstr &To)
Return true if moving From down to To won't cause any physical register reads or writes to be clobber...
std::optional< unsigned > getInverseOpcode(unsigned Opcode) const override
MachineInstr * foldMemoryOperandImpl(MachineFunction &MF, MachineInstr &MI, ArrayRef< unsigned > Ops, MachineBasicBlock::iterator InsertPt, int FrameIndex, MachineInstr *&CopyMI, LiveIntervals *LIS=nullptr, VirtRegMap *VRM=nullptr) const override
bool simplifyInstruction(MachineInstr &MI) const override
ArrayRef< std::pair< unsigned, const char * > > getSerializableDirectMachineOperandTargetFlags() const override
outliner::InstrType getOutliningTypeImpl(const MachineModuleInfo &MMI, MachineBasicBlock::iterator &MBBI, unsigned Flags) const override
MachineTraceStrategy getMachineCombinerTraceStrategy() const override
unsigned getInstSizeInBytes(const MachineInstr &MI) const override
std::optional< RegImmPair > isAddImmediate(const MachineInstr &MI, Register Reg) const override
bool reverseBranchCondition(SmallVectorImpl< MachineOperand > &Cond) const override
ArrayRef< std::pair< MachineMemOperand::Flags, const char * > > getSerializableMachineMemOperandTargetFlags() const override
MCInst getNop() const override
bool analyzeCandidate(outliner::Candidate &C) const
bool isMBBSafeToOutlineFrom(MachineBasicBlock &MBB, unsigned &Flags) const override
bool getMemOperandsWithOffsetWidth(const MachineInstr &MI, SmallVectorImpl< const MachineOperand * > &BaseOps, int64_t &Offset, bool &OffsetIsScalable, LocationSize &Width, const TargetRegisterInfo *TRI) const override
void buildOutlinedFrame(MachineBasicBlock &MBB, MachineFunction &MF, const outliner::OutlinedFunction &OF) const override
bool requiresNTLHint(const MachineInstr &MI) const
Return true if the instruction requires an NTL hint to be emitted.
void finalizeInsInstrs(MachineInstr &Root, unsigned &Pattern, SmallVectorImpl< MachineInstr * > &InsInstrs) const override
std::pair< unsigned, unsigned > decomposeMachineOperandsTargetFlags(unsigned TF) const override
MachineInstr * commuteInstructionImpl(MachineInstr &MI, bool NewMI, unsigned OpIdx1, unsigned OpIdx2) const override
bool hasReassociableOperands(const MachineInstr &Inst, const MachineBasicBlock *MBB) const override
MachineBasicBlock * getBranchDestBlock(const MachineInstr &MI) const override
std::string createMIROperandComment(const MachineInstr &MI, const MachineOperand &Op, unsigned OpIdx, const TargetRegisterInfo *TRI) const override
bool shouldOutlineFromFunctionByDefault(MachineFunction &MF) const override
void copyPhysReg(MachineBasicBlock &MBB, MachineBasicBlock::iterator MBBI, const DebugLoc &DL, Register DstReg, Register SrcReg, bool KillSrc, bool RenamableDest=false, bool RenamableSrc=false) const override
bool findCommutedOpIndices(const MachineInstr &MI, unsigned &SrcOpIdx1, unsigned &SrcOpIdx2) const override
bool analyzeBranch(MachineBasicBlock &MBB, MachineBasicBlock *&TBB, MachineBasicBlock *&FBB, SmallVectorImpl< MachineOperand > &Cond, bool AllowModify) const override
MachineBasicBlock::iterator insertOutlinedCall(Module &M, MachineBasicBlock &MBB, MachineBasicBlock::iterator &It, MachineFunction &MF, outliner::Candidate &C) const override
bool isBranchOffsetInRange(unsigned BranchOpc, int64_t BrOffset) const override
static RISCVCC::CondCode getCondFromBranchOpc(unsigned Opc)
bool isAssociativeAndCommutative(const MachineInstr &Inst, bool Invert) const override
CombinerObjective getCombinerObjective(unsigned Pattern) const override
bool isHighLatencyDef(int Opc) const override
static bool evaluateCondBranch(RISCVCC::CondCode CC, int64_t C0, int64_t C1)
Return the result of the evaluation of C0 CC C1, where CC is a RISCVCC::CondCode.
bool getMachineCombinerPatterns(MachineInstr &Root, SmallVectorImpl< unsigned > &Patterns, bool DoRegPressureReduce) const override
bool optimizeCondBranch(MachineInstr &MI) const override
std::optional< DestSourcePair > isCopyInstrImpl(const MachineInstr &MI) const override
static bool isFromLoadImm(const MachineRegisterInfo &MRI, const MachineOperand &Op, int64_t &Imm)
Return true if the operand is a load immediate instruction and sets Imm to the immediate value.
bool shouldClusterMemOps(ArrayRef< const MachineOperand * > BaseOps1, int64_t Offset1, bool OffsetIsScalable1, ArrayRef< const MachineOperand * > BaseOps2, int64_t Offset2, bool OffsetIsScalable2, unsigned ClusterSize, unsigned NumBytes) const override
bool areMemAccessesTriviallyDisjoint(const MachineInstr &MIa, const MachineInstr &MIb) const override
RISCVMachineFunctionInfo - This class is derived from MachineFunctionInfo and contains private RISCV-...
int getBranchRelaxationScratchFrameIndex() const
const RISCVRegisterInfo * getRegisterInfo() const override
Wrapper class representing virtual and physical registers.
constexpr bool isValid() const
constexpr bool isVirtual() const
Return true if the specified register number is in the virtual register namespace.
SlotIndex - An opaque wrapper around machine indexes.
SlotIndex getRegSlot(bool EC=false) const
Returns the register use/def slot in the current instruction for a normal or early-clobber def.
A templated base class for SmallPtrSet which provides the typesafe interface that is common across al...
bool erase(PtrType Ptr)
Remove pointer from the set.
std::pair< iterator, bool > insert(PtrType Ptr)
Inserts Ptr if and only if there is no element in the container equal to Ptr.
This class consists of common code factored out of the SmallVector class to reduce code duplication b...
void push_back(const T &Elt)
This is a 'vector' (really, a variable-sized array), optimized for the case when the array is small.
MI-level stackmap operands.
uint32_t getNumPatchBytes() const
Return the number of patchable bytes the given stackmap should emit.
MI-level Statepoint operands.
uint32_t getNumPatchBytes() const
Return the number of patchable bytes the given statepoint should emit.
StringRef - Represent a constant reference to a string, i.e.
Object returned by analyzeLoopForPipelining.
TargetInstrInfo - Interface to description of machine instruction set.
virtual bool findCommutedOpIndices(const MachineInstr &MI, unsigned &SrcOpIdx1, unsigned &SrcOpIdx2) const
Returns true iff the routine could find two commutable operands in the given machine instruction.
virtual bool hasReassociableOperands(const MachineInstr &Inst, const MachineBasicBlock *MBB) const
Return true when \P Inst has reassociable operands in the same \P MBB.
virtual void genAlternativeCodeSequence(MachineInstr &Root, unsigned Pattern, SmallVectorImpl< MachineInstr * > &InsInstrs, SmallVectorImpl< MachineInstr * > &DelInstrs, DenseMap< Register, unsigned > &InstIdxForVirtReg) const
When getMachineCombinerPatterns() finds patterns, this function generates the instructions that could...
virtual bool getMachineCombinerPatterns(MachineInstr &Root, SmallVectorImpl< unsigned > &Patterns, bool DoRegPressureReduce) const
Return true when there is potentially a faster code sequence for an instruction chain ending in Root.
virtual bool isReMaterializableImpl(const MachineInstr &MI) const
For instructions with opcodes for which the M_REMATERIALIZABLE flag is set, this hook lets the target...
virtual bool isMBBSafeToOutlineFrom(MachineBasicBlock &MBB, unsigned &Flags) const
Optional target hook that returns true if MBB is safe to outline from, and returns any target-specifi...
virtual void getReassociateOperandIndices(const MachineInstr &Root, unsigned Pattern, std::array< unsigned, 5 > &OperandIndices) const
The returned array encodes the operand index for each parameter because the operands may be commuted;...
virtual CombinerObjective getCombinerObjective(unsigned Pattern) const
Return the objective of a combiner pattern.
virtual MachineInstr * commuteInstructionImpl(MachineInstr &MI, bool NewMI, unsigned OpIdx1, unsigned OpIdx2) const
This method commutes the operands of the given machine instruction MI.
virtual bool hasReassociableSibling(const MachineInstr &Inst, bool &Commuted) const
Return true when \P Inst has reassociable sibling.
virtual std::string createMIROperandComment(const MachineInstr &MI, const MachineOperand &Op, unsigned OpIdx, const TargetRegisterInfo *TRI) const
const MCAsmInfo * getMCAsmInfo() const
Return target specific asm information.
const uint8_t TSFlags
Configurable target specific flags.
TargetRegisterInfo base class - We assume that the target defines a static array of TargetRegisterDes...
TargetSubtargetInfo - Generic base class for all target subtargets.
virtual const TargetInstrInfo * getInstrInfo() const
virtual const TargetRegisterInfo * getRegisterInfo() const =0
Return the target's register information.
Target - Wrapper for Target specific information.
static constexpr TypeSize getFixed(ScalarTy ExactSize)
static constexpr TypeSize getZero()
static constexpr TypeSize getScalable(ScalarTy MinimumSize)
self_iterator getIterator()
A raw_ostream that writes to an std::string.
#define llvm_unreachable(msg)
Marks that the current location is not supposed to be reachable.
@ C
The default llvm calling convention, compatible with C.
CondCode getInverseBranchCondition(CondCode)
unsigned getInverseBranchOpcode(unsigned BCC)
unsigned getBrCond(CondCode CC, unsigned SelectOpc=0)
static bool isValidRoundingMode(unsigned Mode)
static StringRef roundingModeToString(RoundingMode RndMode)
static unsigned getVecPolicyOpNum(const MCInstrDesc &Desc)
static bool usesMaskPolicy(uint64_t TSFlags)
static bool hasRoundModeOp(uint64_t TSFlags)
static unsigned getVLOpNum(const MCInstrDesc &Desc)
static bool hasVLOp(uint64_t TSFlags)
static MCRegister getTailExpandUseRegNo(const FeatureBitset &FeatureBits)
static int getFRMOpNum(const MCInstrDesc &Desc)
static int getVXRMOpNum(const MCInstrDesc &Desc)
static bool hasVecPolicyOp(uint64_t TSFlags)
static bool usesVXRM(uint64_t TSFlags)
static bool isRVVWideningReduction(uint64_t TSFlags)
static unsigned getSEWOpNum(const MCInstrDesc &Desc)
static bool hasSEWOp(uint64_t TSFlags)
static bool isFirstDefTiedToFirstUse(const MCInstrDesc &Desc)
InstSeq generateInstSeq(int64_t Val, const MCSubtargetInfo &STI)
SmallVector< Inst, 8 > InstSeq
@ OPERAND_UIMMLOG2XLEN_NONZERO
@ OPERAND_SIMM12_LSB00000
@ OPERAND_FIRST_RISCV_IMM
@ OPERAND_UIMM10_LSB00_NONZERO
@ OPERAND_SIMM10_LSB0000_NONZERO
@ OPERAND_ATOMIC_ORDERING
static unsigned getNF(uint8_t TSFlags)
static RISCVVType::VLMUL getLMul(uint8_t TSFlags)
static bool isTailAgnostic(unsigned VType)
LLVM_ABI void printXSfmmVType(unsigned VType, raw_ostream &OS)
LLVM_ABI std::pair< unsigned, bool > decodeVLMUL(VLMUL VLMul)
static bool isValidSEW(unsigned SEW)
static bool isValidVType(unsigned VType)
LLVM_ABI void printVType(unsigned VType, raw_ostream &OS)
static bool isValidXSfmmVType(unsigned VTypeI)
static unsigned getSEW(unsigned VType)
static VLMUL getVLMUL(unsigned VType)
static bool isValidRoundingMode(unsigned Mode)
static StringRef roundingModeToString(RoundingMode RndMode)
bool hasEqualFRM(const MachineInstr &MI1, const MachineInstr &MI2)
bool isVLKnownLE(const MachineOperand &LHS, const MachineOperand &RHS)
Given two VL operands, do we know that LHS <= RHS?
unsigned getRVVMCOpcode(unsigned RVVPseudoOpcode)
unsigned getDestLog2EEW(const MCInstrDesc &Desc, unsigned Log2SEW)
std::optional< unsigned > getVectorLowDemandedScalarBits(unsigned Opcode, unsigned Log2SEW)
std::optional< std::pair< unsigned, unsigned > > isRVVSpillForZvlsseg(unsigned Opcode)
static constexpr unsigned RVVBitsPerBlock
bool isRVVSpill(const MachineInstr &MI)
static constexpr unsigned RVVBytesPerBlock
static constexpr int64_t VLMaxSentinel
bool isVectorCopy(const TargetRegisterInfo *TRI, const MachineInstr &MI)
Return true if MI is a copy that will be lowered to one or more vmvNr.vs.
ValuesClass values(OptsTy... Options)
Helper to build a ValuesClass by forwarding a variable number of arguments as an initializer list to ...
initializer< Ty > init(const Ty &Val)
InstrType
Represents how an instruction should be mapped by the outliner.
This is an optimization pass for GlobalISel generic memory operations.
auto drop_begin(T &&RangeOrContainer, size_t N=1)
Return a range covering RangeOrContainer with the first N elements excluded.
MachineTraceStrategy
Strategies for selecting traces.
@ TS_MinInstrCount
Select the trace through a block that has the fewest instructions.
@ TS_Local
Select the trace that contains only the current basic block.
bool all_of(R &&range, UnaryPredicate P)
Provide wrappers to std::all_of which take ranges instead of having to pass begin/end explicitly.
static const MachineMemOperand::Flags MONontemporalBit1
MachineInstrBuilder BuildMI(MachineFunction &MF, const MIMetadata &MIMD, const MCInstrDesc &MCID)
Builder interface. Specify how to create the initial instruction itself.
constexpr bool isInt(int64_t x)
Checks if an integer fits into the given bit width.
RegState
Flags to represent properties of register accesses.
@ Implicit
Not emitted register (e.g. carry, or temporary result).
@ Kill
The last use of a register.
@ Undef
Value of the register doesn't matter.
@ Define
Register definition.
auto enumerate(FirstRange &&First, RestRanges &&...Rest)
Given two or more input ranges, returns a new range whose values are tuples (A, B,...
bool isValidAtomicOrdering(Int I)
constexpr RegState getKillRegState(bool B)
static const MachineMemOperand::Flags MONontemporalBit0
constexpr RegState getDeadRegState(bool B)
unsigned M1(unsigned Val)
constexpr bool has_single_bit(T Value) noexcept
unsigned Log2_32(uint32_t Value)
Return the floor log base 2 of the specified value, -1 if the value is zero.
MachineInstr * getImm(const MachineOperand &MO, const MachineRegisterInfo *MRI)
constexpr RegState getRenamableRegState(bool B)
decltype(auto) get(const PointerIntPair< PointerTy, IntBits, IntType, PtrTraits, Info > &Pair)
LLVM_ABI raw_ostream & dbgs()
dbgs() - This returns a reference to a raw_ostream for debugging messages.
LLVM_ABI void report_fatal_error(Error Err, bool gen_crash_diag=true)
constexpr RegState getDefRegState(bool B)
CombinerObjective
The combiner's goal may differ based on which pattern it is attempting to optimize.
constexpr bool isUInt(uint64_t x)
Checks if an unsigned integer fits into the given bit width.
CodeGenOptLevel
Code generation optimization level.
int isShifted359(T Value, int &Shift)
bool isa(const From &Val)
isa<X> - Return true if the parameter to the template is an instance of one of the template type argu...
uint16_t MCPhysReg
An unsigned integer type large enough to represent all physical registers, but not necessarily virtua...
DWARFExpression::Operation Op
ArrayRef(const T &OneElt) -> ArrayRef< T >
constexpr bool isShiftedInt(int64_t x)
Checks if a signed integer is an N bit number shifted left by S.
void erase_if(Container &C, UnaryPredicate P)
Provide a container algorithm similar to C++ Library Fundamentals v2's erase_if which is equivalent t...
constexpr int64_t SignExtend64(uint64_t x)
Sign-extend the number in the bottom B bits of X to a 64-bit integer.
LLVM_ABI const Value * getUnderlyingObject(const Value *V, unsigned MaxLookup=MaxLookupSearchDepth)
This method strips off any GEP address adjustments, pointer casts or llvm.threadlocal....
constexpr bool isShiftedUInt(uint64_t x)
Checks if a unsigned integer is an N bit number shifted left by S.
void swap(llvm::BitVector &LHS, llvm::BitVector &RHS)
Implement std::swap in terms of BitVector swap.
This struct is a compact representation of a valid (non-zero power of two) alignment.
Used to describe addressing mode similar to ExtAddrMode in CodeGenPrepare.
This represents a simple continuous liveness interval for a value.
static LLVM_ABI MachinePointerInfo getFixedStack(MachineFunction &MF, int FI, int64_t Offset=0)
Return a MachinePointerInfo record that refers to the specified FrameIndex.
static bool isRVVRegClass(const TargetRegisterClass *RC)
Used to describe a register and immediate addition.
An individual sequence of instructions to be replaced with a call to an outlined function.
MachineFunction * getMF() const
The information necessary to create an outlined function for some class of candidate.