65#define GET_INSTRINFO_CTOR_DTOR
66#include "AArch64GenInstrInfo.inc"
70 cl::desc(
"Restrict range of CB instructions (DEBUG)"));
74 cl::desc(
"Restrict range of TB[N]Z instructions (DEBUG)"));
78 cl::desc(
"Restrict range of CB[N]Z instructions (DEBUG)"));
82 cl::desc(
"Restrict range of Bcc instructions (DEBUG)"));
86 cl::desc(
"Restrict range of B instructions (DEBUG)"));
90 cl::desc(
"Restrict range of instructions to search for the "
91 "machine-combiner gather pattern optimization"));
96 RI(STI.getTargetTriple(), STI.getHwMode()), Subtarget(STI) {}
107 auto Op =
MI.getOpcode();
108 if (
Op == AArch64::INLINEASM ||
Op == AArch64::INLINEASM_BR)
109 return getInlineAsmLength(
MI.getOperand(0).getSymbolName(), *MAI);
113 if (
MI.isMetaInstruction())
118 unsigned NumBytes = 0;
122 NumBytes =
Desc.getSize() ?
Desc.getSize() : 4;
125 if (!MFI->shouldSignReturnAddress(MF))
129 auto Method = STI.getAuthenticatedLRCheckMethod(*MF);
137 switch (
Desc.getOpcode()) {
140 return Desc.getSize();
147 case TargetOpcode::STACKMAP:
150 assert(NumBytes % 4 == 0 &&
"Invalid number of NOP bytes requested!");
152 case TargetOpcode::PATCHPOINT:
155 assert(NumBytes % 4 == 0 &&
"Invalid number of NOP bytes requested!");
157 case TargetOpcode::STATEPOINT:
159 assert(NumBytes % 4 == 0 &&
"Invalid number of NOP bytes requested!");
164 case TargetOpcode::PATCHABLE_FUNCTION_ENTER:
169 F.getFnAttributeAsParsedInteger(
"patchable-function-entry", 9) * 4;
171 case TargetOpcode::PATCHABLE_FUNCTION_EXIT:
172 case TargetOpcode::PATCHABLE_TAIL_CALL:
173 case TargetOpcode::PATCHABLE_TYPED_EVENT_CALL:
177 case TargetOpcode::PATCHABLE_EVENT_CALL:
183 NumBytes =
MI.getOperand(1).getImm();
185 case TargetOpcode::BUNDLE:
186 NumBytes = getInstBundleLength(
MI);
193unsigned AArch64InstrInfo::getInstBundleLength(
const MachineInstr &
MI)
const {
197 while (++
I != E &&
I->isInsideBundle()) {
198 assert(!
I->isBundle() &&
"No nested bundle!");
233 case AArch64::CBWPri:
234 case AArch64::CBXPri:
235 case AArch64::CBWPrr:
236 case AArch64::CBXPrr:
265 case AArch64::CBWPri:
266 case AArch64::CBXPri:
267 case AArch64::CBWPrr:
268 case AArch64::CBXPrr:
274 int64_t BrOffset)
const {
276 assert(Bits >= 3 &&
"max branch displacement must be enough to jump"
277 "over conditional branch expansion");
278 return isIntN(Bits, BrOffset / 4);
283 switch (
MI.getOpcode()) {
287 return MI.getOperand(0).getMBB();
292 return MI.getOperand(2).getMBB();
298 return MI.getOperand(1).getMBB();
299 case AArch64::CBWPri:
300 case AArch64::CBXPri:
301 case AArch64::CBWPrr:
302 case AArch64::CBXPrr:
303 return MI.getOperand(3).getMBB();
313 assert(RS &&
"RegScavenger required for long branching");
315 "new block should be inserted for expanding unconditional branch");
318 "restore block should be inserted for restoring clobbered registers");
325 "Branch offsets outside of the signed 33-bit range not supported");
339 constexpr Register Reg = AArch64::X16;
341 insertUnconditionalBranch(
MBB, &NewDestBB,
DL);
349 if (Scavenged != AArch64::NoRegister &&
351 buildIndirectBranch(Scavenged, NewDestBB);
361 "Unable to insert indirect branch inside function that has red zone");
384 bool AllowModify)
const {
391 if (
I->getOpcode() == AArch64::SpeculationBarrierISBDSBEndBB ||
392 I->getOpcode() == AArch64::SpeculationBarrierSBEndBB) {
396 if (!isUnpredicatedTerminator(*
I))
403 unsigned LastOpc = LastInst->
getOpcode();
404 if (
I ==
MBB.begin() || !isUnpredicatedTerminator(*--
I)) {
419 unsigned SecondLastOpc = SecondLastInst->
getOpcode();
426 LastInst = SecondLastInst;
428 if (
I ==
MBB.begin() || !isUnpredicatedTerminator(*--
I)) {
433 SecondLastInst = &*
I;
434 SecondLastOpc = SecondLastInst->
getOpcode();
445 LastInst = SecondLastInst;
447 if (
I ==
MBB.begin() || !isUnpredicatedTerminator(*--
I)) {
449 "unreachable unconditional branches removed above");
458 SecondLastInst = &*
I;
459 SecondLastOpc = SecondLastInst->
getOpcode();
463 if (SecondLastInst &&
I !=
MBB.begin() && isUnpredicatedTerminator(*--
I))
479 I->eraseFromParent();
488 I->eraseFromParent();
497 MachineBranchPredicate &MBP,
498 bool AllowModify)
const {
508 if (
I->getOpcode() == AArch64::SpeculationBarrierISBDSBEndBB ||
509 I->getOpcode() == AArch64::SpeculationBarrierSBEndBB) {
513 if (!isUnpredicatedTerminator(*
I))
518 unsigned LastOpc = LastInst->
getOpcode();
533 assert(MBP.TrueDest &&
"expected!");
534 MBP.FalseDest =
MBB.getNextNode();
536 MBP.ConditionDef =
nullptr;
537 MBP.SingleUseCondition =
false;
541 MBP.Predicate = (LastOpc == AArch64::CBNZX || LastOpc == AArch64::CBNZW)
542 ? MachineBranchPredicate::PRED_NE
543 : MachineBranchPredicate::PRED_EQ;
559 Cond[1].setImm(AArch64::CBNZW);
562 Cond[1].setImm(AArch64::CBZW);
565 Cond[1].setImm(AArch64::CBNZX);
568 Cond[1].setImm(AArch64::CBZX);
571 Cond[1].setImm(AArch64::TBNZW);
574 Cond[1].setImm(AArch64::TBZW);
577 Cond[1].setImm(AArch64::TBNZX);
580 Cond[1].setImm(AArch64::TBZX);
584 case AArch64::CBWPri:
585 case AArch64::CBXPri:
586 case AArch64::CBWPrr:
587 case AArch64::CBXPrr: {
600 int *BytesRemoved)
const {
610 I->eraseFromParent();
614 if (
I ==
MBB.begin()) {
627 I->eraseFromParent();
634void AArch64InstrInfo::instantiateCondBranch(
664 assert(
TBB &&
"insertBranch must not be told to insert a fallthrough");
692 if (!
DefMI->isFullCopy())
694 VReg =
DefMI->getOperand(1).getReg();
703 unsigned *NewVReg =
nullptr) {
708 bool Is64Bit = AArch64::GPR64allRegClass.hasSubClassEq(
MRI.getRegClass(VReg));
711 unsigned SrcOpNum = 0;
712 switch (
DefMI->getOpcode()) {
713 case AArch64::ADDSXri:
714 case AArch64::ADDSWri:
716 if (
DefMI->findRegisterDefOperandIdx(AArch64::NZCV,
nullptr,
721 case AArch64::ADDXri:
722 case AArch64::ADDWri:
724 if (!
DefMI->getOperand(2).isImm() ||
DefMI->getOperand(2).getImm() != 1 ||
725 DefMI->getOperand(3).getImm() != 0)
728 Opc = Is64Bit ? AArch64::CSINCXr : AArch64::CSINCWr;
731 case AArch64::ORNXrr:
732 case AArch64::ORNWrr: {
735 if (ZReg != AArch64::XZR && ZReg != AArch64::WZR)
738 Opc = Is64Bit ? AArch64::CSINVXr : AArch64::CSINVWr;
742 case AArch64::SUBSXrr:
743 case AArch64::SUBSWrr:
745 if (
DefMI->findRegisterDefOperandIdx(AArch64::NZCV,
nullptr,
750 case AArch64::SUBXrr:
751 case AArch64::SUBWrr: {
754 if (ZReg != AArch64::XZR && ZReg != AArch64::WZR)
757 Opc = Is64Bit ? AArch64::CSNEGXr : AArch64::CSNEGWr;
763 assert(
Opc && SrcOpNum &&
"Missing parameters");
766 *NewVReg =
DefMI->getOperand(SrcOpNum).getReg();
775 int &FalseCycles)
const {
779 RI.getCommonSubClass(
MRI.getRegClass(TrueReg),
MRI.getRegClass(FalseReg));
786 if (!RI.getCommonSubClass(RC,
MRI.getRegClass(DstReg)))
790 unsigned ExtraCondLat =
Cond.size() != 1;
794 if (AArch64::GPR64allRegClass.hasSubClassEq(RC) ||
795 AArch64::GPR32allRegClass.hasSubClassEq(RC)) {
797 CondCycles = 1 + ExtraCondLat;
798 TrueCycles = FalseCycles = 1;
808 if (AArch64::FPR64RegClass.hasSubClassEq(RC) ||
809 AArch64::FPR32RegClass.hasSubClassEq(RC)) {
810 CondCycles = 5 + ExtraCondLat;
811 TrueCycles = FalseCycles = 2;
828 switch (
Cond.size()) {
860 MRI.constrainRegClass(SrcReg, &AArch64::GPR64spRegClass);
866 MRI.constrainRegClass(SrcReg, &AArch64::GPR32spRegClass);
905 unsigned SUBSOpC, SUBSDestReg;
911 case AArch64::CBWPri:
912 SUBSOpC = AArch64::SUBSWri;
913 SUBSDestReg = AArch64::WZR;
916 case AArch64::CBXPri:
917 SUBSOpC = AArch64::SUBSXri;
918 SUBSDestReg = AArch64::XZR;
921 case AArch64::CBWPrr:
922 SUBSOpC = AArch64::SUBSWrr;
923 SUBSDestReg = AArch64::WZR;
926 case AArch64::CBXPrr:
927 SUBSOpC = AArch64::SUBSXrr;
928 SUBSDestReg = AArch64::XZR;
947 bool TryFold =
false;
948 if (
MRI.constrainRegClass(DstReg, &AArch64::GPR64RegClass)) {
949 RC = &AArch64::GPR64RegClass;
950 Opc = AArch64::CSELXr;
952 }
else if (
MRI.constrainRegClass(DstReg, &AArch64::GPR32RegClass)) {
953 RC = &AArch64::GPR32RegClass;
954 Opc = AArch64::CSELWr;
956 }
else if (
MRI.constrainRegClass(DstReg, &AArch64::FPR64RegClass)) {
957 RC = &AArch64::FPR64RegClass;
958 Opc = AArch64::FCSELDrrr;
959 }
else if (
MRI.constrainRegClass(DstReg, &AArch64::FPR32RegClass)) {
960 RC = &AArch64::FPR32RegClass;
961 Opc = AArch64::FCSELSrrr;
963 assert(RC &&
"Unsupported regclass");
967 unsigned NewVReg = 0;
982 MRI.clearKillFlags(NewVReg);
987 MRI.constrainRegClass(TrueReg, RC);
988 MRI.constrainRegClass(FalseReg, RC);
1003 assert(BitSize == 64 &&
"Only bit sizes of 32 or 64 allowed");
1008 return Is.
size() <= 2;
1014 if (Subtarget.hasExynosCheapAsMoveHandling()) {
1015 if (isExynosCheapAsMove(
MI))
1017 return MI.isAsCheapAsAMove();
1020 switch (
MI.getOpcode()) {
1022 return MI.isAsCheapAsAMove();
1024 case AArch64::ADDWrs:
1025 case AArch64::ADDXrs:
1026 case AArch64::SUBWrs:
1027 case AArch64::SUBXrs:
1028 return Subtarget.hasALULSLFast() &&
MI.getOperand(3).getImm() <= 4;
1033 case AArch64::MOVi32imm:
1035 case AArch64::MOVi64imm:
1040bool AArch64InstrInfo::isFalkorShiftExtFast(
const MachineInstr &
MI) {
1041 switch (
MI.getOpcode()) {
1045 case AArch64::ADDWrs:
1046 case AArch64::ADDXrs:
1047 case AArch64::ADDSWrs:
1048 case AArch64::ADDSXrs: {
1049 unsigned Imm =
MI.getOperand(3).getImm();
1056 case AArch64::ADDWrx:
1057 case AArch64::ADDXrx:
1058 case AArch64::ADDXrx64:
1059 case AArch64::ADDSWrx:
1060 case AArch64::ADDSXrx:
1061 case AArch64::ADDSXrx64: {
1062 unsigned Imm =
MI.getOperand(3).getImm();
1074 case AArch64::SUBWrs:
1075 case AArch64::SUBSWrs: {
1076 unsigned Imm =
MI.getOperand(3).getImm();
1078 return ShiftVal == 0 ||
1082 case AArch64::SUBXrs:
1083 case AArch64::SUBSXrs: {
1084 unsigned Imm =
MI.getOperand(3).getImm();
1086 return ShiftVal == 0 ||
1090 case AArch64::SUBWrx:
1091 case AArch64::SUBXrx:
1092 case AArch64::SUBXrx64:
1093 case AArch64::SUBSWrx:
1094 case AArch64::SUBSXrx:
1095 case AArch64::SUBSXrx64: {
1096 unsigned Imm =
MI.getOperand(3).getImm();
1108 case AArch64::LDRBBroW:
1109 case AArch64::LDRBBroX:
1110 case AArch64::LDRBroW:
1111 case AArch64::LDRBroX:
1112 case AArch64::LDRDroW:
1113 case AArch64::LDRDroX:
1114 case AArch64::LDRHHroW:
1115 case AArch64::LDRHHroX:
1116 case AArch64::LDRHroW:
1117 case AArch64::LDRHroX:
1118 case AArch64::LDRQroW:
1119 case AArch64::LDRQroX:
1120 case AArch64::LDRSBWroW:
1121 case AArch64::LDRSBWroX:
1122 case AArch64::LDRSBXroW:
1123 case AArch64::LDRSBXroX:
1124 case AArch64::LDRSHWroW:
1125 case AArch64::LDRSHWroX:
1126 case AArch64::LDRSHXroW:
1127 case AArch64::LDRSHXroX:
1128 case AArch64::LDRSWroW:
1129 case AArch64::LDRSWroX:
1130 case AArch64::LDRSroW:
1131 case AArch64::LDRSroX:
1132 case AArch64::LDRWroW:
1133 case AArch64::LDRWroX:
1134 case AArch64::LDRXroW:
1135 case AArch64::LDRXroX:
1136 case AArch64::PRFMroW:
1137 case AArch64::PRFMroX:
1138 case AArch64::STRBBroW:
1139 case AArch64::STRBBroX:
1140 case AArch64::STRBroW:
1141 case AArch64::STRBroX:
1142 case AArch64::STRDroW:
1143 case AArch64::STRDroX:
1144 case AArch64::STRHHroW:
1145 case AArch64::STRHHroX:
1146 case AArch64::STRHroW:
1147 case AArch64::STRHroX:
1148 case AArch64::STRQroW:
1149 case AArch64::STRQroX:
1150 case AArch64::STRSroW:
1151 case AArch64::STRSroX:
1152 case AArch64::STRWroW:
1153 case AArch64::STRWroX:
1154 case AArch64::STRXroW:
1155 case AArch64::STRXroX: {
1156 unsigned IsSigned =
MI.getOperand(3).getImm();
1163 unsigned Opc =
MI.getOpcode();
1167 case AArch64::SEH_StackAlloc:
1168 case AArch64::SEH_SaveFPLR:
1169 case AArch64::SEH_SaveFPLR_X:
1170 case AArch64::SEH_SaveReg:
1171 case AArch64::SEH_SaveReg_X:
1172 case AArch64::SEH_SaveRegP:
1173 case AArch64::SEH_SaveRegP_X:
1174 case AArch64::SEH_SaveFReg:
1175 case AArch64::SEH_SaveFReg_X:
1176 case AArch64::SEH_SaveFRegP:
1177 case AArch64::SEH_SaveFRegP_X:
1178 case AArch64::SEH_SetFP:
1179 case AArch64::SEH_AddFP:
1180 case AArch64::SEH_Nop:
1181 case AArch64::SEH_PrologEnd:
1182 case AArch64::SEH_EpilogStart:
1183 case AArch64::SEH_EpilogEnd:
1184 case AArch64::SEH_PACSignLR:
1185 case AArch64::SEH_SaveAnyRegQP:
1186 case AArch64::SEH_SaveAnyRegQPX:
1187 case AArch64::SEH_AllocZ:
1188 case AArch64::SEH_SaveZReg:
1189 case AArch64::SEH_SavePReg:
1196 unsigned &SubIdx)
const {
1197 switch (
MI.getOpcode()) {
1200 case AArch64::SBFMXri:
1201 case AArch64::UBFMXri:
1204 if (
MI.getOperand(2).getImm() != 0 ||
MI.getOperand(3).getImm() != 31)
1207 SrcReg =
MI.getOperand(1).getReg();
1208 DstReg =
MI.getOperand(0).getReg();
1209 SubIdx = AArch64::sub_32;
1218 int64_t OffsetA = 0, OffsetB = 0;
1219 TypeSize WidthA(0,
false), WidthB(0,
false);
1220 bool OffsetAIsScalable =
false, OffsetBIsScalable =
false;
1241 OffsetAIsScalable == OffsetBIsScalable) {
1242 int LowOffset = OffsetA < OffsetB ? OffsetA : OffsetB;
1243 int HighOffset = OffsetA < OffsetB ? OffsetB : OffsetA;
1244 TypeSize LowWidth = (LowOffset == OffsetA) ? WidthA : WidthB;
1245 if (LowWidth.
isScalable() == OffsetAIsScalable &&
1263 switch (
MI.getOpcode()) {
1266 if (
MI.getOperand(0).getImm() == 0x14)
1273 case AArch64::MSRpstatesvcrImm1:
1280 auto Next = std::next(
MI.getIterator());
1281 return Next !=
MBB->end() &&
Next->isCFIInstruction();
1288 Register &SrcReg2, int64_t &CmpMask,
1289 int64_t &CmpValue)
const {
1293 assert(
MI.getNumOperands() >= 2 &&
"All AArch64 cmps should have 2 operands");
1294 if (!
MI.getOperand(1).isReg() ||
MI.getOperand(1).getSubReg())
1297 switch (
MI.getOpcode()) {
1300 case AArch64::PTEST_PP:
1301 case AArch64::PTEST_PP_ANY:
1302 case AArch64::PTEST_PP_FIRST:
1303 SrcReg =
MI.getOperand(0).getReg();
1304 SrcReg2 =
MI.getOperand(1).getReg();
1305 if (
MI.getOperand(2).getSubReg())
1312 case AArch64::SUBSWrr:
1313 case AArch64::SUBSWrs:
1314 case AArch64::SUBSWrx:
1315 case AArch64::SUBSXrr:
1316 case AArch64::SUBSXrs:
1317 case AArch64::SUBSXrx:
1318 case AArch64::ADDSWrr:
1319 case AArch64::ADDSWrs:
1320 case AArch64::ADDSWrx:
1321 case AArch64::ADDSXrr:
1322 case AArch64::ADDSXrs:
1323 case AArch64::ADDSXrx:
1325 SrcReg =
MI.getOperand(1).getReg();
1326 SrcReg2 =
MI.getOperand(2).getReg();
1329 if (
MI.getOperand(2).getSubReg())
1335 case AArch64::SUBSWri:
1336 case AArch64::ADDSWri:
1337 case AArch64::SUBSXri:
1338 case AArch64::ADDSXri:
1339 SrcReg =
MI.getOperand(1).getReg();
1342 CmpValue =
MI.getOperand(2).getImm();
1344 case AArch64::ANDSWri:
1345 case AArch64::ANDSXri:
1348 SrcReg =
MI.getOperand(1).getReg();
1352 MI.getOperand(2).getImm(),
1353 MI.getOpcode() == AArch64::ANDSWri ? 32 : 64);
1362 assert(
MBB &&
"Can't get MachineBasicBlock here");
1364 assert(MF &&
"Can't get MachineFunction here");
1369 for (
unsigned OpIdx = 0, EndIdx = Instr.getNumOperands();
OpIdx < EndIdx;
1376 if (!OpRegCstraints)
1384 "Operand has register constraints without being a register!");
1387 if (
Reg.isPhysical()) {
1391 !
MRI->constrainRegClass(
Reg, OpRegCstraints))
1404 bool MIDefinesZeroReg =
false;
1405 if (
MI.definesRegister(AArch64::WZR,
nullptr) ||
1406 MI.definesRegister(AArch64::XZR,
nullptr))
1407 MIDefinesZeroReg =
true;
1409 switch (
MI.getOpcode()) {
1411 return MI.getOpcode();
1412 case AArch64::ADDSWrr:
1413 return AArch64::ADDWrr;
1414 case AArch64::ADDSWri:
1415 return MIDefinesZeroReg ? AArch64::ADDSWri : AArch64::ADDWri;
1416 case AArch64::ADDSWrs:
1417 return MIDefinesZeroReg ? AArch64::ADDSWrs : AArch64::ADDWrs;
1418 case AArch64::ADDSWrx:
1419 return AArch64::ADDWrx;
1420 case AArch64::ADDSXrr:
1421 return AArch64::ADDXrr;
1422 case AArch64::ADDSXri:
1423 return MIDefinesZeroReg ? AArch64::ADDSXri : AArch64::ADDXri;
1424 case AArch64::ADDSXrs:
1425 return MIDefinesZeroReg ? AArch64::ADDSXrs : AArch64::ADDXrs;
1426 case AArch64::ADDSXrx:
1427 return AArch64::ADDXrx;
1428 case AArch64::SUBSWrr:
1429 return AArch64::SUBWrr;
1430 case AArch64::SUBSWri:
1431 return MIDefinesZeroReg ? AArch64::SUBSWri : AArch64::SUBWri;
1432 case AArch64::SUBSWrs:
1433 return MIDefinesZeroReg ? AArch64::SUBSWrs : AArch64::SUBWrs;
1434 case AArch64::SUBSWrx:
1435 return AArch64::SUBWrx;
1436 case AArch64::SUBSXrr:
1437 return AArch64::SUBXrr;
1438 case AArch64::SUBSXri:
1439 return MIDefinesZeroReg ? AArch64::SUBSXri : AArch64::SUBXri;
1440 case AArch64::SUBSXrs:
1441 return MIDefinesZeroReg ? AArch64::SUBSXrs : AArch64::SUBXrs;
1442 case AArch64::SUBSXrx:
1443 return AArch64::SUBXrx;
1458 if (To == To->getParent()->begin())
1463 if (To->getParent() != From->getParent())
1475 Instr.modifiesRegister(AArch64::NZCV,
TRI)) ||
1476 ((AccessToCheck &
AK_Read) && Instr.readsRegister(AArch64::NZCV,
TRI)))
1482std::optional<unsigned>
1486 unsigned MaskOpcode =
Mask->getOpcode();
1487 unsigned PredOpcode = Pred->
getOpcode();
1488 bool PredIsPTestLike = isPTestLikeOpcode(PredOpcode);
1489 bool PredIsWhileLike = isWhileOpcode(PredOpcode);
1491 if (PredIsWhileLike) {
1495 if ((Mask == Pred) && PTest->
getOpcode() == AArch64::PTEST_PP_ANY)
1502 getElementSizeForOpcode(MaskOpcode) ==
1503 getElementSizeForOpcode(PredOpcode))
1509 if (PTest->
getOpcode() == AArch64::PTEST_PP_FIRST &&
1516 if (PredIsPTestLike) {
1521 if ((Mask == Pred) && PTest->
getOpcode() == AArch64::PTEST_PP_ANY)
1529 if (Mask != PTestLikeMask && PTestLikeMask->isFullCopy() &&
1530 PTestLikeMask->getOperand(1).getReg().isVirtual())
1532 MRI->getUniqueVRegDef(PTestLikeMask->getOperand(1).getReg());
1538 getElementSizeForOpcode(MaskOpcode) ==
1539 getElementSizeForOpcode(PredOpcode)) {
1540 if (Mask == PTestLikeMask || PTest->
getOpcode() == AArch64::PTEST_PP_ANY)
1566 uint64_t PredElementSize = getElementSizeForOpcode(PredOpcode);
1568 PTest->
getOpcode() == AArch64::PTEST_PP_ANY))
1576 switch (PredOpcode) {
1577 case AArch64::AND_PPzPP:
1578 case AArch64::BIC_PPzPP:
1579 case AArch64::EOR_PPzPP:
1580 case AArch64::NAND_PPzPP:
1581 case AArch64::NOR_PPzPP:
1582 case AArch64::ORN_PPzPP:
1583 case AArch64::ORR_PPzPP:
1584 case AArch64::BRKA_PPzP:
1585 case AArch64::BRKPA_PPzPP:
1586 case AArch64::BRKB_PPzP:
1587 case AArch64::BRKPB_PPzPP:
1588 case AArch64::RDFFR_PPz: {
1592 if (Mask != PredMask)
1596 case AArch64::BRKN_PPzP: {
1600 if ((MaskOpcode != AArch64::PTRUE_B) ||
1601 (
Mask->getOperand(1).getImm() != 31))
1605 case AArch64::PTRUE_B:
1618bool AArch64InstrInfo::optimizePTestInstr(
1619 MachineInstr *PTest,
unsigned MaskReg,
unsigned PredReg,
1621 auto *
Mask =
MRI->getUniqueVRegDef(MaskReg);
1622 auto *Pred =
MRI->getUniqueVRegDef(PredReg);
1624 if (Pred->
isCopy() && PTest->
getOpcode() == AArch64::PTEST_PP_FIRST) {
1628 if (
Op.isReg() &&
Op.getReg().isVirtual() &&
1629 Op.getSubReg() == AArch64::psub0)
1630 Pred =
MRI->getUniqueVRegDef(
Op.getReg());
1633 unsigned PredOpcode = Pred->
getOpcode();
1634 auto NewOp = canRemovePTestInstr(PTest, Mask, Pred,
MRI);
1650 if (*NewOp != PredOpcode) {
1661 for (; i !=
e; ++i) {
1692 if (DeadNZCVIdx != -1) {
1711 if (CmpInstr.
getOpcode() == AArch64::PTEST_PP ||
1712 CmpInstr.
getOpcode() == AArch64::PTEST_PP_ANY ||
1713 CmpInstr.
getOpcode() == AArch64::PTEST_PP_FIRST)
1714 return optimizePTestInstr(&CmpInstr, SrcReg, SrcReg2,
MRI);
1723 if (CmpValue == 0 && substituteCmpToZero(CmpInstr, SrcReg, *
MRI))
1725 return (CmpValue == 0 || CmpValue == 1) &&
1726 removeCmpToZeroOrOne(CmpInstr, SrcReg, CmpValue, *
MRI);
1734 switch (Instr.getOpcode()) {
1736 return AArch64::INSTRUCTION_LIST_END;
1738 case AArch64::ADDSWrr:
1739 case AArch64::ADDSWri:
1740 case AArch64::ADDSXrr:
1741 case AArch64::ADDSXri:
1742 case AArch64::SUBSWrr:
1743 case AArch64::SUBSWri:
1744 case AArch64::SUBSXrr:
1745 case AArch64::SUBSXri:
1746 return Instr.getOpcode();
1748 case AArch64::ADDWrr:
1749 return AArch64::ADDSWrr;
1750 case AArch64::ADDWri:
1751 return AArch64::ADDSWri;
1752 case AArch64::ADDXrr:
1753 return AArch64::ADDSXrr;
1754 case AArch64::ADDXri:
1755 return AArch64::ADDSXri;
1756 case AArch64::ADCWr:
1757 return AArch64::ADCSWr;
1758 case AArch64::ADCXr:
1759 return AArch64::ADCSXr;
1760 case AArch64::SUBWrr:
1761 return AArch64::SUBSWrr;
1762 case AArch64::SUBWri:
1763 return AArch64::SUBSWri;
1764 case AArch64::SUBXrr:
1765 return AArch64::SUBSXrr;
1766 case AArch64::SUBXri:
1767 return AArch64::SUBSXri;
1768 case AArch64::SBCWr:
1769 return AArch64::SBCSWr;
1770 case AArch64::SBCXr:
1771 return AArch64::SBCSXr;
1772 case AArch64::ANDWri:
1773 return AArch64::ANDSWri;
1774 case AArch64::ANDXri:
1775 return AArch64::ANDSXri;
1781 for (
auto *BB :
MBB->successors())
1782 if (BB->isLiveIn(AArch64::NZCV))
1791 switch (Instr.getOpcode()) {
1795 case AArch64::Bcc: {
1796 int Idx = Instr.findRegisterUseOperandIdx(AArch64::NZCV,
nullptr);
1801 case AArch64::CSINVWr:
1802 case AArch64::CSINVXr:
1803 case AArch64::CSINCWr:
1804 case AArch64::CSINCXr:
1805 case AArch64::CSELWr:
1806 case AArch64::CSELXr:
1807 case AArch64::CSNEGWr:
1808 case AArch64::CSNEGXr:
1809 case AArch64::FCSELSrrr:
1810 case AArch64::FCSELDrrr: {
1811 int Idx = Instr.findRegisterUseOperandIdx(AArch64::NZCV,
nullptr);
1824 Instr.getOperand(CCIdx).
getImm())
1877std::optional<UsedNZCV>
1882 if (
MI.getParent() != CmpParent)
1883 return std::nullopt;
1886 return std::nullopt;
1891 if (Instr.readsRegister(AArch64::NZCV, &
TRI)) {
1894 return std::nullopt;
1899 if (Instr.modifiesRegister(AArch64::NZCV, &
TRI))
1902 return NZCVUsedAfterCmp;
1906 return Opcode == AArch64::ADDSWri || Opcode == AArch64::ADDSXri;
1910 return Opcode == AArch64::SUBSWri || Opcode == AArch64::SUBSXri;
1932 const unsigned CmpOpcode = CmpInstr.
getOpcode();
1938 "Caller guarantees that CmpInstr compares with constant 0");
1941 if (!NZVCUsed || NZVCUsed->C)
1963bool AArch64InstrInfo::substituteCmpToZero(
1967 MachineInstr *
MI =
MRI.getUniqueVRegDef(SrcReg);
1974 if (NewOpc == AArch64::INSTRUCTION_LIST_END)
1981 MI->setDesc(
get(NewOpc));
1986 MI->addRegisterDefined(AArch64::NZCV, &
TRI);
1998 assert((CmpValue == 0 || CmpValue == 1) &&
1999 "Only comparisons to 0 or 1 considered for removal!");
2002 unsigned MIOpc =
MI.getOpcode();
2003 if (MIOpc == AArch64::CSINCWr) {
2004 if (
MI.getOperand(1).getReg() != AArch64::WZR ||
2005 MI.getOperand(2).getReg() != AArch64::WZR)
2007 }
else if (MIOpc == AArch64::CSINCXr) {
2008 if (
MI.getOperand(1).getReg() != AArch64::XZR ||
2009 MI.getOperand(2).getReg() != AArch64::XZR)
2019 if (
MI.findRegisterDefOperandIdx(AArch64::NZCV,
nullptr,
true) != -1)
2023 const unsigned CmpOpcode = CmpInstr.
getOpcode();
2025 if (CmpValue && !IsSubsRegImm)
2027 if (!CmpValue && !IsSubsRegImm && !
isADDSRegImm(CmpOpcode))
2032 if (MIUsedNZCV.
C || MIUsedNZCV.
V)
2035 std::optional<UsedNZCV> NZCVUsedAfterCmp =
2039 if (!NZCVUsedAfterCmp || NZCVUsedAfterCmp->C || NZCVUsedAfterCmp->V)
2042 if ((MIUsedNZCV.
Z && NZCVUsedAfterCmp->N) ||
2043 (MIUsedNZCV.
N && NZCVUsedAfterCmp->Z))
2046 if (MIUsedNZCV.
N && !CmpValue)
2088bool AArch64InstrInfo::removeCmpToZeroOrOne(
2091 MachineInstr *
MI =
MRI.getUniqueVRegDef(SrcReg);
2095 SmallVector<MachineInstr *, 4> CCUseInstrs;
2096 bool IsInvertCC =
false;
2104 for (MachineInstr *CCUseInstr : CCUseInstrs) {
2106 assert(Idx >= 0 &&
"Unexpected instruction using CC.");
2107 MachineOperand &CCOperand = CCUseInstr->getOperand(Idx);
2116bool AArch64InstrInfo::expandPostRAPseudo(
MachineInstr &
MI)
const {
2117 if (
MI.getOpcode() != TargetOpcode::LOAD_STACK_GUARD &&
2118 MI.getOpcode() != AArch64::CATCHRET)
2121 MachineBasicBlock &
MBB = *
MI.getParent();
2123 auto TRI = Subtarget.getRegisterInfo();
2126 if (
MI.getOpcode() == AArch64::CATCHRET) {
2128 const TargetInstrInfo *
TII =
2130 MachineBasicBlock *TargetMBB =
MI.getOperand(0).getMBB();
2135 FirstEpilogSEH = std::prev(FirstEpilogSEH);
2137 FirstEpilogSEH = std::next(FirstEpilogSEH);
2152 if (
M.getStackProtectorGuard() ==
"sysreg") {
2153 const AArch64SysReg::SysReg *SrcReg =
2154 AArch64SysReg::lookupSysRegByName(
M.getStackProtectorGuardReg());
2162 int Offset =
M.getStackProtectorGuardOffset();
2213 const GlobalValue *GV =
2216 unsigned OpFlags = Subtarget.ClassifyGlobalReference(GV, TM);
2222 if (Subtarget.isTargetILP32()) {
2223 unsigned Reg32 =
TRI->getSubReg(
Reg, AArch64::sub_32);
2237 assert(!Subtarget.isTargetILP32() &&
"how can large exist in ILP32?");
2264 if (Subtarget.isTargetILP32()) {
2265 unsigned Reg32 =
TRI->getSubReg(
Reg, AArch64::sub_32);
2288 switch (
MI.getOpcode()) {
2291 case AArch64::MOVZWi:
2292 case AArch64::MOVZXi:
2293 if (
MI.getOperand(1).isImm() &&
MI.getOperand(1).getImm() == 0) {
2294 assert(
MI.getDesc().getNumOperands() == 3 &&
2295 MI.getOperand(2).getImm() == 0 &&
"invalid MOVZi operands");
2299 case AArch64::ANDWri:
2300 return MI.getOperand(1).getReg() == AArch64::WZR;
2301 case AArch64::ANDXri:
2302 return MI.getOperand(1).getReg() == AArch64::XZR;
2303 case TargetOpcode::COPY:
2304 return MI.getOperand(1).getReg() == AArch64::WZR;
2312 switch (
MI.getOpcode()) {
2315 case TargetOpcode::COPY: {
2318 return (AArch64::GPR32RegClass.
contains(DstReg) ||
2319 AArch64::GPR64RegClass.
contains(DstReg));
2321 case AArch64::ORRXrs:
2322 if (
MI.getOperand(1).getReg() == AArch64::XZR) {
2323 assert(
MI.getDesc().getNumOperands() == 4 &&
2324 MI.getOperand(3).getImm() == 0 &&
"invalid ORRrs operands");
2328 case AArch64::ADDXri:
2329 if (
MI.getOperand(2).getImm() == 0) {
2330 assert(
MI.getDesc().getNumOperands() == 4 &&
2331 MI.getOperand(3).getImm() == 0 &&
"invalid ADDXri operands");
2342 switch (
MI.getOpcode()) {
2345 case TargetOpcode::COPY: {
2347 return AArch64::FPR128RegClass.contains(DstReg);
2349 case AArch64::ORRv16i8:
2350 if (
MI.getOperand(1).getReg() ==
MI.getOperand(2).getReg()) {
2351 assert(
MI.getDesc().getNumOperands() == 3 &&
MI.getOperand(0).isReg() &&
2352 "invalid ORRv16i8 operands");
2361 int &FrameIndex)
const {
2362 switch (
MI.getOpcode()) {
2365 case AArch64::LDRWui:
2366 case AArch64::LDRXui:
2367 case AArch64::LDRBui:
2368 case AArch64::LDRHui:
2369 case AArch64::LDRSui:
2370 case AArch64::LDRDui:
2371 case AArch64::LDRQui:
2372 case AArch64::LDR_PXI:
2373 if (
MI.getOperand(0).getSubReg() == 0 &&
MI.getOperand(1).isFI() &&
2374 MI.getOperand(2).isImm() &&
MI.getOperand(2).getImm() == 0) {
2375 FrameIndex =
MI.getOperand(1).getIndex();
2376 return MI.getOperand(0).getReg();
2385 int &FrameIndex)
const {
2386 switch (
MI.getOpcode()) {
2389 case AArch64::STRWui:
2390 case AArch64::STRXui:
2391 case AArch64::STRBui:
2392 case AArch64::STRHui:
2393 case AArch64::STRSui:
2394 case AArch64::STRDui:
2395 case AArch64::STRQui:
2396 case AArch64::STR_PXI:
2397 if (
MI.getOperand(0).getSubReg() == 0 &&
MI.getOperand(1).isFI() &&
2398 MI.getOperand(2).isImm() &&
MI.getOperand(2).getImm() == 0) {
2399 FrameIndex =
MI.getOperand(1).getIndex();
2400 return MI.getOperand(0).getReg();
2410 return MMO->getFlags() & MOSuppressPair;
2416 if (
MI.memoperands_empty())
2424 return MMO->getFlags() & MOStridedAccess;
2432 case AArch64::STURSi:
2433 case AArch64::STRSpre:
2434 case AArch64::STURDi:
2435 case AArch64::STRDpre:
2436 case AArch64::STURQi:
2437 case AArch64::STRQpre:
2438 case AArch64::STURBBi:
2439 case AArch64::STURHHi:
2440 case AArch64::STURWi:
2441 case AArch64::STRWpre:
2442 case AArch64::STURXi:
2443 case AArch64::STRXpre:
2444 case AArch64::LDURSi:
2445 case AArch64::LDRSpre:
2446 case AArch64::LDURDi:
2447 case AArch64::LDRDpre:
2448 case AArch64::LDURQi:
2449 case AArch64::LDRQpre:
2450 case AArch64::LDURWi:
2451 case AArch64::LDRWpre:
2452 case AArch64::LDURXi:
2453 case AArch64::LDRXpre:
2454 case AArch64::LDRSWpre:
2455 case AArch64::LDURSWi:
2456 case AArch64::LDURHHi:
2457 case AArch64::LDURBBi:
2458 case AArch64::LDURSBWi:
2459 case AArch64::LDURSHWi:
2467 case AArch64::PRFMui:
return AArch64::PRFUMi;
2468 case AArch64::LDRXui:
return AArch64::LDURXi;
2469 case AArch64::LDRWui:
return AArch64::LDURWi;
2470 case AArch64::LDRBui:
return AArch64::LDURBi;
2471 case AArch64::LDRHui:
return AArch64::LDURHi;
2472 case AArch64::LDRSui:
return AArch64::LDURSi;
2473 case AArch64::LDRDui:
return AArch64::LDURDi;
2474 case AArch64::LDRQui:
return AArch64::LDURQi;
2475 case AArch64::LDRBBui:
return AArch64::LDURBBi;
2476 case AArch64::LDRHHui:
return AArch64::LDURHHi;
2477 case AArch64::LDRSBXui:
return AArch64::LDURSBXi;
2478 case AArch64::LDRSBWui:
return AArch64::LDURSBWi;
2479 case AArch64::LDRSHXui:
return AArch64::LDURSHXi;
2480 case AArch64::LDRSHWui:
return AArch64::LDURSHWi;
2481 case AArch64::LDRSWui:
return AArch64::LDURSWi;
2482 case AArch64::STRXui:
return AArch64::STURXi;
2483 case AArch64::STRWui:
return AArch64::STURWi;
2484 case AArch64::STRBui:
return AArch64::STURBi;
2485 case AArch64::STRHui:
return AArch64::STURHi;
2486 case AArch64::STRSui:
return AArch64::STURSi;
2487 case AArch64::STRDui:
return AArch64::STURDi;
2488 case AArch64::STRQui:
return AArch64::STURQi;
2489 case AArch64::STRBBui:
return AArch64::STURBBi;
2490 case AArch64::STRHHui:
return AArch64::STURHHi;
2499 case AArch64::LDAPURBi:
2500 case AArch64::LDAPURHi:
2501 case AArch64::LDAPURi:
2502 case AArch64::LDAPURSBWi:
2503 case AArch64::LDAPURSBXi:
2504 case AArch64::LDAPURSHWi:
2505 case AArch64::LDAPURSHXi:
2506 case AArch64::LDAPURSWi:
2507 case AArch64::LDAPURXi:
2508 case AArch64::LDR_PPXI:
2509 case AArch64::LDR_PXI:
2510 case AArch64::LDR_ZXI:
2511 case AArch64::LDR_ZZXI:
2512 case AArch64::LDR_ZZXI_STRIDED_CONTIGUOUS:
2513 case AArch64::LDR_ZZZXI:
2514 case AArch64::LDR_ZZZZXI:
2515 case AArch64::LDR_ZZZZXI_STRIDED_CONTIGUOUS:
2516 case AArch64::LDRBBui:
2517 case AArch64::LDRBui:
2518 case AArch64::LDRDui:
2519 case AArch64::LDRHHui:
2520 case AArch64::LDRHui:
2521 case AArch64::LDRQui:
2522 case AArch64::LDRSBWui:
2523 case AArch64::LDRSBXui:
2524 case AArch64::LDRSHWui:
2525 case AArch64::LDRSHXui:
2526 case AArch64::LDRSui:
2527 case AArch64::LDRSWui:
2528 case AArch64::LDRWui:
2529 case AArch64::LDRXui:
2530 case AArch64::LDURBBi:
2531 case AArch64::LDURBi:
2532 case AArch64::LDURDi:
2533 case AArch64::LDURHHi:
2534 case AArch64::LDURHi:
2535 case AArch64::LDURQi:
2536 case AArch64::LDURSBWi:
2537 case AArch64::LDURSBXi:
2538 case AArch64::LDURSHWi:
2539 case AArch64::LDURSHXi:
2540 case AArch64::LDURSi:
2541 case AArch64::LDURSWi:
2542 case AArch64::LDURWi:
2543 case AArch64::LDURXi:
2544 case AArch64::PRFMui:
2545 case AArch64::PRFUMi:
2546 case AArch64::ST2Gi:
2548 case AArch64::STLURBi:
2549 case AArch64::STLURHi:
2550 case AArch64::STLURWi:
2551 case AArch64::STLURXi:
2552 case AArch64::StoreSwiftAsyncContext:
2553 case AArch64::STR_PPXI:
2554 case AArch64::STR_PXI:
2555 case AArch64::STR_ZXI:
2556 case AArch64::STR_ZZXI:
2557 case AArch64::STR_ZZXI_STRIDED_CONTIGUOUS:
2558 case AArch64::STR_ZZZXI:
2559 case AArch64::STR_ZZZZXI:
2560 case AArch64::STR_ZZZZXI_STRIDED_CONTIGUOUS:
2561 case AArch64::STRBBui:
2562 case AArch64::STRBui:
2563 case AArch64::STRDui:
2564 case AArch64::STRHHui:
2565 case AArch64::STRHui:
2566 case AArch64::STRQui:
2567 case AArch64::STRSui:
2568 case AArch64::STRWui:
2569 case AArch64::STRXui:
2570 case AArch64::STURBBi:
2571 case AArch64::STURBi:
2572 case AArch64::STURDi:
2573 case AArch64::STURHHi:
2574 case AArch64::STURHi:
2575 case AArch64::STURQi:
2576 case AArch64::STURSi:
2577 case AArch64::STURWi:
2578 case AArch64::STURXi:
2579 case AArch64::STZ2Gi:
2580 case AArch64::STZGi:
2581 case AArch64::TAGPstack:
2582 case AArch64::SPILL_PPR_TO_ZPR_SLOT_PSEUDO:
2583 case AArch64::FILL_PPR_FROM_ZPR_SLOT_PSEUDO:
2585 case AArch64::LD1B_D_IMM:
2586 case AArch64::LD1B_H_IMM:
2587 case AArch64::LD1B_IMM:
2588 case AArch64::LD1B_S_IMM:
2589 case AArch64::LD1D_IMM:
2590 case AArch64::LD1H_D_IMM:
2591 case AArch64::LD1H_IMM:
2592 case AArch64::LD1H_S_IMM:
2593 case AArch64::LD1RB_D_IMM:
2594 case AArch64::LD1RB_H_IMM:
2595 case AArch64::LD1RB_IMM:
2596 case AArch64::LD1RB_S_IMM:
2597 case AArch64::LD1RD_IMM:
2598 case AArch64::LD1RH_D_IMM:
2599 case AArch64::LD1RH_IMM:
2600 case AArch64::LD1RH_S_IMM:
2601 case AArch64::LD1RSB_D_IMM:
2602 case AArch64::LD1RSB_H_IMM:
2603 case AArch64::LD1RSB_S_IMM:
2604 case AArch64::LD1RSH_D_IMM:
2605 case AArch64::LD1RSH_S_IMM:
2606 case AArch64::LD1RSW_IMM:
2607 case AArch64::LD1RW_D_IMM:
2608 case AArch64::LD1RW_IMM:
2609 case AArch64::LD1SB_D_IMM:
2610 case AArch64::LD1SB_H_IMM:
2611 case AArch64::LD1SB_S_IMM:
2612 case AArch64::LD1SH_D_IMM:
2613 case AArch64::LD1SH_S_IMM:
2614 case AArch64::LD1SW_D_IMM:
2615 case AArch64::LD1W_D_IMM:
2616 case AArch64::LD1W_IMM:
2617 case AArch64::LD2B_IMM:
2618 case AArch64::LD2D_IMM:
2619 case AArch64::LD2H_IMM:
2620 case AArch64::LD2W_IMM:
2621 case AArch64::LD3B_IMM:
2622 case AArch64::LD3D_IMM:
2623 case AArch64::LD3H_IMM:
2624 case AArch64::LD3W_IMM:
2625 case AArch64::LD4B_IMM:
2626 case AArch64::LD4D_IMM:
2627 case AArch64::LD4H_IMM:
2628 case AArch64::LD4W_IMM:
2630 case AArch64::LDNF1B_D_IMM:
2631 case AArch64::LDNF1B_H_IMM:
2632 case AArch64::LDNF1B_IMM:
2633 case AArch64::LDNF1B_S_IMM:
2634 case AArch64::LDNF1D_IMM:
2635 case AArch64::LDNF1H_D_IMM:
2636 case AArch64::LDNF1H_IMM:
2637 case AArch64::LDNF1H_S_IMM:
2638 case AArch64::LDNF1SB_D_IMM:
2639 case AArch64::LDNF1SB_H_IMM:
2640 case AArch64::LDNF1SB_S_IMM:
2641 case AArch64::LDNF1SH_D_IMM:
2642 case AArch64::LDNF1SH_S_IMM:
2643 case AArch64::LDNF1SW_D_IMM:
2644 case AArch64::LDNF1W_D_IMM:
2645 case AArch64::LDNF1W_IMM:
2646 case AArch64::LDNPDi:
2647 case AArch64::LDNPQi:
2648 case AArch64::LDNPSi:
2649 case AArch64::LDNPWi:
2650 case AArch64::LDNPXi:
2651 case AArch64::LDNT1B_ZRI:
2652 case AArch64::LDNT1D_ZRI:
2653 case AArch64::LDNT1H_ZRI:
2654 case AArch64::LDNT1W_ZRI:
2655 case AArch64::LDPDi:
2656 case AArch64::LDPQi:
2657 case AArch64::LDPSi:
2658 case AArch64::LDPWi:
2659 case AArch64::LDPXi:
2660 case AArch64::LDRBBpost:
2661 case AArch64::LDRBBpre:
2662 case AArch64::LDRBpost:
2663 case AArch64::LDRBpre:
2664 case AArch64::LDRDpost:
2665 case AArch64::LDRDpre:
2666 case AArch64::LDRHHpost:
2667 case AArch64::LDRHHpre:
2668 case AArch64::LDRHpost:
2669 case AArch64::LDRHpre:
2670 case AArch64::LDRQpost:
2671 case AArch64::LDRQpre:
2672 case AArch64::LDRSpost:
2673 case AArch64::LDRSpre:
2674 case AArch64::LDRWpost:
2675 case AArch64::LDRWpre:
2676 case AArch64::LDRXpost:
2677 case AArch64::LDRXpre:
2678 case AArch64::ST1B_D_IMM:
2679 case AArch64::ST1B_H_IMM:
2680 case AArch64::ST1B_IMM:
2681 case AArch64::ST1B_S_IMM:
2682 case AArch64::ST1D_IMM:
2683 case AArch64::ST1H_D_IMM:
2684 case AArch64::ST1H_IMM:
2685 case AArch64::ST1H_S_IMM:
2686 case AArch64::ST1W_D_IMM:
2687 case AArch64::ST1W_IMM:
2688 case AArch64::ST2B_IMM:
2689 case AArch64::ST2D_IMM:
2690 case AArch64::ST2H_IMM:
2691 case AArch64::ST2W_IMM:
2692 case AArch64::ST3B_IMM:
2693 case AArch64::ST3D_IMM:
2694 case AArch64::ST3H_IMM:
2695 case AArch64::ST3W_IMM:
2696 case AArch64::ST4B_IMM:
2697 case AArch64::ST4D_IMM:
2698 case AArch64::ST4H_IMM:
2699 case AArch64::ST4W_IMM:
2700 case AArch64::STGPi:
2701 case AArch64::STGPreIndex:
2702 case AArch64::STZGPreIndex:
2703 case AArch64::ST2GPreIndex:
2704 case AArch64::STZ2GPreIndex:
2705 case AArch64::STGPostIndex:
2706 case AArch64::STZGPostIndex:
2707 case AArch64::ST2GPostIndex:
2708 case AArch64::STZ2GPostIndex:
2709 case AArch64::STNPDi:
2710 case AArch64::STNPQi:
2711 case AArch64::STNPSi:
2712 case AArch64::STNPWi:
2713 case AArch64::STNPXi:
2714 case AArch64::STNT1B_ZRI:
2715 case AArch64::STNT1D_ZRI:
2716 case AArch64::STNT1H_ZRI:
2717 case AArch64::STNT1W_ZRI:
2718 case AArch64::STPDi:
2719 case AArch64::STPQi:
2720 case AArch64::STPSi:
2721 case AArch64::STPWi:
2722 case AArch64::STPXi:
2723 case AArch64::STRBBpost:
2724 case AArch64::STRBBpre:
2725 case AArch64::STRBpost:
2726 case AArch64::STRBpre:
2727 case AArch64::STRDpost:
2728 case AArch64::STRDpre:
2729 case AArch64::STRHHpost:
2730 case AArch64::STRHHpre:
2731 case AArch64::STRHpost:
2732 case AArch64::STRHpre:
2733 case AArch64::STRQpost:
2734 case AArch64::STRQpre:
2735 case AArch64::STRSpost:
2736 case AArch64::STRSpre:
2737 case AArch64::STRWpost:
2738 case AArch64::STRWpre:
2739 case AArch64::STRXpost:
2740 case AArch64::STRXpre:
2742 case AArch64::LDPDpost:
2743 case AArch64::LDPDpre:
2744 case AArch64::LDPQpost:
2745 case AArch64::LDPQpre:
2746 case AArch64::LDPSpost:
2747 case AArch64::LDPSpre:
2748 case AArch64::LDPWpost:
2749 case AArch64::LDPWpre:
2750 case AArch64::LDPXpost:
2751 case AArch64::LDPXpre:
2752 case AArch64::STGPpre:
2753 case AArch64::STGPpost:
2754 case AArch64::STPDpost:
2755 case AArch64::STPDpre:
2756 case AArch64::STPQpost:
2757 case AArch64::STPQpre:
2758 case AArch64::STPSpost:
2759 case AArch64::STPSpre:
2760 case AArch64::STPWpost:
2761 case AArch64::STPWpre:
2762 case AArch64::STPXpost:
2763 case AArch64::STPXpre:
2769 switch (
MI.getOpcode()) {
2773 case AArch64::STRSui:
2774 case AArch64::STRDui:
2775 case AArch64::STRQui:
2776 case AArch64::STRXui:
2777 case AArch64::STRWui:
2778 case AArch64::LDRSui:
2779 case AArch64::LDRDui:
2780 case AArch64::LDRQui:
2781 case AArch64::LDRXui:
2782 case AArch64::LDRWui:
2783 case AArch64::LDRSWui:
2785 case AArch64::STURSi:
2786 case AArch64::STRSpre:
2787 case AArch64::STURDi:
2788 case AArch64::STRDpre:
2789 case AArch64::STURQi:
2790 case AArch64::STRQpre:
2791 case AArch64::STURWi:
2792 case AArch64::STRWpre:
2793 case AArch64::STURXi:
2794 case AArch64::STRXpre:
2795 case AArch64::LDURSi:
2796 case AArch64::LDRSpre:
2797 case AArch64::LDURDi:
2798 case AArch64::LDRDpre:
2799 case AArch64::LDURQi:
2800 case AArch64::LDRQpre:
2801 case AArch64::LDURWi:
2802 case AArch64::LDRWpre:
2803 case AArch64::LDURXi:
2804 case AArch64::LDRXpre:
2805 case AArch64::LDURSWi:
2806 case AArch64::LDRSWpre:
2808 case AArch64::LDR_ZXI:
2809 case AArch64::STR_ZXI:
2815 switch (
MI.getOpcode()) {
2818 "Unexpected instruction - was a new tail call opcode introduced?");
2820 case AArch64::TCRETURNdi:
2821 case AArch64::TCRETURNri:
2822 case AArch64::TCRETURNrix16x17:
2823 case AArch64::TCRETURNrix17:
2824 case AArch64::TCRETURNrinotx16:
2825 case AArch64::TCRETURNriALL:
2826 case AArch64::AUTH_TCRETURN:
2827 case AArch64::AUTH_TCRETURN_BTI:
2837 case AArch64::ADDWri:
2838 return AArch64::ADDSWri;
2839 case AArch64::ADDWrr:
2840 return AArch64::ADDSWrr;
2841 case AArch64::ADDWrs:
2842 return AArch64::ADDSWrs;
2843 case AArch64::ADDWrx:
2844 return AArch64::ADDSWrx;
2845 case AArch64::ANDWri:
2846 return AArch64::ANDSWri;
2847 case AArch64::ANDWrr:
2848 return AArch64::ANDSWrr;
2849 case AArch64::ANDWrs:
2850 return AArch64::ANDSWrs;
2851 case AArch64::BICWrr:
2852 return AArch64::BICSWrr;
2853 case AArch64::BICWrs:
2854 return AArch64::BICSWrs;
2855 case AArch64::SUBWri:
2856 return AArch64::SUBSWri;
2857 case AArch64::SUBWrr:
2858 return AArch64::SUBSWrr;
2859 case AArch64::SUBWrs:
2860 return AArch64::SUBSWrs;
2861 case AArch64::SUBWrx:
2862 return AArch64::SUBSWrx;
2864 case AArch64::ADDXri:
2865 return AArch64::ADDSXri;
2866 case AArch64::ADDXrr:
2867 return AArch64::ADDSXrr;
2868 case AArch64::ADDXrs:
2869 return AArch64::ADDSXrs;
2870 case AArch64::ADDXrx:
2871 return AArch64::ADDSXrx;
2872 case AArch64::ANDXri:
2873 return AArch64::ANDSXri;
2874 case AArch64::ANDXrr:
2875 return AArch64::ANDSXrr;
2876 case AArch64::ANDXrs:
2877 return AArch64::ANDSXrs;
2878 case AArch64::BICXrr:
2879 return AArch64::BICSXrr;
2880 case AArch64::BICXrs:
2881 return AArch64::BICSXrs;
2882 case AArch64::SUBXri:
2883 return AArch64::SUBSXri;
2884 case AArch64::SUBXrr:
2885 return AArch64::SUBSXrr;
2886 case AArch64::SUBXrs:
2887 return AArch64::SUBSXrs;
2888 case AArch64::SUBXrx:
2889 return AArch64::SUBSXrx;
2891 case AArch64::AND_PPzPP:
2892 return AArch64::ANDS_PPzPP;
2893 case AArch64::BIC_PPzPP:
2894 return AArch64::BICS_PPzPP;
2895 case AArch64::EOR_PPzPP:
2896 return AArch64::EORS_PPzPP;
2897 case AArch64::NAND_PPzPP:
2898 return AArch64::NANDS_PPzPP;
2899 case AArch64::NOR_PPzPP:
2900 return AArch64::NORS_PPzPP;
2901 case AArch64::ORN_PPzPP:
2902 return AArch64::ORNS_PPzPP;
2903 case AArch64::ORR_PPzPP:
2904 return AArch64::ORRS_PPzPP;
2905 case AArch64::BRKA_PPzP:
2906 return AArch64::BRKAS_PPzP;
2907 case AArch64::BRKPA_PPzPP:
2908 return AArch64::BRKPAS_PPzPP;
2909 case AArch64::BRKB_PPzP:
2910 return AArch64::BRKBS_PPzP;
2911 case AArch64::BRKPB_PPzPP:
2912 return AArch64::BRKPBS_PPzPP;
2913 case AArch64::BRKN_PPzP:
2914 return AArch64::BRKNS_PPzP;
2915 case AArch64::RDFFR_PPz:
2916 return AArch64::RDFFRS_PPz;
2917 case AArch64::PTRUE_B:
2918 return AArch64::PTRUES_B;
2929 if (
MI.hasOrderedMemoryRef())
2934 assert((
MI.getOperand(IsPreLdSt ? 2 : 1).isReg() ||
2935 MI.getOperand(IsPreLdSt ? 2 : 1).isFI()) &&
2936 "Expected a reg or frame index operand.");
2940 bool IsImmPreLdSt = IsPreLdSt &&
MI.getOperand(3).isImm();
2942 if (!
MI.getOperand(2).isImm() && !IsImmPreLdSt)
2955 if (
MI.getOperand(1).isReg() && !IsPreLdSt) {
2956 Register BaseReg =
MI.getOperand(1).getReg();
2958 if (
MI.modifiesRegister(BaseReg,
TRI))
2964 switch (
MI.getOpcode()) {
2967 case AArch64::LDR_ZXI:
2968 case AArch64::STR_ZXI:
2969 if (!Subtarget.isLittleEndian() ||
2970 Subtarget.getSVEVectorSizeInBits() != 128)
2983 const MCAsmInfo *MAI =
MI.getMF()->getTarget().getMCAsmInfo();
2985 MI.getMF()->getFunction().needsUnwindTableEntry();
2991 if (Subtarget.isPaired128Slow()) {
2992 switch (
MI.getOpcode()) {
2995 case AArch64::LDURQi:
2996 case AArch64::STURQi:
2997 case AArch64::LDRQui:
2998 case AArch64::STRQui:
3025std::optional<ExtAddrMode>
3030 bool OffsetIsScalable;
3031 if (!getMemOperandWithOffset(MemI,
Base,
Offset, OffsetIsScalable,
TRI))
3032 return std::nullopt;
3035 return std::nullopt;
3050 int64_t OffsetScale = 1;
3055 case AArch64::LDURQi:
3056 case AArch64::STURQi:
3060 case AArch64::LDURDi:
3061 case AArch64::STURDi:
3062 case AArch64::LDURXi:
3063 case AArch64::STURXi:
3067 case AArch64::LDURWi:
3068 case AArch64::LDURSWi:
3069 case AArch64::STURWi:
3073 case AArch64::LDURHi:
3074 case AArch64::STURHi:
3075 case AArch64::LDURHHi:
3076 case AArch64::STURHHi:
3077 case AArch64::LDURSHXi:
3078 case AArch64::LDURSHWi:
3082 case AArch64::LDRBroX:
3083 case AArch64::LDRBBroX:
3084 case AArch64::LDRSBXroX:
3085 case AArch64::LDRSBWroX:
3086 case AArch64::STRBroX:
3087 case AArch64::STRBBroX:
3088 case AArch64::LDURBi:
3089 case AArch64::LDURBBi:
3090 case AArch64::LDURSBXi:
3091 case AArch64::LDURSBWi:
3092 case AArch64::STURBi:
3093 case AArch64::STURBBi:
3094 case AArch64::LDRBui:
3095 case AArch64::LDRBBui:
3096 case AArch64::LDRSBXui:
3097 case AArch64::LDRSBWui:
3098 case AArch64::STRBui:
3099 case AArch64::STRBBui:
3103 case AArch64::LDRQroX:
3104 case AArch64::STRQroX:
3105 case AArch64::LDRQui:
3106 case AArch64::STRQui:
3111 case AArch64::LDRDroX:
3112 case AArch64::STRDroX:
3113 case AArch64::LDRXroX:
3114 case AArch64::STRXroX:
3115 case AArch64::LDRDui:
3116 case AArch64::STRDui:
3117 case AArch64::LDRXui:
3118 case AArch64::STRXui:
3123 case AArch64::LDRWroX:
3124 case AArch64::LDRSWroX:
3125 case AArch64::STRWroX:
3126 case AArch64::LDRWui:
3127 case AArch64::LDRSWui:
3128 case AArch64::STRWui:
3133 case AArch64::LDRHroX:
3134 case AArch64::STRHroX:
3135 case AArch64::LDRHHroX:
3136 case AArch64::STRHHroX:
3137 case AArch64::LDRSHXroX:
3138 case AArch64::LDRSHWroX:
3139 case AArch64::LDRHui:
3140 case AArch64::STRHui:
3141 case AArch64::LDRHHui:
3142 case AArch64::STRHHui:
3143 case AArch64::LDRSHXui:
3144 case AArch64::LDRSHWui:
3152 if (BaseRegOp.
isReg() && BaseRegOp.
getReg() == Reg)
3176 case AArch64::SBFMXri:
3189 AM.
Scale = OffsetScale;
3194 case TargetOpcode::SUBREG_TO_REG: {
3207 if (!OffsetReg.
isVirtual() || !
MRI.hasOneNonDBGUse(OffsetReg))
3211 if (
DefMI.getOpcode() != AArch64::ORRWrs ||
3212 DefMI.getOperand(1).getReg() != AArch64::WZR ||
3213 DefMI.getOperand(3).getImm() != 0)
3220 AM.
Scale = OffsetScale;
3231 auto validateOffsetForLDP = [](
unsigned NumBytes, int64_t OldOffset,
3232 int64_t NewOffset) ->
bool {
3233 int64_t MinOffset, MaxOffset;
3250 return OldOffset < MinOffset || OldOffset > MaxOffset ||
3251 (NewOffset >= MinOffset && NewOffset <= MaxOffset);
3253 auto canFoldAddSubImmIntoAddrMode = [&](int64_t Disp) ->
bool {
3255 int64_t NewOffset = OldOffset + Disp;
3256 if (!isLegalAddressingMode(NumBytes, NewOffset, 0))
3260 if (!validateOffsetForLDP(NumBytes, OldOffset, NewOffset))
3270 auto canFoldAddRegIntoAddrMode =
3275 if ((
unsigned)Scale != Scale)
3277 if (!isLegalAddressingMode(NumBytes, 0, Scale))
3289 return (Opcode == AArch64::STURQi || Opcode == AArch64::STRQui) &&
3290 Subtarget.isSTRQroSlow();
3299 case AArch64::ADDXri:
3305 return canFoldAddSubImmIntoAddrMode(Disp);
3307 case AArch64::SUBXri:
3313 return canFoldAddSubImmIntoAddrMode(-Disp);
3315 case AArch64::ADDXrs: {
3328 if (Shift != 2 && Shift != 3 && Subtarget.hasAddrLSLSlow14())
3330 if (avoidSlowSTRQ(MemI))
3333 return canFoldAddRegIntoAddrMode(1ULL << Shift);
3336 case AArch64::ADDXrr:
3344 if (!OptSize && avoidSlowSTRQ(MemI))
3346 return canFoldAddRegIntoAddrMode(1);
3348 case AArch64::ADDXrx:
3356 if (!OptSize && avoidSlowSTRQ(MemI))
3365 return canFoldAddRegIntoAddrMode(
3380 case AArch64::LDURQi:
3381 case AArch64::LDRQui:
3382 return AArch64::LDRQroX;
3383 case AArch64::STURQi:
3384 case AArch64::STRQui:
3385 return AArch64::STRQroX;
3386 case AArch64::LDURDi:
3387 case AArch64::LDRDui:
3388 return AArch64::LDRDroX;
3389 case AArch64::STURDi:
3390 case AArch64::STRDui:
3391 return AArch64::STRDroX;
3392 case AArch64::LDURXi:
3393 case AArch64::LDRXui:
3394 return AArch64::LDRXroX;
3395 case AArch64::STURXi:
3396 case AArch64::STRXui:
3397 return AArch64::STRXroX;
3398 case AArch64::LDURWi:
3399 case AArch64::LDRWui:
3400 return AArch64::LDRWroX;
3401 case AArch64::LDURSWi:
3402 case AArch64::LDRSWui:
3403 return AArch64::LDRSWroX;
3404 case AArch64::STURWi:
3405 case AArch64::STRWui:
3406 return AArch64::STRWroX;
3407 case AArch64::LDURHi:
3408 case AArch64::LDRHui:
3409 return AArch64::LDRHroX;
3410 case AArch64::STURHi:
3411 case AArch64::STRHui:
3412 return AArch64::STRHroX;
3413 case AArch64::LDURHHi:
3414 case AArch64::LDRHHui:
3415 return AArch64::LDRHHroX;
3416 case AArch64::STURHHi:
3417 case AArch64::STRHHui:
3418 return AArch64::STRHHroX;
3419 case AArch64::LDURSHXi:
3420 case AArch64::LDRSHXui:
3421 return AArch64::LDRSHXroX;
3422 case AArch64::LDURSHWi:
3423 case AArch64::LDRSHWui:
3424 return AArch64::LDRSHWroX;
3425 case AArch64::LDURBi:
3426 case AArch64::LDRBui:
3427 return AArch64::LDRBroX;
3428 case AArch64::LDURBBi:
3429 case AArch64::LDRBBui:
3430 return AArch64::LDRBBroX;
3431 case AArch64::LDURSBXi:
3432 case AArch64::LDRSBXui:
3433 return AArch64::LDRSBXroX;
3434 case AArch64::LDURSBWi:
3435 case AArch64::LDRSBWui:
3436 return AArch64::LDRSBWroX;
3437 case AArch64::STURBi:
3438 case AArch64::STRBui:
3439 return AArch64::STRBroX;
3440 case AArch64::STURBBi:
3441 case AArch64::STRBBui:
3442 return AArch64::STRBBroX;
3454 case AArch64::LDURQi:
3456 return AArch64::LDRQui;
3457 case AArch64::STURQi:
3459 return AArch64::STRQui;
3460 case AArch64::LDURDi:
3462 return AArch64::LDRDui;
3463 case AArch64::STURDi:
3465 return AArch64::STRDui;
3466 case AArch64::LDURXi:
3468 return AArch64::LDRXui;
3469 case AArch64::STURXi:
3471 return AArch64::STRXui;
3472 case AArch64::LDURWi:
3474 return AArch64::LDRWui;
3475 case AArch64::LDURSWi:
3477 return AArch64::LDRSWui;
3478 case AArch64::STURWi:
3480 return AArch64::STRWui;
3481 case AArch64::LDURHi:
3483 return AArch64::LDRHui;
3484 case AArch64::STURHi:
3486 return AArch64::STRHui;
3487 case AArch64::LDURHHi:
3489 return AArch64::LDRHHui;
3490 case AArch64::STURHHi:
3492 return AArch64::STRHHui;
3493 case AArch64::LDURSHXi:
3495 return AArch64::LDRSHXui;
3496 case AArch64::LDURSHWi:
3498 return AArch64::LDRSHWui;
3499 case AArch64::LDURBi:
3501 return AArch64::LDRBui;
3502 case AArch64::LDURBBi:
3504 return AArch64::LDRBBui;
3505 case AArch64::LDURSBXi:
3507 return AArch64::LDRSBXui;
3508 case AArch64::LDURSBWi:
3510 return AArch64::LDRSBWui;
3511 case AArch64::STURBi:
3513 return AArch64::STRBui;
3514 case AArch64::STURBBi:
3516 return AArch64::STRBBui;
3517 case AArch64::LDRQui:
3518 case AArch64::STRQui:
3521 case AArch64::LDRDui:
3522 case AArch64::STRDui:
3523 case AArch64::LDRXui:
3524 case AArch64::STRXui:
3527 case AArch64::LDRWui:
3528 case AArch64::LDRSWui:
3529 case AArch64::STRWui:
3532 case AArch64::LDRHui:
3533 case AArch64::STRHui:
3534 case AArch64::LDRHHui:
3535 case AArch64::STRHHui:
3536 case AArch64::LDRSHXui:
3537 case AArch64::LDRSHWui:
3540 case AArch64::LDRBui:
3541 case AArch64::LDRBBui:
3542 case AArch64::LDRSBXui:
3543 case AArch64::LDRSBWui:
3544 case AArch64::STRBui:
3545 case AArch64::STRBBui:
3559 case AArch64::LDURQi:
3560 case AArch64::STURQi:
3561 case AArch64::LDURDi:
3562 case AArch64::STURDi:
3563 case AArch64::LDURXi:
3564 case AArch64::STURXi:
3565 case AArch64::LDURWi:
3566 case AArch64::LDURSWi:
3567 case AArch64::STURWi:
3568 case AArch64::LDURHi:
3569 case AArch64::STURHi:
3570 case AArch64::LDURHHi:
3571 case AArch64::STURHHi:
3572 case AArch64::LDURSHXi:
3573 case AArch64::LDURSHWi:
3574 case AArch64::LDURBi:
3575 case AArch64::STURBi:
3576 case AArch64::LDURBBi:
3577 case AArch64::STURBBi:
3578 case AArch64::LDURSBWi:
3579 case AArch64::LDURSBXi:
3581 case AArch64::LDRQui:
3582 return AArch64::LDURQi;
3583 case AArch64::STRQui:
3584 return AArch64::STURQi;
3585 case AArch64::LDRDui:
3586 return AArch64::LDURDi;
3587 case AArch64::STRDui:
3588 return AArch64::STURDi;
3589 case AArch64::LDRXui:
3590 return AArch64::LDURXi;
3591 case AArch64::STRXui:
3592 return AArch64::STURXi;
3593 case AArch64::LDRWui:
3594 return AArch64::LDURWi;
3595 case AArch64::LDRSWui:
3596 return AArch64::LDURSWi;
3597 case AArch64::STRWui:
3598 return AArch64::STURWi;
3599 case AArch64::LDRHui:
3600 return AArch64::LDURHi;
3601 case AArch64::STRHui:
3602 return AArch64::STURHi;
3603 case AArch64::LDRHHui:
3604 return AArch64::LDURHHi;
3605 case AArch64::STRHHui:
3606 return AArch64::STURHHi;
3607 case AArch64::LDRSHXui:
3608 return AArch64::LDURSHXi;
3609 case AArch64::LDRSHWui:
3610 return AArch64::LDURSHWi;
3611 case AArch64::LDRBBui:
3612 return AArch64::LDURBBi;
3613 case AArch64::LDRBui:
3614 return AArch64::LDURBi;
3615 case AArch64::STRBBui:
3616 return AArch64::STURBBi;
3617 case AArch64::STRBui:
3618 return AArch64::STURBi;
3619 case AArch64::LDRSBWui:
3620 return AArch64::LDURSBWi;
3621 case AArch64::LDRSBXui:
3622 return AArch64::LDURSBXi;
3635 case AArch64::LDRQroX:
3636 case AArch64::LDURQi:
3637 case AArch64::LDRQui:
3638 return AArch64::LDRQroW;
3639 case AArch64::STRQroX:
3640 case AArch64::STURQi:
3641 case AArch64::STRQui:
3642 return AArch64::STRQroW;
3643 case AArch64::LDRDroX:
3644 case AArch64::LDURDi:
3645 case AArch64::LDRDui:
3646 return AArch64::LDRDroW;
3647 case AArch64::STRDroX:
3648 case AArch64::STURDi:
3649 case AArch64::STRDui:
3650 return AArch64::STRDroW;
3651 case AArch64::LDRXroX:
3652 case AArch64::LDURXi:
3653 case AArch64::LDRXui:
3654 return AArch64::LDRXroW;
3655 case AArch64::STRXroX:
3656 case AArch64::STURXi:
3657 case AArch64::STRXui:
3658 return AArch64::STRXroW;
3659 case AArch64::LDRWroX:
3660 case AArch64::LDURWi:
3661 case AArch64::LDRWui:
3662 return AArch64::LDRWroW;
3663 case AArch64::LDRSWroX:
3664 case AArch64::LDURSWi:
3665 case AArch64::LDRSWui:
3666 return AArch64::LDRSWroW;
3667 case AArch64::STRWroX:
3668 case AArch64::STURWi:
3669 case AArch64::STRWui:
3670 return AArch64::STRWroW;
3671 case AArch64::LDRHroX:
3672 case AArch64::LDURHi:
3673 case AArch64::LDRHui:
3674 return AArch64::LDRHroW;
3675 case AArch64::STRHroX:
3676 case AArch64::STURHi:
3677 case AArch64::STRHui:
3678 return AArch64::STRHroW;
3679 case AArch64::LDRHHroX:
3680 case AArch64::LDURHHi:
3681 case AArch64::LDRHHui:
3682 return AArch64::LDRHHroW;
3683 case AArch64::STRHHroX:
3684 case AArch64::STURHHi:
3685 case AArch64::STRHHui:
3686 return AArch64::STRHHroW;
3687 case AArch64::LDRSHXroX:
3688 case AArch64::LDURSHXi:
3689 case AArch64::LDRSHXui:
3690 return AArch64::LDRSHXroW;
3691 case AArch64::LDRSHWroX:
3692 case AArch64::LDURSHWi:
3693 case AArch64::LDRSHWui:
3694 return AArch64::LDRSHWroW;
3695 case AArch64::LDRBroX:
3696 case AArch64::LDURBi:
3697 case AArch64::LDRBui:
3698 return AArch64::LDRBroW;
3699 case AArch64::LDRBBroX:
3700 case AArch64::LDURBBi:
3701 case AArch64::LDRBBui:
3702 return AArch64::LDRBBroW;
3703 case AArch64::LDRSBXroX:
3704 case AArch64::LDURSBXi:
3705 case AArch64::LDRSBXui:
3706 return AArch64::LDRSBXroW;
3707 case AArch64::LDRSBWroX:
3708 case AArch64::LDURSBWi:
3709 case AArch64::LDRSBWui:
3710 return AArch64::LDRSBWroW;
3711 case AArch64::STRBroX:
3712 case AArch64::STURBi:
3713 case AArch64::STRBui:
3714 return AArch64::STRBroW;
3715 case AArch64::STRBBroX:
3716 case AArch64::STURBBi:
3717 case AArch64::STRBBui:
3718 return AArch64::STRBBroW;
3733 MRI.constrainRegClass(AM.
BaseReg, &AArch64::GPR64spRegClass);
3743 return B.getInstr();
3747 "Addressing mode not supported for folding");
3764 return B.getInstr();
3771 "Address offset can be a register or an immediate, but not both");
3773 MRI.constrainRegClass(AM.
BaseReg, &AArch64::GPR64spRegClass);
3778 OffsetReg =
MRI.createVirtualRegister(&AArch64::GPR32RegClass);
3792 return B.getInstr();
3796 "Function must not be called with an addressing mode it can't handle");
3805 case AArch64::LD1Fourv16b_POST:
3806 case AArch64::LD1Fourv1d_POST:
3807 case AArch64::LD1Fourv2d_POST:
3808 case AArch64::LD1Fourv2s_POST:
3809 case AArch64::LD1Fourv4h_POST:
3810 case AArch64::LD1Fourv4s_POST:
3811 case AArch64::LD1Fourv8b_POST:
3812 case AArch64::LD1Fourv8h_POST:
3813 case AArch64::LD1Onev16b_POST:
3814 case AArch64::LD1Onev1d_POST:
3815 case AArch64::LD1Onev2d_POST:
3816 case AArch64::LD1Onev2s_POST:
3817 case AArch64::LD1Onev4h_POST:
3818 case AArch64::LD1Onev4s_POST:
3819 case AArch64::LD1Onev8b_POST:
3820 case AArch64::LD1Onev8h_POST:
3821 case AArch64::LD1Rv16b_POST:
3822 case AArch64::LD1Rv1d_POST:
3823 case AArch64::LD1Rv2d_POST:
3824 case AArch64::LD1Rv2s_POST:
3825 case AArch64::LD1Rv4h_POST:
3826 case AArch64::LD1Rv4s_POST:
3827 case AArch64::LD1Rv8b_POST:
3828 case AArch64::LD1Rv8h_POST:
3829 case AArch64::LD1Threev16b_POST:
3830 case AArch64::LD1Threev1d_POST:
3831 case AArch64::LD1Threev2d_POST:
3832 case AArch64::LD1Threev2s_POST:
3833 case AArch64::LD1Threev4h_POST:
3834 case AArch64::LD1Threev4s_POST:
3835 case AArch64::LD1Threev8b_POST:
3836 case AArch64::LD1Threev8h_POST:
3837 case AArch64::LD1Twov16b_POST:
3838 case AArch64::LD1Twov1d_POST:
3839 case AArch64::LD1Twov2d_POST:
3840 case AArch64::LD1Twov2s_POST:
3841 case AArch64::LD1Twov4h_POST:
3842 case AArch64::LD1Twov4s_POST:
3843 case AArch64::LD1Twov8b_POST:
3844 case AArch64::LD1Twov8h_POST:
3845 case AArch64::LD1i16_POST:
3846 case AArch64::LD1i32_POST:
3847 case AArch64::LD1i64_POST:
3848 case AArch64::LD1i8_POST:
3849 case AArch64::LD2Rv16b_POST:
3850 case AArch64::LD2Rv1d_POST:
3851 case AArch64::LD2Rv2d_POST:
3852 case AArch64::LD2Rv2s_POST:
3853 case AArch64::LD2Rv4h_POST:
3854 case AArch64::LD2Rv4s_POST:
3855 case AArch64::LD2Rv8b_POST:
3856 case AArch64::LD2Rv8h_POST:
3857 case AArch64::LD2Twov16b_POST:
3858 case AArch64::LD2Twov2d_POST:
3859 case AArch64::LD2Twov2s_POST:
3860 case AArch64::LD2Twov4h_POST:
3861 case AArch64::LD2Twov4s_POST:
3862 case AArch64::LD2Twov8b_POST:
3863 case AArch64::LD2Twov8h_POST:
3864 case AArch64::LD2i16_POST:
3865 case AArch64::LD2i32_POST:
3866 case AArch64::LD2i64_POST:
3867 case AArch64::LD2i8_POST:
3868 case AArch64::LD3Rv16b_POST:
3869 case AArch64::LD3Rv1d_POST:
3870 case AArch64::LD3Rv2d_POST:
3871 case AArch64::LD3Rv2s_POST:
3872 case AArch64::LD3Rv4h_POST:
3873 case AArch64::LD3Rv4s_POST:
3874 case AArch64::LD3Rv8b_POST:
3875 case AArch64::LD3Rv8h_POST:
3876 case AArch64::LD3Threev16b_POST:
3877 case AArch64::LD3Threev2d_POST:
3878 case AArch64::LD3Threev2s_POST:
3879 case AArch64::LD3Threev4h_POST:
3880 case AArch64::LD3Threev4s_POST:
3881 case AArch64::LD3Threev8b_POST:
3882 case AArch64::LD3Threev8h_POST:
3883 case AArch64::LD3i16_POST:
3884 case AArch64::LD3i32_POST:
3885 case AArch64::LD3i64_POST:
3886 case AArch64::LD3i8_POST:
3887 case AArch64::LD4Fourv16b_POST:
3888 case AArch64::LD4Fourv2d_POST:
3889 case AArch64::LD4Fourv2s_POST:
3890 case AArch64::LD4Fourv4h_POST:
3891 case AArch64::LD4Fourv4s_POST:
3892 case AArch64::LD4Fourv8b_POST:
3893 case AArch64::LD4Fourv8h_POST:
3894 case AArch64::LD4Rv16b_POST:
3895 case AArch64::LD4Rv1d_POST:
3896 case AArch64::LD4Rv2d_POST:
3897 case AArch64::LD4Rv2s_POST:
3898 case AArch64::LD4Rv4h_POST:
3899 case AArch64::LD4Rv4s_POST:
3900 case AArch64::LD4Rv8b_POST:
3901 case AArch64::LD4Rv8h_POST:
3902 case AArch64::LD4i16_POST:
3903 case AArch64::LD4i32_POST:
3904 case AArch64::LD4i64_POST:
3905 case AArch64::LD4i8_POST:
3906 case AArch64::LDAPRWpost:
3907 case AArch64::LDAPRXpost:
3908 case AArch64::LDIAPPWpost:
3909 case AArch64::LDIAPPXpost:
3910 case AArch64::LDPDpost:
3911 case AArch64::LDPQpost:
3912 case AArch64::LDPSWpost:
3913 case AArch64::LDPSpost:
3914 case AArch64::LDPWpost:
3915 case AArch64::LDPXpost:
3916 case AArch64::LDRBBpost:
3917 case AArch64::LDRBpost:
3918 case AArch64::LDRDpost:
3919 case AArch64::LDRHHpost:
3920 case AArch64::LDRHpost:
3921 case AArch64::LDRQpost:
3922 case AArch64::LDRSBWpost:
3923 case AArch64::LDRSBXpost:
3924 case AArch64::LDRSHWpost:
3925 case AArch64::LDRSHXpost:
3926 case AArch64::LDRSWpost:
3927 case AArch64::LDRSpost:
3928 case AArch64::LDRWpost:
3929 case AArch64::LDRXpost:
3930 case AArch64::ST1Fourv16b_POST:
3931 case AArch64::ST1Fourv1d_POST:
3932 case AArch64::ST1Fourv2d_POST:
3933 case AArch64::ST1Fourv2s_POST:
3934 case AArch64::ST1Fourv4h_POST:
3935 case AArch64::ST1Fourv4s_POST:
3936 case AArch64::ST1Fourv8b_POST:
3937 case AArch64::ST1Fourv8h_POST:
3938 case AArch64::ST1Onev16b_POST:
3939 case AArch64::ST1Onev1d_POST:
3940 case AArch64::ST1Onev2d_POST:
3941 case AArch64::ST1Onev2s_POST:
3942 case AArch64::ST1Onev4h_POST:
3943 case AArch64::ST1Onev4s_POST:
3944 case AArch64::ST1Onev8b_POST:
3945 case AArch64::ST1Onev8h_POST:
3946 case AArch64::ST1Threev16b_POST:
3947 case AArch64::ST1Threev1d_POST:
3948 case AArch64::ST1Threev2d_POST:
3949 case AArch64::ST1Threev2s_POST:
3950 case AArch64::ST1Threev4h_POST:
3951 case AArch64::ST1Threev4s_POST:
3952 case AArch64::ST1Threev8b_POST:
3953 case AArch64::ST1Threev8h_POST:
3954 case AArch64::ST1Twov16b_POST:
3955 case AArch64::ST1Twov1d_POST:
3956 case AArch64::ST1Twov2d_POST:
3957 case AArch64::ST1Twov2s_POST:
3958 case AArch64::ST1Twov4h_POST:
3959 case AArch64::ST1Twov4s_POST:
3960 case AArch64::ST1Twov8b_POST:
3961 case AArch64::ST1Twov8h_POST:
3962 case AArch64::ST1i16_POST:
3963 case AArch64::ST1i32_POST:
3964 case AArch64::ST1i64_POST:
3965 case AArch64::ST1i8_POST:
3966 case AArch64::ST2GPostIndex:
3967 case AArch64::ST2Twov16b_POST:
3968 case AArch64::ST2Twov2d_POST:
3969 case AArch64::ST2Twov2s_POST:
3970 case AArch64::ST2Twov4h_POST:
3971 case AArch64::ST2Twov4s_POST:
3972 case AArch64::ST2Twov8b_POST:
3973 case AArch64::ST2Twov8h_POST:
3974 case AArch64::ST2i16_POST:
3975 case AArch64::ST2i32_POST:
3976 case AArch64::ST2i64_POST:
3977 case AArch64::ST2i8_POST:
3978 case AArch64::ST3Threev16b_POST:
3979 case AArch64::ST3Threev2d_POST:
3980 case AArch64::ST3Threev2s_POST:
3981 case AArch64::ST3Threev4h_POST:
3982 case AArch64::ST3Threev4s_POST:
3983 case AArch64::ST3Threev8b_POST:
3984 case AArch64::ST3Threev8h_POST:
3985 case AArch64::ST3i16_POST:
3986 case AArch64::ST3i32_POST:
3987 case AArch64::ST3i64_POST:
3988 case AArch64::ST3i8_POST:
3989 case AArch64::ST4Fourv16b_POST:
3990 case AArch64::ST4Fourv2d_POST:
3991 case AArch64::ST4Fourv2s_POST:
3992 case AArch64::ST4Fourv4h_POST:
3993 case AArch64::ST4Fourv4s_POST:
3994 case AArch64::ST4Fourv8b_POST:
3995 case AArch64::ST4Fourv8h_POST:
3996 case AArch64::ST4i16_POST:
3997 case AArch64::ST4i32_POST:
3998 case AArch64::ST4i64_POST:
3999 case AArch64::ST4i8_POST:
4000 case AArch64::STGPostIndex:
4001 case AArch64::STGPpost:
4002 case AArch64::STPDpost:
4003 case AArch64::STPQpost:
4004 case AArch64::STPSpost:
4005 case AArch64::STPWpost:
4006 case AArch64::STPXpost:
4007 case AArch64::STRBBpost:
4008 case AArch64::STRBpost:
4009 case AArch64::STRDpost:
4010 case AArch64::STRHHpost:
4011 case AArch64::STRHpost:
4012 case AArch64::STRQpost:
4013 case AArch64::STRSpost:
4014 case AArch64::STRWpost:
4015 case AArch64::STRXpost:
4016 case AArch64::STZ2GPostIndex:
4017 case AArch64::STZGPostIndex:
4024 bool &OffsetIsScalable,
TypeSize &Width,
4045 int64_t Dummy1, Dummy2;
4067 return BaseOp->
isReg() || BaseOp->
isFI();
4074 assert(OfsOp.
isImm() &&
"Offset operand wasn't immediate.");
4079 TypeSize &Width, int64_t &MinOffset,
4080 int64_t &MaxOffset) {
4086 MinOffset = MaxOffset = 0;
4089 case AArch64::LDRQui:
4090 case AArch64::STRQui:
4096 case AArch64::LDRXui:
4097 case AArch64::LDRDui:
4098 case AArch64::STRXui:
4099 case AArch64::STRDui:
4100 case AArch64::PRFMui:
4106 case AArch64::LDRWui:
4107 case AArch64::LDRSui:
4108 case AArch64::LDRSWui:
4109 case AArch64::STRWui:
4110 case AArch64::STRSui:
4116 case AArch64::LDRHui:
4117 case AArch64::LDRHHui:
4118 case AArch64::LDRSHWui:
4119 case AArch64::LDRSHXui:
4120 case AArch64::STRHui:
4121 case AArch64::STRHHui:
4127 case AArch64::LDRBui:
4128 case AArch64::LDRBBui:
4129 case AArch64::LDRSBWui:
4130 case AArch64::LDRSBXui:
4131 case AArch64::STRBui:
4132 case AArch64::STRBBui:
4139 case AArch64::STRQpre:
4140 case AArch64::LDRQpost:
4146 case AArch64::LDRDpost:
4147 case AArch64::LDRDpre:
4148 case AArch64::LDRXpost:
4149 case AArch64::LDRXpre:
4150 case AArch64::STRDpost:
4151 case AArch64::STRDpre:
4152 case AArch64::STRXpost:
4153 case AArch64::STRXpre:
4159 case AArch64::STRWpost:
4160 case AArch64::STRWpre:
4161 case AArch64::LDRWpost:
4162 case AArch64::LDRWpre:
4163 case AArch64::STRSpost:
4164 case AArch64::STRSpre:
4165 case AArch64::LDRSpost:
4166 case AArch64::LDRSpre:
4172 case AArch64::LDRHpost:
4173 case AArch64::LDRHpre:
4174 case AArch64::STRHpost:
4175 case AArch64::STRHpre:
4176 case AArch64::LDRHHpost:
4177 case AArch64::LDRHHpre:
4178 case AArch64::STRHHpost:
4179 case AArch64::STRHHpre:
4185 case AArch64::LDRBpost:
4186 case AArch64::LDRBpre:
4187 case AArch64::STRBpost:
4188 case AArch64::STRBpre:
4189 case AArch64::LDRBBpost:
4190 case AArch64::LDRBBpre:
4191 case AArch64::STRBBpost:
4192 case AArch64::STRBBpre:
4199 case AArch64::LDURQi:
4200 case AArch64::STURQi:
4206 case AArch64::LDURXi:
4207 case AArch64::LDURDi:
4208 case AArch64::LDAPURXi:
4209 case AArch64::STURXi:
4210 case AArch64::STURDi:
4211 case AArch64::STLURXi:
4212 case AArch64::PRFUMi:
4218 case AArch64::LDURWi:
4219 case AArch64::LDURSi:
4220 case AArch64::LDURSWi:
4221 case AArch64::LDAPURi:
4222 case AArch64::LDAPURSWi:
4223 case AArch64::STURWi:
4224 case AArch64::STURSi:
4225 case AArch64::STLURWi:
4231 case AArch64::LDURHi:
4232 case AArch64::LDURHHi:
4233 case AArch64::LDURSHXi:
4234 case AArch64::LDURSHWi:
4235 case AArch64::LDAPURHi:
4236 case AArch64::LDAPURSHWi:
4237 case AArch64::LDAPURSHXi:
4238 case AArch64::STURHi:
4239 case AArch64::STURHHi:
4240 case AArch64::STLURHi:
4246 case AArch64::LDURBi:
4247 case AArch64::LDURBBi:
4248 case AArch64::LDURSBXi:
4249 case AArch64::LDURSBWi:
4250 case AArch64::LDAPURBi:
4251 case AArch64::LDAPURSBWi:
4252 case AArch64::LDAPURSBXi:
4253 case AArch64::STURBi:
4254 case AArch64::STURBBi:
4255 case AArch64::STLURBi:
4262 case AArch64::LDPQi:
4263 case AArch64::LDNPQi:
4264 case AArch64::STPQi:
4265 case AArch64::STNPQi:
4266 case AArch64::LDPQpost:
4267 case AArch64::LDPQpre:
4268 case AArch64::STPQpost:
4269 case AArch64::STPQpre:
4275 case AArch64::LDPXi:
4276 case AArch64::LDPDi:
4277 case AArch64::LDNPXi:
4278 case AArch64::LDNPDi:
4279 case AArch64::STPXi:
4280 case AArch64::STPDi:
4281 case AArch64::STNPXi:
4282 case AArch64::STNPDi:
4283 case AArch64::LDPDpost:
4284 case AArch64::LDPDpre:
4285 case AArch64::LDPXpost:
4286 case AArch64::LDPXpre:
4287 case AArch64::STPDpost:
4288 case AArch64::STPDpre:
4289 case AArch64::STPXpost:
4290 case AArch64::STPXpre:
4296 case AArch64::LDPWi:
4297 case AArch64::LDPSi:
4298 case AArch64::LDNPWi:
4299 case AArch64::LDNPSi:
4300 case AArch64::STPWi:
4301 case AArch64::STPSi:
4302 case AArch64::STNPWi:
4303 case AArch64::STNPSi:
4304 case AArch64::LDPSpost:
4305 case AArch64::LDPSpre:
4306 case AArch64::LDPWpost:
4307 case AArch64::LDPWpre:
4308 case AArch64::STPSpost:
4309 case AArch64::STPSpre:
4310 case AArch64::STPWpost:
4311 case AArch64::STPWpre:
4317 case AArch64::StoreSwiftAsyncContext:
4330 case AArch64::TAGPstack:
4340 case AArch64::STGPreIndex:
4341 case AArch64::STGPostIndex:
4342 case AArch64::STZGi:
4343 case AArch64::STZGPreIndex:
4344 case AArch64::STZGPostIndex:
4351 case AArch64::STR_ZZZZXI:
4352 case AArch64::STR_ZZZZXI_STRIDED_CONTIGUOUS:
4353 case AArch64::LDR_ZZZZXI:
4354 case AArch64::LDR_ZZZZXI_STRIDED_CONTIGUOUS:
4360 case AArch64::STR_ZZZXI:
4361 case AArch64::LDR_ZZZXI:
4367 case AArch64::STR_ZZXI:
4368 case AArch64::STR_ZZXI_STRIDED_CONTIGUOUS:
4369 case AArch64::LDR_ZZXI:
4370 case AArch64::LDR_ZZXI_STRIDED_CONTIGUOUS:
4376 case AArch64::LDR_PXI:
4377 case AArch64::STR_PXI:
4383 case AArch64::LDR_PPXI:
4384 case AArch64::STR_PPXI:
4390 case AArch64::SPILL_PPR_TO_ZPR_SLOT_PSEUDO:
4391 case AArch64::FILL_PPR_FROM_ZPR_SLOT_PSEUDO:
4392 case AArch64::LDR_ZXI:
4393 case AArch64::STR_ZXI:
4399 case AArch64::LD1B_IMM:
4400 case AArch64::LD1H_IMM:
4401 case AArch64::LD1W_IMM:
4402 case AArch64::LD1D_IMM:
4403 case AArch64::LDNT1B_ZRI:
4404 case AArch64::LDNT1H_ZRI:
4405 case AArch64::LDNT1W_ZRI:
4406 case AArch64::LDNT1D_ZRI:
4407 case AArch64::ST1B_IMM:
4408 case AArch64::ST1H_IMM:
4409 case AArch64::ST1W_IMM:
4410 case AArch64::ST1D_IMM:
4411 case AArch64::STNT1B_ZRI:
4412 case AArch64::STNT1H_ZRI:
4413 case AArch64::STNT1W_ZRI:
4414 case AArch64::STNT1D_ZRI:
4415 case AArch64::LDNF1B_IMM:
4416 case AArch64::LDNF1H_IMM:
4417 case AArch64::LDNF1W_IMM:
4418 case AArch64::LDNF1D_IMM:
4426 case AArch64::LD2B_IMM:
4427 case AArch64::LD2H_IMM:
4428 case AArch64::LD2W_IMM:
4429 case AArch64::LD2D_IMM:
4430 case AArch64::ST2B_IMM:
4431 case AArch64::ST2H_IMM:
4432 case AArch64::ST2W_IMM:
4433 case AArch64::ST2D_IMM:
4439 case AArch64::LD3B_IMM:
4440 case AArch64::LD3H_IMM:
4441 case AArch64::LD3W_IMM:
4442 case AArch64::LD3D_IMM:
4443 case AArch64::ST3B_IMM:
4444 case AArch64::ST3H_IMM:
4445 case AArch64::ST3W_IMM:
4446 case AArch64::ST3D_IMM:
4452 case AArch64::LD4B_IMM:
4453 case AArch64::LD4H_IMM:
4454 case AArch64::LD4W_IMM:
4455 case AArch64::LD4D_IMM:
4456 case AArch64::ST4B_IMM:
4457 case AArch64::ST4H_IMM:
4458 case AArch64::ST4W_IMM:
4459 case AArch64::ST4D_IMM:
4465 case AArch64::LD1B_H_IMM:
4466 case AArch64::LD1SB_H_IMM:
4467 case AArch64::LD1H_S_IMM:
4468 case AArch64::LD1SH_S_IMM:
4469 case AArch64::LD1W_D_IMM:
4470 case AArch64::LD1SW_D_IMM:
4471 case AArch64::ST1B_H_IMM:
4472 case AArch64::ST1H_S_IMM:
4473 case AArch64::ST1W_D_IMM:
4474 case AArch64::LDNF1B_H_IMM:
4475 case AArch64::LDNF1SB_H_IMM:
4476 case AArch64::LDNF1H_S_IMM:
4477 case AArch64::LDNF1SH_S_IMM:
4478 case AArch64::LDNF1W_D_IMM:
4479 case AArch64::LDNF1SW_D_IMM:
4487 case AArch64::LD1B_S_IMM:
4488 case AArch64::LD1SB_S_IMM:
4489 case AArch64::LD1H_D_IMM:
4490 case AArch64::LD1SH_D_IMM:
4491 case AArch64::ST1B_S_IMM:
4492 case AArch64::ST1H_D_IMM:
4493 case AArch64::LDNF1B_S_IMM:
4494 case AArch64::LDNF1SB_S_IMM:
4495 case AArch64::LDNF1H_D_IMM:
4496 case AArch64::LDNF1SH_D_IMM:
4504 case AArch64::LD1B_D_IMM:
4505 case AArch64::LD1SB_D_IMM:
4506 case AArch64::ST1B_D_IMM:
4507 case AArch64::LDNF1B_D_IMM:
4508 case AArch64::LDNF1SB_D_IMM:
4516 case AArch64::ST2Gi:
4517 case AArch64::ST2GPreIndex:
4518 case AArch64::ST2GPostIndex:
4519 case AArch64::STZ2Gi:
4520 case AArch64::STZ2GPreIndex:
4521 case AArch64::STZ2GPostIndex:
4527 case AArch64::STGPi:
4528 case AArch64::STGPpost:
4529 case AArch64::STGPpre:
4535 case AArch64::LD1RB_IMM:
4536 case AArch64::LD1RB_H_IMM:
4537 case AArch64::LD1RB_S_IMM:
4538 case AArch64::LD1RB_D_IMM:
4539 case AArch64::LD1RSB_H_IMM:
4540 case AArch64::LD1RSB_S_IMM:
4541 case AArch64::LD1RSB_D_IMM:
4547 case AArch64::LD1RH_IMM:
4548 case AArch64::LD1RH_S_IMM:
4549 case AArch64::LD1RH_D_IMM:
4550 case AArch64::LD1RSH_S_IMM:
4551 case AArch64::LD1RSH_D_IMM:
4557 case AArch64::LD1RW_IMM:
4558 case AArch64::LD1RW_D_IMM:
4559 case AArch64::LD1RSW_IMM:
4565 case AArch64::LD1RD_IMM:
4581 case AArch64::LDRBBui:
4582 case AArch64::LDURBBi:
4583 case AArch64::LDRSBWui:
4584 case AArch64::LDURSBWi:
4585 case AArch64::STRBBui:
4586 case AArch64::STURBBi:
4588 case AArch64::LDRHHui:
4589 case AArch64::LDURHHi:
4590 case AArch64::LDRSHWui:
4591 case AArch64::LDURSHWi:
4592 case AArch64::STRHHui:
4593 case AArch64::STURHHi:
4595 case AArch64::LDRSui:
4596 case AArch64::LDURSi:
4597 case AArch64::LDRSpre:
4598 case AArch64::LDRSWui:
4599 case AArch64::LDURSWi:
4600 case AArch64::LDRSWpre:
4601 case AArch64::LDRWpre:
4602 case AArch64::LDRWui:
4603 case AArch64::LDURWi:
4604 case AArch64::STRSui:
4605 case AArch64::STURSi:
4606 case AArch64::STRSpre:
4607 case AArch64::STRWui:
4608 case AArch64::STURWi:
4609 case AArch64::STRWpre:
4610 case AArch64::LDPSi:
4611 case AArch64::LDPSWi:
4612 case AArch64::LDPWi:
4613 case AArch64::STPSi:
4614 case AArch64::STPWi:
4616 case AArch64::LDRDui:
4617 case AArch64::LDURDi:
4618 case AArch64::LDRDpre:
4619 case AArch64::LDRXui:
4620 case AArch64::LDURXi:
4621 case AArch64::LDRXpre:
4622 case AArch64::STRDui:
4623 case AArch64::STURDi:
4624 case AArch64::STRDpre:
4625 case AArch64::STRXui:
4626 case AArch64::STURXi:
4627 case AArch64::STRXpre:
4628 case AArch64::LDPDi:
4629 case AArch64::LDPXi:
4630 case AArch64::STPDi:
4631 case AArch64::STPXi:
4633 case AArch64::LDRQui:
4634 case AArch64::LDURQi:
4635 case AArch64::STRQui:
4636 case AArch64::STURQi:
4637 case AArch64::STRQpre:
4638 case AArch64::LDPQi:
4639 case AArch64::LDRQpre:
4640 case AArch64::STPQi:
4642 case AArch64::STZGi:
4643 case AArch64::ST2Gi:
4644 case AArch64::STZ2Gi:
4645 case AArch64::STGPi:
4651 switch (
MI.getOpcode()) {
4654 case AArch64::LDRWpre:
4655 case AArch64::LDRXpre:
4656 case AArch64::LDRSWpre:
4657 case AArch64::LDRSpre:
4658 case AArch64::LDRDpre:
4659 case AArch64::LDRQpre:
4665 switch (
MI.getOpcode()) {
4668 case AArch64::STRWpre:
4669 case AArch64::STRXpre:
4670 case AArch64::STRSpre:
4671 case AArch64::STRDpre:
4672 case AArch64::STRQpre:
4682 switch (
MI.getOpcode()) {
4685 case AArch64::LDPSi:
4686 case AArch64::LDPSWi:
4687 case AArch64::LDPDi:
4688 case AArch64::LDPQi:
4689 case AArch64::LDPWi:
4690 case AArch64::LDPXi:
4691 case AArch64::STPSi:
4692 case AArch64::STPDi:
4693 case AArch64::STPQi:
4694 case AArch64::STPWi:
4695 case AArch64::STPXi:
4696 case AArch64::STGPi:
4702 assert(
MI.mayLoadOrStore() &&
"Load or store instruction expected");
4706 return MI.getOperand(Idx);
4711 assert(
MI.mayLoadOrStore() &&
"Load or store instruction expected");
4715 return MI.getOperand(Idx);
4720 switch (
MI.getOpcode()) {
4723 case AArch64::LDRBroX:
4724 case AArch64::LDRBBroX:
4725 case AArch64::LDRSBXroX:
4726 case AArch64::LDRSBWroX:
4727 case AArch64::LDRHroX:
4728 case AArch64::LDRHHroX:
4729 case AArch64::LDRSHXroX:
4730 case AArch64::LDRSHWroX:
4731 case AArch64::LDRWroX:
4732 case AArch64::LDRSroX:
4733 case AArch64::LDRSWroX:
4734 case AArch64::LDRDroX:
4735 case AArch64::LDRXroX:
4736 case AArch64::LDRQroX:
4737 return MI.getOperand(4);
4743 if (
MI.getParent() ==
nullptr)
4753 auto Reg =
Op.getReg();
4754 if (Reg.isPhysical())
4755 return AArch64::FPR16RegClass.contains(Reg);
4757 return TRC == &AArch64::FPR16RegClass ||
4758 TRC == &AArch64::FPR16_loRegClass;
4767 auto Reg =
Op.getReg();
4768 if (Reg.isPhysical())
4769 return AArch64::FPR128RegClass.contains(Reg);
4771 return TRC == &AArch64::FPR128RegClass ||
4772 TRC == &AArch64::FPR128_loRegClass;
4778 switch (
MI.getOpcode()) {
4781 case AArch64::PACIASP:
4782 case AArch64::PACIBSP:
4785 case AArch64::PAUTH_PROLOGUE:
4788 case AArch64::HINT: {
4789 unsigned Imm =
MI.getOperand(0).getImm();
4791 if (Imm == 32 || Imm == 34 || Imm == 36 || Imm == 38)
4794 if (Imm == 25 || Imm == 27)
4806 assert(Reg.isPhysical() &&
"Expected physical register in isFpOrNEON");
4807 return AArch64::FPR128RegClass.contains(Reg) ||
4808 AArch64::FPR64RegClass.contains(Reg) ||
4809 AArch64::FPR32RegClass.contains(Reg) ||
4810 AArch64::FPR16RegClass.contains(Reg) ||
4811 AArch64::FPR8RegClass.contains(Reg);
4818 auto Reg =
Op.getReg();
4819 if (Reg.isPhysical())
4823 return TRC == &AArch64::FPR128RegClass ||
4824 TRC == &AArch64::FPR128_loRegClass ||
4825 TRC == &AArch64::FPR64RegClass ||
4826 TRC == &AArch64::FPR64_loRegClass ||
4827 TRC == &AArch64::FPR32RegClass || TRC == &AArch64::FPR16RegClass ||
4828 TRC == &AArch64::FPR8RegClass;
4850 if (FirstOpc == SecondOpc)
4856 case AArch64::STRSui:
4857 case AArch64::STURSi:
4858 return SecondOpc == AArch64::STRSui || SecondOpc == AArch64::STURSi;
4859 case AArch64::STRDui:
4860 case AArch64::STURDi:
4861 return SecondOpc == AArch64::STRDui || SecondOpc == AArch64::STURDi;
4862 case AArch64::STRQui:
4863 case AArch64::STURQi:
4864 return SecondOpc == AArch64::STRQui || SecondOpc == AArch64::STURQi;
4865 case AArch64::STRWui:
4866 case AArch64::STURWi:
4867 return SecondOpc == AArch64::STRWui || SecondOpc == AArch64::STURWi;
4868 case AArch64::STRXui:
4869 case AArch64::STURXi:
4870 return SecondOpc == AArch64::STRXui || SecondOpc == AArch64::STURXi;
4871 case AArch64::LDRSui:
4872 case AArch64::LDURSi:
4873 return SecondOpc == AArch64::LDRSui || SecondOpc == AArch64::LDURSi;
4874 case AArch64::LDRDui:
4875 case AArch64::LDURDi:
4876 return SecondOpc == AArch64::LDRDui || SecondOpc == AArch64::LDURDi;
4877 case AArch64::LDRQui:
4878 case AArch64::LDURQi:
4879 return SecondOpc == AArch64::LDRQui || SecondOpc == AArch64::LDURQi;
4880 case AArch64::LDRWui:
4881 case AArch64::LDURWi:
4882 return SecondOpc == AArch64::LDRSWui || SecondOpc == AArch64::LDURSWi;
4883 case AArch64::LDRSWui:
4884 case AArch64::LDURSWi:
4885 return SecondOpc == AArch64::LDRWui || SecondOpc == AArch64::LDURWi;
4886 case AArch64::LDRXui:
4887 case AArch64::LDURXi:
4888 return SecondOpc == AArch64::LDRXui || SecondOpc == AArch64::LDURXi;
4895 int64_t Offset1,
unsigned Opcode1,
int FI2,
4896 int64_t Offset2,
unsigned Opcode2) {
4902 assert(ObjectOffset1 <= ObjectOffset2 &&
"Object offsets are not ordered.");
4905 if (ObjectOffset1 % Scale1 != 0)
4907 ObjectOffset1 /= Scale1;
4909 if (ObjectOffset2 % Scale2 != 0)
4911 ObjectOffset2 /= Scale2;
4912 ObjectOffset1 += Offset1;
4913 ObjectOffset2 += Offset2;
4914 return ObjectOffset1 + 1 == ObjectOffset2;
4926 int64_t OpOffset2,
bool OffsetIsScalable2,
unsigned ClusterSize,
4927 unsigned NumBytes)
const {
4937 "Only base registers and frame indices are supported.");
4944 if (ClusterSize > 2)
4951 unsigned FirstOpc = FirstLdSt.
getOpcode();
4952 unsigned SecondOpc = SecondLdSt.
getOpcode();
4972 if (Offset1 > 63 || Offset1 < -64)
4977 if (BaseOp1.
isFI()) {
4979 "Caller should have ordered offsets.");
4984 BaseOp2.
getIndex(), Offset2, SecondOpc);
4987 assert(Offset1 <= Offset2 &&
"Caller should have ordered offsets.");
4989 return Offset1 + 1 == Offset2;
4999 if (
Reg.isPhysical())
5008 return ((DestReg - SrcReg) & 0x1f) < NumRegs;
5017 assert(Subtarget.hasNEON() &&
"Unexpected register copy without NEON");
5019 uint16_t DestEncoding =
TRI->getEncodingValue(DestReg);
5020 uint16_t SrcEncoding =
TRI->getEncodingValue(SrcReg);
5021 unsigned NumRegs = Indices.
size();
5023 int SubReg = 0, End = NumRegs, Incr = 1;
5042 unsigned Opcode,
unsigned ZeroReg,
5045 unsigned NumRegs = Indices.
size();
5048 uint16_t DestEncoding =
TRI->getEncodingValue(DestReg);
5049 uint16_t SrcEncoding =
TRI->getEncodingValue(SrcReg);
5050 assert(DestEncoding % NumRegs == 0 && SrcEncoding % NumRegs == 0 &&
5051 "GPR reg sequences should not be able to overlap");
5068 bool RenamableSrc)
const {
5069 if (AArch64::GPR32spRegClass.
contains(DestReg) &&
5070 (AArch64::GPR32spRegClass.
contains(SrcReg) || SrcReg == AArch64::WZR)) {
5073 if (DestReg == AArch64::WSP || SrcReg == AArch64::WSP) {
5075 if (Subtarget.hasZeroCycleRegMoveGPR64() &&
5076 !Subtarget.hasZeroCycleRegMoveGPR32()) {
5079 DestReg, AArch64::sub_32, &AArch64::GPR64spRegClass);
5081 SrcReg, AArch64::sub_32, &AArch64::GPR64spRegClass);
5097 }
else if (SrcReg == AArch64::WZR && Subtarget.hasZeroCycleZeroingGPR32()) {
5102 if (Subtarget.hasZeroCycleRegMoveGPR64() &&
5103 !Subtarget.hasZeroCycleRegMoveGPR32()) {
5106 DestReg, AArch64::sub_32, &AArch64::GPR64spRegClass);
5107 assert(DestRegX.
isValid() &&
"Destination super-reg not valid");
5109 SrcReg == AArch64::WZR
5111 :
TRI->getMatchingSuperReg(SrcReg, AArch64::sub_32,
5112 &AArch64::GPR64spRegClass);
5133 if (AArch64::PPRRegClass.
contains(DestReg) &&
5134 AArch64::PPRRegClass.
contains(SrcReg)) {
5135 assert(Subtarget.isSVEorStreamingSVEAvailable() &&
5136 "Unexpected SVE register.");
5146 bool DestIsPNR = AArch64::PNRRegClass.contains(DestReg);
5147 bool SrcIsPNR = AArch64::PNRRegClass.contains(SrcReg);
5148 if (DestIsPNR || SrcIsPNR) {
5150 return (R - AArch64::PN0) + AArch64::P0;
5155 if (PPRSrcReg != PPRDestReg) {
5167 if (AArch64::ZPRRegClass.
contains(DestReg) &&
5168 AArch64::ZPRRegClass.
contains(SrcReg)) {
5169 assert(Subtarget.isSVEorStreamingSVEAvailable() &&
5170 "Unexpected SVE register.");
5178 if ((AArch64::ZPR2RegClass.
contains(DestReg) ||
5179 AArch64::ZPR2StridedOrContiguousRegClass.
contains(DestReg)) &&
5180 (AArch64::ZPR2RegClass.
contains(SrcReg) ||
5181 AArch64::ZPR2StridedOrContiguousRegClass.
contains(SrcReg))) {
5182 assert(Subtarget.isSVEorStreamingSVEAvailable() &&
5183 "Unexpected SVE register.");
5184 static const unsigned Indices[] = {AArch64::zsub0, AArch64::zsub1};
5191 if (AArch64::ZPR3RegClass.
contains(DestReg) &&
5192 AArch64::ZPR3RegClass.
contains(SrcReg)) {
5193 assert(Subtarget.isSVEorStreamingSVEAvailable() &&
5194 "Unexpected SVE register.");
5195 static const unsigned Indices[] = {AArch64::zsub0, AArch64::zsub1,
5203 if ((AArch64::ZPR4RegClass.
contains(DestReg) ||
5204 AArch64::ZPR4StridedOrContiguousRegClass.
contains(DestReg)) &&
5205 (AArch64::ZPR4RegClass.
contains(SrcReg) ||
5206 AArch64::ZPR4StridedOrContiguousRegClass.
contains(SrcReg))) {
5207 assert(Subtarget.isSVEorStreamingSVEAvailable() &&
5208 "Unexpected SVE register.");
5209 static const unsigned Indices[] = {AArch64::zsub0, AArch64::zsub1,
5210 AArch64::zsub2, AArch64::zsub3};
5216 if (AArch64::GPR64spRegClass.
contains(DestReg) &&
5217 (AArch64::GPR64spRegClass.
contains(SrcReg) || SrcReg == AArch64::XZR)) {
5218 if (DestReg == AArch64::SP || SrcReg == AArch64::SP) {
5224 }
else if (SrcReg == AArch64::XZR && Subtarget.hasZeroCycleZeroingGPR64()) {
5238 if (AArch64::DDDDRegClass.
contains(DestReg) &&
5239 AArch64::DDDDRegClass.
contains(SrcReg)) {
5240 static const unsigned Indices[] = {AArch64::dsub0, AArch64::dsub1,
5241 AArch64::dsub2, AArch64::dsub3};
5248 if (AArch64::DDDRegClass.
contains(DestReg) &&
5249 AArch64::DDDRegClass.
contains(SrcReg)) {
5250 static const unsigned Indices[] = {AArch64::dsub0, AArch64::dsub1,
5258 if (AArch64::DDRegClass.
contains(DestReg) &&
5259 AArch64::DDRegClass.
contains(SrcReg)) {
5260 static const unsigned Indices[] = {AArch64::dsub0, AArch64::dsub1};
5267 if (AArch64::QQQQRegClass.
contains(DestReg) &&
5268 AArch64::QQQQRegClass.
contains(SrcReg)) {
5269 static const unsigned Indices[] = {AArch64::qsub0, AArch64::qsub1,
5270 AArch64::qsub2, AArch64::qsub3};
5277 if (AArch64::QQQRegClass.
contains(DestReg) &&
5278 AArch64::QQQRegClass.
contains(SrcReg)) {
5279 static const unsigned Indices[] = {AArch64::qsub0, AArch64::qsub1,
5287 if (AArch64::QQRegClass.
contains(DestReg) &&
5288 AArch64::QQRegClass.
contains(SrcReg)) {
5289 static const unsigned Indices[] = {AArch64::qsub0, AArch64::qsub1};
5295 if (AArch64::XSeqPairsClassRegClass.
contains(DestReg) &&
5296 AArch64::XSeqPairsClassRegClass.
contains(SrcReg)) {
5297 static const unsigned Indices[] = {AArch64::sube64, AArch64::subo64};
5299 AArch64::XZR, Indices);
5303 if (AArch64::WSeqPairsClassRegClass.
contains(DestReg) &&
5304 AArch64::WSeqPairsClassRegClass.
contains(SrcReg)) {
5305 static const unsigned Indices[] = {AArch64::sube32, AArch64::subo32};
5307 AArch64::WZR, Indices);
5311 if (AArch64::FPR128RegClass.
contains(DestReg) &&
5312 AArch64::FPR128RegClass.
contains(SrcReg)) {
5313 if (Subtarget.isSVEorStreamingSVEAvailable() &&
5314 !Subtarget.isNeonAvailable())
5317 .
addReg(AArch64::Z0 + (SrcReg - AArch64::Q0))
5318 .
addReg(AArch64::Z0 + (SrcReg - AArch64::Q0));
5319 else if (Subtarget.isNeonAvailable())
5338 if (AArch64::FPR64RegClass.
contains(DestReg) &&
5339 AArch64::FPR64RegClass.
contains(SrcReg)) {
5340 if (Subtarget.hasZeroCycleRegMoveFPR128() &&
5341 !Subtarget.hasZeroCycleRegMoveFPR64() &&
5342 !Subtarget.hasZeroCycleRegMoveFPR32() && Subtarget.isNeonAvailable()) {
5344 MCRegister DestRegQ =
TRI->getMatchingSuperReg(DestReg, AArch64::dsub,
5345 &AArch64::FPR128RegClass);
5346 MCRegister SrcRegQ =
TRI->getMatchingSuperReg(SrcReg, AArch64::dsub,
5347 &AArch64::FPR128RegClass);
5363 if (AArch64::FPR32RegClass.
contains(DestReg) &&
5364 AArch64::FPR32RegClass.
contains(SrcReg)) {
5365 if (Subtarget.hasZeroCycleRegMoveFPR128() &&
5366 !Subtarget.hasZeroCycleRegMoveFPR64() &&
5367 !Subtarget.hasZeroCycleRegMoveFPR32() && Subtarget.isNeonAvailable()) {
5369 MCRegister DestRegQ =
TRI->getMatchingSuperReg(DestReg, AArch64::ssub,
5370 &AArch64::FPR128RegClass);
5371 MCRegister SrcRegQ =
TRI->getMatchingSuperReg(SrcReg, AArch64::ssub,
5372 &AArch64::FPR128RegClass);
5381 }
else if (Subtarget.hasZeroCycleRegMoveFPR64() &&
5382 !Subtarget.hasZeroCycleRegMoveFPR32()) {
5384 MCRegister DestRegD =
TRI->getMatchingSuperReg(DestReg, AArch64::ssub,
5385 &AArch64::FPR64RegClass);
5386 MCRegister SrcRegD =
TRI->getMatchingSuperReg(SrcReg, AArch64::ssub,
5387 &AArch64::FPR64RegClass);
5402 if (AArch64::FPR16RegClass.
contains(DestReg) &&
5403 AArch64::FPR16RegClass.
contains(SrcReg)) {
5404 if (Subtarget.hasZeroCycleRegMoveFPR128() &&
5405 !Subtarget.hasZeroCycleRegMoveFPR64() &&
5406 !Subtarget.hasZeroCycleRegMoveFPR32() && Subtarget.isNeonAvailable()) {
5408 MCRegister DestRegQ =
TRI->getMatchingSuperReg(DestReg, AArch64::hsub,
5409 &AArch64::FPR128RegClass);
5410 MCRegister SrcRegQ =
TRI->getMatchingSuperReg(SrcReg, AArch64::hsub,
5411 &AArch64::FPR128RegClass);
5420 }
else if (Subtarget.hasZeroCycleRegMoveFPR64() &&
5421 !Subtarget.hasZeroCycleRegMoveFPR32()) {
5423 MCRegister DestRegD =
TRI->getMatchingSuperReg(DestReg, AArch64::hsub,
5424 &AArch64::FPR64RegClass);
5425 MCRegister SrcRegD =
TRI->getMatchingSuperReg(SrcReg, AArch64::hsub,
5426 &AArch64::FPR64RegClass);
5435 DestReg = RI.getMatchingSuperReg(DestReg, AArch64::hsub,
5436 &AArch64::FPR32RegClass);
5437 SrcReg = RI.getMatchingSuperReg(SrcReg, AArch64::hsub,
5438 &AArch64::FPR32RegClass);
5445 if (AArch64::FPR8RegClass.
contains(DestReg) &&
5446 AArch64::FPR8RegClass.
contains(SrcReg)) {
5447 if (Subtarget.hasZeroCycleRegMoveFPR128() &&
5448 !Subtarget.hasZeroCycleRegMoveFPR64() &&
5449 !Subtarget.hasZeroCycleRegMoveFPR64() && Subtarget.isNeonAvailable()) {
5451 MCRegister DestRegQ =
TRI->getMatchingSuperReg(DestReg, AArch64::bsub,
5452 &AArch64::FPR128RegClass);
5453 MCRegister SrcRegQ =
TRI->getMatchingSuperReg(SrcReg, AArch64::bsub,
5454 &AArch64::FPR128RegClass);
5463 }
else if (Subtarget.hasZeroCycleRegMoveFPR64() &&
5464 !Subtarget.hasZeroCycleRegMoveFPR32()) {
5466 MCRegister DestRegD =
TRI->getMatchingSuperReg(DestReg, AArch64::bsub,
5467 &AArch64::FPR64RegClass);
5468 MCRegister SrcRegD =
TRI->getMatchingSuperReg(SrcReg, AArch64::bsub,
5469 &AArch64::FPR64RegClass);
5478 DestReg = RI.getMatchingSuperReg(DestReg, AArch64::bsub,
5479 &AArch64::FPR32RegClass);
5480 SrcReg = RI.getMatchingSuperReg(SrcReg, AArch64::bsub,
5481 &AArch64::FPR32RegClass);
5489 if (AArch64::FPR64RegClass.
contains(DestReg) &&
5490 AArch64::GPR64RegClass.
contains(SrcReg)) {
5491 if (AArch64::XZR == SrcReg) {
5499 if (AArch64::GPR64RegClass.
contains(DestReg) &&
5500 AArch64::FPR64RegClass.
contains(SrcReg)) {
5506 if (AArch64::FPR32RegClass.
contains(DestReg) &&
5507 AArch64::GPR32RegClass.
contains(SrcReg)) {
5508 if (AArch64::WZR == SrcReg) {
5516 if (AArch64::GPR32RegClass.
contains(DestReg) &&
5517 AArch64::FPR32RegClass.
contains(SrcReg)) {
5523 if (DestReg == AArch64::NZCV) {
5524 assert(AArch64::GPR64RegClass.
contains(SrcReg) &&
"Invalid NZCV copy");
5526 .
addImm(AArch64SysReg::NZCV)
5532 if (SrcReg == AArch64::NZCV) {
5533 assert(AArch64::GPR64RegClass.
contains(DestReg) &&
"Invalid NZCV copy");
5535 .
addImm(AArch64SysReg::NZCV)
5542 errs() <<
TRI.getRegAsmName(DestReg) <<
" = COPY "
5543 <<
TRI.getRegAsmName(SrcReg) <<
"\n";
5553 unsigned SubIdx0,
unsigned SubIdx1,
int FI,
5558 SrcReg0 =
TRI.getSubReg(SrcReg, SubIdx0);
5560 SrcReg1 =
TRI.getSubReg(SrcReg, SubIdx1);
5573 Register SrcReg,
bool isKill,
int FI,
5589 switch (
TRI->getSpillSize(*RC)) {
5591 if (AArch64::FPR8RegClass.hasSubClassEq(RC))
5592 Opc = AArch64::STRBui;
5595 if (AArch64::FPR16RegClass.hasSubClassEq(RC))
5596 Opc = AArch64::STRHui;
5597 else if (AArch64::PNRRegClass.hasSubClassEq(RC) ||
5598 AArch64::PPRRegClass.hasSubClassEq(RC)) {
5599 assert(Subtarget.isSVEorStreamingSVEAvailable() &&
5600 "Unexpected register store without SVE store instructions");
5601 Opc = AArch64::STR_PXI;
5607 if (AArch64::GPR32allRegClass.hasSubClassEq(RC)) {
5608 Opc = AArch64::STRWui;
5612 assert(SrcReg != AArch64::WSP);
5613 }
else if (AArch64::FPR32RegClass.hasSubClassEq(RC))
5614 Opc = AArch64::STRSui;
5615 else if (AArch64::PPR2RegClass.hasSubClassEq(RC)) {
5616 Opc = AArch64::STR_PPXI;
5621 if (AArch64::GPR64allRegClass.hasSubClassEq(RC)) {
5622 Opc = AArch64::STRXui;
5626 assert(SrcReg != AArch64::SP);
5627 }
else if (AArch64::FPR64RegClass.hasSubClassEq(RC)) {
5628 Opc = AArch64::STRDui;
5629 }
else if (AArch64::WSeqPairsClassRegClass.hasSubClassEq(RC)) {
5631 get(AArch64::STPWi), SrcReg, isKill,
5632 AArch64::sube32, AArch64::subo32, FI, MMO);
5637 if (AArch64::FPR128RegClass.hasSubClassEq(RC))
5638 Opc = AArch64::STRQui;
5639 else if (AArch64::DDRegClass.hasSubClassEq(RC)) {
5640 assert(Subtarget.hasNEON() &&
"Unexpected register store without NEON");
5641 Opc = AArch64::ST1Twov1d;
5643 }
else if (AArch64::XSeqPairsClassRegClass.hasSubClassEq(RC)) {
5645 get(AArch64::STPXi), SrcReg, isKill,
5646 AArch64::sube64, AArch64::subo64, FI, MMO);
5648 }
else if (AArch64::ZPRRegClass.hasSubClassEq(RC)) {
5649 assert(Subtarget.isSVEorStreamingSVEAvailable() &&
5650 "Unexpected register store without SVE store instructions");
5651 Opc = AArch64::STR_ZXI;
5653 }
else if (AArch64::PPRRegClass.hasSubClassEq(RC)) {
5654 assert(Subtarget.isSVEorStreamingSVEAvailable() &&
5655 "Unexpected predicate store without SVE store instructions");
5656 Opc = AArch64::SPILL_PPR_TO_ZPR_SLOT_PSEUDO;
5661 if (AArch64::DDDRegClass.hasSubClassEq(RC)) {
5662 assert(Subtarget.hasNEON() &&
"Unexpected register store without NEON");
5663 Opc = AArch64::ST1Threev1d;
5668 if (AArch64::DDDDRegClass.hasSubClassEq(RC)) {
5669 assert(Subtarget.hasNEON() &&
"Unexpected register store without NEON");
5670 Opc = AArch64::ST1Fourv1d;
5672 }
else if (AArch64::QQRegClass.hasSubClassEq(RC)) {
5673 assert(Subtarget.hasNEON() &&
"Unexpected register store without NEON");
5674 Opc = AArch64::ST1Twov2d;
5676 }
else if (AArch64::ZPR2StridedOrContiguousRegClass.hasSubClassEq(RC)) {
5677 assert(Subtarget.isSVEorStreamingSVEAvailable() &&
5678 "Unexpected register store without SVE store instructions");
5679 Opc = AArch64::STR_ZZXI_STRIDED_CONTIGUOUS;
5681 }
else if (AArch64::ZPR2RegClass.hasSubClassEq(RC)) {
5682 assert(Subtarget.isSVEorStreamingSVEAvailable() &&
5683 "Unexpected register store without SVE store instructions");
5684 Opc = AArch64::STR_ZZXI;
5689 if (AArch64::QQQRegClass.hasSubClassEq(RC)) {
5690 assert(Subtarget.hasNEON() &&
"Unexpected register store without NEON");
5691 Opc = AArch64::ST1Threev2d;
5693 }
else if (AArch64::ZPR3RegClass.hasSubClassEq(RC)) {
5694 assert(Subtarget.isSVEorStreamingSVEAvailable() &&
5695 "Unexpected register store without SVE store instructions");
5696 Opc = AArch64::STR_ZZZXI;
5701 if (AArch64::QQQQRegClass.hasSubClassEq(RC)) {
5702 assert(Subtarget.hasNEON() &&
"Unexpected register store without NEON");
5703 Opc = AArch64::ST1Fourv2d;
5705 }
else if (AArch64::ZPR4StridedOrContiguousRegClass.hasSubClassEq(RC)) {
5706 assert(Subtarget.isSVEorStreamingSVEAvailable() &&
5707 "Unexpected register store without SVE store instructions");
5708 Opc = AArch64::STR_ZZZZXI_STRIDED_CONTIGUOUS;
5710 }
else if (AArch64::ZPR4RegClass.hasSubClassEq(RC)) {
5711 assert(Subtarget.isSVEorStreamingSVEAvailable() &&
5712 "Unexpected register store without SVE store instructions");
5713 Opc = AArch64::STR_ZZZZXI;
5718 assert(
Opc &&
"Unknown register class");
5729 MI.addMemOperand(MMO);
5736 Register DestReg,
unsigned SubIdx0,
5737 unsigned SubIdx1,
int FI,
5741 bool IsUndef =
true;
5743 DestReg0 =
TRI.getSubReg(DestReg, SubIdx0);
5745 DestReg1 =
TRI.getSubReg(DestReg, SubIdx1);
5772 switch (
TRI->getSpillSize(*RC)) {
5774 if (AArch64::FPR8RegClass.hasSubClassEq(RC))
5775 Opc = AArch64::LDRBui;
5778 bool IsPNR = AArch64::PNRRegClass.hasSubClassEq(RC);
5779 if (AArch64::FPR16RegClass.hasSubClassEq(RC))
5780 Opc = AArch64::LDRHui;
5781 else if (IsPNR || AArch64::PPRRegClass.hasSubClassEq(RC)) {
5782 assert(Subtarget.isSVEorStreamingSVEAvailable() &&
5783 "Unexpected register load without SVE load instructions");
5786 Opc = AArch64::LDR_PXI;
5792 if (AArch64::GPR32allRegClass.hasSubClassEq(RC)) {
5793 Opc = AArch64::LDRWui;
5797 assert(DestReg != AArch64::WSP);
5798 }
else if (AArch64::FPR32RegClass.hasSubClassEq(RC))
5799 Opc = AArch64::LDRSui;
5800 else if (AArch64::PPR2RegClass.hasSubClassEq(RC)) {
5801 Opc = AArch64::LDR_PPXI;
5806 if (AArch64::GPR64allRegClass.hasSubClassEq(RC)) {
5807 Opc = AArch64::LDRXui;
5811 assert(DestReg != AArch64::SP);
5812 }
else if (AArch64::FPR64RegClass.hasSubClassEq(RC)) {
5813 Opc = AArch64::LDRDui;
5814 }
else if (AArch64::WSeqPairsClassRegClass.hasSubClassEq(RC)) {
5816 get(AArch64::LDPWi), DestReg, AArch64::sube32,
5817 AArch64::subo32, FI, MMO);
5822 if (AArch64::FPR128RegClass.hasSubClassEq(RC))
5823 Opc = AArch64::LDRQui;
5824 else if (AArch64::DDRegClass.hasSubClassEq(RC)) {
5825 assert(Subtarget.hasNEON() &&
"Unexpected register load without NEON");
5826 Opc = AArch64::LD1Twov1d;
5828 }
else if (AArch64::XSeqPairsClassRegClass.hasSubClassEq(RC)) {
5830 get(AArch64::LDPXi), DestReg, AArch64::sube64,
5831 AArch64::subo64, FI, MMO);
5833 }
else if (AArch64::ZPRRegClass.hasSubClassEq(RC)) {
5834 assert(Subtarget.isSVEorStreamingSVEAvailable() &&
5835 "Unexpected register load without SVE load instructions");
5836 Opc = AArch64::LDR_ZXI;
5838 }
else if (AArch64::PPRRegClass.hasSubClassEq(RC)) {
5839 assert(Subtarget.isSVEorStreamingSVEAvailable() &&
5840 "Unexpected predicate load without SVE load instructions");
5841 Opc = AArch64::FILL_PPR_FROM_ZPR_SLOT_PSEUDO;
5846 if (AArch64::DDDRegClass.hasSubClassEq(RC)) {
5847 assert(Subtarget.hasNEON() &&
"Unexpected register load without NEON");
5848 Opc = AArch64::LD1Threev1d;
5853 if (AArch64::DDDDRegClass.hasSubClassEq(RC)) {
5854 assert(Subtarget.hasNEON() &&
"Unexpected register load without NEON");
5855 Opc = AArch64::LD1Fourv1d;
5857 }
else if (AArch64::QQRegClass.hasSubClassEq(RC)) {
5858 assert(Subtarget.hasNEON() &&
"Unexpected register load without NEON");
5859 Opc = AArch64::LD1Twov2d;
5861 }
else if (AArch64::ZPR2StridedOrContiguousRegClass.hasSubClassEq(RC)) {
5862 assert(Subtarget.isSVEorStreamingSVEAvailable() &&
5863 "Unexpected register load without SVE load instructions");
5864 Opc = AArch64::LDR_ZZXI_STRIDED_CONTIGUOUS;
5866 }
else if (AArch64::ZPR2RegClass.hasSubClassEq(RC)) {
5867 assert(Subtarget.isSVEorStreamingSVEAvailable() &&
5868 "Unexpected register load without SVE load instructions");
5869 Opc = AArch64::LDR_ZZXI;
5874 if (AArch64::QQQRegClass.hasSubClassEq(RC)) {
5875 assert(Subtarget.hasNEON() &&
"Unexpected register load without NEON");
5876 Opc = AArch64::LD1Threev2d;
5878 }
else if (AArch64::ZPR3RegClass.hasSubClassEq(RC)) {
5879 assert(Subtarget.isSVEorStreamingSVEAvailable() &&
5880 "Unexpected register load without SVE load instructions");
5881 Opc = AArch64::LDR_ZZZXI;
5886 if (AArch64::QQQQRegClass.hasSubClassEq(RC)) {
5887 assert(Subtarget.hasNEON() &&
"Unexpected register load without NEON");
5888 Opc = AArch64::LD1Fourv2d;
5890 }
else if (AArch64::ZPR4StridedOrContiguousRegClass.hasSubClassEq(RC)) {
5891 assert(Subtarget.isSVEorStreamingSVEAvailable() &&
5892 "Unexpected register load without SVE load instructions");
5893 Opc = AArch64::LDR_ZZZZXI_STRIDED_CONTIGUOUS;
5895 }
else if (AArch64::ZPR4RegClass.hasSubClassEq(RC)) {
5896 assert(Subtarget.isSVEorStreamingSVEAvailable() &&
5897 "Unexpected register load without SVE load instructions");
5898 Opc = AArch64::LDR_ZZZZXI;
5904 assert(
Opc &&
"Unknown register class");
5914 MI.addMemOperand(MMO);
5921 UseMI.getIterator()),
5923 return I.modifiesRegister(AArch64::NZCV, TRI) ||
5924 I.readsRegister(AArch64::NZCV, TRI);
5928void AArch64InstrInfo::decomposeStackOffsetForDwarfOffsets(
5933 assert(
Offset.getScalable() % 2 == 0 &&
"Invalid frame offset");
5940 ByteSized =
Offset.getFixed();
5941 VGSized =
Offset.getScalable() / 2;
5947void AArch64InstrInfo::decomposeStackOffsetForFrameOffsets(
5949 int64_t &NumDataVectors) {
5953 assert(
Offset.getScalable() % 2 == 0 &&
"Invalid frame offset");
5955 NumBytes =
Offset.getFixed();
5957 NumPredicateVectors =
Offset.getScalable() / 2;
5962 if (NumPredicateVectors % 8 == 0 || NumPredicateVectors < -64 ||
5963 NumPredicateVectors > 62) {
5964 NumDataVectors = NumPredicateVectors / 8;
5965 NumPredicateVectors -= NumDataVectors * 8;
5991 Expr.
push_back((
char)dwarf::DW_OP_bregx);
5999 int64_t OffsetFromDefCFA) {
6013 Comment << (NumBytes < 0 ?
" - " :
" + ") << std::abs(NumBytes);
6014 if (!RegScale.empty())
6024 int64_t NumBytes, NumVGScaledBytes;
6025 AArch64InstrInfo::decomposeStackOffsetForDwarfOffsets(
Offset, NumBytes,
6027 std::string CommentBuffer;
6030 if (
Reg == AArch64::SP)
6032 else if (
Reg == AArch64::FP)
6039 unsigned DwarfReg =
TRI.getDwarfRegNum(
Reg,
true);
6040 assert(DwarfReg <= 31 &&
"DwarfReg out of bounds (0..31)");
6042 Expr.
push_back(dwarf::DW_OP_breg0 + DwarfReg);
6045 if (NumVGScaledBytes) {
6055 DefCfaExpr.
push_back(dwarf::DW_CFA_def_cfa_expression);
6063 unsigned FrameReg,
unsigned Reg,
6065 bool LastAdjustmentWasScalable) {
6066 if (
Offset.getScalable())
6069 if (FrameReg == Reg && !LastAdjustmentWasScalable)
6072 unsigned DwarfReg =
TRI.getDwarfRegNum(Reg,
true);
6079 std::optional<int64_t> IncomingVGOffsetFromDefCFA) {
6080 int64_t NumBytes, NumVGScaledBytes;
6081 AArch64InstrInfo::decomposeStackOffsetForDwarfOffsets(
6082 OffsetFromDefCFA, NumBytes, NumVGScaledBytes);
6084 unsigned DwarfReg =
TRI.getDwarfRegNum(Reg,
true);
6087 if (!NumVGScaledBytes)
6090 std::string CommentBuffer;
6095 assert(NumVGScaledBytes &&
"Expected scalable offset");
6099 if (IncomingVGOffsetFromDefCFA) {
6101 VGRegScale =
"* IncomingVG";
6104 VGRegScale =
"* VG";
6108 OffsetExpr.
push_back(dwarf::DW_OP_plus);
6117 CfaExpr.
push_back(dwarf::DW_CFA_expression);
6132 unsigned SrcReg, int64_t
Offset,
unsigned Opc,
6135 bool *HasWinCFI,
bool EmitCFAOffset,
6138 unsigned MaxEncoding, ShiftSize;
6140 case AArch64::ADDXri:
6141 case AArch64::ADDSXri:
6142 case AArch64::SUBXri:
6143 case AArch64::SUBSXri:
6144 MaxEncoding = 0xfff;
6147 case AArch64::ADDVL_XXI:
6148 case AArch64::ADDPL_XXI:
6149 case AArch64::ADDSVL_XXI:
6150 case AArch64::ADDSPL_XXI:
6165 if (
Opc == AArch64::ADDVL_XXI ||
Opc == AArch64::ADDSVL_XXI)
6167 else if (
Opc == AArch64::ADDPL_XXI ||
Opc == AArch64::ADDSPL_XXI)
6181 const unsigned MaxEncodableValue = MaxEncoding << ShiftSize;
6183 if (TmpReg == AArch64::XZR)
6184 TmpReg =
MBB.getParent()->getRegInfo().createVirtualRegister(
6185 &AArch64::GPR64RegClass);
6187 uint64_t ThisVal = std::min<uint64_t>(
Offset, MaxEncodableValue);
6188 unsigned LocalShiftSize = 0;
6189 if (ThisVal > MaxEncoding) {
6190 ThisVal = ThisVal >> ShiftSize;
6191 LocalShiftSize = ShiftSize;
6193 assert((ThisVal >> ShiftSize) <= MaxEncoding &&
6194 "Encoding cannot handle value that big");
6196 Offset -= ThisVal << LocalShiftSize;
6201 .
addImm(Sign * (
int)ThisVal);
6211 if (Sign == -1 ||
Opc == AArch64::SUBXri ||
Opc == AArch64::SUBSXri)
6212 CFAOffset += Change;
6214 CFAOffset -= Change;
6215 if (EmitCFAOffset && DestReg == TmpReg) {
6228 int Imm = (int)(ThisVal << LocalShiftSize);
6229 if (VScale != 1 && DestReg == AArch64::SP) {
6235 }
else if ((DestReg == AArch64::FP && SrcReg == AArch64::SP) ||
6236 (SrcReg == AArch64::FP && DestReg == AArch64::SP)) {
6237 assert(VScale == 1 &&
"Expected non-scalable operation");
6246 assert(
Offset == 0 &&
"Expected remaining offset to be zero to "
6247 "emit a single SEH directive");
6248 }
else if (DestReg == AArch64::SP) {
6249 assert(VScale == 1 &&
"Expected non-scalable operation");
6252 assert(SrcReg == AArch64::SP &&
"Unexpected SrcReg for SEH_StackAlloc");
6265 unsigned DestReg,
unsigned SrcReg,
6268 bool NeedsWinCFI,
bool *HasWinCFI,
6270 unsigned FrameReg) {
6277 bool UseSVL =
F.hasFnAttribute(
"aarch64_pstate_sm_body");
6279 int64_t Bytes, NumPredicateVectors, NumDataVectors;
6280 AArch64InstrInfo::decomposeStackOffsetForFrameOffsets(
6281 Offset, Bytes, NumPredicateVectors, NumDataVectors);
6284 bool NeedsFinalDefNZCV = SetNZCV && (NumPredicateVectors || NumDataVectors);
6285 if (NeedsFinalDefNZCV)
6289 if (Bytes || (!
Offset && SrcReg != DestReg)) {
6290 assert((DestReg != AArch64::SP || Bytes % 8 == 0) &&
6291 "SP increment/decrement not 8-byte aligned");
6292 unsigned Opc = SetNZCV ? AArch64::ADDSXri : AArch64::ADDXri;
6295 Opc = SetNZCV ? AArch64::SUBSXri : AArch64::SUBXri;
6298 NeedsWinCFI, HasWinCFI, EmitCFAOffset, CFAOffset,
6300 CFAOffset += (
Opc == AArch64::ADDXri ||
Opc == AArch64::ADDSXri)
6307 assert(!(NeedsWinCFI && NumPredicateVectors) &&
6308 "WinCFI can't allocate fractions of an SVE data vector");
6310 if (NumDataVectors) {
6312 UseSVL ? AArch64::ADDSVL_XXI : AArch64::ADDVL_XXI,
TII,
6313 Flag, NeedsWinCFI, HasWinCFI, EmitCFAOffset, CFAOffset,
6319 if (NumPredicateVectors) {
6320 assert(DestReg != AArch64::SP &&
"Unaligned access to SP");
6322 UseSVL ? AArch64::ADDSPL_XXI : AArch64::ADDPL_XXI,
TII,
6323 Flag, NeedsWinCFI, HasWinCFI, EmitCFAOffset, CFAOffset,
6327 if (NeedsFinalDefNZCV)
6348 if (
MI.isFullCopy()) {
6351 if (SrcReg == AArch64::SP && DstReg.
isVirtual()) {
6355 if (DstReg == AArch64::SP && SrcReg.
isVirtual()) {
6360 if (SrcReg == AArch64::NZCV || DstReg == AArch64::NZCV)
6388 if (
MI.isCopy() &&
Ops.size() == 1 &&
6390 (
Ops[0] == 0 ||
Ops[0] == 1)) {
6391 bool IsSpill =
Ops[0] == 0;
6392 bool IsFill = !IsSpill;
6404 :
TRI.getMinimalPhysRegClass(Reg);
6410 "Mismatched register size in non subreg COPY");
6417 return &*--InsertPt;
6429 if (IsSpill && DstMO.
isUndef() && SrcReg == AArch64::WZR &&
6432 "Unexpected subreg on physical register");
6434 FrameIndex, &AArch64::GPR64RegClass, &
TRI,
6436 return &*--InsertPt;
6453 case AArch64::sub_32:
6454 if (AArch64::GPR64RegClass.hasSubClassEq(
getRegClass(DstReg)))
6455 FillRC = &AArch64::GPR32RegClass;
6458 FillRC = &AArch64::FPR32RegClass;
6461 FillRC = &AArch64::FPR64RegClass;
6467 TRI.getRegSizeInBits(*FillRC) &&
6468 "Mismatched regclass size on folded subreg COPY");
6487 bool *OutUseUnscaledOp,
6488 unsigned *OutUnscaledOp,
6489 int64_t *EmittableOffset) {
6491 if (EmittableOffset)
6492 *EmittableOffset = 0;
6493 if (OutUseUnscaledOp)
6494 *OutUseUnscaledOp =
false;
6500 switch (
MI.getOpcode()) {
6503 case AArch64::LD1Rv1d:
6504 case AArch64::LD1Rv2s:
6505 case AArch64::LD1Rv2d:
6506 case AArch64::LD1Rv4h:
6507 case AArch64::LD1Rv4s:
6508 case AArch64::LD1Rv8b:
6509 case AArch64::LD1Rv8h:
6510 case AArch64::LD1Rv16b:
6511 case AArch64::LD1Twov2d:
6512 case AArch64::LD1Threev2d:
6513 case AArch64::LD1Fourv2d:
6514 case AArch64::LD1Twov1d:
6515 case AArch64::LD1Threev1d:
6516 case AArch64::LD1Fourv1d:
6517 case AArch64::ST1Twov2d:
6518 case AArch64::ST1Threev2d:
6519 case AArch64::ST1Fourv2d:
6520 case AArch64::ST1Twov1d:
6521 case AArch64::ST1Threev1d:
6522 case AArch64::ST1Fourv1d:
6523 case AArch64::ST1i8:
6524 case AArch64::ST1i16:
6525 case AArch64::ST1i32:
6526 case AArch64::ST1i64:
6528 case AArch64::IRGstack:
6529 case AArch64::STGloop:
6530 case AArch64::STZGloop:
6535 TypeSize ScaleValue(0U,
false), Width(0U,
false);
6536 int64_t MinOff, MaxOff;
6542 bool IsMulVL = ScaleValue.isScalable();
6543 unsigned Scale = ScaleValue.getKnownMinValue();
6553 std::optional<unsigned> UnscaledOp =
6555 bool useUnscaledOp = UnscaledOp && (
Offset % Scale ||
Offset < 0);
6556 if (useUnscaledOp &&
6561 Scale = ScaleValue.getKnownMinValue();
6562 assert(IsMulVL == ScaleValue.isScalable() &&
6563 "Unscaled opcode has different value for scalable");
6565 int64_t Remainder =
Offset % Scale;
6566 assert(!(Remainder && useUnscaledOp) &&
6567 "Cannot have remainder when using unscaled op");
6569 assert(MinOff < MaxOff &&
"Unexpected Min/Max offsets");
6570 int64_t NewOffset =
Offset / Scale;
6571 if (MinOff <= NewOffset && NewOffset <= MaxOff)
6574 NewOffset = NewOffset < 0 ? MinOff : MaxOff;
6578 if (EmittableOffset)
6579 *EmittableOffset = NewOffset;
6580 if (OutUseUnscaledOp)
6581 *OutUseUnscaledOp = useUnscaledOp;
6582 if (OutUnscaledOp && UnscaledOp)
6583 *OutUnscaledOp = *UnscaledOp;
6596 unsigned Opcode =
MI.getOpcode();
6597 unsigned ImmIdx = FrameRegIdx + 1;
6599 if (Opcode == AArch64::ADDSXri || Opcode == AArch64::ADDXri) {
6604 MI.eraseFromParent();
6610 unsigned UnscaledOp;
6613 &UnscaledOp, &NewOffset);
6617 MI.getOperand(FrameRegIdx).ChangeToRegister(FrameReg,
false);
6619 MI.setDesc(
TII->get(UnscaledOp));
6621 MI.getOperand(ImmIdx).ChangeToImmediate(NewOffset);
6639bool AArch64InstrInfo::useMachineCombiner()
const {
return true; }
6644 case AArch64::ADDSWrr:
6645 case AArch64::ADDSWri:
6646 case AArch64::ADDSXrr:
6647 case AArch64::ADDSXri:
6648 case AArch64::SUBSWrr:
6649 case AArch64::SUBSXrr:
6651 case AArch64::SUBSWri:
6652 case AArch64::SUBSXri:
6663 case AArch64::ADDWrr:
6664 case AArch64::ADDWri:
6665 case AArch64::SUBWrr:
6666 case AArch64::ADDSWrr:
6667 case AArch64::ADDSWri:
6668 case AArch64::SUBSWrr:
6670 case AArch64::SUBWri:
6671 case AArch64::SUBSWri:
6682 case AArch64::ADDXrr:
6683 case AArch64::ADDXri:
6684 case AArch64::SUBXrr:
6685 case AArch64::ADDSXrr:
6686 case AArch64::ADDSXri:
6687 case AArch64::SUBSXrr:
6689 case AArch64::SUBXri:
6690 case AArch64::SUBSXri:
6691 case AArch64::ADDv8i8:
6692 case AArch64::ADDv16i8:
6693 case AArch64::ADDv4i16:
6694 case AArch64::ADDv8i16:
6695 case AArch64::ADDv2i32:
6696 case AArch64::ADDv4i32:
6697 case AArch64::SUBv8i8:
6698 case AArch64::SUBv16i8:
6699 case AArch64::SUBv4i16:
6700 case AArch64::SUBv8i16:
6701 case AArch64::SUBv2i32:
6702 case AArch64::SUBv4i32:
6715 case AArch64::FADDHrr:
6716 case AArch64::FADDSrr:
6717 case AArch64::FADDDrr:
6718 case AArch64::FADDv4f16:
6719 case AArch64::FADDv8f16:
6720 case AArch64::FADDv2f32:
6721 case AArch64::FADDv2f64:
6722 case AArch64::FADDv4f32:
6723 case AArch64::FSUBHrr:
6724 case AArch64::FSUBSrr:
6725 case AArch64::FSUBDrr:
6726 case AArch64::FSUBv4f16:
6727 case AArch64::FSUBv8f16:
6728 case AArch64::FSUBv2f32:
6729 case AArch64::FSUBv2f64:
6730 case AArch64::FSUBv4f32:
6749 unsigned CombineOpc,
unsigned ZeroReg = 0,
6750 bool CheckZeroReg =
false) {
6757 if (!
MI ||
MI->getParent() != &
MBB ||
MI->getOpcode() != CombineOpc)
6760 if (!
MRI.hasOneNonDBGUse(
MI->getOperand(0).getReg()))
6764 assert(
MI->getNumOperands() >= 4 &&
MI->getOperand(0).isReg() &&
6765 MI->getOperand(1).isReg() &&
MI->getOperand(2).isReg() &&
6766 MI->getOperand(3).isReg() &&
"MAdd/MSub must have a least 4 regs");
6768 if (
MI->getOperand(3).getReg() != ZeroReg)
6773 MI->findRegisterDefOperandIdx(AArch64::NZCV,
nullptr,
true) == -1)
6782 unsigned MulOpc,
unsigned ZeroReg) {
6797bool AArch64InstrInfo::isAssociativeAndCommutative(
const MachineInstr &Inst,
6798 bool Invert)
const {
6804 case AArch64::FADDHrr:
6805 case AArch64::FADDSrr:
6806 case AArch64::FADDDrr:
6807 case AArch64::FMULHrr:
6808 case AArch64::FMULSrr:
6809 case AArch64::FMULDrr:
6810 case AArch64::FMULX16:
6811 case AArch64::FMULX32:
6812 case AArch64::FMULX64:
6814 case AArch64::FADDv4f16:
6815 case AArch64::FADDv8f16:
6816 case AArch64::FADDv2f32:
6817 case AArch64::FADDv4f32:
6818 case AArch64::FADDv2f64:
6819 case AArch64::FMULv4f16:
6820 case AArch64::FMULv8f16:
6821 case AArch64::FMULv2f32:
6822 case AArch64::FMULv4f32:
6823 case AArch64::FMULv2f64:
6824 case AArch64::FMULXv4f16:
6825 case AArch64::FMULXv8f16:
6826 case AArch64::FMULXv2f32:
6827 case AArch64::FMULXv4f32:
6828 case AArch64::FMULXv2f64:
6832 case AArch64::FADD_ZZZ_H:
6833 case AArch64::FADD_ZZZ_S:
6834 case AArch64::FADD_ZZZ_D:
6835 case AArch64::FMUL_ZZZ_H:
6836 case AArch64::FMUL_ZZZ_S:
6837 case AArch64::FMUL_ZZZ_D:
6848 case AArch64::ADDWrr:
6849 case AArch64::ADDXrr:
6850 case AArch64::ANDWrr:
6851 case AArch64::ANDXrr:
6852 case AArch64::ORRWrr:
6853 case AArch64::ORRXrr:
6854 case AArch64::EORWrr:
6855 case AArch64::EORXrr:
6856 case AArch64::EONWrr:
6857 case AArch64::EONXrr:
6861 case AArch64::ADDv8i8:
6862 case AArch64::ADDv16i8:
6863 case AArch64::ADDv4i16:
6864 case AArch64::ADDv8i16:
6865 case AArch64::ADDv2i32:
6866 case AArch64::ADDv4i32:
6867 case AArch64::ADDv1i64:
6868 case AArch64::ADDv2i64:
6869 case AArch64::MULv8i8:
6870 case AArch64::MULv16i8:
6871 case AArch64::MULv4i16:
6872 case AArch64::MULv8i16:
6873 case AArch64::MULv2i32:
6874 case AArch64::MULv4i32:
6875 case AArch64::ANDv8i8:
6876 case AArch64::ANDv16i8:
6877 case AArch64::ORRv8i8:
6878 case AArch64::ORRv16i8:
6879 case AArch64::EORv8i8:
6880 case AArch64::EORv16i8:
6882 case AArch64::ADD_ZZZ_B:
6883 case AArch64::ADD_ZZZ_H:
6884 case AArch64::ADD_ZZZ_S:
6885 case AArch64::ADD_ZZZ_D:
6886 case AArch64::MUL_ZZZ_B:
6887 case AArch64::MUL_ZZZ_H:
6888 case AArch64::MUL_ZZZ_S:
6889 case AArch64::MUL_ZZZ_D:
6890 case AArch64::AND_ZZZ:
6891 case AArch64::ORR_ZZZ:
6892 case AArch64::EOR_ZZZ:
6923 auto setFound = [&](
int Opcode,
int Operand,
unsigned ZeroReg,
6931 auto setVFound = [&](
int Opcode,
int Operand,
unsigned Pattern) {
6943 case AArch64::ADDWrr:
6945 "ADDWrr does not have register operands");
6946 setFound(AArch64::MADDWrrr, 1, AArch64::WZR, MCP::MULADDW_OP1);
6947 setFound(AArch64::MADDWrrr, 2, AArch64::WZR, MCP::MULADDW_OP2);
6949 case AArch64::ADDXrr:
6950 setFound(AArch64::MADDXrrr, 1, AArch64::XZR, MCP::MULADDX_OP1);
6951 setFound(AArch64::MADDXrrr, 2, AArch64::XZR, MCP::MULADDX_OP2);
6953 case AArch64::SUBWrr:
6954 setFound(AArch64::MADDWrrr, 2, AArch64::WZR, MCP::MULSUBW_OP2);
6955 setFound(AArch64::MADDWrrr, 1, AArch64::WZR, MCP::MULSUBW_OP1);
6957 case AArch64::SUBXrr:
6958 setFound(AArch64::MADDXrrr, 2, AArch64::XZR, MCP::MULSUBX_OP2);
6959 setFound(AArch64::MADDXrrr, 1, AArch64::XZR, MCP::MULSUBX_OP1);
6961 case AArch64::ADDWri:
6962 setFound(AArch64::MADDWrrr, 1, AArch64::WZR, MCP::MULADDWI_OP1);
6964 case AArch64::ADDXri:
6965 setFound(AArch64::MADDXrrr, 1, AArch64::XZR, MCP::MULADDXI_OP1);
6967 case AArch64::SUBWri:
6968 setFound(AArch64::MADDWrrr, 1, AArch64::WZR, MCP::MULSUBWI_OP1);
6970 case AArch64::SUBXri:
6971 setFound(AArch64::MADDXrrr, 1, AArch64::XZR, MCP::MULSUBXI_OP1);
6973 case AArch64::ADDv8i8:
6974 setVFound(AArch64::MULv8i8, 1, MCP::MULADDv8i8_OP1);
6975 setVFound(AArch64::MULv8i8, 2, MCP::MULADDv8i8_OP2);
6977 case AArch64::ADDv16i8:
6978 setVFound(AArch64::MULv16i8, 1, MCP::MULADDv16i8_OP1);
6979 setVFound(AArch64::MULv16i8, 2, MCP::MULADDv16i8_OP2);
6981 case AArch64::ADDv4i16:
6982 setVFound(AArch64::MULv4i16, 1, MCP::MULADDv4i16_OP1);
6983 setVFound(AArch64::MULv4i16, 2, MCP::MULADDv4i16_OP2);
6984 setVFound(AArch64::MULv4i16_indexed, 1, MCP::MULADDv4i16_indexed_OP1);
6985 setVFound(AArch64::MULv4i16_indexed, 2, MCP::MULADDv4i16_indexed_OP2);
6987 case AArch64::ADDv8i16:
6988 setVFound(AArch64::MULv8i16, 1, MCP::MULADDv8i16_OP1);
6989 setVFound(AArch64::MULv8i16, 2, MCP::MULADDv8i16_OP2);
6990 setVFound(AArch64::MULv8i16_indexed, 1, MCP::MULADDv8i16_indexed_OP1);
6991 setVFound(AArch64::MULv8i16_indexed, 2, MCP::MULADDv8i16_indexed_OP2);
6993 case AArch64::ADDv2i32:
6994 setVFound(AArch64::MULv2i32, 1, MCP::MULADDv2i32_OP1);
6995 setVFound(AArch64::MULv2i32, 2, MCP::MULADDv2i32_OP2);
6996 setVFound(AArch64::MULv2i32_indexed, 1, MCP::MULADDv2i32_indexed_OP1);
6997 setVFound(AArch64::MULv2i32_indexed, 2, MCP::MULADDv2i32_indexed_OP2);
6999 case AArch64::ADDv4i32:
7000 setVFound(AArch64::MULv4i32, 1, MCP::MULADDv4i32_OP1);
7001 setVFound(AArch64::MULv4i32, 2, MCP::MULADDv4i32_OP2);
7002 setVFound(AArch64::MULv4i32_indexed, 1, MCP::MULADDv4i32_indexed_OP1);
7003 setVFound(AArch64::MULv4i32_indexed, 2, MCP::MULADDv4i32_indexed_OP2);
7005 case AArch64::SUBv8i8:
7006 setVFound(AArch64::MULv8i8, 1, MCP::MULSUBv8i8_OP1);
7007 setVFound(AArch64::MULv8i8, 2, MCP::MULSUBv8i8_OP2);
7009 case AArch64::SUBv16i8:
7010 setVFound(AArch64::MULv16i8, 1, MCP::MULSUBv16i8_OP1);
7011 setVFound(AArch64::MULv16i8, 2, MCP::MULSUBv16i8_OP2);
7013 case AArch64::SUBv4i16:
7014 setVFound(AArch64::MULv4i16, 1, MCP::MULSUBv4i16_OP1);
7015 setVFound(AArch64::MULv4i16, 2, MCP::MULSUBv4i16_OP2);
7016 setVFound(AArch64::MULv4i16_indexed, 1, MCP::MULSUBv4i16_indexed_OP1);
7017 setVFound(AArch64::MULv4i16_indexed, 2, MCP::MULSUBv4i16_indexed_OP2);
7019 case AArch64::SUBv8i16:
7020 setVFound(AArch64::MULv8i16, 1, MCP::MULSUBv8i16_OP1);
7021 setVFound(AArch64::MULv8i16, 2, MCP::MULSUBv8i16_OP2);
7022 setVFound(AArch64::MULv8i16_indexed, 1, MCP::MULSUBv8i16_indexed_OP1);
7023 setVFound(AArch64::MULv8i16_indexed, 2, MCP::MULSUBv8i16_indexed_OP2);
7025 case AArch64::SUBv2i32:
7026 setVFound(AArch64::MULv2i32, 1, MCP::MULSUBv2i32_OP1);
7027 setVFound(AArch64::MULv2i32, 2, MCP::MULSUBv2i32_OP2);
7028 setVFound(AArch64::MULv2i32_indexed, 1, MCP::MULSUBv2i32_indexed_OP1);
7029 setVFound(AArch64::MULv2i32_indexed, 2, MCP::MULSUBv2i32_indexed_OP2);
7031 case AArch64::SUBv4i32:
7032 setVFound(AArch64::MULv4i32, 1, MCP::MULSUBv4i32_OP1);
7033 setVFound(AArch64::MULv4i32, 2, MCP::MULSUBv4i32_OP2);
7034 setVFound(AArch64::MULv4i32_indexed, 1, MCP::MULSUBv4i32_indexed_OP1);
7035 setVFound(AArch64::MULv4i32_indexed, 2, MCP::MULSUBv4i32_indexed_OP2);
7041bool AArch64InstrInfo::isAccumulationOpcode(
unsigned Opcode)
const {
7045 case AArch64::UABALB_ZZZ_D:
7046 case AArch64::UABALB_ZZZ_H:
7047 case AArch64::UABALB_ZZZ_S:
7048 case AArch64::UABALT_ZZZ_D:
7049 case AArch64::UABALT_ZZZ_H:
7050 case AArch64::UABALT_ZZZ_S:
7051 case AArch64::SABALB_ZZZ_D:
7052 case AArch64::SABALB_ZZZ_S:
7053 case AArch64::SABALB_ZZZ_H:
7054 case AArch64::SABALT_ZZZ_D:
7055 case AArch64::SABALT_ZZZ_S:
7056 case AArch64::SABALT_ZZZ_H:
7057 case AArch64::UABALv16i8_v8i16:
7058 case AArch64::UABALv2i32_v2i64:
7059 case AArch64::UABALv4i16_v4i32:
7060 case AArch64::UABALv4i32_v2i64:
7061 case AArch64::UABALv8i16_v4i32:
7062 case AArch64::UABALv8i8_v8i16:
7063 case AArch64::UABAv16i8:
7064 case AArch64::UABAv2i32:
7065 case AArch64::UABAv4i16:
7066 case AArch64::UABAv4i32:
7067 case AArch64::UABAv8i16:
7068 case AArch64::UABAv8i8:
7069 case AArch64::SABALv16i8_v8i16:
7070 case AArch64::SABALv2i32_v2i64:
7071 case AArch64::SABALv4i16_v4i32:
7072 case AArch64::SABALv4i32_v2i64:
7073 case AArch64::SABALv8i16_v4i32:
7074 case AArch64::SABALv8i8_v8i16:
7075 case AArch64::SABAv16i8:
7076 case AArch64::SABAv2i32:
7077 case AArch64::SABAv4i16:
7078 case AArch64::SABAv4i32:
7079 case AArch64::SABAv8i16:
7080 case AArch64::SABAv8i8:
7087unsigned AArch64InstrInfo::getAccumulationStartOpcode(
7088 unsigned AccumulationOpcode)
const {
7089 switch (AccumulationOpcode) {
7092 case AArch64::UABALB_ZZZ_D:
7093 return AArch64::UABDLB_ZZZ_D;
7094 case AArch64::UABALB_ZZZ_H:
7095 return AArch64::UABDLB_ZZZ_H;
7096 case AArch64::UABALB_ZZZ_S:
7097 return AArch64::UABDLB_ZZZ_S;
7098 case AArch64::UABALT_ZZZ_D:
7099 return AArch64::UABDLT_ZZZ_D;
7100 case AArch64::UABALT_ZZZ_H:
7101 return AArch64::UABDLT_ZZZ_H;
7102 case AArch64::UABALT_ZZZ_S:
7103 return AArch64::UABDLT_ZZZ_S;
7104 case AArch64::UABALv16i8_v8i16:
7105 return AArch64::UABDLv16i8_v8i16;
7106 case AArch64::UABALv2i32_v2i64:
7107 return AArch64::UABDLv2i32_v2i64;
7108 case AArch64::UABALv4i16_v4i32:
7109 return AArch64::UABDLv4i16_v4i32;
7110 case AArch64::UABALv4i32_v2i64:
7111 return AArch64::UABDLv4i32_v2i64;
7112 case AArch64::UABALv8i16_v4i32:
7113 return AArch64::UABDLv8i16_v4i32;
7114 case AArch64::UABALv8i8_v8i16:
7115 return AArch64::UABDLv8i8_v8i16;
7116 case AArch64::UABAv16i8:
7117 return AArch64::UABDv16i8;
7118 case AArch64::UABAv2i32:
7119 return AArch64::UABDv2i32;
7120 case AArch64::UABAv4i16:
7121 return AArch64::UABDv4i16;
7122 case AArch64::UABAv4i32:
7123 return AArch64::UABDv4i32;
7124 case AArch64::UABAv8i16:
7125 return AArch64::UABDv8i16;
7126 case AArch64::UABAv8i8:
7127 return AArch64::UABDv8i8;
7128 case AArch64::SABALB_ZZZ_D:
7129 return AArch64::SABDLB_ZZZ_D;
7130 case AArch64::SABALB_ZZZ_S:
7131 return AArch64::SABDLB_ZZZ_S;
7132 case AArch64::SABALB_ZZZ_H:
7133 return AArch64::SABDLB_ZZZ_H;
7134 case AArch64::SABALT_ZZZ_D:
7135 return AArch64::SABDLT_ZZZ_D;
7136 case AArch64::SABALT_ZZZ_S:
7137 return AArch64::SABDLT_ZZZ_S;
7138 case AArch64::SABALT_ZZZ_H:
7139 return AArch64::SABDLT_ZZZ_H;
7140 case AArch64::SABALv16i8_v8i16:
7141 return AArch64::SABDLv16i8_v8i16;
7142 case AArch64::SABALv2i32_v2i64:
7143 return AArch64::SABDLv2i32_v2i64;
7144 case AArch64::SABALv4i16_v4i32:
7145 return AArch64::SABDLv4i16_v4i32;
7146 case AArch64::SABALv4i32_v2i64:
7147 return AArch64::SABDLv4i32_v2i64;
7148 case AArch64::SABALv8i16_v4i32:
7149 return AArch64::SABDLv8i16_v4i32;
7150 case AArch64::SABALv8i8_v8i16:
7151 return AArch64::SABDLv8i8_v8i16;
7152 case AArch64::SABAv16i8:
7153 return AArch64::SABDv16i8;
7154 case AArch64::SABAv2i32:
7155 return AArch64::SABAv2i32;
7156 case AArch64::SABAv4i16:
7157 return AArch64::SABDv4i16;
7158 case AArch64::SABAv4i32:
7159 return AArch64::SABDv4i32;
7160 case AArch64::SABAv8i16:
7161 return AArch64::SABDv8i16;
7162 case AArch64::SABAv8i8:
7163 return AArch64::SABDv8i8;
7179 auto Match = [&](
int Opcode,
int Operand,
unsigned Pattern) ->
bool {
7191 assert(
false &&
"Unsupported FP instruction in combiner\n");
7193 case AArch64::FADDHrr:
7195 "FADDHrr does not have register operands");
7197 Found = Match(AArch64::FMULHrr, 1, MCP::FMULADDH_OP1);
7198 Found |= Match(AArch64::FMULHrr, 2, MCP::FMULADDH_OP2);
7200 case AArch64::FADDSrr:
7202 "FADDSrr does not have register operands");
7204 Found |= Match(AArch64::FMULSrr, 1, MCP::FMULADDS_OP1) ||
7205 Match(AArch64::FMULv1i32_indexed, 1, MCP::FMLAv1i32_indexed_OP1);
7207 Found |= Match(AArch64::FMULSrr, 2, MCP::FMULADDS_OP2) ||
7208 Match(AArch64::FMULv1i32_indexed, 2, MCP::FMLAv1i32_indexed_OP2);
7210 case AArch64::FADDDrr:
7211 Found |= Match(AArch64::FMULDrr, 1, MCP::FMULADDD_OP1) ||
7212 Match(AArch64::FMULv1i64_indexed, 1, MCP::FMLAv1i64_indexed_OP1);
7214 Found |= Match(AArch64::FMULDrr, 2, MCP::FMULADDD_OP2) ||
7215 Match(AArch64::FMULv1i64_indexed, 2, MCP::FMLAv1i64_indexed_OP2);
7217 case AArch64::FADDv4f16:
7218 Found |= Match(AArch64::FMULv4i16_indexed, 1, MCP::FMLAv4i16_indexed_OP1) ||
7219 Match(AArch64::FMULv4f16, 1, MCP::FMLAv4f16_OP1);
7221 Found |= Match(AArch64::FMULv4i16_indexed, 2, MCP::FMLAv4i16_indexed_OP2) ||
7222 Match(AArch64::FMULv4f16, 2, MCP::FMLAv4f16_OP2);
7224 case AArch64::FADDv8f16:
7225 Found |= Match(AArch64::FMULv8i16_indexed, 1, MCP::FMLAv8i16_indexed_OP1) ||
7226 Match(AArch64::FMULv8f16, 1, MCP::FMLAv8f16_OP1);
7228 Found |= Match(AArch64::FMULv8i16_indexed, 2, MCP::FMLAv8i16_indexed_OP2) ||
7229 Match(AArch64::FMULv8f16, 2, MCP::FMLAv8f16_OP2);
7231 case AArch64::FADDv2f32:
7232 Found |= Match(AArch64::FMULv2i32_indexed, 1, MCP::FMLAv2i32_indexed_OP1) ||
7233 Match(AArch64::FMULv2f32, 1, MCP::FMLAv2f32_OP1);
7235 Found |= Match(AArch64::FMULv2i32_indexed, 2, MCP::FMLAv2i32_indexed_OP2) ||
7236 Match(AArch64::FMULv2f32, 2, MCP::FMLAv2f32_OP2);
7238 case AArch64::FADDv2f64:
7239 Found |= Match(AArch64::FMULv2i64_indexed, 1, MCP::FMLAv2i64_indexed_OP1) ||
7240 Match(AArch64::FMULv2f64, 1, MCP::FMLAv2f64_OP1);
7242 Found |= Match(AArch64::FMULv2i64_indexed, 2, MCP::FMLAv2i64_indexed_OP2) ||
7243 Match(AArch64::FMULv2f64, 2, MCP::FMLAv2f64_OP2);
7245 case AArch64::FADDv4f32:
7246 Found |= Match(AArch64::FMULv4i32_indexed, 1, MCP::FMLAv4i32_indexed_OP1) ||
7247 Match(AArch64::FMULv4f32, 1, MCP::FMLAv4f32_OP1);
7249 Found |= Match(AArch64::FMULv4i32_indexed, 2, MCP::FMLAv4i32_indexed_OP2) ||
7250 Match(AArch64::FMULv4f32, 2, MCP::FMLAv4f32_OP2);
7252 case AArch64::FSUBHrr:
7253 Found = Match(AArch64::FMULHrr, 1, MCP::FMULSUBH_OP1);
7254 Found |= Match(AArch64::FMULHrr, 2, MCP::FMULSUBH_OP2);
7255 Found |= Match(AArch64::FNMULHrr, 1, MCP::FNMULSUBH_OP1);
7257 case AArch64::FSUBSrr:
7258 Found = Match(AArch64::FMULSrr, 1, MCP::FMULSUBS_OP1);
7260 Found |= Match(AArch64::FMULSrr, 2, MCP::FMULSUBS_OP2) ||
7261 Match(AArch64::FMULv1i32_indexed, 2, MCP::FMLSv1i32_indexed_OP2);
7263 Found |= Match(AArch64::FNMULSrr, 1, MCP::FNMULSUBS_OP1);
7265 case AArch64::FSUBDrr:
7266 Found = Match(AArch64::FMULDrr, 1, MCP::FMULSUBD_OP1);
7268 Found |= Match(AArch64::FMULDrr, 2, MCP::FMULSUBD_OP2) ||
7269 Match(AArch64::FMULv1i64_indexed, 2, MCP::FMLSv1i64_indexed_OP2);
7271 Found |= Match(AArch64::FNMULDrr, 1, MCP::FNMULSUBD_OP1);
7273 case AArch64::FSUBv4f16:
7274 Found |= Match(AArch64::FMULv4i16_indexed, 2, MCP::FMLSv4i16_indexed_OP2) ||
7275 Match(AArch64::FMULv4f16, 2, MCP::FMLSv4f16_OP2);
7277 Found |= Match(AArch64::FMULv4i16_indexed, 1, MCP::FMLSv4i16_indexed_OP1) ||
7278 Match(AArch64::FMULv4f16, 1, MCP::FMLSv4f16_OP1);
7280 case AArch64::FSUBv8f16:
7281 Found |= Match(AArch64::FMULv8i16_indexed, 2, MCP::FMLSv8i16_indexed_OP2) ||
7282 Match(AArch64::FMULv8f16, 2, MCP::FMLSv8f16_OP2);
7284 Found |= Match(AArch64::FMULv8i16_indexed, 1, MCP::FMLSv8i16_indexed_OP1) ||
7285 Match(AArch64::FMULv8f16, 1, MCP::FMLSv8f16_OP1);
7287 case AArch64::FSUBv2f32:
7288 Found |= Match(AArch64::FMULv2i32_indexed, 2, MCP::FMLSv2i32_indexed_OP2) ||
7289 Match(AArch64::FMULv2f32, 2, MCP::FMLSv2f32_OP2);
7291 Found |= Match(AArch64::FMULv2i32_indexed, 1, MCP::FMLSv2i32_indexed_OP1) ||
7292 Match(AArch64::FMULv2f32, 1, MCP::FMLSv2f32_OP1);
7294 case AArch64::FSUBv2f64:
7295 Found |= Match(AArch64::FMULv2i64_indexed, 2, MCP::FMLSv2i64_indexed_OP2) ||
7296 Match(AArch64::FMULv2f64, 2, MCP::FMLSv2f64_OP2);
7298 Found |= Match(AArch64::FMULv2i64_indexed, 1, MCP::FMLSv2i64_indexed_OP1) ||
7299 Match(AArch64::FMULv2f64, 1, MCP::FMLSv2f64_OP1);
7301 case AArch64::FSUBv4f32:
7302 Found |= Match(AArch64::FMULv4i32_indexed, 2, MCP::FMLSv4i32_indexed_OP2) ||
7303 Match(AArch64::FMULv4f32, 2, MCP::FMLSv4f32_OP2);
7305 Found |= Match(AArch64::FMULv4i32_indexed, 1, MCP::FMLSv4i32_indexed_OP1) ||
7306 Match(AArch64::FMULv4f32, 1, MCP::FMLSv4f32_OP1);
7317 auto Match = [&](
unsigned Opcode,
int Operand,
unsigned Pattern) ->
bool {
7324 if (
MI &&
MI->getOpcode() == TargetOpcode::COPY &&
7325 MI->getOperand(1).getReg().isVirtual())
7326 MI =
MRI.getUniqueVRegDef(
MI->getOperand(1).getReg());
7327 if (
MI &&
MI->getOpcode() == Opcode) {
7339 case AArch64::FMULv2f32:
7340 Found = Match(AArch64::DUPv2i32lane, 1, MCP::FMULv2i32_indexed_OP1);
7341 Found |= Match(AArch64::DUPv2i32lane, 2, MCP::FMULv2i32_indexed_OP2);
7343 case AArch64::FMULv2f64:
7344 Found = Match(AArch64::DUPv2i64lane, 1, MCP::FMULv2i64_indexed_OP1);
7345 Found |= Match(AArch64::DUPv2i64lane, 2, MCP::FMULv2i64_indexed_OP2);
7347 case AArch64::FMULv4f16:
7348 Found = Match(AArch64::DUPv4i16lane, 1, MCP::FMULv4i16_indexed_OP1);
7349 Found |= Match(AArch64::DUPv4i16lane, 2, MCP::FMULv4i16_indexed_OP2);
7351 case AArch64::FMULv4f32:
7352 Found = Match(AArch64::DUPv4i32lane, 1, MCP::FMULv4i32_indexed_OP1);
7353 Found |= Match(AArch64::DUPv4i32lane, 2, MCP::FMULv4i32_indexed_OP2);
7355 case AArch64::FMULv8f16:
7356 Found = Match(AArch64::DUPv8i16lane, 1, MCP::FMULv8i16_indexed_OP1);
7357 Found |= Match(AArch64::DUPv8i16lane, 2, MCP::FMULv8i16_indexed_OP2);
7370 auto Match = [&](
unsigned Opcode,
unsigned Pattern) ->
bool {
7373 if (
MI !=
nullptr && (
MI->getOpcode() == Opcode) &&
7374 MRI.hasOneNonDBGUse(
MI->getOperand(0).getReg()) &&
7388 case AArch64::FNEGDr:
7390 case AArch64::FNEGSr:
7522 case AArch64::SUBWrr:
7523 case AArch64::SUBSWrr:
7524 case AArch64::SUBXrr:
7525 case AArch64::SUBSXrr:
7570 unsigned LoadLaneOpCode,
unsigned NumLanes) {
7593 while (!RemainingLanes.
empty() && CurrInstr &&
7594 CurrInstr->getOpcode() == LoadLaneOpCode &&
7595 MRI.hasOneNonDBGUse(CurrInstr->getOperand(0).getReg()) &&
7596 CurrInstr->getNumOperands() == 4) {
7597 RemainingLanes.
erase(CurrInstr->getOperand(2).getImm());
7599 CurrInstr =
MRI.getUniqueVRegDef(CurrInstr->getOperand(1).getReg());
7603 if (!RemainingLanes.
empty())
7607 if (CurrInstr->getOpcode() != TargetOpcode::SUBREG_TO_REG)
7611 auto Lane0LoadReg = CurrInstr->getOperand(2).getReg();
7612 unsigned SingleLaneSizeInBits = 128 / NumLanes;
7613 if (
TRI->getRegSizeInBits(Lane0LoadReg,
MRI) != SingleLaneSizeInBits)
7617 if (!
MRI.hasOneNonDBGUse(Lane0LoadReg))
7620 LoadInstrs.
push_back(
MRI.getUniqueVRegDef(Lane0LoadReg));
7629 RemainingLoadInstrs.
insert(LoadInstrs.
begin(), LoadInstrs.
end());
7632 for (; MBBItr !=
MBB->begin() && RemainingSteps > 0 &&
7633 !RemainingLoadInstrs.
empty();
7634 --MBBItr, --RemainingSteps) {
7638 RemainingLoadInstrs.
erase(&CurrInstr);
7648 if (RemainingSteps == 0 && !RemainingLoadInstrs.
empty())
7674 case AArch64::LD1i32:
7676 case AArch64::LD1i16:
7678 case AArch64::LD1i8:
7694 unsigned Pattern,
unsigned NumLanes) {
7702 for (
unsigned i = 0; i < NumLanes - 1; ++i) {
7710 return A->getOperand(2).getImm() >
B->getOperand(2).getImm();
7715 MRI.getUniqueVRegDef(SubregToReg->getOperand(2).getReg()));
7716 auto LoadToLaneInstrsAscending =
llvm::reverse(LoadToLaneInstrs);
7722 auto CreateLD1Instruction = [&](
MachineInstr *OriginalInstr,
7723 Register SrcRegister,
unsigned Lane,
7725 bool OffsetRegisterKillState) {
7726 auto NewRegister =
MRI.createVirtualRegister(FPR128RegClass);
7733 InstrIdxForVirtReg.
insert(std::make_pair(NewRegister, InsInstrs.
size()));
7734 InsInstrs.
push_back(LoadIndexIntoRegister);
7740 auto CreateLDRInstruction = [&](
unsigned NumLanes,
Register DestReg,
7746 Opcode = AArch64::LDRSui;
7749 Opcode = AArch64::LDRHui;
7752 Opcode = AArch64::LDRBui;
7756 "Got unsupported number of lanes in machine-combiner gather pattern");
7765 auto LanesToLoadToReg0 =
7767 LoadToLaneInstrsAscending.begin() + NumLanes / 2);
7768 Register PrevReg = SubregToReg->getOperand(0).getReg();
7770 const MachineOperand &OffsetRegOperand = LoadInstr->getOperand(3);
7771 PrevReg = CreateLD1Instruction(LoadInstr, PrevReg, Index + 1,
7772 OffsetRegOperand.
getReg(),
7773 OffsetRegOperand.
isKill());
7780 MachineInstr *Lane0Load = *LoadToLaneInstrsAscending.begin();
7782 *std::next(LoadToLaneInstrsAscending.begin(), NumLanes / 2);
7783 Register DestRegForMiddleIndex =
MRI.createVirtualRegister(
7789 CreateLDRInstruction(NumLanes, DestRegForMiddleIndex,
7790 OriginalSplitToLoadOffsetOperand.
getReg(),
7791 OriginalSplitToLoadOffsetOperand.
isKill());
7793 InstrIdxForVirtReg.
insert(
7794 std::make_pair(DestRegForMiddleIndex, InsInstrs.
size()));
7795 InsInstrs.
push_back(MiddleIndexLoadInstr);
7799 Register DestRegForSubregToReg =
MRI.createVirtualRegister(FPR128RegClass);
7800 unsigned SubregType;
7803 SubregType = AArch64::ssub;
7806 SubregType = AArch64::hsub;
7809 SubregType = AArch64::bsub;
7813 "Got invalid NumLanes for machine-combiner gather pattern");
7816 auto SubRegToRegInstr =
7818 DestRegForSubregToReg)
7822 InstrIdxForVirtReg.
insert(
7823 std::make_pair(DestRegForSubregToReg, InsInstrs.
size()));
7827 auto LanesToLoadToReg1 =
7829 LoadToLaneInstrsAscending.end());
7830 PrevReg = SubRegToRegInstr->getOperand(0).getReg();
7832 const MachineOperand &OffsetRegOperand = LoadInstr->getOperand(3);
7833 PrevReg = CreateLD1Instruction(LoadInstr, PrevReg, Index + 1,
7834 OffsetRegOperand.
getReg(),
7835 OffsetRegOperand.
isKill());
7838 if (Index == NumLanes / 2 - 2) {
7873bool AArch64InstrInfo::getMachineCombinerPatterns(
7875 bool DoRegPressureReduce)
const {
7896 DoRegPressureReduce);
7925 const Register *ReplacedAddend =
nullptr) {
7926 assert(IdxMulOpd == 1 || IdxMulOpd == 2);
7928 unsigned IdxOtherOpd = IdxMulOpd == 1 ? 2 : 1;
7931 Register SrcReg0 = MUL->getOperand(1).getReg();
7932 bool Src0IsKill = MUL->getOperand(1).isKill();
7933 Register SrcReg1 = MUL->getOperand(2).getReg();
7934 bool Src1IsKill = MUL->getOperand(2).isKill();
7938 if (ReplacedAddend) {
7940 SrcReg2 = *ReplacedAddend;
7948 MRI.constrainRegClass(ResultReg, RC);
7950 MRI.constrainRegClass(SrcReg0, RC);
7952 MRI.constrainRegClass(SrcReg1, RC);
7954 MRI.constrainRegClass(SrcReg2, RC);
7967 .
addImm(MUL->getOperand(3).getImm());
7974 assert(
false &&
"Invalid FMA instruction kind \n");
7988 if (AArch64::FPR32RegClass.hasSubClassEq(RC))
7989 Opc = AArch64::FNMADDSrrr;
7990 else if (AArch64::FPR64RegClass.hasSubClassEq(RC))
7991 Opc = AArch64::FNMADDDrrr;
8003 MRI.constrainRegClass(ResultReg, RC);
8005 MRI.constrainRegClass(SrcReg0, RC);
8007 MRI.constrainRegClass(SrcReg1, RC);
8009 MRI.constrainRegClass(SrcReg2, RC);
8025 unsigned IdxDupOp,
unsigned MulOpc,
8027 assert(((IdxDupOp == 1) || (IdxDupOp == 2)) &&
8028 "Invalid index of FMUL operand");
8036 if (Dup->
getOpcode() == TargetOpcode::COPY)
8040 MRI.clearKillFlags(DupSrcReg);
8041 MRI.constrainRegClass(DupSrcReg, RC);
8045 unsigned IdxMulOp = IdxDupOp == 1 ? 2 : 1;
8086 InstrIdxForVirtReg.
insert(std::make_pair(NewVR, 0));
8101 genNeg(MF,
MRI,
TII, Root, InsInstrs, InstrIdxForVirtReg, MnegOpc, RC);
8128 genNeg(MF,
MRI,
TII, Root, InsInstrs, InstrIdxForVirtReg, MnegOpc, RC);
8156 unsigned IdxMulOpd,
unsigned MaddOpc,
unsigned VR,
8158 assert(IdxMulOpd == 1 || IdxMulOpd == 2);
8162 Register SrcReg0 = MUL->getOperand(1).getReg();
8163 bool Src0IsKill = MUL->getOperand(1).isKill();
8164 Register SrcReg1 = MUL->getOperand(2).getReg();
8165 bool Src1IsKill = MUL->getOperand(2).isKill();
8168 MRI.constrainRegClass(ResultReg, RC);
8170 MRI.constrainRegClass(SrcReg0, RC);
8172 MRI.constrainRegClass(SrcReg1, RC);
8174 MRI.constrainRegClass(VR, RC);
8195 assert(IdxOpd1 == 1 || IdxOpd1 == 2);
8196 unsigned IdxOtherOpd = IdxOpd1 == 1 ? 2 : 1;
8210 if (Opcode == AArch64::SUBSWrr)
8211 Opcode = AArch64::SUBWrr;
8212 else if (Opcode == AArch64::SUBSXrr)
8213 Opcode = AArch64::SUBXrr;
8215 assert((Opcode == AArch64::SUBWrr || Opcode == AArch64::SUBXrr) &&
8216 "Unexpected instruction opcode.");
8233 InstrIdxForVirtReg.
insert(std::make_pair(NewVR, 0));
8240unsigned AArch64InstrInfo::getReduceOpcodeForAccumulator(
8241 unsigned int AccumulatorOpCode)
const {
8242 switch (AccumulatorOpCode) {
8243 case AArch64::UABALB_ZZZ_D:
8244 case AArch64::SABALB_ZZZ_D:
8245 case AArch64::UABALT_ZZZ_D:
8246 case AArch64::SABALT_ZZZ_D:
8247 return AArch64::ADD_ZZZ_D;
8248 case AArch64::UABALB_ZZZ_H:
8249 case AArch64::SABALB_ZZZ_H:
8250 case AArch64::UABALT_ZZZ_H:
8251 case AArch64::SABALT_ZZZ_H:
8252 return AArch64::ADD_ZZZ_H;
8253 case AArch64::UABALB_ZZZ_S:
8254 case AArch64::SABALB_ZZZ_S:
8255 case AArch64::UABALT_ZZZ_S:
8256 case AArch64::SABALT_ZZZ_S:
8257 return AArch64::ADD_ZZZ_S;
8258 case AArch64::UABALv16i8_v8i16:
8259 case AArch64::SABALv8i8_v8i16:
8260 case AArch64::SABAv8i16:
8261 case AArch64::UABAv8i16:
8262 return AArch64::ADDv8i16;
8263 case AArch64::SABALv2i32_v2i64:
8264 case AArch64::UABALv2i32_v2i64:
8265 case AArch64::SABALv4i32_v2i64:
8266 return AArch64::ADDv2i64;
8267 case AArch64::UABALv4i16_v4i32:
8268 case AArch64::SABALv4i16_v4i32:
8269 case AArch64::SABALv8i16_v4i32:
8270 case AArch64::SABAv4i32:
8271 case AArch64::UABAv4i32:
8272 return AArch64::ADDv4i32;
8273 case AArch64::UABALv4i32_v2i64:
8274 return AArch64::ADDv2i64;
8275 case AArch64::UABALv8i16_v4i32:
8276 return AArch64::ADDv4i32;
8277 case AArch64::UABALv8i8_v8i16:
8278 case AArch64::SABALv16i8_v8i16:
8279 return AArch64::ADDv8i16;
8280 case AArch64::UABAv16i8:
8281 case AArch64::SABAv16i8:
8282 return AArch64::ADDv16i8;
8283 case AArch64::UABAv4i16:
8284 case AArch64::SABAv4i16:
8285 return AArch64::ADDv4i16;
8286 case AArch64::UABAv2i32:
8287 case AArch64::SABAv2i32:
8288 return AArch64::ADDv2i32;
8289 case AArch64::UABAv8i8:
8290 case AArch64::SABAv8i8:
8291 return AArch64::ADDv8i8;
8300void AArch64InstrInfo::genAlternativeCodeSequence(
8310 MachineInstr *
MUL =
nullptr;
8311 const TargetRegisterClass *RC;
8317 DelInstrs, InstrIdxForVirtReg);
8323 InstrIdxForVirtReg);
8329 InstrIdxForVirtReg);
8338 Opc = AArch64::MADDWrrr;
8339 RC = &AArch64::GPR32RegClass;
8341 Opc = AArch64::MADDXrrr;
8342 RC = &AArch64::GPR64RegClass;
8353 Opc = AArch64::MADDWrrr;
8354 RC = &AArch64::GPR32RegClass;
8356 Opc = AArch64::MADDXrrr;
8357 RC = &AArch64::GPR64RegClass;
8370 const TargetRegisterClass *RC;
8371 unsigned BitSize, MovImm;
8374 MovImm = AArch64::MOVi32imm;
8375 RC = &AArch64::GPR32spRegClass;
8377 Opc = AArch64::MADDWrrr;
8378 RC = &AArch64::GPR32RegClass;
8380 MovImm = AArch64::MOVi64imm;
8381 RC = &AArch64::GPR64spRegClass;
8383 Opc = AArch64::MADDXrrr;
8384 RC = &AArch64::GPR64RegClass;
8395 uint64_t UImm =
SignExtend64(IsSub ? -Imm : Imm, BitSize);
8399 if (Insn.
size() != 1)
8401 MachineInstrBuilder MIB1 =
8402 BuildMI(MF, MIMetadata(Root),
TII->get(MovImm), NewVR)
8403 .
addImm(IsSub ? -Imm : Imm);
8405 InstrIdxForVirtReg.
insert(std::make_pair(NewVR, 0));
8416 const TargetRegisterClass *SubRC;
8417 unsigned SubOpc, ZeroReg;
8419 SubOpc = AArch64::SUBWrr;
8420 SubRC = &AArch64::GPR32spRegClass;
8421 ZeroReg = AArch64::WZR;
8422 Opc = AArch64::MADDWrrr;
8423 RC = &AArch64::GPR32RegClass;
8425 SubOpc = AArch64::SUBXrr;
8426 SubRC = &AArch64::GPR64spRegClass;
8427 ZeroReg = AArch64::XZR;
8428 Opc = AArch64::MADDXrrr;
8429 RC = &AArch64::GPR64RegClass;
8431 Register NewVR =
MRI.createVirtualRegister(SubRC);
8433 MachineInstrBuilder MIB1 =
8434 BuildMI(MF, MIMetadata(Root),
TII->get(SubOpc), NewVR)
8438 InstrIdxForVirtReg.
insert(std::make_pair(NewVR, 0));
8449 Opc = AArch64::MSUBWrrr;
8450 RC = &AArch64::GPR32RegClass;
8452 Opc = AArch64::MSUBXrrr;
8453 RC = &AArch64::GPR64RegClass;
8458 Opc = AArch64::MLAv8i8;
8459 RC = &AArch64::FPR64RegClass;
8463 Opc = AArch64::MLAv8i8;
8464 RC = &AArch64::FPR64RegClass;
8468 Opc = AArch64::MLAv16i8;
8469 RC = &AArch64::FPR128RegClass;
8473 Opc = AArch64::MLAv16i8;
8474 RC = &AArch64::FPR128RegClass;
8478 Opc = AArch64::MLAv4i16;
8479 RC = &AArch64::FPR64RegClass;
8483 Opc = AArch64::MLAv4i16;
8484 RC = &AArch64::FPR64RegClass;
8488 Opc = AArch64::MLAv8i16;
8489 RC = &AArch64::FPR128RegClass;
8493 Opc = AArch64::MLAv8i16;
8494 RC = &AArch64::FPR128RegClass;
8498 Opc = AArch64::MLAv2i32;
8499 RC = &AArch64::FPR64RegClass;
8503 Opc = AArch64::MLAv2i32;
8504 RC = &AArch64::FPR64RegClass;
8508 Opc = AArch64::MLAv4i32;
8509 RC = &AArch64::FPR128RegClass;
8513 Opc = AArch64::MLAv4i32;
8514 RC = &AArch64::FPR128RegClass;
8519 Opc = AArch64::MLAv8i8;
8520 RC = &AArch64::FPR64RegClass;
8522 InstrIdxForVirtReg, 1,
Opc, AArch64::NEGv8i8,
8526 Opc = AArch64::MLSv8i8;
8527 RC = &AArch64::FPR64RegClass;
8531 Opc = AArch64::MLAv16i8;
8532 RC = &AArch64::FPR128RegClass;
8534 InstrIdxForVirtReg, 1,
Opc, AArch64::NEGv16i8,
8538 Opc = AArch64::MLSv16i8;
8539 RC = &AArch64::FPR128RegClass;
8543 Opc = AArch64::MLAv4i16;
8544 RC = &AArch64::FPR64RegClass;
8546 InstrIdxForVirtReg, 1,
Opc, AArch64::NEGv4i16,
8550 Opc = AArch64::MLSv4i16;
8551 RC = &AArch64::FPR64RegClass;
8555 Opc = AArch64::MLAv8i16;
8556 RC = &AArch64::FPR128RegClass;
8558 InstrIdxForVirtReg, 1,
Opc, AArch64::NEGv8i16,
8562 Opc = AArch64::MLSv8i16;
8563 RC = &AArch64::FPR128RegClass;
8567 Opc = AArch64::MLAv2i32;
8568 RC = &AArch64::FPR64RegClass;
8570 InstrIdxForVirtReg, 1,
Opc, AArch64::NEGv2i32,
8574 Opc = AArch64::MLSv2i32;
8575 RC = &AArch64::FPR64RegClass;
8579 Opc = AArch64::MLAv4i32;
8580 RC = &AArch64::FPR128RegClass;
8582 InstrIdxForVirtReg, 1,
Opc, AArch64::NEGv4i32,
8586 Opc = AArch64::MLSv4i32;
8587 RC = &AArch64::FPR128RegClass;
8592 Opc = AArch64::MLAv4i16_indexed;
8593 RC = &AArch64::FPR64RegClass;
8597 Opc = AArch64::MLAv4i16_indexed;
8598 RC = &AArch64::FPR64RegClass;
8602 Opc = AArch64::MLAv8i16_indexed;
8603 RC = &AArch64::FPR128RegClass;
8607 Opc = AArch64::MLAv8i16_indexed;
8608 RC = &AArch64::FPR128RegClass;
8612 Opc = AArch64::MLAv2i32_indexed;
8613 RC = &AArch64::FPR64RegClass;
8617 Opc = AArch64::MLAv2i32_indexed;
8618 RC = &AArch64::FPR64RegClass;
8622 Opc = AArch64::MLAv4i32_indexed;
8623 RC = &AArch64::FPR128RegClass;
8627 Opc = AArch64::MLAv4i32_indexed;
8628 RC = &AArch64::FPR128RegClass;
8633 Opc = AArch64::MLAv4i16_indexed;
8634 RC = &AArch64::FPR64RegClass;
8636 InstrIdxForVirtReg, 1,
Opc, AArch64::NEGv4i16,
8640 Opc = AArch64::MLSv4i16_indexed;
8641 RC = &AArch64::FPR64RegClass;
8645 Opc = AArch64::MLAv8i16_indexed;
8646 RC = &AArch64::FPR128RegClass;
8648 InstrIdxForVirtReg, 1,
Opc, AArch64::NEGv8i16,
8652 Opc = AArch64::MLSv8i16_indexed;
8653 RC = &AArch64::FPR128RegClass;
8657 Opc = AArch64::MLAv2i32_indexed;
8658 RC = &AArch64::FPR64RegClass;
8660 InstrIdxForVirtReg, 1,
Opc, AArch64::NEGv2i32,
8664 Opc = AArch64::MLSv2i32_indexed;
8665 RC = &AArch64::FPR64RegClass;
8669 Opc = AArch64::MLAv4i32_indexed;
8670 RC = &AArch64::FPR128RegClass;
8672 InstrIdxForVirtReg, 1,
Opc, AArch64::NEGv4i32,
8676 Opc = AArch64::MLSv4i32_indexed;
8677 RC = &AArch64::FPR128RegClass;
8683 Opc = AArch64::FMADDHrrr;
8684 RC = &AArch64::FPR16RegClass;
8688 Opc = AArch64::FMADDSrrr;
8689 RC = &AArch64::FPR32RegClass;
8693 Opc = AArch64::FMADDDrrr;
8694 RC = &AArch64::FPR64RegClass;
8699 Opc = AArch64::FMADDHrrr;
8700 RC = &AArch64::FPR16RegClass;
8704 Opc = AArch64::FMADDSrrr;
8705 RC = &AArch64::FPR32RegClass;
8709 Opc = AArch64::FMADDDrrr;
8710 RC = &AArch64::FPR64RegClass;
8715 Opc = AArch64::FMLAv1i32_indexed;
8716 RC = &AArch64::FPR32RegClass;
8721 Opc = AArch64::FMLAv1i32_indexed;
8722 RC = &AArch64::FPR32RegClass;
8728 Opc = AArch64::FMLAv1i64_indexed;
8729 RC = &AArch64::FPR64RegClass;
8734 Opc = AArch64::FMLAv1i64_indexed;
8735 RC = &AArch64::FPR64RegClass;
8741 RC = &AArch64::FPR64RegClass;
8742 Opc = AArch64::FMLAv4i16_indexed;
8747 RC = &AArch64::FPR64RegClass;
8748 Opc = AArch64::FMLAv4f16;
8753 RC = &AArch64::FPR64RegClass;
8754 Opc = AArch64::FMLAv4i16_indexed;
8759 RC = &AArch64::FPR64RegClass;
8760 Opc = AArch64::FMLAv4f16;
8767 RC = &AArch64::FPR64RegClass;
8769 Opc = AArch64::FMLAv2i32_indexed;
8773 Opc = AArch64::FMLAv2f32;
8780 RC = &AArch64::FPR64RegClass;
8782 Opc = AArch64::FMLAv2i32_indexed;
8786 Opc = AArch64::FMLAv2f32;
8793 RC = &AArch64::FPR128RegClass;
8794 Opc = AArch64::FMLAv8i16_indexed;
8799 RC = &AArch64::FPR128RegClass;
8800 Opc = AArch64::FMLAv8f16;
8805 RC = &AArch64::FPR128RegClass;
8806 Opc = AArch64::FMLAv8i16_indexed;
8811 RC = &AArch64::FPR128RegClass;
8812 Opc = AArch64::FMLAv8f16;
8819 RC = &AArch64::FPR128RegClass;
8821 Opc = AArch64::FMLAv2i64_indexed;
8825 Opc = AArch64::FMLAv2f64;
8832 RC = &AArch64::FPR128RegClass;
8834 Opc = AArch64::FMLAv2i64_indexed;
8838 Opc = AArch64::FMLAv2f64;
8846 RC = &AArch64::FPR128RegClass;
8848 Opc = AArch64::FMLAv4i32_indexed;
8852 Opc = AArch64::FMLAv4f32;
8860 RC = &AArch64::FPR128RegClass;
8862 Opc = AArch64::FMLAv4i32_indexed;
8866 Opc = AArch64::FMLAv4f32;
8873 Opc = AArch64::FNMSUBHrrr;
8874 RC = &AArch64::FPR16RegClass;
8878 Opc = AArch64::FNMSUBSrrr;
8879 RC = &AArch64::FPR32RegClass;
8883 Opc = AArch64::FNMSUBDrrr;
8884 RC = &AArch64::FPR64RegClass;
8889 Opc = AArch64::FNMADDHrrr;
8890 RC = &AArch64::FPR16RegClass;
8894 Opc = AArch64::FNMADDSrrr;
8895 RC = &AArch64::FPR32RegClass;
8899 Opc = AArch64::FNMADDDrrr;
8900 RC = &AArch64::FPR64RegClass;
8905 Opc = AArch64::FMSUBHrrr;
8906 RC = &AArch64::FPR16RegClass;
8910 Opc = AArch64::FMSUBSrrr;
8911 RC = &AArch64::FPR32RegClass;
8915 Opc = AArch64::FMSUBDrrr;
8916 RC = &AArch64::FPR64RegClass;
8921 Opc = AArch64::FMLSv1i32_indexed;
8922 RC = &AArch64::FPR32RegClass;
8928 Opc = AArch64::FMLSv1i64_indexed;
8929 RC = &AArch64::FPR64RegClass;
8936 RC = &AArch64::FPR64RegClass;
8938 MachineInstrBuilder MIB1 =
8939 BuildMI(MF, MIMetadata(Root),
TII->get(AArch64::FNEGv4f16), NewVR)
8942 InstrIdxForVirtReg.
insert(std::make_pair(NewVR, 0));
8944 Opc = AArch64::FMLAv4f16;
8948 Opc = AArch64::FMLAv4i16_indexed;
8955 RC = &AArch64::FPR64RegClass;
8956 Opc = AArch64::FMLSv4f16;
8961 RC = &AArch64::FPR64RegClass;
8962 Opc = AArch64::FMLSv4i16_indexed;
8969 RC = &AArch64::FPR64RegClass;
8971 Opc = AArch64::FMLSv2i32_indexed;
8975 Opc = AArch64::FMLSv2f32;
8983 RC = &AArch64::FPR128RegClass;
8985 MachineInstrBuilder MIB1 =
8986 BuildMI(MF, MIMetadata(Root),
TII->get(AArch64::FNEGv8f16), NewVR)
8989 InstrIdxForVirtReg.
insert(std::make_pair(NewVR, 0));
8991 Opc = AArch64::FMLAv8f16;
8995 Opc = AArch64::FMLAv8i16_indexed;
9002 RC = &AArch64::FPR128RegClass;
9003 Opc = AArch64::FMLSv8f16;
9008 RC = &AArch64::FPR128RegClass;
9009 Opc = AArch64::FMLSv8i16_indexed;
9016 RC = &AArch64::FPR128RegClass;
9018 Opc = AArch64::FMLSv2i64_indexed;
9022 Opc = AArch64::FMLSv2f64;
9030 RC = &AArch64::FPR128RegClass;
9032 Opc = AArch64::FMLSv4i32_indexed;
9036 Opc = AArch64::FMLSv4f32;
9043 RC = &AArch64::FPR64RegClass;
9045 MachineInstrBuilder MIB1 =
9046 BuildMI(MF, MIMetadata(Root),
TII->get(AArch64::FNEGv2f32), NewVR)
9049 InstrIdxForVirtReg.
insert(std::make_pair(NewVR, 0));
9051 Opc = AArch64::FMLAv2i32_indexed;
9055 Opc = AArch64::FMLAv2f32;
9063 RC = &AArch64::FPR128RegClass;
9065 MachineInstrBuilder MIB1 =
9066 BuildMI(MF, MIMetadata(Root),
TII->get(AArch64::FNEGv4f32), NewVR)
9069 InstrIdxForVirtReg.
insert(std::make_pair(NewVR, 0));
9071 Opc = AArch64::FMLAv4i32_indexed;
9075 Opc = AArch64::FMLAv4f32;
9083 RC = &AArch64::FPR128RegClass;
9085 MachineInstrBuilder MIB1 =
9086 BuildMI(MF, MIMetadata(Root),
TII->get(AArch64::FNEGv2f64), NewVR)
9089 InstrIdxForVirtReg.
insert(std::make_pair(NewVR, 0));
9091 Opc = AArch64::FMLAv2i64_indexed;
9095 Opc = AArch64::FMLAv2f64;
9107 &AArch64::FPR128RegClass,
MRI);
9116 &AArch64::FPR128RegClass,
MRI);
9125 &AArch64::FPR128_loRegClass,
MRI);
9134 &AArch64::FPR128RegClass,
MRI);
9143 &AArch64::FPR128_loRegClass,
MRI);
9177 for (
auto *
MI : InsInstrs)
9178 MI->setFlags(Flags);
9219 bool IsNegativeBranch =
false;
9220 bool IsTestAndBranch =
false;
9221 unsigned TargetBBInMI = 0;
9222 switch (
MI.getOpcode()) {
9226 case AArch64::CBWPri:
9227 case AArch64::CBXPri:
9228 case AArch64::CBWPrr:
9229 case AArch64::CBXPrr:
9235 case AArch64::CBNZW:
9236 case AArch64::CBNZX:
9238 IsNegativeBranch =
true;
9243 IsTestAndBranch =
true;
9245 case AArch64::TBNZW:
9246 case AArch64::TBNZX:
9248 IsNegativeBranch =
true;
9249 IsTestAndBranch =
true;
9255 if (IsTestAndBranch &&
MI.getOperand(1).getImm())
9259 assert(
MI.getParent() &&
"Incomplete machine instruction\n");
9270 while (
DefMI->isCopy()) {
9272 if (!
MRI->hasOneNonDBGUse(CopyVReg))
9274 if (!
MRI->hasOneDef(CopyVReg))
9279 switch (
DefMI->getOpcode()) {
9283 case AArch64::ANDWri:
9284 case AArch64::ANDXri: {
9285 if (IsTestAndBranch)
9289 if (!
MRI->hasOneNonDBGUse(VReg))
9292 bool Is32Bit = (
DefMI->getOpcode() == AArch64::ANDWri);
9294 DefMI->getOperand(2).getImm(), Is32Bit ? 32 : 64);
9303 assert(!
MRI->def_empty(NewReg) &&
"Register must be defined.");
9309 unsigned Opc = (Imm < 32)
9310 ? (IsNegativeBranch ? AArch64::TBNZW : AArch64::TBZW)
9311 : (IsNegativeBranch ? AArch64::TBNZX : AArch64::TBZX);
9324 if (!Is32Bit && Imm < 32)
9326 MI.eraseFromParent();
9330 case AArch64::CSINCWr:
9331 case AArch64::CSINCXr: {
9332 if (!(
DefMI->getOperand(1).getReg() == AArch64::WZR &&
9333 DefMI->getOperand(2).getReg() == AArch64::WZR) &&
9334 !(
DefMI->getOperand(1).getReg() == AArch64::XZR &&
9335 DefMI->getOperand(2).getReg() == AArch64::XZR))
9338 if (
DefMI->findRegisterDefOperandIdx(AArch64::NZCV,
nullptr,
9351 if (IsNegativeBranch)
9354 MI.eraseFromParent();
9360std::pair<unsigned, unsigned>
9361AArch64InstrInfo::decomposeMachineOperandsTargetFlags(
unsigned TF)
const {
9363 return std::make_pair(TF & Mask, TF & ~Mask);
9367AArch64InstrInfo::getSerializableDirectMachineOperandTargetFlags()
const {
9370 static const std::pair<unsigned, const char *> TargetFlags[] = {
9371 {MO_PAGE,
"aarch64-page"}, {
MO_PAGEOFF,
"aarch64-pageoff"},
9372 {
MO_G3,
"aarch64-g3"}, {
MO_G2,
"aarch64-g2"},
9373 {
MO_G1,
"aarch64-g1"}, {
MO_G0,
"aarch64-g0"},
9379AArch64InstrInfo::getSerializableBitmaskMachineOperandTargetFlags()
const {
9380 using namespace AArch64II;
9382 static const std::pair<unsigned, const char *> TargetFlags[] = {
9385 {
MO_NC,
"aarch64-nc"},
9386 {
MO_S,
"aarch64-s"},
9397AArch64InstrInfo::getSerializableMachineMemOperandTargetFlags()
const {
9398 static const std::pair<MachineMemOperand::Flags, const char *> TargetFlags[] =
9500 MachineFunction *MF =
C.getMF();
9502 const AArch64RegisterInfo *ARI =
9503 static_cast<const AArch64RegisterInfo *
>(&
TRI);
9506 for (
unsigned Reg : AArch64::GPR64RegClass) {
9508 Reg != AArch64::LR &&
9509 Reg != AArch64::X16 &&
9510 Reg != AArch64::X17 &&
9511 C.isAvailableAcrossAndOutOfSeq(
Reg,
TRI) &&
9512 C.isAvailableInsideSeq(
Reg,
TRI))
9543 return SubtargetA.hasV8_3aOps() == SubtargetB.hasV8_3aOps();
9546std::optional<std::unique_ptr<outliner::OutlinedFunction>>
9547AArch64InstrInfo::getOutliningCandidateInfo(
9549 std::vector<outliner::Candidate> &RepeatedSequenceLocs,
9550 unsigned MinRepeats)
const {
9551 unsigned SequenceSize = 0;
9552 for (
auto &
MI : RepeatedSequenceLocs[0])
9555 unsigned NumBytesToCreateFrame = 0;
9565 if (std::adjacent_find(
9566 RepeatedSequenceLocs.begin(), RepeatedSequenceLocs.end(),
9567 [](
const outliner::Candidate &a,
const outliner::Candidate &b) {
9570 if (outliningCandidatesSigningScopeConsensus(a, b) &&
9571 outliningCandidatesSigningKeyConsensus(a, b) &&
9572 outliningCandidatesV8_3OpsConsensus(a, b)) {
9576 }) != RepeatedSequenceLocs.end()) {
9577 return std::nullopt;
9594 unsigned NumBytesToCheckLRInTCEpilogue = 0;
9595 if (RepeatedSequenceLocs[0]
9597 ->getInfo<AArch64FunctionInfo>()
9598 ->shouldSignReturnAddress(
true)) {
9600 NumBytesToCreateFrame += 8;
9603 auto LRCheckMethod = Subtarget.getAuthenticatedLRCheckMethod(
9604 *RepeatedSequenceLocs[0].getMF());
9605 NumBytesToCheckLRInTCEpilogue =
9609 if (isTailCallReturnInst(RepeatedSequenceLocs[0].
back()))
9610 SequenceSize += NumBytesToCheckLRInTCEpilogue;
9618 for (
auto &
MI :
C) {
9619 if (
MI.modifiesRegister(AArch64::SP, &
TRI)) {
9620 switch (
MI.getOpcode()) {
9621 case AArch64::ADDXri:
9622 case AArch64::ADDWri:
9623 assert(
MI.getNumOperands() == 4 &&
"Wrong number of operands");
9625 "Expected operand to be immediate");
9627 "Expected operand to be a register");
9631 if (
MI.getOperand(1).getReg() == AArch64::SP)
9632 SPValue +=
MI.getOperand(2).getImm();
9636 case AArch64::SUBXri:
9637 case AArch64::SUBWri:
9638 assert(
MI.getNumOperands() == 4 &&
"Wrong number of operands");
9640 "Expected operand to be immediate");
9642 "Expected operand to be a register");
9646 if (
MI.getOperand(1).getReg() == AArch64::SP)
9647 SPValue -=
MI.getOperand(2).getImm();
9664 if (RepeatedSequenceLocs.size() < MinRepeats)
9665 return std::nullopt;
9669 unsigned FlagsSetInAll = 0xF;
9673 FlagsSetInAll &=
C.Flags;
9675 unsigned LastInstrOpcode = RepeatedSequenceLocs[0].back().getOpcode();
9678 auto SetCandidateCallInfo =
9679 [&RepeatedSequenceLocs](
unsigned CallID,
unsigned NumBytesForCall) {
9681 C.setCallInfo(CallID, NumBytesForCall);
9685 NumBytesToCreateFrame += 4;
9693 unsigned CFICount = 0;
9694 for (
auto &
I : RepeatedSequenceLocs[0]) {
9695 if (
I.isCFIInstruction())
9705 std::vector<MCCFIInstruction> CFIInstructions =
9706 C.getMF()->getFrameInstructions();
9708 if (CFICount > 0 && CFICount != CFIInstructions.size())
9709 return std::nullopt;
9717 if (!
MI.modifiesRegister(AArch64::SP, &
TRI) &&
9718 !
MI.readsRegister(AArch64::SP, &
TRI))
9724 if (
MI.modifiesRegister(AArch64::SP, &
TRI))
9729 if (
MI.mayLoadOrStore()) {
9732 bool OffsetIsScalable;
9736 if (!getMemOperandWithOffset(
MI,
Base,
Offset, OffsetIsScalable, &
TRI) ||
9737 !
Base->isReg() ||
Base->getReg() != AArch64::SP)
9741 if (OffsetIsScalable)
9749 TypeSize Scale(0U,
false), DummyWidth(0U,
false);
9750 getMemOpInfo(
MI.getOpcode(), Scale, DummyWidth, MinOffset, MaxOffset);
9753 if (
Offset < MinOffset * (int64_t)Scale.getFixedValue() ||
9754 Offset > MaxOffset * (int64_t)Scale.getFixedValue())
9769 bool AllStackInstrsSafe =
9774 if (RepeatedSequenceLocs[0].
back().isTerminator()) {
9776 NumBytesToCreateFrame = 0;
9777 unsigned NumBytesForCall = 4 + NumBytesToCheckLRInTCEpilogue;
9781 else if (LastInstrOpcode == AArch64::BL ||
9782 ((LastInstrOpcode == AArch64::BLR ||
9783 LastInstrOpcode == AArch64::BLRNoIP) &&
9787 NumBytesToCreateFrame = NumBytesToCheckLRInTCEpilogue;
9795 unsigned NumBytesNoStackCalls = 0;
9796 std::vector<outliner::Candidate> CandidatesWithoutStackFixups;
9802 ?
C.isAvailableAcrossAndOutOfSeq(AArch64::LR,
TRI)
9811 C.getMF()->getFunction().hasFnAttribute(Attribute::NoReturn);
9814 if (LRAvailable && !IsNoReturn) {
9815 NumBytesNoStackCalls += 4;
9817 CandidatesWithoutStackFixups.push_back(
C);
9822 else if (findRegisterToSaveLRTo(
C)) {
9823 NumBytesNoStackCalls += 12;
9825 CandidatesWithoutStackFixups.push_back(
C);
9830 else if (
C.isAvailableInsideSeq(AArch64::SP,
TRI)) {
9831 NumBytesNoStackCalls += 12;
9833 CandidatesWithoutStackFixups.push_back(
C);
9839 NumBytesNoStackCalls += SequenceSize;
9846 if (!AllStackInstrsSafe ||
9847 NumBytesNoStackCalls <= RepeatedSequenceLocs.size() * 12) {
9848 RepeatedSequenceLocs = CandidatesWithoutStackFixups;
9850 if (RepeatedSequenceLocs.size() < MinRepeats)
9851 return std::nullopt;
9904 (!
C.isAvailableAcrossAndOutOfSeq(AArch64::LR,
TRI) ||
9905 !findRegisterToSaveLRTo(
C));
9911 if (RepeatedSequenceLocs.size() < MinRepeats)
9912 return std::nullopt;
9921 bool ModStackToSaveLR =
false;
9924 ModStackToSaveLR =
true;
9933 ModStackToSaveLR =
true;
9935 if (ModStackToSaveLR) {
9937 if (!AllStackInstrsSafe)
9938 return std::nullopt;
9941 NumBytesToCreateFrame += 8;
9948 return std::nullopt;
9950 return std::make_unique<outliner::OutlinedFunction>(
9951 RepeatedSequenceLocs, SequenceSize, NumBytesToCreateFrame, FrameID);
9954void AArch64InstrInfo::mergeOutliningCandidateAttributes(
9955 Function &
F, std::vector<outliner::Candidate> &Candidates)
const {
9959 const auto &CFn = Candidates.front().getMF()->getFunction();
9961 if (CFn.hasFnAttribute(
"ptrauth-returns"))
9962 F.addFnAttr(CFn.getFnAttribute(
"ptrauth-returns"));
9963 if (CFn.hasFnAttribute(
"ptrauth-auth-traps"))
9964 F.addFnAttr(CFn.getFnAttribute(
"ptrauth-auth-traps"));
9967 if (CFn.hasFnAttribute(
"sign-return-address"))
9968 F.addFnAttr(CFn.getFnAttribute(
"sign-return-address"));
9969 if (CFn.hasFnAttribute(
"sign-return-address-key"))
9970 F.addFnAttr(CFn.getFnAttribute(
"sign-return-address-key"));
9972 AArch64GenInstrInfo::mergeOutliningCandidateAttributes(
F, Candidates);
9975bool AArch64InstrInfo::isFunctionSafeToOutlineFrom(
9980 if (!OutlineFromLinkOnceODRs &&
F.hasLinkOnceODRLinkage())
9993 AArch64FunctionInfo *AFI = MF.
getInfo<AArch64FunctionInfo>();
9994 if (!AFI || AFI->
hasRedZone().value_or(
true))
10014 unsigned &Flags)
const {
10016 "Must track liveness!");
10018 std::pair<MachineBasicBlock::iterator, MachineBasicBlock::iterator>>
10033 auto AreAllUnsafeRegsDead = [&LRU]() {
10034 return LRU.available(AArch64::W16) && LRU.available(AArch64::W17) &&
10035 LRU.available(AArch64::NZCV);
10050 bool LRAvailableEverywhere =
true;
10052 LRU.addLiveOuts(
MBB);
10054 auto UpdateWholeMBBFlags = [&
Flags](
const MachineInstr &
MI) {
10055 if (
MI.isCall() && !
MI.isTerminator())
10061 auto CreateNewRangeStartingAt =
10062 [&RangeBegin, &RangeEnd,
10064 RangeBegin = NewBegin;
10065 RangeEnd = std::next(RangeBegin);
10068 auto SaveRangeIfNonEmpty = [&RangeLen, &
Ranges, &RangeBegin, &RangeEnd]() {
10074 if (!RangeBegin.isEnd() && RangeBegin->isBundledWithPred())
10076 if (!RangeEnd.isEnd() && RangeEnd->isBundledWithPred())
10078 Ranges.emplace_back(RangeBegin, RangeEnd);
10086 for (; FirstPossibleEndPt !=
MBB.
instr_rend(); ++FirstPossibleEndPt) {
10087 LRU.stepBackward(*FirstPossibleEndPt);
10090 UpdateWholeMBBFlags(*FirstPossibleEndPt);
10091 if (AreAllUnsafeRegsDead())
10098 CreateNewRangeStartingAt(FirstPossibleEndPt->getIterator());
10103 LRU.stepBackward(
MI);
10104 UpdateWholeMBBFlags(
MI);
10105 if (!AreAllUnsafeRegsDead()) {
10106 SaveRangeIfNonEmpty();
10107 CreateNewRangeStartingAt(
MI.getIterator());
10110 LRAvailableEverywhere &= LRU.available(AArch64::LR);
10111 RangeBegin =
MI.getIterator();
10116 if (AreAllUnsafeRegsDead())
10117 SaveRangeIfNonEmpty();
10125 if (!LRAvailableEverywhere)
10133 unsigned Flags)
const {
10134 MachineInstr &
MI = *MIT;
10138 switch (
MI.getOpcode()) {
10139 case AArch64::PACM:
10140 case AArch64::PACIASP:
10141 case AArch64::PACIBSP:
10142 case AArch64::PACIASPPC:
10143 case AArch64::PACIBSPPC:
10144 case AArch64::AUTIASP:
10145 case AArch64::AUTIBSP:
10146 case AArch64::AUTIASPPCi:
10147 case AArch64::AUTIASPPCr:
10148 case AArch64::AUTIBSPPCi:
10149 case AArch64::AUTIBSPPCr:
10150 case AArch64::RETAA:
10151 case AArch64::RETAB:
10152 case AArch64::RETAASPPCi:
10153 case AArch64::RETAASPPCr:
10154 case AArch64::RETABSPPCi:
10155 case AArch64::RETABSPPCr:
10156 case AArch64::EMITBKEY:
10157 case AArch64::PAUTH_PROLOGUE:
10158 case AArch64::PAUTH_EPILOGUE:
10168 if (
MI.isCFIInstruction())
10172 if (
MI.isTerminator())
10178 for (
const MachineOperand &MOP :
MI.operands()) {
10181 assert(!MOP.isCFIIndex());
10184 if (MOP.isReg() && !MOP.isImplicit() &&
10185 (MOP.getReg() == AArch64::LR || MOP.getReg() == AArch64::W30))
10192 if (
MI.getOpcode() == AArch64::ADRP)
10212 for (
const MachineOperand &MOP :
MI.operands()) {
10213 if (MOP.isGlobal()) {
10221 if (Callee &&
Callee->getName() ==
"\01_mcount")
10229 if (
MI.getOpcode() == AArch64::BLR ||
10230 MI.getOpcode() == AArch64::BLRNoIP ||
MI.getOpcode() == AArch64::BL)
10234 return UnknownCallOutlineType;
10242 return UnknownCallOutlineType;
10250 return UnknownCallOutlineType;
10271 for (MachineInstr &
MI :
MBB) {
10272 const MachineOperand *
Base;
10273 TypeSize Width(0,
false);
10275 bool OffsetIsScalable;
10278 if (!
MI.mayLoadOrStore() ||
10281 (
Base->isReg() &&
Base->getReg() != AArch64::SP))
10285 TypeSize Scale(0U,
false);
10286 int64_t Dummy1, Dummy2;
10289 assert(StackOffsetOperand.
isImm() &&
"Stack offset wasn't immediate!");
10291 assert(Scale != 0 &&
"Unexpected opcode!");
10292 assert(!OffsetIsScalable &&
"Expected offset to be a byte offset");
10297 int64_t NewImm = (
Offset + 16) / (int64_t)Scale.getFixedValue();
10298 StackOffsetOperand.
setImm(NewImm);
10304 bool ShouldSignReturnAddr) {
10305 if (!ShouldSignReturnAddr)
10311 TII->get(AArch64::PAUTH_EPILOGUE))
10315void AArch64InstrInfo::buildOutlinedFrame(
10319 AArch64FunctionInfo *FI = MF.
getInfo<AArch64FunctionInfo>();
10327 unsigned TailOpcode;
10329 TailOpcode = AArch64::TCRETURNdi;
10333 TailOpcode = AArch64::TCRETURNriALL;
10344 bool IsLeafFunction =
true;
10347 auto IsNonTailCall = [](
const MachineInstr &
MI) {
10348 return MI.isCall() && !
MI.isReturn();
10358 "Can only fix up stack references once");
10359 fixupPostOutline(
MBB);
10361 IsLeafFunction =
false;
10372 Et = std::prev(
MBB.
end());
10382 if (MF.
getInfo<AArch64FunctionInfo>()->needsDwarfUnwindInfo(MF)) {
10386 CFIBuilder.buildDefCFAOffset(16);
10390 CFIBuilder.buildOffset(AArch64::LR, -16);
10431 fixupPostOutline(
MBB);
10442 .addGlobalAddress(
M.getNamedValue(MF.
getName()))
10452 .addGlobalAddress(
M.getNamedValue(MF.
getName())));
10461 MachineInstr *Save;
10462 MachineInstr *Restore;
10468 assert(
Reg &&
"No callee-saved register available?");
10502 .addGlobalAddress(
M.getNamedValue(MF.
getName())));
10510bool AArch64InstrInfo::shouldOutlineFromFunctionByDefault(
10518 bool AllowSideEffects)
const {
10520 const AArch64Subtarget &STI = MF.
getSubtarget<AArch64Subtarget>();
10523 if (
TRI.isGeneralPurposeRegister(MF,
Reg)) {
10536 assert(STI.hasNEON() &&
"Expected to have NEON.");
10542std::optional<DestSourcePair>
10547 if (((
MI.getOpcode() == AArch64::ORRWrs &&
10548 MI.getOperand(1).getReg() == AArch64::WZR &&
10549 MI.getOperand(3).getImm() == 0x0) ||
10550 (
MI.getOpcode() == AArch64::ORRWrr &&
10551 MI.getOperand(1).getReg() == AArch64::WZR)) &&
10553 (!
MI.getOperand(0).getReg().isVirtual() ||
10554 MI.getOperand(0).getSubReg() == 0) &&
10555 (!
MI.getOperand(0).getReg().isPhysical() ||
10560 if (
MI.getOpcode() == AArch64::ORRXrs &&
10561 MI.getOperand(1).getReg() == AArch64::XZR &&
10562 MI.getOperand(3).getImm() == 0x0)
10565 return std::nullopt;
10568std::optional<DestSourcePair>
10570 if ((
MI.getOpcode() == AArch64::ORRWrs &&
10571 MI.getOperand(1).getReg() == AArch64::WZR &&
10572 MI.getOperand(3).getImm() == 0x0) ||
10573 (
MI.getOpcode() == AArch64::ORRWrr &&
10574 MI.getOperand(1).getReg() == AArch64::WZR))
10576 return std::nullopt;
10579std::optional<RegImmPair>
10588 return std::nullopt;
10590 switch (
MI.getOpcode()) {
10592 return std::nullopt;
10593 case AArch64::SUBWri:
10594 case AArch64::SUBXri:
10595 case AArch64::SUBSWri:
10596 case AArch64::SUBSXri:
10599 case AArch64::ADDSWri:
10600 case AArch64::ADDSXri:
10601 case AArch64::ADDWri:
10602 case AArch64::ADDXri: {
10604 if (!
MI.getOperand(0).isReg() || !
MI.getOperand(1).isReg() ||
10605 !
MI.getOperand(2).isImm())
10606 return std::nullopt;
10607 int Shift =
MI.getOperand(3).getImm();
10608 assert((Shift == 0 || Shift == 12) &&
"Shift can be either 0 or 12");
10612 return RegImmPair{
MI.getOperand(1).getReg(),
Offset};
10618static std::optional<ParamLoadedValue>
10622 auto DestSrc =
TII->isCopyLikeInstr(
MI);
10624 return std::nullopt;
10626 Register DestReg = DestSrc->Destination->getReg();
10627 Register SrcReg = DestSrc->Source->getReg();
10632 if (DestReg == DescribedReg)
10636 if (
MI.getOpcode() == AArch64::ORRWrs &&
10637 TRI->isSuperRegister(DestReg, DescribedReg))
10641 if (
MI.getOpcode() == AArch64::ORRXrs &&
10642 TRI->isSubRegister(DestReg, DescribedReg)) {
10643 Register SrcSubReg =
TRI->getSubReg(SrcReg, AArch64::sub_32);
10647 assert(!
TRI->isSuperOrSubRegisterEq(DestReg, DescribedReg) &&
10648 "Unhandled ORR[XW]rs copy case");
10650 return std::nullopt;
10653bool AArch64InstrInfo::isFunctionSafeToSplit(
const MachineFunction &MF)
const {
10658 if (MF.
getInfo<AArch64FunctionInfo>()->hasRedZone().value_or(
true))
10664bool AArch64InstrInfo::isMBBSafeToSplitToCold(
10668 auto isAsmGoto = [](
const MachineInstr &
MI) {
10669 return MI.getOpcode() == AArch64::INLINEASM_BR;
10679 auto containsMBB = [&
MBB](
const MachineJumpTableEntry &JTE) {
10686 for (
const MachineInstr &
MI :
MBB) {
10687 switch (
MI.getOpcode()) {
10688 case TargetOpcode::G_BRJT:
10689 case AArch64::JumpTableDest32:
10690 case AArch64::JumpTableDest16:
10691 case AArch64::JumpTableDest8:
10702std::optional<ParamLoadedValue>
10705 const MachineFunction *MF =
MI.getMF();
10707 switch (
MI.getOpcode()) {
10708 case AArch64::MOVZWi:
10709 case AArch64::MOVZXi: {
10712 if (!
TRI->isSuperRegisterEq(
MI.getOperand(0).getReg(),
Reg))
10713 return std::nullopt;
10715 if (!
MI.getOperand(1).isImm())
10716 return std::nullopt;
10717 int64_t Immediate =
MI.getOperand(1).getImm();
10718 int Shift =
MI.getOperand(2).getImm();
10722 case AArch64::ORRWrs:
10723 case AArch64::ORRXrs:
10730bool AArch64InstrInfo::isExtendLikelyToBeFolded(
10733 ExtMI.
getOpcode() == TargetOpcode::G_ZEXT ||
10734 ExtMI.
getOpcode() == TargetOpcode::G_ANYEXT);
10737 if (ExtMI.
getOpcode() == TargetOpcode::G_ANYEXT)
10741 if (!
MRI.hasOneNonDBGUse(DefReg))
10746 auto *UserMI = &*
MRI.use_instr_nodbg_begin(DefReg);
10747 return UserMI->getOpcode() == TargetOpcode::G_PTR_ADD;
10750uint64_t AArch64InstrInfo::getElementSizeForOpcode(
unsigned Opc)
const {
10754bool AArch64InstrInfo::isPTestLikeOpcode(
unsigned Opc)
const {
10758bool AArch64InstrInfo::isWhileOpcode(
unsigned Opc)
const {
10763AArch64InstrInfo::getTailDuplicateSize(
CodeGenOptLevel OptLevel)
const {
10767bool AArch64InstrInfo::isLegalAddressingMode(
unsigned NumBytes, int64_t
Offset,
10768 unsigned Scale)
const {
10779 unsigned Shift =
Log2_64(NumBytes);
10780 if (NumBytes &&
Offset > 0 && (
Offset / NumBytes) <= (1LL << 12) - 1 &&
10788 return Scale == 1 || (Scale > 0 && Scale == NumBytes);
10793 return AArch64::BLRNoIP;
10795 return AArch64::BLR;
10800 Register TargetReg,
bool FrameSetup)
const {
10801 assert(TargetReg != AArch64::SP &&
"New top of stack cannot already be in SP");
10813 MF.
insert(MBBInsertPoint, LoopTestMBB);
10816 MF.
insert(MBBInsertPoint, LoopBodyMBB);
10818 MF.
insert(MBBInsertPoint, ExitMBB);
10828 BuildMI(*LoopTestMBB, LoopTestMBB->
end(),
DL,
TII->get(AArch64::SUBSXrx64),
10836 BuildMI(*LoopTestMBB, LoopTestMBB->
end(),
DL,
TII->get(AArch64::Bcc))
10842 BuildMI(*LoopBodyMBB, LoopBodyMBB->
end(),
DL,
TII->get(AArch64::STRXui))
10855 BuildMI(*ExitMBB, ExitMBB->
end(),
DL,
TII->get(AArch64::ADDXri), AArch64::SP)
10874 MBB.addSuccessor(LoopTestMBB);
10880 return ExitMBB->
begin();
10897 unsigned CompCounterOprNum;
10901 unsigned UpdateCounterOprNum;
10905 bool IsUpdatePriorComp;
10917 TII(MF->getSubtarget().getInstrInfo()),
10918 TRI(MF->getSubtarget().getRegisterInfo()),
MRI(MF->getRegInfo()),
10919 LoopBB(LoopBB), CondBranch(CondBranch), Comp(Comp),
10920 CompCounterOprNum(CompCounterOprNum), Update(Update),
10921 UpdateCounterOprNum(UpdateCounterOprNum),
Init(
Init),
10922 IsUpdatePriorComp(IsUpdatePriorComp),
Cond(
Cond.begin(),
Cond.end()) {}
10924 bool shouldIgnoreForPipelining(
const MachineInstr *
MI)
const override {
10930 std::optional<bool> createTripCountGreaterCondition(
10931 int TC, MachineBasicBlock &
MBB,
10932 SmallVectorImpl<MachineOperand> &CondParam)
override {
10940 void createRemainingIterationsGreaterCondition(
10941 int TC, MachineBasicBlock &
MBB, SmallVectorImpl<MachineOperand> &
Cond,
10942 DenseMap<MachineInstr *, MachineInstr *> &LastStage0Insts)
override;
10944 void setPreheader(MachineBasicBlock *NewPreheader)
override {}
10946 void adjustTripCount(
int TripCountAdjust)
override {}
10948 bool isMVEExpanderSupported()
override {
return true; }
10961 MBB.getParent()->getSubtarget().getRegisterInfo();
10966 Result =
MRI.createVirtualRegister(
10969 }
else if (
I == ReplaceOprNum) {
10970 MRI.constrainRegClass(ReplaceReg,
10975 MBB.insert(InsertTo, NewMI);
10979void AArch64PipelinerLoopInfo::createRemainingIterationsGreaterCondition(
10995 assert(CondBranch->getOpcode() == AArch64::Bcc);
10999 if (CondBranch->getOperand(1).getMBB() == LoopBB)
11006 auto AccumulateCond = [&](
Register CurCond,
11008 Register NewCond =
MRI.createVirtualRegister(&AArch64::GPR64commonRegClass);
11017 if (!LastStage0Insts.
empty() && LastStage0Insts[Comp]->getParent() == &
MBB) {
11021 for (
int I = 0;
I <= TC; ++
I) {
11027 AccCond = AccumulateCond(AccCond, CC);
11031 if (Update != Comp && IsUpdatePriorComp) {
11033 LastStage0Insts[Comp]->getOperand(CompCounterOprNum).getReg();
11034 NextCounter =
cloneInstr(Update, UpdateCounterOprNum, Counter,
MBB,
11038 NextCounter = LastStage0Insts[Update]->getOperand(0).getReg();
11040 }
else if (Update != Comp) {
11045 Counter = NextCounter;
11049 if (LastStage0Insts.
empty()) {
11053 if (IsUpdatePriorComp)
11058 Counter = LastStage0Insts[Comp]->getOperand(CompCounterOprNum).getReg();
11061 for (
int I = 0;
I <= TC; ++
I) {
11065 AccCond = AccumulateCond(AccCond, CC);
11066 if (
I != TC && Update != Comp)
11069 Counter = NextCounter;
11085 assert(Phi.getNumOperands() == 5);
11086 if (Phi.getOperand(2).getMBB() ==
MBB) {
11087 RegMBB = Phi.getOperand(1).getReg();
11088 RegOther = Phi.getOperand(3).getReg();
11090 assert(Phi.getOperand(4).getMBB() ==
MBB);
11091 RegMBB = Phi.getOperand(3).getReg();
11092 RegOther = Phi.getOperand(1).getReg();
11097 if (!
Reg.isVirtual())
11100 return MRI.getVRegDef(
Reg)->getParent() != BB;
11106 unsigned &UpdateCounterOprNum,
Register &InitReg,
11107 bool &IsUpdatePriorComp) {
11121 if (!
Reg.isVirtual())
11124 UpdateInst =
nullptr;
11125 UpdateCounterOprNum = 0;
11127 IsUpdatePriorComp =
true;
11131 if (Def->getParent() != LoopBB)
11133 if (Def->isCopy()) {
11135 if (Def->getOperand(0).getSubReg() || Def->getOperand(1).getSubReg())
11137 CurReg = Def->getOperand(1).getReg();
11138 }
else if (Def->isPHI()) {
11142 IsUpdatePriorComp =
false;
11147 switch (Def->getOpcode()) {
11148 case AArch64::ADDSXri:
11149 case AArch64::ADDSWri:
11150 case AArch64::SUBSXri:
11151 case AArch64::SUBSWri:
11152 case AArch64::ADDXri:
11153 case AArch64::ADDWri:
11154 case AArch64::SUBXri:
11155 case AArch64::SUBWri:
11157 UpdateCounterOprNum = 1;
11159 case AArch64::ADDSXrr:
11160 case AArch64::ADDSWrr:
11161 case AArch64::SUBSXrr:
11162 case AArch64::SUBSWrr:
11163 case AArch64::ADDXrr:
11164 case AArch64::ADDWrr:
11165 case AArch64::SUBXrr:
11166 case AArch64::SUBWrr:
11169 UpdateCounterOprNum = 1;
11171 UpdateCounterOprNum = 2;
11178 CurReg = Def->getOperand(UpdateCounterOprNum).getReg();
11193std::unique_ptr<TargetInstrInfo::PipelinerLoopInfo>
11204 if (
MI.isCall() ||
MI.hasUnmodeledSideEffects())
11215 if (
TBB == LoopBB && FBB == LoopBB)
11219 if (
TBB != LoopBB && FBB ==
nullptr)
11222 assert((
TBB == LoopBB || FBB == LoopBB) &&
11223 "The Loop must be a single-basic-block loop");
11228 if (CondBranch->
getOpcode() != AArch64::Bcc)
11236 unsigned CompCounterOprNum = 0;
11238 if (
MI.modifiesRegister(AArch64::NZCV, &
TRI)) {
11242 switch (
MI.getOpcode()) {
11243 case AArch64::SUBSXri:
11244 case AArch64::SUBSWri:
11245 case AArch64::ADDSXri:
11246 case AArch64::ADDSWri:
11248 CompCounterOprNum = 1;
11250 case AArch64::ADDSWrr:
11251 case AArch64::ADDSXrr:
11252 case AArch64::SUBSWrr:
11253 case AArch64::SUBSXrr:
11257 if (isWhileOpcode(
MI.getOpcode())) {
11264 if (CompCounterOprNum == 0) {
11266 CompCounterOprNum = 2;
11268 CompCounterOprNum = 1;
11280 bool IsUpdatePriorComp;
11281 unsigned UpdateCounterOprNum;
11283 Update, UpdateCounterOprNum,
Init, IsUpdatePriorComp))
11286 return std::make_unique<AArch64PipelinerLoopInfo>(
11287 LoopBB, CondBranch, Comp, CompCounterOprNum, Update, UpdateCounterOprNum,
11297 TypeSize Scale(0U,
false), Width(0U,
false);
11298 int64_t MinOffset, MaxOffset;
11299 if (
getMemOpInfo(
MI.getOpcode(), Scale, Width, MinOffset, MaxOffset)) {
11301 if (
MI.getOperand(ImmIdx).isImm() && !
MI.getOperand(ImmIdx - 1).isFI()) {
11302 int64_t Imm =
MI.getOperand(ImmIdx).getImm();
11303 if (Imm < MinOffset || Imm > MaxOffset) {
11304 ErrInfo =
"Unexpected immediate on load/store instruction";
11310 const MCInstrDesc &MCID =
MI.getDesc();
11312 const MachineOperand &MO =
MI.getOperand(
Op);
11316 ErrInfo =
"OPERAND_IMPLICIT_IMM_0 should be 0";
11325 ErrInfo =
"OPERAND_SHIFT_MSL should be msl shift of 8 or 16";
11336#define GET_INSTRINFO_HELPERS
11337#define GET_INSTRMAP_INFO
11338#include "AArch64GenInstrInfo.inc"
unsigned const MachineRegisterInfo * MRI
MachineInstrBuilder & UseMI
MachineInstrBuilder MachineInstrBuilder & DefMI
static bool forwardCopyWillClobberTuple(unsigned DestReg, unsigned SrcReg, unsigned NumRegs)
static cl::opt< unsigned > BCCDisplacementBits("aarch64-bcc-offset-bits", cl::Hidden, cl::init(19), cl::desc("Restrict range of Bcc instructions (DEBUG)"))
static Register genNeg(MachineFunction &MF, MachineRegisterInfo &MRI, const TargetInstrInfo *TII, MachineInstr &Root, SmallVectorImpl< MachineInstr * > &InsInstrs, DenseMap< Register, unsigned > &InstrIdxForVirtReg, unsigned MnegOpc, const TargetRegisterClass *RC)
genNeg - Helper to generate an intermediate negation of the second operand of Root
static cl::opt< unsigned > GatherOptSearchLimit("aarch64-search-limit", cl::Hidden, cl::init(2048), cl::desc("Restrict range of instructions to search for the " "machine-combiner gather pattern optimization"))
static bool getMaddPatterns(MachineInstr &Root, SmallVectorImpl< unsigned > &Patterns)
Find instructions that can be turned into madd.
static AArch64CC::CondCode findCondCodeUsedByInstr(const MachineInstr &Instr)
Find a condition code used by the instruction.
static MachineInstr * genFusedMultiplyAcc(MachineFunction &MF, MachineRegisterInfo &MRI, const TargetInstrInfo *TII, MachineInstr &Root, SmallVectorImpl< MachineInstr * > &InsInstrs, unsigned IdxMulOpd, unsigned MaddOpc, const TargetRegisterClass *RC)
genFusedMultiplyAcc - Helper to generate fused multiply accumulate instructions.
static MachineInstr * genFusedMultiplyAccNeg(MachineFunction &MF, MachineRegisterInfo &MRI, const TargetInstrInfo *TII, MachineInstr &Root, SmallVectorImpl< MachineInstr * > &InsInstrs, DenseMap< Register, unsigned > &InstrIdxForVirtReg, unsigned IdxMulOpd, unsigned MaddOpc, unsigned MnegOpc, const TargetRegisterClass *RC)
genFusedMultiplyAccNeg - Helper to generate fused multiply accumulate instructions with an additional...
static bool isCombineInstrCandidate64(unsigned Opc)
static unsigned removeCopies(const MachineRegisterInfo &MRI, unsigned VReg)
static bool areCFlagsAccessedBetweenInstrs(MachineBasicBlock::iterator From, MachineBasicBlock::iterator To, const TargetRegisterInfo *TRI, const AccessKind AccessToCheck=AK_All)
True when condition flags are accessed (either by writing or reading) on the instruction trace starti...
static bool getFMAPatterns(MachineInstr &Root, SmallVectorImpl< unsigned > &Patterns)
Floating-Point Support.
static bool isADDSRegImm(unsigned Opcode)
static void appendOffsetComment(int NumBytes, llvm::raw_string_ostream &Comment, StringRef RegScale={})
static unsigned sForm(MachineInstr &Instr)
Get opcode of S version of Instr.
static bool isCombineInstrSettingFlag(unsigned Opc)
static bool getFNEGPatterns(MachineInstr &Root, SmallVectorImpl< unsigned > &Patterns)
static bool getIndVarInfo(Register Reg, const MachineBasicBlock *LoopBB, MachineInstr *&UpdateInst, unsigned &UpdateCounterOprNum, Register &InitReg, bool &IsUpdatePriorComp)
If Reg is an induction variable, return true and set some parameters.
static const MachineInstrBuilder & AddSubReg(const MachineInstrBuilder &MIB, MCRegister Reg, unsigned SubIdx, unsigned State, const TargetRegisterInfo *TRI)
static bool canPairLdStOpc(unsigned FirstOpc, unsigned SecondOpc)
static int findCondCodeUseOperandIdxForBranchOrSelect(const MachineInstr &Instr)
static bool isPostIndexLdStOpcode(unsigned Opcode)
Return true if the opcode is a post-index ld/st instruction, which really loads from base+0.
static unsigned getBranchDisplacementBits(unsigned Opc)
static cl::opt< unsigned > CBDisplacementBits("aarch64-cb-offset-bits", cl::Hidden, cl::init(9), cl::desc("Restrict range of CB instructions (DEBUG)"))
static std::optional< ParamLoadedValue > describeORRLoadedValue(const MachineInstr &MI, Register DescribedReg, const TargetInstrInfo *TII, const TargetRegisterInfo *TRI)
If the given ORR instruction is a copy, and DescribedReg overlaps with the destination register then,...
static bool getFMULPatterns(MachineInstr &Root, SmallVectorImpl< unsigned > &Patterns)
static void appendReadRegExpr(SmallVectorImpl< char > &Expr, unsigned RegNum)
static MachineInstr * genMaddR(MachineFunction &MF, MachineRegisterInfo &MRI, const TargetInstrInfo *TII, MachineInstr &Root, SmallVectorImpl< MachineInstr * > &InsInstrs, unsigned IdxMulOpd, unsigned MaddOpc, unsigned VR, const TargetRegisterClass *RC)
genMaddR - Generate madd instruction and combine mul and add using an extra virtual register Example ...
static Register cloneInstr(const MachineInstr *MI, unsigned ReplaceOprNum, Register ReplaceReg, MachineBasicBlock &MBB, MachineBasicBlock::iterator InsertTo)
Clone an instruction from MI.
static bool scaleOffset(unsigned Opc, int64_t &Offset)
static bool canCombineWithFMUL(MachineBasicBlock &MBB, MachineOperand &MO, unsigned MulOpc)
unsigned scaledOffsetOpcode(unsigned Opcode, unsigned &Scale)
static MachineInstr * genFusedMultiplyIdx(MachineFunction &MF, MachineRegisterInfo &MRI, const TargetInstrInfo *TII, MachineInstr &Root, SmallVectorImpl< MachineInstr * > &InsInstrs, unsigned IdxMulOpd, unsigned MaddOpc, const TargetRegisterClass *RC)
genFusedMultiplyIdx - Helper to generate fused multiply accumulate instructions.
static MachineInstr * genIndexedMultiply(MachineInstr &Root, SmallVectorImpl< MachineInstr * > &InsInstrs, unsigned IdxDupOp, unsigned MulOpc, const TargetRegisterClass *RC, MachineRegisterInfo &MRI)
Fold (FMUL x (DUP y lane)) into (FMUL_indexed x y lane)
static bool isSUBSRegImm(unsigned Opcode)
static bool UpdateOperandRegClass(MachineInstr &Instr)
static const TargetRegisterClass * getRegClass(const MachineInstr &MI, Register Reg)
static bool canCmpInstrBeRemoved(MachineInstr &MI, MachineInstr &CmpInstr, int CmpValue, const TargetRegisterInfo &TRI, SmallVectorImpl< MachineInstr * > &CCUseInstrs, bool &IsInvertCC)
unsigned unscaledOffsetOpcode(unsigned Opcode)
static bool getLoadPatterns(MachineInstr &Root, SmallVectorImpl< unsigned > &Patterns)
Search for patterns of LD instructions we can optimize.
static bool canInstrSubstituteCmpInstr(MachineInstr &MI, MachineInstr &CmpInstr, const TargetRegisterInfo &TRI)
Check if CmpInstr can be substituted by MI.
static UsedNZCV getUsedNZCV(AArch64CC::CondCode CC)
static bool isCombineInstrCandidateFP(const MachineInstr &Inst)
static void appendLoadRegExpr(SmallVectorImpl< char > &Expr, int64_t OffsetFromDefCFA)
static void appendConstantExpr(SmallVectorImpl< char > &Expr, int64_t Constant, dwarf::LocationAtom Operation)
static unsigned convertToNonFlagSettingOpc(const MachineInstr &MI)
Return the opcode that does not set flags when possible - otherwise return the original opcode.
static bool outliningCandidatesV8_3OpsConsensus(const outliner::Candidate &a, const outliner::Candidate &b)
static bool isCombineInstrCandidate32(unsigned Opc)
static void parseCondBranch(MachineInstr *LastInst, MachineBasicBlock *&Target, SmallVectorImpl< MachineOperand > &Cond)
static unsigned offsetExtendOpcode(unsigned Opcode)
static void loadRegPairFromStackSlot(const TargetRegisterInfo &TRI, MachineBasicBlock &MBB, MachineBasicBlock::iterator InsertBefore, const MCInstrDesc &MCID, Register DestReg, unsigned SubIdx0, unsigned SubIdx1, int FI, MachineMemOperand *MMO)
static void generateGatherLanePattern(MachineInstr &Root, SmallVectorImpl< MachineInstr * > &InsInstrs, SmallVectorImpl< MachineInstr * > &DelInstrs, DenseMap< Register, unsigned > &InstrIdxForVirtReg, unsigned Pattern, unsigned NumLanes)
Generate optimized instruction sequence for gather load patterns to improve Memory-Level Parallelism ...
static bool getMiscPatterns(MachineInstr &Root, SmallVectorImpl< unsigned > &Patterns)
Find other MI combine patterns.
static bool outliningCandidatesSigningKeyConsensus(const outliner::Candidate &a, const outliner::Candidate &b)
static bool outliningCandidatesSigningScopeConsensus(const outliner::Candidate &a, const outliner::Candidate &b)
static bool shouldClusterFI(const MachineFrameInfo &MFI, int FI1, int64_t Offset1, unsigned Opcode1, int FI2, int64_t Offset2, unsigned Opcode2)
static cl::opt< unsigned > TBZDisplacementBits("aarch64-tbz-offset-bits", cl::Hidden, cl::init(14), cl::desc("Restrict range of TB[N]Z instructions (DEBUG)"))
static void extractPhiReg(const MachineInstr &Phi, const MachineBasicBlock *MBB, Register &RegMBB, Register &RegOther)
static MCCFIInstruction createDefCFAExpression(const TargetRegisterInfo &TRI, unsigned Reg, const StackOffset &Offset)
static bool isDefinedOutside(Register Reg, const MachineBasicBlock *BB)
static MachineInstr * genFusedMultiply(MachineFunction &MF, MachineRegisterInfo &MRI, const TargetInstrInfo *TII, MachineInstr &Root, SmallVectorImpl< MachineInstr * > &InsInstrs, unsigned IdxMulOpd, unsigned MaddOpc, const TargetRegisterClass *RC, FMAInstKind kind=FMAInstKind::Default, const Register *ReplacedAddend=nullptr)
genFusedMultiply - Generate fused multiply instructions.
static bool getGatherLanePattern(MachineInstr &Root, SmallVectorImpl< unsigned > &Patterns, unsigned LoadLaneOpCode, unsigned NumLanes)
Check if the given instruction forms a gather load pattern that can be optimized for better Memory-Le...
static MachineInstr * genFusedMultiplyIdxNeg(MachineFunction &MF, MachineRegisterInfo &MRI, const TargetInstrInfo *TII, MachineInstr &Root, SmallVectorImpl< MachineInstr * > &InsInstrs, DenseMap< Register, unsigned > &InstrIdxForVirtReg, unsigned IdxMulOpd, unsigned MaddOpc, unsigned MnegOpc, const TargetRegisterClass *RC)
genFusedMultiplyAccNeg - Helper to generate fused multiply accumulate instructions with an additional...
static bool isCombineInstrCandidate(unsigned Opc)
static unsigned regOffsetOpcode(unsigned Opcode)
MachineOutlinerClass
Constants defining how certain sequences should be outlined.
@ MachineOutlinerTailCall
Emit a save, restore, call, and return.
@ MachineOutlinerRegSave
Emit a call and tail-call.
@ MachineOutlinerNoLRSave
Only emit a branch.
@ MachineOutlinerThunk
Emit a call and return.
static cl::opt< unsigned > BDisplacementBits("aarch64-b-offset-bits", cl::Hidden, cl::init(26), cl::desc("Restrict range of B instructions (DEBUG)"))
static bool areCFlagsAliveInSuccessors(const MachineBasicBlock *MBB)
Check if AArch64::NZCV should be alive in successors of MBB.
static void emitFrameOffsetAdj(MachineBasicBlock &MBB, MachineBasicBlock::iterator MBBI, const DebugLoc &DL, unsigned DestReg, unsigned SrcReg, int64_t Offset, unsigned Opc, const TargetInstrInfo *TII, MachineInstr::MIFlag Flag, bool NeedsWinCFI, bool *HasWinCFI, bool EmitCFAOffset, StackOffset CFAOffset, unsigned FrameReg)
static bool isCheapImmediate(const MachineInstr &MI, unsigned BitSize)
static cl::opt< unsigned > CBZDisplacementBits("aarch64-cbz-offset-bits", cl::Hidden, cl::init(19), cl::desc("Restrict range of CB[N]Z instructions (DEBUG)"))
static unsigned canFoldIntoCSel(const MachineRegisterInfo &MRI, unsigned VReg, unsigned *NewVReg=nullptr)
static void genSubAdd2SubSub(MachineFunction &MF, MachineRegisterInfo &MRI, const TargetInstrInfo *TII, MachineInstr &Root, SmallVectorImpl< MachineInstr * > &InsInstrs, SmallVectorImpl< MachineInstr * > &DelInstrs, unsigned IdxOpd1, DenseMap< Register, unsigned > &InstrIdxForVirtReg)
Do the following transformation A - (B + C) ==> (A - B) - C A - (B + C) ==> (A - C) - B.
static void signOutlinedFunction(MachineFunction &MF, MachineBasicBlock &MBB, const AArch64InstrInfo *TII, bool ShouldSignReturnAddr)
static MachineInstr * genFNegatedMAD(MachineFunction &MF, MachineRegisterInfo &MRI, const TargetInstrInfo *TII, MachineInstr &Root, SmallVectorImpl< MachineInstr * > &InsInstrs)
static bool canCombineWithMUL(MachineBasicBlock &MBB, MachineOperand &MO, unsigned MulOpc, unsigned ZeroReg)
static void storeRegPairToStackSlot(const TargetRegisterInfo &TRI, MachineBasicBlock &MBB, MachineBasicBlock::iterator InsertBefore, const MCInstrDesc &MCID, Register SrcReg, bool IsKill, unsigned SubIdx0, unsigned SubIdx1, int FI, MachineMemOperand *MMO)
assert(UImm &&(UImm !=~static_cast< T >(0)) &&"Invalid immediate!")
MachineBasicBlock MachineBasicBlock::iterator DebugLoc DL
MachineBasicBlock MachineBasicBlock::iterator MBBI
static const Function * getParent(const Value *V)
static GCRegistry::Add< ErlangGC > A("erlang", "erlang-compatible garbage collector")
static GCRegistry::Add< OcamlGC > B("ocaml", "ocaml 3.10-compatible GC")
const HexagonInstrInfo * TII
Module.h This file contains the declarations for the Module class.
const AbstractManglingParser< Derived, Alloc >::OperatorInfo AbstractManglingParser< Derived, Alloc >::Ops[]
This file implements the LivePhysRegs utility for tracking liveness of physical registers.
Machine Check Debug Module
Register const TargetRegisterInfo * TRI
Promote Memory to Register
static unsigned getReg(const MCDisassembler *D, unsigned RC, unsigned RegNo)
MachineInstr unsigned OpIdx
ConstantRange Range(APInt(BitWidth, Low), APInt(BitWidth, High))
PowerPC Reduce CR logical Operation
const SmallVectorImpl< MachineOperand > MachineBasicBlock * TBB
const SmallVectorImpl< MachineOperand > & Cond
This file declares the machine register scavenger class.
static bool contains(SmallPtrSetImpl< ConstantExpr * > &Cache, ConstantExpr *Expr, Constant *C)
This file defines the SmallSet class.
This file defines the SmallVector class.
static bool canCombine(MachineBasicBlock &MBB, MachineOperand &MO, unsigned CombineOpc=0)
AArch64FunctionInfo - This class is derived from MachineFunctionInfo and contains private AArch64-spe...
bool hasStreamingModeChanges() const
bool shouldSignReturnAddress(const MachineFunction &MF) const
void setOutliningStyle(const std::string &Style)
std::optional< bool > hasRedZone() const
bool shouldSignWithBKey() const
static bool isHForm(const MachineInstr &MI)
Returns whether the instruction is in H form (16 bit operands)
void insertSelect(MachineBasicBlock &MBB, MachineBasicBlock::iterator MI, const DebugLoc &DL, Register DstReg, ArrayRef< MachineOperand > Cond, Register TrueReg, Register FalseReg) const override
static bool hasBTISemantics(const MachineInstr &MI)
Returns whether the instruction can be compatible with non-zero BTYPE.
static bool isQForm(const MachineInstr &MI)
Returns whether the instruction is in Q form (128 bit operands)
static bool getMemOpInfo(unsigned Opcode, TypeSize &Scale, TypeSize &Width, int64_t &MinOffset, int64_t &MaxOffset)
Returns true if opcode Opc is a memory operation.
static bool isTailCallReturnInst(const MachineInstr &MI)
Returns true if MI is one of the TCRETURN* instructions.
static bool isFPRCopy(const MachineInstr &MI)
Does this instruction rename an FPR without modifying bits?
MachineInstr * emitLdStWithAddr(MachineInstr &MemI, const ExtAddrMode &AM) const override
std::optional< DestSourcePair > isCopyInstrImpl(const MachineInstr &MI) const override
If the specific machine instruction is an instruction that moves/copies value from one register to an...
MachineBasicBlock * getBranchDestBlock(const MachineInstr &MI) const override
unsigned getInstSizeInBytes(const MachineInstr &MI) const override
GetInstSize - Return the number of bytes of code the specified instruction may be.
bool areMemAccessesTriviallyDisjoint(const MachineInstr &MIa, const MachineInstr &MIb) const override
void copyPhysReg(MachineBasicBlock &MBB, MachineBasicBlock::iterator I, const DebugLoc &DL, Register DestReg, Register SrcReg, bool KillSrc, bool RenamableDest=false, bool RenamableSrc=false) const override
static bool isGPRCopy(const MachineInstr &MI)
Does this instruction rename a GPR without modifying bits?
static unsigned convertToFlagSettingOpc(unsigned Opc)
Return the opcode that set flags when possible.
bool isBranchOffsetInRange(unsigned BranchOpc, int64_t BrOffset) const override
bool canInsertSelect(const MachineBasicBlock &, ArrayRef< MachineOperand > Cond, Register, Register, Register, int &, int &, int &) const override
static const MachineOperand & getLdStOffsetOp(const MachineInstr &MI)
Returns the immediate offset operator of a load/store.
bool isCoalescableExtInstr(const MachineInstr &MI, Register &SrcReg, Register &DstReg, unsigned &SubIdx) const override
static std::optional< unsigned > getUnscaledLdSt(unsigned Opc)
Returns the unscaled load/store for the scaled load/store opcode, if there is a corresponding unscale...
static bool hasUnscaledLdStOffset(unsigned Opc)
Return true if it has an unscaled load/store offset.
static const MachineOperand & getLdStAmountOp(const MachineInstr &MI)
Returns the shift amount operator of a load/store.
static bool isPreLdSt(const MachineInstr &MI)
Returns whether the instruction is a pre-indexed load/store.
std::optional< ExtAddrMode > getAddrModeFromMemoryOp(const MachineInstr &MemI, const TargetRegisterInfo *TRI) const override
bool getMemOperandsWithOffsetWidth(const MachineInstr &MI, SmallVectorImpl< const MachineOperand * > &BaseOps, int64_t &Offset, bool &OffsetIsScalable, LocationSize &Width, const TargetRegisterInfo *TRI) const override
bool analyzeBranchPredicate(MachineBasicBlock &MBB, MachineBranchPredicate &MBP, bool AllowModify) const override
void insertIndirectBranch(MachineBasicBlock &MBB, MachineBasicBlock &NewDestBB, MachineBasicBlock &RestoreBB, const DebugLoc &DL, int64_t BrOffset, RegScavenger *RS) const override
static bool isPairableLdStInst(const MachineInstr &MI)
Return true if pairing the given load or store may be paired with another.
const AArch64RegisterInfo & getRegisterInfo() const
getRegisterInfo - TargetInstrInfo is a superset of MRegister info.
static bool isPreSt(const MachineInstr &MI)
Returns whether the instruction is a pre-indexed store.
MachineInstr * foldMemoryOperandImpl(MachineFunction &MF, MachineInstr &MI, ArrayRef< unsigned > Ops, MachineBasicBlock::iterator InsertPt, int FrameIndex, LiveIntervals *LIS=nullptr, VirtRegMap *VRM=nullptr) const override
void insertNoop(MachineBasicBlock &MBB, MachineBasicBlock::iterator MI) const override
AArch64InstrInfo(const AArch64Subtarget &STI)
static bool isPairedLdSt(const MachineInstr &MI)
Returns whether the instruction is a paired load/store.
bool getMemOperandWithOffsetWidth(const MachineInstr &MI, const MachineOperand *&BaseOp, int64_t &Offset, bool &OffsetIsScalable, TypeSize &Width, const TargetRegisterInfo *TRI) const
If OffsetIsScalable is set to 'true', the offset is scaled by vscale.
Register isLoadFromStackSlot(const MachineInstr &MI, int &FrameIndex) const override
bool reverseBranchCondition(SmallVectorImpl< MachineOperand > &Cond) const override
static bool isStridedAccess(const MachineInstr &MI)
Return true if the given load or store is a strided memory access.
bool shouldClusterMemOps(ArrayRef< const MachineOperand * > BaseOps1, int64_t Offset1, bool OffsetIsScalable1, ArrayRef< const MachineOperand * > BaseOps2, int64_t Offset2, bool OffsetIsScalable2, unsigned ClusterSize, unsigned NumBytes) const override
Detect opportunities for ldp/stp formation.
unsigned removeBranch(MachineBasicBlock &MBB, int *BytesRemoved=nullptr) const override
bool isThroughputPattern(unsigned Pattern) const override
Return true when a code sequence can improve throughput.
MachineOperand & getMemOpBaseRegImmOfsOffsetOperand(MachineInstr &LdSt) const
Return the immediate offset of the base register in a load/store LdSt.
bool analyzeBranch(MachineBasicBlock &MBB, MachineBasicBlock *&TBB, MachineBasicBlock *&FBB, SmallVectorImpl< MachineOperand > &Cond, bool AllowModify=false) const override
bool canFoldIntoAddrMode(const MachineInstr &MemI, Register Reg, const MachineInstr &AddrI, ExtAddrMode &AM) const override
static bool isLdStPairSuppressed(const MachineInstr &MI)
Return true if pairing the given load or store is hinted to be unprofitable.
std::unique_ptr< TargetInstrInfo::PipelinerLoopInfo > analyzeLoopForPipelining(MachineBasicBlock *LoopBB) const override
bool isSchedulingBoundary(const MachineInstr &MI, const MachineBasicBlock *MBB, const MachineFunction &MF) const override
unsigned insertBranch(MachineBasicBlock &MBB, MachineBasicBlock *TBB, MachineBasicBlock *FBB, ArrayRef< MachineOperand > Cond, const DebugLoc &DL, int *BytesAdded=nullptr) const override
MachineBasicBlock::iterator probedStackAlloc(MachineBasicBlock::iterator MBBI, Register TargetReg, bool FrameSetup) const
Return true when there is potentially a faster code sequence for an instruction chain ending in Root.
bool optimizeCompareInstr(MachineInstr &CmpInstr, Register SrcReg, Register SrcReg2, int64_t CmpMask, int64_t CmpValue, const MachineRegisterInfo *MRI) const override
optimizeCompareInstr - Convert the instruction supplying the argument to the comparison into one that...
static unsigned getLoadStoreImmIdx(unsigned Opc)
Returns the index for the immediate for a given instruction.
static bool isGPRZero(const MachineInstr &MI)
Does this instruction set its full destination register to zero?
void copyGPRRegTuple(MachineBasicBlock &MBB, MachineBasicBlock::iterator I, const DebugLoc &DL, MCRegister DestReg, MCRegister SrcReg, bool KillSrc, unsigned Opcode, unsigned ZeroReg, llvm::ArrayRef< unsigned > Indices) const
bool analyzeCompare(const MachineInstr &MI, Register &SrcReg, Register &SrcReg2, int64_t &CmpMask, int64_t &CmpValue) const override
analyzeCompare - For a comparison instruction, return the source registers in SrcReg and SrcReg2,...
CombinerObjective getCombinerObjective(unsigned Pattern) const override
static bool isFpOrNEON(Register Reg)
Returns whether the physical register is FP or NEON.
bool isAsCheapAsAMove(const MachineInstr &MI) const override
std::optional< DestSourcePair > isCopyLikeInstrImpl(const MachineInstr &MI) const override
static void suppressLdStPair(MachineInstr &MI)
Hint that pairing the given load or store is unprofitable.
Register isStoreToStackSlot(const MachineInstr &MI, int &FrameIndex) const override
static bool isPreLd(const MachineInstr &MI)
Returns whether the instruction is a pre-indexed load.
void storeRegToStackSlot(MachineBasicBlock &MBB, MachineBasicBlock::iterator MBBI, Register SrcReg, bool isKill, int FrameIndex, const TargetRegisterClass *RC, const TargetRegisterInfo *TRI, Register VReg, MachineInstr::MIFlag Flags=MachineInstr::NoFlags) const override
void loadRegFromStackSlot(MachineBasicBlock &MBB, MachineBasicBlock::iterator MBBI, Register DestReg, int FrameIndex, const TargetRegisterClass *RC, const TargetRegisterInfo *TRI, Register VReg, MachineInstr::MIFlag Flags=MachineInstr::NoFlags) const override
void copyPhysRegTuple(MachineBasicBlock &MBB, MachineBasicBlock::iterator I, const DebugLoc &DL, MCRegister DestReg, MCRegister SrcReg, bool KillSrc, unsigned Opcode, llvm::ArrayRef< unsigned > Indices) const
bool optimizeCondBranch(MachineInstr &MI) const override
Replace csincr-branch sequence by simple conditional branch.
static int getMemScale(unsigned Opc)
Scaling factor for (scaled or unscaled) load or store.
bool isCandidateToMergeOrPair(const MachineInstr &MI) const
Return true if this is a load/store that can be potentially paired/merged.
MCInst getNop() const override
static const MachineOperand & getLdStBaseOp(const MachineInstr &MI)
Returns the base register operator of a load/store.
bool isReservedReg(const MachineFunction &MF, MCRegister Reg) const
const AArch64RegisterInfo * getRegisterInfo() const override
bool isNeonAvailable() const
Returns true if the target has NEON and the function at runtime is known to have NEON enabled (e....
bool isSVEorStreamingSVEAvailable() const
Returns true if the target has access to either the full range of SVE instructions,...
ArrayRef - Represent a constant reference to an array (0 or more elements consecutively in memory),...
const T & front() const
front - Get the first element.
size_t size() const
size - Get the array size.
This is an important base class in LLVM.
std::pair< iterator, bool > insert(const std::pair< KeyT, ValueT > &KV)
bool hasOptSize() const
Optimize this function for size (-Os) or minimum size (-Oz).
bool hasMinSize() const
Optimize this function for minimum size (-Oz).
Module * getParent()
Get the module that this global value is contained inside of...
LLVM_ABI InstListType::iterator eraseFromParent()
This method unlinks 'this' from the containing basic block and deletes it.
unsigned getOpcode() const
Returns a member of one of the enums like Instruction::Add.
static LocationSize precise(uint64_t Value)
This class is intended to be used as a base class for asm properties and features specific to the tar...
bool usesWindowsCFI() const
static MCCFIInstruction cfiDefCfa(MCSymbol *L, unsigned Register, int64_t Offset, SMLoc Loc={})
.cfi_def_cfa defines a rule for computing CFA as: take address from Register and add Offset to it.
static MCCFIInstruction createOffset(MCSymbol *L, unsigned Register, int64_t Offset, SMLoc Loc={})
.cfi_offset Previous value of Register is saved at offset Offset from CFA.
static MCCFIInstruction cfiDefCfaOffset(MCSymbol *L, int64_t Offset, SMLoc Loc={})
.cfi_def_cfa_offset modifies a rule for computing CFA.
static MCCFIInstruction createEscape(MCSymbol *L, StringRef Vals, SMLoc Loc={}, StringRef Comment="")
.cfi_escape Allows the user to add arbitrary bytes to the unwind info.
MCInstBuilder & addImm(int64_t Val)
Add a new integer immediate operand.
Instances of this class represent a single low-level machine instruction.
Describe properties that are true of each instruction in the target description file.
unsigned getNumOperands() const
Return the number of declared MachineOperands for this MachineInstruction.
ArrayRef< MCOperandInfo > operands() const
Wrapper class representing physical registers. Should be passed by value.
constexpr bool isValid() const
static constexpr unsigned NoRegister
static MDTuple * get(LLVMContext &Context, ArrayRef< Metadata * > MDs)
bool isInlineAsmBrIndirectTarget() const
Returns true if this is the indirect dest of an INLINEASM_BR.
unsigned pred_size() const
LLVM_ABI void transferSuccessorsAndUpdatePHIs(MachineBasicBlock *FromMBB)
Transfers all the successors, as in transferSuccessors, and update PHI operands in the successor bloc...
LLVM_ABI instr_iterator insert(instr_iterator I, MachineInstr *M)
Insert MI into the instruction list before I, possibly inside a bundle.
reverse_instr_iterator instr_rbegin()
LLVM_ABI iterator getFirstTerminator()
Returns an iterator to the first terminator instruction of this basic block.
LLVM_ABI void addSuccessor(MachineBasicBlock *Succ, BranchProbability Prob=BranchProbability::getUnknown())
Add Succ as a successor of this MachineBasicBlock.
reverse_instr_iterator instr_rend()
Instructions::iterator instr_iterator
instr_iterator instr_end()
Instructions::const_iterator const_instr_iterator
void addLiveIn(MCRegister PhysReg, LaneBitmask LaneMask=LaneBitmask::getAll())
Adds the specified register as a live in.
const MachineFunction * getParent() const
Return the MachineFunction containing this basic block.
LLVM_ABI instr_iterator erase(instr_iterator I)
Remove an instruction from the instruction list and delete it.
void splice(iterator Where, MachineBasicBlock *Other, iterator From)
Take an instruction from MBB 'Other' at the position From, and insert it into this MBB right before '...
MachineInstrBundleIterator< MachineInstr > iterator
void setMachineBlockAddressTaken()
Set this block to indicate that its address is used as something other than the target of a terminato...
LLVM_ABI bool isLiveIn(MCRegister Reg, LaneBitmask LaneMask=LaneBitmask::getAll()) const
Return true if the specified register is in the live in set.
The MachineFrameInfo class represents an abstract stack frame until prolog/epilog code is inserted.
uint64_t getStackSize() const
Return the number of bytes that must be allocated to hold all of the fixed size frame objects.
void setStackID(int ObjectIdx, uint8_t ID)
bool isCalleeSavedInfoValid() const
Has the callee saved info been calculated yet?
Align getObjectAlign(int ObjectIdx) const
Return the alignment of the specified stack object.
int64_t getObjectSize(int ObjectIdx) const
Return the size of the specified object.
unsigned getNumObjects() const
Return the number of objects.
int64_t getObjectOffset(int ObjectIdx) const
Return the assigned stack offset of the specified object from the incoming stack pointer.
bool isFixedObjectIndex(int ObjectIdx) const
Returns true if the specified index corresponds to a fixed stack object.
unsigned addFrameInst(const MCCFIInstruction &Inst)
const TargetSubtargetInfo & getSubtarget() const
getSubtarget - Return the subtarget for which this machine code is being compiled.
StringRef getName() const
getName - Return the name of the corresponding LLVM function.
MachineMemOperand * getMachineMemOperand(MachinePointerInfo PtrInfo, MachineMemOperand::Flags f, LLT MemTy, Align base_alignment, const AAMDNodes &AAInfo=AAMDNodes(), const MDNode *Ranges=nullptr, SyncScope::ID SSID=SyncScope::System, AtomicOrdering Ordering=AtomicOrdering::NotAtomic, AtomicOrdering FailureOrdering=AtomicOrdering::NotAtomic)
getMachineMemOperand - Allocate a new MachineMemOperand.
MachineFrameInfo & getFrameInfo()
getFrameInfo - Return the frame info object for the current function.
MachineRegisterInfo & getRegInfo()
getRegInfo - Return information about the registers currently in use.
Function & getFunction()
Return the LLVM function that this machine code represents.
BasicBlockListType::iterator iterator
Ty * getInfo()
getInfo - Keep track of various per-function pieces of information for backends that would like to do...
const MachineJumpTableInfo * getJumpTableInfo() const
getJumpTableInfo - Return the jump table info object for the current function.
MachineBasicBlock * CreateMachineBasicBlock(const BasicBlock *BB=nullptr, std::optional< UniqueBBID > BBID=std::nullopt)
CreateMachineInstr - Allocate a new MachineInstr.
void insert(iterator MBBI, MachineBasicBlock *MBB)
const TargetMachine & getTarget() const
getTarget - Return the target machine this machine code is compiled with
const MachineInstrBuilder & setMemRefs(ArrayRef< MachineMemOperand * > MMOs) const
const MachineInstrBuilder & addCFIIndex(unsigned CFIIndex) const
const MachineInstrBuilder & setMIFlag(MachineInstr::MIFlag Flag) const
const MachineInstrBuilder & addImm(int64_t Val) const
Add a new immediate operand.
const MachineInstrBuilder & add(const MachineOperand &MO) const
const MachineInstrBuilder & addSym(MCSymbol *Sym, unsigned char TargetFlags=0) const
const MachineInstrBuilder & addFrameIndex(int Idx) const
const MachineInstrBuilder & addGlobalAddress(const GlobalValue *GV, int64_t Offset=0, unsigned TargetFlags=0) const
const MachineInstrBuilder & addReg(Register RegNo, unsigned flags=0, unsigned SubReg=0) const
Add a new virtual register operand.
const MachineInstrBuilder & addMBB(MachineBasicBlock *MBB, unsigned TargetFlags=0) const
const MachineInstrBuilder & addUse(Register RegNo, unsigned Flags=0, unsigned SubReg=0) const
Add a virtual register use operand.
const MachineInstrBuilder & setMIFlags(unsigned Flags) const
const MachineInstrBuilder & addMemOperand(MachineMemOperand *MMO) const
const MachineInstrBuilder & addDef(Register RegNo, unsigned Flags=0, unsigned SubReg=0) const
Add a virtual register definition operand.
reverse_iterator getReverse() const
Get a reverse iterator to the same node.
Representation of each machine instruction.
unsigned getOpcode() const
Returns the opcode of this MachineInstr.
bool mayLoadOrStore(QueryType Type=AnyInBundle) const
Return true if this instruction could possibly read or modify memory.
const MachineBasicBlock * getParent() const
bool isCall(QueryType Type=AnyInBundle) const
bool getFlag(MIFlag Flag) const
Return whether an MI flag is set.
LLVM_ABI uint32_t mergeFlagsWith(const MachineInstr &Other) const
Return the MIFlags which represent both MachineInstrs.
unsigned getNumOperands() const
Retuns the total number of operands.
LLVM_ABI unsigned getNumExplicitOperands() const
Returns the number of non-implicit operands.
bool mayLoad(QueryType Type=AnyInBundle) const
Return true if this instruction could possibly read memory.
const MCInstrDesc & getDesc() const
Returns the target instruction descriptor of this MachineInstr.
LLVM_ABI bool hasUnmodeledSideEffects() const
Return true if this instruction has side effects that are not modeled by mayLoad / mayStore,...
bool registerDefIsDead(Register Reg, const TargetRegisterInfo *TRI) const
Returns true if the register is dead in this machine instruction.
bool definesRegister(Register Reg, const TargetRegisterInfo *TRI) const
Return true if the MachineInstr fully defines the specified register.
LLVM_ABI void setDesc(const MCInstrDesc &TID)
Replace the instruction descriptor (thus opcode) of the current instruction with a new one.
LLVM_ABI bool hasOrderedMemoryRef() const
Return true if this instruction may have an ordered or volatile memory reference, or if the informati...
LLVM_ABI const MachineFunction * getMF() const
Return the function that contains the basic block that this instruction belongs to.
ArrayRef< MachineMemOperand * > memoperands() const
Access to memory operands of the instruction.
LLVM_ABI bool isLoadFoldBarrier() const
Returns true if it is illegal to fold a load across this instruction.
const DebugLoc & getDebugLoc() const
Returns the debug location id of this MachineInstr.
LLVM_ABI void eraseFromParent()
Unlink 'this' from the containing basic block and delete it.
LLVM_ABI void removeOperand(unsigned OpNo)
Erase an operand from an instruction, leaving it with one fewer operand than it started with.
LLVM_ABI void addRegisterDefined(Register Reg, const TargetRegisterInfo *RegInfo=nullptr)
We have determined MI defines a register.
const MachineOperand & getOperand(unsigned i) const
uint32_t getFlags() const
Return the MI flags bitvector.
LLVM_ABI int findRegisterDefOperandIdx(Register Reg, const TargetRegisterInfo *TRI, bool isDead=false, bool Overlap=false) const
Returns the operand index that is a def of the specified register or -1 if it is not found.
const std::vector< MachineJumpTableEntry > & getJumpTables() const
A description of a memory reference used in the backend.
@ MOLoad
The memory access reads data.
@ MOStore
The memory access writes data.
This class contains meta information specific to a module.
LLVM_ABI MachineFunction * getMachineFunction(const Function &F) const
Returns the MachineFunction associated to IR function F if there is one, otherwise nullptr.
MachineOperand class - Representation of each machine instruction operand.
void setSubReg(unsigned subReg)
unsigned getSubReg() const
void setImm(int64_t immVal)
bool isReg() const
isReg - Tests if this is a MO_Register operand.
MachineBasicBlock * getMBB() const
void setIsDead(bool Val=true)
LLVM_ABI void setReg(Register Reg)
Change the register this operand corresponds to.
bool isImm() const
isImm - Tests if this is a MO_Immediate operand.
void setIsKill(bool Val=true)
MachineInstr * getParent()
getParent - Return the instruction that this operand belongs to.
static MachineOperand CreateImm(int64_t Val)
MachineOperandType getType() const
getType - Returns the MachineOperandType for this operand.
void setIsUndef(bool Val=true)
Register getReg() const
getReg - Returns the register number.
bool isFI() const
isFI - Tests if this is a MO_FrameIndex operand.
LLVM_ABI bool isIdenticalTo(const MachineOperand &Other) const
Returns true if this operand is identical to the specified operand except for liveness related flags ...
static MachineOperand CreateReg(Register Reg, bool isDef, bool isImp=false, bool isKill=false, bool isDead=false, bool isUndef=false, bool isEarlyClobber=false, unsigned SubReg=0, bool isDebug=false, bool isInternalRead=false, bool isRenamable=false)
MachineRegisterInfo - Keep track of information for virtual and physical registers,...
bool tracksLiveness() const
tracksLiveness - Returns true when tracking register liveness accurately.
bool reservedRegsFrozen() const
reservedRegsFrozen - Returns true after freezeReservedRegs() was called to ensure the set of reserved...
const TargetRegisterClass * getRegClassOrNull(Register Reg) const
Return the register class of Reg, or null if Reg has not been assigned a register class yet.
LLVM_ABI const TargetRegisterClass * constrainRegClass(Register Reg, const TargetRegisterClass *RC, unsigned MinNumRegs=0)
constrainRegClass - Constrain the register class of the specified virtual register to be a common sub...
LLVM_ABI MachineInstr * getUniqueVRegDef(Register Reg) const
getUniqueVRegDef - Return the unique machine instr that defines the specified virtual register or nul...
A Module instance is used to store all the information related to an LLVM module.
MI-level patchpoint operands.
uint32_t getNumPatchBytes() const
Return the number of patchable bytes the given patchpoint should emit.
void enterBasicBlockEnd(MachineBasicBlock &MBB)
Start tracking liveness from the end of basic block MBB.
bool isRegUsed(Register Reg, bool includeReserved=true) const
Return if a specific register is currently used.
Register FindUnusedReg(const TargetRegisterClass *RC) const
Find an unused register of the specified register class.
void setRegUsed(Register Reg, LaneBitmask LaneMask=LaneBitmask::getAll())
Tell the scavenger a register is used.
Wrapper class representing virtual and physical registers.
MCRegister asMCReg() const
Utility to check-convert this value to a MCRegister.
constexpr bool isValid() const
constexpr bool isVirtual() const
Return true if the specified register number is in the virtual register namespace.
static constexpr bool isVirtualRegister(unsigned Reg)
Return true if the specified register number is in the virtual register namespace.
constexpr bool isPhysical() const
Return true if the specified register number is in the physical register namespace.
Represents a location in source code.
bool erase(PtrType Ptr)
Remove pointer from the set.
std::pair< iterator, bool > insert(PtrType Ptr)
Inserts Ptr if and only if there is no element in the container equal to Ptr.
SmallPtrSet - This class implements a set which is optimized for holding SmallSize or less elements.
SmallSet - This maintains a set of unique values, optimizing for the case when the set is small (less...
SmallString - A SmallString is just a SmallVector with methods and accessors that make it work better...
void append(StringRef RHS)
Append from a StringRef.
StringRef str() const
Explicit conversion to StringRef.
This class consists of common code factored out of the SmallVector class to reduce code duplication b...
void push_back(const T &Elt)
This is a 'vector' (really, a variable-sized array), optimized for the case when the array is small.
MI-level stackmap operands.
uint32_t getNumPatchBytes() const
Return the number of patchable bytes the given stackmap should emit.
StackOffset holds a fixed and a scalable offset in bytes.
int64_t getFixed() const
Returns the fixed component of the stack.
int64_t getScalable() const
Returns the scalable component of the stack.
static StackOffset get(int64_t Fixed, int64_t Scalable)
static StackOffset getScalable(int64_t Scalable)
static StackOffset getFixed(int64_t Fixed)
MI-level Statepoint operands.
uint32_t getNumPatchBytes() const
Return the number of patchable bytes the given statepoint should emit.
StringRef - Represent a constant reference to a string, i.e.
Object returned by analyzeLoopForPipelining.
TargetInstrInfo - Interface to description of machine instruction set.
virtual void genAlternativeCodeSequence(MachineInstr &Root, unsigned Pattern, SmallVectorImpl< MachineInstr * > &InsInstrs, SmallVectorImpl< MachineInstr * > &DelInstrs, DenseMap< Register, unsigned > &InstIdxForVirtReg) const
When getMachineCombinerPatterns() finds patterns, this function generates the instructions that could...
virtual std::optional< ParamLoadedValue > describeLoadedValue(const MachineInstr &MI, Register Reg) const
Produce the expression describing the MI loading a value into the physical register Reg.
virtual bool getMachineCombinerPatterns(MachineInstr &Root, SmallVectorImpl< unsigned > &Patterns, bool DoRegPressureReduce) const
Return true when there is potentially a faster code sequence for an instruction chain ending in Root.
virtual bool isSchedulingBoundary(const MachineInstr &MI, const MachineBasicBlock *MBB, const MachineFunction &MF) const
Test if the given instruction should be considered a scheduling boundary.
virtual CombinerObjective getCombinerObjective(unsigned Pattern) const
Return the objective of a combiner pattern.
virtual bool isFunctionSafeToSplit(const MachineFunction &MF) const
Return true if the function is a viable candidate for machine function splitting.
const MCAsmInfo * getMCAsmInfo() const
Return target specific asm information.
bool contains(Register Reg) const
Return true if the specified register is included in this register class.
bool hasSubClassEq(const TargetRegisterClass *RC) const
Returns true if RC is a sub-class of or equal to this class.
bool hasSuperClassEq(const TargetRegisterClass *RC) const
Returns true if RC is a super-class of or equal to this class.
TargetRegisterInfo base class - We assume that the target defines a static array of TargetRegisterDes...
TargetSubtargetInfo - Generic base class for all target subtargets.
virtual const TargetInstrInfo * getInstrInfo() const
virtual const TargetRegisterInfo * getRegisterInfo() const =0
Return the target's register information.
Target - Wrapper for Target specific information.
static constexpr TypeSize getFixed(ScalarTy ExactSize)
static constexpr TypeSize getScalable(ScalarTy MinimumSize)
Value * getOperand(unsigned i) const
constexpr bool isScalable() const
Returns whether the quantity is scaled by a runtime quantity (vscale).
constexpr ScalarTy getKnownMinValue() const
Returns the minimum value this quantity can represent.
self_iterator getIterator()
A raw_ostream that writes to an std::string.
#define llvm_unreachable(msg)
Marks that the current location is not supposed to be reachable.
static CondCode getInvertedCondCode(CondCode Code)
@ MO_DLLIMPORT
MO_DLLIMPORT - On a symbol operand, this represents that the reference to the symbol is for an import...
@ MO_NC
MO_NC - Indicates whether the linker is expected to check the symbol reference for overflow.
@ MO_G1
MO_G1 - A symbol operand with this flag (granule 1) represents the bits 16-31 of a 64-bit address,...
@ MO_S
MO_S - Indicates that the bits of the symbol operand represented by MO_G0 etc are signed.
@ MO_PAGEOFF
MO_PAGEOFF - A symbol operand with this flag represents the offset of that symbol within a 4K page.
@ MO_GOT
MO_GOT - This flag indicates that a symbol operand represents the address of the GOT entry for the sy...
@ MO_PREL
MO_PREL - Indicates that the bits of the symbol operand represented by MO_G0 etc are PC relative.
@ MO_G0
MO_G0 - A symbol operand with this flag (granule 0) represents the bits 0-15 of a 64-bit address,...
@ MO_ARM64EC_CALLMANGLE
MO_ARM64EC_CALLMANGLE - Operand refers to the Arm64EC-mangled version of a symbol,...
@ MO_PAGE
MO_PAGE - A symbol operand with this flag represents the pc-relative offset of the 4K page containing...
@ MO_HI12
MO_HI12 - This flag indicates that a symbol operand represents the bits 13-24 of a 64-bit address,...
@ MO_TLS
MO_TLS - Indicates that the operand being accessed is some kind of thread-local symbol.
@ MO_G2
MO_G2 - A symbol operand with this flag (granule 2) represents the bits 32-47 of a 64-bit address,...
@ MO_TAGGED
MO_TAGGED - With MO_PAGE, indicates that the page includes a memory tag in bits 56-63.
@ MO_G3
MO_G3 - A symbol operand with this flag (granule 3) represents the high 16-bits of a 64-bit address,...
@ MO_COFFSTUB
MO_COFFSTUB - On a symbol operand "FOO", this indicates that the reference is actually to the "....
unsigned getCheckerSizeInBytes(AuthCheckMethod Method)
Returns the number of bytes added by checkAuthenticatedRegister.
static uint64_t decodeLogicalImmediate(uint64_t val, unsigned regSize)
decodeLogicalImmediate - Decode a logical immediate value in the form "N:immr:imms" (where the immr a...
static unsigned getShiftValue(unsigned Imm)
getShiftValue - Extract the shift value.
static unsigned getArithExtendImm(AArch64_AM::ShiftExtendType ET, unsigned Imm)
getArithExtendImm - Encode the extend type and shift amount for an arithmetic instruction: imm: 3-bit...
static unsigned getArithShiftValue(unsigned Imm)
getArithShiftValue - get the arithmetic shift value.
static uint64_t encodeLogicalImmediate(uint64_t imm, unsigned regSize)
encodeLogicalImmediate - Return the encoded immediate value for a logical immediate instruction of th...
static AArch64_AM::ShiftExtendType getArithExtendType(unsigned Imm)
static AArch64_AM::ShiftExtendType getShiftType(unsigned Imm)
getShiftType - Extract the shift type.
static unsigned getShifterImm(AArch64_AM::ShiftExtendType ST, unsigned Imm)
getShifterImm - Encode the shift type and amount: imm: 6-bit shift amount shifter: 000 ==> lsl 001 ==...
void expandMOVImm(uint64_t Imm, unsigned BitSize, SmallVectorImpl< ImmInsnModel > &Insn)
Expand a MOVi32imm or MOVi64imm pseudo instruction to one or more real move-immediate instructions to...
static const uint64_t InstrFlagIsWhile
static const uint64_t InstrFlagIsPTestLike
constexpr std::underlying_type_t< E > Mask()
Get a bitmask with 1s in all places up to the high-order bit of E's largest value.
@ C
The default llvm calling convention, compatible with C.
@ Implicit
Not emitted register (e.g. carry, or temporary result).
@ Renamable
Register that may be renamed.
@ Define
Register definition.
@ Kill
The last use of a register.
@ Undef
Value of the register doesn't matter.
initializer< Ty > init(const Ty &Val)
InstrType
Represents how an instruction should be mapped by the outliner.
LLVM_ABI Instruction & back() const
This is an optimization pass for GlobalISel generic memory operations.
bool all_of(R &&range, UnaryPredicate P)
Provide wrappers to std::all_of which take ranges instead of having to pass begin/end explicitly.
static bool isCondBranchOpcode(int Opc)
MCCFIInstruction createDefCFA(const TargetRegisterInfo &TRI, unsigned FrameReg, unsigned Reg, const StackOffset &Offset, bool LastAdjustmentWasScalable=true)
static bool isPTrueOpcode(unsigned Opc)
MachineInstrBuilder BuildMI(MachineFunction &MF, const MIMetadata &MIMD, const MCInstrDesc &MCID)
Builder interface. Specify how to create the initial instruction itself.
bool succeeded(LogicalResult Result)
Utility function that returns true if the provided LogicalResult corresponds to a success value.
int isAArch64FrameOffsetLegal(const MachineInstr &MI, StackOffset &Offset, bool *OutUseUnscaledOp=nullptr, unsigned *OutUnscaledOp=nullptr, int64_t *EmittableOffset=nullptr)
Check if the Offset is a valid frame offset for MI.
constexpr bool isInt(int64_t x)
Checks if an integer fits into the given bit width.
auto enumerate(FirstRange &&First, RestRanges &&...Rest)
Given two or more input ranges, returns a new range whose values are tuples (A, B,...
decltype(auto) dyn_cast(const From &Val)
dyn_cast<X> - Return the argument parameter cast to the specified type.
static bool isIndirectBranchOpcode(int Opc)
iterator_range< T > make_range(T x, T y)
Convenience function for iterating over sub-ranges.
unsigned getBLRCallOpcode(const MachineFunction &MF)
Return opcode to be used for indirect calls.
@ AArch64FrameOffsetIsLegal
Offset is legal.
@ AArch64FrameOffsetCanUpdate
Offset can apply, at least partly.
@ AArch64FrameOffsetCannotUpdate
Offset cannot apply.
constexpr bool isPowerOf2_64(uint64_t Value)
Return true if the argument is a power of two > 0 (64 bit edition.)
unsigned Log2_64(uint64_t Value)
Return the floor log base 2 of the specified value, -1 if the value is zero.
static bool isSEHInstruction(const MachineInstr &MI)
bool any_of(R &&range, UnaryPredicate P)
Provide wrappers to std::any_of which take ranges instead of having to pass begin/end explicitly.
auto reverse(ContainerTy &&C)
MachineInstr * getImm(const MachineOperand &MO, const MachineRegisterInfo *MRI)
decltype(auto) get(const PointerIntPair< PointerTy, IntBits, IntType, PtrTraits, Info > &Pair)
void sort(IteratorTy Start, IteratorTy End)
AArch64MachineCombinerPattern
@ MULSUBv2i32_indexed_OP1
@ MULADDv4i16_indexed_OP2
@ MULSUBv8i16_indexed_OP2
@ MULSUBv4i16_indexed_OP2
@ MULSUBv4i32_indexed_OP2
@ MULADDv2i32_indexed_OP1
@ MULADDv4i32_indexed_OP1
@ MULADDv2i32_indexed_OP2
@ MULSUBv4i16_indexed_OP1
@ MULADDv4i32_indexed_OP2
@ MULSUBv8i16_indexed_OP1
@ MULSUBv2i32_indexed_OP2
@ MULADDv8i16_indexed_OP1
@ MULSUBv4i32_indexed_OP1
@ MULADDv8i16_indexed_OP2
@ MULADDv4i16_indexed_OP1
void emitFrameOffset(MachineBasicBlock &MBB, MachineBasicBlock::iterator MBBI, const DebugLoc &DL, unsigned DestReg, unsigned SrcReg, StackOffset Offset, const TargetInstrInfo *TII, MachineInstr::MIFlag=MachineInstr::NoFlags, bool SetNZCV=false, bool NeedsWinCFI=false, bool *HasWinCFI=nullptr, bool EmitCFAOffset=false, StackOffset InitialOffset={}, unsigned FrameReg=AArch64::SP)
emitFrameOffset - Emit instructions as needed to set DestReg to SrcReg plus Offset.
LLVM_ABI void report_fatal_error(Error Err, bool gen_crash_diag=true)
CombinerObjective
The combiner's goal may differ based on which pattern it is attempting to optimize.
std::optional< UsedNZCV > examineCFlagsUse(MachineInstr &MI, MachineInstr &CmpInstr, const TargetRegisterInfo &TRI, SmallVectorImpl< MachineInstr * > *CCUseInstrs=nullptr)
CodeGenOptLevel
Code generation optimization level.
class LLVM_GSL_OWNER SmallVector
Forward declaration of SmallVector so that calculateSmallVectorDefaultInlinedElements can reference s...
auto instructionsWithoutDebug(IterT It, IterT End, bool SkipPseudoOp=true)
Construct a range iterator which begins at It and moves forwards until End is reached,...
LLVM_ABI raw_fd_ostream & errs()
This returns a reference to a raw_ostream for standard error.
auto drop_end(T &&RangeOrContainer, size_t N=1)
Return a range covering RangeOrContainer with the last N elements excluded.
unsigned getUndefRegState(bool B)
static MCRegister getXRegFromWReg(MCRegister Reg)
unsigned getDefRegState(bool B)
MCCFIInstruction createCFAOffset(const TargetRegisterInfo &MRI, unsigned Reg, const StackOffset &OffsetFromDefCFA, std::optional< int64_t > IncomingVGOffsetFromDefCFA)
unsigned getKillRegState(bool B)
FunctionAddr VTableAddr Next
DWARFExpression::Operation Op
ArrayRef(const T &OneElt) -> ArrayRef< T >
static bool isUncondBranchOpcode(int Opc)
decltype(auto) cast(const From &Val)
cast<X> - Return the argument parameter cast to the specified type.
void erase_if(Container &C, UnaryPredicate P)
Provide a container algorithm similar to C++ Library Fundamentals v2's erase_if which is equivalent t...
constexpr bool isIntN(unsigned N, int64_t x)
Checks if an signed integer fits into the given (dynamic) bit width.
bool rewriteAArch64FrameIndex(MachineInstr &MI, unsigned FrameRegIdx, unsigned FrameReg, StackOffset &Offset, const AArch64InstrInfo *TII)
rewriteAArch64FrameIndex - Rewrite MI to access 'Offset' bytes from the FP.
bool is_contained(R &&Range, const E &Element)
Returns true if Element is found in Range.
static const MachineMemOperand::Flags MOSuppressPair
constexpr int64_t SignExtend64(uint64_t x)
Sign-extend the number in the bottom B bits of X to a 64-bit integer.
void appendLEB128(SmallVectorImpl< U > &Buffer, T Value)
std::pair< MachineOperand, DIExpression * > ParamLoadedValue
bool isNZCVTouchedInInstructionRange(const MachineInstr &DefMI, const MachineInstr &UseMI, const TargetRegisterInfo *TRI)
Return true if there is an instruction /after/ DefMI and before UseMI which either reads or clobbers ...
auto seq(T Begin, T End)
Iterate over an integral type from Begin up to - but not including - End.
static const MachineMemOperand::Flags MOStridedAccess
void fullyRecomputeLiveIns(ArrayRef< MachineBasicBlock * > MBBs)
Convenience function for recomputing live-in's for a set of MBBs until the computation converges.
LLVM_ABI Printable printReg(Register Reg, const TargetRegisterInfo *TRI=nullptr, unsigned SubIdx=0, const MachineRegisterInfo *MRI=nullptr)
Prints virtual and physical registers with or without a TRI instance.
Used to describe addressing mode similar to ExtAddrMode in CodeGenPrepare.
LLVM_ABI static const MBBSectionID ColdSectionID
This class contains a discriminated union of information about pointers in memory operands,...
static LLVM_ABI MachinePointerInfo getFixedStack(MachineFunction &MF, int FI, int64_t Offset=0)
Return a MachinePointerInfo record that refers to the specified FrameIndex.
An individual sequence of instructions to be replaced with a call to an outlined function.
MachineFunction * getMF() const
The information necessary to create an outlined function for some class of candidate.