66#define GET_INSTRINFO_CTOR_DTOR
67#include "AArch64GenInstrInfo.inc"
69#define DEBUG_TYPE "AArch64InstrInfo"
71STATISTIC(NumCopyInstrs,
"Number of COPY instructions expanded");
72STATISTIC(NumZCRegMoveInstrsGPR,
"Number of zero-cycle GPR register move "
73 "instructions expanded from canonical COPY");
74STATISTIC(NumZCRegMoveInstrsFPR,
"Number of zero-cycle FPR register move "
75 "instructions expanded from canonical COPY");
76STATISTIC(NumZCZeroingInstrsGPR,
"Number of zero-cycle GPR zeroing "
77 "instructions expanded from canonical COPY");
82 cl::desc(
"Restrict range of CB instructions (DEBUG)"));
86 cl::desc(
"Restrict range of TB[N]Z instructions (DEBUG)"));
90 cl::desc(
"Restrict range of CB[N]Z instructions (DEBUG)"));
94 cl::desc(
"Restrict range of Bcc instructions (DEBUG)"));
98 cl::desc(
"Restrict range of B instructions (DEBUG)"));
102 cl::desc(
"Restrict range of instructions to search for the "
103 "machine-combiner gather pattern optimization"));
108 RI(STI.getTargetTriple(), STI.getHwMode()), Subtarget(STI) {}
119 auto Op =
MI.getOpcode();
120 if (
Op == AArch64::INLINEASM ||
Op == AArch64::INLINEASM_BR)
121 return getInlineAsmLength(
MI.getOperand(0).getSymbolName(), *MAI);
125 if (
MI.isMetaInstruction())
130 unsigned NumBytes = 0;
134 NumBytes =
Desc.getSize() ?
Desc.getSize() : 4;
137 if (!MFI->shouldSignReturnAddress(*MF))
141 auto Method = STI.getAuthenticatedLRCheckMethod(*MF);
149 switch (
Desc.getOpcode()) {
152 return Desc.getSize();
159 case TargetOpcode::STACKMAP:
162 assert(NumBytes % 4 == 0 &&
"Invalid number of NOP bytes requested!");
164 case TargetOpcode::PATCHPOINT:
167 assert(NumBytes % 4 == 0 &&
"Invalid number of NOP bytes requested!");
169 case TargetOpcode::STATEPOINT:
171 assert(NumBytes % 4 == 0 &&
"Invalid number of NOP bytes requested!");
176 case TargetOpcode::PATCHABLE_FUNCTION_ENTER:
181 F.getFnAttributeAsParsedInteger(
"patchable-function-entry", 9) * 4;
183 case TargetOpcode::PATCHABLE_FUNCTION_EXIT:
184 case TargetOpcode::PATCHABLE_TAIL_CALL:
185 case TargetOpcode::PATCHABLE_TYPED_EVENT_CALL:
189 case TargetOpcode::PATCHABLE_EVENT_CALL:
195 NumBytes =
MI.getOperand(1).getImm();
197 case TargetOpcode::BUNDLE:
198 NumBytes = getInstBundleLength(
MI);
205unsigned AArch64InstrInfo::getInstBundleLength(
const MachineInstr &
MI)
const {
209 while (++
I != E &&
I->isInsideBundle()) {
210 assert(!
I->isBundle() &&
"No nested bundle!");
245 case AArch64::CBWPri:
246 case AArch64::CBXPri:
247 case AArch64::CBWPrr:
248 case AArch64::CBXPrr:
256 case AArch64::CBBAssertExt:
257 case AArch64::CBHAssertExt:
288 case AArch64::CBWPri:
289 case AArch64::CBXPri:
290 case AArch64::CBBAssertExt:
291 case AArch64::CBHAssertExt:
292 case AArch64::CBWPrr:
293 case AArch64::CBXPrr:
299 int64_t BrOffset)
const {
301 assert(Bits >= 3 &&
"max branch displacement must be enough to jump"
302 "over conditional branch expansion");
303 return isIntN(Bits, BrOffset / 4);
308 switch (
MI.getOpcode()) {
312 return MI.getOperand(0).getMBB();
317 return MI.getOperand(2).getMBB();
323 return MI.getOperand(1).getMBB();
324 case AArch64::CBWPri:
325 case AArch64::CBXPri:
326 case AArch64::CBBAssertExt:
327 case AArch64::CBHAssertExt:
328 case AArch64::CBWPrr:
329 case AArch64::CBXPrr:
330 return MI.getOperand(3).getMBB();
340 assert(RS &&
"RegScavenger required for long branching");
342 "new block should be inserted for expanding unconditional branch");
345 "restore block should be inserted for restoring clobbered registers");
352 "Branch offsets outside of the signed 33-bit range not supported");
363 RS->enterBasicBlockEnd(
MBB);
366 constexpr Register Reg = AArch64::X16;
367 if (!RS->isRegUsed(Reg)) {
368 insertUnconditionalBranch(
MBB, &NewDestBB,
DL);
375 Register Scavenged = RS->FindUnusedReg(&AArch64::GPR64RegClass);
376 if (Scavenged != AArch64::NoRegister &&
378 buildIndirectBranch(Scavenged, NewDestBB);
379 RS->setRegUsed(Scavenged);
388 "Unable to insert indirect branch inside function that has red zone");
411 bool AllowModify)
const {
418 if (
I->getOpcode() == AArch64::SpeculationBarrierISBDSBEndBB ||
419 I->getOpcode() == AArch64::SpeculationBarrierSBEndBB) {
423 if (!isUnpredicatedTerminator(*
I))
430 unsigned LastOpc = LastInst->
getOpcode();
431 if (
I ==
MBB.begin() || !isUnpredicatedTerminator(*--
I)) {
446 unsigned SecondLastOpc = SecondLastInst->
getOpcode();
453 LastInst = SecondLastInst;
455 if (
I ==
MBB.begin() || !isUnpredicatedTerminator(*--
I)) {
460 SecondLastInst = &*
I;
461 SecondLastOpc = SecondLastInst->
getOpcode();
472 LastInst = SecondLastInst;
474 if (
I ==
MBB.begin() || !isUnpredicatedTerminator(*--
I)) {
476 "unreachable unconditional branches removed above");
485 SecondLastInst = &*
I;
486 SecondLastOpc = SecondLastInst->
getOpcode();
490 if (SecondLastInst &&
I !=
MBB.begin() && isUnpredicatedTerminator(*--
I))
506 I->eraseFromParent();
515 I->eraseFromParent();
524 MachineBranchPredicate &MBP,
525 bool AllowModify)
const {
535 if (
I->getOpcode() == AArch64::SpeculationBarrierISBDSBEndBB ||
536 I->getOpcode() == AArch64::SpeculationBarrierSBEndBB) {
540 if (!isUnpredicatedTerminator(*
I))
545 unsigned LastOpc = LastInst->
getOpcode();
560 assert(MBP.TrueDest &&
"expected!");
561 MBP.FalseDest =
MBB.getNextNode();
563 MBP.ConditionDef =
nullptr;
564 MBP.SingleUseCondition =
false;
568 MBP.Predicate = (LastOpc == AArch64::CBNZX || LastOpc == AArch64::CBNZW)
569 ? MachineBranchPredicate::PRED_NE
570 : MachineBranchPredicate::PRED_EQ;
586 Cond[1].setImm(AArch64::CBNZW);
589 Cond[1].setImm(AArch64::CBZW);
592 Cond[1].setImm(AArch64::CBNZX);
595 Cond[1].setImm(AArch64::CBZX);
598 Cond[1].setImm(AArch64::TBNZW);
601 Cond[1].setImm(AArch64::TBZW);
604 Cond[1].setImm(AArch64::TBNZX);
607 Cond[1].setImm(AArch64::TBZX);
611 case AArch64::CBWPri:
612 case AArch64::CBXPri:
613 case AArch64::CBBAssertExt:
614 case AArch64::CBHAssertExt:
615 case AArch64::CBWPrr:
616 case AArch64::CBXPrr: {
629 int *BytesRemoved)
const {
639 I->eraseFromParent();
643 if (
I ==
MBB.begin()) {
656 I->eraseFromParent();
663void AArch64InstrInfo::instantiateCondBranch(
688 if (
Cond.size() > 5) {
699 assert(
TBB &&
"insertBranch must not be told to insert a fallthrough");
726 unsigned Opc =
MI.getOpcode();
733 if (
MI.getOperand(0).getReg() == AArch64::WZR ||
734 MI.getOperand(0).getReg() == AArch64::XZR) {
736 dbgs() <<
"Removing always taken branch: " <<
MI);
739 for (
auto *S : Succs)
741 MBB->removeSuccessor(S);
743 while (
MBB->rbegin() != &
MI)
744 MBB->rbegin()->eraseFromParent();
745 MI.eraseFromParent();
755 if (
MI.getOperand(0).getReg() == AArch64::WZR ||
756 MI.getOperand(0).getReg() == AArch64::XZR) {
758 dbgs() <<
"Removing never taken branch: " <<
MI);
760 MI.getParent()->removeSuccessor(
Target);
761 MI.eraseFromParent();
774 if (!
DefMI->isFullCopy())
776 VReg =
DefMI->getOperand(1).getReg();
785 unsigned *NewReg =
nullptr) {
790 bool Is64Bit = AArch64::GPR64allRegClass.hasSubClassEq(
MRI.getRegClass(VReg));
794 switch (
DefMI->getOpcode()) {
795 case AArch64::SUBREG_TO_REG:
799 if (!
DefMI->getOperand(1).isImm() ||
DefMI->getOperand(1).getImm() != 0)
801 if (!
DefMI->getOperand(2).isReg())
803 if (!
DefMI->getOperand(3).isImm() ||
804 DefMI->getOperand(3).getImm() != AArch64::sub_32)
807 if (
DefMI->getOpcode() != AArch64::MOVi32imm)
809 if (!
DefMI->getOperand(1).isImm() ||
DefMI->getOperand(1).getImm() != 1)
812 SrcReg = AArch64::XZR;
813 Opc = AArch64::CSINCXr;
816 case AArch64::MOVi32imm:
817 case AArch64::MOVi64imm:
818 if (!
DefMI->getOperand(1).isImm() ||
DefMI->getOperand(1).getImm() != 1)
820 SrcReg = Is64Bit ? AArch64::XZR : AArch64::WZR;
821 Opc = Is64Bit ? AArch64::CSINCXr : AArch64::CSINCWr;
824 case AArch64::ADDSXri:
825 case AArch64::ADDSWri:
827 if (
DefMI->findRegisterDefOperandIdx(AArch64::NZCV,
nullptr,
832 case AArch64::ADDXri:
833 case AArch64::ADDWri:
835 if (!
DefMI->getOperand(2).isImm() ||
DefMI->getOperand(2).getImm() != 1 ||
836 DefMI->getOperand(3).getImm() != 0)
838 SrcReg =
DefMI->getOperand(1).getReg();
839 Opc = Is64Bit ? AArch64::CSINCXr : AArch64::CSINCWr;
842 case AArch64::ORNXrr:
843 case AArch64::ORNWrr: {
846 if (ZReg != AArch64::XZR && ZReg != AArch64::WZR)
848 SrcReg =
DefMI->getOperand(2).getReg();
849 Opc = Is64Bit ? AArch64::CSINVXr : AArch64::CSINVWr;
853 case AArch64::SUBSXrr:
854 case AArch64::SUBSWrr:
856 if (
DefMI->findRegisterDefOperandIdx(AArch64::NZCV,
nullptr,
861 case AArch64::SUBXrr:
862 case AArch64::SUBWrr: {
865 if (ZReg != AArch64::XZR && ZReg != AArch64::WZR)
867 SrcReg =
DefMI->getOperand(2).getReg();
868 Opc = Is64Bit ? AArch64::CSNEGXr : AArch64::CSNEGWr;
874 assert(
Opc && SrcReg &&
"Missing parameters");
886 int &FalseCycles)
const {
890 RI.getCommonSubClass(
MRI.getRegClass(TrueReg),
MRI.getRegClass(FalseReg));
897 if (!RI.getCommonSubClass(RC,
MRI.getRegClass(DstReg)))
901 unsigned ExtraCondLat =
Cond.size() != 1;
905 if (AArch64::GPR64allRegClass.hasSubClassEq(RC) ||
906 AArch64::GPR32allRegClass.hasSubClassEq(RC)) {
908 CondCycles = 1 + ExtraCondLat;
909 TrueCycles = FalseCycles = 1;
919 if (AArch64::FPR64RegClass.hasSubClassEq(RC) ||
920 AArch64::FPR32RegClass.hasSubClassEq(RC)) {
921 CondCycles = 5 + ExtraCondLat;
922 TrueCycles = FalseCycles = 2;
939 switch (
Cond.size()) {
971 MRI.constrainRegClass(SrcReg, &AArch64::GPR64spRegClass);
977 MRI.constrainRegClass(SrcReg, &AArch64::GPR32spRegClass);
1017 unsigned SubsOpc, SubsDestReg;
1023 case AArch64::CBWPri:
1024 SubsOpc = AArch64::SUBSWri;
1025 SubsDestReg = AArch64::WZR;
1028 case AArch64::CBXPri:
1029 SubsOpc = AArch64::SUBSXri;
1030 SubsDestReg = AArch64::XZR;
1033 case AArch64::CBWPrr:
1034 SubsOpc = AArch64::SUBSWrr;
1035 SubsDestReg = AArch64::WZR;
1038 case AArch64::CBXPrr:
1039 SubsOpc = AArch64::SUBSXrr;
1040 SubsDestReg = AArch64::XZR;
1069 switch (ExtendType) {
1075 "Unexpected compare-and-branch instruction for SXTB shift-extend");
1076 ExtOpc = AArch64::SBFMWri;
1082 "Unexpected compare-and-branch instruction for SXTH shift-extend");
1083 ExtOpc = AArch64::SBFMWri;
1089 "Unexpected compare-and-branch instruction for UXTB shift-extend");
1090 ExtOpc = AArch64::ANDWri;
1096 "Unexpected compare-and-branch instruction for UXTH shift-extend");
1097 ExtOpc = AArch64::ANDWri;
1103 Reg =
MRI.createVirtualRegister(&AArch64::GPR32spRegClass);
1106 if (ExtOpc != AArch64::ANDWri)
1108 MBBI.addImm(ExtBits);
1115 MRI.constrainRegClass(Reg,
MRI.getRegClass(
Cond[3].getReg()));
1116 MRI.constrainRegClass(
Cond[3].
getReg(), &AArch64::GPR32spRegClass);
1123 MRI.constrainRegClass(Reg,
MRI.getRegClass(
Cond[3].getReg()));
1124 MRI.constrainRegClass(
Cond[3].
getReg(), &AArch64::GPR32spRegClass);
1136 bool TryFold =
false;
1137 if (
MRI.constrainRegClass(DstReg, &AArch64::GPR64RegClass)) {
1138 RC = &AArch64::GPR64RegClass;
1139 Opc = AArch64::CSELXr;
1141 }
else if (
MRI.constrainRegClass(DstReg, &AArch64::GPR32RegClass)) {
1142 RC = &AArch64::GPR32RegClass;
1143 Opc = AArch64::CSELWr;
1145 }
else if (
MRI.constrainRegClass(DstReg, &AArch64::FPR64RegClass)) {
1146 RC = &AArch64::FPR64RegClass;
1147 Opc = AArch64::FCSELDrrr;
1148 }
else if (
MRI.constrainRegClass(DstReg, &AArch64::FPR32RegClass)) {
1149 RC = &AArch64::FPR32RegClass;
1150 Opc = AArch64::FCSELSrrr;
1152 assert(RC &&
"Unsupported regclass");
1156 unsigned NewReg = 0;
1171 MRI.clearKillFlags(NewReg);
1176 MRI.constrainRegClass(TrueReg, RC);
1179 (FalseReg.
isVirtual() || FalseReg == AArch64::WZR ||
1180 FalseReg == AArch64::XZR) &&
1181 "FalseReg was folded into a non-virtual register other than WZR or XZR");
1183 MRI.constrainRegClass(FalseReg, RC);
1198 assert(BitSize == 64 &&
"Only bit sizes of 32 or 64 allowed");
1203 return Is.
size() <= 2;
1208 assert(
MI.isCopy() &&
"Expected COPY instruction");
1214 if (
Reg.isVirtual())
1215 return MRI.getRegClass(
Reg);
1216 if (
Reg.isPhysical())
1217 return RI.getMinimalPhysRegClass(
Reg);
1222 if (DstRC && SrcRC && !RI.getCommonSubClass(DstRC, SrcRC))
1225 return MI.isAsCheapAsAMove();
1231 if (Subtarget.hasExynosCheapAsMoveHandling()) {
1232 if (isExynosCheapAsMove(
MI))
1234 return MI.isAsCheapAsAMove();
1237 switch (
MI.getOpcode()) {
1239 return MI.isAsCheapAsAMove();
1241 case TargetOpcode::COPY:
1244 case AArch64::ADDWrs:
1245 case AArch64::ADDXrs:
1246 case AArch64::SUBWrs:
1247 case AArch64::SUBXrs:
1248 return Subtarget.hasALULSLFast() &&
MI.getOperand(3).getImm() <= 4;
1253 case AArch64::MOVi32imm:
1255 case AArch64::MOVi64imm:
1260bool AArch64InstrInfo::isFalkorShiftExtFast(
const MachineInstr &
MI) {
1261 switch (
MI.getOpcode()) {
1265 case AArch64::ADDWrs:
1266 case AArch64::ADDXrs:
1267 case AArch64::ADDSWrs:
1268 case AArch64::ADDSXrs: {
1269 unsigned Imm =
MI.getOperand(3).getImm();
1276 case AArch64::ADDWrx:
1277 case AArch64::ADDXrx:
1278 case AArch64::ADDXrx64:
1279 case AArch64::ADDSWrx:
1280 case AArch64::ADDSXrx:
1281 case AArch64::ADDSXrx64: {
1282 unsigned Imm =
MI.getOperand(3).getImm();
1294 case AArch64::SUBWrs:
1295 case AArch64::SUBSWrs: {
1296 unsigned Imm =
MI.getOperand(3).getImm();
1298 return ShiftVal == 0 ||
1302 case AArch64::SUBXrs:
1303 case AArch64::SUBSXrs: {
1304 unsigned Imm =
MI.getOperand(3).getImm();
1306 return ShiftVal == 0 ||
1310 case AArch64::SUBWrx:
1311 case AArch64::SUBXrx:
1312 case AArch64::SUBXrx64:
1313 case AArch64::SUBSWrx:
1314 case AArch64::SUBSXrx:
1315 case AArch64::SUBSXrx64: {
1316 unsigned Imm =
MI.getOperand(3).getImm();
1328 case AArch64::LDRBBroW:
1329 case AArch64::LDRBBroX:
1330 case AArch64::LDRBroW:
1331 case AArch64::LDRBroX:
1332 case AArch64::LDRDroW:
1333 case AArch64::LDRDroX:
1334 case AArch64::LDRHHroW:
1335 case AArch64::LDRHHroX:
1336 case AArch64::LDRHroW:
1337 case AArch64::LDRHroX:
1338 case AArch64::LDRQroW:
1339 case AArch64::LDRQroX:
1340 case AArch64::LDRSBWroW:
1341 case AArch64::LDRSBWroX:
1342 case AArch64::LDRSBXroW:
1343 case AArch64::LDRSBXroX:
1344 case AArch64::LDRSHWroW:
1345 case AArch64::LDRSHWroX:
1346 case AArch64::LDRSHXroW:
1347 case AArch64::LDRSHXroX:
1348 case AArch64::LDRSWroW:
1349 case AArch64::LDRSWroX:
1350 case AArch64::LDRSroW:
1351 case AArch64::LDRSroX:
1352 case AArch64::LDRWroW:
1353 case AArch64::LDRWroX:
1354 case AArch64::LDRXroW:
1355 case AArch64::LDRXroX:
1356 case AArch64::PRFMroW:
1357 case AArch64::PRFMroX:
1358 case AArch64::STRBBroW:
1359 case AArch64::STRBBroX:
1360 case AArch64::STRBroW:
1361 case AArch64::STRBroX:
1362 case AArch64::STRDroW:
1363 case AArch64::STRDroX:
1364 case AArch64::STRHHroW:
1365 case AArch64::STRHHroX:
1366 case AArch64::STRHroW:
1367 case AArch64::STRHroX:
1368 case AArch64::STRQroW:
1369 case AArch64::STRQroX:
1370 case AArch64::STRSroW:
1371 case AArch64::STRSroX:
1372 case AArch64::STRWroW:
1373 case AArch64::STRWroX:
1374 case AArch64::STRXroW:
1375 case AArch64::STRXroX: {
1376 unsigned IsSigned =
MI.getOperand(3).getImm();
1383 unsigned Opc =
MI.getOpcode();
1387 case AArch64::SEH_StackAlloc:
1388 case AArch64::SEH_SaveFPLR:
1389 case AArch64::SEH_SaveFPLR_X:
1390 case AArch64::SEH_SaveReg:
1391 case AArch64::SEH_SaveReg_X:
1392 case AArch64::SEH_SaveRegP:
1393 case AArch64::SEH_SaveRegP_X:
1394 case AArch64::SEH_SaveFReg:
1395 case AArch64::SEH_SaveFReg_X:
1396 case AArch64::SEH_SaveFRegP:
1397 case AArch64::SEH_SaveFRegP_X:
1398 case AArch64::SEH_SetFP:
1399 case AArch64::SEH_AddFP:
1400 case AArch64::SEH_Nop:
1401 case AArch64::SEH_PrologEnd:
1402 case AArch64::SEH_EpilogStart:
1403 case AArch64::SEH_EpilogEnd:
1404 case AArch64::SEH_PACSignLR:
1405 case AArch64::SEH_SaveAnyRegI:
1406 case AArch64::SEH_SaveAnyRegIP:
1407 case AArch64::SEH_SaveAnyRegQP:
1408 case AArch64::SEH_SaveAnyRegQPX:
1409 case AArch64::SEH_AllocZ:
1410 case AArch64::SEH_SaveZReg:
1411 case AArch64::SEH_SavePReg:
1418 unsigned &SubIdx)
const {
1419 switch (
MI.getOpcode()) {
1422 case AArch64::SBFMXri:
1423 case AArch64::UBFMXri:
1426 if (
MI.getOperand(2).getImm() != 0 ||
MI.getOperand(3).getImm() != 31)
1429 SrcReg =
MI.getOperand(1).getReg();
1430 DstReg =
MI.getOperand(0).getReg();
1431 SubIdx = AArch64::sub_32;
1440 int64_t OffsetA = 0, OffsetB = 0;
1441 TypeSize WidthA(0,
false), WidthB(0,
false);
1442 bool OffsetAIsScalable =
false, OffsetBIsScalable =
false;
1463 OffsetAIsScalable == OffsetBIsScalable) {
1464 int LowOffset = OffsetA < OffsetB ? OffsetA : OffsetB;
1465 int HighOffset = OffsetA < OffsetB ? OffsetB : OffsetA;
1466 TypeSize LowWidth = (LowOffset == OffsetA) ? WidthA : WidthB;
1467 if (LowWidth.
isScalable() == OffsetAIsScalable &&
1485 switch (
MI.getOpcode()) {
1488 if (
MI.getOperand(0).getImm() == 0x14)
1495 case AArch64::MSRpstatesvcrImm1:
1502 auto Next = std::next(
MI.getIterator());
1503 return Next !=
MBB->end() &&
Next->isCFIInstruction();
1510 Register &SrcReg2, int64_t &CmpMask,
1511 int64_t &CmpValue)
const {
1515 assert(
MI.getNumOperands() >= 2 &&
"All AArch64 cmps should have 2 operands");
1516 if (!
MI.getOperand(1).isReg() ||
MI.getOperand(1).getSubReg())
1519 switch (
MI.getOpcode()) {
1522 case AArch64::PTEST_PP:
1523 case AArch64::PTEST_PP_ANY:
1524 case AArch64::PTEST_PP_FIRST:
1525 SrcReg =
MI.getOperand(0).getReg();
1526 SrcReg2 =
MI.getOperand(1).getReg();
1527 if (
MI.getOperand(2).getSubReg())
1534 case AArch64::SUBSWrr:
1535 case AArch64::SUBSWrs:
1536 case AArch64::SUBSWrx:
1537 case AArch64::SUBSXrr:
1538 case AArch64::SUBSXrs:
1539 case AArch64::SUBSXrx:
1540 case AArch64::ADDSWrr:
1541 case AArch64::ADDSWrs:
1542 case AArch64::ADDSWrx:
1543 case AArch64::ADDSXrr:
1544 case AArch64::ADDSXrs:
1545 case AArch64::ADDSXrx:
1547 SrcReg =
MI.getOperand(1).getReg();
1548 SrcReg2 =
MI.getOperand(2).getReg();
1551 if (
MI.getOperand(2).getSubReg())
1557 case AArch64::SUBSWri:
1558 case AArch64::ADDSWri:
1559 case AArch64::SUBSXri:
1560 case AArch64::ADDSXri:
1561 SrcReg =
MI.getOperand(1).getReg();
1564 CmpValue =
MI.getOperand(2).getImm();
1566 case AArch64::ANDSWri:
1567 case AArch64::ANDSXri:
1570 SrcReg =
MI.getOperand(1).getReg();
1574 MI.getOperand(2).getImm(),
1575 MI.getOpcode() == AArch64::ANDSWri ? 32 : 64);
1584 assert(
MBB &&
"Can't get MachineBasicBlock here");
1586 assert(MF &&
"Can't get MachineFunction here");
1591 for (
unsigned OpIdx = 0, EndIdx = Instr.getNumOperands();
OpIdx < EndIdx;
1598 if (!OpRegCstraints)
1606 "Operand has register constraints without being a register!");
1609 if (
Reg.isPhysical()) {
1613 !
MRI->constrainRegClass(
Reg, OpRegCstraints))
1626 bool MIDefinesZeroReg =
false;
1627 if (
MI.definesRegister(AArch64::WZR,
nullptr) ||
1628 MI.definesRegister(AArch64::XZR,
nullptr))
1629 MIDefinesZeroReg =
true;
1631 switch (
MI.getOpcode()) {
1633 return MI.getOpcode();
1634 case AArch64::ADDSWrr:
1635 return AArch64::ADDWrr;
1636 case AArch64::ADDSWri:
1637 return MIDefinesZeroReg ? AArch64::ADDSWri : AArch64::ADDWri;
1638 case AArch64::ADDSWrs:
1639 return MIDefinesZeroReg ? AArch64::ADDSWrs : AArch64::ADDWrs;
1640 case AArch64::ADDSWrx:
1641 return AArch64::ADDWrx;
1642 case AArch64::ADDSXrr:
1643 return AArch64::ADDXrr;
1644 case AArch64::ADDSXri:
1645 return MIDefinesZeroReg ? AArch64::ADDSXri : AArch64::ADDXri;
1646 case AArch64::ADDSXrs:
1647 return MIDefinesZeroReg ? AArch64::ADDSXrs : AArch64::ADDXrs;
1648 case AArch64::ADDSXrx:
1649 return AArch64::ADDXrx;
1650 case AArch64::SUBSWrr:
1651 return AArch64::SUBWrr;
1652 case AArch64::SUBSWri:
1653 return MIDefinesZeroReg ? AArch64::SUBSWri : AArch64::SUBWri;
1654 case AArch64::SUBSWrs:
1655 return MIDefinesZeroReg ? AArch64::SUBSWrs : AArch64::SUBWrs;
1656 case AArch64::SUBSWrx:
1657 return AArch64::SUBWrx;
1658 case AArch64::SUBSXrr:
1659 return AArch64::SUBXrr;
1660 case AArch64::SUBSXri:
1661 return MIDefinesZeroReg ? AArch64::SUBSXri : AArch64::SUBXri;
1662 case AArch64::SUBSXrs:
1663 return MIDefinesZeroReg ? AArch64::SUBSXrs : AArch64::SUBXrs;
1664 case AArch64::SUBSXrx:
1665 return AArch64::SUBXrx;
1680 if (To == To->getParent()->begin())
1685 if (To->getParent() != From->getParent())
1697 Instr.modifiesRegister(AArch64::NZCV,
TRI)) ||
1698 ((AccessToCheck &
AK_Read) && Instr.readsRegister(AArch64::NZCV,
TRI)))
1704std::optional<unsigned>
1708 unsigned MaskOpcode =
Mask->getOpcode();
1709 unsigned PredOpcode = Pred->
getOpcode();
1710 bool PredIsPTestLike = isPTestLikeOpcode(PredOpcode);
1711 bool PredIsWhileLike = isWhileOpcode(PredOpcode);
1713 if (PredIsWhileLike) {
1717 if ((Mask == Pred) && PTest->
getOpcode() == AArch64::PTEST_PP_ANY)
1724 getElementSizeForOpcode(MaskOpcode) ==
1725 getElementSizeForOpcode(PredOpcode))
1731 if (PTest->
getOpcode() == AArch64::PTEST_PP_FIRST &&
1738 if (PredIsPTestLike) {
1743 if ((Mask == Pred) && PTest->
getOpcode() == AArch64::PTEST_PP_ANY)
1751 if (Mask != PTestLikeMask && PTestLikeMask->isFullCopy() &&
1752 PTestLikeMask->getOperand(1).getReg().isVirtual())
1754 MRI->getUniqueVRegDef(PTestLikeMask->getOperand(1).getReg());
1760 getElementSizeForOpcode(MaskOpcode) ==
1761 getElementSizeForOpcode(PredOpcode)) {
1762 if (Mask == PTestLikeMask || PTest->
getOpcode() == AArch64::PTEST_PP_ANY)
1788 uint64_t PredElementSize = getElementSizeForOpcode(PredOpcode);
1790 PTest->
getOpcode() == AArch64::PTEST_PP_ANY))
1798 switch (PredOpcode) {
1799 case AArch64::AND_PPzPP:
1800 case AArch64::BIC_PPzPP:
1801 case AArch64::EOR_PPzPP:
1802 case AArch64::NAND_PPzPP:
1803 case AArch64::NOR_PPzPP:
1804 case AArch64::ORN_PPzPP:
1805 case AArch64::ORR_PPzPP:
1806 case AArch64::BRKA_PPzP:
1807 case AArch64::BRKPA_PPzPP:
1808 case AArch64::BRKB_PPzP:
1809 case AArch64::BRKPB_PPzPP:
1810 case AArch64::RDFFR_PPz: {
1814 if (Mask != PredMask)
1818 case AArch64::BRKN_PPzP: {
1822 if ((MaskOpcode != AArch64::PTRUE_B) ||
1823 (
Mask->getOperand(1).getImm() != 31))
1827 case AArch64::PTRUE_B:
1840bool AArch64InstrInfo::optimizePTestInstr(
1841 MachineInstr *PTest,
unsigned MaskReg,
unsigned PredReg,
1843 auto *
Mask =
MRI->getUniqueVRegDef(MaskReg);
1844 auto *Pred =
MRI->getUniqueVRegDef(PredReg);
1846 if (Pred->
isCopy() && PTest->
getOpcode() == AArch64::PTEST_PP_FIRST) {
1850 if (
Op.isReg() &&
Op.getReg().isVirtual() &&
1851 Op.getSubReg() == AArch64::psub0)
1852 Pred =
MRI->getUniqueVRegDef(
Op.getReg());
1855 unsigned PredOpcode = Pred->
getOpcode();
1856 auto NewOp = canRemovePTestInstr(PTest, Mask, Pred,
MRI);
1872 if (*NewOp != PredOpcode) {
1883 for (; i !=
e; ++i) {
1914 if (DeadNZCVIdx != -1) {
1933 if (CmpInstr.
getOpcode() == AArch64::PTEST_PP ||
1934 CmpInstr.
getOpcode() == AArch64::PTEST_PP_ANY ||
1935 CmpInstr.
getOpcode() == AArch64::PTEST_PP_FIRST)
1936 return optimizePTestInstr(&CmpInstr, SrcReg, SrcReg2,
MRI);
1945 if (CmpValue == 0 && substituteCmpToZero(CmpInstr, SrcReg, *
MRI))
1947 return (CmpValue == 0 || CmpValue == 1) &&
1948 removeCmpToZeroOrOne(CmpInstr, SrcReg, CmpValue, *
MRI);
1956 switch (Instr.getOpcode()) {
1958 return AArch64::INSTRUCTION_LIST_END;
1960 case AArch64::ADDSWrr:
1961 case AArch64::ADDSWri:
1962 case AArch64::ADDSXrr:
1963 case AArch64::ADDSXri:
1964 case AArch64::ADDSWrx:
1965 case AArch64::ADDSXrx:
1966 case AArch64::SUBSWrr:
1967 case AArch64::SUBSWri:
1968 case AArch64::SUBSWrx:
1969 case AArch64::SUBSXrr:
1970 case AArch64::SUBSXri:
1971 case AArch64::SUBSXrx:
1972 case AArch64::ANDSWri:
1973 case AArch64::ANDSWrr:
1974 case AArch64::ANDSWrs:
1975 case AArch64::ANDSXri:
1976 case AArch64::ANDSXrr:
1977 case AArch64::ANDSXrs:
1978 case AArch64::BICSWrr:
1979 case AArch64::BICSXrr:
1980 case AArch64::BICSWrs:
1981 case AArch64::BICSXrs:
1982 return Instr.getOpcode();
1984 case AArch64::ADDWrr:
1985 return AArch64::ADDSWrr;
1986 case AArch64::ADDWri:
1987 return AArch64::ADDSWri;
1988 case AArch64::ADDXrr:
1989 return AArch64::ADDSXrr;
1990 case AArch64::ADDXri:
1991 return AArch64::ADDSXri;
1992 case AArch64::ADDWrx:
1993 return AArch64::ADDSWrx;
1994 case AArch64::ADDXrx:
1995 return AArch64::ADDSXrx;
1996 case AArch64::ADCWr:
1997 return AArch64::ADCSWr;
1998 case AArch64::ADCXr:
1999 return AArch64::ADCSXr;
2000 case AArch64::SUBWrr:
2001 return AArch64::SUBSWrr;
2002 case AArch64::SUBWri:
2003 return AArch64::SUBSWri;
2004 case AArch64::SUBXrr:
2005 return AArch64::SUBSXrr;
2006 case AArch64::SUBXri:
2007 return AArch64::SUBSXri;
2008 case AArch64::SUBWrx:
2009 return AArch64::SUBSWrx;
2010 case AArch64::SUBXrx:
2011 return AArch64::SUBSXrx;
2012 case AArch64::SBCWr:
2013 return AArch64::SBCSWr;
2014 case AArch64::SBCXr:
2015 return AArch64::SBCSXr;
2016 case AArch64::ANDWri:
2017 return AArch64::ANDSWri;
2018 case AArch64::ANDXri:
2019 return AArch64::ANDSXri;
2020 case AArch64::ANDWrr:
2021 return AArch64::ANDSWrr;
2022 case AArch64::ANDWrs:
2023 return AArch64::ANDSWrs;
2024 case AArch64::ANDXrr:
2025 return AArch64::ANDSXrr;
2026 case AArch64::ANDXrs:
2027 return AArch64::ANDSXrs;
2028 case AArch64::BICWrr:
2029 return AArch64::BICSWrr;
2030 case AArch64::BICXrr:
2031 return AArch64::BICSXrr;
2032 case AArch64::BICWrs:
2033 return AArch64::BICSWrs;
2034 case AArch64::BICXrs:
2035 return AArch64::BICSXrs;
2041 for (
auto *BB :
MBB->successors())
2042 if (BB->isLiveIn(AArch64::NZCV))
2051 switch (Instr.getOpcode()) {
2055 case AArch64::Bcc: {
2056 int Idx = Instr.findRegisterUseOperandIdx(AArch64::NZCV,
nullptr);
2061 case AArch64::CSINVWr:
2062 case AArch64::CSINVXr:
2063 case AArch64::CSINCWr:
2064 case AArch64::CSINCXr:
2065 case AArch64::CSELWr:
2066 case AArch64::CSELXr:
2067 case AArch64::CSNEGWr:
2068 case AArch64::CSNEGXr:
2069 case AArch64::FCSELSrrr:
2070 case AArch64::FCSELDrrr: {
2071 int Idx = Instr.findRegisterUseOperandIdx(AArch64::NZCV,
nullptr);
2084 Instr.getOperand(CCIdx).
getImm())
2137std::optional<UsedNZCV>
2142 if (
MI.getParent() != CmpParent)
2143 return std::nullopt;
2146 return std::nullopt;
2151 if (Instr.readsRegister(AArch64::NZCV, &
TRI)) {
2154 return std::nullopt;
2159 if (Instr.modifiesRegister(AArch64::NZCV, &
TRI))
2162 return NZCVUsedAfterCmp;
2166 return Opcode == AArch64::ADDSWri || Opcode == AArch64::ADDSXri;
2170 return Opcode == AArch64::SUBSWri || Opcode == AArch64::SUBSXri;
2176 case AArch64::ANDSWri:
2177 case AArch64::ANDSWrr:
2178 case AArch64::ANDSWrs:
2179 case AArch64::ANDSXri:
2180 case AArch64::ANDSXrr:
2181 case AArch64::ANDSXrs:
2182 case AArch64::BICSWrr:
2183 case AArch64::BICSXrr:
2184 case AArch64::BICSWrs:
2185 case AArch64::BICSXrs:
2211 const unsigned CmpOpcode = CmpInstr.
getOpcode();
2217 "Caller guarantees that CmpInstr compares with constant 0");
2220 if (!NZVCUsed || NZVCUsed->C)
2243bool AArch64InstrInfo::substituteCmpToZero(
2247 MachineInstr *
MI =
MRI.getUniqueVRegDef(SrcReg);
2254 if (NewOpc == AArch64::INSTRUCTION_LIST_END)
2261 MI->setDesc(
get(NewOpc));
2266 MI->addRegisterDefined(AArch64::NZCV, &
TRI);
2278 assert((CmpValue == 0 || CmpValue == 1) &&
2279 "Only comparisons to 0 or 1 considered for removal!");
2282 unsigned MIOpc =
MI.getOpcode();
2283 if (MIOpc == AArch64::CSINCWr) {
2284 if (
MI.getOperand(1).getReg() != AArch64::WZR ||
2285 MI.getOperand(2).getReg() != AArch64::WZR)
2287 }
else if (MIOpc == AArch64::CSINCXr) {
2288 if (
MI.getOperand(1).getReg() != AArch64::XZR ||
2289 MI.getOperand(2).getReg() != AArch64::XZR)
2299 if (
MI.findRegisterDefOperandIdx(AArch64::NZCV,
nullptr,
true) != -1)
2303 const unsigned CmpOpcode = CmpInstr.
getOpcode();
2305 if (CmpValue && !IsSubsRegImm)
2307 if (!CmpValue && !IsSubsRegImm && !
isADDSRegImm(CmpOpcode))
2312 if (MIUsedNZCV.
C || MIUsedNZCV.
V)
2315 std::optional<UsedNZCV> NZCVUsedAfterCmp =
2319 if (!NZCVUsedAfterCmp || NZCVUsedAfterCmp->C || NZCVUsedAfterCmp->V)
2322 if ((MIUsedNZCV.
Z && NZCVUsedAfterCmp->N) ||
2323 (MIUsedNZCV.
N && NZCVUsedAfterCmp->Z))
2326 if (MIUsedNZCV.
N && !CmpValue)
2368bool AArch64InstrInfo::removeCmpToZeroOrOne(
2371 MachineInstr *
MI =
MRI.getUniqueVRegDef(SrcReg);
2375 SmallVector<MachineInstr *, 4> CCUseInstrs;
2376 bool IsInvertCC =
false;
2384 for (MachineInstr *CCUseInstr : CCUseInstrs) {
2386 assert(Idx >= 0 &&
"Unexpected instruction using CC.");
2387 MachineOperand &CCOperand = CCUseInstr->getOperand(Idx);
2396bool AArch64InstrInfo::expandPostRAPseudo(
MachineInstr &
MI)
const {
2397 if (
MI.getOpcode() != TargetOpcode::LOAD_STACK_GUARD &&
2398 MI.getOpcode() != AArch64::CATCHRET)
2401 MachineBasicBlock &
MBB = *
MI.getParent();
2403 auto TRI = Subtarget.getRegisterInfo();
2406 if (
MI.getOpcode() == AArch64::CATCHRET) {
2408 const TargetInstrInfo *
TII =
2410 MachineBasicBlock *TargetMBB =
MI.getOperand(0).getMBB();
2415 FirstEpilogSEH = std::prev(FirstEpilogSEH);
2417 FirstEpilogSEH = std::next(FirstEpilogSEH);
2432 if (
M.getStackProtectorGuard() ==
"sysreg") {
2433 const AArch64SysReg::SysReg *SrcReg =
2434 AArch64SysReg::lookupSysRegByName(
M.getStackProtectorGuardReg());
2442 int Offset =
M.getStackProtectorGuardOffset();
2493 const GlobalValue *GV =
2496 unsigned OpFlags = Subtarget.ClassifyGlobalReference(GV, TM);
2502 if (Subtarget.isTargetILP32()) {
2503 unsigned Reg32 =
TRI->getSubReg(
Reg, AArch64::sub_32);
2517 assert(!Subtarget.isTargetILP32() &&
"how can large exist in ILP32?");
2544 if (Subtarget.isTargetILP32()) {
2545 unsigned Reg32 =
TRI->getSubReg(
Reg, AArch64::sub_32);
2568 switch (
MI.getOpcode()) {
2571 case AArch64::MOVZWi:
2572 case AArch64::MOVZXi:
2573 if (
MI.getOperand(1).isImm() &&
MI.getOperand(1).getImm() == 0) {
2574 assert(
MI.getDesc().getNumOperands() == 3 &&
2575 MI.getOperand(2).getImm() == 0 &&
"invalid MOVZi operands");
2579 case AArch64::ANDWri:
2580 return MI.getOperand(1).getReg() == AArch64::WZR;
2581 case AArch64::ANDXri:
2582 return MI.getOperand(1).getReg() == AArch64::XZR;
2583 case TargetOpcode::COPY:
2584 return MI.getOperand(1).getReg() == AArch64::WZR;
2592 switch (
MI.getOpcode()) {
2595 case TargetOpcode::COPY: {
2598 return (AArch64::GPR32RegClass.
contains(DstReg) ||
2599 AArch64::GPR64RegClass.
contains(DstReg));
2601 case AArch64::ORRXrs:
2602 if (
MI.getOperand(1).getReg() == AArch64::XZR) {
2603 assert(
MI.getDesc().getNumOperands() == 4 &&
2604 MI.getOperand(3).getImm() == 0 &&
"invalid ORRrs operands");
2608 case AArch64::ADDXri:
2609 if (
MI.getOperand(2).getImm() == 0) {
2610 assert(
MI.getDesc().getNumOperands() == 4 &&
2611 MI.getOperand(3).getImm() == 0 &&
"invalid ADDXri operands");
2622 switch (
MI.getOpcode()) {
2625 case TargetOpcode::COPY: {
2627 return AArch64::FPR128RegClass.contains(DstReg);
2629 case AArch64::ORRv16i8:
2630 if (
MI.getOperand(1).getReg() ==
MI.getOperand(2).getReg()) {
2631 assert(
MI.getDesc().getNumOperands() == 3 &&
MI.getOperand(0).isReg() &&
2632 "invalid ORRv16i8 operands");
2644 case AArch64::LDRWui:
2645 case AArch64::LDRXui:
2646 case AArch64::LDRBui:
2647 case AArch64::LDRHui:
2648 case AArch64::LDRSui:
2649 case AArch64::LDRDui:
2650 case AArch64::LDRQui:
2651 case AArch64::LDR_PXI:
2657 int &FrameIndex)
const {
2661 if (
MI.getOperand(0).getSubReg() == 0 &&
MI.getOperand(1).isFI() &&
2662 MI.getOperand(2).isImm() &&
MI.getOperand(2).getImm() == 0) {
2663 FrameIndex =
MI.getOperand(1).getIndex();
2664 return MI.getOperand(0).getReg();
2673 case AArch64::STRWui:
2674 case AArch64::STRXui:
2675 case AArch64::STRBui:
2676 case AArch64::STRHui:
2677 case AArch64::STRSui:
2678 case AArch64::STRDui:
2679 case AArch64::STRQui:
2680 case AArch64::STR_PXI:
2686 int &FrameIndex)
const {
2690 if (
MI.getOperand(0).getSubReg() == 0 &&
MI.getOperand(1).isFI() &&
2691 MI.getOperand(2).isImm() &&
MI.getOperand(2).getImm() == 0) {
2692 FrameIndex =
MI.getOperand(1).getIndex();
2693 return MI.getOperand(0).getReg();
2699 int &FrameIndex)
const {
2714 return MI.getOperand(0).getReg();
2720 int &FrameIndex)
const {
2735 return MI.getOperand(0).getReg();
2743 return MMO->getFlags() & MOSuppressPair;
2749 if (
MI.memoperands_empty())
2757 return MMO->getFlags() & MOStridedAccess;
2765 case AArch64::STURSi:
2766 case AArch64::STRSpre:
2767 case AArch64::STURDi:
2768 case AArch64::STRDpre:
2769 case AArch64::STURQi:
2770 case AArch64::STRQpre:
2771 case AArch64::STURBBi:
2772 case AArch64::STURHHi:
2773 case AArch64::STURWi:
2774 case AArch64::STRWpre:
2775 case AArch64::STURXi:
2776 case AArch64::STRXpre:
2777 case AArch64::LDURSi:
2778 case AArch64::LDRSpre:
2779 case AArch64::LDURDi:
2780 case AArch64::LDRDpre:
2781 case AArch64::LDURQi:
2782 case AArch64::LDRQpre:
2783 case AArch64::LDURWi:
2784 case AArch64::LDRWpre:
2785 case AArch64::LDURXi:
2786 case AArch64::LDRXpre:
2787 case AArch64::LDRSWpre:
2788 case AArch64::LDURSWi:
2789 case AArch64::LDURHHi:
2790 case AArch64::LDURBBi:
2791 case AArch64::LDURSBWi:
2792 case AArch64::LDURSHWi:
2800 case AArch64::PRFMui:
return AArch64::PRFUMi;
2801 case AArch64::LDRXui:
return AArch64::LDURXi;
2802 case AArch64::LDRWui:
return AArch64::LDURWi;
2803 case AArch64::LDRBui:
return AArch64::LDURBi;
2804 case AArch64::LDRHui:
return AArch64::LDURHi;
2805 case AArch64::LDRSui:
return AArch64::LDURSi;
2806 case AArch64::LDRDui:
return AArch64::LDURDi;
2807 case AArch64::LDRQui:
return AArch64::LDURQi;
2808 case AArch64::LDRBBui:
return AArch64::LDURBBi;
2809 case AArch64::LDRHHui:
return AArch64::LDURHHi;
2810 case AArch64::LDRSBXui:
return AArch64::LDURSBXi;
2811 case AArch64::LDRSBWui:
return AArch64::LDURSBWi;
2812 case AArch64::LDRSHXui:
return AArch64::LDURSHXi;
2813 case AArch64::LDRSHWui:
return AArch64::LDURSHWi;
2814 case AArch64::LDRSWui:
return AArch64::LDURSWi;
2815 case AArch64::STRXui:
return AArch64::STURXi;
2816 case AArch64::STRWui:
return AArch64::STURWi;
2817 case AArch64::STRBui:
return AArch64::STURBi;
2818 case AArch64::STRHui:
return AArch64::STURHi;
2819 case AArch64::STRSui:
return AArch64::STURSi;
2820 case AArch64::STRDui:
return AArch64::STURDi;
2821 case AArch64::STRQui:
return AArch64::STURQi;
2822 case AArch64::STRBBui:
return AArch64::STURBBi;
2823 case AArch64::STRHHui:
return AArch64::STURHHi;
2832 case AArch64::LDAPURBi:
2833 case AArch64::LDAPURHi:
2834 case AArch64::LDAPURi:
2835 case AArch64::LDAPURSBWi:
2836 case AArch64::LDAPURSBXi:
2837 case AArch64::LDAPURSHWi:
2838 case AArch64::LDAPURSHXi:
2839 case AArch64::LDAPURSWi:
2840 case AArch64::LDAPURXi:
2841 case AArch64::LDR_PPXI:
2842 case AArch64::LDR_PXI:
2843 case AArch64::LDR_ZXI:
2844 case AArch64::LDR_ZZXI:
2845 case AArch64::LDR_ZZXI_STRIDED_CONTIGUOUS:
2846 case AArch64::LDR_ZZZXI:
2847 case AArch64::LDR_ZZZZXI:
2848 case AArch64::LDR_ZZZZXI_STRIDED_CONTIGUOUS:
2849 case AArch64::LDRBBui:
2850 case AArch64::LDRBui:
2851 case AArch64::LDRDui:
2852 case AArch64::LDRHHui:
2853 case AArch64::LDRHui:
2854 case AArch64::LDRQui:
2855 case AArch64::LDRSBWui:
2856 case AArch64::LDRSBXui:
2857 case AArch64::LDRSHWui:
2858 case AArch64::LDRSHXui:
2859 case AArch64::LDRSui:
2860 case AArch64::LDRSWui:
2861 case AArch64::LDRWui:
2862 case AArch64::LDRXui:
2863 case AArch64::LDURBBi:
2864 case AArch64::LDURBi:
2865 case AArch64::LDURDi:
2866 case AArch64::LDURHHi:
2867 case AArch64::LDURHi:
2868 case AArch64::LDURQi:
2869 case AArch64::LDURSBWi:
2870 case AArch64::LDURSBXi:
2871 case AArch64::LDURSHWi:
2872 case AArch64::LDURSHXi:
2873 case AArch64::LDURSi:
2874 case AArch64::LDURSWi:
2875 case AArch64::LDURWi:
2876 case AArch64::LDURXi:
2877 case AArch64::PRFMui:
2878 case AArch64::PRFUMi:
2879 case AArch64::ST2Gi:
2881 case AArch64::STLURBi:
2882 case AArch64::STLURHi:
2883 case AArch64::STLURWi:
2884 case AArch64::STLURXi:
2885 case AArch64::StoreSwiftAsyncContext:
2886 case AArch64::STR_PPXI:
2887 case AArch64::STR_PXI:
2888 case AArch64::STR_ZXI:
2889 case AArch64::STR_ZZXI:
2890 case AArch64::STR_ZZXI_STRIDED_CONTIGUOUS:
2891 case AArch64::STR_ZZZXI:
2892 case AArch64::STR_ZZZZXI:
2893 case AArch64::STR_ZZZZXI_STRIDED_CONTIGUOUS:
2894 case AArch64::STRBBui:
2895 case AArch64::STRBui:
2896 case AArch64::STRDui:
2897 case AArch64::STRHHui:
2898 case AArch64::STRHui:
2899 case AArch64::STRQui:
2900 case AArch64::STRSui:
2901 case AArch64::STRWui:
2902 case AArch64::STRXui:
2903 case AArch64::STURBBi:
2904 case AArch64::STURBi:
2905 case AArch64::STURDi:
2906 case AArch64::STURHHi:
2907 case AArch64::STURHi:
2908 case AArch64::STURQi:
2909 case AArch64::STURSi:
2910 case AArch64::STURWi:
2911 case AArch64::STURXi:
2912 case AArch64::STZ2Gi:
2913 case AArch64::STZGi:
2914 case AArch64::TAGPstack:
2916 case AArch64::LD1B_D_IMM:
2917 case AArch64::LD1B_H_IMM:
2918 case AArch64::LD1B_IMM:
2919 case AArch64::LD1B_S_IMM:
2920 case AArch64::LD1D_IMM:
2921 case AArch64::LD1H_D_IMM:
2922 case AArch64::LD1H_IMM:
2923 case AArch64::LD1H_S_IMM:
2924 case AArch64::LD1RB_D_IMM:
2925 case AArch64::LD1RB_H_IMM:
2926 case AArch64::LD1RB_IMM:
2927 case AArch64::LD1RB_S_IMM:
2928 case AArch64::LD1RD_IMM:
2929 case AArch64::LD1RH_D_IMM:
2930 case AArch64::LD1RH_IMM:
2931 case AArch64::LD1RH_S_IMM:
2932 case AArch64::LD1RSB_D_IMM:
2933 case AArch64::LD1RSB_H_IMM:
2934 case AArch64::LD1RSB_S_IMM:
2935 case AArch64::LD1RSH_D_IMM:
2936 case AArch64::LD1RSH_S_IMM:
2937 case AArch64::LD1RSW_IMM:
2938 case AArch64::LD1RW_D_IMM:
2939 case AArch64::LD1RW_IMM:
2940 case AArch64::LD1SB_D_IMM:
2941 case AArch64::LD1SB_H_IMM:
2942 case AArch64::LD1SB_S_IMM:
2943 case AArch64::LD1SH_D_IMM:
2944 case AArch64::LD1SH_S_IMM:
2945 case AArch64::LD1SW_D_IMM:
2946 case AArch64::LD1W_D_IMM:
2947 case AArch64::LD1W_IMM:
2948 case AArch64::LD2B_IMM:
2949 case AArch64::LD2D_IMM:
2950 case AArch64::LD2H_IMM:
2951 case AArch64::LD2W_IMM:
2952 case AArch64::LD3B_IMM:
2953 case AArch64::LD3D_IMM:
2954 case AArch64::LD3H_IMM:
2955 case AArch64::LD3W_IMM:
2956 case AArch64::LD4B_IMM:
2957 case AArch64::LD4D_IMM:
2958 case AArch64::LD4H_IMM:
2959 case AArch64::LD4W_IMM:
2961 case AArch64::LDNF1B_D_IMM:
2962 case AArch64::LDNF1B_H_IMM:
2963 case AArch64::LDNF1B_IMM:
2964 case AArch64::LDNF1B_S_IMM:
2965 case AArch64::LDNF1D_IMM:
2966 case AArch64::LDNF1H_D_IMM:
2967 case AArch64::LDNF1H_IMM:
2968 case AArch64::LDNF1H_S_IMM:
2969 case AArch64::LDNF1SB_D_IMM:
2970 case AArch64::LDNF1SB_H_IMM:
2971 case AArch64::LDNF1SB_S_IMM:
2972 case AArch64::LDNF1SH_D_IMM:
2973 case AArch64::LDNF1SH_S_IMM:
2974 case AArch64::LDNF1SW_D_IMM:
2975 case AArch64::LDNF1W_D_IMM:
2976 case AArch64::LDNF1W_IMM:
2977 case AArch64::LDNPDi:
2978 case AArch64::LDNPQi:
2979 case AArch64::LDNPSi:
2980 case AArch64::LDNPWi:
2981 case AArch64::LDNPXi:
2982 case AArch64::LDNT1B_ZRI:
2983 case AArch64::LDNT1D_ZRI:
2984 case AArch64::LDNT1H_ZRI:
2985 case AArch64::LDNT1W_ZRI:
2986 case AArch64::LDPDi:
2987 case AArch64::LDPQi:
2988 case AArch64::LDPSi:
2989 case AArch64::LDPWi:
2990 case AArch64::LDPXi:
2991 case AArch64::LDRBBpost:
2992 case AArch64::LDRBBpre:
2993 case AArch64::LDRBpost:
2994 case AArch64::LDRBpre:
2995 case AArch64::LDRDpost:
2996 case AArch64::LDRDpre:
2997 case AArch64::LDRHHpost:
2998 case AArch64::LDRHHpre:
2999 case AArch64::LDRHpost:
3000 case AArch64::LDRHpre:
3001 case AArch64::LDRQpost:
3002 case AArch64::LDRQpre:
3003 case AArch64::LDRSpost:
3004 case AArch64::LDRSpre:
3005 case AArch64::LDRWpost:
3006 case AArch64::LDRWpre:
3007 case AArch64::LDRXpost:
3008 case AArch64::LDRXpre:
3009 case AArch64::ST1B_D_IMM:
3010 case AArch64::ST1B_H_IMM:
3011 case AArch64::ST1B_IMM:
3012 case AArch64::ST1B_S_IMM:
3013 case AArch64::ST1D_IMM:
3014 case AArch64::ST1H_D_IMM:
3015 case AArch64::ST1H_IMM:
3016 case AArch64::ST1H_S_IMM:
3017 case AArch64::ST1W_D_IMM:
3018 case AArch64::ST1W_IMM:
3019 case AArch64::ST2B_IMM:
3020 case AArch64::ST2D_IMM:
3021 case AArch64::ST2H_IMM:
3022 case AArch64::ST2W_IMM:
3023 case AArch64::ST3B_IMM:
3024 case AArch64::ST3D_IMM:
3025 case AArch64::ST3H_IMM:
3026 case AArch64::ST3W_IMM:
3027 case AArch64::ST4B_IMM:
3028 case AArch64::ST4D_IMM:
3029 case AArch64::ST4H_IMM:
3030 case AArch64::ST4W_IMM:
3031 case AArch64::STGPi:
3032 case AArch64::STGPreIndex:
3033 case AArch64::STZGPreIndex:
3034 case AArch64::ST2GPreIndex:
3035 case AArch64::STZ2GPreIndex:
3036 case AArch64::STGPostIndex:
3037 case AArch64::STZGPostIndex:
3038 case AArch64::ST2GPostIndex:
3039 case AArch64::STZ2GPostIndex:
3040 case AArch64::STNPDi:
3041 case AArch64::STNPQi:
3042 case AArch64::STNPSi:
3043 case AArch64::STNPWi:
3044 case AArch64::STNPXi:
3045 case AArch64::STNT1B_ZRI:
3046 case AArch64::STNT1D_ZRI:
3047 case AArch64::STNT1H_ZRI:
3048 case AArch64::STNT1W_ZRI:
3049 case AArch64::STPDi:
3050 case AArch64::STPQi:
3051 case AArch64::STPSi:
3052 case AArch64::STPWi:
3053 case AArch64::STPXi:
3054 case AArch64::STRBBpost:
3055 case AArch64::STRBBpre:
3056 case AArch64::STRBpost:
3057 case AArch64::STRBpre:
3058 case AArch64::STRDpost:
3059 case AArch64::STRDpre:
3060 case AArch64::STRHHpost:
3061 case AArch64::STRHHpre:
3062 case AArch64::STRHpost:
3063 case AArch64::STRHpre:
3064 case AArch64::STRQpost:
3065 case AArch64::STRQpre:
3066 case AArch64::STRSpost:
3067 case AArch64::STRSpre:
3068 case AArch64::STRWpost:
3069 case AArch64::STRWpre:
3070 case AArch64::STRXpost:
3071 case AArch64::STRXpre:
3073 case AArch64::LDPDpost:
3074 case AArch64::LDPDpre:
3075 case AArch64::LDPQpost:
3076 case AArch64::LDPQpre:
3077 case AArch64::LDPSpost:
3078 case AArch64::LDPSpre:
3079 case AArch64::LDPWpost:
3080 case AArch64::LDPWpre:
3081 case AArch64::LDPXpost:
3082 case AArch64::LDPXpre:
3083 case AArch64::STGPpre:
3084 case AArch64::STGPpost:
3085 case AArch64::STPDpost:
3086 case AArch64::STPDpre:
3087 case AArch64::STPQpost:
3088 case AArch64::STPQpre:
3089 case AArch64::STPSpost:
3090 case AArch64::STPSpre:
3091 case AArch64::STPWpost:
3092 case AArch64::STPWpre:
3093 case AArch64::STPXpost:
3094 case AArch64::STPXpre:
3100 switch (
MI.getOpcode()) {
3104 case AArch64::STRSui:
3105 case AArch64::STRDui:
3106 case AArch64::STRQui:
3107 case AArch64::STRXui:
3108 case AArch64::STRWui:
3109 case AArch64::LDRSui:
3110 case AArch64::LDRDui:
3111 case AArch64::LDRQui:
3112 case AArch64::LDRXui:
3113 case AArch64::LDRWui:
3114 case AArch64::LDRSWui:
3116 case AArch64::STURSi:
3117 case AArch64::STRSpre:
3118 case AArch64::STURDi:
3119 case AArch64::STRDpre:
3120 case AArch64::STURQi:
3121 case AArch64::STRQpre:
3122 case AArch64::STURWi:
3123 case AArch64::STRWpre:
3124 case AArch64::STURXi:
3125 case AArch64::STRXpre:
3126 case AArch64::LDURSi:
3127 case AArch64::LDRSpre:
3128 case AArch64::LDURDi:
3129 case AArch64::LDRDpre:
3130 case AArch64::LDURQi:
3131 case AArch64::LDRQpre:
3132 case AArch64::LDURWi:
3133 case AArch64::LDRWpre:
3134 case AArch64::LDURXi:
3135 case AArch64::LDRXpre:
3136 case AArch64::LDURSWi:
3137 case AArch64::LDRSWpre:
3139 case AArch64::LDR_ZXI:
3140 case AArch64::STR_ZXI:
3146 switch (
MI.getOpcode()) {
3149 "Unexpected instruction - was a new tail call opcode introduced?");
3151 case AArch64::TCRETURNdi:
3152 case AArch64::TCRETURNri:
3153 case AArch64::TCRETURNrix16x17:
3154 case AArch64::TCRETURNrix17:
3155 case AArch64::TCRETURNrinotx16:
3156 case AArch64::TCRETURNriALL:
3157 case AArch64::AUTH_TCRETURN:
3158 case AArch64::AUTH_TCRETURN_BTI:
3168 case AArch64::ADDWri:
3169 return AArch64::ADDSWri;
3170 case AArch64::ADDWrr:
3171 return AArch64::ADDSWrr;
3172 case AArch64::ADDWrs:
3173 return AArch64::ADDSWrs;
3174 case AArch64::ADDWrx:
3175 return AArch64::ADDSWrx;
3176 case AArch64::ANDWri:
3177 return AArch64::ANDSWri;
3178 case AArch64::ANDWrr:
3179 return AArch64::ANDSWrr;
3180 case AArch64::ANDWrs:
3181 return AArch64::ANDSWrs;
3182 case AArch64::BICWrr:
3183 return AArch64::BICSWrr;
3184 case AArch64::BICWrs:
3185 return AArch64::BICSWrs;
3186 case AArch64::SUBWri:
3187 return AArch64::SUBSWri;
3188 case AArch64::SUBWrr:
3189 return AArch64::SUBSWrr;
3190 case AArch64::SUBWrs:
3191 return AArch64::SUBSWrs;
3192 case AArch64::SUBWrx:
3193 return AArch64::SUBSWrx;
3195 case AArch64::ADDXri:
3196 return AArch64::ADDSXri;
3197 case AArch64::ADDXrr:
3198 return AArch64::ADDSXrr;
3199 case AArch64::ADDXrs:
3200 return AArch64::ADDSXrs;
3201 case AArch64::ADDXrx:
3202 return AArch64::ADDSXrx;
3203 case AArch64::ANDXri:
3204 return AArch64::ANDSXri;
3205 case AArch64::ANDXrr:
3206 return AArch64::ANDSXrr;
3207 case AArch64::ANDXrs:
3208 return AArch64::ANDSXrs;
3209 case AArch64::BICXrr:
3210 return AArch64::BICSXrr;
3211 case AArch64::BICXrs:
3212 return AArch64::BICSXrs;
3213 case AArch64::SUBXri:
3214 return AArch64::SUBSXri;
3215 case AArch64::SUBXrr:
3216 return AArch64::SUBSXrr;
3217 case AArch64::SUBXrs:
3218 return AArch64::SUBSXrs;
3219 case AArch64::SUBXrx:
3220 return AArch64::SUBSXrx;
3222 case AArch64::AND_PPzPP:
3223 return AArch64::ANDS_PPzPP;
3224 case AArch64::BIC_PPzPP:
3225 return AArch64::BICS_PPzPP;
3226 case AArch64::EOR_PPzPP:
3227 return AArch64::EORS_PPzPP;
3228 case AArch64::NAND_PPzPP:
3229 return AArch64::NANDS_PPzPP;
3230 case AArch64::NOR_PPzPP:
3231 return AArch64::NORS_PPzPP;
3232 case AArch64::ORN_PPzPP:
3233 return AArch64::ORNS_PPzPP;
3234 case AArch64::ORR_PPzPP:
3235 return AArch64::ORRS_PPzPP;
3236 case AArch64::BRKA_PPzP:
3237 return AArch64::BRKAS_PPzP;
3238 case AArch64::BRKPA_PPzPP:
3239 return AArch64::BRKPAS_PPzPP;
3240 case AArch64::BRKB_PPzP:
3241 return AArch64::BRKBS_PPzP;
3242 case AArch64::BRKPB_PPzPP:
3243 return AArch64::BRKPBS_PPzPP;
3244 case AArch64::BRKN_PPzP:
3245 return AArch64::BRKNS_PPzP;
3246 case AArch64::RDFFR_PPz:
3247 return AArch64::RDFFRS_PPz;
3248 case AArch64::PTRUE_B:
3249 return AArch64::PTRUES_B;
3260 if (
MI.hasOrderedMemoryRef())
3265 assert((
MI.getOperand(IsPreLdSt ? 2 : 1).isReg() ||
3266 MI.getOperand(IsPreLdSt ? 2 : 1).isFI()) &&
3267 "Expected a reg or frame index operand.");
3271 bool IsImmPreLdSt = IsPreLdSt &&
MI.getOperand(3).isImm();
3273 if (!
MI.getOperand(2).isImm() && !IsImmPreLdSt)
3286 if (
MI.getOperand(1).isReg() && !IsPreLdSt) {
3287 Register BaseReg =
MI.getOperand(1).getReg();
3289 if (
MI.modifiesRegister(BaseReg,
TRI))
3295 switch (
MI.getOpcode()) {
3298 case AArch64::LDR_ZXI:
3299 case AArch64::STR_ZXI:
3300 if (!Subtarget.isLittleEndian() ||
3301 Subtarget.getSVEVectorSizeInBits() != 128)
3314 const MCAsmInfo *MAI =
MI.getMF()->getTarget().getMCAsmInfo();
3316 MI.getMF()->getFunction().needsUnwindTableEntry();
3322 if (Subtarget.isPaired128Slow()) {
3323 switch (
MI.getOpcode()) {
3326 case AArch64::LDURQi:
3327 case AArch64::STURQi:
3328 case AArch64::LDRQui:
3329 case AArch64::STRQui:
3356std::optional<ExtAddrMode>
3361 bool OffsetIsScalable;
3362 if (!getMemOperandWithOffset(MemI,
Base,
Offset, OffsetIsScalable,
TRI))
3363 return std::nullopt;
3366 return std::nullopt;
3381 int64_t OffsetScale = 1;
3386 case AArch64::LDURQi:
3387 case AArch64::STURQi:
3391 case AArch64::LDURDi:
3392 case AArch64::STURDi:
3393 case AArch64::LDURXi:
3394 case AArch64::STURXi:
3398 case AArch64::LDURWi:
3399 case AArch64::LDURSWi:
3400 case AArch64::STURWi:
3404 case AArch64::LDURHi:
3405 case AArch64::STURHi:
3406 case AArch64::LDURHHi:
3407 case AArch64::STURHHi:
3408 case AArch64::LDURSHXi:
3409 case AArch64::LDURSHWi:
3413 case AArch64::LDRBroX:
3414 case AArch64::LDRBBroX:
3415 case AArch64::LDRSBXroX:
3416 case AArch64::LDRSBWroX:
3417 case AArch64::STRBroX:
3418 case AArch64::STRBBroX:
3419 case AArch64::LDURBi:
3420 case AArch64::LDURBBi:
3421 case AArch64::LDURSBXi:
3422 case AArch64::LDURSBWi:
3423 case AArch64::STURBi:
3424 case AArch64::STURBBi:
3425 case AArch64::LDRBui:
3426 case AArch64::LDRBBui:
3427 case AArch64::LDRSBXui:
3428 case AArch64::LDRSBWui:
3429 case AArch64::STRBui:
3430 case AArch64::STRBBui:
3434 case AArch64::LDRQroX:
3435 case AArch64::STRQroX:
3436 case AArch64::LDRQui:
3437 case AArch64::STRQui:
3442 case AArch64::LDRDroX:
3443 case AArch64::STRDroX:
3444 case AArch64::LDRXroX:
3445 case AArch64::STRXroX:
3446 case AArch64::LDRDui:
3447 case AArch64::STRDui:
3448 case AArch64::LDRXui:
3449 case AArch64::STRXui:
3454 case AArch64::LDRWroX:
3455 case AArch64::LDRSWroX:
3456 case AArch64::STRWroX:
3457 case AArch64::LDRWui:
3458 case AArch64::LDRSWui:
3459 case AArch64::STRWui:
3464 case AArch64::LDRHroX:
3465 case AArch64::STRHroX:
3466 case AArch64::LDRHHroX:
3467 case AArch64::STRHHroX:
3468 case AArch64::LDRSHXroX:
3469 case AArch64::LDRSHWroX:
3470 case AArch64::LDRHui:
3471 case AArch64::STRHui:
3472 case AArch64::LDRHHui:
3473 case AArch64::STRHHui:
3474 case AArch64::LDRSHXui:
3475 case AArch64::LDRSHWui:
3483 if (BaseRegOp.
isReg() && BaseRegOp.
getReg() == Reg)
3507 case AArch64::SBFMXri:
3520 AM.
Scale = OffsetScale;
3525 case TargetOpcode::SUBREG_TO_REG: {
3538 if (!OffsetReg.
isVirtual() || !
MRI.hasOneNonDBGUse(OffsetReg))
3542 if (
DefMI.getOpcode() != AArch64::ORRWrs ||
3543 DefMI.getOperand(1).getReg() != AArch64::WZR ||
3544 DefMI.getOperand(3).getImm() != 0)
3551 AM.
Scale = OffsetScale;
3562 auto validateOffsetForLDP = [](
unsigned NumBytes, int64_t OldOffset,
3563 int64_t NewOffset) ->
bool {
3564 int64_t MinOffset, MaxOffset;
3581 return OldOffset < MinOffset || OldOffset > MaxOffset ||
3582 (NewOffset >= MinOffset && NewOffset <= MaxOffset);
3584 auto canFoldAddSubImmIntoAddrMode = [&](int64_t Disp) ->
bool {
3586 int64_t NewOffset = OldOffset + Disp;
3587 if (!isLegalAddressingMode(NumBytes, NewOffset, 0))
3591 if (!validateOffsetForLDP(NumBytes, OldOffset, NewOffset))
3601 auto canFoldAddRegIntoAddrMode =
3606 if ((
unsigned)Scale != Scale)
3608 if (!isLegalAddressingMode(NumBytes, 0, Scale))
3620 return (Opcode == AArch64::STURQi || Opcode == AArch64::STRQui) &&
3621 Subtarget.isSTRQroSlow();
3630 case AArch64::ADDXri:
3636 return canFoldAddSubImmIntoAddrMode(Disp);
3638 case AArch64::SUBXri:
3644 return canFoldAddSubImmIntoAddrMode(-Disp);
3646 case AArch64::ADDXrs: {
3659 if (Shift != 2 && Shift != 3 && Subtarget.hasAddrLSLSlow14())
3661 if (avoidSlowSTRQ(MemI))
3664 return canFoldAddRegIntoAddrMode(1ULL << Shift);
3667 case AArch64::ADDXrr:
3675 if (!OptSize && avoidSlowSTRQ(MemI))
3677 return canFoldAddRegIntoAddrMode(1);
3679 case AArch64::ADDXrx:
3687 if (!OptSize && avoidSlowSTRQ(MemI))
3696 return canFoldAddRegIntoAddrMode(
3711 case AArch64::LDURQi:
3712 case AArch64::LDRQui:
3713 return AArch64::LDRQroX;
3714 case AArch64::STURQi:
3715 case AArch64::STRQui:
3716 return AArch64::STRQroX;
3717 case AArch64::LDURDi:
3718 case AArch64::LDRDui:
3719 return AArch64::LDRDroX;
3720 case AArch64::STURDi:
3721 case AArch64::STRDui:
3722 return AArch64::STRDroX;
3723 case AArch64::LDURXi:
3724 case AArch64::LDRXui:
3725 return AArch64::LDRXroX;
3726 case AArch64::STURXi:
3727 case AArch64::STRXui:
3728 return AArch64::STRXroX;
3729 case AArch64::LDURWi:
3730 case AArch64::LDRWui:
3731 return AArch64::LDRWroX;
3732 case AArch64::LDURSWi:
3733 case AArch64::LDRSWui:
3734 return AArch64::LDRSWroX;
3735 case AArch64::STURWi:
3736 case AArch64::STRWui:
3737 return AArch64::STRWroX;
3738 case AArch64::LDURHi:
3739 case AArch64::LDRHui:
3740 return AArch64::LDRHroX;
3741 case AArch64::STURHi:
3742 case AArch64::STRHui:
3743 return AArch64::STRHroX;
3744 case AArch64::LDURHHi:
3745 case AArch64::LDRHHui:
3746 return AArch64::LDRHHroX;
3747 case AArch64::STURHHi:
3748 case AArch64::STRHHui:
3749 return AArch64::STRHHroX;
3750 case AArch64::LDURSHXi:
3751 case AArch64::LDRSHXui:
3752 return AArch64::LDRSHXroX;
3753 case AArch64::LDURSHWi:
3754 case AArch64::LDRSHWui:
3755 return AArch64::LDRSHWroX;
3756 case AArch64::LDURBi:
3757 case AArch64::LDRBui:
3758 return AArch64::LDRBroX;
3759 case AArch64::LDURBBi:
3760 case AArch64::LDRBBui:
3761 return AArch64::LDRBBroX;
3762 case AArch64::LDURSBXi:
3763 case AArch64::LDRSBXui:
3764 return AArch64::LDRSBXroX;
3765 case AArch64::LDURSBWi:
3766 case AArch64::LDRSBWui:
3767 return AArch64::LDRSBWroX;
3768 case AArch64::STURBi:
3769 case AArch64::STRBui:
3770 return AArch64::STRBroX;
3771 case AArch64::STURBBi:
3772 case AArch64::STRBBui:
3773 return AArch64::STRBBroX;
3785 case AArch64::LDURQi:
3787 return AArch64::LDRQui;
3788 case AArch64::STURQi:
3790 return AArch64::STRQui;
3791 case AArch64::LDURDi:
3793 return AArch64::LDRDui;
3794 case AArch64::STURDi:
3796 return AArch64::STRDui;
3797 case AArch64::LDURXi:
3799 return AArch64::LDRXui;
3800 case AArch64::STURXi:
3802 return AArch64::STRXui;
3803 case AArch64::LDURWi:
3805 return AArch64::LDRWui;
3806 case AArch64::LDURSWi:
3808 return AArch64::LDRSWui;
3809 case AArch64::STURWi:
3811 return AArch64::STRWui;
3812 case AArch64::LDURHi:
3814 return AArch64::LDRHui;
3815 case AArch64::STURHi:
3817 return AArch64::STRHui;
3818 case AArch64::LDURHHi:
3820 return AArch64::LDRHHui;
3821 case AArch64::STURHHi:
3823 return AArch64::STRHHui;
3824 case AArch64::LDURSHXi:
3826 return AArch64::LDRSHXui;
3827 case AArch64::LDURSHWi:
3829 return AArch64::LDRSHWui;
3830 case AArch64::LDURBi:
3832 return AArch64::LDRBui;
3833 case AArch64::LDURBBi:
3835 return AArch64::LDRBBui;
3836 case AArch64::LDURSBXi:
3838 return AArch64::LDRSBXui;
3839 case AArch64::LDURSBWi:
3841 return AArch64::LDRSBWui;
3842 case AArch64::STURBi:
3844 return AArch64::STRBui;
3845 case AArch64::STURBBi:
3847 return AArch64::STRBBui;
3848 case AArch64::LDRQui:
3849 case AArch64::STRQui:
3852 case AArch64::LDRDui:
3853 case AArch64::STRDui:
3854 case AArch64::LDRXui:
3855 case AArch64::STRXui:
3858 case AArch64::LDRWui:
3859 case AArch64::LDRSWui:
3860 case AArch64::STRWui:
3863 case AArch64::LDRHui:
3864 case AArch64::STRHui:
3865 case AArch64::LDRHHui:
3866 case AArch64::STRHHui:
3867 case AArch64::LDRSHXui:
3868 case AArch64::LDRSHWui:
3871 case AArch64::LDRBui:
3872 case AArch64::LDRBBui:
3873 case AArch64::LDRSBXui:
3874 case AArch64::LDRSBWui:
3875 case AArch64::STRBui:
3876 case AArch64::STRBBui:
3890 case AArch64::LDURQi:
3891 case AArch64::STURQi:
3892 case AArch64::LDURDi:
3893 case AArch64::STURDi:
3894 case AArch64::LDURXi:
3895 case AArch64::STURXi:
3896 case AArch64::LDURWi:
3897 case AArch64::LDURSWi:
3898 case AArch64::STURWi:
3899 case AArch64::LDURHi:
3900 case AArch64::STURHi:
3901 case AArch64::LDURHHi:
3902 case AArch64::STURHHi:
3903 case AArch64::LDURSHXi:
3904 case AArch64::LDURSHWi:
3905 case AArch64::LDURBi:
3906 case AArch64::STURBi:
3907 case AArch64::LDURBBi:
3908 case AArch64::STURBBi:
3909 case AArch64::LDURSBWi:
3910 case AArch64::LDURSBXi:
3912 case AArch64::LDRQui:
3913 return AArch64::LDURQi;
3914 case AArch64::STRQui:
3915 return AArch64::STURQi;
3916 case AArch64::LDRDui:
3917 return AArch64::LDURDi;
3918 case AArch64::STRDui:
3919 return AArch64::STURDi;
3920 case AArch64::LDRXui:
3921 return AArch64::LDURXi;
3922 case AArch64::STRXui:
3923 return AArch64::STURXi;
3924 case AArch64::LDRWui:
3925 return AArch64::LDURWi;
3926 case AArch64::LDRSWui:
3927 return AArch64::LDURSWi;
3928 case AArch64::STRWui:
3929 return AArch64::STURWi;
3930 case AArch64::LDRHui:
3931 return AArch64::LDURHi;
3932 case AArch64::STRHui:
3933 return AArch64::STURHi;
3934 case AArch64::LDRHHui:
3935 return AArch64::LDURHHi;
3936 case AArch64::STRHHui:
3937 return AArch64::STURHHi;
3938 case AArch64::LDRSHXui:
3939 return AArch64::LDURSHXi;
3940 case AArch64::LDRSHWui:
3941 return AArch64::LDURSHWi;
3942 case AArch64::LDRBBui:
3943 return AArch64::LDURBBi;
3944 case AArch64::LDRBui:
3945 return AArch64::LDURBi;
3946 case AArch64::STRBBui:
3947 return AArch64::STURBBi;
3948 case AArch64::STRBui:
3949 return AArch64::STURBi;
3950 case AArch64::LDRSBWui:
3951 return AArch64::LDURSBWi;
3952 case AArch64::LDRSBXui:
3953 return AArch64::LDURSBXi;
3966 case AArch64::LDRQroX:
3967 case AArch64::LDURQi:
3968 case AArch64::LDRQui:
3969 return AArch64::LDRQroW;
3970 case AArch64::STRQroX:
3971 case AArch64::STURQi:
3972 case AArch64::STRQui:
3973 return AArch64::STRQroW;
3974 case AArch64::LDRDroX:
3975 case AArch64::LDURDi:
3976 case AArch64::LDRDui:
3977 return AArch64::LDRDroW;
3978 case AArch64::STRDroX:
3979 case AArch64::STURDi:
3980 case AArch64::STRDui:
3981 return AArch64::STRDroW;
3982 case AArch64::LDRXroX:
3983 case AArch64::LDURXi:
3984 case AArch64::LDRXui:
3985 return AArch64::LDRXroW;
3986 case AArch64::STRXroX:
3987 case AArch64::STURXi:
3988 case AArch64::STRXui:
3989 return AArch64::STRXroW;
3990 case AArch64::LDRWroX:
3991 case AArch64::LDURWi:
3992 case AArch64::LDRWui:
3993 return AArch64::LDRWroW;
3994 case AArch64::LDRSWroX:
3995 case AArch64::LDURSWi:
3996 case AArch64::LDRSWui:
3997 return AArch64::LDRSWroW;
3998 case AArch64::STRWroX:
3999 case AArch64::STURWi:
4000 case AArch64::STRWui:
4001 return AArch64::STRWroW;
4002 case AArch64::LDRHroX:
4003 case AArch64::LDURHi:
4004 case AArch64::LDRHui:
4005 return AArch64::LDRHroW;
4006 case AArch64::STRHroX:
4007 case AArch64::STURHi:
4008 case AArch64::STRHui:
4009 return AArch64::STRHroW;
4010 case AArch64::LDRHHroX:
4011 case AArch64::LDURHHi:
4012 case AArch64::LDRHHui:
4013 return AArch64::LDRHHroW;
4014 case AArch64::STRHHroX:
4015 case AArch64::STURHHi:
4016 case AArch64::STRHHui:
4017 return AArch64::STRHHroW;
4018 case AArch64::LDRSHXroX:
4019 case AArch64::LDURSHXi:
4020 case AArch64::LDRSHXui:
4021 return AArch64::LDRSHXroW;
4022 case AArch64::LDRSHWroX:
4023 case AArch64::LDURSHWi:
4024 case AArch64::LDRSHWui:
4025 return AArch64::LDRSHWroW;
4026 case AArch64::LDRBroX:
4027 case AArch64::LDURBi:
4028 case AArch64::LDRBui:
4029 return AArch64::LDRBroW;
4030 case AArch64::LDRBBroX:
4031 case AArch64::LDURBBi:
4032 case AArch64::LDRBBui:
4033 return AArch64::LDRBBroW;
4034 case AArch64::LDRSBXroX:
4035 case AArch64::LDURSBXi:
4036 case AArch64::LDRSBXui:
4037 return AArch64::LDRSBXroW;
4038 case AArch64::LDRSBWroX:
4039 case AArch64::LDURSBWi:
4040 case AArch64::LDRSBWui:
4041 return AArch64::LDRSBWroW;
4042 case AArch64::STRBroX:
4043 case AArch64::STURBi:
4044 case AArch64::STRBui:
4045 return AArch64::STRBroW;
4046 case AArch64::STRBBroX:
4047 case AArch64::STURBBi:
4048 case AArch64::STRBBui:
4049 return AArch64::STRBBroW;
4064 MRI.constrainRegClass(AM.
BaseReg, &AArch64::GPR64spRegClass);
4074 return B.getInstr();
4078 "Addressing mode not supported for folding");
4095 return B.getInstr();
4102 "Address offset can be a register or an immediate, but not both");
4104 MRI.constrainRegClass(AM.
BaseReg, &AArch64::GPR64spRegClass);
4109 OffsetReg =
MRI.createVirtualRegister(&AArch64::GPR32RegClass);
4123 return B.getInstr();
4127 "Function must not be called with an addressing mode it can't handle");
4136 case AArch64::LD1Fourv16b_POST:
4137 case AArch64::LD1Fourv1d_POST:
4138 case AArch64::LD1Fourv2d_POST:
4139 case AArch64::LD1Fourv2s_POST:
4140 case AArch64::LD1Fourv4h_POST:
4141 case AArch64::LD1Fourv4s_POST:
4142 case AArch64::LD1Fourv8b_POST:
4143 case AArch64::LD1Fourv8h_POST:
4144 case AArch64::LD1Onev16b_POST:
4145 case AArch64::LD1Onev1d_POST:
4146 case AArch64::LD1Onev2d_POST:
4147 case AArch64::LD1Onev2s_POST:
4148 case AArch64::LD1Onev4h_POST:
4149 case AArch64::LD1Onev4s_POST:
4150 case AArch64::LD1Onev8b_POST:
4151 case AArch64::LD1Onev8h_POST:
4152 case AArch64::LD1Rv16b_POST:
4153 case AArch64::LD1Rv1d_POST:
4154 case AArch64::LD1Rv2d_POST:
4155 case AArch64::LD1Rv2s_POST:
4156 case AArch64::LD1Rv4h_POST:
4157 case AArch64::LD1Rv4s_POST:
4158 case AArch64::LD1Rv8b_POST:
4159 case AArch64::LD1Rv8h_POST:
4160 case AArch64::LD1Threev16b_POST:
4161 case AArch64::LD1Threev1d_POST:
4162 case AArch64::LD1Threev2d_POST:
4163 case AArch64::LD1Threev2s_POST:
4164 case AArch64::LD1Threev4h_POST:
4165 case AArch64::LD1Threev4s_POST:
4166 case AArch64::LD1Threev8b_POST:
4167 case AArch64::LD1Threev8h_POST:
4168 case AArch64::LD1Twov16b_POST:
4169 case AArch64::LD1Twov1d_POST:
4170 case AArch64::LD1Twov2d_POST:
4171 case AArch64::LD1Twov2s_POST:
4172 case AArch64::LD1Twov4h_POST:
4173 case AArch64::LD1Twov4s_POST:
4174 case AArch64::LD1Twov8b_POST:
4175 case AArch64::LD1Twov8h_POST:
4176 case AArch64::LD1i16_POST:
4177 case AArch64::LD1i32_POST:
4178 case AArch64::LD1i64_POST:
4179 case AArch64::LD1i8_POST:
4180 case AArch64::LD2Rv16b_POST:
4181 case AArch64::LD2Rv1d_POST:
4182 case AArch64::LD2Rv2d_POST:
4183 case AArch64::LD2Rv2s_POST:
4184 case AArch64::LD2Rv4h_POST:
4185 case AArch64::LD2Rv4s_POST:
4186 case AArch64::LD2Rv8b_POST:
4187 case AArch64::LD2Rv8h_POST:
4188 case AArch64::LD2Twov16b_POST:
4189 case AArch64::LD2Twov2d_POST:
4190 case AArch64::LD2Twov2s_POST:
4191 case AArch64::LD2Twov4h_POST:
4192 case AArch64::LD2Twov4s_POST:
4193 case AArch64::LD2Twov8b_POST:
4194 case AArch64::LD2Twov8h_POST:
4195 case AArch64::LD2i16_POST:
4196 case AArch64::LD2i32_POST:
4197 case AArch64::LD2i64_POST:
4198 case AArch64::LD2i8_POST:
4199 case AArch64::LD3Rv16b_POST:
4200 case AArch64::LD3Rv1d_POST:
4201 case AArch64::LD3Rv2d_POST:
4202 case AArch64::LD3Rv2s_POST:
4203 case AArch64::LD3Rv4h_POST:
4204 case AArch64::LD3Rv4s_POST:
4205 case AArch64::LD3Rv8b_POST:
4206 case AArch64::LD3Rv8h_POST:
4207 case AArch64::LD3Threev16b_POST:
4208 case AArch64::LD3Threev2d_POST:
4209 case AArch64::LD3Threev2s_POST:
4210 case AArch64::LD3Threev4h_POST:
4211 case AArch64::LD3Threev4s_POST:
4212 case AArch64::LD3Threev8b_POST:
4213 case AArch64::LD3Threev8h_POST:
4214 case AArch64::LD3i16_POST:
4215 case AArch64::LD3i32_POST:
4216 case AArch64::LD3i64_POST:
4217 case AArch64::LD3i8_POST:
4218 case AArch64::LD4Fourv16b_POST:
4219 case AArch64::LD4Fourv2d_POST:
4220 case AArch64::LD4Fourv2s_POST:
4221 case AArch64::LD4Fourv4h_POST:
4222 case AArch64::LD4Fourv4s_POST:
4223 case AArch64::LD4Fourv8b_POST:
4224 case AArch64::LD4Fourv8h_POST:
4225 case AArch64::LD4Rv16b_POST:
4226 case AArch64::LD4Rv1d_POST:
4227 case AArch64::LD4Rv2d_POST:
4228 case AArch64::LD4Rv2s_POST:
4229 case AArch64::LD4Rv4h_POST:
4230 case AArch64::LD4Rv4s_POST:
4231 case AArch64::LD4Rv8b_POST:
4232 case AArch64::LD4Rv8h_POST:
4233 case AArch64::LD4i16_POST:
4234 case AArch64::LD4i32_POST:
4235 case AArch64::LD4i64_POST:
4236 case AArch64::LD4i8_POST:
4237 case AArch64::LDAPRWpost:
4238 case AArch64::LDAPRXpost:
4239 case AArch64::LDIAPPWpost:
4240 case AArch64::LDIAPPXpost:
4241 case AArch64::LDPDpost:
4242 case AArch64::LDPQpost:
4243 case AArch64::LDPSWpost:
4244 case AArch64::LDPSpost:
4245 case AArch64::LDPWpost:
4246 case AArch64::LDPXpost:
4247 case AArch64::LDRBBpost:
4248 case AArch64::LDRBpost:
4249 case AArch64::LDRDpost:
4250 case AArch64::LDRHHpost:
4251 case AArch64::LDRHpost:
4252 case AArch64::LDRQpost:
4253 case AArch64::LDRSBWpost:
4254 case AArch64::LDRSBXpost:
4255 case AArch64::LDRSHWpost:
4256 case AArch64::LDRSHXpost:
4257 case AArch64::LDRSWpost:
4258 case AArch64::LDRSpost:
4259 case AArch64::LDRWpost:
4260 case AArch64::LDRXpost:
4261 case AArch64::ST1Fourv16b_POST:
4262 case AArch64::ST1Fourv1d_POST:
4263 case AArch64::ST1Fourv2d_POST:
4264 case AArch64::ST1Fourv2s_POST:
4265 case AArch64::ST1Fourv4h_POST:
4266 case AArch64::ST1Fourv4s_POST:
4267 case AArch64::ST1Fourv8b_POST:
4268 case AArch64::ST1Fourv8h_POST:
4269 case AArch64::ST1Onev16b_POST:
4270 case AArch64::ST1Onev1d_POST:
4271 case AArch64::ST1Onev2d_POST:
4272 case AArch64::ST1Onev2s_POST:
4273 case AArch64::ST1Onev4h_POST:
4274 case AArch64::ST1Onev4s_POST:
4275 case AArch64::ST1Onev8b_POST:
4276 case AArch64::ST1Onev8h_POST:
4277 case AArch64::ST1Threev16b_POST:
4278 case AArch64::ST1Threev1d_POST:
4279 case AArch64::ST1Threev2d_POST:
4280 case AArch64::ST1Threev2s_POST:
4281 case AArch64::ST1Threev4h_POST:
4282 case AArch64::ST1Threev4s_POST:
4283 case AArch64::ST1Threev8b_POST:
4284 case AArch64::ST1Threev8h_POST:
4285 case AArch64::ST1Twov16b_POST:
4286 case AArch64::ST1Twov1d_POST:
4287 case AArch64::ST1Twov2d_POST:
4288 case AArch64::ST1Twov2s_POST:
4289 case AArch64::ST1Twov4h_POST:
4290 case AArch64::ST1Twov4s_POST:
4291 case AArch64::ST1Twov8b_POST:
4292 case AArch64::ST1Twov8h_POST:
4293 case AArch64::ST1i16_POST:
4294 case AArch64::ST1i32_POST:
4295 case AArch64::ST1i64_POST:
4296 case AArch64::ST1i8_POST:
4297 case AArch64::ST2GPostIndex:
4298 case AArch64::ST2Twov16b_POST:
4299 case AArch64::ST2Twov2d_POST:
4300 case AArch64::ST2Twov2s_POST:
4301 case AArch64::ST2Twov4h_POST:
4302 case AArch64::ST2Twov4s_POST:
4303 case AArch64::ST2Twov8b_POST:
4304 case AArch64::ST2Twov8h_POST:
4305 case AArch64::ST2i16_POST:
4306 case AArch64::ST2i32_POST:
4307 case AArch64::ST2i64_POST:
4308 case AArch64::ST2i8_POST:
4309 case AArch64::ST3Threev16b_POST:
4310 case AArch64::ST3Threev2d_POST:
4311 case AArch64::ST3Threev2s_POST:
4312 case AArch64::ST3Threev4h_POST:
4313 case AArch64::ST3Threev4s_POST:
4314 case AArch64::ST3Threev8b_POST:
4315 case AArch64::ST3Threev8h_POST:
4316 case AArch64::ST3i16_POST:
4317 case AArch64::ST3i32_POST:
4318 case AArch64::ST3i64_POST:
4319 case AArch64::ST3i8_POST:
4320 case AArch64::ST4Fourv16b_POST:
4321 case AArch64::ST4Fourv2d_POST:
4322 case AArch64::ST4Fourv2s_POST:
4323 case AArch64::ST4Fourv4h_POST:
4324 case AArch64::ST4Fourv4s_POST:
4325 case AArch64::ST4Fourv8b_POST:
4326 case AArch64::ST4Fourv8h_POST:
4327 case AArch64::ST4i16_POST:
4328 case AArch64::ST4i32_POST:
4329 case AArch64::ST4i64_POST:
4330 case AArch64::ST4i8_POST:
4331 case AArch64::STGPostIndex:
4332 case AArch64::STGPpost:
4333 case AArch64::STPDpost:
4334 case AArch64::STPQpost:
4335 case AArch64::STPSpost:
4336 case AArch64::STPWpost:
4337 case AArch64::STPXpost:
4338 case AArch64::STRBBpost:
4339 case AArch64::STRBpost:
4340 case AArch64::STRDpost:
4341 case AArch64::STRHHpost:
4342 case AArch64::STRHpost:
4343 case AArch64::STRQpost:
4344 case AArch64::STRSpost:
4345 case AArch64::STRWpost:
4346 case AArch64::STRXpost:
4347 case AArch64::STZ2GPostIndex:
4348 case AArch64::STZGPostIndex:
4355 bool &OffsetIsScalable,
TypeSize &Width,
4376 int64_t Dummy1, Dummy2;
4398 return BaseOp->
isReg() || BaseOp->
isFI();
4405 assert(OfsOp.
isImm() &&
"Offset operand wasn't immediate.");
4410 TypeSize &Width, int64_t &MinOffset,
4411 int64_t &MaxOffset) {
4417 MinOffset = MaxOffset = 0;
4420 case AArch64::LDRQui:
4421 case AArch64::STRQui:
4427 case AArch64::LDRXui:
4428 case AArch64::LDRDui:
4429 case AArch64::STRXui:
4430 case AArch64::STRDui:
4431 case AArch64::PRFMui:
4437 case AArch64::LDRWui:
4438 case AArch64::LDRSui:
4439 case AArch64::LDRSWui:
4440 case AArch64::STRWui:
4441 case AArch64::STRSui:
4447 case AArch64::LDRHui:
4448 case AArch64::LDRHHui:
4449 case AArch64::LDRSHWui:
4450 case AArch64::LDRSHXui:
4451 case AArch64::STRHui:
4452 case AArch64::STRHHui:
4458 case AArch64::LDRBui:
4459 case AArch64::LDRBBui:
4460 case AArch64::LDRSBWui:
4461 case AArch64::LDRSBXui:
4462 case AArch64::STRBui:
4463 case AArch64::STRBBui:
4470 case AArch64::STRQpre:
4471 case AArch64::LDRQpost:
4477 case AArch64::LDRDpost:
4478 case AArch64::LDRDpre:
4479 case AArch64::LDRXpost:
4480 case AArch64::LDRXpre:
4481 case AArch64::STRDpost:
4482 case AArch64::STRDpre:
4483 case AArch64::STRXpost:
4484 case AArch64::STRXpre:
4490 case AArch64::STRWpost:
4491 case AArch64::STRWpre:
4492 case AArch64::LDRWpost:
4493 case AArch64::LDRWpre:
4494 case AArch64::STRSpost:
4495 case AArch64::STRSpre:
4496 case AArch64::LDRSpost:
4497 case AArch64::LDRSpre:
4503 case AArch64::LDRHpost:
4504 case AArch64::LDRHpre:
4505 case AArch64::STRHpost:
4506 case AArch64::STRHpre:
4507 case AArch64::LDRHHpost:
4508 case AArch64::LDRHHpre:
4509 case AArch64::STRHHpost:
4510 case AArch64::STRHHpre:
4516 case AArch64::LDRBpost:
4517 case AArch64::LDRBpre:
4518 case AArch64::STRBpost:
4519 case AArch64::STRBpre:
4520 case AArch64::LDRBBpost:
4521 case AArch64::LDRBBpre:
4522 case AArch64::STRBBpost:
4523 case AArch64::STRBBpre:
4530 case AArch64::LDURQi:
4531 case AArch64::STURQi:
4537 case AArch64::LDURXi:
4538 case AArch64::LDURDi:
4539 case AArch64::LDAPURXi:
4540 case AArch64::STURXi:
4541 case AArch64::STURDi:
4542 case AArch64::STLURXi:
4543 case AArch64::PRFUMi:
4549 case AArch64::LDURWi:
4550 case AArch64::LDURSi:
4551 case AArch64::LDURSWi:
4552 case AArch64::LDAPURi:
4553 case AArch64::LDAPURSWi:
4554 case AArch64::STURWi:
4555 case AArch64::STURSi:
4556 case AArch64::STLURWi:
4562 case AArch64::LDURHi:
4563 case AArch64::LDURHHi:
4564 case AArch64::LDURSHXi:
4565 case AArch64::LDURSHWi:
4566 case AArch64::LDAPURHi:
4567 case AArch64::LDAPURSHWi:
4568 case AArch64::LDAPURSHXi:
4569 case AArch64::STURHi:
4570 case AArch64::STURHHi:
4571 case AArch64::STLURHi:
4577 case AArch64::LDURBi:
4578 case AArch64::LDURBBi:
4579 case AArch64::LDURSBXi:
4580 case AArch64::LDURSBWi:
4581 case AArch64::LDAPURBi:
4582 case AArch64::LDAPURSBWi:
4583 case AArch64::LDAPURSBXi:
4584 case AArch64::STURBi:
4585 case AArch64::STURBBi:
4586 case AArch64::STLURBi:
4593 case AArch64::LDPQi:
4594 case AArch64::LDNPQi:
4595 case AArch64::STPQi:
4596 case AArch64::STNPQi:
4597 case AArch64::LDPQpost:
4598 case AArch64::LDPQpre:
4599 case AArch64::STPQpost:
4600 case AArch64::STPQpre:
4606 case AArch64::LDPXi:
4607 case AArch64::LDPDi:
4608 case AArch64::LDNPXi:
4609 case AArch64::LDNPDi:
4610 case AArch64::STPXi:
4611 case AArch64::STPDi:
4612 case AArch64::STNPXi:
4613 case AArch64::STNPDi:
4614 case AArch64::LDPDpost:
4615 case AArch64::LDPDpre:
4616 case AArch64::LDPXpost:
4617 case AArch64::LDPXpre:
4618 case AArch64::STPDpost:
4619 case AArch64::STPDpre:
4620 case AArch64::STPXpost:
4621 case AArch64::STPXpre:
4627 case AArch64::LDPWi:
4628 case AArch64::LDPSi:
4629 case AArch64::LDNPWi:
4630 case AArch64::LDNPSi:
4631 case AArch64::STPWi:
4632 case AArch64::STPSi:
4633 case AArch64::STNPWi:
4634 case AArch64::STNPSi:
4635 case AArch64::LDPSpost:
4636 case AArch64::LDPSpre:
4637 case AArch64::LDPWpost:
4638 case AArch64::LDPWpre:
4639 case AArch64::STPSpost:
4640 case AArch64::STPSpre:
4641 case AArch64::STPWpost:
4642 case AArch64::STPWpre:
4648 case AArch64::StoreSwiftAsyncContext:
4661 case AArch64::TAGPstack:
4671 case AArch64::STGPreIndex:
4672 case AArch64::STGPostIndex:
4673 case AArch64::STZGi:
4674 case AArch64::STZGPreIndex:
4675 case AArch64::STZGPostIndex:
4682 case AArch64::STR_ZZZZXI:
4683 case AArch64::STR_ZZZZXI_STRIDED_CONTIGUOUS:
4684 case AArch64::LDR_ZZZZXI:
4685 case AArch64::LDR_ZZZZXI_STRIDED_CONTIGUOUS:
4691 case AArch64::STR_ZZZXI:
4692 case AArch64::LDR_ZZZXI:
4698 case AArch64::STR_ZZXI:
4699 case AArch64::STR_ZZXI_STRIDED_CONTIGUOUS:
4700 case AArch64::LDR_ZZXI:
4701 case AArch64::LDR_ZZXI_STRIDED_CONTIGUOUS:
4707 case AArch64::LDR_PXI:
4708 case AArch64::STR_PXI:
4714 case AArch64::LDR_PPXI:
4715 case AArch64::STR_PPXI:
4721 case AArch64::LDR_ZXI:
4722 case AArch64::STR_ZXI:
4728 case AArch64::LD1B_IMM:
4729 case AArch64::LD1H_IMM:
4730 case AArch64::LD1W_IMM:
4731 case AArch64::LD1D_IMM:
4732 case AArch64::LDNT1B_ZRI:
4733 case AArch64::LDNT1H_ZRI:
4734 case AArch64::LDNT1W_ZRI:
4735 case AArch64::LDNT1D_ZRI:
4736 case AArch64::ST1B_IMM:
4737 case AArch64::ST1H_IMM:
4738 case AArch64::ST1W_IMM:
4739 case AArch64::ST1D_IMM:
4740 case AArch64::STNT1B_ZRI:
4741 case AArch64::STNT1H_ZRI:
4742 case AArch64::STNT1W_ZRI:
4743 case AArch64::STNT1D_ZRI:
4744 case AArch64::LDNF1B_IMM:
4745 case AArch64::LDNF1H_IMM:
4746 case AArch64::LDNF1W_IMM:
4747 case AArch64::LDNF1D_IMM:
4755 case AArch64::LD2B_IMM:
4756 case AArch64::LD2H_IMM:
4757 case AArch64::LD2W_IMM:
4758 case AArch64::LD2D_IMM:
4759 case AArch64::ST2B_IMM:
4760 case AArch64::ST2H_IMM:
4761 case AArch64::ST2W_IMM:
4762 case AArch64::ST2D_IMM:
4768 case AArch64::LD3B_IMM:
4769 case AArch64::LD3H_IMM:
4770 case AArch64::LD3W_IMM:
4771 case AArch64::LD3D_IMM:
4772 case AArch64::ST3B_IMM:
4773 case AArch64::ST3H_IMM:
4774 case AArch64::ST3W_IMM:
4775 case AArch64::ST3D_IMM:
4781 case AArch64::LD4B_IMM:
4782 case AArch64::LD4H_IMM:
4783 case AArch64::LD4W_IMM:
4784 case AArch64::LD4D_IMM:
4785 case AArch64::ST4B_IMM:
4786 case AArch64::ST4H_IMM:
4787 case AArch64::ST4W_IMM:
4788 case AArch64::ST4D_IMM:
4794 case AArch64::LD1B_H_IMM:
4795 case AArch64::LD1SB_H_IMM:
4796 case AArch64::LD1H_S_IMM:
4797 case AArch64::LD1SH_S_IMM:
4798 case AArch64::LD1W_D_IMM:
4799 case AArch64::LD1SW_D_IMM:
4800 case AArch64::ST1B_H_IMM:
4801 case AArch64::ST1H_S_IMM:
4802 case AArch64::ST1W_D_IMM:
4803 case AArch64::LDNF1B_H_IMM:
4804 case AArch64::LDNF1SB_H_IMM:
4805 case AArch64::LDNF1H_S_IMM:
4806 case AArch64::LDNF1SH_S_IMM:
4807 case AArch64::LDNF1W_D_IMM:
4808 case AArch64::LDNF1SW_D_IMM:
4816 case AArch64::LD1B_S_IMM:
4817 case AArch64::LD1SB_S_IMM:
4818 case AArch64::LD1H_D_IMM:
4819 case AArch64::LD1SH_D_IMM:
4820 case AArch64::ST1B_S_IMM:
4821 case AArch64::ST1H_D_IMM:
4822 case AArch64::LDNF1B_S_IMM:
4823 case AArch64::LDNF1SB_S_IMM:
4824 case AArch64::LDNF1H_D_IMM:
4825 case AArch64::LDNF1SH_D_IMM:
4833 case AArch64::LD1B_D_IMM:
4834 case AArch64::LD1SB_D_IMM:
4835 case AArch64::ST1B_D_IMM:
4836 case AArch64::LDNF1B_D_IMM:
4837 case AArch64::LDNF1SB_D_IMM:
4845 case AArch64::ST2Gi:
4846 case AArch64::ST2GPreIndex:
4847 case AArch64::ST2GPostIndex:
4848 case AArch64::STZ2Gi:
4849 case AArch64::STZ2GPreIndex:
4850 case AArch64::STZ2GPostIndex:
4856 case AArch64::STGPi:
4857 case AArch64::STGPpost:
4858 case AArch64::STGPpre:
4864 case AArch64::LD1RB_IMM:
4865 case AArch64::LD1RB_H_IMM:
4866 case AArch64::LD1RB_S_IMM:
4867 case AArch64::LD1RB_D_IMM:
4868 case AArch64::LD1RSB_H_IMM:
4869 case AArch64::LD1RSB_S_IMM:
4870 case AArch64::LD1RSB_D_IMM:
4876 case AArch64::LD1RH_IMM:
4877 case AArch64::LD1RH_S_IMM:
4878 case AArch64::LD1RH_D_IMM:
4879 case AArch64::LD1RSH_S_IMM:
4880 case AArch64::LD1RSH_D_IMM:
4886 case AArch64::LD1RW_IMM:
4887 case AArch64::LD1RW_D_IMM:
4888 case AArch64::LD1RSW_IMM:
4894 case AArch64::LD1RD_IMM:
4910 case AArch64::LDRBBui:
4911 case AArch64::LDURBBi:
4912 case AArch64::LDRSBWui:
4913 case AArch64::LDURSBWi:
4914 case AArch64::STRBBui:
4915 case AArch64::STURBBi:
4917 case AArch64::LDRHHui:
4918 case AArch64::LDURHHi:
4919 case AArch64::LDRSHWui:
4920 case AArch64::LDURSHWi:
4921 case AArch64::STRHHui:
4922 case AArch64::STURHHi:
4924 case AArch64::LDRSui:
4925 case AArch64::LDURSi:
4926 case AArch64::LDRSpre:
4927 case AArch64::LDRSWui:
4928 case AArch64::LDURSWi:
4929 case AArch64::LDRSWpre:
4930 case AArch64::LDRWpre:
4931 case AArch64::LDRWui:
4932 case AArch64::LDURWi:
4933 case AArch64::STRSui:
4934 case AArch64::STURSi:
4935 case AArch64::STRSpre:
4936 case AArch64::STRWui:
4937 case AArch64::STURWi:
4938 case AArch64::STRWpre:
4939 case AArch64::LDPSi:
4940 case AArch64::LDPSWi:
4941 case AArch64::LDPWi:
4942 case AArch64::STPSi:
4943 case AArch64::STPWi:
4945 case AArch64::LDRDui:
4946 case AArch64::LDURDi:
4947 case AArch64::LDRDpre:
4948 case AArch64::LDRXui:
4949 case AArch64::LDURXi:
4950 case AArch64::LDRXpre:
4951 case AArch64::STRDui:
4952 case AArch64::STURDi:
4953 case AArch64::STRDpre:
4954 case AArch64::STRXui:
4955 case AArch64::STURXi:
4956 case AArch64::STRXpre:
4957 case AArch64::LDPDi:
4958 case AArch64::LDPXi:
4959 case AArch64::STPDi:
4960 case AArch64::STPXi:
4962 case AArch64::LDRQui:
4963 case AArch64::LDURQi:
4964 case AArch64::STRQui:
4965 case AArch64::STURQi:
4966 case AArch64::STRQpre:
4967 case AArch64::LDPQi:
4968 case AArch64::LDRQpre:
4969 case AArch64::STPQi:
4971 case AArch64::STZGi:
4972 case AArch64::ST2Gi:
4973 case AArch64::STZ2Gi:
4974 case AArch64::STGPi:
4980 switch (
MI.getOpcode()) {
4983 case AArch64::LDRWpre:
4984 case AArch64::LDRXpre:
4985 case AArch64::LDRSWpre:
4986 case AArch64::LDRSpre:
4987 case AArch64::LDRDpre:
4988 case AArch64::LDRQpre:
4994 switch (
MI.getOpcode()) {
4997 case AArch64::STRWpre:
4998 case AArch64::STRXpre:
4999 case AArch64::STRSpre:
5000 case AArch64::STRDpre:
5001 case AArch64::STRQpre:
5011 switch (
MI.getOpcode()) {
5014 case AArch64::LDPSi:
5015 case AArch64::LDPSWi:
5016 case AArch64::LDPDi:
5017 case AArch64::LDPQi:
5018 case AArch64::LDPWi:
5019 case AArch64::LDPXi:
5020 case AArch64::STPSi:
5021 case AArch64::STPDi:
5022 case AArch64::STPQi:
5023 case AArch64::STPWi:
5024 case AArch64::STPXi:
5025 case AArch64::STGPi:
5031 assert(
MI.mayLoadOrStore() &&
"Load or store instruction expected");
5035 return MI.getOperand(Idx);
5040 assert(
MI.mayLoadOrStore() &&
"Load or store instruction expected");
5044 return MI.getOperand(Idx);
5049 switch (
MI.getOpcode()) {
5052 case AArch64::LDRBroX:
5053 case AArch64::LDRBBroX:
5054 case AArch64::LDRSBXroX:
5055 case AArch64::LDRSBWroX:
5056 case AArch64::LDRHroX:
5057 case AArch64::LDRHHroX:
5058 case AArch64::LDRSHXroX:
5059 case AArch64::LDRSHWroX:
5060 case AArch64::LDRWroX:
5061 case AArch64::LDRSroX:
5062 case AArch64::LDRSWroX:
5063 case AArch64::LDRDroX:
5064 case AArch64::LDRXroX:
5065 case AArch64::LDRQroX:
5066 return MI.getOperand(4);
5072 if (
MI.getParent() ==
nullptr)
5082 auto Reg =
Op.getReg();
5083 if (Reg.isPhysical())
5084 return AArch64::FPR16RegClass.contains(Reg);
5086 return TRC == &AArch64::FPR16RegClass ||
5087 TRC == &AArch64::FPR16_loRegClass;
5096 auto Reg =
Op.getReg();
5097 if (Reg.isPhysical())
5098 return AArch64::FPR128RegClass.contains(Reg);
5100 return TRC == &AArch64::FPR128RegClass ||
5101 TRC == &AArch64::FPR128_loRegClass;
5107 switch (
MI.getOpcode()) {
5110 case AArch64::PACIASP:
5111 case AArch64::PACIBSP:
5114 case AArch64::PAUTH_PROLOGUE:
5117 case AArch64::HINT: {
5118 unsigned Imm =
MI.getOperand(0).getImm();
5120 if (Imm == 32 || Imm == 34 || Imm == 36 || Imm == 38)
5123 if (Imm == 25 || Imm == 27)
5135 assert(Reg.isPhysical() &&
"Expected physical register in isFpOrNEON");
5136 return AArch64::FPR128RegClass.contains(Reg) ||
5137 AArch64::FPR64RegClass.contains(Reg) ||
5138 AArch64::FPR32RegClass.contains(Reg) ||
5139 AArch64::FPR16RegClass.contains(Reg) ||
5140 AArch64::FPR8RegClass.contains(Reg);
5147 auto Reg =
Op.getReg();
5148 if (Reg.isPhysical())
5152 return TRC == &AArch64::FPR128RegClass ||
5153 TRC == &AArch64::FPR128_loRegClass ||
5154 TRC == &AArch64::FPR64RegClass ||
5155 TRC == &AArch64::FPR64_loRegClass ||
5156 TRC == &AArch64::FPR32RegClass || TRC == &AArch64::FPR16RegClass ||
5157 TRC == &AArch64::FPR8RegClass;
5179 if (FirstOpc == SecondOpc)
5185 case AArch64::STRSui:
5186 case AArch64::STURSi:
5187 return SecondOpc == AArch64::STRSui || SecondOpc == AArch64::STURSi;
5188 case AArch64::STRDui:
5189 case AArch64::STURDi:
5190 return SecondOpc == AArch64::STRDui || SecondOpc == AArch64::STURDi;
5191 case AArch64::STRQui:
5192 case AArch64::STURQi:
5193 return SecondOpc == AArch64::STRQui || SecondOpc == AArch64::STURQi;
5194 case AArch64::STRWui:
5195 case AArch64::STURWi:
5196 return SecondOpc == AArch64::STRWui || SecondOpc == AArch64::STURWi;
5197 case AArch64::STRXui:
5198 case AArch64::STURXi:
5199 return SecondOpc == AArch64::STRXui || SecondOpc == AArch64::STURXi;
5200 case AArch64::LDRSui:
5201 case AArch64::LDURSi:
5202 return SecondOpc == AArch64::LDRSui || SecondOpc == AArch64::LDURSi;
5203 case AArch64::LDRDui:
5204 case AArch64::LDURDi:
5205 return SecondOpc == AArch64::LDRDui || SecondOpc == AArch64::LDURDi;
5206 case AArch64::LDRQui:
5207 case AArch64::LDURQi:
5208 return SecondOpc == AArch64::LDRQui || SecondOpc == AArch64::LDURQi;
5209 case AArch64::LDRWui:
5210 case AArch64::LDURWi:
5211 return SecondOpc == AArch64::LDRSWui || SecondOpc == AArch64::LDURSWi;
5212 case AArch64::LDRSWui:
5213 case AArch64::LDURSWi:
5214 return SecondOpc == AArch64::LDRWui || SecondOpc == AArch64::LDURWi;
5215 case AArch64::LDRXui:
5216 case AArch64::LDURXi:
5217 return SecondOpc == AArch64::LDRXui || SecondOpc == AArch64::LDURXi;
5224 int64_t Offset1,
unsigned Opcode1,
int FI2,
5225 int64_t Offset2,
unsigned Opcode2) {
5231 assert(ObjectOffset1 <= ObjectOffset2 &&
"Object offsets are not ordered.");
5234 if (ObjectOffset1 % Scale1 != 0)
5236 ObjectOffset1 /= Scale1;
5238 if (ObjectOffset2 % Scale2 != 0)
5240 ObjectOffset2 /= Scale2;
5241 ObjectOffset1 += Offset1;
5242 ObjectOffset2 += Offset2;
5243 return ObjectOffset1 + 1 == ObjectOffset2;
5255 int64_t OpOffset2,
bool OffsetIsScalable2,
unsigned ClusterSize,
5256 unsigned NumBytes)
const {
5266 "Only base registers and frame indices are supported.");
5273 if (ClusterSize > 2)
5280 unsigned FirstOpc = FirstLdSt.
getOpcode();
5281 unsigned SecondOpc = SecondLdSt.
getOpcode();
5301 if (Offset1 > 63 || Offset1 < -64)
5306 if (BaseOp1.
isFI()) {
5308 "Caller should have ordered offsets.");
5313 BaseOp2.
getIndex(), Offset2, SecondOpc);
5316 assert(Offset1 <= Offset2 &&
"Caller should have ordered offsets.");
5318 return Offset1 + 1 == Offset2;
5328 if (
Reg.isPhysical())
5337 return ((DestReg - SrcReg) & 0x1f) < NumRegs;
5346 assert(Subtarget.hasNEON() &&
"Unexpected register copy without NEON");
5348 uint16_t DestEncoding =
TRI->getEncodingValue(DestReg);
5349 uint16_t SrcEncoding =
TRI->getEncodingValue(SrcReg);
5350 unsigned NumRegs = Indices.
size();
5352 int SubReg = 0, End = NumRegs, Incr = 1;
5371 unsigned Opcode,
unsigned ZeroReg,
5374 unsigned NumRegs = Indices.
size();
5377 uint16_t DestEncoding =
TRI->getEncodingValue(DestReg);
5378 uint16_t SrcEncoding =
TRI->getEncodingValue(SrcReg);
5379 assert(DestEncoding % NumRegs == 0 && SrcEncoding % NumRegs == 0 &&
5380 "GPR reg sequences should not be able to overlap");
5404 unsigned Opc =
MI.getOpcode();
5405 if (
Opc == AArch64::MSRpstatesvcrImm1 ||
Opc == AArch64::MSRpstatePseudo) {
5407 int64_t PState =
MI.getOperand(0).getImm();
5408 if (PState == AArch64SVCR::SVCRSM || PState == AArch64SVCR::SVCRSMZA) {
5410 return MI.getOperand(1).getImm() == 1;
5429 bool RenamableSrc)
const {
5431 if (AArch64::GPR32spRegClass.
contains(DestReg) &&
5432 AArch64::GPR32spRegClass.
contains(SrcReg)) {
5433 if (DestReg == AArch64::WSP || SrcReg == AArch64::WSP) {
5435 if (Subtarget.hasZeroCycleRegMoveGPR64() &&
5436 !Subtarget.hasZeroCycleRegMoveGPR32()) {
5438 MCRegister DestRegX = RI.getMatchingSuperReg(DestReg, AArch64::sub_32,
5439 &AArch64::GPR64spRegClass);
5440 MCRegister SrcRegX = RI.getMatchingSuperReg(SrcReg, AArch64::sub_32,
5441 &AArch64::GPR64spRegClass);
5451 ++NumZCRegMoveInstrsGPR;
5457 if (Subtarget.hasZeroCycleRegMoveGPR32())
5458 ++NumZCRegMoveInstrsGPR;
5460 }
else if (Subtarget.hasZeroCycleRegMoveGPR64() &&
5461 !Subtarget.hasZeroCycleRegMoveGPR32()) {
5463 MCRegister DestRegX = RI.getMatchingSuperReg(DestReg, AArch64::sub_32,
5464 &AArch64::GPR64spRegClass);
5465 assert(DestRegX.
isValid() &&
"Destination super-reg not valid");
5466 MCRegister SrcRegX = RI.getMatchingSuperReg(SrcReg, AArch64::sub_32,
5467 &AArch64::GPR64spRegClass);
5477 ++NumZCRegMoveInstrsGPR;
5483 if (Subtarget.hasZeroCycleRegMoveGPR32())
5484 ++NumZCRegMoveInstrsGPR;
5490 if (AArch64::GPR32spRegClass.
contains(DestReg) && SrcReg == AArch64::WZR) {
5491 if (Subtarget.hasZeroCycleZeroingGPR64() &&
5492 !Subtarget.hasZeroCycleZeroingGPR32()) {
5493 MCRegister DestRegX = RI.getMatchingSuperReg(DestReg, AArch64::sub_32,
5494 &AArch64::GPR64spRegClass);
5495 assert(DestRegX.
isValid() &&
"Destination super-reg not valid");
5499 ++NumZCZeroingInstrsGPR;
5500 }
else if (Subtarget.hasZeroCycleZeroingGPR32()) {
5504 ++NumZCZeroingInstrsGPR;
5513 if (AArch64::GPR64spRegClass.
contains(DestReg) &&
5514 AArch64::GPR64spRegClass.
contains(SrcReg)) {
5515 if (DestReg == AArch64::SP || SrcReg == AArch64::SP) {
5521 if (Subtarget.hasZeroCycleRegMoveGPR64())
5522 ++NumZCRegMoveInstrsGPR;
5528 if (Subtarget.hasZeroCycleRegMoveGPR64())
5529 ++NumZCRegMoveInstrsGPR;
5535 if (AArch64::GPR64spRegClass.
contains(DestReg) && SrcReg == AArch64::XZR) {
5536 if (Subtarget.hasZeroCycleZeroingGPR64()) {
5540 ++NumZCZeroingInstrsGPR;
5550 if (AArch64::PPRRegClass.
contains(DestReg) &&
5551 AArch64::PPRRegClass.
contains(SrcReg)) {
5552 assert(Subtarget.isSVEorStreamingSVEAvailable() &&
5553 "Unexpected SVE register.");
5563 bool DestIsPNR = AArch64::PNRRegClass.contains(DestReg);
5564 bool SrcIsPNR = AArch64::PNRRegClass.contains(SrcReg);
5565 if (DestIsPNR || SrcIsPNR) {
5567 return (R - AArch64::PN0) + AArch64::P0;
5572 if (PPRSrcReg != PPRDestReg) {
5584 if (AArch64::ZPRRegClass.
contains(DestReg) &&
5585 AArch64::ZPRRegClass.
contains(SrcReg)) {
5586 assert(Subtarget.isSVEorStreamingSVEAvailable() &&
5587 "Unexpected SVE register.");
5595 if ((AArch64::ZPR2RegClass.
contains(DestReg) ||
5596 AArch64::ZPR2StridedOrContiguousRegClass.
contains(DestReg)) &&
5597 (AArch64::ZPR2RegClass.
contains(SrcReg) ||
5598 AArch64::ZPR2StridedOrContiguousRegClass.
contains(SrcReg))) {
5599 assert(Subtarget.isSVEorStreamingSVEAvailable() &&
5600 "Unexpected SVE register.");
5601 static const unsigned Indices[] = {AArch64::zsub0, AArch64::zsub1};
5608 if (AArch64::ZPR3RegClass.
contains(DestReg) &&
5609 AArch64::ZPR3RegClass.
contains(SrcReg)) {
5610 assert(Subtarget.isSVEorStreamingSVEAvailable() &&
5611 "Unexpected SVE register.");
5612 static const unsigned Indices[] = {AArch64::zsub0, AArch64::zsub1,
5620 if ((AArch64::ZPR4RegClass.
contains(DestReg) ||
5621 AArch64::ZPR4StridedOrContiguousRegClass.
contains(DestReg)) &&
5622 (AArch64::ZPR4RegClass.
contains(SrcReg) ||
5623 AArch64::ZPR4StridedOrContiguousRegClass.
contains(SrcReg))) {
5624 assert(Subtarget.isSVEorStreamingSVEAvailable() &&
5625 "Unexpected SVE register.");
5626 static const unsigned Indices[] = {AArch64::zsub0, AArch64::zsub1,
5627 AArch64::zsub2, AArch64::zsub3};
5634 if (AArch64::DDDDRegClass.
contains(DestReg) &&
5635 AArch64::DDDDRegClass.
contains(SrcReg)) {
5636 static const unsigned Indices[] = {AArch64::dsub0, AArch64::dsub1,
5637 AArch64::dsub2, AArch64::dsub3};
5644 if (AArch64::DDDRegClass.
contains(DestReg) &&
5645 AArch64::DDDRegClass.
contains(SrcReg)) {
5646 static const unsigned Indices[] = {AArch64::dsub0, AArch64::dsub1,
5654 if (AArch64::DDRegClass.
contains(DestReg) &&
5655 AArch64::DDRegClass.
contains(SrcReg)) {
5656 static const unsigned Indices[] = {AArch64::dsub0, AArch64::dsub1};
5663 if (AArch64::QQQQRegClass.
contains(DestReg) &&
5664 AArch64::QQQQRegClass.
contains(SrcReg)) {
5665 static const unsigned Indices[] = {AArch64::qsub0, AArch64::qsub1,
5666 AArch64::qsub2, AArch64::qsub3};
5673 if (AArch64::QQQRegClass.
contains(DestReg) &&
5674 AArch64::QQQRegClass.
contains(SrcReg)) {
5675 static const unsigned Indices[] = {AArch64::qsub0, AArch64::qsub1,
5683 if (AArch64::QQRegClass.
contains(DestReg) &&
5684 AArch64::QQRegClass.
contains(SrcReg)) {
5685 static const unsigned Indices[] = {AArch64::qsub0, AArch64::qsub1};
5691 if (AArch64::XSeqPairsClassRegClass.
contains(DestReg) &&
5692 AArch64::XSeqPairsClassRegClass.
contains(SrcReg)) {
5693 static const unsigned Indices[] = {AArch64::sube64, AArch64::subo64};
5695 AArch64::XZR, Indices);
5699 if (AArch64::WSeqPairsClassRegClass.
contains(DestReg) &&
5700 AArch64::WSeqPairsClassRegClass.
contains(SrcReg)) {
5701 static const unsigned Indices[] = {AArch64::sube32, AArch64::subo32};
5703 AArch64::WZR, Indices);
5707 if (AArch64::FPR128RegClass.
contains(DestReg) &&
5708 AArch64::FPR128RegClass.
contains(SrcReg)) {
5712 if ((Subtarget.isSVEorStreamingSVEAvailable() &&
5713 !Subtarget.isNeonAvailable()) ||
5717 .
addReg(AArch64::Z0 + (SrcReg - AArch64::Q0))
5718 .
addReg(AArch64::Z0 + (SrcReg - AArch64::Q0));
5719 }
else if (Subtarget.isNeonAvailable()) {
5723 if (Subtarget.hasZeroCycleRegMoveFPR128())
5724 ++NumZCRegMoveInstrsFPR;
5740 if (AArch64::FPR64RegClass.
contains(DestReg) &&
5741 AArch64::FPR64RegClass.
contains(SrcReg)) {
5742 if (Subtarget.hasZeroCycleRegMoveFPR128() &&
5743 !Subtarget.hasZeroCycleRegMoveFPR64() &&
5744 !Subtarget.hasZeroCycleRegMoveFPR32() && Subtarget.isNeonAvailable() &&
5746 MCRegister DestRegQ = RI.getMatchingSuperReg(DestReg, AArch64::dsub,
5747 &AArch64::FPR128RegClass);
5748 MCRegister SrcRegQ = RI.getMatchingSuperReg(SrcReg, AArch64::dsub,
5749 &AArch64::FPR128RegClass);
5758 ++NumZCRegMoveInstrsFPR;
5762 if (Subtarget.hasZeroCycleRegMoveFPR64())
5763 ++NumZCRegMoveInstrsFPR;
5768 if (AArch64::FPR32RegClass.
contains(DestReg) &&
5769 AArch64::FPR32RegClass.
contains(SrcReg)) {
5770 if (Subtarget.hasZeroCycleRegMoveFPR128() &&
5771 !Subtarget.hasZeroCycleRegMoveFPR64() &&
5772 !Subtarget.hasZeroCycleRegMoveFPR32() && Subtarget.isNeonAvailable() &&
5774 MCRegister DestRegQ = RI.getMatchingSuperReg(DestReg, AArch64::ssub,
5775 &AArch64::FPR128RegClass);
5776 MCRegister SrcRegQ = RI.getMatchingSuperReg(SrcReg, AArch64::ssub,
5777 &AArch64::FPR128RegClass);
5786 ++NumZCRegMoveInstrsFPR;
5787 }
else if (Subtarget.hasZeroCycleRegMoveFPR64() &&
5788 !Subtarget.hasZeroCycleRegMoveFPR32()) {
5789 MCRegister DestRegD = RI.getMatchingSuperReg(DestReg, AArch64::ssub,
5790 &AArch64::FPR64RegClass);
5791 MCRegister SrcRegD = RI.getMatchingSuperReg(SrcReg, AArch64::ssub,
5792 &AArch64::FPR64RegClass);
5800 ++NumZCRegMoveInstrsFPR;
5804 if (Subtarget.hasZeroCycleRegMoveFPR32())
5805 ++NumZCRegMoveInstrsFPR;
5810 if (AArch64::FPR16RegClass.
contains(DestReg) &&
5811 AArch64::FPR16RegClass.
contains(SrcReg)) {
5812 if (Subtarget.hasZeroCycleRegMoveFPR128() &&
5813 !Subtarget.hasZeroCycleRegMoveFPR64() &&
5814 !Subtarget.hasZeroCycleRegMoveFPR32() && Subtarget.isNeonAvailable() &&
5816 MCRegister DestRegQ = RI.getMatchingSuperReg(DestReg, AArch64::hsub,
5817 &AArch64::FPR128RegClass);
5818 MCRegister SrcRegQ = RI.getMatchingSuperReg(SrcReg, AArch64::hsub,
5819 &AArch64::FPR128RegClass);
5828 }
else if (Subtarget.hasZeroCycleRegMoveFPR64() &&
5829 !Subtarget.hasZeroCycleRegMoveFPR32()) {
5830 MCRegister DestRegD = RI.getMatchingSuperReg(DestReg, AArch64::hsub,
5831 &AArch64::FPR64RegClass);
5832 MCRegister SrcRegD = RI.getMatchingSuperReg(SrcReg, AArch64::hsub,
5833 &AArch64::FPR64RegClass);
5842 DestReg = RI.getMatchingSuperReg(DestReg, AArch64::hsub,
5843 &AArch64::FPR32RegClass);
5844 SrcReg = RI.getMatchingSuperReg(SrcReg, AArch64::hsub,
5845 &AArch64::FPR32RegClass);
5852 if (AArch64::FPR8RegClass.
contains(DestReg) &&
5853 AArch64::FPR8RegClass.
contains(SrcReg)) {
5854 if (Subtarget.hasZeroCycleRegMoveFPR128() &&
5855 !Subtarget.hasZeroCycleRegMoveFPR64() &&
5856 !Subtarget.hasZeroCycleRegMoveFPR64() && Subtarget.isNeonAvailable() &&
5858 MCRegister DestRegQ = RI.getMatchingSuperReg(DestReg, AArch64::bsub,
5859 &AArch64::FPR128RegClass);
5860 MCRegister SrcRegQ = RI.getMatchingSuperReg(SrcReg, AArch64::bsub,
5861 &AArch64::FPR128RegClass);
5870 }
else if (Subtarget.hasZeroCycleRegMoveFPR64() &&
5871 !Subtarget.hasZeroCycleRegMoveFPR32()) {
5872 MCRegister DestRegD = RI.getMatchingSuperReg(DestReg, AArch64::bsub,
5873 &AArch64::FPR64RegClass);
5874 MCRegister SrcRegD = RI.getMatchingSuperReg(SrcReg, AArch64::bsub,
5875 &AArch64::FPR64RegClass);
5884 DestReg = RI.getMatchingSuperReg(DestReg, AArch64::bsub,
5885 &AArch64::FPR32RegClass);
5886 SrcReg = RI.getMatchingSuperReg(SrcReg, AArch64::bsub,
5887 &AArch64::FPR32RegClass);
5895 if (AArch64::FPR64RegClass.
contains(DestReg) &&
5896 AArch64::GPR64RegClass.
contains(SrcReg)) {
5897 if (AArch64::XZR == SrcReg) {
5905 if (AArch64::GPR64RegClass.
contains(DestReg) &&
5906 AArch64::FPR64RegClass.
contains(SrcReg)) {
5912 if (AArch64::FPR32RegClass.
contains(DestReg) &&
5913 AArch64::GPR32RegClass.
contains(SrcReg)) {
5914 if (AArch64::WZR == SrcReg) {
5922 if (AArch64::GPR32RegClass.
contains(DestReg) &&
5923 AArch64::FPR32RegClass.
contains(SrcReg)) {
5929 if (DestReg == AArch64::NZCV) {
5930 assert(AArch64::GPR64RegClass.
contains(SrcReg) &&
"Invalid NZCV copy");
5932 .
addImm(AArch64SysReg::NZCV)
5938 if (SrcReg == AArch64::NZCV) {
5939 assert(AArch64::GPR64RegClass.
contains(DestReg) &&
"Invalid NZCV copy");
5941 .
addImm(AArch64SysReg::NZCV)
5947 errs() << RI.getRegAsmName(DestReg) <<
" = COPY " << RI.getRegAsmName(SrcReg)
5958 unsigned SubIdx0,
unsigned SubIdx1,
int FI,
5963 SrcReg0 =
TRI.getSubReg(SrcReg, SubIdx0);
5965 SrcReg1 =
TRI.getSubReg(SrcReg, SubIdx1);
5978 Register SrcReg,
bool isKill,
int FI,
5993 switch (RI.getSpillSize(*RC)) {
5995 if (AArch64::FPR8RegClass.hasSubClassEq(RC))
5996 Opc = AArch64::STRBui;
5999 if (AArch64::FPR16RegClass.hasSubClassEq(RC))
6000 Opc = AArch64::STRHui;
6001 else if (AArch64::PNRRegClass.hasSubClassEq(RC) ||
6002 AArch64::PPRRegClass.hasSubClassEq(RC)) {
6003 assert(Subtarget.isSVEorStreamingSVEAvailable() &&
6004 "Unexpected register store without SVE store instructions");
6005 Opc = AArch64::STR_PXI;
6011 if (AArch64::GPR32allRegClass.hasSubClassEq(RC)) {
6012 Opc = AArch64::STRWui;
6016 assert(SrcReg != AArch64::WSP);
6017 }
else if (AArch64::FPR32RegClass.hasSubClassEq(RC))
6018 Opc = AArch64::STRSui;
6019 else if (AArch64::PPR2RegClass.hasSubClassEq(RC)) {
6020 Opc = AArch64::STR_PPXI;
6025 if (AArch64::GPR64allRegClass.hasSubClassEq(RC)) {
6026 Opc = AArch64::STRXui;
6030 assert(SrcReg != AArch64::SP);
6031 }
else if (AArch64::FPR64RegClass.hasSubClassEq(RC)) {
6032 Opc = AArch64::STRDui;
6033 }
else if (AArch64::WSeqPairsClassRegClass.hasSubClassEq(RC)) {
6035 get(AArch64::STPWi), SrcReg, isKill,
6036 AArch64::sube32, AArch64::subo32, FI, MMO);
6041 if (AArch64::FPR128RegClass.hasSubClassEq(RC))
6042 Opc = AArch64::STRQui;
6043 else if (AArch64::DDRegClass.hasSubClassEq(RC)) {
6044 assert(Subtarget.hasNEON() &&
"Unexpected register store without NEON");
6045 Opc = AArch64::ST1Twov1d;
6047 }
else if (AArch64::XSeqPairsClassRegClass.hasSubClassEq(RC)) {
6049 get(AArch64::STPXi), SrcReg, isKill,
6050 AArch64::sube64, AArch64::subo64, FI, MMO);
6052 }
else if (AArch64::ZPRRegClass.hasSubClassEq(RC)) {
6053 assert(Subtarget.isSVEorStreamingSVEAvailable() &&
6054 "Unexpected register store without SVE store instructions");
6055 Opc = AArch64::STR_ZXI;
6060 if (AArch64::DDDRegClass.hasSubClassEq(RC)) {
6061 assert(Subtarget.hasNEON() &&
"Unexpected register store without NEON");
6062 Opc = AArch64::ST1Threev1d;
6067 if (AArch64::DDDDRegClass.hasSubClassEq(RC)) {
6068 assert(Subtarget.hasNEON() &&
"Unexpected register store without NEON");
6069 Opc = AArch64::ST1Fourv1d;
6071 }
else if (AArch64::QQRegClass.hasSubClassEq(RC)) {
6072 assert(Subtarget.hasNEON() &&
"Unexpected register store without NEON");
6073 Opc = AArch64::ST1Twov2d;
6075 }
else if (AArch64::ZPR2StridedOrContiguousRegClass.hasSubClassEq(RC)) {
6076 assert(Subtarget.isSVEorStreamingSVEAvailable() &&
6077 "Unexpected register store without SVE store instructions");
6078 Opc = AArch64::STR_ZZXI_STRIDED_CONTIGUOUS;
6080 }
else if (AArch64::ZPR2RegClass.hasSubClassEq(RC)) {
6081 assert(Subtarget.isSVEorStreamingSVEAvailable() &&
6082 "Unexpected register store without SVE store instructions");
6083 Opc = AArch64::STR_ZZXI;
6088 if (AArch64::QQQRegClass.hasSubClassEq(RC)) {
6089 assert(Subtarget.hasNEON() &&
"Unexpected register store without NEON");
6090 Opc = AArch64::ST1Threev2d;
6092 }
else if (AArch64::ZPR3RegClass.hasSubClassEq(RC)) {
6093 assert(Subtarget.isSVEorStreamingSVEAvailable() &&
6094 "Unexpected register store without SVE store instructions");
6095 Opc = AArch64::STR_ZZZXI;
6100 if (AArch64::QQQQRegClass.hasSubClassEq(RC)) {
6101 assert(Subtarget.hasNEON() &&
"Unexpected register store without NEON");
6102 Opc = AArch64::ST1Fourv2d;
6104 }
else if (AArch64::ZPR4StridedOrContiguousRegClass.hasSubClassEq(RC)) {
6105 assert(Subtarget.isSVEorStreamingSVEAvailable() &&
6106 "Unexpected register store without SVE store instructions");
6107 Opc = AArch64::STR_ZZZZXI_STRIDED_CONTIGUOUS;
6109 }
else if (AArch64::ZPR4RegClass.hasSubClassEq(RC)) {
6110 assert(Subtarget.isSVEorStreamingSVEAvailable() &&
6111 "Unexpected register store without SVE store instructions");
6112 Opc = AArch64::STR_ZZZZXI;
6117 assert(
Opc &&
"Unknown register class");
6128 MI.addMemOperand(MMO);
6135 Register DestReg,
unsigned SubIdx0,
6136 unsigned SubIdx1,
int FI,
6140 bool IsUndef =
true;
6142 DestReg0 =
TRI.getSubReg(DestReg, SubIdx0);
6144 DestReg1 =
TRI.getSubReg(DestReg, SubIdx1);
6173 switch (
TRI.getSpillSize(*RC)) {
6175 if (AArch64::FPR8RegClass.hasSubClassEq(RC))
6176 Opc = AArch64::LDRBui;
6179 bool IsPNR = AArch64::PNRRegClass.hasSubClassEq(RC);
6180 if (AArch64::FPR16RegClass.hasSubClassEq(RC))
6181 Opc = AArch64::LDRHui;
6182 else if (IsPNR || AArch64::PPRRegClass.hasSubClassEq(RC)) {
6183 assert(Subtarget.isSVEorStreamingSVEAvailable() &&
6184 "Unexpected register load without SVE load instructions");
6187 Opc = AArch64::LDR_PXI;
6193 if (AArch64::GPR32allRegClass.hasSubClassEq(RC)) {
6194 Opc = AArch64::LDRWui;
6198 assert(DestReg != AArch64::WSP);
6199 }
else if (AArch64::FPR32RegClass.hasSubClassEq(RC))
6200 Opc = AArch64::LDRSui;
6201 else if (AArch64::PPR2RegClass.hasSubClassEq(RC)) {
6202 Opc = AArch64::LDR_PPXI;
6207 if (AArch64::GPR64allRegClass.hasSubClassEq(RC)) {
6208 Opc = AArch64::LDRXui;
6212 assert(DestReg != AArch64::SP);
6213 }
else if (AArch64::FPR64RegClass.hasSubClassEq(RC)) {
6214 Opc = AArch64::LDRDui;
6215 }
else if (AArch64::WSeqPairsClassRegClass.hasSubClassEq(RC)) {
6217 get(AArch64::LDPWi), DestReg, AArch64::sube32,
6218 AArch64::subo32, FI, MMO);
6223 if (AArch64::FPR128RegClass.hasSubClassEq(RC))
6224 Opc = AArch64::LDRQui;
6225 else if (AArch64::DDRegClass.hasSubClassEq(RC)) {
6226 assert(Subtarget.hasNEON() &&
"Unexpected register load without NEON");
6227 Opc = AArch64::LD1Twov1d;
6229 }
else if (AArch64::XSeqPairsClassRegClass.hasSubClassEq(RC)) {
6231 get(AArch64::LDPXi), DestReg, AArch64::sube64,
6232 AArch64::subo64, FI, MMO);
6234 }
else if (AArch64::ZPRRegClass.hasSubClassEq(RC)) {
6235 assert(Subtarget.isSVEorStreamingSVEAvailable() &&
6236 "Unexpected register load without SVE load instructions");
6237 Opc = AArch64::LDR_ZXI;
6242 if (AArch64::DDDRegClass.hasSubClassEq(RC)) {
6243 assert(Subtarget.hasNEON() &&
"Unexpected register load without NEON");
6244 Opc = AArch64::LD1Threev1d;
6249 if (AArch64::DDDDRegClass.hasSubClassEq(RC)) {
6250 assert(Subtarget.hasNEON() &&
"Unexpected register load without NEON");
6251 Opc = AArch64::LD1Fourv1d;
6253 }
else if (AArch64::QQRegClass.hasSubClassEq(RC)) {
6254 assert(Subtarget.hasNEON() &&
"Unexpected register load without NEON");
6255 Opc = AArch64::LD1Twov2d;
6257 }
else if (AArch64::ZPR2StridedOrContiguousRegClass.hasSubClassEq(RC)) {
6258 assert(Subtarget.isSVEorStreamingSVEAvailable() &&
6259 "Unexpected register load without SVE load instructions");
6260 Opc = AArch64::LDR_ZZXI_STRIDED_CONTIGUOUS;
6262 }
else if (AArch64::ZPR2RegClass.hasSubClassEq(RC)) {
6263 assert(Subtarget.isSVEorStreamingSVEAvailable() &&
6264 "Unexpected register load without SVE load instructions");
6265 Opc = AArch64::LDR_ZZXI;
6270 if (AArch64::QQQRegClass.hasSubClassEq(RC)) {
6271 assert(Subtarget.hasNEON() &&
"Unexpected register load without NEON");
6272 Opc = AArch64::LD1Threev2d;
6274 }
else if (AArch64::ZPR3RegClass.hasSubClassEq(RC)) {
6275 assert(Subtarget.isSVEorStreamingSVEAvailable() &&
6276 "Unexpected register load without SVE load instructions");
6277 Opc = AArch64::LDR_ZZZXI;
6282 if (AArch64::QQQQRegClass.hasSubClassEq(RC)) {
6283 assert(Subtarget.hasNEON() &&
"Unexpected register load without NEON");
6284 Opc = AArch64::LD1Fourv2d;
6286 }
else if (AArch64::ZPR4StridedOrContiguousRegClass.hasSubClassEq(RC)) {
6287 assert(Subtarget.isSVEorStreamingSVEAvailable() &&
6288 "Unexpected register load without SVE load instructions");
6289 Opc = AArch64::LDR_ZZZZXI_STRIDED_CONTIGUOUS;
6291 }
else if (AArch64::ZPR4RegClass.hasSubClassEq(RC)) {
6292 assert(Subtarget.isSVEorStreamingSVEAvailable() &&
6293 "Unexpected register load without SVE load instructions");
6294 Opc = AArch64::LDR_ZZZZXI;
6300 assert(
Opc &&
"Unknown register class");
6310 MI.addMemOperand(MMO);
6317 UseMI.getIterator()),
6319 return I.modifiesRegister(AArch64::NZCV, TRI) ||
6320 I.readsRegister(AArch64::NZCV, TRI);
6324void AArch64InstrInfo::decomposeStackOffsetForDwarfOffsets(
6329 assert(
Offset.getScalable() % 2 == 0 &&
"Invalid frame offset");
6336 ByteSized =
Offset.getFixed();
6337 VGSized =
Offset.getScalable() / 2;
6343void AArch64InstrInfo::decomposeStackOffsetForFrameOffsets(
6345 int64_t &NumDataVectors) {
6349 assert(
Offset.getScalable() % 2 == 0 &&
"Invalid frame offset");
6351 NumBytes =
Offset.getFixed();
6353 NumPredicateVectors =
Offset.getScalable() / 2;
6358 if (NumPredicateVectors % 8 == 0 || NumPredicateVectors < -64 ||
6359 NumPredicateVectors > 62) {
6360 NumDataVectors = NumPredicateVectors / 8;
6361 NumPredicateVectors -= NumDataVectors * 8;
6387 Expr.
push_back((
char)dwarf::DW_OP_bregx);
6395 int64_t OffsetFromDefCFA) {
6409 Comment << (NumBytes < 0 ?
" - " :
" + ") << std::abs(NumBytes);
6410 if (!RegScale.empty())
6420 int64_t NumBytes, NumVGScaledBytes;
6421 AArch64InstrInfo::decomposeStackOffsetForDwarfOffsets(
Offset, NumBytes,
6423 std::string CommentBuffer;
6426 if (
Reg == AArch64::SP)
6428 else if (
Reg == AArch64::FP)
6435 unsigned DwarfReg =
TRI.getDwarfRegNum(
Reg,
true);
6436 assert(DwarfReg <= 31 &&
"DwarfReg out of bounds (0..31)");
6438 Expr.
push_back(dwarf::DW_OP_breg0 + DwarfReg);
6441 if (NumVGScaledBytes) {
6451 DefCfaExpr.
push_back(dwarf::DW_CFA_def_cfa_expression);
6459 unsigned FrameReg,
unsigned Reg,
6461 bool LastAdjustmentWasScalable) {
6462 if (
Offset.getScalable())
6465 if (FrameReg == Reg && !LastAdjustmentWasScalable)
6468 unsigned DwarfReg =
TRI.getDwarfRegNum(Reg,
true);
6475 std::optional<int64_t> IncomingVGOffsetFromDefCFA) {
6476 int64_t NumBytes, NumVGScaledBytes;
6477 AArch64InstrInfo::decomposeStackOffsetForDwarfOffsets(
6478 OffsetFromDefCFA, NumBytes, NumVGScaledBytes);
6480 unsigned DwarfReg =
TRI.getDwarfRegNum(Reg,
true);
6483 if (!NumVGScaledBytes)
6486 std::string CommentBuffer;
6491 assert(NumVGScaledBytes &&
"Expected scalable offset");
6495 if (IncomingVGOffsetFromDefCFA) {
6497 VGRegScale =
"* IncomingVG";
6500 VGRegScale =
"* VG";
6504 OffsetExpr.
push_back(dwarf::DW_OP_plus);
6513 CfaExpr.
push_back(dwarf::DW_CFA_expression);
6528 unsigned SrcReg, int64_t
Offset,
unsigned Opc,
6531 bool *HasWinCFI,
bool EmitCFAOffset,
6534 unsigned MaxEncoding, ShiftSize;
6536 case AArch64::ADDXri:
6537 case AArch64::ADDSXri:
6538 case AArch64::SUBXri:
6539 case AArch64::SUBSXri:
6540 MaxEncoding = 0xfff;
6543 case AArch64::ADDVL_XXI:
6544 case AArch64::ADDPL_XXI:
6545 case AArch64::ADDSVL_XXI:
6546 case AArch64::ADDSPL_XXI:
6561 if (
Opc == AArch64::ADDVL_XXI ||
Opc == AArch64::ADDSVL_XXI)
6563 else if (
Opc == AArch64::ADDPL_XXI ||
Opc == AArch64::ADDSPL_XXI)
6577 const unsigned MaxEncodableValue = MaxEncoding << ShiftSize;
6579 if (TmpReg == AArch64::XZR)
6580 TmpReg =
MBB.getParent()->getRegInfo().createVirtualRegister(
6581 &AArch64::GPR64RegClass);
6583 uint64_t ThisVal = std::min<uint64_t>(
Offset, MaxEncodableValue);
6584 unsigned LocalShiftSize = 0;
6585 if (ThisVal > MaxEncoding) {
6586 ThisVal = ThisVal >> ShiftSize;
6587 LocalShiftSize = ShiftSize;
6589 assert((ThisVal >> ShiftSize) <= MaxEncoding &&
6590 "Encoding cannot handle value that big");
6592 Offset -= ThisVal << LocalShiftSize;
6597 .
addImm(Sign * (
int)ThisVal);
6607 if (Sign == -1 ||
Opc == AArch64::SUBXri ||
Opc == AArch64::SUBSXri)
6608 CFAOffset += Change;
6610 CFAOffset -= Change;
6611 if (EmitCFAOffset && DestReg == TmpReg) {
6624 int Imm = (int)(ThisVal << LocalShiftSize);
6625 if (VScale != 1 && DestReg == AArch64::SP) {
6631 }
else if ((DestReg == AArch64::FP && SrcReg == AArch64::SP) ||
6632 (SrcReg == AArch64::FP && DestReg == AArch64::SP)) {
6633 assert(VScale == 1 &&
"Expected non-scalable operation");
6642 assert(
Offset == 0 &&
"Expected remaining offset to be zero to "
6643 "emit a single SEH directive");
6644 }
else if (DestReg == AArch64::SP) {
6645 assert(VScale == 1 &&
"Expected non-scalable operation");
6648 assert(SrcReg == AArch64::SP &&
"Unexpected SrcReg for SEH_StackAlloc");
6661 unsigned DestReg,
unsigned SrcReg,
6664 bool NeedsWinCFI,
bool *HasWinCFI,
6666 unsigned FrameReg) {
6673 bool UseSVL =
F.hasFnAttribute(
"aarch64_pstate_sm_body");
6675 int64_t Bytes, NumPredicateVectors, NumDataVectors;
6676 AArch64InstrInfo::decomposeStackOffsetForFrameOffsets(
6677 Offset, Bytes, NumPredicateVectors, NumDataVectors);
6680 bool NeedsFinalDefNZCV = SetNZCV && (NumPredicateVectors || NumDataVectors);
6681 if (NeedsFinalDefNZCV)
6685 if (Bytes || (!
Offset && SrcReg != DestReg)) {
6686 assert((DestReg != AArch64::SP || Bytes % 8 == 0) &&
6687 "SP increment/decrement not 8-byte aligned");
6688 unsigned Opc = SetNZCV ? AArch64::ADDSXri : AArch64::ADDXri;
6691 Opc = SetNZCV ? AArch64::SUBSXri : AArch64::SUBXri;
6694 NeedsWinCFI, HasWinCFI, EmitCFAOffset, CFAOffset,
6696 CFAOffset += (
Opc == AArch64::ADDXri ||
Opc == AArch64::ADDSXri)
6703 assert(!(NeedsWinCFI && NumPredicateVectors) &&
6704 "WinCFI can't allocate fractions of an SVE data vector");
6706 if (NumDataVectors) {
6708 UseSVL ? AArch64::ADDSVL_XXI : AArch64::ADDVL_XXI,
TII,
6709 Flag, NeedsWinCFI, HasWinCFI, EmitCFAOffset, CFAOffset,
6715 if (NumPredicateVectors) {
6716 assert(DestReg != AArch64::SP &&
"Unaligned access to SP");
6718 UseSVL ? AArch64::ADDSPL_XXI : AArch64::ADDPL_XXI,
TII,
6719 Flag, NeedsWinCFI, HasWinCFI, EmitCFAOffset, CFAOffset,
6723 if (NeedsFinalDefNZCV)
6744 if (
MI.isFullCopy()) {
6747 if (SrcReg == AArch64::SP && DstReg.
isVirtual()) {
6751 if (DstReg == AArch64::SP && SrcReg.
isVirtual()) {
6756 if (SrcReg == AArch64::NZCV || DstReg == AArch64::NZCV)
6784 if (
MI.isCopy() &&
Ops.size() == 1 &&
6786 (
Ops[0] == 0 ||
Ops[0] == 1)) {
6787 bool IsSpill =
Ops[0] == 0;
6788 bool IsFill = !IsSpill;
6800 :
TRI.getMinimalPhysRegClass(Reg);
6806 "Mismatched register size in non subreg COPY");
6813 return &*--InsertPt;
6825 if (IsSpill && DstMO.
isUndef() && SrcReg == AArch64::WZR &&
6828 "Unexpected subreg on physical register");
6830 FrameIndex, &AArch64::GPR64RegClass,
Register());
6831 return &*--InsertPt;
6848 case AArch64::sub_32:
6849 if (AArch64::GPR64RegClass.hasSubClassEq(
getRegClass(DstReg)))
6850 FillRC = &AArch64::GPR32RegClass;
6853 FillRC = &AArch64::FPR32RegClass;
6856 FillRC = &AArch64::FPR64RegClass;
6862 TRI.getRegSizeInBits(*FillRC) &&
6863 "Mismatched regclass size on folded subreg COPY");
6882 bool *OutUseUnscaledOp,
6883 unsigned *OutUnscaledOp,
6884 int64_t *EmittableOffset) {
6886 if (EmittableOffset)
6887 *EmittableOffset = 0;
6888 if (OutUseUnscaledOp)
6889 *OutUseUnscaledOp =
false;
6895 switch (
MI.getOpcode()) {
6898 case AArch64::LD1Rv1d:
6899 case AArch64::LD1Rv2s:
6900 case AArch64::LD1Rv2d:
6901 case AArch64::LD1Rv4h:
6902 case AArch64::LD1Rv4s:
6903 case AArch64::LD1Rv8b:
6904 case AArch64::LD1Rv8h:
6905 case AArch64::LD1Rv16b:
6906 case AArch64::LD1Twov2d:
6907 case AArch64::LD1Threev2d:
6908 case AArch64::LD1Fourv2d:
6909 case AArch64::LD1Twov1d:
6910 case AArch64::LD1Threev1d:
6911 case AArch64::LD1Fourv1d:
6912 case AArch64::ST1Twov2d:
6913 case AArch64::ST1Threev2d:
6914 case AArch64::ST1Fourv2d:
6915 case AArch64::ST1Twov1d:
6916 case AArch64::ST1Threev1d:
6917 case AArch64::ST1Fourv1d:
6918 case AArch64::ST1i8:
6919 case AArch64::ST1i16:
6920 case AArch64::ST1i32:
6921 case AArch64::ST1i64:
6923 case AArch64::IRGstack:
6924 case AArch64::STGloop:
6925 case AArch64::STZGloop:
6930 TypeSize ScaleValue(0U,
false), Width(0U,
false);
6931 int64_t MinOff, MaxOff;
6937 bool IsMulVL = ScaleValue.isScalable();
6938 unsigned Scale = ScaleValue.getKnownMinValue();
6948 std::optional<unsigned> UnscaledOp =
6950 bool useUnscaledOp = UnscaledOp && (
Offset % Scale ||
Offset < 0);
6951 if (useUnscaledOp &&
6956 Scale = ScaleValue.getKnownMinValue();
6957 assert(IsMulVL == ScaleValue.isScalable() &&
6958 "Unscaled opcode has different value for scalable");
6960 int64_t Remainder =
Offset % Scale;
6961 assert(!(Remainder && useUnscaledOp) &&
6962 "Cannot have remainder when using unscaled op");
6964 assert(MinOff < MaxOff &&
"Unexpected Min/Max offsets");
6965 int64_t NewOffset =
Offset / Scale;
6966 if (MinOff <= NewOffset && NewOffset <= MaxOff)
6969 NewOffset = NewOffset < 0 ? MinOff : MaxOff;
6973 if (EmittableOffset)
6974 *EmittableOffset = NewOffset;
6975 if (OutUseUnscaledOp)
6976 *OutUseUnscaledOp = useUnscaledOp;
6977 if (OutUnscaledOp && UnscaledOp)
6978 *OutUnscaledOp = *UnscaledOp;
6991 unsigned Opcode =
MI.getOpcode();
6992 unsigned ImmIdx = FrameRegIdx + 1;
6994 if (Opcode == AArch64::ADDSXri || Opcode == AArch64::ADDXri) {
6999 MI.eraseFromParent();
7005 unsigned UnscaledOp;
7008 &UnscaledOp, &NewOffset);
7012 MI.getOperand(FrameRegIdx).ChangeToRegister(FrameReg,
false);
7014 MI.setDesc(
TII->get(UnscaledOp));
7016 MI.getOperand(ImmIdx).ChangeToImmediate(NewOffset);
7032bool AArch64InstrInfo::useMachineCombiner()
const {
return true; }
7037 case AArch64::ADDSWrr:
7038 case AArch64::ADDSWri:
7039 case AArch64::ADDSXrr:
7040 case AArch64::ADDSXri:
7041 case AArch64::SUBSWrr:
7042 case AArch64::SUBSXrr:
7044 case AArch64::SUBSWri:
7045 case AArch64::SUBSXri:
7056 case AArch64::ADDWrr:
7057 case AArch64::ADDWri:
7058 case AArch64::SUBWrr:
7059 case AArch64::ADDSWrr:
7060 case AArch64::ADDSWri:
7061 case AArch64::SUBSWrr:
7063 case AArch64::SUBWri:
7064 case AArch64::SUBSWri:
7075 case AArch64::ADDXrr:
7076 case AArch64::ADDXri:
7077 case AArch64::SUBXrr:
7078 case AArch64::ADDSXrr:
7079 case AArch64::ADDSXri:
7080 case AArch64::SUBSXrr:
7082 case AArch64::SUBXri:
7083 case AArch64::SUBSXri:
7084 case AArch64::ADDv8i8:
7085 case AArch64::ADDv16i8:
7086 case AArch64::ADDv4i16:
7087 case AArch64::ADDv8i16:
7088 case AArch64::ADDv2i32:
7089 case AArch64::ADDv4i32:
7090 case AArch64::SUBv8i8:
7091 case AArch64::SUBv16i8:
7092 case AArch64::SUBv4i16:
7093 case AArch64::SUBv8i16:
7094 case AArch64::SUBv2i32:
7095 case AArch64::SUBv4i32:
7108 case AArch64::FADDHrr:
7109 case AArch64::FADDSrr:
7110 case AArch64::FADDDrr:
7111 case AArch64::FADDv4f16:
7112 case AArch64::FADDv8f16:
7113 case AArch64::FADDv2f32:
7114 case AArch64::FADDv2f64:
7115 case AArch64::FADDv4f32:
7116 case AArch64::FSUBHrr:
7117 case AArch64::FSUBSrr:
7118 case AArch64::FSUBDrr:
7119 case AArch64::FSUBv4f16:
7120 case AArch64::FSUBv8f16:
7121 case AArch64::FSUBv2f32:
7122 case AArch64::FSUBv2f64:
7123 case AArch64::FSUBv4f32:
7142 unsigned CombineOpc,
unsigned ZeroReg = 0,
7143 bool CheckZeroReg =
false) {
7150 if (!
MI ||
MI->getParent() != &
MBB ||
MI->getOpcode() != CombineOpc)
7153 if (!
MRI.hasOneNonDBGUse(
MI->getOperand(0).getReg()))
7157 assert(
MI->getNumOperands() >= 4 &&
MI->getOperand(0).isReg() &&
7158 MI->getOperand(1).isReg() &&
MI->getOperand(2).isReg() &&
7159 MI->getOperand(3).isReg() &&
"MAdd/MSub must have a least 4 regs");
7161 if (
MI->getOperand(3).getReg() != ZeroReg)
7166 MI->findRegisterDefOperandIdx(AArch64::NZCV,
nullptr,
true) == -1)
7175 unsigned MulOpc,
unsigned ZeroReg) {
7190bool AArch64InstrInfo::isAssociativeAndCommutative(
const MachineInstr &Inst,
7191 bool Invert)
const {
7197 case AArch64::FADDHrr:
7198 case AArch64::FADDSrr:
7199 case AArch64::FADDDrr:
7200 case AArch64::FMULHrr:
7201 case AArch64::FMULSrr:
7202 case AArch64::FMULDrr:
7203 case AArch64::FMULX16:
7204 case AArch64::FMULX32:
7205 case AArch64::FMULX64:
7207 case AArch64::FADDv4f16:
7208 case AArch64::FADDv8f16:
7209 case AArch64::FADDv2f32:
7210 case AArch64::FADDv4f32:
7211 case AArch64::FADDv2f64:
7212 case AArch64::FMULv4f16:
7213 case AArch64::FMULv8f16:
7214 case AArch64::FMULv2f32:
7215 case AArch64::FMULv4f32:
7216 case AArch64::FMULv2f64:
7217 case AArch64::FMULXv4f16:
7218 case AArch64::FMULXv8f16:
7219 case AArch64::FMULXv2f32:
7220 case AArch64::FMULXv4f32:
7221 case AArch64::FMULXv2f64:
7225 case AArch64::FADD_ZZZ_H:
7226 case AArch64::FADD_ZZZ_S:
7227 case AArch64::FADD_ZZZ_D:
7228 case AArch64::FMUL_ZZZ_H:
7229 case AArch64::FMUL_ZZZ_S:
7230 case AArch64::FMUL_ZZZ_D:
7241 case AArch64::ADDWrr:
7242 case AArch64::ADDXrr:
7243 case AArch64::ANDWrr:
7244 case AArch64::ANDXrr:
7245 case AArch64::ORRWrr:
7246 case AArch64::ORRXrr:
7247 case AArch64::EORWrr:
7248 case AArch64::EORXrr:
7249 case AArch64::EONWrr:
7250 case AArch64::EONXrr:
7254 case AArch64::ADDv8i8:
7255 case AArch64::ADDv16i8:
7256 case AArch64::ADDv4i16:
7257 case AArch64::ADDv8i16:
7258 case AArch64::ADDv2i32:
7259 case AArch64::ADDv4i32:
7260 case AArch64::ADDv1i64:
7261 case AArch64::ADDv2i64:
7262 case AArch64::MULv8i8:
7263 case AArch64::MULv16i8:
7264 case AArch64::MULv4i16:
7265 case AArch64::MULv8i16:
7266 case AArch64::MULv2i32:
7267 case AArch64::MULv4i32:
7268 case AArch64::ANDv8i8:
7269 case AArch64::ANDv16i8:
7270 case AArch64::ORRv8i8:
7271 case AArch64::ORRv16i8:
7272 case AArch64::EORv8i8:
7273 case AArch64::EORv16i8:
7275 case AArch64::ADD_ZZZ_B:
7276 case AArch64::ADD_ZZZ_H:
7277 case AArch64::ADD_ZZZ_S:
7278 case AArch64::ADD_ZZZ_D:
7279 case AArch64::MUL_ZZZ_B:
7280 case AArch64::MUL_ZZZ_H:
7281 case AArch64::MUL_ZZZ_S:
7282 case AArch64::MUL_ZZZ_D:
7283 case AArch64::AND_ZZZ:
7284 case AArch64::ORR_ZZZ:
7285 case AArch64::EOR_ZZZ:
7316 auto setFound = [&](
int Opcode,
int Operand,
unsigned ZeroReg,
7324 auto setVFound = [&](
int Opcode,
int Operand,
unsigned Pattern) {
7336 case AArch64::ADDWrr:
7338 "ADDWrr does not have register operands");
7339 setFound(AArch64::MADDWrrr, 1, AArch64::WZR, MCP::MULADDW_OP1);
7340 setFound(AArch64::MADDWrrr, 2, AArch64::WZR, MCP::MULADDW_OP2);
7342 case AArch64::ADDXrr:
7343 setFound(AArch64::MADDXrrr, 1, AArch64::XZR, MCP::MULADDX_OP1);
7344 setFound(AArch64::MADDXrrr, 2, AArch64::XZR, MCP::MULADDX_OP2);
7346 case AArch64::SUBWrr:
7347 setFound(AArch64::MADDWrrr, 2, AArch64::WZR, MCP::MULSUBW_OP2);
7348 setFound(AArch64::MADDWrrr, 1, AArch64::WZR, MCP::MULSUBW_OP1);
7350 case AArch64::SUBXrr:
7351 setFound(AArch64::MADDXrrr, 2, AArch64::XZR, MCP::MULSUBX_OP2);
7352 setFound(AArch64::MADDXrrr, 1, AArch64::XZR, MCP::MULSUBX_OP1);
7354 case AArch64::ADDWri:
7355 setFound(AArch64::MADDWrrr, 1, AArch64::WZR, MCP::MULADDWI_OP1);
7357 case AArch64::ADDXri:
7358 setFound(AArch64::MADDXrrr, 1, AArch64::XZR, MCP::MULADDXI_OP1);
7360 case AArch64::SUBWri:
7361 setFound(AArch64::MADDWrrr, 1, AArch64::WZR, MCP::MULSUBWI_OP1);
7363 case AArch64::SUBXri:
7364 setFound(AArch64::MADDXrrr, 1, AArch64::XZR, MCP::MULSUBXI_OP1);
7366 case AArch64::ADDv8i8:
7367 setVFound(AArch64::MULv8i8, 1, MCP::MULADDv8i8_OP1);
7368 setVFound(AArch64::MULv8i8, 2, MCP::MULADDv8i8_OP2);
7370 case AArch64::ADDv16i8:
7371 setVFound(AArch64::MULv16i8, 1, MCP::MULADDv16i8_OP1);
7372 setVFound(AArch64::MULv16i8, 2, MCP::MULADDv16i8_OP2);
7374 case AArch64::ADDv4i16:
7375 setVFound(AArch64::MULv4i16, 1, MCP::MULADDv4i16_OP1);
7376 setVFound(AArch64::MULv4i16, 2, MCP::MULADDv4i16_OP2);
7377 setVFound(AArch64::MULv4i16_indexed, 1, MCP::MULADDv4i16_indexed_OP1);
7378 setVFound(AArch64::MULv4i16_indexed, 2, MCP::MULADDv4i16_indexed_OP2);
7380 case AArch64::ADDv8i16:
7381 setVFound(AArch64::MULv8i16, 1, MCP::MULADDv8i16_OP1);
7382 setVFound(AArch64::MULv8i16, 2, MCP::MULADDv8i16_OP2);
7383 setVFound(AArch64::MULv8i16_indexed, 1, MCP::MULADDv8i16_indexed_OP1);
7384 setVFound(AArch64::MULv8i16_indexed, 2, MCP::MULADDv8i16_indexed_OP2);
7386 case AArch64::ADDv2i32:
7387 setVFound(AArch64::MULv2i32, 1, MCP::MULADDv2i32_OP1);
7388 setVFound(AArch64::MULv2i32, 2, MCP::MULADDv2i32_OP2);
7389 setVFound(AArch64::MULv2i32_indexed, 1, MCP::MULADDv2i32_indexed_OP1);
7390 setVFound(AArch64::MULv2i32_indexed, 2, MCP::MULADDv2i32_indexed_OP2);
7392 case AArch64::ADDv4i32:
7393 setVFound(AArch64::MULv4i32, 1, MCP::MULADDv4i32_OP1);
7394 setVFound(AArch64::MULv4i32, 2, MCP::MULADDv4i32_OP2);
7395 setVFound(AArch64::MULv4i32_indexed, 1, MCP::MULADDv4i32_indexed_OP1);
7396 setVFound(AArch64::MULv4i32_indexed, 2, MCP::MULADDv4i32_indexed_OP2);
7398 case AArch64::SUBv8i8:
7399 setVFound(AArch64::MULv8i8, 1, MCP::MULSUBv8i8_OP1);
7400 setVFound(AArch64::MULv8i8, 2, MCP::MULSUBv8i8_OP2);
7402 case AArch64::SUBv16i8:
7403 setVFound(AArch64::MULv16i8, 1, MCP::MULSUBv16i8_OP1);
7404 setVFound(AArch64::MULv16i8, 2, MCP::MULSUBv16i8_OP2);
7406 case AArch64::SUBv4i16:
7407 setVFound(AArch64::MULv4i16, 1, MCP::MULSUBv4i16_OP1);
7408 setVFound(AArch64::MULv4i16, 2, MCP::MULSUBv4i16_OP2);
7409 setVFound(AArch64::MULv4i16_indexed, 1, MCP::MULSUBv4i16_indexed_OP1);
7410 setVFound(AArch64::MULv4i16_indexed, 2, MCP::MULSUBv4i16_indexed_OP2);
7412 case AArch64::SUBv8i16:
7413 setVFound(AArch64::MULv8i16, 1, MCP::MULSUBv8i16_OP1);
7414 setVFound(AArch64::MULv8i16, 2, MCP::MULSUBv8i16_OP2);
7415 setVFound(AArch64::MULv8i16_indexed, 1, MCP::MULSUBv8i16_indexed_OP1);
7416 setVFound(AArch64::MULv8i16_indexed, 2, MCP::MULSUBv8i16_indexed_OP2);
7418 case AArch64::SUBv2i32:
7419 setVFound(AArch64::MULv2i32, 1, MCP::MULSUBv2i32_OP1);
7420 setVFound(AArch64::MULv2i32, 2, MCP::MULSUBv2i32_OP2);
7421 setVFound(AArch64::MULv2i32_indexed, 1, MCP::MULSUBv2i32_indexed_OP1);
7422 setVFound(AArch64::MULv2i32_indexed, 2, MCP::MULSUBv2i32_indexed_OP2);
7424 case AArch64::SUBv4i32:
7425 setVFound(AArch64::MULv4i32, 1, MCP::MULSUBv4i32_OP1);
7426 setVFound(AArch64::MULv4i32, 2, MCP::MULSUBv4i32_OP2);
7427 setVFound(AArch64::MULv4i32_indexed, 1, MCP::MULSUBv4i32_indexed_OP1);
7428 setVFound(AArch64::MULv4i32_indexed, 2, MCP::MULSUBv4i32_indexed_OP2);
7434bool AArch64InstrInfo::isAccumulationOpcode(
unsigned Opcode)
const {
7438 case AArch64::UABALB_ZZZ_D:
7439 case AArch64::UABALB_ZZZ_H:
7440 case AArch64::UABALB_ZZZ_S:
7441 case AArch64::UABALT_ZZZ_D:
7442 case AArch64::UABALT_ZZZ_H:
7443 case AArch64::UABALT_ZZZ_S:
7444 case AArch64::SABALB_ZZZ_D:
7445 case AArch64::SABALB_ZZZ_S:
7446 case AArch64::SABALB_ZZZ_H:
7447 case AArch64::SABALT_ZZZ_D:
7448 case AArch64::SABALT_ZZZ_S:
7449 case AArch64::SABALT_ZZZ_H:
7450 case AArch64::UABALv16i8_v8i16:
7451 case AArch64::UABALv2i32_v2i64:
7452 case AArch64::UABALv4i16_v4i32:
7453 case AArch64::UABALv4i32_v2i64:
7454 case AArch64::UABALv8i16_v4i32:
7455 case AArch64::UABALv8i8_v8i16:
7456 case AArch64::UABAv16i8:
7457 case AArch64::UABAv2i32:
7458 case AArch64::UABAv4i16:
7459 case AArch64::UABAv4i32:
7460 case AArch64::UABAv8i16:
7461 case AArch64::UABAv8i8:
7462 case AArch64::SABALv16i8_v8i16:
7463 case AArch64::SABALv2i32_v2i64:
7464 case AArch64::SABALv4i16_v4i32:
7465 case AArch64::SABALv4i32_v2i64:
7466 case AArch64::SABALv8i16_v4i32:
7467 case AArch64::SABALv8i8_v8i16:
7468 case AArch64::SABAv16i8:
7469 case AArch64::SABAv2i32:
7470 case AArch64::SABAv4i16:
7471 case AArch64::SABAv4i32:
7472 case AArch64::SABAv8i16:
7473 case AArch64::SABAv8i8:
7480unsigned AArch64InstrInfo::getAccumulationStartOpcode(
7481 unsigned AccumulationOpcode)
const {
7482 switch (AccumulationOpcode) {
7485 case AArch64::UABALB_ZZZ_D:
7486 return AArch64::UABDLB_ZZZ_D;
7487 case AArch64::UABALB_ZZZ_H:
7488 return AArch64::UABDLB_ZZZ_H;
7489 case AArch64::UABALB_ZZZ_S:
7490 return AArch64::UABDLB_ZZZ_S;
7491 case AArch64::UABALT_ZZZ_D:
7492 return AArch64::UABDLT_ZZZ_D;
7493 case AArch64::UABALT_ZZZ_H:
7494 return AArch64::UABDLT_ZZZ_H;
7495 case AArch64::UABALT_ZZZ_S:
7496 return AArch64::UABDLT_ZZZ_S;
7497 case AArch64::UABALv16i8_v8i16:
7498 return AArch64::UABDLv16i8_v8i16;
7499 case AArch64::UABALv2i32_v2i64:
7500 return AArch64::UABDLv2i32_v2i64;
7501 case AArch64::UABALv4i16_v4i32:
7502 return AArch64::UABDLv4i16_v4i32;
7503 case AArch64::UABALv4i32_v2i64:
7504 return AArch64::UABDLv4i32_v2i64;
7505 case AArch64::UABALv8i16_v4i32:
7506 return AArch64::UABDLv8i16_v4i32;
7507 case AArch64::UABALv8i8_v8i16:
7508 return AArch64::UABDLv8i8_v8i16;
7509 case AArch64::UABAv16i8:
7510 return AArch64::UABDv16i8;
7511 case AArch64::UABAv2i32:
7512 return AArch64::UABDv2i32;
7513 case AArch64::UABAv4i16:
7514 return AArch64::UABDv4i16;
7515 case AArch64::UABAv4i32:
7516 return AArch64::UABDv4i32;
7517 case AArch64::UABAv8i16:
7518 return AArch64::UABDv8i16;
7519 case AArch64::UABAv8i8:
7520 return AArch64::UABDv8i8;
7521 case AArch64::SABALB_ZZZ_D:
7522 return AArch64::SABDLB_ZZZ_D;
7523 case AArch64::SABALB_ZZZ_S:
7524 return AArch64::SABDLB_ZZZ_S;
7525 case AArch64::SABALB_ZZZ_H:
7526 return AArch64::SABDLB_ZZZ_H;
7527 case AArch64::SABALT_ZZZ_D:
7528 return AArch64::SABDLT_ZZZ_D;
7529 case AArch64::SABALT_ZZZ_S:
7530 return AArch64::SABDLT_ZZZ_S;
7531 case AArch64::SABALT_ZZZ_H:
7532 return AArch64::SABDLT_ZZZ_H;
7533 case AArch64::SABALv16i8_v8i16:
7534 return AArch64::SABDLv16i8_v8i16;
7535 case AArch64::SABALv2i32_v2i64:
7536 return AArch64::SABDLv2i32_v2i64;
7537 case AArch64::SABALv4i16_v4i32:
7538 return AArch64::SABDLv4i16_v4i32;
7539 case AArch64::SABALv4i32_v2i64:
7540 return AArch64::SABDLv4i32_v2i64;
7541 case AArch64::SABALv8i16_v4i32:
7542 return AArch64::SABDLv8i16_v4i32;
7543 case AArch64::SABALv8i8_v8i16:
7544 return AArch64::SABDLv8i8_v8i16;
7545 case AArch64::SABAv16i8:
7546 return AArch64::SABDv16i8;
7547 case AArch64::SABAv2i32:
7548 return AArch64::SABAv2i32;
7549 case AArch64::SABAv4i16:
7550 return AArch64::SABDv4i16;
7551 case AArch64::SABAv4i32:
7552 return AArch64::SABDv4i32;
7553 case AArch64::SABAv8i16:
7554 return AArch64::SABDv8i16;
7555 case AArch64::SABAv8i8:
7556 return AArch64::SABDv8i8;
7572 auto Match = [&](
int Opcode,
int Operand,
unsigned Pattern) ->
bool {
7584 assert(
false &&
"Unsupported FP instruction in combiner\n");
7586 case AArch64::FADDHrr:
7588 "FADDHrr does not have register operands");
7590 Found = Match(AArch64::FMULHrr, 1, MCP::FMULADDH_OP1);
7591 Found |= Match(AArch64::FMULHrr, 2, MCP::FMULADDH_OP2);
7593 case AArch64::FADDSrr:
7595 "FADDSrr does not have register operands");
7597 Found |= Match(AArch64::FMULSrr, 1, MCP::FMULADDS_OP1) ||
7598 Match(AArch64::FMULv1i32_indexed, 1, MCP::FMLAv1i32_indexed_OP1);
7600 Found |= Match(AArch64::FMULSrr, 2, MCP::FMULADDS_OP2) ||
7601 Match(AArch64::FMULv1i32_indexed, 2, MCP::FMLAv1i32_indexed_OP2);
7603 case AArch64::FADDDrr:
7604 Found |= Match(AArch64::FMULDrr, 1, MCP::FMULADDD_OP1) ||
7605 Match(AArch64::FMULv1i64_indexed, 1, MCP::FMLAv1i64_indexed_OP1);
7607 Found |= Match(AArch64::FMULDrr, 2, MCP::FMULADDD_OP2) ||
7608 Match(AArch64::FMULv1i64_indexed, 2, MCP::FMLAv1i64_indexed_OP2);
7610 case AArch64::FADDv4f16:
7611 Found |= Match(AArch64::FMULv4i16_indexed, 1, MCP::FMLAv4i16_indexed_OP1) ||
7612 Match(AArch64::FMULv4f16, 1, MCP::FMLAv4f16_OP1);
7614 Found |= Match(AArch64::FMULv4i16_indexed, 2, MCP::FMLAv4i16_indexed_OP2) ||
7615 Match(AArch64::FMULv4f16, 2, MCP::FMLAv4f16_OP2);
7617 case AArch64::FADDv8f16:
7618 Found |= Match(AArch64::FMULv8i16_indexed, 1, MCP::FMLAv8i16_indexed_OP1) ||
7619 Match(AArch64::FMULv8f16, 1, MCP::FMLAv8f16_OP1);
7621 Found |= Match(AArch64::FMULv8i16_indexed, 2, MCP::FMLAv8i16_indexed_OP2) ||
7622 Match(AArch64::FMULv8f16, 2, MCP::FMLAv8f16_OP2);
7624 case AArch64::FADDv2f32:
7625 Found |= Match(AArch64::FMULv2i32_indexed, 1, MCP::FMLAv2i32_indexed_OP1) ||
7626 Match(AArch64::FMULv2f32, 1, MCP::FMLAv2f32_OP1);
7628 Found |= Match(AArch64::FMULv2i32_indexed, 2, MCP::FMLAv2i32_indexed_OP2) ||
7629 Match(AArch64::FMULv2f32, 2, MCP::FMLAv2f32_OP2);
7631 case AArch64::FADDv2f64:
7632 Found |= Match(AArch64::FMULv2i64_indexed, 1, MCP::FMLAv2i64_indexed_OP1) ||
7633 Match(AArch64::FMULv2f64, 1, MCP::FMLAv2f64_OP1);
7635 Found |= Match(AArch64::FMULv2i64_indexed, 2, MCP::FMLAv2i64_indexed_OP2) ||
7636 Match(AArch64::FMULv2f64, 2, MCP::FMLAv2f64_OP2);
7638 case AArch64::FADDv4f32:
7639 Found |= Match(AArch64::FMULv4i32_indexed, 1, MCP::FMLAv4i32_indexed_OP1) ||
7640 Match(AArch64::FMULv4f32, 1, MCP::FMLAv4f32_OP1);
7642 Found |= Match(AArch64::FMULv4i32_indexed, 2, MCP::FMLAv4i32_indexed_OP2) ||
7643 Match(AArch64::FMULv4f32, 2, MCP::FMLAv4f32_OP2);
7645 case AArch64::FSUBHrr:
7646 Found = Match(AArch64::FMULHrr, 1, MCP::FMULSUBH_OP1);
7647 Found |= Match(AArch64::FMULHrr, 2, MCP::FMULSUBH_OP2);
7648 Found |= Match(AArch64::FNMULHrr, 1, MCP::FNMULSUBH_OP1);
7650 case AArch64::FSUBSrr:
7651 Found = Match(AArch64::FMULSrr, 1, MCP::FMULSUBS_OP1);
7653 Found |= Match(AArch64::FMULSrr, 2, MCP::FMULSUBS_OP2) ||
7654 Match(AArch64::FMULv1i32_indexed, 2, MCP::FMLSv1i32_indexed_OP2);
7656 Found |= Match(AArch64::FNMULSrr, 1, MCP::FNMULSUBS_OP1);
7658 case AArch64::FSUBDrr:
7659 Found = Match(AArch64::FMULDrr, 1, MCP::FMULSUBD_OP1);
7661 Found |= Match(AArch64::FMULDrr, 2, MCP::FMULSUBD_OP2) ||
7662 Match(AArch64::FMULv1i64_indexed, 2, MCP::FMLSv1i64_indexed_OP2);
7664 Found |= Match(AArch64::FNMULDrr, 1, MCP::FNMULSUBD_OP1);
7666 case AArch64::FSUBv4f16:
7667 Found |= Match(AArch64::FMULv4i16_indexed, 2, MCP::FMLSv4i16_indexed_OP2) ||
7668 Match(AArch64::FMULv4f16, 2, MCP::FMLSv4f16_OP2);
7670 Found |= Match(AArch64::FMULv4i16_indexed, 1, MCP::FMLSv4i16_indexed_OP1) ||
7671 Match(AArch64::FMULv4f16, 1, MCP::FMLSv4f16_OP1);
7673 case AArch64::FSUBv8f16:
7674 Found |= Match(AArch64::FMULv8i16_indexed, 2, MCP::FMLSv8i16_indexed_OP2) ||
7675 Match(AArch64::FMULv8f16, 2, MCP::FMLSv8f16_OP2);
7677 Found |= Match(AArch64::FMULv8i16_indexed, 1, MCP::FMLSv8i16_indexed_OP1) ||
7678 Match(AArch64::FMULv8f16, 1, MCP::FMLSv8f16_OP1);
7680 case AArch64::FSUBv2f32:
7681 Found |= Match(AArch64::FMULv2i32_indexed, 2, MCP::FMLSv2i32_indexed_OP2) ||
7682 Match(AArch64::FMULv2f32, 2, MCP::FMLSv2f32_OP2);
7684 Found |= Match(AArch64::FMULv2i32_indexed, 1, MCP::FMLSv2i32_indexed_OP1) ||
7685 Match(AArch64::FMULv2f32, 1, MCP::FMLSv2f32_OP1);
7687 case AArch64::FSUBv2f64:
7688 Found |= Match(AArch64::FMULv2i64_indexed, 2, MCP::FMLSv2i64_indexed_OP2) ||
7689 Match(AArch64::FMULv2f64, 2, MCP::FMLSv2f64_OP2);
7691 Found |= Match(AArch64::FMULv2i64_indexed, 1, MCP::FMLSv2i64_indexed_OP1) ||
7692 Match(AArch64::FMULv2f64, 1, MCP::FMLSv2f64_OP1);
7694 case AArch64::FSUBv4f32:
7695 Found |= Match(AArch64::FMULv4i32_indexed, 2, MCP::FMLSv4i32_indexed_OP2) ||
7696 Match(AArch64::FMULv4f32, 2, MCP::FMLSv4f32_OP2);
7698 Found |= Match(AArch64::FMULv4i32_indexed, 1, MCP::FMLSv4i32_indexed_OP1) ||
7699 Match(AArch64::FMULv4f32, 1, MCP::FMLSv4f32_OP1);
7710 auto Match = [&](
unsigned Opcode,
int Operand,
unsigned Pattern) ->
bool {
7717 if (
MI &&
MI->getOpcode() == TargetOpcode::COPY &&
7718 MI->getOperand(1).getReg().isVirtual())
7719 MI =
MRI.getUniqueVRegDef(
MI->getOperand(1).getReg());
7720 if (
MI &&
MI->getOpcode() == Opcode) {
7732 case AArch64::FMULv2f32:
7733 Found = Match(AArch64::DUPv2i32lane, 1, MCP::FMULv2i32_indexed_OP1);
7734 Found |= Match(AArch64::DUPv2i32lane, 2, MCP::FMULv2i32_indexed_OP2);
7736 case AArch64::FMULv2f64:
7737 Found = Match(AArch64::DUPv2i64lane, 1, MCP::FMULv2i64_indexed_OP1);
7738 Found |= Match(AArch64::DUPv2i64lane, 2, MCP::FMULv2i64_indexed_OP2);
7740 case AArch64::FMULv4f16:
7741 Found = Match(AArch64::DUPv4i16lane, 1, MCP::FMULv4i16_indexed_OP1);
7742 Found |= Match(AArch64::DUPv4i16lane, 2, MCP::FMULv4i16_indexed_OP2);
7744 case AArch64::FMULv4f32:
7745 Found = Match(AArch64::DUPv4i32lane, 1, MCP::FMULv4i32_indexed_OP1);
7746 Found |= Match(AArch64::DUPv4i32lane, 2, MCP::FMULv4i32_indexed_OP2);
7748 case AArch64::FMULv8f16:
7749 Found = Match(AArch64::DUPv8i16lane, 1, MCP::FMULv8i16_indexed_OP1);
7750 Found |= Match(AArch64::DUPv8i16lane, 2, MCP::FMULv8i16_indexed_OP2);
7763 auto Match = [&](
unsigned Opcode,
unsigned Pattern) ->
bool {
7766 if (
MI !=
nullptr && (
MI->getOpcode() == Opcode) &&
7767 MRI.hasOneNonDBGUse(
MI->getOperand(0).getReg()) &&
7781 case AArch64::FNEGDr:
7783 case AArch64::FNEGSr:
7915 case AArch64::SUBWrr:
7916 case AArch64::SUBSWrr:
7917 case AArch64::SUBXrr:
7918 case AArch64::SUBSXrr:
7963 unsigned LoadLaneOpCode,
unsigned NumLanes) {
7986 while (!RemainingLanes.
empty() && CurrInstr &&
7987 CurrInstr->getOpcode() == LoadLaneOpCode &&
7988 MRI.hasOneNonDBGUse(CurrInstr->getOperand(0).getReg()) &&
7989 CurrInstr->getNumOperands() == 4) {
7990 RemainingLanes.
erase(CurrInstr->getOperand(2).getImm());
7992 CurrInstr =
MRI.getUniqueVRegDef(CurrInstr->getOperand(1).getReg());
7996 if (!RemainingLanes.
empty())
8000 if (CurrInstr->getOpcode() != TargetOpcode::SUBREG_TO_REG)
8004 auto Lane0LoadReg = CurrInstr->getOperand(2).getReg();
8005 unsigned SingleLaneSizeInBits = 128 / NumLanes;
8006 if (
TRI->getRegSizeInBits(Lane0LoadReg,
MRI) != SingleLaneSizeInBits)
8010 if (!
MRI.hasOneNonDBGUse(Lane0LoadReg))
8013 LoadInstrs.
push_back(
MRI.getUniqueVRegDef(Lane0LoadReg));
8022 RemainingLoadInstrs.
insert(LoadInstrs.
begin(), LoadInstrs.
end());
8025 for (; MBBItr !=
MBB->begin() && RemainingSteps > 0 &&
8026 !RemainingLoadInstrs.
empty();
8027 --MBBItr, --RemainingSteps) {
8031 RemainingLoadInstrs.
erase(&CurrInstr);
8041 if (RemainingSteps == 0 && !RemainingLoadInstrs.
empty())
8067 case AArch64::LD1i32:
8069 case AArch64::LD1i16:
8071 case AArch64::LD1i8:
8087 unsigned Pattern,
unsigned NumLanes) {
8095 for (
unsigned i = 0; i < NumLanes - 1; ++i) {
8103 return A->getOperand(2).getImm() >
B->getOperand(2).getImm();
8108 MRI.getUniqueVRegDef(SubregToReg->getOperand(2).getReg()));
8109 auto LoadToLaneInstrsAscending =
llvm::reverse(LoadToLaneInstrs);
8115 auto CreateLD1Instruction = [&](
MachineInstr *OriginalInstr,
8116 Register SrcRegister,
unsigned Lane,
8118 bool OffsetRegisterKillState) {
8119 auto NewRegister =
MRI.createVirtualRegister(FPR128RegClass);
8126 InstrIdxForVirtReg.
insert(std::make_pair(NewRegister, InsInstrs.
size()));
8127 InsInstrs.
push_back(LoadIndexIntoRegister);
8133 auto CreateLDRInstruction = [&](
unsigned NumLanes,
Register DestReg,
8139 Opcode = AArch64::LDRSui;
8142 Opcode = AArch64::LDRHui;
8145 Opcode = AArch64::LDRBui;
8149 "Got unsupported number of lanes in machine-combiner gather pattern");
8158 auto LanesToLoadToReg0 =
8160 LoadToLaneInstrsAscending.begin() + NumLanes / 2);
8161 Register PrevReg = SubregToReg->getOperand(0).getReg();
8163 const MachineOperand &OffsetRegOperand = LoadInstr->getOperand(3);
8164 PrevReg = CreateLD1Instruction(LoadInstr, PrevReg, Index + 1,
8165 OffsetRegOperand.
getReg(),
8166 OffsetRegOperand.
isKill());
8173 MachineInstr *Lane0Load = *LoadToLaneInstrsAscending.begin();
8175 *std::next(LoadToLaneInstrsAscending.begin(), NumLanes / 2);
8176 Register DestRegForMiddleIndex =
MRI.createVirtualRegister(
8182 CreateLDRInstruction(NumLanes, DestRegForMiddleIndex,
8183 OriginalSplitToLoadOffsetOperand.
getReg(),
8184 OriginalSplitToLoadOffsetOperand.
isKill());
8186 InstrIdxForVirtReg.
insert(
8187 std::make_pair(DestRegForMiddleIndex, InsInstrs.
size()));
8188 InsInstrs.
push_back(MiddleIndexLoadInstr);
8192 Register DestRegForSubregToReg =
MRI.createVirtualRegister(FPR128RegClass);
8193 unsigned SubregType;
8196 SubregType = AArch64::ssub;
8199 SubregType = AArch64::hsub;
8202 SubregType = AArch64::bsub;
8206 "Got invalid NumLanes for machine-combiner gather pattern");
8209 auto SubRegToRegInstr =
8211 DestRegForSubregToReg)
8215 InstrIdxForVirtReg.
insert(
8216 std::make_pair(DestRegForSubregToReg, InsInstrs.
size()));
8220 auto LanesToLoadToReg1 =
8222 LoadToLaneInstrsAscending.end());
8223 PrevReg = SubRegToRegInstr->getOperand(0).getReg();
8225 const MachineOperand &OffsetRegOperand = LoadInstr->getOperand(3);
8226 PrevReg = CreateLD1Instruction(LoadInstr, PrevReg, Index + 1,
8227 OffsetRegOperand.
getReg(),
8228 OffsetRegOperand.
isKill());
8231 if (Index == NumLanes / 2 - 2) {
8266bool AArch64InstrInfo::getMachineCombinerPatterns(
8268 bool DoRegPressureReduce)
const {
8289 DoRegPressureReduce);
8318 const Register *ReplacedAddend =
nullptr) {
8319 assert(IdxMulOpd == 1 || IdxMulOpd == 2);
8321 unsigned IdxOtherOpd = IdxMulOpd == 1 ? 2 : 1;
8324 Register SrcReg0 = MUL->getOperand(1).getReg();
8325 bool Src0IsKill = MUL->getOperand(1).isKill();
8326 Register SrcReg1 = MUL->getOperand(2).getReg();
8327 bool Src1IsKill = MUL->getOperand(2).isKill();
8331 if (ReplacedAddend) {
8333 SrcReg2 = *ReplacedAddend;
8341 MRI.constrainRegClass(ResultReg, RC);
8343 MRI.constrainRegClass(SrcReg0, RC);
8345 MRI.constrainRegClass(SrcReg1, RC);
8347 MRI.constrainRegClass(SrcReg2, RC);
8360 .
addImm(MUL->getOperand(3).getImm());
8367 assert(
false &&
"Invalid FMA instruction kind \n");
8381 if (AArch64::FPR32RegClass.hasSubClassEq(RC))
8382 Opc = AArch64::FNMADDSrrr;
8383 else if (AArch64::FPR64RegClass.hasSubClassEq(RC))
8384 Opc = AArch64::FNMADDDrrr;
8396 MRI.constrainRegClass(ResultReg, RC);
8398 MRI.constrainRegClass(SrcReg0, RC);
8400 MRI.constrainRegClass(SrcReg1, RC);
8402 MRI.constrainRegClass(SrcReg2, RC);
8418 unsigned IdxDupOp,
unsigned MulOpc,
8420 assert(((IdxDupOp == 1) || (IdxDupOp == 2)) &&
8421 "Invalid index of FMUL operand");
8429 if (Dup->
getOpcode() == TargetOpcode::COPY)
8433 MRI.clearKillFlags(DupSrcReg);
8434 MRI.constrainRegClass(DupSrcReg, RC);
8438 unsigned IdxMulOp = IdxDupOp == 1 ? 2 : 1;
8479 InstrIdxForVirtReg.
insert(std::make_pair(NewVR, 0));
8494 genNeg(MF,
MRI,
TII, Root, InsInstrs, InstrIdxForVirtReg, MnegOpc, RC);
8521 genNeg(MF,
MRI,
TII, Root, InsInstrs, InstrIdxForVirtReg, MnegOpc, RC);
8549 unsigned IdxMulOpd,
unsigned MaddOpc,
unsigned VR,
8551 assert(IdxMulOpd == 1 || IdxMulOpd == 2);
8555 Register SrcReg0 = MUL->getOperand(1).getReg();
8556 bool Src0IsKill = MUL->getOperand(1).isKill();
8557 Register SrcReg1 = MUL->getOperand(2).getReg();
8558 bool Src1IsKill = MUL->getOperand(2).isKill();
8561 MRI.constrainRegClass(ResultReg, RC);
8563 MRI.constrainRegClass(SrcReg0, RC);
8565 MRI.constrainRegClass(SrcReg1, RC);
8567 MRI.constrainRegClass(VR, RC);
8588 assert(IdxOpd1 == 1 || IdxOpd1 == 2);
8589 unsigned IdxOtherOpd = IdxOpd1 == 1 ? 2 : 1;
8603 if (Opcode == AArch64::SUBSWrr)
8604 Opcode = AArch64::SUBWrr;
8605 else if (Opcode == AArch64::SUBSXrr)
8606 Opcode = AArch64::SUBXrr;
8608 assert((Opcode == AArch64::SUBWrr || Opcode == AArch64::SUBXrr) &&
8609 "Unexpected instruction opcode.");
8626 InstrIdxForVirtReg.
insert(std::make_pair(NewVR, 0));
8633unsigned AArch64InstrInfo::getReduceOpcodeForAccumulator(
8634 unsigned int AccumulatorOpCode)
const {
8635 switch (AccumulatorOpCode) {
8636 case AArch64::UABALB_ZZZ_D:
8637 case AArch64::SABALB_ZZZ_D:
8638 case AArch64::UABALT_ZZZ_D:
8639 case AArch64::SABALT_ZZZ_D:
8640 return AArch64::ADD_ZZZ_D;
8641 case AArch64::UABALB_ZZZ_H:
8642 case AArch64::SABALB_ZZZ_H:
8643 case AArch64::UABALT_ZZZ_H:
8644 case AArch64::SABALT_ZZZ_H:
8645 return AArch64::ADD_ZZZ_H;
8646 case AArch64::UABALB_ZZZ_S:
8647 case AArch64::SABALB_ZZZ_S:
8648 case AArch64::UABALT_ZZZ_S:
8649 case AArch64::SABALT_ZZZ_S:
8650 return AArch64::ADD_ZZZ_S;
8651 case AArch64::UABALv16i8_v8i16:
8652 case AArch64::SABALv8i8_v8i16:
8653 case AArch64::SABAv8i16:
8654 case AArch64::UABAv8i16:
8655 return AArch64::ADDv8i16;
8656 case AArch64::SABALv2i32_v2i64:
8657 case AArch64::UABALv2i32_v2i64:
8658 case AArch64::SABALv4i32_v2i64:
8659 return AArch64::ADDv2i64;
8660 case AArch64::UABALv4i16_v4i32:
8661 case AArch64::SABALv4i16_v4i32:
8662 case AArch64::SABALv8i16_v4i32:
8663 case AArch64::SABAv4i32:
8664 case AArch64::UABAv4i32:
8665 return AArch64::ADDv4i32;
8666 case AArch64::UABALv4i32_v2i64:
8667 return AArch64::ADDv2i64;
8668 case AArch64::UABALv8i16_v4i32:
8669 return AArch64::ADDv4i32;
8670 case AArch64::UABALv8i8_v8i16:
8671 case AArch64::SABALv16i8_v8i16:
8672 return AArch64::ADDv8i16;
8673 case AArch64::UABAv16i8:
8674 case AArch64::SABAv16i8:
8675 return AArch64::ADDv16i8;
8676 case AArch64::UABAv4i16:
8677 case AArch64::SABAv4i16:
8678 return AArch64::ADDv4i16;
8679 case AArch64::UABAv2i32:
8680 case AArch64::SABAv2i32:
8681 return AArch64::ADDv2i32;
8682 case AArch64::UABAv8i8:
8683 case AArch64::SABAv8i8:
8684 return AArch64::ADDv8i8;
8693void AArch64InstrInfo::genAlternativeCodeSequence(
8703 MachineInstr *
MUL =
nullptr;
8704 const TargetRegisterClass *RC;
8710 DelInstrs, InstrIdxForVirtReg);
8716 InstrIdxForVirtReg);
8722 InstrIdxForVirtReg);
8731 Opc = AArch64::MADDWrrr;
8732 RC = &AArch64::GPR32RegClass;
8734 Opc = AArch64::MADDXrrr;
8735 RC = &AArch64::GPR64RegClass;
8746 Opc = AArch64::MADDWrrr;
8747 RC = &AArch64::GPR32RegClass;
8749 Opc = AArch64::MADDXrrr;
8750 RC = &AArch64::GPR64RegClass;
8763 const TargetRegisterClass *RC;
8764 unsigned BitSize, MovImm;
8767 MovImm = AArch64::MOVi32imm;
8768 RC = &AArch64::GPR32spRegClass;
8770 Opc = AArch64::MADDWrrr;
8771 RC = &AArch64::GPR32RegClass;
8773 MovImm = AArch64::MOVi64imm;
8774 RC = &AArch64::GPR64spRegClass;
8776 Opc = AArch64::MADDXrrr;
8777 RC = &AArch64::GPR64RegClass;
8788 uint64_t UImm =
SignExtend64(IsSub ? -Imm : Imm, BitSize);
8792 if (Insn.
size() != 1)
8794 MachineInstrBuilder MIB1 =
8795 BuildMI(MF, MIMetadata(Root),
TII->get(MovImm), NewVR)
8796 .
addImm(IsSub ? -Imm : Imm);
8798 InstrIdxForVirtReg.
insert(std::make_pair(NewVR, 0));
8809 const TargetRegisterClass *SubRC;
8810 unsigned SubOpc, ZeroReg;
8812 SubOpc = AArch64::SUBWrr;
8813 SubRC = &AArch64::GPR32spRegClass;
8814 ZeroReg = AArch64::WZR;
8815 Opc = AArch64::MADDWrrr;
8816 RC = &AArch64::GPR32RegClass;
8818 SubOpc = AArch64::SUBXrr;
8819 SubRC = &AArch64::GPR64spRegClass;
8820 ZeroReg = AArch64::XZR;
8821 Opc = AArch64::MADDXrrr;
8822 RC = &AArch64::GPR64RegClass;
8824 Register NewVR =
MRI.createVirtualRegister(SubRC);
8826 MachineInstrBuilder MIB1 =
8827 BuildMI(MF, MIMetadata(Root),
TII->get(SubOpc), NewVR)
8831 InstrIdxForVirtReg.
insert(std::make_pair(NewVR, 0));
8842 Opc = AArch64::MSUBWrrr;
8843 RC = &AArch64::GPR32RegClass;
8845 Opc = AArch64::MSUBXrrr;
8846 RC = &AArch64::GPR64RegClass;
8851 Opc = AArch64::MLAv8i8;
8852 RC = &AArch64::FPR64RegClass;
8856 Opc = AArch64::MLAv8i8;
8857 RC = &AArch64::FPR64RegClass;
8861 Opc = AArch64::MLAv16i8;
8862 RC = &AArch64::FPR128RegClass;
8866 Opc = AArch64::MLAv16i8;
8867 RC = &AArch64::FPR128RegClass;
8871 Opc = AArch64::MLAv4i16;
8872 RC = &AArch64::FPR64RegClass;
8876 Opc = AArch64::MLAv4i16;
8877 RC = &AArch64::FPR64RegClass;
8881 Opc = AArch64::MLAv8i16;
8882 RC = &AArch64::FPR128RegClass;
8886 Opc = AArch64::MLAv8i16;
8887 RC = &AArch64::FPR128RegClass;
8891 Opc = AArch64::MLAv2i32;
8892 RC = &AArch64::FPR64RegClass;
8896 Opc = AArch64::MLAv2i32;
8897 RC = &AArch64::FPR64RegClass;
8901 Opc = AArch64::MLAv4i32;
8902 RC = &AArch64::FPR128RegClass;
8906 Opc = AArch64::MLAv4i32;
8907 RC = &AArch64::FPR128RegClass;
8912 Opc = AArch64::MLAv8i8;
8913 RC = &AArch64::FPR64RegClass;
8915 InstrIdxForVirtReg, 1,
Opc, AArch64::NEGv8i8,
8919 Opc = AArch64::MLSv8i8;
8920 RC = &AArch64::FPR64RegClass;
8924 Opc = AArch64::MLAv16i8;
8925 RC = &AArch64::FPR128RegClass;
8927 InstrIdxForVirtReg, 1,
Opc, AArch64::NEGv16i8,
8931 Opc = AArch64::MLSv16i8;
8932 RC = &AArch64::FPR128RegClass;
8936 Opc = AArch64::MLAv4i16;
8937 RC = &AArch64::FPR64RegClass;
8939 InstrIdxForVirtReg, 1,
Opc, AArch64::NEGv4i16,
8943 Opc = AArch64::MLSv4i16;
8944 RC = &AArch64::FPR64RegClass;
8948 Opc = AArch64::MLAv8i16;
8949 RC = &AArch64::FPR128RegClass;
8951 InstrIdxForVirtReg, 1,
Opc, AArch64::NEGv8i16,
8955 Opc = AArch64::MLSv8i16;
8956 RC = &AArch64::FPR128RegClass;
8960 Opc = AArch64::MLAv2i32;
8961 RC = &AArch64::FPR64RegClass;
8963 InstrIdxForVirtReg, 1,
Opc, AArch64::NEGv2i32,
8967 Opc = AArch64::MLSv2i32;
8968 RC = &AArch64::FPR64RegClass;
8972 Opc = AArch64::MLAv4i32;
8973 RC = &AArch64::FPR128RegClass;
8975 InstrIdxForVirtReg, 1,
Opc, AArch64::NEGv4i32,
8979 Opc = AArch64::MLSv4i32;
8980 RC = &AArch64::FPR128RegClass;
8985 Opc = AArch64::MLAv4i16_indexed;
8986 RC = &AArch64::FPR64RegClass;
8990 Opc = AArch64::MLAv4i16_indexed;
8991 RC = &AArch64::FPR64RegClass;
8995 Opc = AArch64::MLAv8i16_indexed;
8996 RC = &AArch64::FPR128RegClass;
9000 Opc = AArch64::MLAv8i16_indexed;
9001 RC = &AArch64::FPR128RegClass;
9005 Opc = AArch64::MLAv2i32_indexed;
9006 RC = &AArch64::FPR64RegClass;
9010 Opc = AArch64::MLAv2i32_indexed;
9011 RC = &AArch64::FPR64RegClass;
9015 Opc = AArch64::MLAv4i32_indexed;
9016 RC = &AArch64::FPR128RegClass;
9020 Opc = AArch64::MLAv4i32_indexed;
9021 RC = &AArch64::FPR128RegClass;
9026 Opc = AArch64::MLAv4i16_indexed;
9027 RC = &AArch64::FPR64RegClass;
9029 InstrIdxForVirtReg, 1,
Opc, AArch64::NEGv4i16,
9033 Opc = AArch64::MLSv4i16_indexed;
9034 RC = &AArch64::FPR64RegClass;
9038 Opc = AArch64::MLAv8i16_indexed;
9039 RC = &AArch64::FPR128RegClass;
9041 InstrIdxForVirtReg, 1,
Opc, AArch64::NEGv8i16,
9045 Opc = AArch64::MLSv8i16_indexed;
9046 RC = &AArch64::FPR128RegClass;
9050 Opc = AArch64::MLAv2i32_indexed;
9051 RC = &AArch64::FPR64RegClass;
9053 InstrIdxForVirtReg, 1,
Opc, AArch64::NEGv2i32,
9057 Opc = AArch64::MLSv2i32_indexed;
9058 RC = &AArch64::FPR64RegClass;
9062 Opc = AArch64::MLAv4i32_indexed;
9063 RC = &AArch64::FPR128RegClass;
9065 InstrIdxForVirtReg, 1,
Opc, AArch64::NEGv4i32,
9069 Opc = AArch64::MLSv4i32_indexed;
9070 RC = &AArch64::FPR128RegClass;
9076 Opc = AArch64::FMADDHrrr;
9077 RC = &AArch64::FPR16RegClass;
9081 Opc = AArch64::FMADDSrrr;
9082 RC = &AArch64::FPR32RegClass;
9086 Opc = AArch64::FMADDDrrr;
9087 RC = &AArch64::FPR64RegClass;
9092 Opc = AArch64::FMADDHrrr;
9093 RC = &AArch64::FPR16RegClass;
9097 Opc = AArch64::FMADDSrrr;
9098 RC = &AArch64::FPR32RegClass;
9102 Opc = AArch64::FMADDDrrr;
9103 RC = &AArch64::FPR64RegClass;
9108 Opc = AArch64::FMLAv1i32_indexed;
9109 RC = &AArch64::FPR32RegClass;
9114 Opc = AArch64::FMLAv1i32_indexed;
9115 RC = &AArch64::FPR32RegClass;
9121 Opc = AArch64::FMLAv1i64_indexed;
9122 RC = &AArch64::FPR64RegClass;
9127 Opc = AArch64::FMLAv1i64_indexed;
9128 RC = &AArch64::FPR64RegClass;
9134 RC = &AArch64::FPR64RegClass;
9135 Opc = AArch64::FMLAv4i16_indexed;
9140 RC = &AArch64::FPR64RegClass;
9141 Opc = AArch64::FMLAv4f16;
9146 RC = &AArch64::FPR64RegClass;
9147 Opc = AArch64::FMLAv4i16_indexed;
9152 RC = &AArch64::FPR64RegClass;
9153 Opc = AArch64::FMLAv4f16;
9160 RC = &AArch64::FPR64RegClass;
9162 Opc = AArch64::FMLAv2i32_indexed;
9166 Opc = AArch64::FMLAv2f32;
9173 RC = &AArch64::FPR64RegClass;
9175 Opc = AArch64::FMLAv2i32_indexed;
9179 Opc = AArch64::FMLAv2f32;
9186 RC = &AArch64::FPR128RegClass;
9187 Opc = AArch64::FMLAv8i16_indexed;
9192 RC = &AArch64::FPR128RegClass;
9193 Opc = AArch64::FMLAv8f16;
9198 RC = &AArch64::FPR128RegClass;
9199 Opc = AArch64::FMLAv8i16_indexed;
9204 RC = &AArch64::FPR128RegClass;
9205 Opc = AArch64::FMLAv8f16;
9212 RC = &AArch64::FPR128RegClass;
9214 Opc = AArch64::FMLAv2i64_indexed;
9218 Opc = AArch64::FMLAv2f64;
9225 RC = &AArch64::FPR128RegClass;
9227 Opc = AArch64::FMLAv2i64_indexed;
9231 Opc = AArch64::FMLAv2f64;
9239 RC = &AArch64::FPR128RegClass;
9241 Opc = AArch64::FMLAv4i32_indexed;
9245 Opc = AArch64::FMLAv4f32;
9253 RC = &AArch64::FPR128RegClass;
9255 Opc = AArch64::FMLAv4i32_indexed;
9259 Opc = AArch64::FMLAv4f32;
9266 Opc = AArch64::FNMSUBHrrr;
9267 RC = &AArch64::FPR16RegClass;
9271 Opc = AArch64::FNMSUBSrrr;
9272 RC = &AArch64::FPR32RegClass;
9276 Opc = AArch64::FNMSUBDrrr;
9277 RC = &AArch64::FPR64RegClass;
9282 Opc = AArch64::FNMADDHrrr;
9283 RC = &AArch64::FPR16RegClass;
9287 Opc = AArch64::FNMADDSrrr;
9288 RC = &AArch64::FPR32RegClass;
9292 Opc = AArch64::FNMADDDrrr;
9293 RC = &AArch64::FPR64RegClass;
9298 Opc = AArch64::FMSUBHrrr;
9299 RC = &AArch64::FPR16RegClass;
9303 Opc = AArch64::FMSUBSrrr;
9304 RC = &AArch64::FPR32RegClass;
9308 Opc = AArch64::FMSUBDrrr;
9309 RC = &AArch64::FPR64RegClass;
9314 Opc = AArch64::FMLSv1i32_indexed;
9315 RC = &AArch64::FPR32RegClass;
9321 Opc = AArch64::FMLSv1i64_indexed;
9322 RC = &AArch64::FPR64RegClass;
9329 RC = &AArch64::FPR64RegClass;
9331 MachineInstrBuilder MIB1 =
9332 BuildMI(MF, MIMetadata(Root),
TII->get(AArch64::FNEGv4f16), NewVR)
9335 InstrIdxForVirtReg.
insert(std::make_pair(NewVR, 0));
9337 Opc = AArch64::FMLAv4f16;
9341 Opc = AArch64::FMLAv4i16_indexed;
9348 RC = &AArch64::FPR64RegClass;
9349 Opc = AArch64::FMLSv4f16;
9354 RC = &AArch64::FPR64RegClass;
9355 Opc = AArch64::FMLSv4i16_indexed;
9362 RC = &AArch64::FPR64RegClass;
9364 Opc = AArch64::FMLSv2i32_indexed;
9368 Opc = AArch64::FMLSv2f32;
9376 RC = &AArch64::FPR128RegClass;
9378 MachineInstrBuilder MIB1 =
9379 BuildMI(MF, MIMetadata(Root),
TII->get(AArch64::FNEGv8f16), NewVR)
9382 InstrIdxForVirtReg.
insert(std::make_pair(NewVR, 0));
9384 Opc = AArch64::FMLAv8f16;
9388 Opc = AArch64::FMLAv8i16_indexed;
9395 RC = &AArch64::FPR128RegClass;
9396 Opc = AArch64::FMLSv8f16;
9401 RC = &AArch64::FPR128RegClass;
9402 Opc = AArch64::FMLSv8i16_indexed;
9409 RC = &AArch64::FPR128RegClass;
9411 Opc = AArch64::FMLSv2i64_indexed;
9415 Opc = AArch64::FMLSv2f64;
9423 RC = &AArch64::FPR128RegClass;
9425 Opc = AArch64::FMLSv4i32_indexed;
9429 Opc = AArch64::FMLSv4f32;
9436 RC = &AArch64::FPR64RegClass;
9438 MachineInstrBuilder MIB1 =
9439 BuildMI(MF, MIMetadata(Root),
TII->get(AArch64::FNEGv2f32), NewVR)
9442 InstrIdxForVirtReg.
insert(std::make_pair(NewVR, 0));
9444 Opc = AArch64::FMLAv2i32_indexed;
9448 Opc = AArch64::FMLAv2f32;
9456 RC = &AArch64::FPR128RegClass;
9458 MachineInstrBuilder MIB1 =
9459 BuildMI(MF, MIMetadata(Root),
TII->get(AArch64::FNEGv4f32), NewVR)
9462 InstrIdxForVirtReg.
insert(std::make_pair(NewVR, 0));
9464 Opc = AArch64::FMLAv4i32_indexed;
9468 Opc = AArch64::FMLAv4f32;
9476 RC = &AArch64::FPR128RegClass;
9478 MachineInstrBuilder MIB1 =
9479 BuildMI(MF, MIMetadata(Root),
TII->get(AArch64::FNEGv2f64), NewVR)
9482 InstrIdxForVirtReg.
insert(std::make_pair(NewVR, 0));
9484 Opc = AArch64::FMLAv2i64_indexed;
9488 Opc = AArch64::FMLAv2f64;
9500 &AArch64::FPR128RegClass,
MRI);
9509 &AArch64::FPR128RegClass,
MRI);
9518 &AArch64::FPR128_loRegClass,
MRI);
9527 &AArch64::FPR128RegClass,
MRI);
9536 &AArch64::FPR128_loRegClass,
MRI);
9570 for (
auto *
MI : InsInstrs)
9571 MI->setFlags(Flags);
9612 bool IsNegativeBranch =
false;
9613 bool IsTestAndBranch =
false;
9614 unsigned TargetBBInMI = 0;
9615 switch (
MI.getOpcode()) {
9619 case AArch64::CBWPri:
9620 case AArch64::CBXPri:
9621 case AArch64::CBBAssertExt:
9622 case AArch64::CBHAssertExt:
9623 case AArch64::CBWPrr:
9624 case AArch64::CBXPrr:
9630 case AArch64::CBNZW:
9631 case AArch64::CBNZX:
9633 IsNegativeBranch =
true;
9638 IsTestAndBranch =
true;
9640 case AArch64::TBNZW:
9641 case AArch64::TBNZX:
9643 IsNegativeBranch =
true;
9644 IsTestAndBranch =
true;
9650 if (IsTestAndBranch &&
MI.getOperand(1).getImm())
9654 assert(
MI.getParent() &&
"Incomplete machine instruction\n");
9665 while (
DefMI->isCopy()) {
9667 if (!
MRI->hasOneNonDBGUse(CopyVReg))
9669 if (!
MRI->hasOneDef(CopyVReg))
9674 switch (
DefMI->getOpcode()) {
9678 case AArch64::ANDWri:
9679 case AArch64::ANDXri: {
9680 if (IsTestAndBranch)
9684 if (!
MRI->hasOneNonDBGUse(VReg))
9687 bool Is32Bit = (
DefMI->getOpcode() == AArch64::ANDWri);
9689 DefMI->getOperand(2).getImm(), Is32Bit ? 32 : 64);
9698 assert(!
MRI->def_empty(NewReg) &&
"Register must be defined.");
9704 unsigned Opc = (Imm < 32)
9705 ? (IsNegativeBranch ? AArch64::TBNZW : AArch64::TBZW)
9706 : (IsNegativeBranch ? AArch64::TBNZX : AArch64::TBZX);
9719 if (!Is32Bit && Imm < 32)
9721 MI.eraseFromParent();
9725 case AArch64::CSINCWr:
9726 case AArch64::CSINCXr: {
9727 if (!(
DefMI->getOperand(1).getReg() == AArch64::WZR &&
9728 DefMI->getOperand(2).getReg() == AArch64::WZR) &&
9729 !(
DefMI->getOperand(1).getReg() == AArch64::XZR &&
9730 DefMI->getOperand(2).getReg() == AArch64::XZR))
9733 if (
DefMI->findRegisterDefOperandIdx(AArch64::NZCV,
nullptr,
9746 if (IsNegativeBranch)
9749 MI.eraseFromParent();
9755std::pair<unsigned, unsigned>
9756AArch64InstrInfo::decomposeMachineOperandsTargetFlags(
unsigned TF)
const {
9758 return std::make_pair(TF & Mask, TF & ~Mask);
9762AArch64InstrInfo::getSerializableDirectMachineOperandTargetFlags()
const {
9765 static const std::pair<unsigned, const char *> TargetFlags[] = {
9766 {MO_PAGE,
"aarch64-page"}, {
MO_PAGEOFF,
"aarch64-pageoff"},
9767 {
MO_G3,
"aarch64-g3"}, {
MO_G2,
"aarch64-g2"},
9768 {
MO_G1,
"aarch64-g1"}, {
MO_G0,
"aarch64-g0"},
9774AArch64InstrInfo::getSerializableBitmaskMachineOperandTargetFlags()
const {
9775 using namespace AArch64II;
9777 static const std::pair<unsigned, const char *> TargetFlags[] = {
9780 {
MO_NC,
"aarch64-nc"},
9781 {
MO_S,
"aarch64-s"},
9792AArch64InstrInfo::getSerializableMachineMemOperandTargetFlags()
const {
9793 static const std::pair<MachineMemOperand::Flags, const char *> TargetFlags[] =
9895 MachineFunction *MF =
C.getMF();
9897 const AArch64RegisterInfo *ARI =
9898 static_cast<const AArch64RegisterInfo *
>(&
TRI);
9901 for (
unsigned Reg : AArch64::GPR64RegClass) {
9903 Reg != AArch64::LR &&
9904 Reg != AArch64::X16 &&
9905 Reg != AArch64::X17 &&
9906 C.isAvailableAcrossAndOutOfSeq(
Reg,
TRI) &&
9907 C.isAvailableInsideSeq(
Reg,
TRI))
9938 return SubtargetA.hasV8_3aOps() == SubtargetB.hasV8_3aOps();
9941std::optional<std::unique_ptr<outliner::OutlinedFunction>>
9942AArch64InstrInfo::getOutliningCandidateInfo(
9944 std::vector<outliner::Candidate> &RepeatedSequenceLocs,
9945 unsigned MinRepeats)
const {
9946 unsigned SequenceSize = 0;
9947 for (
auto &
MI : RepeatedSequenceLocs[0])
9950 unsigned NumBytesToCreateFrame = 0;
9956 MachineInstr &LastMI = RepeatedSequenceLocs[0].back();
9957 MachineInstr &FirstMI = RepeatedSequenceLocs[0].front();
9958 if (LastMI.
getOpcode() == AArch64::ADRP &&
9961 return std::nullopt;
9966 if ((FirstMI.
getOpcode() == AArch64::ADDXri ||
9967 FirstMI.
getOpcode() == AArch64::LDRXui) &&
9970 return std::nullopt;
9981 if (std::adjacent_find(
9982 RepeatedSequenceLocs.begin(), RepeatedSequenceLocs.end(),
9983 [](
const outliner::Candidate &a,
const outliner::Candidate &b) {
9986 if (outliningCandidatesSigningScopeConsensus(a, b) &&
9987 outliningCandidatesSigningKeyConsensus(a, b) &&
9988 outliningCandidatesV8_3OpsConsensus(a, b)) {
9992 }) != RepeatedSequenceLocs.end()) {
9993 return std::nullopt;
10010 unsigned NumBytesToCheckLRInTCEpilogue = 0;
10011 const auto RASignCondition = RepeatedSequenceLocs[0]
10014 ->getSignReturnAddressCondition();
10017 NumBytesToCreateFrame += 8;
10020 auto LRCheckMethod = Subtarget.getAuthenticatedLRCheckMethod(
10021 *RepeatedSequenceLocs[0].getMF());
10022 NumBytesToCheckLRInTCEpilogue =
10026 if (isTailCallReturnInst(RepeatedSequenceLocs[0].
back()))
10027 SequenceSize += NumBytesToCheckLRInTCEpilogue;
10035 for (
auto &
MI :
C) {
10036 if (
MI.modifiesRegister(AArch64::SP, &
TRI)) {
10037 switch (
MI.getOpcode()) {
10038 case AArch64::ADDXri:
10039 case AArch64::ADDWri:
10040 assert(
MI.getNumOperands() == 4 &&
"Wrong number of operands");
10041 assert(
MI.getOperand(2).isImm() &&
10042 "Expected operand to be immediate");
10043 assert(
MI.getOperand(1).isReg() &&
10044 "Expected operand to be a register");
10048 if (
MI.getOperand(1).getReg() == AArch64::SP)
10049 SPValue +=
MI.getOperand(2).getImm();
10053 case AArch64::SUBXri:
10054 case AArch64::SUBWri:
10055 assert(
MI.getNumOperands() == 4 &&
"Wrong number of operands");
10056 assert(
MI.getOperand(2).isImm() &&
10057 "Expected operand to be immediate");
10058 assert(
MI.getOperand(1).isReg() &&
10059 "Expected operand to be a register");
10063 if (
MI.getOperand(1).getReg() == AArch64::SP)
10064 SPValue -=
MI.getOperand(2).getImm();
10081 if (RepeatedSequenceLocs.size() < MinRepeats)
10082 return std::nullopt;
10086 unsigned FlagsSetInAll = 0xF;
10090 FlagsSetInAll &=
C.Flags;
10092 unsigned LastInstrOpcode = RepeatedSequenceLocs[0].back().getOpcode();
10095 auto SetCandidateCallInfo =
10096 [&RepeatedSequenceLocs](
unsigned CallID,
unsigned NumBytesForCall) {
10098 C.setCallInfo(CallID, NumBytesForCall);
10102 NumBytesToCreateFrame += 4;
10110 unsigned CFICount = 0;
10111 for (
auto &
I : RepeatedSequenceLocs[0]) {
10112 if (
I.isCFIInstruction())
10122 std::vector<MCCFIInstruction> CFIInstructions =
10123 C.getMF()->getFrameInstructions();
10125 if (CFICount > 0 && CFICount != CFIInstructions.size())
10126 return std::nullopt;
10134 if (!
MI.modifiesRegister(AArch64::SP, &
TRI) &&
10135 !
MI.readsRegister(AArch64::SP, &
TRI))
10141 if (
MI.modifiesRegister(AArch64::SP, &
TRI))
10146 if (
MI.mayLoadOrStore()) {
10149 bool OffsetIsScalable;
10153 if (!getMemOperandWithOffset(
MI,
Base,
Offset, OffsetIsScalable, &
TRI) ||
10154 !
Base->isReg() ||
Base->getReg() != AArch64::SP)
10158 if (OffsetIsScalable)
10166 TypeSize Scale(0U,
false), DummyWidth(0U,
false);
10167 getMemOpInfo(
MI.getOpcode(), Scale, DummyWidth, MinOffset, MaxOffset);
10170 if (
Offset < MinOffset * (int64_t)Scale.getFixedValue() ||
10171 Offset > MaxOffset * (int64_t)Scale.getFixedValue())
10186 bool AllStackInstrsSafe =
10191 if (RepeatedSequenceLocs[0].
back().isTerminator()) {
10193 NumBytesToCreateFrame = 0;
10194 unsigned NumBytesForCall = 4 + NumBytesToCheckLRInTCEpilogue;
10198 else if (LastInstrOpcode == AArch64::BL ||
10199 ((LastInstrOpcode == AArch64::BLR ||
10200 LastInstrOpcode == AArch64::BLRNoIP) &&
10204 NumBytesToCreateFrame = NumBytesToCheckLRInTCEpilogue;
10212 unsigned NumBytesNoStackCalls = 0;
10213 std::vector<outliner::Candidate> CandidatesWithoutStackFixups;
10219 ?
C.isAvailableAcrossAndOutOfSeq(AArch64::LR,
TRI)
10228 C.getMF()->getFunction().hasFnAttribute(Attribute::NoReturn);
10231 if (LRAvailable && !IsNoReturn) {
10232 NumBytesNoStackCalls += 4;
10234 CandidatesWithoutStackFixups.push_back(
C);
10239 else if (findRegisterToSaveLRTo(
C)) {
10240 NumBytesNoStackCalls += 12;
10242 CandidatesWithoutStackFixups.push_back(
C);
10247 else if (
C.isAvailableInsideSeq(AArch64::SP,
TRI)) {
10248 NumBytesNoStackCalls += 12;
10250 CandidatesWithoutStackFixups.push_back(
C);
10256 NumBytesNoStackCalls += SequenceSize;
10263 if (!AllStackInstrsSafe ||
10264 NumBytesNoStackCalls <= RepeatedSequenceLocs.size() * 12) {
10265 RepeatedSequenceLocs = CandidatesWithoutStackFixups;
10267 if (RepeatedSequenceLocs.size() < MinRepeats)
10268 return std::nullopt;
10321 (!
C.isAvailableAcrossAndOutOfSeq(AArch64::LR,
TRI) ||
10322 !findRegisterToSaveLRTo(
C));
10328 if (RepeatedSequenceLocs.size() < MinRepeats)
10329 return std::nullopt;
10338 bool ModStackToSaveLR =
false;
10341 ModStackToSaveLR =
true;
10350 ModStackToSaveLR =
true;
10352 if (ModStackToSaveLR) {
10354 if (!AllStackInstrsSafe)
10355 return std::nullopt;
10358 NumBytesToCreateFrame += 8;
10365 return std::nullopt;
10367 return std::make_unique<outliner::OutlinedFunction>(
10368 RepeatedSequenceLocs, SequenceSize, NumBytesToCreateFrame, FrameID);
10371void AArch64InstrInfo::mergeOutliningCandidateAttributes(
10372 Function &
F, std::vector<outliner::Candidate> &Candidates)
const {
10376 const auto &CFn = Candidates.front().getMF()->getFunction();
10378 if (CFn.hasFnAttribute(
"ptrauth-returns"))
10379 F.addFnAttr(CFn.getFnAttribute(
"ptrauth-returns"));
10380 if (CFn.hasFnAttribute(
"ptrauth-auth-traps"))
10381 F.addFnAttr(CFn.getFnAttribute(
"ptrauth-auth-traps"));
10384 if (CFn.hasFnAttribute(
"sign-return-address"))
10385 F.addFnAttr(CFn.getFnAttribute(
"sign-return-address"));
10386 if (CFn.hasFnAttribute(
"sign-return-address-key"))
10387 F.addFnAttr(CFn.getFnAttribute(
"sign-return-address-key"));
10389 AArch64GenInstrInfo::mergeOutliningCandidateAttributes(
F, Candidates);
10392bool AArch64InstrInfo::isFunctionSafeToOutlineFrom(
10397 if (!OutlineFromLinkOnceODRs &&
F.hasLinkOnceODRLinkage())
10404 if (
F.hasSection())
10410 AArch64FunctionInfo *AFI = MF.
getInfo<AArch64FunctionInfo>();
10411 if (!AFI || AFI->
hasRedZone().value_or(
true))
10431 unsigned &Flags)
const {
10433 "Must track liveness!");
10435 std::pair<MachineBasicBlock::iterator, MachineBasicBlock::iterator>>
10450 auto AreAllUnsafeRegsDead = [&LRU]() {
10451 return LRU.available(AArch64::W16) && LRU.available(AArch64::W17) &&
10452 LRU.available(AArch64::NZCV);
10467 bool LRAvailableEverywhere =
true;
10469 LRU.addLiveOuts(
MBB);
10471 auto UpdateWholeMBBFlags = [&
Flags](
const MachineInstr &
MI) {
10472 if (
MI.isCall() && !
MI.isTerminator())
10478 auto CreateNewRangeStartingAt =
10479 [&RangeBegin, &RangeEnd,
10481 RangeBegin = NewBegin;
10482 RangeEnd = std::next(RangeBegin);
10485 auto SaveRangeIfNonEmpty = [&RangeLen, &
Ranges, &RangeBegin, &RangeEnd]() {
10491 if (!RangeBegin.isEnd() && RangeBegin->isBundledWithPred())
10493 if (!RangeEnd.isEnd() && RangeEnd->isBundledWithPred())
10495 Ranges.emplace_back(RangeBegin, RangeEnd);
10503 for (; FirstPossibleEndPt !=
MBB.
instr_rend(); ++FirstPossibleEndPt) {
10504 LRU.stepBackward(*FirstPossibleEndPt);
10507 UpdateWholeMBBFlags(*FirstPossibleEndPt);
10508 if (AreAllUnsafeRegsDead())
10515 CreateNewRangeStartingAt(FirstPossibleEndPt->getIterator());
10520 LRU.stepBackward(
MI);
10521 UpdateWholeMBBFlags(
MI);
10522 if (!AreAllUnsafeRegsDead()) {
10523 SaveRangeIfNonEmpty();
10524 CreateNewRangeStartingAt(
MI.getIterator());
10527 LRAvailableEverywhere &= LRU.available(AArch64::LR);
10528 RangeBegin =
MI.getIterator();
10533 if (AreAllUnsafeRegsDead())
10534 SaveRangeIfNonEmpty();
10542 if (!LRAvailableEverywhere)
10550 unsigned Flags)
const {
10551 MachineInstr &
MI = *MIT;
10555 switch (
MI.getOpcode()) {
10556 case AArch64::PACM:
10557 case AArch64::PACIASP:
10558 case AArch64::PACIBSP:
10559 case AArch64::PACIASPPC:
10560 case AArch64::PACIBSPPC:
10561 case AArch64::AUTIASP:
10562 case AArch64::AUTIBSP:
10563 case AArch64::AUTIASPPCi:
10564 case AArch64::AUTIASPPCr:
10565 case AArch64::AUTIBSPPCi:
10566 case AArch64::AUTIBSPPCr:
10567 case AArch64::RETAA:
10568 case AArch64::RETAB:
10569 case AArch64::RETAASPPCi:
10570 case AArch64::RETAASPPCr:
10571 case AArch64::RETABSPPCi:
10572 case AArch64::RETABSPPCr:
10573 case AArch64::EMITBKEY:
10574 case AArch64::PAUTH_PROLOGUE:
10575 case AArch64::PAUTH_EPILOGUE:
10585 if (
MI.isCFIInstruction())
10589 if (
MI.isTerminator())
10595 for (
const MachineOperand &MOP :
MI.operands()) {
10598 assert(!MOP.isCFIIndex());
10601 if (MOP.isReg() && !MOP.isImplicit() &&
10602 (MOP.getReg() == AArch64::LR || MOP.getReg() == AArch64::W30))
10609 if (
MI.getOpcode() == AArch64::ADRP)
10629 for (
const MachineOperand &MOP :
MI.operands()) {
10630 if (MOP.isGlobal()) {
10638 if (Callee &&
Callee->getName() ==
"\01_mcount")
10646 if (
MI.getOpcode() == AArch64::BLR ||
10647 MI.getOpcode() == AArch64::BLRNoIP ||
MI.getOpcode() == AArch64::BL)
10651 return UnknownCallOutlineType;
10659 return UnknownCallOutlineType;
10667 return UnknownCallOutlineType;
10688 for (MachineInstr &
MI :
MBB) {
10689 const MachineOperand *
Base;
10690 TypeSize Width(0,
false);
10692 bool OffsetIsScalable;
10695 if (!
MI.mayLoadOrStore() ||
10698 (
Base->isReg() &&
Base->getReg() != AArch64::SP))
10702 TypeSize Scale(0U,
false);
10703 int64_t Dummy1, Dummy2;
10706 assert(StackOffsetOperand.
isImm() &&
"Stack offset wasn't immediate!");
10708 assert(Scale != 0 &&
"Unexpected opcode!");
10709 assert(!OffsetIsScalable &&
"Expected offset to be a byte offset");
10714 int64_t NewImm = (
Offset + 16) / (int64_t)Scale.getFixedValue();
10715 StackOffsetOperand.
setImm(NewImm);
10721 bool ShouldSignReturnAddr) {
10722 if (!ShouldSignReturnAddr)
10728 TII->get(AArch64::PAUTH_EPILOGUE))
10732void AArch64InstrInfo::buildOutlinedFrame(
10736 AArch64FunctionInfo *FI = MF.
getInfo<AArch64FunctionInfo>();
10744 unsigned TailOpcode;
10746 TailOpcode = AArch64::TCRETURNdi;
10750 TailOpcode = AArch64::TCRETURNriALL;
10761 bool IsLeafFunction =
true;
10764 auto IsNonTailCall = [](
const MachineInstr &
MI) {
10765 return MI.isCall() && !
MI.isReturn();
10775 "Can only fix up stack references once");
10776 fixupPostOutline(
MBB);
10778 IsLeafFunction =
false;
10789 Et = std::prev(
MBB.
end());
10799 if (MF.
getInfo<AArch64FunctionInfo>()->needsDwarfUnwindInfo(MF)) {
10803 CFIBuilder.buildDefCFAOffset(16);
10807 CFIBuilder.buildOffset(AArch64::LR, -16);
10821 RASignCondition, !IsLeafFunction);
10850 fixupPostOutline(
MBB);
10861 .addGlobalAddress(
M.getNamedValue(MF.
getName()))
10871 .addGlobalAddress(
M.getNamedValue(MF.
getName())));
10880 MachineInstr *Save;
10881 MachineInstr *Restore;
10887 assert(
Reg &&
"No callee-saved register available?");
10921 .addGlobalAddress(
M.getNamedValue(MF.
getName())));
10929bool AArch64InstrInfo::shouldOutlineFromFunctionByDefault(
10937 bool AllowSideEffects)
const {
10939 const AArch64Subtarget &STI = MF.
getSubtarget<AArch64Subtarget>();
10942 if (
TRI.isGeneralPurposeRegister(MF,
Reg)) {
10955 assert(STI.hasNEON() &&
"Expected to have NEON.");
10961std::optional<DestSourcePair>
10966 if (((
MI.getOpcode() == AArch64::ORRWrs &&
10967 MI.getOperand(1).getReg() == AArch64::WZR &&
10968 MI.getOperand(3).getImm() == 0x0) ||
10969 (
MI.getOpcode() == AArch64::ORRWrr &&
10970 MI.getOperand(1).getReg() == AArch64::WZR)) &&
10972 (!
MI.getOperand(0).getReg().isVirtual() ||
10973 MI.getOperand(0).getSubReg() == 0) &&
10974 (!
MI.getOperand(0).getReg().isPhysical() ||
10979 if (
MI.getOpcode() == AArch64::ORRXrs &&
10980 MI.getOperand(1).getReg() == AArch64::XZR &&
10981 MI.getOperand(3).getImm() == 0x0)
10984 return std::nullopt;
10987std::optional<DestSourcePair>
10989 if ((
MI.getOpcode() == AArch64::ORRWrs &&
10990 MI.getOperand(1).getReg() == AArch64::WZR &&
10991 MI.getOperand(3).getImm() == 0x0) ||
10992 (
MI.getOpcode() == AArch64::ORRWrr &&
10993 MI.getOperand(1).getReg() == AArch64::WZR))
10995 return std::nullopt;
10998std::optional<RegImmPair>
11007 return std::nullopt;
11009 switch (
MI.getOpcode()) {
11011 return std::nullopt;
11012 case AArch64::SUBWri:
11013 case AArch64::SUBXri:
11014 case AArch64::SUBSWri:
11015 case AArch64::SUBSXri:
11018 case AArch64::ADDSWri:
11019 case AArch64::ADDSXri:
11020 case AArch64::ADDWri:
11021 case AArch64::ADDXri: {
11023 if (!
MI.getOperand(0).isReg() || !
MI.getOperand(1).isReg() ||
11024 !
MI.getOperand(2).isImm())
11025 return std::nullopt;
11026 int Shift =
MI.getOperand(3).getImm();
11027 assert((Shift == 0 || Shift == 12) &&
"Shift can be either 0 or 12");
11031 return RegImmPair{
MI.getOperand(1).getReg(),
Offset};
11037static std::optional<ParamLoadedValue>
11041 auto DestSrc =
TII->isCopyLikeInstr(
MI);
11043 return std::nullopt;
11045 Register DestReg = DestSrc->Destination->getReg();
11046 Register SrcReg = DestSrc->Source->getReg();
11049 return std::nullopt;
11054 if (DestReg == DescribedReg)
11058 if (
MI.getOpcode() == AArch64::ORRWrs &&
11059 TRI->isSuperRegister(DestReg, DescribedReg))
11063 if (
MI.getOpcode() == AArch64::ORRXrs &&
11064 TRI->isSubRegister(DestReg, DescribedReg)) {
11065 Register SrcSubReg =
TRI->getSubReg(SrcReg, AArch64::sub_32);
11069 assert(!
TRI->isSuperOrSubRegisterEq(DestReg, DescribedReg) &&
11070 "Unhandled ORR[XW]rs copy case");
11072 return std::nullopt;
11075bool AArch64InstrInfo::isFunctionSafeToSplit(
const MachineFunction &MF)
const {
11080 if (MF.
getInfo<AArch64FunctionInfo>()->hasRedZone().value_or(
true))
11086bool AArch64InstrInfo::isMBBSafeToSplitToCold(
11090 auto isAsmGoto = [](
const MachineInstr &
MI) {
11091 return MI.getOpcode() == AArch64::INLINEASM_BR;
11101 auto containsMBB = [&
MBB](
const MachineJumpTableEntry &JTE) {
11108 for (
const MachineInstr &
MI :
MBB) {
11109 switch (
MI.getOpcode()) {
11110 case TargetOpcode::G_BRJT:
11111 case AArch64::JumpTableDest32:
11112 case AArch64::JumpTableDest16:
11113 case AArch64::JumpTableDest8:
11124std::optional<ParamLoadedValue>
11127 const MachineFunction *MF =
MI.getMF();
11129 switch (
MI.getOpcode()) {
11130 case AArch64::MOVZWi:
11131 case AArch64::MOVZXi: {
11134 if (!
TRI->isSuperRegisterEq(
MI.getOperand(0).getReg(),
Reg))
11135 return std::nullopt;
11137 if (!
MI.getOperand(1).isImm())
11138 return std::nullopt;
11139 int64_t Immediate =
MI.getOperand(1).getImm();
11140 int Shift =
MI.getOperand(2).getImm();
11144 case AArch64::ORRWrs:
11145 case AArch64::ORRXrs:
11152bool AArch64InstrInfo::isExtendLikelyToBeFolded(
11155 ExtMI.
getOpcode() == TargetOpcode::G_ZEXT ||
11156 ExtMI.
getOpcode() == TargetOpcode::G_ANYEXT);
11159 if (ExtMI.
getOpcode() == TargetOpcode::G_ANYEXT)
11163 if (!
MRI.hasOneNonDBGUse(DefReg))
11168 auto *UserMI = &*
MRI.use_instr_nodbg_begin(DefReg);
11169 return UserMI->getOpcode() == TargetOpcode::G_PTR_ADD;
11172uint64_t AArch64InstrInfo::getElementSizeForOpcode(
unsigned Opc)
const {
11176bool AArch64InstrInfo::isPTestLikeOpcode(
unsigned Opc)
const {
11180bool AArch64InstrInfo::isWhileOpcode(
unsigned Opc)
const {
11185AArch64InstrInfo::getTailDuplicateSize(
CodeGenOptLevel OptLevel)
const {
11189bool AArch64InstrInfo::isLegalAddressingMode(
unsigned NumBytes, int64_t
Offset,
11190 unsigned Scale)
const {
11201 unsigned Shift =
Log2_64(NumBytes);
11202 if (NumBytes &&
Offset > 0 && (
Offset / NumBytes) <= (1LL << 12) - 1 &&
11210 return Scale == 1 || (Scale > 0 && Scale == NumBytes);
11215 return AArch64::BLRNoIP;
11217 return AArch64::BLR;
11222 Register TargetReg,
bool FrameSetup)
const {
11223 assert(TargetReg != AArch64::SP &&
"New top of stack cannot already be in SP");
11235 MF.
insert(MBBInsertPoint, LoopTestMBB);
11238 MF.
insert(MBBInsertPoint, LoopBodyMBB);
11240 MF.
insert(MBBInsertPoint, ExitMBB);
11250 BuildMI(*LoopTestMBB, LoopTestMBB->
end(),
DL,
TII->get(AArch64::SUBSXrx64),
11258 BuildMI(*LoopTestMBB, LoopTestMBB->
end(),
DL,
TII->get(AArch64::Bcc))
11264 BuildMI(*LoopBodyMBB, LoopBodyMBB->
end(),
DL,
TII->get(AArch64::STRXui))
11277 BuildMI(*ExitMBB, ExitMBB->
end(),
DL,
TII->get(AArch64::ADDXri), AArch64::SP)
11296 MBB.addSuccessor(LoopTestMBB);
11302 return ExitMBB->
begin();
11319 unsigned CompCounterOprNum;
11323 unsigned UpdateCounterOprNum;
11327 bool IsUpdatePriorComp;
11339 TII(MF->getSubtarget().getInstrInfo()),
11340 TRI(MF->getSubtarget().getRegisterInfo()),
MRI(MF->getRegInfo()),
11341 LoopBB(LoopBB), CondBranch(CondBranch), Comp(Comp),
11342 CompCounterOprNum(CompCounterOprNum), Update(Update),
11343 UpdateCounterOprNum(UpdateCounterOprNum),
Init(
Init),
11344 IsUpdatePriorComp(IsUpdatePriorComp),
Cond(
Cond.begin(),
Cond.end()) {}
11346 bool shouldIgnoreForPipelining(
const MachineInstr *
MI)
const override {
11352 std::optional<bool> createTripCountGreaterCondition(
11353 int TC, MachineBasicBlock &
MBB,
11354 SmallVectorImpl<MachineOperand> &CondParam)
override {
11362 void createRemainingIterationsGreaterCondition(
11363 int TC, MachineBasicBlock &
MBB, SmallVectorImpl<MachineOperand> &
Cond,
11364 DenseMap<MachineInstr *, MachineInstr *> &LastStage0Insts)
override;
11366 void setPreheader(MachineBasicBlock *NewPreheader)
override {}
11368 void adjustTripCount(
int TripCountAdjust)
override {}
11370 bool isMVEExpanderSupported()
override {
return true; }
11386 Result =
MRI.createVirtualRegister(
11389 }
else if (
I == ReplaceOprNum) {
11390 MRI.constrainRegClass(ReplaceReg,
TII->getRegClass(NewMI->
getDesc(),
I));
11394 MBB.insert(InsertTo, NewMI);
11398void AArch64PipelinerLoopInfo::createRemainingIterationsGreaterCondition(
11414 assert(CondBranch->getOpcode() == AArch64::Bcc);
11418 if (CondBranch->getOperand(1).getMBB() == LoopBB)
11425 auto AccumulateCond = [&](
Register CurCond,
11427 Register NewCond =
MRI.createVirtualRegister(&AArch64::GPR64commonRegClass);
11436 if (!LastStage0Insts.
empty() && LastStage0Insts[Comp]->getParent() == &
MBB) {
11440 for (
int I = 0;
I <= TC; ++
I) {
11446 AccCond = AccumulateCond(AccCond, CC);
11450 if (Update != Comp && IsUpdatePriorComp) {
11452 LastStage0Insts[Comp]->getOperand(CompCounterOprNum).getReg();
11453 NextCounter =
cloneInstr(Update, UpdateCounterOprNum, Counter,
MBB,
11457 NextCounter = LastStage0Insts[Update]->getOperand(0).getReg();
11459 }
else if (Update != Comp) {
11464 Counter = NextCounter;
11468 if (LastStage0Insts.
empty()) {
11472 if (IsUpdatePriorComp)
11477 Counter = LastStage0Insts[Comp]->getOperand(CompCounterOprNum).getReg();
11480 for (
int I = 0;
I <= TC; ++
I) {
11484 AccCond = AccumulateCond(AccCond, CC);
11485 if (
I != TC && Update != Comp)
11488 Counter = NextCounter;
11504 assert(Phi.getNumOperands() == 5);
11505 if (Phi.getOperand(2).getMBB() ==
MBB) {
11506 RegMBB = Phi.getOperand(1).getReg();
11507 RegOther = Phi.getOperand(3).getReg();
11509 assert(Phi.getOperand(4).getMBB() ==
MBB);
11510 RegMBB = Phi.getOperand(3).getReg();
11511 RegOther = Phi.getOperand(1).getReg();
11516 if (!
Reg.isVirtual())
11519 return MRI.getVRegDef(
Reg)->getParent() != BB;
11525 unsigned &UpdateCounterOprNum,
Register &InitReg,
11526 bool &IsUpdatePriorComp) {
11540 if (!
Reg.isVirtual())
11543 UpdateInst =
nullptr;
11544 UpdateCounterOprNum = 0;
11546 IsUpdatePriorComp =
true;
11550 if (Def->getParent() != LoopBB)
11552 if (Def->isCopy()) {
11554 if (Def->getOperand(0).getSubReg() || Def->getOperand(1).getSubReg())
11556 CurReg = Def->getOperand(1).getReg();
11557 }
else if (Def->isPHI()) {
11561 IsUpdatePriorComp =
false;
11566 switch (Def->getOpcode()) {
11567 case AArch64::ADDSXri:
11568 case AArch64::ADDSWri:
11569 case AArch64::SUBSXri:
11570 case AArch64::SUBSWri:
11571 case AArch64::ADDXri:
11572 case AArch64::ADDWri:
11573 case AArch64::SUBXri:
11574 case AArch64::SUBWri:
11576 UpdateCounterOprNum = 1;
11578 case AArch64::ADDSXrr:
11579 case AArch64::ADDSWrr:
11580 case AArch64::SUBSXrr:
11581 case AArch64::SUBSWrr:
11582 case AArch64::ADDXrr:
11583 case AArch64::ADDWrr:
11584 case AArch64::SUBXrr:
11585 case AArch64::SUBWrr:
11588 UpdateCounterOprNum = 1;
11590 UpdateCounterOprNum = 2;
11597 CurReg = Def->getOperand(UpdateCounterOprNum).getReg();
11612std::unique_ptr<TargetInstrInfo::PipelinerLoopInfo>
11623 if (
MI.isCall() ||
MI.hasUnmodeledSideEffects())
11634 if (
TBB == LoopBB && FBB == LoopBB)
11638 if (
TBB != LoopBB && FBB ==
nullptr)
11641 assert((
TBB == LoopBB || FBB == LoopBB) &&
11642 "The Loop must be a single-basic-block loop");
11647 if (CondBranch->
getOpcode() != AArch64::Bcc)
11655 unsigned CompCounterOprNum = 0;
11657 if (
MI.modifiesRegister(AArch64::NZCV, &
TRI)) {
11661 switch (
MI.getOpcode()) {
11662 case AArch64::SUBSXri:
11663 case AArch64::SUBSWri:
11664 case AArch64::ADDSXri:
11665 case AArch64::ADDSWri:
11667 CompCounterOprNum = 1;
11669 case AArch64::ADDSWrr:
11670 case AArch64::ADDSXrr:
11671 case AArch64::SUBSWrr:
11672 case AArch64::SUBSXrr:
11676 if (isWhileOpcode(
MI.getOpcode())) {
11683 if (CompCounterOprNum == 0) {
11685 CompCounterOprNum = 2;
11687 CompCounterOprNum = 1;
11699 bool IsUpdatePriorComp;
11700 unsigned UpdateCounterOprNum;
11702 Update, UpdateCounterOprNum,
Init, IsUpdatePriorComp))
11705 return std::make_unique<AArch64PipelinerLoopInfo>(
11706 LoopBB, CondBranch, Comp, CompCounterOprNum, Update, UpdateCounterOprNum,
11716 TypeSize Scale(0U,
false), Width(0U,
false);
11717 int64_t MinOffset, MaxOffset;
11718 if (
getMemOpInfo(
MI.getOpcode(), Scale, Width, MinOffset, MaxOffset)) {
11720 if (
MI.getOperand(ImmIdx).isImm() && !
MI.getOperand(ImmIdx - 1).isFI()) {
11721 int64_t Imm =
MI.getOperand(ImmIdx).getImm();
11722 if (Imm < MinOffset || Imm > MaxOffset) {
11723 ErrInfo =
"Unexpected immediate on load/store instruction";
11729 const MCInstrDesc &MCID =
MI.getDesc();
11731 const MachineOperand &MO =
MI.getOperand(
Op);
11735 ErrInfo =
"OPERAND_IMPLICIT_IMM_0 should be 0";
11744 ErrInfo =
"OPERAND_SHIFT_MSL should be msl shift of 8 or 16";
11755#define GET_INSTRINFO_HELPERS
11756#define GET_INSTRMAP_INFO
11757#include "AArch64GenInstrInfo.inc"
unsigned const MachineRegisterInfo * MRI
MachineInstrBuilder & UseMI
MachineInstrBuilder MachineInstrBuilder & DefMI
static bool forwardCopyWillClobberTuple(unsigned DestReg, unsigned SrcReg, unsigned NumRegs)
static cl::opt< unsigned > BCCDisplacementBits("aarch64-bcc-offset-bits", cl::Hidden, cl::init(19), cl::desc("Restrict range of Bcc instructions (DEBUG)"))
static Register genNeg(MachineFunction &MF, MachineRegisterInfo &MRI, const TargetInstrInfo *TII, MachineInstr &Root, SmallVectorImpl< MachineInstr * > &InsInstrs, DenseMap< Register, unsigned > &InstrIdxForVirtReg, unsigned MnegOpc, const TargetRegisterClass *RC)
genNeg - Helper to generate an intermediate negation of the second operand of Root
static bool isFrameStoreOpcode(int Opcode)
static cl::opt< unsigned > GatherOptSearchLimit("aarch64-search-limit", cl::Hidden, cl::init(2048), cl::desc("Restrict range of instructions to search for the " "machine-combiner gather pattern optimization"))
static bool getMaddPatterns(MachineInstr &Root, SmallVectorImpl< unsigned > &Patterns)
Find instructions that can be turned into madd.
static AArch64CC::CondCode findCondCodeUsedByInstr(const MachineInstr &Instr)
Find a condition code used by the instruction.
static MachineInstr * genFusedMultiplyAcc(MachineFunction &MF, MachineRegisterInfo &MRI, const TargetInstrInfo *TII, MachineInstr &Root, SmallVectorImpl< MachineInstr * > &InsInstrs, unsigned IdxMulOpd, unsigned MaddOpc, const TargetRegisterClass *RC)
genFusedMultiplyAcc - Helper to generate fused multiply accumulate instructions.
static MachineInstr * genFusedMultiplyAccNeg(MachineFunction &MF, MachineRegisterInfo &MRI, const TargetInstrInfo *TII, MachineInstr &Root, SmallVectorImpl< MachineInstr * > &InsInstrs, DenseMap< Register, unsigned > &InstrIdxForVirtReg, unsigned IdxMulOpd, unsigned MaddOpc, unsigned MnegOpc, const TargetRegisterClass *RC)
genFusedMultiplyAccNeg - Helper to generate fused multiply accumulate instructions with an additional...
static bool isCombineInstrCandidate64(unsigned Opc)
static bool isFrameLoadOpcode(int Opcode)
static unsigned removeCopies(const MachineRegisterInfo &MRI, unsigned VReg)
static bool areCFlagsAccessedBetweenInstrs(MachineBasicBlock::iterator From, MachineBasicBlock::iterator To, const TargetRegisterInfo *TRI, const AccessKind AccessToCheck=AK_All)
True when condition flags are accessed (either by writing or reading) on the instruction trace starti...
static bool getFMAPatterns(MachineInstr &Root, SmallVectorImpl< unsigned > &Patterns)
Floating-Point Support.
static bool isADDSRegImm(unsigned Opcode)
static bool isCheapCopy(const MachineInstr &MI, const AArch64RegisterInfo &RI)
static bool isANDOpcode(MachineInstr &MI)
static void appendOffsetComment(int NumBytes, llvm::raw_string_ostream &Comment, StringRef RegScale={})
static unsigned sForm(MachineInstr &Instr)
Get opcode of S version of Instr.
static bool isCombineInstrSettingFlag(unsigned Opc)
static bool getFNEGPatterns(MachineInstr &Root, SmallVectorImpl< unsigned > &Patterns)
static bool getIndVarInfo(Register Reg, const MachineBasicBlock *LoopBB, MachineInstr *&UpdateInst, unsigned &UpdateCounterOprNum, Register &InitReg, bool &IsUpdatePriorComp)
If Reg is an induction variable, return true and set some parameters.
static const MachineInstrBuilder & AddSubReg(const MachineInstrBuilder &MIB, MCRegister Reg, unsigned SubIdx, unsigned State, const TargetRegisterInfo *TRI)
static bool canPairLdStOpc(unsigned FirstOpc, unsigned SecondOpc)
static bool mustAvoidNeonAtMBBI(const AArch64Subtarget &Subtarget, MachineBasicBlock &MBB, MachineBasicBlock::iterator I)
Returns true if in a streaming call site region without SME-FA64.
static int findCondCodeUseOperandIdxForBranchOrSelect(const MachineInstr &Instr)
static bool isPostIndexLdStOpcode(unsigned Opcode)
Return true if the opcode is a post-index ld/st instruction, which really loads from base+0.
static unsigned getBranchDisplacementBits(unsigned Opc)
static cl::opt< unsigned > CBDisplacementBits("aarch64-cb-offset-bits", cl::Hidden, cl::init(9), cl::desc("Restrict range of CB instructions (DEBUG)"))
static std::optional< ParamLoadedValue > describeORRLoadedValue(const MachineInstr &MI, Register DescribedReg, const TargetInstrInfo *TII, const TargetRegisterInfo *TRI)
If the given ORR instruction is a copy, and DescribedReg overlaps with the destination register then,...
static bool getFMULPatterns(MachineInstr &Root, SmallVectorImpl< unsigned > &Patterns)
static void appendReadRegExpr(SmallVectorImpl< char > &Expr, unsigned RegNum)
static MachineInstr * genMaddR(MachineFunction &MF, MachineRegisterInfo &MRI, const TargetInstrInfo *TII, MachineInstr &Root, SmallVectorImpl< MachineInstr * > &InsInstrs, unsigned IdxMulOpd, unsigned MaddOpc, unsigned VR, const TargetRegisterClass *RC)
genMaddR - Generate madd instruction and combine mul and add using an extra virtual register Example ...
static Register cloneInstr(const MachineInstr *MI, unsigned ReplaceOprNum, Register ReplaceReg, MachineBasicBlock &MBB, MachineBasicBlock::iterator InsertTo)
Clone an instruction from MI.
static bool scaleOffset(unsigned Opc, int64_t &Offset)
static bool canCombineWithFMUL(MachineBasicBlock &MBB, MachineOperand &MO, unsigned MulOpc)
unsigned scaledOffsetOpcode(unsigned Opcode, unsigned &Scale)
static MachineInstr * genFusedMultiplyIdx(MachineFunction &MF, MachineRegisterInfo &MRI, const TargetInstrInfo *TII, MachineInstr &Root, SmallVectorImpl< MachineInstr * > &InsInstrs, unsigned IdxMulOpd, unsigned MaddOpc, const TargetRegisterClass *RC)
genFusedMultiplyIdx - Helper to generate fused multiply accumulate instructions.
static MachineInstr * genIndexedMultiply(MachineInstr &Root, SmallVectorImpl< MachineInstr * > &InsInstrs, unsigned IdxDupOp, unsigned MulOpc, const TargetRegisterClass *RC, MachineRegisterInfo &MRI)
Fold (FMUL x (DUP y lane)) into (FMUL_indexed x y lane)
static bool isSUBSRegImm(unsigned Opcode)
static bool UpdateOperandRegClass(MachineInstr &Instr)
static const TargetRegisterClass * getRegClass(const MachineInstr &MI, Register Reg)
static bool isInStreamingCallSiteRegion(MachineBasicBlock &MBB, MachineBasicBlock::iterator I)
Returns true if the instruction at I is in a streaming call site region, within a single basic block.
static bool canCmpInstrBeRemoved(MachineInstr &MI, MachineInstr &CmpInstr, int CmpValue, const TargetRegisterInfo &TRI, SmallVectorImpl< MachineInstr * > &CCUseInstrs, bool &IsInvertCC)
unsigned unscaledOffsetOpcode(unsigned Opcode)
static bool getLoadPatterns(MachineInstr &Root, SmallVectorImpl< unsigned > &Patterns)
Search for patterns of LD instructions we can optimize.
static bool canInstrSubstituteCmpInstr(MachineInstr &MI, MachineInstr &CmpInstr, const TargetRegisterInfo &TRI)
Check if CmpInstr can be substituted by MI.
static UsedNZCV getUsedNZCV(AArch64CC::CondCode CC)
static bool isCombineInstrCandidateFP(const MachineInstr &Inst)
static void appendLoadRegExpr(SmallVectorImpl< char > &Expr, int64_t OffsetFromDefCFA)
static void appendConstantExpr(SmallVectorImpl< char > &Expr, int64_t Constant, dwarf::LocationAtom Operation)
static unsigned convertToNonFlagSettingOpc(const MachineInstr &MI)
Return the opcode that does not set flags when possible - otherwise return the original opcode.
static bool outliningCandidatesV8_3OpsConsensus(const outliner::Candidate &a, const outliner::Candidate &b)
static bool isCombineInstrCandidate32(unsigned Opc)
static void parseCondBranch(MachineInstr *LastInst, MachineBasicBlock *&Target, SmallVectorImpl< MachineOperand > &Cond)
static unsigned offsetExtendOpcode(unsigned Opcode)
static void loadRegPairFromStackSlot(const TargetRegisterInfo &TRI, MachineBasicBlock &MBB, MachineBasicBlock::iterator InsertBefore, const MCInstrDesc &MCID, Register DestReg, unsigned SubIdx0, unsigned SubIdx1, int FI, MachineMemOperand *MMO)
static void generateGatherLanePattern(MachineInstr &Root, SmallVectorImpl< MachineInstr * > &InsInstrs, SmallVectorImpl< MachineInstr * > &DelInstrs, DenseMap< Register, unsigned > &InstrIdxForVirtReg, unsigned Pattern, unsigned NumLanes)
Generate optimized instruction sequence for gather load patterns to improve Memory-Level Parallelism ...
static bool getMiscPatterns(MachineInstr &Root, SmallVectorImpl< unsigned > &Patterns)
Find other MI combine patterns.
static bool outliningCandidatesSigningKeyConsensus(const outliner::Candidate &a, const outliner::Candidate &b)
static bool outliningCandidatesSigningScopeConsensus(const outliner::Candidate &a, const outliner::Candidate &b)
static bool shouldClusterFI(const MachineFrameInfo &MFI, int FI1, int64_t Offset1, unsigned Opcode1, int FI2, int64_t Offset2, unsigned Opcode2)
static cl::opt< unsigned > TBZDisplacementBits("aarch64-tbz-offset-bits", cl::Hidden, cl::init(14), cl::desc("Restrict range of TB[N]Z instructions (DEBUG)"))
static void extractPhiReg(const MachineInstr &Phi, const MachineBasicBlock *MBB, Register &RegMBB, Register &RegOther)
static MCCFIInstruction createDefCFAExpression(const TargetRegisterInfo &TRI, unsigned Reg, const StackOffset &Offset)
static bool isDefinedOutside(Register Reg, const MachineBasicBlock *BB)
static MachineInstr * genFusedMultiply(MachineFunction &MF, MachineRegisterInfo &MRI, const TargetInstrInfo *TII, MachineInstr &Root, SmallVectorImpl< MachineInstr * > &InsInstrs, unsigned IdxMulOpd, unsigned MaddOpc, const TargetRegisterClass *RC, FMAInstKind kind=FMAInstKind::Default, const Register *ReplacedAddend=nullptr)
genFusedMultiply - Generate fused multiply instructions.
static bool getGatherLanePattern(MachineInstr &Root, SmallVectorImpl< unsigned > &Patterns, unsigned LoadLaneOpCode, unsigned NumLanes)
Check if the given instruction forms a gather load pattern that can be optimized for better Memory-Le...
static MachineInstr * genFusedMultiplyIdxNeg(MachineFunction &MF, MachineRegisterInfo &MRI, const TargetInstrInfo *TII, MachineInstr &Root, SmallVectorImpl< MachineInstr * > &InsInstrs, DenseMap< Register, unsigned > &InstrIdxForVirtReg, unsigned IdxMulOpd, unsigned MaddOpc, unsigned MnegOpc, const TargetRegisterClass *RC)
genFusedMultiplyAccNeg - Helper to generate fused multiply accumulate instructions with an additional...
static bool isCombineInstrCandidate(unsigned Opc)
static unsigned regOffsetOpcode(unsigned Opcode)
MachineOutlinerClass
Constants defining how certain sequences should be outlined.
@ MachineOutlinerTailCall
Emit a save, restore, call, and return.
@ MachineOutlinerRegSave
Emit a call and tail-call.
@ MachineOutlinerNoLRSave
Only emit a branch.
@ MachineOutlinerThunk
Emit a call and return.
static cl::opt< unsigned > BDisplacementBits("aarch64-b-offset-bits", cl::Hidden, cl::init(26), cl::desc("Restrict range of B instructions (DEBUG)"))
static bool areCFlagsAliveInSuccessors(const MachineBasicBlock *MBB)
Check if AArch64::NZCV should be alive in successors of MBB.
static void emitFrameOffsetAdj(MachineBasicBlock &MBB, MachineBasicBlock::iterator MBBI, const DebugLoc &DL, unsigned DestReg, unsigned SrcReg, int64_t Offset, unsigned Opc, const TargetInstrInfo *TII, MachineInstr::MIFlag Flag, bool NeedsWinCFI, bool *HasWinCFI, bool EmitCFAOffset, StackOffset CFAOffset, unsigned FrameReg)
static bool isCheapImmediate(const MachineInstr &MI, unsigned BitSize)
static cl::opt< unsigned > CBZDisplacementBits("aarch64-cbz-offset-bits", cl::Hidden, cl::init(19), cl::desc("Restrict range of CB[N]Z instructions (DEBUG)"))
static void genSubAdd2SubSub(MachineFunction &MF, MachineRegisterInfo &MRI, const TargetInstrInfo *TII, MachineInstr &Root, SmallVectorImpl< MachineInstr * > &InsInstrs, SmallVectorImpl< MachineInstr * > &DelInstrs, unsigned IdxOpd1, DenseMap< Register, unsigned > &InstrIdxForVirtReg)
Do the following transformation A - (B + C) ==> (A - B) - C A - (B + C) ==> (A - C) - B.
static unsigned canFoldIntoCSel(const MachineRegisterInfo &MRI, unsigned VReg, unsigned *NewReg=nullptr)
static void signOutlinedFunction(MachineFunction &MF, MachineBasicBlock &MBB, const AArch64InstrInfo *TII, bool ShouldSignReturnAddr)
static MachineInstr * genFNegatedMAD(MachineFunction &MF, MachineRegisterInfo &MRI, const TargetInstrInfo *TII, MachineInstr &Root, SmallVectorImpl< MachineInstr * > &InsInstrs)
static bool canCombineWithMUL(MachineBasicBlock &MBB, MachineOperand &MO, unsigned MulOpc, unsigned ZeroReg)
static void storeRegPairToStackSlot(const TargetRegisterInfo &TRI, MachineBasicBlock &MBB, MachineBasicBlock::iterator InsertBefore, const MCInstrDesc &MCID, Register SrcReg, bool IsKill, unsigned SubIdx0, unsigned SubIdx1, int FI, MachineMemOperand *MMO)
assert(UImm &&(UImm !=~static_cast< T >(0)) &&"Invalid immediate!")
MachineBasicBlock MachineBasicBlock::iterator DebugLoc DL
MachineBasicBlock MachineBasicBlock::iterator MBBI
static const Function * getParent(const Value *V)
static GCRegistry::Add< ErlangGC > A("erlang", "erlang-compatible garbage collector")
static GCRegistry::Add< OcamlGC > B("ocaml", "ocaml 3.10-compatible GC")
DXIL Forward Handle Accesses
const HexagonInstrInfo * TII
Module.h This file contains the declarations for the Module class.
const AbstractManglingParser< Derived, Alloc >::OperatorInfo AbstractManglingParser< Derived, Alloc >::Ops[]
This file implements the LivePhysRegs utility for tracking liveness of physical registers.
Machine Check Debug Module
Register const TargetRegisterInfo * TRI
Promote Memory to Register
static MCRegister getReg(const MCDisassembler *D, unsigned RC, unsigned RegNo)
MachineInstr unsigned OpIdx
ConstantRange Range(APInt(BitWidth, Low), APInt(BitWidth, High))
PowerPC Reduce CR logical Operation
const SmallVectorImpl< MachineOperand > MachineBasicBlock * TBB
const SmallVectorImpl< MachineOperand > & Cond
This file declares the machine register scavenger class.
static bool contains(SmallPtrSetImpl< ConstantExpr * > &Cache, ConstantExpr *Expr, Constant *C)
This file defines the SmallSet class.
This file defines the SmallVector class.
This file defines the 'Statistic' class, which is designed to be an easy way to expose various metric...
#define STATISTIC(VARNAME, DESC)
#define DEBUG_WITH_TYPE(TYPE,...)
DEBUG_WITH_TYPE macro - This macro should be used by passes to emit debug information.
static bool canCombine(MachineBasicBlock &MBB, MachineOperand &MO, unsigned CombineOpc=0)
AArch64FunctionInfo - This class is derived from MachineFunctionInfo and contains private AArch64-spe...
SignReturnAddress getSignReturnAddressCondition() const
bool hasStreamingModeChanges() const
void setOutliningStyle(const std::string &Style)
std::optional< bool > hasRedZone() const
static bool shouldSignReturnAddress(SignReturnAddress Condition, bool IsLRSpilled)
bool shouldSignWithBKey() const
static bool isHForm(const MachineInstr &MI)
Returns whether the instruction is in H form (16 bit operands)
void insertSelect(MachineBasicBlock &MBB, MachineBasicBlock::iterator MI, const DebugLoc &DL, Register DstReg, ArrayRef< MachineOperand > Cond, Register TrueReg, Register FalseReg) const override
static bool hasBTISemantics(const MachineInstr &MI)
Returns whether the instruction can be compatible with non-zero BTYPE.
static bool isQForm(const MachineInstr &MI)
Returns whether the instruction is in Q form (128 bit operands)
static bool getMemOpInfo(unsigned Opcode, TypeSize &Scale, TypeSize &Width, int64_t &MinOffset, int64_t &MaxOffset)
Returns true if opcode Opc is a memory operation.
static bool isTailCallReturnInst(const MachineInstr &MI)
Returns true if MI is one of the TCRETURN* instructions.
static bool isFPRCopy(const MachineInstr &MI)
Does this instruction rename an FPR without modifying bits?
MachineInstr * emitLdStWithAddr(MachineInstr &MemI, const ExtAddrMode &AM) const override
std::optional< DestSourcePair > isCopyInstrImpl(const MachineInstr &MI) const override
If the specific machine instruction is an instruction that moves/copies value from one register to an...
MachineBasicBlock * getBranchDestBlock(const MachineInstr &MI) const override
unsigned getInstSizeInBytes(const MachineInstr &MI) const override
GetInstSize - Return the number of bytes of code the specified instruction may be.
bool areMemAccessesTriviallyDisjoint(const MachineInstr &MIa, const MachineInstr &MIb) const override
void copyPhysReg(MachineBasicBlock &MBB, MachineBasicBlock::iterator I, const DebugLoc &DL, Register DestReg, Register SrcReg, bool KillSrc, bool RenamableDest=false, bool RenamableSrc=false) const override
static bool isGPRCopy(const MachineInstr &MI)
Does this instruction rename a GPR without modifying bits?
static unsigned convertToFlagSettingOpc(unsigned Opc)
Return the opcode that set flags when possible.
bool isBranchOffsetInRange(unsigned BranchOpc, int64_t BrOffset) const override
bool canInsertSelect(const MachineBasicBlock &, ArrayRef< MachineOperand > Cond, Register, Register, Register, int &, int &, int &) const override
Register isLoadFromStackSlotPostFE(const MachineInstr &MI, int &FrameIndex) const override
Check for post-frame ptr elimination stack locations as well.
static const MachineOperand & getLdStOffsetOp(const MachineInstr &MI)
Returns the immediate offset operator of a load/store.
bool isCoalescableExtInstr(const MachineInstr &MI, Register &SrcReg, Register &DstReg, unsigned &SubIdx) const override
static std::optional< unsigned > getUnscaledLdSt(unsigned Opc)
Returns the unscaled load/store for the scaled load/store opcode, if there is a corresponding unscale...
static bool hasUnscaledLdStOffset(unsigned Opc)
Return true if it has an unscaled load/store offset.
static const MachineOperand & getLdStAmountOp(const MachineInstr &MI)
Returns the shift amount operator of a load/store.
static bool isPreLdSt(const MachineInstr &MI)
Returns whether the instruction is a pre-indexed load/store.
std::optional< ExtAddrMode > getAddrModeFromMemoryOp(const MachineInstr &MemI, const TargetRegisterInfo *TRI) const override
bool getMemOperandsWithOffsetWidth(const MachineInstr &MI, SmallVectorImpl< const MachineOperand * > &BaseOps, int64_t &Offset, bool &OffsetIsScalable, LocationSize &Width, const TargetRegisterInfo *TRI) const override
bool analyzeBranchPredicate(MachineBasicBlock &MBB, MachineBranchPredicate &MBP, bool AllowModify) const override
void insertIndirectBranch(MachineBasicBlock &MBB, MachineBasicBlock &NewDestBB, MachineBasicBlock &RestoreBB, const DebugLoc &DL, int64_t BrOffset, RegScavenger *RS) const override
void loadRegFromStackSlot(MachineBasicBlock &MBB, MachineBasicBlock::iterator MBBI, Register DestReg, int FrameIndex, const TargetRegisterClass *RC, Register VReg, MachineInstr::MIFlag Flags=MachineInstr::NoFlags) const override
static bool isPairableLdStInst(const MachineInstr &MI)
Return true if pairing the given load or store may be paired with another.
void storeRegToStackSlot(MachineBasicBlock &MBB, MachineBasicBlock::iterator MBBI, Register SrcReg, bool isKill, int FrameIndex, const TargetRegisterClass *RC, Register VReg, MachineInstr::MIFlag Flags=MachineInstr::NoFlags) const override
const AArch64RegisterInfo & getRegisterInfo() const
getRegisterInfo - TargetInstrInfo is a superset of MRegister info.
static bool isPreSt(const MachineInstr &MI)
Returns whether the instruction is a pre-indexed store.
MachineInstr * foldMemoryOperandImpl(MachineFunction &MF, MachineInstr &MI, ArrayRef< unsigned > Ops, MachineBasicBlock::iterator InsertPt, int FrameIndex, LiveIntervals *LIS=nullptr, VirtRegMap *VRM=nullptr) const override
void insertNoop(MachineBasicBlock &MBB, MachineBasicBlock::iterator MI) const override
AArch64InstrInfo(const AArch64Subtarget &STI)
static bool isPairedLdSt(const MachineInstr &MI)
Returns whether the instruction is a paired load/store.
bool getMemOperandWithOffsetWidth(const MachineInstr &MI, const MachineOperand *&BaseOp, int64_t &Offset, bool &OffsetIsScalable, TypeSize &Width, const TargetRegisterInfo *TRI) const
If OffsetIsScalable is set to 'true', the offset is scaled by vscale.
Register isLoadFromStackSlot(const MachineInstr &MI, int &FrameIndex) const override
bool reverseBranchCondition(SmallVectorImpl< MachineOperand > &Cond) const override
static bool isStridedAccess(const MachineInstr &MI)
Return true if the given load or store is a strided memory access.
bool shouldClusterMemOps(ArrayRef< const MachineOperand * > BaseOps1, int64_t Offset1, bool OffsetIsScalable1, ArrayRef< const MachineOperand * > BaseOps2, int64_t Offset2, bool OffsetIsScalable2, unsigned ClusterSize, unsigned NumBytes) const override
Detect opportunities for ldp/stp formation.
unsigned removeBranch(MachineBasicBlock &MBB, int *BytesRemoved=nullptr) const override
bool isThroughputPattern(unsigned Pattern) const override
Return true when a code sequence can improve throughput.
MachineOperand & getMemOpBaseRegImmOfsOffsetOperand(MachineInstr &LdSt) const
Return the immediate offset of the base register in a load/store LdSt.
bool analyzeBranch(MachineBasicBlock &MBB, MachineBasicBlock *&TBB, MachineBasicBlock *&FBB, SmallVectorImpl< MachineOperand > &Cond, bool AllowModify=false) const override
bool canFoldIntoAddrMode(const MachineInstr &MemI, Register Reg, const MachineInstr &AddrI, ExtAddrMode &AM) const override
static bool isLdStPairSuppressed(const MachineInstr &MI)
Return true if pairing the given load or store is hinted to be unprofitable.
Register isStoreToStackSlotPostFE(const MachineInstr &MI, int &FrameIndex) const override
Check for post-frame ptr elimination stack locations as well.
std::unique_ptr< TargetInstrInfo::PipelinerLoopInfo > analyzeLoopForPipelining(MachineBasicBlock *LoopBB) const override
bool isSchedulingBoundary(const MachineInstr &MI, const MachineBasicBlock *MBB, const MachineFunction &MF) const override
unsigned insertBranch(MachineBasicBlock &MBB, MachineBasicBlock *TBB, MachineBasicBlock *FBB, ArrayRef< MachineOperand > Cond, const DebugLoc &DL, int *BytesAdded=nullptr) const override
MachineBasicBlock::iterator probedStackAlloc(MachineBasicBlock::iterator MBBI, Register TargetReg, bool FrameSetup) const
Return true when there is potentially a faster code sequence for an instruction chain ending in Root.
bool optimizeCompareInstr(MachineInstr &CmpInstr, Register SrcReg, Register SrcReg2, int64_t CmpMask, int64_t CmpValue, const MachineRegisterInfo *MRI) const override
optimizeCompareInstr - Convert the instruction supplying the argument to the comparison into one that...
static unsigned getLoadStoreImmIdx(unsigned Opc)
Returns the index for the immediate for a given instruction.
static bool isGPRZero(const MachineInstr &MI)
Does this instruction set its full destination register to zero?
void copyGPRRegTuple(MachineBasicBlock &MBB, MachineBasicBlock::iterator I, const DebugLoc &DL, MCRegister DestReg, MCRegister SrcReg, bool KillSrc, unsigned Opcode, unsigned ZeroReg, llvm::ArrayRef< unsigned > Indices) const
bool analyzeCompare(const MachineInstr &MI, Register &SrcReg, Register &SrcReg2, int64_t &CmpMask, int64_t &CmpValue) const override
analyzeCompare - For a comparison instruction, return the source registers in SrcReg and SrcReg2,...
CombinerObjective getCombinerObjective(unsigned Pattern) const override
static bool isFpOrNEON(Register Reg)
Returns whether the physical register is FP or NEON.
bool isAsCheapAsAMove(const MachineInstr &MI) const override
std::optional< DestSourcePair > isCopyLikeInstrImpl(const MachineInstr &MI) const override
static void suppressLdStPair(MachineInstr &MI)
Hint that pairing the given load or store is unprofitable.
Register isStoreToStackSlot(const MachineInstr &MI, int &FrameIndex) const override
static bool isPreLd(const MachineInstr &MI)
Returns whether the instruction is a pre-indexed load.
void copyPhysRegTuple(MachineBasicBlock &MBB, MachineBasicBlock::iterator I, const DebugLoc &DL, MCRegister DestReg, MCRegister SrcReg, bool KillSrc, unsigned Opcode, llvm::ArrayRef< unsigned > Indices) const
bool optimizeCondBranch(MachineInstr &MI) const override
Replace csincr-branch sequence by simple conditional branch.
static int getMemScale(unsigned Opc)
Scaling factor for (scaled or unscaled) load or store.
bool isCandidateToMergeOrPair(const MachineInstr &MI) const
Return true if this is a load/store that can be potentially paired/merged.
MCInst getNop() const override
static const MachineOperand & getLdStBaseOp(const MachineInstr &MI)
Returns the base register operator of a load/store.
bool isReservedReg(const MachineFunction &MF, MCRegister Reg) const
const AArch64RegisterInfo * getRegisterInfo() const override
bool isNeonAvailable() const
Returns true if the target has NEON and the function at runtime is known to have NEON enabled (e....
bool isSVEorStreamingSVEAvailable() const
Returns true if the target has access to either the full range of SVE instructions,...
ArrayRef - Represent a constant reference to an array (0 or more elements consecutively in memory),...
const T & front() const
front - Get the first element.
size_t size() const
size - Get the array size.
This is an important base class in LLVM.
std::pair< iterator, bool > insert(const std::pair< KeyT, ValueT > &KV)
bool hasOptSize() const
Optimize this function for size (-Os) or minimum size (-Oz).
bool hasMinSize() const
Optimize this function for minimum size (-Oz).
Module * getParent()
Get the module that this global value is contained inside of...
LLVM_ABI InstListType::iterator eraseFromParent()
This method unlinks 'this' from the containing basic block and deletes it.
unsigned getOpcode() const
Returns a member of one of the enums like Instruction::Add.
static LocationSize precise(uint64_t Value)
This class is intended to be used as a base class for asm properties and features specific to the tar...
bool usesWindowsCFI() const
static MCCFIInstruction cfiDefCfa(MCSymbol *L, unsigned Register, int64_t Offset, SMLoc Loc={})
.cfi_def_cfa defines a rule for computing CFA as: take address from Register and add Offset to it.
static MCCFIInstruction createOffset(MCSymbol *L, unsigned Register, int64_t Offset, SMLoc Loc={})
.cfi_offset Previous value of Register is saved at offset Offset from CFA.
static MCCFIInstruction cfiDefCfaOffset(MCSymbol *L, int64_t Offset, SMLoc Loc={})
.cfi_def_cfa_offset modifies a rule for computing CFA.
static MCCFIInstruction createEscape(MCSymbol *L, StringRef Vals, SMLoc Loc={}, StringRef Comment="")
.cfi_escape Allows the user to add arbitrary bytes to the unwind info.
Instances of this class represent a single low-level machine instruction.
Describe properties that are true of each instruction in the target description file.
unsigned getNumOperands() const
Return the number of declared MachineOperands for this MachineInstruction.
ArrayRef< MCOperandInfo > operands() const
Wrapper class representing physical registers. Should be passed by value.
constexpr bool isValid() const
static constexpr unsigned NoRegister
static MDTuple * get(LLVMContext &Context, ArrayRef< Metadata * > MDs)
bool isInlineAsmBrIndirectTarget() const
Returns true if this is the indirect dest of an INLINEASM_BR.
unsigned pred_size() const
LLVM_ABI void transferSuccessorsAndUpdatePHIs(MachineBasicBlock *FromMBB)
Transfers all the successors, as in transferSuccessors, and update PHI operands in the successor bloc...
LLVM_ABI instr_iterator insert(instr_iterator I, MachineInstr *M)
Insert MI into the instruction list before I, possibly inside a bundle.
reverse_instr_iterator instr_rbegin()
LLVM_ABI iterator getFirstTerminator()
Returns an iterator to the first terminator instruction of this basic block.
LLVM_ABI void addSuccessor(MachineBasicBlock *Succ, BranchProbability Prob=BranchProbability::getUnknown())
Add Succ as a successor of this MachineBasicBlock.
reverse_instr_iterator instr_rend()
Instructions::iterator instr_iterator
instr_iterator instr_end()
Instructions::const_iterator const_instr_iterator
void addLiveIn(MCRegister PhysReg, LaneBitmask LaneMask=LaneBitmask::getAll())
Adds the specified register as a live in.
const MachineFunction * getParent() const
Return the MachineFunction containing this basic block.
LLVM_ABI instr_iterator erase(instr_iterator I)
Remove an instruction from the instruction list and delete it.
void splice(iterator Where, MachineBasicBlock *Other, iterator From)
Take an instruction from MBB 'Other' at the position From, and insert it into this MBB right before '...
MachineInstrBundleIterator< MachineInstr > iterator
void setMachineBlockAddressTaken()
Set this block to indicate that its address is used as something other than the target of a terminato...
LLVM_ABI bool isLiveIn(MCRegister Reg, LaneBitmask LaneMask=LaneBitmask::getAll()) const
Return true if the specified register is in the live in set.
The MachineFrameInfo class represents an abstract stack frame until prolog/epilog code is inserted.
uint64_t getStackSize() const
Return the number of bytes that must be allocated to hold all of the fixed size frame objects.
void setStackID(int ObjectIdx, uint8_t ID)
bool isCalleeSavedInfoValid() const
Has the callee saved info been calculated yet?
Align getObjectAlign(int ObjectIdx) const
Return the alignment of the specified stack object.
int64_t getObjectSize(int ObjectIdx) const
Return the size of the specified object.
unsigned getNumObjects() const
Return the number of objects.
int64_t getObjectOffset(int ObjectIdx) const
Return the assigned stack offset of the specified object from the incoming stack pointer.
bool isFixedObjectIndex(int ObjectIdx) const
Returns true if the specified index corresponds to a fixed stack object.
unsigned addFrameInst(const MCCFIInstruction &Inst)
const TargetSubtargetInfo & getSubtarget() const
getSubtarget - Return the subtarget for which this machine code is being compiled.
StringRef getName() const
getName - Return the name of the corresponding LLVM function.
MachineMemOperand * getMachineMemOperand(MachinePointerInfo PtrInfo, MachineMemOperand::Flags f, LLT MemTy, Align base_alignment, const AAMDNodes &AAInfo=AAMDNodes(), const MDNode *Ranges=nullptr, SyncScope::ID SSID=SyncScope::System, AtomicOrdering Ordering=AtomicOrdering::NotAtomic, AtomicOrdering FailureOrdering=AtomicOrdering::NotAtomic)
getMachineMemOperand - Allocate a new MachineMemOperand.
MachineFrameInfo & getFrameInfo()
getFrameInfo - Return the frame info object for the current function.
MachineRegisterInfo & getRegInfo()
getRegInfo - Return information about the registers currently in use.
Function & getFunction()
Return the LLVM function that this machine code represents.
BasicBlockListType::iterator iterator
Ty * getInfo()
getInfo - Keep track of various per-function pieces of information for backends that would like to do...
const MachineJumpTableInfo * getJumpTableInfo() const
getJumpTableInfo - Return the jump table info object for the current function.
MachineBasicBlock * CreateMachineBasicBlock(const BasicBlock *BB=nullptr, std::optional< UniqueBBID > BBID=std::nullopt)
CreateMachineInstr - Allocate a new MachineInstr.
void insert(iterator MBBI, MachineBasicBlock *MBB)
const TargetMachine & getTarget() const
getTarget - Return the target machine this machine code is compiled with
const MachineInstrBuilder & setMemRefs(ArrayRef< MachineMemOperand * > MMOs) const
const MachineInstrBuilder & addCFIIndex(unsigned CFIIndex) const
const MachineInstrBuilder & setMIFlag(MachineInstr::MIFlag Flag) const
const MachineInstrBuilder & addImm(int64_t Val) const
Add a new immediate operand.
const MachineInstrBuilder & add(const MachineOperand &MO) const
const MachineInstrBuilder & addSym(MCSymbol *Sym, unsigned char TargetFlags=0) const
const MachineInstrBuilder & addFrameIndex(int Idx) const
const MachineInstrBuilder & addGlobalAddress(const GlobalValue *GV, int64_t Offset=0, unsigned TargetFlags=0) const
const MachineInstrBuilder & addReg(Register RegNo, unsigned flags=0, unsigned SubReg=0) const
Add a new virtual register operand.
const MachineInstrBuilder & addMBB(MachineBasicBlock *MBB, unsigned TargetFlags=0) const
const MachineInstrBuilder & addUse(Register RegNo, unsigned Flags=0, unsigned SubReg=0) const
Add a virtual register use operand.
const MachineInstrBuilder & setMIFlags(unsigned Flags) const
const MachineInstrBuilder & addMemOperand(MachineMemOperand *MMO) const
const MachineInstrBuilder & addDef(Register RegNo, unsigned Flags=0, unsigned SubReg=0) const
Add a virtual register definition operand.
reverse_iterator getReverse() const
Get a reverse iterator to the same node.
Representation of each machine instruction.
unsigned getOpcode() const
Returns the opcode of this MachineInstr.
bool mayLoadOrStore(QueryType Type=AnyInBundle) const
Return true if this instruction could possibly read or modify memory.
const MachineBasicBlock * getParent() const
bool isCall(QueryType Type=AnyInBundle) const
bool getFlag(MIFlag Flag) const
Return whether an MI flag is set.
LLVM_ABI uint32_t mergeFlagsWith(const MachineInstr &Other) const
Return the MIFlags which represent both MachineInstrs.
unsigned getNumOperands() const
Retuns the total number of operands.
LLVM_ABI unsigned getNumExplicitOperands() const
Returns the number of non-implicit operands.
bool mayLoad(QueryType Type=AnyInBundle) const
Return true if this instruction could possibly read memory.
const MCInstrDesc & getDesc() const
Returns the target instruction descriptor of this MachineInstr.
LLVM_ABI bool hasUnmodeledSideEffects() const
Return true if this instruction has side effects that are not modeled by mayLoad / mayStore,...
bool registerDefIsDead(Register Reg, const TargetRegisterInfo *TRI) const
Returns true if the register is dead in this machine instruction.
bool definesRegister(Register Reg, const TargetRegisterInfo *TRI) const
Return true if the MachineInstr fully defines the specified register.
LLVM_ABI void setDesc(const MCInstrDesc &TID)
Replace the instruction descriptor (thus opcode) of the current instruction with a new one.
LLVM_ABI bool hasOrderedMemoryRef() const
Return true if this instruction may have an ordered or volatile memory reference, or if the informati...
LLVM_ABI const MachineFunction * getMF() const
Return the function that contains the basic block that this instruction belongs to.
ArrayRef< MachineMemOperand * > memoperands() const
Access to memory operands of the instruction.
LLVM_ABI bool isLoadFoldBarrier() const
Returns true if it is illegal to fold a load across this instruction.
const DebugLoc & getDebugLoc() const
Returns the debug location id of this MachineInstr.
LLVM_ABI void eraseFromParent()
Unlink 'this' from the containing basic block and delete it.
LLVM_ABI void removeOperand(unsigned OpNo)
Erase an operand from an instruction, leaving it with one fewer operand than it started with.
LLVM_ABI void addRegisterDefined(Register Reg, const TargetRegisterInfo *RegInfo=nullptr)
We have determined MI defines a register.
const MachineOperand & getOperand(unsigned i) const
uint32_t getFlags() const
Return the MI flags bitvector.
LLVM_ABI int findRegisterDefOperandIdx(Register Reg, const TargetRegisterInfo *TRI, bool isDead=false, bool Overlap=false) const
Returns the operand index that is a def of the specified register or -1 if it is not found.
const std::vector< MachineJumpTableEntry > & getJumpTables() const
A description of a memory reference used in the backend.
@ MOLoad
The memory access reads data.
@ MOStore
The memory access writes data.
This class contains meta information specific to a module.
LLVM_ABI MachineFunction * getMachineFunction(const Function &F) const
Returns the MachineFunction associated to IR function F if there is one, otherwise nullptr.
MachineOperand class - Representation of each machine instruction operand.
void setSubReg(unsigned subReg)
unsigned getSubReg() const
void setImm(int64_t immVal)
bool isReg() const
isReg - Tests if this is a MO_Register operand.
MachineBasicBlock * getMBB() const
void setIsDead(bool Val=true)
LLVM_ABI void setReg(Register Reg)
Change the register this operand corresponds to.
bool isImm() const
isImm - Tests if this is a MO_Immediate operand.
void setIsKill(bool Val=true)
MachineInstr * getParent()
getParent - Return the instruction that this operand belongs to.
unsigned getTargetFlags() const
static MachineOperand CreateImm(int64_t Val)
MachineOperandType getType() const
getType - Returns the MachineOperandType for this operand.
void setIsUndef(bool Val=true)
Register getReg() const
getReg - Returns the register number.
bool isFI() const
isFI - Tests if this is a MO_FrameIndex operand.
LLVM_ABI bool isIdenticalTo(const MachineOperand &Other) const
Returns true if this operand is identical to the specified operand except for liveness related flags ...
static MachineOperand CreateReg(Register Reg, bool isDef, bool isImp=false, bool isKill=false, bool isDead=false, bool isUndef=false, bool isEarlyClobber=false, unsigned SubReg=0, bool isDebug=false, bool isInternalRead=false, bool isRenamable=false)
MachineRegisterInfo - Keep track of information for virtual and physical registers,...
bool tracksLiveness() const
tracksLiveness - Returns true when tracking register liveness accurately.
bool reservedRegsFrozen() const
reservedRegsFrozen - Returns true after freezeReservedRegs() was called to ensure the set of reserved...
const TargetRegisterClass * getRegClassOrNull(Register Reg) const
Return the register class of Reg, or null if Reg has not been assigned a register class yet.
LLVM_ABI const TargetRegisterClass * constrainRegClass(Register Reg, const TargetRegisterClass *RC, unsigned MinNumRegs=0)
constrainRegClass - Constrain the register class of the specified virtual register to be a common sub...
LLVM_ABI MachineInstr * getUniqueVRegDef(Register Reg) const
getUniqueVRegDef - Return the unique machine instr that defines the specified virtual register or nul...
A Module instance is used to store all the information related to an LLVM module.
MI-level patchpoint operands.
uint32_t getNumPatchBytes() const
Return the number of patchable bytes the given patchpoint should emit.
Wrapper class representing virtual and physical registers.
MCRegister asMCReg() const
Utility to check-convert this value to a MCRegister.
constexpr bool isValid() const
constexpr bool isVirtual() const
Return true if the specified register number is in the virtual register namespace.
static constexpr bool isVirtualRegister(unsigned Reg)
Return true if the specified register number is in the virtual register namespace.
constexpr bool isPhysical() const
Return true if the specified register number is in the physical register namespace.
Represents a location in source code.
bool erase(PtrType Ptr)
Remove pointer from the set.
std::pair< iterator, bool > insert(PtrType Ptr)
Inserts Ptr if and only if there is no element in the container equal to Ptr.
SmallPtrSet - This class implements a set which is optimized for holding SmallSize or less elements.
SmallSet - This maintains a set of unique values, optimizing for the case when the set is small (less...
SmallString - A SmallString is just a SmallVector with methods and accessors that make it work better...
void append(StringRef RHS)
Append from a StringRef.
StringRef str() const
Explicit conversion to StringRef.
This class consists of common code factored out of the SmallVector class to reduce code duplication b...
void push_back(const T &Elt)
This is a 'vector' (really, a variable-sized array), optimized for the case when the array is small.
MI-level stackmap operands.
uint32_t getNumPatchBytes() const
Return the number of patchable bytes the given stackmap should emit.
StackOffset holds a fixed and a scalable offset in bytes.
int64_t getFixed() const
Returns the fixed component of the stack.
int64_t getScalable() const
Returns the scalable component of the stack.
static StackOffset get(int64_t Fixed, int64_t Scalable)
static StackOffset getScalable(int64_t Scalable)
static StackOffset getFixed(int64_t Fixed)
MI-level Statepoint operands.
uint32_t getNumPatchBytes() const
Return the number of patchable bytes the given statepoint should emit.
StringRef - Represent a constant reference to a string, i.e.
Object returned by analyzeLoopForPipelining.
TargetInstrInfo - Interface to description of machine instruction set.
virtual void genAlternativeCodeSequence(MachineInstr &Root, unsigned Pattern, SmallVectorImpl< MachineInstr * > &InsInstrs, SmallVectorImpl< MachineInstr * > &DelInstrs, DenseMap< Register, unsigned > &InstIdxForVirtReg) const
When getMachineCombinerPatterns() finds patterns, this function generates the instructions that could...
virtual std::optional< ParamLoadedValue > describeLoadedValue(const MachineInstr &MI, Register Reg) const
Produce the expression describing the MI loading a value into the physical register Reg.
virtual bool getMachineCombinerPatterns(MachineInstr &Root, SmallVectorImpl< unsigned > &Patterns, bool DoRegPressureReduce) const
Return true when there is potentially a faster code sequence for an instruction chain ending in Root.
virtual bool isSchedulingBoundary(const MachineInstr &MI, const MachineBasicBlock *MBB, const MachineFunction &MF) const
Test if the given instruction should be considered a scheduling boundary.
virtual CombinerObjective getCombinerObjective(unsigned Pattern) const
Return the objective of a combiner pattern.
virtual bool isFunctionSafeToSplit(const MachineFunction &MF) const
Return true if the function is a viable candidate for machine function splitting.
CodeModel::Model getCodeModel() const
Returns the code model.
const MCAsmInfo * getMCAsmInfo() const
Return target specific asm information.
bool contains(Register Reg) const
Return true if the specified register is included in this register class.
bool hasSubClassEq(const TargetRegisterClass *RC) const
Returns true if RC is a sub-class of or equal to this class.
bool hasSuperClassEq(const TargetRegisterClass *RC) const
Returns true if RC is a super-class of or equal to this class.
TargetRegisterInfo base class - We assume that the target defines a static array of TargetRegisterDes...
TargetSubtargetInfo - Generic base class for all target subtargets.
virtual const TargetInstrInfo * getInstrInfo() const
virtual const TargetRegisterInfo * getRegisterInfo() const =0
Return the target's register information.
Target - Wrapper for Target specific information.
static constexpr TypeSize getFixed(ScalarTy ExactSize)
static constexpr TypeSize getScalable(ScalarTy MinimumSize)
Value * getOperand(unsigned i) const
constexpr bool isScalable() const
Returns whether the quantity is scaled by a runtime quantity (vscale).
constexpr ScalarTy getKnownMinValue() const
Returns the minimum value this quantity can represent.
self_iterator getIterator()
A raw_ostream that writes to an std::string.
#define llvm_unreachable(msg)
Marks that the current location is not supposed to be reachable.
static CondCode getInvertedCondCode(CondCode Code)
@ MO_DLLIMPORT
MO_DLLIMPORT - On a symbol operand, this represents that the reference to the symbol is for an import...
@ MO_NC
MO_NC - Indicates whether the linker is expected to check the symbol reference for overflow.
@ MO_G1
MO_G1 - A symbol operand with this flag (granule 1) represents the bits 16-31 of a 64-bit address,...
@ MO_S
MO_S - Indicates that the bits of the symbol operand represented by MO_G0 etc are signed.
@ MO_PAGEOFF
MO_PAGEOFF - A symbol operand with this flag represents the offset of that symbol within a 4K page.
@ MO_GOT
MO_GOT - This flag indicates that a symbol operand represents the address of the GOT entry for the sy...
@ MO_PREL
MO_PREL - Indicates that the bits of the symbol operand represented by MO_G0 etc are PC relative.
@ MO_G0
MO_G0 - A symbol operand with this flag (granule 0) represents the bits 0-15 of a 64-bit address,...
@ MO_ARM64EC_CALLMANGLE
MO_ARM64EC_CALLMANGLE - Operand refers to the Arm64EC-mangled version of a symbol,...
@ MO_PAGE
MO_PAGE - A symbol operand with this flag represents the pc-relative offset of the 4K page containing...
@ MO_HI12
MO_HI12 - This flag indicates that a symbol operand represents the bits 13-24 of a 64-bit address,...
@ MO_TLS
MO_TLS - Indicates that the operand being accessed is some kind of thread-local symbol.
@ MO_G2
MO_G2 - A symbol operand with this flag (granule 2) represents the bits 32-47 of a 64-bit address,...
@ MO_TAGGED
MO_TAGGED - With MO_PAGE, indicates that the page includes a memory tag in bits 56-63.
@ MO_G3
MO_G3 - A symbol operand with this flag (granule 3) represents the high 16-bits of a 64-bit address,...
@ MO_COFFSTUB
MO_COFFSTUB - On a symbol operand "FOO", this indicates that the reference is actually to the "....
unsigned getCheckerSizeInBytes(AuthCheckMethod Method)
Returns the number of bytes added by checkAuthenticatedRegister.
static uint64_t decodeLogicalImmediate(uint64_t val, unsigned regSize)
decodeLogicalImmediate - Decode a logical immediate value in the form "N:immr:imms" (where the immr a...
static unsigned getShiftValue(unsigned Imm)
getShiftValue - Extract the shift value.
static unsigned getArithExtendImm(AArch64_AM::ShiftExtendType ET, unsigned Imm)
getArithExtendImm - Encode the extend type and shift amount for an arithmetic instruction: imm: 3-bit...
static unsigned getArithShiftValue(unsigned Imm)
getArithShiftValue - get the arithmetic shift value.
static uint64_t encodeLogicalImmediate(uint64_t imm, unsigned regSize)
encodeLogicalImmediate - Return the encoded immediate value for a logical immediate instruction of th...
static AArch64_AM::ShiftExtendType getExtendType(unsigned Imm)
getExtendType - Extract the extend type for operands of arithmetic ops.
static AArch64_AM::ShiftExtendType getArithExtendType(unsigned Imm)
static AArch64_AM::ShiftExtendType getShiftType(unsigned Imm)
getShiftType - Extract the shift type.
static unsigned getShifterImm(AArch64_AM::ShiftExtendType ST, unsigned Imm)
getShifterImm - Encode the shift type and amount: imm: 6-bit shift amount shifter: 000 ==> lsl 001 ==...
void expandMOVImm(uint64_t Imm, unsigned BitSize, SmallVectorImpl< ImmInsnModel > &Insn)
Expand a MOVi32imm or MOVi64imm pseudo instruction to one or more real move-immediate instructions to...
static const uint64_t InstrFlagIsWhile
static const uint64_t InstrFlagIsPTestLike
constexpr std::underlying_type_t< E > Mask()
Get a bitmask with 1s in all places up to the high-order bit of E's largest value.
@ C
The default llvm calling convention, compatible with C.
@ Implicit
Not emitted register (e.g. carry, or temporary result).
@ Renamable
Register that may be renamed.
@ Define
Register definition.
@ Kill
The last use of a register.
@ Undef
Value of the register doesn't matter.
@ ScalablePredicateVector
initializer< Ty > init(const Ty &Val)
InstrType
Represents how an instruction should be mapped by the outliner.
LLVM_ABI Instruction & back() const
This is an optimization pass for GlobalISel generic memory operations.
bool all_of(R &&range, UnaryPredicate P)
Provide wrappers to std::all_of which take ranges instead of having to pass begin/end explicitly.
static bool isCondBranchOpcode(int Opc)
MCCFIInstruction createDefCFA(const TargetRegisterInfo &TRI, unsigned FrameReg, unsigned Reg, const StackOffset &Offset, bool LastAdjustmentWasScalable=true)
static bool isPTrueOpcode(unsigned Opc)
MachineInstrBuilder BuildMI(MachineFunction &MF, const MIMetadata &MIMD, const MCInstrDesc &MCID)
Builder interface. Specify how to create the initial instruction itself.
bool succeeded(LogicalResult Result)
Utility function that returns true if the provided LogicalResult corresponds to a success value.
int isAArch64FrameOffsetLegal(const MachineInstr &MI, StackOffset &Offset, bool *OutUseUnscaledOp=nullptr, unsigned *OutUnscaledOp=nullptr, int64_t *EmittableOffset=nullptr)
Check if the Offset is a valid frame offset for MI.
constexpr bool isInt(int64_t x)
Checks if an integer fits into the given bit width.
auto enumerate(FirstRange &&First, RestRanges &&...Rest)
Given two or more input ranges, returns a new range whose values are tuples (A, B,...
decltype(auto) dyn_cast(const From &Val)
dyn_cast<X> - Return the argument parameter cast to the specified type.
static bool isIndirectBranchOpcode(int Opc)
iterator_range< T > make_range(T x, T y)
Convenience function for iterating over sub-ranges.
unsigned getBLRCallOpcode(const MachineFunction &MF)
Return opcode to be used for indirect calls.
@ AArch64FrameOffsetIsLegal
Offset is legal.
@ AArch64FrameOffsetCanUpdate
Offset can apply, at least partly.
@ AArch64FrameOffsetCannotUpdate
Offset cannot apply.
constexpr bool isPowerOf2_64(uint64_t Value)
Return true if the argument is a power of two > 0 (64 bit edition.)
unsigned Log2_64(uint64_t Value)
Return the floor log base 2 of the specified value, -1 if the value is zero.
static bool isSEHInstruction(const MachineInstr &MI)
bool any_of(R &&range, UnaryPredicate P)
Provide wrappers to std::any_of which take ranges instead of having to pass begin/end explicitly.
auto reverse(ContainerTy &&C)
MachineInstr * getImm(const MachineOperand &MO, const MachineRegisterInfo *MRI)
decltype(auto) get(const PointerIntPair< PointerTy, IntBits, IntType, PtrTraits, Info > &Pair)
void sort(IteratorTy Start, IteratorTy End)
AArch64MachineCombinerPattern
@ MULSUBv2i32_indexed_OP1
@ MULADDv4i16_indexed_OP2
@ MULSUBv8i16_indexed_OP2
@ MULSUBv4i16_indexed_OP2
@ MULSUBv4i32_indexed_OP2
@ MULADDv2i32_indexed_OP1
@ MULADDv4i32_indexed_OP1
@ MULADDv2i32_indexed_OP2
@ MULSUBv4i16_indexed_OP1
@ MULADDv4i32_indexed_OP2
@ MULSUBv8i16_indexed_OP1
@ MULSUBv2i32_indexed_OP2
@ MULADDv8i16_indexed_OP1
@ MULSUBv4i32_indexed_OP1
@ MULADDv8i16_indexed_OP2
@ MULADDv4i16_indexed_OP1
LLVM_ABI raw_ostream & dbgs()
dbgs() - This returns a reference to a raw_ostream for debugging messages.
void emitFrameOffset(MachineBasicBlock &MBB, MachineBasicBlock::iterator MBBI, const DebugLoc &DL, unsigned DestReg, unsigned SrcReg, StackOffset Offset, const TargetInstrInfo *TII, MachineInstr::MIFlag=MachineInstr::NoFlags, bool SetNZCV=false, bool NeedsWinCFI=false, bool *HasWinCFI=nullptr, bool EmitCFAOffset=false, StackOffset InitialOffset={}, unsigned FrameReg=AArch64::SP)
emitFrameOffset - Emit instructions as needed to set DestReg to SrcReg plus Offset.
LLVM_ABI void report_fatal_error(Error Err, bool gen_crash_diag=true)
CombinerObjective
The combiner's goal may differ based on which pattern it is attempting to optimize.
std::optional< UsedNZCV > examineCFlagsUse(MachineInstr &MI, MachineInstr &CmpInstr, const TargetRegisterInfo &TRI, SmallVectorImpl< MachineInstr * > *CCUseInstrs=nullptr)
CodeGenOptLevel
Code generation optimization level.
class LLVM_GSL_OWNER SmallVector
Forward declaration of SmallVector so that calculateSmallVectorDefaultInlinedElements can reference s...
auto instructionsWithoutDebug(IterT It, IterT End, bool SkipPseudoOp=true)
Construct a range iterator which begins at It and moves forwards until End is reached,...
LLVM_ABI raw_fd_ostream & errs()
This returns a reference to a raw_ostream for standard error.
auto drop_end(T &&RangeOrContainer, size_t N=1)
Return a range covering RangeOrContainer with the last N elements excluded.
unsigned getUndefRegState(bool B)
static MCRegister getXRegFromWReg(MCRegister Reg)
unsigned getDefRegState(bool B)
MCCFIInstruction createCFAOffset(const TargetRegisterInfo &MRI, unsigned Reg, const StackOffset &OffsetFromDefCFA, std::optional< int64_t > IncomingVGOffsetFromDefCFA)
unsigned getKillRegState(bool B)
FunctionAddr VTableAddr Next
DWARFExpression::Operation Op
ArrayRef(const T &OneElt) -> ArrayRef< T >
static bool isUncondBranchOpcode(int Opc)
decltype(auto) cast(const From &Val)
cast<X> - Return the argument parameter cast to the specified type.
void erase_if(Container &C, UnaryPredicate P)
Provide a container algorithm similar to C++ Library Fundamentals v2's erase_if which is equivalent t...
constexpr bool isIntN(unsigned N, int64_t x)
Checks if an signed integer fits into the given (dynamic) bit width.
bool rewriteAArch64FrameIndex(MachineInstr &MI, unsigned FrameRegIdx, unsigned FrameReg, StackOffset &Offset, const AArch64InstrInfo *TII)
rewriteAArch64FrameIndex - Rewrite MI to access 'Offset' bytes from the FP.
bool is_contained(R &&Range, const E &Element)
Returns true if Element is found in Range.
static const MachineMemOperand::Flags MOSuppressPair
constexpr int64_t SignExtend64(uint64_t x)
Sign-extend the number in the bottom B bits of X to a 64-bit integer.
void appendLEB128(SmallVectorImpl< U > &Buffer, T Value)
bool optimizeTerminators(MachineBasicBlock *MBB, const TargetInstrInfo &TII)
std::pair< MachineOperand, DIExpression * > ParamLoadedValue
bool isNZCVTouchedInInstructionRange(const MachineInstr &DefMI, const MachineInstr &UseMI, const TargetRegisterInfo *TRI)
Return true if there is an instruction /after/ DefMI and before UseMI which either reads or clobbers ...
auto seq(T Begin, T End)
Iterate over an integral type from Begin up to - but not including - End.
static const MachineMemOperand::Flags MOStridedAccess
void fullyRecomputeLiveIns(ArrayRef< MachineBasicBlock * > MBBs)
Convenience function for recomputing live-in's for a set of MBBs until the computation converges.
LLVM_ABI Printable printReg(Register Reg, const TargetRegisterInfo *TRI=nullptr, unsigned SubIdx=0, const MachineRegisterInfo *MRI=nullptr)
Prints virtual and physical registers with or without a TRI instance.
Used to describe addressing mode similar to ExtAddrMode in CodeGenPrepare.
LLVM_ABI static const MBBSectionID ColdSectionID
This class contains a discriminated union of information about pointers in memory operands,...
static LLVM_ABI MachinePointerInfo getFixedStack(MachineFunction &MF, int FI, int64_t Offset=0)
Return a MachinePointerInfo record that refers to the specified FrameIndex.
An individual sequence of instructions to be replaced with a call to an outlined function.
MachineFunction * getMF() const
The information necessary to create an outlined function for some class of candidate.