65#define GET_INSTRINFO_CTOR_DTOR
66#include "AArch64GenInstrInfo.inc"
70 cl::desc(
"Restrict range of CB instructions (DEBUG)"));
74 cl::desc(
"Restrict range of TB[N]Z instructions (DEBUG)"));
78 cl::desc(
"Restrict range of CB[N]Z instructions (DEBUG)"));
82 cl::desc(
"Restrict range of Bcc instructions (DEBUG)"));
86 cl::desc(
"Restrict range of B instructions (DEBUG)"));
90 cl::desc(
"Restrict range of instructions to search for the "
91 "machine-combiner gather pattern optimization"));
96 RI(STI.getTargetTriple(), STI.getHwMode()), Subtarget(STI) {}
107 auto Op =
MI.getOpcode();
108 if (
Op == AArch64::INLINEASM ||
Op == AArch64::INLINEASM_BR)
109 return getInlineAsmLength(
MI.getOperand(0).getSymbolName(), *MAI);
113 if (
MI.isMetaInstruction())
118 unsigned NumBytes = 0;
122 NumBytes =
Desc.getSize() ?
Desc.getSize() : 4;
125 if (!MFI->shouldSignReturnAddress(*MF))
129 auto Method = STI.getAuthenticatedLRCheckMethod(*MF);
137 switch (
Desc.getOpcode()) {
140 return Desc.getSize();
147 case TargetOpcode::STACKMAP:
150 assert(NumBytes % 4 == 0 &&
"Invalid number of NOP bytes requested!");
152 case TargetOpcode::PATCHPOINT:
155 assert(NumBytes % 4 == 0 &&
"Invalid number of NOP bytes requested!");
157 case TargetOpcode::STATEPOINT:
159 assert(NumBytes % 4 == 0 &&
"Invalid number of NOP bytes requested!");
164 case TargetOpcode::PATCHABLE_FUNCTION_ENTER:
169 F.getFnAttributeAsParsedInteger(
"patchable-function-entry", 9) * 4;
171 case TargetOpcode::PATCHABLE_FUNCTION_EXIT:
172 case TargetOpcode::PATCHABLE_TAIL_CALL:
173 case TargetOpcode::PATCHABLE_TYPED_EVENT_CALL:
177 case TargetOpcode::PATCHABLE_EVENT_CALL:
183 NumBytes =
MI.getOperand(1).getImm();
185 case TargetOpcode::BUNDLE:
186 NumBytes = getInstBundleLength(
MI);
193unsigned AArch64InstrInfo::getInstBundleLength(
const MachineInstr &
MI)
const {
197 while (++
I != E &&
I->isInsideBundle()) {
198 assert(!
I->isBundle() &&
"No nested bundle!");
233 case AArch64::CBWPri:
234 case AArch64::CBXPri:
235 case AArch64::CBWPrr:
236 case AArch64::CBXPrr:
244 case AArch64::CBBAssertExt:
245 case AArch64::CBHAssertExt:
276 case AArch64::CBWPri:
277 case AArch64::CBXPri:
278 case AArch64::CBBAssertExt:
279 case AArch64::CBHAssertExt:
280 case AArch64::CBWPrr:
281 case AArch64::CBXPrr:
287 int64_t BrOffset)
const {
289 assert(Bits >= 3 &&
"max branch displacement must be enough to jump"
290 "over conditional branch expansion");
291 return isIntN(Bits, BrOffset / 4);
296 switch (
MI.getOpcode()) {
300 return MI.getOperand(0).getMBB();
305 return MI.getOperand(2).getMBB();
311 return MI.getOperand(1).getMBB();
312 case AArch64::CBWPri:
313 case AArch64::CBXPri:
314 case AArch64::CBBAssertExt:
315 case AArch64::CBHAssertExt:
316 case AArch64::CBWPrr:
317 case AArch64::CBXPrr:
318 return MI.getOperand(3).getMBB();
328 assert(RS &&
"RegScavenger required for long branching");
330 "new block should be inserted for expanding unconditional branch");
333 "restore block should be inserted for restoring clobbered registers");
340 "Branch offsets outside of the signed 33-bit range not supported");
351 RS->enterBasicBlockEnd(
MBB);
354 constexpr Register Reg = AArch64::X16;
355 if (!RS->isRegUsed(Reg)) {
356 insertUnconditionalBranch(
MBB, &NewDestBB,
DL);
363 Register Scavenged = RS->FindUnusedReg(&AArch64::GPR64RegClass);
364 if (Scavenged != AArch64::NoRegister &&
366 buildIndirectBranch(Scavenged, NewDestBB);
367 RS->setRegUsed(Scavenged);
376 "Unable to insert indirect branch inside function that has red zone");
399 bool AllowModify)
const {
406 if (
I->getOpcode() == AArch64::SpeculationBarrierISBDSBEndBB ||
407 I->getOpcode() == AArch64::SpeculationBarrierSBEndBB) {
411 if (!isUnpredicatedTerminator(*
I))
418 unsigned LastOpc = LastInst->
getOpcode();
419 if (
I ==
MBB.begin() || !isUnpredicatedTerminator(*--
I)) {
434 unsigned SecondLastOpc = SecondLastInst->
getOpcode();
441 LastInst = SecondLastInst;
443 if (
I ==
MBB.begin() || !isUnpredicatedTerminator(*--
I)) {
448 SecondLastInst = &*
I;
449 SecondLastOpc = SecondLastInst->
getOpcode();
460 LastInst = SecondLastInst;
462 if (
I ==
MBB.begin() || !isUnpredicatedTerminator(*--
I)) {
464 "unreachable unconditional branches removed above");
473 SecondLastInst = &*
I;
474 SecondLastOpc = SecondLastInst->
getOpcode();
478 if (SecondLastInst &&
I !=
MBB.begin() && isUnpredicatedTerminator(*--
I))
494 I->eraseFromParent();
503 I->eraseFromParent();
512 MachineBranchPredicate &MBP,
513 bool AllowModify)
const {
523 if (
I->getOpcode() == AArch64::SpeculationBarrierISBDSBEndBB ||
524 I->getOpcode() == AArch64::SpeculationBarrierSBEndBB) {
528 if (!isUnpredicatedTerminator(*
I))
533 unsigned LastOpc = LastInst->
getOpcode();
548 assert(MBP.TrueDest &&
"expected!");
549 MBP.FalseDest =
MBB.getNextNode();
551 MBP.ConditionDef =
nullptr;
552 MBP.SingleUseCondition =
false;
556 MBP.Predicate = (LastOpc == AArch64::CBNZX || LastOpc == AArch64::CBNZW)
557 ? MachineBranchPredicate::PRED_NE
558 : MachineBranchPredicate::PRED_EQ;
574 Cond[1].setImm(AArch64::CBNZW);
577 Cond[1].setImm(AArch64::CBZW);
580 Cond[1].setImm(AArch64::CBNZX);
583 Cond[1].setImm(AArch64::CBZX);
586 Cond[1].setImm(AArch64::TBNZW);
589 Cond[1].setImm(AArch64::TBZW);
592 Cond[1].setImm(AArch64::TBNZX);
595 Cond[1].setImm(AArch64::TBZX);
599 case AArch64::CBWPri:
600 case AArch64::CBXPri:
601 case AArch64::CBBAssertExt:
602 case AArch64::CBHAssertExt:
603 case AArch64::CBWPrr:
604 case AArch64::CBXPrr: {
617 int *BytesRemoved)
const {
627 I->eraseFromParent();
631 if (
I ==
MBB.begin()) {
644 I->eraseFromParent();
651void AArch64InstrInfo::instantiateCondBranch(
676 if (
Cond.size() > 5) {
687 assert(
TBB &&
"insertBranch must not be told to insert a fallthrough");
714 unsigned Opc =
MI.getOpcode();
721 if (
MI.getOperand(0).getReg() == AArch64::WZR ||
722 MI.getOperand(0).getReg() == AArch64::XZR) {
724 dbgs() <<
"Removing always taken branch: " <<
MI);
727 for (
auto *S : Succs)
729 MBB->removeSuccessor(S);
731 while (
MBB->rbegin() != &
MI)
732 MBB->rbegin()->eraseFromParent();
733 MI.eraseFromParent();
743 if (
MI.getOperand(0).getReg() == AArch64::WZR ||
744 MI.getOperand(0).getReg() == AArch64::XZR) {
746 dbgs() <<
"Removing never taken branch: " <<
MI);
748 MI.getParent()->removeSuccessor(
Target);
749 MI.eraseFromParent();
762 if (!
DefMI->isFullCopy())
764 VReg =
DefMI->getOperand(1).getReg();
773 unsigned *NewReg =
nullptr) {
778 bool Is64Bit = AArch64::GPR64allRegClass.hasSubClassEq(
MRI.getRegClass(VReg));
782 switch (
DefMI->getOpcode()) {
783 case AArch64::SUBREG_TO_REG:
787 if (!
DefMI->getOperand(1).isImm() ||
DefMI->getOperand(1).getImm() != 0)
789 if (!
DefMI->getOperand(2).isReg())
791 if (!
DefMI->getOperand(3).isImm() ||
792 DefMI->getOperand(3).getImm() != AArch64::sub_32)
795 if (
DefMI->getOpcode() != AArch64::MOVi32imm)
797 if (!
DefMI->getOperand(1).isImm() ||
DefMI->getOperand(1).getImm() != 1)
800 SrcReg = AArch64::XZR;
801 Opc = AArch64::CSINCXr;
804 case AArch64::MOVi32imm:
805 case AArch64::MOVi64imm:
806 if (!
DefMI->getOperand(1).isImm() ||
DefMI->getOperand(1).getImm() != 1)
808 SrcReg = Is64Bit ? AArch64::XZR : AArch64::WZR;
809 Opc = Is64Bit ? AArch64::CSINCXr : AArch64::CSINCWr;
812 case AArch64::ADDSXri:
813 case AArch64::ADDSWri:
815 if (
DefMI->findRegisterDefOperandIdx(AArch64::NZCV,
nullptr,
820 case AArch64::ADDXri:
821 case AArch64::ADDWri:
823 if (!
DefMI->getOperand(2).isImm() ||
DefMI->getOperand(2).getImm() != 1 ||
824 DefMI->getOperand(3).getImm() != 0)
826 SrcReg =
DefMI->getOperand(1).getReg();
827 Opc = Is64Bit ? AArch64::CSINCXr : AArch64::CSINCWr;
830 case AArch64::ORNXrr:
831 case AArch64::ORNWrr: {
834 if (ZReg != AArch64::XZR && ZReg != AArch64::WZR)
836 SrcReg =
DefMI->getOperand(2).getReg();
837 Opc = Is64Bit ? AArch64::CSINVXr : AArch64::CSINVWr;
841 case AArch64::SUBSXrr:
842 case AArch64::SUBSWrr:
844 if (
DefMI->findRegisterDefOperandIdx(AArch64::NZCV,
nullptr,
849 case AArch64::SUBXrr:
850 case AArch64::SUBWrr: {
853 if (ZReg != AArch64::XZR && ZReg != AArch64::WZR)
855 SrcReg =
DefMI->getOperand(2).getReg();
856 Opc = Is64Bit ? AArch64::CSNEGXr : AArch64::CSNEGWr;
862 assert(
Opc && SrcReg &&
"Missing parameters");
874 int &FalseCycles)
const {
878 RI.getCommonSubClass(
MRI.getRegClass(TrueReg),
MRI.getRegClass(FalseReg));
885 if (!RI.getCommonSubClass(RC,
MRI.getRegClass(DstReg)))
889 unsigned ExtraCondLat =
Cond.size() != 1;
893 if (AArch64::GPR64allRegClass.hasSubClassEq(RC) ||
894 AArch64::GPR32allRegClass.hasSubClassEq(RC)) {
896 CondCycles = 1 + ExtraCondLat;
897 TrueCycles = FalseCycles = 1;
907 if (AArch64::FPR64RegClass.hasSubClassEq(RC) ||
908 AArch64::FPR32RegClass.hasSubClassEq(RC)) {
909 CondCycles = 5 + ExtraCondLat;
910 TrueCycles = FalseCycles = 2;
927 switch (
Cond.size()) {
959 MRI.constrainRegClass(SrcReg, &AArch64::GPR64spRegClass);
965 MRI.constrainRegClass(SrcReg, &AArch64::GPR32spRegClass);
1005 unsigned SubsOpc, SubsDestReg;
1011 case AArch64::CBWPri:
1012 SubsOpc = AArch64::SUBSWri;
1013 SubsDestReg = AArch64::WZR;
1016 case AArch64::CBXPri:
1017 SubsOpc = AArch64::SUBSXri;
1018 SubsDestReg = AArch64::XZR;
1021 case AArch64::CBWPrr:
1022 SubsOpc = AArch64::SUBSWrr;
1023 SubsDestReg = AArch64::WZR;
1026 case AArch64::CBXPrr:
1027 SubsOpc = AArch64::SUBSXrr;
1028 SubsDestReg = AArch64::XZR;
1057 switch (ExtendType) {
1063 "Unexpected compare-and-branch instruction for SXTB shift-extend");
1064 ExtOpc = AArch64::SBFMWri;
1070 "Unexpected compare-and-branch instruction for SXTH shift-extend");
1071 ExtOpc = AArch64::SBFMWri;
1077 "Unexpected compare-and-branch instruction for UXTB shift-extend");
1078 ExtOpc = AArch64::ANDWri;
1084 "Unexpected compare-and-branch instruction for UXTH shift-extend");
1085 ExtOpc = AArch64::ANDWri;
1091 Reg =
MRI.createVirtualRegister(&AArch64::GPR32spRegClass);
1094 if (ExtOpc != AArch64::ANDWri)
1096 MBBI.addImm(ExtBits);
1103 MRI.constrainRegClass(Reg,
MRI.getRegClass(
Cond[3].getReg()));
1104 MRI.constrainRegClass(
Cond[3].
getReg(), &AArch64::GPR32spRegClass);
1111 MRI.constrainRegClass(Reg,
MRI.getRegClass(
Cond[3].getReg()));
1112 MRI.constrainRegClass(
Cond[3].
getReg(), &AArch64::GPR32spRegClass);
1124 bool TryFold =
false;
1125 if (
MRI.constrainRegClass(DstReg, &AArch64::GPR64RegClass)) {
1126 RC = &AArch64::GPR64RegClass;
1127 Opc = AArch64::CSELXr;
1129 }
else if (
MRI.constrainRegClass(DstReg, &AArch64::GPR32RegClass)) {
1130 RC = &AArch64::GPR32RegClass;
1131 Opc = AArch64::CSELWr;
1133 }
else if (
MRI.constrainRegClass(DstReg, &AArch64::FPR64RegClass)) {
1134 RC = &AArch64::FPR64RegClass;
1135 Opc = AArch64::FCSELDrrr;
1136 }
else if (
MRI.constrainRegClass(DstReg, &AArch64::FPR32RegClass)) {
1137 RC = &AArch64::FPR32RegClass;
1138 Opc = AArch64::FCSELSrrr;
1140 assert(RC &&
"Unsupported regclass");
1144 unsigned NewReg = 0;
1159 MRI.clearKillFlags(NewReg);
1164 MRI.constrainRegClass(TrueReg, RC);
1167 (FalseReg.
isVirtual() || FalseReg == AArch64::WZR ||
1168 FalseReg == AArch64::XZR) &&
1169 "FalseReg was folded into a non-virtual register other than WZR or XZR");
1171 MRI.constrainRegClass(FalseReg, RC);
1186 assert(BitSize == 64 &&
"Only bit sizes of 32 or 64 allowed");
1191 return Is.
size() <= 2;
1196 assert(
MI.isCopy() &&
"Expected COPY instruction");
1202 if (
Reg.isVirtual())
1203 return MRI.getRegClass(
Reg);
1204 if (
Reg.isPhysical())
1205 return RI.getMinimalPhysRegClass(
Reg);
1210 if (DstRC && SrcRC && !RI.getCommonSubClass(DstRC, SrcRC))
1213 return MI.isAsCheapAsAMove();
1219 if (Subtarget.hasExynosCheapAsMoveHandling()) {
1220 if (isExynosCheapAsMove(
MI))
1222 return MI.isAsCheapAsAMove();
1225 switch (
MI.getOpcode()) {
1227 return MI.isAsCheapAsAMove();
1229 case TargetOpcode::COPY:
1232 case AArch64::ADDWrs:
1233 case AArch64::ADDXrs:
1234 case AArch64::SUBWrs:
1235 case AArch64::SUBXrs:
1236 return Subtarget.hasALULSLFast() &&
MI.getOperand(3).getImm() <= 4;
1241 case AArch64::MOVi32imm:
1243 case AArch64::MOVi64imm:
1248bool AArch64InstrInfo::isFalkorShiftExtFast(
const MachineInstr &
MI) {
1249 switch (
MI.getOpcode()) {
1253 case AArch64::ADDWrs:
1254 case AArch64::ADDXrs:
1255 case AArch64::ADDSWrs:
1256 case AArch64::ADDSXrs: {
1257 unsigned Imm =
MI.getOperand(3).getImm();
1264 case AArch64::ADDWrx:
1265 case AArch64::ADDXrx:
1266 case AArch64::ADDXrx64:
1267 case AArch64::ADDSWrx:
1268 case AArch64::ADDSXrx:
1269 case AArch64::ADDSXrx64: {
1270 unsigned Imm =
MI.getOperand(3).getImm();
1282 case AArch64::SUBWrs:
1283 case AArch64::SUBSWrs: {
1284 unsigned Imm =
MI.getOperand(3).getImm();
1286 return ShiftVal == 0 ||
1290 case AArch64::SUBXrs:
1291 case AArch64::SUBSXrs: {
1292 unsigned Imm =
MI.getOperand(3).getImm();
1294 return ShiftVal == 0 ||
1298 case AArch64::SUBWrx:
1299 case AArch64::SUBXrx:
1300 case AArch64::SUBXrx64:
1301 case AArch64::SUBSWrx:
1302 case AArch64::SUBSXrx:
1303 case AArch64::SUBSXrx64: {
1304 unsigned Imm =
MI.getOperand(3).getImm();
1316 case AArch64::LDRBBroW:
1317 case AArch64::LDRBBroX:
1318 case AArch64::LDRBroW:
1319 case AArch64::LDRBroX:
1320 case AArch64::LDRDroW:
1321 case AArch64::LDRDroX:
1322 case AArch64::LDRHHroW:
1323 case AArch64::LDRHHroX:
1324 case AArch64::LDRHroW:
1325 case AArch64::LDRHroX:
1326 case AArch64::LDRQroW:
1327 case AArch64::LDRQroX:
1328 case AArch64::LDRSBWroW:
1329 case AArch64::LDRSBWroX:
1330 case AArch64::LDRSBXroW:
1331 case AArch64::LDRSBXroX:
1332 case AArch64::LDRSHWroW:
1333 case AArch64::LDRSHWroX:
1334 case AArch64::LDRSHXroW:
1335 case AArch64::LDRSHXroX:
1336 case AArch64::LDRSWroW:
1337 case AArch64::LDRSWroX:
1338 case AArch64::LDRSroW:
1339 case AArch64::LDRSroX:
1340 case AArch64::LDRWroW:
1341 case AArch64::LDRWroX:
1342 case AArch64::LDRXroW:
1343 case AArch64::LDRXroX:
1344 case AArch64::PRFMroW:
1345 case AArch64::PRFMroX:
1346 case AArch64::STRBBroW:
1347 case AArch64::STRBBroX:
1348 case AArch64::STRBroW:
1349 case AArch64::STRBroX:
1350 case AArch64::STRDroW:
1351 case AArch64::STRDroX:
1352 case AArch64::STRHHroW:
1353 case AArch64::STRHHroX:
1354 case AArch64::STRHroW:
1355 case AArch64::STRHroX:
1356 case AArch64::STRQroW:
1357 case AArch64::STRQroX:
1358 case AArch64::STRSroW:
1359 case AArch64::STRSroX:
1360 case AArch64::STRWroW:
1361 case AArch64::STRWroX:
1362 case AArch64::STRXroW:
1363 case AArch64::STRXroX: {
1364 unsigned IsSigned =
MI.getOperand(3).getImm();
1371 unsigned Opc =
MI.getOpcode();
1375 case AArch64::SEH_StackAlloc:
1376 case AArch64::SEH_SaveFPLR:
1377 case AArch64::SEH_SaveFPLR_X:
1378 case AArch64::SEH_SaveReg:
1379 case AArch64::SEH_SaveReg_X:
1380 case AArch64::SEH_SaveRegP:
1381 case AArch64::SEH_SaveRegP_X:
1382 case AArch64::SEH_SaveFReg:
1383 case AArch64::SEH_SaveFReg_X:
1384 case AArch64::SEH_SaveFRegP:
1385 case AArch64::SEH_SaveFRegP_X:
1386 case AArch64::SEH_SetFP:
1387 case AArch64::SEH_AddFP:
1388 case AArch64::SEH_Nop:
1389 case AArch64::SEH_PrologEnd:
1390 case AArch64::SEH_EpilogStart:
1391 case AArch64::SEH_EpilogEnd:
1392 case AArch64::SEH_PACSignLR:
1393 case AArch64::SEH_SaveAnyRegI:
1394 case AArch64::SEH_SaveAnyRegIP:
1395 case AArch64::SEH_SaveAnyRegQP:
1396 case AArch64::SEH_SaveAnyRegQPX:
1397 case AArch64::SEH_AllocZ:
1398 case AArch64::SEH_SaveZReg:
1399 case AArch64::SEH_SavePReg:
1406 unsigned &SubIdx)
const {
1407 switch (
MI.getOpcode()) {
1410 case AArch64::SBFMXri:
1411 case AArch64::UBFMXri:
1414 if (
MI.getOperand(2).getImm() != 0 ||
MI.getOperand(3).getImm() != 31)
1417 SrcReg =
MI.getOperand(1).getReg();
1418 DstReg =
MI.getOperand(0).getReg();
1419 SubIdx = AArch64::sub_32;
1428 int64_t OffsetA = 0, OffsetB = 0;
1429 TypeSize WidthA(0,
false), WidthB(0,
false);
1430 bool OffsetAIsScalable =
false, OffsetBIsScalable =
false;
1451 OffsetAIsScalable == OffsetBIsScalable) {
1452 int LowOffset = OffsetA < OffsetB ? OffsetA : OffsetB;
1453 int HighOffset = OffsetA < OffsetB ? OffsetB : OffsetA;
1454 TypeSize LowWidth = (LowOffset == OffsetA) ? WidthA : WidthB;
1455 if (LowWidth.
isScalable() == OffsetAIsScalable &&
1473 switch (
MI.getOpcode()) {
1476 if (
MI.getOperand(0).getImm() == 0x14)
1483 case AArch64::MSRpstatesvcrImm1:
1490 auto Next = std::next(
MI.getIterator());
1491 return Next !=
MBB->end() &&
Next->isCFIInstruction();
1498 Register &SrcReg2, int64_t &CmpMask,
1499 int64_t &CmpValue)
const {
1503 assert(
MI.getNumOperands() >= 2 &&
"All AArch64 cmps should have 2 operands");
1504 if (!
MI.getOperand(1).isReg() ||
MI.getOperand(1).getSubReg())
1507 switch (
MI.getOpcode()) {
1510 case AArch64::PTEST_PP:
1511 case AArch64::PTEST_PP_ANY:
1512 case AArch64::PTEST_PP_FIRST:
1513 SrcReg =
MI.getOperand(0).getReg();
1514 SrcReg2 =
MI.getOperand(1).getReg();
1515 if (
MI.getOperand(2).getSubReg())
1522 case AArch64::SUBSWrr:
1523 case AArch64::SUBSWrs:
1524 case AArch64::SUBSWrx:
1525 case AArch64::SUBSXrr:
1526 case AArch64::SUBSXrs:
1527 case AArch64::SUBSXrx:
1528 case AArch64::ADDSWrr:
1529 case AArch64::ADDSWrs:
1530 case AArch64::ADDSWrx:
1531 case AArch64::ADDSXrr:
1532 case AArch64::ADDSXrs:
1533 case AArch64::ADDSXrx:
1535 SrcReg =
MI.getOperand(1).getReg();
1536 SrcReg2 =
MI.getOperand(2).getReg();
1539 if (
MI.getOperand(2).getSubReg())
1545 case AArch64::SUBSWri:
1546 case AArch64::ADDSWri:
1547 case AArch64::SUBSXri:
1548 case AArch64::ADDSXri:
1549 SrcReg =
MI.getOperand(1).getReg();
1552 CmpValue =
MI.getOperand(2).getImm();
1554 case AArch64::ANDSWri:
1555 case AArch64::ANDSXri:
1558 SrcReg =
MI.getOperand(1).getReg();
1562 MI.getOperand(2).getImm(),
1563 MI.getOpcode() == AArch64::ANDSWri ? 32 : 64);
1572 assert(
MBB &&
"Can't get MachineBasicBlock here");
1574 assert(MF &&
"Can't get MachineFunction here");
1579 for (
unsigned OpIdx = 0, EndIdx = Instr.getNumOperands();
OpIdx < EndIdx;
1586 if (!OpRegCstraints)
1594 "Operand has register constraints without being a register!");
1597 if (
Reg.isPhysical()) {
1601 !
MRI->constrainRegClass(
Reg, OpRegCstraints))
1614 bool MIDefinesZeroReg =
false;
1615 if (
MI.definesRegister(AArch64::WZR,
nullptr) ||
1616 MI.definesRegister(AArch64::XZR,
nullptr))
1617 MIDefinesZeroReg =
true;
1619 switch (
MI.getOpcode()) {
1621 return MI.getOpcode();
1622 case AArch64::ADDSWrr:
1623 return AArch64::ADDWrr;
1624 case AArch64::ADDSWri:
1625 return MIDefinesZeroReg ? AArch64::ADDSWri : AArch64::ADDWri;
1626 case AArch64::ADDSWrs:
1627 return MIDefinesZeroReg ? AArch64::ADDSWrs : AArch64::ADDWrs;
1628 case AArch64::ADDSWrx:
1629 return AArch64::ADDWrx;
1630 case AArch64::ADDSXrr:
1631 return AArch64::ADDXrr;
1632 case AArch64::ADDSXri:
1633 return MIDefinesZeroReg ? AArch64::ADDSXri : AArch64::ADDXri;
1634 case AArch64::ADDSXrs:
1635 return MIDefinesZeroReg ? AArch64::ADDSXrs : AArch64::ADDXrs;
1636 case AArch64::ADDSXrx:
1637 return AArch64::ADDXrx;
1638 case AArch64::SUBSWrr:
1639 return AArch64::SUBWrr;
1640 case AArch64::SUBSWri:
1641 return MIDefinesZeroReg ? AArch64::SUBSWri : AArch64::SUBWri;
1642 case AArch64::SUBSWrs:
1643 return MIDefinesZeroReg ? AArch64::SUBSWrs : AArch64::SUBWrs;
1644 case AArch64::SUBSWrx:
1645 return AArch64::SUBWrx;
1646 case AArch64::SUBSXrr:
1647 return AArch64::SUBXrr;
1648 case AArch64::SUBSXri:
1649 return MIDefinesZeroReg ? AArch64::SUBSXri : AArch64::SUBXri;
1650 case AArch64::SUBSXrs:
1651 return MIDefinesZeroReg ? AArch64::SUBSXrs : AArch64::SUBXrs;
1652 case AArch64::SUBSXrx:
1653 return AArch64::SUBXrx;
1668 if (To == To->getParent()->begin())
1673 if (To->getParent() != From->getParent())
1685 Instr.modifiesRegister(AArch64::NZCV,
TRI)) ||
1686 ((AccessToCheck &
AK_Read) && Instr.readsRegister(AArch64::NZCV,
TRI)))
1692std::optional<unsigned>
1696 unsigned MaskOpcode =
Mask->getOpcode();
1697 unsigned PredOpcode = Pred->
getOpcode();
1698 bool PredIsPTestLike = isPTestLikeOpcode(PredOpcode);
1699 bool PredIsWhileLike = isWhileOpcode(PredOpcode);
1701 if (PredIsWhileLike) {
1705 if ((Mask == Pred) && PTest->
getOpcode() == AArch64::PTEST_PP_ANY)
1712 getElementSizeForOpcode(MaskOpcode) ==
1713 getElementSizeForOpcode(PredOpcode))
1719 if (PTest->
getOpcode() == AArch64::PTEST_PP_FIRST &&
1726 if (PredIsPTestLike) {
1731 if ((Mask == Pred) && PTest->
getOpcode() == AArch64::PTEST_PP_ANY)
1739 if (Mask != PTestLikeMask && PTestLikeMask->isFullCopy() &&
1740 PTestLikeMask->getOperand(1).getReg().isVirtual())
1742 MRI->getUniqueVRegDef(PTestLikeMask->getOperand(1).getReg());
1748 getElementSizeForOpcode(MaskOpcode) ==
1749 getElementSizeForOpcode(PredOpcode)) {
1750 if (Mask == PTestLikeMask || PTest->
getOpcode() == AArch64::PTEST_PP_ANY)
1776 uint64_t PredElementSize = getElementSizeForOpcode(PredOpcode);
1778 PTest->
getOpcode() == AArch64::PTEST_PP_ANY))
1786 switch (PredOpcode) {
1787 case AArch64::AND_PPzPP:
1788 case AArch64::BIC_PPzPP:
1789 case AArch64::EOR_PPzPP:
1790 case AArch64::NAND_PPzPP:
1791 case AArch64::NOR_PPzPP:
1792 case AArch64::ORN_PPzPP:
1793 case AArch64::ORR_PPzPP:
1794 case AArch64::BRKA_PPzP:
1795 case AArch64::BRKPA_PPzPP:
1796 case AArch64::BRKB_PPzP:
1797 case AArch64::BRKPB_PPzPP:
1798 case AArch64::RDFFR_PPz: {
1802 if (Mask != PredMask)
1806 case AArch64::BRKN_PPzP: {
1810 if ((MaskOpcode != AArch64::PTRUE_B) ||
1811 (
Mask->getOperand(1).getImm() != 31))
1815 case AArch64::PTRUE_B:
1828bool AArch64InstrInfo::optimizePTestInstr(
1829 MachineInstr *PTest,
unsigned MaskReg,
unsigned PredReg,
1831 auto *
Mask =
MRI->getUniqueVRegDef(MaskReg);
1832 auto *Pred =
MRI->getUniqueVRegDef(PredReg);
1834 if (Pred->
isCopy() && PTest->
getOpcode() == AArch64::PTEST_PP_FIRST) {
1838 if (
Op.isReg() &&
Op.getReg().isVirtual() &&
1839 Op.getSubReg() == AArch64::psub0)
1840 Pred =
MRI->getUniqueVRegDef(
Op.getReg());
1843 unsigned PredOpcode = Pred->
getOpcode();
1844 auto NewOp = canRemovePTestInstr(PTest, Mask, Pred,
MRI);
1860 if (*NewOp != PredOpcode) {
1871 for (; i !=
e; ++i) {
1902 if (DeadNZCVIdx != -1) {
1921 if (CmpInstr.
getOpcode() == AArch64::PTEST_PP ||
1922 CmpInstr.
getOpcode() == AArch64::PTEST_PP_ANY ||
1923 CmpInstr.
getOpcode() == AArch64::PTEST_PP_FIRST)
1924 return optimizePTestInstr(&CmpInstr, SrcReg, SrcReg2,
MRI);
1933 if (CmpValue == 0 && substituteCmpToZero(CmpInstr, SrcReg, *
MRI))
1935 return (CmpValue == 0 || CmpValue == 1) &&
1936 removeCmpToZeroOrOne(CmpInstr, SrcReg, CmpValue, *
MRI);
1944 switch (Instr.getOpcode()) {
1946 return AArch64::INSTRUCTION_LIST_END;
1948 case AArch64::ADDSWrr:
1949 case AArch64::ADDSWri:
1950 case AArch64::ADDSXrr:
1951 case AArch64::ADDSXri:
1952 case AArch64::ADDSWrx:
1953 case AArch64::ADDSXrx:
1954 case AArch64::SUBSWrr:
1955 case AArch64::SUBSWri:
1956 case AArch64::SUBSWrx:
1957 case AArch64::SUBSXrr:
1958 case AArch64::SUBSXri:
1959 case AArch64::SUBSXrx:
1960 case AArch64::ANDSWri:
1961 case AArch64::ANDSWrr:
1962 case AArch64::ANDSWrs:
1963 case AArch64::ANDSXri:
1964 case AArch64::ANDSXrr:
1965 case AArch64::ANDSXrs:
1966 case AArch64::BICSWrr:
1967 case AArch64::BICSXrr:
1968 case AArch64::BICSWrs:
1969 case AArch64::BICSXrs:
1970 return Instr.getOpcode();
1972 case AArch64::ADDWrr:
1973 return AArch64::ADDSWrr;
1974 case AArch64::ADDWri:
1975 return AArch64::ADDSWri;
1976 case AArch64::ADDXrr:
1977 return AArch64::ADDSXrr;
1978 case AArch64::ADDXri:
1979 return AArch64::ADDSXri;
1980 case AArch64::ADDWrx:
1981 return AArch64::ADDSWrx;
1982 case AArch64::ADDXrx:
1983 return AArch64::ADDSXrx;
1984 case AArch64::ADCWr:
1985 return AArch64::ADCSWr;
1986 case AArch64::ADCXr:
1987 return AArch64::ADCSXr;
1988 case AArch64::SUBWrr:
1989 return AArch64::SUBSWrr;
1990 case AArch64::SUBWri:
1991 return AArch64::SUBSWri;
1992 case AArch64::SUBXrr:
1993 return AArch64::SUBSXrr;
1994 case AArch64::SUBXri:
1995 return AArch64::SUBSXri;
1996 case AArch64::SUBWrx:
1997 return AArch64::SUBSWrx;
1998 case AArch64::SUBXrx:
1999 return AArch64::SUBSXrx;
2000 case AArch64::SBCWr:
2001 return AArch64::SBCSWr;
2002 case AArch64::SBCXr:
2003 return AArch64::SBCSXr;
2004 case AArch64::ANDWri:
2005 return AArch64::ANDSWri;
2006 case AArch64::ANDXri:
2007 return AArch64::ANDSXri;
2008 case AArch64::ANDWrr:
2009 return AArch64::ANDSWrr;
2010 case AArch64::ANDWrs:
2011 return AArch64::ANDSWrs;
2012 case AArch64::ANDXrr:
2013 return AArch64::ANDSXrr;
2014 case AArch64::ANDXrs:
2015 return AArch64::ANDSXrs;
2016 case AArch64::BICWrr:
2017 return AArch64::BICSWrr;
2018 case AArch64::BICXrr:
2019 return AArch64::BICSXrr;
2020 case AArch64::BICWrs:
2021 return AArch64::BICSWrs;
2022 case AArch64::BICXrs:
2023 return AArch64::BICSXrs;
2029 for (
auto *BB :
MBB->successors())
2030 if (BB->isLiveIn(AArch64::NZCV))
2039 switch (Instr.getOpcode()) {
2043 case AArch64::Bcc: {
2044 int Idx = Instr.findRegisterUseOperandIdx(AArch64::NZCV,
nullptr);
2049 case AArch64::CSINVWr:
2050 case AArch64::CSINVXr:
2051 case AArch64::CSINCWr:
2052 case AArch64::CSINCXr:
2053 case AArch64::CSELWr:
2054 case AArch64::CSELXr:
2055 case AArch64::CSNEGWr:
2056 case AArch64::CSNEGXr:
2057 case AArch64::FCSELSrrr:
2058 case AArch64::FCSELDrrr: {
2059 int Idx = Instr.findRegisterUseOperandIdx(AArch64::NZCV,
nullptr);
2072 Instr.getOperand(CCIdx).
getImm())
2125std::optional<UsedNZCV>
2130 if (
MI.getParent() != CmpParent)
2131 return std::nullopt;
2134 return std::nullopt;
2139 if (Instr.readsRegister(AArch64::NZCV, &
TRI)) {
2142 return std::nullopt;
2147 if (Instr.modifiesRegister(AArch64::NZCV, &
TRI))
2150 return NZCVUsedAfterCmp;
2154 return Opcode == AArch64::ADDSWri || Opcode == AArch64::ADDSXri;
2158 return Opcode == AArch64::SUBSWri || Opcode == AArch64::SUBSXri;
2164 case AArch64::ANDSWri:
2165 case AArch64::ANDSWrr:
2166 case AArch64::ANDSWrs:
2167 case AArch64::ANDSXri:
2168 case AArch64::ANDSXrr:
2169 case AArch64::ANDSXrs:
2170 case AArch64::BICSWrr:
2171 case AArch64::BICSXrr:
2172 case AArch64::BICSWrs:
2173 case AArch64::BICSXrs:
2199 const unsigned CmpOpcode = CmpInstr.
getOpcode();
2205 "Caller guarantees that CmpInstr compares with constant 0");
2208 if (!NZVCUsed || NZVCUsed->C)
2231bool AArch64InstrInfo::substituteCmpToZero(
2235 MachineInstr *
MI =
MRI.getUniqueVRegDef(SrcReg);
2242 if (NewOpc == AArch64::INSTRUCTION_LIST_END)
2249 MI->setDesc(
get(NewOpc));
2254 MI->addRegisterDefined(AArch64::NZCV, &
TRI);
2266 assert((CmpValue == 0 || CmpValue == 1) &&
2267 "Only comparisons to 0 or 1 considered for removal!");
2270 unsigned MIOpc =
MI.getOpcode();
2271 if (MIOpc == AArch64::CSINCWr) {
2272 if (
MI.getOperand(1).getReg() != AArch64::WZR ||
2273 MI.getOperand(2).getReg() != AArch64::WZR)
2275 }
else if (MIOpc == AArch64::CSINCXr) {
2276 if (
MI.getOperand(1).getReg() != AArch64::XZR ||
2277 MI.getOperand(2).getReg() != AArch64::XZR)
2287 if (
MI.findRegisterDefOperandIdx(AArch64::NZCV,
nullptr,
true) != -1)
2291 const unsigned CmpOpcode = CmpInstr.
getOpcode();
2293 if (CmpValue && !IsSubsRegImm)
2295 if (!CmpValue && !IsSubsRegImm && !
isADDSRegImm(CmpOpcode))
2300 if (MIUsedNZCV.
C || MIUsedNZCV.
V)
2303 std::optional<UsedNZCV> NZCVUsedAfterCmp =
2307 if (!NZCVUsedAfterCmp || NZCVUsedAfterCmp->C || NZCVUsedAfterCmp->V)
2310 if ((MIUsedNZCV.
Z && NZCVUsedAfterCmp->N) ||
2311 (MIUsedNZCV.
N && NZCVUsedAfterCmp->Z))
2314 if (MIUsedNZCV.
N && !CmpValue)
2356bool AArch64InstrInfo::removeCmpToZeroOrOne(
2359 MachineInstr *
MI =
MRI.getUniqueVRegDef(SrcReg);
2363 SmallVector<MachineInstr *, 4> CCUseInstrs;
2364 bool IsInvertCC =
false;
2372 for (MachineInstr *CCUseInstr : CCUseInstrs) {
2374 assert(Idx >= 0 &&
"Unexpected instruction using CC.");
2375 MachineOperand &CCOperand = CCUseInstr->getOperand(Idx);
2384bool AArch64InstrInfo::expandPostRAPseudo(
MachineInstr &
MI)
const {
2385 if (
MI.getOpcode() != TargetOpcode::LOAD_STACK_GUARD &&
2386 MI.getOpcode() != AArch64::CATCHRET)
2389 MachineBasicBlock &
MBB = *
MI.getParent();
2391 auto TRI = Subtarget.getRegisterInfo();
2394 if (
MI.getOpcode() == AArch64::CATCHRET) {
2396 const TargetInstrInfo *
TII =
2398 MachineBasicBlock *TargetMBB =
MI.getOperand(0).getMBB();
2403 FirstEpilogSEH = std::prev(FirstEpilogSEH);
2405 FirstEpilogSEH = std::next(FirstEpilogSEH);
2420 if (
M.getStackProtectorGuard() ==
"sysreg") {
2421 const AArch64SysReg::SysReg *SrcReg =
2422 AArch64SysReg::lookupSysRegByName(
M.getStackProtectorGuardReg());
2430 int Offset =
M.getStackProtectorGuardOffset();
2481 const GlobalValue *GV =
2484 unsigned OpFlags = Subtarget.ClassifyGlobalReference(GV, TM);
2490 if (Subtarget.isTargetILP32()) {
2491 unsigned Reg32 =
TRI->getSubReg(
Reg, AArch64::sub_32);
2505 assert(!Subtarget.isTargetILP32() &&
"how can large exist in ILP32?");
2532 if (Subtarget.isTargetILP32()) {
2533 unsigned Reg32 =
TRI->getSubReg(
Reg, AArch64::sub_32);
2556 switch (
MI.getOpcode()) {
2559 case AArch64::MOVZWi:
2560 case AArch64::MOVZXi:
2561 if (
MI.getOperand(1).isImm() &&
MI.getOperand(1).getImm() == 0) {
2562 assert(
MI.getDesc().getNumOperands() == 3 &&
2563 MI.getOperand(2).getImm() == 0 &&
"invalid MOVZi operands");
2567 case AArch64::ANDWri:
2568 return MI.getOperand(1).getReg() == AArch64::WZR;
2569 case AArch64::ANDXri:
2570 return MI.getOperand(1).getReg() == AArch64::XZR;
2571 case TargetOpcode::COPY:
2572 return MI.getOperand(1).getReg() == AArch64::WZR;
2580 switch (
MI.getOpcode()) {
2583 case TargetOpcode::COPY: {
2586 return (AArch64::GPR32RegClass.
contains(DstReg) ||
2587 AArch64::GPR64RegClass.
contains(DstReg));
2589 case AArch64::ORRXrs:
2590 if (
MI.getOperand(1).getReg() == AArch64::XZR) {
2591 assert(
MI.getDesc().getNumOperands() == 4 &&
2592 MI.getOperand(3).getImm() == 0 &&
"invalid ORRrs operands");
2596 case AArch64::ADDXri:
2597 if (
MI.getOperand(2).getImm() == 0) {
2598 assert(
MI.getDesc().getNumOperands() == 4 &&
2599 MI.getOperand(3).getImm() == 0 &&
"invalid ADDXri operands");
2610 switch (
MI.getOpcode()) {
2613 case TargetOpcode::COPY: {
2615 return AArch64::FPR128RegClass.contains(DstReg);
2617 case AArch64::ORRv16i8:
2618 if (
MI.getOperand(1).getReg() ==
MI.getOperand(2).getReg()) {
2619 assert(
MI.getDesc().getNumOperands() == 3 &&
MI.getOperand(0).isReg() &&
2620 "invalid ORRv16i8 operands");
2632 case AArch64::LDRWui:
2633 case AArch64::LDRXui:
2634 case AArch64::LDRBui:
2635 case AArch64::LDRHui:
2636 case AArch64::LDRSui:
2637 case AArch64::LDRDui:
2638 case AArch64::LDRQui:
2639 case AArch64::LDR_PXI:
2645 int &FrameIndex)
const {
2649 if (
MI.getOperand(0).getSubReg() == 0 &&
MI.getOperand(1).isFI() &&
2650 MI.getOperand(2).isImm() &&
MI.getOperand(2).getImm() == 0) {
2651 FrameIndex =
MI.getOperand(1).getIndex();
2652 return MI.getOperand(0).getReg();
2661 case AArch64::STRWui:
2662 case AArch64::STRXui:
2663 case AArch64::STRBui:
2664 case AArch64::STRHui:
2665 case AArch64::STRSui:
2666 case AArch64::STRDui:
2667 case AArch64::STRQui:
2668 case AArch64::STR_PXI:
2674 int &FrameIndex)
const {
2678 if (
MI.getOperand(0).getSubReg() == 0 &&
MI.getOperand(1).isFI() &&
2679 MI.getOperand(2).isImm() &&
MI.getOperand(2).getImm() == 0) {
2680 FrameIndex =
MI.getOperand(1).getIndex();
2681 return MI.getOperand(0).getReg();
2687 int &FrameIndex)
const {
2702 return MI.getOperand(0).getReg();
2708 int &FrameIndex)
const {
2723 return MI.getOperand(0).getReg();
2731 return MMO->getFlags() & MOSuppressPair;
2737 if (
MI.memoperands_empty())
2745 return MMO->getFlags() & MOStridedAccess;
2753 case AArch64::STURSi:
2754 case AArch64::STRSpre:
2755 case AArch64::STURDi:
2756 case AArch64::STRDpre:
2757 case AArch64::STURQi:
2758 case AArch64::STRQpre:
2759 case AArch64::STURBBi:
2760 case AArch64::STURHHi:
2761 case AArch64::STURWi:
2762 case AArch64::STRWpre:
2763 case AArch64::STURXi:
2764 case AArch64::STRXpre:
2765 case AArch64::LDURSi:
2766 case AArch64::LDRSpre:
2767 case AArch64::LDURDi:
2768 case AArch64::LDRDpre:
2769 case AArch64::LDURQi:
2770 case AArch64::LDRQpre:
2771 case AArch64::LDURWi:
2772 case AArch64::LDRWpre:
2773 case AArch64::LDURXi:
2774 case AArch64::LDRXpre:
2775 case AArch64::LDRSWpre:
2776 case AArch64::LDURSWi:
2777 case AArch64::LDURHHi:
2778 case AArch64::LDURBBi:
2779 case AArch64::LDURSBWi:
2780 case AArch64::LDURSHWi:
2788 case AArch64::PRFMui:
return AArch64::PRFUMi;
2789 case AArch64::LDRXui:
return AArch64::LDURXi;
2790 case AArch64::LDRWui:
return AArch64::LDURWi;
2791 case AArch64::LDRBui:
return AArch64::LDURBi;
2792 case AArch64::LDRHui:
return AArch64::LDURHi;
2793 case AArch64::LDRSui:
return AArch64::LDURSi;
2794 case AArch64::LDRDui:
return AArch64::LDURDi;
2795 case AArch64::LDRQui:
return AArch64::LDURQi;
2796 case AArch64::LDRBBui:
return AArch64::LDURBBi;
2797 case AArch64::LDRHHui:
return AArch64::LDURHHi;
2798 case AArch64::LDRSBXui:
return AArch64::LDURSBXi;
2799 case AArch64::LDRSBWui:
return AArch64::LDURSBWi;
2800 case AArch64::LDRSHXui:
return AArch64::LDURSHXi;
2801 case AArch64::LDRSHWui:
return AArch64::LDURSHWi;
2802 case AArch64::LDRSWui:
return AArch64::LDURSWi;
2803 case AArch64::STRXui:
return AArch64::STURXi;
2804 case AArch64::STRWui:
return AArch64::STURWi;
2805 case AArch64::STRBui:
return AArch64::STURBi;
2806 case AArch64::STRHui:
return AArch64::STURHi;
2807 case AArch64::STRSui:
return AArch64::STURSi;
2808 case AArch64::STRDui:
return AArch64::STURDi;
2809 case AArch64::STRQui:
return AArch64::STURQi;
2810 case AArch64::STRBBui:
return AArch64::STURBBi;
2811 case AArch64::STRHHui:
return AArch64::STURHHi;
2820 case AArch64::LDAPURBi:
2821 case AArch64::LDAPURHi:
2822 case AArch64::LDAPURi:
2823 case AArch64::LDAPURSBWi:
2824 case AArch64::LDAPURSBXi:
2825 case AArch64::LDAPURSHWi:
2826 case AArch64::LDAPURSHXi:
2827 case AArch64::LDAPURSWi:
2828 case AArch64::LDAPURXi:
2829 case AArch64::LDR_PPXI:
2830 case AArch64::LDR_PXI:
2831 case AArch64::LDR_ZXI:
2832 case AArch64::LDR_ZZXI:
2833 case AArch64::LDR_ZZXI_STRIDED_CONTIGUOUS:
2834 case AArch64::LDR_ZZZXI:
2835 case AArch64::LDR_ZZZZXI:
2836 case AArch64::LDR_ZZZZXI_STRIDED_CONTIGUOUS:
2837 case AArch64::LDRBBui:
2838 case AArch64::LDRBui:
2839 case AArch64::LDRDui:
2840 case AArch64::LDRHHui:
2841 case AArch64::LDRHui:
2842 case AArch64::LDRQui:
2843 case AArch64::LDRSBWui:
2844 case AArch64::LDRSBXui:
2845 case AArch64::LDRSHWui:
2846 case AArch64::LDRSHXui:
2847 case AArch64::LDRSui:
2848 case AArch64::LDRSWui:
2849 case AArch64::LDRWui:
2850 case AArch64::LDRXui:
2851 case AArch64::LDURBBi:
2852 case AArch64::LDURBi:
2853 case AArch64::LDURDi:
2854 case AArch64::LDURHHi:
2855 case AArch64::LDURHi:
2856 case AArch64::LDURQi:
2857 case AArch64::LDURSBWi:
2858 case AArch64::LDURSBXi:
2859 case AArch64::LDURSHWi:
2860 case AArch64::LDURSHXi:
2861 case AArch64::LDURSi:
2862 case AArch64::LDURSWi:
2863 case AArch64::LDURWi:
2864 case AArch64::LDURXi:
2865 case AArch64::PRFMui:
2866 case AArch64::PRFUMi:
2867 case AArch64::ST2Gi:
2869 case AArch64::STLURBi:
2870 case AArch64::STLURHi:
2871 case AArch64::STLURWi:
2872 case AArch64::STLURXi:
2873 case AArch64::StoreSwiftAsyncContext:
2874 case AArch64::STR_PPXI:
2875 case AArch64::STR_PXI:
2876 case AArch64::STR_ZXI:
2877 case AArch64::STR_ZZXI:
2878 case AArch64::STR_ZZXI_STRIDED_CONTIGUOUS:
2879 case AArch64::STR_ZZZXI:
2880 case AArch64::STR_ZZZZXI:
2881 case AArch64::STR_ZZZZXI_STRIDED_CONTIGUOUS:
2882 case AArch64::STRBBui:
2883 case AArch64::STRBui:
2884 case AArch64::STRDui:
2885 case AArch64::STRHHui:
2886 case AArch64::STRHui:
2887 case AArch64::STRQui:
2888 case AArch64::STRSui:
2889 case AArch64::STRWui:
2890 case AArch64::STRXui:
2891 case AArch64::STURBBi:
2892 case AArch64::STURBi:
2893 case AArch64::STURDi:
2894 case AArch64::STURHHi:
2895 case AArch64::STURHi:
2896 case AArch64::STURQi:
2897 case AArch64::STURSi:
2898 case AArch64::STURWi:
2899 case AArch64::STURXi:
2900 case AArch64::STZ2Gi:
2901 case AArch64::STZGi:
2902 case AArch64::TAGPstack:
2904 case AArch64::LD1B_D_IMM:
2905 case AArch64::LD1B_H_IMM:
2906 case AArch64::LD1B_IMM:
2907 case AArch64::LD1B_S_IMM:
2908 case AArch64::LD1D_IMM:
2909 case AArch64::LD1H_D_IMM:
2910 case AArch64::LD1H_IMM:
2911 case AArch64::LD1H_S_IMM:
2912 case AArch64::LD1RB_D_IMM:
2913 case AArch64::LD1RB_H_IMM:
2914 case AArch64::LD1RB_IMM:
2915 case AArch64::LD1RB_S_IMM:
2916 case AArch64::LD1RD_IMM:
2917 case AArch64::LD1RH_D_IMM:
2918 case AArch64::LD1RH_IMM:
2919 case AArch64::LD1RH_S_IMM:
2920 case AArch64::LD1RSB_D_IMM:
2921 case AArch64::LD1RSB_H_IMM:
2922 case AArch64::LD1RSB_S_IMM:
2923 case AArch64::LD1RSH_D_IMM:
2924 case AArch64::LD1RSH_S_IMM:
2925 case AArch64::LD1RSW_IMM:
2926 case AArch64::LD1RW_D_IMM:
2927 case AArch64::LD1RW_IMM:
2928 case AArch64::LD1SB_D_IMM:
2929 case AArch64::LD1SB_H_IMM:
2930 case AArch64::LD1SB_S_IMM:
2931 case AArch64::LD1SH_D_IMM:
2932 case AArch64::LD1SH_S_IMM:
2933 case AArch64::LD1SW_D_IMM:
2934 case AArch64::LD1W_D_IMM:
2935 case AArch64::LD1W_IMM:
2936 case AArch64::LD2B_IMM:
2937 case AArch64::LD2D_IMM:
2938 case AArch64::LD2H_IMM:
2939 case AArch64::LD2W_IMM:
2940 case AArch64::LD3B_IMM:
2941 case AArch64::LD3D_IMM:
2942 case AArch64::LD3H_IMM:
2943 case AArch64::LD3W_IMM:
2944 case AArch64::LD4B_IMM:
2945 case AArch64::LD4D_IMM:
2946 case AArch64::LD4H_IMM:
2947 case AArch64::LD4W_IMM:
2949 case AArch64::LDNF1B_D_IMM:
2950 case AArch64::LDNF1B_H_IMM:
2951 case AArch64::LDNF1B_IMM:
2952 case AArch64::LDNF1B_S_IMM:
2953 case AArch64::LDNF1D_IMM:
2954 case AArch64::LDNF1H_D_IMM:
2955 case AArch64::LDNF1H_IMM:
2956 case AArch64::LDNF1H_S_IMM:
2957 case AArch64::LDNF1SB_D_IMM:
2958 case AArch64::LDNF1SB_H_IMM:
2959 case AArch64::LDNF1SB_S_IMM:
2960 case AArch64::LDNF1SH_D_IMM:
2961 case AArch64::LDNF1SH_S_IMM:
2962 case AArch64::LDNF1SW_D_IMM:
2963 case AArch64::LDNF1W_D_IMM:
2964 case AArch64::LDNF1W_IMM:
2965 case AArch64::LDNPDi:
2966 case AArch64::LDNPQi:
2967 case AArch64::LDNPSi:
2968 case AArch64::LDNPWi:
2969 case AArch64::LDNPXi:
2970 case AArch64::LDNT1B_ZRI:
2971 case AArch64::LDNT1D_ZRI:
2972 case AArch64::LDNT1H_ZRI:
2973 case AArch64::LDNT1W_ZRI:
2974 case AArch64::LDPDi:
2975 case AArch64::LDPQi:
2976 case AArch64::LDPSi:
2977 case AArch64::LDPWi:
2978 case AArch64::LDPXi:
2979 case AArch64::LDRBBpost:
2980 case AArch64::LDRBBpre:
2981 case AArch64::LDRBpost:
2982 case AArch64::LDRBpre:
2983 case AArch64::LDRDpost:
2984 case AArch64::LDRDpre:
2985 case AArch64::LDRHHpost:
2986 case AArch64::LDRHHpre:
2987 case AArch64::LDRHpost:
2988 case AArch64::LDRHpre:
2989 case AArch64::LDRQpost:
2990 case AArch64::LDRQpre:
2991 case AArch64::LDRSpost:
2992 case AArch64::LDRSpre:
2993 case AArch64::LDRWpost:
2994 case AArch64::LDRWpre:
2995 case AArch64::LDRXpost:
2996 case AArch64::LDRXpre:
2997 case AArch64::ST1B_D_IMM:
2998 case AArch64::ST1B_H_IMM:
2999 case AArch64::ST1B_IMM:
3000 case AArch64::ST1B_S_IMM:
3001 case AArch64::ST1D_IMM:
3002 case AArch64::ST1H_D_IMM:
3003 case AArch64::ST1H_IMM:
3004 case AArch64::ST1H_S_IMM:
3005 case AArch64::ST1W_D_IMM:
3006 case AArch64::ST1W_IMM:
3007 case AArch64::ST2B_IMM:
3008 case AArch64::ST2D_IMM:
3009 case AArch64::ST2H_IMM:
3010 case AArch64::ST2W_IMM:
3011 case AArch64::ST3B_IMM:
3012 case AArch64::ST3D_IMM:
3013 case AArch64::ST3H_IMM:
3014 case AArch64::ST3W_IMM:
3015 case AArch64::ST4B_IMM:
3016 case AArch64::ST4D_IMM:
3017 case AArch64::ST4H_IMM:
3018 case AArch64::ST4W_IMM:
3019 case AArch64::STGPi:
3020 case AArch64::STGPreIndex:
3021 case AArch64::STZGPreIndex:
3022 case AArch64::ST2GPreIndex:
3023 case AArch64::STZ2GPreIndex:
3024 case AArch64::STGPostIndex:
3025 case AArch64::STZGPostIndex:
3026 case AArch64::ST2GPostIndex:
3027 case AArch64::STZ2GPostIndex:
3028 case AArch64::STNPDi:
3029 case AArch64::STNPQi:
3030 case AArch64::STNPSi:
3031 case AArch64::STNPWi:
3032 case AArch64::STNPXi:
3033 case AArch64::STNT1B_ZRI:
3034 case AArch64::STNT1D_ZRI:
3035 case AArch64::STNT1H_ZRI:
3036 case AArch64::STNT1W_ZRI:
3037 case AArch64::STPDi:
3038 case AArch64::STPQi:
3039 case AArch64::STPSi:
3040 case AArch64::STPWi:
3041 case AArch64::STPXi:
3042 case AArch64::STRBBpost:
3043 case AArch64::STRBBpre:
3044 case AArch64::STRBpost:
3045 case AArch64::STRBpre:
3046 case AArch64::STRDpost:
3047 case AArch64::STRDpre:
3048 case AArch64::STRHHpost:
3049 case AArch64::STRHHpre:
3050 case AArch64::STRHpost:
3051 case AArch64::STRHpre:
3052 case AArch64::STRQpost:
3053 case AArch64::STRQpre:
3054 case AArch64::STRSpost:
3055 case AArch64::STRSpre:
3056 case AArch64::STRWpost:
3057 case AArch64::STRWpre:
3058 case AArch64::STRXpost:
3059 case AArch64::STRXpre:
3061 case AArch64::LDPDpost:
3062 case AArch64::LDPDpre:
3063 case AArch64::LDPQpost:
3064 case AArch64::LDPQpre:
3065 case AArch64::LDPSpost:
3066 case AArch64::LDPSpre:
3067 case AArch64::LDPWpost:
3068 case AArch64::LDPWpre:
3069 case AArch64::LDPXpost:
3070 case AArch64::LDPXpre:
3071 case AArch64::STGPpre:
3072 case AArch64::STGPpost:
3073 case AArch64::STPDpost:
3074 case AArch64::STPDpre:
3075 case AArch64::STPQpost:
3076 case AArch64::STPQpre:
3077 case AArch64::STPSpost:
3078 case AArch64::STPSpre:
3079 case AArch64::STPWpost:
3080 case AArch64::STPWpre:
3081 case AArch64::STPXpost:
3082 case AArch64::STPXpre:
3088 switch (
MI.getOpcode()) {
3092 case AArch64::STRSui:
3093 case AArch64::STRDui:
3094 case AArch64::STRQui:
3095 case AArch64::STRXui:
3096 case AArch64::STRWui:
3097 case AArch64::LDRSui:
3098 case AArch64::LDRDui:
3099 case AArch64::LDRQui:
3100 case AArch64::LDRXui:
3101 case AArch64::LDRWui:
3102 case AArch64::LDRSWui:
3104 case AArch64::STURSi:
3105 case AArch64::STRSpre:
3106 case AArch64::STURDi:
3107 case AArch64::STRDpre:
3108 case AArch64::STURQi:
3109 case AArch64::STRQpre:
3110 case AArch64::STURWi:
3111 case AArch64::STRWpre:
3112 case AArch64::STURXi:
3113 case AArch64::STRXpre:
3114 case AArch64::LDURSi:
3115 case AArch64::LDRSpre:
3116 case AArch64::LDURDi:
3117 case AArch64::LDRDpre:
3118 case AArch64::LDURQi:
3119 case AArch64::LDRQpre:
3120 case AArch64::LDURWi:
3121 case AArch64::LDRWpre:
3122 case AArch64::LDURXi:
3123 case AArch64::LDRXpre:
3124 case AArch64::LDURSWi:
3125 case AArch64::LDRSWpre:
3127 case AArch64::LDR_ZXI:
3128 case AArch64::STR_ZXI:
3134 switch (
MI.getOpcode()) {
3137 "Unexpected instruction - was a new tail call opcode introduced?");
3139 case AArch64::TCRETURNdi:
3140 case AArch64::TCRETURNri:
3141 case AArch64::TCRETURNrix16x17:
3142 case AArch64::TCRETURNrix17:
3143 case AArch64::TCRETURNrinotx16:
3144 case AArch64::TCRETURNriALL:
3145 case AArch64::AUTH_TCRETURN:
3146 case AArch64::AUTH_TCRETURN_BTI:
3156 case AArch64::ADDWri:
3157 return AArch64::ADDSWri;
3158 case AArch64::ADDWrr:
3159 return AArch64::ADDSWrr;
3160 case AArch64::ADDWrs:
3161 return AArch64::ADDSWrs;
3162 case AArch64::ADDWrx:
3163 return AArch64::ADDSWrx;
3164 case AArch64::ANDWri:
3165 return AArch64::ANDSWri;
3166 case AArch64::ANDWrr:
3167 return AArch64::ANDSWrr;
3168 case AArch64::ANDWrs:
3169 return AArch64::ANDSWrs;
3170 case AArch64::BICWrr:
3171 return AArch64::BICSWrr;
3172 case AArch64::BICWrs:
3173 return AArch64::BICSWrs;
3174 case AArch64::SUBWri:
3175 return AArch64::SUBSWri;
3176 case AArch64::SUBWrr:
3177 return AArch64::SUBSWrr;
3178 case AArch64::SUBWrs:
3179 return AArch64::SUBSWrs;
3180 case AArch64::SUBWrx:
3181 return AArch64::SUBSWrx;
3183 case AArch64::ADDXri:
3184 return AArch64::ADDSXri;
3185 case AArch64::ADDXrr:
3186 return AArch64::ADDSXrr;
3187 case AArch64::ADDXrs:
3188 return AArch64::ADDSXrs;
3189 case AArch64::ADDXrx:
3190 return AArch64::ADDSXrx;
3191 case AArch64::ANDXri:
3192 return AArch64::ANDSXri;
3193 case AArch64::ANDXrr:
3194 return AArch64::ANDSXrr;
3195 case AArch64::ANDXrs:
3196 return AArch64::ANDSXrs;
3197 case AArch64::BICXrr:
3198 return AArch64::BICSXrr;
3199 case AArch64::BICXrs:
3200 return AArch64::BICSXrs;
3201 case AArch64::SUBXri:
3202 return AArch64::SUBSXri;
3203 case AArch64::SUBXrr:
3204 return AArch64::SUBSXrr;
3205 case AArch64::SUBXrs:
3206 return AArch64::SUBSXrs;
3207 case AArch64::SUBXrx:
3208 return AArch64::SUBSXrx;
3210 case AArch64::AND_PPzPP:
3211 return AArch64::ANDS_PPzPP;
3212 case AArch64::BIC_PPzPP:
3213 return AArch64::BICS_PPzPP;
3214 case AArch64::EOR_PPzPP:
3215 return AArch64::EORS_PPzPP;
3216 case AArch64::NAND_PPzPP:
3217 return AArch64::NANDS_PPzPP;
3218 case AArch64::NOR_PPzPP:
3219 return AArch64::NORS_PPzPP;
3220 case AArch64::ORN_PPzPP:
3221 return AArch64::ORNS_PPzPP;
3222 case AArch64::ORR_PPzPP:
3223 return AArch64::ORRS_PPzPP;
3224 case AArch64::BRKA_PPzP:
3225 return AArch64::BRKAS_PPzP;
3226 case AArch64::BRKPA_PPzPP:
3227 return AArch64::BRKPAS_PPzPP;
3228 case AArch64::BRKB_PPzP:
3229 return AArch64::BRKBS_PPzP;
3230 case AArch64::BRKPB_PPzPP:
3231 return AArch64::BRKPBS_PPzPP;
3232 case AArch64::BRKN_PPzP:
3233 return AArch64::BRKNS_PPzP;
3234 case AArch64::RDFFR_PPz:
3235 return AArch64::RDFFRS_PPz;
3236 case AArch64::PTRUE_B:
3237 return AArch64::PTRUES_B;
3248 if (
MI.hasOrderedMemoryRef())
3253 assert((
MI.getOperand(IsPreLdSt ? 2 : 1).isReg() ||
3254 MI.getOperand(IsPreLdSt ? 2 : 1).isFI()) &&
3255 "Expected a reg or frame index operand.");
3259 bool IsImmPreLdSt = IsPreLdSt &&
MI.getOperand(3).isImm();
3261 if (!
MI.getOperand(2).isImm() && !IsImmPreLdSt)
3274 if (
MI.getOperand(1).isReg() && !IsPreLdSt) {
3275 Register BaseReg =
MI.getOperand(1).getReg();
3277 if (
MI.modifiesRegister(BaseReg,
TRI))
3283 switch (
MI.getOpcode()) {
3286 case AArch64::LDR_ZXI:
3287 case AArch64::STR_ZXI:
3288 if (!Subtarget.isLittleEndian() ||
3289 Subtarget.getSVEVectorSizeInBits() != 128)
3302 const MCAsmInfo *MAI =
MI.getMF()->getTarget().getMCAsmInfo();
3304 MI.getMF()->getFunction().needsUnwindTableEntry();
3310 if (Subtarget.isPaired128Slow()) {
3311 switch (
MI.getOpcode()) {
3314 case AArch64::LDURQi:
3315 case AArch64::STURQi:
3316 case AArch64::LDRQui:
3317 case AArch64::STRQui:
3344std::optional<ExtAddrMode>
3349 bool OffsetIsScalable;
3350 if (!getMemOperandWithOffset(MemI,
Base,
Offset, OffsetIsScalable,
TRI))
3351 return std::nullopt;
3354 return std::nullopt;
3369 int64_t OffsetScale = 1;
3374 case AArch64::LDURQi:
3375 case AArch64::STURQi:
3379 case AArch64::LDURDi:
3380 case AArch64::STURDi:
3381 case AArch64::LDURXi:
3382 case AArch64::STURXi:
3386 case AArch64::LDURWi:
3387 case AArch64::LDURSWi:
3388 case AArch64::STURWi:
3392 case AArch64::LDURHi:
3393 case AArch64::STURHi:
3394 case AArch64::LDURHHi:
3395 case AArch64::STURHHi:
3396 case AArch64::LDURSHXi:
3397 case AArch64::LDURSHWi:
3401 case AArch64::LDRBroX:
3402 case AArch64::LDRBBroX:
3403 case AArch64::LDRSBXroX:
3404 case AArch64::LDRSBWroX:
3405 case AArch64::STRBroX:
3406 case AArch64::STRBBroX:
3407 case AArch64::LDURBi:
3408 case AArch64::LDURBBi:
3409 case AArch64::LDURSBXi:
3410 case AArch64::LDURSBWi:
3411 case AArch64::STURBi:
3412 case AArch64::STURBBi:
3413 case AArch64::LDRBui:
3414 case AArch64::LDRBBui:
3415 case AArch64::LDRSBXui:
3416 case AArch64::LDRSBWui:
3417 case AArch64::STRBui:
3418 case AArch64::STRBBui:
3422 case AArch64::LDRQroX:
3423 case AArch64::STRQroX:
3424 case AArch64::LDRQui:
3425 case AArch64::STRQui:
3430 case AArch64::LDRDroX:
3431 case AArch64::STRDroX:
3432 case AArch64::LDRXroX:
3433 case AArch64::STRXroX:
3434 case AArch64::LDRDui:
3435 case AArch64::STRDui:
3436 case AArch64::LDRXui:
3437 case AArch64::STRXui:
3442 case AArch64::LDRWroX:
3443 case AArch64::LDRSWroX:
3444 case AArch64::STRWroX:
3445 case AArch64::LDRWui:
3446 case AArch64::LDRSWui:
3447 case AArch64::STRWui:
3452 case AArch64::LDRHroX:
3453 case AArch64::STRHroX:
3454 case AArch64::LDRHHroX:
3455 case AArch64::STRHHroX:
3456 case AArch64::LDRSHXroX:
3457 case AArch64::LDRSHWroX:
3458 case AArch64::LDRHui:
3459 case AArch64::STRHui:
3460 case AArch64::LDRHHui:
3461 case AArch64::STRHHui:
3462 case AArch64::LDRSHXui:
3463 case AArch64::LDRSHWui:
3471 if (BaseRegOp.
isReg() && BaseRegOp.
getReg() == Reg)
3495 case AArch64::SBFMXri:
3508 AM.
Scale = OffsetScale;
3513 case TargetOpcode::SUBREG_TO_REG: {
3526 if (!OffsetReg.
isVirtual() || !
MRI.hasOneNonDBGUse(OffsetReg))
3530 if (
DefMI.getOpcode() != AArch64::ORRWrs ||
3531 DefMI.getOperand(1).getReg() != AArch64::WZR ||
3532 DefMI.getOperand(3).getImm() != 0)
3539 AM.
Scale = OffsetScale;
3550 auto validateOffsetForLDP = [](
unsigned NumBytes, int64_t OldOffset,
3551 int64_t NewOffset) ->
bool {
3552 int64_t MinOffset, MaxOffset;
3569 return OldOffset < MinOffset || OldOffset > MaxOffset ||
3570 (NewOffset >= MinOffset && NewOffset <= MaxOffset);
3572 auto canFoldAddSubImmIntoAddrMode = [&](int64_t Disp) ->
bool {
3574 int64_t NewOffset = OldOffset + Disp;
3575 if (!isLegalAddressingMode(NumBytes, NewOffset, 0))
3579 if (!validateOffsetForLDP(NumBytes, OldOffset, NewOffset))
3589 auto canFoldAddRegIntoAddrMode =
3594 if ((
unsigned)Scale != Scale)
3596 if (!isLegalAddressingMode(NumBytes, 0, Scale))
3608 return (Opcode == AArch64::STURQi || Opcode == AArch64::STRQui) &&
3609 Subtarget.isSTRQroSlow();
3618 case AArch64::ADDXri:
3624 return canFoldAddSubImmIntoAddrMode(Disp);
3626 case AArch64::SUBXri:
3632 return canFoldAddSubImmIntoAddrMode(-Disp);
3634 case AArch64::ADDXrs: {
3647 if (Shift != 2 && Shift != 3 && Subtarget.hasAddrLSLSlow14())
3649 if (avoidSlowSTRQ(MemI))
3652 return canFoldAddRegIntoAddrMode(1ULL << Shift);
3655 case AArch64::ADDXrr:
3663 if (!OptSize && avoidSlowSTRQ(MemI))
3665 return canFoldAddRegIntoAddrMode(1);
3667 case AArch64::ADDXrx:
3675 if (!OptSize && avoidSlowSTRQ(MemI))
3684 return canFoldAddRegIntoAddrMode(
3699 case AArch64::LDURQi:
3700 case AArch64::LDRQui:
3701 return AArch64::LDRQroX;
3702 case AArch64::STURQi:
3703 case AArch64::STRQui:
3704 return AArch64::STRQroX;
3705 case AArch64::LDURDi:
3706 case AArch64::LDRDui:
3707 return AArch64::LDRDroX;
3708 case AArch64::STURDi:
3709 case AArch64::STRDui:
3710 return AArch64::STRDroX;
3711 case AArch64::LDURXi:
3712 case AArch64::LDRXui:
3713 return AArch64::LDRXroX;
3714 case AArch64::STURXi:
3715 case AArch64::STRXui:
3716 return AArch64::STRXroX;
3717 case AArch64::LDURWi:
3718 case AArch64::LDRWui:
3719 return AArch64::LDRWroX;
3720 case AArch64::LDURSWi:
3721 case AArch64::LDRSWui:
3722 return AArch64::LDRSWroX;
3723 case AArch64::STURWi:
3724 case AArch64::STRWui:
3725 return AArch64::STRWroX;
3726 case AArch64::LDURHi:
3727 case AArch64::LDRHui:
3728 return AArch64::LDRHroX;
3729 case AArch64::STURHi:
3730 case AArch64::STRHui:
3731 return AArch64::STRHroX;
3732 case AArch64::LDURHHi:
3733 case AArch64::LDRHHui:
3734 return AArch64::LDRHHroX;
3735 case AArch64::STURHHi:
3736 case AArch64::STRHHui:
3737 return AArch64::STRHHroX;
3738 case AArch64::LDURSHXi:
3739 case AArch64::LDRSHXui:
3740 return AArch64::LDRSHXroX;
3741 case AArch64::LDURSHWi:
3742 case AArch64::LDRSHWui:
3743 return AArch64::LDRSHWroX;
3744 case AArch64::LDURBi:
3745 case AArch64::LDRBui:
3746 return AArch64::LDRBroX;
3747 case AArch64::LDURBBi:
3748 case AArch64::LDRBBui:
3749 return AArch64::LDRBBroX;
3750 case AArch64::LDURSBXi:
3751 case AArch64::LDRSBXui:
3752 return AArch64::LDRSBXroX;
3753 case AArch64::LDURSBWi:
3754 case AArch64::LDRSBWui:
3755 return AArch64::LDRSBWroX;
3756 case AArch64::STURBi:
3757 case AArch64::STRBui:
3758 return AArch64::STRBroX;
3759 case AArch64::STURBBi:
3760 case AArch64::STRBBui:
3761 return AArch64::STRBBroX;
3773 case AArch64::LDURQi:
3775 return AArch64::LDRQui;
3776 case AArch64::STURQi:
3778 return AArch64::STRQui;
3779 case AArch64::LDURDi:
3781 return AArch64::LDRDui;
3782 case AArch64::STURDi:
3784 return AArch64::STRDui;
3785 case AArch64::LDURXi:
3787 return AArch64::LDRXui;
3788 case AArch64::STURXi:
3790 return AArch64::STRXui;
3791 case AArch64::LDURWi:
3793 return AArch64::LDRWui;
3794 case AArch64::LDURSWi:
3796 return AArch64::LDRSWui;
3797 case AArch64::STURWi:
3799 return AArch64::STRWui;
3800 case AArch64::LDURHi:
3802 return AArch64::LDRHui;
3803 case AArch64::STURHi:
3805 return AArch64::STRHui;
3806 case AArch64::LDURHHi:
3808 return AArch64::LDRHHui;
3809 case AArch64::STURHHi:
3811 return AArch64::STRHHui;
3812 case AArch64::LDURSHXi:
3814 return AArch64::LDRSHXui;
3815 case AArch64::LDURSHWi:
3817 return AArch64::LDRSHWui;
3818 case AArch64::LDURBi:
3820 return AArch64::LDRBui;
3821 case AArch64::LDURBBi:
3823 return AArch64::LDRBBui;
3824 case AArch64::LDURSBXi:
3826 return AArch64::LDRSBXui;
3827 case AArch64::LDURSBWi:
3829 return AArch64::LDRSBWui;
3830 case AArch64::STURBi:
3832 return AArch64::STRBui;
3833 case AArch64::STURBBi:
3835 return AArch64::STRBBui;
3836 case AArch64::LDRQui:
3837 case AArch64::STRQui:
3840 case AArch64::LDRDui:
3841 case AArch64::STRDui:
3842 case AArch64::LDRXui:
3843 case AArch64::STRXui:
3846 case AArch64::LDRWui:
3847 case AArch64::LDRSWui:
3848 case AArch64::STRWui:
3851 case AArch64::LDRHui:
3852 case AArch64::STRHui:
3853 case AArch64::LDRHHui:
3854 case AArch64::STRHHui:
3855 case AArch64::LDRSHXui:
3856 case AArch64::LDRSHWui:
3859 case AArch64::LDRBui:
3860 case AArch64::LDRBBui:
3861 case AArch64::LDRSBXui:
3862 case AArch64::LDRSBWui:
3863 case AArch64::STRBui:
3864 case AArch64::STRBBui:
3878 case AArch64::LDURQi:
3879 case AArch64::STURQi:
3880 case AArch64::LDURDi:
3881 case AArch64::STURDi:
3882 case AArch64::LDURXi:
3883 case AArch64::STURXi:
3884 case AArch64::LDURWi:
3885 case AArch64::LDURSWi:
3886 case AArch64::STURWi:
3887 case AArch64::LDURHi:
3888 case AArch64::STURHi:
3889 case AArch64::LDURHHi:
3890 case AArch64::STURHHi:
3891 case AArch64::LDURSHXi:
3892 case AArch64::LDURSHWi:
3893 case AArch64::LDURBi:
3894 case AArch64::STURBi:
3895 case AArch64::LDURBBi:
3896 case AArch64::STURBBi:
3897 case AArch64::LDURSBWi:
3898 case AArch64::LDURSBXi:
3900 case AArch64::LDRQui:
3901 return AArch64::LDURQi;
3902 case AArch64::STRQui:
3903 return AArch64::STURQi;
3904 case AArch64::LDRDui:
3905 return AArch64::LDURDi;
3906 case AArch64::STRDui:
3907 return AArch64::STURDi;
3908 case AArch64::LDRXui:
3909 return AArch64::LDURXi;
3910 case AArch64::STRXui:
3911 return AArch64::STURXi;
3912 case AArch64::LDRWui:
3913 return AArch64::LDURWi;
3914 case AArch64::LDRSWui:
3915 return AArch64::LDURSWi;
3916 case AArch64::STRWui:
3917 return AArch64::STURWi;
3918 case AArch64::LDRHui:
3919 return AArch64::LDURHi;
3920 case AArch64::STRHui:
3921 return AArch64::STURHi;
3922 case AArch64::LDRHHui:
3923 return AArch64::LDURHHi;
3924 case AArch64::STRHHui:
3925 return AArch64::STURHHi;
3926 case AArch64::LDRSHXui:
3927 return AArch64::LDURSHXi;
3928 case AArch64::LDRSHWui:
3929 return AArch64::LDURSHWi;
3930 case AArch64::LDRBBui:
3931 return AArch64::LDURBBi;
3932 case AArch64::LDRBui:
3933 return AArch64::LDURBi;
3934 case AArch64::STRBBui:
3935 return AArch64::STURBBi;
3936 case AArch64::STRBui:
3937 return AArch64::STURBi;
3938 case AArch64::LDRSBWui:
3939 return AArch64::LDURSBWi;
3940 case AArch64::LDRSBXui:
3941 return AArch64::LDURSBXi;
3954 case AArch64::LDRQroX:
3955 case AArch64::LDURQi:
3956 case AArch64::LDRQui:
3957 return AArch64::LDRQroW;
3958 case AArch64::STRQroX:
3959 case AArch64::STURQi:
3960 case AArch64::STRQui:
3961 return AArch64::STRQroW;
3962 case AArch64::LDRDroX:
3963 case AArch64::LDURDi:
3964 case AArch64::LDRDui:
3965 return AArch64::LDRDroW;
3966 case AArch64::STRDroX:
3967 case AArch64::STURDi:
3968 case AArch64::STRDui:
3969 return AArch64::STRDroW;
3970 case AArch64::LDRXroX:
3971 case AArch64::LDURXi:
3972 case AArch64::LDRXui:
3973 return AArch64::LDRXroW;
3974 case AArch64::STRXroX:
3975 case AArch64::STURXi:
3976 case AArch64::STRXui:
3977 return AArch64::STRXroW;
3978 case AArch64::LDRWroX:
3979 case AArch64::LDURWi:
3980 case AArch64::LDRWui:
3981 return AArch64::LDRWroW;
3982 case AArch64::LDRSWroX:
3983 case AArch64::LDURSWi:
3984 case AArch64::LDRSWui:
3985 return AArch64::LDRSWroW;
3986 case AArch64::STRWroX:
3987 case AArch64::STURWi:
3988 case AArch64::STRWui:
3989 return AArch64::STRWroW;
3990 case AArch64::LDRHroX:
3991 case AArch64::LDURHi:
3992 case AArch64::LDRHui:
3993 return AArch64::LDRHroW;
3994 case AArch64::STRHroX:
3995 case AArch64::STURHi:
3996 case AArch64::STRHui:
3997 return AArch64::STRHroW;
3998 case AArch64::LDRHHroX:
3999 case AArch64::LDURHHi:
4000 case AArch64::LDRHHui:
4001 return AArch64::LDRHHroW;
4002 case AArch64::STRHHroX:
4003 case AArch64::STURHHi:
4004 case AArch64::STRHHui:
4005 return AArch64::STRHHroW;
4006 case AArch64::LDRSHXroX:
4007 case AArch64::LDURSHXi:
4008 case AArch64::LDRSHXui:
4009 return AArch64::LDRSHXroW;
4010 case AArch64::LDRSHWroX:
4011 case AArch64::LDURSHWi:
4012 case AArch64::LDRSHWui:
4013 return AArch64::LDRSHWroW;
4014 case AArch64::LDRBroX:
4015 case AArch64::LDURBi:
4016 case AArch64::LDRBui:
4017 return AArch64::LDRBroW;
4018 case AArch64::LDRBBroX:
4019 case AArch64::LDURBBi:
4020 case AArch64::LDRBBui:
4021 return AArch64::LDRBBroW;
4022 case AArch64::LDRSBXroX:
4023 case AArch64::LDURSBXi:
4024 case AArch64::LDRSBXui:
4025 return AArch64::LDRSBXroW;
4026 case AArch64::LDRSBWroX:
4027 case AArch64::LDURSBWi:
4028 case AArch64::LDRSBWui:
4029 return AArch64::LDRSBWroW;
4030 case AArch64::STRBroX:
4031 case AArch64::STURBi:
4032 case AArch64::STRBui:
4033 return AArch64::STRBroW;
4034 case AArch64::STRBBroX:
4035 case AArch64::STURBBi:
4036 case AArch64::STRBBui:
4037 return AArch64::STRBBroW;
4052 MRI.constrainRegClass(AM.
BaseReg, &AArch64::GPR64spRegClass);
4062 return B.getInstr();
4066 "Addressing mode not supported for folding");
4083 return B.getInstr();
4090 "Address offset can be a register or an immediate, but not both");
4092 MRI.constrainRegClass(AM.
BaseReg, &AArch64::GPR64spRegClass);
4097 OffsetReg =
MRI.createVirtualRegister(&AArch64::GPR32RegClass);
4111 return B.getInstr();
4115 "Function must not be called with an addressing mode it can't handle");
4124 case AArch64::LD1Fourv16b_POST:
4125 case AArch64::LD1Fourv1d_POST:
4126 case AArch64::LD1Fourv2d_POST:
4127 case AArch64::LD1Fourv2s_POST:
4128 case AArch64::LD1Fourv4h_POST:
4129 case AArch64::LD1Fourv4s_POST:
4130 case AArch64::LD1Fourv8b_POST:
4131 case AArch64::LD1Fourv8h_POST:
4132 case AArch64::LD1Onev16b_POST:
4133 case AArch64::LD1Onev1d_POST:
4134 case AArch64::LD1Onev2d_POST:
4135 case AArch64::LD1Onev2s_POST:
4136 case AArch64::LD1Onev4h_POST:
4137 case AArch64::LD1Onev4s_POST:
4138 case AArch64::LD1Onev8b_POST:
4139 case AArch64::LD1Onev8h_POST:
4140 case AArch64::LD1Rv16b_POST:
4141 case AArch64::LD1Rv1d_POST:
4142 case AArch64::LD1Rv2d_POST:
4143 case AArch64::LD1Rv2s_POST:
4144 case AArch64::LD1Rv4h_POST:
4145 case AArch64::LD1Rv4s_POST:
4146 case AArch64::LD1Rv8b_POST:
4147 case AArch64::LD1Rv8h_POST:
4148 case AArch64::LD1Threev16b_POST:
4149 case AArch64::LD1Threev1d_POST:
4150 case AArch64::LD1Threev2d_POST:
4151 case AArch64::LD1Threev2s_POST:
4152 case AArch64::LD1Threev4h_POST:
4153 case AArch64::LD1Threev4s_POST:
4154 case AArch64::LD1Threev8b_POST:
4155 case AArch64::LD1Threev8h_POST:
4156 case AArch64::LD1Twov16b_POST:
4157 case AArch64::LD1Twov1d_POST:
4158 case AArch64::LD1Twov2d_POST:
4159 case AArch64::LD1Twov2s_POST:
4160 case AArch64::LD1Twov4h_POST:
4161 case AArch64::LD1Twov4s_POST:
4162 case AArch64::LD1Twov8b_POST:
4163 case AArch64::LD1Twov8h_POST:
4164 case AArch64::LD1i16_POST:
4165 case AArch64::LD1i32_POST:
4166 case AArch64::LD1i64_POST:
4167 case AArch64::LD1i8_POST:
4168 case AArch64::LD2Rv16b_POST:
4169 case AArch64::LD2Rv1d_POST:
4170 case AArch64::LD2Rv2d_POST:
4171 case AArch64::LD2Rv2s_POST:
4172 case AArch64::LD2Rv4h_POST:
4173 case AArch64::LD2Rv4s_POST:
4174 case AArch64::LD2Rv8b_POST:
4175 case AArch64::LD2Rv8h_POST:
4176 case AArch64::LD2Twov16b_POST:
4177 case AArch64::LD2Twov2d_POST:
4178 case AArch64::LD2Twov2s_POST:
4179 case AArch64::LD2Twov4h_POST:
4180 case AArch64::LD2Twov4s_POST:
4181 case AArch64::LD2Twov8b_POST:
4182 case AArch64::LD2Twov8h_POST:
4183 case AArch64::LD2i16_POST:
4184 case AArch64::LD2i32_POST:
4185 case AArch64::LD2i64_POST:
4186 case AArch64::LD2i8_POST:
4187 case AArch64::LD3Rv16b_POST:
4188 case AArch64::LD3Rv1d_POST:
4189 case AArch64::LD3Rv2d_POST:
4190 case AArch64::LD3Rv2s_POST:
4191 case AArch64::LD3Rv4h_POST:
4192 case AArch64::LD3Rv4s_POST:
4193 case AArch64::LD3Rv8b_POST:
4194 case AArch64::LD3Rv8h_POST:
4195 case AArch64::LD3Threev16b_POST:
4196 case AArch64::LD3Threev2d_POST:
4197 case AArch64::LD3Threev2s_POST:
4198 case AArch64::LD3Threev4h_POST:
4199 case AArch64::LD3Threev4s_POST:
4200 case AArch64::LD3Threev8b_POST:
4201 case AArch64::LD3Threev8h_POST:
4202 case AArch64::LD3i16_POST:
4203 case AArch64::LD3i32_POST:
4204 case AArch64::LD3i64_POST:
4205 case AArch64::LD3i8_POST:
4206 case AArch64::LD4Fourv16b_POST:
4207 case AArch64::LD4Fourv2d_POST:
4208 case AArch64::LD4Fourv2s_POST:
4209 case AArch64::LD4Fourv4h_POST:
4210 case AArch64::LD4Fourv4s_POST:
4211 case AArch64::LD4Fourv8b_POST:
4212 case AArch64::LD4Fourv8h_POST:
4213 case AArch64::LD4Rv16b_POST:
4214 case AArch64::LD4Rv1d_POST:
4215 case AArch64::LD4Rv2d_POST:
4216 case AArch64::LD4Rv2s_POST:
4217 case AArch64::LD4Rv4h_POST:
4218 case AArch64::LD4Rv4s_POST:
4219 case AArch64::LD4Rv8b_POST:
4220 case AArch64::LD4Rv8h_POST:
4221 case AArch64::LD4i16_POST:
4222 case AArch64::LD4i32_POST:
4223 case AArch64::LD4i64_POST:
4224 case AArch64::LD4i8_POST:
4225 case AArch64::LDAPRWpost:
4226 case AArch64::LDAPRXpost:
4227 case AArch64::LDIAPPWpost:
4228 case AArch64::LDIAPPXpost:
4229 case AArch64::LDPDpost:
4230 case AArch64::LDPQpost:
4231 case AArch64::LDPSWpost:
4232 case AArch64::LDPSpost:
4233 case AArch64::LDPWpost:
4234 case AArch64::LDPXpost:
4235 case AArch64::LDRBBpost:
4236 case AArch64::LDRBpost:
4237 case AArch64::LDRDpost:
4238 case AArch64::LDRHHpost:
4239 case AArch64::LDRHpost:
4240 case AArch64::LDRQpost:
4241 case AArch64::LDRSBWpost:
4242 case AArch64::LDRSBXpost:
4243 case AArch64::LDRSHWpost:
4244 case AArch64::LDRSHXpost:
4245 case AArch64::LDRSWpost:
4246 case AArch64::LDRSpost:
4247 case AArch64::LDRWpost:
4248 case AArch64::LDRXpost:
4249 case AArch64::ST1Fourv16b_POST:
4250 case AArch64::ST1Fourv1d_POST:
4251 case AArch64::ST1Fourv2d_POST:
4252 case AArch64::ST1Fourv2s_POST:
4253 case AArch64::ST1Fourv4h_POST:
4254 case AArch64::ST1Fourv4s_POST:
4255 case AArch64::ST1Fourv8b_POST:
4256 case AArch64::ST1Fourv8h_POST:
4257 case AArch64::ST1Onev16b_POST:
4258 case AArch64::ST1Onev1d_POST:
4259 case AArch64::ST1Onev2d_POST:
4260 case AArch64::ST1Onev2s_POST:
4261 case AArch64::ST1Onev4h_POST:
4262 case AArch64::ST1Onev4s_POST:
4263 case AArch64::ST1Onev8b_POST:
4264 case AArch64::ST1Onev8h_POST:
4265 case AArch64::ST1Threev16b_POST:
4266 case AArch64::ST1Threev1d_POST:
4267 case AArch64::ST1Threev2d_POST:
4268 case AArch64::ST1Threev2s_POST:
4269 case AArch64::ST1Threev4h_POST:
4270 case AArch64::ST1Threev4s_POST:
4271 case AArch64::ST1Threev8b_POST:
4272 case AArch64::ST1Threev8h_POST:
4273 case AArch64::ST1Twov16b_POST:
4274 case AArch64::ST1Twov1d_POST:
4275 case AArch64::ST1Twov2d_POST:
4276 case AArch64::ST1Twov2s_POST:
4277 case AArch64::ST1Twov4h_POST:
4278 case AArch64::ST1Twov4s_POST:
4279 case AArch64::ST1Twov8b_POST:
4280 case AArch64::ST1Twov8h_POST:
4281 case AArch64::ST1i16_POST:
4282 case AArch64::ST1i32_POST:
4283 case AArch64::ST1i64_POST:
4284 case AArch64::ST1i8_POST:
4285 case AArch64::ST2GPostIndex:
4286 case AArch64::ST2Twov16b_POST:
4287 case AArch64::ST2Twov2d_POST:
4288 case AArch64::ST2Twov2s_POST:
4289 case AArch64::ST2Twov4h_POST:
4290 case AArch64::ST2Twov4s_POST:
4291 case AArch64::ST2Twov8b_POST:
4292 case AArch64::ST2Twov8h_POST:
4293 case AArch64::ST2i16_POST:
4294 case AArch64::ST2i32_POST:
4295 case AArch64::ST2i64_POST:
4296 case AArch64::ST2i8_POST:
4297 case AArch64::ST3Threev16b_POST:
4298 case AArch64::ST3Threev2d_POST:
4299 case AArch64::ST3Threev2s_POST:
4300 case AArch64::ST3Threev4h_POST:
4301 case AArch64::ST3Threev4s_POST:
4302 case AArch64::ST3Threev8b_POST:
4303 case AArch64::ST3Threev8h_POST:
4304 case AArch64::ST3i16_POST:
4305 case AArch64::ST3i32_POST:
4306 case AArch64::ST3i64_POST:
4307 case AArch64::ST3i8_POST:
4308 case AArch64::ST4Fourv16b_POST:
4309 case AArch64::ST4Fourv2d_POST:
4310 case AArch64::ST4Fourv2s_POST:
4311 case AArch64::ST4Fourv4h_POST:
4312 case AArch64::ST4Fourv4s_POST:
4313 case AArch64::ST4Fourv8b_POST:
4314 case AArch64::ST4Fourv8h_POST:
4315 case AArch64::ST4i16_POST:
4316 case AArch64::ST4i32_POST:
4317 case AArch64::ST4i64_POST:
4318 case AArch64::ST4i8_POST:
4319 case AArch64::STGPostIndex:
4320 case AArch64::STGPpost:
4321 case AArch64::STPDpost:
4322 case AArch64::STPQpost:
4323 case AArch64::STPSpost:
4324 case AArch64::STPWpost:
4325 case AArch64::STPXpost:
4326 case AArch64::STRBBpost:
4327 case AArch64::STRBpost:
4328 case AArch64::STRDpost:
4329 case AArch64::STRHHpost:
4330 case AArch64::STRHpost:
4331 case AArch64::STRQpost:
4332 case AArch64::STRSpost:
4333 case AArch64::STRWpost:
4334 case AArch64::STRXpost:
4335 case AArch64::STZ2GPostIndex:
4336 case AArch64::STZGPostIndex:
4343 bool &OffsetIsScalable,
TypeSize &Width,
4364 int64_t Dummy1, Dummy2;
4386 return BaseOp->
isReg() || BaseOp->
isFI();
4393 assert(OfsOp.
isImm() &&
"Offset operand wasn't immediate.");
4398 TypeSize &Width, int64_t &MinOffset,
4399 int64_t &MaxOffset) {
4405 MinOffset = MaxOffset = 0;
4408 case AArch64::LDRQui:
4409 case AArch64::STRQui:
4415 case AArch64::LDRXui:
4416 case AArch64::LDRDui:
4417 case AArch64::STRXui:
4418 case AArch64::STRDui:
4419 case AArch64::PRFMui:
4425 case AArch64::LDRWui:
4426 case AArch64::LDRSui:
4427 case AArch64::LDRSWui:
4428 case AArch64::STRWui:
4429 case AArch64::STRSui:
4435 case AArch64::LDRHui:
4436 case AArch64::LDRHHui:
4437 case AArch64::LDRSHWui:
4438 case AArch64::LDRSHXui:
4439 case AArch64::STRHui:
4440 case AArch64::STRHHui:
4446 case AArch64::LDRBui:
4447 case AArch64::LDRBBui:
4448 case AArch64::LDRSBWui:
4449 case AArch64::LDRSBXui:
4450 case AArch64::STRBui:
4451 case AArch64::STRBBui:
4458 case AArch64::STRQpre:
4459 case AArch64::LDRQpost:
4465 case AArch64::LDRDpost:
4466 case AArch64::LDRDpre:
4467 case AArch64::LDRXpost:
4468 case AArch64::LDRXpre:
4469 case AArch64::STRDpost:
4470 case AArch64::STRDpre:
4471 case AArch64::STRXpost:
4472 case AArch64::STRXpre:
4478 case AArch64::STRWpost:
4479 case AArch64::STRWpre:
4480 case AArch64::LDRWpost:
4481 case AArch64::LDRWpre:
4482 case AArch64::STRSpost:
4483 case AArch64::STRSpre:
4484 case AArch64::LDRSpost:
4485 case AArch64::LDRSpre:
4491 case AArch64::LDRHpost:
4492 case AArch64::LDRHpre:
4493 case AArch64::STRHpost:
4494 case AArch64::STRHpre:
4495 case AArch64::LDRHHpost:
4496 case AArch64::LDRHHpre:
4497 case AArch64::STRHHpost:
4498 case AArch64::STRHHpre:
4504 case AArch64::LDRBpost:
4505 case AArch64::LDRBpre:
4506 case AArch64::STRBpost:
4507 case AArch64::STRBpre:
4508 case AArch64::LDRBBpost:
4509 case AArch64::LDRBBpre:
4510 case AArch64::STRBBpost:
4511 case AArch64::STRBBpre:
4518 case AArch64::LDURQi:
4519 case AArch64::STURQi:
4525 case AArch64::LDURXi:
4526 case AArch64::LDURDi:
4527 case AArch64::LDAPURXi:
4528 case AArch64::STURXi:
4529 case AArch64::STURDi:
4530 case AArch64::STLURXi:
4531 case AArch64::PRFUMi:
4537 case AArch64::LDURWi:
4538 case AArch64::LDURSi:
4539 case AArch64::LDURSWi:
4540 case AArch64::LDAPURi:
4541 case AArch64::LDAPURSWi:
4542 case AArch64::STURWi:
4543 case AArch64::STURSi:
4544 case AArch64::STLURWi:
4550 case AArch64::LDURHi:
4551 case AArch64::LDURHHi:
4552 case AArch64::LDURSHXi:
4553 case AArch64::LDURSHWi:
4554 case AArch64::LDAPURHi:
4555 case AArch64::LDAPURSHWi:
4556 case AArch64::LDAPURSHXi:
4557 case AArch64::STURHi:
4558 case AArch64::STURHHi:
4559 case AArch64::STLURHi:
4565 case AArch64::LDURBi:
4566 case AArch64::LDURBBi:
4567 case AArch64::LDURSBXi:
4568 case AArch64::LDURSBWi:
4569 case AArch64::LDAPURBi:
4570 case AArch64::LDAPURSBWi:
4571 case AArch64::LDAPURSBXi:
4572 case AArch64::STURBi:
4573 case AArch64::STURBBi:
4574 case AArch64::STLURBi:
4581 case AArch64::LDPQi:
4582 case AArch64::LDNPQi:
4583 case AArch64::STPQi:
4584 case AArch64::STNPQi:
4585 case AArch64::LDPQpost:
4586 case AArch64::LDPQpre:
4587 case AArch64::STPQpost:
4588 case AArch64::STPQpre:
4594 case AArch64::LDPXi:
4595 case AArch64::LDPDi:
4596 case AArch64::LDNPXi:
4597 case AArch64::LDNPDi:
4598 case AArch64::STPXi:
4599 case AArch64::STPDi:
4600 case AArch64::STNPXi:
4601 case AArch64::STNPDi:
4602 case AArch64::LDPDpost:
4603 case AArch64::LDPDpre:
4604 case AArch64::LDPXpost:
4605 case AArch64::LDPXpre:
4606 case AArch64::STPDpost:
4607 case AArch64::STPDpre:
4608 case AArch64::STPXpost:
4609 case AArch64::STPXpre:
4615 case AArch64::LDPWi:
4616 case AArch64::LDPSi:
4617 case AArch64::LDNPWi:
4618 case AArch64::LDNPSi:
4619 case AArch64::STPWi:
4620 case AArch64::STPSi:
4621 case AArch64::STNPWi:
4622 case AArch64::STNPSi:
4623 case AArch64::LDPSpost:
4624 case AArch64::LDPSpre:
4625 case AArch64::LDPWpost:
4626 case AArch64::LDPWpre:
4627 case AArch64::STPSpost:
4628 case AArch64::STPSpre:
4629 case AArch64::STPWpost:
4630 case AArch64::STPWpre:
4636 case AArch64::StoreSwiftAsyncContext:
4649 case AArch64::TAGPstack:
4659 case AArch64::STGPreIndex:
4660 case AArch64::STGPostIndex:
4661 case AArch64::STZGi:
4662 case AArch64::STZGPreIndex:
4663 case AArch64::STZGPostIndex:
4670 case AArch64::STR_ZZZZXI:
4671 case AArch64::STR_ZZZZXI_STRIDED_CONTIGUOUS:
4672 case AArch64::LDR_ZZZZXI:
4673 case AArch64::LDR_ZZZZXI_STRIDED_CONTIGUOUS:
4679 case AArch64::STR_ZZZXI:
4680 case AArch64::LDR_ZZZXI:
4686 case AArch64::STR_ZZXI:
4687 case AArch64::STR_ZZXI_STRIDED_CONTIGUOUS:
4688 case AArch64::LDR_ZZXI:
4689 case AArch64::LDR_ZZXI_STRIDED_CONTIGUOUS:
4695 case AArch64::LDR_PXI:
4696 case AArch64::STR_PXI:
4702 case AArch64::LDR_PPXI:
4703 case AArch64::STR_PPXI:
4709 case AArch64::LDR_ZXI:
4710 case AArch64::STR_ZXI:
4716 case AArch64::LD1B_IMM:
4717 case AArch64::LD1H_IMM:
4718 case AArch64::LD1W_IMM:
4719 case AArch64::LD1D_IMM:
4720 case AArch64::LDNT1B_ZRI:
4721 case AArch64::LDNT1H_ZRI:
4722 case AArch64::LDNT1W_ZRI:
4723 case AArch64::LDNT1D_ZRI:
4724 case AArch64::ST1B_IMM:
4725 case AArch64::ST1H_IMM:
4726 case AArch64::ST1W_IMM:
4727 case AArch64::ST1D_IMM:
4728 case AArch64::STNT1B_ZRI:
4729 case AArch64::STNT1H_ZRI:
4730 case AArch64::STNT1W_ZRI:
4731 case AArch64::STNT1D_ZRI:
4732 case AArch64::LDNF1B_IMM:
4733 case AArch64::LDNF1H_IMM:
4734 case AArch64::LDNF1W_IMM:
4735 case AArch64::LDNF1D_IMM:
4743 case AArch64::LD2B_IMM:
4744 case AArch64::LD2H_IMM:
4745 case AArch64::LD2W_IMM:
4746 case AArch64::LD2D_IMM:
4747 case AArch64::ST2B_IMM:
4748 case AArch64::ST2H_IMM:
4749 case AArch64::ST2W_IMM:
4750 case AArch64::ST2D_IMM:
4756 case AArch64::LD3B_IMM:
4757 case AArch64::LD3H_IMM:
4758 case AArch64::LD3W_IMM:
4759 case AArch64::LD3D_IMM:
4760 case AArch64::ST3B_IMM:
4761 case AArch64::ST3H_IMM:
4762 case AArch64::ST3W_IMM:
4763 case AArch64::ST3D_IMM:
4769 case AArch64::LD4B_IMM:
4770 case AArch64::LD4H_IMM:
4771 case AArch64::LD4W_IMM:
4772 case AArch64::LD4D_IMM:
4773 case AArch64::ST4B_IMM:
4774 case AArch64::ST4H_IMM:
4775 case AArch64::ST4W_IMM:
4776 case AArch64::ST4D_IMM:
4782 case AArch64::LD1B_H_IMM:
4783 case AArch64::LD1SB_H_IMM:
4784 case AArch64::LD1H_S_IMM:
4785 case AArch64::LD1SH_S_IMM:
4786 case AArch64::LD1W_D_IMM:
4787 case AArch64::LD1SW_D_IMM:
4788 case AArch64::ST1B_H_IMM:
4789 case AArch64::ST1H_S_IMM:
4790 case AArch64::ST1W_D_IMM:
4791 case AArch64::LDNF1B_H_IMM:
4792 case AArch64::LDNF1SB_H_IMM:
4793 case AArch64::LDNF1H_S_IMM:
4794 case AArch64::LDNF1SH_S_IMM:
4795 case AArch64::LDNF1W_D_IMM:
4796 case AArch64::LDNF1SW_D_IMM:
4804 case AArch64::LD1B_S_IMM:
4805 case AArch64::LD1SB_S_IMM:
4806 case AArch64::LD1H_D_IMM:
4807 case AArch64::LD1SH_D_IMM:
4808 case AArch64::ST1B_S_IMM:
4809 case AArch64::ST1H_D_IMM:
4810 case AArch64::LDNF1B_S_IMM:
4811 case AArch64::LDNF1SB_S_IMM:
4812 case AArch64::LDNF1H_D_IMM:
4813 case AArch64::LDNF1SH_D_IMM:
4821 case AArch64::LD1B_D_IMM:
4822 case AArch64::LD1SB_D_IMM:
4823 case AArch64::ST1B_D_IMM:
4824 case AArch64::LDNF1B_D_IMM:
4825 case AArch64::LDNF1SB_D_IMM:
4833 case AArch64::ST2Gi:
4834 case AArch64::ST2GPreIndex:
4835 case AArch64::ST2GPostIndex:
4836 case AArch64::STZ2Gi:
4837 case AArch64::STZ2GPreIndex:
4838 case AArch64::STZ2GPostIndex:
4844 case AArch64::STGPi:
4845 case AArch64::STGPpost:
4846 case AArch64::STGPpre:
4852 case AArch64::LD1RB_IMM:
4853 case AArch64::LD1RB_H_IMM:
4854 case AArch64::LD1RB_S_IMM:
4855 case AArch64::LD1RB_D_IMM:
4856 case AArch64::LD1RSB_H_IMM:
4857 case AArch64::LD1RSB_S_IMM:
4858 case AArch64::LD1RSB_D_IMM:
4864 case AArch64::LD1RH_IMM:
4865 case AArch64::LD1RH_S_IMM:
4866 case AArch64::LD1RH_D_IMM:
4867 case AArch64::LD1RSH_S_IMM:
4868 case AArch64::LD1RSH_D_IMM:
4874 case AArch64::LD1RW_IMM:
4875 case AArch64::LD1RW_D_IMM:
4876 case AArch64::LD1RSW_IMM:
4882 case AArch64::LD1RD_IMM:
4898 case AArch64::LDRBBui:
4899 case AArch64::LDURBBi:
4900 case AArch64::LDRSBWui:
4901 case AArch64::LDURSBWi:
4902 case AArch64::STRBBui:
4903 case AArch64::STURBBi:
4905 case AArch64::LDRHHui:
4906 case AArch64::LDURHHi:
4907 case AArch64::LDRSHWui:
4908 case AArch64::LDURSHWi:
4909 case AArch64::STRHHui:
4910 case AArch64::STURHHi:
4912 case AArch64::LDRSui:
4913 case AArch64::LDURSi:
4914 case AArch64::LDRSpre:
4915 case AArch64::LDRSWui:
4916 case AArch64::LDURSWi:
4917 case AArch64::LDRSWpre:
4918 case AArch64::LDRWpre:
4919 case AArch64::LDRWui:
4920 case AArch64::LDURWi:
4921 case AArch64::STRSui:
4922 case AArch64::STURSi:
4923 case AArch64::STRSpre:
4924 case AArch64::STRWui:
4925 case AArch64::STURWi:
4926 case AArch64::STRWpre:
4927 case AArch64::LDPSi:
4928 case AArch64::LDPSWi:
4929 case AArch64::LDPWi:
4930 case AArch64::STPSi:
4931 case AArch64::STPWi:
4933 case AArch64::LDRDui:
4934 case AArch64::LDURDi:
4935 case AArch64::LDRDpre:
4936 case AArch64::LDRXui:
4937 case AArch64::LDURXi:
4938 case AArch64::LDRXpre:
4939 case AArch64::STRDui:
4940 case AArch64::STURDi:
4941 case AArch64::STRDpre:
4942 case AArch64::STRXui:
4943 case AArch64::STURXi:
4944 case AArch64::STRXpre:
4945 case AArch64::LDPDi:
4946 case AArch64::LDPXi:
4947 case AArch64::STPDi:
4948 case AArch64::STPXi:
4950 case AArch64::LDRQui:
4951 case AArch64::LDURQi:
4952 case AArch64::STRQui:
4953 case AArch64::STURQi:
4954 case AArch64::STRQpre:
4955 case AArch64::LDPQi:
4956 case AArch64::LDRQpre:
4957 case AArch64::STPQi:
4959 case AArch64::STZGi:
4960 case AArch64::ST2Gi:
4961 case AArch64::STZ2Gi:
4962 case AArch64::STGPi:
4968 switch (
MI.getOpcode()) {
4971 case AArch64::LDRWpre:
4972 case AArch64::LDRXpre:
4973 case AArch64::LDRSWpre:
4974 case AArch64::LDRSpre:
4975 case AArch64::LDRDpre:
4976 case AArch64::LDRQpre:
4982 switch (
MI.getOpcode()) {
4985 case AArch64::STRWpre:
4986 case AArch64::STRXpre:
4987 case AArch64::STRSpre:
4988 case AArch64::STRDpre:
4989 case AArch64::STRQpre:
4999 switch (
MI.getOpcode()) {
5002 case AArch64::LDPSi:
5003 case AArch64::LDPSWi:
5004 case AArch64::LDPDi:
5005 case AArch64::LDPQi:
5006 case AArch64::LDPWi:
5007 case AArch64::LDPXi:
5008 case AArch64::STPSi:
5009 case AArch64::STPDi:
5010 case AArch64::STPQi:
5011 case AArch64::STPWi:
5012 case AArch64::STPXi:
5013 case AArch64::STGPi:
5019 assert(
MI.mayLoadOrStore() &&
"Load or store instruction expected");
5023 return MI.getOperand(Idx);
5028 assert(
MI.mayLoadOrStore() &&
"Load or store instruction expected");
5032 return MI.getOperand(Idx);
5037 switch (
MI.getOpcode()) {
5040 case AArch64::LDRBroX:
5041 case AArch64::LDRBBroX:
5042 case AArch64::LDRSBXroX:
5043 case AArch64::LDRSBWroX:
5044 case AArch64::LDRHroX:
5045 case AArch64::LDRHHroX:
5046 case AArch64::LDRSHXroX:
5047 case AArch64::LDRSHWroX:
5048 case AArch64::LDRWroX:
5049 case AArch64::LDRSroX:
5050 case AArch64::LDRSWroX:
5051 case AArch64::LDRDroX:
5052 case AArch64::LDRXroX:
5053 case AArch64::LDRQroX:
5054 return MI.getOperand(4);
5060 if (
MI.getParent() ==
nullptr)
5070 auto Reg =
Op.getReg();
5071 if (Reg.isPhysical())
5072 return AArch64::FPR16RegClass.contains(Reg);
5074 return TRC == &AArch64::FPR16RegClass ||
5075 TRC == &AArch64::FPR16_loRegClass;
5084 auto Reg =
Op.getReg();
5085 if (Reg.isPhysical())
5086 return AArch64::FPR128RegClass.contains(Reg);
5088 return TRC == &AArch64::FPR128RegClass ||
5089 TRC == &AArch64::FPR128_loRegClass;
5095 switch (
MI.getOpcode()) {
5098 case AArch64::PACIASP:
5099 case AArch64::PACIBSP:
5102 case AArch64::PAUTH_PROLOGUE:
5105 case AArch64::HINT: {
5106 unsigned Imm =
MI.getOperand(0).getImm();
5108 if (Imm == 32 || Imm == 34 || Imm == 36 || Imm == 38)
5111 if (Imm == 25 || Imm == 27)
5123 assert(Reg.isPhysical() &&
"Expected physical register in isFpOrNEON");
5124 return AArch64::FPR128RegClass.contains(Reg) ||
5125 AArch64::FPR64RegClass.contains(Reg) ||
5126 AArch64::FPR32RegClass.contains(Reg) ||
5127 AArch64::FPR16RegClass.contains(Reg) ||
5128 AArch64::FPR8RegClass.contains(Reg);
5135 auto Reg =
Op.getReg();
5136 if (Reg.isPhysical())
5140 return TRC == &AArch64::FPR128RegClass ||
5141 TRC == &AArch64::FPR128_loRegClass ||
5142 TRC == &AArch64::FPR64RegClass ||
5143 TRC == &AArch64::FPR64_loRegClass ||
5144 TRC == &AArch64::FPR32RegClass || TRC == &AArch64::FPR16RegClass ||
5145 TRC == &AArch64::FPR8RegClass;
5167 if (FirstOpc == SecondOpc)
5173 case AArch64::STRSui:
5174 case AArch64::STURSi:
5175 return SecondOpc == AArch64::STRSui || SecondOpc == AArch64::STURSi;
5176 case AArch64::STRDui:
5177 case AArch64::STURDi:
5178 return SecondOpc == AArch64::STRDui || SecondOpc == AArch64::STURDi;
5179 case AArch64::STRQui:
5180 case AArch64::STURQi:
5181 return SecondOpc == AArch64::STRQui || SecondOpc == AArch64::STURQi;
5182 case AArch64::STRWui:
5183 case AArch64::STURWi:
5184 return SecondOpc == AArch64::STRWui || SecondOpc == AArch64::STURWi;
5185 case AArch64::STRXui:
5186 case AArch64::STURXi:
5187 return SecondOpc == AArch64::STRXui || SecondOpc == AArch64::STURXi;
5188 case AArch64::LDRSui:
5189 case AArch64::LDURSi:
5190 return SecondOpc == AArch64::LDRSui || SecondOpc == AArch64::LDURSi;
5191 case AArch64::LDRDui:
5192 case AArch64::LDURDi:
5193 return SecondOpc == AArch64::LDRDui || SecondOpc == AArch64::LDURDi;
5194 case AArch64::LDRQui:
5195 case AArch64::LDURQi:
5196 return SecondOpc == AArch64::LDRQui || SecondOpc == AArch64::LDURQi;
5197 case AArch64::LDRWui:
5198 case AArch64::LDURWi:
5199 return SecondOpc == AArch64::LDRSWui || SecondOpc == AArch64::LDURSWi;
5200 case AArch64::LDRSWui:
5201 case AArch64::LDURSWi:
5202 return SecondOpc == AArch64::LDRWui || SecondOpc == AArch64::LDURWi;
5203 case AArch64::LDRXui:
5204 case AArch64::LDURXi:
5205 return SecondOpc == AArch64::LDRXui || SecondOpc == AArch64::LDURXi;
5212 int64_t Offset1,
unsigned Opcode1,
int FI2,
5213 int64_t Offset2,
unsigned Opcode2) {
5219 assert(ObjectOffset1 <= ObjectOffset2 &&
"Object offsets are not ordered.");
5222 if (ObjectOffset1 % Scale1 != 0)
5224 ObjectOffset1 /= Scale1;
5226 if (ObjectOffset2 % Scale2 != 0)
5228 ObjectOffset2 /= Scale2;
5229 ObjectOffset1 += Offset1;
5230 ObjectOffset2 += Offset2;
5231 return ObjectOffset1 + 1 == ObjectOffset2;
5243 int64_t OpOffset2,
bool OffsetIsScalable2,
unsigned ClusterSize,
5244 unsigned NumBytes)
const {
5254 "Only base registers and frame indices are supported.");
5261 if (ClusterSize > 2)
5268 unsigned FirstOpc = FirstLdSt.
getOpcode();
5269 unsigned SecondOpc = SecondLdSt.
getOpcode();
5289 if (Offset1 > 63 || Offset1 < -64)
5294 if (BaseOp1.
isFI()) {
5296 "Caller should have ordered offsets.");
5301 BaseOp2.
getIndex(), Offset2, SecondOpc);
5304 assert(Offset1 <= Offset2 &&
"Caller should have ordered offsets.");
5306 return Offset1 + 1 == Offset2;
5316 if (
Reg.isPhysical())
5325 return ((DestReg - SrcReg) & 0x1f) < NumRegs;
5334 assert(Subtarget.hasNEON() &&
"Unexpected register copy without NEON");
5336 uint16_t DestEncoding =
TRI->getEncodingValue(DestReg);
5337 uint16_t SrcEncoding =
TRI->getEncodingValue(SrcReg);
5338 unsigned NumRegs = Indices.
size();
5340 int SubReg = 0, End = NumRegs, Incr = 1;
5359 unsigned Opcode,
unsigned ZeroReg,
5362 unsigned NumRegs = Indices.
size();
5365 uint16_t DestEncoding =
TRI->getEncodingValue(DestReg);
5366 uint16_t SrcEncoding =
TRI->getEncodingValue(SrcReg);
5367 assert(DestEncoding % NumRegs == 0 && SrcEncoding % NumRegs == 0 &&
5368 "GPR reg sequences should not be able to overlap");
5385 bool RenamableSrc)
const {
5386 if (AArch64::GPR32spRegClass.
contains(DestReg) &&
5387 AArch64::GPR32spRegClass.
contains(SrcReg)) {
5388 if (DestReg == AArch64::WSP || SrcReg == AArch64::WSP) {
5390 if (Subtarget.hasZeroCycleRegMoveGPR64() &&
5391 !Subtarget.hasZeroCycleRegMoveGPR32()) {
5393 MCRegister DestRegX = RI.getMatchingSuperReg(DestReg, AArch64::sub_32,
5394 &AArch64::GPR64spRegClass);
5395 MCRegister SrcRegX = RI.getMatchingSuperReg(SrcReg, AArch64::sub_32,
5396 &AArch64::GPR64spRegClass);
5412 }
else if (Subtarget.hasZeroCycleRegMoveGPR64() &&
5413 !Subtarget.hasZeroCycleRegMoveGPR32()) {
5415 MCRegister DestRegX = RI.getMatchingSuperReg(DestReg, AArch64::sub_32,
5416 &AArch64::GPR64spRegClass);
5417 assert(DestRegX.
isValid() &&
"Destination super-reg not valid");
5418 MCRegister SrcRegX = RI.getMatchingSuperReg(SrcReg, AArch64::sub_32,
5419 &AArch64::GPR64spRegClass);
5439 if (AArch64::GPR32spRegClass.
contains(DestReg) && SrcReg == AArch64::WZR) {
5440 if (Subtarget.hasZeroCycleZeroingGPR64() &&
5441 !Subtarget.hasZeroCycleZeroingGPR32()) {
5442 MCRegister DestRegX = RI.getMatchingSuperReg(DestReg, AArch64::sub_32,
5443 &AArch64::GPR64spRegClass);
5444 assert(DestRegX.
isValid() &&
"Destination super-reg not valid");
5448 }
else if (Subtarget.hasZeroCycleZeroingGPR32()) {
5460 if (AArch64::GPR64spRegClass.
contains(DestReg) &&
5461 AArch64::GPR64spRegClass.
contains(SrcReg)) {
5462 if (DestReg == AArch64::SP || SrcReg == AArch64::SP) {
5478 if (AArch64::GPR64spRegClass.
contains(DestReg) && SrcReg == AArch64::XZR) {
5479 if (Subtarget.hasZeroCycleZeroingGPR64()) {
5492 if (AArch64::PPRRegClass.
contains(DestReg) &&
5493 AArch64::PPRRegClass.
contains(SrcReg)) {
5494 assert(Subtarget.isSVEorStreamingSVEAvailable() &&
5495 "Unexpected SVE register.");
5505 bool DestIsPNR = AArch64::PNRRegClass.contains(DestReg);
5506 bool SrcIsPNR = AArch64::PNRRegClass.contains(SrcReg);
5507 if (DestIsPNR || SrcIsPNR) {
5509 return (R - AArch64::PN0) + AArch64::P0;
5514 if (PPRSrcReg != PPRDestReg) {
5526 if (AArch64::ZPRRegClass.
contains(DestReg) &&
5527 AArch64::ZPRRegClass.
contains(SrcReg)) {
5528 assert(Subtarget.isSVEorStreamingSVEAvailable() &&
5529 "Unexpected SVE register.");
5537 if ((AArch64::ZPR2RegClass.
contains(DestReg) ||
5538 AArch64::ZPR2StridedOrContiguousRegClass.
contains(DestReg)) &&
5539 (AArch64::ZPR2RegClass.
contains(SrcReg) ||
5540 AArch64::ZPR2StridedOrContiguousRegClass.
contains(SrcReg))) {
5541 assert(Subtarget.isSVEorStreamingSVEAvailable() &&
5542 "Unexpected SVE register.");
5543 static const unsigned Indices[] = {AArch64::zsub0, AArch64::zsub1};
5550 if (AArch64::ZPR3RegClass.
contains(DestReg) &&
5551 AArch64::ZPR3RegClass.
contains(SrcReg)) {
5552 assert(Subtarget.isSVEorStreamingSVEAvailable() &&
5553 "Unexpected SVE register.");
5554 static const unsigned Indices[] = {AArch64::zsub0, AArch64::zsub1,
5562 if ((AArch64::ZPR4RegClass.
contains(DestReg) ||
5563 AArch64::ZPR4StridedOrContiguousRegClass.
contains(DestReg)) &&
5564 (AArch64::ZPR4RegClass.
contains(SrcReg) ||
5565 AArch64::ZPR4StridedOrContiguousRegClass.
contains(SrcReg))) {
5566 assert(Subtarget.isSVEorStreamingSVEAvailable() &&
5567 "Unexpected SVE register.");
5568 static const unsigned Indices[] = {AArch64::zsub0, AArch64::zsub1,
5569 AArch64::zsub2, AArch64::zsub3};
5576 if (AArch64::DDDDRegClass.
contains(DestReg) &&
5577 AArch64::DDDDRegClass.
contains(SrcReg)) {
5578 static const unsigned Indices[] = {AArch64::dsub0, AArch64::dsub1,
5579 AArch64::dsub2, AArch64::dsub3};
5586 if (AArch64::DDDRegClass.
contains(DestReg) &&
5587 AArch64::DDDRegClass.
contains(SrcReg)) {
5588 static const unsigned Indices[] = {AArch64::dsub0, AArch64::dsub1,
5596 if (AArch64::DDRegClass.
contains(DestReg) &&
5597 AArch64::DDRegClass.
contains(SrcReg)) {
5598 static const unsigned Indices[] = {AArch64::dsub0, AArch64::dsub1};
5605 if (AArch64::QQQQRegClass.
contains(DestReg) &&
5606 AArch64::QQQQRegClass.
contains(SrcReg)) {
5607 static const unsigned Indices[] = {AArch64::qsub0, AArch64::qsub1,
5608 AArch64::qsub2, AArch64::qsub3};
5615 if (AArch64::QQQRegClass.
contains(DestReg) &&
5616 AArch64::QQQRegClass.
contains(SrcReg)) {
5617 static const unsigned Indices[] = {AArch64::qsub0, AArch64::qsub1,
5625 if (AArch64::QQRegClass.
contains(DestReg) &&
5626 AArch64::QQRegClass.
contains(SrcReg)) {
5627 static const unsigned Indices[] = {AArch64::qsub0, AArch64::qsub1};
5633 if (AArch64::XSeqPairsClassRegClass.
contains(DestReg) &&
5634 AArch64::XSeqPairsClassRegClass.
contains(SrcReg)) {
5635 static const unsigned Indices[] = {AArch64::sube64, AArch64::subo64};
5637 AArch64::XZR, Indices);
5641 if (AArch64::WSeqPairsClassRegClass.
contains(DestReg) &&
5642 AArch64::WSeqPairsClassRegClass.
contains(SrcReg)) {
5643 static const unsigned Indices[] = {AArch64::sube32, AArch64::subo32};
5645 AArch64::WZR, Indices);
5649 if (AArch64::FPR128RegClass.
contains(DestReg) &&
5650 AArch64::FPR128RegClass.
contains(SrcReg)) {
5651 if (Subtarget.isSVEorStreamingSVEAvailable() &&
5652 !Subtarget.isNeonAvailable())
5655 .
addReg(AArch64::Z0 + (SrcReg - AArch64::Q0))
5656 .
addReg(AArch64::Z0 + (SrcReg - AArch64::Q0));
5657 else if (Subtarget.isNeonAvailable())
5676 if (AArch64::FPR64RegClass.
contains(DestReg) &&
5677 AArch64::FPR64RegClass.
contains(SrcReg)) {
5678 if (Subtarget.hasZeroCycleRegMoveFPR128() &&
5679 !Subtarget.hasZeroCycleRegMoveFPR64() &&
5680 !Subtarget.hasZeroCycleRegMoveFPR32() && Subtarget.isNeonAvailable()) {
5681 MCRegister DestRegQ = RI.getMatchingSuperReg(DestReg, AArch64::dsub,
5682 &AArch64::FPR128RegClass);
5683 MCRegister SrcRegQ = RI.getMatchingSuperReg(SrcReg, AArch64::dsub,
5684 &AArch64::FPR128RegClass);
5700 if (AArch64::FPR32RegClass.
contains(DestReg) &&
5701 AArch64::FPR32RegClass.
contains(SrcReg)) {
5702 if (Subtarget.hasZeroCycleRegMoveFPR128() &&
5703 !Subtarget.hasZeroCycleRegMoveFPR64() &&
5704 !Subtarget.hasZeroCycleRegMoveFPR32() && Subtarget.isNeonAvailable()) {
5705 MCRegister DestRegQ = RI.getMatchingSuperReg(DestReg, AArch64::ssub,
5706 &AArch64::FPR128RegClass);
5707 MCRegister SrcRegQ = RI.getMatchingSuperReg(SrcReg, AArch64::ssub,
5708 &AArch64::FPR128RegClass);
5717 }
else if (Subtarget.hasZeroCycleRegMoveFPR64() &&
5718 !Subtarget.hasZeroCycleRegMoveFPR32()) {
5719 MCRegister DestRegD = RI.getMatchingSuperReg(DestReg, AArch64::ssub,
5720 &AArch64::FPR64RegClass);
5721 MCRegister SrcRegD = RI.getMatchingSuperReg(SrcReg, AArch64::ssub,
5722 &AArch64::FPR64RegClass);
5737 if (AArch64::FPR16RegClass.
contains(DestReg) &&
5738 AArch64::FPR16RegClass.
contains(SrcReg)) {
5739 if (Subtarget.hasZeroCycleRegMoveFPR128() &&
5740 !Subtarget.hasZeroCycleRegMoveFPR64() &&
5741 !Subtarget.hasZeroCycleRegMoveFPR32() && Subtarget.isNeonAvailable()) {
5742 MCRegister DestRegQ = RI.getMatchingSuperReg(DestReg, AArch64::hsub,
5743 &AArch64::FPR128RegClass);
5744 MCRegister SrcRegQ = RI.getMatchingSuperReg(SrcReg, AArch64::hsub,
5745 &AArch64::FPR128RegClass);
5754 }
else if (Subtarget.hasZeroCycleRegMoveFPR64() &&
5755 !Subtarget.hasZeroCycleRegMoveFPR32()) {
5756 MCRegister DestRegD = RI.getMatchingSuperReg(DestReg, AArch64::hsub,
5757 &AArch64::FPR64RegClass);
5758 MCRegister SrcRegD = RI.getMatchingSuperReg(SrcReg, AArch64::hsub,
5759 &AArch64::FPR64RegClass);
5768 DestReg = RI.getMatchingSuperReg(DestReg, AArch64::hsub,
5769 &AArch64::FPR32RegClass);
5770 SrcReg = RI.getMatchingSuperReg(SrcReg, AArch64::hsub,
5771 &AArch64::FPR32RegClass);
5778 if (AArch64::FPR8RegClass.
contains(DestReg) &&
5779 AArch64::FPR8RegClass.
contains(SrcReg)) {
5780 if (Subtarget.hasZeroCycleRegMoveFPR128() &&
5781 !Subtarget.hasZeroCycleRegMoveFPR64() &&
5782 !Subtarget.hasZeroCycleRegMoveFPR64() && Subtarget.isNeonAvailable()) {
5783 MCRegister DestRegQ = RI.getMatchingSuperReg(DestReg, AArch64::bsub,
5784 &AArch64::FPR128RegClass);
5785 MCRegister SrcRegQ = RI.getMatchingSuperReg(SrcReg, AArch64::bsub,
5786 &AArch64::FPR128RegClass);
5795 }
else if (Subtarget.hasZeroCycleRegMoveFPR64() &&
5796 !Subtarget.hasZeroCycleRegMoveFPR32()) {
5797 MCRegister DestRegD = RI.getMatchingSuperReg(DestReg, AArch64::bsub,
5798 &AArch64::FPR64RegClass);
5799 MCRegister SrcRegD = RI.getMatchingSuperReg(SrcReg, AArch64::bsub,
5800 &AArch64::FPR64RegClass);
5809 DestReg = RI.getMatchingSuperReg(DestReg, AArch64::bsub,
5810 &AArch64::FPR32RegClass);
5811 SrcReg = RI.getMatchingSuperReg(SrcReg, AArch64::bsub,
5812 &AArch64::FPR32RegClass);
5820 if (AArch64::FPR64RegClass.
contains(DestReg) &&
5821 AArch64::GPR64RegClass.
contains(SrcReg)) {
5822 if (AArch64::XZR == SrcReg) {
5830 if (AArch64::GPR64RegClass.
contains(DestReg) &&
5831 AArch64::FPR64RegClass.
contains(SrcReg)) {
5837 if (AArch64::FPR32RegClass.
contains(DestReg) &&
5838 AArch64::GPR32RegClass.
contains(SrcReg)) {
5839 if (AArch64::WZR == SrcReg) {
5847 if (AArch64::GPR32RegClass.
contains(DestReg) &&
5848 AArch64::FPR32RegClass.
contains(SrcReg)) {
5854 if (DestReg == AArch64::NZCV) {
5855 assert(AArch64::GPR64RegClass.
contains(SrcReg) &&
"Invalid NZCV copy");
5857 .
addImm(AArch64SysReg::NZCV)
5863 if (SrcReg == AArch64::NZCV) {
5864 assert(AArch64::GPR64RegClass.
contains(DestReg) &&
"Invalid NZCV copy");
5866 .
addImm(AArch64SysReg::NZCV)
5872 errs() << RI.getRegAsmName(DestReg) <<
" = COPY " << RI.getRegAsmName(SrcReg)
5883 unsigned SubIdx0,
unsigned SubIdx1,
int FI,
5888 SrcReg0 =
TRI.getSubReg(SrcReg, SubIdx0);
5890 SrcReg1 =
TRI.getSubReg(SrcReg, SubIdx1);
5903 Register SrcReg,
bool isKill,
int FI,
5918 switch (RI.getSpillSize(*RC)) {
5920 if (AArch64::FPR8RegClass.hasSubClassEq(RC))
5921 Opc = AArch64::STRBui;
5924 if (AArch64::FPR16RegClass.hasSubClassEq(RC))
5925 Opc = AArch64::STRHui;
5926 else if (AArch64::PNRRegClass.hasSubClassEq(RC) ||
5927 AArch64::PPRRegClass.hasSubClassEq(RC)) {
5928 assert(Subtarget.isSVEorStreamingSVEAvailable() &&
5929 "Unexpected register store without SVE store instructions");
5930 Opc = AArch64::STR_PXI;
5936 if (AArch64::GPR32allRegClass.hasSubClassEq(RC)) {
5937 Opc = AArch64::STRWui;
5941 assert(SrcReg != AArch64::WSP);
5942 }
else if (AArch64::FPR32RegClass.hasSubClassEq(RC))
5943 Opc = AArch64::STRSui;
5944 else if (AArch64::PPR2RegClass.hasSubClassEq(RC)) {
5945 Opc = AArch64::STR_PPXI;
5950 if (AArch64::GPR64allRegClass.hasSubClassEq(RC)) {
5951 Opc = AArch64::STRXui;
5955 assert(SrcReg != AArch64::SP);
5956 }
else if (AArch64::FPR64RegClass.hasSubClassEq(RC)) {
5957 Opc = AArch64::STRDui;
5958 }
else if (AArch64::WSeqPairsClassRegClass.hasSubClassEq(RC)) {
5960 get(AArch64::STPWi), SrcReg, isKill,
5961 AArch64::sube32, AArch64::subo32, FI, MMO);
5966 if (AArch64::FPR128RegClass.hasSubClassEq(RC))
5967 Opc = AArch64::STRQui;
5968 else if (AArch64::DDRegClass.hasSubClassEq(RC)) {
5969 assert(Subtarget.hasNEON() &&
"Unexpected register store without NEON");
5970 Opc = AArch64::ST1Twov1d;
5972 }
else if (AArch64::XSeqPairsClassRegClass.hasSubClassEq(RC)) {
5974 get(AArch64::STPXi), SrcReg, isKill,
5975 AArch64::sube64, AArch64::subo64, FI, MMO);
5977 }
else if (AArch64::ZPRRegClass.hasSubClassEq(RC)) {
5978 assert(Subtarget.isSVEorStreamingSVEAvailable() &&
5979 "Unexpected register store without SVE store instructions");
5980 Opc = AArch64::STR_ZXI;
5985 if (AArch64::DDDRegClass.hasSubClassEq(RC)) {
5986 assert(Subtarget.hasNEON() &&
"Unexpected register store without NEON");
5987 Opc = AArch64::ST1Threev1d;
5992 if (AArch64::DDDDRegClass.hasSubClassEq(RC)) {
5993 assert(Subtarget.hasNEON() &&
"Unexpected register store without NEON");
5994 Opc = AArch64::ST1Fourv1d;
5996 }
else if (AArch64::QQRegClass.hasSubClassEq(RC)) {
5997 assert(Subtarget.hasNEON() &&
"Unexpected register store without NEON");
5998 Opc = AArch64::ST1Twov2d;
6000 }
else if (AArch64::ZPR2StridedOrContiguousRegClass.hasSubClassEq(RC)) {
6001 assert(Subtarget.isSVEorStreamingSVEAvailable() &&
6002 "Unexpected register store without SVE store instructions");
6003 Opc = AArch64::STR_ZZXI_STRIDED_CONTIGUOUS;
6005 }
else if (AArch64::ZPR2RegClass.hasSubClassEq(RC)) {
6006 assert(Subtarget.isSVEorStreamingSVEAvailable() &&
6007 "Unexpected register store without SVE store instructions");
6008 Opc = AArch64::STR_ZZXI;
6013 if (AArch64::QQQRegClass.hasSubClassEq(RC)) {
6014 assert(Subtarget.hasNEON() &&
"Unexpected register store without NEON");
6015 Opc = AArch64::ST1Threev2d;
6017 }
else if (AArch64::ZPR3RegClass.hasSubClassEq(RC)) {
6018 assert(Subtarget.isSVEorStreamingSVEAvailable() &&
6019 "Unexpected register store without SVE store instructions");
6020 Opc = AArch64::STR_ZZZXI;
6025 if (AArch64::QQQQRegClass.hasSubClassEq(RC)) {
6026 assert(Subtarget.hasNEON() &&
"Unexpected register store without NEON");
6027 Opc = AArch64::ST1Fourv2d;
6029 }
else if (AArch64::ZPR4StridedOrContiguousRegClass.hasSubClassEq(RC)) {
6030 assert(Subtarget.isSVEorStreamingSVEAvailable() &&
6031 "Unexpected register store without SVE store instructions");
6032 Opc = AArch64::STR_ZZZZXI_STRIDED_CONTIGUOUS;
6034 }
else if (AArch64::ZPR4RegClass.hasSubClassEq(RC)) {
6035 assert(Subtarget.isSVEorStreamingSVEAvailable() &&
6036 "Unexpected register store without SVE store instructions");
6037 Opc = AArch64::STR_ZZZZXI;
6042 assert(
Opc &&
"Unknown register class");
6053 MI.addMemOperand(MMO);
6060 Register DestReg,
unsigned SubIdx0,
6061 unsigned SubIdx1,
int FI,
6065 bool IsUndef =
true;
6067 DestReg0 =
TRI.getSubReg(DestReg, SubIdx0);
6069 DestReg1 =
TRI.getSubReg(DestReg, SubIdx1);
6098 switch (
TRI.getSpillSize(*RC)) {
6100 if (AArch64::FPR8RegClass.hasSubClassEq(RC))
6101 Opc = AArch64::LDRBui;
6104 bool IsPNR = AArch64::PNRRegClass.hasSubClassEq(RC);
6105 if (AArch64::FPR16RegClass.hasSubClassEq(RC))
6106 Opc = AArch64::LDRHui;
6107 else if (IsPNR || AArch64::PPRRegClass.hasSubClassEq(RC)) {
6108 assert(Subtarget.isSVEorStreamingSVEAvailable() &&
6109 "Unexpected register load without SVE load instructions");
6112 Opc = AArch64::LDR_PXI;
6118 if (AArch64::GPR32allRegClass.hasSubClassEq(RC)) {
6119 Opc = AArch64::LDRWui;
6123 assert(DestReg != AArch64::WSP);
6124 }
else if (AArch64::FPR32RegClass.hasSubClassEq(RC))
6125 Opc = AArch64::LDRSui;
6126 else if (AArch64::PPR2RegClass.hasSubClassEq(RC)) {
6127 Opc = AArch64::LDR_PPXI;
6132 if (AArch64::GPR64allRegClass.hasSubClassEq(RC)) {
6133 Opc = AArch64::LDRXui;
6137 assert(DestReg != AArch64::SP);
6138 }
else if (AArch64::FPR64RegClass.hasSubClassEq(RC)) {
6139 Opc = AArch64::LDRDui;
6140 }
else if (AArch64::WSeqPairsClassRegClass.hasSubClassEq(RC)) {
6142 get(AArch64::LDPWi), DestReg, AArch64::sube32,
6143 AArch64::subo32, FI, MMO);
6148 if (AArch64::FPR128RegClass.hasSubClassEq(RC))
6149 Opc = AArch64::LDRQui;
6150 else if (AArch64::DDRegClass.hasSubClassEq(RC)) {
6151 assert(Subtarget.hasNEON() &&
"Unexpected register load without NEON");
6152 Opc = AArch64::LD1Twov1d;
6154 }
else if (AArch64::XSeqPairsClassRegClass.hasSubClassEq(RC)) {
6156 get(AArch64::LDPXi), DestReg, AArch64::sube64,
6157 AArch64::subo64, FI, MMO);
6159 }
else if (AArch64::ZPRRegClass.hasSubClassEq(RC)) {
6160 assert(Subtarget.isSVEorStreamingSVEAvailable() &&
6161 "Unexpected register load without SVE load instructions");
6162 Opc = AArch64::LDR_ZXI;
6167 if (AArch64::DDDRegClass.hasSubClassEq(RC)) {
6168 assert(Subtarget.hasNEON() &&
"Unexpected register load without NEON");
6169 Opc = AArch64::LD1Threev1d;
6174 if (AArch64::DDDDRegClass.hasSubClassEq(RC)) {
6175 assert(Subtarget.hasNEON() &&
"Unexpected register load without NEON");
6176 Opc = AArch64::LD1Fourv1d;
6178 }
else if (AArch64::QQRegClass.hasSubClassEq(RC)) {
6179 assert(Subtarget.hasNEON() &&
"Unexpected register load without NEON");
6180 Opc = AArch64::LD1Twov2d;
6182 }
else if (AArch64::ZPR2StridedOrContiguousRegClass.hasSubClassEq(RC)) {
6183 assert(Subtarget.isSVEorStreamingSVEAvailable() &&
6184 "Unexpected register load without SVE load instructions");
6185 Opc = AArch64::LDR_ZZXI_STRIDED_CONTIGUOUS;
6187 }
else if (AArch64::ZPR2RegClass.hasSubClassEq(RC)) {
6188 assert(Subtarget.isSVEorStreamingSVEAvailable() &&
6189 "Unexpected register load without SVE load instructions");
6190 Opc = AArch64::LDR_ZZXI;
6195 if (AArch64::QQQRegClass.hasSubClassEq(RC)) {
6196 assert(Subtarget.hasNEON() &&
"Unexpected register load without NEON");
6197 Opc = AArch64::LD1Threev2d;
6199 }
else if (AArch64::ZPR3RegClass.hasSubClassEq(RC)) {
6200 assert(Subtarget.isSVEorStreamingSVEAvailable() &&
6201 "Unexpected register load without SVE load instructions");
6202 Opc = AArch64::LDR_ZZZXI;
6207 if (AArch64::QQQQRegClass.hasSubClassEq(RC)) {
6208 assert(Subtarget.hasNEON() &&
"Unexpected register load without NEON");
6209 Opc = AArch64::LD1Fourv2d;
6211 }
else if (AArch64::ZPR4StridedOrContiguousRegClass.hasSubClassEq(RC)) {
6212 assert(Subtarget.isSVEorStreamingSVEAvailable() &&
6213 "Unexpected register load without SVE load instructions");
6214 Opc = AArch64::LDR_ZZZZXI_STRIDED_CONTIGUOUS;
6216 }
else if (AArch64::ZPR4RegClass.hasSubClassEq(RC)) {
6217 assert(Subtarget.isSVEorStreamingSVEAvailable() &&
6218 "Unexpected register load without SVE load instructions");
6219 Opc = AArch64::LDR_ZZZZXI;
6225 assert(
Opc &&
"Unknown register class");
6235 MI.addMemOperand(MMO);
6242 UseMI.getIterator()),
6244 return I.modifiesRegister(AArch64::NZCV, TRI) ||
6245 I.readsRegister(AArch64::NZCV, TRI);
6249void AArch64InstrInfo::decomposeStackOffsetForDwarfOffsets(
6254 assert(
Offset.getScalable() % 2 == 0 &&
"Invalid frame offset");
6261 ByteSized =
Offset.getFixed();
6262 VGSized =
Offset.getScalable() / 2;
6268void AArch64InstrInfo::decomposeStackOffsetForFrameOffsets(
6270 int64_t &NumDataVectors) {
6274 assert(
Offset.getScalable() % 2 == 0 &&
"Invalid frame offset");
6276 NumBytes =
Offset.getFixed();
6278 NumPredicateVectors =
Offset.getScalable() / 2;
6283 if (NumPredicateVectors % 8 == 0 || NumPredicateVectors < -64 ||
6284 NumPredicateVectors > 62) {
6285 NumDataVectors = NumPredicateVectors / 8;
6286 NumPredicateVectors -= NumDataVectors * 8;
6312 Expr.
push_back((
char)dwarf::DW_OP_bregx);
6320 int64_t OffsetFromDefCFA) {
6334 Comment << (NumBytes < 0 ?
" - " :
" + ") << std::abs(NumBytes);
6335 if (!RegScale.empty())
6345 int64_t NumBytes, NumVGScaledBytes;
6346 AArch64InstrInfo::decomposeStackOffsetForDwarfOffsets(
Offset, NumBytes,
6348 std::string CommentBuffer;
6351 if (
Reg == AArch64::SP)
6353 else if (
Reg == AArch64::FP)
6360 unsigned DwarfReg =
TRI.getDwarfRegNum(
Reg,
true);
6361 assert(DwarfReg <= 31 &&
"DwarfReg out of bounds (0..31)");
6363 Expr.
push_back(dwarf::DW_OP_breg0 + DwarfReg);
6366 if (NumVGScaledBytes) {
6376 DefCfaExpr.
push_back(dwarf::DW_CFA_def_cfa_expression);
6384 unsigned FrameReg,
unsigned Reg,
6386 bool LastAdjustmentWasScalable) {
6387 if (
Offset.getScalable())
6390 if (FrameReg == Reg && !LastAdjustmentWasScalable)
6393 unsigned DwarfReg =
TRI.getDwarfRegNum(Reg,
true);
6400 std::optional<int64_t> IncomingVGOffsetFromDefCFA) {
6401 int64_t NumBytes, NumVGScaledBytes;
6402 AArch64InstrInfo::decomposeStackOffsetForDwarfOffsets(
6403 OffsetFromDefCFA, NumBytes, NumVGScaledBytes);
6405 unsigned DwarfReg =
TRI.getDwarfRegNum(Reg,
true);
6408 if (!NumVGScaledBytes)
6411 std::string CommentBuffer;
6416 assert(NumVGScaledBytes &&
"Expected scalable offset");
6420 if (IncomingVGOffsetFromDefCFA) {
6422 VGRegScale =
"* IncomingVG";
6425 VGRegScale =
"* VG";
6429 OffsetExpr.
push_back(dwarf::DW_OP_plus);
6438 CfaExpr.
push_back(dwarf::DW_CFA_expression);
6453 unsigned SrcReg, int64_t
Offset,
unsigned Opc,
6456 bool *HasWinCFI,
bool EmitCFAOffset,
6459 unsigned MaxEncoding, ShiftSize;
6461 case AArch64::ADDXri:
6462 case AArch64::ADDSXri:
6463 case AArch64::SUBXri:
6464 case AArch64::SUBSXri:
6465 MaxEncoding = 0xfff;
6468 case AArch64::ADDVL_XXI:
6469 case AArch64::ADDPL_XXI:
6470 case AArch64::ADDSVL_XXI:
6471 case AArch64::ADDSPL_XXI:
6486 if (
Opc == AArch64::ADDVL_XXI ||
Opc == AArch64::ADDSVL_XXI)
6488 else if (
Opc == AArch64::ADDPL_XXI ||
Opc == AArch64::ADDSPL_XXI)
6502 const unsigned MaxEncodableValue = MaxEncoding << ShiftSize;
6504 if (TmpReg == AArch64::XZR)
6505 TmpReg =
MBB.getParent()->getRegInfo().createVirtualRegister(
6506 &AArch64::GPR64RegClass);
6508 uint64_t ThisVal = std::min<uint64_t>(
Offset, MaxEncodableValue);
6509 unsigned LocalShiftSize = 0;
6510 if (ThisVal > MaxEncoding) {
6511 ThisVal = ThisVal >> ShiftSize;
6512 LocalShiftSize = ShiftSize;
6514 assert((ThisVal >> ShiftSize) <= MaxEncoding &&
6515 "Encoding cannot handle value that big");
6517 Offset -= ThisVal << LocalShiftSize;
6522 .
addImm(Sign * (
int)ThisVal);
6532 if (Sign == -1 ||
Opc == AArch64::SUBXri ||
Opc == AArch64::SUBSXri)
6533 CFAOffset += Change;
6535 CFAOffset -= Change;
6536 if (EmitCFAOffset && DestReg == TmpReg) {
6549 int Imm = (int)(ThisVal << LocalShiftSize);
6550 if (VScale != 1 && DestReg == AArch64::SP) {
6556 }
else if ((DestReg == AArch64::FP && SrcReg == AArch64::SP) ||
6557 (SrcReg == AArch64::FP && DestReg == AArch64::SP)) {
6558 assert(VScale == 1 &&
"Expected non-scalable operation");
6567 assert(
Offset == 0 &&
"Expected remaining offset to be zero to "
6568 "emit a single SEH directive");
6569 }
else if (DestReg == AArch64::SP) {
6570 assert(VScale == 1 &&
"Expected non-scalable operation");
6573 assert(SrcReg == AArch64::SP &&
"Unexpected SrcReg for SEH_StackAlloc");
6586 unsigned DestReg,
unsigned SrcReg,
6589 bool NeedsWinCFI,
bool *HasWinCFI,
6591 unsigned FrameReg) {
6598 bool UseSVL =
F.hasFnAttribute(
"aarch64_pstate_sm_body");
6600 int64_t Bytes, NumPredicateVectors, NumDataVectors;
6601 AArch64InstrInfo::decomposeStackOffsetForFrameOffsets(
6602 Offset, Bytes, NumPredicateVectors, NumDataVectors);
6605 bool NeedsFinalDefNZCV = SetNZCV && (NumPredicateVectors || NumDataVectors);
6606 if (NeedsFinalDefNZCV)
6610 if (Bytes || (!
Offset && SrcReg != DestReg)) {
6611 assert((DestReg != AArch64::SP || Bytes % 8 == 0) &&
6612 "SP increment/decrement not 8-byte aligned");
6613 unsigned Opc = SetNZCV ? AArch64::ADDSXri : AArch64::ADDXri;
6616 Opc = SetNZCV ? AArch64::SUBSXri : AArch64::SUBXri;
6619 NeedsWinCFI, HasWinCFI, EmitCFAOffset, CFAOffset,
6621 CFAOffset += (
Opc == AArch64::ADDXri ||
Opc == AArch64::ADDSXri)
6628 assert(!(NeedsWinCFI && NumPredicateVectors) &&
6629 "WinCFI can't allocate fractions of an SVE data vector");
6631 if (NumDataVectors) {
6633 UseSVL ? AArch64::ADDSVL_XXI : AArch64::ADDVL_XXI,
TII,
6634 Flag, NeedsWinCFI, HasWinCFI, EmitCFAOffset, CFAOffset,
6640 if (NumPredicateVectors) {
6641 assert(DestReg != AArch64::SP &&
"Unaligned access to SP");
6643 UseSVL ? AArch64::ADDSPL_XXI : AArch64::ADDPL_XXI,
TII,
6644 Flag, NeedsWinCFI, HasWinCFI, EmitCFAOffset, CFAOffset,
6648 if (NeedsFinalDefNZCV)
6669 if (
MI.isFullCopy()) {
6672 if (SrcReg == AArch64::SP && DstReg.
isVirtual()) {
6676 if (DstReg == AArch64::SP && SrcReg.
isVirtual()) {
6681 if (SrcReg == AArch64::NZCV || DstReg == AArch64::NZCV)
6709 if (
MI.isCopy() &&
Ops.size() == 1 &&
6711 (
Ops[0] == 0 ||
Ops[0] == 1)) {
6712 bool IsSpill =
Ops[0] == 0;
6713 bool IsFill = !IsSpill;
6725 :
TRI.getMinimalPhysRegClass(Reg);
6731 "Mismatched register size in non subreg COPY");
6738 return &*--InsertPt;
6750 if (IsSpill && DstMO.
isUndef() && SrcReg == AArch64::WZR &&
6753 "Unexpected subreg on physical register");
6755 FrameIndex, &AArch64::GPR64RegClass,
Register());
6756 return &*--InsertPt;
6773 case AArch64::sub_32:
6774 if (AArch64::GPR64RegClass.hasSubClassEq(
getRegClass(DstReg)))
6775 FillRC = &AArch64::GPR32RegClass;
6778 FillRC = &AArch64::FPR32RegClass;
6781 FillRC = &AArch64::FPR64RegClass;
6787 TRI.getRegSizeInBits(*FillRC) &&
6788 "Mismatched regclass size on folded subreg COPY");
6807 bool *OutUseUnscaledOp,
6808 unsigned *OutUnscaledOp,
6809 int64_t *EmittableOffset) {
6811 if (EmittableOffset)
6812 *EmittableOffset = 0;
6813 if (OutUseUnscaledOp)
6814 *OutUseUnscaledOp =
false;
6820 switch (
MI.getOpcode()) {
6823 case AArch64::LD1Rv1d:
6824 case AArch64::LD1Rv2s:
6825 case AArch64::LD1Rv2d:
6826 case AArch64::LD1Rv4h:
6827 case AArch64::LD1Rv4s:
6828 case AArch64::LD1Rv8b:
6829 case AArch64::LD1Rv8h:
6830 case AArch64::LD1Rv16b:
6831 case AArch64::LD1Twov2d:
6832 case AArch64::LD1Threev2d:
6833 case AArch64::LD1Fourv2d:
6834 case AArch64::LD1Twov1d:
6835 case AArch64::LD1Threev1d:
6836 case AArch64::LD1Fourv1d:
6837 case AArch64::ST1Twov2d:
6838 case AArch64::ST1Threev2d:
6839 case AArch64::ST1Fourv2d:
6840 case AArch64::ST1Twov1d:
6841 case AArch64::ST1Threev1d:
6842 case AArch64::ST1Fourv1d:
6843 case AArch64::ST1i8:
6844 case AArch64::ST1i16:
6845 case AArch64::ST1i32:
6846 case AArch64::ST1i64:
6848 case AArch64::IRGstack:
6849 case AArch64::STGloop:
6850 case AArch64::STZGloop:
6855 TypeSize ScaleValue(0U,
false), Width(0U,
false);
6856 int64_t MinOff, MaxOff;
6862 bool IsMulVL = ScaleValue.isScalable();
6863 unsigned Scale = ScaleValue.getKnownMinValue();
6873 std::optional<unsigned> UnscaledOp =
6875 bool useUnscaledOp = UnscaledOp && (
Offset % Scale ||
Offset < 0);
6876 if (useUnscaledOp &&
6881 Scale = ScaleValue.getKnownMinValue();
6882 assert(IsMulVL == ScaleValue.isScalable() &&
6883 "Unscaled opcode has different value for scalable");
6885 int64_t Remainder =
Offset % Scale;
6886 assert(!(Remainder && useUnscaledOp) &&
6887 "Cannot have remainder when using unscaled op");
6889 assert(MinOff < MaxOff &&
"Unexpected Min/Max offsets");
6890 int64_t NewOffset =
Offset / Scale;
6891 if (MinOff <= NewOffset && NewOffset <= MaxOff)
6894 NewOffset = NewOffset < 0 ? MinOff : MaxOff;
6898 if (EmittableOffset)
6899 *EmittableOffset = NewOffset;
6900 if (OutUseUnscaledOp)
6901 *OutUseUnscaledOp = useUnscaledOp;
6902 if (OutUnscaledOp && UnscaledOp)
6903 *OutUnscaledOp = *UnscaledOp;
6916 unsigned Opcode =
MI.getOpcode();
6917 unsigned ImmIdx = FrameRegIdx + 1;
6919 if (Opcode == AArch64::ADDSXri || Opcode == AArch64::ADDXri) {
6924 MI.eraseFromParent();
6930 unsigned UnscaledOp;
6933 &UnscaledOp, &NewOffset);
6937 MI.getOperand(FrameRegIdx).ChangeToRegister(FrameReg,
false);
6939 MI.setDesc(
TII->get(UnscaledOp));
6941 MI.getOperand(ImmIdx).ChangeToImmediate(NewOffset);
6957bool AArch64InstrInfo::useMachineCombiner()
const {
return true; }
6962 case AArch64::ADDSWrr:
6963 case AArch64::ADDSWri:
6964 case AArch64::ADDSXrr:
6965 case AArch64::ADDSXri:
6966 case AArch64::SUBSWrr:
6967 case AArch64::SUBSXrr:
6969 case AArch64::SUBSWri:
6970 case AArch64::SUBSXri:
6981 case AArch64::ADDWrr:
6982 case AArch64::ADDWri:
6983 case AArch64::SUBWrr:
6984 case AArch64::ADDSWrr:
6985 case AArch64::ADDSWri:
6986 case AArch64::SUBSWrr:
6988 case AArch64::SUBWri:
6989 case AArch64::SUBSWri:
7000 case AArch64::ADDXrr:
7001 case AArch64::ADDXri:
7002 case AArch64::SUBXrr:
7003 case AArch64::ADDSXrr:
7004 case AArch64::ADDSXri:
7005 case AArch64::SUBSXrr:
7007 case AArch64::SUBXri:
7008 case AArch64::SUBSXri:
7009 case AArch64::ADDv8i8:
7010 case AArch64::ADDv16i8:
7011 case AArch64::ADDv4i16:
7012 case AArch64::ADDv8i16:
7013 case AArch64::ADDv2i32:
7014 case AArch64::ADDv4i32:
7015 case AArch64::SUBv8i8:
7016 case AArch64::SUBv16i8:
7017 case AArch64::SUBv4i16:
7018 case AArch64::SUBv8i16:
7019 case AArch64::SUBv2i32:
7020 case AArch64::SUBv4i32:
7033 case AArch64::FADDHrr:
7034 case AArch64::FADDSrr:
7035 case AArch64::FADDDrr:
7036 case AArch64::FADDv4f16:
7037 case AArch64::FADDv8f16:
7038 case AArch64::FADDv2f32:
7039 case AArch64::FADDv2f64:
7040 case AArch64::FADDv4f32:
7041 case AArch64::FSUBHrr:
7042 case AArch64::FSUBSrr:
7043 case AArch64::FSUBDrr:
7044 case AArch64::FSUBv4f16:
7045 case AArch64::FSUBv8f16:
7046 case AArch64::FSUBv2f32:
7047 case AArch64::FSUBv2f64:
7048 case AArch64::FSUBv4f32:
7067 unsigned CombineOpc,
unsigned ZeroReg = 0,
7068 bool CheckZeroReg =
false) {
7075 if (!
MI ||
MI->getParent() != &
MBB ||
MI->getOpcode() != CombineOpc)
7078 if (!
MRI.hasOneNonDBGUse(
MI->getOperand(0).getReg()))
7082 assert(
MI->getNumOperands() >= 4 &&
MI->getOperand(0).isReg() &&
7083 MI->getOperand(1).isReg() &&
MI->getOperand(2).isReg() &&
7084 MI->getOperand(3).isReg() &&
"MAdd/MSub must have a least 4 regs");
7086 if (
MI->getOperand(3).getReg() != ZeroReg)
7091 MI->findRegisterDefOperandIdx(AArch64::NZCV,
nullptr,
true) == -1)
7100 unsigned MulOpc,
unsigned ZeroReg) {
7115bool AArch64InstrInfo::isAssociativeAndCommutative(
const MachineInstr &Inst,
7116 bool Invert)
const {
7122 case AArch64::FADDHrr:
7123 case AArch64::FADDSrr:
7124 case AArch64::FADDDrr:
7125 case AArch64::FMULHrr:
7126 case AArch64::FMULSrr:
7127 case AArch64::FMULDrr:
7128 case AArch64::FMULX16:
7129 case AArch64::FMULX32:
7130 case AArch64::FMULX64:
7132 case AArch64::FADDv4f16:
7133 case AArch64::FADDv8f16:
7134 case AArch64::FADDv2f32:
7135 case AArch64::FADDv4f32:
7136 case AArch64::FADDv2f64:
7137 case AArch64::FMULv4f16:
7138 case AArch64::FMULv8f16:
7139 case AArch64::FMULv2f32:
7140 case AArch64::FMULv4f32:
7141 case AArch64::FMULv2f64:
7142 case AArch64::FMULXv4f16:
7143 case AArch64::FMULXv8f16:
7144 case AArch64::FMULXv2f32:
7145 case AArch64::FMULXv4f32:
7146 case AArch64::FMULXv2f64:
7150 case AArch64::FADD_ZZZ_H:
7151 case AArch64::FADD_ZZZ_S:
7152 case AArch64::FADD_ZZZ_D:
7153 case AArch64::FMUL_ZZZ_H:
7154 case AArch64::FMUL_ZZZ_S:
7155 case AArch64::FMUL_ZZZ_D:
7166 case AArch64::ADDWrr:
7167 case AArch64::ADDXrr:
7168 case AArch64::ANDWrr:
7169 case AArch64::ANDXrr:
7170 case AArch64::ORRWrr:
7171 case AArch64::ORRXrr:
7172 case AArch64::EORWrr:
7173 case AArch64::EORXrr:
7174 case AArch64::EONWrr:
7175 case AArch64::EONXrr:
7179 case AArch64::ADDv8i8:
7180 case AArch64::ADDv16i8:
7181 case AArch64::ADDv4i16:
7182 case AArch64::ADDv8i16:
7183 case AArch64::ADDv2i32:
7184 case AArch64::ADDv4i32:
7185 case AArch64::ADDv1i64:
7186 case AArch64::ADDv2i64:
7187 case AArch64::MULv8i8:
7188 case AArch64::MULv16i8:
7189 case AArch64::MULv4i16:
7190 case AArch64::MULv8i16:
7191 case AArch64::MULv2i32:
7192 case AArch64::MULv4i32:
7193 case AArch64::ANDv8i8:
7194 case AArch64::ANDv16i8:
7195 case AArch64::ORRv8i8:
7196 case AArch64::ORRv16i8:
7197 case AArch64::EORv8i8:
7198 case AArch64::EORv16i8:
7200 case AArch64::ADD_ZZZ_B:
7201 case AArch64::ADD_ZZZ_H:
7202 case AArch64::ADD_ZZZ_S:
7203 case AArch64::ADD_ZZZ_D:
7204 case AArch64::MUL_ZZZ_B:
7205 case AArch64::MUL_ZZZ_H:
7206 case AArch64::MUL_ZZZ_S:
7207 case AArch64::MUL_ZZZ_D:
7208 case AArch64::AND_ZZZ:
7209 case AArch64::ORR_ZZZ:
7210 case AArch64::EOR_ZZZ:
7241 auto setFound = [&](
int Opcode,
int Operand,
unsigned ZeroReg,
7249 auto setVFound = [&](
int Opcode,
int Operand,
unsigned Pattern) {
7261 case AArch64::ADDWrr:
7263 "ADDWrr does not have register operands");
7264 setFound(AArch64::MADDWrrr, 1, AArch64::WZR, MCP::MULADDW_OP1);
7265 setFound(AArch64::MADDWrrr, 2, AArch64::WZR, MCP::MULADDW_OP2);
7267 case AArch64::ADDXrr:
7268 setFound(AArch64::MADDXrrr, 1, AArch64::XZR, MCP::MULADDX_OP1);
7269 setFound(AArch64::MADDXrrr, 2, AArch64::XZR, MCP::MULADDX_OP2);
7271 case AArch64::SUBWrr:
7272 setFound(AArch64::MADDWrrr, 2, AArch64::WZR, MCP::MULSUBW_OP2);
7273 setFound(AArch64::MADDWrrr, 1, AArch64::WZR, MCP::MULSUBW_OP1);
7275 case AArch64::SUBXrr:
7276 setFound(AArch64::MADDXrrr, 2, AArch64::XZR, MCP::MULSUBX_OP2);
7277 setFound(AArch64::MADDXrrr, 1, AArch64::XZR, MCP::MULSUBX_OP1);
7279 case AArch64::ADDWri:
7280 setFound(AArch64::MADDWrrr, 1, AArch64::WZR, MCP::MULADDWI_OP1);
7282 case AArch64::ADDXri:
7283 setFound(AArch64::MADDXrrr, 1, AArch64::XZR, MCP::MULADDXI_OP1);
7285 case AArch64::SUBWri:
7286 setFound(AArch64::MADDWrrr, 1, AArch64::WZR, MCP::MULSUBWI_OP1);
7288 case AArch64::SUBXri:
7289 setFound(AArch64::MADDXrrr, 1, AArch64::XZR, MCP::MULSUBXI_OP1);
7291 case AArch64::ADDv8i8:
7292 setVFound(AArch64::MULv8i8, 1, MCP::MULADDv8i8_OP1);
7293 setVFound(AArch64::MULv8i8, 2, MCP::MULADDv8i8_OP2);
7295 case AArch64::ADDv16i8:
7296 setVFound(AArch64::MULv16i8, 1, MCP::MULADDv16i8_OP1);
7297 setVFound(AArch64::MULv16i8, 2, MCP::MULADDv16i8_OP2);
7299 case AArch64::ADDv4i16:
7300 setVFound(AArch64::MULv4i16, 1, MCP::MULADDv4i16_OP1);
7301 setVFound(AArch64::MULv4i16, 2, MCP::MULADDv4i16_OP2);
7302 setVFound(AArch64::MULv4i16_indexed, 1, MCP::MULADDv4i16_indexed_OP1);
7303 setVFound(AArch64::MULv4i16_indexed, 2, MCP::MULADDv4i16_indexed_OP2);
7305 case AArch64::ADDv8i16:
7306 setVFound(AArch64::MULv8i16, 1, MCP::MULADDv8i16_OP1);
7307 setVFound(AArch64::MULv8i16, 2, MCP::MULADDv8i16_OP2);
7308 setVFound(AArch64::MULv8i16_indexed, 1, MCP::MULADDv8i16_indexed_OP1);
7309 setVFound(AArch64::MULv8i16_indexed, 2, MCP::MULADDv8i16_indexed_OP2);
7311 case AArch64::ADDv2i32:
7312 setVFound(AArch64::MULv2i32, 1, MCP::MULADDv2i32_OP1);
7313 setVFound(AArch64::MULv2i32, 2, MCP::MULADDv2i32_OP2);
7314 setVFound(AArch64::MULv2i32_indexed, 1, MCP::MULADDv2i32_indexed_OP1);
7315 setVFound(AArch64::MULv2i32_indexed, 2, MCP::MULADDv2i32_indexed_OP2);
7317 case AArch64::ADDv4i32:
7318 setVFound(AArch64::MULv4i32, 1, MCP::MULADDv4i32_OP1);
7319 setVFound(AArch64::MULv4i32, 2, MCP::MULADDv4i32_OP2);
7320 setVFound(AArch64::MULv4i32_indexed, 1, MCP::MULADDv4i32_indexed_OP1);
7321 setVFound(AArch64::MULv4i32_indexed, 2, MCP::MULADDv4i32_indexed_OP2);
7323 case AArch64::SUBv8i8:
7324 setVFound(AArch64::MULv8i8, 1, MCP::MULSUBv8i8_OP1);
7325 setVFound(AArch64::MULv8i8, 2, MCP::MULSUBv8i8_OP2);
7327 case AArch64::SUBv16i8:
7328 setVFound(AArch64::MULv16i8, 1, MCP::MULSUBv16i8_OP1);
7329 setVFound(AArch64::MULv16i8, 2, MCP::MULSUBv16i8_OP2);
7331 case AArch64::SUBv4i16:
7332 setVFound(AArch64::MULv4i16, 1, MCP::MULSUBv4i16_OP1);
7333 setVFound(AArch64::MULv4i16, 2, MCP::MULSUBv4i16_OP2);
7334 setVFound(AArch64::MULv4i16_indexed, 1, MCP::MULSUBv4i16_indexed_OP1);
7335 setVFound(AArch64::MULv4i16_indexed, 2, MCP::MULSUBv4i16_indexed_OP2);
7337 case AArch64::SUBv8i16:
7338 setVFound(AArch64::MULv8i16, 1, MCP::MULSUBv8i16_OP1);
7339 setVFound(AArch64::MULv8i16, 2, MCP::MULSUBv8i16_OP2);
7340 setVFound(AArch64::MULv8i16_indexed, 1, MCP::MULSUBv8i16_indexed_OP1);
7341 setVFound(AArch64::MULv8i16_indexed, 2, MCP::MULSUBv8i16_indexed_OP2);
7343 case AArch64::SUBv2i32:
7344 setVFound(AArch64::MULv2i32, 1, MCP::MULSUBv2i32_OP1);
7345 setVFound(AArch64::MULv2i32, 2, MCP::MULSUBv2i32_OP2);
7346 setVFound(AArch64::MULv2i32_indexed, 1, MCP::MULSUBv2i32_indexed_OP1);
7347 setVFound(AArch64::MULv2i32_indexed, 2, MCP::MULSUBv2i32_indexed_OP2);
7349 case AArch64::SUBv4i32:
7350 setVFound(AArch64::MULv4i32, 1, MCP::MULSUBv4i32_OP1);
7351 setVFound(AArch64::MULv4i32, 2, MCP::MULSUBv4i32_OP2);
7352 setVFound(AArch64::MULv4i32_indexed, 1, MCP::MULSUBv4i32_indexed_OP1);
7353 setVFound(AArch64::MULv4i32_indexed, 2, MCP::MULSUBv4i32_indexed_OP2);
7359bool AArch64InstrInfo::isAccumulationOpcode(
unsigned Opcode)
const {
7363 case AArch64::UABALB_ZZZ_D:
7364 case AArch64::UABALB_ZZZ_H:
7365 case AArch64::UABALB_ZZZ_S:
7366 case AArch64::UABALT_ZZZ_D:
7367 case AArch64::UABALT_ZZZ_H:
7368 case AArch64::UABALT_ZZZ_S:
7369 case AArch64::SABALB_ZZZ_D:
7370 case AArch64::SABALB_ZZZ_S:
7371 case AArch64::SABALB_ZZZ_H:
7372 case AArch64::SABALT_ZZZ_D:
7373 case AArch64::SABALT_ZZZ_S:
7374 case AArch64::SABALT_ZZZ_H:
7375 case AArch64::UABALv16i8_v8i16:
7376 case AArch64::UABALv2i32_v2i64:
7377 case AArch64::UABALv4i16_v4i32:
7378 case AArch64::UABALv4i32_v2i64:
7379 case AArch64::UABALv8i16_v4i32:
7380 case AArch64::UABALv8i8_v8i16:
7381 case AArch64::UABAv16i8:
7382 case AArch64::UABAv2i32:
7383 case AArch64::UABAv4i16:
7384 case AArch64::UABAv4i32:
7385 case AArch64::UABAv8i16:
7386 case AArch64::UABAv8i8:
7387 case AArch64::SABALv16i8_v8i16:
7388 case AArch64::SABALv2i32_v2i64:
7389 case AArch64::SABALv4i16_v4i32:
7390 case AArch64::SABALv4i32_v2i64:
7391 case AArch64::SABALv8i16_v4i32:
7392 case AArch64::SABALv8i8_v8i16:
7393 case AArch64::SABAv16i8:
7394 case AArch64::SABAv2i32:
7395 case AArch64::SABAv4i16:
7396 case AArch64::SABAv4i32:
7397 case AArch64::SABAv8i16:
7398 case AArch64::SABAv8i8:
7405unsigned AArch64InstrInfo::getAccumulationStartOpcode(
7406 unsigned AccumulationOpcode)
const {
7407 switch (AccumulationOpcode) {
7410 case AArch64::UABALB_ZZZ_D:
7411 return AArch64::UABDLB_ZZZ_D;
7412 case AArch64::UABALB_ZZZ_H:
7413 return AArch64::UABDLB_ZZZ_H;
7414 case AArch64::UABALB_ZZZ_S:
7415 return AArch64::UABDLB_ZZZ_S;
7416 case AArch64::UABALT_ZZZ_D:
7417 return AArch64::UABDLT_ZZZ_D;
7418 case AArch64::UABALT_ZZZ_H:
7419 return AArch64::UABDLT_ZZZ_H;
7420 case AArch64::UABALT_ZZZ_S:
7421 return AArch64::UABDLT_ZZZ_S;
7422 case AArch64::UABALv16i8_v8i16:
7423 return AArch64::UABDLv16i8_v8i16;
7424 case AArch64::UABALv2i32_v2i64:
7425 return AArch64::UABDLv2i32_v2i64;
7426 case AArch64::UABALv4i16_v4i32:
7427 return AArch64::UABDLv4i16_v4i32;
7428 case AArch64::UABALv4i32_v2i64:
7429 return AArch64::UABDLv4i32_v2i64;
7430 case AArch64::UABALv8i16_v4i32:
7431 return AArch64::UABDLv8i16_v4i32;
7432 case AArch64::UABALv8i8_v8i16:
7433 return AArch64::UABDLv8i8_v8i16;
7434 case AArch64::UABAv16i8:
7435 return AArch64::UABDv16i8;
7436 case AArch64::UABAv2i32:
7437 return AArch64::UABDv2i32;
7438 case AArch64::UABAv4i16:
7439 return AArch64::UABDv4i16;
7440 case AArch64::UABAv4i32:
7441 return AArch64::UABDv4i32;
7442 case AArch64::UABAv8i16:
7443 return AArch64::UABDv8i16;
7444 case AArch64::UABAv8i8:
7445 return AArch64::UABDv8i8;
7446 case AArch64::SABALB_ZZZ_D:
7447 return AArch64::SABDLB_ZZZ_D;
7448 case AArch64::SABALB_ZZZ_S:
7449 return AArch64::SABDLB_ZZZ_S;
7450 case AArch64::SABALB_ZZZ_H:
7451 return AArch64::SABDLB_ZZZ_H;
7452 case AArch64::SABALT_ZZZ_D:
7453 return AArch64::SABDLT_ZZZ_D;
7454 case AArch64::SABALT_ZZZ_S:
7455 return AArch64::SABDLT_ZZZ_S;
7456 case AArch64::SABALT_ZZZ_H:
7457 return AArch64::SABDLT_ZZZ_H;
7458 case AArch64::SABALv16i8_v8i16:
7459 return AArch64::SABDLv16i8_v8i16;
7460 case AArch64::SABALv2i32_v2i64:
7461 return AArch64::SABDLv2i32_v2i64;
7462 case AArch64::SABALv4i16_v4i32:
7463 return AArch64::SABDLv4i16_v4i32;
7464 case AArch64::SABALv4i32_v2i64:
7465 return AArch64::SABDLv4i32_v2i64;
7466 case AArch64::SABALv8i16_v4i32:
7467 return AArch64::SABDLv8i16_v4i32;
7468 case AArch64::SABALv8i8_v8i16:
7469 return AArch64::SABDLv8i8_v8i16;
7470 case AArch64::SABAv16i8:
7471 return AArch64::SABDv16i8;
7472 case AArch64::SABAv2i32:
7473 return AArch64::SABAv2i32;
7474 case AArch64::SABAv4i16:
7475 return AArch64::SABDv4i16;
7476 case AArch64::SABAv4i32:
7477 return AArch64::SABDv4i32;
7478 case AArch64::SABAv8i16:
7479 return AArch64::SABDv8i16;
7480 case AArch64::SABAv8i8:
7481 return AArch64::SABDv8i8;
7497 auto Match = [&](
int Opcode,
int Operand,
unsigned Pattern) ->
bool {
7509 assert(
false &&
"Unsupported FP instruction in combiner\n");
7511 case AArch64::FADDHrr:
7513 "FADDHrr does not have register operands");
7515 Found = Match(AArch64::FMULHrr, 1, MCP::FMULADDH_OP1);
7516 Found |= Match(AArch64::FMULHrr, 2, MCP::FMULADDH_OP2);
7518 case AArch64::FADDSrr:
7520 "FADDSrr does not have register operands");
7522 Found |= Match(AArch64::FMULSrr, 1, MCP::FMULADDS_OP1) ||
7523 Match(AArch64::FMULv1i32_indexed, 1, MCP::FMLAv1i32_indexed_OP1);
7525 Found |= Match(AArch64::FMULSrr, 2, MCP::FMULADDS_OP2) ||
7526 Match(AArch64::FMULv1i32_indexed, 2, MCP::FMLAv1i32_indexed_OP2);
7528 case AArch64::FADDDrr:
7529 Found |= Match(AArch64::FMULDrr, 1, MCP::FMULADDD_OP1) ||
7530 Match(AArch64::FMULv1i64_indexed, 1, MCP::FMLAv1i64_indexed_OP1);
7532 Found |= Match(AArch64::FMULDrr, 2, MCP::FMULADDD_OP2) ||
7533 Match(AArch64::FMULv1i64_indexed, 2, MCP::FMLAv1i64_indexed_OP2);
7535 case AArch64::FADDv4f16:
7536 Found |= Match(AArch64::FMULv4i16_indexed, 1, MCP::FMLAv4i16_indexed_OP1) ||
7537 Match(AArch64::FMULv4f16, 1, MCP::FMLAv4f16_OP1);
7539 Found |= Match(AArch64::FMULv4i16_indexed, 2, MCP::FMLAv4i16_indexed_OP2) ||
7540 Match(AArch64::FMULv4f16, 2, MCP::FMLAv4f16_OP2);
7542 case AArch64::FADDv8f16:
7543 Found |= Match(AArch64::FMULv8i16_indexed, 1, MCP::FMLAv8i16_indexed_OP1) ||
7544 Match(AArch64::FMULv8f16, 1, MCP::FMLAv8f16_OP1);
7546 Found |= Match(AArch64::FMULv8i16_indexed, 2, MCP::FMLAv8i16_indexed_OP2) ||
7547 Match(AArch64::FMULv8f16, 2, MCP::FMLAv8f16_OP2);
7549 case AArch64::FADDv2f32:
7550 Found |= Match(AArch64::FMULv2i32_indexed, 1, MCP::FMLAv2i32_indexed_OP1) ||
7551 Match(AArch64::FMULv2f32, 1, MCP::FMLAv2f32_OP1);
7553 Found |= Match(AArch64::FMULv2i32_indexed, 2, MCP::FMLAv2i32_indexed_OP2) ||
7554 Match(AArch64::FMULv2f32, 2, MCP::FMLAv2f32_OP2);
7556 case AArch64::FADDv2f64:
7557 Found |= Match(AArch64::FMULv2i64_indexed, 1, MCP::FMLAv2i64_indexed_OP1) ||
7558 Match(AArch64::FMULv2f64, 1, MCP::FMLAv2f64_OP1);
7560 Found |= Match(AArch64::FMULv2i64_indexed, 2, MCP::FMLAv2i64_indexed_OP2) ||
7561 Match(AArch64::FMULv2f64, 2, MCP::FMLAv2f64_OP2);
7563 case AArch64::FADDv4f32:
7564 Found |= Match(AArch64::FMULv4i32_indexed, 1, MCP::FMLAv4i32_indexed_OP1) ||
7565 Match(AArch64::FMULv4f32, 1, MCP::FMLAv4f32_OP1);
7567 Found |= Match(AArch64::FMULv4i32_indexed, 2, MCP::FMLAv4i32_indexed_OP2) ||
7568 Match(AArch64::FMULv4f32, 2, MCP::FMLAv4f32_OP2);
7570 case AArch64::FSUBHrr:
7571 Found = Match(AArch64::FMULHrr, 1, MCP::FMULSUBH_OP1);
7572 Found |= Match(AArch64::FMULHrr, 2, MCP::FMULSUBH_OP2);
7573 Found |= Match(AArch64::FNMULHrr, 1, MCP::FNMULSUBH_OP1);
7575 case AArch64::FSUBSrr:
7576 Found = Match(AArch64::FMULSrr, 1, MCP::FMULSUBS_OP1);
7578 Found |= Match(AArch64::FMULSrr, 2, MCP::FMULSUBS_OP2) ||
7579 Match(AArch64::FMULv1i32_indexed, 2, MCP::FMLSv1i32_indexed_OP2);
7581 Found |= Match(AArch64::FNMULSrr, 1, MCP::FNMULSUBS_OP1);
7583 case AArch64::FSUBDrr:
7584 Found = Match(AArch64::FMULDrr, 1, MCP::FMULSUBD_OP1);
7586 Found |= Match(AArch64::FMULDrr, 2, MCP::FMULSUBD_OP2) ||
7587 Match(AArch64::FMULv1i64_indexed, 2, MCP::FMLSv1i64_indexed_OP2);
7589 Found |= Match(AArch64::FNMULDrr, 1, MCP::FNMULSUBD_OP1);
7591 case AArch64::FSUBv4f16:
7592 Found |= Match(AArch64::FMULv4i16_indexed, 2, MCP::FMLSv4i16_indexed_OP2) ||
7593 Match(AArch64::FMULv4f16, 2, MCP::FMLSv4f16_OP2);
7595 Found |= Match(AArch64::FMULv4i16_indexed, 1, MCP::FMLSv4i16_indexed_OP1) ||
7596 Match(AArch64::FMULv4f16, 1, MCP::FMLSv4f16_OP1);
7598 case AArch64::FSUBv8f16:
7599 Found |= Match(AArch64::FMULv8i16_indexed, 2, MCP::FMLSv8i16_indexed_OP2) ||
7600 Match(AArch64::FMULv8f16, 2, MCP::FMLSv8f16_OP2);
7602 Found |= Match(AArch64::FMULv8i16_indexed, 1, MCP::FMLSv8i16_indexed_OP1) ||
7603 Match(AArch64::FMULv8f16, 1, MCP::FMLSv8f16_OP1);
7605 case AArch64::FSUBv2f32:
7606 Found |= Match(AArch64::FMULv2i32_indexed, 2, MCP::FMLSv2i32_indexed_OP2) ||
7607 Match(AArch64::FMULv2f32, 2, MCP::FMLSv2f32_OP2);
7609 Found |= Match(AArch64::FMULv2i32_indexed, 1, MCP::FMLSv2i32_indexed_OP1) ||
7610 Match(AArch64::FMULv2f32, 1, MCP::FMLSv2f32_OP1);
7612 case AArch64::FSUBv2f64:
7613 Found |= Match(AArch64::FMULv2i64_indexed, 2, MCP::FMLSv2i64_indexed_OP2) ||
7614 Match(AArch64::FMULv2f64, 2, MCP::FMLSv2f64_OP2);
7616 Found |= Match(AArch64::FMULv2i64_indexed, 1, MCP::FMLSv2i64_indexed_OP1) ||
7617 Match(AArch64::FMULv2f64, 1, MCP::FMLSv2f64_OP1);
7619 case AArch64::FSUBv4f32:
7620 Found |= Match(AArch64::FMULv4i32_indexed, 2, MCP::FMLSv4i32_indexed_OP2) ||
7621 Match(AArch64::FMULv4f32, 2, MCP::FMLSv4f32_OP2);
7623 Found |= Match(AArch64::FMULv4i32_indexed, 1, MCP::FMLSv4i32_indexed_OP1) ||
7624 Match(AArch64::FMULv4f32, 1, MCP::FMLSv4f32_OP1);
7635 auto Match = [&](
unsigned Opcode,
int Operand,
unsigned Pattern) ->
bool {
7642 if (
MI &&
MI->getOpcode() == TargetOpcode::COPY &&
7643 MI->getOperand(1).getReg().isVirtual())
7644 MI =
MRI.getUniqueVRegDef(
MI->getOperand(1).getReg());
7645 if (
MI &&
MI->getOpcode() == Opcode) {
7657 case AArch64::FMULv2f32:
7658 Found = Match(AArch64::DUPv2i32lane, 1, MCP::FMULv2i32_indexed_OP1);
7659 Found |= Match(AArch64::DUPv2i32lane, 2, MCP::FMULv2i32_indexed_OP2);
7661 case AArch64::FMULv2f64:
7662 Found = Match(AArch64::DUPv2i64lane, 1, MCP::FMULv2i64_indexed_OP1);
7663 Found |= Match(AArch64::DUPv2i64lane, 2, MCP::FMULv2i64_indexed_OP2);
7665 case AArch64::FMULv4f16:
7666 Found = Match(AArch64::DUPv4i16lane, 1, MCP::FMULv4i16_indexed_OP1);
7667 Found |= Match(AArch64::DUPv4i16lane, 2, MCP::FMULv4i16_indexed_OP2);
7669 case AArch64::FMULv4f32:
7670 Found = Match(AArch64::DUPv4i32lane, 1, MCP::FMULv4i32_indexed_OP1);
7671 Found |= Match(AArch64::DUPv4i32lane, 2, MCP::FMULv4i32_indexed_OP2);
7673 case AArch64::FMULv8f16:
7674 Found = Match(AArch64::DUPv8i16lane, 1, MCP::FMULv8i16_indexed_OP1);
7675 Found |= Match(AArch64::DUPv8i16lane, 2, MCP::FMULv8i16_indexed_OP2);
7688 auto Match = [&](
unsigned Opcode,
unsigned Pattern) ->
bool {
7691 if (
MI !=
nullptr && (
MI->getOpcode() == Opcode) &&
7692 MRI.hasOneNonDBGUse(
MI->getOperand(0).getReg()) &&
7706 case AArch64::FNEGDr:
7708 case AArch64::FNEGSr:
7840 case AArch64::SUBWrr:
7841 case AArch64::SUBSWrr:
7842 case AArch64::SUBXrr:
7843 case AArch64::SUBSXrr:
7888 unsigned LoadLaneOpCode,
unsigned NumLanes) {
7911 while (!RemainingLanes.
empty() && CurrInstr &&
7912 CurrInstr->getOpcode() == LoadLaneOpCode &&
7913 MRI.hasOneNonDBGUse(CurrInstr->getOperand(0).getReg()) &&
7914 CurrInstr->getNumOperands() == 4) {
7915 RemainingLanes.
erase(CurrInstr->getOperand(2).getImm());
7917 CurrInstr =
MRI.getUniqueVRegDef(CurrInstr->getOperand(1).getReg());
7921 if (!RemainingLanes.
empty())
7925 if (CurrInstr->getOpcode() != TargetOpcode::SUBREG_TO_REG)
7929 auto Lane0LoadReg = CurrInstr->getOperand(2).getReg();
7930 unsigned SingleLaneSizeInBits = 128 / NumLanes;
7931 if (
TRI->getRegSizeInBits(Lane0LoadReg,
MRI) != SingleLaneSizeInBits)
7935 if (!
MRI.hasOneNonDBGUse(Lane0LoadReg))
7938 LoadInstrs.
push_back(
MRI.getUniqueVRegDef(Lane0LoadReg));
7947 RemainingLoadInstrs.
insert(LoadInstrs.
begin(), LoadInstrs.
end());
7950 for (; MBBItr !=
MBB->begin() && RemainingSteps > 0 &&
7951 !RemainingLoadInstrs.
empty();
7952 --MBBItr, --RemainingSteps) {
7956 RemainingLoadInstrs.
erase(&CurrInstr);
7966 if (RemainingSteps == 0 && !RemainingLoadInstrs.
empty())
7992 case AArch64::LD1i32:
7994 case AArch64::LD1i16:
7996 case AArch64::LD1i8:
8012 unsigned Pattern,
unsigned NumLanes) {
8020 for (
unsigned i = 0; i < NumLanes - 1; ++i) {
8028 return A->getOperand(2).getImm() >
B->getOperand(2).getImm();
8033 MRI.getUniqueVRegDef(SubregToReg->getOperand(2).getReg()));
8034 auto LoadToLaneInstrsAscending =
llvm::reverse(LoadToLaneInstrs);
8040 auto CreateLD1Instruction = [&](
MachineInstr *OriginalInstr,
8041 Register SrcRegister,
unsigned Lane,
8043 bool OffsetRegisterKillState) {
8044 auto NewRegister =
MRI.createVirtualRegister(FPR128RegClass);
8051 InstrIdxForVirtReg.
insert(std::make_pair(NewRegister, InsInstrs.
size()));
8052 InsInstrs.
push_back(LoadIndexIntoRegister);
8058 auto CreateLDRInstruction = [&](
unsigned NumLanes,
Register DestReg,
8064 Opcode = AArch64::LDRSui;
8067 Opcode = AArch64::LDRHui;
8070 Opcode = AArch64::LDRBui;
8074 "Got unsupported number of lanes in machine-combiner gather pattern");
8083 auto LanesToLoadToReg0 =
8085 LoadToLaneInstrsAscending.begin() + NumLanes / 2);
8086 Register PrevReg = SubregToReg->getOperand(0).getReg();
8088 const MachineOperand &OffsetRegOperand = LoadInstr->getOperand(3);
8089 PrevReg = CreateLD1Instruction(LoadInstr, PrevReg, Index + 1,
8090 OffsetRegOperand.
getReg(),
8091 OffsetRegOperand.
isKill());
8098 MachineInstr *Lane0Load = *LoadToLaneInstrsAscending.begin();
8100 *std::next(LoadToLaneInstrsAscending.begin(), NumLanes / 2);
8101 Register DestRegForMiddleIndex =
MRI.createVirtualRegister(
8107 CreateLDRInstruction(NumLanes, DestRegForMiddleIndex,
8108 OriginalSplitToLoadOffsetOperand.
getReg(),
8109 OriginalSplitToLoadOffsetOperand.
isKill());
8111 InstrIdxForVirtReg.
insert(
8112 std::make_pair(DestRegForMiddleIndex, InsInstrs.
size()));
8113 InsInstrs.
push_back(MiddleIndexLoadInstr);
8117 Register DestRegForSubregToReg =
MRI.createVirtualRegister(FPR128RegClass);
8118 unsigned SubregType;
8121 SubregType = AArch64::ssub;
8124 SubregType = AArch64::hsub;
8127 SubregType = AArch64::bsub;
8131 "Got invalid NumLanes for machine-combiner gather pattern");
8134 auto SubRegToRegInstr =
8136 DestRegForSubregToReg)
8140 InstrIdxForVirtReg.
insert(
8141 std::make_pair(DestRegForSubregToReg, InsInstrs.
size()));
8145 auto LanesToLoadToReg1 =
8147 LoadToLaneInstrsAscending.end());
8148 PrevReg = SubRegToRegInstr->getOperand(0).getReg();
8150 const MachineOperand &OffsetRegOperand = LoadInstr->getOperand(3);
8151 PrevReg = CreateLD1Instruction(LoadInstr, PrevReg, Index + 1,
8152 OffsetRegOperand.
getReg(),
8153 OffsetRegOperand.
isKill());
8156 if (Index == NumLanes / 2 - 2) {
8191bool AArch64InstrInfo::getMachineCombinerPatterns(
8193 bool DoRegPressureReduce)
const {
8214 DoRegPressureReduce);
8243 const Register *ReplacedAddend =
nullptr) {
8244 assert(IdxMulOpd == 1 || IdxMulOpd == 2);
8246 unsigned IdxOtherOpd = IdxMulOpd == 1 ? 2 : 1;
8249 Register SrcReg0 = MUL->getOperand(1).getReg();
8250 bool Src0IsKill = MUL->getOperand(1).isKill();
8251 Register SrcReg1 = MUL->getOperand(2).getReg();
8252 bool Src1IsKill = MUL->getOperand(2).isKill();
8256 if (ReplacedAddend) {
8258 SrcReg2 = *ReplacedAddend;
8266 MRI.constrainRegClass(ResultReg, RC);
8268 MRI.constrainRegClass(SrcReg0, RC);
8270 MRI.constrainRegClass(SrcReg1, RC);
8272 MRI.constrainRegClass(SrcReg2, RC);
8285 .
addImm(MUL->getOperand(3).getImm());
8292 assert(
false &&
"Invalid FMA instruction kind \n");
8306 if (AArch64::FPR32RegClass.hasSubClassEq(RC))
8307 Opc = AArch64::FNMADDSrrr;
8308 else if (AArch64::FPR64RegClass.hasSubClassEq(RC))
8309 Opc = AArch64::FNMADDDrrr;
8321 MRI.constrainRegClass(ResultReg, RC);
8323 MRI.constrainRegClass(SrcReg0, RC);
8325 MRI.constrainRegClass(SrcReg1, RC);
8327 MRI.constrainRegClass(SrcReg2, RC);
8343 unsigned IdxDupOp,
unsigned MulOpc,
8345 assert(((IdxDupOp == 1) || (IdxDupOp == 2)) &&
8346 "Invalid index of FMUL operand");
8354 if (Dup->
getOpcode() == TargetOpcode::COPY)
8358 MRI.clearKillFlags(DupSrcReg);
8359 MRI.constrainRegClass(DupSrcReg, RC);
8363 unsigned IdxMulOp = IdxDupOp == 1 ? 2 : 1;
8404 InstrIdxForVirtReg.
insert(std::make_pair(NewVR, 0));
8419 genNeg(MF,
MRI,
TII, Root, InsInstrs, InstrIdxForVirtReg, MnegOpc, RC);
8446 genNeg(MF,
MRI,
TII, Root, InsInstrs, InstrIdxForVirtReg, MnegOpc, RC);
8474 unsigned IdxMulOpd,
unsigned MaddOpc,
unsigned VR,
8476 assert(IdxMulOpd == 1 || IdxMulOpd == 2);
8480 Register SrcReg0 = MUL->getOperand(1).getReg();
8481 bool Src0IsKill = MUL->getOperand(1).isKill();
8482 Register SrcReg1 = MUL->getOperand(2).getReg();
8483 bool Src1IsKill = MUL->getOperand(2).isKill();
8486 MRI.constrainRegClass(ResultReg, RC);
8488 MRI.constrainRegClass(SrcReg0, RC);
8490 MRI.constrainRegClass(SrcReg1, RC);
8492 MRI.constrainRegClass(VR, RC);
8513 assert(IdxOpd1 == 1 || IdxOpd1 == 2);
8514 unsigned IdxOtherOpd = IdxOpd1 == 1 ? 2 : 1;
8528 if (Opcode == AArch64::SUBSWrr)
8529 Opcode = AArch64::SUBWrr;
8530 else if (Opcode == AArch64::SUBSXrr)
8531 Opcode = AArch64::SUBXrr;
8533 assert((Opcode == AArch64::SUBWrr || Opcode == AArch64::SUBXrr) &&
8534 "Unexpected instruction opcode.");
8551 InstrIdxForVirtReg.
insert(std::make_pair(NewVR, 0));
8558unsigned AArch64InstrInfo::getReduceOpcodeForAccumulator(
8559 unsigned int AccumulatorOpCode)
const {
8560 switch (AccumulatorOpCode) {
8561 case AArch64::UABALB_ZZZ_D:
8562 case AArch64::SABALB_ZZZ_D:
8563 case AArch64::UABALT_ZZZ_D:
8564 case AArch64::SABALT_ZZZ_D:
8565 return AArch64::ADD_ZZZ_D;
8566 case AArch64::UABALB_ZZZ_H:
8567 case AArch64::SABALB_ZZZ_H:
8568 case AArch64::UABALT_ZZZ_H:
8569 case AArch64::SABALT_ZZZ_H:
8570 return AArch64::ADD_ZZZ_H;
8571 case AArch64::UABALB_ZZZ_S:
8572 case AArch64::SABALB_ZZZ_S:
8573 case AArch64::UABALT_ZZZ_S:
8574 case AArch64::SABALT_ZZZ_S:
8575 return AArch64::ADD_ZZZ_S;
8576 case AArch64::UABALv16i8_v8i16:
8577 case AArch64::SABALv8i8_v8i16:
8578 case AArch64::SABAv8i16:
8579 case AArch64::UABAv8i16:
8580 return AArch64::ADDv8i16;
8581 case AArch64::SABALv2i32_v2i64:
8582 case AArch64::UABALv2i32_v2i64:
8583 case AArch64::SABALv4i32_v2i64:
8584 return AArch64::ADDv2i64;
8585 case AArch64::UABALv4i16_v4i32:
8586 case AArch64::SABALv4i16_v4i32:
8587 case AArch64::SABALv8i16_v4i32:
8588 case AArch64::SABAv4i32:
8589 case AArch64::UABAv4i32:
8590 return AArch64::ADDv4i32;
8591 case AArch64::UABALv4i32_v2i64:
8592 return AArch64::ADDv2i64;
8593 case AArch64::UABALv8i16_v4i32:
8594 return AArch64::ADDv4i32;
8595 case AArch64::UABALv8i8_v8i16:
8596 case AArch64::SABALv16i8_v8i16:
8597 return AArch64::ADDv8i16;
8598 case AArch64::UABAv16i8:
8599 case AArch64::SABAv16i8:
8600 return AArch64::ADDv16i8;
8601 case AArch64::UABAv4i16:
8602 case AArch64::SABAv4i16:
8603 return AArch64::ADDv4i16;
8604 case AArch64::UABAv2i32:
8605 case AArch64::SABAv2i32:
8606 return AArch64::ADDv2i32;
8607 case AArch64::UABAv8i8:
8608 case AArch64::SABAv8i8:
8609 return AArch64::ADDv8i8;
8618void AArch64InstrInfo::genAlternativeCodeSequence(
8628 MachineInstr *
MUL =
nullptr;
8629 const TargetRegisterClass *RC;
8635 DelInstrs, InstrIdxForVirtReg);
8641 InstrIdxForVirtReg);
8647 InstrIdxForVirtReg);
8656 Opc = AArch64::MADDWrrr;
8657 RC = &AArch64::GPR32RegClass;
8659 Opc = AArch64::MADDXrrr;
8660 RC = &AArch64::GPR64RegClass;
8671 Opc = AArch64::MADDWrrr;
8672 RC = &AArch64::GPR32RegClass;
8674 Opc = AArch64::MADDXrrr;
8675 RC = &AArch64::GPR64RegClass;
8688 const TargetRegisterClass *RC;
8689 unsigned BitSize, MovImm;
8692 MovImm = AArch64::MOVi32imm;
8693 RC = &AArch64::GPR32spRegClass;
8695 Opc = AArch64::MADDWrrr;
8696 RC = &AArch64::GPR32RegClass;
8698 MovImm = AArch64::MOVi64imm;
8699 RC = &AArch64::GPR64spRegClass;
8701 Opc = AArch64::MADDXrrr;
8702 RC = &AArch64::GPR64RegClass;
8713 uint64_t UImm =
SignExtend64(IsSub ? -Imm : Imm, BitSize);
8717 if (Insn.
size() != 1)
8719 MachineInstrBuilder MIB1 =
8720 BuildMI(MF, MIMetadata(Root),
TII->get(MovImm), NewVR)
8721 .
addImm(IsSub ? -Imm : Imm);
8723 InstrIdxForVirtReg.
insert(std::make_pair(NewVR, 0));
8734 const TargetRegisterClass *SubRC;
8735 unsigned SubOpc, ZeroReg;
8737 SubOpc = AArch64::SUBWrr;
8738 SubRC = &AArch64::GPR32spRegClass;
8739 ZeroReg = AArch64::WZR;
8740 Opc = AArch64::MADDWrrr;
8741 RC = &AArch64::GPR32RegClass;
8743 SubOpc = AArch64::SUBXrr;
8744 SubRC = &AArch64::GPR64spRegClass;
8745 ZeroReg = AArch64::XZR;
8746 Opc = AArch64::MADDXrrr;
8747 RC = &AArch64::GPR64RegClass;
8749 Register NewVR =
MRI.createVirtualRegister(SubRC);
8751 MachineInstrBuilder MIB1 =
8752 BuildMI(MF, MIMetadata(Root),
TII->get(SubOpc), NewVR)
8756 InstrIdxForVirtReg.
insert(std::make_pair(NewVR, 0));
8767 Opc = AArch64::MSUBWrrr;
8768 RC = &AArch64::GPR32RegClass;
8770 Opc = AArch64::MSUBXrrr;
8771 RC = &AArch64::GPR64RegClass;
8776 Opc = AArch64::MLAv8i8;
8777 RC = &AArch64::FPR64RegClass;
8781 Opc = AArch64::MLAv8i8;
8782 RC = &AArch64::FPR64RegClass;
8786 Opc = AArch64::MLAv16i8;
8787 RC = &AArch64::FPR128RegClass;
8791 Opc = AArch64::MLAv16i8;
8792 RC = &AArch64::FPR128RegClass;
8796 Opc = AArch64::MLAv4i16;
8797 RC = &AArch64::FPR64RegClass;
8801 Opc = AArch64::MLAv4i16;
8802 RC = &AArch64::FPR64RegClass;
8806 Opc = AArch64::MLAv8i16;
8807 RC = &AArch64::FPR128RegClass;
8811 Opc = AArch64::MLAv8i16;
8812 RC = &AArch64::FPR128RegClass;
8816 Opc = AArch64::MLAv2i32;
8817 RC = &AArch64::FPR64RegClass;
8821 Opc = AArch64::MLAv2i32;
8822 RC = &AArch64::FPR64RegClass;
8826 Opc = AArch64::MLAv4i32;
8827 RC = &AArch64::FPR128RegClass;
8831 Opc = AArch64::MLAv4i32;
8832 RC = &AArch64::FPR128RegClass;
8837 Opc = AArch64::MLAv8i8;
8838 RC = &AArch64::FPR64RegClass;
8840 InstrIdxForVirtReg, 1,
Opc, AArch64::NEGv8i8,
8844 Opc = AArch64::MLSv8i8;
8845 RC = &AArch64::FPR64RegClass;
8849 Opc = AArch64::MLAv16i8;
8850 RC = &AArch64::FPR128RegClass;
8852 InstrIdxForVirtReg, 1,
Opc, AArch64::NEGv16i8,
8856 Opc = AArch64::MLSv16i8;
8857 RC = &AArch64::FPR128RegClass;
8861 Opc = AArch64::MLAv4i16;
8862 RC = &AArch64::FPR64RegClass;
8864 InstrIdxForVirtReg, 1,
Opc, AArch64::NEGv4i16,
8868 Opc = AArch64::MLSv4i16;
8869 RC = &AArch64::FPR64RegClass;
8873 Opc = AArch64::MLAv8i16;
8874 RC = &AArch64::FPR128RegClass;
8876 InstrIdxForVirtReg, 1,
Opc, AArch64::NEGv8i16,
8880 Opc = AArch64::MLSv8i16;
8881 RC = &AArch64::FPR128RegClass;
8885 Opc = AArch64::MLAv2i32;
8886 RC = &AArch64::FPR64RegClass;
8888 InstrIdxForVirtReg, 1,
Opc, AArch64::NEGv2i32,
8892 Opc = AArch64::MLSv2i32;
8893 RC = &AArch64::FPR64RegClass;
8897 Opc = AArch64::MLAv4i32;
8898 RC = &AArch64::FPR128RegClass;
8900 InstrIdxForVirtReg, 1,
Opc, AArch64::NEGv4i32,
8904 Opc = AArch64::MLSv4i32;
8905 RC = &AArch64::FPR128RegClass;
8910 Opc = AArch64::MLAv4i16_indexed;
8911 RC = &AArch64::FPR64RegClass;
8915 Opc = AArch64::MLAv4i16_indexed;
8916 RC = &AArch64::FPR64RegClass;
8920 Opc = AArch64::MLAv8i16_indexed;
8921 RC = &AArch64::FPR128RegClass;
8925 Opc = AArch64::MLAv8i16_indexed;
8926 RC = &AArch64::FPR128RegClass;
8930 Opc = AArch64::MLAv2i32_indexed;
8931 RC = &AArch64::FPR64RegClass;
8935 Opc = AArch64::MLAv2i32_indexed;
8936 RC = &AArch64::FPR64RegClass;
8940 Opc = AArch64::MLAv4i32_indexed;
8941 RC = &AArch64::FPR128RegClass;
8945 Opc = AArch64::MLAv4i32_indexed;
8946 RC = &AArch64::FPR128RegClass;
8951 Opc = AArch64::MLAv4i16_indexed;
8952 RC = &AArch64::FPR64RegClass;
8954 InstrIdxForVirtReg, 1,
Opc, AArch64::NEGv4i16,
8958 Opc = AArch64::MLSv4i16_indexed;
8959 RC = &AArch64::FPR64RegClass;
8963 Opc = AArch64::MLAv8i16_indexed;
8964 RC = &AArch64::FPR128RegClass;
8966 InstrIdxForVirtReg, 1,
Opc, AArch64::NEGv8i16,
8970 Opc = AArch64::MLSv8i16_indexed;
8971 RC = &AArch64::FPR128RegClass;
8975 Opc = AArch64::MLAv2i32_indexed;
8976 RC = &AArch64::FPR64RegClass;
8978 InstrIdxForVirtReg, 1,
Opc, AArch64::NEGv2i32,
8982 Opc = AArch64::MLSv2i32_indexed;
8983 RC = &AArch64::FPR64RegClass;
8987 Opc = AArch64::MLAv4i32_indexed;
8988 RC = &AArch64::FPR128RegClass;
8990 InstrIdxForVirtReg, 1,
Opc, AArch64::NEGv4i32,
8994 Opc = AArch64::MLSv4i32_indexed;
8995 RC = &AArch64::FPR128RegClass;
9001 Opc = AArch64::FMADDHrrr;
9002 RC = &AArch64::FPR16RegClass;
9006 Opc = AArch64::FMADDSrrr;
9007 RC = &AArch64::FPR32RegClass;
9011 Opc = AArch64::FMADDDrrr;
9012 RC = &AArch64::FPR64RegClass;
9017 Opc = AArch64::FMADDHrrr;
9018 RC = &AArch64::FPR16RegClass;
9022 Opc = AArch64::FMADDSrrr;
9023 RC = &AArch64::FPR32RegClass;
9027 Opc = AArch64::FMADDDrrr;
9028 RC = &AArch64::FPR64RegClass;
9033 Opc = AArch64::FMLAv1i32_indexed;
9034 RC = &AArch64::FPR32RegClass;
9039 Opc = AArch64::FMLAv1i32_indexed;
9040 RC = &AArch64::FPR32RegClass;
9046 Opc = AArch64::FMLAv1i64_indexed;
9047 RC = &AArch64::FPR64RegClass;
9052 Opc = AArch64::FMLAv1i64_indexed;
9053 RC = &AArch64::FPR64RegClass;
9059 RC = &AArch64::FPR64RegClass;
9060 Opc = AArch64::FMLAv4i16_indexed;
9065 RC = &AArch64::FPR64RegClass;
9066 Opc = AArch64::FMLAv4f16;
9071 RC = &AArch64::FPR64RegClass;
9072 Opc = AArch64::FMLAv4i16_indexed;
9077 RC = &AArch64::FPR64RegClass;
9078 Opc = AArch64::FMLAv4f16;
9085 RC = &AArch64::FPR64RegClass;
9087 Opc = AArch64::FMLAv2i32_indexed;
9091 Opc = AArch64::FMLAv2f32;
9098 RC = &AArch64::FPR64RegClass;
9100 Opc = AArch64::FMLAv2i32_indexed;
9104 Opc = AArch64::FMLAv2f32;
9111 RC = &AArch64::FPR128RegClass;
9112 Opc = AArch64::FMLAv8i16_indexed;
9117 RC = &AArch64::FPR128RegClass;
9118 Opc = AArch64::FMLAv8f16;
9123 RC = &AArch64::FPR128RegClass;
9124 Opc = AArch64::FMLAv8i16_indexed;
9129 RC = &AArch64::FPR128RegClass;
9130 Opc = AArch64::FMLAv8f16;
9137 RC = &AArch64::FPR128RegClass;
9139 Opc = AArch64::FMLAv2i64_indexed;
9143 Opc = AArch64::FMLAv2f64;
9150 RC = &AArch64::FPR128RegClass;
9152 Opc = AArch64::FMLAv2i64_indexed;
9156 Opc = AArch64::FMLAv2f64;
9164 RC = &AArch64::FPR128RegClass;
9166 Opc = AArch64::FMLAv4i32_indexed;
9170 Opc = AArch64::FMLAv4f32;
9178 RC = &AArch64::FPR128RegClass;
9180 Opc = AArch64::FMLAv4i32_indexed;
9184 Opc = AArch64::FMLAv4f32;
9191 Opc = AArch64::FNMSUBHrrr;
9192 RC = &AArch64::FPR16RegClass;
9196 Opc = AArch64::FNMSUBSrrr;
9197 RC = &AArch64::FPR32RegClass;
9201 Opc = AArch64::FNMSUBDrrr;
9202 RC = &AArch64::FPR64RegClass;
9207 Opc = AArch64::FNMADDHrrr;
9208 RC = &AArch64::FPR16RegClass;
9212 Opc = AArch64::FNMADDSrrr;
9213 RC = &AArch64::FPR32RegClass;
9217 Opc = AArch64::FNMADDDrrr;
9218 RC = &AArch64::FPR64RegClass;
9223 Opc = AArch64::FMSUBHrrr;
9224 RC = &AArch64::FPR16RegClass;
9228 Opc = AArch64::FMSUBSrrr;
9229 RC = &AArch64::FPR32RegClass;
9233 Opc = AArch64::FMSUBDrrr;
9234 RC = &AArch64::FPR64RegClass;
9239 Opc = AArch64::FMLSv1i32_indexed;
9240 RC = &AArch64::FPR32RegClass;
9246 Opc = AArch64::FMLSv1i64_indexed;
9247 RC = &AArch64::FPR64RegClass;
9254 RC = &AArch64::FPR64RegClass;
9256 MachineInstrBuilder MIB1 =
9257 BuildMI(MF, MIMetadata(Root),
TII->get(AArch64::FNEGv4f16), NewVR)
9260 InstrIdxForVirtReg.
insert(std::make_pair(NewVR, 0));
9262 Opc = AArch64::FMLAv4f16;
9266 Opc = AArch64::FMLAv4i16_indexed;
9273 RC = &AArch64::FPR64RegClass;
9274 Opc = AArch64::FMLSv4f16;
9279 RC = &AArch64::FPR64RegClass;
9280 Opc = AArch64::FMLSv4i16_indexed;
9287 RC = &AArch64::FPR64RegClass;
9289 Opc = AArch64::FMLSv2i32_indexed;
9293 Opc = AArch64::FMLSv2f32;
9301 RC = &AArch64::FPR128RegClass;
9303 MachineInstrBuilder MIB1 =
9304 BuildMI(MF, MIMetadata(Root),
TII->get(AArch64::FNEGv8f16), NewVR)
9307 InstrIdxForVirtReg.
insert(std::make_pair(NewVR, 0));
9309 Opc = AArch64::FMLAv8f16;
9313 Opc = AArch64::FMLAv8i16_indexed;
9320 RC = &AArch64::FPR128RegClass;
9321 Opc = AArch64::FMLSv8f16;
9326 RC = &AArch64::FPR128RegClass;
9327 Opc = AArch64::FMLSv8i16_indexed;
9334 RC = &AArch64::FPR128RegClass;
9336 Opc = AArch64::FMLSv2i64_indexed;
9340 Opc = AArch64::FMLSv2f64;
9348 RC = &AArch64::FPR128RegClass;
9350 Opc = AArch64::FMLSv4i32_indexed;
9354 Opc = AArch64::FMLSv4f32;
9361 RC = &AArch64::FPR64RegClass;
9363 MachineInstrBuilder MIB1 =
9364 BuildMI(MF, MIMetadata(Root),
TII->get(AArch64::FNEGv2f32), NewVR)
9367 InstrIdxForVirtReg.
insert(std::make_pair(NewVR, 0));
9369 Opc = AArch64::FMLAv2i32_indexed;
9373 Opc = AArch64::FMLAv2f32;
9381 RC = &AArch64::FPR128RegClass;
9383 MachineInstrBuilder MIB1 =
9384 BuildMI(MF, MIMetadata(Root),
TII->get(AArch64::FNEGv4f32), NewVR)
9387 InstrIdxForVirtReg.
insert(std::make_pair(NewVR, 0));
9389 Opc = AArch64::FMLAv4i32_indexed;
9393 Opc = AArch64::FMLAv4f32;
9401 RC = &AArch64::FPR128RegClass;
9403 MachineInstrBuilder MIB1 =
9404 BuildMI(MF, MIMetadata(Root),
TII->get(AArch64::FNEGv2f64), NewVR)
9407 InstrIdxForVirtReg.
insert(std::make_pair(NewVR, 0));
9409 Opc = AArch64::FMLAv2i64_indexed;
9413 Opc = AArch64::FMLAv2f64;
9425 &AArch64::FPR128RegClass,
MRI);
9434 &AArch64::FPR128RegClass,
MRI);
9443 &AArch64::FPR128_loRegClass,
MRI);
9452 &AArch64::FPR128RegClass,
MRI);
9461 &AArch64::FPR128_loRegClass,
MRI);
9495 for (
auto *
MI : InsInstrs)
9496 MI->setFlags(Flags);
9537 bool IsNegativeBranch =
false;
9538 bool IsTestAndBranch =
false;
9539 unsigned TargetBBInMI = 0;
9540 switch (
MI.getOpcode()) {
9544 case AArch64::CBWPri:
9545 case AArch64::CBXPri:
9546 case AArch64::CBBAssertExt:
9547 case AArch64::CBHAssertExt:
9548 case AArch64::CBWPrr:
9549 case AArch64::CBXPrr:
9555 case AArch64::CBNZW:
9556 case AArch64::CBNZX:
9558 IsNegativeBranch =
true;
9563 IsTestAndBranch =
true;
9565 case AArch64::TBNZW:
9566 case AArch64::TBNZX:
9568 IsNegativeBranch =
true;
9569 IsTestAndBranch =
true;
9575 if (IsTestAndBranch &&
MI.getOperand(1).getImm())
9579 assert(
MI.getParent() &&
"Incomplete machine instruction\n");
9590 while (
DefMI->isCopy()) {
9592 if (!
MRI->hasOneNonDBGUse(CopyVReg))
9594 if (!
MRI->hasOneDef(CopyVReg))
9599 switch (
DefMI->getOpcode()) {
9603 case AArch64::ANDWri:
9604 case AArch64::ANDXri: {
9605 if (IsTestAndBranch)
9609 if (!
MRI->hasOneNonDBGUse(VReg))
9612 bool Is32Bit = (
DefMI->getOpcode() == AArch64::ANDWri);
9614 DefMI->getOperand(2).getImm(), Is32Bit ? 32 : 64);
9623 assert(!
MRI->def_empty(NewReg) &&
"Register must be defined.");
9629 unsigned Opc = (Imm < 32)
9630 ? (IsNegativeBranch ? AArch64::TBNZW : AArch64::TBZW)
9631 : (IsNegativeBranch ? AArch64::TBNZX : AArch64::TBZX);
9644 if (!Is32Bit && Imm < 32)
9646 MI.eraseFromParent();
9650 case AArch64::CSINCWr:
9651 case AArch64::CSINCXr: {
9652 if (!(
DefMI->getOperand(1).getReg() == AArch64::WZR &&
9653 DefMI->getOperand(2).getReg() == AArch64::WZR) &&
9654 !(
DefMI->getOperand(1).getReg() == AArch64::XZR &&
9655 DefMI->getOperand(2).getReg() == AArch64::XZR))
9658 if (
DefMI->findRegisterDefOperandIdx(AArch64::NZCV,
nullptr,
9671 if (IsNegativeBranch)
9674 MI.eraseFromParent();
9680std::pair<unsigned, unsigned>
9681AArch64InstrInfo::decomposeMachineOperandsTargetFlags(
unsigned TF)
const {
9683 return std::make_pair(TF & Mask, TF & ~Mask);
9687AArch64InstrInfo::getSerializableDirectMachineOperandTargetFlags()
const {
9690 static const std::pair<unsigned, const char *> TargetFlags[] = {
9691 {MO_PAGE,
"aarch64-page"}, {
MO_PAGEOFF,
"aarch64-pageoff"},
9692 {
MO_G3,
"aarch64-g3"}, {
MO_G2,
"aarch64-g2"},
9693 {
MO_G1,
"aarch64-g1"}, {
MO_G0,
"aarch64-g0"},
9699AArch64InstrInfo::getSerializableBitmaskMachineOperandTargetFlags()
const {
9700 using namespace AArch64II;
9702 static const std::pair<unsigned, const char *> TargetFlags[] = {
9705 {
MO_NC,
"aarch64-nc"},
9706 {
MO_S,
"aarch64-s"},
9717AArch64InstrInfo::getSerializableMachineMemOperandTargetFlags()
const {
9718 static const std::pair<MachineMemOperand::Flags, const char *> TargetFlags[] =
9820 MachineFunction *MF =
C.getMF();
9822 const AArch64RegisterInfo *ARI =
9823 static_cast<const AArch64RegisterInfo *
>(&
TRI);
9826 for (
unsigned Reg : AArch64::GPR64RegClass) {
9828 Reg != AArch64::LR &&
9829 Reg != AArch64::X16 &&
9830 Reg != AArch64::X17 &&
9831 C.isAvailableAcrossAndOutOfSeq(
Reg,
TRI) &&
9832 C.isAvailableInsideSeq(
Reg,
TRI))
9863 return SubtargetA.hasV8_3aOps() == SubtargetB.hasV8_3aOps();
9866std::optional<std::unique_ptr<outliner::OutlinedFunction>>
9867AArch64InstrInfo::getOutliningCandidateInfo(
9869 std::vector<outliner::Candidate> &RepeatedSequenceLocs,
9870 unsigned MinRepeats)
const {
9871 unsigned SequenceSize = 0;
9872 for (
auto &
MI : RepeatedSequenceLocs[0])
9875 unsigned NumBytesToCreateFrame = 0;
9881 MachineInstr &LastMI = RepeatedSequenceLocs[0].back();
9882 MachineInstr &FirstMI = RepeatedSequenceLocs[0].front();
9883 if (LastMI.
getOpcode() == AArch64::ADRP &&
9886 return std::nullopt;
9891 if ((FirstMI.
getOpcode() == AArch64::ADDXri ||
9892 FirstMI.
getOpcode() == AArch64::LDRXui) &&
9895 return std::nullopt;
9906 if (std::adjacent_find(
9907 RepeatedSequenceLocs.begin(), RepeatedSequenceLocs.end(),
9908 [](
const outliner::Candidate &a,
const outliner::Candidate &b) {
9911 if (outliningCandidatesSigningScopeConsensus(a, b) &&
9912 outliningCandidatesSigningKeyConsensus(a, b) &&
9913 outliningCandidatesV8_3OpsConsensus(a, b)) {
9917 }) != RepeatedSequenceLocs.end()) {
9918 return std::nullopt;
9935 unsigned NumBytesToCheckLRInTCEpilogue = 0;
9936 const auto RASignCondition = RepeatedSequenceLocs[0]
9939 ->getSignReturnAddressCondition();
9942 NumBytesToCreateFrame += 8;
9945 auto LRCheckMethod = Subtarget.getAuthenticatedLRCheckMethod(
9946 *RepeatedSequenceLocs[0].getMF());
9947 NumBytesToCheckLRInTCEpilogue =
9951 if (isTailCallReturnInst(RepeatedSequenceLocs[0].
back()))
9952 SequenceSize += NumBytesToCheckLRInTCEpilogue;
9960 for (
auto &
MI :
C) {
9961 if (
MI.modifiesRegister(AArch64::SP, &
TRI)) {
9962 switch (
MI.getOpcode()) {
9963 case AArch64::ADDXri:
9964 case AArch64::ADDWri:
9965 assert(
MI.getNumOperands() == 4 &&
"Wrong number of operands");
9967 "Expected operand to be immediate");
9969 "Expected operand to be a register");
9973 if (
MI.getOperand(1).getReg() == AArch64::SP)
9974 SPValue +=
MI.getOperand(2).getImm();
9978 case AArch64::SUBXri:
9979 case AArch64::SUBWri:
9980 assert(
MI.getNumOperands() == 4 &&
"Wrong number of operands");
9982 "Expected operand to be immediate");
9984 "Expected operand to be a register");
9988 if (
MI.getOperand(1).getReg() == AArch64::SP)
9989 SPValue -=
MI.getOperand(2).getImm();
10006 if (RepeatedSequenceLocs.size() < MinRepeats)
10007 return std::nullopt;
10011 unsigned FlagsSetInAll = 0xF;
10015 FlagsSetInAll &=
C.Flags;
10017 unsigned LastInstrOpcode = RepeatedSequenceLocs[0].back().getOpcode();
10020 auto SetCandidateCallInfo =
10021 [&RepeatedSequenceLocs](
unsigned CallID,
unsigned NumBytesForCall) {
10023 C.setCallInfo(CallID, NumBytesForCall);
10027 NumBytesToCreateFrame += 4;
10035 unsigned CFICount = 0;
10036 for (
auto &
I : RepeatedSequenceLocs[0]) {
10037 if (
I.isCFIInstruction())
10047 std::vector<MCCFIInstruction> CFIInstructions =
10048 C.getMF()->getFrameInstructions();
10050 if (CFICount > 0 && CFICount != CFIInstructions.size())
10051 return std::nullopt;
10059 if (!
MI.modifiesRegister(AArch64::SP, &
TRI) &&
10060 !
MI.readsRegister(AArch64::SP, &
TRI))
10066 if (
MI.modifiesRegister(AArch64::SP, &
TRI))
10071 if (
MI.mayLoadOrStore()) {
10074 bool OffsetIsScalable;
10078 if (!getMemOperandWithOffset(
MI,
Base,
Offset, OffsetIsScalable, &
TRI) ||
10079 !
Base->isReg() ||
Base->getReg() != AArch64::SP)
10083 if (OffsetIsScalable)
10091 TypeSize Scale(0U,
false), DummyWidth(0U,
false);
10092 getMemOpInfo(
MI.getOpcode(), Scale, DummyWidth, MinOffset, MaxOffset);
10095 if (
Offset < MinOffset * (int64_t)Scale.getFixedValue() ||
10096 Offset > MaxOffset * (int64_t)Scale.getFixedValue())
10111 bool AllStackInstrsSafe =
10116 if (RepeatedSequenceLocs[0].
back().isTerminator()) {
10118 NumBytesToCreateFrame = 0;
10119 unsigned NumBytesForCall = 4 + NumBytesToCheckLRInTCEpilogue;
10123 else if (LastInstrOpcode == AArch64::BL ||
10124 ((LastInstrOpcode == AArch64::BLR ||
10125 LastInstrOpcode == AArch64::BLRNoIP) &&
10129 NumBytesToCreateFrame = NumBytesToCheckLRInTCEpilogue;
10137 unsigned NumBytesNoStackCalls = 0;
10138 std::vector<outliner::Candidate> CandidatesWithoutStackFixups;
10144 ?
C.isAvailableAcrossAndOutOfSeq(AArch64::LR,
TRI)
10153 C.getMF()->getFunction().hasFnAttribute(Attribute::NoReturn);
10156 if (LRAvailable && !IsNoReturn) {
10157 NumBytesNoStackCalls += 4;
10159 CandidatesWithoutStackFixups.push_back(
C);
10164 else if (findRegisterToSaveLRTo(
C)) {
10165 NumBytesNoStackCalls += 12;
10167 CandidatesWithoutStackFixups.push_back(
C);
10172 else if (
C.isAvailableInsideSeq(AArch64::SP,
TRI)) {
10173 NumBytesNoStackCalls += 12;
10175 CandidatesWithoutStackFixups.push_back(
C);
10181 NumBytesNoStackCalls += SequenceSize;
10188 if (!AllStackInstrsSafe ||
10189 NumBytesNoStackCalls <= RepeatedSequenceLocs.size() * 12) {
10190 RepeatedSequenceLocs = CandidatesWithoutStackFixups;
10192 if (RepeatedSequenceLocs.size() < MinRepeats)
10193 return std::nullopt;
10246 (!
C.isAvailableAcrossAndOutOfSeq(AArch64::LR,
TRI) ||
10247 !findRegisterToSaveLRTo(
C));
10253 if (RepeatedSequenceLocs.size() < MinRepeats)
10254 return std::nullopt;
10263 bool ModStackToSaveLR =
false;
10266 ModStackToSaveLR =
true;
10275 ModStackToSaveLR =
true;
10277 if (ModStackToSaveLR) {
10279 if (!AllStackInstrsSafe)
10280 return std::nullopt;
10283 NumBytesToCreateFrame += 8;
10290 return std::nullopt;
10292 return std::make_unique<outliner::OutlinedFunction>(
10293 RepeatedSequenceLocs, SequenceSize, NumBytesToCreateFrame, FrameID);
10296void AArch64InstrInfo::mergeOutliningCandidateAttributes(
10297 Function &
F, std::vector<outliner::Candidate> &Candidates)
const {
10301 const auto &CFn = Candidates.front().getMF()->getFunction();
10303 if (CFn.hasFnAttribute(
"ptrauth-returns"))
10304 F.addFnAttr(CFn.getFnAttribute(
"ptrauth-returns"));
10305 if (CFn.hasFnAttribute(
"ptrauth-auth-traps"))
10306 F.addFnAttr(CFn.getFnAttribute(
"ptrauth-auth-traps"));
10309 if (CFn.hasFnAttribute(
"sign-return-address"))
10310 F.addFnAttr(CFn.getFnAttribute(
"sign-return-address"));
10311 if (CFn.hasFnAttribute(
"sign-return-address-key"))
10312 F.addFnAttr(CFn.getFnAttribute(
"sign-return-address-key"));
10314 AArch64GenInstrInfo::mergeOutliningCandidateAttributes(
F, Candidates);
10317bool AArch64InstrInfo::isFunctionSafeToOutlineFrom(
10322 if (!OutlineFromLinkOnceODRs &&
F.hasLinkOnceODRLinkage())
10329 if (
F.hasSection())
10335 AArch64FunctionInfo *AFI = MF.
getInfo<AArch64FunctionInfo>();
10336 if (!AFI || AFI->
hasRedZone().value_or(
true))
10356 unsigned &Flags)
const {
10358 "Must track liveness!");
10360 std::pair<MachineBasicBlock::iterator, MachineBasicBlock::iterator>>
10375 auto AreAllUnsafeRegsDead = [&LRU]() {
10376 return LRU.available(AArch64::W16) && LRU.available(AArch64::W17) &&
10377 LRU.available(AArch64::NZCV);
10392 bool LRAvailableEverywhere =
true;
10394 LRU.addLiveOuts(
MBB);
10396 auto UpdateWholeMBBFlags = [&
Flags](
const MachineInstr &
MI) {
10397 if (
MI.isCall() && !
MI.isTerminator())
10403 auto CreateNewRangeStartingAt =
10404 [&RangeBegin, &RangeEnd,
10406 RangeBegin = NewBegin;
10407 RangeEnd = std::next(RangeBegin);
10410 auto SaveRangeIfNonEmpty = [&RangeLen, &
Ranges, &RangeBegin, &RangeEnd]() {
10416 if (!RangeBegin.isEnd() && RangeBegin->isBundledWithPred())
10418 if (!RangeEnd.isEnd() && RangeEnd->isBundledWithPred())
10420 Ranges.emplace_back(RangeBegin, RangeEnd);
10428 for (; FirstPossibleEndPt !=
MBB.
instr_rend(); ++FirstPossibleEndPt) {
10429 LRU.stepBackward(*FirstPossibleEndPt);
10432 UpdateWholeMBBFlags(*FirstPossibleEndPt);
10433 if (AreAllUnsafeRegsDead())
10440 CreateNewRangeStartingAt(FirstPossibleEndPt->getIterator());
10445 LRU.stepBackward(
MI);
10446 UpdateWholeMBBFlags(
MI);
10447 if (!AreAllUnsafeRegsDead()) {
10448 SaveRangeIfNonEmpty();
10449 CreateNewRangeStartingAt(
MI.getIterator());
10452 LRAvailableEverywhere &= LRU.available(AArch64::LR);
10453 RangeBegin =
MI.getIterator();
10458 if (AreAllUnsafeRegsDead())
10459 SaveRangeIfNonEmpty();
10467 if (!LRAvailableEverywhere)
10475 unsigned Flags)
const {
10476 MachineInstr &
MI = *MIT;
10480 switch (
MI.getOpcode()) {
10481 case AArch64::PACM:
10482 case AArch64::PACIASP:
10483 case AArch64::PACIBSP:
10484 case AArch64::PACIASPPC:
10485 case AArch64::PACIBSPPC:
10486 case AArch64::AUTIASP:
10487 case AArch64::AUTIBSP:
10488 case AArch64::AUTIASPPCi:
10489 case AArch64::AUTIASPPCr:
10490 case AArch64::AUTIBSPPCi:
10491 case AArch64::AUTIBSPPCr:
10492 case AArch64::RETAA:
10493 case AArch64::RETAB:
10494 case AArch64::RETAASPPCi:
10495 case AArch64::RETAASPPCr:
10496 case AArch64::RETABSPPCi:
10497 case AArch64::RETABSPPCr:
10498 case AArch64::EMITBKEY:
10499 case AArch64::PAUTH_PROLOGUE:
10500 case AArch64::PAUTH_EPILOGUE:
10510 if (
MI.isCFIInstruction())
10514 if (
MI.isTerminator())
10520 for (
const MachineOperand &MOP :
MI.operands()) {
10523 assert(!MOP.isCFIIndex());
10526 if (MOP.isReg() && !MOP.isImplicit() &&
10527 (MOP.getReg() == AArch64::LR || MOP.getReg() == AArch64::W30))
10534 if (
MI.getOpcode() == AArch64::ADRP)
10554 for (
const MachineOperand &MOP :
MI.operands()) {
10555 if (MOP.isGlobal()) {
10563 if (Callee &&
Callee->getName() ==
"\01_mcount")
10571 if (
MI.getOpcode() == AArch64::BLR ||
10572 MI.getOpcode() == AArch64::BLRNoIP ||
MI.getOpcode() == AArch64::BL)
10576 return UnknownCallOutlineType;
10584 return UnknownCallOutlineType;
10592 return UnknownCallOutlineType;
10613 for (MachineInstr &
MI :
MBB) {
10614 const MachineOperand *
Base;
10615 TypeSize Width(0,
false);
10617 bool OffsetIsScalable;
10620 if (!
MI.mayLoadOrStore() ||
10623 (
Base->isReg() &&
Base->getReg() != AArch64::SP))
10627 TypeSize Scale(0U,
false);
10628 int64_t Dummy1, Dummy2;
10631 assert(StackOffsetOperand.
isImm() &&
"Stack offset wasn't immediate!");
10633 assert(Scale != 0 &&
"Unexpected opcode!");
10634 assert(!OffsetIsScalable &&
"Expected offset to be a byte offset");
10639 int64_t NewImm = (
Offset + 16) / (int64_t)Scale.getFixedValue();
10640 StackOffsetOperand.
setImm(NewImm);
10646 bool ShouldSignReturnAddr) {
10647 if (!ShouldSignReturnAddr)
10653 TII->get(AArch64::PAUTH_EPILOGUE))
10657void AArch64InstrInfo::buildOutlinedFrame(
10661 AArch64FunctionInfo *FI = MF.
getInfo<AArch64FunctionInfo>();
10669 unsigned TailOpcode;
10671 TailOpcode = AArch64::TCRETURNdi;
10675 TailOpcode = AArch64::TCRETURNriALL;
10686 bool IsLeafFunction =
true;
10689 auto IsNonTailCall = [](
const MachineInstr &
MI) {
10690 return MI.isCall() && !
MI.isReturn();
10700 "Can only fix up stack references once");
10701 fixupPostOutline(
MBB);
10703 IsLeafFunction =
false;
10714 Et = std::prev(
MBB.
end());
10724 if (MF.
getInfo<AArch64FunctionInfo>()->needsDwarfUnwindInfo(MF)) {
10728 CFIBuilder.buildDefCFAOffset(16);
10732 CFIBuilder.buildOffset(AArch64::LR, -16);
10746 RASignCondition, !IsLeafFunction);
10775 fixupPostOutline(
MBB);
10786 .addGlobalAddress(
M.getNamedValue(MF.
getName()))
10796 .addGlobalAddress(
M.getNamedValue(MF.
getName())));
10805 MachineInstr *Save;
10806 MachineInstr *Restore;
10812 assert(
Reg &&
"No callee-saved register available?");
10846 .addGlobalAddress(
M.getNamedValue(MF.
getName())));
10854bool AArch64InstrInfo::shouldOutlineFromFunctionByDefault(
10862 bool AllowSideEffects)
const {
10864 const AArch64Subtarget &STI = MF.
getSubtarget<AArch64Subtarget>();
10867 if (
TRI.isGeneralPurposeRegister(MF,
Reg)) {
10880 assert(STI.hasNEON() &&
"Expected to have NEON.");
10886std::optional<DestSourcePair>
10891 if (((
MI.getOpcode() == AArch64::ORRWrs &&
10892 MI.getOperand(1).getReg() == AArch64::WZR &&
10893 MI.getOperand(3).getImm() == 0x0) ||
10894 (
MI.getOpcode() == AArch64::ORRWrr &&
10895 MI.getOperand(1).getReg() == AArch64::WZR)) &&
10897 (!
MI.getOperand(0).getReg().isVirtual() ||
10898 MI.getOperand(0).getSubReg() == 0) &&
10899 (!
MI.getOperand(0).getReg().isPhysical() ||
10904 if (
MI.getOpcode() == AArch64::ORRXrs &&
10905 MI.getOperand(1).getReg() == AArch64::XZR &&
10906 MI.getOperand(3).getImm() == 0x0)
10909 return std::nullopt;
10912std::optional<DestSourcePair>
10914 if ((
MI.getOpcode() == AArch64::ORRWrs &&
10915 MI.getOperand(1).getReg() == AArch64::WZR &&
10916 MI.getOperand(3).getImm() == 0x0) ||
10917 (
MI.getOpcode() == AArch64::ORRWrr &&
10918 MI.getOperand(1).getReg() == AArch64::WZR))
10920 return std::nullopt;
10923std::optional<RegImmPair>
10932 return std::nullopt;
10934 switch (
MI.getOpcode()) {
10936 return std::nullopt;
10937 case AArch64::SUBWri:
10938 case AArch64::SUBXri:
10939 case AArch64::SUBSWri:
10940 case AArch64::SUBSXri:
10943 case AArch64::ADDSWri:
10944 case AArch64::ADDSXri:
10945 case AArch64::ADDWri:
10946 case AArch64::ADDXri: {
10948 if (!
MI.getOperand(0).isReg() || !
MI.getOperand(1).isReg() ||
10949 !
MI.getOperand(2).isImm())
10950 return std::nullopt;
10951 int Shift =
MI.getOperand(3).getImm();
10952 assert((Shift == 0 || Shift == 12) &&
"Shift can be either 0 or 12");
10956 return RegImmPair{
MI.getOperand(1).getReg(),
Offset};
10962static std::optional<ParamLoadedValue>
10966 auto DestSrc =
TII->isCopyLikeInstr(
MI);
10968 return std::nullopt;
10970 Register DestReg = DestSrc->Destination->getReg();
10971 Register SrcReg = DestSrc->Source->getReg();
10974 return std::nullopt;
10979 if (DestReg == DescribedReg)
10983 if (
MI.getOpcode() == AArch64::ORRWrs &&
10984 TRI->isSuperRegister(DestReg, DescribedReg))
10988 if (
MI.getOpcode() == AArch64::ORRXrs &&
10989 TRI->isSubRegister(DestReg, DescribedReg)) {
10990 Register SrcSubReg =
TRI->getSubReg(SrcReg, AArch64::sub_32);
10994 assert(!
TRI->isSuperOrSubRegisterEq(DestReg, DescribedReg) &&
10995 "Unhandled ORR[XW]rs copy case");
10997 return std::nullopt;
11000bool AArch64InstrInfo::isFunctionSafeToSplit(
const MachineFunction &MF)
const {
11005 if (MF.
getInfo<AArch64FunctionInfo>()->hasRedZone().value_or(
true))
11011bool AArch64InstrInfo::isMBBSafeToSplitToCold(
11015 auto isAsmGoto = [](
const MachineInstr &
MI) {
11016 return MI.getOpcode() == AArch64::INLINEASM_BR;
11026 auto containsMBB = [&
MBB](
const MachineJumpTableEntry &JTE) {
11033 for (
const MachineInstr &
MI :
MBB) {
11034 switch (
MI.getOpcode()) {
11035 case TargetOpcode::G_BRJT:
11036 case AArch64::JumpTableDest32:
11037 case AArch64::JumpTableDest16:
11038 case AArch64::JumpTableDest8:
11049std::optional<ParamLoadedValue>
11052 const MachineFunction *MF =
MI.getMF();
11054 switch (
MI.getOpcode()) {
11055 case AArch64::MOVZWi:
11056 case AArch64::MOVZXi: {
11059 if (!
TRI->isSuperRegisterEq(
MI.getOperand(0).getReg(),
Reg))
11060 return std::nullopt;
11062 if (!
MI.getOperand(1).isImm())
11063 return std::nullopt;
11064 int64_t Immediate =
MI.getOperand(1).getImm();
11065 int Shift =
MI.getOperand(2).getImm();
11069 case AArch64::ORRWrs:
11070 case AArch64::ORRXrs:
11077bool AArch64InstrInfo::isExtendLikelyToBeFolded(
11080 ExtMI.
getOpcode() == TargetOpcode::G_ZEXT ||
11081 ExtMI.
getOpcode() == TargetOpcode::G_ANYEXT);
11084 if (ExtMI.
getOpcode() == TargetOpcode::G_ANYEXT)
11088 if (!
MRI.hasOneNonDBGUse(DefReg))
11093 auto *UserMI = &*
MRI.use_instr_nodbg_begin(DefReg);
11094 return UserMI->getOpcode() == TargetOpcode::G_PTR_ADD;
11097uint64_t AArch64InstrInfo::getElementSizeForOpcode(
unsigned Opc)
const {
11101bool AArch64InstrInfo::isPTestLikeOpcode(
unsigned Opc)
const {
11105bool AArch64InstrInfo::isWhileOpcode(
unsigned Opc)
const {
11110AArch64InstrInfo::getTailDuplicateSize(
CodeGenOptLevel OptLevel)
const {
11114bool AArch64InstrInfo::isLegalAddressingMode(
unsigned NumBytes, int64_t
Offset,
11115 unsigned Scale)
const {
11126 unsigned Shift =
Log2_64(NumBytes);
11127 if (NumBytes &&
Offset > 0 && (
Offset / NumBytes) <= (1LL << 12) - 1 &&
11135 return Scale == 1 || (Scale > 0 && Scale == NumBytes);
11140 return AArch64::BLRNoIP;
11142 return AArch64::BLR;
11147 Register TargetReg,
bool FrameSetup)
const {
11148 assert(TargetReg != AArch64::SP &&
"New top of stack cannot already be in SP");
11160 MF.
insert(MBBInsertPoint, LoopTestMBB);
11163 MF.
insert(MBBInsertPoint, LoopBodyMBB);
11165 MF.
insert(MBBInsertPoint, ExitMBB);
11175 BuildMI(*LoopTestMBB, LoopTestMBB->
end(),
DL,
TII->get(AArch64::SUBSXrx64),
11183 BuildMI(*LoopTestMBB, LoopTestMBB->
end(),
DL,
TII->get(AArch64::Bcc))
11189 BuildMI(*LoopBodyMBB, LoopBodyMBB->
end(),
DL,
TII->get(AArch64::STRXui))
11202 BuildMI(*ExitMBB, ExitMBB->
end(),
DL,
TII->get(AArch64::ADDXri), AArch64::SP)
11221 MBB.addSuccessor(LoopTestMBB);
11227 return ExitMBB->
begin();
11244 unsigned CompCounterOprNum;
11248 unsigned UpdateCounterOprNum;
11252 bool IsUpdatePriorComp;
11264 TII(MF->getSubtarget().getInstrInfo()),
11265 TRI(MF->getSubtarget().getRegisterInfo()),
MRI(MF->getRegInfo()),
11266 LoopBB(LoopBB), CondBranch(CondBranch), Comp(Comp),
11267 CompCounterOprNum(CompCounterOprNum), Update(Update),
11268 UpdateCounterOprNum(UpdateCounterOprNum),
Init(
Init),
11269 IsUpdatePriorComp(IsUpdatePriorComp),
Cond(
Cond.begin(),
Cond.end()) {}
11271 bool shouldIgnoreForPipelining(
const MachineInstr *
MI)
const override {
11277 std::optional<bool> createTripCountGreaterCondition(
11278 int TC, MachineBasicBlock &
MBB,
11279 SmallVectorImpl<MachineOperand> &CondParam)
override {
11287 void createRemainingIterationsGreaterCondition(
11288 int TC, MachineBasicBlock &
MBB, SmallVectorImpl<MachineOperand> &
Cond,
11289 DenseMap<MachineInstr *, MachineInstr *> &LastStage0Insts)
override;
11291 void setPreheader(MachineBasicBlock *NewPreheader)
override {}
11293 void adjustTripCount(
int TripCountAdjust)
override {}
11295 bool isMVEExpanderSupported()
override {
return true; }
11311 Result =
MRI.createVirtualRegister(
11314 }
else if (
I == ReplaceOprNum) {
11315 MRI.constrainRegClass(ReplaceReg,
TII->getRegClass(NewMI->
getDesc(),
I));
11319 MBB.insert(InsertTo, NewMI);
11323void AArch64PipelinerLoopInfo::createRemainingIterationsGreaterCondition(
11339 assert(CondBranch->getOpcode() == AArch64::Bcc);
11343 if (CondBranch->getOperand(1).getMBB() == LoopBB)
11350 auto AccumulateCond = [&](
Register CurCond,
11352 Register NewCond =
MRI.createVirtualRegister(&AArch64::GPR64commonRegClass);
11361 if (!LastStage0Insts.
empty() && LastStage0Insts[Comp]->getParent() == &
MBB) {
11365 for (
int I = 0;
I <= TC; ++
I) {
11371 AccCond = AccumulateCond(AccCond, CC);
11375 if (Update != Comp && IsUpdatePriorComp) {
11377 LastStage0Insts[Comp]->getOperand(CompCounterOprNum).getReg();
11378 NextCounter =
cloneInstr(Update, UpdateCounterOprNum, Counter,
MBB,
11382 NextCounter = LastStage0Insts[Update]->getOperand(0).getReg();
11384 }
else if (Update != Comp) {
11389 Counter = NextCounter;
11393 if (LastStage0Insts.
empty()) {
11397 if (IsUpdatePriorComp)
11402 Counter = LastStage0Insts[Comp]->getOperand(CompCounterOprNum).getReg();
11405 for (
int I = 0;
I <= TC; ++
I) {
11409 AccCond = AccumulateCond(AccCond, CC);
11410 if (
I != TC && Update != Comp)
11413 Counter = NextCounter;
11429 assert(Phi.getNumOperands() == 5);
11430 if (Phi.getOperand(2).getMBB() ==
MBB) {
11431 RegMBB = Phi.getOperand(1).getReg();
11432 RegOther = Phi.getOperand(3).getReg();
11434 assert(Phi.getOperand(4).getMBB() ==
MBB);
11435 RegMBB = Phi.getOperand(3).getReg();
11436 RegOther = Phi.getOperand(1).getReg();
11441 if (!
Reg.isVirtual())
11444 return MRI.getVRegDef(
Reg)->getParent() != BB;
11450 unsigned &UpdateCounterOprNum,
Register &InitReg,
11451 bool &IsUpdatePriorComp) {
11465 if (!
Reg.isVirtual())
11468 UpdateInst =
nullptr;
11469 UpdateCounterOprNum = 0;
11471 IsUpdatePriorComp =
true;
11475 if (Def->getParent() != LoopBB)
11477 if (Def->isCopy()) {
11479 if (Def->getOperand(0).getSubReg() || Def->getOperand(1).getSubReg())
11481 CurReg = Def->getOperand(1).getReg();
11482 }
else if (Def->isPHI()) {
11486 IsUpdatePriorComp =
false;
11491 switch (Def->getOpcode()) {
11492 case AArch64::ADDSXri:
11493 case AArch64::ADDSWri:
11494 case AArch64::SUBSXri:
11495 case AArch64::SUBSWri:
11496 case AArch64::ADDXri:
11497 case AArch64::ADDWri:
11498 case AArch64::SUBXri:
11499 case AArch64::SUBWri:
11501 UpdateCounterOprNum = 1;
11503 case AArch64::ADDSXrr:
11504 case AArch64::ADDSWrr:
11505 case AArch64::SUBSXrr:
11506 case AArch64::SUBSWrr:
11507 case AArch64::ADDXrr:
11508 case AArch64::ADDWrr:
11509 case AArch64::SUBXrr:
11510 case AArch64::SUBWrr:
11513 UpdateCounterOprNum = 1;
11515 UpdateCounterOprNum = 2;
11522 CurReg = Def->getOperand(UpdateCounterOprNum).getReg();
11537std::unique_ptr<TargetInstrInfo::PipelinerLoopInfo>
11548 if (
MI.isCall() ||
MI.hasUnmodeledSideEffects())
11559 if (
TBB == LoopBB && FBB == LoopBB)
11563 if (
TBB != LoopBB && FBB ==
nullptr)
11566 assert((
TBB == LoopBB || FBB == LoopBB) &&
11567 "The Loop must be a single-basic-block loop");
11572 if (CondBranch->
getOpcode() != AArch64::Bcc)
11580 unsigned CompCounterOprNum = 0;
11582 if (
MI.modifiesRegister(AArch64::NZCV, &
TRI)) {
11586 switch (
MI.getOpcode()) {
11587 case AArch64::SUBSXri:
11588 case AArch64::SUBSWri:
11589 case AArch64::ADDSXri:
11590 case AArch64::ADDSWri:
11592 CompCounterOprNum = 1;
11594 case AArch64::ADDSWrr:
11595 case AArch64::ADDSXrr:
11596 case AArch64::SUBSWrr:
11597 case AArch64::SUBSXrr:
11601 if (isWhileOpcode(
MI.getOpcode())) {
11608 if (CompCounterOprNum == 0) {
11610 CompCounterOprNum = 2;
11612 CompCounterOprNum = 1;
11624 bool IsUpdatePriorComp;
11625 unsigned UpdateCounterOprNum;
11627 Update, UpdateCounterOprNum,
Init, IsUpdatePriorComp))
11630 return std::make_unique<AArch64PipelinerLoopInfo>(
11631 LoopBB, CondBranch, Comp, CompCounterOprNum, Update, UpdateCounterOprNum,
11641 TypeSize Scale(0U,
false), Width(0U,
false);
11642 int64_t MinOffset, MaxOffset;
11643 if (
getMemOpInfo(
MI.getOpcode(), Scale, Width, MinOffset, MaxOffset)) {
11645 if (
MI.getOperand(ImmIdx).isImm() && !
MI.getOperand(ImmIdx - 1).isFI()) {
11646 int64_t Imm =
MI.getOperand(ImmIdx).getImm();
11647 if (Imm < MinOffset || Imm > MaxOffset) {
11648 ErrInfo =
"Unexpected immediate on load/store instruction";
11654 const MCInstrDesc &MCID =
MI.getDesc();
11656 const MachineOperand &MO =
MI.getOperand(
Op);
11660 ErrInfo =
"OPERAND_IMPLICIT_IMM_0 should be 0";
11669 ErrInfo =
"OPERAND_SHIFT_MSL should be msl shift of 8 or 16";
11680#define GET_INSTRINFO_HELPERS
11681#define GET_INSTRMAP_INFO
11682#include "AArch64GenInstrInfo.inc"
unsigned const MachineRegisterInfo * MRI
MachineInstrBuilder & UseMI
MachineInstrBuilder MachineInstrBuilder & DefMI
static bool forwardCopyWillClobberTuple(unsigned DestReg, unsigned SrcReg, unsigned NumRegs)
static cl::opt< unsigned > BCCDisplacementBits("aarch64-bcc-offset-bits", cl::Hidden, cl::init(19), cl::desc("Restrict range of Bcc instructions (DEBUG)"))
static Register genNeg(MachineFunction &MF, MachineRegisterInfo &MRI, const TargetInstrInfo *TII, MachineInstr &Root, SmallVectorImpl< MachineInstr * > &InsInstrs, DenseMap< Register, unsigned > &InstrIdxForVirtReg, unsigned MnegOpc, const TargetRegisterClass *RC)
genNeg - Helper to generate an intermediate negation of the second operand of Root
static bool isFrameStoreOpcode(int Opcode)
static cl::opt< unsigned > GatherOptSearchLimit("aarch64-search-limit", cl::Hidden, cl::init(2048), cl::desc("Restrict range of instructions to search for the " "machine-combiner gather pattern optimization"))
static bool getMaddPatterns(MachineInstr &Root, SmallVectorImpl< unsigned > &Patterns)
Find instructions that can be turned into madd.
static AArch64CC::CondCode findCondCodeUsedByInstr(const MachineInstr &Instr)
Find a condition code used by the instruction.
static MachineInstr * genFusedMultiplyAcc(MachineFunction &MF, MachineRegisterInfo &MRI, const TargetInstrInfo *TII, MachineInstr &Root, SmallVectorImpl< MachineInstr * > &InsInstrs, unsigned IdxMulOpd, unsigned MaddOpc, const TargetRegisterClass *RC)
genFusedMultiplyAcc - Helper to generate fused multiply accumulate instructions.
static MachineInstr * genFusedMultiplyAccNeg(MachineFunction &MF, MachineRegisterInfo &MRI, const TargetInstrInfo *TII, MachineInstr &Root, SmallVectorImpl< MachineInstr * > &InsInstrs, DenseMap< Register, unsigned > &InstrIdxForVirtReg, unsigned IdxMulOpd, unsigned MaddOpc, unsigned MnegOpc, const TargetRegisterClass *RC)
genFusedMultiplyAccNeg - Helper to generate fused multiply accumulate instructions with an additional...
static bool isCombineInstrCandidate64(unsigned Opc)
static bool isFrameLoadOpcode(int Opcode)
static unsigned removeCopies(const MachineRegisterInfo &MRI, unsigned VReg)
static bool areCFlagsAccessedBetweenInstrs(MachineBasicBlock::iterator From, MachineBasicBlock::iterator To, const TargetRegisterInfo *TRI, const AccessKind AccessToCheck=AK_All)
True when condition flags are accessed (either by writing or reading) on the instruction trace starti...
static bool getFMAPatterns(MachineInstr &Root, SmallVectorImpl< unsigned > &Patterns)
Floating-Point Support.
static bool isADDSRegImm(unsigned Opcode)
static bool isCheapCopy(const MachineInstr &MI, const AArch64RegisterInfo &RI)
static bool isANDOpcode(MachineInstr &MI)
static void appendOffsetComment(int NumBytes, llvm::raw_string_ostream &Comment, StringRef RegScale={})
static unsigned sForm(MachineInstr &Instr)
Get opcode of S version of Instr.
static bool isCombineInstrSettingFlag(unsigned Opc)
static bool getFNEGPatterns(MachineInstr &Root, SmallVectorImpl< unsigned > &Patterns)
static bool getIndVarInfo(Register Reg, const MachineBasicBlock *LoopBB, MachineInstr *&UpdateInst, unsigned &UpdateCounterOprNum, Register &InitReg, bool &IsUpdatePriorComp)
If Reg is an induction variable, return true and set some parameters.
static const MachineInstrBuilder & AddSubReg(const MachineInstrBuilder &MIB, MCRegister Reg, unsigned SubIdx, unsigned State, const TargetRegisterInfo *TRI)
static bool canPairLdStOpc(unsigned FirstOpc, unsigned SecondOpc)
static int findCondCodeUseOperandIdxForBranchOrSelect(const MachineInstr &Instr)
static bool isPostIndexLdStOpcode(unsigned Opcode)
Return true if the opcode is a post-index ld/st instruction, which really loads from base+0.
static unsigned getBranchDisplacementBits(unsigned Opc)
static cl::opt< unsigned > CBDisplacementBits("aarch64-cb-offset-bits", cl::Hidden, cl::init(9), cl::desc("Restrict range of CB instructions (DEBUG)"))
static std::optional< ParamLoadedValue > describeORRLoadedValue(const MachineInstr &MI, Register DescribedReg, const TargetInstrInfo *TII, const TargetRegisterInfo *TRI)
If the given ORR instruction is a copy, and DescribedReg overlaps with the destination register then,...
static bool getFMULPatterns(MachineInstr &Root, SmallVectorImpl< unsigned > &Patterns)
static void appendReadRegExpr(SmallVectorImpl< char > &Expr, unsigned RegNum)
static MachineInstr * genMaddR(MachineFunction &MF, MachineRegisterInfo &MRI, const TargetInstrInfo *TII, MachineInstr &Root, SmallVectorImpl< MachineInstr * > &InsInstrs, unsigned IdxMulOpd, unsigned MaddOpc, unsigned VR, const TargetRegisterClass *RC)
genMaddR - Generate madd instruction and combine mul and add using an extra virtual register Example ...
static Register cloneInstr(const MachineInstr *MI, unsigned ReplaceOprNum, Register ReplaceReg, MachineBasicBlock &MBB, MachineBasicBlock::iterator InsertTo)
Clone an instruction from MI.
static bool scaleOffset(unsigned Opc, int64_t &Offset)
static bool canCombineWithFMUL(MachineBasicBlock &MBB, MachineOperand &MO, unsigned MulOpc)
unsigned scaledOffsetOpcode(unsigned Opcode, unsigned &Scale)
static MachineInstr * genFusedMultiplyIdx(MachineFunction &MF, MachineRegisterInfo &MRI, const TargetInstrInfo *TII, MachineInstr &Root, SmallVectorImpl< MachineInstr * > &InsInstrs, unsigned IdxMulOpd, unsigned MaddOpc, const TargetRegisterClass *RC)
genFusedMultiplyIdx - Helper to generate fused multiply accumulate instructions.
static MachineInstr * genIndexedMultiply(MachineInstr &Root, SmallVectorImpl< MachineInstr * > &InsInstrs, unsigned IdxDupOp, unsigned MulOpc, const TargetRegisterClass *RC, MachineRegisterInfo &MRI)
Fold (FMUL x (DUP y lane)) into (FMUL_indexed x y lane)
static bool isSUBSRegImm(unsigned Opcode)
static bool UpdateOperandRegClass(MachineInstr &Instr)
static const TargetRegisterClass * getRegClass(const MachineInstr &MI, Register Reg)
static bool canCmpInstrBeRemoved(MachineInstr &MI, MachineInstr &CmpInstr, int CmpValue, const TargetRegisterInfo &TRI, SmallVectorImpl< MachineInstr * > &CCUseInstrs, bool &IsInvertCC)
unsigned unscaledOffsetOpcode(unsigned Opcode)
static bool getLoadPatterns(MachineInstr &Root, SmallVectorImpl< unsigned > &Patterns)
Search for patterns of LD instructions we can optimize.
static bool canInstrSubstituteCmpInstr(MachineInstr &MI, MachineInstr &CmpInstr, const TargetRegisterInfo &TRI)
Check if CmpInstr can be substituted by MI.
static UsedNZCV getUsedNZCV(AArch64CC::CondCode CC)
static bool isCombineInstrCandidateFP(const MachineInstr &Inst)
static void appendLoadRegExpr(SmallVectorImpl< char > &Expr, int64_t OffsetFromDefCFA)
static void appendConstantExpr(SmallVectorImpl< char > &Expr, int64_t Constant, dwarf::LocationAtom Operation)
static unsigned convertToNonFlagSettingOpc(const MachineInstr &MI)
Return the opcode that does not set flags when possible - otherwise return the original opcode.
static bool outliningCandidatesV8_3OpsConsensus(const outliner::Candidate &a, const outliner::Candidate &b)
static bool isCombineInstrCandidate32(unsigned Opc)
static void parseCondBranch(MachineInstr *LastInst, MachineBasicBlock *&Target, SmallVectorImpl< MachineOperand > &Cond)
static unsigned offsetExtendOpcode(unsigned Opcode)
static void loadRegPairFromStackSlot(const TargetRegisterInfo &TRI, MachineBasicBlock &MBB, MachineBasicBlock::iterator InsertBefore, const MCInstrDesc &MCID, Register DestReg, unsigned SubIdx0, unsigned SubIdx1, int FI, MachineMemOperand *MMO)
static void generateGatherLanePattern(MachineInstr &Root, SmallVectorImpl< MachineInstr * > &InsInstrs, SmallVectorImpl< MachineInstr * > &DelInstrs, DenseMap< Register, unsigned > &InstrIdxForVirtReg, unsigned Pattern, unsigned NumLanes)
Generate optimized instruction sequence for gather load patterns to improve Memory-Level Parallelism ...
static bool getMiscPatterns(MachineInstr &Root, SmallVectorImpl< unsigned > &Patterns)
Find other MI combine patterns.
static bool outliningCandidatesSigningKeyConsensus(const outliner::Candidate &a, const outliner::Candidate &b)
static bool outliningCandidatesSigningScopeConsensus(const outliner::Candidate &a, const outliner::Candidate &b)
static bool shouldClusterFI(const MachineFrameInfo &MFI, int FI1, int64_t Offset1, unsigned Opcode1, int FI2, int64_t Offset2, unsigned Opcode2)
static cl::opt< unsigned > TBZDisplacementBits("aarch64-tbz-offset-bits", cl::Hidden, cl::init(14), cl::desc("Restrict range of TB[N]Z instructions (DEBUG)"))
static void extractPhiReg(const MachineInstr &Phi, const MachineBasicBlock *MBB, Register &RegMBB, Register &RegOther)
static MCCFIInstruction createDefCFAExpression(const TargetRegisterInfo &TRI, unsigned Reg, const StackOffset &Offset)
static bool isDefinedOutside(Register Reg, const MachineBasicBlock *BB)
static MachineInstr * genFusedMultiply(MachineFunction &MF, MachineRegisterInfo &MRI, const TargetInstrInfo *TII, MachineInstr &Root, SmallVectorImpl< MachineInstr * > &InsInstrs, unsigned IdxMulOpd, unsigned MaddOpc, const TargetRegisterClass *RC, FMAInstKind kind=FMAInstKind::Default, const Register *ReplacedAddend=nullptr)
genFusedMultiply - Generate fused multiply instructions.
static bool getGatherLanePattern(MachineInstr &Root, SmallVectorImpl< unsigned > &Patterns, unsigned LoadLaneOpCode, unsigned NumLanes)
Check if the given instruction forms a gather load pattern that can be optimized for better Memory-Le...
static MachineInstr * genFusedMultiplyIdxNeg(MachineFunction &MF, MachineRegisterInfo &MRI, const TargetInstrInfo *TII, MachineInstr &Root, SmallVectorImpl< MachineInstr * > &InsInstrs, DenseMap< Register, unsigned > &InstrIdxForVirtReg, unsigned IdxMulOpd, unsigned MaddOpc, unsigned MnegOpc, const TargetRegisterClass *RC)
genFusedMultiplyAccNeg - Helper to generate fused multiply accumulate instructions with an additional...
static bool isCombineInstrCandidate(unsigned Opc)
static unsigned regOffsetOpcode(unsigned Opcode)
MachineOutlinerClass
Constants defining how certain sequences should be outlined.
@ MachineOutlinerTailCall
Emit a save, restore, call, and return.
@ MachineOutlinerRegSave
Emit a call and tail-call.
@ MachineOutlinerNoLRSave
Only emit a branch.
@ MachineOutlinerThunk
Emit a call and return.
static cl::opt< unsigned > BDisplacementBits("aarch64-b-offset-bits", cl::Hidden, cl::init(26), cl::desc("Restrict range of B instructions (DEBUG)"))
static bool areCFlagsAliveInSuccessors(const MachineBasicBlock *MBB)
Check if AArch64::NZCV should be alive in successors of MBB.
static void emitFrameOffsetAdj(MachineBasicBlock &MBB, MachineBasicBlock::iterator MBBI, const DebugLoc &DL, unsigned DestReg, unsigned SrcReg, int64_t Offset, unsigned Opc, const TargetInstrInfo *TII, MachineInstr::MIFlag Flag, bool NeedsWinCFI, bool *HasWinCFI, bool EmitCFAOffset, StackOffset CFAOffset, unsigned FrameReg)
static bool isCheapImmediate(const MachineInstr &MI, unsigned BitSize)
static cl::opt< unsigned > CBZDisplacementBits("aarch64-cbz-offset-bits", cl::Hidden, cl::init(19), cl::desc("Restrict range of CB[N]Z instructions (DEBUG)"))
static void genSubAdd2SubSub(MachineFunction &MF, MachineRegisterInfo &MRI, const TargetInstrInfo *TII, MachineInstr &Root, SmallVectorImpl< MachineInstr * > &InsInstrs, SmallVectorImpl< MachineInstr * > &DelInstrs, unsigned IdxOpd1, DenseMap< Register, unsigned > &InstrIdxForVirtReg)
Do the following transformation A - (B + C) ==> (A - B) - C A - (B + C) ==> (A - C) - B.
static unsigned canFoldIntoCSel(const MachineRegisterInfo &MRI, unsigned VReg, unsigned *NewReg=nullptr)
static void signOutlinedFunction(MachineFunction &MF, MachineBasicBlock &MBB, const AArch64InstrInfo *TII, bool ShouldSignReturnAddr)
static MachineInstr * genFNegatedMAD(MachineFunction &MF, MachineRegisterInfo &MRI, const TargetInstrInfo *TII, MachineInstr &Root, SmallVectorImpl< MachineInstr * > &InsInstrs)
static bool canCombineWithMUL(MachineBasicBlock &MBB, MachineOperand &MO, unsigned MulOpc, unsigned ZeroReg)
static void storeRegPairToStackSlot(const TargetRegisterInfo &TRI, MachineBasicBlock &MBB, MachineBasicBlock::iterator InsertBefore, const MCInstrDesc &MCID, Register SrcReg, bool IsKill, unsigned SubIdx0, unsigned SubIdx1, int FI, MachineMemOperand *MMO)
assert(UImm &&(UImm !=~static_cast< T >(0)) &&"Invalid immediate!")
MachineBasicBlock MachineBasicBlock::iterator DebugLoc DL
MachineBasicBlock MachineBasicBlock::iterator MBBI
static const Function * getParent(const Value *V)
static GCRegistry::Add< ErlangGC > A("erlang", "erlang-compatible garbage collector")
static GCRegistry::Add< OcamlGC > B("ocaml", "ocaml 3.10-compatible GC")
DXIL Forward Handle Accesses
const HexagonInstrInfo * TII
Module.h This file contains the declarations for the Module class.
const AbstractManglingParser< Derived, Alloc >::OperatorInfo AbstractManglingParser< Derived, Alloc >::Ops[]
This file implements the LivePhysRegs utility for tracking liveness of physical registers.
Machine Check Debug Module
Register const TargetRegisterInfo * TRI
Promote Memory to Register
static MCRegister getReg(const MCDisassembler *D, unsigned RC, unsigned RegNo)
MachineInstr unsigned OpIdx
ConstantRange Range(APInt(BitWidth, Low), APInt(BitWidth, High))
PowerPC Reduce CR logical Operation
const SmallVectorImpl< MachineOperand > MachineBasicBlock * TBB
const SmallVectorImpl< MachineOperand > & Cond
This file declares the machine register scavenger class.
static bool contains(SmallPtrSetImpl< ConstantExpr * > &Cache, ConstantExpr *Expr, Constant *C)
This file defines the SmallSet class.
This file defines the SmallVector class.
#define DEBUG_WITH_TYPE(TYPE,...)
DEBUG_WITH_TYPE macro - This macro should be used by passes to emit debug information.
static bool canCombine(MachineBasicBlock &MBB, MachineOperand &MO, unsigned CombineOpc=0)
AArch64FunctionInfo - This class is derived from MachineFunctionInfo and contains private AArch64-spe...
SignReturnAddress getSignReturnAddressCondition() const
bool hasStreamingModeChanges() const
void setOutliningStyle(const std::string &Style)
std::optional< bool > hasRedZone() const
static bool shouldSignReturnAddress(SignReturnAddress Condition, bool IsLRSpilled)
bool shouldSignWithBKey() const
static bool isHForm(const MachineInstr &MI)
Returns whether the instruction is in H form (16 bit operands)
void insertSelect(MachineBasicBlock &MBB, MachineBasicBlock::iterator MI, const DebugLoc &DL, Register DstReg, ArrayRef< MachineOperand > Cond, Register TrueReg, Register FalseReg) const override
static bool hasBTISemantics(const MachineInstr &MI)
Returns whether the instruction can be compatible with non-zero BTYPE.
static bool isQForm(const MachineInstr &MI)
Returns whether the instruction is in Q form (128 bit operands)
static bool getMemOpInfo(unsigned Opcode, TypeSize &Scale, TypeSize &Width, int64_t &MinOffset, int64_t &MaxOffset)
Returns true if opcode Opc is a memory operation.
static bool isTailCallReturnInst(const MachineInstr &MI)
Returns true if MI is one of the TCRETURN* instructions.
static bool isFPRCopy(const MachineInstr &MI)
Does this instruction rename an FPR without modifying bits?
MachineInstr * emitLdStWithAddr(MachineInstr &MemI, const ExtAddrMode &AM) const override
std::optional< DestSourcePair > isCopyInstrImpl(const MachineInstr &MI) const override
If the specific machine instruction is an instruction that moves/copies value from one register to an...
MachineBasicBlock * getBranchDestBlock(const MachineInstr &MI) const override
unsigned getInstSizeInBytes(const MachineInstr &MI) const override
GetInstSize - Return the number of bytes of code the specified instruction may be.
bool areMemAccessesTriviallyDisjoint(const MachineInstr &MIa, const MachineInstr &MIb) const override
void copyPhysReg(MachineBasicBlock &MBB, MachineBasicBlock::iterator I, const DebugLoc &DL, Register DestReg, Register SrcReg, bool KillSrc, bool RenamableDest=false, bool RenamableSrc=false) const override
static bool isGPRCopy(const MachineInstr &MI)
Does this instruction rename a GPR without modifying bits?
static unsigned convertToFlagSettingOpc(unsigned Opc)
Return the opcode that set flags when possible.
bool isBranchOffsetInRange(unsigned BranchOpc, int64_t BrOffset) const override
bool canInsertSelect(const MachineBasicBlock &, ArrayRef< MachineOperand > Cond, Register, Register, Register, int &, int &, int &) const override
Register isLoadFromStackSlotPostFE(const MachineInstr &MI, int &FrameIndex) const override
Check for post-frame ptr elimination stack locations as well.
static const MachineOperand & getLdStOffsetOp(const MachineInstr &MI)
Returns the immediate offset operator of a load/store.
bool isCoalescableExtInstr(const MachineInstr &MI, Register &SrcReg, Register &DstReg, unsigned &SubIdx) const override
static std::optional< unsigned > getUnscaledLdSt(unsigned Opc)
Returns the unscaled load/store for the scaled load/store opcode, if there is a corresponding unscale...
static bool hasUnscaledLdStOffset(unsigned Opc)
Return true if it has an unscaled load/store offset.
static const MachineOperand & getLdStAmountOp(const MachineInstr &MI)
Returns the shift amount operator of a load/store.
static bool isPreLdSt(const MachineInstr &MI)
Returns whether the instruction is a pre-indexed load/store.
std::optional< ExtAddrMode > getAddrModeFromMemoryOp(const MachineInstr &MemI, const TargetRegisterInfo *TRI) const override
bool getMemOperandsWithOffsetWidth(const MachineInstr &MI, SmallVectorImpl< const MachineOperand * > &BaseOps, int64_t &Offset, bool &OffsetIsScalable, LocationSize &Width, const TargetRegisterInfo *TRI) const override
bool analyzeBranchPredicate(MachineBasicBlock &MBB, MachineBranchPredicate &MBP, bool AllowModify) const override
void insertIndirectBranch(MachineBasicBlock &MBB, MachineBasicBlock &NewDestBB, MachineBasicBlock &RestoreBB, const DebugLoc &DL, int64_t BrOffset, RegScavenger *RS) const override
void loadRegFromStackSlot(MachineBasicBlock &MBB, MachineBasicBlock::iterator MBBI, Register DestReg, int FrameIndex, const TargetRegisterClass *RC, Register VReg, MachineInstr::MIFlag Flags=MachineInstr::NoFlags) const override
static bool isPairableLdStInst(const MachineInstr &MI)
Return true if pairing the given load or store may be paired with another.
void storeRegToStackSlot(MachineBasicBlock &MBB, MachineBasicBlock::iterator MBBI, Register SrcReg, bool isKill, int FrameIndex, const TargetRegisterClass *RC, Register VReg, MachineInstr::MIFlag Flags=MachineInstr::NoFlags) const override
const AArch64RegisterInfo & getRegisterInfo() const
getRegisterInfo - TargetInstrInfo is a superset of MRegister info.
static bool isPreSt(const MachineInstr &MI)
Returns whether the instruction is a pre-indexed store.
MachineInstr * foldMemoryOperandImpl(MachineFunction &MF, MachineInstr &MI, ArrayRef< unsigned > Ops, MachineBasicBlock::iterator InsertPt, int FrameIndex, LiveIntervals *LIS=nullptr, VirtRegMap *VRM=nullptr) const override
void insertNoop(MachineBasicBlock &MBB, MachineBasicBlock::iterator MI) const override
AArch64InstrInfo(const AArch64Subtarget &STI)
static bool isPairedLdSt(const MachineInstr &MI)
Returns whether the instruction is a paired load/store.
bool getMemOperandWithOffsetWidth(const MachineInstr &MI, const MachineOperand *&BaseOp, int64_t &Offset, bool &OffsetIsScalable, TypeSize &Width, const TargetRegisterInfo *TRI) const
If OffsetIsScalable is set to 'true', the offset is scaled by vscale.
Register isLoadFromStackSlot(const MachineInstr &MI, int &FrameIndex) const override
bool reverseBranchCondition(SmallVectorImpl< MachineOperand > &Cond) const override
static bool isStridedAccess(const MachineInstr &MI)
Return true if the given load or store is a strided memory access.
bool shouldClusterMemOps(ArrayRef< const MachineOperand * > BaseOps1, int64_t Offset1, bool OffsetIsScalable1, ArrayRef< const MachineOperand * > BaseOps2, int64_t Offset2, bool OffsetIsScalable2, unsigned ClusterSize, unsigned NumBytes) const override
Detect opportunities for ldp/stp formation.
unsigned removeBranch(MachineBasicBlock &MBB, int *BytesRemoved=nullptr) const override
bool isThroughputPattern(unsigned Pattern) const override
Return true when a code sequence can improve throughput.
MachineOperand & getMemOpBaseRegImmOfsOffsetOperand(MachineInstr &LdSt) const
Return the immediate offset of the base register in a load/store LdSt.
bool analyzeBranch(MachineBasicBlock &MBB, MachineBasicBlock *&TBB, MachineBasicBlock *&FBB, SmallVectorImpl< MachineOperand > &Cond, bool AllowModify=false) const override
bool canFoldIntoAddrMode(const MachineInstr &MemI, Register Reg, const MachineInstr &AddrI, ExtAddrMode &AM) const override
static bool isLdStPairSuppressed(const MachineInstr &MI)
Return true if pairing the given load or store is hinted to be unprofitable.
Register isStoreToStackSlotPostFE(const MachineInstr &MI, int &FrameIndex) const override
Check for post-frame ptr elimination stack locations as well.
std::unique_ptr< TargetInstrInfo::PipelinerLoopInfo > analyzeLoopForPipelining(MachineBasicBlock *LoopBB) const override
bool isSchedulingBoundary(const MachineInstr &MI, const MachineBasicBlock *MBB, const MachineFunction &MF) const override
unsigned insertBranch(MachineBasicBlock &MBB, MachineBasicBlock *TBB, MachineBasicBlock *FBB, ArrayRef< MachineOperand > Cond, const DebugLoc &DL, int *BytesAdded=nullptr) const override
MachineBasicBlock::iterator probedStackAlloc(MachineBasicBlock::iterator MBBI, Register TargetReg, bool FrameSetup) const
Return true when there is potentially a faster code sequence for an instruction chain ending in Root.
bool optimizeCompareInstr(MachineInstr &CmpInstr, Register SrcReg, Register SrcReg2, int64_t CmpMask, int64_t CmpValue, const MachineRegisterInfo *MRI) const override
optimizeCompareInstr - Convert the instruction supplying the argument to the comparison into one that...
static unsigned getLoadStoreImmIdx(unsigned Opc)
Returns the index for the immediate for a given instruction.
static bool isGPRZero(const MachineInstr &MI)
Does this instruction set its full destination register to zero?
void copyGPRRegTuple(MachineBasicBlock &MBB, MachineBasicBlock::iterator I, const DebugLoc &DL, MCRegister DestReg, MCRegister SrcReg, bool KillSrc, unsigned Opcode, unsigned ZeroReg, llvm::ArrayRef< unsigned > Indices) const
bool analyzeCompare(const MachineInstr &MI, Register &SrcReg, Register &SrcReg2, int64_t &CmpMask, int64_t &CmpValue) const override
analyzeCompare - For a comparison instruction, return the source registers in SrcReg and SrcReg2,...
CombinerObjective getCombinerObjective(unsigned Pattern) const override
static bool isFpOrNEON(Register Reg)
Returns whether the physical register is FP or NEON.
bool isAsCheapAsAMove(const MachineInstr &MI) const override
std::optional< DestSourcePair > isCopyLikeInstrImpl(const MachineInstr &MI) const override
static void suppressLdStPair(MachineInstr &MI)
Hint that pairing the given load or store is unprofitable.
Register isStoreToStackSlot(const MachineInstr &MI, int &FrameIndex) const override
static bool isPreLd(const MachineInstr &MI)
Returns whether the instruction is a pre-indexed load.
void copyPhysRegTuple(MachineBasicBlock &MBB, MachineBasicBlock::iterator I, const DebugLoc &DL, MCRegister DestReg, MCRegister SrcReg, bool KillSrc, unsigned Opcode, llvm::ArrayRef< unsigned > Indices) const
bool optimizeCondBranch(MachineInstr &MI) const override
Replace csincr-branch sequence by simple conditional branch.
static int getMemScale(unsigned Opc)
Scaling factor for (scaled or unscaled) load or store.
bool isCandidateToMergeOrPair(const MachineInstr &MI) const
Return true if this is a load/store that can be potentially paired/merged.
MCInst getNop() const override
static const MachineOperand & getLdStBaseOp(const MachineInstr &MI)
Returns the base register operator of a load/store.
bool isReservedReg(const MachineFunction &MF, MCRegister Reg) const
const AArch64RegisterInfo * getRegisterInfo() const override
bool isNeonAvailable() const
Returns true if the target has NEON and the function at runtime is known to have NEON enabled (e....
bool isSVEorStreamingSVEAvailable() const
Returns true if the target has access to either the full range of SVE instructions,...
ArrayRef - Represent a constant reference to an array (0 or more elements consecutively in memory),...
const T & front() const
front - Get the first element.
size_t size() const
size - Get the array size.
This is an important base class in LLVM.
std::pair< iterator, bool > insert(const std::pair< KeyT, ValueT > &KV)
bool hasOptSize() const
Optimize this function for size (-Os) or minimum size (-Oz).
bool hasMinSize() const
Optimize this function for minimum size (-Oz).
Module * getParent()
Get the module that this global value is contained inside of...
LLVM_ABI InstListType::iterator eraseFromParent()
This method unlinks 'this' from the containing basic block and deletes it.
unsigned getOpcode() const
Returns a member of one of the enums like Instruction::Add.
static LocationSize precise(uint64_t Value)
This class is intended to be used as a base class for asm properties and features specific to the tar...
bool usesWindowsCFI() const
static MCCFIInstruction cfiDefCfa(MCSymbol *L, unsigned Register, int64_t Offset, SMLoc Loc={})
.cfi_def_cfa defines a rule for computing CFA as: take address from Register and add Offset to it.
static MCCFIInstruction createOffset(MCSymbol *L, unsigned Register, int64_t Offset, SMLoc Loc={})
.cfi_offset Previous value of Register is saved at offset Offset from CFA.
static MCCFIInstruction cfiDefCfaOffset(MCSymbol *L, int64_t Offset, SMLoc Loc={})
.cfi_def_cfa_offset modifies a rule for computing CFA.
static MCCFIInstruction createEscape(MCSymbol *L, StringRef Vals, SMLoc Loc={}, StringRef Comment="")
.cfi_escape Allows the user to add arbitrary bytes to the unwind info.
Instances of this class represent a single low-level machine instruction.
Describe properties that are true of each instruction in the target description file.
unsigned getNumOperands() const
Return the number of declared MachineOperands for this MachineInstruction.
ArrayRef< MCOperandInfo > operands() const
Wrapper class representing physical registers. Should be passed by value.
constexpr bool isValid() const
static constexpr unsigned NoRegister
static MDTuple * get(LLVMContext &Context, ArrayRef< Metadata * > MDs)
bool isInlineAsmBrIndirectTarget() const
Returns true if this is the indirect dest of an INLINEASM_BR.
unsigned pred_size() const
LLVM_ABI void transferSuccessorsAndUpdatePHIs(MachineBasicBlock *FromMBB)
Transfers all the successors, as in transferSuccessors, and update PHI operands in the successor bloc...
LLVM_ABI instr_iterator insert(instr_iterator I, MachineInstr *M)
Insert MI into the instruction list before I, possibly inside a bundle.
reverse_instr_iterator instr_rbegin()
LLVM_ABI iterator getFirstTerminator()
Returns an iterator to the first terminator instruction of this basic block.
LLVM_ABI void addSuccessor(MachineBasicBlock *Succ, BranchProbability Prob=BranchProbability::getUnknown())
Add Succ as a successor of this MachineBasicBlock.
reverse_instr_iterator instr_rend()
Instructions::iterator instr_iterator
instr_iterator instr_end()
Instructions::const_iterator const_instr_iterator
void addLiveIn(MCRegister PhysReg, LaneBitmask LaneMask=LaneBitmask::getAll())
Adds the specified register as a live in.
const MachineFunction * getParent() const
Return the MachineFunction containing this basic block.
LLVM_ABI instr_iterator erase(instr_iterator I)
Remove an instruction from the instruction list and delete it.
void splice(iterator Where, MachineBasicBlock *Other, iterator From)
Take an instruction from MBB 'Other' at the position From, and insert it into this MBB right before '...
MachineInstrBundleIterator< MachineInstr > iterator
void setMachineBlockAddressTaken()
Set this block to indicate that its address is used as something other than the target of a terminato...
LLVM_ABI bool isLiveIn(MCRegister Reg, LaneBitmask LaneMask=LaneBitmask::getAll()) const
Return true if the specified register is in the live in set.
The MachineFrameInfo class represents an abstract stack frame until prolog/epilog code is inserted.
uint64_t getStackSize() const
Return the number of bytes that must be allocated to hold all of the fixed size frame objects.
void setStackID(int ObjectIdx, uint8_t ID)
bool isCalleeSavedInfoValid() const
Has the callee saved info been calculated yet?
Align getObjectAlign(int ObjectIdx) const
Return the alignment of the specified stack object.
int64_t getObjectSize(int ObjectIdx) const
Return the size of the specified object.
unsigned getNumObjects() const
Return the number of objects.
int64_t getObjectOffset(int ObjectIdx) const
Return the assigned stack offset of the specified object from the incoming stack pointer.
bool isFixedObjectIndex(int ObjectIdx) const
Returns true if the specified index corresponds to a fixed stack object.
unsigned addFrameInst(const MCCFIInstruction &Inst)
const TargetSubtargetInfo & getSubtarget() const
getSubtarget - Return the subtarget for which this machine code is being compiled.
StringRef getName() const
getName - Return the name of the corresponding LLVM function.
MachineMemOperand * getMachineMemOperand(MachinePointerInfo PtrInfo, MachineMemOperand::Flags f, LLT MemTy, Align base_alignment, const AAMDNodes &AAInfo=AAMDNodes(), const MDNode *Ranges=nullptr, SyncScope::ID SSID=SyncScope::System, AtomicOrdering Ordering=AtomicOrdering::NotAtomic, AtomicOrdering FailureOrdering=AtomicOrdering::NotAtomic)
getMachineMemOperand - Allocate a new MachineMemOperand.
MachineFrameInfo & getFrameInfo()
getFrameInfo - Return the frame info object for the current function.
MachineRegisterInfo & getRegInfo()
getRegInfo - Return information about the registers currently in use.
Function & getFunction()
Return the LLVM function that this machine code represents.
BasicBlockListType::iterator iterator
Ty * getInfo()
getInfo - Keep track of various per-function pieces of information for backends that would like to do...
const MachineJumpTableInfo * getJumpTableInfo() const
getJumpTableInfo - Return the jump table info object for the current function.
MachineBasicBlock * CreateMachineBasicBlock(const BasicBlock *BB=nullptr, std::optional< UniqueBBID > BBID=std::nullopt)
CreateMachineInstr - Allocate a new MachineInstr.
void insert(iterator MBBI, MachineBasicBlock *MBB)
const TargetMachine & getTarget() const
getTarget - Return the target machine this machine code is compiled with
const MachineInstrBuilder & setMemRefs(ArrayRef< MachineMemOperand * > MMOs) const
const MachineInstrBuilder & addCFIIndex(unsigned CFIIndex) const
const MachineInstrBuilder & setMIFlag(MachineInstr::MIFlag Flag) const
const MachineInstrBuilder & addImm(int64_t Val) const
Add a new immediate operand.
const MachineInstrBuilder & add(const MachineOperand &MO) const
const MachineInstrBuilder & addSym(MCSymbol *Sym, unsigned char TargetFlags=0) const
const MachineInstrBuilder & addFrameIndex(int Idx) const
const MachineInstrBuilder & addGlobalAddress(const GlobalValue *GV, int64_t Offset=0, unsigned TargetFlags=0) const
const MachineInstrBuilder & addReg(Register RegNo, unsigned flags=0, unsigned SubReg=0) const
Add a new virtual register operand.
const MachineInstrBuilder & addMBB(MachineBasicBlock *MBB, unsigned TargetFlags=0) const
const MachineInstrBuilder & addUse(Register RegNo, unsigned Flags=0, unsigned SubReg=0) const
Add a virtual register use operand.
const MachineInstrBuilder & setMIFlags(unsigned Flags) const
const MachineInstrBuilder & addMemOperand(MachineMemOperand *MMO) const
const MachineInstrBuilder & addDef(Register RegNo, unsigned Flags=0, unsigned SubReg=0) const
Add a virtual register definition operand.
reverse_iterator getReverse() const
Get a reverse iterator to the same node.
Representation of each machine instruction.
unsigned getOpcode() const
Returns the opcode of this MachineInstr.
bool mayLoadOrStore(QueryType Type=AnyInBundle) const
Return true if this instruction could possibly read or modify memory.
const MachineBasicBlock * getParent() const
bool isCall(QueryType Type=AnyInBundle) const
bool getFlag(MIFlag Flag) const
Return whether an MI flag is set.
LLVM_ABI uint32_t mergeFlagsWith(const MachineInstr &Other) const
Return the MIFlags which represent both MachineInstrs.
unsigned getNumOperands() const
Retuns the total number of operands.
LLVM_ABI unsigned getNumExplicitOperands() const
Returns the number of non-implicit operands.
bool mayLoad(QueryType Type=AnyInBundle) const
Return true if this instruction could possibly read memory.
const MCInstrDesc & getDesc() const
Returns the target instruction descriptor of this MachineInstr.
LLVM_ABI bool hasUnmodeledSideEffects() const
Return true if this instruction has side effects that are not modeled by mayLoad / mayStore,...
bool registerDefIsDead(Register Reg, const TargetRegisterInfo *TRI) const
Returns true if the register is dead in this machine instruction.
bool definesRegister(Register Reg, const TargetRegisterInfo *TRI) const
Return true if the MachineInstr fully defines the specified register.
LLVM_ABI void setDesc(const MCInstrDesc &TID)
Replace the instruction descriptor (thus opcode) of the current instruction with a new one.
LLVM_ABI bool hasOrderedMemoryRef() const
Return true if this instruction may have an ordered or volatile memory reference, or if the informati...
LLVM_ABI const MachineFunction * getMF() const
Return the function that contains the basic block that this instruction belongs to.
ArrayRef< MachineMemOperand * > memoperands() const
Access to memory operands of the instruction.
LLVM_ABI bool isLoadFoldBarrier() const
Returns true if it is illegal to fold a load across this instruction.
const DebugLoc & getDebugLoc() const
Returns the debug location id of this MachineInstr.
LLVM_ABI void eraseFromParent()
Unlink 'this' from the containing basic block and delete it.
LLVM_ABI void removeOperand(unsigned OpNo)
Erase an operand from an instruction, leaving it with one fewer operand than it started with.
LLVM_ABI void addRegisterDefined(Register Reg, const TargetRegisterInfo *RegInfo=nullptr)
We have determined MI defines a register.
const MachineOperand & getOperand(unsigned i) const
uint32_t getFlags() const
Return the MI flags bitvector.
LLVM_ABI int findRegisterDefOperandIdx(Register Reg, const TargetRegisterInfo *TRI, bool isDead=false, bool Overlap=false) const
Returns the operand index that is a def of the specified register or -1 if it is not found.
const std::vector< MachineJumpTableEntry > & getJumpTables() const
A description of a memory reference used in the backend.
@ MOLoad
The memory access reads data.
@ MOStore
The memory access writes data.
This class contains meta information specific to a module.
LLVM_ABI MachineFunction * getMachineFunction(const Function &F) const
Returns the MachineFunction associated to IR function F if there is one, otherwise nullptr.
MachineOperand class - Representation of each machine instruction operand.
void setSubReg(unsigned subReg)
unsigned getSubReg() const
void setImm(int64_t immVal)
bool isReg() const
isReg - Tests if this is a MO_Register operand.
MachineBasicBlock * getMBB() const
void setIsDead(bool Val=true)
LLVM_ABI void setReg(Register Reg)
Change the register this operand corresponds to.
bool isImm() const
isImm - Tests if this is a MO_Immediate operand.
void setIsKill(bool Val=true)
MachineInstr * getParent()
getParent - Return the instruction that this operand belongs to.
unsigned getTargetFlags() const
static MachineOperand CreateImm(int64_t Val)
MachineOperandType getType() const
getType - Returns the MachineOperandType for this operand.
void setIsUndef(bool Val=true)
Register getReg() const
getReg - Returns the register number.
bool isFI() const
isFI - Tests if this is a MO_FrameIndex operand.
LLVM_ABI bool isIdenticalTo(const MachineOperand &Other) const
Returns true if this operand is identical to the specified operand except for liveness related flags ...
static MachineOperand CreateReg(Register Reg, bool isDef, bool isImp=false, bool isKill=false, bool isDead=false, bool isUndef=false, bool isEarlyClobber=false, unsigned SubReg=0, bool isDebug=false, bool isInternalRead=false, bool isRenamable=false)
MachineRegisterInfo - Keep track of information for virtual and physical registers,...
bool tracksLiveness() const
tracksLiveness - Returns true when tracking register liveness accurately.
bool reservedRegsFrozen() const
reservedRegsFrozen - Returns true after freezeReservedRegs() was called to ensure the set of reserved...
const TargetRegisterClass * getRegClassOrNull(Register Reg) const
Return the register class of Reg, or null if Reg has not been assigned a register class yet.
LLVM_ABI const TargetRegisterClass * constrainRegClass(Register Reg, const TargetRegisterClass *RC, unsigned MinNumRegs=0)
constrainRegClass - Constrain the register class of the specified virtual register to be a common sub...
LLVM_ABI MachineInstr * getUniqueVRegDef(Register Reg) const
getUniqueVRegDef - Return the unique machine instr that defines the specified virtual register or nul...
A Module instance is used to store all the information related to an LLVM module.
MI-level patchpoint operands.
uint32_t getNumPatchBytes() const
Return the number of patchable bytes the given patchpoint should emit.
Wrapper class representing virtual and physical registers.
MCRegister asMCReg() const
Utility to check-convert this value to a MCRegister.
constexpr bool isValid() const
constexpr bool isVirtual() const
Return true if the specified register number is in the virtual register namespace.
static constexpr bool isVirtualRegister(unsigned Reg)
Return true if the specified register number is in the virtual register namespace.
constexpr bool isPhysical() const
Return true if the specified register number is in the physical register namespace.
Represents a location in source code.
bool erase(PtrType Ptr)
Remove pointer from the set.
std::pair< iterator, bool > insert(PtrType Ptr)
Inserts Ptr if and only if there is no element in the container equal to Ptr.
SmallPtrSet - This class implements a set which is optimized for holding SmallSize or less elements.
SmallSet - This maintains a set of unique values, optimizing for the case when the set is small (less...
SmallString - A SmallString is just a SmallVector with methods and accessors that make it work better...
void append(StringRef RHS)
Append from a StringRef.
StringRef str() const
Explicit conversion to StringRef.
This class consists of common code factored out of the SmallVector class to reduce code duplication b...
void push_back(const T &Elt)
This is a 'vector' (really, a variable-sized array), optimized for the case when the array is small.
MI-level stackmap operands.
uint32_t getNumPatchBytes() const
Return the number of patchable bytes the given stackmap should emit.
StackOffset holds a fixed and a scalable offset in bytes.
int64_t getFixed() const
Returns the fixed component of the stack.
int64_t getScalable() const
Returns the scalable component of the stack.
static StackOffset get(int64_t Fixed, int64_t Scalable)
static StackOffset getScalable(int64_t Scalable)
static StackOffset getFixed(int64_t Fixed)
MI-level Statepoint operands.
uint32_t getNumPatchBytes() const
Return the number of patchable bytes the given statepoint should emit.
StringRef - Represent a constant reference to a string, i.e.
Object returned by analyzeLoopForPipelining.
TargetInstrInfo - Interface to description of machine instruction set.
virtual void genAlternativeCodeSequence(MachineInstr &Root, unsigned Pattern, SmallVectorImpl< MachineInstr * > &InsInstrs, SmallVectorImpl< MachineInstr * > &DelInstrs, DenseMap< Register, unsigned > &InstIdxForVirtReg) const
When getMachineCombinerPatterns() finds patterns, this function generates the instructions that could...
virtual std::optional< ParamLoadedValue > describeLoadedValue(const MachineInstr &MI, Register Reg) const
Produce the expression describing the MI loading a value into the physical register Reg.
virtual bool getMachineCombinerPatterns(MachineInstr &Root, SmallVectorImpl< unsigned > &Patterns, bool DoRegPressureReduce) const
Return true when there is potentially a faster code sequence for an instruction chain ending in Root.
virtual bool isSchedulingBoundary(const MachineInstr &MI, const MachineBasicBlock *MBB, const MachineFunction &MF) const
Test if the given instruction should be considered a scheduling boundary.
virtual CombinerObjective getCombinerObjective(unsigned Pattern) const
Return the objective of a combiner pattern.
virtual bool isFunctionSafeToSplit(const MachineFunction &MF) const
Return true if the function is a viable candidate for machine function splitting.
CodeModel::Model getCodeModel() const
Returns the code model.
const MCAsmInfo * getMCAsmInfo() const
Return target specific asm information.
bool contains(Register Reg) const
Return true if the specified register is included in this register class.
bool hasSubClassEq(const TargetRegisterClass *RC) const
Returns true if RC is a sub-class of or equal to this class.
bool hasSuperClassEq(const TargetRegisterClass *RC) const
Returns true if RC is a super-class of or equal to this class.
TargetRegisterInfo base class - We assume that the target defines a static array of TargetRegisterDes...
TargetSubtargetInfo - Generic base class for all target subtargets.
virtual const TargetInstrInfo * getInstrInfo() const
virtual const TargetRegisterInfo * getRegisterInfo() const =0
Return the target's register information.
Target - Wrapper for Target specific information.
static constexpr TypeSize getFixed(ScalarTy ExactSize)
static constexpr TypeSize getScalable(ScalarTy MinimumSize)
Value * getOperand(unsigned i) const
constexpr bool isScalable() const
Returns whether the quantity is scaled by a runtime quantity (vscale).
constexpr ScalarTy getKnownMinValue() const
Returns the minimum value this quantity can represent.
self_iterator getIterator()
A raw_ostream that writes to an std::string.
#define llvm_unreachable(msg)
Marks that the current location is not supposed to be reachable.
static CondCode getInvertedCondCode(CondCode Code)
@ MO_DLLIMPORT
MO_DLLIMPORT - On a symbol operand, this represents that the reference to the symbol is for an import...
@ MO_NC
MO_NC - Indicates whether the linker is expected to check the symbol reference for overflow.
@ MO_G1
MO_G1 - A symbol operand with this flag (granule 1) represents the bits 16-31 of a 64-bit address,...
@ MO_S
MO_S - Indicates that the bits of the symbol operand represented by MO_G0 etc are signed.
@ MO_PAGEOFF
MO_PAGEOFF - A symbol operand with this flag represents the offset of that symbol within a 4K page.
@ MO_GOT
MO_GOT - This flag indicates that a symbol operand represents the address of the GOT entry for the sy...
@ MO_PREL
MO_PREL - Indicates that the bits of the symbol operand represented by MO_G0 etc are PC relative.
@ MO_G0
MO_G0 - A symbol operand with this flag (granule 0) represents the bits 0-15 of a 64-bit address,...
@ MO_ARM64EC_CALLMANGLE
MO_ARM64EC_CALLMANGLE - Operand refers to the Arm64EC-mangled version of a symbol,...
@ MO_PAGE
MO_PAGE - A symbol operand with this flag represents the pc-relative offset of the 4K page containing...
@ MO_HI12
MO_HI12 - This flag indicates that a symbol operand represents the bits 13-24 of a 64-bit address,...
@ MO_TLS
MO_TLS - Indicates that the operand being accessed is some kind of thread-local symbol.
@ MO_G2
MO_G2 - A symbol operand with this flag (granule 2) represents the bits 32-47 of a 64-bit address,...
@ MO_TAGGED
MO_TAGGED - With MO_PAGE, indicates that the page includes a memory tag in bits 56-63.
@ MO_G3
MO_G3 - A symbol operand with this flag (granule 3) represents the high 16-bits of a 64-bit address,...
@ MO_COFFSTUB
MO_COFFSTUB - On a symbol operand "FOO", this indicates that the reference is actually to the "....
unsigned getCheckerSizeInBytes(AuthCheckMethod Method)
Returns the number of bytes added by checkAuthenticatedRegister.
static uint64_t decodeLogicalImmediate(uint64_t val, unsigned regSize)
decodeLogicalImmediate - Decode a logical immediate value in the form "N:immr:imms" (where the immr a...
static unsigned getShiftValue(unsigned Imm)
getShiftValue - Extract the shift value.
static unsigned getArithExtendImm(AArch64_AM::ShiftExtendType ET, unsigned Imm)
getArithExtendImm - Encode the extend type and shift amount for an arithmetic instruction: imm: 3-bit...
static unsigned getArithShiftValue(unsigned Imm)
getArithShiftValue - get the arithmetic shift value.
static uint64_t encodeLogicalImmediate(uint64_t imm, unsigned regSize)
encodeLogicalImmediate - Return the encoded immediate value for a logical immediate instruction of th...
static AArch64_AM::ShiftExtendType getExtendType(unsigned Imm)
getExtendType - Extract the extend type for operands of arithmetic ops.
static AArch64_AM::ShiftExtendType getArithExtendType(unsigned Imm)
static AArch64_AM::ShiftExtendType getShiftType(unsigned Imm)
getShiftType - Extract the shift type.
static unsigned getShifterImm(AArch64_AM::ShiftExtendType ST, unsigned Imm)
getShifterImm - Encode the shift type and amount: imm: 6-bit shift amount shifter: 000 ==> lsl 001 ==...
void expandMOVImm(uint64_t Imm, unsigned BitSize, SmallVectorImpl< ImmInsnModel > &Insn)
Expand a MOVi32imm or MOVi64imm pseudo instruction to one or more real move-immediate instructions to...
static const uint64_t InstrFlagIsWhile
static const uint64_t InstrFlagIsPTestLike
constexpr std::underlying_type_t< E > Mask()
Get a bitmask with 1s in all places up to the high-order bit of E's largest value.
@ C
The default llvm calling convention, compatible with C.
@ Implicit
Not emitted register (e.g. carry, or temporary result).
@ Renamable
Register that may be renamed.
@ Define
Register definition.
@ Kill
The last use of a register.
@ Undef
Value of the register doesn't matter.
@ ScalablePredicateVector
initializer< Ty > init(const Ty &Val)
InstrType
Represents how an instruction should be mapped by the outliner.
LLVM_ABI Instruction & back() const
This is an optimization pass for GlobalISel generic memory operations.
bool all_of(R &&range, UnaryPredicate P)
Provide wrappers to std::all_of which take ranges instead of having to pass begin/end explicitly.
static bool isCondBranchOpcode(int Opc)
MCCFIInstruction createDefCFA(const TargetRegisterInfo &TRI, unsigned FrameReg, unsigned Reg, const StackOffset &Offset, bool LastAdjustmentWasScalable=true)
static bool isPTrueOpcode(unsigned Opc)
MachineInstrBuilder BuildMI(MachineFunction &MF, const MIMetadata &MIMD, const MCInstrDesc &MCID)
Builder interface. Specify how to create the initial instruction itself.
bool succeeded(LogicalResult Result)
Utility function that returns true if the provided LogicalResult corresponds to a success value.
int isAArch64FrameOffsetLegal(const MachineInstr &MI, StackOffset &Offset, bool *OutUseUnscaledOp=nullptr, unsigned *OutUnscaledOp=nullptr, int64_t *EmittableOffset=nullptr)
Check if the Offset is a valid frame offset for MI.
constexpr bool isInt(int64_t x)
Checks if an integer fits into the given bit width.
auto enumerate(FirstRange &&First, RestRanges &&...Rest)
Given two or more input ranges, returns a new range whose values are tuples (A, B,...
decltype(auto) dyn_cast(const From &Val)
dyn_cast<X> - Return the argument parameter cast to the specified type.
static bool isIndirectBranchOpcode(int Opc)
iterator_range< T > make_range(T x, T y)
Convenience function for iterating over sub-ranges.
unsigned getBLRCallOpcode(const MachineFunction &MF)
Return opcode to be used for indirect calls.
@ AArch64FrameOffsetIsLegal
Offset is legal.
@ AArch64FrameOffsetCanUpdate
Offset can apply, at least partly.
@ AArch64FrameOffsetCannotUpdate
Offset cannot apply.
constexpr bool isPowerOf2_64(uint64_t Value)
Return true if the argument is a power of two > 0 (64 bit edition.)
unsigned Log2_64(uint64_t Value)
Return the floor log base 2 of the specified value, -1 if the value is zero.
static bool isSEHInstruction(const MachineInstr &MI)
bool any_of(R &&range, UnaryPredicate P)
Provide wrappers to std::any_of which take ranges instead of having to pass begin/end explicitly.
auto reverse(ContainerTy &&C)
MachineInstr * getImm(const MachineOperand &MO, const MachineRegisterInfo *MRI)
decltype(auto) get(const PointerIntPair< PointerTy, IntBits, IntType, PtrTraits, Info > &Pair)
void sort(IteratorTy Start, IteratorTy End)
AArch64MachineCombinerPattern
@ MULSUBv2i32_indexed_OP1
@ MULADDv4i16_indexed_OP2
@ MULSUBv8i16_indexed_OP2
@ MULSUBv4i16_indexed_OP2
@ MULSUBv4i32_indexed_OP2
@ MULADDv2i32_indexed_OP1
@ MULADDv4i32_indexed_OP1
@ MULADDv2i32_indexed_OP2
@ MULSUBv4i16_indexed_OP1
@ MULADDv4i32_indexed_OP2
@ MULSUBv8i16_indexed_OP1
@ MULSUBv2i32_indexed_OP2
@ MULADDv8i16_indexed_OP1
@ MULSUBv4i32_indexed_OP1
@ MULADDv8i16_indexed_OP2
@ MULADDv4i16_indexed_OP1
LLVM_ABI raw_ostream & dbgs()
dbgs() - This returns a reference to a raw_ostream for debugging messages.
void emitFrameOffset(MachineBasicBlock &MBB, MachineBasicBlock::iterator MBBI, const DebugLoc &DL, unsigned DestReg, unsigned SrcReg, StackOffset Offset, const TargetInstrInfo *TII, MachineInstr::MIFlag=MachineInstr::NoFlags, bool SetNZCV=false, bool NeedsWinCFI=false, bool *HasWinCFI=nullptr, bool EmitCFAOffset=false, StackOffset InitialOffset={}, unsigned FrameReg=AArch64::SP)
emitFrameOffset - Emit instructions as needed to set DestReg to SrcReg plus Offset.
LLVM_ABI void report_fatal_error(Error Err, bool gen_crash_diag=true)
CombinerObjective
The combiner's goal may differ based on which pattern it is attempting to optimize.
std::optional< UsedNZCV > examineCFlagsUse(MachineInstr &MI, MachineInstr &CmpInstr, const TargetRegisterInfo &TRI, SmallVectorImpl< MachineInstr * > *CCUseInstrs=nullptr)
CodeGenOptLevel
Code generation optimization level.
class LLVM_GSL_OWNER SmallVector
Forward declaration of SmallVector so that calculateSmallVectorDefaultInlinedElements can reference s...
auto instructionsWithoutDebug(IterT It, IterT End, bool SkipPseudoOp=true)
Construct a range iterator which begins at It and moves forwards until End is reached,...
LLVM_ABI raw_fd_ostream & errs()
This returns a reference to a raw_ostream for standard error.
auto drop_end(T &&RangeOrContainer, size_t N=1)
Return a range covering RangeOrContainer with the last N elements excluded.
unsigned getUndefRegState(bool B)
static MCRegister getXRegFromWReg(MCRegister Reg)
unsigned getDefRegState(bool B)
MCCFIInstruction createCFAOffset(const TargetRegisterInfo &MRI, unsigned Reg, const StackOffset &OffsetFromDefCFA, std::optional< int64_t > IncomingVGOffsetFromDefCFA)
unsigned getKillRegState(bool B)
FunctionAddr VTableAddr Next
DWARFExpression::Operation Op
ArrayRef(const T &OneElt) -> ArrayRef< T >
static bool isUncondBranchOpcode(int Opc)
decltype(auto) cast(const From &Val)
cast<X> - Return the argument parameter cast to the specified type.
void erase_if(Container &C, UnaryPredicate P)
Provide a container algorithm similar to C++ Library Fundamentals v2's erase_if which is equivalent t...
constexpr bool isIntN(unsigned N, int64_t x)
Checks if an signed integer fits into the given (dynamic) bit width.
bool rewriteAArch64FrameIndex(MachineInstr &MI, unsigned FrameRegIdx, unsigned FrameReg, StackOffset &Offset, const AArch64InstrInfo *TII)
rewriteAArch64FrameIndex - Rewrite MI to access 'Offset' bytes from the FP.
bool is_contained(R &&Range, const E &Element)
Returns true if Element is found in Range.
static const MachineMemOperand::Flags MOSuppressPair
constexpr int64_t SignExtend64(uint64_t x)
Sign-extend the number in the bottom B bits of X to a 64-bit integer.
void appendLEB128(SmallVectorImpl< U > &Buffer, T Value)
bool optimizeTerminators(MachineBasicBlock *MBB, const TargetInstrInfo &TII)
std::pair< MachineOperand, DIExpression * > ParamLoadedValue
bool isNZCVTouchedInInstructionRange(const MachineInstr &DefMI, const MachineInstr &UseMI, const TargetRegisterInfo *TRI)
Return true if there is an instruction /after/ DefMI and before UseMI which either reads or clobbers ...
auto seq(T Begin, T End)
Iterate over an integral type from Begin up to - but not including - End.
static const MachineMemOperand::Flags MOStridedAccess
void fullyRecomputeLiveIns(ArrayRef< MachineBasicBlock * > MBBs)
Convenience function for recomputing live-in's for a set of MBBs until the computation converges.
LLVM_ABI Printable printReg(Register Reg, const TargetRegisterInfo *TRI=nullptr, unsigned SubIdx=0, const MachineRegisterInfo *MRI=nullptr)
Prints virtual and physical registers with or without a TRI instance.
Used to describe addressing mode similar to ExtAddrMode in CodeGenPrepare.
LLVM_ABI static const MBBSectionID ColdSectionID
This class contains a discriminated union of information about pointers in memory operands,...
static LLVM_ABI MachinePointerInfo getFixedStack(MachineFunction &MF, int FI, int64_t Offset=0)
Return a MachinePointerInfo record that refers to the specified FrameIndex.
An individual sequence of instructions to be replaced with a call to an outlined function.
MachineFunction * getMF() const
The information necessary to create an outlined function for some class of candidate.