66#define GET_INSTRINFO_CTOR_DTOR
67#include "AArch64GenInstrInfo.inc"
69#define DEBUG_TYPE "AArch64InstrInfo"
71STATISTIC(NumCopyInstrs,
"Number of COPY instructions expanded");
72STATISTIC(NumZCRegMoveInstrsGPR,
"Number of zero-cycle GPR register move "
73 "instructions expanded from canonical COPY");
74STATISTIC(NumZCRegMoveInstrsFPR,
"Number of zero-cycle FPR register move "
75 "instructions expanded from canonical COPY");
76STATISTIC(NumZCZeroingInstrsGPR,
"Number of zero-cycle GPR zeroing "
77 "instructions expanded from canonical COPY");
82 cl::desc(
"Restrict range of CB instructions (DEBUG)"));
86 cl::desc(
"Restrict range of TB[N]Z instructions (DEBUG)"));
90 cl::desc(
"Restrict range of CB[N]Z instructions (DEBUG)"));
94 cl::desc(
"Restrict range of Bcc instructions (DEBUG)"));
98 cl::desc(
"Restrict range of B instructions (DEBUG)"));
102 cl::desc(
"Restrict range of instructions to search for the "
103 "machine-combiner gather pattern optimization"));
108 RI(STI.getTargetTriple(), STI.getHwMode()), Subtarget(STI) {}
118 switch (
MI.getOpcode()) {
129 if (
MI.getOperand(0).getReg() != AArch64::LR)
138 if (MO.isReg() && MO.isDef() && MO.getReg() == AArch64::LR)
156 auto Op =
MI.getOpcode();
157 if (
Op == AArch64::INLINEASM ||
Op == AArch64::INLINEASM_BR)
158 return getInlineAsmLength(
MI.getOperand(0).getSymbolName(), *MAI);
162 if (
MI.isMetaInstruction())
167 unsigned NumBytes = 0;
177 NumBytes =
Desc.getSize() ?
Desc.getSize() : 4;
180 if (!MFI->shouldSignReturnAddress(*MF))
183 auto Method = STI.getAuthenticatedLRCheckMethod(*MF);
191 switch (
Desc.getOpcode()) {
194 return Desc.getSize();
201 case TargetOpcode::STACKMAP:
204 assert(NumBytes % 4 == 0 &&
"Invalid number of NOP bytes requested!");
206 case TargetOpcode::PATCHPOINT:
209 assert(NumBytes % 4 == 0 &&
"Invalid number of NOP bytes requested!");
211 case TargetOpcode::STATEPOINT:
213 assert(NumBytes % 4 == 0 &&
"Invalid number of NOP bytes requested!");
218 case TargetOpcode::PATCHABLE_FUNCTION_ENTER:
223 F.getFnAttributeAsParsedInteger(
"patchable-function-entry", 9) * 4;
225 case TargetOpcode::PATCHABLE_FUNCTION_EXIT:
226 case TargetOpcode::PATCHABLE_TAIL_CALL:
227 case TargetOpcode::PATCHABLE_TYPED_EVENT_CALL:
231 case TargetOpcode::PATCHABLE_EVENT_CALL:
237 NumBytes =
MI.getOperand(1).getImm();
239 case TargetOpcode::BUNDLE:
240 NumBytes = getInstBundleSize(
MI);
276 case AArch64::CBWPri:
277 case AArch64::CBXPri:
278 case AArch64::CBWPrr:
279 case AArch64::CBXPrr:
287 case AArch64::CBBAssertExt:
288 case AArch64::CBHAssertExt:
319 case AArch64::CBWPri:
320 case AArch64::CBXPri:
321 case AArch64::CBBAssertExt:
322 case AArch64::CBHAssertExt:
323 case AArch64::CBWPrr:
324 case AArch64::CBXPrr:
330 int64_t BrOffset)
const {
332 assert(Bits >= 3 &&
"max branch displacement must be enough to jump"
333 "over conditional branch expansion");
334 return isIntN(Bits, BrOffset / 4);
339 switch (
MI.getOpcode()) {
343 return MI.getOperand(0).getMBB();
348 return MI.getOperand(2).getMBB();
354 return MI.getOperand(1).getMBB();
355 case AArch64::CBWPri:
356 case AArch64::CBXPri:
357 case AArch64::CBBAssertExt:
358 case AArch64::CBHAssertExt:
359 case AArch64::CBWPrr:
360 case AArch64::CBXPrr:
361 return MI.getOperand(3).getMBB();
371 assert(RS &&
"RegScavenger required for long branching");
373 "new block should be inserted for expanding unconditional branch");
376 "restore block should be inserted for restoring clobbered registers");
383 "Branch offsets outside of the signed 33-bit range not supported");
394 RS->enterBasicBlockEnd(
MBB);
397 constexpr Register Reg = AArch64::X16;
398 if (!RS->isRegUsed(Reg)) {
399 insertUnconditionalBranch(
MBB, &NewDestBB,
DL);
406 Register Scavenged = RS->FindUnusedReg(&AArch64::GPR64RegClass);
407 if (Scavenged != AArch64::NoRegister &&
409 buildIndirectBranch(Scavenged, NewDestBB);
410 RS->setRegUsed(Scavenged);
419 "Unable to insert indirect branch inside function that has red zone");
442 bool AllowModify)
const {
449 if (
I->getOpcode() == AArch64::SpeculationBarrierISBDSBEndBB ||
450 I->getOpcode() == AArch64::SpeculationBarrierSBEndBB) {
454 if (!isUnpredicatedTerminator(*
I))
461 unsigned LastOpc = LastInst->
getOpcode();
462 if (
I ==
MBB.begin() || !isUnpredicatedTerminator(*--
I)) {
477 unsigned SecondLastOpc = SecondLastInst->
getOpcode();
484 LastInst = SecondLastInst;
486 if (
I ==
MBB.begin() || !isUnpredicatedTerminator(*--
I)) {
491 SecondLastInst = &*
I;
492 SecondLastOpc = SecondLastInst->
getOpcode();
503 LastInst = SecondLastInst;
505 if (
I ==
MBB.begin() || !isUnpredicatedTerminator(*--
I)) {
507 "unreachable unconditional branches removed above");
516 SecondLastInst = &*
I;
517 SecondLastOpc = SecondLastInst->
getOpcode();
521 if (SecondLastInst &&
I !=
MBB.begin() && isUnpredicatedTerminator(*--
I))
537 I->eraseFromParent();
546 I->eraseFromParent();
555 MachineBranchPredicate &MBP,
556 bool AllowModify)
const {
568 assert(MBP.TrueDest &&
"expected!");
569 MBP.FalseDest = FBB ? FBB :
MBB.getNextNode();
571 MBP.ConditionDef =
nullptr;
572 MBP.SingleUseCondition =
false;
582 if (
I ==
MBB.begin())
598 if (
MI.modifiesRegister(AArch64::NZCV,
nullptr)) {
599 MBP.ConditionDef = &
MI;
608 case AArch64::CBNZX: {
612 MBP.Predicate = (
Opc == AArch64::CBNZX ||
Opc == AArch64::CBNZW)
613 ? MachineBranchPredicate::PRED_NE
614 : MachineBranchPredicate::PRED_EQ;
615 Register CondReg = MBP.LHS.getReg();
624 case AArch64::TBNZX: {
645 Cond[1].setImm(AArch64::CBNZW);
648 Cond[1].setImm(AArch64::CBZW);
651 Cond[1].setImm(AArch64::CBNZX);
654 Cond[1].setImm(AArch64::CBZX);
657 Cond[1].setImm(AArch64::TBNZW);
660 Cond[1].setImm(AArch64::TBZW);
663 Cond[1].setImm(AArch64::TBNZX);
666 Cond[1].setImm(AArch64::TBZX);
670 case AArch64::CBWPri:
671 case AArch64::CBXPri:
672 case AArch64::CBBAssertExt:
673 case AArch64::CBHAssertExt:
674 case AArch64::CBWPrr:
675 case AArch64::CBXPrr: {
688 int *BytesRemoved)
const {
698 I->eraseFromParent();
702 if (
I ==
MBB.begin()) {
715 I->eraseFromParent();
722void AArch64InstrInfo::instantiateCondBranch(
747 if (
Cond.size() > 5) {
758 assert(
TBB &&
"insertBranch must not be told to insert a fallthrough");
785 unsigned Opc =
MI.getOpcode();
792 if (
MI.getOperand(0).getReg() == AArch64::WZR ||
793 MI.getOperand(0).getReg() == AArch64::XZR) {
795 dbgs() <<
"Removing always taken branch: " <<
MI);
798 for (
auto *S : Succs)
800 MBB->removeSuccessor(S);
802 while (
MBB->rbegin() != &
MI)
803 MBB->rbegin()->eraseFromParent();
804 MI.eraseFromParent();
814 if (
MI.getOperand(0).getReg() == AArch64::WZR ||
815 MI.getOperand(0).getReg() == AArch64::XZR) {
817 dbgs() <<
"Removing never taken branch: " <<
MI);
819 MI.getParent()->removeSuccessor(
Target);
820 MI.eraseFromParent();
833 if (!
DefMI->isFullCopy())
835 VReg =
DefMI->getOperand(1).getReg();
844 unsigned *NewReg =
nullptr) {
849 bool Is64Bit = AArch64::GPR64allRegClass.hasSubClassEq(MRI.
getRegClass(VReg));
853 switch (
DefMI->getOpcode()) {
854 case AArch64::SUBREG_TO_REG:
858 if (!
DefMI->getOperand(1).isReg())
860 if (!
DefMI->getOperand(2).isImm() ||
861 DefMI->getOperand(2).getImm() != AArch64::sub_32)
864 if (
DefMI->getOpcode() != AArch64::MOVi32imm)
866 if (!
DefMI->getOperand(1).isImm() ||
DefMI->getOperand(1).getImm() != 1)
869 SrcReg = AArch64::XZR;
870 Opc = AArch64::CSINCXr;
873 case AArch64::MOVi32imm:
874 case AArch64::MOVi64imm:
875 if (!
DefMI->getOperand(1).isImm() ||
DefMI->getOperand(1).getImm() != 1)
877 SrcReg = Is64Bit ? AArch64::XZR : AArch64::WZR;
878 Opc = Is64Bit ? AArch64::CSINCXr : AArch64::CSINCWr;
881 case AArch64::ADDSXri:
882 case AArch64::ADDSWri:
884 if (
DefMI->findRegisterDefOperandIdx(AArch64::NZCV,
nullptr,
889 case AArch64::ADDXri:
890 case AArch64::ADDWri:
892 if (!
DefMI->getOperand(2).isImm() ||
DefMI->getOperand(2).getImm() != 1 ||
893 DefMI->getOperand(3).getImm() != 0)
895 SrcReg =
DefMI->getOperand(1).getReg();
896 Opc = Is64Bit ? AArch64::CSINCXr : AArch64::CSINCWr;
899 case AArch64::ORNXrr:
900 case AArch64::ORNWrr: {
903 if (ZReg != AArch64::XZR && ZReg != AArch64::WZR)
905 SrcReg =
DefMI->getOperand(2).getReg();
906 Opc = Is64Bit ? AArch64::CSINVXr : AArch64::CSINVWr;
910 case AArch64::SUBSXrr:
911 case AArch64::SUBSWrr:
913 if (
DefMI->findRegisterDefOperandIdx(AArch64::NZCV,
nullptr,
918 case AArch64::SUBXrr:
919 case AArch64::SUBWrr: {
922 if (ZReg != AArch64::XZR && ZReg != AArch64::WZR)
924 SrcReg =
DefMI->getOperand(2).getReg();
925 Opc = Is64Bit ? AArch64::CSNEGXr : AArch64::CSNEGWr;
931 assert(
Opc && SrcReg &&
"Missing parameters");
943 int &FalseCycles)
const {
954 if (!RI.getCommonSubClass(RC, MRI.
getRegClass(DstReg)))
958 unsigned ExtraCondLat =
Cond.size() != 1;
962 if (AArch64::GPR64allRegClass.hasSubClassEq(RC) ||
963 AArch64::GPR32allRegClass.hasSubClassEq(RC)) {
965 CondCycles = 1 + ExtraCondLat;
966 TrueCycles = FalseCycles = 1;
976 if (AArch64::FPR64RegClass.hasSubClassEq(RC) ||
977 AArch64::FPR32RegClass.hasSubClassEq(RC)) {
978 CondCycles = 5 + ExtraCondLat;
979 TrueCycles = FalseCycles = 2;
996 switch (
Cond.size()) {
1016 case AArch64::CBNZW:
1020 case AArch64::CBNZX:
1051 case AArch64::TBNZW:
1052 case AArch64::TBNZX:
1074 unsigned SubsOpc, SubsDestReg;
1080 case AArch64::CBWPri:
1081 SubsOpc = AArch64::SUBSWri;
1082 SubsDestReg = AArch64::WZR;
1085 case AArch64::CBXPri:
1086 SubsOpc = AArch64::SUBSXri;
1087 SubsDestReg = AArch64::XZR;
1090 case AArch64::CBWPrr:
1091 SubsOpc = AArch64::SUBSWrr;
1092 SubsDestReg = AArch64::WZR;
1095 case AArch64::CBXPrr:
1096 SubsOpc = AArch64::SUBSXrr;
1097 SubsDestReg = AArch64::XZR;
1126 switch (ExtendType) {
1132 "Unexpected compare-and-branch instruction for SXTB shift-extend");
1133 ExtOpc = AArch64::SBFMWri;
1139 "Unexpected compare-and-branch instruction for SXTH shift-extend");
1140 ExtOpc = AArch64::SBFMWri;
1146 "Unexpected compare-and-branch instruction for UXTB shift-extend");
1147 ExtOpc = AArch64::ANDWri;
1153 "Unexpected compare-and-branch instruction for UXTH shift-extend");
1154 ExtOpc = AArch64::ANDWri;
1163 if (ExtOpc != AArch64::ANDWri)
1165 MBBI.addImm(ExtBits);
1193 bool TryFold =
false;
1195 RC = &AArch64::GPR64RegClass;
1196 Opc = AArch64::CSELXr;
1199 RC = &AArch64::GPR32RegClass;
1200 Opc = AArch64::CSELWr;
1203 RC = &AArch64::FPR64RegClass;
1204 Opc = AArch64::FCSELDrrr;
1206 RC = &AArch64::FPR32RegClass;
1207 Opc = AArch64::FCSELSrrr;
1209 assert(RC &&
"Unsupported regclass");
1213 unsigned NewReg = 0;
1236 (FalseReg.
isVirtual() || FalseReg == AArch64::WZR ||
1237 FalseReg == AArch64::XZR) &&
1238 "FalseReg was folded into a non-virtual register other than WZR or XZR");
1255 assert(BitSize == 64 &&
"Only bit sizes of 32 or 64 allowed");
1260 return Is.
size() <= 2;
1265 assert(
MI.isCopy() &&
"Expected COPY instruction");
1271 if (
Reg.isVirtual())
1273 if (
Reg.isPhysical())
1274 return RI.getMinimalPhysRegClass(
Reg);
1279 if (DstRC && SrcRC && !RI.getCommonSubClass(DstRC, SrcRC))
1282 return MI.isAsCheapAsAMove();
1288 if (Subtarget.hasExynosCheapAsMoveHandling()) {
1289 if (isExynosCheapAsMove(
MI))
1291 return MI.isAsCheapAsAMove();
1294 switch (
MI.getOpcode()) {
1296 return MI.isAsCheapAsAMove();
1298 case TargetOpcode::COPY:
1301 case AArch64::ADDWrs:
1302 case AArch64::ADDXrs:
1303 case AArch64::SUBWrs:
1304 case AArch64::SUBXrs:
1305 return Subtarget.hasALULSLFast() &&
MI.getOperand(3).getImm() <= 4;
1310 case AArch64::MOVi32imm:
1312 case AArch64::MOVi64imm:
1317bool AArch64InstrInfo::isFalkorShiftExtFast(
const MachineInstr &
MI) {
1318 switch (
MI.getOpcode()) {
1322 case AArch64::ADDWrs:
1323 case AArch64::ADDXrs:
1324 case AArch64::ADDSWrs:
1325 case AArch64::ADDSXrs: {
1326 unsigned Imm =
MI.getOperand(3).getImm();
1333 case AArch64::ADDWrx:
1334 case AArch64::ADDXrx:
1335 case AArch64::ADDXrx64:
1336 case AArch64::ADDSWrx:
1337 case AArch64::ADDSXrx:
1338 case AArch64::ADDSXrx64: {
1339 unsigned Imm =
MI.getOperand(3).getImm();
1351 case AArch64::SUBWrs:
1352 case AArch64::SUBSWrs: {
1353 unsigned Imm =
MI.getOperand(3).getImm();
1355 return ShiftVal == 0 ||
1359 case AArch64::SUBXrs:
1360 case AArch64::SUBSXrs: {
1361 unsigned Imm =
MI.getOperand(3).getImm();
1363 return ShiftVal == 0 ||
1367 case AArch64::SUBWrx:
1368 case AArch64::SUBXrx:
1369 case AArch64::SUBXrx64:
1370 case AArch64::SUBSWrx:
1371 case AArch64::SUBSXrx:
1372 case AArch64::SUBSXrx64: {
1373 unsigned Imm =
MI.getOperand(3).getImm();
1385 case AArch64::LDRBBroW:
1386 case AArch64::LDRBBroX:
1387 case AArch64::LDRBroW:
1388 case AArch64::LDRBroX:
1389 case AArch64::LDRDroW:
1390 case AArch64::LDRDroX:
1391 case AArch64::LDRHHroW:
1392 case AArch64::LDRHHroX:
1393 case AArch64::LDRHroW:
1394 case AArch64::LDRHroX:
1395 case AArch64::LDRQroW:
1396 case AArch64::LDRQroX:
1397 case AArch64::LDRSBWroW:
1398 case AArch64::LDRSBWroX:
1399 case AArch64::LDRSBXroW:
1400 case AArch64::LDRSBXroX:
1401 case AArch64::LDRSHWroW:
1402 case AArch64::LDRSHWroX:
1403 case AArch64::LDRSHXroW:
1404 case AArch64::LDRSHXroX:
1405 case AArch64::LDRSWroW:
1406 case AArch64::LDRSWroX:
1407 case AArch64::LDRSroW:
1408 case AArch64::LDRSroX:
1409 case AArch64::LDRWroW:
1410 case AArch64::LDRWroX:
1411 case AArch64::LDRXroW:
1412 case AArch64::LDRXroX:
1413 case AArch64::PRFMroW:
1414 case AArch64::PRFMroX:
1415 case AArch64::STRBBroW:
1416 case AArch64::STRBBroX:
1417 case AArch64::STRBroW:
1418 case AArch64::STRBroX:
1419 case AArch64::STRDroW:
1420 case AArch64::STRDroX:
1421 case AArch64::STRHHroW:
1422 case AArch64::STRHHroX:
1423 case AArch64::STRHroW:
1424 case AArch64::STRHroX:
1425 case AArch64::STRQroW:
1426 case AArch64::STRQroX:
1427 case AArch64::STRSroW:
1428 case AArch64::STRSroX:
1429 case AArch64::STRWroW:
1430 case AArch64::STRWroX:
1431 case AArch64::STRXroW:
1432 case AArch64::STRXroX: {
1433 unsigned IsSigned =
MI.getOperand(3).getImm();
1440 unsigned Opc =
MI.getOpcode();
1444 case AArch64::SEH_StackAlloc:
1445 case AArch64::SEH_SaveFPLR:
1446 case AArch64::SEH_SaveFPLR_X:
1447 case AArch64::SEH_SaveReg:
1448 case AArch64::SEH_SaveReg_X:
1449 case AArch64::SEH_SaveRegP:
1450 case AArch64::SEH_SaveRegP_X:
1451 case AArch64::SEH_SaveFReg:
1452 case AArch64::SEH_SaveFReg_X:
1453 case AArch64::SEH_SaveFRegP:
1454 case AArch64::SEH_SaveFRegP_X:
1455 case AArch64::SEH_SetFP:
1456 case AArch64::SEH_AddFP:
1457 case AArch64::SEH_Nop:
1458 case AArch64::SEH_PrologEnd:
1459 case AArch64::SEH_EpilogStart:
1460 case AArch64::SEH_EpilogEnd:
1461 case AArch64::SEH_PACSignLR:
1462 case AArch64::SEH_SaveAnyRegI:
1463 case AArch64::SEH_SaveAnyRegIP:
1464 case AArch64::SEH_SaveAnyRegQP:
1465 case AArch64::SEH_SaveAnyRegQPX:
1466 case AArch64::SEH_AllocZ:
1467 case AArch64::SEH_SaveZReg:
1468 case AArch64::SEH_SavePReg:
1475 unsigned &SubIdx)
const {
1476 switch (
MI.getOpcode()) {
1479 case AArch64::SBFMXri:
1480 case AArch64::UBFMXri:
1483 if (
MI.getOperand(2).getImm() != 0 ||
MI.getOperand(3).getImm() != 31)
1486 SrcReg =
MI.getOperand(1).getReg();
1487 DstReg =
MI.getOperand(0).getReg();
1488 SubIdx = AArch64::sub_32;
1497 int64_t OffsetA = 0, OffsetB = 0;
1498 TypeSize WidthA(0,
false), WidthB(0,
false);
1499 bool OffsetAIsScalable =
false, OffsetBIsScalable =
false;
1520 OffsetAIsScalable == OffsetBIsScalable) {
1521 int LowOffset = OffsetA < OffsetB ? OffsetA : OffsetB;
1522 int HighOffset = OffsetA < OffsetB ? OffsetB : OffsetA;
1523 TypeSize LowWidth = (LowOffset == OffsetA) ? WidthA : WidthB;
1524 if (LowWidth.
isScalable() == OffsetAIsScalable &&
1542 switch (
MI.getOpcode()) {
1545 if (
MI.getOperand(0).getImm() == 0x14)
1552 case AArch64::MSRpstatesvcrImm1:
1559 auto Next = std::next(
MI.getIterator());
1560 return Next !=
MBB->end() &&
Next->isCFIInstruction();
1567 Register &SrcReg2, int64_t &CmpMask,
1568 int64_t &CmpValue)
const {
1572 assert(
MI.getNumOperands() >= 2 &&
"All AArch64 cmps should have 2 operands");
1573 if (!
MI.getOperand(1).isReg() ||
MI.getOperand(1).getSubReg())
1576 switch (
MI.getOpcode()) {
1579 case AArch64::PTEST_PP:
1580 case AArch64::PTEST_PP_ANY:
1581 case AArch64::PTEST_PP_FIRST:
1582 SrcReg =
MI.getOperand(0).getReg();
1583 SrcReg2 =
MI.getOperand(1).getReg();
1584 if (
MI.getOperand(2).getSubReg())
1591 case AArch64::SUBSWrr:
1592 case AArch64::SUBSWrs:
1593 case AArch64::SUBSWrx:
1594 case AArch64::SUBSXrr:
1595 case AArch64::SUBSXrs:
1596 case AArch64::SUBSXrx:
1597 case AArch64::ADDSWrr:
1598 case AArch64::ADDSWrs:
1599 case AArch64::ADDSWrx:
1600 case AArch64::ADDSXrr:
1601 case AArch64::ADDSXrs:
1602 case AArch64::ADDSXrx:
1604 SrcReg =
MI.getOperand(1).getReg();
1605 SrcReg2 =
MI.getOperand(2).getReg();
1608 if (
MI.getOperand(2).getSubReg())
1614 case AArch64::SUBSWri:
1615 case AArch64::ADDSWri:
1616 case AArch64::SUBSXri:
1617 case AArch64::ADDSXri:
1618 SrcReg =
MI.getOperand(1).getReg();
1621 CmpValue =
MI.getOperand(2).getImm();
1623 case AArch64::ANDSWri:
1624 case AArch64::ANDSXri:
1627 SrcReg =
MI.getOperand(1).getReg();
1631 MI.getOperand(2).getImm(),
1632 MI.getOpcode() == AArch64::ANDSWri ? 32 : 64);
1641 assert(
MBB &&
"Can't get MachineBasicBlock here");
1643 assert(MF &&
"Can't get MachineFunction here");
1648 for (
unsigned OpIdx = 0, EndIdx = Instr.getNumOperands();
OpIdx < EndIdx;
1655 if (!OpRegCstraints)
1663 "Operand has register constraints without being a register!");
1666 if (
Reg.isPhysical()) {
1683 bool MIDefinesZeroReg =
false;
1684 if (
MI.definesRegister(AArch64::WZR,
nullptr) ||
1685 MI.definesRegister(AArch64::XZR,
nullptr))
1686 MIDefinesZeroReg =
true;
1688 switch (
MI.getOpcode()) {
1690 return MI.getOpcode();
1691 case AArch64::ADDSWrr:
1692 return AArch64::ADDWrr;
1693 case AArch64::ADDSWri:
1694 return MIDefinesZeroReg ? AArch64::ADDSWri : AArch64::ADDWri;
1695 case AArch64::ADDSWrs:
1696 return MIDefinesZeroReg ? AArch64::ADDSWrs : AArch64::ADDWrs;
1697 case AArch64::ADDSWrx:
1698 return AArch64::ADDWrx;
1699 case AArch64::ADDSXrr:
1700 return AArch64::ADDXrr;
1701 case AArch64::ADDSXri:
1702 return MIDefinesZeroReg ? AArch64::ADDSXri : AArch64::ADDXri;
1703 case AArch64::ADDSXrs:
1704 return MIDefinesZeroReg ? AArch64::ADDSXrs : AArch64::ADDXrs;
1705 case AArch64::ADDSXrx:
1706 return AArch64::ADDXrx;
1707 case AArch64::SUBSWrr:
1708 return AArch64::SUBWrr;
1709 case AArch64::SUBSWri:
1710 return MIDefinesZeroReg ? AArch64::SUBSWri : AArch64::SUBWri;
1711 case AArch64::SUBSWrs:
1712 return MIDefinesZeroReg ? AArch64::SUBSWrs : AArch64::SUBWrs;
1713 case AArch64::SUBSWrx:
1714 return AArch64::SUBWrx;
1715 case AArch64::SUBSXrr:
1716 return AArch64::SUBXrr;
1717 case AArch64::SUBSXri:
1718 return MIDefinesZeroReg ? AArch64::SUBSXri : AArch64::SUBXri;
1719 case AArch64::SUBSXrs:
1720 return MIDefinesZeroReg ? AArch64::SUBSXrs : AArch64::SUBXrs;
1721 case AArch64::SUBSXrx:
1722 return AArch64::SUBXrx;
1737 if (To == To->getParent()->begin())
1742 if (To->getParent() != From->getParent())
1754 Instr.modifiesRegister(AArch64::NZCV,
TRI)) ||
1755 ((AccessToCheck &
AK_Read) && Instr.readsRegister(AArch64::NZCV,
TRI)))
1761std::optional<unsigned>
1765 unsigned MaskOpcode =
Mask->getOpcode();
1766 unsigned PredOpcode = Pred->
getOpcode();
1767 bool PredIsPTestLike = isPTestLikeOpcode(PredOpcode);
1768 bool PredIsWhileLike = isWhileOpcode(PredOpcode);
1770 if (PredIsWhileLike) {
1774 if ((Mask == Pred) && PTest->
getOpcode() == AArch64::PTEST_PP_ANY)
1781 getElementSizeForOpcode(MaskOpcode) ==
1782 getElementSizeForOpcode(PredOpcode))
1788 if (PTest->
getOpcode() == AArch64::PTEST_PP_FIRST &&
1795 if (PredIsPTestLike) {
1800 if ((Mask == Pred) && PTest->
getOpcode() == AArch64::PTEST_PP_ANY)
1808 if (Mask != PTestLikeMask && PTestLikeMask->isFullCopy() &&
1809 PTestLikeMask->getOperand(1).getReg().isVirtual())
1817 getElementSizeForOpcode(MaskOpcode) ==
1818 getElementSizeForOpcode(PredOpcode)) {
1819 if (Mask == PTestLikeMask || PTest->
getOpcode() == AArch64::PTEST_PP_ANY)
1845 uint64_t PredElementSize = getElementSizeForOpcode(PredOpcode);
1847 PTest->
getOpcode() == AArch64::PTEST_PP_ANY))
1855 switch (PredOpcode) {
1856 case AArch64::AND_PPzPP:
1857 case AArch64::BIC_PPzPP:
1858 case AArch64::EOR_PPzPP:
1859 case AArch64::NAND_PPzPP:
1860 case AArch64::NOR_PPzPP:
1861 case AArch64::ORN_PPzPP:
1862 case AArch64::ORR_PPzPP:
1863 case AArch64::BRKA_PPzP:
1864 case AArch64::BRKPA_PPzPP:
1865 case AArch64::BRKB_PPzP:
1866 case AArch64::BRKPB_PPzPP:
1867 case AArch64::RDFFR_PPz: {
1871 if (Mask != PredMask)
1875 case AArch64::BRKN_PPzP: {
1879 if ((MaskOpcode != AArch64::PTRUE_B) ||
1880 (
Mask->getOperand(1).getImm() != 31))
1884 case AArch64::PTRUE_B:
1897bool AArch64InstrInfo::optimizePTestInstr(
1898 MachineInstr *PTest,
unsigned MaskReg,
unsigned PredReg,
1903 if (Pred->
isCopy() && PTest->
getOpcode() == AArch64::PTEST_PP_FIRST) {
1907 if (
Op.isReg() &&
Op.getReg().isVirtual() &&
1908 Op.getSubReg() == AArch64::psub0)
1912 unsigned PredOpcode = Pred->
getOpcode();
1913 auto NewOp = canRemovePTestInstr(PTest, Mask, Pred, MRI);
1929 if (*NewOp != PredOpcode) {
1940 for (; i !=
e; ++i) {
1971 if (DeadNZCVIdx != -1) {
1990 if (CmpInstr.
getOpcode() == AArch64::PTEST_PP ||
1991 CmpInstr.
getOpcode() == AArch64::PTEST_PP_ANY ||
1992 CmpInstr.
getOpcode() == AArch64::PTEST_PP_FIRST)
1993 return optimizePTestInstr(&CmpInstr, SrcReg, SrcReg2, MRI);
2002 if (CmpValue == 0 && substituteCmpToZero(CmpInstr, SrcReg, *MRI))
2004 return (CmpValue == 0 || CmpValue == 1) &&
2005 removeCmpToZeroOrOne(CmpInstr, SrcReg, CmpValue, *MRI);
2013 switch (Instr.getOpcode()) {
2015 return AArch64::INSTRUCTION_LIST_END;
2017 case AArch64::ADDSWrr:
2018 case AArch64::ADDSWri:
2019 case AArch64::ADDSXrr:
2020 case AArch64::ADDSXri:
2021 case AArch64::ADDSWrx:
2022 case AArch64::ADDSXrx:
2023 case AArch64::SUBSWrr:
2024 case AArch64::SUBSWri:
2025 case AArch64::SUBSWrx:
2026 case AArch64::SUBSXrr:
2027 case AArch64::SUBSXri:
2028 case AArch64::SUBSXrx:
2029 case AArch64::ANDSWri:
2030 case AArch64::ANDSWrr:
2031 case AArch64::ANDSWrs:
2032 case AArch64::ANDSXri:
2033 case AArch64::ANDSXrr:
2034 case AArch64::ANDSXrs:
2035 case AArch64::BICSWrr:
2036 case AArch64::BICSXrr:
2037 case AArch64::BICSWrs:
2038 case AArch64::BICSXrs:
2039 return Instr.getOpcode();
2041 case AArch64::ADDWrr:
2042 return AArch64::ADDSWrr;
2043 case AArch64::ADDWri:
2044 return AArch64::ADDSWri;
2045 case AArch64::ADDXrr:
2046 return AArch64::ADDSXrr;
2047 case AArch64::ADDXri:
2048 return AArch64::ADDSXri;
2049 case AArch64::ADDWrx:
2050 return AArch64::ADDSWrx;
2051 case AArch64::ADDXrx:
2052 return AArch64::ADDSXrx;
2053 case AArch64::ADCWr:
2054 return AArch64::ADCSWr;
2055 case AArch64::ADCXr:
2056 return AArch64::ADCSXr;
2057 case AArch64::SUBWrr:
2058 return AArch64::SUBSWrr;
2059 case AArch64::SUBWri:
2060 return AArch64::SUBSWri;
2061 case AArch64::SUBXrr:
2062 return AArch64::SUBSXrr;
2063 case AArch64::SUBXri:
2064 return AArch64::SUBSXri;
2065 case AArch64::SUBWrx:
2066 return AArch64::SUBSWrx;
2067 case AArch64::SUBXrx:
2068 return AArch64::SUBSXrx;
2069 case AArch64::SBCWr:
2070 return AArch64::SBCSWr;
2071 case AArch64::SBCXr:
2072 return AArch64::SBCSXr;
2073 case AArch64::ANDWri:
2074 return AArch64::ANDSWri;
2075 case AArch64::ANDXri:
2076 return AArch64::ANDSXri;
2077 case AArch64::ANDWrr:
2078 return AArch64::ANDSWrr;
2079 case AArch64::ANDWrs:
2080 return AArch64::ANDSWrs;
2081 case AArch64::ANDXrr:
2082 return AArch64::ANDSXrr;
2083 case AArch64::ANDXrs:
2084 return AArch64::ANDSXrs;
2085 case AArch64::BICWrr:
2086 return AArch64::BICSWrr;
2087 case AArch64::BICXrr:
2088 return AArch64::BICSXrr;
2089 case AArch64::BICWrs:
2090 return AArch64::BICSWrs;
2091 case AArch64::BICXrs:
2092 return AArch64::BICSXrs;
2098 for (
auto *BB :
MBB->successors())
2099 if (BB->isLiveIn(AArch64::NZCV))
2106int AArch64InstrInfo::findCondCodeUseOperandIdxForBranchOrSelect(
2108 switch (
Instr.getOpcode()) {
2112 case AArch64::Bcc: {
2113 int Idx =
Instr.findRegisterUseOperandIdx(AArch64::NZCV,
nullptr);
2118 case AArch64::CSINVWr:
2119 case AArch64::CSINVXr:
2120 case AArch64::CSINCWr:
2121 case AArch64::CSINCXr:
2122 case AArch64::CSELWr:
2123 case AArch64::CSELXr:
2124 case AArch64::CSNEGWr:
2125 case AArch64::CSNEGXr:
2126 case AArch64::FCSELSrrr:
2127 case AArch64::FCSELDrrr: {
2128 int Idx =
Instr.findRegisterUseOperandIdx(AArch64::NZCV,
nullptr);
2140 AArch64InstrInfo::findCondCodeUseOperandIdxForBranchOrSelect(Instr);
2142 Instr.getOperand(CCIdx).
getImm())
2195std::optional<UsedNZCV>
2200 if (
MI.getParent() != CmpParent)
2201 return std::nullopt;
2204 return std::nullopt;
2209 if (Instr.readsRegister(AArch64::NZCV, &
TRI)) {
2212 return std::nullopt;
2217 if (Instr.modifiesRegister(AArch64::NZCV, &
TRI))
2220 return NZCVUsedAfterCmp;
2224 return Opcode == AArch64::ADDSWri || Opcode == AArch64::ADDSXri;
2228 return Opcode == AArch64::SUBSWri || Opcode == AArch64::SUBSXri;
2234 case AArch64::ANDSWri:
2235 case AArch64::ANDSWrr:
2236 case AArch64::ANDSWrs:
2237 case AArch64::ANDSXri:
2238 case AArch64::ANDSXrr:
2239 case AArch64::ANDSXrs:
2240 case AArch64::BICSWrr:
2241 case AArch64::BICSXrr:
2242 case AArch64::BICSWrs:
2243 case AArch64::BICSXrs:
2269 const unsigned CmpOpcode = CmpInstr.
getOpcode();
2275 "Caller guarantees that CmpInstr compares with constant 0");
2278 if (!NZVCUsed || NZVCUsed->C)
2301bool AArch64InstrInfo::substituteCmpToZero(
2312 if (NewOpc == AArch64::INSTRUCTION_LIST_END)
2319 MI->setDesc(
get(NewOpc));
2324 MI->addRegisterDefined(AArch64::NZCV, &
TRI);
2336 assert((CmpValue == 0 || CmpValue == 1) &&
2337 "Only comparisons to 0 or 1 considered for removal!");
2340 unsigned MIOpc =
MI.getOpcode();
2341 if (MIOpc == AArch64::CSINCWr) {
2342 if (
MI.getOperand(1).getReg() != AArch64::WZR ||
2343 MI.getOperand(2).getReg() != AArch64::WZR)
2345 }
else if (MIOpc == AArch64::CSINCXr) {
2346 if (
MI.getOperand(1).getReg() != AArch64::XZR ||
2347 MI.getOperand(2).getReg() != AArch64::XZR)
2357 if (
MI.findRegisterDefOperandIdx(AArch64::NZCV,
nullptr,
true) != -1)
2361 const unsigned CmpOpcode = CmpInstr.
getOpcode();
2363 if (CmpValue && !IsSubsRegImm)
2365 if (!CmpValue && !IsSubsRegImm && !
isADDSRegImm(CmpOpcode))
2370 if (MIUsedNZCV.
C || MIUsedNZCV.
V)
2373 std::optional<UsedNZCV> NZCVUsedAfterCmp =
2377 if (!NZCVUsedAfterCmp || NZCVUsedAfterCmp->C || NZCVUsedAfterCmp->V)
2380 if ((MIUsedNZCV.
Z && NZCVUsedAfterCmp->N) ||
2381 (MIUsedNZCV.
N && NZCVUsedAfterCmp->Z))
2384 if (MIUsedNZCV.
N && !CmpValue)
2426bool AArch64InstrInfo::removeCmpToZeroOrOne(
2433 SmallVector<MachineInstr *, 4> CCUseInstrs;
2434 bool IsInvertCC =
false;
2442 for (MachineInstr *CCUseInstr : CCUseInstrs) {
2443 int Idx = findCondCodeUseOperandIdxForBranchOrSelect(*CCUseInstr);
2444 assert(Idx >= 0 &&
"Unexpected instruction using CC.");
2445 MachineOperand &CCOperand = CCUseInstr->getOperand(Idx);
2454bool AArch64InstrInfo::expandPostRAPseudo(
MachineInstr &
MI)
const {
2455 if (
MI.getOpcode() != TargetOpcode::LOAD_STACK_GUARD &&
2456 MI.getOpcode() != AArch64::CATCHRET)
2459 MachineBasicBlock &
MBB = *
MI.getParent();
2461 auto TRI = Subtarget.getRegisterInfo();
2464 if (
MI.getOpcode() == AArch64::CATCHRET) {
2466 const TargetInstrInfo *
TII =
2468 MachineBasicBlock *TargetMBB =
MI.getOperand(0).getMBB();
2473 FirstEpilogSEH = std::prev(FirstEpilogSEH);
2475 FirstEpilogSEH = std::next(FirstEpilogSEH);
2490 if (
M.getStackProtectorGuard() ==
"sysreg") {
2491 const AArch64SysReg::SysReg *SrcReg =
2492 AArch64SysReg::lookupSysRegByName(
M.getStackProtectorGuardReg());
2500 int Offset =
M.getStackProtectorGuardOffset();
2551 const GlobalValue *GV =
2554 unsigned OpFlags = Subtarget.ClassifyGlobalReference(GV, TM);
2560 if (Subtarget.isTargetILP32()) {
2561 unsigned Reg32 =
TRI->getSubReg(
Reg, AArch64::sub_32);
2575 assert(!Subtarget.isTargetILP32() &&
"how can large exist in ILP32?");
2602 if (Subtarget.isTargetILP32()) {
2603 unsigned Reg32 =
TRI->getSubReg(
Reg, AArch64::sub_32);
2626 switch (
MI.getOpcode()) {
2629 case AArch64::MOVZWi:
2630 case AArch64::MOVZXi:
2631 if (
MI.getOperand(1).isImm() &&
MI.getOperand(1).getImm() == 0) {
2632 assert(
MI.getDesc().getNumOperands() == 3 &&
2633 MI.getOperand(2).getImm() == 0 &&
"invalid MOVZi operands");
2637 case AArch64::ANDWri:
2638 return MI.getOperand(1).getReg() == AArch64::WZR;
2639 case AArch64::ANDXri:
2640 return MI.getOperand(1).getReg() == AArch64::XZR;
2641 case TargetOpcode::COPY:
2642 return MI.getOperand(1).getReg() == AArch64::WZR;
2650 switch (
MI.getOpcode()) {
2653 case TargetOpcode::COPY: {
2656 return (AArch64::GPR32RegClass.
contains(DstReg) ||
2657 AArch64::GPR64RegClass.
contains(DstReg));
2659 case AArch64::ORRXrs:
2660 if (
MI.getOperand(1).getReg() == AArch64::XZR) {
2661 assert(
MI.getDesc().getNumOperands() == 4 &&
2662 MI.getOperand(3).getImm() == 0 &&
"invalid ORRrs operands");
2666 case AArch64::ADDXri:
2667 if (
MI.getOperand(2).getImm() == 0) {
2668 assert(
MI.getDesc().getNumOperands() == 4 &&
2669 MI.getOperand(3).getImm() == 0 &&
"invalid ADDXri operands");
2680 switch (
MI.getOpcode()) {
2683 case TargetOpcode::COPY: {
2685 return AArch64::FPR128RegClass.contains(DstReg);
2687 case AArch64::ORRv16i8:
2688 if (
MI.getOperand(1).getReg() ==
MI.getOperand(2).getReg()) {
2689 assert(
MI.getDesc().getNumOperands() == 3 &&
MI.getOperand(0).isReg() &&
2690 "invalid ORRv16i8 operands");
2702 case AArch64::LDRWui:
2703 case AArch64::LDRXui:
2704 case AArch64::LDRBui:
2705 case AArch64::LDRHui:
2706 case AArch64::LDRSui:
2707 case AArch64::LDRDui:
2708 case AArch64::LDRQui:
2709 case AArch64::LDR_PXI:
2715 int &FrameIndex)
const {
2719 if (
MI.getOperand(0).getSubReg() == 0 &&
MI.getOperand(1).isFI() &&
2720 MI.getOperand(2).isImm() &&
MI.getOperand(2).getImm() == 0) {
2721 FrameIndex =
MI.getOperand(1).getIndex();
2722 return MI.getOperand(0).getReg();
2731 case AArch64::STRWui:
2732 case AArch64::STRXui:
2733 case AArch64::STRBui:
2734 case AArch64::STRHui:
2735 case AArch64::STRSui:
2736 case AArch64::STRDui:
2737 case AArch64::STRQui:
2738 case AArch64::STR_PXI:
2744 int &FrameIndex)
const {
2748 if (
MI.getOperand(0).getSubReg() == 0 &&
MI.getOperand(1).isFI() &&
2749 MI.getOperand(2).isImm() &&
MI.getOperand(2).getImm() == 0) {
2750 FrameIndex =
MI.getOperand(1).getIndex();
2751 return MI.getOperand(0).getReg();
2757 int &FrameIndex)
const {
2772 return MI.getOperand(0).getReg();
2778 int &FrameIndex)
const {
2793 return MI.getOperand(0).getReg();
2801 return MMO->getFlags() & MOSuppressPair;
2807 if (
MI.memoperands_empty())
2815 return MMO->getFlags() & MOStridedAccess;
2823 case AArch64::STURSi:
2824 case AArch64::STRSpre:
2825 case AArch64::STURDi:
2826 case AArch64::STRDpre:
2827 case AArch64::STURQi:
2828 case AArch64::STRQpre:
2829 case AArch64::STURBBi:
2830 case AArch64::STURHHi:
2831 case AArch64::STURWi:
2832 case AArch64::STRWpre:
2833 case AArch64::STURXi:
2834 case AArch64::STRXpre:
2835 case AArch64::LDURSi:
2836 case AArch64::LDRSpre:
2837 case AArch64::LDURDi:
2838 case AArch64::LDRDpre:
2839 case AArch64::LDURQi:
2840 case AArch64::LDRQpre:
2841 case AArch64::LDURWi:
2842 case AArch64::LDRWpre:
2843 case AArch64::LDURXi:
2844 case AArch64::LDRXpre:
2845 case AArch64::LDRSWpre:
2846 case AArch64::LDURSWi:
2847 case AArch64::LDURHHi:
2848 case AArch64::LDURBBi:
2849 case AArch64::LDURSBWi:
2850 case AArch64::LDURSHWi:
2858 case AArch64::PRFMui:
return AArch64::PRFUMi;
2859 case AArch64::LDRXui:
return AArch64::LDURXi;
2860 case AArch64::LDRWui:
return AArch64::LDURWi;
2861 case AArch64::LDRBui:
return AArch64::LDURBi;
2862 case AArch64::LDRHui:
return AArch64::LDURHi;
2863 case AArch64::LDRSui:
return AArch64::LDURSi;
2864 case AArch64::LDRDui:
return AArch64::LDURDi;
2865 case AArch64::LDRQui:
return AArch64::LDURQi;
2866 case AArch64::LDRBBui:
return AArch64::LDURBBi;
2867 case AArch64::LDRHHui:
return AArch64::LDURHHi;
2868 case AArch64::LDRSBXui:
return AArch64::LDURSBXi;
2869 case AArch64::LDRSBWui:
return AArch64::LDURSBWi;
2870 case AArch64::LDRSHXui:
return AArch64::LDURSHXi;
2871 case AArch64::LDRSHWui:
return AArch64::LDURSHWi;
2872 case AArch64::LDRSWui:
return AArch64::LDURSWi;
2873 case AArch64::STRXui:
return AArch64::STURXi;
2874 case AArch64::STRWui:
return AArch64::STURWi;
2875 case AArch64::STRBui:
return AArch64::STURBi;
2876 case AArch64::STRHui:
return AArch64::STURHi;
2877 case AArch64::STRSui:
return AArch64::STURSi;
2878 case AArch64::STRDui:
return AArch64::STURDi;
2879 case AArch64::STRQui:
return AArch64::STURQi;
2880 case AArch64::STRBBui:
return AArch64::STURBBi;
2881 case AArch64::STRHHui:
return AArch64::STURHHi;
2890 case AArch64::LDAPURBi:
2891 case AArch64::LDAPURHi:
2892 case AArch64::LDAPURi:
2893 case AArch64::LDAPURSBWi:
2894 case AArch64::LDAPURSBXi:
2895 case AArch64::LDAPURSHWi:
2896 case AArch64::LDAPURSHXi:
2897 case AArch64::LDAPURSWi:
2898 case AArch64::LDAPURXi:
2899 case AArch64::LDR_PPXI:
2900 case AArch64::LDR_PXI:
2901 case AArch64::LDR_ZXI:
2902 case AArch64::LDR_ZZXI:
2903 case AArch64::LDR_ZZXI_STRIDED_CONTIGUOUS:
2904 case AArch64::LDR_ZZZXI:
2905 case AArch64::LDR_ZZZZXI:
2906 case AArch64::LDR_ZZZZXI_STRIDED_CONTIGUOUS:
2907 case AArch64::LDRBBui:
2908 case AArch64::LDRBui:
2909 case AArch64::LDRDui:
2910 case AArch64::LDRHHui:
2911 case AArch64::LDRHui:
2912 case AArch64::LDRQui:
2913 case AArch64::LDRSBWui:
2914 case AArch64::LDRSBXui:
2915 case AArch64::LDRSHWui:
2916 case AArch64::LDRSHXui:
2917 case AArch64::LDRSui:
2918 case AArch64::LDRSWui:
2919 case AArch64::LDRWui:
2920 case AArch64::LDRXui:
2921 case AArch64::LDURBBi:
2922 case AArch64::LDURBi:
2923 case AArch64::LDURDi:
2924 case AArch64::LDURHHi:
2925 case AArch64::LDURHi:
2926 case AArch64::LDURQi:
2927 case AArch64::LDURSBWi:
2928 case AArch64::LDURSBXi:
2929 case AArch64::LDURSHWi:
2930 case AArch64::LDURSHXi:
2931 case AArch64::LDURSi:
2932 case AArch64::LDURSWi:
2933 case AArch64::LDURWi:
2934 case AArch64::LDURXi:
2935 case AArch64::PRFMui:
2936 case AArch64::PRFUMi:
2937 case AArch64::ST2Gi:
2939 case AArch64::STLURBi:
2940 case AArch64::STLURHi:
2941 case AArch64::STLURWi:
2942 case AArch64::STLURXi:
2943 case AArch64::StoreSwiftAsyncContext:
2944 case AArch64::STR_PPXI:
2945 case AArch64::STR_PXI:
2946 case AArch64::STR_ZXI:
2947 case AArch64::STR_ZZXI:
2948 case AArch64::STR_ZZXI_STRIDED_CONTIGUOUS:
2949 case AArch64::STR_ZZZXI:
2950 case AArch64::STR_ZZZZXI:
2951 case AArch64::STR_ZZZZXI_STRIDED_CONTIGUOUS:
2952 case AArch64::STRBBui:
2953 case AArch64::STRBui:
2954 case AArch64::STRDui:
2955 case AArch64::STRHHui:
2956 case AArch64::STRHui:
2957 case AArch64::STRQui:
2958 case AArch64::STRSui:
2959 case AArch64::STRWui:
2960 case AArch64::STRXui:
2961 case AArch64::STURBBi:
2962 case AArch64::STURBi:
2963 case AArch64::STURDi:
2964 case AArch64::STURHHi:
2965 case AArch64::STURHi:
2966 case AArch64::STURQi:
2967 case AArch64::STURSi:
2968 case AArch64::STURWi:
2969 case AArch64::STURXi:
2970 case AArch64::STZ2Gi:
2971 case AArch64::STZGi:
2972 case AArch64::TAGPstack:
2974 case AArch64::LD1B_D_IMM:
2975 case AArch64::LD1B_H_IMM:
2976 case AArch64::LD1B_IMM:
2977 case AArch64::LD1B_S_IMM:
2978 case AArch64::LD1D_IMM:
2979 case AArch64::LD1H_D_IMM:
2980 case AArch64::LD1H_IMM:
2981 case AArch64::LD1H_S_IMM:
2982 case AArch64::LD1RB_D_IMM:
2983 case AArch64::LD1RB_H_IMM:
2984 case AArch64::LD1RB_IMM:
2985 case AArch64::LD1RB_S_IMM:
2986 case AArch64::LD1RD_IMM:
2987 case AArch64::LD1RH_D_IMM:
2988 case AArch64::LD1RH_IMM:
2989 case AArch64::LD1RH_S_IMM:
2990 case AArch64::LD1RSB_D_IMM:
2991 case AArch64::LD1RSB_H_IMM:
2992 case AArch64::LD1RSB_S_IMM:
2993 case AArch64::LD1RSH_D_IMM:
2994 case AArch64::LD1RSH_S_IMM:
2995 case AArch64::LD1RSW_IMM:
2996 case AArch64::LD1RW_D_IMM:
2997 case AArch64::LD1RW_IMM:
2998 case AArch64::LD1SB_D_IMM:
2999 case AArch64::LD1SB_H_IMM:
3000 case AArch64::LD1SB_S_IMM:
3001 case AArch64::LD1SH_D_IMM:
3002 case AArch64::LD1SH_S_IMM:
3003 case AArch64::LD1SW_D_IMM:
3004 case AArch64::LD1W_D_IMM:
3005 case AArch64::LD1W_IMM:
3006 case AArch64::LD2B_IMM:
3007 case AArch64::LD2D_IMM:
3008 case AArch64::LD2H_IMM:
3009 case AArch64::LD2W_IMM:
3010 case AArch64::LD3B_IMM:
3011 case AArch64::LD3D_IMM:
3012 case AArch64::LD3H_IMM:
3013 case AArch64::LD3W_IMM:
3014 case AArch64::LD4B_IMM:
3015 case AArch64::LD4D_IMM:
3016 case AArch64::LD4H_IMM:
3017 case AArch64::LD4W_IMM:
3019 case AArch64::LDNF1B_D_IMM:
3020 case AArch64::LDNF1B_H_IMM:
3021 case AArch64::LDNF1B_IMM:
3022 case AArch64::LDNF1B_S_IMM:
3023 case AArch64::LDNF1D_IMM:
3024 case AArch64::LDNF1H_D_IMM:
3025 case AArch64::LDNF1H_IMM:
3026 case AArch64::LDNF1H_S_IMM:
3027 case AArch64::LDNF1SB_D_IMM:
3028 case AArch64::LDNF1SB_H_IMM:
3029 case AArch64::LDNF1SB_S_IMM:
3030 case AArch64::LDNF1SH_D_IMM:
3031 case AArch64::LDNF1SH_S_IMM:
3032 case AArch64::LDNF1SW_D_IMM:
3033 case AArch64::LDNF1W_D_IMM:
3034 case AArch64::LDNF1W_IMM:
3035 case AArch64::LDNPDi:
3036 case AArch64::LDNPQi:
3037 case AArch64::LDNPSi:
3038 case AArch64::LDNPWi:
3039 case AArch64::LDNPXi:
3040 case AArch64::LDNT1B_ZRI:
3041 case AArch64::LDNT1D_ZRI:
3042 case AArch64::LDNT1H_ZRI:
3043 case AArch64::LDNT1W_ZRI:
3044 case AArch64::LDPDi:
3045 case AArch64::LDPQi:
3046 case AArch64::LDPSi:
3047 case AArch64::LDPWi:
3048 case AArch64::LDPXi:
3049 case AArch64::LDRBBpost:
3050 case AArch64::LDRBBpre:
3051 case AArch64::LDRBpost:
3052 case AArch64::LDRBpre:
3053 case AArch64::LDRDpost:
3054 case AArch64::LDRDpre:
3055 case AArch64::LDRHHpost:
3056 case AArch64::LDRHHpre:
3057 case AArch64::LDRHpost:
3058 case AArch64::LDRHpre:
3059 case AArch64::LDRQpost:
3060 case AArch64::LDRQpre:
3061 case AArch64::LDRSpost:
3062 case AArch64::LDRSpre:
3063 case AArch64::LDRWpost:
3064 case AArch64::LDRWpre:
3065 case AArch64::LDRXpost:
3066 case AArch64::LDRXpre:
3067 case AArch64::ST1B_D_IMM:
3068 case AArch64::ST1B_H_IMM:
3069 case AArch64::ST1B_IMM:
3070 case AArch64::ST1B_S_IMM:
3071 case AArch64::ST1D_IMM:
3072 case AArch64::ST1H_D_IMM:
3073 case AArch64::ST1H_IMM:
3074 case AArch64::ST1H_S_IMM:
3075 case AArch64::ST1W_D_IMM:
3076 case AArch64::ST1W_IMM:
3077 case AArch64::ST2B_IMM:
3078 case AArch64::ST2D_IMM:
3079 case AArch64::ST2H_IMM:
3080 case AArch64::ST2W_IMM:
3081 case AArch64::ST3B_IMM:
3082 case AArch64::ST3D_IMM:
3083 case AArch64::ST3H_IMM:
3084 case AArch64::ST3W_IMM:
3085 case AArch64::ST4B_IMM:
3086 case AArch64::ST4D_IMM:
3087 case AArch64::ST4H_IMM:
3088 case AArch64::ST4W_IMM:
3089 case AArch64::STGPi:
3090 case AArch64::STGPreIndex:
3091 case AArch64::STZGPreIndex:
3092 case AArch64::ST2GPreIndex:
3093 case AArch64::STZ2GPreIndex:
3094 case AArch64::STGPostIndex:
3095 case AArch64::STZGPostIndex:
3096 case AArch64::ST2GPostIndex:
3097 case AArch64::STZ2GPostIndex:
3098 case AArch64::STNPDi:
3099 case AArch64::STNPQi:
3100 case AArch64::STNPSi:
3101 case AArch64::STNPWi:
3102 case AArch64::STNPXi:
3103 case AArch64::STNT1B_ZRI:
3104 case AArch64::STNT1D_ZRI:
3105 case AArch64::STNT1H_ZRI:
3106 case AArch64::STNT1W_ZRI:
3107 case AArch64::STPDi:
3108 case AArch64::STPQi:
3109 case AArch64::STPSi:
3110 case AArch64::STPWi:
3111 case AArch64::STPXi:
3112 case AArch64::STRBBpost:
3113 case AArch64::STRBBpre:
3114 case AArch64::STRBpost:
3115 case AArch64::STRBpre:
3116 case AArch64::STRDpost:
3117 case AArch64::STRDpre:
3118 case AArch64::STRHHpost:
3119 case AArch64::STRHHpre:
3120 case AArch64::STRHpost:
3121 case AArch64::STRHpre:
3122 case AArch64::STRQpost:
3123 case AArch64::STRQpre:
3124 case AArch64::STRSpost:
3125 case AArch64::STRSpre:
3126 case AArch64::STRWpost:
3127 case AArch64::STRWpre:
3128 case AArch64::STRXpost:
3129 case AArch64::STRXpre:
3131 case AArch64::LDPDpost:
3132 case AArch64::LDPDpre:
3133 case AArch64::LDPQpost:
3134 case AArch64::LDPQpre:
3135 case AArch64::LDPSpost:
3136 case AArch64::LDPSpre:
3137 case AArch64::LDPWpost:
3138 case AArch64::LDPWpre:
3139 case AArch64::LDPXpost:
3140 case AArch64::LDPXpre:
3141 case AArch64::STGPpre:
3142 case AArch64::STGPpost:
3143 case AArch64::STPDpost:
3144 case AArch64::STPDpre:
3145 case AArch64::STPQpost:
3146 case AArch64::STPQpre:
3147 case AArch64::STPSpost:
3148 case AArch64::STPSpre:
3149 case AArch64::STPWpost:
3150 case AArch64::STPWpre:
3151 case AArch64::STPXpost:
3152 case AArch64::STPXpre:
3158 switch (
MI.getOpcode()) {
3162 case AArch64::STRSui:
3163 case AArch64::STRDui:
3164 case AArch64::STRQui:
3165 case AArch64::STRXui:
3166 case AArch64::STRWui:
3167 case AArch64::LDRSui:
3168 case AArch64::LDRDui:
3169 case AArch64::LDRQui:
3170 case AArch64::LDRXui:
3171 case AArch64::LDRWui:
3172 case AArch64::LDRSWui:
3174 case AArch64::STURSi:
3175 case AArch64::STRSpre:
3176 case AArch64::STURDi:
3177 case AArch64::STRDpre:
3178 case AArch64::STURQi:
3179 case AArch64::STRQpre:
3180 case AArch64::STURWi:
3181 case AArch64::STRWpre:
3182 case AArch64::STURXi:
3183 case AArch64::STRXpre:
3184 case AArch64::LDURSi:
3185 case AArch64::LDRSpre:
3186 case AArch64::LDURDi:
3187 case AArch64::LDRDpre:
3188 case AArch64::LDURQi:
3189 case AArch64::LDRQpre:
3190 case AArch64::LDURWi:
3191 case AArch64::LDRWpre:
3192 case AArch64::LDURXi:
3193 case AArch64::LDRXpre:
3194 case AArch64::LDURSWi:
3195 case AArch64::LDRSWpre:
3197 case AArch64::LDR_ZXI:
3198 case AArch64::STR_ZXI:
3204 switch (
MI.getOpcode()) {
3207 "Unexpected instruction - was a new tail call opcode introduced?");
3209 case AArch64::TCRETURNdi:
3210 case AArch64::TCRETURNri:
3211 case AArch64::TCRETURNrix16x17:
3212 case AArch64::TCRETURNrix17:
3213 case AArch64::TCRETURNrinotx16:
3214 case AArch64::TCRETURNriALL:
3215 case AArch64::AUTH_TCRETURN:
3216 case AArch64::AUTH_TCRETURN_BTI:
3226 case AArch64::ADDWri:
3227 return AArch64::ADDSWri;
3228 case AArch64::ADDWrr:
3229 return AArch64::ADDSWrr;
3230 case AArch64::ADDWrs:
3231 return AArch64::ADDSWrs;
3232 case AArch64::ADDWrx:
3233 return AArch64::ADDSWrx;
3234 case AArch64::ANDWri:
3235 return AArch64::ANDSWri;
3236 case AArch64::ANDWrr:
3237 return AArch64::ANDSWrr;
3238 case AArch64::ANDWrs:
3239 return AArch64::ANDSWrs;
3240 case AArch64::BICWrr:
3241 return AArch64::BICSWrr;
3242 case AArch64::BICWrs:
3243 return AArch64::BICSWrs;
3244 case AArch64::SUBWri:
3245 return AArch64::SUBSWri;
3246 case AArch64::SUBWrr:
3247 return AArch64::SUBSWrr;
3248 case AArch64::SUBWrs:
3249 return AArch64::SUBSWrs;
3250 case AArch64::SUBWrx:
3251 return AArch64::SUBSWrx;
3253 case AArch64::ADDXri:
3254 return AArch64::ADDSXri;
3255 case AArch64::ADDXrr:
3256 return AArch64::ADDSXrr;
3257 case AArch64::ADDXrs:
3258 return AArch64::ADDSXrs;
3259 case AArch64::ADDXrx:
3260 return AArch64::ADDSXrx;
3261 case AArch64::ANDXri:
3262 return AArch64::ANDSXri;
3263 case AArch64::ANDXrr:
3264 return AArch64::ANDSXrr;
3265 case AArch64::ANDXrs:
3266 return AArch64::ANDSXrs;
3267 case AArch64::BICXrr:
3268 return AArch64::BICSXrr;
3269 case AArch64::BICXrs:
3270 return AArch64::BICSXrs;
3271 case AArch64::SUBXri:
3272 return AArch64::SUBSXri;
3273 case AArch64::SUBXrr:
3274 return AArch64::SUBSXrr;
3275 case AArch64::SUBXrs:
3276 return AArch64::SUBSXrs;
3277 case AArch64::SUBXrx:
3278 return AArch64::SUBSXrx;
3280 case AArch64::AND_PPzPP:
3281 return AArch64::ANDS_PPzPP;
3282 case AArch64::BIC_PPzPP:
3283 return AArch64::BICS_PPzPP;
3284 case AArch64::EOR_PPzPP:
3285 return AArch64::EORS_PPzPP;
3286 case AArch64::NAND_PPzPP:
3287 return AArch64::NANDS_PPzPP;
3288 case AArch64::NOR_PPzPP:
3289 return AArch64::NORS_PPzPP;
3290 case AArch64::ORN_PPzPP:
3291 return AArch64::ORNS_PPzPP;
3292 case AArch64::ORR_PPzPP:
3293 return AArch64::ORRS_PPzPP;
3294 case AArch64::BRKA_PPzP:
3295 return AArch64::BRKAS_PPzP;
3296 case AArch64::BRKPA_PPzPP:
3297 return AArch64::BRKPAS_PPzPP;
3298 case AArch64::BRKB_PPzP:
3299 return AArch64::BRKBS_PPzP;
3300 case AArch64::BRKPB_PPzPP:
3301 return AArch64::BRKPBS_PPzPP;
3302 case AArch64::BRKN_PPzP:
3303 return AArch64::BRKNS_PPzP;
3304 case AArch64::RDFFR_PPz:
3305 return AArch64::RDFFRS_PPz;
3306 case AArch64::PTRUE_B:
3307 return AArch64::PTRUES_B;
3318 if (
MI.hasOrderedMemoryRef())
3323 assert((
MI.getOperand(IsPreLdSt ? 2 : 1).isReg() ||
3324 MI.getOperand(IsPreLdSt ? 2 : 1).isFI()) &&
3325 "Expected a reg or frame index operand.");
3329 bool IsImmPreLdSt = IsPreLdSt &&
MI.getOperand(3).isImm();
3331 if (!
MI.getOperand(2).isImm() && !IsImmPreLdSt)
3344 if (
MI.getOperand(1).isReg() && !IsPreLdSt) {
3345 Register BaseReg =
MI.getOperand(1).getReg();
3347 if (
MI.modifiesRegister(BaseReg,
TRI))
3353 switch (
MI.getOpcode()) {
3356 case AArch64::LDR_ZXI:
3357 case AArch64::STR_ZXI:
3358 if (!Subtarget.isLittleEndian() ||
3359 Subtarget.getSVEVectorSizeInBits() != 128)
3372 const MCAsmInfo *MAI =
MI.getMF()->getTarget().getMCAsmInfo();
3374 MI.getMF()->getFunction().needsUnwindTableEntry();
3380 if (Subtarget.isPaired128Slow()) {
3381 switch (
MI.getOpcode()) {
3384 case AArch64::LDURQi:
3385 case AArch64::STURQi:
3386 case AArch64::LDRQui:
3387 case AArch64::STRQui:
3414std::optional<ExtAddrMode>
3419 bool OffsetIsScalable;
3420 if (!getMemOperandWithOffset(MemI,
Base,
Offset, OffsetIsScalable,
TRI))
3421 return std::nullopt;
3424 return std::nullopt;
3439 int64_t OffsetScale = 1;
3444 case AArch64::LDURQi:
3445 case AArch64::STURQi:
3449 case AArch64::LDURDi:
3450 case AArch64::STURDi:
3451 case AArch64::LDURXi:
3452 case AArch64::STURXi:
3456 case AArch64::LDURWi:
3457 case AArch64::LDURSWi:
3458 case AArch64::STURWi:
3462 case AArch64::LDURHi:
3463 case AArch64::STURHi:
3464 case AArch64::LDURHHi:
3465 case AArch64::STURHHi:
3466 case AArch64::LDURSHXi:
3467 case AArch64::LDURSHWi:
3471 case AArch64::LDRBroX:
3472 case AArch64::LDRBBroX:
3473 case AArch64::LDRSBXroX:
3474 case AArch64::LDRSBWroX:
3475 case AArch64::STRBroX:
3476 case AArch64::STRBBroX:
3477 case AArch64::LDURBi:
3478 case AArch64::LDURBBi:
3479 case AArch64::LDURSBXi:
3480 case AArch64::LDURSBWi:
3481 case AArch64::STURBi:
3482 case AArch64::STURBBi:
3483 case AArch64::LDRBui:
3484 case AArch64::LDRBBui:
3485 case AArch64::LDRSBXui:
3486 case AArch64::LDRSBWui:
3487 case AArch64::STRBui:
3488 case AArch64::STRBBui:
3492 case AArch64::LDRQroX:
3493 case AArch64::STRQroX:
3494 case AArch64::LDRQui:
3495 case AArch64::STRQui:
3500 case AArch64::LDRDroX:
3501 case AArch64::STRDroX:
3502 case AArch64::LDRXroX:
3503 case AArch64::STRXroX:
3504 case AArch64::LDRDui:
3505 case AArch64::STRDui:
3506 case AArch64::LDRXui:
3507 case AArch64::STRXui:
3512 case AArch64::LDRWroX:
3513 case AArch64::LDRSWroX:
3514 case AArch64::STRWroX:
3515 case AArch64::LDRWui:
3516 case AArch64::LDRSWui:
3517 case AArch64::STRWui:
3522 case AArch64::LDRHroX:
3523 case AArch64::STRHroX:
3524 case AArch64::LDRHHroX:
3525 case AArch64::STRHHroX:
3526 case AArch64::LDRSHXroX:
3527 case AArch64::LDRSHWroX:
3528 case AArch64::LDRHui:
3529 case AArch64::STRHui:
3530 case AArch64::LDRHHui:
3531 case AArch64::STRHHui:
3532 case AArch64::LDRSHXui:
3533 case AArch64::LDRSHWui:
3541 if (BaseRegOp.
isReg() && BaseRegOp.
getReg() == Reg)
3565 case AArch64::SBFMXri:
3578 AM.
Scale = OffsetScale;
3583 case TargetOpcode::SUBREG_TO_REG: {
3599 if (
DefMI.getOpcode() != AArch64::ORRWrs ||
3600 DefMI.getOperand(1).getReg() != AArch64::WZR ||
3601 DefMI.getOperand(3).getImm() != 0)
3608 AM.
Scale = OffsetScale;
3619 auto validateOffsetForLDP = [](
unsigned NumBytes, int64_t OldOffset,
3620 int64_t NewOffset) ->
bool {
3621 int64_t MinOffset, MaxOffset;
3638 return OldOffset < MinOffset || OldOffset > MaxOffset ||
3639 (NewOffset >= MinOffset && NewOffset <= MaxOffset);
3641 auto canFoldAddSubImmIntoAddrMode = [&](int64_t Disp) ->
bool {
3643 int64_t NewOffset = OldOffset + Disp;
3644 if (!isLegalAddressingMode(NumBytes, NewOffset, 0))
3648 if (!validateOffsetForLDP(NumBytes, OldOffset, NewOffset))
3658 auto canFoldAddRegIntoAddrMode =
3663 if ((
unsigned)Scale != Scale)
3665 if (!isLegalAddressingMode(NumBytes, 0, Scale))
3677 return (Opcode == AArch64::STURQi || Opcode == AArch64::STRQui) &&
3678 Subtarget.isSTRQroSlow();
3687 case AArch64::ADDXri:
3693 return canFoldAddSubImmIntoAddrMode(Disp);
3695 case AArch64::SUBXri:
3701 return canFoldAddSubImmIntoAddrMode(-Disp);
3703 case AArch64::ADDXrs: {
3716 if (Shift != 2 && Shift != 3 && Subtarget.hasAddrLSLSlow14())
3718 if (avoidSlowSTRQ(MemI))
3721 return canFoldAddRegIntoAddrMode(1ULL << Shift);
3724 case AArch64::ADDXrr:
3732 if (!OptSize && avoidSlowSTRQ(MemI))
3734 return canFoldAddRegIntoAddrMode(1);
3736 case AArch64::ADDXrx:
3744 if (!OptSize && avoidSlowSTRQ(MemI))
3753 return canFoldAddRegIntoAddrMode(
3768 case AArch64::LDURQi:
3769 case AArch64::LDRQui:
3770 return AArch64::LDRQroX;
3771 case AArch64::STURQi:
3772 case AArch64::STRQui:
3773 return AArch64::STRQroX;
3774 case AArch64::LDURDi:
3775 case AArch64::LDRDui:
3776 return AArch64::LDRDroX;
3777 case AArch64::STURDi:
3778 case AArch64::STRDui:
3779 return AArch64::STRDroX;
3780 case AArch64::LDURXi:
3781 case AArch64::LDRXui:
3782 return AArch64::LDRXroX;
3783 case AArch64::STURXi:
3784 case AArch64::STRXui:
3785 return AArch64::STRXroX;
3786 case AArch64::LDURWi:
3787 case AArch64::LDRWui:
3788 return AArch64::LDRWroX;
3789 case AArch64::LDURSWi:
3790 case AArch64::LDRSWui:
3791 return AArch64::LDRSWroX;
3792 case AArch64::STURWi:
3793 case AArch64::STRWui:
3794 return AArch64::STRWroX;
3795 case AArch64::LDURHi:
3796 case AArch64::LDRHui:
3797 return AArch64::LDRHroX;
3798 case AArch64::STURHi:
3799 case AArch64::STRHui:
3800 return AArch64::STRHroX;
3801 case AArch64::LDURHHi:
3802 case AArch64::LDRHHui:
3803 return AArch64::LDRHHroX;
3804 case AArch64::STURHHi:
3805 case AArch64::STRHHui:
3806 return AArch64::STRHHroX;
3807 case AArch64::LDURSHXi:
3808 case AArch64::LDRSHXui:
3809 return AArch64::LDRSHXroX;
3810 case AArch64::LDURSHWi:
3811 case AArch64::LDRSHWui:
3812 return AArch64::LDRSHWroX;
3813 case AArch64::LDURBi:
3814 case AArch64::LDRBui:
3815 return AArch64::LDRBroX;
3816 case AArch64::LDURBBi:
3817 case AArch64::LDRBBui:
3818 return AArch64::LDRBBroX;
3819 case AArch64::LDURSBXi:
3820 case AArch64::LDRSBXui:
3821 return AArch64::LDRSBXroX;
3822 case AArch64::LDURSBWi:
3823 case AArch64::LDRSBWui:
3824 return AArch64::LDRSBWroX;
3825 case AArch64::STURBi:
3826 case AArch64::STRBui:
3827 return AArch64::STRBroX;
3828 case AArch64::STURBBi:
3829 case AArch64::STRBBui:
3830 return AArch64::STRBBroX;
3842 case AArch64::LDURQi:
3844 return AArch64::LDRQui;
3845 case AArch64::STURQi:
3847 return AArch64::STRQui;
3848 case AArch64::LDURDi:
3850 return AArch64::LDRDui;
3851 case AArch64::STURDi:
3853 return AArch64::STRDui;
3854 case AArch64::LDURXi:
3856 return AArch64::LDRXui;
3857 case AArch64::STURXi:
3859 return AArch64::STRXui;
3860 case AArch64::LDURWi:
3862 return AArch64::LDRWui;
3863 case AArch64::LDURSWi:
3865 return AArch64::LDRSWui;
3866 case AArch64::STURWi:
3868 return AArch64::STRWui;
3869 case AArch64::LDURHi:
3871 return AArch64::LDRHui;
3872 case AArch64::STURHi:
3874 return AArch64::STRHui;
3875 case AArch64::LDURHHi:
3877 return AArch64::LDRHHui;
3878 case AArch64::STURHHi:
3880 return AArch64::STRHHui;
3881 case AArch64::LDURSHXi:
3883 return AArch64::LDRSHXui;
3884 case AArch64::LDURSHWi:
3886 return AArch64::LDRSHWui;
3887 case AArch64::LDURBi:
3889 return AArch64::LDRBui;
3890 case AArch64::LDURBBi:
3892 return AArch64::LDRBBui;
3893 case AArch64::LDURSBXi:
3895 return AArch64::LDRSBXui;
3896 case AArch64::LDURSBWi:
3898 return AArch64::LDRSBWui;
3899 case AArch64::STURBi:
3901 return AArch64::STRBui;
3902 case AArch64::STURBBi:
3904 return AArch64::STRBBui;
3905 case AArch64::LDRQui:
3906 case AArch64::STRQui:
3909 case AArch64::LDRDui:
3910 case AArch64::STRDui:
3911 case AArch64::LDRXui:
3912 case AArch64::STRXui:
3915 case AArch64::LDRWui:
3916 case AArch64::LDRSWui:
3917 case AArch64::STRWui:
3920 case AArch64::LDRHui:
3921 case AArch64::STRHui:
3922 case AArch64::LDRHHui:
3923 case AArch64::STRHHui:
3924 case AArch64::LDRSHXui:
3925 case AArch64::LDRSHWui:
3928 case AArch64::LDRBui:
3929 case AArch64::LDRBBui:
3930 case AArch64::LDRSBXui:
3931 case AArch64::LDRSBWui:
3932 case AArch64::STRBui:
3933 case AArch64::STRBBui:
3947 case AArch64::LDURQi:
3948 case AArch64::STURQi:
3949 case AArch64::LDURDi:
3950 case AArch64::STURDi:
3951 case AArch64::LDURXi:
3952 case AArch64::STURXi:
3953 case AArch64::LDURWi:
3954 case AArch64::LDURSWi:
3955 case AArch64::STURWi:
3956 case AArch64::LDURHi:
3957 case AArch64::STURHi:
3958 case AArch64::LDURHHi:
3959 case AArch64::STURHHi:
3960 case AArch64::LDURSHXi:
3961 case AArch64::LDURSHWi:
3962 case AArch64::LDURBi:
3963 case AArch64::STURBi:
3964 case AArch64::LDURBBi:
3965 case AArch64::STURBBi:
3966 case AArch64::LDURSBWi:
3967 case AArch64::LDURSBXi:
3969 case AArch64::LDRQui:
3970 return AArch64::LDURQi;
3971 case AArch64::STRQui:
3972 return AArch64::STURQi;
3973 case AArch64::LDRDui:
3974 return AArch64::LDURDi;
3975 case AArch64::STRDui:
3976 return AArch64::STURDi;
3977 case AArch64::LDRXui:
3978 return AArch64::LDURXi;
3979 case AArch64::STRXui:
3980 return AArch64::STURXi;
3981 case AArch64::LDRWui:
3982 return AArch64::LDURWi;
3983 case AArch64::LDRSWui:
3984 return AArch64::LDURSWi;
3985 case AArch64::STRWui:
3986 return AArch64::STURWi;
3987 case AArch64::LDRHui:
3988 return AArch64::LDURHi;
3989 case AArch64::STRHui:
3990 return AArch64::STURHi;
3991 case AArch64::LDRHHui:
3992 return AArch64::LDURHHi;
3993 case AArch64::STRHHui:
3994 return AArch64::STURHHi;
3995 case AArch64::LDRSHXui:
3996 return AArch64::LDURSHXi;
3997 case AArch64::LDRSHWui:
3998 return AArch64::LDURSHWi;
3999 case AArch64::LDRBBui:
4000 return AArch64::LDURBBi;
4001 case AArch64::LDRBui:
4002 return AArch64::LDURBi;
4003 case AArch64::STRBBui:
4004 return AArch64::STURBBi;
4005 case AArch64::STRBui:
4006 return AArch64::STURBi;
4007 case AArch64::LDRSBWui:
4008 return AArch64::LDURSBWi;
4009 case AArch64::LDRSBXui:
4010 return AArch64::LDURSBXi;
4023 case AArch64::LDRQroX:
4024 case AArch64::LDURQi:
4025 case AArch64::LDRQui:
4026 return AArch64::LDRQroW;
4027 case AArch64::STRQroX:
4028 case AArch64::STURQi:
4029 case AArch64::STRQui:
4030 return AArch64::STRQroW;
4031 case AArch64::LDRDroX:
4032 case AArch64::LDURDi:
4033 case AArch64::LDRDui:
4034 return AArch64::LDRDroW;
4035 case AArch64::STRDroX:
4036 case AArch64::STURDi:
4037 case AArch64::STRDui:
4038 return AArch64::STRDroW;
4039 case AArch64::LDRXroX:
4040 case AArch64::LDURXi:
4041 case AArch64::LDRXui:
4042 return AArch64::LDRXroW;
4043 case AArch64::STRXroX:
4044 case AArch64::STURXi:
4045 case AArch64::STRXui:
4046 return AArch64::STRXroW;
4047 case AArch64::LDRWroX:
4048 case AArch64::LDURWi:
4049 case AArch64::LDRWui:
4050 return AArch64::LDRWroW;
4051 case AArch64::LDRSWroX:
4052 case AArch64::LDURSWi:
4053 case AArch64::LDRSWui:
4054 return AArch64::LDRSWroW;
4055 case AArch64::STRWroX:
4056 case AArch64::STURWi:
4057 case AArch64::STRWui:
4058 return AArch64::STRWroW;
4059 case AArch64::LDRHroX:
4060 case AArch64::LDURHi:
4061 case AArch64::LDRHui:
4062 return AArch64::LDRHroW;
4063 case AArch64::STRHroX:
4064 case AArch64::STURHi:
4065 case AArch64::STRHui:
4066 return AArch64::STRHroW;
4067 case AArch64::LDRHHroX:
4068 case AArch64::LDURHHi:
4069 case AArch64::LDRHHui:
4070 return AArch64::LDRHHroW;
4071 case AArch64::STRHHroX:
4072 case AArch64::STURHHi:
4073 case AArch64::STRHHui:
4074 return AArch64::STRHHroW;
4075 case AArch64::LDRSHXroX:
4076 case AArch64::LDURSHXi:
4077 case AArch64::LDRSHXui:
4078 return AArch64::LDRSHXroW;
4079 case AArch64::LDRSHWroX:
4080 case AArch64::LDURSHWi:
4081 case AArch64::LDRSHWui:
4082 return AArch64::LDRSHWroW;
4083 case AArch64::LDRBroX:
4084 case AArch64::LDURBi:
4085 case AArch64::LDRBui:
4086 return AArch64::LDRBroW;
4087 case AArch64::LDRBBroX:
4088 case AArch64::LDURBBi:
4089 case AArch64::LDRBBui:
4090 return AArch64::LDRBBroW;
4091 case AArch64::LDRSBXroX:
4092 case AArch64::LDURSBXi:
4093 case AArch64::LDRSBXui:
4094 return AArch64::LDRSBXroW;
4095 case AArch64::LDRSBWroX:
4096 case AArch64::LDURSBWi:
4097 case AArch64::LDRSBWui:
4098 return AArch64::LDRSBWroW;
4099 case AArch64::STRBroX:
4100 case AArch64::STURBi:
4101 case AArch64::STRBui:
4102 return AArch64::STRBroW;
4103 case AArch64::STRBBroX:
4104 case AArch64::STURBBi:
4105 case AArch64::STRBBui:
4106 return AArch64::STRBBroW;
4131 return B.getInstr();
4135 "Addressing mode not supported for folding");
4152 return B.getInstr();
4159 "Address offset can be a register or an immediate, but not both");
4180 return B.getInstr();
4184 "Function must not be called with an addressing mode it can't handle");
4193 case AArch64::LD1Fourv16b_POST:
4194 case AArch64::LD1Fourv1d_POST:
4195 case AArch64::LD1Fourv2d_POST:
4196 case AArch64::LD1Fourv2s_POST:
4197 case AArch64::LD1Fourv4h_POST:
4198 case AArch64::LD1Fourv4s_POST:
4199 case AArch64::LD1Fourv8b_POST:
4200 case AArch64::LD1Fourv8h_POST:
4201 case AArch64::LD1Onev16b_POST:
4202 case AArch64::LD1Onev1d_POST:
4203 case AArch64::LD1Onev2d_POST:
4204 case AArch64::LD1Onev2s_POST:
4205 case AArch64::LD1Onev4h_POST:
4206 case AArch64::LD1Onev4s_POST:
4207 case AArch64::LD1Onev8b_POST:
4208 case AArch64::LD1Onev8h_POST:
4209 case AArch64::LD1Rv16b_POST:
4210 case AArch64::LD1Rv1d_POST:
4211 case AArch64::LD1Rv2d_POST:
4212 case AArch64::LD1Rv2s_POST:
4213 case AArch64::LD1Rv4h_POST:
4214 case AArch64::LD1Rv4s_POST:
4215 case AArch64::LD1Rv8b_POST:
4216 case AArch64::LD1Rv8h_POST:
4217 case AArch64::LD1Threev16b_POST:
4218 case AArch64::LD1Threev1d_POST:
4219 case AArch64::LD1Threev2d_POST:
4220 case AArch64::LD1Threev2s_POST:
4221 case AArch64::LD1Threev4h_POST:
4222 case AArch64::LD1Threev4s_POST:
4223 case AArch64::LD1Threev8b_POST:
4224 case AArch64::LD1Threev8h_POST:
4225 case AArch64::LD1Twov16b_POST:
4226 case AArch64::LD1Twov1d_POST:
4227 case AArch64::LD1Twov2d_POST:
4228 case AArch64::LD1Twov2s_POST:
4229 case AArch64::LD1Twov4h_POST:
4230 case AArch64::LD1Twov4s_POST:
4231 case AArch64::LD1Twov8b_POST:
4232 case AArch64::LD1Twov8h_POST:
4233 case AArch64::LD1i16_POST:
4234 case AArch64::LD1i32_POST:
4235 case AArch64::LD1i64_POST:
4236 case AArch64::LD1i8_POST:
4237 case AArch64::LD2Rv16b_POST:
4238 case AArch64::LD2Rv1d_POST:
4239 case AArch64::LD2Rv2d_POST:
4240 case AArch64::LD2Rv2s_POST:
4241 case AArch64::LD2Rv4h_POST:
4242 case AArch64::LD2Rv4s_POST:
4243 case AArch64::LD2Rv8b_POST:
4244 case AArch64::LD2Rv8h_POST:
4245 case AArch64::LD2Twov16b_POST:
4246 case AArch64::LD2Twov2d_POST:
4247 case AArch64::LD2Twov2s_POST:
4248 case AArch64::LD2Twov4h_POST:
4249 case AArch64::LD2Twov4s_POST:
4250 case AArch64::LD2Twov8b_POST:
4251 case AArch64::LD2Twov8h_POST:
4252 case AArch64::LD2i16_POST:
4253 case AArch64::LD2i32_POST:
4254 case AArch64::LD2i64_POST:
4255 case AArch64::LD2i8_POST:
4256 case AArch64::LD3Rv16b_POST:
4257 case AArch64::LD3Rv1d_POST:
4258 case AArch64::LD3Rv2d_POST:
4259 case AArch64::LD3Rv2s_POST:
4260 case AArch64::LD3Rv4h_POST:
4261 case AArch64::LD3Rv4s_POST:
4262 case AArch64::LD3Rv8b_POST:
4263 case AArch64::LD3Rv8h_POST:
4264 case AArch64::LD3Threev16b_POST:
4265 case AArch64::LD3Threev2d_POST:
4266 case AArch64::LD3Threev2s_POST:
4267 case AArch64::LD3Threev4h_POST:
4268 case AArch64::LD3Threev4s_POST:
4269 case AArch64::LD3Threev8b_POST:
4270 case AArch64::LD3Threev8h_POST:
4271 case AArch64::LD3i16_POST:
4272 case AArch64::LD3i32_POST:
4273 case AArch64::LD3i64_POST:
4274 case AArch64::LD3i8_POST:
4275 case AArch64::LD4Fourv16b_POST:
4276 case AArch64::LD4Fourv2d_POST:
4277 case AArch64::LD4Fourv2s_POST:
4278 case AArch64::LD4Fourv4h_POST:
4279 case AArch64::LD4Fourv4s_POST:
4280 case AArch64::LD4Fourv8b_POST:
4281 case AArch64::LD4Fourv8h_POST:
4282 case AArch64::LD4Rv16b_POST:
4283 case AArch64::LD4Rv1d_POST:
4284 case AArch64::LD4Rv2d_POST:
4285 case AArch64::LD4Rv2s_POST:
4286 case AArch64::LD4Rv4h_POST:
4287 case AArch64::LD4Rv4s_POST:
4288 case AArch64::LD4Rv8b_POST:
4289 case AArch64::LD4Rv8h_POST:
4290 case AArch64::LD4i16_POST:
4291 case AArch64::LD4i32_POST:
4292 case AArch64::LD4i64_POST:
4293 case AArch64::LD4i8_POST:
4294 case AArch64::LDAPRWpost:
4295 case AArch64::LDAPRXpost:
4296 case AArch64::LDIAPPWpost:
4297 case AArch64::LDIAPPXpost:
4298 case AArch64::LDPDpost:
4299 case AArch64::LDPQpost:
4300 case AArch64::LDPSWpost:
4301 case AArch64::LDPSpost:
4302 case AArch64::LDPWpost:
4303 case AArch64::LDPXpost:
4304 case AArch64::LDRBBpost:
4305 case AArch64::LDRBpost:
4306 case AArch64::LDRDpost:
4307 case AArch64::LDRHHpost:
4308 case AArch64::LDRHpost:
4309 case AArch64::LDRQpost:
4310 case AArch64::LDRSBWpost:
4311 case AArch64::LDRSBXpost:
4312 case AArch64::LDRSHWpost:
4313 case AArch64::LDRSHXpost:
4314 case AArch64::LDRSWpost:
4315 case AArch64::LDRSpost:
4316 case AArch64::LDRWpost:
4317 case AArch64::LDRXpost:
4318 case AArch64::ST1Fourv16b_POST:
4319 case AArch64::ST1Fourv1d_POST:
4320 case AArch64::ST1Fourv2d_POST:
4321 case AArch64::ST1Fourv2s_POST:
4322 case AArch64::ST1Fourv4h_POST:
4323 case AArch64::ST1Fourv4s_POST:
4324 case AArch64::ST1Fourv8b_POST:
4325 case AArch64::ST1Fourv8h_POST:
4326 case AArch64::ST1Onev16b_POST:
4327 case AArch64::ST1Onev1d_POST:
4328 case AArch64::ST1Onev2d_POST:
4329 case AArch64::ST1Onev2s_POST:
4330 case AArch64::ST1Onev4h_POST:
4331 case AArch64::ST1Onev4s_POST:
4332 case AArch64::ST1Onev8b_POST:
4333 case AArch64::ST1Onev8h_POST:
4334 case AArch64::ST1Threev16b_POST:
4335 case AArch64::ST1Threev1d_POST:
4336 case AArch64::ST1Threev2d_POST:
4337 case AArch64::ST1Threev2s_POST:
4338 case AArch64::ST1Threev4h_POST:
4339 case AArch64::ST1Threev4s_POST:
4340 case AArch64::ST1Threev8b_POST:
4341 case AArch64::ST1Threev8h_POST:
4342 case AArch64::ST1Twov16b_POST:
4343 case AArch64::ST1Twov1d_POST:
4344 case AArch64::ST1Twov2d_POST:
4345 case AArch64::ST1Twov2s_POST:
4346 case AArch64::ST1Twov4h_POST:
4347 case AArch64::ST1Twov4s_POST:
4348 case AArch64::ST1Twov8b_POST:
4349 case AArch64::ST1Twov8h_POST:
4350 case AArch64::ST1i16_POST:
4351 case AArch64::ST1i32_POST:
4352 case AArch64::ST1i64_POST:
4353 case AArch64::ST1i8_POST:
4354 case AArch64::ST2GPostIndex:
4355 case AArch64::ST2Twov16b_POST:
4356 case AArch64::ST2Twov2d_POST:
4357 case AArch64::ST2Twov2s_POST:
4358 case AArch64::ST2Twov4h_POST:
4359 case AArch64::ST2Twov4s_POST:
4360 case AArch64::ST2Twov8b_POST:
4361 case AArch64::ST2Twov8h_POST:
4362 case AArch64::ST2i16_POST:
4363 case AArch64::ST2i32_POST:
4364 case AArch64::ST2i64_POST:
4365 case AArch64::ST2i8_POST:
4366 case AArch64::ST3Threev16b_POST:
4367 case AArch64::ST3Threev2d_POST:
4368 case AArch64::ST3Threev2s_POST:
4369 case AArch64::ST3Threev4h_POST:
4370 case AArch64::ST3Threev4s_POST:
4371 case AArch64::ST3Threev8b_POST:
4372 case AArch64::ST3Threev8h_POST:
4373 case AArch64::ST3i16_POST:
4374 case AArch64::ST3i32_POST:
4375 case AArch64::ST3i64_POST:
4376 case AArch64::ST3i8_POST:
4377 case AArch64::ST4Fourv16b_POST:
4378 case AArch64::ST4Fourv2d_POST:
4379 case AArch64::ST4Fourv2s_POST:
4380 case AArch64::ST4Fourv4h_POST:
4381 case AArch64::ST4Fourv4s_POST:
4382 case AArch64::ST4Fourv8b_POST:
4383 case AArch64::ST4Fourv8h_POST:
4384 case AArch64::ST4i16_POST:
4385 case AArch64::ST4i32_POST:
4386 case AArch64::ST4i64_POST:
4387 case AArch64::ST4i8_POST:
4388 case AArch64::STGPostIndex:
4389 case AArch64::STGPpost:
4390 case AArch64::STPDpost:
4391 case AArch64::STPQpost:
4392 case AArch64::STPSpost:
4393 case AArch64::STPWpost:
4394 case AArch64::STPXpost:
4395 case AArch64::STRBBpost:
4396 case AArch64::STRBpost:
4397 case AArch64::STRDpost:
4398 case AArch64::STRHHpost:
4399 case AArch64::STRHpost:
4400 case AArch64::STRQpost:
4401 case AArch64::STRSpost:
4402 case AArch64::STRWpost:
4403 case AArch64::STRXpost:
4404 case AArch64::STZ2GPostIndex:
4405 case AArch64::STZGPostIndex:
4412 bool &OffsetIsScalable,
TypeSize &Width,
4433 int64_t Dummy1, Dummy2;
4455 return BaseOp->
isReg() || BaseOp->
isFI();
4462 assert(OfsOp.
isImm() &&
"Offset operand wasn't immediate.");
4467 TypeSize &Width, int64_t &MinOffset,
4468 int64_t &MaxOffset) {
4474 MinOffset = MaxOffset = 0;
4477 case AArch64::LDRQui:
4478 case AArch64::STRQui:
4484 case AArch64::LDRXui:
4485 case AArch64::LDRDui:
4486 case AArch64::STRXui:
4487 case AArch64::STRDui:
4488 case AArch64::PRFMui:
4494 case AArch64::LDRWui:
4495 case AArch64::LDRSui:
4496 case AArch64::LDRSWui:
4497 case AArch64::STRWui:
4498 case AArch64::STRSui:
4504 case AArch64::LDRHui:
4505 case AArch64::LDRHHui:
4506 case AArch64::LDRSHWui:
4507 case AArch64::LDRSHXui:
4508 case AArch64::STRHui:
4509 case AArch64::STRHHui:
4515 case AArch64::LDRBui:
4516 case AArch64::LDRBBui:
4517 case AArch64::LDRSBWui:
4518 case AArch64::LDRSBXui:
4519 case AArch64::STRBui:
4520 case AArch64::STRBBui:
4527 case AArch64::STRQpre:
4528 case AArch64::LDRQpost:
4534 case AArch64::LDRDpost:
4535 case AArch64::LDRDpre:
4536 case AArch64::LDRXpost:
4537 case AArch64::LDRXpre:
4538 case AArch64::STRDpost:
4539 case AArch64::STRDpre:
4540 case AArch64::STRXpost:
4541 case AArch64::STRXpre:
4547 case AArch64::STRWpost:
4548 case AArch64::STRWpre:
4549 case AArch64::LDRWpost:
4550 case AArch64::LDRWpre:
4551 case AArch64::STRSpost:
4552 case AArch64::STRSpre:
4553 case AArch64::LDRSpost:
4554 case AArch64::LDRSpre:
4560 case AArch64::LDRHpost:
4561 case AArch64::LDRHpre:
4562 case AArch64::STRHpost:
4563 case AArch64::STRHpre:
4564 case AArch64::LDRHHpost:
4565 case AArch64::LDRHHpre:
4566 case AArch64::STRHHpost:
4567 case AArch64::STRHHpre:
4573 case AArch64::LDRBpost:
4574 case AArch64::LDRBpre:
4575 case AArch64::STRBpost:
4576 case AArch64::STRBpre:
4577 case AArch64::LDRBBpost:
4578 case AArch64::LDRBBpre:
4579 case AArch64::STRBBpost:
4580 case AArch64::STRBBpre:
4587 case AArch64::LDURQi:
4588 case AArch64::STURQi:
4594 case AArch64::LDURXi:
4595 case AArch64::LDURDi:
4596 case AArch64::LDAPURXi:
4597 case AArch64::STURXi:
4598 case AArch64::STURDi:
4599 case AArch64::STLURXi:
4600 case AArch64::PRFUMi:
4606 case AArch64::LDURWi:
4607 case AArch64::LDURSi:
4608 case AArch64::LDURSWi:
4609 case AArch64::LDAPURi:
4610 case AArch64::LDAPURSWi:
4611 case AArch64::STURWi:
4612 case AArch64::STURSi:
4613 case AArch64::STLURWi:
4619 case AArch64::LDURHi:
4620 case AArch64::LDURHHi:
4621 case AArch64::LDURSHXi:
4622 case AArch64::LDURSHWi:
4623 case AArch64::LDAPURHi:
4624 case AArch64::LDAPURSHWi:
4625 case AArch64::LDAPURSHXi:
4626 case AArch64::STURHi:
4627 case AArch64::STURHHi:
4628 case AArch64::STLURHi:
4634 case AArch64::LDURBi:
4635 case AArch64::LDURBBi:
4636 case AArch64::LDURSBXi:
4637 case AArch64::LDURSBWi:
4638 case AArch64::LDAPURBi:
4639 case AArch64::LDAPURSBWi:
4640 case AArch64::LDAPURSBXi:
4641 case AArch64::STURBi:
4642 case AArch64::STURBBi:
4643 case AArch64::STLURBi:
4650 case AArch64::LDPQi:
4651 case AArch64::LDNPQi:
4652 case AArch64::STPQi:
4653 case AArch64::STNPQi:
4654 case AArch64::LDPQpost:
4655 case AArch64::LDPQpre:
4656 case AArch64::STPQpost:
4657 case AArch64::STPQpre:
4663 case AArch64::LDPXi:
4664 case AArch64::LDPDi:
4665 case AArch64::LDNPXi:
4666 case AArch64::LDNPDi:
4667 case AArch64::STPXi:
4668 case AArch64::STPDi:
4669 case AArch64::STNPXi:
4670 case AArch64::STNPDi:
4671 case AArch64::LDPDpost:
4672 case AArch64::LDPDpre:
4673 case AArch64::LDPXpost:
4674 case AArch64::LDPXpre:
4675 case AArch64::STPDpost:
4676 case AArch64::STPDpre:
4677 case AArch64::STPXpost:
4678 case AArch64::STPXpre:
4684 case AArch64::LDPWi:
4685 case AArch64::LDPSi:
4686 case AArch64::LDNPWi:
4687 case AArch64::LDNPSi:
4688 case AArch64::STPWi:
4689 case AArch64::STPSi:
4690 case AArch64::STNPWi:
4691 case AArch64::STNPSi:
4692 case AArch64::LDPSpost:
4693 case AArch64::LDPSpre:
4694 case AArch64::LDPWpost:
4695 case AArch64::LDPWpre:
4696 case AArch64::STPSpost:
4697 case AArch64::STPSpre:
4698 case AArch64::STPWpost:
4699 case AArch64::STPWpre:
4705 case AArch64::StoreSwiftAsyncContext:
4718 case AArch64::TAGPstack:
4728 case AArch64::STGPreIndex:
4729 case AArch64::STGPostIndex:
4730 case AArch64::STZGi:
4731 case AArch64::STZGPreIndex:
4732 case AArch64::STZGPostIndex:
4739 case AArch64::STR_ZZZZXI:
4740 case AArch64::STR_ZZZZXI_STRIDED_CONTIGUOUS:
4741 case AArch64::LDR_ZZZZXI:
4742 case AArch64::LDR_ZZZZXI_STRIDED_CONTIGUOUS:
4748 case AArch64::STR_ZZZXI:
4749 case AArch64::LDR_ZZZXI:
4755 case AArch64::STR_ZZXI:
4756 case AArch64::STR_ZZXI_STRIDED_CONTIGUOUS:
4757 case AArch64::LDR_ZZXI:
4758 case AArch64::LDR_ZZXI_STRIDED_CONTIGUOUS:
4764 case AArch64::LDR_PXI:
4765 case AArch64::STR_PXI:
4771 case AArch64::LDR_PPXI:
4772 case AArch64::STR_PPXI:
4778 case AArch64::LDR_ZXI:
4779 case AArch64::STR_ZXI:
4785 case AArch64::LD1B_IMM:
4786 case AArch64::LD1H_IMM:
4787 case AArch64::LD1W_IMM:
4788 case AArch64::LD1D_IMM:
4789 case AArch64::LDNT1B_ZRI:
4790 case AArch64::LDNT1H_ZRI:
4791 case AArch64::LDNT1W_ZRI:
4792 case AArch64::LDNT1D_ZRI:
4793 case AArch64::ST1B_IMM:
4794 case AArch64::ST1H_IMM:
4795 case AArch64::ST1W_IMM:
4796 case AArch64::ST1D_IMM:
4797 case AArch64::STNT1B_ZRI:
4798 case AArch64::STNT1H_ZRI:
4799 case AArch64::STNT1W_ZRI:
4800 case AArch64::STNT1D_ZRI:
4801 case AArch64::LDNF1B_IMM:
4802 case AArch64::LDNF1H_IMM:
4803 case AArch64::LDNF1W_IMM:
4804 case AArch64::LDNF1D_IMM:
4812 case AArch64::LD2B_IMM:
4813 case AArch64::LD2H_IMM:
4814 case AArch64::LD2W_IMM:
4815 case AArch64::LD2D_IMM:
4816 case AArch64::ST2B_IMM:
4817 case AArch64::ST2H_IMM:
4818 case AArch64::ST2W_IMM:
4819 case AArch64::ST2D_IMM:
4825 case AArch64::LD3B_IMM:
4826 case AArch64::LD3H_IMM:
4827 case AArch64::LD3W_IMM:
4828 case AArch64::LD3D_IMM:
4829 case AArch64::ST3B_IMM:
4830 case AArch64::ST3H_IMM:
4831 case AArch64::ST3W_IMM:
4832 case AArch64::ST3D_IMM:
4838 case AArch64::LD4B_IMM:
4839 case AArch64::LD4H_IMM:
4840 case AArch64::LD4W_IMM:
4841 case AArch64::LD4D_IMM:
4842 case AArch64::ST4B_IMM:
4843 case AArch64::ST4H_IMM:
4844 case AArch64::ST4W_IMM:
4845 case AArch64::ST4D_IMM:
4851 case AArch64::LD1B_H_IMM:
4852 case AArch64::LD1SB_H_IMM:
4853 case AArch64::LD1H_S_IMM:
4854 case AArch64::LD1SH_S_IMM:
4855 case AArch64::LD1W_D_IMM:
4856 case AArch64::LD1SW_D_IMM:
4857 case AArch64::ST1B_H_IMM:
4858 case AArch64::ST1H_S_IMM:
4859 case AArch64::ST1W_D_IMM:
4860 case AArch64::LDNF1B_H_IMM:
4861 case AArch64::LDNF1SB_H_IMM:
4862 case AArch64::LDNF1H_S_IMM:
4863 case AArch64::LDNF1SH_S_IMM:
4864 case AArch64::LDNF1W_D_IMM:
4865 case AArch64::LDNF1SW_D_IMM:
4873 case AArch64::LD1B_S_IMM:
4874 case AArch64::LD1SB_S_IMM:
4875 case AArch64::LD1H_D_IMM:
4876 case AArch64::LD1SH_D_IMM:
4877 case AArch64::ST1B_S_IMM:
4878 case AArch64::ST1H_D_IMM:
4879 case AArch64::LDNF1B_S_IMM:
4880 case AArch64::LDNF1SB_S_IMM:
4881 case AArch64::LDNF1H_D_IMM:
4882 case AArch64::LDNF1SH_D_IMM:
4890 case AArch64::LD1B_D_IMM:
4891 case AArch64::LD1SB_D_IMM:
4892 case AArch64::ST1B_D_IMM:
4893 case AArch64::LDNF1B_D_IMM:
4894 case AArch64::LDNF1SB_D_IMM:
4902 case AArch64::ST2Gi:
4903 case AArch64::ST2GPreIndex:
4904 case AArch64::ST2GPostIndex:
4905 case AArch64::STZ2Gi:
4906 case AArch64::STZ2GPreIndex:
4907 case AArch64::STZ2GPostIndex:
4913 case AArch64::STGPi:
4914 case AArch64::STGPpost:
4915 case AArch64::STGPpre:
4921 case AArch64::LD1RB_IMM:
4922 case AArch64::LD1RB_H_IMM:
4923 case AArch64::LD1RB_S_IMM:
4924 case AArch64::LD1RB_D_IMM:
4925 case AArch64::LD1RSB_H_IMM:
4926 case AArch64::LD1RSB_S_IMM:
4927 case AArch64::LD1RSB_D_IMM:
4933 case AArch64::LD1RH_IMM:
4934 case AArch64::LD1RH_S_IMM:
4935 case AArch64::LD1RH_D_IMM:
4936 case AArch64::LD1RSH_S_IMM:
4937 case AArch64::LD1RSH_D_IMM:
4943 case AArch64::LD1RW_IMM:
4944 case AArch64::LD1RW_D_IMM:
4945 case AArch64::LD1RSW_IMM:
4951 case AArch64::LD1RD_IMM:
4967 case AArch64::LDRBui:
4968 case AArch64::LDRBBui:
4969 case AArch64::LDURBBi:
4970 case AArch64::LDRSBWui:
4971 case AArch64::LDURSBWi:
4972 case AArch64::STRBui:
4973 case AArch64::STRBBui:
4974 case AArch64::STURBBi:
4976 case AArch64::LDRHui:
4977 case AArch64::LDRHHui:
4978 case AArch64::LDURHHi:
4979 case AArch64::LDRSHWui:
4980 case AArch64::LDURSHWi:
4981 case AArch64::STRHui:
4982 case AArch64::STRHHui:
4983 case AArch64::STURHHi:
4985 case AArch64::LDRSui:
4986 case AArch64::LDURSi:
4987 case AArch64::LDRSpre:
4988 case AArch64::LDRSWui:
4989 case AArch64::LDURSWi:
4990 case AArch64::LDRSWpre:
4991 case AArch64::LDRWpre:
4992 case AArch64::LDRWui:
4993 case AArch64::LDURWi:
4994 case AArch64::STRSui:
4995 case AArch64::STURSi:
4996 case AArch64::STRSpre:
4997 case AArch64::STRWui:
4998 case AArch64::STURWi:
4999 case AArch64::STRWpre:
5000 case AArch64::LDPSi:
5001 case AArch64::LDPSWi:
5002 case AArch64::LDPWi:
5003 case AArch64::STPSi:
5004 case AArch64::STPWi:
5006 case AArch64::LDRDui:
5007 case AArch64::LDURDi:
5008 case AArch64::LDRDpre:
5009 case AArch64::LDRXui:
5010 case AArch64::LDURXi:
5011 case AArch64::LDRXpre:
5012 case AArch64::STRDui:
5013 case AArch64::STURDi:
5014 case AArch64::STRDpre:
5015 case AArch64::STRXui:
5016 case AArch64::STURXi:
5017 case AArch64::STRXpre:
5018 case AArch64::LDPDi:
5019 case AArch64::LDPXi:
5020 case AArch64::STPDi:
5021 case AArch64::STPXi:
5023 case AArch64::LDRQui:
5024 case AArch64::LDURQi:
5025 case AArch64::STRQui:
5026 case AArch64::STURQi:
5027 case AArch64::STRQpre:
5028 case AArch64::LDPQi:
5029 case AArch64::LDRQpre:
5030 case AArch64::STPQi:
5032 case AArch64::STZGi:
5033 case AArch64::ST2Gi:
5034 case AArch64::STZ2Gi:
5035 case AArch64::STGPi:
5041 switch (
MI.getOpcode()) {
5044 case AArch64::LDRWpre:
5045 case AArch64::LDRXpre:
5046 case AArch64::LDRSWpre:
5047 case AArch64::LDRSpre:
5048 case AArch64::LDRDpre:
5049 case AArch64::LDRQpre:
5055 switch (
MI.getOpcode()) {
5058 case AArch64::STRWpre:
5059 case AArch64::STRXpre:
5060 case AArch64::STRSpre:
5061 case AArch64::STRDpre:
5062 case AArch64::STRQpre:
5072 switch (
MI.getOpcode()) {
5075 case AArch64::LDURBBi:
5076 case AArch64::LDURHHi:
5077 case AArch64::LDURWi:
5078 case AArch64::LDRBBui:
5079 case AArch64::LDRHHui:
5080 case AArch64::LDRWui:
5081 case AArch64::LDRBBroX:
5082 case AArch64::LDRHHroX:
5083 case AArch64::LDRWroX:
5084 case AArch64::LDRBBroW:
5085 case AArch64::LDRHHroW:
5086 case AArch64::LDRWroW:
5092 switch (
MI.getOpcode()) {
5095 case AArch64::LDURSBWi:
5096 case AArch64::LDURSHWi:
5097 case AArch64::LDURSBXi:
5098 case AArch64::LDURSHXi:
5099 case AArch64::LDURSWi:
5100 case AArch64::LDRSBWui:
5101 case AArch64::LDRSHWui:
5102 case AArch64::LDRSBXui:
5103 case AArch64::LDRSHXui:
5104 case AArch64::LDRSWui:
5105 case AArch64::LDRSBWroX:
5106 case AArch64::LDRSHWroX:
5107 case AArch64::LDRSBXroX:
5108 case AArch64::LDRSHXroX:
5109 case AArch64::LDRSWroX:
5110 case AArch64::LDRSBWroW:
5111 case AArch64::LDRSHWroW:
5112 case AArch64::LDRSBXroW:
5113 case AArch64::LDRSHXroW:
5114 case AArch64::LDRSWroW:
5120 switch (
MI.getOpcode()) {
5123 case AArch64::LDPSi:
5124 case AArch64::LDPSWi:
5125 case AArch64::LDPDi:
5126 case AArch64::LDPQi:
5127 case AArch64::LDPWi:
5128 case AArch64::LDPXi:
5129 case AArch64::STPSi:
5130 case AArch64::STPDi:
5131 case AArch64::STPQi:
5132 case AArch64::STPWi:
5133 case AArch64::STPXi:
5134 case AArch64::STGPi:
5140 assert(
MI.mayLoadOrStore() &&
"Load or store instruction expected");
5144 return MI.getOperand(Idx);
5149 assert(
MI.mayLoadOrStore() &&
"Load or store instruction expected");
5153 return MI.getOperand(Idx);
5158 switch (
MI.getOpcode()) {
5161 case AArch64::LDRBroX:
5162 case AArch64::LDRBBroX:
5163 case AArch64::LDRSBXroX:
5164 case AArch64::LDRSBWroX:
5165 case AArch64::LDRHroX:
5166 case AArch64::LDRHHroX:
5167 case AArch64::LDRSHXroX:
5168 case AArch64::LDRSHWroX:
5169 case AArch64::LDRWroX:
5170 case AArch64::LDRSroX:
5171 case AArch64::LDRSWroX:
5172 case AArch64::LDRDroX:
5173 case AArch64::LDRXroX:
5174 case AArch64::LDRQroX:
5175 return MI.getOperand(4);
5181 if (
MI.getParent() ==
nullptr)
5191 auto Reg =
Op.getReg();
5192 if (Reg.isPhysical())
5193 return AArch64::FPR16RegClass.contains(Reg);
5195 return TRC == &AArch64::FPR16RegClass ||
5196 TRC == &AArch64::FPR16_loRegClass;
5205 auto Reg =
Op.getReg();
5206 if (Reg.isPhysical())
5207 return AArch64::FPR128RegClass.contains(Reg);
5209 return TRC == &AArch64::FPR128RegClass ||
5210 TRC == &AArch64::FPR128_loRegClass;
5216 switch (
MI.getOpcode()) {
5219 case AArch64::PACIASP:
5220 case AArch64::PACIBSP:
5223 case AArch64::PAUTH_PROLOGUE:
5226 case AArch64::HINT: {
5227 unsigned Imm =
MI.getOperand(0).getImm();
5229 if (Imm == 32 || Imm == 34 || Imm == 36 || Imm == 38)
5232 if (Imm == 25 || Imm == 27)
5244 assert(Reg.isPhysical() &&
"Expected physical register in isFpOrNEON");
5245 return AArch64::FPR128RegClass.contains(Reg) ||
5246 AArch64::FPR64RegClass.contains(Reg) ||
5247 AArch64::FPR32RegClass.contains(Reg) ||
5248 AArch64::FPR16RegClass.contains(Reg) ||
5249 AArch64::FPR8RegClass.contains(Reg);
5256 auto Reg =
Op.getReg();
5257 if (Reg.isPhysical())
5261 return TRC == &AArch64::FPR128RegClass ||
5262 TRC == &AArch64::FPR128_loRegClass ||
5263 TRC == &AArch64::FPR64RegClass ||
5264 TRC == &AArch64::FPR64_loRegClass ||
5265 TRC == &AArch64::FPR32RegClass || TRC == &AArch64::FPR16RegClass ||
5266 TRC == &AArch64::FPR8RegClass;
5288 if (FirstOpc == SecondOpc)
5294 case AArch64::STRSui:
5295 case AArch64::STURSi:
5296 return SecondOpc == AArch64::STRSui || SecondOpc == AArch64::STURSi;
5297 case AArch64::STRDui:
5298 case AArch64::STURDi:
5299 return SecondOpc == AArch64::STRDui || SecondOpc == AArch64::STURDi;
5300 case AArch64::STRQui:
5301 case AArch64::STURQi:
5302 return SecondOpc == AArch64::STRQui || SecondOpc == AArch64::STURQi;
5303 case AArch64::STRWui:
5304 case AArch64::STURWi:
5305 return SecondOpc == AArch64::STRWui || SecondOpc == AArch64::STURWi;
5306 case AArch64::STRXui:
5307 case AArch64::STURXi:
5308 return SecondOpc == AArch64::STRXui || SecondOpc == AArch64::STURXi;
5309 case AArch64::LDRSui:
5310 case AArch64::LDURSi:
5311 return SecondOpc == AArch64::LDRSui || SecondOpc == AArch64::LDURSi;
5312 case AArch64::LDRDui:
5313 case AArch64::LDURDi:
5314 return SecondOpc == AArch64::LDRDui || SecondOpc == AArch64::LDURDi;
5315 case AArch64::LDRQui:
5316 case AArch64::LDURQi:
5317 return SecondOpc == AArch64::LDRQui || SecondOpc == AArch64::LDURQi;
5318 case AArch64::LDRWui:
5319 case AArch64::LDURWi:
5320 return SecondOpc == AArch64::LDRSWui || SecondOpc == AArch64::LDURSWi;
5321 case AArch64::LDRSWui:
5322 case AArch64::LDURSWi:
5323 return SecondOpc == AArch64::LDRWui || SecondOpc == AArch64::LDURWi;
5324 case AArch64::LDRXui:
5325 case AArch64::LDURXi:
5326 return SecondOpc == AArch64::LDRXui || SecondOpc == AArch64::LDURXi;
5333 int64_t Offset1,
unsigned Opcode1,
int FI2,
5334 int64_t Offset2,
unsigned Opcode2) {
5340 assert(ObjectOffset1 <= ObjectOffset2 &&
"Object offsets are not ordered.");
5343 if (ObjectOffset1 % Scale1 != 0)
5345 ObjectOffset1 /= Scale1;
5347 if (ObjectOffset2 % Scale2 != 0)
5349 ObjectOffset2 /= Scale2;
5350 ObjectOffset1 += Offset1;
5351 ObjectOffset2 += Offset2;
5352 return ObjectOffset1 + 1 == ObjectOffset2;
5364 int64_t OpOffset2,
bool OffsetIsScalable2,
unsigned ClusterSize,
5365 unsigned NumBytes)
const {
5375 "Only base registers and frame indices are supported.");
5382 if (ClusterSize > 2)
5389 unsigned FirstOpc = FirstLdSt.
getOpcode();
5390 unsigned SecondOpc = SecondLdSt.
getOpcode();
5410 if (Offset1 > 63 || Offset1 < -64)
5415 if (BaseOp1.
isFI()) {
5417 "Caller should have ordered offsets.");
5422 BaseOp2.
getIndex(), Offset2, SecondOpc);
5425 assert(Offset1 <= Offset2 &&
"Caller should have ordered offsets.");
5427 return Offset1 + 1 == Offset2;
5437 if (
Reg.isPhysical())
5446 return ((DestReg - SrcReg) & 0x1f) < NumRegs;
5455 assert(Subtarget.hasNEON() &&
"Unexpected register copy without NEON");
5457 uint16_t DestEncoding =
TRI->getEncodingValue(DestReg);
5458 uint16_t SrcEncoding =
TRI->getEncodingValue(SrcReg);
5459 unsigned NumRegs = Indices.
size();
5461 int SubReg = 0, End = NumRegs, Incr = 1;
5463 SubReg = NumRegs - 1;
5468 for (; SubReg != End; SubReg += Incr) {
5480 unsigned Opcode,
unsigned ZeroReg,
5483 unsigned NumRegs = Indices.
size();
5486 uint16_t DestEncoding =
TRI->getEncodingValue(DestReg);
5487 uint16_t SrcEncoding =
TRI->getEncodingValue(SrcReg);
5488 assert(DestEncoding % NumRegs == 0 && SrcEncoding % NumRegs == 0 &&
5489 "GPR reg sequences should not be able to overlap");
5492 for (
unsigned SubReg = 0; SubReg != NumRegs; ++SubReg) {
5513 unsigned Opc =
MI.getOpcode();
5514 if (
Opc == AArch64::MSRpstatesvcrImm1 ||
Opc == AArch64::MSRpstatePseudo) {
5516 int64_t PState =
MI.getOperand(0).getImm();
5517 if (PState == AArch64SVCR::SVCRSM || PState == AArch64SVCR::SVCRSMZA) {
5519 return MI.getOperand(1).getImm() == 1;
5538 bool RenamableSrc)
const {
5540 if (AArch64::GPR32spRegClass.
contains(DestReg) &&
5541 AArch64::GPR32spRegClass.
contains(SrcReg)) {
5542 if (DestReg == AArch64::WSP || SrcReg == AArch64::WSP) {
5544 if (Subtarget.hasZeroCycleRegMoveGPR64() &&
5545 !Subtarget.hasZeroCycleRegMoveGPR32()) {
5547 MCRegister DestRegX = RI.getMatchingSuperReg(DestReg, AArch64::sub_32,
5548 &AArch64::GPR64spRegClass);
5549 MCRegister SrcRegX = RI.getMatchingSuperReg(SrcReg, AArch64::sub_32,
5550 &AArch64::GPR64spRegClass);
5560 ++NumZCRegMoveInstrsGPR;
5566 if (Subtarget.hasZeroCycleRegMoveGPR32())
5567 ++NumZCRegMoveInstrsGPR;
5569 }
else if (Subtarget.hasZeroCycleRegMoveGPR64() &&
5570 !Subtarget.hasZeroCycleRegMoveGPR32()) {
5572 MCRegister DestRegX = RI.getMatchingSuperReg(DestReg, AArch64::sub_32,
5573 &AArch64::GPR64spRegClass);
5574 assert(DestRegX.
isValid() &&
"Destination super-reg not valid");
5575 MCRegister SrcRegX = RI.getMatchingSuperReg(SrcReg, AArch64::sub_32,
5576 &AArch64::GPR64spRegClass);
5586 ++NumZCRegMoveInstrsGPR;
5592 if (Subtarget.hasZeroCycleRegMoveGPR32())
5593 ++NumZCRegMoveInstrsGPR;
5599 if (AArch64::GPR32spRegClass.
contains(DestReg) && SrcReg == AArch64::WZR) {
5600 if (Subtarget.hasZeroCycleZeroingGPR64() &&
5601 !Subtarget.hasZeroCycleZeroingGPR32()) {
5602 MCRegister DestRegX = RI.getMatchingSuperReg(DestReg, AArch64::sub_32,
5603 &AArch64::GPR64spRegClass);
5604 assert(DestRegX.
isValid() &&
"Destination super-reg not valid");
5608 ++NumZCZeroingInstrsGPR;
5609 }
else if (Subtarget.hasZeroCycleZeroingGPR32()) {
5613 ++NumZCZeroingInstrsGPR;
5622 if (AArch64::GPR64spRegClass.
contains(DestReg) &&
5623 AArch64::GPR64spRegClass.
contains(SrcReg)) {
5624 if (DestReg == AArch64::SP || SrcReg == AArch64::SP) {
5630 if (Subtarget.hasZeroCycleRegMoveGPR64())
5631 ++NumZCRegMoveInstrsGPR;
5637 if (Subtarget.hasZeroCycleRegMoveGPR64())
5638 ++NumZCRegMoveInstrsGPR;
5644 if (AArch64::GPR64spRegClass.
contains(DestReg) && SrcReg == AArch64::XZR) {
5645 if (Subtarget.hasZeroCycleZeroingGPR64()) {
5649 ++NumZCZeroingInstrsGPR;
5659 if (AArch64::PPRRegClass.
contains(DestReg) &&
5660 AArch64::PPRRegClass.
contains(SrcReg)) {
5661 assert(Subtarget.isSVEorStreamingSVEAvailable() &&
5662 "Unexpected SVE register.");
5672 bool DestIsPNR = AArch64::PNRRegClass.contains(DestReg);
5673 bool SrcIsPNR = AArch64::PNRRegClass.contains(SrcReg);
5674 if (DestIsPNR || SrcIsPNR) {
5676 return (R - AArch64::PN0) + AArch64::P0;
5681 if (PPRSrcReg != PPRDestReg) {
5693 if (AArch64::ZPRRegClass.
contains(DestReg) &&
5694 AArch64::ZPRRegClass.
contains(SrcReg)) {
5695 assert(Subtarget.isSVEorStreamingSVEAvailable() &&
5696 "Unexpected SVE register.");
5704 if ((AArch64::ZPR2RegClass.
contains(DestReg) ||
5705 AArch64::ZPR2StridedOrContiguousRegClass.
contains(DestReg)) &&
5706 (AArch64::ZPR2RegClass.
contains(SrcReg) ||
5707 AArch64::ZPR2StridedOrContiguousRegClass.
contains(SrcReg))) {
5708 assert(Subtarget.isSVEorStreamingSVEAvailable() &&
5709 "Unexpected SVE register.");
5710 static const unsigned Indices[] = {AArch64::zsub0, AArch64::zsub1};
5717 if (AArch64::ZPR3RegClass.
contains(DestReg) &&
5718 AArch64::ZPR3RegClass.
contains(SrcReg)) {
5719 assert(Subtarget.isSVEorStreamingSVEAvailable() &&
5720 "Unexpected SVE register.");
5721 static const unsigned Indices[] = {AArch64::zsub0, AArch64::zsub1,
5729 if ((AArch64::ZPR4RegClass.
contains(DestReg) ||
5730 AArch64::ZPR4StridedOrContiguousRegClass.
contains(DestReg)) &&
5731 (AArch64::ZPR4RegClass.
contains(SrcReg) ||
5732 AArch64::ZPR4StridedOrContiguousRegClass.
contains(SrcReg))) {
5733 assert(Subtarget.isSVEorStreamingSVEAvailable() &&
5734 "Unexpected SVE register.");
5735 static const unsigned Indices[] = {AArch64::zsub0, AArch64::zsub1,
5736 AArch64::zsub2, AArch64::zsub3};
5743 if (AArch64::DDDDRegClass.
contains(DestReg) &&
5744 AArch64::DDDDRegClass.
contains(SrcReg)) {
5745 static const unsigned Indices[] = {AArch64::dsub0, AArch64::dsub1,
5746 AArch64::dsub2, AArch64::dsub3};
5753 if (AArch64::DDDRegClass.
contains(DestReg) &&
5754 AArch64::DDDRegClass.
contains(SrcReg)) {
5755 static const unsigned Indices[] = {AArch64::dsub0, AArch64::dsub1,
5763 if (AArch64::DDRegClass.
contains(DestReg) &&
5764 AArch64::DDRegClass.
contains(SrcReg)) {
5765 static const unsigned Indices[] = {AArch64::dsub0, AArch64::dsub1};
5772 if (AArch64::QQQQRegClass.
contains(DestReg) &&
5773 AArch64::QQQQRegClass.
contains(SrcReg)) {
5774 static const unsigned Indices[] = {AArch64::qsub0, AArch64::qsub1,
5775 AArch64::qsub2, AArch64::qsub3};
5782 if (AArch64::QQQRegClass.
contains(DestReg) &&
5783 AArch64::QQQRegClass.
contains(SrcReg)) {
5784 static const unsigned Indices[] = {AArch64::qsub0, AArch64::qsub1,
5792 if (AArch64::QQRegClass.
contains(DestReg) &&
5793 AArch64::QQRegClass.
contains(SrcReg)) {
5794 static const unsigned Indices[] = {AArch64::qsub0, AArch64::qsub1};
5800 if (AArch64::XSeqPairsClassRegClass.
contains(DestReg) &&
5801 AArch64::XSeqPairsClassRegClass.
contains(SrcReg)) {
5802 static const unsigned Indices[] = {AArch64::sube64, AArch64::subo64};
5804 AArch64::XZR, Indices);
5808 if (AArch64::WSeqPairsClassRegClass.
contains(DestReg) &&
5809 AArch64::WSeqPairsClassRegClass.
contains(SrcReg)) {
5810 static const unsigned Indices[] = {AArch64::sube32, AArch64::subo32};
5812 AArch64::WZR, Indices);
5816 if (AArch64::FPR128RegClass.
contains(DestReg) &&
5817 AArch64::FPR128RegClass.
contains(SrcReg)) {
5821 if ((Subtarget.isSVEorStreamingSVEAvailable() &&
5822 !Subtarget.isNeonAvailable()) ||
5826 .
addReg(AArch64::Z0 + (SrcReg - AArch64::Q0))
5827 .
addReg(AArch64::Z0 + (SrcReg - AArch64::Q0));
5828 }
else if (Subtarget.isNeonAvailable()) {
5832 if (Subtarget.hasZeroCycleRegMoveFPR128())
5833 ++NumZCRegMoveInstrsFPR;
5849 if (AArch64::FPR64RegClass.
contains(DestReg) &&
5850 AArch64::FPR64RegClass.
contains(SrcReg)) {
5851 if (Subtarget.hasZeroCycleRegMoveFPR128() &&
5852 !Subtarget.hasZeroCycleRegMoveFPR64() &&
5853 !Subtarget.hasZeroCycleRegMoveFPR32() && Subtarget.isNeonAvailable() &&
5855 MCRegister DestRegQ = RI.getMatchingSuperReg(DestReg, AArch64::dsub,
5856 &AArch64::FPR128RegClass);
5857 MCRegister SrcRegQ = RI.getMatchingSuperReg(SrcReg, AArch64::dsub,
5858 &AArch64::FPR128RegClass);
5867 ++NumZCRegMoveInstrsFPR;
5871 if (Subtarget.hasZeroCycleRegMoveFPR64())
5872 ++NumZCRegMoveInstrsFPR;
5877 if (AArch64::FPR32RegClass.
contains(DestReg) &&
5878 AArch64::FPR32RegClass.
contains(SrcReg)) {
5879 if (Subtarget.hasZeroCycleRegMoveFPR128() &&
5880 !Subtarget.hasZeroCycleRegMoveFPR64() &&
5881 !Subtarget.hasZeroCycleRegMoveFPR32() && Subtarget.isNeonAvailable() &&
5883 MCRegister DestRegQ = RI.getMatchingSuperReg(DestReg, AArch64::ssub,
5884 &AArch64::FPR128RegClass);
5885 MCRegister SrcRegQ = RI.getMatchingSuperReg(SrcReg, AArch64::ssub,
5886 &AArch64::FPR128RegClass);
5895 ++NumZCRegMoveInstrsFPR;
5896 }
else if (Subtarget.hasZeroCycleRegMoveFPR64() &&
5897 !Subtarget.hasZeroCycleRegMoveFPR32()) {
5898 MCRegister DestRegD = RI.getMatchingSuperReg(DestReg, AArch64::ssub,
5899 &AArch64::FPR64RegClass);
5900 MCRegister SrcRegD = RI.getMatchingSuperReg(SrcReg, AArch64::ssub,
5901 &AArch64::FPR64RegClass);
5909 ++NumZCRegMoveInstrsFPR;
5913 if (Subtarget.hasZeroCycleRegMoveFPR32())
5914 ++NumZCRegMoveInstrsFPR;
5919 if (AArch64::FPR16RegClass.
contains(DestReg) &&
5920 AArch64::FPR16RegClass.
contains(SrcReg)) {
5921 if (Subtarget.hasZeroCycleRegMoveFPR128() &&
5922 !Subtarget.hasZeroCycleRegMoveFPR64() &&
5923 !Subtarget.hasZeroCycleRegMoveFPR32() && Subtarget.isNeonAvailable() &&
5925 MCRegister DestRegQ = RI.getMatchingSuperReg(DestReg, AArch64::hsub,
5926 &AArch64::FPR128RegClass);
5927 MCRegister SrcRegQ = RI.getMatchingSuperReg(SrcReg, AArch64::hsub,
5928 &AArch64::FPR128RegClass);
5937 }
else if (Subtarget.hasZeroCycleRegMoveFPR64() &&
5938 !Subtarget.hasZeroCycleRegMoveFPR32()) {
5939 MCRegister DestRegD = RI.getMatchingSuperReg(DestReg, AArch64::hsub,
5940 &AArch64::FPR64RegClass);
5941 MCRegister SrcRegD = RI.getMatchingSuperReg(SrcReg, AArch64::hsub,
5942 &AArch64::FPR64RegClass);
5951 DestReg = RI.getMatchingSuperReg(DestReg, AArch64::hsub,
5952 &AArch64::FPR32RegClass);
5953 SrcReg = RI.getMatchingSuperReg(SrcReg, AArch64::hsub,
5954 &AArch64::FPR32RegClass);
5961 if (AArch64::FPR8RegClass.
contains(DestReg) &&
5962 AArch64::FPR8RegClass.
contains(SrcReg)) {
5963 if (Subtarget.hasZeroCycleRegMoveFPR128() &&
5964 !Subtarget.hasZeroCycleRegMoveFPR64() &&
5965 !Subtarget.hasZeroCycleRegMoveFPR32() && Subtarget.isNeonAvailable() &&
5967 MCRegister DestRegQ = RI.getMatchingSuperReg(DestReg, AArch64::bsub,
5968 &AArch64::FPR128RegClass);
5969 MCRegister SrcRegQ = RI.getMatchingSuperReg(SrcReg, AArch64::bsub,
5970 &AArch64::FPR128RegClass);
5979 }
else if (Subtarget.hasZeroCycleRegMoveFPR64() &&
5980 !Subtarget.hasZeroCycleRegMoveFPR32()) {
5981 MCRegister DestRegD = RI.getMatchingSuperReg(DestReg, AArch64::bsub,
5982 &AArch64::FPR64RegClass);
5983 MCRegister SrcRegD = RI.getMatchingSuperReg(SrcReg, AArch64::bsub,
5984 &AArch64::FPR64RegClass);
5993 DestReg = RI.getMatchingSuperReg(DestReg, AArch64::bsub,
5994 &AArch64::FPR32RegClass);
5995 SrcReg = RI.getMatchingSuperReg(SrcReg, AArch64::bsub,
5996 &AArch64::FPR32RegClass);
6004 if (AArch64::FPR64RegClass.
contains(DestReg) &&
6005 AArch64::GPR64RegClass.
contains(SrcReg)) {
6006 if (AArch64::XZR == SrcReg) {
6014 if (AArch64::GPR64RegClass.
contains(DestReg) &&
6015 AArch64::FPR64RegClass.
contains(SrcReg)) {
6021 if (AArch64::FPR32RegClass.
contains(DestReg) &&
6022 AArch64::GPR32RegClass.
contains(SrcReg)) {
6023 if (AArch64::WZR == SrcReg) {
6031 if (AArch64::GPR32RegClass.
contains(DestReg) &&
6032 AArch64::FPR32RegClass.
contains(SrcReg)) {
6038 if (DestReg == AArch64::NZCV) {
6039 assert(AArch64::GPR64RegClass.
contains(SrcReg) &&
"Invalid NZCV copy");
6041 .
addImm(AArch64SysReg::NZCV)
6047 if (SrcReg == AArch64::NZCV) {
6048 assert(AArch64::GPR64RegClass.
contains(DestReg) &&
"Invalid NZCV copy");
6050 .
addImm(AArch64SysReg::NZCV)
6056 errs() << RI.getRegAsmName(DestReg) <<
" = COPY " << RI.getRegAsmName(SrcReg)
6067 unsigned SubIdx0,
unsigned SubIdx1,
int FI,
6072 SrcReg0 =
TRI.getSubReg(SrcReg, SubIdx0);
6074 SrcReg1 =
TRI.getSubReg(SrcReg, SubIdx1);
6087 Register SrcReg,
bool isKill,
int FI,
6102 switch (RI.getSpillSize(*RC)) {
6104 if (AArch64::FPR8RegClass.hasSubClassEq(RC))
6105 Opc = AArch64::STRBui;
6108 if (AArch64::FPR16RegClass.hasSubClassEq(RC))
6109 Opc = AArch64::STRHui;
6110 else if (AArch64::PNRRegClass.hasSubClassEq(RC) ||
6111 AArch64::PPRRegClass.hasSubClassEq(RC)) {
6112 assert(Subtarget.isSVEorStreamingSVEAvailable() &&
6113 "Unexpected register store without SVE store instructions");
6114 Opc = AArch64::STR_PXI;
6120 if (AArch64::GPR32allRegClass.hasSubClassEq(RC)) {
6121 Opc = AArch64::STRWui;
6125 assert(SrcReg != AArch64::WSP);
6126 }
else if (AArch64::FPR32RegClass.hasSubClassEq(RC))
6127 Opc = AArch64::STRSui;
6128 else if (AArch64::PPR2RegClass.hasSubClassEq(RC)) {
6129 Opc = AArch64::STR_PPXI;
6134 if (AArch64::GPR64allRegClass.hasSubClassEq(RC)) {
6135 Opc = AArch64::STRXui;
6139 assert(SrcReg != AArch64::SP);
6140 }
else if (AArch64::FPR64RegClass.hasSubClassEq(RC)) {
6141 Opc = AArch64::STRDui;
6142 }
else if (AArch64::WSeqPairsClassRegClass.hasSubClassEq(RC)) {
6144 get(AArch64::STPWi), SrcReg, isKill,
6145 AArch64::sube32, AArch64::subo32, FI, MMO);
6150 if (AArch64::FPR128RegClass.hasSubClassEq(RC))
6151 Opc = AArch64::STRQui;
6152 else if (AArch64::DDRegClass.hasSubClassEq(RC)) {
6153 assert(Subtarget.hasNEON() &&
"Unexpected register store without NEON");
6154 Opc = AArch64::ST1Twov1d;
6156 }
else if (AArch64::XSeqPairsClassRegClass.hasSubClassEq(RC)) {
6158 get(AArch64::STPXi), SrcReg, isKill,
6159 AArch64::sube64, AArch64::subo64, FI, MMO);
6161 }
else if (AArch64::ZPRRegClass.hasSubClassEq(RC)) {
6162 assert(Subtarget.isSVEorStreamingSVEAvailable() &&
6163 "Unexpected register store without SVE store instructions");
6164 Opc = AArch64::STR_ZXI;
6169 if (AArch64::DDDRegClass.hasSubClassEq(RC)) {
6170 assert(Subtarget.hasNEON() &&
"Unexpected register store without NEON");
6171 Opc = AArch64::ST1Threev1d;
6176 if (AArch64::DDDDRegClass.hasSubClassEq(RC)) {
6177 assert(Subtarget.hasNEON() &&
"Unexpected register store without NEON");
6178 Opc = AArch64::ST1Fourv1d;
6180 }
else if (AArch64::QQRegClass.hasSubClassEq(RC)) {
6181 assert(Subtarget.hasNEON() &&
"Unexpected register store without NEON");
6182 Opc = AArch64::ST1Twov2d;
6184 }
else if (AArch64::ZPR2StridedOrContiguousRegClass.hasSubClassEq(RC)) {
6185 assert(Subtarget.isSVEorStreamingSVEAvailable() &&
6186 "Unexpected register store without SVE store instructions");
6187 Opc = AArch64::STR_ZZXI_STRIDED_CONTIGUOUS;
6189 }
else if (AArch64::ZPR2RegClass.hasSubClassEq(RC)) {
6190 assert(Subtarget.isSVEorStreamingSVEAvailable() &&
6191 "Unexpected register store without SVE store instructions");
6192 Opc = AArch64::STR_ZZXI;
6197 if (AArch64::QQQRegClass.hasSubClassEq(RC)) {
6198 assert(Subtarget.hasNEON() &&
"Unexpected register store without NEON");
6199 Opc = AArch64::ST1Threev2d;
6201 }
else if (AArch64::ZPR3RegClass.hasSubClassEq(RC)) {
6202 assert(Subtarget.isSVEorStreamingSVEAvailable() &&
6203 "Unexpected register store without SVE store instructions");
6204 Opc = AArch64::STR_ZZZXI;
6209 if (AArch64::QQQQRegClass.hasSubClassEq(RC)) {
6210 assert(Subtarget.hasNEON() &&
"Unexpected register store without NEON");
6211 Opc = AArch64::ST1Fourv2d;
6213 }
else if (AArch64::ZPR4StridedOrContiguousRegClass.hasSubClassEq(RC)) {
6214 assert(Subtarget.isSVEorStreamingSVEAvailable() &&
6215 "Unexpected register store without SVE store instructions");
6216 Opc = AArch64::STR_ZZZZXI_STRIDED_CONTIGUOUS;
6218 }
else if (AArch64::ZPR4RegClass.hasSubClassEq(RC)) {
6219 assert(Subtarget.isSVEorStreamingSVEAvailable() &&
6220 "Unexpected register store without SVE store instructions");
6221 Opc = AArch64::STR_ZZZZXI;
6226 assert(
Opc &&
"Unknown register class");
6237 MI.addMemOperand(MMO);
6244 Register DestReg,
unsigned SubIdx0,
6245 unsigned SubIdx1,
int FI,
6249 bool IsUndef =
true;
6251 DestReg0 =
TRI.getSubReg(DestReg, SubIdx0);
6253 DestReg1 =
TRI.getSubReg(DestReg, SubIdx1);
6282 switch (
TRI.getSpillSize(*RC)) {
6284 if (AArch64::FPR8RegClass.hasSubClassEq(RC))
6285 Opc = AArch64::LDRBui;
6288 bool IsPNR = AArch64::PNRRegClass.hasSubClassEq(RC);
6289 if (AArch64::FPR16RegClass.hasSubClassEq(RC))
6290 Opc = AArch64::LDRHui;
6291 else if (IsPNR || AArch64::PPRRegClass.hasSubClassEq(RC)) {
6292 assert(Subtarget.isSVEorStreamingSVEAvailable() &&
6293 "Unexpected register load without SVE load instructions");
6296 Opc = AArch64::LDR_PXI;
6302 if (AArch64::GPR32allRegClass.hasSubClassEq(RC)) {
6303 Opc = AArch64::LDRWui;
6307 assert(DestReg != AArch64::WSP);
6308 }
else if (AArch64::FPR32RegClass.hasSubClassEq(RC))
6309 Opc = AArch64::LDRSui;
6310 else if (AArch64::PPR2RegClass.hasSubClassEq(RC)) {
6311 Opc = AArch64::LDR_PPXI;
6316 if (AArch64::GPR64allRegClass.hasSubClassEq(RC)) {
6317 Opc = AArch64::LDRXui;
6321 assert(DestReg != AArch64::SP);
6322 }
else if (AArch64::FPR64RegClass.hasSubClassEq(RC)) {
6323 Opc = AArch64::LDRDui;
6324 }
else if (AArch64::WSeqPairsClassRegClass.hasSubClassEq(RC)) {
6326 get(AArch64::LDPWi), DestReg, AArch64::sube32,
6327 AArch64::subo32, FI, MMO);
6332 if (AArch64::FPR128RegClass.hasSubClassEq(RC))
6333 Opc = AArch64::LDRQui;
6334 else if (AArch64::DDRegClass.hasSubClassEq(RC)) {
6335 assert(Subtarget.hasNEON() &&
"Unexpected register load without NEON");
6336 Opc = AArch64::LD1Twov1d;
6338 }
else if (AArch64::XSeqPairsClassRegClass.hasSubClassEq(RC)) {
6340 get(AArch64::LDPXi), DestReg, AArch64::sube64,
6341 AArch64::subo64, FI, MMO);
6343 }
else if (AArch64::ZPRRegClass.hasSubClassEq(RC)) {
6344 assert(Subtarget.isSVEorStreamingSVEAvailable() &&
6345 "Unexpected register load without SVE load instructions");
6346 Opc = AArch64::LDR_ZXI;
6351 if (AArch64::DDDRegClass.hasSubClassEq(RC)) {
6352 assert(Subtarget.hasNEON() &&
"Unexpected register load without NEON");
6353 Opc = AArch64::LD1Threev1d;
6358 if (AArch64::DDDDRegClass.hasSubClassEq(RC)) {
6359 assert(Subtarget.hasNEON() &&
"Unexpected register load without NEON");
6360 Opc = AArch64::LD1Fourv1d;
6362 }
else if (AArch64::QQRegClass.hasSubClassEq(RC)) {
6363 assert(Subtarget.hasNEON() &&
"Unexpected register load without NEON");
6364 Opc = AArch64::LD1Twov2d;
6366 }
else if (AArch64::ZPR2StridedOrContiguousRegClass.hasSubClassEq(RC)) {
6367 assert(Subtarget.isSVEorStreamingSVEAvailable() &&
6368 "Unexpected register load without SVE load instructions");
6369 Opc = AArch64::LDR_ZZXI_STRIDED_CONTIGUOUS;
6371 }
else if (AArch64::ZPR2RegClass.hasSubClassEq(RC)) {
6372 assert(Subtarget.isSVEorStreamingSVEAvailable() &&
6373 "Unexpected register load without SVE load instructions");
6374 Opc = AArch64::LDR_ZZXI;
6379 if (AArch64::QQQRegClass.hasSubClassEq(RC)) {
6380 assert(Subtarget.hasNEON() &&
"Unexpected register load without NEON");
6381 Opc = AArch64::LD1Threev2d;
6383 }
else if (AArch64::ZPR3RegClass.hasSubClassEq(RC)) {
6384 assert(Subtarget.isSVEorStreamingSVEAvailable() &&
6385 "Unexpected register load without SVE load instructions");
6386 Opc = AArch64::LDR_ZZZXI;
6391 if (AArch64::QQQQRegClass.hasSubClassEq(RC)) {
6392 assert(Subtarget.hasNEON() &&
"Unexpected register load without NEON");
6393 Opc = AArch64::LD1Fourv2d;
6395 }
else if (AArch64::ZPR4StridedOrContiguousRegClass.hasSubClassEq(RC)) {
6396 assert(Subtarget.isSVEorStreamingSVEAvailable() &&
6397 "Unexpected register load without SVE load instructions");
6398 Opc = AArch64::LDR_ZZZZXI_STRIDED_CONTIGUOUS;
6400 }
else if (AArch64::ZPR4RegClass.hasSubClassEq(RC)) {
6401 assert(Subtarget.isSVEorStreamingSVEAvailable() &&
6402 "Unexpected register load without SVE load instructions");
6403 Opc = AArch64::LDR_ZZZZXI;
6409 assert(
Opc &&
"Unknown register class");
6419 MI.addMemOperand(MMO);
6426 UseMI.getIterator()),
6428 return I.modifiesRegister(AArch64::NZCV, TRI) ||
6429 I.readsRegister(AArch64::NZCV, TRI);
6433void AArch64InstrInfo::decomposeStackOffsetForDwarfOffsets(
6438 assert(
Offset.getScalable() % 2 == 0 &&
"Invalid frame offset");
6445 ByteSized =
Offset.getFixed();
6446 VGSized =
Offset.getScalable() / 2;
6452void AArch64InstrInfo::decomposeStackOffsetForFrameOffsets(
6454 int64_t &NumDataVectors) {
6458 assert(
Offset.getScalable() % 2 == 0 &&
"Invalid frame offset");
6460 NumBytes =
Offset.getFixed();
6462 NumPredicateVectors =
Offset.getScalable() / 2;
6467 if (NumPredicateVectors % 8 == 0 || NumPredicateVectors < -64 ||
6468 NumPredicateVectors > 62) {
6469 NumDataVectors = NumPredicateVectors / 8;
6470 NumPredicateVectors -= NumDataVectors * 8;
6496 Expr.
push_back((
char)dwarf::DW_OP_bregx);
6504 int64_t OffsetFromDefCFA) {
6518 Comment << (NumBytes < 0 ?
" - " :
" + ") << std::abs(NumBytes);
6519 if (!RegScale.empty())
6529 int64_t NumBytes, NumVGScaledBytes;
6530 AArch64InstrInfo::decomposeStackOffsetForDwarfOffsets(
Offset, NumBytes,
6532 std::string CommentBuffer;
6535 if (
Reg == AArch64::SP)
6537 else if (
Reg == AArch64::FP)
6544 unsigned DwarfReg =
TRI.getDwarfRegNum(
Reg,
true);
6545 assert(DwarfReg <= 31 &&
"DwarfReg out of bounds (0..31)");
6547 Expr.
push_back(dwarf::DW_OP_breg0 + DwarfReg);
6550 if (NumVGScaledBytes) {
6560 DefCfaExpr.
push_back(dwarf::DW_CFA_def_cfa_expression);
6568 unsigned FrameReg,
unsigned Reg,
6570 bool LastAdjustmentWasScalable) {
6571 if (
Offset.getScalable())
6574 if (FrameReg == Reg && !LastAdjustmentWasScalable)
6577 unsigned DwarfReg =
TRI.getDwarfRegNum(Reg,
true);
6584 std::optional<int64_t> IncomingVGOffsetFromDefCFA) {
6585 int64_t NumBytes, NumVGScaledBytes;
6586 AArch64InstrInfo::decomposeStackOffsetForDwarfOffsets(
6587 OffsetFromDefCFA, NumBytes, NumVGScaledBytes);
6589 unsigned DwarfReg =
TRI.getDwarfRegNum(Reg,
true);
6592 if (!NumVGScaledBytes)
6595 std::string CommentBuffer;
6600 assert(NumVGScaledBytes &&
"Expected scalable offset");
6604 if (IncomingVGOffsetFromDefCFA) {
6606 VGRegScale =
"* IncomingVG";
6609 VGRegScale =
"* VG";
6613 OffsetExpr.
push_back(dwarf::DW_OP_plus);
6622 CfaExpr.
push_back(dwarf::DW_CFA_expression);
6637 unsigned SrcReg, int64_t
Offset,
unsigned Opc,
6640 bool *HasWinCFI,
bool EmitCFAOffset,
6643 unsigned MaxEncoding, ShiftSize;
6645 case AArch64::ADDXri:
6646 case AArch64::ADDSXri:
6647 case AArch64::SUBXri:
6648 case AArch64::SUBSXri:
6649 MaxEncoding = 0xfff;
6652 case AArch64::ADDVL_XXI:
6653 case AArch64::ADDPL_XXI:
6654 case AArch64::ADDSVL_XXI:
6655 case AArch64::ADDSPL_XXI:
6670 if (
Opc == AArch64::ADDVL_XXI ||
Opc == AArch64::ADDSVL_XXI)
6672 else if (
Opc == AArch64::ADDPL_XXI ||
Opc == AArch64::ADDSPL_XXI)
6686 const unsigned MaxEncodableValue = MaxEncoding << ShiftSize;
6688 if (TmpReg == AArch64::XZR)
6689 TmpReg =
MBB.getParent()->getRegInfo().createVirtualRegister(
6690 &AArch64::GPR64RegClass);
6692 uint64_t ThisVal = std::min<uint64_t>(
Offset, MaxEncodableValue);
6693 unsigned LocalShiftSize = 0;
6694 if (ThisVal > MaxEncoding) {
6695 ThisVal = ThisVal >> ShiftSize;
6696 LocalShiftSize = ShiftSize;
6698 assert((ThisVal >> ShiftSize) <= MaxEncoding &&
6699 "Encoding cannot handle value that big");
6701 Offset -= ThisVal << LocalShiftSize;
6706 .
addImm(Sign * (
int)ThisVal);
6716 if (Sign == -1 ||
Opc == AArch64::SUBXri ||
Opc == AArch64::SUBSXri)
6717 CFAOffset += Change;
6719 CFAOffset -= Change;
6720 if (EmitCFAOffset && DestReg == TmpReg) {
6733 int Imm = (int)(ThisVal << LocalShiftSize);
6734 if (VScale != 1 && DestReg == AArch64::SP) {
6740 }
else if ((DestReg == AArch64::FP && SrcReg == AArch64::SP) ||
6741 (SrcReg == AArch64::FP && DestReg == AArch64::SP)) {
6742 assert(VScale == 1 &&
"Expected non-scalable operation");
6751 assert(
Offset == 0 &&
"Expected remaining offset to be zero to "
6752 "emit a single SEH directive");
6753 }
else if (DestReg == AArch64::SP) {
6754 assert(VScale == 1 &&
"Expected non-scalable operation");
6757 assert(SrcReg == AArch64::SP &&
"Unexpected SrcReg for SEH_StackAlloc");
6770 unsigned DestReg,
unsigned SrcReg,
6773 bool NeedsWinCFI,
bool *HasWinCFI,
6775 unsigned FrameReg) {
6782 bool UseSVL =
F.hasFnAttribute(
"aarch64_pstate_sm_body");
6784 int64_t Bytes, NumPredicateVectors, NumDataVectors;
6785 AArch64InstrInfo::decomposeStackOffsetForFrameOffsets(
6786 Offset, Bytes, NumPredicateVectors, NumDataVectors);
6789 bool NeedsFinalDefNZCV = SetNZCV && (NumPredicateVectors || NumDataVectors);
6790 if (NeedsFinalDefNZCV)
6794 if (Bytes || (!
Offset && SrcReg != DestReg)) {
6795 assert((DestReg != AArch64::SP || Bytes % 8 == 0) &&
6796 "SP increment/decrement not 8-byte aligned");
6797 unsigned Opc = SetNZCV ? AArch64::ADDSXri : AArch64::ADDXri;
6800 Opc = SetNZCV ? AArch64::SUBSXri : AArch64::SUBXri;
6803 NeedsWinCFI, HasWinCFI, EmitCFAOffset, CFAOffset,
6805 CFAOffset += (
Opc == AArch64::ADDXri ||
Opc == AArch64::ADDSXri)
6812 assert(!(NeedsWinCFI && NumPredicateVectors) &&
6813 "WinCFI can't allocate fractions of an SVE data vector");
6815 if (NumDataVectors) {
6817 UseSVL ? AArch64::ADDSVL_XXI : AArch64::ADDVL_XXI,
TII,
6818 Flag, NeedsWinCFI, HasWinCFI, EmitCFAOffset, CFAOffset,
6824 if (NumPredicateVectors) {
6825 assert(DestReg != AArch64::SP &&
"Unaligned access to SP");
6827 UseSVL ? AArch64::ADDSPL_XXI : AArch64::ADDPL_XXI,
TII,
6828 Flag, NeedsWinCFI, HasWinCFI, EmitCFAOffset, CFAOffset,
6832 if (NeedsFinalDefNZCV)
6854 if (
MI.isFullCopy()) {
6857 if (SrcReg == AArch64::SP && DstReg.
isVirtual()) {
6861 if (DstReg == AArch64::SP && SrcReg.
isVirtual()) {
6866 if (SrcReg == AArch64::NZCV || DstReg == AArch64::NZCV)
6894 if (
MI.isCopy() &&
Ops.size() == 1 &&
6896 (
Ops[0] == 0 ||
Ops[0] == 1)) {
6897 bool IsSpill =
Ops[0] == 0;
6898 bool IsFill = !IsSpill;
6910 :
TRI.getMinimalPhysRegClass(Reg);
6916 "Mismatched register size in non subreg COPY");
6923 return &*--InsertPt;
6935 if (IsSpill && DstMO.
isUndef() && SrcReg == AArch64::WZR &&
6938 "Unexpected subreg on physical register");
6940 FrameIndex, &AArch64::GPR64RegClass,
Register());
6941 return &*--InsertPt;
6958 case AArch64::sub_32:
6959 if (AArch64::GPR64RegClass.hasSubClassEq(
getRegClass(DstReg)))
6960 FillRC = &AArch64::GPR32RegClass;
6963 FillRC = &AArch64::FPR32RegClass;
6966 FillRC = &AArch64::FPR64RegClass;
6972 TRI.getRegSizeInBits(*FillRC) &&
6973 "Mismatched regclass size on folded subreg COPY");
6992 bool *OutUseUnscaledOp,
6993 unsigned *OutUnscaledOp,
6994 int64_t *EmittableOffset) {
6996 if (EmittableOffset)
6997 *EmittableOffset = 0;
6998 if (OutUseUnscaledOp)
6999 *OutUseUnscaledOp =
false;
7005 switch (
MI.getOpcode()) {
7008 case AArch64::LD1Rv1d:
7009 case AArch64::LD1Rv2s:
7010 case AArch64::LD1Rv2d:
7011 case AArch64::LD1Rv4h:
7012 case AArch64::LD1Rv4s:
7013 case AArch64::LD1Rv8b:
7014 case AArch64::LD1Rv8h:
7015 case AArch64::LD1Rv16b:
7016 case AArch64::LD1Twov2d:
7017 case AArch64::LD1Threev2d:
7018 case AArch64::LD1Fourv2d:
7019 case AArch64::LD1Twov1d:
7020 case AArch64::LD1Threev1d:
7021 case AArch64::LD1Fourv1d:
7022 case AArch64::ST1Twov2d:
7023 case AArch64::ST1Threev2d:
7024 case AArch64::ST1Fourv2d:
7025 case AArch64::ST1Twov1d:
7026 case AArch64::ST1Threev1d:
7027 case AArch64::ST1Fourv1d:
7028 case AArch64::ST1i8:
7029 case AArch64::ST1i16:
7030 case AArch64::ST1i32:
7031 case AArch64::ST1i64:
7033 case AArch64::IRGstack:
7034 case AArch64::STGloop:
7035 case AArch64::STZGloop:
7040 TypeSize ScaleValue(0U,
false), Width(0U,
false);
7041 int64_t MinOff, MaxOff;
7047 bool IsMulVL = ScaleValue.isScalable();
7048 unsigned Scale = ScaleValue.getKnownMinValue();
7058 std::optional<unsigned> UnscaledOp =
7060 bool useUnscaledOp = UnscaledOp && (
Offset % Scale ||
Offset < 0);
7061 if (useUnscaledOp &&
7066 Scale = ScaleValue.getKnownMinValue();
7067 assert(IsMulVL == ScaleValue.isScalable() &&
7068 "Unscaled opcode has different value for scalable");
7070 int64_t Remainder =
Offset % Scale;
7071 assert(!(Remainder && useUnscaledOp) &&
7072 "Cannot have remainder when using unscaled op");
7074 assert(MinOff < MaxOff &&
"Unexpected Min/Max offsets");
7075 int64_t NewOffset =
Offset / Scale;
7076 if (MinOff <= NewOffset && NewOffset <= MaxOff)
7079 NewOffset = NewOffset < 0 ? MinOff : MaxOff;
7083 if (EmittableOffset)
7084 *EmittableOffset = NewOffset;
7085 if (OutUseUnscaledOp)
7086 *OutUseUnscaledOp = useUnscaledOp;
7087 if (OutUnscaledOp && UnscaledOp)
7088 *OutUnscaledOp = *UnscaledOp;
7101 unsigned Opcode =
MI.getOpcode();
7102 unsigned ImmIdx = FrameRegIdx + 1;
7104 if (Opcode == AArch64::ADDSXri || Opcode == AArch64::ADDXri) {
7109 MI.eraseFromParent();
7115 unsigned UnscaledOp;
7118 &UnscaledOp, &NewOffset);
7122 MI.getOperand(FrameRegIdx).ChangeToRegister(FrameReg,
false);
7124 MI.setDesc(
TII->get(UnscaledOp));
7126 MI.getOperand(ImmIdx).ChangeToImmediate(NewOffset);
7142bool AArch64InstrInfo::useMachineCombiner()
const {
return true; }
7147 case AArch64::ADDSWrr:
7148 case AArch64::ADDSWri:
7149 case AArch64::ADDSXrr:
7150 case AArch64::ADDSXri:
7151 case AArch64::SUBSWrr:
7152 case AArch64::SUBSXrr:
7154 case AArch64::SUBSWri:
7155 case AArch64::SUBSXri:
7166 case AArch64::ADDWrr:
7167 case AArch64::ADDWri:
7168 case AArch64::SUBWrr:
7169 case AArch64::ADDSWrr:
7170 case AArch64::ADDSWri:
7171 case AArch64::SUBSWrr:
7173 case AArch64::SUBWri:
7174 case AArch64::SUBSWri:
7185 case AArch64::ADDXrr:
7186 case AArch64::ADDXri:
7187 case AArch64::SUBXrr:
7188 case AArch64::ADDSXrr:
7189 case AArch64::ADDSXri:
7190 case AArch64::SUBSXrr:
7192 case AArch64::SUBXri:
7193 case AArch64::SUBSXri:
7194 case AArch64::ADDv8i8:
7195 case AArch64::ADDv16i8:
7196 case AArch64::ADDv4i16:
7197 case AArch64::ADDv8i16:
7198 case AArch64::ADDv2i32:
7199 case AArch64::ADDv4i32:
7200 case AArch64::SUBv8i8:
7201 case AArch64::SUBv16i8:
7202 case AArch64::SUBv4i16:
7203 case AArch64::SUBv8i16:
7204 case AArch64::SUBv2i32:
7205 case AArch64::SUBv4i32:
7218 case AArch64::FADDHrr:
7219 case AArch64::FADDSrr:
7220 case AArch64::FADDDrr:
7221 case AArch64::FADDv4f16:
7222 case AArch64::FADDv8f16:
7223 case AArch64::FADDv2f32:
7224 case AArch64::FADDv2f64:
7225 case AArch64::FADDv4f32:
7226 case AArch64::FSUBHrr:
7227 case AArch64::FSUBSrr:
7228 case AArch64::FSUBDrr:
7229 case AArch64::FSUBv4f16:
7230 case AArch64::FSUBv8f16:
7231 case AArch64::FSUBv2f32:
7232 case AArch64::FSUBv2f64:
7233 case AArch64::FSUBv4f32:
7252 unsigned CombineOpc,
unsigned ZeroReg = 0,
7253 bool CheckZeroReg =
false) {
7260 if (!
MI ||
MI->getParent() != &
MBB ||
MI->getOpcode() != CombineOpc)
7267 assert(
MI->getNumOperands() >= 4 &&
MI->getOperand(0).isReg() &&
7268 MI->getOperand(1).isReg() &&
MI->getOperand(2).isReg() &&
7269 MI->getOperand(3).isReg() &&
"MAdd/MSub must have a least 4 regs");
7271 if (
MI->getOperand(3).getReg() != ZeroReg)
7276 MI->findRegisterDefOperandIdx(AArch64::NZCV,
nullptr,
true) == -1)
7285 unsigned MulOpc,
unsigned ZeroReg) {
7300bool AArch64InstrInfo::isAssociativeAndCommutative(
const MachineInstr &Inst,
7301 bool Invert)
const {
7307 case AArch64::FADDHrr:
7308 case AArch64::FADDSrr:
7309 case AArch64::FADDDrr:
7310 case AArch64::FMULHrr:
7311 case AArch64::FMULSrr:
7312 case AArch64::FMULDrr:
7313 case AArch64::FMULX16:
7314 case AArch64::FMULX32:
7315 case AArch64::FMULX64:
7317 case AArch64::FADDv4f16:
7318 case AArch64::FADDv8f16:
7319 case AArch64::FADDv2f32:
7320 case AArch64::FADDv4f32:
7321 case AArch64::FADDv2f64:
7322 case AArch64::FMULv4f16:
7323 case AArch64::FMULv8f16:
7324 case AArch64::FMULv2f32:
7325 case AArch64::FMULv4f32:
7326 case AArch64::FMULv2f64:
7327 case AArch64::FMULXv4f16:
7328 case AArch64::FMULXv8f16:
7329 case AArch64::FMULXv2f32:
7330 case AArch64::FMULXv4f32:
7331 case AArch64::FMULXv2f64:
7335 case AArch64::FADD_ZZZ_H:
7336 case AArch64::FADD_ZZZ_S:
7337 case AArch64::FADD_ZZZ_D:
7338 case AArch64::FMUL_ZZZ_H:
7339 case AArch64::FMUL_ZZZ_S:
7340 case AArch64::FMUL_ZZZ_D:
7351 case AArch64::ADDWrr:
7352 case AArch64::ADDXrr:
7353 case AArch64::ANDWrr:
7354 case AArch64::ANDXrr:
7355 case AArch64::ORRWrr:
7356 case AArch64::ORRXrr:
7357 case AArch64::EORWrr:
7358 case AArch64::EORXrr:
7359 case AArch64::EONWrr:
7360 case AArch64::EONXrr:
7364 case AArch64::ADDv8i8:
7365 case AArch64::ADDv16i8:
7366 case AArch64::ADDv4i16:
7367 case AArch64::ADDv8i16:
7368 case AArch64::ADDv2i32:
7369 case AArch64::ADDv4i32:
7370 case AArch64::ADDv1i64:
7371 case AArch64::ADDv2i64:
7372 case AArch64::MULv8i8:
7373 case AArch64::MULv16i8:
7374 case AArch64::MULv4i16:
7375 case AArch64::MULv8i16:
7376 case AArch64::MULv2i32:
7377 case AArch64::MULv4i32:
7378 case AArch64::ANDv8i8:
7379 case AArch64::ANDv16i8:
7380 case AArch64::ORRv8i8:
7381 case AArch64::ORRv16i8:
7382 case AArch64::EORv8i8:
7383 case AArch64::EORv16i8:
7385 case AArch64::ADD_ZZZ_B:
7386 case AArch64::ADD_ZZZ_H:
7387 case AArch64::ADD_ZZZ_S:
7388 case AArch64::ADD_ZZZ_D:
7389 case AArch64::MUL_ZZZ_B:
7390 case AArch64::MUL_ZZZ_H:
7391 case AArch64::MUL_ZZZ_S:
7392 case AArch64::MUL_ZZZ_D:
7393 case AArch64::AND_ZZZ:
7394 case AArch64::ORR_ZZZ:
7395 case AArch64::EOR_ZZZ:
7426 auto setFound = [&](
int Opcode,
int Operand,
unsigned ZeroReg,
7434 auto setVFound = [&](
int Opcode,
int Operand,
unsigned Pattern) {
7446 case AArch64::ADDWrr:
7448 "ADDWrr does not have register operands");
7449 setFound(AArch64::MADDWrrr, 1, AArch64::WZR, MCP::MULADDW_OP1);
7450 setFound(AArch64::MADDWrrr, 2, AArch64::WZR, MCP::MULADDW_OP2);
7452 case AArch64::ADDXrr:
7453 setFound(AArch64::MADDXrrr, 1, AArch64::XZR, MCP::MULADDX_OP1);
7454 setFound(AArch64::MADDXrrr, 2, AArch64::XZR, MCP::MULADDX_OP2);
7456 case AArch64::SUBWrr:
7457 setFound(AArch64::MADDWrrr, 2, AArch64::WZR, MCP::MULSUBW_OP2);
7458 setFound(AArch64::MADDWrrr, 1, AArch64::WZR, MCP::MULSUBW_OP1);
7460 case AArch64::SUBXrr:
7461 setFound(AArch64::MADDXrrr, 2, AArch64::XZR, MCP::MULSUBX_OP2);
7462 setFound(AArch64::MADDXrrr, 1, AArch64::XZR, MCP::MULSUBX_OP1);
7464 case AArch64::ADDWri:
7465 setFound(AArch64::MADDWrrr, 1, AArch64::WZR, MCP::MULADDWI_OP1);
7467 case AArch64::ADDXri:
7468 setFound(AArch64::MADDXrrr, 1, AArch64::XZR, MCP::MULADDXI_OP1);
7470 case AArch64::SUBWri:
7471 setFound(AArch64::MADDWrrr, 1, AArch64::WZR, MCP::MULSUBWI_OP1);
7473 case AArch64::SUBXri:
7474 setFound(AArch64::MADDXrrr, 1, AArch64::XZR, MCP::MULSUBXI_OP1);
7476 case AArch64::ADDv8i8:
7477 setVFound(AArch64::MULv8i8, 1, MCP::MULADDv8i8_OP1);
7478 setVFound(AArch64::MULv8i8, 2, MCP::MULADDv8i8_OP2);
7480 case AArch64::ADDv16i8:
7481 setVFound(AArch64::MULv16i8, 1, MCP::MULADDv16i8_OP1);
7482 setVFound(AArch64::MULv16i8, 2, MCP::MULADDv16i8_OP2);
7484 case AArch64::ADDv4i16:
7485 setVFound(AArch64::MULv4i16, 1, MCP::MULADDv4i16_OP1);
7486 setVFound(AArch64::MULv4i16, 2, MCP::MULADDv4i16_OP2);
7487 setVFound(AArch64::MULv4i16_indexed, 1, MCP::MULADDv4i16_indexed_OP1);
7488 setVFound(AArch64::MULv4i16_indexed, 2, MCP::MULADDv4i16_indexed_OP2);
7490 case AArch64::ADDv8i16:
7491 setVFound(AArch64::MULv8i16, 1, MCP::MULADDv8i16_OP1);
7492 setVFound(AArch64::MULv8i16, 2, MCP::MULADDv8i16_OP2);
7493 setVFound(AArch64::MULv8i16_indexed, 1, MCP::MULADDv8i16_indexed_OP1);
7494 setVFound(AArch64::MULv8i16_indexed, 2, MCP::MULADDv8i16_indexed_OP2);
7496 case AArch64::ADDv2i32:
7497 setVFound(AArch64::MULv2i32, 1, MCP::MULADDv2i32_OP1);
7498 setVFound(AArch64::MULv2i32, 2, MCP::MULADDv2i32_OP2);
7499 setVFound(AArch64::MULv2i32_indexed, 1, MCP::MULADDv2i32_indexed_OP1);
7500 setVFound(AArch64::MULv2i32_indexed, 2, MCP::MULADDv2i32_indexed_OP2);
7502 case AArch64::ADDv4i32:
7503 setVFound(AArch64::MULv4i32, 1, MCP::MULADDv4i32_OP1);
7504 setVFound(AArch64::MULv4i32, 2, MCP::MULADDv4i32_OP2);
7505 setVFound(AArch64::MULv4i32_indexed, 1, MCP::MULADDv4i32_indexed_OP1);
7506 setVFound(AArch64::MULv4i32_indexed, 2, MCP::MULADDv4i32_indexed_OP2);
7508 case AArch64::SUBv8i8:
7509 setVFound(AArch64::MULv8i8, 1, MCP::MULSUBv8i8_OP1);
7510 setVFound(AArch64::MULv8i8, 2, MCP::MULSUBv8i8_OP2);
7512 case AArch64::SUBv16i8:
7513 setVFound(AArch64::MULv16i8, 1, MCP::MULSUBv16i8_OP1);
7514 setVFound(AArch64::MULv16i8, 2, MCP::MULSUBv16i8_OP2);
7516 case AArch64::SUBv4i16:
7517 setVFound(AArch64::MULv4i16, 1, MCP::MULSUBv4i16_OP1);
7518 setVFound(AArch64::MULv4i16, 2, MCP::MULSUBv4i16_OP2);
7519 setVFound(AArch64::MULv4i16_indexed, 1, MCP::MULSUBv4i16_indexed_OP1);
7520 setVFound(AArch64::MULv4i16_indexed, 2, MCP::MULSUBv4i16_indexed_OP2);
7522 case AArch64::SUBv8i16:
7523 setVFound(AArch64::MULv8i16, 1, MCP::MULSUBv8i16_OP1);
7524 setVFound(AArch64::MULv8i16, 2, MCP::MULSUBv8i16_OP2);
7525 setVFound(AArch64::MULv8i16_indexed, 1, MCP::MULSUBv8i16_indexed_OP1);
7526 setVFound(AArch64::MULv8i16_indexed, 2, MCP::MULSUBv8i16_indexed_OP2);
7528 case AArch64::SUBv2i32:
7529 setVFound(AArch64::MULv2i32, 1, MCP::MULSUBv2i32_OP1);
7530 setVFound(AArch64::MULv2i32, 2, MCP::MULSUBv2i32_OP2);
7531 setVFound(AArch64::MULv2i32_indexed, 1, MCP::MULSUBv2i32_indexed_OP1);
7532 setVFound(AArch64::MULv2i32_indexed, 2, MCP::MULSUBv2i32_indexed_OP2);
7534 case AArch64::SUBv4i32:
7535 setVFound(AArch64::MULv4i32, 1, MCP::MULSUBv4i32_OP1);
7536 setVFound(AArch64::MULv4i32, 2, MCP::MULSUBv4i32_OP2);
7537 setVFound(AArch64::MULv4i32_indexed, 1, MCP::MULSUBv4i32_indexed_OP1);
7538 setVFound(AArch64::MULv4i32_indexed, 2, MCP::MULSUBv4i32_indexed_OP2);
7544bool AArch64InstrInfo::isAccumulationOpcode(
unsigned Opcode)
const {
7548 case AArch64::UABALB_ZZZ_D:
7549 case AArch64::UABALB_ZZZ_H:
7550 case AArch64::UABALB_ZZZ_S:
7551 case AArch64::UABALT_ZZZ_D:
7552 case AArch64::UABALT_ZZZ_H:
7553 case AArch64::UABALT_ZZZ_S:
7554 case AArch64::SABALB_ZZZ_D:
7555 case AArch64::SABALB_ZZZ_S:
7556 case AArch64::SABALB_ZZZ_H:
7557 case AArch64::SABALT_ZZZ_D:
7558 case AArch64::SABALT_ZZZ_S:
7559 case AArch64::SABALT_ZZZ_H:
7560 case AArch64::UABALv16i8_v8i16:
7561 case AArch64::UABALv2i32_v2i64:
7562 case AArch64::UABALv4i16_v4i32:
7563 case AArch64::UABALv4i32_v2i64:
7564 case AArch64::UABALv8i16_v4i32:
7565 case AArch64::UABALv8i8_v8i16:
7566 case AArch64::UABAv16i8:
7567 case AArch64::UABAv2i32:
7568 case AArch64::UABAv4i16:
7569 case AArch64::UABAv4i32:
7570 case AArch64::UABAv8i16:
7571 case AArch64::UABAv8i8:
7572 case AArch64::SABALv16i8_v8i16:
7573 case AArch64::SABALv2i32_v2i64:
7574 case AArch64::SABALv4i16_v4i32:
7575 case AArch64::SABALv4i32_v2i64:
7576 case AArch64::SABALv8i16_v4i32:
7577 case AArch64::SABALv8i8_v8i16:
7578 case AArch64::SABAv16i8:
7579 case AArch64::SABAv2i32:
7580 case AArch64::SABAv4i16:
7581 case AArch64::SABAv4i32:
7582 case AArch64::SABAv8i16:
7583 case AArch64::SABAv8i8:
7590unsigned AArch64InstrInfo::getAccumulationStartOpcode(
7591 unsigned AccumulationOpcode)
const {
7592 switch (AccumulationOpcode) {
7595 case AArch64::UABALB_ZZZ_D:
7596 return AArch64::UABDLB_ZZZ_D;
7597 case AArch64::UABALB_ZZZ_H:
7598 return AArch64::UABDLB_ZZZ_H;
7599 case AArch64::UABALB_ZZZ_S:
7600 return AArch64::UABDLB_ZZZ_S;
7601 case AArch64::UABALT_ZZZ_D:
7602 return AArch64::UABDLT_ZZZ_D;
7603 case AArch64::UABALT_ZZZ_H:
7604 return AArch64::UABDLT_ZZZ_H;
7605 case AArch64::UABALT_ZZZ_S:
7606 return AArch64::UABDLT_ZZZ_S;
7607 case AArch64::UABALv16i8_v8i16:
7608 return AArch64::UABDLv16i8_v8i16;
7609 case AArch64::UABALv2i32_v2i64:
7610 return AArch64::UABDLv2i32_v2i64;
7611 case AArch64::UABALv4i16_v4i32:
7612 return AArch64::UABDLv4i16_v4i32;
7613 case AArch64::UABALv4i32_v2i64:
7614 return AArch64::UABDLv4i32_v2i64;
7615 case AArch64::UABALv8i16_v4i32:
7616 return AArch64::UABDLv8i16_v4i32;
7617 case AArch64::UABALv8i8_v8i16:
7618 return AArch64::UABDLv8i8_v8i16;
7619 case AArch64::UABAv16i8:
7620 return AArch64::UABDv16i8;
7621 case AArch64::UABAv2i32:
7622 return AArch64::UABDv2i32;
7623 case AArch64::UABAv4i16:
7624 return AArch64::UABDv4i16;
7625 case AArch64::UABAv4i32:
7626 return AArch64::UABDv4i32;
7627 case AArch64::UABAv8i16:
7628 return AArch64::UABDv8i16;
7629 case AArch64::UABAv8i8:
7630 return AArch64::UABDv8i8;
7631 case AArch64::SABALB_ZZZ_D:
7632 return AArch64::SABDLB_ZZZ_D;
7633 case AArch64::SABALB_ZZZ_S:
7634 return AArch64::SABDLB_ZZZ_S;
7635 case AArch64::SABALB_ZZZ_H:
7636 return AArch64::SABDLB_ZZZ_H;
7637 case AArch64::SABALT_ZZZ_D:
7638 return AArch64::SABDLT_ZZZ_D;
7639 case AArch64::SABALT_ZZZ_S:
7640 return AArch64::SABDLT_ZZZ_S;
7641 case AArch64::SABALT_ZZZ_H:
7642 return AArch64::SABDLT_ZZZ_H;
7643 case AArch64::SABALv16i8_v8i16:
7644 return AArch64::SABDLv16i8_v8i16;
7645 case AArch64::SABALv2i32_v2i64:
7646 return AArch64::SABDLv2i32_v2i64;
7647 case AArch64::SABALv4i16_v4i32:
7648 return AArch64::SABDLv4i16_v4i32;
7649 case AArch64::SABALv4i32_v2i64:
7650 return AArch64::SABDLv4i32_v2i64;
7651 case AArch64::SABALv8i16_v4i32:
7652 return AArch64::SABDLv8i16_v4i32;
7653 case AArch64::SABALv8i8_v8i16:
7654 return AArch64::SABDLv8i8_v8i16;
7655 case AArch64::SABAv16i8:
7656 return AArch64::SABDv16i8;
7657 case AArch64::SABAv2i32:
7658 return AArch64::SABAv2i32;
7659 case AArch64::SABAv4i16:
7660 return AArch64::SABDv4i16;
7661 case AArch64::SABAv4i32:
7662 return AArch64::SABDv4i32;
7663 case AArch64::SABAv8i16:
7664 return AArch64::SABDv8i16;
7665 case AArch64::SABAv8i8:
7666 return AArch64::SABDv8i8;
7682 auto Match = [&](
int Opcode,
int Operand,
unsigned Pattern) ->
bool {
7694 assert(
false &&
"Unsupported FP instruction in combiner\n");
7696 case AArch64::FADDHrr:
7698 "FADDHrr does not have register operands");
7700 Found = Match(AArch64::FMULHrr, 1, MCP::FMULADDH_OP1);
7701 Found |= Match(AArch64::FMULHrr, 2, MCP::FMULADDH_OP2);
7703 case AArch64::FADDSrr:
7705 "FADDSrr does not have register operands");
7707 Found |= Match(AArch64::FMULSrr, 1, MCP::FMULADDS_OP1) ||
7708 Match(AArch64::FMULv1i32_indexed, 1, MCP::FMLAv1i32_indexed_OP1);
7710 Found |= Match(AArch64::FMULSrr, 2, MCP::FMULADDS_OP2) ||
7711 Match(AArch64::FMULv1i32_indexed, 2, MCP::FMLAv1i32_indexed_OP2);
7713 case AArch64::FADDDrr:
7714 Found |= Match(AArch64::FMULDrr, 1, MCP::FMULADDD_OP1) ||
7715 Match(AArch64::FMULv1i64_indexed, 1, MCP::FMLAv1i64_indexed_OP1);
7717 Found |= Match(AArch64::FMULDrr, 2, MCP::FMULADDD_OP2) ||
7718 Match(AArch64::FMULv1i64_indexed, 2, MCP::FMLAv1i64_indexed_OP2);
7720 case AArch64::FADDv4f16:
7721 Found |= Match(AArch64::FMULv4i16_indexed, 1, MCP::FMLAv4i16_indexed_OP1) ||
7722 Match(AArch64::FMULv4f16, 1, MCP::FMLAv4f16_OP1);
7724 Found |= Match(AArch64::FMULv4i16_indexed, 2, MCP::FMLAv4i16_indexed_OP2) ||
7725 Match(AArch64::FMULv4f16, 2, MCP::FMLAv4f16_OP2);
7727 case AArch64::FADDv8f16:
7728 Found |= Match(AArch64::FMULv8i16_indexed, 1, MCP::FMLAv8i16_indexed_OP1) ||
7729 Match(AArch64::FMULv8f16, 1, MCP::FMLAv8f16_OP1);
7731 Found |= Match(AArch64::FMULv8i16_indexed, 2, MCP::FMLAv8i16_indexed_OP2) ||
7732 Match(AArch64::FMULv8f16, 2, MCP::FMLAv8f16_OP2);
7734 case AArch64::FADDv2f32:
7735 Found |= Match(AArch64::FMULv2i32_indexed, 1, MCP::FMLAv2i32_indexed_OP1) ||
7736 Match(AArch64::FMULv2f32, 1, MCP::FMLAv2f32_OP1);
7738 Found |= Match(AArch64::FMULv2i32_indexed, 2, MCP::FMLAv2i32_indexed_OP2) ||
7739 Match(AArch64::FMULv2f32, 2, MCP::FMLAv2f32_OP2);
7741 case AArch64::FADDv2f64:
7742 Found |= Match(AArch64::FMULv2i64_indexed, 1, MCP::FMLAv2i64_indexed_OP1) ||
7743 Match(AArch64::FMULv2f64, 1, MCP::FMLAv2f64_OP1);
7745 Found |= Match(AArch64::FMULv2i64_indexed, 2, MCP::FMLAv2i64_indexed_OP2) ||
7746 Match(AArch64::FMULv2f64, 2, MCP::FMLAv2f64_OP2);
7748 case AArch64::FADDv4f32:
7749 Found |= Match(AArch64::FMULv4i32_indexed, 1, MCP::FMLAv4i32_indexed_OP1) ||
7750 Match(AArch64::FMULv4f32, 1, MCP::FMLAv4f32_OP1);
7752 Found |= Match(AArch64::FMULv4i32_indexed, 2, MCP::FMLAv4i32_indexed_OP2) ||
7753 Match(AArch64::FMULv4f32, 2, MCP::FMLAv4f32_OP2);
7755 case AArch64::FSUBHrr:
7756 Found = Match(AArch64::FMULHrr, 1, MCP::FMULSUBH_OP1);
7757 Found |= Match(AArch64::FMULHrr, 2, MCP::FMULSUBH_OP2);
7758 Found |= Match(AArch64::FNMULHrr, 1, MCP::FNMULSUBH_OP1);
7760 case AArch64::FSUBSrr:
7761 Found = Match(AArch64::FMULSrr, 1, MCP::FMULSUBS_OP1);
7763 Found |= Match(AArch64::FMULSrr, 2, MCP::FMULSUBS_OP2) ||
7764 Match(AArch64::FMULv1i32_indexed, 2, MCP::FMLSv1i32_indexed_OP2);
7766 Found |= Match(AArch64::FNMULSrr, 1, MCP::FNMULSUBS_OP1);
7768 case AArch64::FSUBDrr:
7769 Found = Match(AArch64::FMULDrr, 1, MCP::FMULSUBD_OP1);
7771 Found |= Match(AArch64::FMULDrr, 2, MCP::FMULSUBD_OP2) ||
7772 Match(AArch64::FMULv1i64_indexed, 2, MCP::FMLSv1i64_indexed_OP2);
7774 Found |= Match(AArch64::FNMULDrr, 1, MCP::FNMULSUBD_OP1);
7776 case AArch64::FSUBv4f16:
7777 Found |= Match(AArch64::FMULv4i16_indexed, 2, MCP::FMLSv4i16_indexed_OP2) ||
7778 Match(AArch64::FMULv4f16, 2, MCP::FMLSv4f16_OP2);
7780 Found |= Match(AArch64::FMULv4i16_indexed, 1, MCP::FMLSv4i16_indexed_OP1) ||
7781 Match(AArch64::FMULv4f16, 1, MCP::FMLSv4f16_OP1);
7783 case AArch64::FSUBv8f16:
7784 Found |= Match(AArch64::FMULv8i16_indexed, 2, MCP::FMLSv8i16_indexed_OP2) ||
7785 Match(AArch64::FMULv8f16, 2, MCP::FMLSv8f16_OP2);
7787 Found |= Match(AArch64::FMULv8i16_indexed, 1, MCP::FMLSv8i16_indexed_OP1) ||
7788 Match(AArch64::FMULv8f16, 1, MCP::FMLSv8f16_OP1);
7790 case AArch64::FSUBv2f32:
7791 Found |= Match(AArch64::FMULv2i32_indexed, 2, MCP::FMLSv2i32_indexed_OP2) ||
7792 Match(AArch64::FMULv2f32, 2, MCP::FMLSv2f32_OP2);
7794 Found |= Match(AArch64::FMULv2i32_indexed, 1, MCP::FMLSv2i32_indexed_OP1) ||
7795 Match(AArch64::FMULv2f32, 1, MCP::FMLSv2f32_OP1);
7797 case AArch64::FSUBv2f64:
7798 Found |= Match(AArch64::FMULv2i64_indexed, 2, MCP::FMLSv2i64_indexed_OP2) ||
7799 Match(AArch64::FMULv2f64, 2, MCP::FMLSv2f64_OP2);
7801 Found |= Match(AArch64::FMULv2i64_indexed, 1, MCP::FMLSv2i64_indexed_OP1) ||
7802 Match(AArch64::FMULv2f64, 1, MCP::FMLSv2f64_OP1);
7804 case AArch64::FSUBv4f32:
7805 Found |= Match(AArch64::FMULv4i32_indexed, 2, MCP::FMLSv4i32_indexed_OP2) ||
7806 Match(AArch64::FMULv4f32, 2, MCP::FMLSv4f32_OP2);
7808 Found |= Match(AArch64::FMULv4i32_indexed, 1, MCP::FMLSv4i32_indexed_OP1) ||
7809 Match(AArch64::FMULv4f32, 1, MCP::FMLSv4f32_OP1);
7820 auto Match = [&](
unsigned Opcode,
int Operand,
unsigned Pattern) ->
bool {
7827 if (
MI &&
MI->getOpcode() == TargetOpcode::COPY &&
7828 MI->getOperand(1).getReg().isVirtual())
7830 if (
MI &&
MI->getOpcode() == Opcode) {
7842 case AArch64::FMULv2f32:
7843 Found = Match(AArch64::DUPv2i32lane, 1, MCP::FMULv2i32_indexed_OP1);
7844 Found |= Match(AArch64::DUPv2i32lane, 2, MCP::FMULv2i32_indexed_OP2);
7846 case AArch64::FMULv2f64:
7847 Found = Match(AArch64::DUPv2i64lane, 1, MCP::FMULv2i64_indexed_OP1);
7848 Found |= Match(AArch64::DUPv2i64lane, 2, MCP::FMULv2i64_indexed_OP2);
7850 case AArch64::FMULv4f16:
7851 Found = Match(AArch64::DUPv4i16lane, 1, MCP::FMULv4i16_indexed_OP1);
7852 Found |= Match(AArch64::DUPv4i16lane, 2, MCP::FMULv4i16_indexed_OP2);
7854 case AArch64::FMULv4f32:
7855 Found = Match(AArch64::DUPv4i32lane, 1, MCP::FMULv4i32_indexed_OP1);
7856 Found |= Match(AArch64::DUPv4i32lane, 2, MCP::FMULv4i32_indexed_OP2);
7858 case AArch64::FMULv8f16:
7859 Found = Match(AArch64::DUPv8i16lane, 1, MCP::FMULv8i16_indexed_OP1);
7860 Found |= Match(AArch64::DUPv8i16lane, 2, MCP::FMULv8i16_indexed_OP2);
7873 auto Match = [&](
unsigned Opcode,
unsigned Pattern) ->
bool {
7876 if (
MI !=
nullptr && (
MI->getOpcode() == Opcode) &&
7891 case AArch64::FNEGDr:
7893 case AArch64::FNEGSr:
8025 case AArch64::SUBWrr:
8026 case AArch64::SUBSWrr:
8027 case AArch64::SUBXrr:
8028 case AArch64::SUBSXrr:
8073 unsigned LoadLaneOpCode,
unsigned NumLanes) {
8096 while (!RemainingLanes.
empty() && CurrInstr &&
8097 CurrInstr->getOpcode() == LoadLaneOpCode &&
8099 CurrInstr->getNumOperands() == 4) {
8100 RemainingLanes.
erase(CurrInstr->getOperand(2).getImm());
8106 if (!RemainingLanes.
empty())
8110 if (CurrInstr->getOpcode() != TargetOpcode::SUBREG_TO_REG)
8114 auto Lane0LoadReg = CurrInstr->getOperand(1).getReg();
8115 unsigned SingleLaneSizeInBits = 128 / NumLanes;
8116 if (
TRI->getRegSizeInBits(Lane0LoadReg, MRI) != SingleLaneSizeInBits)
8132 RemainingLoadInstrs.
insert(LoadInstrs.
begin(), LoadInstrs.
end());
8135 for (; MBBItr !=
MBB->begin() && RemainingSteps > 0 &&
8136 !RemainingLoadInstrs.
empty();
8137 --MBBItr, --RemainingSteps) {
8141 RemainingLoadInstrs.
erase(&CurrInstr);
8151 if (RemainingSteps == 0 && !RemainingLoadInstrs.
empty())
8177 case AArch64::LD1i32:
8179 case AArch64::LD1i16:
8181 case AArch64::LD1i8:
8197 unsigned Pattern,
unsigned NumLanes) {
8205 for (
unsigned i = 0; i < NumLanes - 1; ++i) {
8213 return A->getOperand(2).getImm() >
B->getOperand(2).getImm();
8219 auto LoadToLaneInstrsAscending =
llvm::reverse(LoadToLaneInstrs);
8225 auto CreateLD1Instruction = [&](
MachineInstr *OriginalInstr,
8226 Register SrcRegister,
unsigned Lane,
8228 bool OffsetRegisterKillState) {
8236 InstrIdxForVirtReg.
insert(std::make_pair(NewRegister, InsInstrs.
size()));
8237 InsInstrs.
push_back(LoadIndexIntoRegister);
8243 auto CreateLDRInstruction = [&](
unsigned NumLanes,
Register DestReg,
8249 Opcode = AArch64::LDRSui;
8252 Opcode = AArch64::LDRHui;
8255 Opcode = AArch64::LDRBui;
8259 "Got unsupported number of lanes in machine-combiner gather pattern");
8268 auto LanesToLoadToReg0 =
8270 LoadToLaneInstrsAscending.begin() + NumLanes / 2);
8271 Register PrevReg = SubregToReg->getOperand(0).getReg();
8273 const MachineOperand &OffsetRegOperand = LoadInstr->getOperand(3);
8274 PrevReg = CreateLD1Instruction(LoadInstr, PrevReg, Index + 1,
8275 OffsetRegOperand.
getReg(),
8276 OffsetRegOperand.
isKill());
8283 MachineInstr *Lane0Load = *LoadToLaneInstrsAscending.begin();
8285 *std::next(LoadToLaneInstrsAscending.begin(), NumLanes / 2);
8292 CreateLDRInstruction(NumLanes, DestRegForMiddleIndex,
8293 OriginalSplitToLoadOffsetOperand.
getReg(),
8294 OriginalSplitToLoadOffsetOperand.
isKill());
8296 InstrIdxForVirtReg.
insert(
8297 std::make_pair(DestRegForMiddleIndex, InsInstrs.
size()));
8298 InsInstrs.
push_back(MiddleIndexLoadInstr);
8303 unsigned SubregType;
8306 SubregType = AArch64::ssub;
8309 SubregType = AArch64::hsub;
8312 SubregType = AArch64::bsub;
8316 "Got invalid NumLanes for machine-combiner gather pattern");
8319 auto SubRegToRegInstr =
8321 DestRegForSubregToReg)
8324 InstrIdxForVirtReg.
insert(
8325 std::make_pair(DestRegForSubregToReg, InsInstrs.
size()));
8329 auto LanesToLoadToReg1 =
8331 LoadToLaneInstrsAscending.end());
8332 PrevReg = SubRegToRegInstr->getOperand(0).getReg();
8334 const MachineOperand &OffsetRegOperand = LoadInstr->getOperand(3);
8335 PrevReg = CreateLD1Instruction(LoadInstr, PrevReg, Index + 1,
8336 OffsetRegOperand.
getReg(),
8337 OffsetRegOperand.
isKill());
8340 if (Index == NumLanes / 2 - 2) {
8375bool AArch64InstrInfo::getMachineCombinerPatterns(
8377 bool DoRegPressureReduce)
const {
8398 DoRegPressureReduce);
8427 const Register *ReplacedAddend =
nullptr) {
8428 assert(IdxMulOpd == 1 || IdxMulOpd == 2);
8430 unsigned IdxOtherOpd = IdxMulOpd == 1 ? 2 : 1;
8433 Register SrcReg0 = MUL->getOperand(1).getReg();
8434 bool Src0IsKill = MUL->getOperand(1).isKill();
8435 Register SrcReg1 = MUL->getOperand(2).getReg();
8436 bool Src1IsKill = MUL->getOperand(2).isKill();
8440 if (ReplacedAddend) {
8442 SrcReg2 = *ReplacedAddend;
8469 .
addImm(MUL->getOperand(3).getImm());
8476 assert(
false &&
"Invalid FMA instruction kind \n");
8490 if (AArch64::FPR32RegClass.hasSubClassEq(RC))
8491 Opc = AArch64::FNMADDSrrr;
8492 else if (AArch64::FPR64RegClass.hasSubClassEq(RC))
8493 Opc = AArch64::FNMADDDrrr;
8527 unsigned IdxDupOp,
unsigned MulOpc,
8529 assert(((IdxDupOp == 1) || (IdxDupOp == 2)) &&
8530 "Invalid index of FMUL operand");
8538 if (Dup->
getOpcode() == TargetOpcode::COPY)
8547 unsigned IdxMulOp = IdxDupOp == 1 ? 2 : 1;
8588 InstrIdxForVirtReg.
insert(std::make_pair(NewVR, 0));
8603 genNeg(MF, MRI,
TII, Root, InsInstrs, InstrIdxForVirtReg, MnegOpc, RC);
8630 genNeg(MF, MRI,
TII, Root, InsInstrs, InstrIdxForVirtReg, MnegOpc, RC);
8658 unsigned IdxMulOpd,
unsigned MaddOpc,
unsigned VR,
8660 assert(IdxMulOpd == 1 || IdxMulOpd == 2);
8664 Register SrcReg0 = MUL->getOperand(1).getReg();
8665 bool Src0IsKill = MUL->getOperand(1).isKill();
8666 Register SrcReg1 = MUL->getOperand(2).getReg();
8667 bool Src1IsKill = MUL->getOperand(2).isKill();
8697 assert(IdxOpd1 == 1 || IdxOpd1 == 2);
8698 unsigned IdxOtherOpd = IdxOpd1 == 1 ? 2 : 1;
8712 if (Opcode == AArch64::SUBSWrr)
8713 Opcode = AArch64::SUBWrr;
8714 else if (Opcode == AArch64::SUBSXrr)
8715 Opcode = AArch64::SUBXrr;
8717 assert((Opcode == AArch64::SUBWrr || Opcode == AArch64::SUBXrr) &&
8718 "Unexpected instruction opcode.");
8735 InstrIdxForVirtReg.
insert(std::make_pair(NewVR, 0));
8742unsigned AArch64InstrInfo::getReduceOpcodeForAccumulator(
8743 unsigned int AccumulatorOpCode)
const {
8744 switch (AccumulatorOpCode) {
8745 case AArch64::UABALB_ZZZ_D:
8746 case AArch64::SABALB_ZZZ_D:
8747 case AArch64::UABALT_ZZZ_D:
8748 case AArch64::SABALT_ZZZ_D:
8749 return AArch64::ADD_ZZZ_D;
8750 case AArch64::UABALB_ZZZ_H:
8751 case AArch64::SABALB_ZZZ_H:
8752 case AArch64::UABALT_ZZZ_H:
8753 case AArch64::SABALT_ZZZ_H:
8754 return AArch64::ADD_ZZZ_H;
8755 case AArch64::UABALB_ZZZ_S:
8756 case AArch64::SABALB_ZZZ_S:
8757 case AArch64::UABALT_ZZZ_S:
8758 case AArch64::SABALT_ZZZ_S:
8759 return AArch64::ADD_ZZZ_S;
8760 case AArch64::UABALv16i8_v8i16:
8761 case AArch64::SABALv8i8_v8i16:
8762 case AArch64::SABAv8i16:
8763 case AArch64::UABAv8i16:
8764 return AArch64::ADDv8i16;
8765 case AArch64::SABALv2i32_v2i64:
8766 case AArch64::UABALv2i32_v2i64:
8767 case AArch64::SABALv4i32_v2i64:
8768 return AArch64::ADDv2i64;
8769 case AArch64::UABALv4i16_v4i32:
8770 case AArch64::SABALv4i16_v4i32:
8771 case AArch64::SABALv8i16_v4i32:
8772 case AArch64::SABAv4i32:
8773 case AArch64::UABAv4i32:
8774 return AArch64::ADDv4i32;
8775 case AArch64::UABALv4i32_v2i64:
8776 return AArch64::ADDv2i64;
8777 case AArch64::UABALv8i16_v4i32:
8778 return AArch64::ADDv4i32;
8779 case AArch64::UABALv8i8_v8i16:
8780 case AArch64::SABALv16i8_v8i16:
8781 return AArch64::ADDv8i16;
8782 case AArch64::UABAv16i8:
8783 case AArch64::SABAv16i8:
8784 return AArch64::ADDv16i8;
8785 case AArch64::UABAv4i16:
8786 case AArch64::SABAv4i16:
8787 return AArch64::ADDv4i16;
8788 case AArch64::UABAv2i32:
8789 case AArch64::SABAv2i32:
8790 return AArch64::ADDv2i32;
8791 case AArch64::UABAv8i8:
8792 case AArch64::SABAv8i8:
8793 return AArch64::ADDv8i8;
8802void AArch64InstrInfo::genAlternativeCodeSequence(
8812 MachineInstr *
MUL =
nullptr;
8813 const TargetRegisterClass *RC;
8819 DelInstrs, InstrIdxForVirtReg);
8825 InstrIdxForVirtReg);
8831 InstrIdxForVirtReg);
8840 Opc = AArch64::MADDWrrr;
8841 RC = &AArch64::GPR32RegClass;
8843 Opc = AArch64::MADDXrrr;
8844 RC = &AArch64::GPR64RegClass;
8855 Opc = AArch64::MADDWrrr;
8856 RC = &AArch64::GPR32RegClass;
8858 Opc = AArch64::MADDXrrr;
8859 RC = &AArch64::GPR64RegClass;
8872 const TargetRegisterClass *RC;
8873 unsigned BitSize, MovImm;
8876 MovImm = AArch64::MOVi32imm;
8877 RC = &AArch64::GPR32spRegClass;
8879 Opc = AArch64::MADDWrrr;
8880 RC = &AArch64::GPR32RegClass;
8882 MovImm = AArch64::MOVi64imm;
8883 RC = &AArch64::GPR64spRegClass;
8885 Opc = AArch64::MADDXrrr;
8886 RC = &AArch64::GPR64RegClass;
8897 uint64_t UImm =
SignExtend64(IsSub ? -Imm : Imm, BitSize);
8901 if (Insn.
size() != 1)
8903 MachineInstrBuilder MIB1 =
8904 BuildMI(MF, MIMetadata(Root),
TII->get(MovImm), NewVR)
8905 .
addImm(IsSub ? -Imm : Imm);
8907 InstrIdxForVirtReg.
insert(std::make_pair(NewVR, 0));
8918 const TargetRegisterClass *SubRC;
8919 unsigned SubOpc, ZeroReg;
8921 SubOpc = AArch64::SUBWrr;
8922 SubRC = &AArch64::GPR32spRegClass;
8923 ZeroReg = AArch64::WZR;
8924 Opc = AArch64::MADDWrrr;
8925 RC = &AArch64::GPR32RegClass;
8927 SubOpc = AArch64::SUBXrr;
8928 SubRC = &AArch64::GPR64spRegClass;
8929 ZeroReg = AArch64::XZR;
8930 Opc = AArch64::MADDXrrr;
8931 RC = &AArch64::GPR64RegClass;
8935 MachineInstrBuilder MIB1 =
8936 BuildMI(MF, MIMetadata(Root),
TII->get(SubOpc), NewVR)
8940 InstrIdxForVirtReg.
insert(std::make_pair(NewVR, 0));
8951 Opc = AArch64::MSUBWrrr;
8952 RC = &AArch64::GPR32RegClass;
8954 Opc = AArch64::MSUBXrrr;
8955 RC = &AArch64::GPR64RegClass;
8960 Opc = AArch64::MLAv8i8;
8961 RC = &AArch64::FPR64RegClass;
8965 Opc = AArch64::MLAv8i8;
8966 RC = &AArch64::FPR64RegClass;
8970 Opc = AArch64::MLAv16i8;
8971 RC = &AArch64::FPR128RegClass;
8975 Opc = AArch64::MLAv16i8;
8976 RC = &AArch64::FPR128RegClass;
8980 Opc = AArch64::MLAv4i16;
8981 RC = &AArch64::FPR64RegClass;
8985 Opc = AArch64::MLAv4i16;
8986 RC = &AArch64::FPR64RegClass;
8990 Opc = AArch64::MLAv8i16;
8991 RC = &AArch64::FPR128RegClass;
8995 Opc = AArch64::MLAv8i16;
8996 RC = &AArch64::FPR128RegClass;
9000 Opc = AArch64::MLAv2i32;
9001 RC = &AArch64::FPR64RegClass;
9005 Opc = AArch64::MLAv2i32;
9006 RC = &AArch64::FPR64RegClass;
9010 Opc = AArch64::MLAv4i32;
9011 RC = &AArch64::FPR128RegClass;
9015 Opc = AArch64::MLAv4i32;
9016 RC = &AArch64::FPR128RegClass;
9021 Opc = AArch64::MLAv8i8;
9022 RC = &AArch64::FPR64RegClass;
9024 InstrIdxForVirtReg, 1,
Opc, AArch64::NEGv8i8,
9028 Opc = AArch64::MLSv8i8;
9029 RC = &AArch64::FPR64RegClass;
9033 Opc = AArch64::MLAv16i8;
9034 RC = &AArch64::FPR128RegClass;
9036 InstrIdxForVirtReg, 1,
Opc, AArch64::NEGv16i8,
9040 Opc = AArch64::MLSv16i8;
9041 RC = &AArch64::FPR128RegClass;
9045 Opc = AArch64::MLAv4i16;
9046 RC = &AArch64::FPR64RegClass;
9048 InstrIdxForVirtReg, 1,
Opc, AArch64::NEGv4i16,
9052 Opc = AArch64::MLSv4i16;
9053 RC = &AArch64::FPR64RegClass;
9057 Opc = AArch64::MLAv8i16;
9058 RC = &AArch64::FPR128RegClass;
9060 InstrIdxForVirtReg, 1,
Opc, AArch64::NEGv8i16,
9064 Opc = AArch64::MLSv8i16;
9065 RC = &AArch64::FPR128RegClass;
9069 Opc = AArch64::MLAv2i32;
9070 RC = &AArch64::FPR64RegClass;
9072 InstrIdxForVirtReg, 1,
Opc, AArch64::NEGv2i32,
9076 Opc = AArch64::MLSv2i32;
9077 RC = &AArch64::FPR64RegClass;
9081 Opc = AArch64::MLAv4i32;
9082 RC = &AArch64::FPR128RegClass;
9084 InstrIdxForVirtReg, 1,
Opc, AArch64::NEGv4i32,
9088 Opc = AArch64::MLSv4i32;
9089 RC = &AArch64::FPR128RegClass;
9094 Opc = AArch64::MLAv4i16_indexed;
9095 RC = &AArch64::FPR64RegClass;
9099 Opc = AArch64::MLAv4i16_indexed;
9100 RC = &AArch64::FPR64RegClass;
9104 Opc = AArch64::MLAv8i16_indexed;
9105 RC = &AArch64::FPR128RegClass;
9109 Opc = AArch64::MLAv8i16_indexed;
9110 RC = &AArch64::FPR128RegClass;
9114 Opc = AArch64::MLAv2i32_indexed;
9115 RC = &AArch64::FPR64RegClass;
9119 Opc = AArch64::MLAv2i32_indexed;
9120 RC = &AArch64::FPR64RegClass;
9124 Opc = AArch64::MLAv4i32_indexed;
9125 RC = &AArch64::FPR128RegClass;
9129 Opc = AArch64::MLAv4i32_indexed;
9130 RC = &AArch64::FPR128RegClass;
9135 Opc = AArch64::MLAv4i16_indexed;
9136 RC = &AArch64::FPR64RegClass;
9138 InstrIdxForVirtReg, 1,
Opc, AArch64::NEGv4i16,
9142 Opc = AArch64::MLSv4i16_indexed;
9143 RC = &AArch64::FPR64RegClass;
9147 Opc = AArch64::MLAv8i16_indexed;
9148 RC = &AArch64::FPR128RegClass;
9150 InstrIdxForVirtReg, 1,
Opc, AArch64::NEGv8i16,
9154 Opc = AArch64::MLSv8i16_indexed;
9155 RC = &AArch64::FPR128RegClass;
9159 Opc = AArch64::MLAv2i32_indexed;
9160 RC = &AArch64::FPR64RegClass;
9162 InstrIdxForVirtReg, 1,
Opc, AArch64::NEGv2i32,
9166 Opc = AArch64::MLSv2i32_indexed;
9167 RC = &AArch64::FPR64RegClass;
9171 Opc = AArch64::MLAv4i32_indexed;
9172 RC = &AArch64::FPR128RegClass;
9174 InstrIdxForVirtReg, 1,
Opc, AArch64::NEGv4i32,
9178 Opc = AArch64::MLSv4i32_indexed;
9179 RC = &AArch64::FPR128RegClass;
9185 Opc = AArch64::FMADDHrrr;
9186 RC = &AArch64::FPR16RegClass;
9190 Opc = AArch64::FMADDSrrr;
9191 RC = &AArch64::FPR32RegClass;
9195 Opc = AArch64::FMADDDrrr;
9196 RC = &AArch64::FPR64RegClass;
9201 Opc = AArch64::FMADDHrrr;
9202 RC = &AArch64::FPR16RegClass;
9206 Opc = AArch64::FMADDSrrr;
9207 RC = &AArch64::FPR32RegClass;
9211 Opc = AArch64::FMADDDrrr;
9212 RC = &AArch64::FPR64RegClass;
9217 Opc = AArch64::FMLAv1i32_indexed;
9218 RC = &AArch64::FPR32RegClass;
9223 Opc = AArch64::FMLAv1i32_indexed;
9224 RC = &AArch64::FPR32RegClass;
9230 Opc = AArch64::FMLAv1i64_indexed;
9231 RC = &AArch64::FPR64RegClass;
9236 Opc = AArch64::FMLAv1i64_indexed;
9237 RC = &AArch64::FPR64RegClass;
9243 RC = &AArch64::FPR64RegClass;
9244 Opc = AArch64::FMLAv4i16_indexed;
9249 RC = &AArch64::FPR64RegClass;
9250 Opc = AArch64::FMLAv4f16;
9255 RC = &AArch64::FPR64RegClass;
9256 Opc = AArch64::FMLAv4i16_indexed;
9261 RC = &AArch64::FPR64RegClass;
9262 Opc = AArch64::FMLAv4f16;
9269 RC = &AArch64::FPR64RegClass;
9271 Opc = AArch64::FMLAv2i32_indexed;
9275 Opc = AArch64::FMLAv2f32;
9282 RC = &AArch64::FPR64RegClass;
9284 Opc = AArch64::FMLAv2i32_indexed;
9288 Opc = AArch64::FMLAv2f32;
9295 RC = &AArch64::FPR128RegClass;
9296 Opc = AArch64::FMLAv8i16_indexed;
9301 RC = &AArch64::FPR128RegClass;
9302 Opc = AArch64::FMLAv8f16;
9307 RC = &AArch64::FPR128RegClass;
9308 Opc = AArch64::FMLAv8i16_indexed;
9313 RC = &AArch64::FPR128RegClass;
9314 Opc = AArch64::FMLAv8f16;
9321 RC = &AArch64::FPR128RegClass;
9323 Opc = AArch64::FMLAv2i64_indexed;
9327 Opc = AArch64::FMLAv2f64;
9334 RC = &AArch64::FPR128RegClass;
9336 Opc = AArch64::FMLAv2i64_indexed;
9340 Opc = AArch64::FMLAv2f64;
9348 RC = &AArch64::FPR128RegClass;
9350 Opc = AArch64::FMLAv4i32_indexed;
9354 Opc = AArch64::FMLAv4f32;
9362 RC = &AArch64::FPR128RegClass;
9364 Opc = AArch64::FMLAv4i32_indexed;
9368 Opc = AArch64::FMLAv4f32;
9375 Opc = AArch64::FNMSUBHrrr;
9376 RC = &AArch64::FPR16RegClass;
9380 Opc = AArch64::FNMSUBSrrr;
9381 RC = &AArch64::FPR32RegClass;
9385 Opc = AArch64::FNMSUBDrrr;
9386 RC = &AArch64::FPR64RegClass;
9391 Opc = AArch64::FNMADDHrrr;
9392 RC = &AArch64::FPR16RegClass;
9396 Opc = AArch64::FNMADDSrrr;
9397 RC = &AArch64::FPR32RegClass;
9401 Opc = AArch64::FNMADDDrrr;
9402 RC = &AArch64::FPR64RegClass;
9407 Opc = AArch64::FMSUBHrrr;
9408 RC = &AArch64::FPR16RegClass;
9412 Opc = AArch64::FMSUBSrrr;
9413 RC = &AArch64::FPR32RegClass;
9417 Opc = AArch64::FMSUBDrrr;
9418 RC = &AArch64::FPR64RegClass;
9423 Opc = AArch64::FMLSv1i32_indexed;
9424 RC = &AArch64::FPR32RegClass;
9430 Opc = AArch64::FMLSv1i64_indexed;
9431 RC = &AArch64::FPR64RegClass;
9438 RC = &AArch64::FPR64RegClass;
9440 MachineInstrBuilder MIB1 =
9441 BuildMI(MF, MIMetadata(Root),
TII->get(AArch64::FNEGv4f16), NewVR)
9444 InstrIdxForVirtReg.
insert(std::make_pair(NewVR, 0));
9446 Opc = AArch64::FMLAv4f16;
9450 Opc = AArch64::FMLAv4i16_indexed;
9457 RC = &AArch64::FPR64RegClass;
9458 Opc = AArch64::FMLSv4f16;
9463 RC = &AArch64::FPR64RegClass;
9464 Opc = AArch64::FMLSv4i16_indexed;
9471 RC = &AArch64::FPR64RegClass;
9473 Opc = AArch64::FMLSv2i32_indexed;
9477 Opc = AArch64::FMLSv2f32;
9485 RC = &AArch64::FPR128RegClass;
9487 MachineInstrBuilder MIB1 =
9488 BuildMI(MF, MIMetadata(Root),
TII->get(AArch64::FNEGv8f16), NewVR)
9491 InstrIdxForVirtReg.
insert(std::make_pair(NewVR, 0));
9493 Opc = AArch64::FMLAv8f16;
9497 Opc = AArch64::FMLAv8i16_indexed;
9504 RC = &AArch64::FPR128RegClass;
9505 Opc = AArch64::FMLSv8f16;
9510 RC = &AArch64::FPR128RegClass;
9511 Opc = AArch64::FMLSv8i16_indexed;
9518 RC = &AArch64::FPR128RegClass;
9520 Opc = AArch64::FMLSv2i64_indexed;
9524 Opc = AArch64::FMLSv2f64;
9532 RC = &AArch64::FPR128RegClass;
9534 Opc = AArch64::FMLSv4i32_indexed;
9538 Opc = AArch64::FMLSv4f32;
9545 RC = &AArch64::FPR64RegClass;
9547 MachineInstrBuilder MIB1 =
9548 BuildMI(MF, MIMetadata(Root),
TII->get(AArch64::FNEGv2f32), NewVR)
9551 InstrIdxForVirtReg.
insert(std::make_pair(NewVR, 0));
9553 Opc = AArch64::FMLAv2i32_indexed;
9557 Opc = AArch64::FMLAv2f32;
9565 RC = &AArch64::FPR128RegClass;
9567 MachineInstrBuilder MIB1 =
9568 BuildMI(MF, MIMetadata(Root),
TII->get(AArch64::FNEGv4f32), NewVR)
9571 InstrIdxForVirtReg.
insert(std::make_pair(NewVR, 0));
9573 Opc = AArch64::FMLAv4i32_indexed;
9577 Opc = AArch64::FMLAv4f32;
9585 RC = &AArch64::FPR128RegClass;
9587 MachineInstrBuilder MIB1 =
9588 BuildMI(MF, MIMetadata(Root),
TII->get(AArch64::FNEGv2f64), NewVR)
9591 InstrIdxForVirtReg.
insert(std::make_pair(NewVR, 0));
9593 Opc = AArch64::FMLAv2i64_indexed;
9597 Opc = AArch64::FMLAv2f64;
9609 &AArch64::FPR128RegClass, MRI);
9618 &AArch64::FPR128RegClass, MRI);
9627 &AArch64::FPR128_loRegClass, MRI);
9636 &AArch64::FPR128RegClass, MRI);
9645 &AArch64::FPR128_loRegClass, MRI);
9679 for (
auto *
MI : InsInstrs)
9680 MI->setFlags(Flags);
9721 bool IsNegativeBranch =
false;
9722 bool IsTestAndBranch =
false;
9723 unsigned TargetBBInMI = 0;
9724 switch (
MI.getOpcode()) {
9728 case AArch64::CBWPri:
9729 case AArch64::CBXPri:
9730 case AArch64::CBBAssertExt:
9731 case AArch64::CBHAssertExt:
9732 case AArch64::CBWPrr:
9733 case AArch64::CBXPrr:
9739 case AArch64::CBNZW:
9740 case AArch64::CBNZX:
9742 IsNegativeBranch =
true;
9747 IsTestAndBranch =
true;
9749 case AArch64::TBNZW:
9750 case AArch64::TBNZX:
9752 IsNegativeBranch =
true;
9753 IsTestAndBranch =
true;
9759 if (IsTestAndBranch &&
MI.getOperand(1).getImm())
9763 assert(
MI.getParent() &&
"Incomplete machine instruction\n");
9774 while (
DefMI->isCopy()) {
9783 switch (
DefMI->getOpcode()) {
9787 case AArch64::ANDWri:
9788 case AArch64::ANDXri: {
9789 if (IsTestAndBranch)
9796 bool Is32Bit = (
DefMI->getOpcode() == AArch64::ANDWri);
9798 DefMI->getOperand(2).getImm(), Is32Bit ? 32 : 64);
9813 unsigned Opc = (Imm < 32)
9814 ? (IsNegativeBranch ? AArch64::TBNZW : AArch64::TBZW)
9815 : (IsNegativeBranch ? AArch64::TBNZX : AArch64::TBZX);
9828 if (!Is32Bit && Imm < 32)
9830 MI.eraseFromParent();
9834 case AArch64::CSINCWr:
9835 case AArch64::CSINCXr: {
9836 if (!(
DefMI->getOperand(1).getReg() == AArch64::WZR &&
9837 DefMI->getOperand(2).getReg() == AArch64::WZR) &&
9838 !(
DefMI->getOperand(1).getReg() == AArch64::XZR &&
9839 DefMI->getOperand(2).getReg() == AArch64::XZR))
9842 if (
DefMI->findRegisterDefOperandIdx(AArch64::NZCV,
nullptr,
9855 if (IsNegativeBranch)
9858 MI.eraseFromParent();
9864std::pair<unsigned, unsigned>
9865AArch64InstrInfo::decomposeMachineOperandsTargetFlags(
unsigned TF)
const {
9867 return std::make_pair(TF & Mask, TF & ~Mask);
9871AArch64InstrInfo::getSerializableDirectMachineOperandTargetFlags()
const {
9874 static const std::pair<unsigned, const char *> TargetFlags[] = {
9875 {MO_PAGE,
"aarch64-page"}, {
MO_PAGEOFF,
"aarch64-pageoff"},
9876 {
MO_G3,
"aarch64-g3"}, {
MO_G2,
"aarch64-g2"},
9877 {
MO_G1,
"aarch64-g1"}, {
MO_G0,
"aarch64-g0"},
9883AArch64InstrInfo::getSerializableBitmaskMachineOperandTargetFlags()
const {
9884 using namespace AArch64II;
9886 static const std::pair<unsigned, const char *> TargetFlags[] = {
9889 {
MO_NC,
"aarch64-nc"},
9890 {
MO_S,
"aarch64-s"},
9901AArch64InstrInfo::getSerializableMachineMemOperandTargetFlags()
const {
9902 static const std::pair<MachineMemOperand::Flags, const char *> TargetFlags[] =
10004 MachineFunction *MF =
C.getMF();
10006 const AArch64RegisterInfo *ARI =
10007 static_cast<const AArch64RegisterInfo *
>(&
TRI);
10010 for (
unsigned Reg : AArch64::GPR64RegClass) {
10012 Reg != AArch64::LR &&
10013 Reg != AArch64::X16 &&
10014 Reg != AArch64::X17 &&
10015 C.isAvailableAcrossAndOutOfSeq(
Reg,
TRI) &&
10016 C.isAvailableInsideSeq(
Reg,
TRI))
10047 return SubtargetA.hasV8_3aOps() == SubtargetB.hasV8_3aOps();
10050std::optional<std::unique_ptr<outliner::OutlinedFunction>>
10051AArch64InstrInfo::getOutliningCandidateInfo(
10053 std::vector<outliner::Candidate> &RepeatedSequenceLocs,
10054 unsigned MinRepeats)
const {
10055 unsigned SequenceSize = 0;
10056 for (
auto &
MI : RepeatedSequenceLocs[0])
10059 unsigned NumBytesToCreateFrame = 0;
10065 MachineInstr &LastMI = RepeatedSequenceLocs[0].back();
10066 MachineInstr &FirstMI = RepeatedSequenceLocs[0].front();
10067 if (LastMI.
getOpcode() == AArch64::ADRP &&
10070 return std::nullopt;
10075 if ((FirstMI.
getOpcode() == AArch64::ADDXri ||
10076 FirstMI.
getOpcode() == AArch64::LDRXui) &&
10079 return std::nullopt;
10090 if (std::adjacent_find(
10091 RepeatedSequenceLocs.begin(), RepeatedSequenceLocs.end(),
10092 [](
const outliner::Candidate &a,
const outliner::Candidate &b) {
10095 if (outliningCandidatesSigningScopeConsensus(a, b) &&
10096 outliningCandidatesSigningKeyConsensus(a, b) &&
10097 outliningCandidatesV8_3OpsConsensus(a, b)) {
10101 }) != RepeatedSequenceLocs.end()) {
10102 return std::nullopt;
10119 unsigned NumBytesToCheckLRInTCEpilogue = 0;
10120 const auto RASignCondition = RepeatedSequenceLocs[0]
10123 ->getSignReturnAddressCondition();
10126 NumBytesToCreateFrame += 8;
10129 auto LRCheckMethod = Subtarget.getAuthenticatedLRCheckMethod(
10130 *RepeatedSequenceLocs[0].getMF());
10131 NumBytesToCheckLRInTCEpilogue =
10135 if (isTailCallReturnInst(RepeatedSequenceLocs[0].
back()))
10136 SequenceSize += NumBytesToCheckLRInTCEpilogue;
10144 for (
auto &
MI :
C) {
10145 if (
MI.modifiesRegister(AArch64::SP, &
TRI)) {
10146 switch (
MI.getOpcode()) {
10147 case AArch64::ADDXri:
10148 case AArch64::ADDWri:
10149 assert(
MI.getNumOperands() == 4 &&
"Wrong number of operands");
10150 assert(
MI.getOperand(2).isImm() &&
10151 "Expected operand to be immediate");
10152 assert(
MI.getOperand(1).isReg() &&
10153 "Expected operand to be a register");
10157 if (
MI.getOperand(1).getReg() == AArch64::SP)
10158 SPValue +=
MI.getOperand(2).getImm();
10162 case AArch64::SUBXri:
10163 case AArch64::SUBWri:
10164 assert(
MI.getNumOperands() == 4 &&
"Wrong number of operands");
10165 assert(
MI.getOperand(2).isImm() &&
10166 "Expected operand to be immediate");
10167 assert(
MI.getOperand(1).isReg() &&
10168 "Expected operand to be a register");
10172 if (
MI.getOperand(1).getReg() == AArch64::SP)
10173 SPValue -=
MI.getOperand(2).getImm();
10190 if (RepeatedSequenceLocs.size() < MinRepeats)
10191 return std::nullopt;
10195 unsigned FlagsSetInAll = 0xF;
10199 FlagsSetInAll &=
C.Flags;
10201 unsigned LastInstrOpcode = RepeatedSequenceLocs[0].back().getOpcode();
10204 auto SetCandidateCallInfo =
10205 [&RepeatedSequenceLocs](
unsigned CallID,
unsigned NumBytesForCall) {
10207 C.setCallInfo(CallID, NumBytesForCall);
10211 NumBytesToCreateFrame += 4;
10219 unsigned CFICount = 0;
10220 for (
auto &
I : RepeatedSequenceLocs[0]) {
10221 if (
I.isCFIInstruction())
10231 std::vector<MCCFIInstruction> CFIInstructions =
10232 C.getMF()->getFrameInstructions();
10234 if (CFICount > 0 && CFICount != CFIInstructions.size())
10235 return std::nullopt;
10243 if (!
MI.modifiesRegister(AArch64::SP, &
TRI) &&
10244 !
MI.readsRegister(AArch64::SP, &
TRI))
10250 if (
MI.modifiesRegister(AArch64::SP, &
TRI))
10255 if (
MI.mayLoadOrStore()) {
10258 bool OffsetIsScalable;
10262 if (!getMemOperandWithOffset(
MI,
Base,
Offset, OffsetIsScalable, &
TRI) ||
10263 !
Base->isReg() ||
Base->getReg() != AArch64::SP)
10267 if (OffsetIsScalable)
10275 TypeSize Scale(0U,
false), DummyWidth(0U,
false);
10276 getMemOpInfo(
MI.getOpcode(), Scale, DummyWidth, MinOffset, MaxOffset);
10279 if (
Offset < MinOffset * (int64_t)Scale.getFixedValue() ||
10280 Offset > MaxOffset * (int64_t)Scale.getFixedValue())
10295 bool AllStackInstrsSafe =
10300 if (RepeatedSequenceLocs[0].
back().isTerminator()) {
10302 NumBytesToCreateFrame = 0;
10303 unsigned NumBytesForCall = 4 + NumBytesToCheckLRInTCEpilogue;
10307 else if (LastInstrOpcode == AArch64::BL ||
10308 ((LastInstrOpcode == AArch64::BLR ||
10309 LastInstrOpcode == AArch64::BLRNoIP) &&
10313 NumBytesToCreateFrame = NumBytesToCheckLRInTCEpilogue;
10321 unsigned NumBytesNoStackCalls = 0;
10322 std::vector<outliner::Candidate> CandidatesWithoutStackFixups;
10328 ?
C.isAvailableAcrossAndOutOfSeq(AArch64::LR,
TRI)
10337 C.getMF()->getFunction().hasFnAttribute(Attribute::NoReturn);
10340 if (LRAvailable && !IsNoReturn) {
10341 NumBytesNoStackCalls += 4;
10343 CandidatesWithoutStackFixups.push_back(
C);
10348 else if (findRegisterToSaveLRTo(
C)) {
10349 NumBytesNoStackCalls += 12;
10351 CandidatesWithoutStackFixups.push_back(
C);
10356 else if (
C.isAvailableInsideSeq(AArch64::SP,
TRI)) {
10357 NumBytesNoStackCalls += 12;
10359 CandidatesWithoutStackFixups.push_back(
C);
10365 NumBytesNoStackCalls += SequenceSize;
10372 if (!AllStackInstrsSafe ||
10373 NumBytesNoStackCalls <= RepeatedSequenceLocs.size() * 12) {
10374 RepeatedSequenceLocs = CandidatesWithoutStackFixups;
10376 if (RepeatedSequenceLocs.size() < MinRepeats)
10377 return std::nullopt;
10430 (!
C.isAvailableAcrossAndOutOfSeq(AArch64::LR,
TRI) ||
10431 !findRegisterToSaveLRTo(
C));
10437 if (RepeatedSequenceLocs.size() < MinRepeats)
10438 return std::nullopt;
10447 bool ModStackToSaveLR =
false;
10450 ModStackToSaveLR =
true;
10459 ModStackToSaveLR =
true;
10461 if (ModStackToSaveLR) {
10463 if (!AllStackInstrsSafe)
10464 return std::nullopt;
10467 NumBytesToCreateFrame += 8;
10474 return std::nullopt;
10476 return std::make_unique<outliner::OutlinedFunction>(
10477 RepeatedSequenceLocs, SequenceSize, NumBytesToCreateFrame, FrameID);
10480void AArch64InstrInfo::mergeOutliningCandidateAttributes(
10481 Function &
F, std::vector<outliner::Candidate> &Candidates)
const {
10485 const auto &CFn = Candidates.front().getMF()->getFunction();
10487 if (CFn.hasFnAttribute(
"ptrauth-returns"))
10488 F.addFnAttr(CFn.getFnAttribute(
"ptrauth-returns"));
10489 if (CFn.hasFnAttribute(
"ptrauth-auth-traps"))
10490 F.addFnAttr(CFn.getFnAttribute(
"ptrauth-auth-traps"));
10493 if (CFn.hasFnAttribute(
"sign-return-address"))
10494 F.addFnAttr(CFn.getFnAttribute(
"sign-return-address"));
10495 if (CFn.hasFnAttribute(
"sign-return-address-key"))
10496 F.addFnAttr(CFn.getFnAttribute(
"sign-return-address-key"));
10498 AArch64GenInstrInfo::mergeOutliningCandidateAttributes(
F, Candidates);
10501bool AArch64InstrInfo::isFunctionSafeToOutlineFrom(
10506 if (!OutlineFromLinkOnceODRs &&
F.hasLinkOnceODRLinkage())
10513 if (
F.hasSection())
10519 AArch64FunctionInfo *AFI = MF.
getInfo<AArch64FunctionInfo>();
10520 if (!AFI || AFI->
hasRedZone().value_or(
true))
10540 unsigned &Flags)
const {
10542 "Must track liveness!");
10544 std::pair<MachineBasicBlock::iterator, MachineBasicBlock::iterator>>
10559 auto AreAllUnsafeRegsDead = [&LRU]() {
10560 return LRU.available(AArch64::W16) && LRU.available(AArch64::W17) &&
10561 LRU.available(AArch64::NZCV);
10576 bool LRAvailableEverywhere =
true;
10578 LRU.addLiveOuts(
MBB);
10580 auto UpdateWholeMBBFlags = [&
Flags](
const MachineInstr &
MI) {
10581 if (
MI.isCall() && !
MI.isTerminator())
10587 auto CreateNewRangeStartingAt =
10588 [&RangeBegin, &RangeEnd,
10590 RangeBegin = NewBegin;
10591 RangeEnd = std::next(RangeBegin);
10594 auto SaveRangeIfNonEmpty = [&RangeLen, &
Ranges, &RangeBegin, &RangeEnd]() {
10600 if (!RangeBegin.isEnd() && RangeBegin->isBundledWithPred())
10602 if (!RangeEnd.isEnd() && RangeEnd->isBundledWithPred())
10604 Ranges.emplace_back(RangeBegin, RangeEnd);
10612 for (; FirstPossibleEndPt !=
MBB.
instr_rend(); ++FirstPossibleEndPt) {
10613 LRU.stepBackward(*FirstPossibleEndPt);
10616 UpdateWholeMBBFlags(*FirstPossibleEndPt);
10617 if (AreAllUnsafeRegsDead())
10624 CreateNewRangeStartingAt(FirstPossibleEndPt->getIterator());
10629 LRU.stepBackward(
MI);
10630 UpdateWholeMBBFlags(
MI);
10631 if (!AreAllUnsafeRegsDead()) {
10632 SaveRangeIfNonEmpty();
10633 CreateNewRangeStartingAt(
MI.getIterator());
10636 LRAvailableEverywhere &= LRU.available(AArch64::LR);
10637 RangeBegin =
MI.getIterator();
10642 if (AreAllUnsafeRegsDead())
10643 SaveRangeIfNonEmpty();
10651 if (!LRAvailableEverywhere)
10659 unsigned Flags)
const {
10660 MachineInstr &
MI = *MIT;
10664 switch (
MI.getOpcode()) {
10665 case AArch64::PACM:
10666 case AArch64::PACIASP:
10667 case AArch64::PACIBSP:
10668 case AArch64::PACIASPPC:
10669 case AArch64::PACIBSPPC:
10670 case AArch64::AUTIASP:
10671 case AArch64::AUTIBSP:
10672 case AArch64::AUTIASPPCi:
10673 case AArch64::AUTIASPPCr:
10674 case AArch64::AUTIBSPPCi:
10675 case AArch64::AUTIBSPPCr:
10676 case AArch64::RETAA:
10677 case AArch64::RETAB:
10678 case AArch64::RETAASPPCi:
10679 case AArch64::RETAASPPCr:
10680 case AArch64::RETABSPPCi:
10681 case AArch64::RETABSPPCr:
10682 case AArch64::EMITBKEY:
10683 case AArch64::PAUTH_PROLOGUE:
10684 case AArch64::PAUTH_EPILOGUE:
10694 if (
MI.isCFIInstruction())
10698 if (
MI.isTerminator())
10704 for (
const MachineOperand &MOP :
MI.operands()) {
10707 assert(!MOP.isCFIIndex());
10710 if (MOP.isReg() && !MOP.isImplicit() &&
10711 (MOP.getReg() == AArch64::LR || MOP.getReg() == AArch64::W30))
10718 if (
MI.getOpcode() == AArch64::ADRP)
10738 for (
const MachineOperand &MOP :
MI.operands()) {
10739 if (MOP.isGlobal()) {
10747 if (Callee &&
Callee->getName() ==
"\01_mcount")
10755 if (
MI.getOpcode() == AArch64::BLR ||
10756 MI.getOpcode() == AArch64::BLRNoIP ||
MI.getOpcode() == AArch64::BL)
10760 return UnknownCallOutlineType;
10768 return UnknownCallOutlineType;
10776 return UnknownCallOutlineType;
10797 for (MachineInstr &
MI :
MBB) {
10798 const MachineOperand *
Base;
10799 TypeSize Width(0,
false);
10801 bool OffsetIsScalable;
10804 if (!
MI.mayLoadOrStore() ||
10807 (
Base->isReg() &&
Base->getReg() != AArch64::SP))
10811 TypeSize Scale(0U,
false);
10812 int64_t Dummy1, Dummy2;
10815 assert(StackOffsetOperand.
isImm() &&
"Stack offset wasn't immediate!");
10817 assert(Scale != 0 &&
"Unexpected opcode!");
10818 assert(!OffsetIsScalable &&
"Expected offset to be a byte offset");
10823 int64_t NewImm = (
Offset + 16) / (int64_t)Scale.getFixedValue();
10824 StackOffsetOperand.
setImm(NewImm);
10830 bool ShouldSignReturnAddr) {
10831 if (!ShouldSignReturnAddr)
10839void AArch64InstrInfo::buildOutlinedFrame(
10843 AArch64FunctionInfo *FI = MF.
getInfo<AArch64FunctionInfo>();
10851 unsigned TailOpcode;
10853 TailOpcode = AArch64::TCRETURNdi;
10857 TailOpcode = AArch64::TCRETURNriALL;
10868 bool IsLeafFunction =
true;
10871 auto IsNonTailCall = [](
const MachineInstr &
MI) {
10872 return MI.isCall() && !
MI.isReturn();
10882 "Can only fix up stack references once");
10883 fixupPostOutline(
MBB);
10885 IsLeafFunction =
false;
10896 Et = std::prev(
MBB.
end());
10906 if (MF.
getInfo<AArch64FunctionInfo>()->needsDwarfUnwindInfo(MF)) {
10910 CFIBuilder.buildDefCFAOffset(16);
10914 CFIBuilder.buildOffset(AArch64::LR, -16);
10928 RASignCondition, !IsLeafFunction);
10957 fixupPostOutline(
MBB);
10968 .addGlobalAddress(
M.getNamedValue(MF.
getName()))
10978 .addGlobalAddress(
M.getNamedValue(MF.
getName())));
10987 MachineInstr *Save;
10988 MachineInstr *Restore;
10994 assert(
Reg &&
"No callee-saved register available?");
11028 .addGlobalAddress(
M.getNamedValue(MF.
getName())));
11036bool AArch64InstrInfo::shouldOutlineFromFunctionByDefault(
11044 bool AllowSideEffects)
const {
11046 const AArch64Subtarget &STI = MF.
getSubtarget<AArch64Subtarget>();
11049 if (
TRI.isGeneralPurposeRegister(MF,
Reg)) {
11062 assert(STI.hasNEON() &&
"Expected to have NEON.");
11068std::optional<DestSourcePair>
11073 if (((
MI.getOpcode() == AArch64::ORRWrs &&
11074 MI.getOperand(1).getReg() == AArch64::WZR &&
11075 MI.getOperand(3).getImm() == 0x0) ||
11076 (
MI.getOpcode() == AArch64::ORRWrr &&
11077 MI.getOperand(1).getReg() == AArch64::WZR)) &&
11079 (!
MI.getOperand(0).getReg().isVirtual() ||
11080 MI.getOperand(0).getSubReg() == 0) &&
11081 (!
MI.getOperand(0).getReg().isPhysical() ||
11086 if (
MI.getOpcode() == AArch64::ORRXrs &&
11087 MI.getOperand(1).getReg() == AArch64::XZR &&
11088 MI.getOperand(3).getImm() == 0x0)
11091 return std::nullopt;
11094std::optional<DestSourcePair>
11096 if ((
MI.getOpcode() == AArch64::ORRWrs &&
11097 MI.getOperand(1).getReg() == AArch64::WZR &&
11098 MI.getOperand(3).getImm() == 0x0) ||
11099 (
MI.getOpcode() == AArch64::ORRWrr &&
11100 MI.getOperand(1).getReg() == AArch64::WZR))
11102 return std::nullopt;
11105std::optional<RegImmPair>
11114 return std::nullopt;
11116 switch (
MI.getOpcode()) {
11118 return std::nullopt;
11119 case AArch64::SUBWri:
11120 case AArch64::SUBXri:
11121 case AArch64::SUBSWri:
11122 case AArch64::SUBSXri:
11125 case AArch64::ADDSWri:
11126 case AArch64::ADDSXri:
11127 case AArch64::ADDWri:
11128 case AArch64::ADDXri: {
11130 if (!
MI.getOperand(0).isReg() || !
MI.getOperand(1).isReg() ||
11131 !
MI.getOperand(2).isImm())
11132 return std::nullopt;
11133 int Shift =
MI.getOperand(3).getImm();
11134 assert((Shift == 0 || Shift == 12) &&
"Shift can be either 0 or 12");
11138 return RegImmPair{
MI.getOperand(1).getReg(),
Offset};
11144static std::optional<ParamLoadedValue>
11148 auto DestSrc =
TII->isCopyLikeInstr(
MI);
11150 return std::nullopt;
11152 Register DestReg = DestSrc->Destination->getReg();
11153 Register SrcReg = DestSrc->Source->getReg();
11156 return std::nullopt;
11161 if (DestReg == DescribedReg)
11165 if (
MI.getOpcode() == AArch64::ORRWrs &&
11166 TRI->isSuperRegister(DestReg, DescribedReg))
11170 if (
MI.getOpcode() == AArch64::ORRXrs &&
11171 TRI->isSubRegister(DestReg, DescribedReg)) {
11172 Register SrcSubReg =
TRI->getSubReg(SrcReg, AArch64::sub_32);
11176 assert(!
TRI->isSuperOrSubRegisterEq(DestReg, DescribedReg) &&
11177 "Unhandled ORR[XW]rs copy case");
11179 return std::nullopt;
11182bool AArch64InstrInfo::isFunctionSafeToSplit(
const MachineFunction &MF)
const {
11187 if (MF.
getInfo<AArch64FunctionInfo>()->hasRedZone().value_or(
true))
11193bool AArch64InstrInfo::isMBBSafeToSplitToCold(
11197 auto isAsmGoto = [](
const MachineInstr &
MI) {
11198 return MI.getOpcode() == AArch64::INLINEASM_BR;
11208 auto containsMBB = [&
MBB](
const MachineJumpTableEntry &JTE) {
11215 for (
const MachineInstr &
MI :
MBB) {
11216 switch (
MI.getOpcode()) {
11217 case TargetOpcode::G_BRJT:
11218 case AArch64::JumpTableDest32:
11219 case AArch64::JumpTableDest16:
11220 case AArch64::JumpTableDest8:
11231std::optional<ParamLoadedValue>
11234 const MachineFunction *MF =
MI.getMF();
11236 switch (
MI.getOpcode()) {
11237 case AArch64::MOVZWi:
11238 case AArch64::MOVZXi: {
11241 if (!
TRI->isSuperRegisterEq(
MI.getOperand(0).getReg(),
Reg))
11242 return std::nullopt;
11244 if (!
MI.getOperand(1).isImm())
11245 return std::nullopt;
11246 int64_t Immediate =
MI.getOperand(1).getImm();
11247 int Shift =
MI.getOperand(2).getImm();
11251 case AArch64::ORRWrs:
11252 case AArch64::ORRXrs:
11259bool AArch64InstrInfo::isExtendLikelyToBeFolded(
11262 ExtMI.
getOpcode() == TargetOpcode::G_ZEXT ||
11263 ExtMI.
getOpcode() == TargetOpcode::G_ANYEXT);
11266 if (ExtMI.
getOpcode() == TargetOpcode::G_ANYEXT)
11276 return UserMI->
getOpcode() == TargetOpcode::G_PTR_ADD;
11279uint64_t AArch64InstrInfo::getElementSizeForOpcode(
unsigned Opc)
const {
11283bool AArch64InstrInfo::isPTestLikeOpcode(
unsigned Opc)
const {
11287bool AArch64InstrInfo::isWhileOpcode(
unsigned Opc)
const {
11292AArch64InstrInfo::getTailDuplicateSize(
CodeGenOptLevel OptLevel)
const {
11296bool AArch64InstrInfo::isLegalAddressingMode(
unsigned NumBytes, int64_t
Offset,
11297 unsigned Scale)
const {
11308 unsigned Shift =
Log2_64(NumBytes);
11309 if (NumBytes &&
Offset > 0 && (
Offset / NumBytes) <= (1LL << 12) - 1 &&
11317 return Scale == 1 || (Scale > 0 && Scale == NumBytes);
11322 return AArch64::BLRNoIP;
11324 return AArch64::BLR;
11330 auto Builder =
BuildMI(
MBB, InsertPt,
DL,
get(AArch64::PAUTH_EPILOGUE))
11340 Register TargetReg,
bool FrameSetup)
const {
11341 assert(TargetReg != AArch64::SP &&
"New top of stack cannot already be in SP");
11353 MF.
insert(MBBInsertPoint, LoopTestMBB);
11356 MF.
insert(MBBInsertPoint, LoopBodyMBB);
11358 MF.
insert(MBBInsertPoint, ExitMBB);
11368 BuildMI(*LoopTestMBB, LoopTestMBB->
end(),
DL,
TII->get(AArch64::SUBSXrx64),
11376 BuildMI(*LoopTestMBB, LoopTestMBB->
end(),
DL,
TII->get(AArch64::Bcc))
11382 BuildMI(*LoopBodyMBB, LoopBodyMBB->
end(),
DL,
TII->get(AArch64::LDRXui))
11399 BuildMI(*ExitMBB, ExitMBB->
end(),
DL,
TII->get(AArch64::ADDXri), AArch64::SP)
11418 MBB.addSuccessor(LoopTestMBB);
11424 return ExitMBB->
begin();
11429 MachineFunction *MF;
11430 const TargetInstrInfo *
TII;
11431 const TargetRegisterInfo *
TRI;
11432 MachineRegisterInfo &MRI;
11435 MachineBasicBlock *LoopBB;
11437 MachineInstr *CondBranch;
11439 MachineInstr *Comp;
11441 unsigned CompCounterOprNum;
11443 MachineInstr *Update;
11445 unsigned UpdateCounterOprNum;
11449 bool IsUpdatePriorComp;
11455 AArch64PipelinerLoopInfo(MachineBasicBlock *LoopBB, MachineInstr *CondBranch,
11456 MachineInstr *Comp,
unsigned CompCounterOprNum,
11457 MachineInstr *Update,
unsigned UpdateCounterOprNum,
11458 Register Init,
bool IsUpdatePriorComp,
11459 const SmallVectorImpl<MachineOperand> &
Cond)
11461 TII(MF->getSubtarget().getInstrInfo()),
11462 TRI(MF->getSubtarget().getRegisterInfo()), MRI(MF->getRegInfo()),
11463 LoopBB(LoopBB), CondBranch(CondBranch), Comp(Comp),
11464 CompCounterOprNum(CompCounterOprNum), Update(Update),
11465 UpdateCounterOprNum(UpdateCounterOprNum), Init(Init),
11468 bool shouldIgnoreForPipelining(
const MachineInstr *
MI)
const override {
11474 std::optional<bool> createTripCountGreaterCondition(
11475 int TC, MachineBasicBlock &
MBB,
11476 SmallVectorImpl<MachineOperand> &CondParam)
override {
11484 void createRemainingIterationsGreaterCondition(
11485 int TC, MachineBasicBlock &
MBB, SmallVectorImpl<MachineOperand> &
Cond,
11486 DenseMap<MachineInstr *, MachineInstr *> &LastStage0Insts)
override;
11488 void setPreheader(MachineBasicBlock *NewPreheader)
override {}
11490 void adjustTripCount(
int TripCountAdjust)
override {}
11492 bool isMVEExpanderSupported()
override {
return true; }
11511 }
else if (
I == ReplaceOprNum) {
11516 MBB.insert(InsertTo, NewMI);
11520void AArch64PipelinerLoopInfo::createRemainingIterationsGreaterCondition(
11536 assert(CondBranch->getOpcode() == AArch64::Bcc);
11540 if (CondBranch->getOperand(1).getMBB() == LoopBB)
11547 auto AccumulateCond = [&](
Register CurCond,
11558 if (!LastStage0Insts.
empty() && LastStage0Insts[Comp]->getParent() == &
MBB) {
11562 for (
int I = 0;
I <= TC; ++
I) {
11568 AccCond = AccumulateCond(AccCond, CC);
11572 if (Update != Comp && IsUpdatePriorComp) {
11574 LastStage0Insts[Comp]->getOperand(CompCounterOprNum).getReg();
11575 NextCounter =
cloneInstr(Update, UpdateCounterOprNum, Counter,
MBB,
11579 NextCounter = LastStage0Insts[Update]->getOperand(0).getReg();
11581 }
else if (Update != Comp) {
11586 Counter = NextCounter;
11590 if (LastStage0Insts.
empty()) {
11594 if (IsUpdatePriorComp)
11599 Counter = LastStage0Insts[Comp]->getOperand(CompCounterOprNum).getReg();
11602 for (
int I = 0;
I <= TC; ++
I) {
11606 AccCond = AccumulateCond(AccCond, CC);
11607 if (
I != TC && Update != Comp)
11610 Counter = NextCounter;
11626 assert(Phi.getNumOperands() == 5);
11627 if (Phi.getOperand(2).getMBB() ==
MBB) {
11628 RegMBB = Phi.getOperand(1).getReg();
11629 RegOther = Phi.getOperand(3).getReg();
11631 assert(Phi.getOperand(4).getMBB() ==
MBB);
11632 RegMBB = Phi.getOperand(3).getReg();
11633 RegOther = Phi.getOperand(1).getReg();
11638 if (!
Reg.isVirtual())
11647 unsigned &UpdateCounterOprNum,
Register &InitReg,
11648 bool &IsUpdatePriorComp) {
11662 if (!
Reg.isVirtual())
11665 UpdateInst =
nullptr;
11666 UpdateCounterOprNum = 0;
11668 IsUpdatePriorComp =
true;
11672 if (Def->getParent() != LoopBB)
11674 if (Def->isCopy()) {
11676 if (Def->getOperand(0).getSubReg() || Def->getOperand(1).getSubReg())
11678 CurReg = Def->getOperand(1).getReg();
11679 }
else if (Def->isPHI()) {
11683 IsUpdatePriorComp =
false;
11688 switch (Def->getOpcode()) {
11689 case AArch64::ADDSXri:
11690 case AArch64::ADDSWri:
11691 case AArch64::SUBSXri:
11692 case AArch64::SUBSWri:
11693 case AArch64::ADDXri:
11694 case AArch64::ADDWri:
11695 case AArch64::SUBXri:
11696 case AArch64::SUBWri:
11698 UpdateCounterOprNum = 1;
11700 case AArch64::ADDSXrr:
11701 case AArch64::ADDSWrr:
11702 case AArch64::SUBSXrr:
11703 case AArch64::SUBSWrr:
11704 case AArch64::ADDXrr:
11705 case AArch64::ADDWrr:
11706 case AArch64::SUBXrr:
11707 case AArch64::SUBWrr:
11710 UpdateCounterOprNum = 1;
11712 UpdateCounterOprNum = 2;
11719 CurReg = Def->getOperand(UpdateCounterOprNum).getReg();
11734std::unique_ptr<TargetInstrInfo::PipelinerLoopInfo>
11745 if (
MI.isCall() ||
MI.hasUnmodeledSideEffects())
11756 if (
TBB == LoopBB && FBB == LoopBB)
11760 if (
TBB != LoopBB && FBB ==
nullptr)
11763 assert((
TBB == LoopBB || FBB == LoopBB) &&
11764 "The Loop must be a single-basic-block loop");
11769 if (CondBranch->
getOpcode() != AArch64::Bcc)
11777 unsigned CompCounterOprNum = 0;
11779 if (
MI.modifiesRegister(AArch64::NZCV, &
TRI)) {
11783 switch (
MI.getOpcode()) {
11784 case AArch64::SUBSXri:
11785 case AArch64::SUBSWri:
11786 case AArch64::ADDSXri:
11787 case AArch64::ADDSWri:
11789 CompCounterOprNum = 1;
11791 case AArch64::ADDSWrr:
11792 case AArch64::ADDSXrr:
11793 case AArch64::SUBSWrr:
11794 case AArch64::SUBSXrr:
11798 if (isWhileOpcode(
MI.getOpcode())) {
11805 if (CompCounterOprNum == 0) {
11807 CompCounterOprNum = 2;
11809 CompCounterOprNum = 1;
11821 bool IsUpdatePriorComp;
11822 unsigned UpdateCounterOprNum;
11824 Update, UpdateCounterOprNum,
Init, IsUpdatePriorComp))
11827 return std::make_unique<AArch64PipelinerLoopInfo>(
11828 LoopBB, CondBranch, Comp, CompCounterOprNum, Update, UpdateCounterOprNum,
11838 TypeSize Scale(0U,
false), Width(0U,
false);
11839 int64_t MinOffset, MaxOffset;
11840 if (
getMemOpInfo(
MI.getOpcode(), Scale, Width, MinOffset, MaxOffset)) {
11842 if (
MI.getOperand(ImmIdx).isImm() && !
MI.getOperand(ImmIdx - 1).isFI()) {
11843 int64_t Imm =
MI.getOperand(ImmIdx).getImm();
11844 if (Imm < MinOffset || Imm > MaxOffset) {
11845 ErrInfo =
"Unexpected immediate on load/store instruction";
11851 const MCInstrDesc &MCID =
MI.getDesc();
11853 const MachineOperand &MO =
MI.getOperand(
Op);
11857 ErrInfo =
"OPERAND_IMPLICIT_IMM_0 should be 0";
11866 ErrInfo =
"OPERAND_SHIFT_MSL should be msl shift of 8 or 16";
11877#define GET_INSTRINFO_HELPERS
11878#define GET_INSTRMAP_INFO
11879#include "AArch64GenInstrInfo.inc"
MachineInstrBuilder & UseMI
MachineInstrBuilder MachineInstrBuilder & DefMI
static bool forwardCopyWillClobberTuple(unsigned DestReg, unsigned SrcReg, unsigned NumRegs)
static cl::opt< unsigned > BCCDisplacementBits("aarch64-bcc-offset-bits", cl::Hidden, cl::init(19), cl::desc("Restrict range of Bcc instructions (DEBUG)"))
static Register genNeg(MachineFunction &MF, MachineRegisterInfo &MRI, const TargetInstrInfo *TII, MachineInstr &Root, SmallVectorImpl< MachineInstr * > &InsInstrs, DenseMap< Register, unsigned > &InstrIdxForVirtReg, unsigned MnegOpc, const TargetRegisterClass *RC)
genNeg - Helper to generate an intermediate negation of the second operand of Root
static bool isFrameStoreOpcode(int Opcode)
static cl::opt< unsigned > GatherOptSearchLimit("aarch64-search-limit", cl::Hidden, cl::init(2048), cl::desc("Restrict range of instructions to search for the " "machine-combiner gather pattern optimization"))
static bool getMaddPatterns(MachineInstr &Root, SmallVectorImpl< unsigned > &Patterns)
Find instructions that can be turned into madd.
static AArch64CC::CondCode findCondCodeUsedByInstr(const MachineInstr &Instr)
Find a condition code used by the instruction.
static MachineInstr * genFusedMultiplyAcc(MachineFunction &MF, MachineRegisterInfo &MRI, const TargetInstrInfo *TII, MachineInstr &Root, SmallVectorImpl< MachineInstr * > &InsInstrs, unsigned IdxMulOpd, unsigned MaddOpc, const TargetRegisterClass *RC)
genFusedMultiplyAcc - Helper to generate fused multiply accumulate instructions.
static MachineInstr * genFusedMultiplyAccNeg(MachineFunction &MF, MachineRegisterInfo &MRI, const TargetInstrInfo *TII, MachineInstr &Root, SmallVectorImpl< MachineInstr * > &InsInstrs, DenseMap< Register, unsigned > &InstrIdxForVirtReg, unsigned IdxMulOpd, unsigned MaddOpc, unsigned MnegOpc, const TargetRegisterClass *RC)
genFusedMultiplyAccNeg - Helper to generate fused multiply accumulate instructions with an additional...
static bool isCombineInstrCandidate64(unsigned Opc)
static bool isFrameLoadOpcode(int Opcode)
static unsigned removeCopies(const MachineRegisterInfo &MRI, unsigned VReg)
static bool areCFlagsAccessedBetweenInstrs(MachineBasicBlock::iterator From, MachineBasicBlock::iterator To, const TargetRegisterInfo *TRI, const AccessKind AccessToCheck=AK_All)
True when condition flags are accessed (either by writing or reading) on the instruction trace starti...
static bool getFMAPatterns(MachineInstr &Root, SmallVectorImpl< unsigned > &Patterns)
Floating-Point Support.
static bool isADDSRegImm(unsigned Opcode)
static bool isCheapCopy(const MachineInstr &MI, const AArch64RegisterInfo &RI)
static bool isANDOpcode(MachineInstr &MI)
static void appendOffsetComment(int NumBytes, llvm::raw_string_ostream &Comment, StringRef RegScale={})
static unsigned sForm(MachineInstr &Instr)
Get opcode of S version of Instr.
static bool isCombineInstrSettingFlag(unsigned Opc)
static bool getFNEGPatterns(MachineInstr &Root, SmallVectorImpl< unsigned > &Patterns)
static bool getIndVarInfo(Register Reg, const MachineBasicBlock *LoopBB, MachineInstr *&UpdateInst, unsigned &UpdateCounterOprNum, Register &InitReg, bool &IsUpdatePriorComp)
If Reg is an induction variable, return true and set some parameters.
static bool canPairLdStOpc(unsigned FirstOpc, unsigned SecondOpc)
static bool mustAvoidNeonAtMBBI(const AArch64Subtarget &Subtarget, MachineBasicBlock &MBB, MachineBasicBlock::iterator I)
Returns true if in a streaming call site region without SME-FA64.
static bool isPostIndexLdStOpcode(unsigned Opcode)
Return true if the opcode is a post-index ld/st instruction, which really loads from base+0.
static std::optional< unsigned > getLFIInstSizeInBytes(const MachineInstr &MI)
Return the maximum number of bytes of code the specified instruction may be after LFI rewriting.
static unsigned getBranchDisplacementBits(unsigned Opc)
static cl::opt< unsigned > CBDisplacementBits("aarch64-cb-offset-bits", cl::Hidden, cl::init(9), cl::desc("Restrict range of CB instructions (DEBUG)"))
static std::optional< ParamLoadedValue > describeORRLoadedValue(const MachineInstr &MI, Register DescribedReg, const TargetInstrInfo *TII, const TargetRegisterInfo *TRI)
If the given ORR instruction is a copy, and DescribedReg overlaps with the destination register then,...
static bool getFMULPatterns(MachineInstr &Root, SmallVectorImpl< unsigned > &Patterns)
static void appendReadRegExpr(SmallVectorImpl< char > &Expr, unsigned RegNum)
static MachineInstr * genMaddR(MachineFunction &MF, MachineRegisterInfo &MRI, const TargetInstrInfo *TII, MachineInstr &Root, SmallVectorImpl< MachineInstr * > &InsInstrs, unsigned IdxMulOpd, unsigned MaddOpc, unsigned VR, const TargetRegisterClass *RC)
genMaddR - Generate madd instruction and combine mul and add using an extra virtual register Example ...
static Register cloneInstr(const MachineInstr *MI, unsigned ReplaceOprNum, Register ReplaceReg, MachineBasicBlock &MBB, MachineBasicBlock::iterator InsertTo)
Clone an instruction from MI.
static bool scaleOffset(unsigned Opc, int64_t &Offset)
static bool canCombineWithFMUL(MachineBasicBlock &MBB, MachineOperand &MO, unsigned MulOpc)
unsigned scaledOffsetOpcode(unsigned Opcode, unsigned &Scale)
static MachineInstr * genFusedMultiplyIdx(MachineFunction &MF, MachineRegisterInfo &MRI, const TargetInstrInfo *TII, MachineInstr &Root, SmallVectorImpl< MachineInstr * > &InsInstrs, unsigned IdxMulOpd, unsigned MaddOpc, const TargetRegisterClass *RC)
genFusedMultiplyIdx - Helper to generate fused multiply accumulate instructions.
static MachineInstr * genIndexedMultiply(MachineInstr &Root, SmallVectorImpl< MachineInstr * > &InsInstrs, unsigned IdxDupOp, unsigned MulOpc, const TargetRegisterClass *RC, MachineRegisterInfo &MRI)
Fold (FMUL x (DUP y lane)) into (FMUL_indexed x y lane)
static bool isSUBSRegImm(unsigned Opcode)
static bool UpdateOperandRegClass(MachineInstr &Instr)
static const TargetRegisterClass * getRegClass(const MachineInstr &MI, Register Reg)
static bool isInStreamingCallSiteRegion(MachineBasicBlock &MBB, MachineBasicBlock::iterator I)
Returns true if the instruction at I is in a streaming call site region, within a single basic block.
static bool canCmpInstrBeRemoved(MachineInstr &MI, MachineInstr &CmpInstr, int CmpValue, const TargetRegisterInfo &TRI, SmallVectorImpl< MachineInstr * > &CCUseInstrs, bool &IsInvertCC)
unsigned unscaledOffsetOpcode(unsigned Opcode)
static bool getLoadPatterns(MachineInstr &Root, SmallVectorImpl< unsigned > &Patterns)
Search for patterns of LD instructions we can optimize.
static bool canInstrSubstituteCmpInstr(MachineInstr &MI, MachineInstr &CmpInstr, const TargetRegisterInfo &TRI)
Check if CmpInstr can be substituted by MI.
static UsedNZCV getUsedNZCV(AArch64CC::CondCode CC)
static bool isCombineInstrCandidateFP(const MachineInstr &Inst)
static void appendLoadRegExpr(SmallVectorImpl< char > &Expr, int64_t OffsetFromDefCFA)
static void appendConstantExpr(SmallVectorImpl< char > &Expr, int64_t Constant, dwarf::LocationAtom Operation)
static unsigned convertToNonFlagSettingOpc(const MachineInstr &MI)
Return the opcode that does not set flags when possible - otherwise return the original opcode.
static bool outliningCandidatesV8_3OpsConsensus(const outliner::Candidate &a, const outliner::Candidate &b)
static bool isCombineInstrCandidate32(unsigned Opc)
static void parseCondBranch(MachineInstr *LastInst, MachineBasicBlock *&Target, SmallVectorImpl< MachineOperand > &Cond)
static unsigned offsetExtendOpcode(unsigned Opcode)
static void loadRegPairFromStackSlot(const TargetRegisterInfo &TRI, MachineBasicBlock &MBB, MachineBasicBlock::iterator InsertBefore, const MCInstrDesc &MCID, Register DestReg, unsigned SubIdx0, unsigned SubIdx1, int FI, MachineMemOperand *MMO)
static void generateGatherLanePattern(MachineInstr &Root, SmallVectorImpl< MachineInstr * > &InsInstrs, SmallVectorImpl< MachineInstr * > &DelInstrs, DenseMap< Register, unsigned > &InstrIdxForVirtReg, unsigned Pattern, unsigned NumLanes)
Generate optimized instruction sequence for gather load patterns to improve Memory-Level Parallelism ...
static bool getMiscPatterns(MachineInstr &Root, SmallVectorImpl< unsigned > &Patterns)
Find other MI combine patterns.
static bool outliningCandidatesSigningKeyConsensus(const outliner::Candidate &a, const outliner::Candidate &b)
static const MachineInstrBuilder & AddSubReg(const MachineInstrBuilder &MIB, MCRegister Reg, unsigned SubIdx, RegState State, const TargetRegisterInfo *TRI)
static bool outliningCandidatesSigningScopeConsensus(const outliner::Candidate &a, const outliner::Candidate &b)
static bool shouldClusterFI(const MachineFrameInfo &MFI, int FI1, int64_t Offset1, unsigned Opcode1, int FI2, int64_t Offset2, unsigned Opcode2)
static cl::opt< unsigned > TBZDisplacementBits("aarch64-tbz-offset-bits", cl::Hidden, cl::init(14), cl::desc("Restrict range of TB[N]Z instructions (DEBUG)"))
static void extractPhiReg(const MachineInstr &Phi, const MachineBasicBlock *MBB, Register &RegMBB, Register &RegOther)
static MCCFIInstruction createDefCFAExpression(const TargetRegisterInfo &TRI, unsigned Reg, const StackOffset &Offset)
static bool isDefinedOutside(Register Reg, const MachineBasicBlock *BB)
static MachineInstr * genFusedMultiply(MachineFunction &MF, MachineRegisterInfo &MRI, const TargetInstrInfo *TII, MachineInstr &Root, SmallVectorImpl< MachineInstr * > &InsInstrs, unsigned IdxMulOpd, unsigned MaddOpc, const TargetRegisterClass *RC, FMAInstKind kind=FMAInstKind::Default, const Register *ReplacedAddend=nullptr)
genFusedMultiply - Generate fused multiply instructions.
static bool getGatherLanePattern(MachineInstr &Root, SmallVectorImpl< unsigned > &Patterns, unsigned LoadLaneOpCode, unsigned NumLanes)
Check if the given instruction forms a gather load pattern that can be optimized for better Memory-Le...
static MachineInstr * genFusedMultiplyIdxNeg(MachineFunction &MF, MachineRegisterInfo &MRI, const TargetInstrInfo *TII, MachineInstr &Root, SmallVectorImpl< MachineInstr * > &InsInstrs, DenseMap< Register, unsigned > &InstrIdxForVirtReg, unsigned IdxMulOpd, unsigned MaddOpc, unsigned MnegOpc, const TargetRegisterClass *RC)
genFusedMultiplyAccNeg - Helper to generate fused multiply accumulate instructions with an additional...
static bool isCombineInstrCandidate(unsigned Opc)
static unsigned regOffsetOpcode(unsigned Opcode)
MachineOutlinerClass
Constants defining how certain sequences should be outlined.
@ MachineOutlinerTailCall
Emit a save, restore, call, and return.
@ MachineOutlinerRegSave
Emit a call and tail-call.
@ MachineOutlinerNoLRSave
Only emit a branch.
@ MachineOutlinerThunk
Emit a call and return.
static cl::opt< unsigned > BDisplacementBits("aarch64-b-offset-bits", cl::Hidden, cl::init(26), cl::desc("Restrict range of B instructions (DEBUG)"))
static bool areCFlagsAliveInSuccessors(const MachineBasicBlock *MBB)
Check if AArch64::NZCV should be alive in successors of MBB.
static void emitFrameOffsetAdj(MachineBasicBlock &MBB, MachineBasicBlock::iterator MBBI, const DebugLoc &DL, unsigned DestReg, unsigned SrcReg, int64_t Offset, unsigned Opc, const TargetInstrInfo *TII, MachineInstr::MIFlag Flag, bool NeedsWinCFI, bool *HasWinCFI, bool EmitCFAOffset, StackOffset CFAOffset, unsigned FrameReg)
static bool isCheapImmediate(const MachineInstr &MI, unsigned BitSize)
static cl::opt< unsigned > CBZDisplacementBits("aarch64-cbz-offset-bits", cl::Hidden, cl::init(19), cl::desc("Restrict range of CB[N]Z instructions (DEBUG)"))
static void genSubAdd2SubSub(MachineFunction &MF, MachineRegisterInfo &MRI, const TargetInstrInfo *TII, MachineInstr &Root, SmallVectorImpl< MachineInstr * > &InsInstrs, SmallVectorImpl< MachineInstr * > &DelInstrs, unsigned IdxOpd1, DenseMap< Register, unsigned > &InstrIdxForVirtReg)
Do the following transformation A - (B + C) ==> (A - B) - C A - (B + C) ==> (A - C) - B.
static unsigned canFoldIntoCSel(const MachineRegisterInfo &MRI, unsigned VReg, unsigned *NewReg=nullptr)
static void signOutlinedFunction(MachineFunction &MF, MachineBasicBlock &MBB, const AArch64InstrInfo *TII, bool ShouldSignReturnAddr)
static MachineInstr * genFNegatedMAD(MachineFunction &MF, MachineRegisterInfo &MRI, const TargetInstrInfo *TII, MachineInstr &Root, SmallVectorImpl< MachineInstr * > &InsInstrs)
static bool canCombineWithMUL(MachineBasicBlock &MBB, MachineOperand &MO, unsigned MulOpc, unsigned ZeroReg)
static void storeRegPairToStackSlot(const TargetRegisterInfo &TRI, MachineBasicBlock &MBB, MachineBasicBlock::iterator InsertBefore, const MCInstrDesc &MCID, Register SrcReg, bool IsKill, unsigned SubIdx0, unsigned SubIdx1, int FI, MachineMemOperand *MMO)
assert(UImm &&(UImm !=~static_cast< T >(0)) &&"Invalid immediate!")
MachineBasicBlock MachineBasicBlock::iterator DebugLoc DL
MachineBasicBlock MachineBasicBlock::iterator MBBI
static const Function * getParent(const Value *V)
static GCRegistry::Add< ErlangGC > A("erlang", "erlang-compatible garbage collector")
static GCRegistry::Add< OcamlGC > B("ocaml", "ocaml 3.10-compatible GC")
DXIL Forward Handle Accesses
const HexagonInstrInfo * TII
Module.h This file contains the declarations for the Module class.
const AbstractManglingParser< Derived, Alloc >::OperatorInfo AbstractManglingParser< Derived, Alloc >::Ops[]
This file implements the LivePhysRegs utility for tracking liveness of physical registers.
Machine Check Debug Module
Register const TargetRegisterInfo * TRI
Promote Memory to Register
static MCRegister getReg(const MCDisassembler *D, unsigned RC, unsigned RegNo)
MachineInstr unsigned OpIdx
ConstantRange Range(APInt(BitWidth, Low), APInt(BitWidth, High))
PowerPC Reduce CR logical Operation
const SmallVectorImpl< MachineOperand > MachineBasicBlock * TBB
const SmallVectorImpl< MachineOperand > & Cond
This file declares the machine register scavenger class.
static bool contains(SmallPtrSetImpl< ConstantExpr * > &Cache, ConstantExpr *Expr, Constant *C)
This file defines the SmallSet class.
This file defines the SmallVector class.
This file defines the 'Statistic' class, which is designed to be an easy way to expose various metric...
#define STATISTIC(VARNAME, DESC)
#define DEBUG_WITH_TYPE(TYPE,...)
DEBUG_WITH_TYPE macro - This macro should be used by passes to emit debug information.
static bool canCombine(MachineBasicBlock &MBB, MachineOperand &MO, unsigned CombineOpc=0)
AArch64FunctionInfo - This class is derived from MachineFunctionInfo and contains private AArch64-spe...
SignReturnAddress getSignReturnAddressCondition() const
bool hasStreamingModeChanges() const
void setOutliningStyle(const std::string &Style)
bool branchProtectionPAuthLR() const
std::optional< bool > hasRedZone() const
static bool shouldSignReturnAddress(SignReturnAddress Condition, bool IsLRSpilled)
bool shouldSignWithBKey() const
static bool isHForm(const MachineInstr &MI)
Returns whether the instruction is in H form (16 bit operands)
void insertSelect(MachineBasicBlock &MBB, MachineBasicBlock::iterator MI, const DebugLoc &DL, Register DstReg, ArrayRef< MachineOperand > Cond, Register TrueReg, Register FalseReg) const override
static bool hasBTISemantics(const MachineInstr &MI)
Returns whether the instruction can be compatible with non-zero BTYPE.
static bool isQForm(const MachineInstr &MI)
Returns whether the instruction is in Q form (128 bit operands)
static bool getMemOpInfo(unsigned Opcode, TypeSize &Scale, TypeSize &Width, int64_t &MinOffset, int64_t &MaxOffset)
Returns true if opcode Opc is a memory operation.
static bool isTailCallReturnInst(const MachineInstr &MI)
Returns true if MI is one of the TCRETURN* instructions.
static bool isFPRCopy(const MachineInstr &MI)
Does this instruction rename an FPR without modifying bits?
MachineInstr * emitLdStWithAddr(MachineInstr &MemI, const ExtAddrMode &AM) const override
std::optional< DestSourcePair > isCopyInstrImpl(const MachineInstr &MI) const override
If the specific machine instruction is an instruction that moves/copies value from one register to an...
MachineBasicBlock * getBranchDestBlock(const MachineInstr &MI) const override
unsigned getInstSizeInBytes(const MachineInstr &MI) const override
GetInstSize - Return the number of bytes of code the specified instruction may be.
static bool isZExtLoad(const MachineInstr &MI)
Returns whether the instruction is a zero-extending load.
bool areMemAccessesTriviallyDisjoint(const MachineInstr &MIa, const MachineInstr &MIb) const override
void copyPhysReg(MachineBasicBlock &MBB, MachineBasicBlock::iterator I, const DebugLoc &DL, Register DestReg, Register SrcReg, bool KillSrc, bool RenamableDest=false, bool RenamableSrc=false) const override
static bool isGPRCopy(const MachineInstr &MI)
Does this instruction rename a GPR without modifying bits?
static unsigned convertToFlagSettingOpc(unsigned Opc)
Return the opcode that set flags when possible.
void createPauthEpilogueInstr(MachineBasicBlock &MBB, DebugLoc DL) const
Return true when there is potentially a faster code sequence for an instruction chain ending in Root.
bool isBranchOffsetInRange(unsigned BranchOpc, int64_t BrOffset) const override
bool canInsertSelect(const MachineBasicBlock &, ArrayRef< MachineOperand > Cond, Register, Register, Register, int &, int &, int &) const override
Register isLoadFromStackSlotPostFE(const MachineInstr &MI, int &FrameIndex) const override
Check for post-frame ptr elimination stack locations as well.
static const MachineOperand & getLdStOffsetOp(const MachineInstr &MI)
Returns the immediate offset operator of a load/store.
bool isCoalescableExtInstr(const MachineInstr &MI, Register &SrcReg, Register &DstReg, unsigned &SubIdx) const override
static std::optional< unsigned > getUnscaledLdSt(unsigned Opc)
Returns the unscaled load/store for the scaled load/store opcode, if there is a corresponding unscale...
static bool hasUnscaledLdStOffset(unsigned Opc)
Return true if it has an unscaled load/store offset.
static const MachineOperand & getLdStAmountOp(const MachineInstr &MI)
Returns the shift amount operator of a load/store.
static bool isPreLdSt(const MachineInstr &MI)
Returns whether the instruction is a pre-indexed load/store.
std::optional< ExtAddrMode > getAddrModeFromMemoryOp(const MachineInstr &MemI, const TargetRegisterInfo *TRI) const override
bool getMemOperandsWithOffsetWidth(const MachineInstr &MI, SmallVectorImpl< const MachineOperand * > &BaseOps, int64_t &Offset, bool &OffsetIsScalable, LocationSize &Width, const TargetRegisterInfo *TRI) const override
bool analyzeBranchPredicate(MachineBasicBlock &MBB, MachineBranchPredicate &MBP, bool AllowModify) const override
void insertIndirectBranch(MachineBasicBlock &MBB, MachineBasicBlock &NewDestBB, MachineBasicBlock &RestoreBB, const DebugLoc &DL, int64_t BrOffset, RegScavenger *RS) const override
void loadRegFromStackSlot(MachineBasicBlock &MBB, MachineBasicBlock::iterator MBBI, Register DestReg, int FrameIndex, const TargetRegisterClass *RC, Register VReg, unsigned SubReg=0, MachineInstr::MIFlag Flags=MachineInstr::NoFlags) const override
static bool isPairableLdStInst(const MachineInstr &MI)
Return true if pairing the given load or store may be paired with another.
void storeRegToStackSlot(MachineBasicBlock &MBB, MachineBasicBlock::iterator MBBI, Register SrcReg, bool isKill, int FrameIndex, const TargetRegisterClass *RC, Register VReg, MachineInstr::MIFlag Flags=MachineInstr::NoFlags) const override
static bool isSExtLoad(const MachineInstr &MI)
Returns whether the instruction is a sign-extending load.
const AArch64RegisterInfo & getRegisterInfo() const
getRegisterInfo - TargetInstrInfo is a superset of MRegister info.
static bool isPreSt(const MachineInstr &MI)
Returns whether the instruction is a pre-indexed store.
void insertNoop(MachineBasicBlock &MBB, MachineBasicBlock::iterator MI) const override
AArch64InstrInfo(const AArch64Subtarget &STI)
static bool isPairedLdSt(const MachineInstr &MI)
Returns whether the instruction is a paired load/store.
MachineInstr * foldMemoryOperandImpl(MachineFunction &MF, MachineInstr &MI, ArrayRef< unsigned > Ops, int FrameIndex, MachineInstr *&CopyMI, LiveIntervals *LIS=nullptr, VirtRegMap *VRM=nullptr) const override
bool getMemOperandWithOffsetWidth(const MachineInstr &MI, const MachineOperand *&BaseOp, int64_t &Offset, bool &OffsetIsScalable, TypeSize &Width, const TargetRegisterInfo *TRI) const
If OffsetIsScalable is set to 'true', the offset is scaled by vscale.
Register isLoadFromStackSlot(const MachineInstr &MI, int &FrameIndex) const override
bool reverseBranchCondition(SmallVectorImpl< MachineOperand > &Cond) const override
static bool isStridedAccess(const MachineInstr &MI)
Return true if the given load or store is a strided memory access.
bool shouldClusterMemOps(ArrayRef< const MachineOperand * > BaseOps1, int64_t Offset1, bool OffsetIsScalable1, ArrayRef< const MachineOperand * > BaseOps2, int64_t Offset2, bool OffsetIsScalable2, unsigned ClusterSize, unsigned NumBytes) const override
Detect opportunities for ldp/stp formation.
unsigned removeBranch(MachineBasicBlock &MBB, int *BytesRemoved=nullptr) const override
bool isThroughputPattern(unsigned Pattern) const override
Return true when a code sequence can improve throughput.
MachineOperand & getMemOpBaseRegImmOfsOffsetOperand(MachineInstr &LdSt) const
Return the immediate offset of the base register in a load/store LdSt.
bool analyzeBranch(MachineBasicBlock &MBB, MachineBasicBlock *&TBB, MachineBasicBlock *&FBB, SmallVectorImpl< MachineOperand > &Cond, bool AllowModify=false) const override
bool canFoldIntoAddrMode(const MachineInstr &MemI, Register Reg, const MachineInstr &AddrI, ExtAddrMode &AM) const override
static bool isLdStPairSuppressed(const MachineInstr &MI)
Return true if pairing the given load or store is hinted to be unprofitable.
Register isStoreToStackSlotPostFE(const MachineInstr &MI, int &FrameIndex) const override
Check for post-frame ptr elimination stack locations as well.
std::unique_ptr< TargetInstrInfo::PipelinerLoopInfo > analyzeLoopForPipelining(MachineBasicBlock *LoopBB) const override
bool isSchedulingBoundary(const MachineInstr &MI, const MachineBasicBlock *MBB, const MachineFunction &MF) const override
unsigned insertBranch(MachineBasicBlock &MBB, MachineBasicBlock *TBB, MachineBasicBlock *FBB, ArrayRef< MachineOperand > Cond, const DebugLoc &DL, int *BytesAdded=nullptr) const override
bool optimizeCompareInstr(MachineInstr &CmpInstr, Register SrcReg, Register SrcReg2, int64_t CmpMask, int64_t CmpValue, const MachineRegisterInfo *MRI) const override
optimizeCompareInstr - Convert the instruction supplying the argument to the comparison into one that...
static unsigned getLoadStoreImmIdx(unsigned Opc)
Returns the index for the immediate for a given instruction.
static bool isGPRZero(const MachineInstr &MI)
Does this instruction set its full destination register to zero?
void copyGPRRegTuple(MachineBasicBlock &MBB, MachineBasicBlock::iterator I, const DebugLoc &DL, MCRegister DestReg, MCRegister SrcReg, bool KillSrc, unsigned Opcode, unsigned ZeroReg, llvm::ArrayRef< unsigned > Indices) const
bool analyzeCompare(const MachineInstr &MI, Register &SrcReg, Register &SrcReg2, int64_t &CmpMask, int64_t &CmpValue) const override
analyzeCompare - For a comparison instruction, return the source registers in SrcReg and SrcReg2,...
CombinerObjective getCombinerObjective(unsigned Pattern) const override
static bool isFpOrNEON(Register Reg)
Returns whether the physical register is FP or NEON.
bool isAsCheapAsAMove(const MachineInstr &MI) const override
std::optional< DestSourcePair > isCopyLikeInstrImpl(const MachineInstr &MI) const override
static void suppressLdStPair(MachineInstr &MI)
Hint that pairing the given load or store is unprofitable.
Register isStoreToStackSlot(const MachineInstr &MI, int &FrameIndex) const override
static bool isPreLd(const MachineInstr &MI)
Returns whether the instruction is a pre-indexed load.
void copyPhysRegTuple(MachineBasicBlock &MBB, MachineBasicBlock::iterator I, const DebugLoc &DL, MCRegister DestReg, MCRegister SrcReg, bool KillSrc, unsigned Opcode, llvm::ArrayRef< unsigned > Indices) const
bool optimizeCondBranch(MachineInstr &MI) const override
Replace csincr-branch sequence by simple conditional branch.
static int getMemScale(unsigned Opc)
Scaling factor for (scaled or unscaled) load or store.
bool isCandidateToMergeOrPair(const MachineInstr &MI) const
Return true if this is a load/store that can be potentially paired/merged.
MCInst getNop() const override
static const MachineOperand & getLdStBaseOp(const MachineInstr &MI)
Returns the base register operator of a load/store.
bool isReservedReg(const MachineFunction &MF, MCRegister Reg) const
const AArch64RegisterInfo * getRegisterInfo() const override
bool isNeonAvailable() const
Returns true if the target has NEON and the function at runtime is known to have NEON enabled (e....
bool isSVEorStreamingSVEAvailable() const
Returns true if the target has access to either the full range of SVE instructions,...
ArrayRef - Represent a constant reference to an array (0 or more elements consecutively in memory),...
const T & front() const
front - Get the first element.
size_t size() const
size - Get the array size.
This is an important base class in LLVM.
std::pair< iterator, bool > insert(const std::pair< KeyT, ValueT > &KV)
bool hasOptSize() const
Optimize this function for size (-Os) or minimum size (-Oz).
bool hasMinSize() const
Optimize this function for minimum size (-Oz).
Module * getParent()
Get the module that this global value is contained inside of...
LLVM_ABI InstListType::iterator eraseFromParent()
This method unlinks 'this' from the containing basic block and deletes it.
unsigned getOpcode() const
Returns a member of one of the enums like Instruction::Add.
static LocationSize precise(uint64_t Value)
This class is intended to be used as a base class for asm properties and features specific to the tar...
bool usesWindowsCFI() const
static MCCFIInstruction cfiDefCfa(MCSymbol *L, unsigned Register, int64_t Offset, SMLoc Loc={})
.cfi_def_cfa defines a rule for computing CFA as: take address from Register and add Offset to it.
static MCCFIInstruction createOffset(MCSymbol *L, unsigned Register, int64_t Offset, SMLoc Loc={})
.cfi_offset Previous value of Register is saved at offset Offset from CFA.
static MCCFIInstruction cfiDefCfaOffset(MCSymbol *L, int64_t Offset, SMLoc Loc={})
.cfi_def_cfa_offset modifies a rule for computing CFA.
static MCCFIInstruction createEscape(MCSymbol *L, StringRef Vals, SMLoc Loc={}, StringRef Comment="")
.cfi_escape Allows the user to add arbitrary bytes to the unwind info.
Instances of this class represent a single low-level machine instruction.
Describe properties that are true of each instruction in the target description file.
unsigned getNumOperands() const
Return the number of declared MachineOperands for this MachineInstruction.
ArrayRef< MCOperandInfo > operands() const
Wrapper class representing physical registers. Should be passed by value.
constexpr bool isValid() const
static constexpr unsigned NoRegister
static MDTuple * get(LLVMContext &Context, ArrayRef< Metadata * > MDs)
bool isInlineAsmBrIndirectTarget() const
Returns true if this is the indirect dest of an INLINEASM_BR.
unsigned pred_size() const
LLVM_ABI void transferSuccessorsAndUpdatePHIs(MachineBasicBlock *FromMBB)
Transfers all the successors, as in transferSuccessors, and update PHI operands in the successor bloc...
LLVM_ABI instr_iterator insert(instr_iterator I, MachineInstr *M)
Insert MI into the instruction list before I, possibly inside a bundle.
reverse_instr_iterator instr_rbegin()
LLVM_ABI iterator getFirstTerminator()
Returns an iterator to the first terminator instruction of this basic block.
LLVM_ABI void addSuccessor(MachineBasicBlock *Succ, BranchProbability Prob=BranchProbability::getUnknown())
Add Succ as a successor of this MachineBasicBlock.
reverse_instr_iterator instr_rend()
Instructions::iterator instr_iterator
instr_iterator instr_end()
void addLiveIn(MCRegister PhysReg, LaneBitmask LaneMask=LaneBitmask::getAll())
Adds the specified register as a live in.
const MachineFunction * getParent() const
Return the MachineFunction containing this basic block.
LLVM_ABI instr_iterator erase(instr_iterator I)
Remove an instruction from the instruction list and delete it.
void splice(iterator Where, MachineBasicBlock *Other, iterator From)
Take an instruction from MBB 'Other' at the position From, and insert it into this MBB right before '...
MachineInstrBundleIterator< MachineInstr > iterator
void setMachineBlockAddressTaken()
Set this block to indicate that its address is used as something other than the target of a terminato...
LLVM_ABI bool isLiveIn(MCRegister Reg, LaneBitmask LaneMask=LaneBitmask::getAll()) const
Return true if the specified register is in the live in set.
The MachineFrameInfo class represents an abstract stack frame until prolog/epilog code is inserted.
uint64_t getStackSize() const
Return the number of bytes that must be allocated to hold all of the fixed size frame objects.
void setStackID(int ObjectIdx, uint8_t ID)
bool isCalleeSavedInfoValid() const
Has the callee saved info been calculated yet?
Align getObjectAlign(int ObjectIdx) const
Return the alignment of the specified stack object.
int64_t getObjectSize(int ObjectIdx) const
Return the size of the specified object.
unsigned getNumObjects() const
Return the number of objects.
int64_t getObjectOffset(int ObjectIdx) const
Return the assigned stack offset of the specified object from the incoming stack pointer.
bool isFixedObjectIndex(int ObjectIdx) const
Returns true if the specified index corresponds to a fixed stack object.
unsigned addFrameInst(const MCCFIInstruction &Inst)
const TargetSubtargetInfo & getSubtarget() const
getSubtarget - Return the subtarget for which this machine code is being compiled.
StringRef getName() const
getName - Return the name of the corresponding LLVM function.
MachineMemOperand * getMachineMemOperand(MachinePointerInfo PtrInfo, MachineMemOperand::Flags f, LLT MemTy, Align base_alignment, const AAMDNodes &AAInfo=AAMDNodes(), const MDNode *Ranges=nullptr, SyncScope::ID SSID=SyncScope::System, AtomicOrdering Ordering=AtomicOrdering::NotAtomic, AtomicOrdering FailureOrdering=AtomicOrdering::NotAtomic)
getMachineMemOperand - Allocate a new MachineMemOperand.
MachineFrameInfo & getFrameInfo()
getFrameInfo - Return the frame info object for the current function.
MachineRegisterInfo & getRegInfo()
getRegInfo - Return information about the registers currently in use.
Function & getFunction()
Return the LLVM function that this machine code represents.
BasicBlockListType::iterator iterator
Ty * getInfo()
getInfo - Keep track of various per-function pieces of information for backends that would like to do...
const MachineJumpTableInfo * getJumpTableInfo() const
getJumpTableInfo - Return the jump table info object for the current function.
MachineBasicBlock * CreateMachineBasicBlock(const BasicBlock *BB=nullptr, std::optional< UniqueBBID > BBID=std::nullopt)
CreateMachineInstr - Allocate a new MachineInstr.
void insert(iterator MBBI, MachineBasicBlock *MBB)
const TargetMachine & getTarget() const
getTarget - Return the target machine this machine code is compiled with
const MachineInstrBuilder & setMemRefs(ArrayRef< MachineMemOperand * > MMOs) const
const MachineInstrBuilder & addCFIIndex(unsigned CFIIndex) const
const MachineInstrBuilder & addUse(Register RegNo, RegState Flags={}, unsigned SubReg=0) const
Add a virtual register use operand.
const MachineInstrBuilder & addReg(Register RegNo, RegState Flags={}, unsigned SubReg=0) const
Add a new virtual register operand.
const MachineInstrBuilder & setMIFlag(MachineInstr::MIFlag Flag) const
const MachineInstrBuilder & addImm(int64_t Val) const
Add a new immediate operand.
const MachineInstrBuilder & add(const MachineOperand &MO) const
const MachineInstrBuilder & addSym(MCSymbol *Sym, unsigned char TargetFlags=0) const
const MachineInstrBuilder & addFrameIndex(int Idx) const
const MachineInstrBuilder & addGlobalAddress(const GlobalValue *GV, int64_t Offset=0, unsigned TargetFlags=0) const
const MachineInstrBuilder & addMBB(MachineBasicBlock *MBB, unsigned TargetFlags=0) const
const MachineInstrBuilder & addDef(Register RegNo, RegState Flags={}, unsigned SubReg=0) const
Add a virtual register definition operand.
const MachineInstrBuilder & setMIFlags(unsigned Flags) const
const MachineInstrBuilder & addMemOperand(MachineMemOperand *MMO) const
reverse_iterator getReverse() const
Get a reverse iterator to the same node.
Representation of each machine instruction.
unsigned getOpcode() const
Returns the opcode of this MachineInstr.
bool mayLoadOrStore(QueryType Type=AnyInBundle) const
Return true if this instruction could possibly read or modify memory.
const MachineBasicBlock * getParent() const
bool isCall(QueryType Type=AnyInBundle) const
bool getFlag(MIFlag Flag) const
Return whether an MI flag is set.
LLVM_ABI uint32_t mergeFlagsWith(const MachineInstr &Other) const
Return the MIFlags which represent both MachineInstrs.
unsigned getNumOperands() const
Retuns the total number of operands.
LLVM_ABI unsigned getNumExplicitOperands() const
Returns the number of non-implicit operands.
bool mayLoad(QueryType Type=AnyInBundle) const
Return true if this instruction could possibly read memory.
const MCInstrDesc & getDesc() const
Returns the target instruction descriptor of this MachineInstr.
LLVM_ABI bool hasUnmodeledSideEffects() const
Return true if this instruction has side effects that are not modeled by mayLoad / mayStore,...
bool registerDefIsDead(Register Reg, const TargetRegisterInfo *TRI) const
Returns true if the register is dead in this machine instruction.
bool definesRegister(Register Reg, const TargetRegisterInfo *TRI) const
Return true if the MachineInstr fully defines the specified register.
LLVM_ABI void setDesc(const MCInstrDesc &TID)
Replace the instruction descriptor (thus opcode) of the current instruction with a new one.
LLVM_ABI bool hasOrderedMemoryRef() const
Return true if this instruction may have an ordered or volatile memory reference, or if the informati...
LLVM_ABI const MachineFunction * getMF() const
Return the function that contains the basic block that this instruction belongs to.
ArrayRef< MachineMemOperand * > memoperands() const
Access to memory operands of the instruction.
LLVM_ABI bool isLoadFoldBarrier() const
Returns true if it is illegal to fold a load across this instruction.
const DebugLoc & getDebugLoc() const
Returns the debug location id of this MachineInstr.
LLVM_ABI void removeOperand(unsigned OpNo)
Erase an operand from an instruction, leaving it with one fewer operand than it started with.
LLVM_ABI void addRegisterDefined(Register Reg, const TargetRegisterInfo *RegInfo=nullptr)
We have determined MI defines a register.
const MachineOperand & getOperand(unsigned i) const
uint32_t getFlags() const
Return the MI flags bitvector.
LLVM_ABI int findRegisterDefOperandIdx(Register Reg, const TargetRegisterInfo *TRI, bool isDead=false, bool Overlap=false) const
Returns the operand index that is a def of the specified register or -1 if it is not found.
LLVM_ABI MachineInstrBundleIterator< MachineInstr > eraseFromParent()
Unlink 'this' from the containing basic block and delete it.
const std::vector< MachineJumpTableEntry > & getJumpTables() const
A description of a memory reference used in the backend.
@ MOVolatile
The memory access is volatile.
@ MOLoad
The memory access reads data.
@ MOStore
The memory access writes data.
This class contains meta information specific to a module.
LLVM_ABI MachineFunction * getMachineFunction(const Function &F) const
Returns the MachineFunction associated to IR function F if there is one, otherwise nullptr.
MachineOperand class - Representation of each machine instruction operand.
void setSubReg(unsigned subReg)
unsigned getSubReg() const
void setImm(int64_t immVal)
bool isReg() const
isReg - Tests if this is a MO_Register operand.
MachineBasicBlock * getMBB() const
void setIsDead(bool Val=true)
LLVM_ABI void setReg(Register Reg)
Change the register this operand corresponds to.
bool isImm() const
isImm - Tests if this is a MO_Immediate operand.
void setIsKill(bool Val=true)
MachineInstr * getParent()
getParent - Return the instruction that this operand belongs to.
unsigned getTargetFlags() const
static MachineOperand CreateImm(int64_t Val)
MachineOperandType getType() const
getType - Returns the MachineOperandType for this operand.
void setIsUndef(bool Val=true)
Register getReg() const
getReg - Returns the register number.
bool isFI() const
isFI - Tests if this is a MO_FrameIndex operand.
LLVM_ABI bool isIdenticalTo(const MachineOperand &Other) const
Returns true if this operand is identical to the specified operand except for liveness related flags ...
static MachineOperand CreateReg(Register Reg, bool isDef, bool isImp=false, bool isKill=false, bool isDead=false, bool isUndef=false, bool isEarlyClobber=false, unsigned SubReg=0, bool isDebug=false, bool isInternalRead=false, bool isRenamable=false)
MachineRegisterInfo - Keep track of information for virtual and physical registers,...
LLVM_ABI bool hasOneNonDBGUse(Register RegNo) const
hasOneNonDBGUse - Return true if there is exactly one non-Debug use of the specified register.
bool tracksLiveness() const
tracksLiveness - Returns true when tracking register liveness accurately.
const TargetRegisterClass * getRegClass(Register Reg) const
Return the register class of the specified virtual register.
LLVM_ABI void clearKillFlags(Register Reg) const
clearKillFlags - Iterate over all the uses of the given register and clear the kill flag from the Mac...
LLVM_ABI MachineInstr * getVRegDef(Register Reg) const
getVRegDef - Return the machine instr that defines the specified virtual register or null if none is ...
bool use_nodbg_empty(Register RegNo) const
use_nodbg_empty - Return true if there are no non-Debug instructions using the specified register.
LLVM_ABI Register createVirtualRegister(const TargetRegisterClass *RegClass, StringRef Name="")
createVirtualRegister - Create and return a new virtual register in the function with the specified r...
bool reservedRegsFrozen() const
reservedRegsFrozen - Returns true after freezeReservedRegs() was called to ensure the set of reserved...
bool def_empty(Register RegNo) const
def_empty - Return true if there are no instructions defining the specified register (it may be live-...
use_instr_nodbg_iterator use_instr_nodbg_begin(Register RegNo) const
bool hasOneDef(Register RegNo) const
Return true if there is exactly one operand defining the specified register.
const TargetRegisterClass * getRegClassOrNull(Register Reg) const
Return the register class of Reg, or null if Reg has not been assigned a register class yet.
LLVM_ABI const TargetRegisterClass * constrainRegClass(Register Reg, const TargetRegisterClass *RC, unsigned MinNumRegs=0)
constrainRegClass - Constrain the register class of the specified virtual register to be a common sub...
LLVM_ABI MachineInstr * getUniqueVRegDef(Register Reg) const
getUniqueVRegDef - Return the unique machine instr that defines the specified virtual register or nul...
A Module instance is used to store all the information related to an LLVM module.
MI-level patchpoint operands.
uint32_t getNumPatchBytes() const
Return the number of patchable bytes the given patchpoint should emit.
Wrapper class representing virtual and physical registers.
MCRegister asMCReg() const
Utility to check-convert this value to a MCRegister.
constexpr bool isValid() const
constexpr bool isVirtual() const
Return true if the specified register number is in the virtual register namespace.
static constexpr bool isVirtualRegister(unsigned Reg)
Return true if the specified register number is in the virtual register namespace.
constexpr bool isPhysical() const
Return true if the specified register number is in the physical register namespace.
Represents a location in source code.
bool erase(PtrType Ptr)
Remove pointer from the set.
std::pair< iterator, bool > insert(PtrType Ptr)
Inserts Ptr if and only if there is no element in the container equal to Ptr.
SmallPtrSet - This class implements a set which is optimized for holding SmallSize or less elements.
SmallSet - This maintains a set of unique values, optimizing for the case when the set is small (less...
SmallString - A SmallString is just a SmallVector with methods and accessors that make it work better...
void append(StringRef RHS)
Append from a StringRef.
StringRef str() const
Explicit conversion to StringRef.
This class consists of common code factored out of the SmallVector class to reduce code duplication b...
void push_back(const T &Elt)
This is a 'vector' (really, a variable-sized array), optimized for the case when the array is small.
MI-level stackmap operands.
uint32_t getNumPatchBytes() const
Return the number of patchable bytes the given stackmap should emit.
StackOffset holds a fixed and a scalable offset in bytes.
int64_t getFixed() const
Returns the fixed component of the stack.
int64_t getScalable() const
Returns the scalable component of the stack.
static StackOffset get(int64_t Fixed, int64_t Scalable)
static StackOffset getScalable(int64_t Scalable)
static StackOffset getFixed(int64_t Fixed)
MI-level Statepoint operands.
uint32_t getNumPatchBytes() const
Return the number of patchable bytes the given statepoint should emit.
StringRef - Represent a constant reference to a string, i.e.
Object returned by analyzeLoopForPipelining.
TargetInstrInfo - Interface to description of machine instruction set.
virtual void genAlternativeCodeSequence(MachineInstr &Root, unsigned Pattern, SmallVectorImpl< MachineInstr * > &InsInstrs, SmallVectorImpl< MachineInstr * > &DelInstrs, DenseMap< Register, unsigned > &InstIdxForVirtReg) const
When getMachineCombinerPatterns() finds patterns, this function generates the instructions that could...
virtual std::optional< ParamLoadedValue > describeLoadedValue(const MachineInstr &MI, Register Reg) const
Produce the expression describing the MI loading a value into the physical register Reg.
virtual bool getMachineCombinerPatterns(MachineInstr &Root, SmallVectorImpl< unsigned > &Patterns, bool DoRegPressureReduce) const
Return true when there is potentially a faster code sequence for an instruction chain ending in Root.
virtual bool isSchedulingBoundary(const MachineInstr &MI, const MachineBasicBlock *MBB, const MachineFunction &MF) const
Test if the given instruction should be considered a scheduling boundary.
virtual CombinerObjective getCombinerObjective(unsigned Pattern) const
Return the objective of a combiner pattern.
virtual bool isFunctionSafeToSplit(const MachineFunction &MF) const
Return true if the function is a viable candidate for machine function splitting.
CodeModel::Model getCodeModel() const
Returns the code model.
const MCAsmInfo * getMCAsmInfo() const
Return target specific asm information.
bool contains(Register Reg) const
Return true if the specified register is included in this register class.
bool hasSubClassEq(const TargetRegisterClass *RC) const
Returns true if RC is a sub-class of or equal to this class.
bool hasSuperClassEq(const TargetRegisterClass *RC) const
Returns true if RC is a super-class of or equal to this class.
TargetRegisterInfo base class - We assume that the target defines a static array of TargetRegisterDes...
TargetSubtargetInfo - Generic base class for all target subtargets.
virtual const TargetInstrInfo * getInstrInfo() const
virtual const TargetRegisterInfo * getRegisterInfo() const =0
Return the target's register information.
Target - Wrapper for Target specific information.
static constexpr TypeSize getFixed(ScalarTy ExactSize)
static constexpr TypeSize getScalable(ScalarTy MinimumSize)
Value * getOperand(unsigned i) const
constexpr bool isScalable() const
Returns whether the quantity is scaled by a runtime quantity (vscale).
constexpr ScalarTy getKnownMinValue() const
Returns the minimum value this quantity can represent.
self_iterator getIterator()
A raw_ostream that writes to an std::string.
#define llvm_unreachable(msg)
Marks that the current location is not supposed to be reachable.
static CondCode getInvertedCondCode(CondCode Code)
@ MO_DLLIMPORT
MO_DLLIMPORT - On a symbol operand, this represents that the reference to the symbol is for an import...
@ MO_NC
MO_NC - Indicates whether the linker is expected to check the symbol reference for overflow.
@ MO_G1
MO_G1 - A symbol operand with this flag (granule 1) represents the bits 16-31 of a 64-bit address,...
@ MO_S
MO_S - Indicates that the bits of the symbol operand represented by MO_G0 etc are signed.
@ MO_PAGEOFF
MO_PAGEOFF - A symbol operand with this flag represents the offset of that symbol within a 4K page.
@ MO_GOT
MO_GOT - This flag indicates that a symbol operand represents the address of the GOT entry for the sy...
@ MO_PREL
MO_PREL - Indicates that the bits of the symbol operand represented by MO_G0 etc are PC relative.
@ MO_G0
MO_G0 - A symbol operand with this flag (granule 0) represents the bits 0-15 of a 64-bit address,...
@ MO_ARM64EC_CALLMANGLE
MO_ARM64EC_CALLMANGLE - Operand refers to the Arm64EC-mangled version of a symbol,...
@ MO_PAGE
MO_PAGE - A symbol operand with this flag represents the pc-relative offset of the 4K page containing...
@ MO_HI12
MO_HI12 - This flag indicates that a symbol operand represents the bits 13-24 of a 64-bit address,...
@ MO_TLS
MO_TLS - Indicates that the operand being accessed is some kind of thread-local symbol.
@ MO_G2
MO_G2 - A symbol operand with this flag (granule 2) represents the bits 32-47 of a 64-bit address,...
@ MO_TAGGED
MO_TAGGED - With MO_PAGE, indicates that the page includes a memory tag in bits 56-63.
@ MO_G3
MO_G3 - A symbol operand with this flag (granule 3) represents the high 16-bits of a 64-bit address,...
@ MO_COFFSTUB
MO_COFFSTUB - On a symbol operand "FOO", this indicates that the reference is actually to the "....
unsigned getCheckerSizeInBytes(AuthCheckMethod Method)
Returns the number of bytes added by checkAuthenticatedRegister.
static uint64_t decodeLogicalImmediate(uint64_t val, unsigned regSize)
decodeLogicalImmediate - Decode a logical immediate value in the form "N:immr:imms" (where the immr a...
static unsigned getShiftValue(unsigned Imm)
getShiftValue - Extract the shift value.
static unsigned getArithExtendImm(AArch64_AM::ShiftExtendType ET, unsigned Imm)
getArithExtendImm - Encode the extend type and shift amount for an arithmetic instruction: imm: 3-bit...
static unsigned getArithShiftValue(unsigned Imm)
getArithShiftValue - get the arithmetic shift value.
static uint64_t encodeLogicalImmediate(uint64_t imm, unsigned regSize)
encodeLogicalImmediate - Return the encoded immediate value for a logical immediate instruction of th...
static AArch64_AM::ShiftExtendType getExtendType(unsigned Imm)
getExtendType - Extract the extend type for operands of arithmetic ops.
static AArch64_AM::ShiftExtendType getArithExtendType(unsigned Imm)
static AArch64_AM::ShiftExtendType getShiftType(unsigned Imm)
getShiftType - Extract the shift type.
static unsigned getShifterImm(AArch64_AM::ShiftExtendType ST, unsigned Imm)
getShifterImm - Encode the shift type and amount: imm: 6-bit shift amount shifter: 000 ==> lsl 001 ==...
void expandMOVImm(uint64_t Imm, unsigned BitSize, SmallVectorImpl< ImmInsnModel > &Insn)
Expand a MOVi32imm or MOVi64imm pseudo instruction to one or more real move-immediate instructions to...
static const uint64_t InstrFlagIsWhile
static const uint64_t InstrFlagIsPTestLike
constexpr std::underlying_type_t< E > Mask()
Get a bitmask with 1s in all places up to the high-order bit of E's largest value.
@ C
The default llvm calling convention, compatible with C.
@ ScalablePredicateVector
initializer< Ty > init(const Ty &Val)
InstrType
Represents how an instruction should be mapped by the outliner.
NodeAddr< InstrNode * > Instr
LLVM_ABI Instruction & back() const
LLVM_ABI iterator begin() const
This is an optimization pass for GlobalISel generic memory operations.
auto drop_begin(T &&RangeOrContainer, size_t N=1)
Return a range covering RangeOrContainer with the first N elements excluded.
bool all_of(R &&range, UnaryPredicate P)
Provide wrappers to std::all_of which take ranges instead of having to pass begin/end explicitly.
static bool isCondBranchOpcode(int Opc)
MCCFIInstruction createDefCFA(const TargetRegisterInfo &TRI, unsigned FrameReg, unsigned Reg, const StackOffset &Offset, bool LastAdjustmentWasScalable=true)
static bool isPTrueOpcode(unsigned Opc)
MachineInstrBuilder BuildMI(MachineFunction &MF, const MIMetadata &MIMD, const MCInstrDesc &MCID)
Builder interface. Specify how to create the initial instruction itself.
bool succeeded(LogicalResult Result)
Utility function that returns true if the provided LogicalResult corresponds to a success value.
int isAArch64FrameOffsetLegal(const MachineInstr &MI, StackOffset &Offset, bool *OutUseUnscaledOp=nullptr, unsigned *OutUnscaledOp=nullptr, int64_t *EmittableOffset=nullptr)
Check if the Offset is a valid frame offset for MI.
constexpr bool isInt(int64_t x)
Checks if an integer fits into the given bit width.
RegState
Flags to represent properties of register accesses.
@ Implicit
Not emitted register (e.g. carry, or temporary result).
@ Kill
The last use of a register.
@ Undef
Value of the register doesn't matter.
@ Define
Register definition.
@ Renamable
Register that may be renamed.
auto enumerate(FirstRange &&First, RestRanges &&...Rest)
Given two or more input ranges, returns a new range whose values are tuples (A, B,...
constexpr RegState getKillRegState(bool B)
decltype(auto) dyn_cast(const From &Val)
dyn_cast<X> - Return the argument parameter cast to the specified type.
static bool isIndirectBranchOpcode(int Opc)
iterator_range< T > make_range(T x, T y)
Convenience function for iterating over sub-ranges.
unsigned getBLRCallOpcode(const MachineFunction &MF)
Return opcode to be used for indirect calls.
@ AArch64FrameOffsetIsLegal
Offset is legal.
@ AArch64FrameOffsetCanUpdate
Offset can apply, at least partly.
@ AArch64FrameOffsetCannotUpdate
Offset cannot apply.
constexpr bool isPowerOf2_64(uint64_t Value)
Return true if the argument is a power of two > 0 (64 bit edition.)
unsigned Log2_64(uint64_t Value)
Return the floor log base 2 of the specified value, -1 if the value is zero.
static bool isSEHInstruction(const MachineInstr &MI)
bool any_of(R &&range, UnaryPredicate P)
Provide wrappers to std::any_of which take ranges instead of having to pass begin/end explicitly.
auto reverse(ContainerTy &&C)
MachineInstr * getImm(const MachineOperand &MO, const MachineRegisterInfo *MRI)
decltype(auto) get(const PointerIntPair< PointerTy, IntBits, IntType, PtrTraits, Info > &Pair)
void sort(IteratorTy Start, IteratorTy End)
AArch64MachineCombinerPattern
@ MULSUBv2i32_indexed_OP1
@ MULADDv4i16_indexed_OP2
@ MULSUBv8i16_indexed_OP2
@ MULSUBv4i16_indexed_OP2
@ MULSUBv4i32_indexed_OP2
@ MULADDv2i32_indexed_OP1
@ MULADDv4i32_indexed_OP1
@ MULADDv2i32_indexed_OP2
@ MULSUBv4i16_indexed_OP1
@ MULADDv4i32_indexed_OP2
@ MULSUBv8i16_indexed_OP1
@ MULSUBv2i32_indexed_OP2
@ MULADDv8i16_indexed_OP1
@ MULSUBv4i32_indexed_OP1
@ MULADDv8i16_indexed_OP2
@ MULADDv4i16_indexed_OP1
LLVM_ABI raw_ostream & dbgs()
dbgs() - This returns a reference to a raw_ostream for debugging messages.
void emitFrameOffset(MachineBasicBlock &MBB, MachineBasicBlock::iterator MBBI, const DebugLoc &DL, unsigned DestReg, unsigned SrcReg, StackOffset Offset, const TargetInstrInfo *TII, MachineInstr::MIFlag=MachineInstr::NoFlags, bool SetNZCV=false, bool NeedsWinCFI=false, bool *HasWinCFI=nullptr, bool EmitCFAOffset=false, StackOffset InitialOffset={}, unsigned FrameReg=AArch64::SP)
emitFrameOffset - Emit instructions as needed to set DestReg to SrcReg plus Offset.
LLVM_ABI void report_fatal_error(Error Err, bool gen_crash_diag=true)
constexpr RegState getDefRegState(bool B)
CombinerObjective
The combiner's goal may differ based on which pattern it is attempting to optimize.
std::optional< UsedNZCV > examineCFlagsUse(MachineInstr &MI, MachineInstr &CmpInstr, const TargetRegisterInfo &TRI, SmallVectorImpl< MachineInstr * > *CCUseInstrs=nullptr)
CodeGenOptLevel
Code generation optimization level.
class LLVM_GSL_OWNER SmallVector
Forward declaration of SmallVector so that calculateSmallVectorDefaultInlinedElements can reference s...
auto instructionsWithoutDebug(IterT It, IterT End, bool SkipPseudoOp=true)
Construct a range iterator which begins at It and moves forwards until End is reached,...
LLVM_ABI raw_fd_ostream & errs()
This returns a reference to a raw_ostream for standard error.
auto drop_end(T &&RangeOrContainer, size_t N=1)
Return a range covering RangeOrContainer with the last N elements excluded.
static MCRegister getXRegFromWReg(MCRegister Reg)
MCCFIInstruction createCFAOffset(const TargetRegisterInfo &MRI, unsigned Reg, const StackOffset &OffsetFromDefCFA, std::optional< int64_t > IncomingVGOffsetFromDefCFA)
FunctionAddr VTableAddr Next
DWARFExpression::Operation Op
ArrayRef(const T &OneElt) -> ArrayRef< T >
static bool isUncondBranchOpcode(int Opc)
decltype(auto) cast(const From &Val)
cast<X> - Return the argument parameter cast to the specified type.
void erase_if(Container &C, UnaryPredicate P)
Provide a container algorithm similar to C++ Library Fundamentals v2's erase_if which is equivalent t...
constexpr bool isIntN(unsigned N, int64_t x)
Checks if an signed integer fits into the given (dynamic) bit width.
bool rewriteAArch64FrameIndex(MachineInstr &MI, unsigned FrameRegIdx, unsigned FrameReg, StackOffset &Offset, const AArch64InstrInfo *TII)
rewriteAArch64FrameIndex - Rewrite MI to access 'Offset' bytes from the FP.
bool is_contained(R &&Range, const E &Element)
Returns true if Element is found in Range.
static const MachineMemOperand::Flags MOSuppressPair
constexpr int64_t SignExtend64(uint64_t x)
Sign-extend the number in the bottom B bits of X to a 64-bit integer.
void appendLEB128(SmallVectorImpl< U > &Buffer, T Value)
bool optimizeTerminators(MachineBasicBlock *MBB, const TargetInstrInfo &TII)
std::pair< MachineOperand, DIExpression * > ParamLoadedValue
bool isNZCVTouchedInInstructionRange(const MachineInstr &DefMI, const MachineInstr &UseMI, const TargetRegisterInfo *TRI)
Return true if there is an instruction /after/ DefMI and before UseMI which either reads or clobbers ...
auto seq(T Begin, T End)
Iterate over an integral type from Begin up to - but not including - End.
static const MachineMemOperand::Flags MOStridedAccess
constexpr RegState getUndefRegState(bool B)
void fullyRecomputeLiveIns(ArrayRef< MachineBasicBlock * > MBBs)
Convenience function for recomputing live-in's for a set of MBBs until the computation converges.
LLVM_ABI Printable printReg(Register Reg, const TargetRegisterInfo *TRI=nullptr, unsigned SubIdx=0, const MachineRegisterInfo *MRI=nullptr)
Prints virtual and physical registers with or without a TRI instance.
This struct is a compact representation of a valid (non-zero power of two) alignment.
Used to describe addressing mode similar to ExtAddrMode in CodeGenPrepare.
LLVM_ABI static const MBBSectionID ColdSectionID
This class contains a discriminated union of information about pointers in memory operands,...
static LLVM_ABI MachinePointerInfo getUnknownStack(MachineFunction &MF)
Stack memory without other information.
static LLVM_ABI MachinePointerInfo getFixedStack(MachineFunction &MF, int FI, int64_t Offset=0)
Return a MachinePointerInfo record that refers to the specified FrameIndex.
An individual sequence of instructions to be replaced with a call to an outlined function.
MachineFunction * getMF() const
The information necessary to create an outlined function for some class of candidate.