66#define GET_INSTRINFO_CTOR_DTOR
67#include "AArch64GenInstrInfo.inc"
69#define DEBUG_TYPE "AArch64InstrInfo"
71STATISTIC(NumCopyInstrs,
"Number of COPY instructions expanded");
72STATISTIC(NumZCRegMoveInstrsGPR,
"Number of zero-cycle GPR register move "
73 "instructions expanded from canonical COPY");
74STATISTIC(NumZCRegMoveInstrsFPR,
"Number of zero-cycle FPR register move "
75 "instructions expanded from canonical COPY");
76STATISTIC(NumZCZeroingInstrsGPR,
"Number of zero-cycle GPR zeroing "
77 "instructions expanded from canonical COPY");
82 cl::desc(
"Restrict range of CB instructions (DEBUG)"));
86 cl::desc(
"Restrict range of TB[N]Z instructions (DEBUG)"));
90 cl::desc(
"Restrict range of CB[N]Z instructions (DEBUG)"));
94 cl::desc(
"Restrict range of Bcc instructions (DEBUG)"));
98 cl::desc(
"Restrict range of B instructions (DEBUG)"));
102 cl::desc(
"Restrict range of instructions to search for the "
103 "machine-combiner gather pattern optimization"));
108 RI(STI.getTargetTriple(), STI.getHwMode()), Subtarget(STI) {}
119 auto Op =
MI.getOpcode();
120 if (
Op == AArch64::INLINEASM ||
Op == AArch64::INLINEASM_BR)
121 return getInlineAsmLength(
MI.getOperand(0).getSymbolName(), *MAI);
125 if (
MI.isMetaInstruction())
130 unsigned NumBytes = 0;
134 NumBytes =
Desc.getSize() ?
Desc.getSize() : 4;
137 if (!MFI->shouldSignReturnAddress(*MF))
141 auto Method = STI.getAuthenticatedLRCheckMethod(*MF);
149 switch (
Desc.getOpcode()) {
152 return Desc.getSize();
159 case TargetOpcode::STACKMAP:
162 assert(NumBytes % 4 == 0 &&
"Invalid number of NOP bytes requested!");
164 case TargetOpcode::PATCHPOINT:
167 assert(NumBytes % 4 == 0 &&
"Invalid number of NOP bytes requested!");
169 case TargetOpcode::STATEPOINT:
171 assert(NumBytes % 4 == 0 &&
"Invalid number of NOP bytes requested!");
176 case TargetOpcode::PATCHABLE_FUNCTION_ENTER:
181 F.getFnAttributeAsParsedInteger(
"patchable-function-entry", 9) * 4;
183 case TargetOpcode::PATCHABLE_FUNCTION_EXIT:
184 case TargetOpcode::PATCHABLE_TAIL_CALL:
185 case TargetOpcode::PATCHABLE_TYPED_EVENT_CALL:
189 case TargetOpcode::PATCHABLE_EVENT_CALL:
195 NumBytes =
MI.getOperand(1).getImm();
197 case TargetOpcode::BUNDLE:
198 NumBytes = getInstBundleLength(
MI);
205unsigned AArch64InstrInfo::getInstBundleLength(
const MachineInstr &
MI)
const {
209 while (++
I != E &&
I->isInsideBundle()) {
210 assert(!
I->isBundle() &&
"No nested bundle!");
245 case AArch64::CBWPri:
246 case AArch64::CBXPri:
247 case AArch64::CBWPrr:
248 case AArch64::CBXPrr:
256 case AArch64::CBBAssertExt:
257 case AArch64::CBHAssertExt:
288 case AArch64::CBWPri:
289 case AArch64::CBXPri:
290 case AArch64::CBBAssertExt:
291 case AArch64::CBHAssertExt:
292 case AArch64::CBWPrr:
293 case AArch64::CBXPrr:
299 int64_t BrOffset)
const {
301 assert(Bits >= 3 &&
"max branch displacement must be enough to jump"
302 "over conditional branch expansion");
303 return isIntN(Bits, BrOffset / 4);
308 switch (
MI.getOpcode()) {
312 return MI.getOperand(0).getMBB();
317 return MI.getOperand(2).getMBB();
323 return MI.getOperand(1).getMBB();
324 case AArch64::CBWPri:
325 case AArch64::CBXPri:
326 case AArch64::CBBAssertExt:
327 case AArch64::CBHAssertExt:
328 case AArch64::CBWPrr:
329 case AArch64::CBXPrr:
330 return MI.getOperand(3).getMBB();
340 assert(RS &&
"RegScavenger required for long branching");
342 "new block should be inserted for expanding unconditional branch");
345 "restore block should be inserted for restoring clobbered registers");
352 "Branch offsets outside of the signed 33-bit range not supported");
363 RS->enterBasicBlockEnd(
MBB);
366 constexpr Register Reg = AArch64::X16;
367 if (!RS->isRegUsed(Reg)) {
368 insertUnconditionalBranch(
MBB, &NewDestBB,
DL);
375 Register Scavenged = RS->FindUnusedReg(&AArch64::GPR64RegClass);
376 if (Scavenged != AArch64::NoRegister &&
378 buildIndirectBranch(Scavenged, NewDestBB);
379 RS->setRegUsed(Scavenged);
388 "Unable to insert indirect branch inside function that has red zone");
411 bool AllowModify)
const {
418 if (
I->getOpcode() == AArch64::SpeculationBarrierISBDSBEndBB ||
419 I->getOpcode() == AArch64::SpeculationBarrierSBEndBB) {
423 if (!isUnpredicatedTerminator(*
I))
430 unsigned LastOpc = LastInst->
getOpcode();
431 if (
I ==
MBB.begin() || !isUnpredicatedTerminator(*--
I)) {
446 unsigned SecondLastOpc = SecondLastInst->
getOpcode();
453 LastInst = SecondLastInst;
455 if (
I ==
MBB.begin() || !isUnpredicatedTerminator(*--
I)) {
460 SecondLastInst = &*
I;
461 SecondLastOpc = SecondLastInst->
getOpcode();
472 LastInst = SecondLastInst;
474 if (
I ==
MBB.begin() || !isUnpredicatedTerminator(*--
I)) {
476 "unreachable unconditional branches removed above");
485 SecondLastInst = &*
I;
486 SecondLastOpc = SecondLastInst->
getOpcode();
490 if (SecondLastInst &&
I !=
MBB.begin() && isUnpredicatedTerminator(*--
I))
506 I->eraseFromParent();
515 I->eraseFromParent();
524 MachineBranchPredicate &MBP,
525 bool AllowModify)
const {
537 assert(MBP.TrueDest &&
"expected!");
538 MBP.FalseDest = FBB ? FBB :
MBB.getNextNode();
540 MBP.ConditionDef =
nullptr;
541 MBP.SingleUseCondition =
false;
551 if (
I ==
MBB.begin())
567 if (
MI.modifiesRegister(AArch64::NZCV,
nullptr)) {
568 MBP.ConditionDef = &
MI;
577 case AArch64::CBNZX: {
581 MBP.Predicate = (
Opc == AArch64::CBNZX ||
Opc == AArch64::CBNZW)
582 ? MachineBranchPredicate::PRED_NE
583 : MachineBranchPredicate::PRED_EQ;
584 Register CondReg = MBP.LHS.getReg();
586 MBP.ConditionDef =
MRI.getVRegDef(CondReg);
593 case AArch64::TBNZX: {
596 MBP.ConditionDef =
MRI.getVRegDef(CondReg);
614 Cond[1].setImm(AArch64::CBNZW);
617 Cond[1].setImm(AArch64::CBZW);
620 Cond[1].setImm(AArch64::CBNZX);
623 Cond[1].setImm(AArch64::CBZX);
626 Cond[1].setImm(AArch64::TBNZW);
629 Cond[1].setImm(AArch64::TBZW);
632 Cond[1].setImm(AArch64::TBNZX);
635 Cond[1].setImm(AArch64::TBZX);
639 case AArch64::CBWPri:
640 case AArch64::CBXPri:
641 case AArch64::CBBAssertExt:
642 case AArch64::CBHAssertExt:
643 case AArch64::CBWPrr:
644 case AArch64::CBXPrr: {
657 int *BytesRemoved)
const {
667 I->eraseFromParent();
671 if (
I ==
MBB.begin()) {
684 I->eraseFromParent();
691void AArch64InstrInfo::instantiateCondBranch(
716 if (
Cond.size() > 5) {
727 assert(
TBB &&
"insertBranch must not be told to insert a fallthrough");
754 unsigned Opc =
MI.getOpcode();
761 if (
MI.getOperand(0).getReg() == AArch64::WZR ||
762 MI.getOperand(0).getReg() == AArch64::XZR) {
764 dbgs() <<
"Removing always taken branch: " <<
MI);
767 for (
auto *S : Succs)
769 MBB->removeSuccessor(S);
771 while (
MBB->rbegin() != &
MI)
772 MBB->rbegin()->eraseFromParent();
773 MI.eraseFromParent();
783 if (
MI.getOperand(0).getReg() == AArch64::WZR ||
784 MI.getOperand(0).getReg() == AArch64::XZR) {
786 dbgs() <<
"Removing never taken branch: " <<
MI);
788 MI.getParent()->removeSuccessor(
Target);
789 MI.eraseFromParent();
802 if (!
DefMI->isFullCopy())
804 VReg =
DefMI->getOperand(1).getReg();
813 unsigned *NewReg =
nullptr) {
818 bool Is64Bit = AArch64::GPR64allRegClass.hasSubClassEq(
MRI.getRegClass(VReg));
822 switch (
DefMI->getOpcode()) {
823 case AArch64::SUBREG_TO_REG:
827 if (!
DefMI->getOperand(1).isImm() ||
DefMI->getOperand(1).getImm() != 0)
829 if (!
DefMI->getOperand(2).isReg())
831 if (!
DefMI->getOperand(3).isImm() ||
832 DefMI->getOperand(3).getImm() != AArch64::sub_32)
835 if (
DefMI->getOpcode() != AArch64::MOVi32imm)
837 if (!
DefMI->getOperand(1).isImm() ||
DefMI->getOperand(1).getImm() != 1)
840 SrcReg = AArch64::XZR;
841 Opc = AArch64::CSINCXr;
844 case AArch64::MOVi32imm:
845 case AArch64::MOVi64imm:
846 if (!
DefMI->getOperand(1).isImm() ||
DefMI->getOperand(1).getImm() != 1)
848 SrcReg = Is64Bit ? AArch64::XZR : AArch64::WZR;
849 Opc = Is64Bit ? AArch64::CSINCXr : AArch64::CSINCWr;
852 case AArch64::ADDSXri:
853 case AArch64::ADDSWri:
855 if (
DefMI->findRegisterDefOperandIdx(AArch64::NZCV,
nullptr,
860 case AArch64::ADDXri:
861 case AArch64::ADDWri:
863 if (!
DefMI->getOperand(2).isImm() ||
DefMI->getOperand(2).getImm() != 1 ||
864 DefMI->getOperand(3).getImm() != 0)
866 SrcReg =
DefMI->getOperand(1).getReg();
867 Opc = Is64Bit ? AArch64::CSINCXr : AArch64::CSINCWr;
870 case AArch64::ORNXrr:
871 case AArch64::ORNWrr: {
874 if (ZReg != AArch64::XZR && ZReg != AArch64::WZR)
876 SrcReg =
DefMI->getOperand(2).getReg();
877 Opc = Is64Bit ? AArch64::CSINVXr : AArch64::CSINVWr;
881 case AArch64::SUBSXrr:
882 case AArch64::SUBSWrr:
884 if (
DefMI->findRegisterDefOperandIdx(AArch64::NZCV,
nullptr,
889 case AArch64::SUBXrr:
890 case AArch64::SUBWrr: {
893 if (ZReg != AArch64::XZR && ZReg != AArch64::WZR)
895 SrcReg =
DefMI->getOperand(2).getReg();
896 Opc = Is64Bit ? AArch64::CSNEGXr : AArch64::CSNEGWr;
902 assert(
Opc && SrcReg &&
"Missing parameters");
914 int &FalseCycles)
const {
918 RI.getCommonSubClass(
MRI.getRegClass(TrueReg),
MRI.getRegClass(FalseReg));
925 if (!RI.getCommonSubClass(RC,
MRI.getRegClass(DstReg)))
929 unsigned ExtraCondLat =
Cond.size() != 1;
933 if (AArch64::GPR64allRegClass.hasSubClassEq(RC) ||
934 AArch64::GPR32allRegClass.hasSubClassEq(RC)) {
936 CondCycles = 1 + ExtraCondLat;
937 TrueCycles = FalseCycles = 1;
947 if (AArch64::FPR64RegClass.hasSubClassEq(RC) ||
948 AArch64::FPR32RegClass.hasSubClassEq(RC)) {
949 CondCycles = 5 + ExtraCondLat;
950 TrueCycles = FalseCycles = 2;
967 switch (
Cond.size()) {
999 MRI.constrainRegClass(SrcReg, &AArch64::GPR64spRegClass);
1005 MRI.constrainRegClass(SrcReg, &AArch64::GPR32spRegClass);
1022 case AArch64::TBNZW:
1023 case AArch64::TBNZX:
1045 unsigned SubsOpc, SubsDestReg;
1051 case AArch64::CBWPri:
1052 SubsOpc = AArch64::SUBSWri;
1053 SubsDestReg = AArch64::WZR;
1056 case AArch64::CBXPri:
1057 SubsOpc = AArch64::SUBSXri;
1058 SubsDestReg = AArch64::XZR;
1061 case AArch64::CBWPrr:
1062 SubsOpc = AArch64::SUBSWrr;
1063 SubsDestReg = AArch64::WZR;
1066 case AArch64::CBXPrr:
1067 SubsOpc = AArch64::SUBSXrr;
1068 SubsDestReg = AArch64::XZR;
1097 switch (ExtendType) {
1103 "Unexpected compare-and-branch instruction for SXTB shift-extend");
1104 ExtOpc = AArch64::SBFMWri;
1110 "Unexpected compare-and-branch instruction for SXTH shift-extend");
1111 ExtOpc = AArch64::SBFMWri;
1117 "Unexpected compare-and-branch instruction for UXTB shift-extend");
1118 ExtOpc = AArch64::ANDWri;
1124 "Unexpected compare-and-branch instruction for UXTH shift-extend");
1125 ExtOpc = AArch64::ANDWri;
1131 Reg =
MRI.createVirtualRegister(&AArch64::GPR32spRegClass);
1134 if (ExtOpc != AArch64::ANDWri)
1136 MBBI.addImm(ExtBits);
1143 MRI.constrainRegClass(Reg,
MRI.getRegClass(
Cond[3].getReg()));
1144 MRI.constrainRegClass(
Cond[3].
getReg(), &AArch64::GPR32spRegClass);
1151 MRI.constrainRegClass(Reg,
MRI.getRegClass(
Cond[3].getReg()));
1152 MRI.constrainRegClass(
Cond[3].
getReg(), &AArch64::GPR32spRegClass);
1164 bool TryFold =
false;
1165 if (
MRI.constrainRegClass(DstReg, &AArch64::GPR64RegClass)) {
1166 RC = &AArch64::GPR64RegClass;
1167 Opc = AArch64::CSELXr;
1169 }
else if (
MRI.constrainRegClass(DstReg, &AArch64::GPR32RegClass)) {
1170 RC = &AArch64::GPR32RegClass;
1171 Opc = AArch64::CSELWr;
1173 }
else if (
MRI.constrainRegClass(DstReg, &AArch64::FPR64RegClass)) {
1174 RC = &AArch64::FPR64RegClass;
1175 Opc = AArch64::FCSELDrrr;
1176 }
else if (
MRI.constrainRegClass(DstReg, &AArch64::FPR32RegClass)) {
1177 RC = &AArch64::FPR32RegClass;
1178 Opc = AArch64::FCSELSrrr;
1180 assert(RC &&
"Unsupported regclass");
1184 unsigned NewReg = 0;
1199 MRI.clearKillFlags(NewReg);
1204 MRI.constrainRegClass(TrueReg, RC);
1207 (FalseReg.
isVirtual() || FalseReg == AArch64::WZR ||
1208 FalseReg == AArch64::XZR) &&
1209 "FalseReg was folded into a non-virtual register other than WZR or XZR");
1211 MRI.constrainRegClass(FalseReg, RC);
1226 assert(BitSize == 64 &&
"Only bit sizes of 32 or 64 allowed");
1231 return Is.
size() <= 2;
1236 assert(
MI.isCopy() &&
"Expected COPY instruction");
1242 if (
Reg.isVirtual())
1243 return MRI.getRegClass(
Reg);
1244 if (
Reg.isPhysical())
1245 return RI.getMinimalPhysRegClass(
Reg);
1250 if (DstRC && SrcRC && !RI.getCommonSubClass(DstRC, SrcRC))
1253 return MI.isAsCheapAsAMove();
1259 if (Subtarget.hasExynosCheapAsMoveHandling()) {
1260 if (isExynosCheapAsMove(
MI))
1262 return MI.isAsCheapAsAMove();
1265 switch (
MI.getOpcode()) {
1267 return MI.isAsCheapAsAMove();
1269 case TargetOpcode::COPY:
1272 case AArch64::ADDWrs:
1273 case AArch64::ADDXrs:
1274 case AArch64::SUBWrs:
1275 case AArch64::SUBXrs:
1276 return Subtarget.hasALULSLFast() &&
MI.getOperand(3).getImm() <= 4;
1281 case AArch64::MOVi32imm:
1283 case AArch64::MOVi64imm:
1288bool AArch64InstrInfo::isFalkorShiftExtFast(
const MachineInstr &
MI) {
1289 switch (
MI.getOpcode()) {
1293 case AArch64::ADDWrs:
1294 case AArch64::ADDXrs:
1295 case AArch64::ADDSWrs:
1296 case AArch64::ADDSXrs: {
1297 unsigned Imm =
MI.getOperand(3).getImm();
1304 case AArch64::ADDWrx:
1305 case AArch64::ADDXrx:
1306 case AArch64::ADDXrx64:
1307 case AArch64::ADDSWrx:
1308 case AArch64::ADDSXrx:
1309 case AArch64::ADDSXrx64: {
1310 unsigned Imm =
MI.getOperand(3).getImm();
1322 case AArch64::SUBWrs:
1323 case AArch64::SUBSWrs: {
1324 unsigned Imm =
MI.getOperand(3).getImm();
1326 return ShiftVal == 0 ||
1330 case AArch64::SUBXrs:
1331 case AArch64::SUBSXrs: {
1332 unsigned Imm =
MI.getOperand(3).getImm();
1334 return ShiftVal == 0 ||
1338 case AArch64::SUBWrx:
1339 case AArch64::SUBXrx:
1340 case AArch64::SUBXrx64:
1341 case AArch64::SUBSWrx:
1342 case AArch64::SUBSXrx:
1343 case AArch64::SUBSXrx64: {
1344 unsigned Imm =
MI.getOperand(3).getImm();
1356 case AArch64::LDRBBroW:
1357 case AArch64::LDRBBroX:
1358 case AArch64::LDRBroW:
1359 case AArch64::LDRBroX:
1360 case AArch64::LDRDroW:
1361 case AArch64::LDRDroX:
1362 case AArch64::LDRHHroW:
1363 case AArch64::LDRHHroX:
1364 case AArch64::LDRHroW:
1365 case AArch64::LDRHroX:
1366 case AArch64::LDRQroW:
1367 case AArch64::LDRQroX:
1368 case AArch64::LDRSBWroW:
1369 case AArch64::LDRSBWroX:
1370 case AArch64::LDRSBXroW:
1371 case AArch64::LDRSBXroX:
1372 case AArch64::LDRSHWroW:
1373 case AArch64::LDRSHWroX:
1374 case AArch64::LDRSHXroW:
1375 case AArch64::LDRSHXroX:
1376 case AArch64::LDRSWroW:
1377 case AArch64::LDRSWroX:
1378 case AArch64::LDRSroW:
1379 case AArch64::LDRSroX:
1380 case AArch64::LDRWroW:
1381 case AArch64::LDRWroX:
1382 case AArch64::LDRXroW:
1383 case AArch64::LDRXroX:
1384 case AArch64::PRFMroW:
1385 case AArch64::PRFMroX:
1386 case AArch64::STRBBroW:
1387 case AArch64::STRBBroX:
1388 case AArch64::STRBroW:
1389 case AArch64::STRBroX:
1390 case AArch64::STRDroW:
1391 case AArch64::STRDroX:
1392 case AArch64::STRHHroW:
1393 case AArch64::STRHHroX:
1394 case AArch64::STRHroW:
1395 case AArch64::STRHroX:
1396 case AArch64::STRQroW:
1397 case AArch64::STRQroX:
1398 case AArch64::STRSroW:
1399 case AArch64::STRSroX:
1400 case AArch64::STRWroW:
1401 case AArch64::STRWroX:
1402 case AArch64::STRXroW:
1403 case AArch64::STRXroX: {
1404 unsigned IsSigned =
MI.getOperand(3).getImm();
1411 unsigned Opc =
MI.getOpcode();
1415 case AArch64::SEH_StackAlloc:
1416 case AArch64::SEH_SaveFPLR:
1417 case AArch64::SEH_SaveFPLR_X:
1418 case AArch64::SEH_SaveReg:
1419 case AArch64::SEH_SaveReg_X:
1420 case AArch64::SEH_SaveRegP:
1421 case AArch64::SEH_SaveRegP_X:
1422 case AArch64::SEH_SaveFReg:
1423 case AArch64::SEH_SaveFReg_X:
1424 case AArch64::SEH_SaveFRegP:
1425 case AArch64::SEH_SaveFRegP_X:
1426 case AArch64::SEH_SetFP:
1427 case AArch64::SEH_AddFP:
1428 case AArch64::SEH_Nop:
1429 case AArch64::SEH_PrologEnd:
1430 case AArch64::SEH_EpilogStart:
1431 case AArch64::SEH_EpilogEnd:
1432 case AArch64::SEH_PACSignLR:
1433 case AArch64::SEH_SaveAnyRegI:
1434 case AArch64::SEH_SaveAnyRegIP:
1435 case AArch64::SEH_SaveAnyRegQP:
1436 case AArch64::SEH_SaveAnyRegQPX:
1437 case AArch64::SEH_AllocZ:
1438 case AArch64::SEH_SaveZReg:
1439 case AArch64::SEH_SavePReg:
1446 unsigned &SubIdx)
const {
1447 switch (
MI.getOpcode()) {
1450 case AArch64::SBFMXri:
1451 case AArch64::UBFMXri:
1454 if (
MI.getOperand(2).getImm() != 0 ||
MI.getOperand(3).getImm() != 31)
1457 SrcReg =
MI.getOperand(1).getReg();
1458 DstReg =
MI.getOperand(0).getReg();
1459 SubIdx = AArch64::sub_32;
1468 int64_t OffsetA = 0, OffsetB = 0;
1469 TypeSize WidthA(0,
false), WidthB(0,
false);
1470 bool OffsetAIsScalable =
false, OffsetBIsScalable =
false;
1491 OffsetAIsScalable == OffsetBIsScalable) {
1492 int LowOffset = OffsetA < OffsetB ? OffsetA : OffsetB;
1493 int HighOffset = OffsetA < OffsetB ? OffsetB : OffsetA;
1494 TypeSize LowWidth = (LowOffset == OffsetA) ? WidthA : WidthB;
1495 if (LowWidth.
isScalable() == OffsetAIsScalable &&
1513 switch (
MI.getOpcode()) {
1516 if (
MI.getOperand(0).getImm() == 0x14)
1523 case AArch64::MSRpstatesvcrImm1:
1530 auto Next = std::next(
MI.getIterator());
1531 return Next !=
MBB->end() &&
Next->isCFIInstruction();
1538 Register &SrcReg2, int64_t &CmpMask,
1539 int64_t &CmpValue)
const {
1543 assert(
MI.getNumOperands() >= 2 &&
"All AArch64 cmps should have 2 operands");
1544 if (!
MI.getOperand(1).isReg() ||
MI.getOperand(1).getSubReg())
1547 switch (
MI.getOpcode()) {
1550 case AArch64::PTEST_PP:
1551 case AArch64::PTEST_PP_ANY:
1552 case AArch64::PTEST_PP_FIRST:
1553 SrcReg =
MI.getOperand(0).getReg();
1554 SrcReg2 =
MI.getOperand(1).getReg();
1555 if (
MI.getOperand(2).getSubReg())
1562 case AArch64::SUBSWrr:
1563 case AArch64::SUBSWrs:
1564 case AArch64::SUBSWrx:
1565 case AArch64::SUBSXrr:
1566 case AArch64::SUBSXrs:
1567 case AArch64::SUBSXrx:
1568 case AArch64::ADDSWrr:
1569 case AArch64::ADDSWrs:
1570 case AArch64::ADDSWrx:
1571 case AArch64::ADDSXrr:
1572 case AArch64::ADDSXrs:
1573 case AArch64::ADDSXrx:
1575 SrcReg =
MI.getOperand(1).getReg();
1576 SrcReg2 =
MI.getOperand(2).getReg();
1579 if (
MI.getOperand(2).getSubReg())
1585 case AArch64::SUBSWri:
1586 case AArch64::ADDSWri:
1587 case AArch64::SUBSXri:
1588 case AArch64::ADDSXri:
1589 SrcReg =
MI.getOperand(1).getReg();
1592 CmpValue =
MI.getOperand(2).getImm();
1594 case AArch64::ANDSWri:
1595 case AArch64::ANDSXri:
1598 SrcReg =
MI.getOperand(1).getReg();
1602 MI.getOperand(2).getImm(),
1603 MI.getOpcode() == AArch64::ANDSWri ? 32 : 64);
1612 assert(
MBB &&
"Can't get MachineBasicBlock here");
1614 assert(MF &&
"Can't get MachineFunction here");
1619 for (
unsigned OpIdx = 0, EndIdx = Instr.getNumOperands();
OpIdx < EndIdx;
1626 if (!OpRegCstraints)
1634 "Operand has register constraints without being a register!");
1637 if (
Reg.isPhysical()) {
1641 !
MRI->constrainRegClass(
Reg, OpRegCstraints))
1654 bool MIDefinesZeroReg =
false;
1655 if (
MI.definesRegister(AArch64::WZR,
nullptr) ||
1656 MI.definesRegister(AArch64::XZR,
nullptr))
1657 MIDefinesZeroReg =
true;
1659 switch (
MI.getOpcode()) {
1661 return MI.getOpcode();
1662 case AArch64::ADDSWrr:
1663 return AArch64::ADDWrr;
1664 case AArch64::ADDSWri:
1665 return MIDefinesZeroReg ? AArch64::ADDSWri : AArch64::ADDWri;
1666 case AArch64::ADDSWrs:
1667 return MIDefinesZeroReg ? AArch64::ADDSWrs : AArch64::ADDWrs;
1668 case AArch64::ADDSWrx:
1669 return AArch64::ADDWrx;
1670 case AArch64::ADDSXrr:
1671 return AArch64::ADDXrr;
1672 case AArch64::ADDSXri:
1673 return MIDefinesZeroReg ? AArch64::ADDSXri : AArch64::ADDXri;
1674 case AArch64::ADDSXrs:
1675 return MIDefinesZeroReg ? AArch64::ADDSXrs : AArch64::ADDXrs;
1676 case AArch64::ADDSXrx:
1677 return AArch64::ADDXrx;
1678 case AArch64::SUBSWrr:
1679 return AArch64::SUBWrr;
1680 case AArch64::SUBSWri:
1681 return MIDefinesZeroReg ? AArch64::SUBSWri : AArch64::SUBWri;
1682 case AArch64::SUBSWrs:
1683 return MIDefinesZeroReg ? AArch64::SUBSWrs : AArch64::SUBWrs;
1684 case AArch64::SUBSWrx:
1685 return AArch64::SUBWrx;
1686 case AArch64::SUBSXrr:
1687 return AArch64::SUBXrr;
1688 case AArch64::SUBSXri:
1689 return MIDefinesZeroReg ? AArch64::SUBSXri : AArch64::SUBXri;
1690 case AArch64::SUBSXrs:
1691 return MIDefinesZeroReg ? AArch64::SUBSXrs : AArch64::SUBXrs;
1692 case AArch64::SUBSXrx:
1693 return AArch64::SUBXrx;
1708 if (To == To->getParent()->begin())
1713 if (To->getParent() != From->getParent())
1725 Instr.modifiesRegister(AArch64::NZCV,
TRI)) ||
1726 ((AccessToCheck &
AK_Read) && Instr.readsRegister(AArch64::NZCV,
TRI)))
1732std::optional<unsigned>
1736 unsigned MaskOpcode =
Mask->getOpcode();
1737 unsigned PredOpcode = Pred->
getOpcode();
1738 bool PredIsPTestLike = isPTestLikeOpcode(PredOpcode);
1739 bool PredIsWhileLike = isWhileOpcode(PredOpcode);
1741 if (PredIsWhileLike) {
1745 if ((Mask == Pred) && PTest->
getOpcode() == AArch64::PTEST_PP_ANY)
1752 getElementSizeForOpcode(MaskOpcode) ==
1753 getElementSizeForOpcode(PredOpcode))
1759 if (PTest->
getOpcode() == AArch64::PTEST_PP_FIRST &&
1766 if (PredIsPTestLike) {
1771 if ((Mask == Pred) && PTest->
getOpcode() == AArch64::PTEST_PP_ANY)
1779 if (Mask != PTestLikeMask && PTestLikeMask->isFullCopy() &&
1780 PTestLikeMask->getOperand(1).getReg().isVirtual())
1782 MRI->getUniqueVRegDef(PTestLikeMask->getOperand(1).getReg());
1788 getElementSizeForOpcode(MaskOpcode) ==
1789 getElementSizeForOpcode(PredOpcode)) {
1790 if (Mask == PTestLikeMask || PTest->
getOpcode() == AArch64::PTEST_PP_ANY)
1816 uint64_t PredElementSize = getElementSizeForOpcode(PredOpcode);
1818 PTest->
getOpcode() == AArch64::PTEST_PP_ANY))
1826 switch (PredOpcode) {
1827 case AArch64::AND_PPzPP:
1828 case AArch64::BIC_PPzPP:
1829 case AArch64::EOR_PPzPP:
1830 case AArch64::NAND_PPzPP:
1831 case AArch64::NOR_PPzPP:
1832 case AArch64::ORN_PPzPP:
1833 case AArch64::ORR_PPzPP:
1834 case AArch64::BRKA_PPzP:
1835 case AArch64::BRKPA_PPzPP:
1836 case AArch64::BRKB_PPzP:
1837 case AArch64::BRKPB_PPzPP:
1838 case AArch64::RDFFR_PPz: {
1842 if (Mask != PredMask)
1846 case AArch64::BRKN_PPzP: {
1850 if ((MaskOpcode != AArch64::PTRUE_B) ||
1851 (
Mask->getOperand(1).getImm() != 31))
1855 case AArch64::PTRUE_B:
1868bool AArch64InstrInfo::optimizePTestInstr(
1869 MachineInstr *PTest,
unsigned MaskReg,
unsigned PredReg,
1871 auto *
Mask =
MRI->getUniqueVRegDef(MaskReg);
1872 auto *Pred =
MRI->getUniqueVRegDef(PredReg);
1874 if (Pred->
isCopy() && PTest->
getOpcode() == AArch64::PTEST_PP_FIRST) {
1878 if (
Op.isReg() &&
Op.getReg().isVirtual() &&
1879 Op.getSubReg() == AArch64::psub0)
1880 Pred =
MRI->getUniqueVRegDef(
Op.getReg());
1883 unsigned PredOpcode = Pred->
getOpcode();
1884 auto NewOp = canRemovePTestInstr(PTest, Mask, Pred,
MRI);
1900 if (*NewOp != PredOpcode) {
1911 for (; i !=
e; ++i) {
1942 if (DeadNZCVIdx != -1) {
1961 if (CmpInstr.
getOpcode() == AArch64::PTEST_PP ||
1962 CmpInstr.
getOpcode() == AArch64::PTEST_PP_ANY ||
1963 CmpInstr.
getOpcode() == AArch64::PTEST_PP_FIRST)
1964 return optimizePTestInstr(&CmpInstr, SrcReg, SrcReg2,
MRI);
1973 if (CmpValue == 0 && substituteCmpToZero(CmpInstr, SrcReg, *
MRI))
1975 return (CmpValue == 0 || CmpValue == 1) &&
1976 removeCmpToZeroOrOne(CmpInstr, SrcReg, CmpValue, *
MRI);
1984 switch (Instr.getOpcode()) {
1986 return AArch64::INSTRUCTION_LIST_END;
1988 case AArch64::ADDSWrr:
1989 case AArch64::ADDSWri:
1990 case AArch64::ADDSXrr:
1991 case AArch64::ADDSXri:
1992 case AArch64::ADDSWrx:
1993 case AArch64::ADDSXrx:
1994 case AArch64::SUBSWrr:
1995 case AArch64::SUBSWri:
1996 case AArch64::SUBSWrx:
1997 case AArch64::SUBSXrr:
1998 case AArch64::SUBSXri:
1999 case AArch64::SUBSXrx:
2000 case AArch64::ANDSWri:
2001 case AArch64::ANDSWrr:
2002 case AArch64::ANDSWrs:
2003 case AArch64::ANDSXri:
2004 case AArch64::ANDSXrr:
2005 case AArch64::ANDSXrs:
2006 case AArch64::BICSWrr:
2007 case AArch64::BICSXrr:
2008 case AArch64::BICSWrs:
2009 case AArch64::BICSXrs:
2010 return Instr.getOpcode();
2012 case AArch64::ADDWrr:
2013 return AArch64::ADDSWrr;
2014 case AArch64::ADDWri:
2015 return AArch64::ADDSWri;
2016 case AArch64::ADDXrr:
2017 return AArch64::ADDSXrr;
2018 case AArch64::ADDXri:
2019 return AArch64::ADDSXri;
2020 case AArch64::ADDWrx:
2021 return AArch64::ADDSWrx;
2022 case AArch64::ADDXrx:
2023 return AArch64::ADDSXrx;
2024 case AArch64::ADCWr:
2025 return AArch64::ADCSWr;
2026 case AArch64::ADCXr:
2027 return AArch64::ADCSXr;
2028 case AArch64::SUBWrr:
2029 return AArch64::SUBSWrr;
2030 case AArch64::SUBWri:
2031 return AArch64::SUBSWri;
2032 case AArch64::SUBXrr:
2033 return AArch64::SUBSXrr;
2034 case AArch64::SUBXri:
2035 return AArch64::SUBSXri;
2036 case AArch64::SUBWrx:
2037 return AArch64::SUBSWrx;
2038 case AArch64::SUBXrx:
2039 return AArch64::SUBSXrx;
2040 case AArch64::SBCWr:
2041 return AArch64::SBCSWr;
2042 case AArch64::SBCXr:
2043 return AArch64::SBCSXr;
2044 case AArch64::ANDWri:
2045 return AArch64::ANDSWri;
2046 case AArch64::ANDXri:
2047 return AArch64::ANDSXri;
2048 case AArch64::ANDWrr:
2049 return AArch64::ANDSWrr;
2050 case AArch64::ANDWrs:
2051 return AArch64::ANDSWrs;
2052 case AArch64::ANDXrr:
2053 return AArch64::ANDSXrr;
2054 case AArch64::ANDXrs:
2055 return AArch64::ANDSXrs;
2056 case AArch64::BICWrr:
2057 return AArch64::BICSWrr;
2058 case AArch64::BICXrr:
2059 return AArch64::BICSXrr;
2060 case AArch64::BICWrs:
2061 return AArch64::BICSWrs;
2062 case AArch64::BICXrs:
2063 return AArch64::BICSXrs;
2069 for (
auto *BB :
MBB->successors())
2070 if (BB->isLiveIn(AArch64::NZCV))
2079 switch (Instr.getOpcode()) {
2083 case AArch64::Bcc: {
2084 int Idx = Instr.findRegisterUseOperandIdx(AArch64::NZCV,
nullptr);
2089 case AArch64::CSINVWr:
2090 case AArch64::CSINVXr:
2091 case AArch64::CSINCWr:
2092 case AArch64::CSINCXr:
2093 case AArch64::CSELWr:
2094 case AArch64::CSELXr:
2095 case AArch64::CSNEGWr:
2096 case AArch64::CSNEGXr:
2097 case AArch64::FCSELSrrr:
2098 case AArch64::FCSELDrrr: {
2099 int Idx = Instr.findRegisterUseOperandIdx(AArch64::NZCV,
nullptr);
2112 Instr.getOperand(CCIdx).
getImm())
2165std::optional<UsedNZCV>
2170 if (
MI.getParent() != CmpParent)
2171 return std::nullopt;
2174 return std::nullopt;
2179 if (Instr.readsRegister(AArch64::NZCV, &
TRI)) {
2182 return std::nullopt;
2187 if (Instr.modifiesRegister(AArch64::NZCV, &
TRI))
2190 return NZCVUsedAfterCmp;
2194 return Opcode == AArch64::ADDSWri || Opcode == AArch64::ADDSXri;
2198 return Opcode == AArch64::SUBSWri || Opcode == AArch64::SUBSXri;
2204 case AArch64::ANDSWri:
2205 case AArch64::ANDSWrr:
2206 case AArch64::ANDSWrs:
2207 case AArch64::ANDSXri:
2208 case AArch64::ANDSXrr:
2209 case AArch64::ANDSXrs:
2210 case AArch64::BICSWrr:
2211 case AArch64::BICSXrr:
2212 case AArch64::BICSWrs:
2213 case AArch64::BICSXrs:
2239 const unsigned CmpOpcode = CmpInstr.
getOpcode();
2245 "Caller guarantees that CmpInstr compares with constant 0");
2248 if (!NZVCUsed || NZVCUsed->C)
2271bool AArch64InstrInfo::substituteCmpToZero(
2275 MachineInstr *
MI =
MRI.getUniqueVRegDef(SrcReg);
2282 if (NewOpc == AArch64::INSTRUCTION_LIST_END)
2289 MI->setDesc(
get(NewOpc));
2294 MI->addRegisterDefined(AArch64::NZCV, &
TRI);
2306 assert((CmpValue == 0 || CmpValue == 1) &&
2307 "Only comparisons to 0 or 1 considered for removal!");
2310 unsigned MIOpc =
MI.getOpcode();
2311 if (MIOpc == AArch64::CSINCWr) {
2312 if (
MI.getOperand(1).getReg() != AArch64::WZR ||
2313 MI.getOperand(2).getReg() != AArch64::WZR)
2315 }
else if (MIOpc == AArch64::CSINCXr) {
2316 if (
MI.getOperand(1).getReg() != AArch64::XZR ||
2317 MI.getOperand(2).getReg() != AArch64::XZR)
2327 if (
MI.findRegisterDefOperandIdx(AArch64::NZCV,
nullptr,
true) != -1)
2331 const unsigned CmpOpcode = CmpInstr.
getOpcode();
2333 if (CmpValue && !IsSubsRegImm)
2335 if (!CmpValue && !IsSubsRegImm && !
isADDSRegImm(CmpOpcode))
2340 if (MIUsedNZCV.
C || MIUsedNZCV.
V)
2343 std::optional<UsedNZCV> NZCVUsedAfterCmp =
2347 if (!NZCVUsedAfterCmp || NZCVUsedAfterCmp->C || NZCVUsedAfterCmp->V)
2350 if ((MIUsedNZCV.
Z && NZCVUsedAfterCmp->N) ||
2351 (MIUsedNZCV.
N && NZCVUsedAfterCmp->Z))
2354 if (MIUsedNZCV.
N && !CmpValue)
2396bool AArch64InstrInfo::removeCmpToZeroOrOne(
2399 MachineInstr *
MI =
MRI.getUniqueVRegDef(SrcReg);
2403 SmallVector<MachineInstr *, 4> CCUseInstrs;
2404 bool IsInvertCC =
false;
2412 for (MachineInstr *CCUseInstr : CCUseInstrs) {
2414 assert(Idx >= 0 &&
"Unexpected instruction using CC.");
2415 MachineOperand &CCOperand = CCUseInstr->getOperand(Idx);
2424bool AArch64InstrInfo::expandPostRAPseudo(
MachineInstr &
MI)
const {
2425 if (
MI.getOpcode() != TargetOpcode::LOAD_STACK_GUARD &&
2426 MI.getOpcode() != AArch64::CATCHRET)
2429 MachineBasicBlock &
MBB = *
MI.getParent();
2431 auto TRI = Subtarget.getRegisterInfo();
2434 if (
MI.getOpcode() == AArch64::CATCHRET) {
2436 const TargetInstrInfo *
TII =
2438 MachineBasicBlock *TargetMBB =
MI.getOperand(0).getMBB();
2443 FirstEpilogSEH = std::prev(FirstEpilogSEH);
2445 FirstEpilogSEH = std::next(FirstEpilogSEH);
2460 if (
M.getStackProtectorGuard() ==
"sysreg") {
2461 const AArch64SysReg::SysReg *SrcReg =
2462 AArch64SysReg::lookupSysRegByName(
M.getStackProtectorGuardReg());
2470 int Offset =
M.getStackProtectorGuardOffset();
2521 const GlobalValue *GV =
2524 unsigned OpFlags = Subtarget.ClassifyGlobalReference(GV, TM);
2530 if (Subtarget.isTargetILP32()) {
2531 unsigned Reg32 =
TRI->getSubReg(
Reg, AArch64::sub_32);
2545 assert(!Subtarget.isTargetILP32() &&
"how can large exist in ILP32?");
2572 if (Subtarget.isTargetILP32()) {
2573 unsigned Reg32 =
TRI->getSubReg(
Reg, AArch64::sub_32);
2596 switch (
MI.getOpcode()) {
2599 case AArch64::MOVZWi:
2600 case AArch64::MOVZXi:
2601 if (
MI.getOperand(1).isImm() &&
MI.getOperand(1).getImm() == 0) {
2602 assert(
MI.getDesc().getNumOperands() == 3 &&
2603 MI.getOperand(2).getImm() == 0 &&
"invalid MOVZi operands");
2607 case AArch64::ANDWri:
2608 return MI.getOperand(1).getReg() == AArch64::WZR;
2609 case AArch64::ANDXri:
2610 return MI.getOperand(1).getReg() == AArch64::XZR;
2611 case TargetOpcode::COPY:
2612 return MI.getOperand(1).getReg() == AArch64::WZR;
2620 switch (
MI.getOpcode()) {
2623 case TargetOpcode::COPY: {
2626 return (AArch64::GPR32RegClass.
contains(DstReg) ||
2627 AArch64::GPR64RegClass.
contains(DstReg));
2629 case AArch64::ORRXrs:
2630 if (
MI.getOperand(1).getReg() == AArch64::XZR) {
2631 assert(
MI.getDesc().getNumOperands() == 4 &&
2632 MI.getOperand(3).getImm() == 0 &&
"invalid ORRrs operands");
2636 case AArch64::ADDXri:
2637 if (
MI.getOperand(2).getImm() == 0) {
2638 assert(
MI.getDesc().getNumOperands() == 4 &&
2639 MI.getOperand(3).getImm() == 0 &&
"invalid ADDXri operands");
2650 switch (
MI.getOpcode()) {
2653 case TargetOpcode::COPY: {
2655 return AArch64::FPR128RegClass.contains(DstReg);
2657 case AArch64::ORRv16i8:
2658 if (
MI.getOperand(1).getReg() ==
MI.getOperand(2).getReg()) {
2659 assert(
MI.getDesc().getNumOperands() == 3 &&
MI.getOperand(0).isReg() &&
2660 "invalid ORRv16i8 operands");
2672 case AArch64::LDRWui:
2673 case AArch64::LDRXui:
2674 case AArch64::LDRBui:
2675 case AArch64::LDRHui:
2676 case AArch64::LDRSui:
2677 case AArch64::LDRDui:
2678 case AArch64::LDRQui:
2679 case AArch64::LDR_PXI:
2685 int &FrameIndex)
const {
2689 if (
MI.getOperand(0).getSubReg() == 0 &&
MI.getOperand(1).isFI() &&
2690 MI.getOperand(2).isImm() &&
MI.getOperand(2).getImm() == 0) {
2691 FrameIndex =
MI.getOperand(1).getIndex();
2692 return MI.getOperand(0).getReg();
2701 case AArch64::STRWui:
2702 case AArch64::STRXui:
2703 case AArch64::STRBui:
2704 case AArch64::STRHui:
2705 case AArch64::STRSui:
2706 case AArch64::STRDui:
2707 case AArch64::STRQui:
2708 case AArch64::STR_PXI:
2714 int &FrameIndex)
const {
2718 if (
MI.getOperand(0).getSubReg() == 0 &&
MI.getOperand(1).isFI() &&
2719 MI.getOperand(2).isImm() &&
MI.getOperand(2).getImm() == 0) {
2720 FrameIndex =
MI.getOperand(1).getIndex();
2721 return MI.getOperand(0).getReg();
2727 int &FrameIndex)
const {
2742 return MI.getOperand(0).getReg();
2748 int &FrameIndex)
const {
2763 return MI.getOperand(0).getReg();
2771 return MMO->getFlags() & MOSuppressPair;
2777 if (
MI.memoperands_empty())
2785 return MMO->getFlags() & MOStridedAccess;
2793 case AArch64::STURSi:
2794 case AArch64::STRSpre:
2795 case AArch64::STURDi:
2796 case AArch64::STRDpre:
2797 case AArch64::STURQi:
2798 case AArch64::STRQpre:
2799 case AArch64::STURBBi:
2800 case AArch64::STURHHi:
2801 case AArch64::STURWi:
2802 case AArch64::STRWpre:
2803 case AArch64::STURXi:
2804 case AArch64::STRXpre:
2805 case AArch64::LDURSi:
2806 case AArch64::LDRSpre:
2807 case AArch64::LDURDi:
2808 case AArch64::LDRDpre:
2809 case AArch64::LDURQi:
2810 case AArch64::LDRQpre:
2811 case AArch64::LDURWi:
2812 case AArch64::LDRWpre:
2813 case AArch64::LDURXi:
2814 case AArch64::LDRXpre:
2815 case AArch64::LDRSWpre:
2816 case AArch64::LDURSWi:
2817 case AArch64::LDURHHi:
2818 case AArch64::LDURBBi:
2819 case AArch64::LDURSBWi:
2820 case AArch64::LDURSHWi:
2828 case AArch64::PRFMui:
return AArch64::PRFUMi;
2829 case AArch64::LDRXui:
return AArch64::LDURXi;
2830 case AArch64::LDRWui:
return AArch64::LDURWi;
2831 case AArch64::LDRBui:
return AArch64::LDURBi;
2832 case AArch64::LDRHui:
return AArch64::LDURHi;
2833 case AArch64::LDRSui:
return AArch64::LDURSi;
2834 case AArch64::LDRDui:
return AArch64::LDURDi;
2835 case AArch64::LDRQui:
return AArch64::LDURQi;
2836 case AArch64::LDRBBui:
return AArch64::LDURBBi;
2837 case AArch64::LDRHHui:
return AArch64::LDURHHi;
2838 case AArch64::LDRSBXui:
return AArch64::LDURSBXi;
2839 case AArch64::LDRSBWui:
return AArch64::LDURSBWi;
2840 case AArch64::LDRSHXui:
return AArch64::LDURSHXi;
2841 case AArch64::LDRSHWui:
return AArch64::LDURSHWi;
2842 case AArch64::LDRSWui:
return AArch64::LDURSWi;
2843 case AArch64::STRXui:
return AArch64::STURXi;
2844 case AArch64::STRWui:
return AArch64::STURWi;
2845 case AArch64::STRBui:
return AArch64::STURBi;
2846 case AArch64::STRHui:
return AArch64::STURHi;
2847 case AArch64::STRSui:
return AArch64::STURSi;
2848 case AArch64::STRDui:
return AArch64::STURDi;
2849 case AArch64::STRQui:
return AArch64::STURQi;
2850 case AArch64::STRBBui:
return AArch64::STURBBi;
2851 case AArch64::STRHHui:
return AArch64::STURHHi;
2860 case AArch64::LDAPURBi:
2861 case AArch64::LDAPURHi:
2862 case AArch64::LDAPURi:
2863 case AArch64::LDAPURSBWi:
2864 case AArch64::LDAPURSBXi:
2865 case AArch64::LDAPURSHWi:
2866 case AArch64::LDAPURSHXi:
2867 case AArch64::LDAPURSWi:
2868 case AArch64::LDAPURXi:
2869 case AArch64::LDR_PPXI:
2870 case AArch64::LDR_PXI:
2871 case AArch64::LDR_ZXI:
2872 case AArch64::LDR_ZZXI:
2873 case AArch64::LDR_ZZXI_STRIDED_CONTIGUOUS:
2874 case AArch64::LDR_ZZZXI:
2875 case AArch64::LDR_ZZZZXI:
2876 case AArch64::LDR_ZZZZXI_STRIDED_CONTIGUOUS:
2877 case AArch64::LDRBBui:
2878 case AArch64::LDRBui:
2879 case AArch64::LDRDui:
2880 case AArch64::LDRHHui:
2881 case AArch64::LDRHui:
2882 case AArch64::LDRQui:
2883 case AArch64::LDRSBWui:
2884 case AArch64::LDRSBXui:
2885 case AArch64::LDRSHWui:
2886 case AArch64::LDRSHXui:
2887 case AArch64::LDRSui:
2888 case AArch64::LDRSWui:
2889 case AArch64::LDRWui:
2890 case AArch64::LDRXui:
2891 case AArch64::LDURBBi:
2892 case AArch64::LDURBi:
2893 case AArch64::LDURDi:
2894 case AArch64::LDURHHi:
2895 case AArch64::LDURHi:
2896 case AArch64::LDURQi:
2897 case AArch64::LDURSBWi:
2898 case AArch64::LDURSBXi:
2899 case AArch64::LDURSHWi:
2900 case AArch64::LDURSHXi:
2901 case AArch64::LDURSi:
2902 case AArch64::LDURSWi:
2903 case AArch64::LDURWi:
2904 case AArch64::LDURXi:
2905 case AArch64::PRFMui:
2906 case AArch64::PRFUMi:
2907 case AArch64::ST2Gi:
2909 case AArch64::STLURBi:
2910 case AArch64::STLURHi:
2911 case AArch64::STLURWi:
2912 case AArch64::STLURXi:
2913 case AArch64::StoreSwiftAsyncContext:
2914 case AArch64::STR_PPXI:
2915 case AArch64::STR_PXI:
2916 case AArch64::STR_ZXI:
2917 case AArch64::STR_ZZXI:
2918 case AArch64::STR_ZZXI_STRIDED_CONTIGUOUS:
2919 case AArch64::STR_ZZZXI:
2920 case AArch64::STR_ZZZZXI:
2921 case AArch64::STR_ZZZZXI_STRIDED_CONTIGUOUS:
2922 case AArch64::STRBBui:
2923 case AArch64::STRBui:
2924 case AArch64::STRDui:
2925 case AArch64::STRHHui:
2926 case AArch64::STRHui:
2927 case AArch64::STRQui:
2928 case AArch64::STRSui:
2929 case AArch64::STRWui:
2930 case AArch64::STRXui:
2931 case AArch64::STURBBi:
2932 case AArch64::STURBi:
2933 case AArch64::STURDi:
2934 case AArch64::STURHHi:
2935 case AArch64::STURHi:
2936 case AArch64::STURQi:
2937 case AArch64::STURSi:
2938 case AArch64::STURWi:
2939 case AArch64::STURXi:
2940 case AArch64::STZ2Gi:
2941 case AArch64::STZGi:
2942 case AArch64::TAGPstack:
2944 case AArch64::LD1B_D_IMM:
2945 case AArch64::LD1B_H_IMM:
2946 case AArch64::LD1B_IMM:
2947 case AArch64::LD1B_S_IMM:
2948 case AArch64::LD1D_IMM:
2949 case AArch64::LD1H_D_IMM:
2950 case AArch64::LD1H_IMM:
2951 case AArch64::LD1H_S_IMM:
2952 case AArch64::LD1RB_D_IMM:
2953 case AArch64::LD1RB_H_IMM:
2954 case AArch64::LD1RB_IMM:
2955 case AArch64::LD1RB_S_IMM:
2956 case AArch64::LD1RD_IMM:
2957 case AArch64::LD1RH_D_IMM:
2958 case AArch64::LD1RH_IMM:
2959 case AArch64::LD1RH_S_IMM:
2960 case AArch64::LD1RSB_D_IMM:
2961 case AArch64::LD1RSB_H_IMM:
2962 case AArch64::LD1RSB_S_IMM:
2963 case AArch64::LD1RSH_D_IMM:
2964 case AArch64::LD1RSH_S_IMM:
2965 case AArch64::LD1RSW_IMM:
2966 case AArch64::LD1RW_D_IMM:
2967 case AArch64::LD1RW_IMM:
2968 case AArch64::LD1SB_D_IMM:
2969 case AArch64::LD1SB_H_IMM:
2970 case AArch64::LD1SB_S_IMM:
2971 case AArch64::LD1SH_D_IMM:
2972 case AArch64::LD1SH_S_IMM:
2973 case AArch64::LD1SW_D_IMM:
2974 case AArch64::LD1W_D_IMM:
2975 case AArch64::LD1W_IMM:
2976 case AArch64::LD2B_IMM:
2977 case AArch64::LD2D_IMM:
2978 case AArch64::LD2H_IMM:
2979 case AArch64::LD2W_IMM:
2980 case AArch64::LD3B_IMM:
2981 case AArch64::LD3D_IMM:
2982 case AArch64::LD3H_IMM:
2983 case AArch64::LD3W_IMM:
2984 case AArch64::LD4B_IMM:
2985 case AArch64::LD4D_IMM:
2986 case AArch64::LD4H_IMM:
2987 case AArch64::LD4W_IMM:
2989 case AArch64::LDNF1B_D_IMM:
2990 case AArch64::LDNF1B_H_IMM:
2991 case AArch64::LDNF1B_IMM:
2992 case AArch64::LDNF1B_S_IMM:
2993 case AArch64::LDNF1D_IMM:
2994 case AArch64::LDNF1H_D_IMM:
2995 case AArch64::LDNF1H_IMM:
2996 case AArch64::LDNF1H_S_IMM:
2997 case AArch64::LDNF1SB_D_IMM:
2998 case AArch64::LDNF1SB_H_IMM:
2999 case AArch64::LDNF1SB_S_IMM:
3000 case AArch64::LDNF1SH_D_IMM:
3001 case AArch64::LDNF1SH_S_IMM:
3002 case AArch64::LDNF1SW_D_IMM:
3003 case AArch64::LDNF1W_D_IMM:
3004 case AArch64::LDNF1W_IMM:
3005 case AArch64::LDNPDi:
3006 case AArch64::LDNPQi:
3007 case AArch64::LDNPSi:
3008 case AArch64::LDNPWi:
3009 case AArch64::LDNPXi:
3010 case AArch64::LDNT1B_ZRI:
3011 case AArch64::LDNT1D_ZRI:
3012 case AArch64::LDNT1H_ZRI:
3013 case AArch64::LDNT1W_ZRI:
3014 case AArch64::LDPDi:
3015 case AArch64::LDPQi:
3016 case AArch64::LDPSi:
3017 case AArch64::LDPWi:
3018 case AArch64::LDPXi:
3019 case AArch64::LDRBBpost:
3020 case AArch64::LDRBBpre:
3021 case AArch64::LDRBpost:
3022 case AArch64::LDRBpre:
3023 case AArch64::LDRDpost:
3024 case AArch64::LDRDpre:
3025 case AArch64::LDRHHpost:
3026 case AArch64::LDRHHpre:
3027 case AArch64::LDRHpost:
3028 case AArch64::LDRHpre:
3029 case AArch64::LDRQpost:
3030 case AArch64::LDRQpre:
3031 case AArch64::LDRSpost:
3032 case AArch64::LDRSpre:
3033 case AArch64::LDRWpost:
3034 case AArch64::LDRWpre:
3035 case AArch64::LDRXpost:
3036 case AArch64::LDRXpre:
3037 case AArch64::ST1B_D_IMM:
3038 case AArch64::ST1B_H_IMM:
3039 case AArch64::ST1B_IMM:
3040 case AArch64::ST1B_S_IMM:
3041 case AArch64::ST1D_IMM:
3042 case AArch64::ST1H_D_IMM:
3043 case AArch64::ST1H_IMM:
3044 case AArch64::ST1H_S_IMM:
3045 case AArch64::ST1W_D_IMM:
3046 case AArch64::ST1W_IMM:
3047 case AArch64::ST2B_IMM:
3048 case AArch64::ST2D_IMM:
3049 case AArch64::ST2H_IMM:
3050 case AArch64::ST2W_IMM:
3051 case AArch64::ST3B_IMM:
3052 case AArch64::ST3D_IMM:
3053 case AArch64::ST3H_IMM:
3054 case AArch64::ST3W_IMM:
3055 case AArch64::ST4B_IMM:
3056 case AArch64::ST4D_IMM:
3057 case AArch64::ST4H_IMM:
3058 case AArch64::ST4W_IMM:
3059 case AArch64::STGPi:
3060 case AArch64::STGPreIndex:
3061 case AArch64::STZGPreIndex:
3062 case AArch64::ST2GPreIndex:
3063 case AArch64::STZ2GPreIndex:
3064 case AArch64::STGPostIndex:
3065 case AArch64::STZGPostIndex:
3066 case AArch64::ST2GPostIndex:
3067 case AArch64::STZ2GPostIndex:
3068 case AArch64::STNPDi:
3069 case AArch64::STNPQi:
3070 case AArch64::STNPSi:
3071 case AArch64::STNPWi:
3072 case AArch64::STNPXi:
3073 case AArch64::STNT1B_ZRI:
3074 case AArch64::STNT1D_ZRI:
3075 case AArch64::STNT1H_ZRI:
3076 case AArch64::STNT1W_ZRI:
3077 case AArch64::STPDi:
3078 case AArch64::STPQi:
3079 case AArch64::STPSi:
3080 case AArch64::STPWi:
3081 case AArch64::STPXi:
3082 case AArch64::STRBBpost:
3083 case AArch64::STRBBpre:
3084 case AArch64::STRBpost:
3085 case AArch64::STRBpre:
3086 case AArch64::STRDpost:
3087 case AArch64::STRDpre:
3088 case AArch64::STRHHpost:
3089 case AArch64::STRHHpre:
3090 case AArch64::STRHpost:
3091 case AArch64::STRHpre:
3092 case AArch64::STRQpost:
3093 case AArch64::STRQpre:
3094 case AArch64::STRSpost:
3095 case AArch64::STRSpre:
3096 case AArch64::STRWpost:
3097 case AArch64::STRWpre:
3098 case AArch64::STRXpost:
3099 case AArch64::STRXpre:
3101 case AArch64::LDPDpost:
3102 case AArch64::LDPDpre:
3103 case AArch64::LDPQpost:
3104 case AArch64::LDPQpre:
3105 case AArch64::LDPSpost:
3106 case AArch64::LDPSpre:
3107 case AArch64::LDPWpost:
3108 case AArch64::LDPWpre:
3109 case AArch64::LDPXpost:
3110 case AArch64::LDPXpre:
3111 case AArch64::STGPpre:
3112 case AArch64::STGPpost:
3113 case AArch64::STPDpost:
3114 case AArch64::STPDpre:
3115 case AArch64::STPQpost:
3116 case AArch64::STPQpre:
3117 case AArch64::STPSpost:
3118 case AArch64::STPSpre:
3119 case AArch64::STPWpost:
3120 case AArch64::STPWpre:
3121 case AArch64::STPXpost:
3122 case AArch64::STPXpre:
3128 switch (
MI.getOpcode()) {
3132 case AArch64::STRSui:
3133 case AArch64::STRDui:
3134 case AArch64::STRQui:
3135 case AArch64::STRXui:
3136 case AArch64::STRWui:
3137 case AArch64::LDRSui:
3138 case AArch64::LDRDui:
3139 case AArch64::LDRQui:
3140 case AArch64::LDRXui:
3141 case AArch64::LDRWui:
3142 case AArch64::LDRSWui:
3144 case AArch64::STURSi:
3145 case AArch64::STRSpre:
3146 case AArch64::STURDi:
3147 case AArch64::STRDpre:
3148 case AArch64::STURQi:
3149 case AArch64::STRQpre:
3150 case AArch64::STURWi:
3151 case AArch64::STRWpre:
3152 case AArch64::STURXi:
3153 case AArch64::STRXpre:
3154 case AArch64::LDURSi:
3155 case AArch64::LDRSpre:
3156 case AArch64::LDURDi:
3157 case AArch64::LDRDpre:
3158 case AArch64::LDURQi:
3159 case AArch64::LDRQpre:
3160 case AArch64::LDURWi:
3161 case AArch64::LDRWpre:
3162 case AArch64::LDURXi:
3163 case AArch64::LDRXpre:
3164 case AArch64::LDURSWi:
3165 case AArch64::LDRSWpre:
3167 case AArch64::LDR_ZXI:
3168 case AArch64::STR_ZXI:
3174 switch (
MI.getOpcode()) {
3177 "Unexpected instruction - was a new tail call opcode introduced?");
3179 case AArch64::TCRETURNdi:
3180 case AArch64::TCRETURNri:
3181 case AArch64::TCRETURNrix16x17:
3182 case AArch64::TCRETURNrix17:
3183 case AArch64::TCRETURNrinotx16:
3184 case AArch64::TCRETURNriALL:
3185 case AArch64::AUTH_TCRETURN:
3186 case AArch64::AUTH_TCRETURN_BTI:
3196 case AArch64::ADDWri:
3197 return AArch64::ADDSWri;
3198 case AArch64::ADDWrr:
3199 return AArch64::ADDSWrr;
3200 case AArch64::ADDWrs:
3201 return AArch64::ADDSWrs;
3202 case AArch64::ADDWrx:
3203 return AArch64::ADDSWrx;
3204 case AArch64::ANDWri:
3205 return AArch64::ANDSWri;
3206 case AArch64::ANDWrr:
3207 return AArch64::ANDSWrr;
3208 case AArch64::ANDWrs:
3209 return AArch64::ANDSWrs;
3210 case AArch64::BICWrr:
3211 return AArch64::BICSWrr;
3212 case AArch64::BICWrs:
3213 return AArch64::BICSWrs;
3214 case AArch64::SUBWri:
3215 return AArch64::SUBSWri;
3216 case AArch64::SUBWrr:
3217 return AArch64::SUBSWrr;
3218 case AArch64::SUBWrs:
3219 return AArch64::SUBSWrs;
3220 case AArch64::SUBWrx:
3221 return AArch64::SUBSWrx;
3223 case AArch64::ADDXri:
3224 return AArch64::ADDSXri;
3225 case AArch64::ADDXrr:
3226 return AArch64::ADDSXrr;
3227 case AArch64::ADDXrs:
3228 return AArch64::ADDSXrs;
3229 case AArch64::ADDXrx:
3230 return AArch64::ADDSXrx;
3231 case AArch64::ANDXri:
3232 return AArch64::ANDSXri;
3233 case AArch64::ANDXrr:
3234 return AArch64::ANDSXrr;
3235 case AArch64::ANDXrs:
3236 return AArch64::ANDSXrs;
3237 case AArch64::BICXrr:
3238 return AArch64::BICSXrr;
3239 case AArch64::BICXrs:
3240 return AArch64::BICSXrs;
3241 case AArch64::SUBXri:
3242 return AArch64::SUBSXri;
3243 case AArch64::SUBXrr:
3244 return AArch64::SUBSXrr;
3245 case AArch64::SUBXrs:
3246 return AArch64::SUBSXrs;
3247 case AArch64::SUBXrx:
3248 return AArch64::SUBSXrx;
3250 case AArch64::AND_PPzPP:
3251 return AArch64::ANDS_PPzPP;
3252 case AArch64::BIC_PPzPP:
3253 return AArch64::BICS_PPzPP;
3254 case AArch64::EOR_PPzPP:
3255 return AArch64::EORS_PPzPP;
3256 case AArch64::NAND_PPzPP:
3257 return AArch64::NANDS_PPzPP;
3258 case AArch64::NOR_PPzPP:
3259 return AArch64::NORS_PPzPP;
3260 case AArch64::ORN_PPzPP:
3261 return AArch64::ORNS_PPzPP;
3262 case AArch64::ORR_PPzPP:
3263 return AArch64::ORRS_PPzPP;
3264 case AArch64::BRKA_PPzP:
3265 return AArch64::BRKAS_PPzP;
3266 case AArch64::BRKPA_PPzPP:
3267 return AArch64::BRKPAS_PPzPP;
3268 case AArch64::BRKB_PPzP:
3269 return AArch64::BRKBS_PPzP;
3270 case AArch64::BRKPB_PPzPP:
3271 return AArch64::BRKPBS_PPzPP;
3272 case AArch64::BRKN_PPzP:
3273 return AArch64::BRKNS_PPzP;
3274 case AArch64::RDFFR_PPz:
3275 return AArch64::RDFFRS_PPz;
3276 case AArch64::PTRUE_B:
3277 return AArch64::PTRUES_B;
3288 if (
MI.hasOrderedMemoryRef())
3293 assert((
MI.getOperand(IsPreLdSt ? 2 : 1).isReg() ||
3294 MI.getOperand(IsPreLdSt ? 2 : 1).isFI()) &&
3295 "Expected a reg or frame index operand.");
3299 bool IsImmPreLdSt = IsPreLdSt &&
MI.getOperand(3).isImm();
3301 if (!
MI.getOperand(2).isImm() && !IsImmPreLdSt)
3314 if (
MI.getOperand(1).isReg() && !IsPreLdSt) {
3315 Register BaseReg =
MI.getOperand(1).getReg();
3317 if (
MI.modifiesRegister(BaseReg,
TRI))
3323 switch (
MI.getOpcode()) {
3326 case AArch64::LDR_ZXI:
3327 case AArch64::STR_ZXI:
3328 if (!Subtarget.isLittleEndian() ||
3329 Subtarget.getSVEVectorSizeInBits() != 128)
3342 const MCAsmInfo *MAI =
MI.getMF()->getTarget().getMCAsmInfo();
3344 MI.getMF()->getFunction().needsUnwindTableEntry();
3350 if (Subtarget.isPaired128Slow()) {
3351 switch (
MI.getOpcode()) {
3354 case AArch64::LDURQi:
3355 case AArch64::STURQi:
3356 case AArch64::LDRQui:
3357 case AArch64::STRQui:
3384std::optional<ExtAddrMode>
3389 bool OffsetIsScalable;
3390 if (!getMemOperandWithOffset(MemI,
Base,
Offset, OffsetIsScalable,
TRI))
3391 return std::nullopt;
3394 return std::nullopt;
3409 int64_t OffsetScale = 1;
3414 case AArch64::LDURQi:
3415 case AArch64::STURQi:
3419 case AArch64::LDURDi:
3420 case AArch64::STURDi:
3421 case AArch64::LDURXi:
3422 case AArch64::STURXi:
3426 case AArch64::LDURWi:
3427 case AArch64::LDURSWi:
3428 case AArch64::STURWi:
3432 case AArch64::LDURHi:
3433 case AArch64::STURHi:
3434 case AArch64::LDURHHi:
3435 case AArch64::STURHHi:
3436 case AArch64::LDURSHXi:
3437 case AArch64::LDURSHWi:
3441 case AArch64::LDRBroX:
3442 case AArch64::LDRBBroX:
3443 case AArch64::LDRSBXroX:
3444 case AArch64::LDRSBWroX:
3445 case AArch64::STRBroX:
3446 case AArch64::STRBBroX:
3447 case AArch64::LDURBi:
3448 case AArch64::LDURBBi:
3449 case AArch64::LDURSBXi:
3450 case AArch64::LDURSBWi:
3451 case AArch64::STURBi:
3452 case AArch64::STURBBi:
3453 case AArch64::LDRBui:
3454 case AArch64::LDRBBui:
3455 case AArch64::LDRSBXui:
3456 case AArch64::LDRSBWui:
3457 case AArch64::STRBui:
3458 case AArch64::STRBBui:
3462 case AArch64::LDRQroX:
3463 case AArch64::STRQroX:
3464 case AArch64::LDRQui:
3465 case AArch64::STRQui:
3470 case AArch64::LDRDroX:
3471 case AArch64::STRDroX:
3472 case AArch64::LDRXroX:
3473 case AArch64::STRXroX:
3474 case AArch64::LDRDui:
3475 case AArch64::STRDui:
3476 case AArch64::LDRXui:
3477 case AArch64::STRXui:
3482 case AArch64::LDRWroX:
3483 case AArch64::LDRSWroX:
3484 case AArch64::STRWroX:
3485 case AArch64::LDRWui:
3486 case AArch64::LDRSWui:
3487 case AArch64::STRWui:
3492 case AArch64::LDRHroX:
3493 case AArch64::STRHroX:
3494 case AArch64::LDRHHroX:
3495 case AArch64::STRHHroX:
3496 case AArch64::LDRSHXroX:
3497 case AArch64::LDRSHWroX:
3498 case AArch64::LDRHui:
3499 case AArch64::STRHui:
3500 case AArch64::LDRHHui:
3501 case AArch64::STRHHui:
3502 case AArch64::LDRSHXui:
3503 case AArch64::LDRSHWui:
3511 if (BaseRegOp.
isReg() && BaseRegOp.
getReg() == Reg)
3535 case AArch64::SBFMXri:
3548 AM.
Scale = OffsetScale;
3553 case TargetOpcode::SUBREG_TO_REG: {
3566 if (!OffsetReg.
isVirtual() || !
MRI.hasOneNonDBGUse(OffsetReg))
3570 if (
DefMI.getOpcode() != AArch64::ORRWrs ||
3571 DefMI.getOperand(1).getReg() != AArch64::WZR ||
3572 DefMI.getOperand(3).getImm() != 0)
3579 AM.
Scale = OffsetScale;
3590 auto validateOffsetForLDP = [](
unsigned NumBytes, int64_t OldOffset,
3591 int64_t NewOffset) ->
bool {
3592 int64_t MinOffset, MaxOffset;
3609 return OldOffset < MinOffset || OldOffset > MaxOffset ||
3610 (NewOffset >= MinOffset && NewOffset <= MaxOffset);
3612 auto canFoldAddSubImmIntoAddrMode = [&](int64_t Disp) ->
bool {
3614 int64_t NewOffset = OldOffset + Disp;
3615 if (!isLegalAddressingMode(NumBytes, NewOffset, 0))
3619 if (!validateOffsetForLDP(NumBytes, OldOffset, NewOffset))
3629 auto canFoldAddRegIntoAddrMode =
3634 if ((
unsigned)Scale != Scale)
3636 if (!isLegalAddressingMode(NumBytes, 0, Scale))
3648 return (Opcode == AArch64::STURQi || Opcode == AArch64::STRQui) &&
3649 Subtarget.isSTRQroSlow();
3658 case AArch64::ADDXri:
3664 return canFoldAddSubImmIntoAddrMode(Disp);
3666 case AArch64::SUBXri:
3672 return canFoldAddSubImmIntoAddrMode(-Disp);
3674 case AArch64::ADDXrs: {
3687 if (Shift != 2 && Shift != 3 && Subtarget.hasAddrLSLSlow14())
3689 if (avoidSlowSTRQ(MemI))
3692 return canFoldAddRegIntoAddrMode(1ULL << Shift);
3695 case AArch64::ADDXrr:
3703 if (!OptSize && avoidSlowSTRQ(MemI))
3705 return canFoldAddRegIntoAddrMode(1);
3707 case AArch64::ADDXrx:
3715 if (!OptSize && avoidSlowSTRQ(MemI))
3724 return canFoldAddRegIntoAddrMode(
3739 case AArch64::LDURQi:
3740 case AArch64::LDRQui:
3741 return AArch64::LDRQroX;
3742 case AArch64::STURQi:
3743 case AArch64::STRQui:
3744 return AArch64::STRQroX;
3745 case AArch64::LDURDi:
3746 case AArch64::LDRDui:
3747 return AArch64::LDRDroX;
3748 case AArch64::STURDi:
3749 case AArch64::STRDui:
3750 return AArch64::STRDroX;
3751 case AArch64::LDURXi:
3752 case AArch64::LDRXui:
3753 return AArch64::LDRXroX;
3754 case AArch64::STURXi:
3755 case AArch64::STRXui:
3756 return AArch64::STRXroX;
3757 case AArch64::LDURWi:
3758 case AArch64::LDRWui:
3759 return AArch64::LDRWroX;
3760 case AArch64::LDURSWi:
3761 case AArch64::LDRSWui:
3762 return AArch64::LDRSWroX;
3763 case AArch64::STURWi:
3764 case AArch64::STRWui:
3765 return AArch64::STRWroX;
3766 case AArch64::LDURHi:
3767 case AArch64::LDRHui:
3768 return AArch64::LDRHroX;
3769 case AArch64::STURHi:
3770 case AArch64::STRHui:
3771 return AArch64::STRHroX;
3772 case AArch64::LDURHHi:
3773 case AArch64::LDRHHui:
3774 return AArch64::LDRHHroX;
3775 case AArch64::STURHHi:
3776 case AArch64::STRHHui:
3777 return AArch64::STRHHroX;
3778 case AArch64::LDURSHXi:
3779 case AArch64::LDRSHXui:
3780 return AArch64::LDRSHXroX;
3781 case AArch64::LDURSHWi:
3782 case AArch64::LDRSHWui:
3783 return AArch64::LDRSHWroX;
3784 case AArch64::LDURBi:
3785 case AArch64::LDRBui:
3786 return AArch64::LDRBroX;
3787 case AArch64::LDURBBi:
3788 case AArch64::LDRBBui:
3789 return AArch64::LDRBBroX;
3790 case AArch64::LDURSBXi:
3791 case AArch64::LDRSBXui:
3792 return AArch64::LDRSBXroX;
3793 case AArch64::LDURSBWi:
3794 case AArch64::LDRSBWui:
3795 return AArch64::LDRSBWroX;
3796 case AArch64::STURBi:
3797 case AArch64::STRBui:
3798 return AArch64::STRBroX;
3799 case AArch64::STURBBi:
3800 case AArch64::STRBBui:
3801 return AArch64::STRBBroX;
3813 case AArch64::LDURQi:
3815 return AArch64::LDRQui;
3816 case AArch64::STURQi:
3818 return AArch64::STRQui;
3819 case AArch64::LDURDi:
3821 return AArch64::LDRDui;
3822 case AArch64::STURDi:
3824 return AArch64::STRDui;
3825 case AArch64::LDURXi:
3827 return AArch64::LDRXui;
3828 case AArch64::STURXi:
3830 return AArch64::STRXui;
3831 case AArch64::LDURWi:
3833 return AArch64::LDRWui;
3834 case AArch64::LDURSWi:
3836 return AArch64::LDRSWui;
3837 case AArch64::STURWi:
3839 return AArch64::STRWui;
3840 case AArch64::LDURHi:
3842 return AArch64::LDRHui;
3843 case AArch64::STURHi:
3845 return AArch64::STRHui;
3846 case AArch64::LDURHHi:
3848 return AArch64::LDRHHui;
3849 case AArch64::STURHHi:
3851 return AArch64::STRHHui;
3852 case AArch64::LDURSHXi:
3854 return AArch64::LDRSHXui;
3855 case AArch64::LDURSHWi:
3857 return AArch64::LDRSHWui;
3858 case AArch64::LDURBi:
3860 return AArch64::LDRBui;
3861 case AArch64::LDURBBi:
3863 return AArch64::LDRBBui;
3864 case AArch64::LDURSBXi:
3866 return AArch64::LDRSBXui;
3867 case AArch64::LDURSBWi:
3869 return AArch64::LDRSBWui;
3870 case AArch64::STURBi:
3872 return AArch64::STRBui;
3873 case AArch64::STURBBi:
3875 return AArch64::STRBBui;
3876 case AArch64::LDRQui:
3877 case AArch64::STRQui:
3880 case AArch64::LDRDui:
3881 case AArch64::STRDui:
3882 case AArch64::LDRXui:
3883 case AArch64::STRXui:
3886 case AArch64::LDRWui:
3887 case AArch64::LDRSWui:
3888 case AArch64::STRWui:
3891 case AArch64::LDRHui:
3892 case AArch64::STRHui:
3893 case AArch64::LDRHHui:
3894 case AArch64::STRHHui:
3895 case AArch64::LDRSHXui:
3896 case AArch64::LDRSHWui:
3899 case AArch64::LDRBui:
3900 case AArch64::LDRBBui:
3901 case AArch64::LDRSBXui:
3902 case AArch64::LDRSBWui:
3903 case AArch64::STRBui:
3904 case AArch64::STRBBui:
3918 case AArch64::LDURQi:
3919 case AArch64::STURQi:
3920 case AArch64::LDURDi:
3921 case AArch64::STURDi:
3922 case AArch64::LDURXi:
3923 case AArch64::STURXi:
3924 case AArch64::LDURWi:
3925 case AArch64::LDURSWi:
3926 case AArch64::STURWi:
3927 case AArch64::LDURHi:
3928 case AArch64::STURHi:
3929 case AArch64::LDURHHi:
3930 case AArch64::STURHHi:
3931 case AArch64::LDURSHXi:
3932 case AArch64::LDURSHWi:
3933 case AArch64::LDURBi:
3934 case AArch64::STURBi:
3935 case AArch64::LDURBBi:
3936 case AArch64::STURBBi:
3937 case AArch64::LDURSBWi:
3938 case AArch64::LDURSBXi:
3940 case AArch64::LDRQui:
3941 return AArch64::LDURQi;
3942 case AArch64::STRQui:
3943 return AArch64::STURQi;
3944 case AArch64::LDRDui:
3945 return AArch64::LDURDi;
3946 case AArch64::STRDui:
3947 return AArch64::STURDi;
3948 case AArch64::LDRXui:
3949 return AArch64::LDURXi;
3950 case AArch64::STRXui:
3951 return AArch64::STURXi;
3952 case AArch64::LDRWui:
3953 return AArch64::LDURWi;
3954 case AArch64::LDRSWui:
3955 return AArch64::LDURSWi;
3956 case AArch64::STRWui:
3957 return AArch64::STURWi;
3958 case AArch64::LDRHui:
3959 return AArch64::LDURHi;
3960 case AArch64::STRHui:
3961 return AArch64::STURHi;
3962 case AArch64::LDRHHui:
3963 return AArch64::LDURHHi;
3964 case AArch64::STRHHui:
3965 return AArch64::STURHHi;
3966 case AArch64::LDRSHXui:
3967 return AArch64::LDURSHXi;
3968 case AArch64::LDRSHWui:
3969 return AArch64::LDURSHWi;
3970 case AArch64::LDRBBui:
3971 return AArch64::LDURBBi;
3972 case AArch64::LDRBui:
3973 return AArch64::LDURBi;
3974 case AArch64::STRBBui:
3975 return AArch64::STURBBi;
3976 case AArch64::STRBui:
3977 return AArch64::STURBi;
3978 case AArch64::LDRSBWui:
3979 return AArch64::LDURSBWi;
3980 case AArch64::LDRSBXui:
3981 return AArch64::LDURSBXi;
3994 case AArch64::LDRQroX:
3995 case AArch64::LDURQi:
3996 case AArch64::LDRQui:
3997 return AArch64::LDRQroW;
3998 case AArch64::STRQroX:
3999 case AArch64::STURQi:
4000 case AArch64::STRQui:
4001 return AArch64::STRQroW;
4002 case AArch64::LDRDroX:
4003 case AArch64::LDURDi:
4004 case AArch64::LDRDui:
4005 return AArch64::LDRDroW;
4006 case AArch64::STRDroX:
4007 case AArch64::STURDi:
4008 case AArch64::STRDui:
4009 return AArch64::STRDroW;
4010 case AArch64::LDRXroX:
4011 case AArch64::LDURXi:
4012 case AArch64::LDRXui:
4013 return AArch64::LDRXroW;
4014 case AArch64::STRXroX:
4015 case AArch64::STURXi:
4016 case AArch64::STRXui:
4017 return AArch64::STRXroW;
4018 case AArch64::LDRWroX:
4019 case AArch64::LDURWi:
4020 case AArch64::LDRWui:
4021 return AArch64::LDRWroW;
4022 case AArch64::LDRSWroX:
4023 case AArch64::LDURSWi:
4024 case AArch64::LDRSWui:
4025 return AArch64::LDRSWroW;
4026 case AArch64::STRWroX:
4027 case AArch64::STURWi:
4028 case AArch64::STRWui:
4029 return AArch64::STRWroW;
4030 case AArch64::LDRHroX:
4031 case AArch64::LDURHi:
4032 case AArch64::LDRHui:
4033 return AArch64::LDRHroW;
4034 case AArch64::STRHroX:
4035 case AArch64::STURHi:
4036 case AArch64::STRHui:
4037 return AArch64::STRHroW;
4038 case AArch64::LDRHHroX:
4039 case AArch64::LDURHHi:
4040 case AArch64::LDRHHui:
4041 return AArch64::LDRHHroW;
4042 case AArch64::STRHHroX:
4043 case AArch64::STURHHi:
4044 case AArch64::STRHHui:
4045 return AArch64::STRHHroW;
4046 case AArch64::LDRSHXroX:
4047 case AArch64::LDURSHXi:
4048 case AArch64::LDRSHXui:
4049 return AArch64::LDRSHXroW;
4050 case AArch64::LDRSHWroX:
4051 case AArch64::LDURSHWi:
4052 case AArch64::LDRSHWui:
4053 return AArch64::LDRSHWroW;
4054 case AArch64::LDRBroX:
4055 case AArch64::LDURBi:
4056 case AArch64::LDRBui:
4057 return AArch64::LDRBroW;
4058 case AArch64::LDRBBroX:
4059 case AArch64::LDURBBi:
4060 case AArch64::LDRBBui:
4061 return AArch64::LDRBBroW;
4062 case AArch64::LDRSBXroX:
4063 case AArch64::LDURSBXi:
4064 case AArch64::LDRSBXui:
4065 return AArch64::LDRSBXroW;
4066 case AArch64::LDRSBWroX:
4067 case AArch64::LDURSBWi:
4068 case AArch64::LDRSBWui:
4069 return AArch64::LDRSBWroW;
4070 case AArch64::STRBroX:
4071 case AArch64::STURBi:
4072 case AArch64::STRBui:
4073 return AArch64::STRBroW;
4074 case AArch64::STRBBroX:
4075 case AArch64::STURBBi:
4076 case AArch64::STRBBui:
4077 return AArch64::STRBBroW;
4092 MRI.constrainRegClass(AM.
BaseReg, &AArch64::GPR64spRegClass);
4102 return B.getInstr();
4106 "Addressing mode not supported for folding");
4123 return B.getInstr();
4130 "Address offset can be a register or an immediate, but not both");
4132 MRI.constrainRegClass(AM.
BaseReg, &AArch64::GPR64spRegClass);
4137 OffsetReg =
MRI.createVirtualRegister(&AArch64::GPR32RegClass);
4151 return B.getInstr();
4155 "Function must not be called with an addressing mode it can't handle");
4164 case AArch64::LD1Fourv16b_POST:
4165 case AArch64::LD1Fourv1d_POST:
4166 case AArch64::LD1Fourv2d_POST:
4167 case AArch64::LD1Fourv2s_POST:
4168 case AArch64::LD1Fourv4h_POST:
4169 case AArch64::LD1Fourv4s_POST:
4170 case AArch64::LD1Fourv8b_POST:
4171 case AArch64::LD1Fourv8h_POST:
4172 case AArch64::LD1Onev16b_POST:
4173 case AArch64::LD1Onev1d_POST:
4174 case AArch64::LD1Onev2d_POST:
4175 case AArch64::LD1Onev2s_POST:
4176 case AArch64::LD1Onev4h_POST:
4177 case AArch64::LD1Onev4s_POST:
4178 case AArch64::LD1Onev8b_POST:
4179 case AArch64::LD1Onev8h_POST:
4180 case AArch64::LD1Rv16b_POST:
4181 case AArch64::LD1Rv1d_POST:
4182 case AArch64::LD1Rv2d_POST:
4183 case AArch64::LD1Rv2s_POST:
4184 case AArch64::LD1Rv4h_POST:
4185 case AArch64::LD1Rv4s_POST:
4186 case AArch64::LD1Rv8b_POST:
4187 case AArch64::LD1Rv8h_POST:
4188 case AArch64::LD1Threev16b_POST:
4189 case AArch64::LD1Threev1d_POST:
4190 case AArch64::LD1Threev2d_POST:
4191 case AArch64::LD1Threev2s_POST:
4192 case AArch64::LD1Threev4h_POST:
4193 case AArch64::LD1Threev4s_POST:
4194 case AArch64::LD1Threev8b_POST:
4195 case AArch64::LD1Threev8h_POST:
4196 case AArch64::LD1Twov16b_POST:
4197 case AArch64::LD1Twov1d_POST:
4198 case AArch64::LD1Twov2d_POST:
4199 case AArch64::LD1Twov2s_POST:
4200 case AArch64::LD1Twov4h_POST:
4201 case AArch64::LD1Twov4s_POST:
4202 case AArch64::LD1Twov8b_POST:
4203 case AArch64::LD1Twov8h_POST:
4204 case AArch64::LD1i16_POST:
4205 case AArch64::LD1i32_POST:
4206 case AArch64::LD1i64_POST:
4207 case AArch64::LD1i8_POST:
4208 case AArch64::LD2Rv16b_POST:
4209 case AArch64::LD2Rv1d_POST:
4210 case AArch64::LD2Rv2d_POST:
4211 case AArch64::LD2Rv2s_POST:
4212 case AArch64::LD2Rv4h_POST:
4213 case AArch64::LD2Rv4s_POST:
4214 case AArch64::LD2Rv8b_POST:
4215 case AArch64::LD2Rv8h_POST:
4216 case AArch64::LD2Twov16b_POST:
4217 case AArch64::LD2Twov2d_POST:
4218 case AArch64::LD2Twov2s_POST:
4219 case AArch64::LD2Twov4h_POST:
4220 case AArch64::LD2Twov4s_POST:
4221 case AArch64::LD2Twov8b_POST:
4222 case AArch64::LD2Twov8h_POST:
4223 case AArch64::LD2i16_POST:
4224 case AArch64::LD2i32_POST:
4225 case AArch64::LD2i64_POST:
4226 case AArch64::LD2i8_POST:
4227 case AArch64::LD3Rv16b_POST:
4228 case AArch64::LD3Rv1d_POST:
4229 case AArch64::LD3Rv2d_POST:
4230 case AArch64::LD3Rv2s_POST:
4231 case AArch64::LD3Rv4h_POST:
4232 case AArch64::LD3Rv4s_POST:
4233 case AArch64::LD3Rv8b_POST:
4234 case AArch64::LD3Rv8h_POST:
4235 case AArch64::LD3Threev16b_POST:
4236 case AArch64::LD3Threev2d_POST:
4237 case AArch64::LD3Threev2s_POST:
4238 case AArch64::LD3Threev4h_POST:
4239 case AArch64::LD3Threev4s_POST:
4240 case AArch64::LD3Threev8b_POST:
4241 case AArch64::LD3Threev8h_POST:
4242 case AArch64::LD3i16_POST:
4243 case AArch64::LD3i32_POST:
4244 case AArch64::LD3i64_POST:
4245 case AArch64::LD3i8_POST:
4246 case AArch64::LD4Fourv16b_POST:
4247 case AArch64::LD4Fourv2d_POST:
4248 case AArch64::LD4Fourv2s_POST:
4249 case AArch64::LD4Fourv4h_POST:
4250 case AArch64::LD4Fourv4s_POST:
4251 case AArch64::LD4Fourv8b_POST:
4252 case AArch64::LD4Fourv8h_POST:
4253 case AArch64::LD4Rv16b_POST:
4254 case AArch64::LD4Rv1d_POST:
4255 case AArch64::LD4Rv2d_POST:
4256 case AArch64::LD4Rv2s_POST:
4257 case AArch64::LD4Rv4h_POST:
4258 case AArch64::LD4Rv4s_POST:
4259 case AArch64::LD4Rv8b_POST:
4260 case AArch64::LD4Rv8h_POST:
4261 case AArch64::LD4i16_POST:
4262 case AArch64::LD4i32_POST:
4263 case AArch64::LD4i64_POST:
4264 case AArch64::LD4i8_POST:
4265 case AArch64::LDAPRWpost:
4266 case AArch64::LDAPRXpost:
4267 case AArch64::LDIAPPWpost:
4268 case AArch64::LDIAPPXpost:
4269 case AArch64::LDPDpost:
4270 case AArch64::LDPQpost:
4271 case AArch64::LDPSWpost:
4272 case AArch64::LDPSpost:
4273 case AArch64::LDPWpost:
4274 case AArch64::LDPXpost:
4275 case AArch64::LDRBBpost:
4276 case AArch64::LDRBpost:
4277 case AArch64::LDRDpost:
4278 case AArch64::LDRHHpost:
4279 case AArch64::LDRHpost:
4280 case AArch64::LDRQpost:
4281 case AArch64::LDRSBWpost:
4282 case AArch64::LDRSBXpost:
4283 case AArch64::LDRSHWpost:
4284 case AArch64::LDRSHXpost:
4285 case AArch64::LDRSWpost:
4286 case AArch64::LDRSpost:
4287 case AArch64::LDRWpost:
4288 case AArch64::LDRXpost:
4289 case AArch64::ST1Fourv16b_POST:
4290 case AArch64::ST1Fourv1d_POST:
4291 case AArch64::ST1Fourv2d_POST:
4292 case AArch64::ST1Fourv2s_POST:
4293 case AArch64::ST1Fourv4h_POST:
4294 case AArch64::ST1Fourv4s_POST:
4295 case AArch64::ST1Fourv8b_POST:
4296 case AArch64::ST1Fourv8h_POST:
4297 case AArch64::ST1Onev16b_POST:
4298 case AArch64::ST1Onev1d_POST:
4299 case AArch64::ST1Onev2d_POST:
4300 case AArch64::ST1Onev2s_POST:
4301 case AArch64::ST1Onev4h_POST:
4302 case AArch64::ST1Onev4s_POST:
4303 case AArch64::ST1Onev8b_POST:
4304 case AArch64::ST1Onev8h_POST:
4305 case AArch64::ST1Threev16b_POST:
4306 case AArch64::ST1Threev1d_POST:
4307 case AArch64::ST1Threev2d_POST:
4308 case AArch64::ST1Threev2s_POST:
4309 case AArch64::ST1Threev4h_POST:
4310 case AArch64::ST1Threev4s_POST:
4311 case AArch64::ST1Threev8b_POST:
4312 case AArch64::ST1Threev8h_POST:
4313 case AArch64::ST1Twov16b_POST:
4314 case AArch64::ST1Twov1d_POST:
4315 case AArch64::ST1Twov2d_POST:
4316 case AArch64::ST1Twov2s_POST:
4317 case AArch64::ST1Twov4h_POST:
4318 case AArch64::ST1Twov4s_POST:
4319 case AArch64::ST1Twov8b_POST:
4320 case AArch64::ST1Twov8h_POST:
4321 case AArch64::ST1i16_POST:
4322 case AArch64::ST1i32_POST:
4323 case AArch64::ST1i64_POST:
4324 case AArch64::ST1i8_POST:
4325 case AArch64::ST2GPostIndex:
4326 case AArch64::ST2Twov16b_POST:
4327 case AArch64::ST2Twov2d_POST:
4328 case AArch64::ST2Twov2s_POST:
4329 case AArch64::ST2Twov4h_POST:
4330 case AArch64::ST2Twov4s_POST:
4331 case AArch64::ST2Twov8b_POST:
4332 case AArch64::ST2Twov8h_POST:
4333 case AArch64::ST2i16_POST:
4334 case AArch64::ST2i32_POST:
4335 case AArch64::ST2i64_POST:
4336 case AArch64::ST2i8_POST:
4337 case AArch64::ST3Threev16b_POST:
4338 case AArch64::ST3Threev2d_POST:
4339 case AArch64::ST3Threev2s_POST:
4340 case AArch64::ST3Threev4h_POST:
4341 case AArch64::ST3Threev4s_POST:
4342 case AArch64::ST3Threev8b_POST:
4343 case AArch64::ST3Threev8h_POST:
4344 case AArch64::ST3i16_POST:
4345 case AArch64::ST3i32_POST:
4346 case AArch64::ST3i64_POST:
4347 case AArch64::ST3i8_POST:
4348 case AArch64::ST4Fourv16b_POST:
4349 case AArch64::ST4Fourv2d_POST:
4350 case AArch64::ST4Fourv2s_POST:
4351 case AArch64::ST4Fourv4h_POST:
4352 case AArch64::ST4Fourv4s_POST:
4353 case AArch64::ST4Fourv8b_POST:
4354 case AArch64::ST4Fourv8h_POST:
4355 case AArch64::ST4i16_POST:
4356 case AArch64::ST4i32_POST:
4357 case AArch64::ST4i64_POST:
4358 case AArch64::ST4i8_POST:
4359 case AArch64::STGPostIndex:
4360 case AArch64::STGPpost:
4361 case AArch64::STPDpost:
4362 case AArch64::STPQpost:
4363 case AArch64::STPSpost:
4364 case AArch64::STPWpost:
4365 case AArch64::STPXpost:
4366 case AArch64::STRBBpost:
4367 case AArch64::STRBpost:
4368 case AArch64::STRDpost:
4369 case AArch64::STRHHpost:
4370 case AArch64::STRHpost:
4371 case AArch64::STRQpost:
4372 case AArch64::STRSpost:
4373 case AArch64::STRWpost:
4374 case AArch64::STRXpost:
4375 case AArch64::STZ2GPostIndex:
4376 case AArch64::STZGPostIndex:
4383 bool &OffsetIsScalable,
TypeSize &Width,
4404 int64_t Dummy1, Dummy2;
4426 return BaseOp->
isReg() || BaseOp->
isFI();
4433 assert(OfsOp.
isImm() &&
"Offset operand wasn't immediate.");
4438 TypeSize &Width, int64_t &MinOffset,
4439 int64_t &MaxOffset) {
4445 MinOffset = MaxOffset = 0;
4448 case AArch64::LDRQui:
4449 case AArch64::STRQui:
4455 case AArch64::LDRXui:
4456 case AArch64::LDRDui:
4457 case AArch64::STRXui:
4458 case AArch64::STRDui:
4459 case AArch64::PRFMui:
4465 case AArch64::LDRWui:
4466 case AArch64::LDRSui:
4467 case AArch64::LDRSWui:
4468 case AArch64::STRWui:
4469 case AArch64::STRSui:
4475 case AArch64::LDRHui:
4476 case AArch64::LDRHHui:
4477 case AArch64::LDRSHWui:
4478 case AArch64::LDRSHXui:
4479 case AArch64::STRHui:
4480 case AArch64::STRHHui:
4486 case AArch64::LDRBui:
4487 case AArch64::LDRBBui:
4488 case AArch64::LDRSBWui:
4489 case AArch64::LDRSBXui:
4490 case AArch64::STRBui:
4491 case AArch64::STRBBui:
4498 case AArch64::STRQpre:
4499 case AArch64::LDRQpost:
4505 case AArch64::LDRDpost:
4506 case AArch64::LDRDpre:
4507 case AArch64::LDRXpost:
4508 case AArch64::LDRXpre:
4509 case AArch64::STRDpost:
4510 case AArch64::STRDpre:
4511 case AArch64::STRXpost:
4512 case AArch64::STRXpre:
4518 case AArch64::STRWpost:
4519 case AArch64::STRWpre:
4520 case AArch64::LDRWpost:
4521 case AArch64::LDRWpre:
4522 case AArch64::STRSpost:
4523 case AArch64::STRSpre:
4524 case AArch64::LDRSpost:
4525 case AArch64::LDRSpre:
4531 case AArch64::LDRHpost:
4532 case AArch64::LDRHpre:
4533 case AArch64::STRHpost:
4534 case AArch64::STRHpre:
4535 case AArch64::LDRHHpost:
4536 case AArch64::LDRHHpre:
4537 case AArch64::STRHHpost:
4538 case AArch64::STRHHpre:
4544 case AArch64::LDRBpost:
4545 case AArch64::LDRBpre:
4546 case AArch64::STRBpost:
4547 case AArch64::STRBpre:
4548 case AArch64::LDRBBpost:
4549 case AArch64::LDRBBpre:
4550 case AArch64::STRBBpost:
4551 case AArch64::STRBBpre:
4558 case AArch64::LDURQi:
4559 case AArch64::STURQi:
4565 case AArch64::LDURXi:
4566 case AArch64::LDURDi:
4567 case AArch64::LDAPURXi:
4568 case AArch64::STURXi:
4569 case AArch64::STURDi:
4570 case AArch64::STLURXi:
4571 case AArch64::PRFUMi:
4577 case AArch64::LDURWi:
4578 case AArch64::LDURSi:
4579 case AArch64::LDURSWi:
4580 case AArch64::LDAPURi:
4581 case AArch64::LDAPURSWi:
4582 case AArch64::STURWi:
4583 case AArch64::STURSi:
4584 case AArch64::STLURWi:
4590 case AArch64::LDURHi:
4591 case AArch64::LDURHHi:
4592 case AArch64::LDURSHXi:
4593 case AArch64::LDURSHWi:
4594 case AArch64::LDAPURHi:
4595 case AArch64::LDAPURSHWi:
4596 case AArch64::LDAPURSHXi:
4597 case AArch64::STURHi:
4598 case AArch64::STURHHi:
4599 case AArch64::STLURHi:
4605 case AArch64::LDURBi:
4606 case AArch64::LDURBBi:
4607 case AArch64::LDURSBXi:
4608 case AArch64::LDURSBWi:
4609 case AArch64::LDAPURBi:
4610 case AArch64::LDAPURSBWi:
4611 case AArch64::LDAPURSBXi:
4612 case AArch64::STURBi:
4613 case AArch64::STURBBi:
4614 case AArch64::STLURBi:
4621 case AArch64::LDPQi:
4622 case AArch64::LDNPQi:
4623 case AArch64::STPQi:
4624 case AArch64::STNPQi:
4625 case AArch64::LDPQpost:
4626 case AArch64::LDPQpre:
4627 case AArch64::STPQpost:
4628 case AArch64::STPQpre:
4634 case AArch64::LDPXi:
4635 case AArch64::LDPDi:
4636 case AArch64::LDNPXi:
4637 case AArch64::LDNPDi:
4638 case AArch64::STPXi:
4639 case AArch64::STPDi:
4640 case AArch64::STNPXi:
4641 case AArch64::STNPDi:
4642 case AArch64::LDPDpost:
4643 case AArch64::LDPDpre:
4644 case AArch64::LDPXpost:
4645 case AArch64::LDPXpre:
4646 case AArch64::STPDpost:
4647 case AArch64::STPDpre:
4648 case AArch64::STPXpost:
4649 case AArch64::STPXpre:
4655 case AArch64::LDPWi:
4656 case AArch64::LDPSi:
4657 case AArch64::LDNPWi:
4658 case AArch64::LDNPSi:
4659 case AArch64::STPWi:
4660 case AArch64::STPSi:
4661 case AArch64::STNPWi:
4662 case AArch64::STNPSi:
4663 case AArch64::LDPSpost:
4664 case AArch64::LDPSpre:
4665 case AArch64::LDPWpost:
4666 case AArch64::LDPWpre:
4667 case AArch64::STPSpost:
4668 case AArch64::STPSpre:
4669 case AArch64::STPWpost:
4670 case AArch64::STPWpre:
4676 case AArch64::StoreSwiftAsyncContext:
4689 case AArch64::TAGPstack:
4699 case AArch64::STGPreIndex:
4700 case AArch64::STGPostIndex:
4701 case AArch64::STZGi:
4702 case AArch64::STZGPreIndex:
4703 case AArch64::STZGPostIndex:
4710 case AArch64::STR_ZZZZXI:
4711 case AArch64::STR_ZZZZXI_STRIDED_CONTIGUOUS:
4712 case AArch64::LDR_ZZZZXI:
4713 case AArch64::LDR_ZZZZXI_STRIDED_CONTIGUOUS:
4719 case AArch64::STR_ZZZXI:
4720 case AArch64::LDR_ZZZXI:
4726 case AArch64::STR_ZZXI:
4727 case AArch64::STR_ZZXI_STRIDED_CONTIGUOUS:
4728 case AArch64::LDR_ZZXI:
4729 case AArch64::LDR_ZZXI_STRIDED_CONTIGUOUS:
4735 case AArch64::LDR_PXI:
4736 case AArch64::STR_PXI:
4742 case AArch64::LDR_PPXI:
4743 case AArch64::STR_PPXI:
4749 case AArch64::LDR_ZXI:
4750 case AArch64::STR_ZXI:
4756 case AArch64::LD1B_IMM:
4757 case AArch64::LD1H_IMM:
4758 case AArch64::LD1W_IMM:
4759 case AArch64::LD1D_IMM:
4760 case AArch64::LDNT1B_ZRI:
4761 case AArch64::LDNT1H_ZRI:
4762 case AArch64::LDNT1W_ZRI:
4763 case AArch64::LDNT1D_ZRI:
4764 case AArch64::ST1B_IMM:
4765 case AArch64::ST1H_IMM:
4766 case AArch64::ST1W_IMM:
4767 case AArch64::ST1D_IMM:
4768 case AArch64::STNT1B_ZRI:
4769 case AArch64::STNT1H_ZRI:
4770 case AArch64::STNT1W_ZRI:
4771 case AArch64::STNT1D_ZRI:
4772 case AArch64::LDNF1B_IMM:
4773 case AArch64::LDNF1H_IMM:
4774 case AArch64::LDNF1W_IMM:
4775 case AArch64::LDNF1D_IMM:
4783 case AArch64::LD2B_IMM:
4784 case AArch64::LD2H_IMM:
4785 case AArch64::LD2W_IMM:
4786 case AArch64::LD2D_IMM:
4787 case AArch64::ST2B_IMM:
4788 case AArch64::ST2H_IMM:
4789 case AArch64::ST2W_IMM:
4790 case AArch64::ST2D_IMM:
4796 case AArch64::LD3B_IMM:
4797 case AArch64::LD3H_IMM:
4798 case AArch64::LD3W_IMM:
4799 case AArch64::LD3D_IMM:
4800 case AArch64::ST3B_IMM:
4801 case AArch64::ST3H_IMM:
4802 case AArch64::ST3W_IMM:
4803 case AArch64::ST3D_IMM:
4809 case AArch64::LD4B_IMM:
4810 case AArch64::LD4H_IMM:
4811 case AArch64::LD4W_IMM:
4812 case AArch64::LD4D_IMM:
4813 case AArch64::ST4B_IMM:
4814 case AArch64::ST4H_IMM:
4815 case AArch64::ST4W_IMM:
4816 case AArch64::ST4D_IMM:
4822 case AArch64::LD1B_H_IMM:
4823 case AArch64::LD1SB_H_IMM:
4824 case AArch64::LD1H_S_IMM:
4825 case AArch64::LD1SH_S_IMM:
4826 case AArch64::LD1W_D_IMM:
4827 case AArch64::LD1SW_D_IMM:
4828 case AArch64::ST1B_H_IMM:
4829 case AArch64::ST1H_S_IMM:
4830 case AArch64::ST1W_D_IMM:
4831 case AArch64::LDNF1B_H_IMM:
4832 case AArch64::LDNF1SB_H_IMM:
4833 case AArch64::LDNF1H_S_IMM:
4834 case AArch64::LDNF1SH_S_IMM:
4835 case AArch64::LDNF1W_D_IMM:
4836 case AArch64::LDNF1SW_D_IMM:
4844 case AArch64::LD1B_S_IMM:
4845 case AArch64::LD1SB_S_IMM:
4846 case AArch64::LD1H_D_IMM:
4847 case AArch64::LD1SH_D_IMM:
4848 case AArch64::ST1B_S_IMM:
4849 case AArch64::ST1H_D_IMM:
4850 case AArch64::LDNF1B_S_IMM:
4851 case AArch64::LDNF1SB_S_IMM:
4852 case AArch64::LDNF1H_D_IMM:
4853 case AArch64::LDNF1SH_D_IMM:
4861 case AArch64::LD1B_D_IMM:
4862 case AArch64::LD1SB_D_IMM:
4863 case AArch64::ST1B_D_IMM:
4864 case AArch64::LDNF1B_D_IMM:
4865 case AArch64::LDNF1SB_D_IMM:
4873 case AArch64::ST2Gi:
4874 case AArch64::ST2GPreIndex:
4875 case AArch64::ST2GPostIndex:
4876 case AArch64::STZ2Gi:
4877 case AArch64::STZ2GPreIndex:
4878 case AArch64::STZ2GPostIndex:
4884 case AArch64::STGPi:
4885 case AArch64::STGPpost:
4886 case AArch64::STGPpre:
4892 case AArch64::LD1RB_IMM:
4893 case AArch64::LD1RB_H_IMM:
4894 case AArch64::LD1RB_S_IMM:
4895 case AArch64::LD1RB_D_IMM:
4896 case AArch64::LD1RSB_H_IMM:
4897 case AArch64::LD1RSB_S_IMM:
4898 case AArch64::LD1RSB_D_IMM:
4904 case AArch64::LD1RH_IMM:
4905 case AArch64::LD1RH_S_IMM:
4906 case AArch64::LD1RH_D_IMM:
4907 case AArch64::LD1RSH_S_IMM:
4908 case AArch64::LD1RSH_D_IMM:
4914 case AArch64::LD1RW_IMM:
4915 case AArch64::LD1RW_D_IMM:
4916 case AArch64::LD1RSW_IMM:
4922 case AArch64::LD1RD_IMM:
4938 case AArch64::LDRBBui:
4939 case AArch64::LDURBBi:
4940 case AArch64::LDRSBWui:
4941 case AArch64::LDURSBWi:
4942 case AArch64::STRBBui:
4943 case AArch64::STURBBi:
4945 case AArch64::LDRHHui:
4946 case AArch64::LDURHHi:
4947 case AArch64::LDRSHWui:
4948 case AArch64::LDURSHWi:
4949 case AArch64::STRHHui:
4950 case AArch64::STURHHi:
4952 case AArch64::LDRSui:
4953 case AArch64::LDURSi:
4954 case AArch64::LDRSpre:
4955 case AArch64::LDRSWui:
4956 case AArch64::LDURSWi:
4957 case AArch64::LDRSWpre:
4958 case AArch64::LDRWpre:
4959 case AArch64::LDRWui:
4960 case AArch64::LDURWi:
4961 case AArch64::STRSui:
4962 case AArch64::STURSi:
4963 case AArch64::STRSpre:
4964 case AArch64::STRWui:
4965 case AArch64::STURWi:
4966 case AArch64::STRWpre:
4967 case AArch64::LDPSi:
4968 case AArch64::LDPSWi:
4969 case AArch64::LDPWi:
4970 case AArch64::STPSi:
4971 case AArch64::STPWi:
4973 case AArch64::LDRDui:
4974 case AArch64::LDURDi:
4975 case AArch64::LDRDpre:
4976 case AArch64::LDRXui:
4977 case AArch64::LDURXi:
4978 case AArch64::LDRXpre:
4979 case AArch64::STRDui:
4980 case AArch64::STURDi:
4981 case AArch64::STRDpre:
4982 case AArch64::STRXui:
4983 case AArch64::STURXi:
4984 case AArch64::STRXpre:
4985 case AArch64::LDPDi:
4986 case AArch64::LDPXi:
4987 case AArch64::STPDi:
4988 case AArch64::STPXi:
4990 case AArch64::LDRQui:
4991 case AArch64::LDURQi:
4992 case AArch64::STRQui:
4993 case AArch64::STURQi:
4994 case AArch64::STRQpre:
4995 case AArch64::LDPQi:
4996 case AArch64::LDRQpre:
4997 case AArch64::STPQi:
4999 case AArch64::STZGi:
5000 case AArch64::ST2Gi:
5001 case AArch64::STZ2Gi:
5002 case AArch64::STGPi:
5008 switch (
MI.getOpcode()) {
5011 case AArch64::LDRWpre:
5012 case AArch64::LDRXpre:
5013 case AArch64::LDRSWpre:
5014 case AArch64::LDRSpre:
5015 case AArch64::LDRDpre:
5016 case AArch64::LDRQpre:
5022 switch (
MI.getOpcode()) {
5025 case AArch64::STRWpre:
5026 case AArch64::STRXpre:
5027 case AArch64::STRSpre:
5028 case AArch64::STRDpre:
5029 case AArch64::STRQpre:
5039 switch (
MI.getOpcode()) {
5042 case AArch64::LDPSi:
5043 case AArch64::LDPSWi:
5044 case AArch64::LDPDi:
5045 case AArch64::LDPQi:
5046 case AArch64::LDPWi:
5047 case AArch64::LDPXi:
5048 case AArch64::STPSi:
5049 case AArch64::STPDi:
5050 case AArch64::STPQi:
5051 case AArch64::STPWi:
5052 case AArch64::STPXi:
5053 case AArch64::STGPi:
5059 assert(
MI.mayLoadOrStore() &&
"Load or store instruction expected");
5063 return MI.getOperand(Idx);
5068 assert(
MI.mayLoadOrStore() &&
"Load or store instruction expected");
5072 return MI.getOperand(Idx);
5077 switch (
MI.getOpcode()) {
5080 case AArch64::LDRBroX:
5081 case AArch64::LDRBBroX:
5082 case AArch64::LDRSBXroX:
5083 case AArch64::LDRSBWroX:
5084 case AArch64::LDRHroX:
5085 case AArch64::LDRHHroX:
5086 case AArch64::LDRSHXroX:
5087 case AArch64::LDRSHWroX:
5088 case AArch64::LDRWroX:
5089 case AArch64::LDRSroX:
5090 case AArch64::LDRSWroX:
5091 case AArch64::LDRDroX:
5092 case AArch64::LDRXroX:
5093 case AArch64::LDRQroX:
5094 return MI.getOperand(4);
5100 if (
MI.getParent() ==
nullptr)
5110 auto Reg =
Op.getReg();
5111 if (Reg.isPhysical())
5112 return AArch64::FPR16RegClass.contains(Reg);
5114 return TRC == &AArch64::FPR16RegClass ||
5115 TRC == &AArch64::FPR16_loRegClass;
5124 auto Reg =
Op.getReg();
5125 if (Reg.isPhysical())
5126 return AArch64::FPR128RegClass.contains(Reg);
5128 return TRC == &AArch64::FPR128RegClass ||
5129 TRC == &AArch64::FPR128_loRegClass;
5135 switch (
MI.getOpcode()) {
5138 case AArch64::PACIASP:
5139 case AArch64::PACIBSP:
5142 case AArch64::PAUTH_PROLOGUE:
5145 case AArch64::HINT: {
5146 unsigned Imm =
MI.getOperand(0).getImm();
5148 if (Imm == 32 || Imm == 34 || Imm == 36 || Imm == 38)
5151 if (Imm == 25 || Imm == 27)
5163 assert(Reg.isPhysical() &&
"Expected physical register in isFpOrNEON");
5164 return AArch64::FPR128RegClass.contains(Reg) ||
5165 AArch64::FPR64RegClass.contains(Reg) ||
5166 AArch64::FPR32RegClass.contains(Reg) ||
5167 AArch64::FPR16RegClass.contains(Reg) ||
5168 AArch64::FPR8RegClass.contains(Reg);
5175 auto Reg =
Op.getReg();
5176 if (Reg.isPhysical())
5180 return TRC == &AArch64::FPR128RegClass ||
5181 TRC == &AArch64::FPR128_loRegClass ||
5182 TRC == &AArch64::FPR64RegClass ||
5183 TRC == &AArch64::FPR64_loRegClass ||
5184 TRC == &AArch64::FPR32RegClass || TRC == &AArch64::FPR16RegClass ||
5185 TRC == &AArch64::FPR8RegClass;
5207 if (FirstOpc == SecondOpc)
5213 case AArch64::STRSui:
5214 case AArch64::STURSi:
5215 return SecondOpc == AArch64::STRSui || SecondOpc == AArch64::STURSi;
5216 case AArch64::STRDui:
5217 case AArch64::STURDi:
5218 return SecondOpc == AArch64::STRDui || SecondOpc == AArch64::STURDi;
5219 case AArch64::STRQui:
5220 case AArch64::STURQi:
5221 return SecondOpc == AArch64::STRQui || SecondOpc == AArch64::STURQi;
5222 case AArch64::STRWui:
5223 case AArch64::STURWi:
5224 return SecondOpc == AArch64::STRWui || SecondOpc == AArch64::STURWi;
5225 case AArch64::STRXui:
5226 case AArch64::STURXi:
5227 return SecondOpc == AArch64::STRXui || SecondOpc == AArch64::STURXi;
5228 case AArch64::LDRSui:
5229 case AArch64::LDURSi:
5230 return SecondOpc == AArch64::LDRSui || SecondOpc == AArch64::LDURSi;
5231 case AArch64::LDRDui:
5232 case AArch64::LDURDi:
5233 return SecondOpc == AArch64::LDRDui || SecondOpc == AArch64::LDURDi;
5234 case AArch64::LDRQui:
5235 case AArch64::LDURQi:
5236 return SecondOpc == AArch64::LDRQui || SecondOpc == AArch64::LDURQi;
5237 case AArch64::LDRWui:
5238 case AArch64::LDURWi:
5239 return SecondOpc == AArch64::LDRSWui || SecondOpc == AArch64::LDURSWi;
5240 case AArch64::LDRSWui:
5241 case AArch64::LDURSWi:
5242 return SecondOpc == AArch64::LDRWui || SecondOpc == AArch64::LDURWi;
5243 case AArch64::LDRXui:
5244 case AArch64::LDURXi:
5245 return SecondOpc == AArch64::LDRXui || SecondOpc == AArch64::LDURXi;
5252 int64_t Offset1,
unsigned Opcode1,
int FI2,
5253 int64_t Offset2,
unsigned Opcode2) {
5259 assert(ObjectOffset1 <= ObjectOffset2 &&
"Object offsets are not ordered.");
5262 if (ObjectOffset1 % Scale1 != 0)
5264 ObjectOffset1 /= Scale1;
5266 if (ObjectOffset2 % Scale2 != 0)
5268 ObjectOffset2 /= Scale2;
5269 ObjectOffset1 += Offset1;
5270 ObjectOffset2 += Offset2;
5271 return ObjectOffset1 + 1 == ObjectOffset2;
5283 int64_t OpOffset2,
bool OffsetIsScalable2,
unsigned ClusterSize,
5284 unsigned NumBytes)
const {
5294 "Only base registers and frame indices are supported.");
5301 if (ClusterSize > 2)
5308 unsigned FirstOpc = FirstLdSt.
getOpcode();
5309 unsigned SecondOpc = SecondLdSt.
getOpcode();
5329 if (Offset1 > 63 || Offset1 < -64)
5334 if (BaseOp1.
isFI()) {
5336 "Caller should have ordered offsets.");
5341 BaseOp2.
getIndex(), Offset2, SecondOpc);
5344 assert(Offset1 <= Offset2 &&
"Caller should have ordered offsets.");
5346 return Offset1 + 1 == Offset2;
5356 if (
Reg.isPhysical())
5365 return ((DestReg - SrcReg) & 0x1f) < NumRegs;
5374 assert(Subtarget.hasNEON() &&
"Unexpected register copy without NEON");
5376 uint16_t DestEncoding =
TRI->getEncodingValue(DestReg);
5377 uint16_t SrcEncoding =
TRI->getEncodingValue(SrcReg);
5378 unsigned NumRegs = Indices.
size();
5380 int SubReg = 0, End = NumRegs, Incr = 1;
5399 unsigned Opcode,
unsigned ZeroReg,
5402 unsigned NumRegs = Indices.
size();
5405 uint16_t DestEncoding =
TRI->getEncodingValue(DestReg);
5406 uint16_t SrcEncoding =
TRI->getEncodingValue(SrcReg);
5407 assert(DestEncoding % NumRegs == 0 && SrcEncoding % NumRegs == 0 &&
5408 "GPR reg sequences should not be able to overlap");
5432 unsigned Opc =
MI.getOpcode();
5433 if (
Opc == AArch64::MSRpstatesvcrImm1 ||
Opc == AArch64::MSRpstatePseudo) {
5435 int64_t PState =
MI.getOperand(0).getImm();
5436 if (PState == AArch64SVCR::SVCRSM || PState == AArch64SVCR::SVCRSMZA) {
5438 return MI.getOperand(1).getImm() == 1;
5457 bool RenamableSrc)
const {
5459 if (AArch64::GPR32spRegClass.
contains(DestReg) &&
5460 AArch64::GPR32spRegClass.
contains(SrcReg)) {
5461 if (DestReg == AArch64::WSP || SrcReg == AArch64::WSP) {
5463 if (Subtarget.hasZeroCycleRegMoveGPR64() &&
5464 !Subtarget.hasZeroCycleRegMoveGPR32()) {
5466 MCRegister DestRegX = RI.getMatchingSuperReg(DestReg, AArch64::sub_32,
5467 &AArch64::GPR64spRegClass);
5468 MCRegister SrcRegX = RI.getMatchingSuperReg(SrcReg, AArch64::sub_32,
5469 &AArch64::GPR64spRegClass);
5479 ++NumZCRegMoveInstrsGPR;
5485 if (Subtarget.hasZeroCycleRegMoveGPR32())
5486 ++NumZCRegMoveInstrsGPR;
5488 }
else if (Subtarget.hasZeroCycleRegMoveGPR64() &&
5489 !Subtarget.hasZeroCycleRegMoveGPR32()) {
5491 MCRegister DestRegX = RI.getMatchingSuperReg(DestReg, AArch64::sub_32,
5492 &AArch64::GPR64spRegClass);
5493 assert(DestRegX.
isValid() &&
"Destination super-reg not valid");
5494 MCRegister SrcRegX = RI.getMatchingSuperReg(SrcReg, AArch64::sub_32,
5495 &AArch64::GPR64spRegClass);
5505 ++NumZCRegMoveInstrsGPR;
5511 if (Subtarget.hasZeroCycleRegMoveGPR32())
5512 ++NumZCRegMoveInstrsGPR;
5518 if (AArch64::GPR32spRegClass.
contains(DestReg) && SrcReg == AArch64::WZR) {
5519 if (Subtarget.hasZeroCycleZeroingGPR64() &&
5520 !Subtarget.hasZeroCycleZeroingGPR32()) {
5521 MCRegister DestRegX = RI.getMatchingSuperReg(DestReg, AArch64::sub_32,
5522 &AArch64::GPR64spRegClass);
5523 assert(DestRegX.
isValid() &&
"Destination super-reg not valid");
5527 ++NumZCZeroingInstrsGPR;
5528 }
else if (Subtarget.hasZeroCycleZeroingGPR32()) {
5532 ++NumZCZeroingInstrsGPR;
5541 if (AArch64::GPR64spRegClass.
contains(DestReg) &&
5542 AArch64::GPR64spRegClass.
contains(SrcReg)) {
5543 if (DestReg == AArch64::SP || SrcReg == AArch64::SP) {
5549 if (Subtarget.hasZeroCycleRegMoveGPR64())
5550 ++NumZCRegMoveInstrsGPR;
5556 if (Subtarget.hasZeroCycleRegMoveGPR64())
5557 ++NumZCRegMoveInstrsGPR;
5563 if (AArch64::GPR64spRegClass.
contains(DestReg) && SrcReg == AArch64::XZR) {
5564 if (Subtarget.hasZeroCycleZeroingGPR64()) {
5568 ++NumZCZeroingInstrsGPR;
5578 if (AArch64::PPRRegClass.
contains(DestReg) &&
5579 AArch64::PPRRegClass.
contains(SrcReg)) {
5580 assert(Subtarget.isSVEorStreamingSVEAvailable() &&
5581 "Unexpected SVE register.");
5591 bool DestIsPNR = AArch64::PNRRegClass.contains(DestReg);
5592 bool SrcIsPNR = AArch64::PNRRegClass.contains(SrcReg);
5593 if (DestIsPNR || SrcIsPNR) {
5595 return (R - AArch64::PN0) + AArch64::P0;
5600 if (PPRSrcReg != PPRDestReg) {
5612 if (AArch64::ZPRRegClass.
contains(DestReg) &&
5613 AArch64::ZPRRegClass.
contains(SrcReg)) {
5614 assert(Subtarget.isSVEorStreamingSVEAvailable() &&
5615 "Unexpected SVE register.");
5623 if ((AArch64::ZPR2RegClass.
contains(DestReg) ||
5624 AArch64::ZPR2StridedOrContiguousRegClass.
contains(DestReg)) &&
5625 (AArch64::ZPR2RegClass.
contains(SrcReg) ||
5626 AArch64::ZPR2StridedOrContiguousRegClass.
contains(SrcReg))) {
5627 assert(Subtarget.isSVEorStreamingSVEAvailable() &&
5628 "Unexpected SVE register.");
5629 static const unsigned Indices[] = {AArch64::zsub0, AArch64::zsub1};
5636 if (AArch64::ZPR3RegClass.
contains(DestReg) &&
5637 AArch64::ZPR3RegClass.
contains(SrcReg)) {
5638 assert(Subtarget.isSVEorStreamingSVEAvailable() &&
5639 "Unexpected SVE register.");
5640 static const unsigned Indices[] = {AArch64::zsub0, AArch64::zsub1,
5648 if ((AArch64::ZPR4RegClass.
contains(DestReg) ||
5649 AArch64::ZPR4StridedOrContiguousRegClass.
contains(DestReg)) &&
5650 (AArch64::ZPR4RegClass.
contains(SrcReg) ||
5651 AArch64::ZPR4StridedOrContiguousRegClass.
contains(SrcReg))) {
5652 assert(Subtarget.isSVEorStreamingSVEAvailable() &&
5653 "Unexpected SVE register.");
5654 static const unsigned Indices[] = {AArch64::zsub0, AArch64::zsub1,
5655 AArch64::zsub2, AArch64::zsub3};
5662 if (AArch64::DDDDRegClass.
contains(DestReg) &&
5663 AArch64::DDDDRegClass.
contains(SrcReg)) {
5664 static const unsigned Indices[] = {AArch64::dsub0, AArch64::dsub1,
5665 AArch64::dsub2, AArch64::dsub3};
5672 if (AArch64::DDDRegClass.
contains(DestReg) &&
5673 AArch64::DDDRegClass.
contains(SrcReg)) {
5674 static const unsigned Indices[] = {AArch64::dsub0, AArch64::dsub1,
5682 if (AArch64::DDRegClass.
contains(DestReg) &&
5683 AArch64::DDRegClass.
contains(SrcReg)) {
5684 static const unsigned Indices[] = {AArch64::dsub0, AArch64::dsub1};
5691 if (AArch64::QQQQRegClass.
contains(DestReg) &&
5692 AArch64::QQQQRegClass.
contains(SrcReg)) {
5693 static const unsigned Indices[] = {AArch64::qsub0, AArch64::qsub1,
5694 AArch64::qsub2, AArch64::qsub3};
5701 if (AArch64::QQQRegClass.
contains(DestReg) &&
5702 AArch64::QQQRegClass.
contains(SrcReg)) {
5703 static const unsigned Indices[] = {AArch64::qsub0, AArch64::qsub1,
5711 if (AArch64::QQRegClass.
contains(DestReg) &&
5712 AArch64::QQRegClass.
contains(SrcReg)) {
5713 static const unsigned Indices[] = {AArch64::qsub0, AArch64::qsub1};
5719 if (AArch64::XSeqPairsClassRegClass.
contains(DestReg) &&
5720 AArch64::XSeqPairsClassRegClass.
contains(SrcReg)) {
5721 static const unsigned Indices[] = {AArch64::sube64, AArch64::subo64};
5723 AArch64::XZR, Indices);
5727 if (AArch64::WSeqPairsClassRegClass.
contains(DestReg) &&
5728 AArch64::WSeqPairsClassRegClass.
contains(SrcReg)) {
5729 static const unsigned Indices[] = {AArch64::sube32, AArch64::subo32};
5731 AArch64::WZR, Indices);
5735 if (AArch64::FPR128RegClass.
contains(DestReg) &&
5736 AArch64::FPR128RegClass.
contains(SrcReg)) {
5740 if ((Subtarget.isSVEorStreamingSVEAvailable() &&
5741 !Subtarget.isNeonAvailable()) ||
5745 .
addReg(AArch64::Z0 + (SrcReg - AArch64::Q0))
5746 .
addReg(AArch64::Z0 + (SrcReg - AArch64::Q0));
5747 }
else if (Subtarget.isNeonAvailable()) {
5751 if (Subtarget.hasZeroCycleRegMoveFPR128())
5752 ++NumZCRegMoveInstrsFPR;
5768 if (AArch64::FPR64RegClass.
contains(DestReg) &&
5769 AArch64::FPR64RegClass.
contains(SrcReg)) {
5770 if (Subtarget.hasZeroCycleRegMoveFPR128() &&
5771 !Subtarget.hasZeroCycleRegMoveFPR64() &&
5772 !Subtarget.hasZeroCycleRegMoveFPR32() && Subtarget.isNeonAvailable() &&
5774 MCRegister DestRegQ = RI.getMatchingSuperReg(DestReg, AArch64::dsub,
5775 &AArch64::FPR128RegClass);
5776 MCRegister SrcRegQ = RI.getMatchingSuperReg(SrcReg, AArch64::dsub,
5777 &AArch64::FPR128RegClass);
5786 ++NumZCRegMoveInstrsFPR;
5790 if (Subtarget.hasZeroCycleRegMoveFPR64())
5791 ++NumZCRegMoveInstrsFPR;
5796 if (AArch64::FPR32RegClass.
contains(DestReg) &&
5797 AArch64::FPR32RegClass.
contains(SrcReg)) {
5798 if (Subtarget.hasZeroCycleRegMoveFPR128() &&
5799 !Subtarget.hasZeroCycleRegMoveFPR64() &&
5800 !Subtarget.hasZeroCycleRegMoveFPR32() && Subtarget.isNeonAvailable() &&
5802 MCRegister DestRegQ = RI.getMatchingSuperReg(DestReg, AArch64::ssub,
5803 &AArch64::FPR128RegClass);
5804 MCRegister SrcRegQ = RI.getMatchingSuperReg(SrcReg, AArch64::ssub,
5805 &AArch64::FPR128RegClass);
5814 ++NumZCRegMoveInstrsFPR;
5815 }
else if (Subtarget.hasZeroCycleRegMoveFPR64() &&
5816 !Subtarget.hasZeroCycleRegMoveFPR32()) {
5817 MCRegister DestRegD = RI.getMatchingSuperReg(DestReg, AArch64::ssub,
5818 &AArch64::FPR64RegClass);
5819 MCRegister SrcRegD = RI.getMatchingSuperReg(SrcReg, AArch64::ssub,
5820 &AArch64::FPR64RegClass);
5828 ++NumZCRegMoveInstrsFPR;
5832 if (Subtarget.hasZeroCycleRegMoveFPR32())
5833 ++NumZCRegMoveInstrsFPR;
5838 if (AArch64::FPR16RegClass.
contains(DestReg) &&
5839 AArch64::FPR16RegClass.
contains(SrcReg)) {
5840 if (Subtarget.hasZeroCycleRegMoveFPR128() &&
5841 !Subtarget.hasZeroCycleRegMoveFPR64() &&
5842 !Subtarget.hasZeroCycleRegMoveFPR32() && Subtarget.isNeonAvailable() &&
5844 MCRegister DestRegQ = RI.getMatchingSuperReg(DestReg, AArch64::hsub,
5845 &AArch64::FPR128RegClass);
5846 MCRegister SrcRegQ = RI.getMatchingSuperReg(SrcReg, AArch64::hsub,
5847 &AArch64::FPR128RegClass);
5856 }
else if (Subtarget.hasZeroCycleRegMoveFPR64() &&
5857 !Subtarget.hasZeroCycleRegMoveFPR32()) {
5858 MCRegister DestRegD = RI.getMatchingSuperReg(DestReg, AArch64::hsub,
5859 &AArch64::FPR64RegClass);
5860 MCRegister SrcRegD = RI.getMatchingSuperReg(SrcReg, AArch64::hsub,
5861 &AArch64::FPR64RegClass);
5870 DestReg = RI.getMatchingSuperReg(DestReg, AArch64::hsub,
5871 &AArch64::FPR32RegClass);
5872 SrcReg = RI.getMatchingSuperReg(SrcReg, AArch64::hsub,
5873 &AArch64::FPR32RegClass);
5880 if (AArch64::FPR8RegClass.
contains(DestReg) &&
5881 AArch64::FPR8RegClass.
contains(SrcReg)) {
5882 if (Subtarget.hasZeroCycleRegMoveFPR128() &&
5883 !Subtarget.hasZeroCycleRegMoveFPR64() &&
5884 !Subtarget.hasZeroCycleRegMoveFPR32() && Subtarget.isNeonAvailable() &&
5886 MCRegister DestRegQ = RI.getMatchingSuperReg(DestReg, AArch64::bsub,
5887 &AArch64::FPR128RegClass);
5888 MCRegister SrcRegQ = RI.getMatchingSuperReg(SrcReg, AArch64::bsub,
5889 &AArch64::FPR128RegClass);
5898 }
else if (Subtarget.hasZeroCycleRegMoveFPR64() &&
5899 !Subtarget.hasZeroCycleRegMoveFPR32()) {
5900 MCRegister DestRegD = RI.getMatchingSuperReg(DestReg, AArch64::bsub,
5901 &AArch64::FPR64RegClass);
5902 MCRegister SrcRegD = RI.getMatchingSuperReg(SrcReg, AArch64::bsub,
5903 &AArch64::FPR64RegClass);
5912 DestReg = RI.getMatchingSuperReg(DestReg, AArch64::bsub,
5913 &AArch64::FPR32RegClass);
5914 SrcReg = RI.getMatchingSuperReg(SrcReg, AArch64::bsub,
5915 &AArch64::FPR32RegClass);
5923 if (AArch64::FPR64RegClass.
contains(DestReg) &&
5924 AArch64::GPR64RegClass.
contains(SrcReg)) {
5925 if (AArch64::XZR == SrcReg) {
5933 if (AArch64::GPR64RegClass.
contains(DestReg) &&
5934 AArch64::FPR64RegClass.
contains(SrcReg)) {
5940 if (AArch64::FPR32RegClass.
contains(DestReg) &&
5941 AArch64::GPR32RegClass.
contains(SrcReg)) {
5942 if (AArch64::WZR == SrcReg) {
5950 if (AArch64::GPR32RegClass.
contains(DestReg) &&
5951 AArch64::FPR32RegClass.
contains(SrcReg)) {
5957 if (DestReg == AArch64::NZCV) {
5958 assert(AArch64::GPR64RegClass.
contains(SrcReg) &&
"Invalid NZCV copy");
5960 .
addImm(AArch64SysReg::NZCV)
5966 if (SrcReg == AArch64::NZCV) {
5967 assert(AArch64::GPR64RegClass.
contains(DestReg) &&
"Invalid NZCV copy");
5969 .
addImm(AArch64SysReg::NZCV)
5975 errs() << RI.getRegAsmName(DestReg) <<
" = COPY " << RI.getRegAsmName(SrcReg)
5986 unsigned SubIdx0,
unsigned SubIdx1,
int FI,
5991 SrcReg0 =
TRI.getSubReg(SrcReg, SubIdx0);
5993 SrcReg1 =
TRI.getSubReg(SrcReg, SubIdx1);
6006 Register SrcReg,
bool isKill,
int FI,
6021 switch (RI.getSpillSize(*RC)) {
6023 if (AArch64::FPR8RegClass.hasSubClassEq(RC))
6024 Opc = AArch64::STRBui;
6027 if (AArch64::FPR16RegClass.hasSubClassEq(RC))
6028 Opc = AArch64::STRHui;
6029 else if (AArch64::PNRRegClass.hasSubClassEq(RC) ||
6030 AArch64::PPRRegClass.hasSubClassEq(RC)) {
6031 assert(Subtarget.isSVEorStreamingSVEAvailable() &&
6032 "Unexpected register store without SVE store instructions");
6033 Opc = AArch64::STR_PXI;
6039 if (AArch64::GPR32allRegClass.hasSubClassEq(RC)) {
6040 Opc = AArch64::STRWui;
6044 assert(SrcReg != AArch64::WSP);
6045 }
else if (AArch64::FPR32RegClass.hasSubClassEq(RC))
6046 Opc = AArch64::STRSui;
6047 else if (AArch64::PPR2RegClass.hasSubClassEq(RC)) {
6048 Opc = AArch64::STR_PPXI;
6053 if (AArch64::GPR64allRegClass.hasSubClassEq(RC)) {
6054 Opc = AArch64::STRXui;
6058 assert(SrcReg != AArch64::SP);
6059 }
else if (AArch64::FPR64RegClass.hasSubClassEq(RC)) {
6060 Opc = AArch64::STRDui;
6061 }
else if (AArch64::WSeqPairsClassRegClass.hasSubClassEq(RC)) {
6063 get(AArch64::STPWi), SrcReg, isKill,
6064 AArch64::sube32, AArch64::subo32, FI, MMO);
6069 if (AArch64::FPR128RegClass.hasSubClassEq(RC))
6070 Opc = AArch64::STRQui;
6071 else if (AArch64::DDRegClass.hasSubClassEq(RC)) {
6072 assert(Subtarget.hasNEON() &&
"Unexpected register store without NEON");
6073 Opc = AArch64::ST1Twov1d;
6075 }
else if (AArch64::XSeqPairsClassRegClass.hasSubClassEq(RC)) {
6077 get(AArch64::STPXi), SrcReg, isKill,
6078 AArch64::sube64, AArch64::subo64, FI, MMO);
6080 }
else if (AArch64::ZPRRegClass.hasSubClassEq(RC)) {
6081 assert(Subtarget.isSVEorStreamingSVEAvailable() &&
6082 "Unexpected register store without SVE store instructions");
6083 Opc = AArch64::STR_ZXI;
6088 if (AArch64::DDDRegClass.hasSubClassEq(RC)) {
6089 assert(Subtarget.hasNEON() &&
"Unexpected register store without NEON");
6090 Opc = AArch64::ST1Threev1d;
6095 if (AArch64::DDDDRegClass.hasSubClassEq(RC)) {
6096 assert(Subtarget.hasNEON() &&
"Unexpected register store without NEON");
6097 Opc = AArch64::ST1Fourv1d;
6099 }
else if (AArch64::QQRegClass.hasSubClassEq(RC)) {
6100 assert(Subtarget.hasNEON() &&
"Unexpected register store without NEON");
6101 Opc = AArch64::ST1Twov2d;
6103 }
else if (AArch64::ZPR2StridedOrContiguousRegClass.hasSubClassEq(RC)) {
6104 assert(Subtarget.isSVEorStreamingSVEAvailable() &&
6105 "Unexpected register store without SVE store instructions");
6106 Opc = AArch64::STR_ZZXI_STRIDED_CONTIGUOUS;
6108 }
else if (AArch64::ZPR2RegClass.hasSubClassEq(RC)) {
6109 assert(Subtarget.isSVEorStreamingSVEAvailable() &&
6110 "Unexpected register store without SVE store instructions");
6111 Opc = AArch64::STR_ZZXI;
6116 if (AArch64::QQQRegClass.hasSubClassEq(RC)) {
6117 assert(Subtarget.hasNEON() &&
"Unexpected register store without NEON");
6118 Opc = AArch64::ST1Threev2d;
6120 }
else if (AArch64::ZPR3RegClass.hasSubClassEq(RC)) {
6121 assert(Subtarget.isSVEorStreamingSVEAvailable() &&
6122 "Unexpected register store without SVE store instructions");
6123 Opc = AArch64::STR_ZZZXI;
6128 if (AArch64::QQQQRegClass.hasSubClassEq(RC)) {
6129 assert(Subtarget.hasNEON() &&
"Unexpected register store without NEON");
6130 Opc = AArch64::ST1Fourv2d;
6132 }
else if (AArch64::ZPR4StridedOrContiguousRegClass.hasSubClassEq(RC)) {
6133 assert(Subtarget.isSVEorStreamingSVEAvailable() &&
6134 "Unexpected register store without SVE store instructions");
6135 Opc = AArch64::STR_ZZZZXI_STRIDED_CONTIGUOUS;
6137 }
else if (AArch64::ZPR4RegClass.hasSubClassEq(RC)) {
6138 assert(Subtarget.isSVEorStreamingSVEAvailable() &&
6139 "Unexpected register store without SVE store instructions");
6140 Opc = AArch64::STR_ZZZZXI;
6145 assert(
Opc &&
"Unknown register class");
6156 MI.addMemOperand(MMO);
6163 Register DestReg,
unsigned SubIdx0,
6164 unsigned SubIdx1,
int FI,
6168 bool IsUndef =
true;
6170 DestReg0 =
TRI.getSubReg(DestReg, SubIdx0);
6172 DestReg1 =
TRI.getSubReg(DestReg, SubIdx1);
6201 switch (
TRI.getSpillSize(*RC)) {
6203 if (AArch64::FPR8RegClass.hasSubClassEq(RC))
6204 Opc = AArch64::LDRBui;
6207 bool IsPNR = AArch64::PNRRegClass.hasSubClassEq(RC);
6208 if (AArch64::FPR16RegClass.hasSubClassEq(RC))
6209 Opc = AArch64::LDRHui;
6210 else if (IsPNR || AArch64::PPRRegClass.hasSubClassEq(RC)) {
6211 assert(Subtarget.isSVEorStreamingSVEAvailable() &&
6212 "Unexpected register load without SVE load instructions");
6215 Opc = AArch64::LDR_PXI;
6221 if (AArch64::GPR32allRegClass.hasSubClassEq(RC)) {
6222 Opc = AArch64::LDRWui;
6226 assert(DestReg != AArch64::WSP);
6227 }
else if (AArch64::FPR32RegClass.hasSubClassEq(RC))
6228 Opc = AArch64::LDRSui;
6229 else if (AArch64::PPR2RegClass.hasSubClassEq(RC)) {
6230 Opc = AArch64::LDR_PPXI;
6235 if (AArch64::GPR64allRegClass.hasSubClassEq(RC)) {
6236 Opc = AArch64::LDRXui;
6240 assert(DestReg != AArch64::SP);
6241 }
else if (AArch64::FPR64RegClass.hasSubClassEq(RC)) {
6242 Opc = AArch64::LDRDui;
6243 }
else if (AArch64::WSeqPairsClassRegClass.hasSubClassEq(RC)) {
6245 get(AArch64::LDPWi), DestReg, AArch64::sube32,
6246 AArch64::subo32, FI, MMO);
6251 if (AArch64::FPR128RegClass.hasSubClassEq(RC))
6252 Opc = AArch64::LDRQui;
6253 else if (AArch64::DDRegClass.hasSubClassEq(RC)) {
6254 assert(Subtarget.hasNEON() &&
"Unexpected register load without NEON");
6255 Opc = AArch64::LD1Twov1d;
6257 }
else if (AArch64::XSeqPairsClassRegClass.hasSubClassEq(RC)) {
6259 get(AArch64::LDPXi), DestReg, AArch64::sube64,
6260 AArch64::subo64, FI, MMO);
6262 }
else if (AArch64::ZPRRegClass.hasSubClassEq(RC)) {
6263 assert(Subtarget.isSVEorStreamingSVEAvailable() &&
6264 "Unexpected register load without SVE load instructions");
6265 Opc = AArch64::LDR_ZXI;
6270 if (AArch64::DDDRegClass.hasSubClassEq(RC)) {
6271 assert(Subtarget.hasNEON() &&
"Unexpected register load without NEON");
6272 Opc = AArch64::LD1Threev1d;
6277 if (AArch64::DDDDRegClass.hasSubClassEq(RC)) {
6278 assert(Subtarget.hasNEON() &&
"Unexpected register load without NEON");
6279 Opc = AArch64::LD1Fourv1d;
6281 }
else if (AArch64::QQRegClass.hasSubClassEq(RC)) {
6282 assert(Subtarget.hasNEON() &&
"Unexpected register load without NEON");
6283 Opc = AArch64::LD1Twov2d;
6285 }
else if (AArch64::ZPR2StridedOrContiguousRegClass.hasSubClassEq(RC)) {
6286 assert(Subtarget.isSVEorStreamingSVEAvailable() &&
6287 "Unexpected register load without SVE load instructions");
6288 Opc = AArch64::LDR_ZZXI_STRIDED_CONTIGUOUS;
6290 }
else if (AArch64::ZPR2RegClass.hasSubClassEq(RC)) {
6291 assert(Subtarget.isSVEorStreamingSVEAvailable() &&
6292 "Unexpected register load without SVE load instructions");
6293 Opc = AArch64::LDR_ZZXI;
6298 if (AArch64::QQQRegClass.hasSubClassEq(RC)) {
6299 assert(Subtarget.hasNEON() &&
"Unexpected register load without NEON");
6300 Opc = AArch64::LD1Threev2d;
6302 }
else if (AArch64::ZPR3RegClass.hasSubClassEq(RC)) {
6303 assert(Subtarget.isSVEorStreamingSVEAvailable() &&
6304 "Unexpected register load without SVE load instructions");
6305 Opc = AArch64::LDR_ZZZXI;
6310 if (AArch64::QQQQRegClass.hasSubClassEq(RC)) {
6311 assert(Subtarget.hasNEON() &&
"Unexpected register load without NEON");
6312 Opc = AArch64::LD1Fourv2d;
6314 }
else if (AArch64::ZPR4StridedOrContiguousRegClass.hasSubClassEq(RC)) {
6315 assert(Subtarget.isSVEorStreamingSVEAvailable() &&
6316 "Unexpected register load without SVE load instructions");
6317 Opc = AArch64::LDR_ZZZZXI_STRIDED_CONTIGUOUS;
6319 }
else if (AArch64::ZPR4RegClass.hasSubClassEq(RC)) {
6320 assert(Subtarget.isSVEorStreamingSVEAvailable() &&
6321 "Unexpected register load without SVE load instructions");
6322 Opc = AArch64::LDR_ZZZZXI;
6328 assert(
Opc &&
"Unknown register class");
6338 MI.addMemOperand(MMO);
6345 UseMI.getIterator()),
6347 return I.modifiesRegister(AArch64::NZCV, TRI) ||
6348 I.readsRegister(AArch64::NZCV, TRI);
6352void AArch64InstrInfo::decomposeStackOffsetForDwarfOffsets(
6357 assert(
Offset.getScalable() % 2 == 0 &&
"Invalid frame offset");
6364 ByteSized =
Offset.getFixed();
6365 VGSized =
Offset.getScalable() / 2;
6371void AArch64InstrInfo::decomposeStackOffsetForFrameOffsets(
6373 int64_t &NumDataVectors) {
6377 assert(
Offset.getScalable() % 2 == 0 &&
"Invalid frame offset");
6379 NumBytes =
Offset.getFixed();
6381 NumPredicateVectors =
Offset.getScalable() / 2;
6386 if (NumPredicateVectors % 8 == 0 || NumPredicateVectors < -64 ||
6387 NumPredicateVectors > 62) {
6388 NumDataVectors = NumPredicateVectors / 8;
6389 NumPredicateVectors -= NumDataVectors * 8;
6415 Expr.
push_back((
char)dwarf::DW_OP_bregx);
6423 int64_t OffsetFromDefCFA) {
6437 Comment << (NumBytes < 0 ?
" - " :
" + ") << std::abs(NumBytes);
6438 if (!RegScale.empty())
6448 int64_t NumBytes, NumVGScaledBytes;
6449 AArch64InstrInfo::decomposeStackOffsetForDwarfOffsets(
Offset, NumBytes,
6451 std::string CommentBuffer;
6454 if (
Reg == AArch64::SP)
6456 else if (
Reg == AArch64::FP)
6463 unsigned DwarfReg =
TRI.getDwarfRegNum(
Reg,
true);
6464 assert(DwarfReg <= 31 &&
"DwarfReg out of bounds (0..31)");
6466 Expr.
push_back(dwarf::DW_OP_breg0 + DwarfReg);
6469 if (NumVGScaledBytes) {
6479 DefCfaExpr.
push_back(dwarf::DW_CFA_def_cfa_expression);
6487 unsigned FrameReg,
unsigned Reg,
6489 bool LastAdjustmentWasScalable) {
6490 if (
Offset.getScalable())
6493 if (FrameReg == Reg && !LastAdjustmentWasScalable)
6496 unsigned DwarfReg =
TRI.getDwarfRegNum(Reg,
true);
6503 std::optional<int64_t> IncomingVGOffsetFromDefCFA) {
6504 int64_t NumBytes, NumVGScaledBytes;
6505 AArch64InstrInfo::decomposeStackOffsetForDwarfOffsets(
6506 OffsetFromDefCFA, NumBytes, NumVGScaledBytes);
6508 unsigned DwarfReg =
TRI.getDwarfRegNum(Reg,
true);
6511 if (!NumVGScaledBytes)
6514 std::string CommentBuffer;
6519 assert(NumVGScaledBytes &&
"Expected scalable offset");
6523 if (IncomingVGOffsetFromDefCFA) {
6525 VGRegScale =
"* IncomingVG";
6528 VGRegScale =
"* VG";
6532 OffsetExpr.
push_back(dwarf::DW_OP_plus);
6541 CfaExpr.
push_back(dwarf::DW_CFA_expression);
6556 unsigned SrcReg, int64_t
Offset,
unsigned Opc,
6559 bool *HasWinCFI,
bool EmitCFAOffset,
6562 unsigned MaxEncoding, ShiftSize;
6564 case AArch64::ADDXri:
6565 case AArch64::ADDSXri:
6566 case AArch64::SUBXri:
6567 case AArch64::SUBSXri:
6568 MaxEncoding = 0xfff;
6571 case AArch64::ADDVL_XXI:
6572 case AArch64::ADDPL_XXI:
6573 case AArch64::ADDSVL_XXI:
6574 case AArch64::ADDSPL_XXI:
6589 if (
Opc == AArch64::ADDVL_XXI ||
Opc == AArch64::ADDSVL_XXI)
6591 else if (
Opc == AArch64::ADDPL_XXI ||
Opc == AArch64::ADDSPL_XXI)
6605 const unsigned MaxEncodableValue = MaxEncoding << ShiftSize;
6607 if (TmpReg == AArch64::XZR)
6608 TmpReg =
MBB.getParent()->getRegInfo().createVirtualRegister(
6609 &AArch64::GPR64RegClass);
6611 uint64_t ThisVal = std::min<uint64_t>(
Offset, MaxEncodableValue);
6612 unsigned LocalShiftSize = 0;
6613 if (ThisVal > MaxEncoding) {
6614 ThisVal = ThisVal >> ShiftSize;
6615 LocalShiftSize = ShiftSize;
6617 assert((ThisVal >> ShiftSize) <= MaxEncoding &&
6618 "Encoding cannot handle value that big");
6620 Offset -= ThisVal << LocalShiftSize;
6625 .
addImm(Sign * (
int)ThisVal);
6635 if (Sign == -1 ||
Opc == AArch64::SUBXri ||
Opc == AArch64::SUBSXri)
6636 CFAOffset += Change;
6638 CFAOffset -= Change;
6639 if (EmitCFAOffset && DestReg == TmpReg) {
6652 int Imm = (int)(ThisVal << LocalShiftSize);
6653 if (VScale != 1 && DestReg == AArch64::SP) {
6659 }
else if ((DestReg == AArch64::FP && SrcReg == AArch64::SP) ||
6660 (SrcReg == AArch64::FP && DestReg == AArch64::SP)) {
6661 assert(VScale == 1 &&
"Expected non-scalable operation");
6670 assert(
Offset == 0 &&
"Expected remaining offset to be zero to "
6671 "emit a single SEH directive");
6672 }
else if (DestReg == AArch64::SP) {
6673 assert(VScale == 1 &&
"Expected non-scalable operation");
6676 assert(SrcReg == AArch64::SP &&
"Unexpected SrcReg for SEH_StackAlloc");
6689 unsigned DestReg,
unsigned SrcReg,
6692 bool NeedsWinCFI,
bool *HasWinCFI,
6694 unsigned FrameReg) {
6701 bool UseSVL =
F.hasFnAttribute(
"aarch64_pstate_sm_body");
6703 int64_t Bytes, NumPredicateVectors, NumDataVectors;
6704 AArch64InstrInfo::decomposeStackOffsetForFrameOffsets(
6705 Offset, Bytes, NumPredicateVectors, NumDataVectors);
6708 bool NeedsFinalDefNZCV = SetNZCV && (NumPredicateVectors || NumDataVectors);
6709 if (NeedsFinalDefNZCV)
6713 if (Bytes || (!
Offset && SrcReg != DestReg)) {
6714 assert((DestReg != AArch64::SP || Bytes % 8 == 0) &&
6715 "SP increment/decrement not 8-byte aligned");
6716 unsigned Opc = SetNZCV ? AArch64::ADDSXri : AArch64::ADDXri;
6719 Opc = SetNZCV ? AArch64::SUBSXri : AArch64::SUBXri;
6722 NeedsWinCFI, HasWinCFI, EmitCFAOffset, CFAOffset,
6724 CFAOffset += (
Opc == AArch64::ADDXri ||
Opc == AArch64::ADDSXri)
6731 assert(!(NeedsWinCFI && NumPredicateVectors) &&
6732 "WinCFI can't allocate fractions of an SVE data vector");
6734 if (NumDataVectors) {
6736 UseSVL ? AArch64::ADDSVL_XXI : AArch64::ADDVL_XXI,
TII,
6737 Flag, NeedsWinCFI, HasWinCFI, EmitCFAOffset, CFAOffset,
6743 if (NumPredicateVectors) {
6744 assert(DestReg != AArch64::SP &&
"Unaligned access to SP");
6746 UseSVL ? AArch64::ADDSPL_XXI : AArch64::ADDPL_XXI,
TII,
6747 Flag, NeedsWinCFI, HasWinCFI, EmitCFAOffset, CFAOffset,
6751 if (NeedsFinalDefNZCV)
6772 if (
MI.isFullCopy()) {
6775 if (SrcReg == AArch64::SP && DstReg.
isVirtual()) {
6779 if (DstReg == AArch64::SP && SrcReg.
isVirtual()) {
6784 if (SrcReg == AArch64::NZCV || DstReg == AArch64::NZCV)
6812 if (
MI.isCopy() &&
Ops.size() == 1 &&
6814 (
Ops[0] == 0 ||
Ops[0] == 1)) {
6815 bool IsSpill =
Ops[0] == 0;
6816 bool IsFill = !IsSpill;
6828 :
TRI.getMinimalPhysRegClass(Reg);
6834 "Mismatched register size in non subreg COPY");
6841 return &*--InsertPt;
6853 if (IsSpill && DstMO.
isUndef() && SrcReg == AArch64::WZR &&
6856 "Unexpected subreg on physical register");
6858 FrameIndex, &AArch64::GPR64RegClass,
Register());
6859 return &*--InsertPt;
6876 case AArch64::sub_32:
6877 if (AArch64::GPR64RegClass.hasSubClassEq(
getRegClass(DstReg)))
6878 FillRC = &AArch64::GPR32RegClass;
6881 FillRC = &AArch64::FPR32RegClass;
6884 FillRC = &AArch64::FPR64RegClass;
6890 TRI.getRegSizeInBits(*FillRC) &&
6891 "Mismatched regclass size on folded subreg COPY");
6910 bool *OutUseUnscaledOp,
6911 unsigned *OutUnscaledOp,
6912 int64_t *EmittableOffset) {
6914 if (EmittableOffset)
6915 *EmittableOffset = 0;
6916 if (OutUseUnscaledOp)
6917 *OutUseUnscaledOp =
false;
6923 switch (
MI.getOpcode()) {
6926 case AArch64::LD1Rv1d:
6927 case AArch64::LD1Rv2s:
6928 case AArch64::LD1Rv2d:
6929 case AArch64::LD1Rv4h:
6930 case AArch64::LD1Rv4s:
6931 case AArch64::LD1Rv8b:
6932 case AArch64::LD1Rv8h:
6933 case AArch64::LD1Rv16b:
6934 case AArch64::LD1Twov2d:
6935 case AArch64::LD1Threev2d:
6936 case AArch64::LD1Fourv2d:
6937 case AArch64::LD1Twov1d:
6938 case AArch64::LD1Threev1d:
6939 case AArch64::LD1Fourv1d:
6940 case AArch64::ST1Twov2d:
6941 case AArch64::ST1Threev2d:
6942 case AArch64::ST1Fourv2d:
6943 case AArch64::ST1Twov1d:
6944 case AArch64::ST1Threev1d:
6945 case AArch64::ST1Fourv1d:
6946 case AArch64::ST1i8:
6947 case AArch64::ST1i16:
6948 case AArch64::ST1i32:
6949 case AArch64::ST1i64:
6951 case AArch64::IRGstack:
6952 case AArch64::STGloop:
6953 case AArch64::STZGloop:
6958 TypeSize ScaleValue(0U,
false), Width(0U,
false);
6959 int64_t MinOff, MaxOff;
6965 bool IsMulVL = ScaleValue.isScalable();
6966 unsigned Scale = ScaleValue.getKnownMinValue();
6976 std::optional<unsigned> UnscaledOp =
6978 bool useUnscaledOp = UnscaledOp && (
Offset % Scale ||
Offset < 0);
6979 if (useUnscaledOp &&
6984 Scale = ScaleValue.getKnownMinValue();
6985 assert(IsMulVL == ScaleValue.isScalable() &&
6986 "Unscaled opcode has different value for scalable");
6988 int64_t Remainder =
Offset % Scale;
6989 assert(!(Remainder && useUnscaledOp) &&
6990 "Cannot have remainder when using unscaled op");
6992 assert(MinOff < MaxOff &&
"Unexpected Min/Max offsets");
6993 int64_t NewOffset =
Offset / Scale;
6994 if (MinOff <= NewOffset && NewOffset <= MaxOff)
6997 NewOffset = NewOffset < 0 ? MinOff : MaxOff;
7001 if (EmittableOffset)
7002 *EmittableOffset = NewOffset;
7003 if (OutUseUnscaledOp)
7004 *OutUseUnscaledOp = useUnscaledOp;
7005 if (OutUnscaledOp && UnscaledOp)
7006 *OutUnscaledOp = *UnscaledOp;
7019 unsigned Opcode =
MI.getOpcode();
7020 unsigned ImmIdx = FrameRegIdx + 1;
7022 if (Opcode == AArch64::ADDSXri || Opcode == AArch64::ADDXri) {
7027 MI.eraseFromParent();
7033 unsigned UnscaledOp;
7036 &UnscaledOp, &NewOffset);
7040 MI.getOperand(FrameRegIdx).ChangeToRegister(FrameReg,
false);
7042 MI.setDesc(
TII->get(UnscaledOp));
7044 MI.getOperand(ImmIdx).ChangeToImmediate(NewOffset);
7060bool AArch64InstrInfo::useMachineCombiner()
const {
return true; }
7065 case AArch64::ADDSWrr:
7066 case AArch64::ADDSWri:
7067 case AArch64::ADDSXrr:
7068 case AArch64::ADDSXri:
7069 case AArch64::SUBSWrr:
7070 case AArch64::SUBSXrr:
7072 case AArch64::SUBSWri:
7073 case AArch64::SUBSXri:
7084 case AArch64::ADDWrr:
7085 case AArch64::ADDWri:
7086 case AArch64::SUBWrr:
7087 case AArch64::ADDSWrr:
7088 case AArch64::ADDSWri:
7089 case AArch64::SUBSWrr:
7091 case AArch64::SUBWri:
7092 case AArch64::SUBSWri:
7103 case AArch64::ADDXrr:
7104 case AArch64::ADDXri:
7105 case AArch64::SUBXrr:
7106 case AArch64::ADDSXrr:
7107 case AArch64::ADDSXri:
7108 case AArch64::SUBSXrr:
7110 case AArch64::SUBXri:
7111 case AArch64::SUBSXri:
7112 case AArch64::ADDv8i8:
7113 case AArch64::ADDv16i8:
7114 case AArch64::ADDv4i16:
7115 case AArch64::ADDv8i16:
7116 case AArch64::ADDv2i32:
7117 case AArch64::ADDv4i32:
7118 case AArch64::SUBv8i8:
7119 case AArch64::SUBv16i8:
7120 case AArch64::SUBv4i16:
7121 case AArch64::SUBv8i16:
7122 case AArch64::SUBv2i32:
7123 case AArch64::SUBv4i32:
7136 case AArch64::FADDHrr:
7137 case AArch64::FADDSrr:
7138 case AArch64::FADDDrr:
7139 case AArch64::FADDv4f16:
7140 case AArch64::FADDv8f16:
7141 case AArch64::FADDv2f32:
7142 case AArch64::FADDv2f64:
7143 case AArch64::FADDv4f32:
7144 case AArch64::FSUBHrr:
7145 case AArch64::FSUBSrr:
7146 case AArch64::FSUBDrr:
7147 case AArch64::FSUBv4f16:
7148 case AArch64::FSUBv8f16:
7149 case AArch64::FSUBv2f32:
7150 case AArch64::FSUBv2f64:
7151 case AArch64::FSUBv4f32:
7170 unsigned CombineOpc,
unsigned ZeroReg = 0,
7171 bool CheckZeroReg =
false) {
7178 if (!
MI ||
MI->getParent() != &
MBB ||
MI->getOpcode() != CombineOpc)
7181 if (!
MRI.hasOneNonDBGUse(
MI->getOperand(0).getReg()))
7185 assert(
MI->getNumOperands() >= 4 &&
MI->getOperand(0).isReg() &&
7186 MI->getOperand(1).isReg() &&
MI->getOperand(2).isReg() &&
7187 MI->getOperand(3).isReg() &&
"MAdd/MSub must have a least 4 regs");
7189 if (
MI->getOperand(3).getReg() != ZeroReg)
7194 MI->findRegisterDefOperandIdx(AArch64::NZCV,
nullptr,
true) == -1)
7203 unsigned MulOpc,
unsigned ZeroReg) {
7218bool AArch64InstrInfo::isAssociativeAndCommutative(
const MachineInstr &Inst,
7219 bool Invert)
const {
7225 case AArch64::FADDHrr:
7226 case AArch64::FADDSrr:
7227 case AArch64::FADDDrr:
7228 case AArch64::FMULHrr:
7229 case AArch64::FMULSrr:
7230 case AArch64::FMULDrr:
7231 case AArch64::FMULX16:
7232 case AArch64::FMULX32:
7233 case AArch64::FMULX64:
7235 case AArch64::FADDv4f16:
7236 case AArch64::FADDv8f16:
7237 case AArch64::FADDv2f32:
7238 case AArch64::FADDv4f32:
7239 case AArch64::FADDv2f64:
7240 case AArch64::FMULv4f16:
7241 case AArch64::FMULv8f16:
7242 case AArch64::FMULv2f32:
7243 case AArch64::FMULv4f32:
7244 case AArch64::FMULv2f64:
7245 case AArch64::FMULXv4f16:
7246 case AArch64::FMULXv8f16:
7247 case AArch64::FMULXv2f32:
7248 case AArch64::FMULXv4f32:
7249 case AArch64::FMULXv2f64:
7253 case AArch64::FADD_ZZZ_H:
7254 case AArch64::FADD_ZZZ_S:
7255 case AArch64::FADD_ZZZ_D:
7256 case AArch64::FMUL_ZZZ_H:
7257 case AArch64::FMUL_ZZZ_S:
7258 case AArch64::FMUL_ZZZ_D:
7269 case AArch64::ADDWrr:
7270 case AArch64::ADDXrr:
7271 case AArch64::ANDWrr:
7272 case AArch64::ANDXrr:
7273 case AArch64::ORRWrr:
7274 case AArch64::ORRXrr:
7275 case AArch64::EORWrr:
7276 case AArch64::EORXrr:
7277 case AArch64::EONWrr:
7278 case AArch64::EONXrr:
7282 case AArch64::ADDv8i8:
7283 case AArch64::ADDv16i8:
7284 case AArch64::ADDv4i16:
7285 case AArch64::ADDv8i16:
7286 case AArch64::ADDv2i32:
7287 case AArch64::ADDv4i32:
7288 case AArch64::ADDv1i64:
7289 case AArch64::ADDv2i64:
7290 case AArch64::MULv8i8:
7291 case AArch64::MULv16i8:
7292 case AArch64::MULv4i16:
7293 case AArch64::MULv8i16:
7294 case AArch64::MULv2i32:
7295 case AArch64::MULv4i32:
7296 case AArch64::ANDv8i8:
7297 case AArch64::ANDv16i8:
7298 case AArch64::ORRv8i8:
7299 case AArch64::ORRv16i8:
7300 case AArch64::EORv8i8:
7301 case AArch64::EORv16i8:
7303 case AArch64::ADD_ZZZ_B:
7304 case AArch64::ADD_ZZZ_H:
7305 case AArch64::ADD_ZZZ_S:
7306 case AArch64::ADD_ZZZ_D:
7307 case AArch64::MUL_ZZZ_B:
7308 case AArch64::MUL_ZZZ_H:
7309 case AArch64::MUL_ZZZ_S:
7310 case AArch64::MUL_ZZZ_D:
7311 case AArch64::AND_ZZZ:
7312 case AArch64::ORR_ZZZ:
7313 case AArch64::EOR_ZZZ:
7344 auto setFound = [&](
int Opcode,
int Operand,
unsigned ZeroReg,
7352 auto setVFound = [&](
int Opcode,
int Operand,
unsigned Pattern) {
7364 case AArch64::ADDWrr:
7366 "ADDWrr does not have register operands");
7367 setFound(AArch64::MADDWrrr, 1, AArch64::WZR, MCP::MULADDW_OP1);
7368 setFound(AArch64::MADDWrrr, 2, AArch64::WZR, MCP::MULADDW_OP2);
7370 case AArch64::ADDXrr:
7371 setFound(AArch64::MADDXrrr, 1, AArch64::XZR, MCP::MULADDX_OP1);
7372 setFound(AArch64::MADDXrrr, 2, AArch64::XZR, MCP::MULADDX_OP2);
7374 case AArch64::SUBWrr:
7375 setFound(AArch64::MADDWrrr, 2, AArch64::WZR, MCP::MULSUBW_OP2);
7376 setFound(AArch64::MADDWrrr, 1, AArch64::WZR, MCP::MULSUBW_OP1);
7378 case AArch64::SUBXrr:
7379 setFound(AArch64::MADDXrrr, 2, AArch64::XZR, MCP::MULSUBX_OP2);
7380 setFound(AArch64::MADDXrrr, 1, AArch64::XZR, MCP::MULSUBX_OP1);
7382 case AArch64::ADDWri:
7383 setFound(AArch64::MADDWrrr, 1, AArch64::WZR, MCP::MULADDWI_OP1);
7385 case AArch64::ADDXri:
7386 setFound(AArch64::MADDXrrr, 1, AArch64::XZR, MCP::MULADDXI_OP1);
7388 case AArch64::SUBWri:
7389 setFound(AArch64::MADDWrrr, 1, AArch64::WZR, MCP::MULSUBWI_OP1);
7391 case AArch64::SUBXri:
7392 setFound(AArch64::MADDXrrr, 1, AArch64::XZR, MCP::MULSUBXI_OP1);
7394 case AArch64::ADDv8i8:
7395 setVFound(AArch64::MULv8i8, 1, MCP::MULADDv8i8_OP1);
7396 setVFound(AArch64::MULv8i8, 2, MCP::MULADDv8i8_OP2);
7398 case AArch64::ADDv16i8:
7399 setVFound(AArch64::MULv16i8, 1, MCP::MULADDv16i8_OP1);
7400 setVFound(AArch64::MULv16i8, 2, MCP::MULADDv16i8_OP2);
7402 case AArch64::ADDv4i16:
7403 setVFound(AArch64::MULv4i16, 1, MCP::MULADDv4i16_OP1);
7404 setVFound(AArch64::MULv4i16, 2, MCP::MULADDv4i16_OP2);
7405 setVFound(AArch64::MULv4i16_indexed, 1, MCP::MULADDv4i16_indexed_OP1);
7406 setVFound(AArch64::MULv4i16_indexed, 2, MCP::MULADDv4i16_indexed_OP2);
7408 case AArch64::ADDv8i16:
7409 setVFound(AArch64::MULv8i16, 1, MCP::MULADDv8i16_OP1);
7410 setVFound(AArch64::MULv8i16, 2, MCP::MULADDv8i16_OP2);
7411 setVFound(AArch64::MULv8i16_indexed, 1, MCP::MULADDv8i16_indexed_OP1);
7412 setVFound(AArch64::MULv8i16_indexed, 2, MCP::MULADDv8i16_indexed_OP2);
7414 case AArch64::ADDv2i32:
7415 setVFound(AArch64::MULv2i32, 1, MCP::MULADDv2i32_OP1);
7416 setVFound(AArch64::MULv2i32, 2, MCP::MULADDv2i32_OP2);
7417 setVFound(AArch64::MULv2i32_indexed, 1, MCP::MULADDv2i32_indexed_OP1);
7418 setVFound(AArch64::MULv2i32_indexed, 2, MCP::MULADDv2i32_indexed_OP2);
7420 case AArch64::ADDv4i32:
7421 setVFound(AArch64::MULv4i32, 1, MCP::MULADDv4i32_OP1);
7422 setVFound(AArch64::MULv4i32, 2, MCP::MULADDv4i32_OP2);
7423 setVFound(AArch64::MULv4i32_indexed, 1, MCP::MULADDv4i32_indexed_OP1);
7424 setVFound(AArch64::MULv4i32_indexed, 2, MCP::MULADDv4i32_indexed_OP2);
7426 case AArch64::SUBv8i8:
7427 setVFound(AArch64::MULv8i8, 1, MCP::MULSUBv8i8_OP1);
7428 setVFound(AArch64::MULv8i8, 2, MCP::MULSUBv8i8_OP2);
7430 case AArch64::SUBv16i8:
7431 setVFound(AArch64::MULv16i8, 1, MCP::MULSUBv16i8_OP1);
7432 setVFound(AArch64::MULv16i8, 2, MCP::MULSUBv16i8_OP2);
7434 case AArch64::SUBv4i16:
7435 setVFound(AArch64::MULv4i16, 1, MCP::MULSUBv4i16_OP1);
7436 setVFound(AArch64::MULv4i16, 2, MCP::MULSUBv4i16_OP2);
7437 setVFound(AArch64::MULv4i16_indexed, 1, MCP::MULSUBv4i16_indexed_OP1);
7438 setVFound(AArch64::MULv4i16_indexed, 2, MCP::MULSUBv4i16_indexed_OP2);
7440 case AArch64::SUBv8i16:
7441 setVFound(AArch64::MULv8i16, 1, MCP::MULSUBv8i16_OP1);
7442 setVFound(AArch64::MULv8i16, 2, MCP::MULSUBv8i16_OP2);
7443 setVFound(AArch64::MULv8i16_indexed, 1, MCP::MULSUBv8i16_indexed_OP1);
7444 setVFound(AArch64::MULv8i16_indexed, 2, MCP::MULSUBv8i16_indexed_OP2);
7446 case AArch64::SUBv2i32:
7447 setVFound(AArch64::MULv2i32, 1, MCP::MULSUBv2i32_OP1);
7448 setVFound(AArch64::MULv2i32, 2, MCP::MULSUBv2i32_OP2);
7449 setVFound(AArch64::MULv2i32_indexed, 1, MCP::MULSUBv2i32_indexed_OP1);
7450 setVFound(AArch64::MULv2i32_indexed, 2, MCP::MULSUBv2i32_indexed_OP2);
7452 case AArch64::SUBv4i32:
7453 setVFound(AArch64::MULv4i32, 1, MCP::MULSUBv4i32_OP1);
7454 setVFound(AArch64::MULv4i32, 2, MCP::MULSUBv4i32_OP2);
7455 setVFound(AArch64::MULv4i32_indexed, 1, MCP::MULSUBv4i32_indexed_OP1);
7456 setVFound(AArch64::MULv4i32_indexed, 2, MCP::MULSUBv4i32_indexed_OP2);
7462bool AArch64InstrInfo::isAccumulationOpcode(
unsigned Opcode)
const {
7466 case AArch64::UABALB_ZZZ_D:
7467 case AArch64::UABALB_ZZZ_H:
7468 case AArch64::UABALB_ZZZ_S:
7469 case AArch64::UABALT_ZZZ_D:
7470 case AArch64::UABALT_ZZZ_H:
7471 case AArch64::UABALT_ZZZ_S:
7472 case AArch64::SABALB_ZZZ_D:
7473 case AArch64::SABALB_ZZZ_S:
7474 case AArch64::SABALB_ZZZ_H:
7475 case AArch64::SABALT_ZZZ_D:
7476 case AArch64::SABALT_ZZZ_S:
7477 case AArch64::SABALT_ZZZ_H:
7478 case AArch64::UABALv16i8_v8i16:
7479 case AArch64::UABALv2i32_v2i64:
7480 case AArch64::UABALv4i16_v4i32:
7481 case AArch64::UABALv4i32_v2i64:
7482 case AArch64::UABALv8i16_v4i32:
7483 case AArch64::UABALv8i8_v8i16:
7484 case AArch64::UABAv16i8:
7485 case AArch64::UABAv2i32:
7486 case AArch64::UABAv4i16:
7487 case AArch64::UABAv4i32:
7488 case AArch64::UABAv8i16:
7489 case AArch64::UABAv8i8:
7490 case AArch64::SABALv16i8_v8i16:
7491 case AArch64::SABALv2i32_v2i64:
7492 case AArch64::SABALv4i16_v4i32:
7493 case AArch64::SABALv4i32_v2i64:
7494 case AArch64::SABALv8i16_v4i32:
7495 case AArch64::SABALv8i8_v8i16:
7496 case AArch64::SABAv16i8:
7497 case AArch64::SABAv2i32:
7498 case AArch64::SABAv4i16:
7499 case AArch64::SABAv4i32:
7500 case AArch64::SABAv8i16:
7501 case AArch64::SABAv8i8:
7508unsigned AArch64InstrInfo::getAccumulationStartOpcode(
7509 unsigned AccumulationOpcode)
const {
7510 switch (AccumulationOpcode) {
7513 case AArch64::UABALB_ZZZ_D:
7514 return AArch64::UABDLB_ZZZ_D;
7515 case AArch64::UABALB_ZZZ_H:
7516 return AArch64::UABDLB_ZZZ_H;
7517 case AArch64::UABALB_ZZZ_S:
7518 return AArch64::UABDLB_ZZZ_S;
7519 case AArch64::UABALT_ZZZ_D:
7520 return AArch64::UABDLT_ZZZ_D;
7521 case AArch64::UABALT_ZZZ_H:
7522 return AArch64::UABDLT_ZZZ_H;
7523 case AArch64::UABALT_ZZZ_S:
7524 return AArch64::UABDLT_ZZZ_S;
7525 case AArch64::UABALv16i8_v8i16:
7526 return AArch64::UABDLv16i8_v8i16;
7527 case AArch64::UABALv2i32_v2i64:
7528 return AArch64::UABDLv2i32_v2i64;
7529 case AArch64::UABALv4i16_v4i32:
7530 return AArch64::UABDLv4i16_v4i32;
7531 case AArch64::UABALv4i32_v2i64:
7532 return AArch64::UABDLv4i32_v2i64;
7533 case AArch64::UABALv8i16_v4i32:
7534 return AArch64::UABDLv8i16_v4i32;
7535 case AArch64::UABALv8i8_v8i16:
7536 return AArch64::UABDLv8i8_v8i16;
7537 case AArch64::UABAv16i8:
7538 return AArch64::UABDv16i8;
7539 case AArch64::UABAv2i32:
7540 return AArch64::UABDv2i32;
7541 case AArch64::UABAv4i16:
7542 return AArch64::UABDv4i16;
7543 case AArch64::UABAv4i32:
7544 return AArch64::UABDv4i32;
7545 case AArch64::UABAv8i16:
7546 return AArch64::UABDv8i16;
7547 case AArch64::UABAv8i8:
7548 return AArch64::UABDv8i8;
7549 case AArch64::SABALB_ZZZ_D:
7550 return AArch64::SABDLB_ZZZ_D;
7551 case AArch64::SABALB_ZZZ_S:
7552 return AArch64::SABDLB_ZZZ_S;
7553 case AArch64::SABALB_ZZZ_H:
7554 return AArch64::SABDLB_ZZZ_H;
7555 case AArch64::SABALT_ZZZ_D:
7556 return AArch64::SABDLT_ZZZ_D;
7557 case AArch64::SABALT_ZZZ_S:
7558 return AArch64::SABDLT_ZZZ_S;
7559 case AArch64::SABALT_ZZZ_H:
7560 return AArch64::SABDLT_ZZZ_H;
7561 case AArch64::SABALv16i8_v8i16:
7562 return AArch64::SABDLv16i8_v8i16;
7563 case AArch64::SABALv2i32_v2i64:
7564 return AArch64::SABDLv2i32_v2i64;
7565 case AArch64::SABALv4i16_v4i32:
7566 return AArch64::SABDLv4i16_v4i32;
7567 case AArch64::SABALv4i32_v2i64:
7568 return AArch64::SABDLv4i32_v2i64;
7569 case AArch64::SABALv8i16_v4i32:
7570 return AArch64::SABDLv8i16_v4i32;
7571 case AArch64::SABALv8i8_v8i16:
7572 return AArch64::SABDLv8i8_v8i16;
7573 case AArch64::SABAv16i8:
7574 return AArch64::SABDv16i8;
7575 case AArch64::SABAv2i32:
7576 return AArch64::SABAv2i32;
7577 case AArch64::SABAv4i16:
7578 return AArch64::SABDv4i16;
7579 case AArch64::SABAv4i32:
7580 return AArch64::SABDv4i32;
7581 case AArch64::SABAv8i16:
7582 return AArch64::SABDv8i16;
7583 case AArch64::SABAv8i8:
7584 return AArch64::SABDv8i8;
7600 auto Match = [&](
int Opcode,
int Operand,
unsigned Pattern) ->
bool {
7612 assert(
false &&
"Unsupported FP instruction in combiner\n");
7614 case AArch64::FADDHrr:
7616 "FADDHrr does not have register operands");
7618 Found = Match(AArch64::FMULHrr, 1, MCP::FMULADDH_OP1);
7619 Found |= Match(AArch64::FMULHrr, 2, MCP::FMULADDH_OP2);
7621 case AArch64::FADDSrr:
7623 "FADDSrr does not have register operands");
7625 Found |= Match(AArch64::FMULSrr, 1, MCP::FMULADDS_OP1) ||
7626 Match(AArch64::FMULv1i32_indexed, 1, MCP::FMLAv1i32_indexed_OP1);
7628 Found |= Match(AArch64::FMULSrr, 2, MCP::FMULADDS_OP2) ||
7629 Match(AArch64::FMULv1i32_indexed, 2, MCP::FMLAv1i32_indexed_OP2);
7631 case AArch64::FADDDrr:
7632 Found |= Match(AArch64::FMULDrr, 1, MCP::FMULADDD_OP1) ||
7633 Match(AArch64::FMULv1i64_indexed, 1, MCP::FMLAv1i64_indexed_OP1);
7635 Found |= Match(AArch64::FMULDrr, 2, MCP::FMULADDD_OP2) ||
7636 Match(AArch64::FMULv1i64_indexed, 2, MCP::FMLAv1i64_indexed_OP2);
7638 case AArch64::FADDv4f16:
7639 Found |= Match(AArch64::FMULv4i16_indexed, 1, MCP::FMLAv4i16_indexed_OP1) ||
7640 Match(AArch64::FMULv4f16, 1, MCP::FMLAv4f16_OP1);
7642 Found |= Match(AArch64::FMULv4i16_indexed, 2, MCP::FMLAv4i16_indexed_OP2) ||
7643 Match(AArch64::FMULv4f16, 2, MCP::FMLAv4f16_OP2);
7645 case AArch64::FADDv8f16:
7646 Found |= Match(AArch64::FMULv8i16_indexed, 1, MCP::FMLAv8i16_indexed_OP1) ||
7647 Match(AArch64::FMULv8f16, 1, MCP::FMLAv8f16_OP1);
7649 Found |= Match(AArch64::FMULv8i16_indexed, 2, MCP::FMLAv8i16_indexed_OP2) ||
7650 Match(AArch64::FMULv8f16, 2, MCP::FMLAv8f16_OP2);
7652 case AArch64::FADDv2f32:
7653 Found |= Match(AArch64::FMULv2i32_indexed, 1, MCP::FMLAv2i32_indexed_OP1) ||
7654 Match(AArch64::FMULv2f32, 1, MCP::FMLAv2f32_OP1);
7656 Found |= Match(AArch64::FMULv2i32_indexed, 2, MCP::FMLAv2i32_indexed_OP2) ||
7657 Match(AArch64::FMULv2f32, 2, MCP::FMLAv2f32_OP2);
7659 case AArch64::FADDv2f64:
7660 Found |= Match(AArch64::FMULv2i64_indexed, 1, MCP::FMLAv2i64_indexed_OP1) ||
7661 Match(AArch64::FMULv2f64, 1, MCP::FMLAv2f64_OP1);
7663 Found |= Match(AArch64::FMULv2i64_indexed, 2, MCP::FMLAv2i64_indexed_OP2) ||
7664 Match(AArch64::FMULv2f64, 2, MCP::FMLAv2f64_OP2);
7666 case AArch64::FADDv4f32:
7667 Found |= Match(AArch64::FMULv4i32_indexed, 1, MCP::FMLAv4i32_indexed_OP1) ||
7668 Match(AArch64::FMULv4f32, 1, MCP::FMLAv4f32_OP1);
7670 Found |= Match(AArch64::FMULv4i32_indexed, 2, MCP::FMLAv4i32_indexed_OP2) ||
7671 Match(AArch64::FMULv4f32, 2, MCP::FMLAv4f32_OP2);
7673 case AArch64::FSUBHrr:
7674 Found = Match(AArch64::FMULHrr, 1, MCP::FMULSUBH_OP1);
7675 Found |= Match(AArch64::FMULHrr, 2, MCP::FMULSUBH_OP2);
7676 Found |= Match(AArch64::FNMULHrr, 1, MCP::FNMULSUBH_OP1);
7678 case AArch64::FSUBSrr:
7679 Found = Match(AArch64::FMULSrr, 1, MCP::FMULSUBS_OP1);
7681 Found |= Match(AArch64::FMULSrr, 2, MCP::FMULSUBS_OP2) ||
7682 Match(AArch64::FMULv1i32_indexed, 2, MCP::FMLSv1i32_indexed_OP2);
7684 Found |= Match(AArch64::FNMULSrr, 1, MCP::FNMULSUBS_OP1);
7686 case AArch64::FSUBDrr:
7687 Found = Match(AArch64::FMULDrr, 1, MCP::FMULSUBD_OP1);
7689 Found |= Match(AArch64::FMULDrr, 2, MCP::FMULSUBD_OP2) ||
7690 Match(AArch64::FMULv1i64_indexed, 2, MCP::FMLSv1i64_indexed_OP2);
7692 Found |= Match(AArch64::FNMULDrr, 1, MCP::FNMULSUBD_OP1);
7694 case AArch64::FSUBv4f16:
7695 Found |= Match(AArch64::FMULv4i16_indexed, 2, MCP::FMLSv4i16_indexed_OP2) ||
7696 Match(AArch64::FMULv4f16, 2, MCP::FMLSv4f16_OP2);
7698 Found |= Match(AArch64::FMULv4i16_indexed, 1, MCP::FMLSv4i16_indexed_OP1) ||
7699 Match(AArch64::FMULv4f16, 1, MCP::FMLSv4f16_OP1);
7701 case AArch64::FSUBv8f16:
7702 Found |= Match(AArch64::FMULv8i16_indexed, 2, MCP::FMLSv8i16_indexed_OP2) ||
7703 Match(AArch64::FMULv8f16, 2, MCP::FMLSv8f16_OP2);
7705 Found |= Match(AArch64::FMULv8i16_indexed, 1, MCP::FMLSv8i16_indexed_OP1) ||
7706 Match(AArch64::FMULv8f16, 1, MCP::FMLSv8f16_OP1);
7708 case AArch64::FSUBv2f32:
7709 Found |= Match(AArch64::FMULv2i32_indexed, 2, MCP::FMLSv2i32_indexed_OP2) ||
7710 Match(AArch64::FMULv2f32, 2, MCP::FMLSv2f32_OP2);
7712 Found |= Match(AArch64::FMULv2i32_indexed, 1, MCP::FMLSv2i32_indexed_OP1) ||
7713 Match(AArch64::FMULv2f32, 1, MCP::FMLSv2f32_OP1);
7715 case AArch64::FSUBv2f64:
7716 Found |= Match(AArch64::FMULv2i64_indexed, 2, MCP::FMLSv2i64_indexed_OP2) ||
7717 Match(AArch64::FMULv2f64, 2, MCP::FMLSv2f64_OP2);
7719 Found |= Match(AArch64::FMULv2i64_indexed, 1, MCP::FMLSv2i64_indexed_OP1) ||
7720 Match(AArch64::FMULv2f64, 1, MCP::FMLSv2f64_OP1);
7722 case AArch64::FSUBv4f32:
7723 Found |= Match(AArch64::FMULv4i32_indexed, 2, MCP::FMLSv4i32_indexed_OP2) ||
7724 Match(AArch64::FMULv4f32, 2, MCP::FMLSv4f32_OP2);
7726 Found |= Match(AArch64::FMULv4i32_indexed, 1, MCP::FMLSv4i32_indexed_OP1) ||
7727 Match(AArch64::FMULv4f32, 1, MCP::FMLSv4f32_OP1);
7738 auto Match = [&](
unsigned Opcode,
int Operand,
unsigned Pattern) ->
bool {
7745 if (
MI &&
MI->getOpcode() == TargetOpcode::COPY &&
7746 MI->getOperand(1).getReg().isVirtual())
7747 MI =
MRI.getUniqueVRegDef(
MI->getOperand(1).getReg());
7748 if (
MI &&
MI->getOpcode() == Opcode) {
7760 case AArch64::FMULv2f32:
7761 Found = Match(AArch64::DUPv2i32lane, 1, MCP::FMULv2i32_indexed_OP1);
7762 Found |= Match(AArch64::DUPv2i32lane, 2, MCP::FMULv2i32_indexed_OP2);
7764 case AArch64::FMULv2f64:
7765 Found = Match(AArch64::DUPv2i64lane, 1, MCP::FMULv2i64_indexed_OP1);
7766 Found |= Match(AArch64::DUPv2i64lane, 2, MCP::FMULv2i64_indexed_OP2);
7768 case AArch64::FMULv4f16:
7769 Found = Match(AArch64::DUPv4i16lane, 1, MCP::FMULv4i16_indexed_OP1);
7770 Found |= Match(AArch64::DUPv4i16lane, 2, MCP::FMULv4i16_indexed_OP2);
7772 case AArch64::FMULv4f32:
7773 Found = Match(AArch64::DUPv4i32lane, 1, MCP::FMULv4i32_indexed_OP1);
7774 Found |= Match(AArch64::DUPv4i32lane, 2, MCP::FMULv4i32_indexed_OP2);
7776 case AArch64::FMULv8f16:
7777 Found = Match(AArch64::DUPv8i16lane, 1, MCP::FMULv8i16_indexed_OP1);
7778 Found |= Match(AArch64::DUPv8i16lane, 2, MCP::FMULv8i16_indexed_OP2);
7791 auto Match = [&](
unsigned Opcode,
unsigned Pattern) ->
bool {
7794 if (
MI !=
nullptr && (
MI->getOpcode() == Opcode) &&
7795 MRI.hasOneNonDBGUse(
MI->getOperand(0).getReg()) &&
7809 case AArch64::FNEGDr:
7811 case AArch64::FNEGSr:
7943 case AArch64::SUBWrr:
7944 case AArch64::SUBSWrr:
7945 case AArch64::SUBXrr:
7946 case AArch64::SUBSXrr:
7991 unsigned LoadLaneOpCode,
unsigned NumLanes) {
8014 while (!RemainingLanes.
empty() && CurrInstr &&
8015 CurrInstr->getOpcode() == LoadLaneOpCode &&
8016 MRI.hasOneNonDBGUse(CurrInstr->getOperand(0).getReg()) &&
8017 CurrInstr->getNumOperands() == 4) {
8018 RemainingLanes.
erase(CurrInstr->getOperand(2).getImm());
8020 CurrInstr =
MRI.getUniqueVRegDef(CurrInstr->getOperand(1).getReg());
8024 if (!RemainingLanes.
empty())
8028 if (CurrInstr->getOpcode() != TargetOpcode::SUBREG_TO_REG)
8032 auto Lane0LoadReg = CurrInstr->getOperand(2).getReg();
8033 unsigned SingleLaneSizeInBits = 128 / NumLanes;
8034 if (
TRI->getRegSizeInBits(Lane0LoadReg,
MRI) != SingleLaneSizeInBits)
8038 if (!
MRI.hasOneNonDBGUse(Lane0LoadReg))
8041 LoadInstrs.
push_back(
MRI.getUniqueVRegDef(Lane0LoadReg));
8050 RemainingLoadInstrs.
insert(LoadInstrs.
begin(), LoadInstrs.
end());
8053 for (; MBBItr !=
MBB->begin() && RemainingSteps > 0 &&
8054 !RemainingLoadInstrs.
empty();
8055 --MBBItr, --RemainingSteps) {
8059 RemainingLoadInstrs.
erase(&CurrInstr);
8069 if (RemainingSteps == 0 && !RemainingLoadInstrs.
empty())
8095 case AArch64::LD1i32:
8097 case AArch64::LD1i16:
8099 case AArch64::LD1i8:
8115 unsigned Pattern,
unsigned NumLanes) {
8123 for (
unsigned i = 0; i < NumLanes - 1; ++i) {
8131 return A->getOperand(2).getImm() >
B->getOperand(2).getImm();
8136 MRI.getUniqueVRegDef(SubregToReg->getOperand(2).getReg()));
8137 auto LoadToLaneInstrsAscending =
llvm::reverse(LoadToLaneInstrs);
8143 auto CreateLD1Instruction = [&](
MachineInstr *OriginalInstr,
8144 Register SrcRegister,
unsigned Lane,
8146 bool OffsetRegisterKillState) {
8147 auto NewRegister =
MRI.createVirtualRegister(FPR128RegClass);
8154 InstrIdxForVirtReg.
insert(std::make_pair(NewRegister, InsInstrs.
size()));
8155 InsInstrs.
push_back(LoadIndexIntoRegister);
8161 auto CreateLDRInstruction = [&](
unsigned NumLanes,
Register DestReg,
8167 Opcode = AArch64::LDRSui;
8170 Opcode = AArch64::LDRHui;
8173 Opcode = AArch64::LDRBui;
8177 "Got unsupported number of lanes in machine-combiner gather pattern");
8186 auto LanesToLoadToReg0 =
8188 LoadToLaneInstrsAscending.begin() + NumLanes / 2);
8189 Register PrevReg = SubregToReg->getOperand(0).getReg();
8191 const MachineOperand &OffsetRegOperand = LoadInstr->getOperand(3);
8192 PrevReg = CreateLD1Instruction(LoadInstr, PrevReg, Index + 1,
8193 OffsetRegOperand.
getReg(),
8194 OffsetRegOperand.
isKill());
8201 MachineInstr *Lane0Load = *LoadToLaneInstrsAscending.begin();
8203 *std::next(LoadToLaneInstrsAscending.begin(), NumLanes / 2);
8204 Register DestRegForMiddleIndex =
MRI.createVirtualRegister(
8210 CreateLDRInstruction(NumLanes, DestRegForMiddleIndex,
8211 OriginalSplitToLoadOffsetOperand.
getReg(),
8212 OriginalSplitToLoadOffsetOperand.
isKill());
8214 InstrIdxForVirtReg.
insert(
8215 std::make_pair(DestRegForMiddleIndex, InsInstrs.
size()));
8216 InsInstrs.
push_back(MiddleIndexLoadInstr);
8220 Register DestRegForSubregToReg =
MRI.createVirtualRegister(FPR128RegClass);
8221 unsigned SubregType;
8224 SubregType = AArch64::ssub;
8227 SubregType = AArch64::hsub;
8230 SubregType = AArch64::bsub;
8234 "Got invalid NumLanes for machine-combiner gather pattern");
8237 auto SubRegToRegInstr =
8239 DestRegForSubregToReg)
8243 InstrIdxForVirtReg.
insert(
8244 std::make_pair(DestRegForSubregToReg, InsInstrs.
size()));
8248 auto LanesToLoadToReg1 =
8250 LoadToLaneInstrsAscending.end());
8251 PrevReg = SubRegToRegInstr->getOperand(0).getReg();
8253 const MachineOperand &OffsetRegOperand = LoadInstr->getOperand(3);
8254 PrevReg = CreateLD1Instruction(LoadInstr, PrevReg, Index + 1,
8255 OffsetRegOperand.
getReg(),
8256 OffsetRegOperand.
isKill());
8259 if (Index == NumLanes / 2 - 2) {
8294bool AArch64InstrInfo::getMachineCombinerPatterns(
8296 bool DoRegPressureReduce)
const {
8317 DoRegPressureReduce);
8346 const Register *ReplacedAddend =
nullptr) {
8347 assert(IdxMulOpd == 1 || IdxMulOpd == 2);
8349 unsigned IdxOtherOpd = IdxMulOpd == 1 ? 2 : 1;
8352 Register SrcReg0 = MUL->getOperand(1).getReg();
8353 bool Src0IsKill = MUL->getOperand(1).isKill();
8354 Register SrcReg1 = MUL->getOperand(2).getReg();
8355 bool Src1IsKill = MUL->getOperand(2).isKill();
8359 if (ReplacedAddend) {
8361 SrcReg2 = *ReplacedAddend;
8369 MRI.constrainRegClass(ResultReg, RC);
8371 MRI.constrainRegClass(SrcReg0, RC);
8373 MRI.constrainRegClass(SrcReg1, RC);
8375 MRI.constrainRegClass(SrcReg2, RC);
8388 .
addImm(MUL->getOperand(3).getImm());
8395 assert(
false &&
"Invalid FMA instruction kind \n");
8409 if (AArch64::FPR32RegClass.hasSubClassEq(RC))
8410 Opc = AArch64::FNMADDSrrr;
8411 else if (AArch64::FPR64RegClass.hasSubClassEq(RC))
8412 Opc = AArch64::FNMADDDrrr;
8424 MRI.constrainRegClass(ResultReg, RC);
8426 MRI.constrainRegClass(SrcReg0, RC);
8428 MRI.constrainRegClass(SrcReg1, RC);
8430 MRI.constrainRegClass(SrcReg2, RC);
8446 unsigned IdxDupOp,
unsigned MulOpc,
8448 assert(((IdxDupOp == 1) || (IdxDupOp == 2)) &&
8449 "Invalid index of FMUL operand");
8457 if (Dup->
getOpcode() == TargetOpcode::COPY)
8461 MRI.clearKillFlags(DupSrcReg);
8462 MRI.constrainRegClass(DupSrcReg, RC);
8466 unsigned IdxMulOp = IdxDupOp == 1 ? 2 : 1;
8507 InstrIdxForVirtReg.
insert(std::make_pair(NewVR, 0));
8522 genNeg(MF,
MRI,
TII, Root, InsInstrs, InstrIdxForVirtReg, MnegOpc, RC);
8549 genNeg(MF,
MRI,
TII, Root, InsInstrs, InstrIdxForVirtReg, MnegOpc, RC);
8577 unsigned IdxMulOpd,
unsigned MaddOpc,
unsigned VR,
8579 assert(IdxMulOpd == 1 || IdxMulOpd == 2);
8583 Register SrcReg0 = MUL->getOperand(1).getReg();
8584 bool Src0IsKill = MUL->getOperand(1).isKill();
8585 Register SrcReg1 = MUL->getOperand(2).getReg();
8586 bool Src1IsKill = MUL->getOperand(2).isKill();
8589 MRI.constrainRegClass(ResultReg, RC);
8591 MRI.constrainRegClass(SrcReg0, RC);
8593 MRI.constrainRegClass(SrcReg1, RC);
8595 MRI.constrainRegClass(VR, RC);
8616 assert(IdxOpd1 == 1 || IdxOpd1 == 2);
8617 unsigned IdxOtherOpd = IdxOpd1 == 1 ? 2 : 1;
8631 if (Opcode == AArch64::SUBSWrr)
8632 Opcode = AArch64::SUBWrr;
8633 else if (Opcode == AArch64::SUBSXrr)
8634 Opcode = AArch64::SUBXrr;
8636 assert((Opcode == AArch64::SUBWrr || Opcode == AArch64::SUBXrr) &&
8637 "Unexpected instruction opcode.");
8654 InstrIdxForVirtReg.
insert(std::make_pair(NewVR, 0));
8661unsigned AArch64InstrInfo::getReduceOpcodeForAccumulator(
8662 unsigned int AccumulatorOpCode)
const {
8663 switch (AccumulatorOpCode) {
8664 case AArch64::UABALB_ZZZ_D:
8665 case AArch64::SABALB_ZZZ_D:
8666 case AArch64::UABALT_ZZZ_D:
8667 case AArch64::SABALT_ZZZ_D:
8668 return AArch64::ADD_ZZZ_D;
8669 case AArch64::UABALB_ZZZ_H:
8670 case AArch64::SABALB_ZZZ_H:
8671 case AArch64::UABALT_ZZZ_H:
8672 case AArch64::SABALT_ZZZ_H:
8673 return AArch64::ADD_ZZZ_H;
8674 case AArch64::UABALB_ZZZ_S:
8675 case AArch64::SABALB_ZZZ_S:
8676 case AArch64::UABALT_ZZZ_S:
8677 case AArch64::SABALT_ZZZ_S:
8678 return AArch64::ADD_ZZZ_S;
8679 case AArch64::UABALv16i8_v8i16:
8680 case AArch64::SABALv8i8_v8i16:
8681 case AArch64::SABAv8i16:
8682 case AArch64::UABAv8i16:
8683 return AArch64::ADDv8i16;
8684 case AArch64::SABALv2i32_v2i64:
8685 case AArch64::UABALv2i32_v2i64:
8686 case AArch64::SABALv4i32_v2i64:
8687 return AArch64::ADDv2i64;
8688 case AArch64::UABALv4i16_v4i32:
8689 case AArch64::SABALv4i16_v4i32:
8690 case AArch64::SABALv8i16_v4i32:
8691 case AArch64::SABAv4i32:
8692 case AArch64::UABAv4i32:
8693 return AArch64::ADDv4i32;
8694 case AArch64::UABALv4i32_v2i64:
8695 return AArch64::ADDv2i64;
8696 case AArch64::UABALv8i16_v4i32:
8697 return AArch64::ADDv4i32;
8698 case AArch64::UABALv8i8_v8i16:
8699 case AArch64::SABALv16i8_v8i16:
8700 return AArch64::ADDv8i16;
8701 case AArch64::UABAv16i8:
8702 case AArch64::SABAv16i8:
8703 return AArch64::ADDv16i8;
8704 case AArch64::UABAv4i16:
8705 case AArch64::SABAv4i16:
8706 return AArch64::ADDv4i16;
8707 case AArch64::UABAv2i32:
8708 case AArch64::SABAv2i32:
8709 return AArch64::ADDv2i32;
8710 case AArch64::UABAv8i8:
8711 case AArch64::SABAv8i8:
8712 return AArch64::ADDv8i8;
8721void AArch64InstrInfo::genAlternativeCodeSequence(
8731 MachineInstr *
MUL =
nullptr;
8732 const TargetRegisterClass *RC;
8738 DelInstrs, InstrIdxForVirtReg);
8744 InstrIdxForVirtReg);
8750 InstrIdxForVirtReg);
8759 Opc = AArch64::MADDWrrr;
8760 RC = &AArch64::GPR32RegClass;
8762 Opc = AArch64::MADDXrrr;
8763 RC = &AArch64::GPR64RegClass;
8774 Opc = AArch64::MADDWrrr;
8775 RC = &AArch64::GPR32RegClass;
8777 Opc = AArch64::MADDXrrr;
8778 RC = &AArch64::GPR64RegClass;
8791 const TargetRegisterClass *RC;
8792 unsigned BitSize, MovImm;
8795 MovImm = AArch64::MOVi32imm;
8796 RC = &AArch64::GPR32spRegClass;
8798 Opc = AArch64::MADDWrrr;
8799 RC = &AArch64::GPR32RegClass;
8801 MovImm = AArch64::MOVi64imm;
8802 RC = &AArch64::GPR64spRegClass;
8804 Opc = AArch64::MADDXrrr;
8805 RC = &AArch64::GPR64RegClass;
8816 uint64_t UImm =
SignExtend64(IsSub ? -Imm : Imm, BitSize);
8820 if (Insn.
size() != 1)
8822 MachineInstrBuilder MIB1 =
8823 BuildMI(MF, MIMetadata(Root),
TII->get(MovImm), NewVR)
8824 .
addImm(IsSub ? -Imm : Imm);
8826 InstrIdxForVirtReg.
insert(std::make_pair(NewVR, 0));
8837 const TargetRegisterClass *SubRC;
8838 unsigned SubOpc, ZeroReg;
8840 SubOpc = AArch64::SUBWrr;
8841 SubRC = &AArch64::GPR32spRegClass;
8842 ZeroReg = AArch64::WZR;
8843 Opc = AArch64::MADDWrrr;
8844 RC = &AArch64::GPR32RegClass;
8846 SubOpc = AArch64::SUBXrr;
8847 SubRC = &AArch64::GPR64spRegClass;
8848 ZeroReg = AArch64::XZR;
8849 Opc = AArch64::MADDXrrr;
8850 RC = &AArch64::GPR64RegClass;
8852 Register NewVR =
MRI.createVirtualRegister(SubRC);
8854 MachineInstrBuilder MIB1 =
8855 BuildMI(MF, MIMetadata(Root),
TII->get(SubOpc), NewVR)
8859 InstrIdxForVirtReg.
insert(std::make_pair(NewVR, 0));
8870 Opc = AArch64::MSUBWrrr;
8871 RC = &AArch64::GPR32RegClass;
8873 Opc = AArch64::MSUBXrrr;
8874 RC = &AArch64::GPR64RegClass;
8879 Opc = AArch64::MLAv8i8;
8880 RC = &AArch64::FPR64RegClass;
8884 Opc = AArch64::MLAv8i8;
8885 RC = &AArch64::FPR64RegClass;
8889 Opc = AArch64::MLAv16i8;
8890 RC = &AArch64::FPR128RegClass;
8894 Opc = AArch64::MLAv16i8;
8895 RC = &AArch64::FPR128RegClass;
8899 Opc = AArch64::MLAv4i16;
8900 RC = &AArch64::FPR64RegClass;
8904 Opc = AArch64::MLAv4i16;
8905 RC = &AArch64::FPR64RegClass;
8909 Opc = AArch64::MLAv8i16;
8910 RC = &AArch64::FPR128RegClass;
8914 Opc = AArch64::MLAv8i16;
8915 RC = &AArch64::FPR128RegClass;
8919 Opc = AArch64::MLAv2i32;
8920 RC = &AArch64::FPR64RegClass;
8924 Opc = AArch64::MLAv2i32;
8925 RC = &AArch64::FPR64RegClass;
8929 Opc = AArch64::MLAv4i32;
8930 RC = &AArch64::FPR128RegClass;
8934 Opc = AArch64::MLAv4i32;
8935 RC = &AArch64::FPR128RegClass;
8940 Opc = AArch64::MLAv8i8;
8941 RC = &AArch64::FPR64RegClass;
8943 InstrIdxForVirtReg, 1,
Opc, AArch64::NEGv8i8,
8947 Opc = AArch64::MLSv8i8;
8948 RC = &AArch64::FPR64RegClass;
8952 Opc = AArch64::MLAv16i8;
8953 RC = &AArch64::FPR128RegClass;
8955 InstrIdxForVirtReg, 1,
Opc, AArch64::NEGv16i8,
8959 Opc = AArch64::MLSv16i8;
8960 RC = &AArch64::FPR128RegClass;
8964 Opc = AArch64::MLAv4i16;
8965 RC = &AArch64::FPR64RegClass;
8967 InstrIdxForVirtReg, 1,
Opc, AArch64::NEGv4i16,
8971 Opc = AArch64::MLSv4i16;
8972 RC = &AArch64::FPR64RegClass;
8976 Opc = AArch64::MLAv8i16;
8977 RC = &AArch64::FPR128RegClass;
8979 InstrIdxForVirtReg, 1,
Opc, AArch64::NEGv8i16,
8983 Opc = AArch64::MLSv8i16;
8984 RC = &AArch64::FPR128RegClass;
8988 Opc = AArch64::MLAv2i32;
8989 RC = &AArch64::FPR64RegClass;
8991 InstrIdxForVirtReg, 1,
Opc, AArch64::NEGv2i32,
8995 Opc = AArch64::MLSv2i32;
8996 RC = &AArch64::FPR64RegClass;
9000 Opc = AArch64::MLAv4i32;
9001 RC = &AArch64::FPR128RegClass;
9003 InstrIdxForVirtReg, 1,
Opc, AArch64::NEGv4i32,
9007 Opc = AArch64::MLSv4i32;
9008 RC = &AArch64::FPR128RegClass;
9013 Opc = AArch64::MLAv4i16_indexed;
9014 RC = &AArch64::FPR64RegClass;
9018 Opc = AArch64::MLAv4i16_indexed;
9019 RC = &AArch64::FPR64RegClass;
9023 Opc = AArch64::MLAv8i16_indexed;
9024 RC = &AArch64::FPR128RegClass;
9028 Opc = AArch64::MLAv8i16_indexed;
9029 RC = &AArch64::FPR128RegClass;
9033 Opc = AArch64::MLAv2i32_indexed;
9034 RC = &AArch64::FPR64RegClass;
9038 Opc = AArch64::MLAv2i32_indexed;
9039 RC = &AArch64::FPR64RegClass;
9043 Opc = AArch64::MLAv4i32_indexed;
9044 RC = &AArch64::FPR128RegClass;
9048 Opc = AArch64::MLAv4i32_indexed;
9049 RC = &AArch64::FPR128RegClass;
9054 Opc = AArch64::MLAv4i16_indexed;
9055 RC = &AArch64::FPR64RegClass;
9057 InstrIdxForVirtReg, 1,
Opc, AArch64::NEGv4i16,
9061 Opc = AArch64::MLSv4i16_indexed;
9062 RC = &AArch64::FPR64RegClass;
9066 Opc = AArch64::MLAv8i16_indexed;
9067 RC = &AArch64::FPR128RegClass;
9069 InstrIdxForVirtReg, 1,
Opc, AArch64::NEGv8i16,
9073 Opc = AArch64::MLSv8i16_indexed;
9074 RC = &AArch64::FPR128RegClass;
9078 Opc = AArch64::MLAv2i32_indexed;
9079 RC = &AArch64::FPR64RegClass;
9081 InstrIdxForVirtReg, 1,
Opc, AArch64::NEGv2i32,
9085 Opc = AArch64::MLSv2i32_indexed;
9086 RC = &AArch64::FPR64RegClass;
9090 Opc = AArch64::MLAv4i32_indexed;
9091 RC = &AArch64::FPR128RegClass;
9093 InstrIdxForVirtReg, 1,
Opc, AArch64::NEGv4i32,
9097 Opc = AArch64::MLSv4i32_indexed;
9098 RC = &AArch64::FPR128RegClass;
9104 Opc = AArch64::FMADDHrrr;
9105 RC = &AArch64::FPR16RegClass;
9109 Opc = AArch64::FMADDSrrr;
9110 RC = &AArch64::FPR32RegClass;
9114 Opc = AArch64::FMADDDrrr;
9115 RC = &AArch64::FPR64RegClass;
9120 Opc = AArch64::FMADDHrrr;
9121 RC = &AArch64::FPR16RegClass;
9125 Opc = AArch64::FMADDSrrr;
9126 RC = &AArch64::FPR32RegClass;
9130 Opc = AArch64::FMADDDrrr;
9131 RC = &AArch64::FPR64RegClass;
9136 Opc = AArch64::FMLAv1i32_indexed;
9137 RC = &AArch64::FPR32RegClass;
9142 Opc = AArch64::FMLAv1i32_indexed;
9143 RC = &AArch64::FPR32RegClass;
9149 Opc = AArch64::FMLAv1i64_indexed;
9150 RC = &AArch64::FPR64RegClass;
9155 Opc = AArch64::FMLAv1i64_indexed;
9156 RC = &AArch64::FPR64RegClass;
9162 RC = &AArch64::FPR64RegClass;
9163 Opc = AArch64::FMLAv4i16_indexed;
9168 RC = &AArch64::FPR64RegClass;
9169 Opc = AArch64::FMLAv4f16;
9174 RC = &AArch64::FPR64RegClass;
9175 Opc = AArch64::FMLAv4i16_indexed;
9180 RC = &AArch64::FPR64RegClass;
9181 Opc = AArch64::FMLAv4f16;
9188 RC = &AArch64::FPR64RegClass;
9190 Opc = AArch64::FMLAv2i32_indexed;
9194 Opc = AArch64::FMLAv2f32;
9201 RC = &AArch64::FPR64RegClass;
9203 Opc = AArch64::FMLAv2i32_indexed;
9207 Opc = AArch64::FMLAv2f32;
9214 RC = &AArch64::FPR128RegClass;
9215 Opc = AArch64::FMLAv8i16_indexed;
9220 RC = &AArch64::FPR128RegClass;
9221 Opc = AArch64::FMLAv8f16;
9226 RC = &AArch64::FPR128RegClass;
9227 Opc = AArch64::FMLAv8i16_indexed;
9232 RC = &AArch64::FPR128RegClass;
9233 Opc = AArch64::FMLAv8f16;
9240 RC = &AArch64::FPR128RegClass;
9242 Opc = AArch64::FMLAv2i64_indexed;
9246 Opc = AArch64::FMLAv2f64;
9253 RC = &AArch64::FPR128RegClass;
9255 Opc = AArch64::FMLAv2i64_indexed;
9259 Opc = AArch64::FMLAv2f64;
9267 RC = &AArch64::FPR128RegClass;
9269 Opc = AArch64::FMLAv4i32_indexed;
9273 Opc = AArch64::FMLAv4f32;
9281 RC = &AArch64::FPR128RegClass;
9283 Opc = AArch64::FMLAv4i32_indexed;
9287 Opc = AArch64::FMLAv4f32;
9294 Opc = AArch64::FNMSUBHrrr;
9295 RC = &AArch64::FPR16RegClass;
9299 Opc = AArch64::FNMSUBSrrr;
9300 RC = &AArch64::FPR32RegClass;
9304 Opc = AArch64::FNMSUBDrrr;
9305 RC = &AArch64::FPR64RegClass;
9310 Opc = AArch64::FNMADDHrrr;
9311 RC = &AArch64::FPR16RegClass;
9315 Opc = AArch64::FNMADDSrrr;
9316 RC = &AArch64::FPR32RegClass;
9320 Opc = AArch64::FNMADDDrrr;
9321 RC = &AArch64::FPR64RegClass;
9326 Opc = AArch64::FMSUBHrrr;
9327 RC = &AArch64::FPR16RegClass;
9331 Opc = AArch64::FMSUBSrrr;
9332 RC = &AArch64::FPR32RegClass;
9336 Opc = AArch64::FMSUBDrrr;
9337 RC = &AArch64::FPR64RegClass;
9342 Opc = AArch64::FMLSv1i32_indexed;
9343 RC = &AArch64::FPR32RegClass;
9349 Opc = AArch64::FMLSv1i64_indexed;
9350 RC = &AArch64::FPR64RegClass;
9357 RC = &AArch64::FPR64RegClass;
9359 MachineInstrBuilder MIB1 =
9360 BuildMI(MF, MIMetadata(Root),
TII->get(AArch64::FNEGv4f16), NewVR)
9363 InstrIdxForVirtReg.
insert(std::make_pair(NewVR, 0));
9365 Opc = AArch64::FMLAv4f16;
9369 Opc = AArch64::FMLAv4i16_indexed;
9376 RC = &AArch64::FPR64RegClass;
9377 Opc = AArch64::FMLSv4f16;
9382 RC = &AArch64::FPR64RegClass;
9383 Opc = AArch64::FMLSv4i16_indexed;
9390 RC = &AArch64::FPR64RegClass;
9392 Opc = AArch64::FMLSv2i32_indexed;
9396 Opc = AArch64::FMLSv2f32;
9404 RC = &AArch64::FPR128RegClass;
9406 MachineInstrBuilder MIB1 =
9407 BuildMI(MF, MIMetadata(Root),
TII->get(AArch64::FNEGv8f16), NewVR)
9410 InstrIdxForVirtReg.
insert(std::make_pair(NewVR, 0));
9412 Opc = AArch64::FMLAv8f16;
9416 Opc = AArch64::FMLAv8i16_indexed;
9423 RC = &AArch64::FPR128RegClass;
9424 Opc = AArch64::FMLSv8f16;
9429 RC = &AArch64::FPR128RegClass;
9430 Opc = AArch64::FMLSv8i16_indexed;
9437 RC = &AArch64::FPR128RegClass;
9439 Opc = AArch64::FMLSv2i64_indexed;
9443 Opc = AArch64::FMLSv2f64;
9451 RC = &AArch64::FPR128RegClass;
9453 Opc = AArch64::FMLSv4i32_indexed;
9457 Opc = AArch64::FMLSv4f32;
9464 RC = &AArch64::FPR64RegClass;
9466 MachineInstrBuilder MIB1 =
9467 BuildMI(MF, MIMetadata(Root),
TII->get(AArch64::FNEGv2f32), NewVR)
9470 InstrIdxForVirtReg.
insert(std::make_pair(NewVR, 0));
9472 Opc = AArch64::FMLAv2i32_indexed;
9476 Opc = AArch64::FMLAv2f32;
9484 RC = &AArch64::FPR128RegClass;
9486 MachineInstrBuilder MIB1 =
9487 BuildMI(MF, MIMetadata(Root),
TII->get(AArch64::FNEGv4f32), NewVR)
9490 InstrIdxForVirtReg.
insert(std::make_pair(NewVR, 0));
9492 Opc = AArch64::FMLAv4i32_indexed;
9496 Opc = AArch64::FMLAv4f32;
9504 RC = &AArch64::FPR128RegClass;
9506 MachineInstrBuilder MIB1 =
9507 BuildMI(MF, MIMetadata(Root),
TII->get(AArch64::FNEGv2f64), NewVR)
9510 InstrIdxForVirtReg.
insert(std::make_pair(NewVR, 0));
9512 Opc = AArch64::FMLAv2i64_indexed;
9516 Opc = AArch64::FMLAv2f64;
9528 &AArch64::FPR128RegClass,
MRI);
9537 &AArch64::FPR128RegClass,
MRI);
9546 &AArch64::FPR128_loRegClass,
MRI);
9555 &AArch64::FPR128RegClass,
MRI);
9564 &AArch64::FPR128_loRegClass,
MRI);
9598 for (
auto *
MI : InsInstrs)
9599 MI->setFlags(Flags);
9640 bool IsNegativeBranch =
false;
9641 bool IsTestAndBranch =
false;
9642 unsigned TargetBBInMI = 0;
9643 switch (
MI.getOpcode()) {
9647 case AArch64::CBWPri:
9648 case AArch64::CBXPri:
9649 case AArch64::CBBAssertExt:
9650 case AArch64::CBHAssertExt:
9651 case AArch64::CBWPrr:
9652 case AArch64::CBXPrr:
9658 case AArch64::CBNZW:
9659 case AArch64::CBNZX:
9661 IsNegativeBranch =
true;
9666 IsTestAndBranch =
true;
9668 case AArch64::TBNZW:
9669 case AArch64::TBNZX:
9671 IsNegativeBranch =
true;
9672 IsTestAndBranch =
true;
9678 if (IsTestAndBranch &&
MI.getOperand(1).getImm())
9682 assert(
MI.getParent() &&
"Incomplete machine instruction\n");
9693 while (
DefMI->isCopy()) {
9695 if (!
MRI->hasOneNonDBGUse(CopyVReg))
9697 if (!
MRI->hasOneDef(CopyVReg))
9702 switch (
DefMI->getOpcode()) {
9706 case AArch64::ANDWri:
9707 case AArch64::ANDXri: {
9708 if (IsTestAndBranch)
9712 if (!
MRI->hasOneNonDBGUse(VReg))
9715 bool Is32Bit = (
DefMI->getOpcode() == AArch64::ANDWri);
9717 DefMI->getOperand(2).getImm(), Is32Bit ? 32 : 64);
9726 assert(!
MRI->def_empty(NewReg) &&
"Register must be defined.");
9732 unsigned Opc = (Imm < 32)
9733 ? (IsNegativeBranch ? AArch64::TBNZW : AArch64::TBZW)
9734 : (IsNegativeBranch ? AArch64::TBNZX : AArch64::TBZX);
9747 if (!Is32Bit && Imm < 32)
9749 MI.eraseFromParent();
9753 case AArch64::CSINCWr:
9754 case AArch64::CSINCXr: {
9755 if (!(
DefMI->getOperand(1).getReg() == AArch64::WZR &&
9756 DefMI->getOperand(2).getReg() == AArch64::WZR) &&
9757 !(
DefMI->getOperand(1).getReg() == AArch64::XZR &&
9758 DefMI->getOperand(2).getReg() == AArch64::XZR))
9761 if (
DefMI->findRegisterDefOperandIdx(AArch64::NZCV,
nullptr,
9774 if (IsNegativeBranch)
9777 MI.eraseFromParent();
9783std::pair<unsigned, unsigned>
9784AArch64InstrInfo::decomposeMachineOperandsTargetFlags(
unsigned TF)
const {
9786 return std::make_pair(TF & Mask, TF & ~Mask);
9790AArch64InstrInfo::getSerializableDirectMachineOperandTargetFlags()
const {
9793 static const std::pair<unsigned, const char *> TargetFlags[] = {
9794 {MO_PAGE,
"aarch64-page"}, {
MO_PAGEOFF,
"aarch64-pageoff"},
9795 {
MO_G3,
"aarch64-g3"}, {
MO_G2,
"aarch64-g2"},
9796 {
MO_G1,
"aarch64-g1"}, {
MO_G0,
"aarch64-g0"},
9802AArch64InstrInfo::getSerializableBitmaskMachineOperandTargetFlags()
const {
9803 using namespace AArch64II;
9805 static const std::pair<unsigned, const char *> TargetFlags[] = {
9808 {
MO_NC,
"aarch64-nc"},
9809 {
MO_S,
"aarch64-s"},
9820AArch64InstrInfo::getSerializableMachineMemOperandTargetFlags()
const {
9821 static const std::pair<MachineMemOperand::Flags, const char *> TargetFlags[] =
9923 MachineFunction *MF =
C.getMF();
9925 const AArch64RegisterInfo *ARI =
9926 static_cast<const AArch64RegisterInfo *
>(&
TRI);
9929 for (
unsigned Reg : AArch64::GPR64RegClass) {
9931 Reg != AArch64::LR &&
9932 Reg != AArch64::X16 &&
9933 Reg != AArch64::X17 &&
9934 C.isAvailableAcrossAndOutOfSeq(
Reg,
TRI) &&
9935 C.isAvailableInsideSeq(
Reg,
TRI))
9966 return SubtargetA.hasV8_3aOps() == SubtargetB.hasV8_3aOps();
9969std::optional<std::unique_ptr<outliner::OutlinedFunction>>
9970AArch64InstrInfo::getOutliningCandidateInfo(
9972 std::vector<outliner::Candidate> &RepeatedSequenceLocs,
9973 unsigned MinRepeats)
const {
9974 unsigned SequenceSize = 0;
9975 for (
auto &
MI : RepeatedSequenceLocs[0])
9978 unsigned NumBytesToCreateFrame = 0;
9984 MachineInstr &LastMI = RepeatedSequenceLocs[0].back();
9985 MachineInstr &FirstMI = RepeatedSequenceLocs[0].front();
9986 if (LastMI.
getOpcode() == AArch64::ADRP &&
9989 return std::nullopt;
9994 if ((FirstMI.
getOpcode() == AArch64::ADDXri ||
9995 FirstMI.
getOpcode() == AArch64::LDRXui) &&
9998 return std::nullopt;
10009 if (std::adjacent_find(
10010 RepeatedSequenceLocs.begin(), RepeatedSequenceLocs.end(),
10011 [](
const outliner::Candidate &a,
const outliner::Candidate &b) {
10014 if (outliningCandidatesSigningScopeConsensus(a, b) &&
10015 outliningCandidatesSigningKeyConsensus(a, b) &&
10016 outliningCandidatesV8_3OpsConsensus(a, b)) {
10020 }) != RepeatedSequenceLocs.end()) {
10021 return std::nullopt;
10038 unsigned NumBytesToCheckLRInTCEpilogue = 0;
10039 const auto RASignCondition = RepeatedSequenceLocs[0]
10042 ->getSignReturnAddressCondition();
10045 NumBytesToCreateFrame += 8;
10048 auto LRCheckMethod = Subtarget.getAuthenticatedLRCheckMethod(
10049 *RepeatedSequenceLocs[0].getMF());
10050 NumBytesToCheckLRInTCEpilogue =
10054 if (isTailCallReturnInst(RepeatedSequenceLocs[0].
back()))
10055 SequenceSize += NumBytesToCheckLRInTCEpilogue;
10063 for (
auto &
MI :
C) {
10064 if (
MI.modifiesRegister(AArch64::SP, &
TRI)) {
10065 switch (
MI.getOpcode()) {
10066 case AArch64::ADDXri:
10067 case AArch64::ADDWri:
10068 assert(
MI.getNumOperands() == 4 &&
"Wrong number of operands");
10069 assert(
MI.getOperand(2).isImm() &&
10070 "Expected operand to be immediate");
10071 assert(
MI.getOperand(1).isReg() &&
10072 "Expected operand to be a register");
10076 if (
MI.getOperand(1).getReg() == AArch64::SP)
10077 SPValue +=
MI.getOperand(2).getImm();
10081 case AArch64::SUBXri:
10082 case AArch64::SUBWri:
10083 assert(
MI.getNumOperands() == 4 &&
"Wrong number of operands");
10084 assert(
MI.getOperand(2).isImm() &&
10085 "Expected operand to be immediate");
10086 assert(
MI.getOperand(1).isReg() &&
10087 "Expected operand to be a register");
10091 if (
MI.getOperand(1).getReg() == AArch64::SP)
10092 SPValue -=
MI.getOperand(2).getImm();
10109 if (RepeatedSequenceLocs.size() < MinRepeats)
10110 return std::nullopt;
10114 unsigned FlagsSetInAll = 0xF;
10118 FlagsSetInAll &=
C.Flags;
10120 unsigned LastInstrOpcode = RepeatedSequenceLocs[0].back().getOpcode();
10123 auto SetCandidateCallInfo =
10124 [&RepeatedSequenceLocs](
unsigned CallID,
unsigned NumBytesForCall) {
10126 C.setCallInfo(CallID, NumBytesForCall);
10130 NumBytesToCreateFrame += 4;
10138 unsigned CFICount = 0;
10139 for (
auto &
I : RepeatedSequenceLocs[0]) {
10140 if (
I.isCFIInstruction())
10150 std::vector<MCCFIInstruction> CFIInstructions =
10151 C.getMF()->getFrameInstructions();
10153 if (CFICount > 0 && CFICount != CFIInstructions.size())
10154 return std::nullopt;
10162 if (!
MI.modifiesRegister(AArch64::SP, &
TRI) &&
10163 !
MI.readsRegister(AArch64::SP, &
TRI))
10169 if (
MI.modifiesRegister(AArch64::SP, &
TRI))
10174 if (
MI.mayLoadOrStore()) {
10177 bool OffsetIsScalable;
10181 if (!getMemOperandWithOffset(
MI,
Base,
Offset, OffsetIsScalable, &
TRI) ||
10182 !
Base->isReg() ||
Base->getReg() != AArch64::SP)
10186 if (OffsetIsScalable)
10194 TypeSize Scale(0U,
false), DummyWidth(0U,
false);
10195 getMemOpInfo(
MI.getOpcode(), Scale, DummyWidth, MinOffset, MaxOffset);
10198 if (
Offset < MinOffset * (int64_t)Scale.getFixedValue() ||
10199 Offset > MaxOffset * (int64_t)Scale.getFixedValue())
10214 bool AllStackInstrsSafe =
10219 if (RepeatedSequenceLocs[0].
back().isTerminator()) {
10221 NumBytesToCreateFrame = 0;
10222 unsigned NumBytesForCall = 4 + NumBytesToCheckLRInTCEpilogue;
10226 else if (LastInstrOpcode == AArch64::BL ||
10227 ((LastInstrOpcode == AArch64::BLR ||
10228 LastInstrOpcode == AArch64::BLRNoIP) &&
10232 NumBytesToCreateFrame = NumBytesToCheckLRInTCEpilogue;
10240 unsigned NumBytesNoStackCalls = 0;
10241 std::vector<outliner::Candidate> CandidatesWithoutStackFixups;
10247 ?
C.isAvailableAcrossAndOutOfSeq(AArch64::LR,
TRI)
10256 C.getMF()->getFunction().hasFnAttribute(Attribute::NoReturn);
10259 if (LRAvailable && !IsNoReturn) {
10260 NumBytesNoStackCalls += 4;
10262 CandidatesWithoutStackFixups.push_back(
C);
10267 else if (findRegisterToSaveLRTo(
C)) {
10268 NumBytesNoStackCalls += 12;
10270 CandidatesWithoutStackFixups.push_back(
C);
10275 else if (
C.isAvailableInsideSeq(AArch64::SP,
TRI)) {
10276 NumBytesNoStackCalls += 12;
10278 CandidatesWithoutStackFixups.push_back(
C);
10284 NumBytesNoStackCalls += SequenceSize;
10291 if (!AllStackInstrsSafe ||
10292 NumBytesNoStackCalls <= RepeatedSequenceLocs.size() * 12) {
10293 RepeatedSequenceLocs = CandidatesWithoutStackFixups;
10295 if (RepeatedSequenceLocs.size() < MinRepeats)
10296 return std::nullopt;
10349 (!
C.isAvailableAcrossAndOutOfSeq(AArch64::LR,
TRI) ||
10350 !findRegisterToSaveLRTo(
C));
10356 if (RepeatedSequenceLocs.size() < MinRepeats)
10357 return std::nullopt;
10366 bool ModStackToSaveLR =
false;
10369 ModStackToSaveLR =
true;
10378 ModStackToSaveLR =
true;
10380 if (ModStackToSaveLR) {
10382 if (!AllStackInstrsSafe)
10383 return std::nullopt;
10386 NumBytesToCreateFrame += 8;
10393 return std::nullopt;
10395 return std::make_unique<outliner::OutlinedFunction>(
10396 RepeatedSequenceLocs, SequenceSize, NumBytesToCreateFrame, FrameID);
10399void AArch64InstrInfo::mergeOutliningCandidateAttributes(
10400 Function &
F, std::vector<outliner::Candidate> &Candidates)
const {
10404 const auto &CFn = Candidates.front().getMF()->getFunction();
10406 if (CFn.hasFnAttribute(
"ptrauth-returns"))
10407 F.addFnAttr(CFn.getFnAttribute(
"ptrauth-returns"));
10408 if (CFn.hasFnAttribute(
"ptrauth-auth-traps"))
10409 F.addFnAttr(CFn.getFnAttribute(
"ptrauth-auth-traps"));
10412 if (CFn.hasFnAttribute(
"sign-return-address"))
10413 F.addFnAttr(CFn.getFnAttribute(
"sign-return-address"));
10414 if (CFn.hasFnAttribute(
"sign-return-address-key"))
10415 F.addFnAttr(CFn.getFnAttribute(
"sign-return-address-key"));
10417 AArch64GenInstrInfo::mergeOutliningCandidateAttributes(
F, Candidates);
10420bool AArch64InstrInfo::isFunctionSafeToOutlineFrom(
10425 if (!OutlineFromLinkOnceODRs &&
F.hasLinkOnceODRLinkage())
10432 if (
F.hasSection())
10438 AArch64FunctionInfo *AFI = MF.
getInfo<AArch64FunctionInfo>();
10439 if (!AFI || AFI->
hasRedZone().value_or(
true))
10459 unsigned &Flags)
const {
10461 "Must track liveness!");
10463 std::pair<MachineBasicBlock::iterator, MachineBasicBlock::iterator>>
10478 auto AreAllUnsafeRegsDead = [&LRU]() {
10479 return LRU.available(AArch64::W16) && LRU.available(AArch64::W17) &&
10480 LRU.available(AArch64::NZCV);
10495 bool LRAvailableEverywhere =
true;
10497 LRU.addLiveOuts(
MBB);
10499 auto UpdateWholeMBBFlags = [&
Flags](
const MachineInstr &
MI) {
10500 if (
MI.isCall() && !
MI.isTerminator())
10506 auto CreateNewRangeStartingAt =
10507 [&RangeBegin, &RangeEnd,
10509 RangeBegin = NewBegin;
10510 RangeEnd = std::next(RangeBegin);
10513 auto SaveRangeIfNonEmpty = [&RangeLen, &
Ranges, &RangeBegin, &RangeEnd]() {
10519 if (!RangeBegin.isEnd() && RangeBegin->isBundledWithPred())
10521 if (!RangeEnd.isEnd() && RangeEnd->isBundledWithPred())
10523 Ranges.emplace_back(RangeBegin, RangeEnd);
10531 for (; FirstPossibleEndPt !=
MBB.
instr_rend(); ++FirstPossibleEndPt) {
10532 LRU.stepBackward(*FirstPossibleEndPt);
10535 UpdateWholeMBBFlags(*FirstPossibleEndPt);
10536 if (AreAllUnsafeRegsDead())
10543 CreateNewRangeStartingAt(FirstPossibleEndPt->getIterator());
10548 LRU.stepBackward(
MI);
10549 UpdateWholeMBBFlags(
MI);
10550 if (!AreAllUnsafeRegsDead()) {
10551 SaveRangeIfNonEmpty();
10552 CreateNewRangeStartingAt(
MI.getIterator());
10555 LRAvailableEverywhere &= LRU.available(AArch64::LR);
10556 RangeBegin =
MI.getIterator();
10561 if (AreAllUnsafeRegsDead())
10562 SaveRangeIfNonEmpty();
10570 if (!LRAvailableEverywhere)
10578 unsigned Flags)
const {
10579 MachineInstr &
MI = *MIT;
10583 switch (
MI.getOpcode()) {
10584 case AArch64::PACM:
10585 case AArch64::PACIASP:
10586 case AArch64::PACIBSP:
10587 case AArch64::PACIASPPC:
10588 case AArch64::PACIBSPPC:
10589 case AArch64::AUTIASP:
10590 case AArch64::AUTIBSP:
10591 case AArch64::AUTIASPPCi:
10592 case AArch64::AUTIASPPCr:
10593 case AArch64::AUTIBSPPCi:
10594 case AArch64::AUTIBSPPCr:
10595 case AArch64::RETAA:
10596 case AArch64::RETAB:
10597 case AArch64::RETAASPPCi:
10598 case AArch64::RETAASPPCr:
10599 case AArch64::RETABSPPCi:
10600 case AArch64::RETABSPPCr:
10601 case AArch64::EMITBKEY:
10602 case AArch64::PAUTH_PROLOGUE:
10603 case AArch64::PAUTH_EPILOGUE:
10613 if (
MI.isCFIInstruction())
10617 if (
MI.isTerminator())
10623 for (
const MachineOperand &MOP :
MI.operands()) {
10626 assert(!MOP.isCFIIndex());
10629 if (MOP.isReg() && !MOP.isImplicit() &&
10630 (MOP.getReg() == AArch64::LR || MOP.getReg() == AArch64::W30))
10637 if (
MI.getOpcode() == AArch64::ADRP)
10657 for (
const MachineOperand &MOP :
MI.operands()) {
10658 if (MOP.isGlobal()) {
10666 if (Callee &&
Callee->getName() ==
"\01_mcount")
10674 if (
MI.getOpcode() == AArch64::BLR ||
10675 MI.getOpcode() == AArch64::BLRNoIP ||
MI.getOpcode() == AArch64::BL)
10679 return UnknownCallOutlineType;
10687 return UnknownCallOutlineType;
10695 return UnknownCallOutlineType;
10716 for (MachineInstr &
MI :
MBB) {
10717 const MachineOperand *
Base;
10718 TypeSize Width(0,
false);
10720 bool OffsetIsScalable;
10723 if (!
MI.mayLoadOrStore() ||
10726 (
Base->isReg() &&
Base->getReg() != AArch64::SP))
10730 TypeSize Scale(0U,
false);
10731 int64_t Dummy1, Dummy2;
10734 assert(StackOffsetOperand.
isImm() &&
"Stack offset wasn't immediate!");
10736 assert(Scale != 0 &&
"Unexpected opcode!");
10737 assert(!OffsetIsScalable &&
"Expected offset to be a byte offset");
10742 int64_t NewImm = (
Offset + 16) / (int64_t)Scale.getFixedValue();
10743 StackOffsetOperand.
setImm(NewImm);
10749 bool ShouldSignReturnAddr) {
10750 if (!ShouldSignReturnAddr)
10756 TII->get(AArch64::PAUTH_EPILOGUE))
10760void AArch64InstrInfo::buildOutlinedFrame(
10764 AArch64FunctionInfo *FI = MF.
getInfo<AArch64FunctionInfo>();
10772 unsigned TailOpcode;
10774 TailOpcode = AArch64::TCRETURNdi;
10778 TailOpcode = AArch64::TCRETURNriALL;
10789 bool IsLeafFunction =
true;
10792 auto IsNonTailCall = [](
const MachineInstr &
MI) {
10793 return MI.isCall() && !
MI.isReturn();
10803 "Can only fix up stack references once");
10804 fixupPostOutline(
MBB);
10806 IsLeafFunction =
false;
10817 Et = std::prev(
MBB.
end());
10827 if (MF.
getInfo<AArch64FunctionInfo>()->needsDwarfUnwindInfo(MF)) {
10831 CFIBuilder.buildDefCFAOffset(16);
10835 CFIBuilder.buildOffset(AArch64::LR, -16);
10849 RASignCondition, !IsLeafFunction);
10878 fixupPostOutline(
MBB);
10889 .addGlobalAddress(
M.getNamedValue(MF.
getName()))
10899 .addGlobalAddress(
M.getNamedValue(MF.
getName())));
10908 MachineInstr *Save;
10909 MachineInstr *Restore;
10915 assert(
Reg &&
"No callee-saved register available?");
10949 .addGlobalAddress(
M.getNamedValue(MF.
getName())));
10957bool AArch64InstrInfo::shouldOutlineFromFunctionByDefault(
10965 bool AllowSideEffects)
const {
10967 const AArch64Subtarget &STI = MF.
getSubtarget<AArch64Subtarget>();
10970 if (
TRI.isGeneralPurposeRegister(MF,
Reg)) {
10983 assert(STI.hasNEON() &&
"Expected to have NEON.");
10989std::optional<DestSourcePair>
10994 if (((
MI.getOpcode() == AArch64::ORRWrs &&
10995 MI.getOperand(1).getReg() == AArch64::WZR &&
10996 MI.getOperand(3).getImm() == 0x0) ||
10997 (
MI.getOpcode() == AArch64::ORRWrr &&
10998 MI.getOperand(1).getReg() == AArch64::WZR)) &&
11000 (!
MI.getOperand(0).getReg().isVirtual() ||
11001 MI.getOperand(0).getSubReg() == 0) &&
11002 (!
MI.getOperand(0).getReg().isPhysical() ||
11007 if (
MI.getOpcode() == AArch64::ORRXrs &&
11008 MI.getOperand(1).getReg() == AArch64::XZR &&
11009 MI.getOperand(3).getImm() == 0x0)
11012 return std::nullopt;
11015std::optional<DestSourcePair>
11017 if ((
MI.getOpcode() == AArch64::ORRWrs &&
11018 MI.getOperand(1).getReg() == AArch64::WZR &&
11019 MI.getOperand(3).getImm() == 0x0) ||
11020 (
MI.getOpcode() == AArch64::ORRWrr &&
11021 MI.getOperand(1).getReg() == AArch64::WZR))
11023 return std::nullopt;
11026std::optional<RegImmPair>
11035 return std::nullopt;
11037 switch (
MI.getOpcode()) {
11039 return std::nullopt;
11040 case AArch64::SUBWri:
11041 case AArch64::SUBXri:
11042 case AArch64::SUBSWri:
11043 case AArch64::SUBSXri:
11046 case AArch64::ADDSWri:
11047 case AArch64::ADDSXri:
11048 case AArch64::ADDWri:
11049 case AArch64::ADDXri: {
11051 if (!
MI.getOperand(0).isReg() || !
MI.getOperand(1).isReg() ||
11052 !
MI.getOperand(2).isImm())
11053 return std::nullopt;
11054 int Shift =
MI.getOperand(3).getImm();
11055 assert((Shift == 0 || Shift == 12) &&
"Shift can be either 0 or 12");
11059 return RegImmPair{
MI.getOperand(1).getReg(),
Offset};
11065static std::optional<ParamLoadedValue>
11069 auto DestSrc =
TII->isCopyLikeInstr(
MI);
11071 return std::nullopt;
11073 Register DestReg = DestSrc->Destination->getReg();
11074 Register SrcReg = DestSrc->Source->getReg();
11077 return std::nullopt;
11082 if (DestReg == DescribedReg)
11086 if (
MI.getOpcode() == AArch64::ORRWrs &&
11087 TRI->isSuperRegister(DestReg, DescribedReg))
11091 if (
MI.getOpcode() == AArch64::ORRXrs &&
11092 TRI->isSubRegister(DestReg, DescribedReg)) {
11093 Register SrcSubReg =
TRI->getSubReg(SrcReg, AArch64::sub_32);
11097 assert(!
TRI->isSuperOrSubRegisterEq(DestReg, DescribedReg) &&
11098 "Unhandled ORR[XW]rs copy case");
11100 return std::nullopt;
11103bool AArch64InstrInfo::isFunctionSafeToSplit(
const MachineFunction &MF)
const {
11108 if (MF.
getInfo<AArch64FunctionInfo>()->hasRedZone().value_or(
true))
11114bool AArch64InstrInfo::isMBBSafeToSplitToCold(
11118 auto isAsmGoto = [](
const MachineInstr &
MI) {
11119 return MI.getOpcode() == AArch64::INLINEASM_BR;
11129 auto containsMBB = [&
MBB](
const MachineJumpTableEntry &JTE) {
11136 for (
const MachineInstr &
MI :
MBB) {
11137 switch (
MI.getOpcode()) {
11138 case TargetOpcode::G_BRJT:
11139 case AArch64::JumpTableDest32:
11140 case AArch64::JumpTableDest16:
11141 case AArch64::JumpTableDest8:
11152std::optional<ParamLoadedValue>
11155 const MachineFunction *MF =
MI.getMF();
11157 switch (
MI.getOpcode()) {
11158 case AArch64::MOVZWi:
11159 case AArch64::MOVZXi: {
11162 if (!
TRI->isSuperRegisterEq(
MI.getOperand(0).getReg(),
Reg))
11163 return std::nullopt;
11165 if (!
MI.getOperand(1).isImm())
11166 return std::nullopt;
11167 int64_t Immediate =
MI.getOperand(1).getImm();
11168 int Shift =
MI.getOperand(2).getImm();
11172 case AArch64::ORRWrs:
11173 case AArch64::ORRXrs:
11180bool AArch64InstrInfo::isExtendLikelyToBeFolded(
11183 ExtMI.
getOpcode() == TargetOpcode::G_ZEXT ||
11184 ExtMI.
getOpcode() == TargetOpcode::G_ANYEXT);
11187 if (ExtMI.
getOpcode() == TargetOpcode::G_ANYEXT)
11191 if (!
MRI.hasOneNonDBGUse(DefReg))
11196 auto *UserMI = &*
MRI.use_instr_nodbg_begin(DefReg);
11197 return UserMI->getOpcode() == TargetOpcode::G_PTR_ADD;
11200uint64_t AArch64InstrInfo::getElementSizeForOpcode(
unsigned Opc)
const {
11204bool AArch64InstrInfo::isPTestLikeOpcode(
unsigned Opc)
const {
11208bool AArch64InstrInfo::isWhileOpcode(
unsigned Opc)
const {
11213AArch64InstrInfo::getTailDuplicateSize(
CodeGenOptLevel OptLevel)
const {
11217bool AArch64InstrInfo::isLegalAddressingMode(
unsigned NumBytes, int64_t
Offset,
11218 unsigned Scale)
const {
11229 unsigned Shift =
Log2_64(NumBytes);
11230 if (NumBytes &&
Offset > 0 && (
Offset / NumBytes) <= (1LL << 12) - 1 &&
11238 return Scale == 1 || (Scale > 0 && Scale == NumBytes);
11243 return AArch64::BLRNoIP;
11245 return AArch64::BLR;
11250 Register TargetReg,
bool FrameSetup)
const {
11251 assert(TargetReg != AArch64::SP &&
"New top of stack cannot already be in SP");
11263 MF.
insert(MBBInsertPoint, LoopTestMBB);
11266 MF.
insert(MBBInsertPoint, LoopBodyMBB);
11268 MF.
insert(MBBInsertPoint, ExitMBB);
11278 BuildMI(*LoopTestMBB, LoopTestMBB->
end(),
DL,
TII->get(AArch64::SUBSXrx64),
11286 BuildMI(*LoopTestMBB, LoopTestMBB->
end(),
DL,
TII->get(AArch64::Bcc))
11292 BuildMI(*LoopBodyMBB, LoopBodyMBB->
end(),
DL,
TII->get(AArch64::LDRXui))
11309 BuildMI(*ExitMBB, ExitMBB->
end(),
DL,
TII->get(AArch64::ADDXri), AArch64::SP)
11328 MBB.addSuccessor(LoopTestMBB);
11334 return ExitMBB->
begin();
11351 unsigned CompCounterOprNum;
11355 unsigned UpdateCounterOprNum;
11359 bool IsUpdatePriorComp;
11371 TII(MF->getSubtarget().getInstrInfo()),
11372 TRI(MF->getSubtarget().getRegisterInfo()),
MRI(MF->getRegInfo()),
11373 LoopBB(LoopBB), CondBranch(CondBranch), Comp(Comp),
11374 CompCounterOprNum(CompCounterOprNum), Update(Update),
11375 UpdateCounterOprNum(UpdateCounterOprNum),
Init(
Init),
11376 IsUpdatePriorComp(IsUpdatePriorComp),
Cond(
Cond.begin(),
Cond.end()) {}
11378 bool shouldIgnoreForPipelining(
const MachineInstr *
MI)
const override {
11384 std::optional<bool> createTripCountGreaterCondition(
11385 int TC, MachineBasicBlock &
MBB,
11386 SmallVectorImpl<MachineOperand> &CondParam)
override {
11394 void createRemainingIterationsGreaterCondition(
11395 int TC, MachineBasicBlock &
MBB, SmallVectorImpl<MachineOperand> &
Cond,
11396 DenseMap<MachineInstr *, MachineInstr *> &LastStage0Insts)
override;
11398 void setPreheader(MachineBasicBlock *NewPreheader)
override {}
11400 void adjustTripCount(
int TripCountAdjust)
override {}
11402 bool isMVEExpanderSupported()
override {
return true; }
11418 Result =
MRI.createVirtualRegister(
11421 }
else if (
I == ReplaceOprNum) {
11422 MRI.constrainRegClass(ReplaceReg,
TII->getRegClass(NewMI->
getDesc(),
I));
11426 MBB.insert(InsertTo, NewMI);
11430void AArch64PipelinerLoopInfo::createRemainingIterationsGreaterCondition(
11446 assert(CondBranch->getOpcode() == AArch64::Bcc);
11450 if (CondBranch->getOperand(1).getMBB() == LoopBB)
11457 auto AccumulateCond = [&](
Register CurCond,
11459 Register NewCond =
MRI.createVirtualRegister(&AArch64::GPR64commonRegClass);
11468 if (!LastStage0Insts.
empty() && LastStage0Insts[Comp]->getParent() == &
MBB) {
11472 for (
int I = 0;
I <= TC; ++
I) {
11478 AccCond = AccumulateCond(AccCond, CC);
11482 if (Update != Comp && IsUpdatePriorComp) {
11484 LastStage0Insts[Comp]->getOperand(CompCounterOprNum).getReg();
11485 NextCounter =
cloneInstr(Update, UpdateCounterOprNum, Counter,
MBB,
11489 NextCounter = LastStage0Insts[Update]->getOperand(0).getReg();
11491 }
else if (Update != Comp) {
11496 Counter = NextCounter;
11500 if (LastStage0Insts.
empty()) {
11504 if (IsUpdatePriorComp)
11509 Counter = LastStage0Insts[Comp]->getOperand(CompCounterOprNum).getReg();
11512 for (
int I = 0;
I <= TC; ++
I) {
11516 AccCond = AccumulateCond(AccCond, CC);
11517 if (
I != TC && Update != Comp)
11520 Counter = NextCounter;
11536 assert(Phi.getNumOperands() == 5);
11537 if (Phi.getOperand(2).getMBB() ==
MBB) {
11538 RegMBB = Phi.getOperand(1).getReg();
11539 RegOther = Phi.getOperand(3).getReg();
11541 assert(Phi.getOperand(4).getMBB() ==
MBB);
11542 RegMBB = Phi.getOperand(3).getReg();
11543 RegOther = Phi.getOperand(1).getReg();
11548 if (!
Reg.isVirtual())
11551 return MRI.getVRegDef(
Reg)->getParent() != BB;
11557 unsigned &UpdateCounterOprNum,
Register &InitReg,
11558 bool &IsUpdatePriorComp) {
11572 if (!
Reg.isVirtual())
11575 UpdateInst =
nullptr;
11576 UpdateCounterOprNum = 0;
11578 IsUpdatePriorComp =
true;
11582 if (Def->getParent() != LoopBB)
11584 if (Def->isCopy()) {
11586 if (Def->getOperand(0).getSubReg() || Def->getOperand(1).getSubReg())
11588 CurReg = Def->getOperand(1).getReg();
11589 }
else if (Def->isPHI()) {
11593 IsUpdatePriorComp =
false;
11598 switch (Def->getOpcode()) {
11599 case AArch64::ADDSXri:
11600 case AArch64::ADDSWri:
11601 case AArch64::SUBSXri:
11602 case AArch64::SUBSWri:
11603 case AArch64::ADDXri:
11604 case AArch64::ADDWri:
11605 case AArch64::SUBXri:
11606 case AArch64::SUBWri:
11608 UpdateCounterOprNum = 1;
11610 case AArch64::ADDSXrr:
11611 case AArch64::ADDSWrr:
11612 case AArch64::SUBSXrr:
11613 case AArch64::SUBSWrr:
11614 case AArch64::ADDXrr:
11615 case AArch64::ADDWrr:
11616 case AArch64::SUBXrr:
11617 case AArch64::SUBWrr:
11620 UpdateCounterOprNum = 1;
11622 UpdateCounterOprNum = 2;
11629 CurReg = Def->getOperand(UpdateCounterOprNum).getReg();
11644std::unique_ptr<TargetInstrInfo::PipelinerLoopInfo>
11655 if (
MI.isCall() ||
MI.hasUnmodeledSideEffects())
11666 if (
TBB == LoopBB && FBB == LoopBB)
11670 if (
TBB != LoopBB && FBB ==
nullptr)
11673 assert((
TBB == LoopBB || FBB == LoopBB) &&
11674 "The Loop must be a single-basic-block loop");
11679 if (CondBranch->
getOpcode() != AArch64::Bcc)
11687 unsigned CompCounterOprNum = 0;
11689 if (
MI.modifiesRegister(AArch64::NZCV, &
TRI)) {
11693 switch (
MI.getOpcode()) {
11694 case AArch64::SUBSXri:
11695 case AArch64::SUBSWri:
11696 case AArch64::ADDSXri:
11697 case AArch64::ADDSWri:
11699 CompCounterOprNum = 1;
11701 case AArch64::ADDSWrr:
11702 case AArch64::ADDSXrr:
11703 case AArch64::SUBSWrr:
11704 case AArch64::SUBSXrr:
11708 if (isWhileOpcode(
MI.getOpcode())) {
11715 if (CompCounterOprNum == 0) {
11717 CompCounterOprNum = 2;
11719 CompCounterOprNum = 1;
11731 bool IsUpdatePriorComp;
11732 unsigned UpdateCounterOprNum;
11734 Update, UpdateCounterOprNum,
Init, IsUpdatePriorComp))
11737 return std::make_unique<AArch64PipelinerLoopInfo>(
11738 LoopBB, CondBranch, Comp, CompCounterOprNum, Update, UpdateCounterOprNum,
11748 TypeSize Scale(0U,
false), Width(0U,
false);
11749 int64_t MinOffset, MaxOffset;
11750 if (
getMemOpInfo(
MI.getOpcode(), Scale, Width, MinOffset, MaxOffset)) {
11752 if (
MI.getOperand(ImmIdx).isImm() && !
MI.getOperand(ImmIdx - 1).isFI()) {
11753 int64_t Imm =
MI.getOperand(ImmIdx).getImm();
11754 if (Imm < MinOffset || Imm > MaxOffset) {
11755 ErrInfo =
"Unexpected immediate on load/store instruction";
11761 const MCInstrDesc &MCID =
MI.getDesc();
11763 const MachineOperand &MO =
MI.getOperand(
Op);
11767 ErrInfo =
"OPERAND_IMPLICIT_IMM_0 should be 0";
11776 ErrInfo =
"OPERAND_SHIFT_MSL should be msl shift of 8 or 16";
11787#define GET_INSTRINFO_HELPERS
11788#define GET_INSTRMAP_INFO
11789#include "AArch64GenInstrInfo.inc"
unsigned const MachineRegisterInfo * MRI
MachineInstrBuilder & UseMI
MachineInstrBuilder MachineInstrBuilder & DefMI
static bool forwardCopyWillClobberTuple(unsigned DestReg, unsigned SrcReg, unsigned NumRegs)
static cl::opt< unsigned > BCCDisplacementBits("aarch64-bcc-offset-bits", cl::Hidden, cl::init(19), cl::desc("Restrict range of Bcc instructions (DEBUG)"))
static Register genNeg(MachineFunction &MF, MachineRegisterInfo &MRI, const TargetInstrInfo *TII, MachineInstr &Root, SmallVectorImpl< MachineInstr * > &InsInstrs, DenseMap< Register, unsigned > &InstrIdxForVirtReg, unsigned MnegOpc, const TargetRegisterClass *RC)
genNeg - Helper to generate an intermediate negation of the second operand of Root
static bool isFrameStoreOpcode(int Opcode)
static cl::opt< unsigned > GatherOptSearchLimit("aarch64-search-limit", cl::Hidden, cl::init(2048), cl::desc("Restrict range of instructions to search for the " "machine-combiner gather pattern optimization"))
static bool getMaddPatterns(MachineInstr &Root, SmallVectorImpl< unsigned > &Patterns)
Find instructions that can be turned into madd.
static AArch64CC::CondCode findCondCodeUsedByInstr(const MachineInstr &Instr)
Find a condition code used by the instruction.
static MachineInstr * genFusedMultiplyAcc(MachineFunction &MF, MachineRegisterInfo &MRI, const TargetInstrInfo *TII, MachineInstr &Root, SmallVectorImpl< MachineInstr * > &InsInstrs, unsigned IdxMulOpd, unsigned MaddOpc, const TargetRegisterClass *RC)
genFusedMultiplyAcc - Helper to generate fused multiply accumulate instructions.
static MachineInstr * genFusedMultiplyAccNeg(MachineFunction &MF, MachineRegisterInfo &MRI, const TargetInstrInfo *TII, MachineInstr &Root, SmallVectorImpl< MachineInstr * > &InsInstrs, DenseMap< Register, unsigned > &InstrIdxForVirtReg, unsigned IdxMulOpd, unsigned MaddOpc, unsigned MnegOpc, const TargetRegisterClass *RC)
genFusedMultiplyAccNeg - Helper to generate fused multiply accumulate instructions with an additional...
static bool isCombineInstrCandidate64(unsigned Opc)
static bool isFrameLoadOpcode(int Opcode)
static unsigned removeCopies(const MachineRegisterInfo &MRI, unsigned VReg)
static bool areCFlagsAccessedBetweenInstrs(MachineBasicBlock::iterator From, MachineBasicBlock::iterator To, const TargetRegisterInfo *TRI, const AccessKind AccessToCheck=AK_All)
True when condition flags are accessed (either by writing or reading) on the instruction trace starti...
static bool getFMAPatterns(MachineInstr &Root, SmallVectorImpl< unsigned > &Patterns)
Floating-Point Support.
static bool isADDSRegImm(unsigned Opcode)
static bool isCheapCopy(const MachineInstr &MI, const AArch64RegisterInfo &RI)
static bool isANDOpcode(MachineInstr &MI)
static void appendOffsetComment(int NumBytes, llvm::raw_string_ostream &Comment, StringRef RegScale={})
static unsigned sForm(MachineInstr &Instr)
Get opcode of S version of Instr.
static bool isCombineInstrSettingFlag(unsigned Opc)
static bool getFNEGPatterns(MachineInstr &Root, SmallVectorImpl< unsigned > &Patterns)
static bool getIndVarInfo(Register Reg, const MachineBasicBlock *LoopBB, MachineInstr *&UpdateInst, unsigned &UpdateCounterOprNum, Register &InitReg, bool &IsUpdatePriorComp)
If Reg is an induction variable, return true and set some parameters.
static bool canPairLdStOpc(unsigned FirstOpc, unsigned SecondOpc)
static bool mustAvoidNeonAtMBBI(const AArch64Subtarget &Subtarget, MachineBasicBlock &MBB, MachineBasicBlock::iterator I)
Returns true if in a streaming call site region without SME-FA64.
static int findCondCodeUseOperandIdxForBranchOrSelect(const MachineInstr &Instr)
static bool isPostIndexLdStOpcode(unsigned Opcode)
Return true if the opcode is a post-index ld/st instruction, which really loads from base+0.
static unsigned getBranchDisplacementBits(unsigned Opc)
static cl::opt< unsigned > CBDisplacementBits("aarch64-cb-offset-bits", cl::Hidden, cl::init(9), cl::desc("Restrict range of CB instructions (DEBUG)"))
static std::optional< ParamLoadedValue > describeORRLoadedValue(const MachineInstr &MI, Register DescribedReg, const TargetInstrInfo *TII, const TargetRegisterInfo *TRI)
If the given ORR instruction is a copy, and DescribedReg overlaps with the destination register then,...
static bool getFMULPatterns(MachineInstr &Root, SmallVectorImpl< unsigned > &Patterns)
static void appendReadRegExpr(SmallVectorImpl< char > &Expr, unsigned RegNum)
static MachineInstr * genMaddR(MachineFunction &MF, MachineRegisterInfo &MRI, const TargetInstrInfo *TII, MachineInstr &Root, SmallVectorImpl< MachineInstr * > &InsInstrs, unsigned IdxMulOpd, unsigned MaddOpc, unsigned VR, const TargetRegisterClass *RC)
genMaddR - Generate madd instruction and combine mul and add using an extra virtual register Example ...
static Register cloneInstr(const MachineInstr *MI, unsigned ReplaceOprNum, Register ReplaceReg, MachineBasicBlock &MBB, MachineBasicBlock::iterator InsertTo)
Clone an instruction from MI.
static bool scaleOffset(unsigned Opc, int64_t &Offset)
static bool canCombineWithFMUL(MachineBasicBlock &MBB, MachineOperand &MO, unsigned MulOpc)
unsigned scaledOffsetOpcode(unsigned Opcode, unsigned &Scale)
static MachineInstr * genFusedMultiplyIdx(MachineFunction &MF, MachineRegisterInfo &MRI, const TargetInstrInfo *TII, MachineInstr &Root, SmallVectorImpl< MachineInstr * > &InsInstrs, unsigned IdxMulOpd, unsigned MaddOpc, const TargetRegisterClass *RC)
genFusedMultiplyIdx - Helper to generate fused multiply accumulate instructions.
static MachineInstr * genIndexedMultiply(MachineInstr &Root, SmallVectorImpl< MachineInstr * > &InsInstrs, unsigned IdxDupOp, unsigned MulOpc, const TargetRegisterClass *RC, MachineRegisterInfo &MRI)
Fold (FMUL x (DUP y lane)) into (FMUL_indexed x y lane)
static bool isSUBSRegImm(unsigned Opcode)
static bool UpdateOperandRegClass(MachineInstr &Instr)
static const TargetRegisterClass * getRegClass(const MachineInstr &MI, Register Reg)
static bool isInStreamingCallSiteRegion(MachineBasicBlock &MBB, MachineBasicBlock::iterator I)
Returns true if the instruction at I is in a streaming call site region, within a single basic block.
static bool canCmpInstrBeRemoved(MachineInstr &MI, MachineInstr &CmpInstr, int CmpValue, const TargetRegisterInfo &TRI, SmallVectorImpl< MachineInstr * > &CCUseInstrs, bool &IsInvertCC)
unsigned unscaledOffsetOpcode(unsigned Opcode)
static bool getLoadPatterns(MachineInstr &Root, SmallVectorImpl< unsigned > &Patterns)
Search for patterns of LD instructions we can optimize.
static bool canInstrSubstituteCmpInstr(MachineInstr &MI, MachineInstr &CmpInstr, const TargetRegisterInfo &TRI)
Check if CmpInstr can be substituted by MI.
static UsedNZCV getUsedNZCV(AArch64CC::CondCode CC)
static bool isCombineInstrCandidateFP(const MachineInstr &Inst)
static void appendLoadRegExpr(SmallVectorImpl< char > &Expr, int64_t OffsetFromDefCFA)
static void appendConstantExpr(SmallVectorImpl< char > &Expr, int64_t Constant, dwarf::LocationAtom Operation)
static unsigned convertToNonFlagSettingOpc(const MachineInstr &MI)
Return the opcode that does not set flags when possible - otherwise return the original opcode.
static bool outliningCandidatesV8_3OpsConsensus(const outliner::Candidate &a, const outliner::Candidate &b)
static bool isCombineInstrCandidate32(unsigned Opc)
static void parseCondBranch(MachineInstr *LastInst, MachineBasicBlock *&Target, SmallVectorImpl< MachineOperand > &Cond)
static unsigned offsetExtendOpcode(unsigned Opcode)
static void loadRegPairFromStackSlot(const TargetRegisterInfo &TRI, MachineBasicBlock &MBB, MachineBasicBlock::iterator InsertBefore, const MCInstrDesc &MCID, Register DestReg, unsigned SubIdx0, unsigned SubIdx1, int FI, MachineMemOperand *MMO)
static void generateGatherLanePattern(MachineInstr &Root, SmallVectorImpl< MachineInstr * > &InsInstrs, SmallVectorImpl< MachineInstr * > &DelInstrs, DenseMap< Register, unsigned > &InstrIdxForVirtReg, unsigned Pattern, unsigned NumLanes)
Generate optimized instruction sequence for gather load patterns to improve Memory-Level Parallelism ...
static bool getMiscPatterns(MachineInstr &Root, SmallVectorImpl< unsigned > &Patterns)
Find other MI combine patterns.
static bool outliningCandidatesSigningKeyConsensus(const outliner::Candidate &a, const outliner::Candidate &b)
static const MachineInstrBuilder & AddSubReg(const MachineInstrBuilder &MIB, MCRegister Reg, unsigned SubIdx, RegState State, const TargetRegisterInfo *TRI)
static bool outliningCandidatesSigningScopeConsensus(const outliner::Candidate &a, const outliner::Candidate &b)
static bool shouldClusterFI(const MachineFrameInfo &MFI, int FI1, int64_t Offset1, unsigned Opcode1, int FI2, int64_t Offset2, unsigned Opcode2)
static cl::opt< unsigned > TBZDisplacementBits("aarch64-tbz-offset-bits", cl::Hidden, cl::init(14), cl::desc("Restrict range of TB[N]Z instructions (DEBUG)"))
static void extractPhiReg(const MachineInstr &Phi, const MachineBasicBlock *MBB, Register &RegMBB, Register &RegOther)
static MCCFIInstruction createDefCFAExpression(const TargetRegisterInfo &TRI, unsigned Reg, const StackOffset &Offset)
static bool isDefinedOutside(Register Reg, const MachineBasicBlock *BB)
static MachineInstr * genFusedMultiply(MachineFunction &MF, MachineRegisterInfo &MRI, const TargetInstrInfo *TII, MachineInstr &Root, SmallVectorImpl< MachineInstr * > &InsInstrs, unsigned IdxMulOpd, unsigned MaddOpc, const TargetRegisterClass *RC, FMAInstKind kind=FMAInstKind::Default, const Register *ReplacedAddend=nullptr)
genFusedMultiply - Generate fused multiply instructions.
static bool getGatherLanePattern(MachineInstr &Root, SmallVectorImpl< unsigned > &Patterns, unsigned LoadLaneOpCode, unsigned NumLanes)
Check if the given instruction forms a gather load pattern that can be optimized for better Memory-Le...
static MachineInstr * genFusedMultiplyIdxNeg(MachineFunction &MF, MachineRegisterInfo &MRI, const TargetInstrInfo *TII, MachineInstr &Root, SmallVectorImpl< MachineInstr * > &InsInstrs, DenseMap< Register, unsigned > &InstrIdxForVirtReg, unsigned IdxMulOpd, unsigned MaddOpc, unsigned MnegOpc, const TargetRegisterClass *RC)
genFusedMultiplyAccNeg - Helper to generate fused multiply accumulate instructions with an additional...
static bool isCombineInstrCandidate(unsigned Opc)
static unsigned regOffsetOpcode(unsigned Opcode)
MachineOutlinerClass
Constants defining how certain sequences should be outlined.
@ MachineOutlinerTailCall
Emit a save, restore, call, and return.
@ MachineOutlinerRegSave
Emit a call and tail-call.
@ MachineOutlinerNoLRSave
Only emit a branch.
@ MachineOutlinerThunk
Emit a call and return.
static cl::opt< unsigned > BDisplacementBits("aarch64-b-offset-bits", cl::Hidden, cl::init(26), cl::desc("Restrict range of B instructions (DEBUG)"))
static bool areCFlagsAliveInSuccessors(const MachineBasicBlock *MBB)
Check if AArch64::NZCV should be alive in successors of MBB.
static void emitFrameOffsetAdj(MachineBasicBlock &MBB, MachineBasicBlock::iterator MBBI, const DebugLoc &DL, unsigned DestReg, unsigned SrcReg, int64_t Offset, unsigned Opc, const TargetInstrInfo *TII, MachineInstr::MIFlag Flag, bool NeedsWinCFI, bool *HasWinCFI, bool EmitCFAOffset, StackOffset CFAOffset, unsigned FrameReg)
static bool isCheapImmediate(const MachineInstr &MI, unsigned BitSize)
static cl::opt< unsigned > CBZDisplacementBits("aarch64-cbz-offset-bits", cl::Hidden, cl::init(19), cl::desc("Restrict range of CB[N]Z instructions (DEBUG)"))
static void genSubAdd2SubSub(MachineFunction &MF, MachineRegisterInfo &MRI, const TargetInstrInfo *TII, MachineInstr &Root, SmallVectorImpl< MachineInstr * > &InsInstrs, SmallVectorImpl< MachineInstr * > &DelInstrs, unsigned IdxOpd1, DenseMap< Register, unsigned > &InstrIdxForVirtReg)
Do the following transformation A - (B + C) ==> (A - B) - C A - (B + C) ==> (A - C) - B.
static unsigned canFoldIntoCSel(const MachineRegisterInfo &MRI, unsigned VReg, unsigned *NewReg=nullptr)
static void signOutlinedFunction(MachineFunction &MF, MachineBasicBlock &MBB, const AArch64InstrInfo *TII, bool ShouldSignReturnAddr)
static MachineInstr * genFNegatedMAD(MachineFunction &MF, MachineRegisterInfo &MRI, const TargetInstrInfo *TII, MachineInstr &Root, SmallVectorImpl< MachineInstr * > &InsInstrs)
static bool canCombineWithMUL(MachineBasicBlock &MBB, MachineOperand &MO, unsigned MulOpc, unsigned ZeroReg)
static void storeRegPairToStackSlot(const TargetRegisterInfo &TRI, MachineBasicBlock &MBB, MachineBasicBlock::iterator InsertBefore, const MCInstrDesc &MCID, Register SrcReg, bool IsKill, unsigned SubIdx0, unsigned SubIdx1, int FI, MachineMemOperand *MMO)
assert(UImm &&(UImm !=~static_cast< T >(0)) &&"Invalid immediate!")
MachineBasicBlock MachineBasicBlock::iterator DebugLoc DL
MachineBasicBlock MachineBasicBlock::iterator MBBI
static const Function * getParent(const Value *V)
static GCRegistry::Add< ErlangGC > A("erlang", "erlang-compatible garbage collector")
static GCRegistry::Add< OcamlGC > B("ocaml", "ocaml 3.10-compatible GC")
DXIL Forward Handle Accesses
const HexagonInstrInfo * TII
Module.h This file contains the declarations for the Module class.
const AbstractManglingParser< Derived, Alloc >::OperatorInfo AbstractManglingParser< Derived, Alloc >::Ops[]
This file implements the LivePhysRegs utility for tracking liveness of physical registers.
Machine Check Debug Module
Register const TargetRegisterInfo * TRI
Promote Memory to Register
static MCRegister getReg(const MCDisassembler *D, unsigned RC, unsigned RegNo)
MachineInstr unsigned OpIdx
ConstantRange Range(APInt(BitWidth, Low), APInt(BitWidth, High))
PowerPC Reduce CR logical Operation
const SmallVectorImpl< MachineOperand > MachineBasicBlock * TBB
const SmallVectorImpl< MachineOperand > & Cond
This file declares the machine register scavenger class.
static bool contains(SmallPtrSetImpl< ConstantExpr * > &Cache, ConstantExpr *Expr, Constant *C)
This file defines the SmallSet class.
This file defines the SmallVector class.
This file defines the 'Statistic' class, which is designed to be an easy way to expose various metric...
#define STATISTIC(VARNAME, DESC)
#define DEBUG_WITH_TYPE(TYPE,...)
DEBUG_WITH_TYPE macro - This macro should be used by passes to emit debug information.
static bool canCombine(MachineBasicBlock &MBB, MachineOperand &MO, unsigned CombineOpc=0)
AArch64FunctionInfo - This class is derived from MachineFunctionInfo and contains private AArch64-spe...
SignReturnAddress getSignReturnAddressCondition() const
bool hasStreamingModeChanges() const
void setOutliningStyle(const std::string &Style)
std::optional< bool > hasRedZone() const
static bool shouldSignReturnAddress(SignReturnAddress Condition, bool IsLRSpilled)
bool shouldSignWithBKey() const
static bool isHForm(const MachineInstr &MI)
Returns whether the instruction is in H form (16 bit operands)
void insertSelect(MachineBasicBlock &MBB, MachineBasicBlock::iterator MI, const DebugLoc &DL, Register DstReg, ArrayRef< MachineOperand > Cond, Register TrueReg, Register FalseReg) const override
static bool hasBTISemantics(const MachineInstr &MI)
Returns whether the instruction can be compatible with non-zero BTYPE.
static bool isQForm(const MachineInstr &MI)
Returns whether the instruction is in Q form (128 bit operands)
static bool getMemOpInfo(unsigned Opcode, TypeSize &Scale, TypeSize &Width, int64_t &MinOffset, int64_t &MaxOffset)
Returns true if opcode Opc is a memory operation.
static bool isTailCallReturnInst(const MachineInstr &MI)
Returns true if MI is one of the TCRETURN* instructions.
static bool isFPRCopy(const MachineInstr &MI)
Does this instruction rename an FPR without modifying bits?
MachineInstr * emitLdStWithAddr(MachineInstr &MemI, const ExtAddrMode &AM) const override
std::optional< DestSourcePair > isCopyInstrImpl(const MachineInstr &MI) const override
If the specific machine instruction is an instruction that moves/copies value from one register to an...
MachineBasicBlock * getBranchDestBlock(const MachineInstr &MI) const override
unsigned getInstSizeInBytes(const MachineInstr &MI) const override
GetInstSize - Return the number of bytes of code the specified instruction may be.
bool areMemAccessesTriviallyDisjoint(const MachineInstr &MIa, const MachineInstr &MIb) const override
void copyPhysReg(MachineBasicBlock &MBB, MachineBasicBlock::iterator I, const DebugLoc &DL, Register DestReg, Register SrcReg, bool KillSrc, bool RenamableDest=false, bool RenamableSrc=false) const override
static bool isGPRCopy(const MachineInstr &MI)
Does this instruction rename a GPR without modifying bits?
static unsigned convertToFlagSettingOpc(unsigned Opc)
Return the opcode that set flags when possible.
bool isBranchOffsetInRange(unsigned BranchOpc, int64_t BrOffset) const override
bool canInsertSelect(const MachineBasicBlock &, ArrayRef< MachineOperand > Cond, Register, Register, Register, int &, int &, int &) const override
Register isLoadFromStackSlotPostFE(const MachineInstr &MI, int &FrameIndex) const override
Check for post-frame ptr elimination stack locations as well.
static const MachineOperand & getLdStOffsetOp(const MachineInstr &MI)
Returns the immediate offset operator of a load/store.
bool isCoalescableExtInstr(const MachineInstr &MI, Register &SrcReg, Register &DstReg, unsigned &SubIdx) const override
static std::optional< unsigned > getUnscaledLdSt(unsigned Opc)
Returns the unscaled load/store for the scaled load/store opcode, if there is a corresponding unscale...
static bool hasUnscaledLdStOffset(unsigned Opc)
Return true if it has an unscaled load/store offset.
static const MachineOperand & getLdStAmountOp(const MachineInstr &MI)
Returns the shift amount operator of a load/store.
static bool isPreLdSt(const MachineInstr &MI)
Returns whether the instruction is a pre-indexed load/store.
std::optional< ExtAddrMode > getAddrModeFromMemoryOp(const MachineInstr &MemI, const TargetRegisterInfo *TRI) const override
bool getMemOperandsWithOffsetWidth(const MachineInstr &MI, SmallVectorImpl< const MachineOperand * > &BaseOps, int64_t &Offset, bool &OffsetIsScalable, LocationSize &Width, const TargetRegisterInfo *TRI) const override
bool analyzeBranchPredicate(MachineBasicBlock &MBB, MachineBranchPredicate &MBP, bool AllowModify) const override
void insertIndirectBranch(MachineBasicBlock &MBB, MachineBasicBlock &NewDestBB, MachineBasicBlock &RestoreBB, const DebugLoc &DL, int64_t BrOffset, RegScavenger *RS) const override
void loadRegFromStackSlot(MachineBasicBlock &MBB, MachineBasicBlock::iterator MBBI, Register DestReg, int FrameIndex, const TargetRegisterClass *RC, Register VReg, unsigned SubReg=0, MachineInstr::MIFlag Flags=MachineInstr::NoFlags) const override
static bool isPairableLdStInst(const MachineInstr &MI)
Return true if pairing the given load or store may be paired with another.
void storeRegToStackSlot(MachineBasicBlock &MBB, MachineBasicBlock::iterator MBBI, Register SrcReg, bool isKill, int FrameIndex, const TargetRegisterClass *RC, Register VReg, MachineInstr::MIFlag Flags=MachineInstr::NoFlags) const override
const AArch64RegisterInfo & getRegisterInfo() const
getRegisterInfo - TargetInstrInfo is a superset of MRegister info.
static bool isPreSt(const MachineInstr &MI)
Returns whether the instruction is a pre-indexed store.
MachineInstr * foldMemoryOperandImpl(MachineFunction &MF, MachineInstr &MI, ArrayRef< unsigned > Ops, MachineBasicBlock::iterator InsertPt, int FrameIndex, LiveIntervals *LIS=nullptr, VirtRegMap *VRM=nullptr) const override
void insertNoop(MachineBasicBlock &MBB, MachineBasicBlock::iterator MI) const override
AArch64InstrInfo(const AArch64Subtarget &STI)
static bool isPairedLdSt(const MachineInstr &MI)
Returns whether the instruction is a paired load/store.
bool getMemOperandWithOffsetWidth(const MachineInstr &MI, const MachineOperand *&BaseOp, int64_t &Offset, bool &OffsetIsScalable, TypeSize &Width, const TargetRegisterInfo *TRI) const
If OffsetIsScalable is set to 'true', the offset is scaled by vscale.
Register isLoadFromStackSlot(const MachineInstr &MI, int &FrameIndex) const override
bool reverseBranchCondition(SmallVectorImpl< MachineOperand > &Cond) const override
static bool isStridedAccess(const MachineInstr &MI)
Return true if the given load or store is a strided memory access.
bool shouldClusterMemOps(ArrayRef< const MachineOperand * > BaseOps1, int64_t Offset1, bool OffsetIsScalable1, ArrayRef< const MachineOperand * > BaseOps2, int64_t Offset2, bool OffsetIsScalable2, unsigned ClusterSize, unsigned NumBytes) const override
Detect opportunities for ldp/stp formation.
unsigned removeBranch(MachineBasicBlock &MBB, int *BytesRemoved=nullptr) const override
bool isThroughputPattern(unsigned Pattern) const override
Return true when a code sequence can improve throughput.
MachineOperand & getMemOpBaseRegImmOfsOffsetOperand(MachineInstr &LdSt) const
Return the immediate offset of the base register in a load/store LdSt.
bool analyzeBranch(MachineBasicBlock &MBB, MachineBasicBlock *&TBB, MachineBasicBlock *&FBB, SmallVectorImpl< MachineOperand > &Cond, bool AllowModify=false) const override
bool canFoldIntoAddrMode(const MachineInstr &MemI, Register Reg, const MachineInstr &AddrI, ExtAddrMode &AM) const override
static bool isLdStPairSuppressed(const MachineInstr &MI)
Return true if pairing the given load or store is hinted to be unprofitable.
Register isStoreToStackSlotPostFE(const MachineInstr &MI, int &FrameIndex) const override
Check for post-frame ptr elimination stack locations as well.
std::unique_ptr< TargetInstrInfo::PipelinerLoopInfo > analyzeLoopForPipelining(MachineBasicBlock *LoopBB) const override
bool isSchedulingBoundary(const MachineInstr &MI, const MachineBasicBlock *MBB, const MachineFunction &MF) const override
unsigned insertBranch(MachineBasicBlock &MBB, MachineBasicBlock *TBB, MachineBasicBlock *FBB, ArrayRef< MachineOperand > Cond, const DebugLoc &DL, int *BytesAdded=nullptr) const override
MachineBasicBlock::iterator probedStackAlloc(MachineBasicBlock::iterator MBBI, Register TargetReg, bool FrameSetup) const
Return true when there is potentially a faster code sequence for an instruction chain ending in Root.
bool optimizeCompareInstr(MachineInstr &CmpInstr, Register SrcReg, Register SrcReg2, int64_t CmpMask, int64_t CmpValue, const MachineRegisterInfo *MRI) const override
optimizeCompareInstr - Convert the instruction supplying the argument to the comparison into one that...
static unsigned getLoadStoreImmIdx(unsigned Opc)
Returns the index for the immediate for a given instruction.
static bool isGPRZero(const MachineInstr &MI)
Does this instruction set its full destination register to zero?
void copyGPRRegTuple(MachineBasicBlock &MBB, MachineBasicBlock::iterator I, const DebugLoc &DL, MCRegister DestReg, MCRegister SrcReg, bool KillSrc, unsigned Opcode, unsigned ZeroReg, llvm::ArrayRef< unsigned > Indices) const
bool analyzeCompare(const MachineInstr &MI, Register &SrcReg, Register &SrcReg2, int64_t &CmpMask, int64_t &CmpValue) const override
analyzeCompare - For a comparison instruction, return the source registers in SrcReg and SrcReg2,...
CombinerObjective getCombinerObjective(unsigned Pattern) const override
static bool isFpOrNEON(Register Reg)
Returns whether the physical register is FP or NEON.
bool isAsCheapAsAMove(const MachineInstr &MI) const override
std::optional< DestSourcePair > isCopyLikeInstrImpl(const MachineInstr &MI) const override
static void suppressLdStPair(MachineInstr &MI)
Hint that pairing the given load or store is unprofitable.
Register isStoreToStackSlot(const MachineInstr &MI, int &FrameIndex) const override
static bool isPreLd(const MachineInstr &MI)
Returns whether the instruction is a pre-indexed load.
void copyPhysRegTuple(MachineBasicBlock &MBB, MachineBasicBlock::iterator I, const DebugLoc &DL, MCRegister DestReg, MCRegister SrcReg, bool KillSrc, unsigned Opcode, llvm::ArrayRef< unsigned > Indices) const
bool optimizeCondBranch(MachineInstr &MI) const override
Replace csincr-branch sequence by simple conditional branch.
static int getMemScale(unsigned Opc)
Scaling factor for (scaled or unscaled) load or store.
bool isCandidateToMergeOrPair(const MachineInstr &MI) const
Return true if this is a load/store that can be potentially paired/merged.
MCInst getNop() const override
static const MachineOperand & getLdStBaseOp(const MachineInstr &MI)
Returns the base register operator of a load/store.
bool isReservedReg(const MachineFunction &MF, MCRegister Reg) const
const AArch64RegisterInfo * getRegisterInfo() const override
bool isNeonAvailable() const
Returns true if the target has NEON and the function at runtime is known to have NEON enabled (e....
bool isSVEorStreamingSVEAvailable() const
Returns true if the target has access to either the full range of SVE instructions,...
ArrayRef - Represent a constant reference to an array (0 or more elements consecutively in memory),...
const T & front() const
front - Get the first element.
size_t size() const
size - Get the array size.
This is an important base class in LLVM.
std::pair< iterator, bool > insert(const std::pair< KeyT, ValueT > &KV)
bool hasOptSize() const
Optimize this function for size (-Os) or minimum size (-Oz).
bool hasMinSize() const
Optimize this function for minimum size (-Oz).
Module * getParent()
Get the module that this global value is contained inside of...
LLVM_ABI InstListType::iterator eraseFromParent()
This method unlinks 'this' from the containing basic block and deletes it.
unsigned getOpcode() const
Returns a member of one of the enums like Instruction::Add.
static LocationSize precise(uint64_t Value)
This class is intended to be used as a base class for asm properties and features specific to the tar...
bool usesWindowsCFI() const
static MCCFIInstruction cfiDefCfa(MCSymbol *L, unsigned Register, int64_t Offset, SMLoc Loc={})
.cfi_def_cfa defines a rule for computing CFA as: take address from Register and add Offset to it.
static MCCFIInstruction createOffset(MCSymbol *L, unsigned Register, int64_t Offset, SMLoc Loc={})
.cfi_offset Previous value of Register is saved at offset Offset from CFA.
static MCCFIInstruction cfiDefCfaOffset(MCSymbol *L, int64_t Offset, SMLoc Loc={})
.cfi_def_cfa_offset modifies a rule for computing CFA.
static MCCFIInstruction createEscape(MCSymbol *L, StringRef Vals, SMLoc Loc={}, StringRef Comment="")
.cfi_escape Allows the user to add arbitrary bytes to the unwind info.
Instances of this class represent a single low-level machine instruction.
Describe properties that are true of each instruction in the target description file.
unsigned getNumOperands() const
Return the number of declared MachineOperands for this MachineInstruction.
ArrayRef< MCOperandInfo > operands() const
Wrapper class representing physical registers. Should be passed by value.
constexpr bool isValid() const
static constexpr unsigned NoRegister
static MDTuple * get(LLVMContext &Context, ArrayRef< Metadata * > MDs)
bool isInlineAsmBrIndirectTarget() const
Returns true if this is the indirect dest of an INLINEASM_BR.
unsigned pred_size() const
LLVM_ABI void transferSuccessorsAndUpdatePHIs(MachineBasicBlock *FromMBB)
Transfers all the successors, as in transferSuccessors, and update PHI operands in the successor bloc...
LLVM_ABI instr_iterator insert(instr_iterator I, MachineInstr *M)
Insert MI into the instruction list before I, possibly inside a bundle.
reverse_instr_iterator instr_rbegin()
LLVM_ABI iterator getFirstTerminator()
Returns an iterator to the first terminator instruction of this basic block.
LLVM_ABI void addSuccessor(MachineBasicBlock *Succ, BranchProbability Prob=BranchProbability::getUnknown())
Add Succ as a successor of this MachineBasicBlock.
reverse_instr_iterator instr_rend()
Instructions::iterator instr_iterator
instr_iterator instr_end()
Instructions::const_iterator const_instr_iterator
void addLiveIn(MCRegister PhysReg, LaneBitmask LaneMask=LaneBitmask::getAll())
Adds the specified register as a live in.
const MachineFunction * getParent() const
Return the MachineFunction containing this basic block.
LLVM_ABI instr_iterator erase(instr_iterator I)
Remove an instruction from the instruction list and delete it.
void splice(iterator Where, MachineBasicBlock *Other, iterator From)
Take an instruction from MBB 'Other' at the position From, and insert it into this MBB right before '...
MachineInstrBundleIterator< MachineInstr > iterator
void setMachineBlockAddressTaken()
Set this block to indicate that its address is used as something other than the target of a terminato...
LLVM_ABI bool isLiveIn(MCRegister Reg, LaneBitmask LaneMask=LaneBitmask::getAll()) const
Return true if the specified register is in the live in set.
The MachineFrameInfo class represents an abstract stack frame until prolog/epilog code is inserted.
uint64_t getStackSize() const
Return the number of bytes that must be allocated to hold all of the fixed size frame objects.
void setStackID(int ObjectIdx, uint8_t ID)
bool isCalleeSavedInfoValid() const
Has the callee saved info been calculated yet?
Align getObjectAlign(int ObjectIdx) const
Return the alignment of the specified stack object.
int64_t getObjectSize(int ObjectIdx) const
Return the size of the specified object.
unsigned getNumObjects() const
Return the number of objects.
int64_t getObjectOffset(int ObjectIdx) const
Return the assigned stack offset of the specified object from the incoming stack pointer.
bool isFixedObjectIndex(int ObjectIdx) const
Returns true if the specified index corresponds to a fixed stack object.
unsigned addFrameInst(const MCCFIInstruction &Inst)
const TargetSubtargetInfo & getSubtarget() const
getSubtarget - Return the subtarget for which this machine code is being compiled.
StringRef getName() const
getName - Return the name of the corresponding LLVM function.
MachineMemOperand * getMachineMemOperand(MachinePointerInfo PtrInfo, MachineMemOperand::Flags f, LLT MemTy, Align base_alignment, const AAMDNodes &AAInfo=AAMDNodes(), const MDNode *Ranges=nullptr, SyncScope::ID SSID=SyncScope::System, AtomicOrdering Ordering=AtomicOrdering::NotAtomic, AtomicOrdering FailureOrdering=AtomicOrdering::NotAtomic)
getMachineMemOperand - Allocate a new MachineMemOperand.
MachineFrameInfo & getFrameInfo()
getFrameInfo - Return the frame info object for the current function.
MachineRegisterInfo & getRegInfo()
getRegInfo - Return information about the registers currently in use.
Function & getFunction()
Return the LLVM function that this machine code represents.
BasicBlockListType::iterator iterator
Ty * getInfo()
getInfo - Keep track of various per-function pieces of information for backends that would like to do...
const MachineJumpTableInfo * getJumpTableInfo() const
getJumpTableInfo - Return the jump table info object for the current function.
MachineBasicBlock * CreateMachineBasicBlock(const BasicBlock *BB=nullptr, std::optional< UniqueBBID > BBID=std::nullopt)
CreateMachineInstr - Allocate a new MachineInstr.
void insert(iterator MBBI, MachineBasicBlock *MBB)
const TargetMachine & getTarget() const
getTarget - Return the target machine this machine code is compiled with
const MachineInstrBuilder & setMemRefs(ArrayRef< MachineMemOperand * > MMOs) const
const MachineInstrBuilder & addCFIIndex(unsigned CFIIndex) const
const MachineInstrBuilder & addUse(Register RegNo, RegState Flags={}, unsigned SubReg=0) const
Add a virtual register use operand.
const MachineInstrBuilder & addReg(Register RegNo, RegState Flags={}, unsigned SubReg=0) const
Add a new virtual register operand.
const MachineInstrBuilder & setMIFlag(MachineInstr::MIFlag Flag) const
const MachineInstrBuilder & addImm(int64_t Val) const
Add a new immediate operand.
const MachineInstrBuilder & add(const MachineOperand &MO) const
const MachineInstrBuilder & addSym(MCSymbol *Sym, unsigned char TargetFlags=0) const
const MachineInstrBuilder & addFrameIndex(int Idx) const
const MachineInstrBuilder & addGlobalAddress(const GlobalValue *GV, int64_t Offset=0, unsigned TargetFlags=0) const
const MachineInstrBuilder & addMBB(MachineBasicBlock *MBB, unsigned TargetFlags=0) const
const MachineInstrBuilder & addDef(Register RegNo, RegState Flags={}, unsigned SubReg=0) const
Add a virtual register definition operand.
const MachineInstrBuilder & setMIFlags(unsigned Flags) const
const MachineInstrBuilder & addMemOperand(MachineMemOperand *MMO) const
reverse_iterator getReverse() const
Get a reverse iterator to the same node.
Representation of each machine instruction.
unsigned getOpcode() const
Returns the opcode of this MachineInstr.
bool mayLoadOrStore(QueryType Type=AnyInBundle) const
Return true if this instruction could possibly read or modify memory.
const MachineBasicBlock * getParent() const
bool isCall(QueryType Type=AnyInBundle) const
bool getFlag(MIFlag Flag) const
Return whether an MI flag is set.
LLVM_ABI uint32_t mergeFlagsWith(const MachineInstr &Other) const
Return the MIFlags which represent both MachineInstrs.
unsigned getNumOperands() const
Retuns the total number of operands.
LLVM_ABI unsigned getNumExplicitOperands() const
Returns the number of non-implicit operands.
bool mayLoad(QueryType Type=AnyInBundle) const
Return true if this instruction could possibly read memory.
const MCInstrDesc & getDesc() const
Returns the target instruction descriptor of this MachineInstr.
LLVM_ABI bool hasUnmodeledSideEffects() const
Return true if this instruction has side effects that are not modeled by mayLoad / mayStore,...
bool registerDefIsDead(Register Reg, const TargetRegisterInfo *TRI) const
Returns true if the register is dead in this machine instruction.
bool definesRegister(Register Reg, const TargetRegisterInfo *TRI) const
Return true if the MachineInstr fully defines the specified register.
LLVM_ABI void setDesc(const MCInstrDesc &TID)
Replace the instruction descriptor (thus opcode) of the current instruction with a new one.
LLVM_ABI bool hasOrderedMemoryRef() const
Return true if this instruction may have an ordered or volatile memory reference, or if the informati...
LLVM_ABI const MachineFunction * getMF() const
Return the function that contains the basic block that this instruction belongs to.
ArrayRef< MachineMemOperand * > memoperands() const
Access to memory operands of the instruction.
LLVM_ABI bool isLoadFoldBarrier() const
Returns true if it is illegal to fold a load across this instruction.
const DebugLoc & getDebugLoc() const
Returns the debug location id of this MachineInstr.
LLVM_ABI void eraseFromParent()
Unlink 'this' from the containing basic block and delete it.
LLVM_ABI void removeOperand(unsigned OpNo)
Erase an operand from an instruction, leaving it with one fewer operand than it started with.
LLVM_ABI void addRegisterDefined(Register Reg, const TargetRegisterInfo *RegInfo=nullptr)
We have determined MI defines a register.
const MachineOperand & getOperand(unsigned i) const
uint32_t getFlags() const
Return the MI flags bitvector.
LLVM_ABI int findRegisterDefOperandIdx(Register Reg, const TargetRegisterInfo *TRI, bool isDead=false, bool Overlap=false) const
Returns the operand index that is a def of the specified register or -1 if it is not found.
const std::vector< MachineJumpTableEntry > & getJumpTables() const
A description of a memory reference used in the backend.
@ MOVolatile
The memory access is volatile.
@ MOLoad
The memory access reads data.
@ MOStore
The memory access writes data.
This class contains meta information specific to a module.
LLVM_ABI MachineFunction * getMachineFunction(const Function &F) const
Returns the MachineFunction associated to IR function F if there is one, otherwise nullptr.
MachineOperand class - Representation of each machine instruction operand.
void setSubReg(unsigned subReg)
unsigned getSubReg() const
void setImm(int64_t immVal)
bool isReg() const
isReg - Tests if this is a MO_Register operand.
MachineBasicBlock * getMBB() const
void setIsDead(bool Val=true)
LLVM_ABI void setReg(Register Reg)
Change the register this operand corresponds to.
bool isImm() const
isImm - Tests if this is a MO_Immediate operand.
void setIsKill(bool Val=true)
MachineInstr * getParent()
getParent - Return the instruction that this operand belongs to.
unsigned getTargetFlags() const
static MachineOperand CreateImm(int64_t Val)
MachineOperandType getType() const
getType - Returns the MachineOperandType for this operand.
void setIsUndef(bool Val=true)
Register getReg() const
getReg - Returns the register number.
bool isFI() const
isFI - Tests if this is a MO_FrameIndex operand.
LLVM_ABI bool isIdenticalTo(const MachineOperand &Other) const
Returns true if this operand is identical to the specified operand except for liveness related flags ...
static MachineOperand CreateReg(Register Reg, bool isDef, bool isImp=false, bool isKill=false, bool isDead=false, bool isUndef=false, bool isEarlyClobber=false, unsigned SubReg=0, bool isDebug=false, bool isInternalRead=false, bool isRenamable=false)
MachineRegisterInfo - Keep track of information for virtual and physical registers,...
bool tracksLiveness() const
tracksLiveness - Returns true when tracking register liveness accurately.
bool reservedRegsFrozen() const
reservedRegsFrozen - Returns true after freezeReservedRegs() was called to ensure the set of reserved...
const TargetRegisterClass * getRegClassOrNull(Register Reg) const
Return the register class of Reg, or null if Reg has not been assigned a register class yet.
LLVM_ABI const TargetRegisterClass * constrainRegClass(Register Reg, const TargetRegisterClass *RC, unsigned MinNumRegs=0)
constrainRegClass - Constrain the register class of the specified virtual register to be a common sub...
LLVM_ABI MachineInstr * getUniqueVRegDef(Register Reg) const
getUniqueVRegDef - Return the unique machine instr that defines the specified virtual register or nul...
A Module instance is used to store all the information related to an LLVM module.
MI-level patchpoint operands.
uint32_t getNumPatchBytes() const
Return the number of patchable bytes the given patchpoint should emit.
Wrapper class representing virtual and physical registers.
MCRegister asMCReg() const
Utility to check-convert this value to a MCRegister.
constexpr bool isValid() const
constexpr bool isVirtual() const
Return true if the specified register number is in the virtual register namespace.
static constexpr bool isVirtualRegister(unsigned Reg)
Return true if the specified register number is in the virtual register namespace.
constexpr bool isPhysical() const
Return true if the specified register number is in the physical register namespace.
Represents a location in source code.
bool erase(PtrType Ptr)
Remove pointer from the set.
std::pair< iterator, bool > insert(PtrType Ptr)
Inserts Ptr if and only if there is no element in the container equal to Ptr.
SmallPtrSet - This class implements a set which is optimized for holding SmallSize or less elements.
SmallSet - This maintains a set of unique values, optimizing for the case when the set is small (less...
SmallString - A SmallString is just a SmallVector with methods and accessors that make it work better...
void append(StringRef RHS)
Append from a StringRef.
StringRef str() const
Explicit conversion to StringRef.
This class consists of common code factored out of the SmallVector class to reduce code duplication b...
void push_back(const T &Elt)
This is a 'vector' (really, a variable-sized array), optimized for the case when the array is small.
MI-level stackmap operands.
uint32_t getNumPatchBytes() const
Return the number of patchable bytes the given stackmap should emit.
StackOffset holds a fixed and a scalable offset in bytes.
int64_t getFixed() const
Returns the fixed component of the stack.
int64_t getScalable() const
Returns the scalable component of the stack.
static StackOffset get(int64_t Fixed, int64_t Scalable)
static StackOffset getScalable(int64_t Scalable)
static StackOffset getFixed(int64_t Fixed)
MI-level Statepoint operands.
uint32_t getNumPatchBytes() const
Return the number of patchable bytes the given statepoint should emit.
StringRef - Represent a constant reference to a string, i.e.
Object returned by analyzeLoopForPipelining.
TargetInstrInfo - Interface to description of machine instruction set.
virtual void genAlternativeCodeSequence(MachineInstr &Root, unsigned Pattern, SmallVectorImpl< MachineInstr * > &InsInstrs, SmallVectorImpl< MachineInstr * > &DelInstrs, DenseMap< Register, unsigned > &InstIdxForVirtReg) const
When getMachineCombinerPatterns() finds patterns, this function generates the instructions that could...
virtual std::optional< ParamLoadedValue > describeLoadedValue(const MachineInstr &MI, Register Reg) const
Produce the expression describing the MI loading a value into the physical register Reg.
virtual bool getMachineCombinerPatterns(MachineInstr &Root, SmallVectorImpl< unsigned > &Patterns, bool DoRegPressureReduce) const
Return true when there is potentially a faster code sequence for an instruction chain ending in Root.
virtual bool isSchedulingBoundary(const MachineInstr &MI, const MachineBasicBlock *MBB, const MachineFunction &MF) const
Test if the given instruction should be considered a scheduling boundary.
virtual CombinerObjective getCombinerObjective(unsigned Pattern) const
Return the objective of a combiner pattern.
virtual bool isFunctionSafeToSplit(const MachineFunction &MF) const
Return true if the function is a viable candidate for machine function splitting.
CodeModel::Model getCodeModel() const
Returns the code model.
const MCAsmInfo * getMCAsmInfo() const
Return target specific asm information.
bool contains(Register Reg) const
Return true if the specified register is included in this register class.
bool hasSubClassEq(const TargetRegisterClass *RC) const
Returns true if RC is a sub-class of or equal to this class.
bool hasSuperClassEq(const TargetRegisterClass *RC) const
Returns true if RC is a super-class of or equal to this class.
TargetRegisterInfo base class - We assume that the target defines a static array of TargetRegisterDes...
TargetSubtargetInfo - Generic base class for all target subtargets.
virtual const TargetInstrInfo * getInstrInfo() const
virtual const TargetRegisterInfo * getRegisterInfo() const =0
Return the target's register information.
Target - Wrapper for Target specific information.
static constexpr TypeSize getFixed(ScalarTy ExactSize)
static constexpr TypeSize getScalable(ScalarTy MinimumSize)
Value * getOperand(unsigned i) const
constexpr bool isScalable() const
Returns whether the quantity is scaled by a runtime quantity (vscale).
constexpr ScalarTy getKnownMinValue() const
Returns the minimum value this quantity can represent.
self_iterator getIterator()
A raw_ostream that writes to an std::string.
#define llvm_unreachable(msg)
Marks that the current location is not supposed to be reachable.
static CondCode getInvertedCondCode(CondCode Code)
@ MO_DLLIMPORT
MO_DLLIMPORT - On a symbol operand, this represents that the reference to the symbol is for an import...
@ MO_NC
MO_NC - Indicates whether the linker is expected to check the symbol reference for overflow.
@ MO_G1
MO_G1 - A symbol operand with this flag (granule 1) represents the bits 16-31 of a 64-bit address,...
@ MO_S
MO_S - Indicates that the bits of the symbol operand represented by MO_G0 etc are signed.
@ MO_PAGEOFF
MO_PAGEOFF - A symbol operand with this flag represents the offset of that symbol within a 4K page.
@ MO_GOT
MO_GOT - This flag indicates that a symbol operand represents the address of the GOT entry for the sy...
@ MO_PREL
MO_PREL - Indicates that the bits of the symbol operand represented by MO_G0 etc are PC relative.
@ MO_G0
MO_G0 - A symbol operand with this flag (granule 0) represents the bits 0-15 of a 64-bit address,...
@ MO_ARM64EC_CALLMANGLE
MO_ARM64EC_CALLMANGLE - Operand refers to the Arm64EC-mangled version of a symbol,...
@ MO_PAGE
MO_PAGE - A symbol operand with this flag represents the pc-relative offset of the 4K page containing...
@ MO_HI12
MO_HI12 - This flag indicates that a symbol operand represents the bits 13-24 of a 64-bit address,...
@ MO_TLS
MO_TLS - Indicates that the operand being accessed is some kind of thread-local symbol.
@ MO_G2
MO_G2 - A symbol operand with this flag (granule 2) represents the bits 32-47 of a 64-bit address,...
@ MO_TAGGED
MO_TAGGED - With MO_PAGE, indicates that the page includes a memory tag in bits 56-63.
@ MO_G3
MO_G3 - A symbol operand with this flag (granule 3) represents the high 16-bits of a 64-bit address,...
@ MO_COFFSTUB
MO_COFFSTUB - On a symbol operand "FOO", this indicates that the reference is actually to the "....
unsigned getCheckerSizeInBytes(AuthCheckMethod Method)
Returns the number of bytes added by checkAuthenticatedRegister.
static uint64_t decodeLogicalImmediate(uint64_t val, unsigned regSize)
decodeLogicalImmediate - Decode a logical immediate value in the form "N:immr:imms" (where the immr a...
static unsigned getShiftValue(unsigned Imm)
getShiftValue - Extract the shift value.
static unsigned getArithExtendImm(AArch64_AM::ShiftExtendType ET, unsigned Imm)
getArithExtendImm - Encode the extend type and shift amount for an arithmetic instruction: imm: 3-bit...
static unsigned getArithShiftValue(unsigned Imm)
getArithShiftValue - get the arithmetic shift value.
static uint64_t encodeLogicalImmediate(uint64_t imm, unsigned regSize)
encodeLogicalImmediate - Return the encoded immediate value for a logical immediate instruction of th...
static AArch64_AM::ShiftExtendType getExtendType(unsigned Imm)
getExtendType - Extract the extend type for operands of arithmetic ops.
static AArch64_AM::ShiftExtendType getArithExtendType(unsigned Imm)
static AArch64_AM::ShiftExtendType getShiftType(unsigned Imm)
getShiftType - Extract the shift type.
static unsigned getShifterImm(AArch64_AM::ShiftExtendType ST, unsigned Imm)
getShifterImm - Encode the shift type and amount: imm: 6-bit shift amount shifter: 000 ==> lsl 001 ==...
void expandMOVImm(uint64_t Imm, unsigned BitSize, SmallVectorImpl< ImmInsnModel > &Insn)
Expand a MOVi32imm or MOVi64imm pseudo instruction to one or more real move-immediate instructions to...
static const uint64_t InstrFlagIsWhile
static const uint64_t InstrFlagIsPTestLike
constexpr std::underlying_type_t< E > Mask()
Get a bitmask with 1s in all places up to the high-order bit of E's largest value.
@ C
The default llvm calling convention, compatible with C.
@ ScalablePredicateVector
initializer< Ty > init(const Ty &Val)
InstrType
Represents how an instruction should be mapped by the outliner.
LLVM_ABI Instruction & back() const
This is an optimization pass for GlobalISel generic memory operations.
auto drop_begin(T &&RangeOrContainer, size_t N=1)
Return a range covering RangeOrContainer with the first N elements excluded.
bool all_of(R &&range, UnaryPredicate P)
Provide wrappers to std::all_of which take ranges instead of having to pass begin/end explicitly.
static bool isCondBranchOpcode(int Opc)
MCCFIInstruction createDefCFA(const TargetRegisterInfo &TRI, unsigned FrameReg, unsigned Reg, const StackOffset &Offset, bool LastAdjustmentWasScalable=true)
static bool isPTrueOpcode(unsigned Opc)
MachineInstrBuilder BuildMI(MachineFunction &MF, const MIMetadata &MIMD, const MCInstrDesc &MCID)
Builder interface. Specify how to create the initial instruction itself.
bool succeeded(LogicalResult Result)
Utility function that returns true if the provided LogicalResult corresponds to a success value.
int isAArch64FrameOffsetLegal(const MachineInstr &MI, StackOffset &Offset, bool *OutUseUnscaledOp=nullptr, unsigned *OutUnscaledOp=nullptr, int64_t *EmittableOffset=nullptr)
Check if the Offset is a valid frame offset for MI.
constexpr bool isInt(int64_t x)
Checks if an integer fits into the given bit width.
RegState
Flags to represent properties of register accesses.
@ Implicit
Not emitted register (e.g. carry, or temporary result).
@ Kill
The last use of a register.
@ Undef
Value of the register doesn't matter.
@ Define
Register definition.
@ Renamable
Register that may be renamed.
auto enumerate(FirstRange &&First, RestRanges &&...Rest)
Given two or more input ranges, returns a new range whose values are tuples (A, B,...
constexpr RegState getKillRegState(bool B)
decltype(auto) dyn_cast(const From &Val)
dyn_cast<X> - Return the argument parameter cast to the specified type.
static bool isIndirectBranchOpcode(int Opc)
iterator_range< T > make_range(T x, T y)
Convenience function for iterating over sub-ranges.
unsigned getBLRCallOpcode(const MachineFunction &MF)
Return opcode to be used for indirect calls.
@ AArch64FrameOffsetIsLegal
Offset is legal.
@ AArch64FrameOffsetCanUpdate
Offset can apply, at least partly.
@ AArch64FrameOffsetCannotUpdate
Offset cannot apply.
constexpr bool isPowerOf2_64(uint64_t Value)
Return true if the argument is a power of two > 0 (64 bit edition.)
unsigned Log2_64(uint64_t Value)
Return the floor log base 2 of the specified value, -1 if the value is zero.
static bool isSEHInstruction(const MachineInstr &MI)
bool any_of(R &&range, UnaryPredicate P)
Provide wrappers to std::any_of which take ranges instead of having to pass begin/end explicitly.
auto reverse(ContainerTy &&C)
MachineInstr * getImm(const MachineOperand &MO, const MachineRegisterInfo *MRI)
decltype(auto) get(const PointerIntPair< PointerTy, IntBits, IntType, PtrTraits, Info > &Pair)
void sort(IteratorTy Start, IteratorTy End)
AArch64MachineCombinerPattern
@ MULSUBv2i32_indexed_OP1
@ MULADDv4i16_indexed_OP2
@ MULSUBv8i16_indexed_OP2
@ MULSUBv4i16_indexed_OP2
@ MULSUBv4i32_indexed_OP2
@ MULADDv2i32_indexed_OP1
@ MULADDv4i32_indexed_OP1
@ MULADDv2i32_indexed_OP2
@ MULSUBv4i16_indexed_OP1
@ MULADDv4i32_indexed_OP2
@ MULSUBv8i16_indexed_OP1
@ MULSUBv2i32_indexed_OP2
@ MULADDv8i16_indexed_OP1
@ MULSUBv4i32_indexed_OP1
@ MULADDv8i16_indexed_OP2
@ MULADDv4i16_indexed_OP1
LLVM_ABI raw_ostream & dbgs()
dbgs() - This returns a reference to a raw_ostream for debugging messages.
void emitFrameOffset(MachineBasicBlock &MBB, MachineBasicBlock::iterator MBBI, const DebugLoc &DL, unsigned DestReg, unsigned SrcReg, StackOffset Offset, const TargetInstrInfo *TII, MachineInstr::MIFlag=MachineInstr::NoFlags, bool SetNZCV=false, bool NeedsWinCFI=false, bool *HasWinCFI=nullptr, bool EmitCFAOffset=false, StackOffset InitialOffset={}, unsigned FrameReg=AArch64::SP)
emitFrameOffset - Emit instructions as needed to set DestReg to SrcReg plus Offset.
LLVM_ABI void report_fatal_error(Error Err, bool gen_crash_diag=true)
constexpr RegState getDefRegState(bool B)
CombinerObjective
The combiner's goal may differ based on which pattern it is attempting to optimize.
std::optional< UsedNZCV > examineCFlagsUse(MachineInstr &MI, MachineInstr &CmpInstr, const TargetRegisterInfo &TRI, SmallVectorImpl< MachineInstr * > *CCUseInstrs=nullptr)
CodeGenOptLevel
Code generation optimization level.
class LLVM_GSL_OWNER SmallVector
Forward declaration of SmallVector so that calculateSmallVectorDefaultInlinedElements can reference s...
auto instructionsWithoutDebug(IterT It, IterT End, bool SkipPseudoOp=true)
Construct a range iterator which begins at It and moves forwards until End is reached,...
LLVM_ABI raw_fd_ostream & errs()
This returns a reference to a raw_ostream for standard error.
auto drop_end(T &&RangeOrContainer, size_t N=1)
Return a range covering RangeOrContainer with the last N elements excluded.
static MCRegister getXRegFromWReg(MCRegister Reg)
MCCFIInstruction createCFAOffset(const TargetRegisterInfo &MRI, unsigned Reg, const StackOffset &OffsetFromDefCFA, std::optional< int64_t > IncomingVGOffsetFromDefCFA)
FunctionAddr VTableAddr Next
DWARFExpression::Operation Op
ArrayRef(const T &OneElt) -> ArrayRef< T >
static bool isUncondBranchOpcode(int Opc)
decltype(auto) cast(const From &Val)
cast<X> - Return the argument parameter cast to the specified type.
void erase_if(Container &C, UnaryPredicate P)
Provide a container algorithm similar to C++ Library Fundamentals v2's erase_if which is equivalent t...
constexpr bool isIntN(unsigned N, int64_t x)
Checks if an signed integer fits into the given (dynamic) bit width.
bool rewriteAArch64FrameIndex(MachineInstr &MI, unsigned FrameRegIdx, unsigned FrameReg, StackOffset &Offset, const AArch64InstrInfo *TII)
rewriteAArch64FrameIndex - Rewrite MI to access 'Offset' bytes from the FP.
bool is_contained(R &&Range, const E &Element)
Returns true if Element is found in Range.
static const MachineMemOperand::Flags MOSuppressPair
constexpr int64_t SignExtend64(uint64_t x)
Sign-extend the number in the bottom B bits of X to a 64-bit integer.
void appendLEB128(SmallVectorImpl< U > &Buffer, T Value)
bool optimizeTerminators(MachineBasicBlock *MBB, const TargetInstrInfo &TII)
std::pair< MachineOperand, DIExpression * > ParamLoadedValue
bool isNZCVTouchedInInstructionRange(const MachineInstr &DefMI, const MachineInstr &UseMI, const TargetRegisterInfo *TRI)
Return true if there is an instruction /after/ DefMI and before UseMI which either reads or clobbers ...
auto seq(T Begin, T End)
Iterate over an integral type from Begin up to - but not including - End.
static const MachineMemOperand::Flags MOStridedAccess
constexpr RegState getUndefRegState(bool B)
void fullyRecomputeLiveIns(ArrayRef< MachineBasicBlock * > MBBs)
Convenience function for recomputing live-in's for a set of MBBs until the computation converges.
LLVM_ABI Printable printReg(Register Reg, const TargetRegisterInfo *TRI=nullptr, unsigned SubIdx=0, const MachineRegisterInfo *MRI=nullptr)
Prints virtual and physical registers with or without a TRI instance.
This struct is a compact representation of a valid (non-zero power of two) alignment.
Used to describe addressing mode similar to ExtAddrMode in CodeGenPrepare.
LLVM_ABI static const MBBSectionID ColdSectionID
This class contains a discriminated union of information about pointers in memory operands,...
static LLVM_ABI MachinePointerInfo getUnknownStack(MachineFunction &MF)
Stack memory without other information.
static LLVM_ABI MachinePointerInfo getFixedStack(MachineFunction &MF, int FI, int64_t Offset=0)
Return a MachinePointerInfo record that refers to the specified FrameIndex.
An individual sequence of instructions to be replaced with a call to an outlined function.
MachineFunction * getMF() const
The information necessary to create an outlined function for some class of candidate.