38#include "llvm/IR/IntrinsicsAArch64.h"
43#define GET_TARGET_REGBANK_IMPL
44#include "AArch64GenRegisterBank.inc"
47#include "AArch64GenRegisterBankInfo.def"
56 static auto InitializeRegisterBankOnce = [&]() {
65 assert(&AArch64::GPRRegBank == &RBGPR &&
66 "The order in RegBanks is messed up");
70 assert(&AArch64::FPRRegBank == &RBFPR &&
71 "The order in RegBanks is messed up");
75 assert(&AArch64::CCRegBank == &RBCCR &&
76 "The order in RegBanks is messed up");
81 "Subclass not added?");
83 "GPRs should hold up to 128-bit");
88 "Subclass not added?");
90 "Subclass not added?");
92 "FPRs should hold up to 512-bit via QQQQ sequence");
97 "CCR should hold up to 32-bit");
103 "PartialMappingIdx's are incorrectly ordered");
107 "PartialMappingIdx's are incorrectly ordered");
110#define CHECK_PARTIALMAP(Idx, ValStartIdx, ValLength, RB) \
113 checkPartialMap(PartialMappingIdx::Idx, ValStartIdx, ValLength, RB) && \
114 #Idx " is incorrectly initialized"); \
128#define CHECK_VALUEMAP_IMPL(RBName, Size, Offset) \
130 assert(checkValueMapImpl(PartialMappingIdx::PMI_##RBName##Size, \
131 PartialMappingIdx::PMI_First##RBName, Size, \
133 #RBName #Size " " #Offset " is incorrectly initialized"); \
136#define CHECK_VALUEMAP(RBName, Size) CHECK_VALUEMAP_IMPL(RBName, Size, 0)
150#define CHECK_VALUEMAP_3OPS(RBName, Size) \
152 CHECK_VALUEMAP_IMPL(RBName, Size, 0); \
153 CHECK_VALUEMAP_IMPL(RBName, Size, 1); \
154 CHECK_VALUEMAP_IMPL(RBName, Size, 2); \
166#define CHECK_VALUEMAP_CROSSREGCPY(RBNameDst, RBNameSrc, Size) \
168 unsigned PartialMapDstIdx = PMI_##RBNameDst##Size - PMI_Min; \
169 unsigned PartialMapSrcIdx = PMI_##RBNameSrc##Size - PMI_Min; \
170 (void)PartialMapDstIdx; \
171 (void)PartialMapSrcIdx; \
172 const ValueMapping *Map = getCopyMapping(AArch64::RBNameDst##RegBankID, \
173 AArch64::RBNameSrc##RegBankID, \
174 TypeSize::getFixed(Size)); \
176 assert(Map[0].BreakDown == \
177 &AArch64GenRegisterBankInfo::PartMappings[PartialMapDstIdx] && \
178 Map[0].NumBreakDowns == 1 && \
179 #RBNameDst #Size " Dst is incorrectly initialized"); \
180 assert(Map[1].BreakDown == \
181 &AArch64GenRegisterBankInfo::PartMappings[PartialMapSrcIdx] && \
182 Map[1].NumBreakDowns == 1 && \
183 #RBNameSrc #Size " Src is incorrectly initialized"); \
196#define CHECK_VALUEMAP_FPEXT(DstSize, SrcSize) \
198 unsigned PartialMapDstIdx = PMI_FPR##DstSize - PMI_Min; \
199 unsigned PartialMapSrcIdx = PMI_FPR##SrcSize - PMI_Min; \
200 (void)PartialMapDstIdx; \
201 (void)PartialMapSrcIdx; \
202 const ValueMapping *Map = getFPExtMapping(DstSize, SrcSize); \
204 assert(Map[0].BreakDown == \
205 &AArch64GenRegisterBankInfo::PartMappings[PartialMapDstIdx] && \
206 Map[0].NumBreakDowns == 1 && "FPR" #DstSize \
207 " Dst is incorrectly initialized"); \
208 assert(Map[1].BreakDown == \
209 &AArch64GenRegisterBankInfo::PartMappings[PartialMapSrcIdx] && \
210 Map[1].NumBreakDowns == 1 && "FPR" #SrcSize \
211 " Src is incorrectly initialized"); \
223 llvm::call_once(InitializeRegisterBankFlag, InitializeRegisterBankOnce);
237 if (&
A == &AArch64::GPRRegBank && &
B == &AArch64::FPRRegBank)
240 if (&
A == &AArch64::FPRRegBank && &
B == &AArch64::GPRRegBank)
250 switch (RC.
getID()) {
251 case AArch64::GPR64sponlyRegClassID:
266 switch (
MI.getOpcode()) {
267 case TargetOpcode::G_OR: {
276 if (
MI.getNumOperands() != 3)
290 case TargetOpcode::G_BITCAST: {
297 if (
MI.getNumOperands() != 2)
312 copyCost(AArch64::GPRRegBank, AArch64::FPRRegBank,
319 copyCost(AArch64::GPRRegBank, AArch64::FPRRegBank,
330 case TargetOpcode::G_LOAD: {
337 if (
MI.getNumOperands() != 2)
369 assert(
MI.getOpcode() == TargetOpcode::G_FCONSTANT);
373 unsigned Size = Ty.getSizeInBits();
380 const APFloat Imm =
MI.getOperand(1).getFPImm()->getValueAPF();
381 const APInt ImmBits = Imm.bitcastToAPInt();
385 return UseMI.getOpcode() == TargetOpcode::G_STORE &&
386 UseMI.getOperand(0).getReg() == Dst;
407 return !IsFMov && IsLegal;
422 if (!TruncMI || TruncMI->
getOpcode() != TargetOpcode::G_TRUNC)
427 if (!SrcDef || SrcDef->
getOpcode() != TargetOpcode::G_CONSTANT)
444void AArch64RegisterBankInfo::applyMappingImpl(
446 MachineInstr &
MI = OpdMapper.getMI();
447 MachineRegisterInfo &MRI = OpdMapper.getMRI();
449 switch (
MI.getOpcode()) {
450 case TargetOpcode::G_CONSTANT: {
452 [[maybe_unused]] LLT DstTy = MRI.
getType(Dst);
455 "Expected a scalar smaller than 32 bits on a GPR.");
460 APInt Val =
MI.getOperand(1).getCImm()->getValue().zext(32);
462 MI.getOperand(1).setCImm(ConstantInt::get(Ctx, Val));
463 MI.getOperand(0).setReg(ExtReg);
468 case TargetOpcode::G_FCONSTANT: {
471 "Expected Dst to be on a GPR.");
472 const APFloat &
Imm =
MI.getOperand(1).getFPImm()->getValueAPF();
473 APInt
Bits =
Imm.bitcastToAPInt();
475 if (
Bits.getBitWidth() < 32) {
483 MI.eraseFromParent();
486 case TargetOpcode::G_STORE: {
498 MI.getOperand(0).setReg(Ext.getReg(0));
499 MRI.
setRegBank(Ext.getReg(0), AArch64::GPRRegBank);
503 case TargetOpcode::G_LOAD: {
511 MI.getOperand(0).setReg(ExtReg);
516 case TargetOpcode::G_OR:
517 case TargetOpcode::G_BITCAST:
519 assert((OpdMapper.getInstrMapping().getID() >= 1 &&
520 OpdMapper.getInstrMapping().getID() <= 4) &&
521 "Don't know how to handle that ID");
523 case TargetOpcode::G_INSERT_VECTOR_ELT: {
531 MI.getOperand(2).getReg());
533 MI.getOperand(2).setReg(Ext.getReg(0));
536 case AArch64::G_DUP: {
542 "Expected sources smaller than 32-bits");
549 MI.getOperand(1).setReg(ConstReg);
559AArch64RegisterBankInfo::getSameKindOfOperandsMapping(
561 const unsigned Opc =
MI.getOpcode();
562 const MachineFunction &MF = *
MI.getParent()->getParent();
563 const MachineRegisterInfo &MRI = MF.
getRegInfo();
565 unsigned NumOperands =
MI.getNumOperands();
566 assert(NumOperands <= 3 &&
567 "This code is for instructions with 3 or less operands");
569 LLT Ty = MRI.
getType(
MI.getOperand(0).getReg());
584 for (
unsigned Idx = 1; Idx != NumOperands; ++Idx) {
585 LLT OpTy = MRI.
getType(
MI.getOperand(Idx).getReg());
590 "Operand has incompatible size");
593 assert(IsFPR == OpIsFPR &&
"Operand has incompatible type");
608 case Intrinsic::aarch64_neon_uaddlv:
609 case Intrinsic::aarch64_neon_uaddv:
610 case Intrinsic::aarch64_neon_saddv:
611 case Intrinsic::aarch64_neon_umaxv:
612 case Intrinsic::aarch64_neon_smaxv:
613 case Intrinsic::aarch64_neon_uminv:
614 case Intrinsic::aarch64_neon_sminv:
615 case Intrinsic::aarch64_neon_faddv:
616 case Intrinsic::aarch64_neon_fmaxv:
617 case Intrinsic::aarch64_neon_fminv:
618 case Intrinsic::aarch64_neon_fmaxnmv:
619 case Intrinsic::aarch64_neon_fminnmv:
620 case Intrinsic::aarch64_neon_fmulx:
621 case Intrinsic::aarch64_neon_frecpe:
622 case Intrinsic::aarch64_neon_frecps:
623 case Intrinsic::aarch64_neon_frecpx:
624 case Intrinsic::aarch64_neon_frsqrte:
625 case Intrinsic::aarch64_neon_frsqrts:
626 case Intrinsic::aarch64_neon_facge:
627 case Intrinsic::aarch64_neon_facgt:
628 case Intrinsic::aarch64_neon_fabd:
629 case Intrinsic::aarch64_neon_sqrdmlah:
630 case Intrinsic::aarch64_neon_sqrdmlsh:
631 case Intrinsic::aarch64_neon_sqrdmulh:
632 case Intrinsic::aarch64_neon_suqadd:
633 case Intrinsic::aarch64_neon_usqadd:
634 case Intrinsic::aarch64_neon_uqadd:
635 case Intrinsic::aarch64_neon_sqadd:
636 case Intrinsic::aarch64_neon_uqsub:
637 case Intrinsic::aarch64_neon_sqsub:
638 case Intrinsic::aarch64_neon_sqdmulls_scalar:
639 case Intrinsic::aarch64_neon_srshl:
640 case Intrinsic::aarch64_neon_urshl:
641 case Intrinsic::aarch64_neon_sqshl:
642 case Intrinsic::aarch64_neon_uqshl:
643 case Intrinsic::aarch64_neon_sqrshl:
644 case Intrinsic::aarch64_neon_uqrshl:
645 case Intrinsic::aarch64_neon_ushl:
646 case Intrinsic::aarch64_neon_sshl:
647 case Intrinsic::aarch64_neon_sqshrn:
648 case Intrinsic::aarch64_neon_sqshrun:
649 case Intrinsic::aarch64_neon_sqrshrn:
650 case Intrinsic::aarch64_neon_sqrshrun:
651 case Intrinsic::aarch64_neon_uqshrn:
652 case Intrinsic::aarch64_neon_uqrshrn:
653 case Intrinsic::aarch64_crypto_sha1h:
654 case Intrinsic::aarch64_crypto_sha1c:
655 case Intrinsic::aarch64_crypto_sha1p:
656 case Intrinsic::aarch64_crypto_sha1m:
657 case Intrinsic::aarch64_sisd_fcvtxn:
658 case Intrinsic::aarch64_sisd_fabd:
660 case Intrinsic::aarch64_neon_saddlv: {
661 const LLT SrcTy = MRI.
getType(
MI.getOperand(2).getReg());
668bool AArch64RegisterBankInfo::isPHIWithFPConstraints(
671 if (!
MI.isPHI() ||
Depth > MaxFPRSearchDepth)
675 [&](
const MachineInstr &
UseMI) {
676 if (onlyUsesFP(UseMI, MRI, TRI, Depth + 1))
678 return isPHIWithFPConstraints(UseMI, MRI, TRI, Depth + 1);
682bool AArch64RegisterBankInfo::hasFPConstraints(
const MachineInstr &
MI,
685 unsigned Depth)
const {
686 unsigned Op =
MI.getOpcode();
696 if (
Op != TargetOpcode::COPY && !
MI.isPHI() &&
702 if (RB == &AArch64::FPRRegBank)
704 if (RB == &AArch64::GPRRegBank)
711 if (!
MI.isPHI() ||
Depth > MaxFPRSearchDepth)
714 return any_of(
MI.explicit_uses(), [&](
const MachineOperand &
Op) {
716 onlyDefinesFP(*MRI.getVRegDef(Op.getReg()), MRI, TRI, Depth + 1);
723 unsigned Depth)
const {
724 switch (
MI.getOpcode()) {
725 case TargetOpcode::G_BITCAST: {
728 [&](
const MachineInstr &
UseMI) {
729 return onlyUsesFP(UseMI, MRI, TRI, Depth + 1) ||
730 prefersFPUse(UseMI, MRI, TRI);
734 case TargetOpcode::G_FPTOSI:
735 case TargetOpcode::G_FPTOUI:
736 case TargetOpcode::G_FPTOSI_SAT:
737 case TargetOpcode::G_FPTOUI_SAT:
738 case TargetOpcode::G_FCMP:
739 case TargetOpcode::G_LROUND:
740 case TargetOpcode::G_LLROUND:
741 case AArch64::G_PMULL:
744 case AArch64::G_FPTRUNC_ODD:
746 case TargetOpcode::G_INTRINSIC:
748 case Intrinsic::aarch64_neon_fcvtas:
749 case Intrinsic::aarch64_neon_fcvtau:
750 case Intrinsic::aarch64_neon_fcvtzs:
751 case Intrinsic::aarch64_neon_fcvtzu:
752 case Intrinsic::aarch64_neon_fcvtms:
753 case Intrinsic::aarch64_neon_fcvtmu:
754 case Intrinsic::aarch64_neon_fcvtns:
755 case Intrinsic::aarch64_neon_fcvtnu:
756 case Intrinsic::aarch64_neon_fcvtps:
757 case Intrinsic::aarch64_neon_fcvtpu:
769bool AArch64RegisterBankInfo::onlyDefinesFP(
const MachineInstr &
MI,
772 unsigned Depth)
const {
773 switch (
MI.getOpcode()) {
775 case AArch64::G_SADDLP:
776 case AArch64::G_UADDLP:
777 case TargetOpcode::G_SITOFP:
778 case TargetOpcode::G_UITOFP:
779 case TargetOpcode::G_EXTRACT_VECTOR_ELT:
780 case TargetOpcode::G_INSERT_VECTOR_ELT:
781 case TargetOpcode::G_BUILD_VECTOR:
782 case TargetOpcode::G_BUILD_VECTOR_TRUNC:
785 case AArch64::G_FPTRUNC_ODD:
787 case TargetOpcode::G_INTRINSIC_W_SIDE_EFFECTS:
789 case Intrinsic::aarch64_neon_ld1x2:
790 case Intrinsic::aarch64_neon_ld1x3:
791 case Intrinsic::aarch64_neon_ld1x4:
792 case Intrinsic::aarch64_neon_ld2:
793 case Intrinsic::aarch64_neon_ld2lane:
794 case Intrinsic::aarch64_neon_ld2r:
795 case Intrinsic::aarch64_neon_ld3:
796 case Intrinsic::aarch64_neon_ld3lane:
797 case Intrinsic::aarch64_neon_ld3r:
798 case Intrinsic::aarch64_neon_ld4:
799 case Intrinsic::aarch64_neon_ld4lane:
800 case Intrinsic::aarch64_neon_ld4r:
815 unsigned Depth)
const {
816 switch (
MI.getOpcode()) {
817 case TargetOpcode::G_SITOFP:
818 case TargetOpcode::G_UITOFP:
825bool AArch64RegisterBankInfo::isLoadFromFPType(
const MachineInstr &
MI)
const {
828 const Value *LdVal = MemOp->getMMO().getValue();
832 Type *EltTy =
nullptr;
834 EltTy = GV->getValueType();
838 if (StructEltTy->getNumElements() == 0)
840 EltTy = StructEltTy->getTypeAtIndex(0U);
848 for (
const auto *LdUser : LdVal->
users()) {
850 EltTy = LdUser->getType();
854 EltTy = LdUser->getOperand(0)->getType();
864 const unsigned Opc =
MI.getOpcode();
869 Opc == TargetOpcode::G_PHI) {
884 case TargetOpcode::G_ADD:
885 case TargetOpcode::G_SUB:
886 case TargetOpcode::G_PTR_ADD:
887 case TargetOpcode::G_MUL:
888 case TargetOpcode::G_SDIV:
889 case TargetOpcode::G_UDIV:
891 case TargetOpcode::G_AND:
892 case TargetOpcode::G_OR:
893 case TargetOpcode::G_XOR:
895 case TargetOpcode::G_FADD:
896 case TargetOpcode::G_FSUB:
897 case TargetOpcode::G_FMUL:
898 case TargetOpcode::G_FDIV:
899 case TargetOpcode::G_FMAXIMUM:
900 case TargetOpcode::G_FMINIMUM:
901 return getSameKindOfOperandsMapping(
MI);
902 case TargetOpcode::G_FPEXT: {
911 case TargetOpcode::G_SHL:
912 case TargetOpcode::G_LSHR:
913 case TargetOpcode::G_ASHR: {
914 LLT ShiftAmtTy = MRI.
getType(
MI.getOperand(2).getReg());
916 if (ShiftAmtTy.
getSizeInBits() == 64 && SrcTy.getSizeInBits() == 32)
919 return getSameKindOfOperandsMapping(
MI);
921 case TargetOpcode::G_BITCAST: {
934 case TargetOpcode::COPY: {
948 assert(DstRB && SrcRB &&
"Both RegBank were nullptr");
961 bool SrcIsGPR = !SrcTy.isVector() && SrcTy.getSizeInBits() <= 64;
963 DstIsGPR ? AArch64::GPRRegBank : AArch64::FPRRegBank;
965 SrcIsGPR ? AArch64::GPRRegBank : AArch64::FPRRegBank;
970 Opc == TargetOpcode::G_BITCAST ? 2 : 1);
976 unsigned NumOperands =
MI.getNumOperands();
982 for (
unsigned Idx = 0; Idx < NumOperands; ++Idx) {
983 auto &MO =
MI.getOperand(Idx);
984 if (!MO.isReg() || !MO.getReg())
990 OpSize[Idx] = Ty.getSizeInBits().getKnownMinValue();
998 (MO.isDef() && onlyDefinesFP(
MI, MRI,
TRI)) ||
999 (MO.isUse() && onlyUsesFP(
MI, MRI,
TRI)) ||
1000 Ty.getSizeInBits() > 64)
1010 case TargetOpcode::G_CONSTANT: {
1017 case TargetOpcode::G_FCONSTANT: {
1025 case AArch64::G_DUP: {
1026 Register ScalarReg =
MI.getOperand(1).getReg();
1030 if (ScalarDef->getOpcode() == TargetOpcode::G_LOAD)
1035 onlyDefinesFP(*ScalarDef, MRI,
TRI)))
1047 case TargetOpcode::G_TRUNC: {
1049 if (!SrcTy.isVector() && SrcTy.getSizeInBits() == 128)
1053 case TargetOpcode::G_SITOFP:
1054 case TargetOpcode::G_UITOFP: {
1068 case TargetOpcode::G_FPTOSI_SAT:
1069 case TargetOpcode::G_FPTOUI_SAT:
1070 case TargetOpcode::G_FPTOSI:
1071 case TargetOpcode::G_FPTOUI:
1072 case TargetOpcode::G_INTRINSIC_LRINT:
1073 case TargetOpcode::G_INTRINSIC_LLRINT:
1074 case TargetOpcode::G_LROUND:
1075 case TargetOpcode::G_LLROUND: {
1077 if (DstType.isVector())
1085 if (((DstSize == SrcSize) || STI.hasFeature(AArch64::FeatureFPRCVT)) &&
1088 return onlyUsesFP(UseMI, MRI, TRI) ||
1089 prefersFPUse(UseMI, MRI, TRI);
1096 case TargetOpcode::G_FCMP: {
1101 OpRegBankIdx = {Idx0,
1105 case TargetOpcode::G_BITCAST:
1107 if (OpRegBankIdx[0] != OpRegBankIdx[1])
1113 case TargetOpcode::G_LOAD: {
1134 if (isLoadFromFPType(
MI)) {
1153 if (isPHIWithFPConstraints(UseMI, MRI, TRI))
1156 return onlyUsesFP(UseMI, MRI, TRI) ||
1157 prefersFPUse(UseMI, MRI, TRI);
1163 if (Ty.isScalar() && Ty.getSizeInBits() < 32)
1167 case TargetOpcode::G_STORE:
1173 if (onlyDefinesFP(*
DefMI, MRI,
TRI)) {
1181 if (Ty.isScalar() && Ty.getSizeInBits() < 32)
1185 case TargetOpcode::G_INDEXED_STORE:
1191 if (onlyDefinesFP(*
DefMI, MRI,
TRI))
1196 case TargetOpcode::G_INDEXED_SEXTLOAD:
1197 case TargetOpcode::G_INDEXED_ZEXTLOAD:
1201 case TargetOpcode::G_INDEXED_LOAD: {
1202 if (isLoadFromFPType(
MI))
1206 case TargetOpcode::G_SELECT: {
1214 if (SrcTy.isVector()) {
1247 for (
unsigned Idx = 2; Idx < 4; ++Idx) {
1262 case TargetOpcode::G_UNMERGE_VALUES: {
1268 LLT SrcTy = MRI.
getType(
MI.getOperand(
MI.getNumOperands()-1).getReg());
1271 if (SrcTy.isVector() || SrcTy ==
LLT::scalar(128) ||
1275 for (
unsigned Idx = 0, NumOperands =
MI.getNumOperands();
1276 Idx < NumOperands; ++Idx)
1281 case TargetOpcode::G_EXTRACT_VECTOR_ELT:
1289 case AArch64::G_SQSHLU_I:
1298 case TargetOpcode::G_INSERT_VECTOR_ELT:
1303 if (
getRegBank(
MI.getOperand(2).getReg(), MRI,
TRI) == &AArch64::FPRRegBank)
1309 if (Ty.getSizeInBits() == 8 || Ty.getSizeInBits() == 16) {
1319 case TargetOpcode::G_EXTRACT: {
1321 auto Src =
MI.getOperand(1).getReg();
1323 if (SrcTy.getSizeInBits() != 128)
1328 OpRegBankIdx[0] = Idx;
1329 OpRegBankIdx[1] = Idx;
1332 case TargetOpcode::G_BUILD_VECTOR: {
1347 unsigned DefOpc =
DefMI->getOpcode();
1350 return Op.isDef() || MRI.getVRegDef(Op.getReg())->getOpcode() ==
1351 TargetOpcode::G_CONSTANT;
1355 SrcTy.getSizeInBits() < 32 ||
1359 unsigned NumOperands =
MI.getNumOperands();
1360 for (
unsigned Idx = 0; Idx < NumOperands; ++Idx)
1365 case TargetOpcode::G_VECREDUCE_FADD:
1366 case TargetOpcode::G_VECREDUCE_FMUL:
1367 case TargetOpcode::G_VECREDUCE_FMAX:
1368 case TargetOpcode::G_VECREDUCE_FMIN:
1369 case TargetOpcode::G_VECREDUCE_FMAXIMUM:
1370 case TargetOpcode::G_VECREDUCE_FMINIMUM:
1371 case TargetOpcode::G_VECREDUCE_ADD:
1372 case TargetOpcode::G_VECREDUCE_MUL:
1373 case TargetOpcode::G_VECREDUCE_AND:
1374 case TargetOpcode::G_VECREDUCE_OR:
1375 case TargetOpcode::G_VECREDUCE_XOR:
1376 case TargetOpcode::G_VECREDUCE_SMAX:
1377 case TargetOpcode::G_VECREDUCE_SMIN:
1378 case TargetOpcode::G_VECREDUCE_UMAX:
1379 case TargetOpcode::G_VECREDUCE_UMIN:
1384 case TargetOpcode::G_VECREDUCE_SEQ_FADD:
1385 case TargetOpcode::G_VECREDUCE_SEQ_FMUL:
1390 case TargetOpcode::G_INTRINSIC:
1391 case TargetOpcode::G_INTRINSIC_W_SIDE_EFFECTS: {
1393 case Intrinsic::aarch64_neon_fcvtas:
1394 case Intrinsic::aarch64_neon_fcvtau:
1395 case Intrinsic::aarch64_neon_fcvtzs:
1396 case Intrinsic::aarch64_neon_fcvtzu:
1397 case Intrinsic::aarch64_neon_fcvtms:
1398 case Intrinsic::aarch64_neon_fcvtmu:
1399 case Intrinsic::aarch64_neon_fcvtns:
1400 case Intrinsic::aarch64_neon_fcvtnu:
1401 case Intrinsic::aarch64_neon_fcvtps:
1402 case Intrinsic::aarch64_neon_fcvtpu: {
1414 if (DstSize == 16 ||
1415 ((DstSize == SrcSize || STI.hasFeature(AArch64::FeatureFPRCVT)) &&
1418 return onlyUsesFP(UseMI, MRI, TRI) ||
1419 prefersFPUse(UseMI, MRI, TRI);
1426 case Intrinsic::aarch64_neon_vcvtfxs2fp:
1427 case Intrinsic::aarch64_neon_vcvtfxu2fp:
1428 case Intrinsic::aarch64_neon_vcvtfp2fxs:
1429 case Intrinsic::aarch64_neon_vcvtfp2fxu:
1442 if (onlyDefinesFP(
MI, MRI,
TRI))
1443 for (
const auto &
Op :
MI.defs()) {
1449 Idx +=
MI.getNumExplicitDefs();
1451 if (onlyUsesFP(
MI, MRI,
TRI))
1452 for (
const auto &
Op :
MI.explicit_uses()) {
1466 for (
unsigned Idx = 0; Idx < NumOperands; ++Idx) {
1467 if (
MI.getOperand(Idx).isReg() &&
MI.getOperand(Idx).getReg()) {
1473 if (!Mapping->isValid())
1476 OpdsMapping[Idx] = Mapping;
MachineInstrBuilder & UseMI
MachineInstrBuilder MachineInstrBuilder & DefMI
static unsigned getIntrinsicID(const SDNode *N)
assert(UImm &&(UImm !=~static_cast< T >(0)) &&"Invalid immediate!")
#define CHECK_VALUEMAP(RBName, Size)
static bool isFPIntrinsic(const MachineRegisterInfo &MRI, const MachineInstr &MI)
#define CHECK_VALUEMAP_3OPS(RBName, Size)
static bool foldTruncOfI32Constant(MachineInstr &MI, unsigned OpIdx, MachineRegisterInfo &MRI, const AArch64RegisterBankInfo &RBI)
static const unsigned CustomMappingID
#define CHECK_PARTIALMAP(Idx, ValStartIdx, ValLength, RB)
#define CHECK_VALUEMAP_CROSSREGCPY(RBNameDst, RBNameSrc, Size)
#define CHECK_VALUEMAP_FPEXT(DstSize, SrcSize)
static bool preferGPRForFPImm(const MachineInstr &MI, const MachineRegisterInfo &MRI, const AArch64Subtarget &STI)
This file declares the targeting of the RegisterBankInfo class for AArch64.
This file implements a class to represent arbitrary precision integral constant values and operations...
static GCRegistry::Add< ErlangGC > A("erlang", "erlang-compatible garbage collector")
static GCRegistry::Add< OcamlGC > B("ocaml", "ocaml 3.10-compatible GC")
This file contains the declarations for the subclasses of Constant, which represent the different fla...
Declares convenience wrapper classes for interpreting MachineInstr instances as specific generic oper...
Implement a low-level type suitable for MachineInstr level instruction selection.
This file declares the MachineIRBuilder class.
Register const TargetRegisterInfo * TRI
Promote Memory to Register
MachineInstr unsigned OpIdx
static const MCPhysReg FPR[]
FPR - The set of FP registers that should be allocated for arguments on Darwin and AIX.
This file defines the SmallVector class.
static unsigned getRegBankBaseIdxOffset(unsigned RBIdx, TypeSize Size)
static const RegisterBankInfo::ValueMapping * getCopyMapping(unsigned DstBankID, unsigned SrcBankID, TypeSize Size)
Get the pointer to the ValueMapping of the operands of a copy instruction from the SrcBankID register...
static bool checkPartialMappingIdx(PartialMappingIdx FirstAlias, PartialMappingIdx LastAlias, ArrayRef< PartialMappingIdx > Order)
static const RegisterBankInfo::PartialMapping PartMappings[]
static const RegisterBankInfo::ValueMapping * getFPExtMapping(unsigned DstSize, unsigned SrcSize)
Get the instruction mapping for G_FPEXT.
static const RegisterBankInfo::ValueMapping * getValueMapping(PartialMappingIdx RBIdx, TypeSize Size)
Get the pointer to the ValueMapping representing the RegisterBank at RBIdx with a size of Size.
static const RegisterBankInfo::ValueMapping ValMappings[]
This class provides the information for the target register banks.
InstructionMappings getInstrAlternativeMappings(const MachineInstr &MI) const override
Get the alternative mappings for MI.
unsigned copyCost(const RegisterBank &A, const RegisterBank &B, TypeSize Size) const override
Get the cost of a copy from B to A, or put differently, get the cost of A = COPY B.
const RegisterBank & getRegBankFromRegClass(const TargetRegisterClass &RC, LLT Ty) const override
Get a register bank that covers RC.
AArch64RegisterBankInfo(const TargetRegisterInfo &TRI)
const InstructionMapping & getInstrMapping(const MachineInstr &MI) const override
Get the mapping of the different operands of MI on the register bank.
const AArch64RegisterInfo * getRegisterInfo() const override
const AArch64TargetLowering * getTargetLowering() const override
bool isFPImmLegal(const APFloat &Imm, EVT VT, bool ForCodeSize) const override
Returns true if the target can instruction select the specified FP immediate natively.
bool isFPImmLegalAsFMov(const APFloat &Imm, EVT VT) const
Class for arbitrary precision integers.
uint64_t getZExtValue() const
Get zero extended value.
static LLVM_ABI APInt getSplat(unsigned NewLen, const APInt &V)
Return a value containing V broadcasted over NewLen bits.
LLVMContext & getContext() const
getContext - Return a reference to the LLVMContext associated with this function.
constexpr bool isScalar() const
static constexpr LLT scalar(unsigned SizeInBits)
Get a low-level scalar or aggregate "bag of bits".
constexpr bool isValid() const
constexpr bool isVector() const
constexpr TypeSize getSizeInBits() const
Returns the total size of the type. Must only be called on sized types.
static LLT integer(unsigned SizeInBits)
LLT getElementType() const
Returns the vector's element type. Only valid for vector types.
const TargetSubtargetInfo & getSubtarget() const
getSubtarget - Return the subtarget for which this machine code is being compiled.
MachineRegisterInfo & getRegInfo()
getRegInfo - Return information about the registers currently in use.
Function & getFunction()
Return the LLVM function that this machine code represents.
Helper class to build MachineInstr.
void setInsertPt(MachineBasicBlock &MBB, MachineBasicBlock::iterator II)
Set the insertion point before the specified position.
MachineFunction & getMF()
Getter for the function we currently build.
MachineInstrBuilder buildTrunc(const DstOp &Res, const SrcOp &Op, std::optional< unsigned > Flags=std::nullopt)
Build and insert Res = G_TRUNC Op.
MachineInstrBuilder buildAnyExt(const DstOp &Res, const SrcOp &Op)
Build and insert Res = G_ANYEXT Op0.
virtual MachineInstrBuilder buildConstant(const DstOp &Res, const ConstantInt &Val)
Build and insert Res = G_CONSTANT Val.
Register getReg(unsigned Idx) const
Get the register for the operand index.
Representation of each machine instruction.
unsigned getOpcode() const
Returns the opcode of this MachineInstr.
const MachineOperand & getOperand(unsigned i) const
LLVM_ABI MachineInstrBundleIterator< MachineInstr > eraseFromParent()
Unlink 'this' from the containing basic block and delete it.
MachineOperand class - Representation of each machine instruction operand.
Register getReg() const
getReg - Returns the register number.
MachineRegisterInfo - Keep track of information for virtual and physical registers,...
LLVM_ABI MachineInstr * getVRegDef(Register Reg) const
getVRegDef - Return the machine instr that defines the specified virtual register or null if none is ...
const RegisterBank * getRegBank(Register Reg) const
Return the register bank of Reg.
LLT getType(Register Reg) const
Get the low-level type of Reg or LLT{} if Reg is not a generic (target independent) virtual register.
LLVM_ABI void setRegBank(Register Reg, const RegisterBank &RegBank)
Set the register bank to RegBank for Reg.
iterator_range< use_instr_nodbg_iterator > use_nodbg_instructions(Register Reg) const
LLVM_ABI Register createGenericVirtualRegister(LLT Ty, StringRef Name="")
Create and return a new generic virtual register with low-level type Ty.
const TargetRegisterClass * getRegClassOrNull(Register Reg) const
Return the register class of Reg, or null if Reg has not been assigned a register class yet.
bool use_empty(Register RegNo) const
use_empty - Return true if there are no instructions using the specified register.
Helper class that represents how the value of an instruction may be mapped and what is the related co...
bool isValid() const
Check whether this object is valid.
virtual InstructionMappings getInstrAlternativeMappings(const MachineInstr &MI) const
Get the alternative mappings for MI.
const InstructionMapping & getInstructionMapping(unsigned ID, unsigned Cost, const ValueMapping *OperandsMapping, unsigned NumOperands) const
Method to get a uniquely generated InstructionMapping.
static void applyDefaultMapping(const OperandsMapper &OpdMapper)
Helper method to apply something that is like the default mapping.
const InstructionMapping & getInvalidInstructionMapping() const
Method to get a uniquely generated invalid InstructionMapping.
const RegisterBank & getRegBank(unsigned ID)
Get the register bank identified by ID.
unsigned getMaximumSize(unsigned RegBankID) const
Get the maximum size in bits that fits in the given register bank.
TypeSize getSizeInBits(Register Reg, const MachineRegisterInfo &MRI, const TargetRegisterInfo &TRI) const
Get the size in bits of Reg.
virtual const RegisterBank & getRegBankFromRegClass(const TargetRegisterClass &RC, LLT Ty) const
Get a register bank that covers RC.
const ValueMapping * getOperandsMapping(Iterator Begin, Iterator End) const
Get the uniquely generated array of ValueMapping for the elements of between Begin and End.
static const unsigned DefaultMappingID
Identifier used when the related instruction mapping instance is generated by target independent code...
SmallVector< const InstructionMapping *, 4 > InstructionMappings
Convenient type to represent the alternatives for mapping an instruction.
virtual unsigned copyCost(const RegisterBank &A, const RegisterBank &B, TypeSize Size) const
Get the cost of a copy from B to A, or put differently, get the cost of A = COPY B.
const InstructionMapping & getInstrMappingImpl(const MachineInstr &MI) const
Try to get the mapping of MI.
This class implements the register bank concept.
LLVM_ABI bool covers(const TargetRegisterClass &RC) const
Check whether this register bank covers RC.
unsigned getID() const
Get the identifier of this register bank.
Wrapper class representing virtual and physical registers.
constexpr bool isPhysical() const
Return true if the specified register number is in the physical register namespace.
void push_back(const T &Elt)
This is a 'vector' (really, a variable-sized array), optimized for the case when the array is small.
unsigned getID() const
Return the register class ID number.
TargetRegisterInfo base class - We assume that the target defines a static array of TargetRegisterDes...
TargetSubtargetInfo - Generic base class for all target subtargets.
virtual const TargetRegisterInfo * getRegisterInfo() const =0
Return the target's register information.
static constexpr TypeSize getFixed(ScalarTy ExactSize)
Type * getArrayElementType() const
bool isFPOrFPVectorTy() const
Return true if this is a FP type or a vector of FP.
iterator_range< user_iterator > users()
constexpr ScalarTy getFixedValue() const
#define llvm_unreachable(msg)
Marks that the current location is not supposed to be reachable.
static bool isAdvSIMDModImmType4(uint64_t Imm)
OperandType
Operands are tagged with one of the values of this enum.
This is an optimization pass for GlobalISel generic memory operations.
FunctionAddr VTableAddr Value
bool all_of(R &&range, UnaryPredicate P)
Provide wrappers to std::all_of which take ranges instead of having to pass begin/end explicitly.
decltype(auto) dyn_cast(const From &Val)
dyn_cast<X> - Return the argument parameter cast to the specified type.
bool isPreISelGenericOpcode(unsigned Opcode)
Check whether the given Opcode is a generic opcode that is not supposed to appear after ISel.
LLVM_ABI bool shouldOptimizeForSize(const MachineFunction *MF, ProfileSummaryInfo *PSI, const MachineBlockFrequencyInfo *BFI, PGSOQueryType QueryType=PGSOQueryType::Other)
Returns true if machine function MF is suggested to be size-optimized based on the profile.
bool isPreISelGenericOptimizationHint(unsigned Opcode)
bool any_of(R &&range, UnaryPredicate P)
Provide wrappers to std::any_of which take ranges instead of having to pass begin/end explicitly.
bool isa(const From &Val)
isa<X> - Return true if the parameter to the template is an instance of one of the template type argu...
DWARFExpression::Operation Op
void call_once(once_flag &flag, Function &&F, Args &&... ArgList)
Execute the function specified as a parameter once.
decltype(auto) cast(const From &Val)
cast<X> - Return the argument parameter cast to the specified type.
LLVM_ABI bool isPreISelGenericFloatingPointOpcode(unsigned Opc)
Returns whether opcode Opc is a pre-isel generic floating-point opcode, having only floating-point op...
static EVT getFloatingPointVT(unsigned BitWidth)
Returns the EVT that represents a floating-point type with the given number of bits.
The llvm::once_flag structure.