2198 B.setInstrAndDebugLoc(
MI);
2199 unsigned Opc =
MI.getOpcode();
2202 case AMDGPU::G_CONSTANT:
2203 case AMDGPU::G_IMPLICIT_DEF: {
2205 LLT DstTy =
MRI.getType(DstReg);
2211 if (DstBank == &AMDGPU::VCCRegBank)
2214 if (DefRegs.
empty())
2217 B.setInsertPt(*
MI.getParent(), ++
MI.getIterator());
2220 LLVMContext &Ctx =
B.getMF().getFunction().getContext();
2222 MI.getOperand(0).setReg(NewDstReg);
2223 if (
Opc != AMDGPU::G_IMPLICIT_DEF) {
2224 uint64_t ConstVal =
MI.getOperand(1).getCImm()->getZExtValue();
2225 MI.getOperand(1).setCImm(
2229 MRI.setRegBank(NewDstReg, *DstBank);
2230 B.buildTrunc(DefRegs[0], NewDstReg);
2233 case AMDGPU::G_PHI: {
2235 LLT DstTy =
MRI.getType(DstReg);
2242 if (DstBank == &AMDGPU::VCCRegBank) {
2249 for (
unsigned I = 1, E =
MI.getNumOperands();
I != E;
I += 2) {
2253 if (SrcBank != &AMDGPU::VCCRegBank) {
2258 MRI.setRegBank(Copy.getReg(0), AMDGPU::VCCRegBank);
2259 MI.getOperand(
I).setReg(Copy.getReg(0));
2270 ApplyRegBankMapping ApplyBank(
B, *
this,
MRI, DstBank);
2271 B.setInsertPt(
B.getMBB(),
MI);
2279 case AMDGPU::G_FCMP:
2283 case AMDGPU::G_ICMP:
2284 case AMDGPU::G_UADDO:
2285 case AMDGPU::G_USUBO:
2286 case AMDGPU::G_UADDE:
2287 case AMDGPU::G_SADDE:
2288 case AMDGPU::G_USUBE:
2289 case AMDGPU::G_SSUBE: {
2290 unsigned BoolDstOp =
2291 (
Opc == AMDGPU::G_ICMP ||
Opc == AMDGPU::G_FCMP) ? 0 : 1;
2292 Register DstReg =
MI.getOperand(BoolDstOp).getReg();
2296 if (DstBank != &AMDGPU::SGPRRegBank)
2299 const bool HasCarryIn =
MI.getNumOperands() == 5;
2305 MRI.setRegBank(NewDstReg, AMDGPU::SGPRRegBank);
2306 MI.getOperand(BoolDstOp).setReg(NewDstReg);
2310 MRI.setRegBank(NewSrcReg, AMDGPU::SGPRRegBank);
2311 B.buildZExt(NewSrcReg,
MI.getOperand(4).getReg());
2312 MI.getOperand(4).setReg(NewSrcReg);
2316 B.setInsertPt(*
MBB, std::next(
MI.getIterator()));
2321 if (DefRegs.
empty())
2323 B.buildTrunc(DefRegs[0], NewDstReg);
2326 case AMDGPU::G_SELECT: {
2328 LLT DstTy =
MRI.getType(DstReg);
2331 if (CondRegs.
empty())
2338 if (CondBank == &AMDGPU::SGPRRegBank) {
2341 MRI.setRegBank(NewCondReg, AMDGPU::SGPRRegBank);
2343 MI.getOperand(1).setReg(NewCondReg);
2344 B.buildZExt(NewCondReg, CondRegs[0]);
2357 if (DefRegs.
empty()) {
2362 if (Src1Regs.
empty())
2368 if (Src2Regs.
empty())
2375 auto Flags =
MI.getFlags();
2376 B.buildSelect(DefRegs[0], CondRegs[0], Src1Regs[0], Src2Regs[0], Flags);
2377 B.buildSelect(DefRegs[1], CondRegs[0], Src1Regs[1], Src2Regs[1], Flags);
2379 MRI.setRegBank(DstReg, AMDGPU::VGPRRegBank);
2380 MI.eraseFromParent();
2383 case AMDGPU::G_BRCOND: {
2384 Register CondReg =
MI.getOperand(0).getReg();
2389 if (CondBank == &AMDGPU::SGPRRegBank) {
2392 MRI.setRegBank(NewCondReg, AMDGPU::SGPRRegBank);
2394 MI.getOperand(0).setReg(NewCondReg);
2395 B.buildZExt(NewCondReg, CondReg);
2403 case AMDGPU::G_XOR: {
2407 LLT DstTy =
MRI.getType(DstReg);
2413 if (DstBank == &AMDGPU::VCCRegBank)
2417 ApplyRegBankMapping ApplyBank(
B, *
this,
MRI, DstBank);
2426 if (DstTy.
getSizeInBits() == 16 && DstBank == &AMDGPU::SGPRRegBank) {
2430 ApplyRegBankMapping ApplySALU(
B, *
this,
MRI, &AMDGPU::SGPRRegBank);
2435 if (
MI.getOpcode() == AMDGPU::G_XOR &&
2456 if (DefRegs.
empty()) {
2463 (Src0Regs.
empty() || Src0Regs.
size() == 2));
2469 if (Src0Regs.
empty())
2474 if (Src1Regs.
empty())
2481 auto Flags =
MI.getFlags();
2482 B.buildInstr(
Opc, {DefRegs[0]}, {Src0Regs[0], Src1Regs[0]}, Flags);
2483 B.buildInstr(
Opc, {DefRegs[1]}, {Src0Regs[1], Src1Regs[1]}, Flags);
2485 MRI.setRegBank(DstReg, AMDGPU::VGPRRegBank);
2486 MI.eraseFromParent();
2489 case AMDGPU::G_ABS: {
2495 if (SrcBank && SrcBank == &AMDGPU::VGPRRegBank) {
2497 ApplyRegBankMapping Apply(
B, *
this,
MRI, &AMDGPU::VGPRRegBank);
2510 case AMDGPU::G_LSHR:
2511 case AMDGPU::G_ASHR:
2512 case AMDGPU::G_SMIN:
2513 case AMDGPU::G_SMAX:
2514 case AMDGPU::G_UMIN:
2515 case AMDGPU::G_UMAX: {
2517 LLT DstTy =
MRI.getType(DstReg);
2522 if (!
Subtarget.hasVectorMulU64() &&
Opc == AMDGPU::G_MUL &&
2535 if (DstBank == &AMDGPU::VGPRRegBank)
2541 ApplyRegBankMapping ApplySALU(
B, *
this,
MRI, &AMDGPU::SGPRRegBank);
2546 std::tie(WideSrcLo, WideSrcHi) =
2548 auto Lo =
B.buildInstr(AMDGPU::G_ABS, {
S32}, {WideSrcLo});
2549 auto Hi =
B.buildInstr(AMDGPU::G_ABS, {
S32}, {WideSrcHi});
2550 B.buildBuildVectorTrunc(DstReg, {
Lo.getReg(0),
Hi.getReg(0)});
2551 MI.eraseFromParent();
2560 std::tie(WideSrc0Lo, WideSrc0Hi)
2562 std::tie(WideSrc1Lo, WideSrc1Hi)
2564 auto Lo =
B.buildInstr(
MI.getOpcode(), {S32}, {WideSrc0Lo, WideSrc1Lo});
2565 auto Hi =
B.buildInstr(
MI.getOpcode(), {S32}, {WideSrc0Hi, WideSrc1Hi});
2566 B.buildBuildVectorTrunc(DstReg, {
Lo.getReg(0),
Hi.getReg(0)});
2567 MI.eraseFromParent();
2575 if (
Opc == AMDGPU::G_SHL ||
Opc == AMDGPU::G_LSHR ||
2576 Opc == AMDGPU::G_ASHR) {
2577 B.setInsertPt(*
MBB,
MI.getIterator());
2585 case AMDGPU::G_AMDGPU_S_MUL_I64_I32:
2586 case AMDGPU::G_AMDGPU_S_MUL_U64_U32: {
2600 Register SrcReg0 =
MI.getOperand(1).getReg();
2601 Register SrcReg1 =
MI.getOperand(2).getReg();
2604 assert(
MRI.getType(DstReg) ==
S64 &&
"This is a special case for s_mul_u64 "
2605 "that handles only 64-bit operands.");
2611 if (DstBank == &AMDGPU::SGPRRegBank) {
2612 MI.setDesc(
TII->get(AMDGPU::S_MUL_U64));
2613 MRI.setRegClass(DstReg, &AMDGPU::SGPR_64RegClass);
2614 MRI.setRegClass(SrcReg0, &AMDGPU::SGPR_64RegClass);
2615 MRI.setRegClass(SrcReg1, &AMDGPU::SGPR_64RegClass);
2621 assert(
MRI.getRegBankOrNull(DstReg) == &AMDGPU::VGPRRegBank &&
2622 "The destination operand should be in vector registers.");
2625 Register Op0L =
MRI.createVirtualRegister(&AMDGPU::VGPR_32RegClass);
2626 MRI.setRegClass(Op0L, &AMDGPU::VGPR_32RegClass);
2628 B.buildTrunc(Op0L, SrcReg0);
2631 Register Op1L =
MRI.createVirtualRegister(&AMDGPU::VGPR_32RegClass);
2632 MRI.setRegClass(Op1L, &AMDGPU::VGPR_32RegClass);
2634 B.buildTrunc(Op1L, SrcReg1);
2636 unsigned NewOpc =
Opc == AMDGPU::G_AMDGPU_S_MUL_U64_U32
2637 ? AMDGPU::G_AMDGPU_MAD_U64_U32
2638 : AMDGPU::G_AMDGPU_MAD_I64_I32;
2642 MRI.setRegClass(Zero64, &AMDGPU::VReg_64RegClass);
2643 Register CarryOut =
MRI.createVirtualRegister(&AMDGPU::VReg_64RegClass);
2644 MRI.setRegClass(CarryOut, &AMDGPU::VReg_64RegClass);
2645 B.buildInstr(NewOpc, {DstReg, CarryOut}, {Op0L, Op1L, Zero64});
2646 MI.eraseFromParent();
2649 case AMDGPU::G_SEXT_INREG: {
2651 if (SrcRegs.
empty())
2655 ApplyRegBankMapping O(
B, *
this,
MRI, &AMDGPU::VGPRRegBank);
2662 int Amt =
MI.getOperand(2).getImm();
2668 B.buildFreeze(DstRegs[0], SrcRegs[0]);
2670 auto Freeze =
B.buildFreeze(
S32, SrcRegs[0]);
2672 B.buildSExtInReg(DstRegs[0], Freeze, Amt);
2675 B.buildAShr(DstRegs[1], DstRegs[0],
B.buildConstant(
S32, 31));
2679 B.buildCopy(DstRegs[0], SrcRegs[0]);
2680 B.buildSExtInReg(DstRegs[1], DstRegs[0], Amt - 32);
2684 MRI.setRegBank(DstReg, AMDGPU::VGPRRegBank);
2685 MI.eraseFromParent();
2688 case AMDGPU::G_CTPOP:
2689 case AMDGPU::G_BITREVERSE: {
2692 if (DstBank == &AMDGPU::SGPRRegBank)
2697 LLT Ty =
MRI.getType(SrcReg);
2701 ApplyRegBankMapping ApplyVALU(
B, *
this,
MRI, &AMDGPU::VGPRRegBank);
2710 case AMDGPU::G_AMDGPU_FFBH_U32:
2711 case AMDGPU::G_AMDGPU_FFBL_B32:
2712 case AMDGPU::G_CTLZ_ZERO_UNDEF:
2713 case AMDGPU::G_CTTZ_ZERO_UNDEF: {
2716 if (DstBank == &AMDGPU::SGPRRegBank)
2721 LLT Ty =
MRI.getType(SrcReg);
2731 ApplyRegBankMapping ApplyVALU(
B, *
this,
MRI, &AMDGPU::VGPRRegBank);
2733 unsigned NewOpc =
Opc == AMDGPU::G_CTLZ_ZERO_UNDEF
2734 ? (
unsigned)AMDGPU::G_AMDGPU_FFBH_U32
2735 :
Opc == AMDGPU::G_CTTZ_ZERO_UNDEF
2736 ? (
unsigned)AMDGPU::G_AMDGPU_FFBL_B32
2738 unsigned Idx = NewOpc == AMDGPU::G_AMDGPU_FFBH_U32;
2739 auto X =
B.buildInstr(NewOpc, {
S32}, {SrcRegs[Idx]});
2740 auto Y =
B.buildInstr(NewOpc, {
S32}, {SrcRegs[Idx ^ 1]});
2742 Opc == AMDGPU::G_CTLZ_ZERO_UNDEF ||
Opc == AMDGPU::G_CTTZ_ZERO_UNDEF
2744 : AMDGPU::G_UADDSAT;
2745 Y =
B.buildInstr(AddOpc, {
S32}, {
Y,
B.buildConstant(
S32, 32)});
2747 B.buildUMin(DstReg,
X,
Y);
2748 MI.eraseFromParent();
2751 case AMDGPU::G_SEXT:
2752 case AMDGPU::G_ZEXT:
2753 case AMDGPU::G_ANYEXT: {
2755 LLT SrcTy =
MRI.getType(SrcReg);
2756 const bool Signed =
Opc == AMDGPU::G_SEXT;
2764 LLT DstTy =
MRI.getType(DstReg);
2766 SrcBank != &AMDGPU::SGPRRegBank &&
2767 SrcBank != &AMDGPU::VCCRegBank &&
2771 SrcTy.getSizeInBits() <= 32) {
2777 B.buildSExtOrTrunc(DefRegs[0], SrcReg);
2778 }
else if (
Opc == AMDGPU::G_ZEXT) {
2779 B.buildZExtOrTrunc(DefRegs[0], SrcReg);
2781 B.buildAnyExtOrTrunc(DefRegs[0], SrcReg);
2785 MRI.setRegBank(DstReg, *SrcBank);
2786 MI.eraseFromParent();
2796 if (SrcBank == &AMDGPU::VCCRegBank) {
2803 const bool UseSel64 = DstSize > 32 &&
2804 SrcBank->
getID() == AMDGPU::SGPRRegBankID;
2808 auto True =
B.buildConstant(SelType,
Signed ? -1 : 1);
2809 auto False =
B.buildConstant(SelType, 0);
2811 MRI.setRegBank(True.getReg(0), *DstBank);
2812 MRI.setRegBank(False.getReg(0), *DstBank);
2813 MRI.setRegBank(DstReg, *DstBank);
2816 B.buildSelect(DefRegs[0], SrcReg, True, False);
2818 }
else if (DstSize < 32) {
2819 auto Sel =
B.buildSelect(SelType, SrcReg, True, False);
2820 MRI.setRegBank(Sel.getReg(0), *DstBank);
2821 B.buildTrunc(DstReg, Sel);
2823 B.buildSelect(DstReg, SrcReg, True, False);
2826 MI.eraseFromParent();
2832 case AMDGPU::G_EXTRACT_VECTOR_ELT: {
2841 LLT DstTy =
MRI.getType(DstReg);
2842 LLT SrcTy =
MRI.getType(SrcReg);
2844 if (foldExtractEltToCmpSelect(
B,
MI, OpdMapper))
2856 unsigned ConstOffset;
2857 std::tie(BaseIdxReg, ConstOffset) =
2864 bool ShouldMoveIndexIntoLoop = IdxBank != &AMDGPU::SGPRRegBank &&
2866 ConstOffset < SrcTy.getNumElements();
2869 if (ShouldMoveIndexIntoLoop)
2870 MI.getOperand(2).setReg(BaseIdxReg);
2876 const bool NeedCopyToVGPR = DstBank == &AMDGPU::VGPRRegBank &&
2877 SrcBank == &AMDGPU::SGPRRegBank;
2878 if (DstRegs.
empty()) {
2883 if (NeedCopyToVGPR) {
2885 Register TmpReg =
MRI.createGenericVirtualRegister(DstTy);
2886 MRI.setRegBank(TmpReg, AMDGPU::SGPRRegBank);
2887 MI.getOperand(0).setReg(TmpReg);
2888 B.setInsertPt(*
MI.getParent(), ++
MI.getIterator());
2895 if (ShouldMoveIndexIntoLoop)
2905 auto CastSrc =
B.buildBitcast(Vec32, SrcReg);
2906 auto One =
B.buildConstant(
S32, 1);
2917 auto IdxLo =
B.buildShl(
S32, BaseIdxReg, One);
2918 auto IdxHi =
B.buildAdd(
S32, IdxLo, One);
2920 auto Extract0 =
B.buildExtractVectorElement(DstRegs[0], CastSrc, IdxLo);
2921 auto Extract1 =
B.buildExtractVectorElement(DstRegs[1], CastSrc, IdxHi);
2923 MRI.setRegBank(DstReg, *DstBank);
2924 MRI.setRegBank(CastSrc.getReg(0), *SrcBank);
2925 MRI.setRegBank(One.getReg(0), AMDGPU::SGPRRegBank);
2926 MRI.setRegBank(IdxLo.getReg(0), AMDGPU::SGPRRegBank);
2927 MRI.setRegBank(IdxHi.getReg(0), AMDGPU::SGPRRegBank);
2931 MI.eraseFromParent();
2937 B.setInstr(*Span.
begin());
2938 MI.eraseFromParent();
2942 if (NeedCopyToVGPR) {
2946 MRI.setRegBank(TmpReg0, AMDGPU::SGPRRegBank);
2947 MRI.setRegBank(TmpReg1, AMDGPU::SGPRRegBank);
2949 Extract0->getOperand(0).setReg(TmpReg0);
2950 Extract1->getOperand(0).setReg(TmpReg1);
2958 if (ShouldMoveIndexIntoLoop)
2963 case AMDGPU::G_INSERT_VECTOR_ELT: {
2967 LLT VecTy =
MRI.getType(DstReg);
2973 MRI.setType(
MI.getOperand(1).getReg(), VecTy);
2975 if (foldInsertEltToCmpSelect(
B,
MI, OpdMapper))
2983 LLT InsTy =
MRI.getType(InsReg);
2987 unsigned ConstOffset;
2988 std::tie(BaseIdxReg, ConstOffset) =
2995 bool ShouldMoveIndexIntoLoop = IdxBank != &AMDGPU::SGPRRegBank &&
3000 if (ShouldMoveIndexIntoLoop)
3001 MI.getOperand(3).setReg(BaseIdxReg);
3004 if (InsRegs.
empty()) {
3008 if (ShouldMoveIndexIntoLoop) {
3020 auto CastSrc =
B.buildBitcast(Vec32, SrcReg);
3021 auto One =
B.buildConstant(
S32, 1);
3030 auto IdxLo =
B.buildShl(
S32, BaseIdxReg, One);
3031 auto IdxHi =
B.buildAdd(
S32, IdxLo, One);
3033 auto InsLo =
B.buildInsertVectorElement(Vec32, CastSrc, InsRegs[0], IdxLo);
3034 auto InsHi =
B.buildInsertVectorElement(Vec32, InsLo, InsRegs[1], IdxHi);
3043 MRI.setRegBank(InsReg, *InsSrcBank);
3044 MRI.setRegBank(CastSrc.getReg(0), *SrcBank);
3045 MRI.setRegBank(InsLo.getReg(0), *DstBank);
3046 MRI.setRegBank(InsHi.getReg(0), *DstBank);
3047 MRI.setRegBank(One.getReg(0), AMDGPU::SGPRRegBank);
3048 MRI.setRegBank(IdxLo.getReg(0), AMDGPU::SGPRRegBank);
3049 MRI.setRegBank(IdxHi.getReg(0), AMDGPU::SGPRRegBank);
3054 B.setInsertPt(
B.getMBB(),
MI);
3055 B.buildBitcast(DstReg, InsHi);
3056 MI.eraseFromParent();
3060 B.setInstr(*Span.
begin());
3061 MI.eraseFromParent();
3072 B.buildBitcast(DstReg, InsHi);
3075 if (ShouldMoveIndexIntoLoop)
3080 case AMDGPU::G_AMDGPU_BUFFER_LOAD:
3081 case AMDGPU::G_AMDGPU_BUFFER_LOAD_USHORT:
3082 case AMDGPU::G_AMDGPU_BUFFER_LOAD_SSHORT:
3083 case AMDGPU::G_AMDGPU_BUFFER_LOAD_UBYTE:
3084 case AMDGPU::G_AMDGPU_BUFFER_LOAD_SBYTE:
3085 case AMDGPU::G_AMDGPU_BUFFER_LOAD_TFE:
3086 case AMDGPU::G_AMDGPU_BUFFER_LOAD_USHORT_TFE:
3087 case AMDGPU::G_AMDGPU_BUFFER_LOAD_SSHORT_TFE:
3088 case AMDGPU::G_AMDGPU_BUFFER_LOAD_UBYTE_TFE:
3089 case AMDGPU::G_AMDGPU_BUFFER_LOAD_SBYTE_TFE:
3090 case AMDGPU::G_AMDGPU_BUFFER_LOAD_FORMAT:
3091 case AMDGPU::G_AMDGPU_BUFFER_LOAD_FORMAT_TFE:
3092 case AMDGPU::G_AMDGPU_BUFFER_LOAD_FORMAT_D16:
3093 case AMDGPU::G_AMDGPU_TBUFFER_LOAD_FORMAT:
3094 case AMDGPU::G_AMDGPU_TBUFFER_LOAD_FORMAT_D16:
3095 case AMDGPU::G_AMDGPU_BUFFER_STORE:
3096 case AMDGPU::G_AMDGPU_BUFFER_STORE_BYTE:
3097 case AMDGPU::G_AMDGPU_BUFFER_STORE_SHORT:
3098 case AMDGPU::G_AMDGPU_BUFFER_STORE_FORMAT:
3099 case AMDGPU::G_AMDGPU_BUFFER_STORE_FORMAT_D16:
3100 case AMDGPU::G_AMDGPU_TBUFFER_STORE_FORMAT:
3101 case AMDGPU::G_AMDGPU_TBUFFER_STORE_FORMAT_D16: {
3106 case AMDGPU::G_AMDGPU_BUFFER_ATOMIC_SWAP:
3107 case AMDGPU::G_AMDGPU_BUFFER_ATOMIC_ADD:
3108 case AMDGPU::G_AMDGPU_BUFFER_ATOMIC_SUB:
3109 case AMDGPU::G_AMDGPU_BUFFER_ATOMIC_SMIN:
3110 case AMDGPU::G_AMDGPU_BUFFER_ATOMIC_UMIN:
3111 case AMDGPU::G_AMDGPU_BUFFER_ATOMIC_SMAX:
3112 case AMDGPU::G_AMDGPU_BUFFER_ATOMIC_UMAX:
3113 case AMDGPU::G_AMDGPU_BUFFER_ATOMIC_AND:
3114 case AMDGPU::G_AMDGPU_BUFFER_ATOMIC_OR:
3115 case AMDGPU::G_AMDGPU_BUFFER_ATOMIC_XOR:
3116 case AMDGPU::G_AMDGPU_BUFFER_ATOMIC_INC:
3117 case AMDGPU::G_AMDGPU_BUFFER_ATOMIC_DEC:
3118 case AMDGPU::G_AMDGPU_BUFFER_ATOMIC_SUB_CLAMP_U32:
3119 case AMDGPU::G_AMDGPU_BUFFER_ATOMIC_COND_SUB_U32:
3120 case AMDGPU::G_AMDGPU_BUFFER_ATOMIC_FADD:
3121 case AMDGPU::G_AMDGPU_BUFFER_ATOMIC_FMIN:
3122 case AMDGPU::G_AMDGPU_BUFFER_ATOMIC_FMAX: {
3127 case AMDGPU::G_AMDGPU_BUFFER_ATOMIC_CMPSWAP: {
3132 case AMDGPU::G_AMDGPU_S_BUFFER_LOAD:
3133 case AMDGPU::G_AMDGPU_S_BUFFER_LOAD_UBYTE:
3134 case AMDGPU::G_AMDGPU_S_BUFFER_LOAD_SBYTE:
3135 case AMDGPU::G_AMDGPU_S_BUFFER_LOAD_USHORT:
3136 case AMDGPU::G_AMDGPU_S_BUFFER_LOAD_SSHORT: {
3140 case AMDGPU::G_AMDGPU_S_BUFFER_PREFETCH:
3144 case AMDGPU::G_INTRINSIC:
3145 case AMDGPU::G_INTRINSIC_CONVERGENT: {
3147 case Intrinsic::amdgcn_readlane: {
3158 case Intrinsic::amdgcn_writelane: {
3168 case Intrinsic::amdgcn_interp_p1:
3169 case Intrinsic::amdgcn_interp_p2:
3170 case Intrinsic::amdgcn_interp_mov:
3171 case Intrinsic::amdgcn_interp_p1_f16:
3172 case Intrinsic::amdgcn_interp_p2_f16:
3173 case Intrinsic::amdgcn_lds_param_load: {
3181 case Intrinsic::amdgcn_interp_inreg_p10:
3182 case Intrinsic::amdgcn_interp_inreg_p2:
3183 case Intrinsic::amdgcn_interp_inreg_p10_f16:
3184 case Intrinsic::amdgcn_interp_inreg_p2_f16:
3185 case Intrinsic::amdgcn_interp_p10_rtz_f16:
3186 case Intrinsic::amdgcn_interp_p2_rtz_f16:
3187 case Intrinsic::amdgcn_permlane16_swap:
3188 case Intrinsic::amdgcn_permlane32_swap:
3191 case Intrinsic::amdgcn_permlane16:
3192 case Intrinsic::amdgcn_permlanex16: {
3200 case Intrinsic::amdgcn_permlane_bcast:
3201 case Intrinsic::amdgcn_permlane_up:
3202 case Intrinsic::amdgcn_permlane_down:
3203 case Intrinsic::amdgcn_permlane_xor:
3208 case Intrinsic::amdgcn_permlane_idx_gen: {
3212 case Intrinsic::amdgcn_sbfe:
3215 case Intrinsic::amdgcn_ubfe:
3218 case Intrinsic::amdgcn_inverse_ballot:
3219 case Intrinsic::amdgcn_s_bitreplicate:
3220 case Intrinsic::amdgcn_s_quadmask:
3221 case Intrinsic::amdgcn_s_wqm:
3225 case Intrinsic::amdgcn_ballot:
3231 case AMDGPU::G_AMDGPU_INTRIN_IMAGE_LOAD:
3232 case AMDGPU::G_AMDGPU_INTRIN_IMAGE_LOAD_D16:
3233 case AMDGPU::G_AMDGPU_INTRIN_IMAGE_LOAD_NORET:
3234 case AMDGPU::G_AMDGPU_INTRIN_IMAGE_STORE:
3235 case AMDGPU::G_AMDGPU_INTRIN_IMAGE_STORE_D16: {
3245 case AMDGPU::G_AMDGPU_BVH_INTERSECT_RAY:
3246 case AMDGPU::G_AMDGPU_BVH8_INTERSECT_RAY:
3247 case AMDGPU::G_AMDGPU_BVH_DUAL_INTERSECT_RAY: {
3249 MI.getOpcode() == AMDGPU::G_AMDGPU_BVH_DUAL_INTERSECT_RAY ||
3250 MI.getOpcode() == AMDGPU::G_AMDGPU_BVH8_INTERSECT_RAY;
3251 unsigned NumMods = IsDualOrBVH8 ? 0 : 1;
3252 unsigned LastRegOpIdx =
MI.getNumExplicitOperands() - 1 - NumMods;
3257 case AMDGPU::G_INTRINSIC_W_SIDE_EFFECTS:
3258 case AMDGPU::G_INTRINSIC_CONVERGENT_W_SIDE_EFFECTS: {
3261 case Intrinsic::amdgcn_ds_ordered_add:
3262 case Intrinsic::amdgcn_ds_ordered_swap: {
3269 case Intrinsic::amdgcn_ds_gws_init:
3270 case Intrinsic::amdgcn_ds_gws_barrier:
3271 case Intrinsic::amdgcn_ds_gws_sema_br: {
3277 case Intrinsic::amdgcn_ds_gws_sema_v:
3278 case Intrinsic::amdgcn_ds_gws_sema_p:
3279 case Intrinsic::amdgcn_ds_gws_sema_release_all: {
3284 case Intrinsic::amdgcn_ds_append:
3285 case Intrinsic::amdgcn_ds_consume: {
3289 case Intrinsic::amdgcn_s_sendmsg:
3290 case Intrinsic::amdgcn_s_sendmsghalt: {
3295 case Intrinsic::amdgcn_s_setreg: {
3299 case Intrinsic::amdgcn_s_ttracedata:
3302 case Intrinsic::amdgcn_raw_buffer_load_lds:
3303 case Intrinsic::amdgcn_raw_ptr_buffer_load_lds: {
3310 case Intrinsic::amdgcn_struct_buffer_load_lds:
3311 case Intrinsic::amdgcn_struct_ptr_buffer_load_lds: {
3318 case Intrinsic::amdgcn_cluster_load_async_to_lds_b8:
3319 case Intrinsic::amdgcn_cluster_load_async_to_lds_b32:
3320 case Intrinsic::amdgcn_cluster_load_async_to_lds_b64:
3321 case Intrinsic::amdgcn_cluster_load_async_to_lds_b128: {
3326 case Intrinsic::amdgcn_load_to_lds:
3327 case Intrinsic::amdgcn_global_load_lds: {
3332 case Intrinsic::amdgcn_lds_direct_load: {
3338 case Intrinsic::amdgcn_exp_row:
3342 case Intrinsic::amdgcn_cluster_load_b32:
3343 case Intrinsic::amdgcn_cluster_load_b64:
3344 case Intrinsic::amdgcn_cluster_load_b128: {
3349 case Intrinsic::amdgcn_s_sleep_var:
3353 case Intrinsic::amdgcn_s_barrier_join:
3354 case Intrinsic::amdgcn_s_wakeup_barrier:
3357 case Intrinsic::amdgcn_s_barrier_init:
3358 case Intrinsic::amdgcn_s_barrier_signal_var:
3362 case Intrinsic::amdgcn_s_get_barrier_state:
3363 case Intrinsic::amdgcn_s_get_named_barrier_state: {
3367 case Intrinsic::amdgcn_s_prefetch_data: {
3369 unsigned AS =
MRI.getType(PtrReg).getAddressSpace();
3374 MI.eraseFromParent();
3377 case Intrinsic::amdgcn_tensor_load_to_lds:
3378 case Intrinsic::amdgcn_tensor_store_from_lds: {
3385 case Intrinsic::amdgcn_tensor_load_to_lds_d2:
3386 case Intrinsic::amdgcn_tensor_store_from_lds_d2: {
3397 if (RSrcIntrin->IsImage) {
3408 case AMDGPU::G_SI_CALL: {
3419 unsigned FrameSetupOpcode = AMDGPU::ADJCALLSTACKUP;
3420 unsigned FrameDestroyOpcode = AMDGPU::ADJCALLSTACKDOWN;
3426 unsigned NonCopyInstrsLen = 0;
3432 while (Start->getOpcode() != FrameSetupOpcode) {
3434 bool IsCopy =
false;
3435 if (Start->getOpcode() == AMDGPU::COPY) {
3436 auto &Dst = Start->getOperand(0);
3439 if (Reg.isPhysical() &&
MI.readsRegister(Reg,
TRI)) {
3444 auto &Src = Start->getOperand(1);
3447 IsCopy = Info->getScratchRSrcReg() == Reg;
3455 NonCopyInstrsLen = NonCopyInstrs.
size();
3460 NonCopyInstrs.
resize(NonCopyInstrsLen);
3462 for (
auto *NonCopy :
reverse(NonCopyInstrs)) {
3463 MBB->splice(LastCopy,
MBB, NonCopy->getIterator());
3468 NonCopyInstrs.
clear();
3469 NonCopyInstrsLen = 0;
3472 while (End->getOpcode() != FrameDestroyOpcode) {
3474 bool IsCopy =
false;
3475 if (End->getOpcode() == AMDGPU::COPY) {
3476 auto &Src = End->getOperand(1);
3479 IsCopy = Reg.isPhysical() &&
MI.modifiesRegister(Reg,
TRI);
3485 NonCopyInstrsLen = NonCopyInstrs.
size();
3490 NonCopyInstrs.
resize(NonCopyInstrsLen);
3494 for (
auto *NonCopy :
reverse(NonCopyInstrs)) {
3495 MBB->splice(LastCopy,
MBB, NonCopy->getIterator());
3499 B.setInsertPt(
B.getMBB(), Start);
3503 case AMDGPU::G_LOAD:
3504 case AMDGPU::G_ZEXTLOAD:
3505 case AMDGPU::G_SEXTLOAD: {
3510 case AMDGPU::G_DYN_STACKALLOC:
3513 case AMDGPU::G_STACKRESTORE: {
3518 case AMDGPU::G_SBFX:
3521 case AMDGPU::G_UBFX:
3524 case AMDGPU::G_AMDGPU_MAD_U64_U32:
3525 case AMDGPU::G_AMDGPU_MAD_I64_I32:
3528 case AMDGPU::G_PREFETCH: {
3530 MI.eraseFromParent();
3535 if (PtrBank == AMDGPU::VGPRRegBankID &&
3536 (!
Subtarget.hasVmemPrefInsts() || !
MI.getOperand(3).getImm())) {
3538 MI.eraseFromParent();
3541 unsigned AS =
MRI.getType(PtrReg).getAddressSpace();
3546 !
MI.getOperand(3).getImm() ))) {
3547 MI.eraseFromParent();
3844 if (
MI.isCopy() ||
MI.getOpcode() == AMDGPU::G_FREEZE) {
3859 DstBank = &AMDGPU::VCCRegBank;
3862 DstBank = &AMDGPU::VCCRegBank;
3873 if (
MI.getOpcode() != AMDGPU::G_FREEZE &&
3878 unsigned OpdsMappingSize =
MI.isCopy() ? 1 : 2;
3880 OpdsMapping[0] = &ValMap;
3881 if (
MI.getOpcode() == AMDGPU::G_FREEZE)
3882 OpdsMapping[1] = &ValMap;
3889 if (
MI.isRegSequence()) {
3892 unsigned BankID = AMDGPU::SGPRRegBankID;
3894 for (
unsigned I = 1, E =
MI.getNumOperands();
I != E;
I += 2) {
3898 if (OpBank != AMDGPU::SGPRRegBankID) {
3899 BankID = AMDGPU::VGPRRegBankID;
3916 unsigned ResultBank = AMDGPU::InvalidRegBankID;
3921 ResultBank = DstBank->
getID();
3923 for (
unsigned I = 0;
I <
PHI->getNumIncomingValues(); ++
I) {
3928 if (!Bank || Bank->
getID() == AMDGPU::VGPRRegBankID) {
3929 ResultBank = AMDGPU::VGPRRegBankID;
3934 unsigned OpBank = Bank->
getID();
3938 assert(ResultBank != AMDGPU::InvalidRegBankID);
3940 unsigned Size =
MRI.getType(DstReg).getSizeInBits();
3955 switch (
MI.getOpcode()) {
3962 case AMDGPU::G_MUL: {
3963 unsigned Size =
MRI.getType(
MI.getOperand(0).getReg()).getSizeInBits();
3968 unsigned TargetBankID = AMDGPU::InvalidRegBankID;
3969 unsigned BankLHS = AMDGPU::InvalidRegBankID;
3970 unsigned BankRHS = AMDGPU::InvalidRegBankID;
3972 TargetBankID = DstBank->
getID();
3973 if (DstBank == &AMDGPU::VCCRegBank) {
3974 TargetBankID = AMDGPU::VCCRegBankID;
3975 BankLHS = AMDGPU::VCCRegBankID;
3976 BankRHS = AMDGPU::VCCRegBankID;
3979 AMDGPU::SGPRRegBankID);
3981 AMDGPU::SGPRRegBankID);
3985 AMDGPU::VCCRegBankID);
3987 AMDGPU::VCCRegBankID);
3990 if (BankLHS == AMDGPU::VGPRRegBankID || BankRHS == AMDGPU::VGPRRegBankID) {
3991 TargetBankID = AMDGPU::VGPRRegBankID;
3992 }
else if (BankLHS == AMDGPU::VCCRegBankID || BankRHS == AMDGPU::VCCRegBankID) {
3993 TargetBankID = AMDGPU::VCCRegBankID;
3994 BankLHS = AMDGPU::VCCRegBankID;
3995 BankRHS = AMDGPU::VCCRegBankID;
3996 }
else if (BankLHS == AMDGPU::SGPRRegBankID && BankRHS == AMDGPU::SGPRRegBankID) {
3997 TargetBankID = AMDGPU::SGPRRegBankID;
4001 OpdsMapping[0] = AMDGPU::getValueMapping(TargetBankID,
Size);
4002 OpdsMapping[1] = AMDGPU::getValueMapping(BankLHS,
Size);
4003 OpdsMapping[2] = AMDGPU::getValueMapping(BankRHS,
Size);
4010 OpdsMapping[0] = getValueMappingSGPR64Only(AMDGPU::SGPRRegBankID,
Size);
4011 OpdsMapping[1] = OpdsMapping[2] = OpdsMapping[0];
4013 if (
MI.getOpcode() == AMDGPU::G_MUL &&
Subtarget.hasVectorMulU64())
4014 OpdsMapping[0] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID,
Size);
4017 getValueMappingSGPR64Only(AMDGPU::VGPRRegBankID,
Size);
4019 OpdsMapping[1] = AMDGPU::getValueMapping(Bank1,
Size);
4022 OpdsMapping[2] = AMDGPU::getValueMapping(Bank2,
Size);
4030 case AMDGPU::G_PTR_ADD:
4031 case AMDGPU::G_PTRMASK:
4035 case AMDGPU::G_LSHR:
4036 case AMDGPU::G_ASHR:
4037 case AMDGPU::G_UADDO:
4038 case AMDGPU::G_USUBO:
4039 case AMDGPU::G_UADDE:
4040 case AMDGPU::G_SADDE:
4041 case AMDGPU::G_USUBE:
4042 case AMDGPU::G_SSUBE:
4044 case AMDGPU::G_SHUFFLE_VECTOR:
4045 case AMDGPU::G_SBFX:
4046 case AMDGPU::G_UBFX:
4047 case AMDGPU::G_AMDGPU_S_MUL_I64_I32:
4048 case AMDGPU::G_AMDGPU_S_MUL_U64_U32:
4052 case AMDGPU::G_SMIN:
4053 case AMDGPU::G_SMAX:
4054 case AMDGPU::G_UMIN:
4055 case AMDGPU::G_UMAX:
4058 if (
MRI.getType(
MI.getOperand(0).getReg()).getSizeInBits() == 64 &&
4064 case AMDGPU::G_FADD:
4065 case AMDGPU::G_FSUB:
4066 case AMDGPU::G_FMUL:
4068 case AMDGPU::G_FFLOOR:
4069 case AMDGPU::G_FCEIL:
4070 case AMDGPU::G_INTRINSIC_ROUNDEVEN:
4071 case AMDGPU::G_FMINNUM:
4072 case AMDGPU::G_FMAXNUM:
4073 case AMDGPU::G_FMINIMUM:
4074 case AMDGPU::G_FMAXIMUM:
4075 case AMDGPU::G_FMINIMUMNUM:
4076 case AMDGPU::G_FMAXIMUMNUM:
4077 case AMDGPU::G_INTRINSIC_TRUNC:
4078 case AMDGPU::G_STRICT_FADD:
4079 case AMDGPU::G_STRICT_FSUB:
4080 case AMDGPU::G_STRICT_FMUL:
4081 case AMDGPU::G_STRICT_FMA: {
4082 LLT Ty =
MRI.getType(
MI.getOperand(0).getReg());
4083 unsigned Size = Ty.getSizeInBits();
4084 if (
Subtarget.hasSALUFloatInsts() && Ty.isScalar() &&
4089 case AMDGPU::G_FPTOSI:
4090 case AMDGPU::G_FPTOUI:
4091 case AMDGPU::G_FPTOSI_SAT:
4092 case AMDGPU::G_FPTOUI_SAT:
4093 case AMDGPU::G_SITOFP:
4094 case AMDGPU::G_UITOFP: {
4095 unsigned SizeDst =
MRI.getType(
MI.getOperand(0).getReg()).getSizeInBits();
4096 unsigned SizeSrc =
MRI.getType(
MI.getOperand(1).getReg()).getSizeInBits();
4097 if (
Subtarget.hasSALUFloatInsts() && SizeDst == 32 && SizeSrc == 32 &&
4102 case AMDGPU::G_FPTRUNC:
4103 case AMDGPU::G_FPEXT: {
4104 unsigned SizeDst =
MRI.getType(
MI.getOperand(0).getReg()).getSizeInBits();
4105 unsigned SizeSrc =
MRI.getType(
MI.getOperand(1).getReg()).getSizeInBits();
4106 if (
Subtarget.hasSALUFloatInsts() && SizeDst != 64 && SizeSrc != 64 &&
4111 case AMDGPU::G_FSQRT:
4112 case AMDGPU::G_FEXP2:
4113 case AMDGPU::G_FLOG2: {
4114 unsigned Size =
MRI.getType(
MI.getOperand(0).getReg()).getSizeInBits();
4120 case AMDGPU::G_SADDSAT:
4121 case AMDGPU::G_SSUBSAT:
4122 case AMDGPU::G_UADDSAT:
4123 case AMDGPU::G_USUBSAT:
4124 case AMDGPU::G_FMAD:
4125 case AMDGPU::G_FLDEXP:
4126 case AMDGPU::G_FMINNUM_IEEE:
4127 case AMDGPU::G_FMAXNUM_IEEE:
4128 case AMDGPU::G_FCANONICALIZE:
4129 case AMDGPU::G_STRICT_FLDEXP:
4130 case AMDGPU::G_BSWAP:
4131 case AMDGPU::G_FSHR:
4132 case AMDGPU::G_AMDGPU_FMIN_LEGACY:
4133 case AMDGPU::G_AMDGPU_FMAX_LEGACY:
4134 case AMDGPU::G_AMDGPU_RCP_IFLAG:
4135 case AMDGPU::G_AMDGPU_CVT_F32_UBYTE0:
4136 case AMDGPU::G_AMDGPU_CVT_F32_UBYTE1:
4137 case AMDGPU::G_AMDGPU_CVT_F32_UBYTE2:
4138 case AMDGPU::G_AMDGPU_CVT_F32_UBYTE3:
4139 case AMDGPU::G_AMDGPU_CVT_PK_I16_I32:
4140 case AMDGPU::G_AMDGPU_SMED3:
4141 case AMDGPU::G_AMDGPU_FMED3:
4143 case AMDGPU::G_UMULH:
4144 case AMDGPU::G_SMULH: {
4149 case AMDGPU::G_AMDGPU_MAD_U64_U32:
4150 case AMDGPU::G_AMDGPU_MAD_I64_I32: {
4159 bool AllSalu =
true;
4160 bool MulSalu =
true;
4161 for (
unsigned i = 0; i < 5; ++i) {
4164 if (Bank->getID() != AMDGPU::SGPRRegBankID) {
4166 if (i == 2 || i == 3) {
4180 if (!MulSalu ||
Subtarget.hasFullRate64Ops())
4184 OpdsMapping[0] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID, 64);
4185 OpdsMapping[1] = AMDGPU::getValueMapping(AMDGPU::VCCRegBankID, 1);
4186 OpdsMapping[2] = AMDGPU::getValueMapping(AMDGPU::SGPRRegBankID, 32);
4187 OpdsMapping[3] = AMDGPU::getValueMapping(AMDGPU::SGPRRegBankID, 32);
4188 OpdsMapping[4] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID, 64);
4191 case AMDGPU::G_IMPLICIT_DEF: {
4192 unsigned Size =
MRI.getType(
MI.getOperand(0).getReg()).getSizeInBits();
4193 OpdsMapping[0] = AMDGPU::getValueMapping(AMDGPU::SGPRRegBankID,
Size);
4196 case AMDGPU::G_FCONSTANT:
4197 case AMDGPU::G_CONSTANT:
4198 case AMDGPU::G_GLOBAL_VALUE:
4199 case AMDGPU::G_FRAME_INDEX:
4200 case AMDGPU::G_BLOCK_ADDR:
4201 case AMDGPU::G_READSTEADYCOUNTER:
4202 case AMDGPU::G_READCYCLECOUNTER: {
4203 unsigned Size =
MRI.getType(
MI.getOperand(0).getReg()).getSizeInBits();
4204 OpdsMapping[0] = AMDGPU::getValueMapping(AMDGPU::SGPRRegBankID,
Size);
4207 case AMDGPU::G_DYN_STACKALLOC: {
4209 OpdsMapping[0] = AMDGPU::getValueMapping(AMDGPU::SGPRRegBankID, 32);
4211 OpdsMapping[1] = AMDGPU::getValueMapping(SrcBankID, 32);
4214 case AMDGPU::G_AMDGPU_WAVE_ADDRESS: {
4219 OpdsMapping[0] = AMDGPU::getValueMapping(AMDGPU::SGPRRegBankID, 32);
4220 OpdsMapping[1] = AMDGPU::getValueMapping(AMDGPU::SGPRRegBankID, 32);
4223 case AMDGPU::G_INSERT: {
4228 OpdsMapping[0] = AMDGPU::getValueMapping(BankID, DstSize);
4229 OpdsMapping[1] = AMDGPU::getValueMapping(BankID, SrcSize);
4230 OpdsMapping[2] = AMDGPU::getValueMapping(BankID, EltSize);
4231 OpdsMapping[3] =
nullptr;
4234 case AMDGPU::G_EXTRACT: {
4238 OpdsMapping[0] = AMDGPU::getValueMapping(BankID, DstSize);
4239 OpdsMapping[1] = AMDGPU::getValueMapping(BankID, SrcSize);
4240 OpdsMapping[2] =
nullptr;
4243 case AMDGPU::G_BUILD_VECTOR:
4244 case AMDGPU::G_BUILD_VECTOR_TRUNC: {
4245 LLT DstTy =
MRI.getType(
MI.getOperand(0).getReg());
4248 unsigned SrcSize =
MRI.getType(
MI.getOperand(1).getReg()).getSizeInBits();
4251 unsigned DstBankID =
regBankUnion(Src0BankID, Src1BankID);
4253 OpdsMapping[0] = AMDGPU::getValueMapping(DstBankID, DstSize);
4254 OpdsMapping[1] = AMDGPU::getValueMapping(Src0BankID, SrcSize);
4255 OpdsMapping[2] = AMDGPU::getValueMapping(Src1BankID, SrcSize);
4261 case AMDGPU::G_MERGE_VALUES:
4262 case AMDGPU::G_CONCAT_VECTORS: {
4264 unsigned DstSize =
MRI.getType(
MI.getOperand(0).getReg()).getSizeInBits();
4265 unsigned SrcSize =
MRI.getType(
MI.getOperand(1).getReg()).getSizeInBits();
4267 OpdsMapping[0] = AMDGPU::getValueMapping(Bank, DstSize);
4269 for (
unsigned i = 1, e =
MI.getNumOperands(); i != e; ++i)
4270 OpdsMapping[i] = AMDGPU::getValueMapping(Bank, SrcSize);
4273 case AMDGPU::G_BITREVERSE:
4274 case AMDGPU::G_BITCAST:
4275 case AMDGPU::G_INTTOPTR:
4276 case AMDGPU::G_PTRTOINT:
4277 case AMDGPU::G_FABS:
4278 case AMDGPU::G_FNEG: {
4279 unsigned Size =
MRI.getType(
MI.getOperand(0).getReg()).getSizeInBits();
4281 OpdsMapping[0] = OpdsMapping[1] = AMDGPU::getValueMapping(BankID,
Size);
4284 case AMDGPU::G_AMDGPU_FFBH_U32:
4285 case AMDGPU::G_AMDGPU_FFBL_B32:
4286 case AMDGPU::G_CTLZ_ZERO_UNDEF:
4287 case AMDGPU::G_CTTZ_ZERO_UNDEF: {
4288 unsigned Size =
MRI.getType(
MI.getOperand(1).getReg()).getSizeInBits();
4290 OpdsMapping[0] = AMDGPU::getValueMapping(BankID, 32);
4291 OpdsMapping[1] = AMDGPU::getValueMappingSGPR64Only(BankID,
Size);
4294 case AMDGPU::G_CTPOP: {
4295 unsigned Size =
MRI.getType(
MI.getOperand(1).getReg()).getSizeInBits();
4297 OpdsMapping[0] = AMDGPU::getValueMapping(BankID, 32);
4302 OpdsMapping[1] = AMDGPU::getValueMapping(BankID,
Size);
4305 case AMDGPU::G_TRUNC: {
4311 OpdsMapping[0] = AMDGPU::getValueMapping(Bank, DstSize);
4312 OpdsMapping[1] = AMDGPU::getValueMapping(Bank, SrcSize);
4315 case AMDGPU::G_ZEXT:
4316 case AMDGPU::G_SEXT:
4317 case AMDGPU::G_ANYEXT:
4318 case AMDGPU::G_SEXT_INREG: {
4327 switch (SrcBank->
getID()) {
4328 case AMDGPU::SGPRRegBankID:
4329 DstBank = AMDGPU::SGPRRegBankID;
4332 DstBank = AMDGPU::VGPRRegBankID;
4338 OpdsMapping[0] = AMDGPU::getValueMappingSGPR64Only(DstBank, DstSize);
4339 OpdsMapping[1] = AMDGPU::getValueMappingSGPR64Only(SrcBank->
getID(),
4343 case AMDGPU::G_IS_FPCLASS: {
4345 unsigned SrcSize =
MRI.getType(SrcReg).getSizeInBits();
4346 unsigned DstSize =
MRI.getType(
MI.getOperand(0).getReg()).getSizeInBits();
4347 OpdsMapping[0] = AMDGPU::getValueMapping(AMDGPU::VCCRegBankID, DstSize);
4348 OpdsMapping[1] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID, SrcSize);
4351 case AMDGPU::G_STORE: {
4353 unsigned Size =
MRI.getType(
MI.getOperand(0).getReg()).getSizeInBits();
4358 AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID,
Size);
4359 OpdsMapping[0] = ValMapping;
4363 case AMDGPU::G_ICMP:
4364 case AMDGPU::G_FCMP: {
4365 unsigned Size =
MRI.getType(
MI.getOperand(2).getReg()).getSizeInBits();
4370 AMDGPU::SGPRRegBankID);
4374 auto canUseSCCICMP = [&]() {
4377 return Size == 32 ||
4382 auto canUseSCCFCMP = [&]() {
4386 bool isICMP =
MI.getOpcode() == AMDGPU::G_ICMP;
4387 bool CanUseSCC = DstBank == AMDGPU::SGPRRegBankID &&
4388 Op2Bank == AMDGPU::SGPRRegBankID &&
4389 Op3Bank == AMDGPU::SGPRRegBankID &&
4390 (isICMP ? canUseSCCICMP() : canUseSCCFCMP());
4392 DstBank = CanUseSCC ? AMDGPU::SGPRRegBankID : AMDGPU::VCCRegBankID;
4393 unsigned SrcBank = CanUseSCC ? AMDGPU::SGPRRegBankID : AMDGPU::VGPRRegBankID;
4397 const unsigned ResultSize = 1;
4399 OpdsMapping[0] = AMDGPU::getValueMapping(DstBank, ResultSize);
4400 OpdsMapping[1] =
nullptr;
4401 OpdsMapping[2] = AMDGPU::getValueMapping(SrcBank,
Size);
4402 OpdsMapping[3] = AMDGPU::getValueMapping(SrcBank,
Size);
4405 case AMDGPU::G_EXTRACT_VECTOR_ELT: {
4408 unsigned DstSize =
MRI.getType(
MI.getOperand(0).getReg()).getSizeInBits();
4409 unsigned SrcSize =
MRI.getType(
MI.getOperand(1).getReg()).getSizeInBits();
4410 unsigned IdxSize =
MRI.getType(
MI.getOperand(2).getReg()).getSizeInBits();
4412 unsigned OutputBankID =
regBankUnion(SrcBankID, IdxBank);
4414 OpdsMapping[0] = AMDGPU::getValueMappingSGPR64Only(OutputBankID, DstSize);
4415 OpdsMapping[1] = AMDGPU::getValueMapping(SrcBankID, SrcSize);
4418 OpdsMapping[2] = AMDGPU::getValueMapping(IdxBank, IdxSize);
4421 case AMDGPU::G_INSERT_VECTOR_ELT: {
4423 AMDGPU::SGPRRegBankID : AMDGPU::VGPRRegBankID;
4425 unsigned VecSize =
MRI.getType(
MI.getOperand(0).getReg()).getSizeInBits();
4426 unsigned InsertSize =
MRI.getType(
MI.getOperand(2).getReg()).getSizeInBits();
4427 unsigned IdxSize =
MRI.getType(
MI.getOperand(3).getReg()).getSizeInBits();
4431 OpdsMapping[0] = AMDGPU::getValueMapping(OutputBankID, VecSize);
4432 OpdsMapping[1] = AMDGPU::getValueMapping(OutputBankID, VecSize);
4436 if (InsertSize == 64 && OutputBankID == AMDGPU::VGPRRegBankID) {
4437 OpdsMapping[2] = AMDGPU::getValueMappingSplit64(InsertEltBankID,
4440 assert(InsertSize == 32 || InsertSize == 64);
4441 OpdsMapping[2] = AMDGPU::getValueMapping(InsertEltBankID, InsertSize);
4445 OpdsMapping[3] = AMDGPU::getValueMapping(IdxBankID, IdxSize);
4448 case AMDGPU::G_UNMERGE_VALUES: {
4453 for (
unsigned i = 0, e =
MI.getNumOperands(); i != e; ++i) {
4455 OpdsMapping[i] = AMDGPU::getValueMapping(Bank,
Size);
4459 case AMDGPU::G_AMDGPU_BUFFER_LOAD:
4460 case AMDGPU::G_AMDGPU_BUFFER_LOAD_UBYTE:
4461 case AMDGPU::G_AMDGPU_BUFFER_LOAD_SBYTE:
4462 case AMDGPU::G_AMDGPU_BUFFER_LOAD_USHORT:
4463 case AMDGPU::G_AMDGPU_BUFFER_LOAD_SSHORT:
4464 case AMDGPU::G_AMDGPU_BUFFER_LOAD_TFE:
4465 case AMDGPU::G_AMDGPU_BUFFER_LOAD_UBYTE_TFE:
4466 case AMDGPU::G_AMDGPU_BUFFER_LOAD_SBYTE_TFE:
4467 case AMDGPU::G_AMDGPU_BUFFER_LOAD_USHORT_TFE:
4468 case AMDGPU::G_AMDGPU_BUFFER_LOAD_SSHORT_TFE:
4469 case AMDGPU::G_AMDGPU_BUFFER_LOAD_FORMAT:
4470 case AMDGPU::G_AMDGPU_BUFFER_LOAD_FORMAT_TFE:
4471 case AMDGPU::G_AMDGPU_BUFFER_LOAD_FORMAT_D16:
4472 case AMDGPU::G_AMDGPU_TBUFFER_LOAD_FORMAT:
4473 case AMDGPU::G_AMDGPU_TBUFFER_LOAD_FORMAT_D16:
4474 case AMDGPU::G_AMDGPU_TBUFFER_STORE_FORMAT:
4475 case AMDGPU::G_AMDGPU_TBUFFER_STORE_FORMAT_D16:
4476 case AMDGPU::G_AMDGPU_BUFFER_STORE:
4477 case AMDGPU::G_AMDGPU_BUFFER_STORE_BYTE:
4478 case AMDGPU::G_AMDGPU_BUFFER_STORE_SHORT:
4479 case AMDGPU::G_AMDGPU_BUFFER_STORE_FORMAT:
4480 case AMDGPU::G_AMDGPU_BUFFER_STORE_FORMAT_D16: {
4499 case AMDGPU::G_AMDGPU_BUFFER_ATOMIC_SWAP:
4500 case AMDGPU::G_AMDGPU_BUFFER_ATOMIC_ADD:
4501 case AMDGPU::G_AMDGPU_BUFFER_ATOMIC_SUB:
4502 case AMDGPU::G_AMDGPU_BUFFER_ATOMIC_SMIN:
4503 case AMDGPU::G_AMDGPU_BUFFER_ATOMIC_UMIN:
4504 case AMDGPU::G_AMDGPU_BUFFER_ATOMIC_SMAX:
4505 case AMDGPU::G_AMDGPU_BUFFER_ATOMIC_UMAX:
4506 case AMDGPU::G_AMDGPU_BUFFER_ATOMIC_AND:
4507 case AMDGPU::G_AMDGPU_BUFFER_ATOMIC_OR:
4508 case AMDGPU::G_AMDGPU_BUFFER_ATOMIC_XOR:
4509 case AMDGPU::G_AMDGPU_BUFFER_ATOMIC_INC:
4510 case AMDGPU::G_AMDGPU_BUFFER_ATOMIC_DEC:
4511 case AMDGPU::G_AMDGPU_BUFFER_ATOMIC_SUB_CLAMP_U32:
4512 case AMDGPU::G_AMDGPU_BUFFER_ATOMIC_COND_SUB_U32:
4513 case AMDGPU::G_AMDGPU_BUFFER_ATOMIC_FADD:
4514 case AMDGPU::G_AMDGPU_BUFFER_ATOMIC_FMIN:
4515 case AMDGPU::G_AMDGPU_BUFFER_ATOMIC_FMAX: {
4538 case AMDGPU::G_AMDGPU_BUFFER_ATOMIC_CMPSWAP: {
4564 case AMDGPU::G_AMDGPU_S_BUFFER_LOAD:
4565 case AMDGPU::G_AMDGPU_S_BUFFER_LOAD_UBYTE:
4566 case AMDGPU::G_AMDGPU_S_BUFFER_LOAD_SBYTE:
4567 case AMDGPU::G_AMDGPU_S_BUFFER_LOAD_USHORT:
4568 case AMDGPU::G_AMDGPU_S_BUFFER_LOAD_SSHORT: {
4576 unsigned RSrcBank = OpdsMapping[1]->BreakDown[0].RegBank->getID();
4577 unsigned OffsetBank = OpdsMapping[2]->BreakDown[0].RegBank->getID();
4578 unsigned ResultBank =
regBankUnion(RSrcBank, OffsetBank);
4580 unsigned Size0 =
MRI.getType(
MI.getOperand(0).getReg()).getSizeInBits();
4581 OpdsMapping[0] = AMDGPU::getValueMapping(ResultBank, Size0);
4584 case AMDGPU::G_AMDGPU_S_BUFFER_PREFETCH:
4588 case AMDGPU::G_INTRINSIC:
4589 case AMDGPU::G_INTRINSIC_CONVERGENT: {
4593 case Intrinsic::amdgcn_div_fmas:
4594 case Intrinsic::amdgcn_div_fixup:
4595 case Intrinsic::amdgcn_trig_preop:
4596 case Intrinsic::amdgcn_sin:
4597 case Intrinsic::amdgcn_cos:
4598 case Intrinsic::amdgcn_log_clamp:
4599 case Intrinsic::amdgcn_rcp_legacy:
4600 case Intrinsic::amdgcn_rsq_legacy:
4601 case Intrinsic::amdgcn_rsq_clamp:
4602 case Intrinsic::amdgcn_tanh:
4603 case Intrinsic::amdgcn_fmul_legacy:
4604 case Intrinsic::amdgcn_fma_legacy:
4605 case Intrinsic::amdgcn_frexp_mant:
4606 case Intrinsic::amdgcn_frexp_exp:
4607 case Intrinsic::amdgcn_fract:
4608 case Intrinsic::amdgcn_cvt_pknorm_i16:
4609 case Intrinsic::amdgcn_cvt_pknorm_u16:
4610 case Intrinsic::amdgcn_cvt_pk_i16:
4611 case Intrinsic::amdgcn_cvt_pk_u16:
4612 case Intrinsic::amdgcn_cvt_sr_pk_f16_f32:
4613 case Intrinsic::amdgcn_cvt_sr_pk_bf16_f32:
4614 case Intrinsic::amdgcn_cvt_pk_f16_fp8:
4615 case Intrinsic::amdgcn_cvt_pk_f16_bf8:
4616 case Intrinsic::amdgcn_cvt_pk_fp8_f16:
4617 case Intrinsic::amdgcn_cvt_pk_bf8_f16:
4618 case Intrinsic::amdgcn_cvt_sr_fp8_f16:
4619 case Intrinsic::amdgcn_cvt_sr_bf8_f16:
4620 case Intrinsic::amdgcn_cvt_scale_pk8_f16_fp8:
4621 case Intrinsic::amdgcn_cvt_scale_pk8_bf16_fp8:
4622 case Intrinsic::amdgcn_cvt_scale_pk8_f16_bf8:
4623 case Intrinsic::amdgcn_cvt_scale_pk8_bf16_bf8:
4624 case Intrinsic::amdgcn_cvt_scale_pk8_f16_fp4:
4625 case Intrinsic::amdgcn_cvt_scale_pk8_bf16_fp4:
4626 case Intrinsic::amdgcn_cvt_scale_pk8_f32_fp8:
4627 case Intrinsic::amdgcn_cvt_scale_pk8_f32_bf8:
4628 case Intrinsic::amdgcn_cvt_scale_pk8_f32_fp4:
4629 case Intrinsic::amdgcn_cvt_scale_pk16_f16_fp6:
4630 case Intrinsic::amdgcn_cvt_scale_pk16_bf16_fp6:
4631 case Intrinsic::amdgcn_cvt_scale_pk16_f16_bf6:
4632 case Intrinsic::amdgcn_cvt_scale_pk16_bf16_bf6:
4633 case Intrinsic::amdgcn_cvt_scale_pk16_f32_fp6:
4634 case Intrinsic::amdgcn_cvt_scale_pk16_f32_bf6:
4635 case Intrinsic::amdgcn_cvt_scalef32_pk8_fp8_bf16:
4636 case Intrinsic::amdgcn_cvt_scalef32_pk8_bf8_bf16:
4637 case Intrinsic::amdgcn_cvt_scalef32_pk8_fp8_f16:
4638 case Intrinsic::amdgcn_cvt_scalef32_pk8_bf8_f16:
4639 case Intrinsic::amdgcn_cvt_scalef32_pk8_fp8_f32:
4640 case Intrinsic::amdgcn_cvt_scalef32_pk8_bf8_f32:
4641 case Intrinsic::amdgcn_cvt_scalef32_pk8_fp4_f32:
4642 case Intrinsic::amdgcn_cvt_scalef32_pk8_fp4_f16:
4643 case Intrinsic::amdgcn_cvt_scalef32_pk8_fp4_bf16:
4644 case Intrinsic::amdgcn_cvt_scalef32_pk16_fp6_f32:
4645 case Intrinsic::amdgcn_cvt_scalef32_pk16_bf6_f32:
4646 case Intrinsic::amdgcn_cvt_scalef32_pk16_fp6_f16:
4647 case Intrinsic::amdgcn_cvt_scalef32_pk16_bf6_f16:
4648 case Intrinsic::amdgcn_cvt_scalef32_pk16_fp6_bf16:
4649 case Intrinsic::amdgcn_cvt_scalef32_pk16_bf6_bf16:
4650 case Intrinsic::amdgcn_cvt_scalef32_sr_pk8_fp8_bf16:
4651 case Intrinsic::amdgcn_cvt_scalef32_sr_pk8_bf8_bf16:
4652 case Intrinsic::amdgcn_cvt_scalef32_sr_pk8_fp8_f16:
4653 case Intrinsic::amdgcn_cvt_scalef32_sr_pk8_bf8_f16:
4654 case Intrinsic::amdgcn_cvt_scalef32_sr_pk8_fp8_f32:
4655 case Intrinsic::amdgcn_cvt_scalef32_sr_pk8_bf8_f32:
4656 case Intrinsic::amdgcn_cvt_scalef32_sr_pk8_fp4_f32:
4657 case Intrinsic::amdgcn_cvt_scalef32_sr_pk8_fp4_f16:
4658 case Intrinsic::amdgcn_cvt_scalef32_sr_pk8_fp4_bf16:
4659 case Intrinsic::amdgcn_cvt_scalef32_sr_pk16_fp6_f32:
4660 case Intrinsic::amdgcn_cvt_scalef32_sr_pk16_bf6_f32:
4661 case Intrinsic::amdgcn_cvt_scalef32_sr_pk16_fp6_f16:
4662 case Intrinsic::amdgcn_cvt_scalef32_sr_pk16_bf6_f16:
4663 case Intrinsic::amdgcn_cvt_scalef32_sr_pk16_fp6_bf16:
4664 case Intrinsic::amdgcn_cvt_scalef32_sr_pk16_bf6_bf16:
4665 case Intrinsic::amdgcn_sat_pk4_i4_i8:
4666 case Intrinsic::amdgcn_sat_pk4_u4_u8:
4667 case Intrinsic::amdgcn_fmed3:
4668 case Intrinsic::amdgcn_cubeid:
4669 case Intrinsic::amdgcn_cubema:
4670 case Intrinsic::amdgcn_cubesc:
4671 case Intrinsic::amdgcn_cubetc:
4672 case Intrinsic::amdgcn_sffbh:
4673 case Intrinsic::amdgcn_fmad_ftz:
4674 case Intrinsic::amdgcn_mbcnt_lo:
4675 case Intrinsic::amdgcn_mbcnt_hi:
4676 case Intrinsic::amdgcn_mul_u24:
4677 case Intrinsic::amdgcn_mul_i24:
4678 case Intrinsic::amdgcn_mulhi_u24:
4679 case Intrinsic::amdgcn_mulhi_i24:
4680 case Intrinsic::amdgcn_lerp:
4681 case Intrinsic::amdgcn_sad_u8:
4682 case Intrinsic::amdgcn_msad_u8:
4683 case Intrinsic::amdgcn_sad_hi_u8:
4684 case Intrinsic::amdgcn_sad_u16:
4685 case Intrinsic::amdgcn_qsad_pk_u16_u8:
4686 case Intrinsic::amdgcn_mqsad_pk_u16_u8:
4687 case Intrinsic::amdgcn_mqsad_u32_u8:
4688 case Intrinsic::amdgcn_cvt_pk_u8_f32:
4689 case Intrinsic::amdgcn_alignbyte:
4690 case Intrinsic::amdgcn_perm:
4691 case Intrinsic::amdgcn_prng_b32:
4692 case Intrinsic::amdgcn_fdot2:
4693 case Intrinsic::amdgcn_sdot2:
4694 case Intrinsic::amdgcn_udot2:
4695 case Intrinsic::amdgcn_sdot4:
4696 case Intrinsic::amdgcn_udot4:
4697 case Intrinsic::amdgcn_sdot8:
4698 case Intrinsic::amdgcn_udot8:
4699 case Intrinsic::amdgcn_fdot2_bf16_bf16:
4700 case Intrinsic::amdgcn_fdot2_f16_f16:
4701 case Intrinsic::amdgcn_fdot2_f32_bf16:
4702 case Intrinsic::amdgcn_fdot2c_f32_bf16:
4703 case Intrinsic::amdgcn_sudot4:
4704 case Intrinsic::amdgcn_sudot8:
4705 case Intrinsic::amdgcn_dot4_f32_fp8_bf8:
4706 case Intrinsic::amdgcn_dot4_f32_bf8_fp8:
4707 case Intrinsic::amdgcn_dot4_f32_fp8_fp8:
4708 case Intrinsic::amdgcn_dot4_f32_bf8_bf8:
4709 case Intrinsic::amdgcn_cvt_f32_fp8:
4710 case Intrinsic::amdgcn_cvt_f32_fp8_e5m3:
4711 case Intrinsic::amdgcn_cvt_f32_bf8:
4712 case Intrinsic::amdgcn_cvt_off_f32_i4:
4713 case Intrinsic::amdgcn_cvt_pk_f32_fp8:
4714 case Intrinsic::amdgcn_cvt_pk_f32_bf8:
4715 case Intrinsic::amdgcn_cvt_pk_fp8_f32:
4716 case Intrinsic::amdgcn_cvt_pk_fp8_f32_e5m3:
4717 case Intrinsic::amdgcn_cvt_pk_bf8_f32:
4718 case Intrinsic::amdgcn_cvt_sr_fp8_f32:
4719 case Intrinsic::amdgcn_cvt_sr_fp8_f32_e5m3:
4720 case Intrinsic::amdgcn_cvt_sr_bf8_f32:
4721 case Intrinsic::amdgcn_cvt_sr_bf16_f32:
4722 case Intrinsic::amdgcn_cvt_sr_f16_f32:
4723 case Intrinsic::amdgcn_cvt_f16_fp8:
4724 case Intrinsic::amdgcn_cvt_f16_bf8:
4725 case Intrinsic::amdgcn_cvt_scalef32_pk32_fp6_f16:
4726 case Intrinsic::amdgcn_cvt_scalef32_pk32_bf6_f16:
4727 case Intrinsic::amdgcn_cvt_scalef32_pk32_fp6_bf16:
4728 case Intrinsic::amdgcn_cvt_scalef32_pk32_bf6_bf16:
4729 case Intrinsic::amdgcn_cvt_scalef32_f16_fp8:
4730 case Intrinsic::amdgcn_cvt_scalef32_f16_bf8:
4731 case Intrinsic::amdgcn_cvt_scalef32_f32_fp8:
4732 case Intrinsic::amdgcn_cvt_scalef32_f32_bf8:
4733 case Intrinsic::amdgcn_cvt_scalef32_pk_fp8_f32:
4734 case Intrinsic::amdgcn_cvt_scalef32_pk_bf8_f32:
4735 case Intrinsic::amdgcn_cvt_scalef32_pk_f32_fp8:
4736 case Intrinsic::amdgcn_cvt_scalef32_pk_f32_bf8:
4737 case Intrinsic::amdgcn_cvt_scalef32_pk_fp8_f16:
4738 case Intrinsic::amdgcn_cvt_scalef32_pk_fp8_bf16:
4739 case Intrinsic::amdgcn_cvt_scalef32_pk_bf8_f16:
4740 case Intrinsic::amdgcn_cvt_scalef32_pk_bf8_bf16:
4741 case Intrinsic::amdgcn_cvt_scalef32_pk_f32_fp4:
4742 case Intrinsic::amdgcn_cvt_scalef32_pk_fp4_f32:
4743 case Intrinsic::amdgcn_cvt_scalef32_pk_f16_fp4:
4744 case Intrinsic::amdgcn_cvt_scalef32_pk_bf16_fp4:
4745 case Intrinsic::amdgcn_cvt_scalef32_pk32_f32_fp6:
4746 case Intrinsic::amdgcn_cvt_scalef32_pk32_f32_bf6:
4747 case Intrinsic::amdgcn_cvt_scalef32_pk32_f16_bf6:
4748 case Intrinsic::amdgcn_cvt_scalef32_pk32_bf16_bf6:
4749 case Intrinsic::amdgcn_cvt_scalef32_pk32_f16_fp6:
4750 case Intrinsic::amdgcn_cvt_scalef32_pk32_bf16_fp6:
4751 case Intrinsic::amdgcn_cvt_scalef32_pk_f16_bf8:
4752 case Intrinsic::amdgcn_cvt_scalef32_pk_bf16_bf8:
4753 case Intrinsic::amdgcn_cvt_scalef32_pk_f16_fp8:
4754 case Intrinsic::amdgcn_cvt_scalef32_pk_bf16_fp8:
4755 case Intrinsic::amdgcn_cvt_scalef32_pk_fp4_f16:
4756 case Intrinsic::amdgcn_cvt_scalef32_pk_fp4_bf16:
4757 case Intrinsic::amdgcn_cvt_scalef32_sr_pk_fp4_f16:
4758 case Intrinsic::amdgcn_cvt_scalef32_sr_pk_fp4_bf16:
4759 case Intrinsic::amdgcn_cvt_scalef32_sr_pk_fp4_f32:
4760 case Intrinsic::amdgcn_cvt_scalef32_sr_pk32_bf6_bf16:
4761 case Intrinsic::amdgcn_cvt_scalef32_sr_pk32_bf6_f16:
4762 case Intrinsic::amdgcn_cvt_scalef32_sr_pk32_bf6_f32:
4763 case Intrinsic::amdgcn_cvt_scalef32_sr_pk32_fp6_bf16:
4764 case Intrinsic::amdgcn_cvt_scalef32_sr_pk32_fp6_f16:
4765 case Intrinsic::amdgcn_cvt_scalef32_sr_pk32_fp6_f32:
4766 case Intrinsic::amdgcn_cvt_scalef32_sr_bf8_bf16:
4767 case Intrinsic::amdgcn_cvt_scalef32_sr_bf8_f16:
4768 case Intrinsic::amdgcn_cvt_scalef32_sr_bf8_f32:
4769 case Intrinsic::amdgcn_cvt_scalef32_sr_fp8_bf16:
4770 case Intrinsic::amdgcn_cvt_scalef32_sr_fp8_f16:
4771 case Intrinsic::amdgcn_cvt_scalef32_sr_fp8_f32:
4772 case Intrinsic::amdgcn_ashr_pk_i8_i32:
4773 case Intrinsic::amdgcn_ashr_pk_u8_i32:
4774 case Intrinsic::amdgcn_cvt_scalef32_2xpk16_fp6_f32:
4775 case Intrinsic::amdgcn_cvt_scalef32_2xpk16_bf6_f32:
4776 case Intrinsic::amdgcn_wmma_bf16_16x16x16_bf16:
4777 case Intrinsic::amdgcn_wmma_f16_16x16x16_f16:
4778 case Intrinsic::amdgcn_wmma_bf16_16x16x16_bf16_tied:
4779 case Intrinsic::amdgcn_wmma_f16_16x16x16_f16_tied:
4780 case Intrinsic::amdgcn_wmma_f32_16x16x16_bf16:
4781 case Intrinsic::amdgcn_wmma_f32_16x16x16_f16:
4782 case Intrinsic::amdgcn_wmma_i32_16x16x16_iu4:
4783 case Intrinsic::amdgcn_wmma_i32_16x16x16_iu8:
4784 case Intrinsic::amdgcn_wmma_f32_16x16x16_fp8_fp8:
4785 case Intrinsic::amdgcn_wmma_f32_16x16x16_fp8_bf8:
4786 case Intrinsic::amdgcn_wmma_f32_16x16x16_bf8_fp8:
4787 case Intrinsic::amdgcn_wmma_f32_16x16x16_bf8_bf8:
4788 case Intrinsic::amdgcn_wmma_i32_16x16x32_iu4:
4789 case Intrinsic::amdgcn_swmmac_f32_16x16x32_f16:
4790 case Intrinsic::amdgcn_swmmac_f32_16x16x32_bf16:
4791 case Intrinsic::amdgcn_swmmac_f16_16x16x32_f16:
4792 case Intrinsic::amdgcn_swmmac_bf16_16x16x32_bf16:
4793 case Intrinsic::amdgcn_swmmac_i32_16x16x32_iu8:
4794 case Intrinsic::amdgcn_swmmac_i32_16x16x32_iu4:
4795 case Intrinsic::amdgcn_swmmac_i32_16x16x64_iu4:
4796 case Intrinsic::amdgcn_swmmac_f32_16x16x32_fp8_fp8:
4797 case Intrinsic::amdgcn_swmmac_f32_16x16x32_fp8_bf8:
4798 case Intrinsic::amdgcn_swmmac_f32_16x16x32_bf8_fp8:
4799 case Intrinsic::amdgcn_swmmac_f32_16x16x32_bf8_bf8:
4800 case Intrinsic::amdgcn_wmma_f32_16x16x4_f32:
4801 case Intrinsic::amdgcn_wmma_f32_16x16x32_bf16:
4802 case Intrinsic::amdgcn_wmma_f32_16x16x32_f16:
4803 case Intrinsic::amdgcn_wmma_f16_16x16x32_f16:
4804 case Intrinsic::amdgcn_wmma_bf16_16x16x32_bf16:
4805 case Intrinsic::amdgcn_wmma_bf16f32_16x16x32_bf16:
4806 case Intrinsic::amdgcn_wmma_f32_16x16x64_fp8_fp8:
4807 case Intrinsic::amdgcn_wmma_f32_16x16x64_fp8_bf8:
4808 case Intrinsic::amdgcn_wmma_f32_16x16x64_bf8_fp8:
4809 case Intrinsic::amdgcn_wmma_f32_16x16x64_bf8_bf8:
4810 case Intrinsic::amdgcn_wmma_f16_16x16x64_fp8_fp8:
4811 case Intrinsic::amdgcn_wmma_f16_16x16x64_fp8_bf8:
4812 case Intrinsic::amdgcn_wmma_f16_16x16x64_bf8_fp8:
4813 case Intrinsic::amdgcn_wmma_f16_16x16x64_bf8_bf8:
4814 case Intrinsic::amdgcn_wmma_f16_16x16x128_fp8_fp8:
4815 case Intrinsic::amdgcn_wmma_f16_16x16x128_fp8_bf8:
4816 case Intrinsic::amdgcn_wmma_f16_16x16x128_bf8_fp8:
4817 case Intrinsic::amdgcn_wmma_f16_16x16x128_bf8_bf8:
4818 case Intrinsic::amdgcn_wmma_f32_16x16x128_fp8_fp8:
4819 case Intrinsic::amdgcn_wmma_f32_16x16x128_fp8_bf8:
4820 case Intrinsic::amdgcn_wmma_f32_16x16x128_bf8_fp8:
4821 case Intrinsic::amdgcn_wmma_f32_16x16x128_bf8_bf8:
4822 case Intrinsic::amdgcn_wmma_i32_16x16x64_iu8:
4823 case Intrinsic::amdgcn_wmma_f32_16x16x128_f8f6f4:
4824 case Intrinsic::amdgcn_wmma_scale_f32_16x16x128_f8f6f4:
4825 case Intrinsic::amdgcn_wmma_scale16_f32_16x16x128_f8f6f4:
4826 case Intrinsic::amdgcn_wmma_f32_32x16x128_f4:
4827 case Intrinsic::amdgcn_wmma_scale_f32_32x16x128_f4:
4828 case Intrinsic::amdgcn_wmma_scale16_f32_32x16x128_f4:
4829 case Intrinsic::amdgcn_swmmac_f16_16x16x64_f16:
4830 case Intrinsic::amdgcn_swmmac_bf16_16x16x64_bf16:
4831 case Intrinsic::amdgcn_swmmac_f32_16x16x64_bf16:
4832 case Intrinsic::amdgcn_swmmac_bf16f32_16x16x64_bf16:
4833 case Intrinsic::amdgcn_swmmac_f32_16x16x64_f16:
4834 case Intrinsic::amdgcn_swmmac_f32_16x16x128_fp8_fp8:
4835 case Intrinsic::amdgcn_swmmac_f32_16x16x128_fp8_bf8:
4836 case Intrinsic::amdgcn_swmmac_f32_16x16x128_bf8_fp8:
4837 case Intrinsic::amdgcn_swmmac_f32_16x16x128_bf8_bf8:
4838 case Intrinsic::amdgcn_swmmac_f16_16x16x128_fp8_fp8:
4839 case Intrinsic::amdgcn_swmmac_f16_16x16x128_fp8_bf8:
4840 case Intrinsic::amdgcn_swmmac_f16_16x16x128_bf8_fp8:
4841 case Intrinsic::amdgcn_swmmac_f16_16x16x128_bf8_bf8:
4842 case Intrinsic::amdgcn_swmmac_i32_16x16x128_iu8:
4843 case Intrinsic::amdgcn_perm_pk16_b4_u4:
4844 case Intrinsic::amdgcn_perm_pk16_b6_u4:
4845 case Intrinsic::amdgcn_perm_pk16_b8_u4:
4846 case Intrinsic::amdgcn_add_max_i32:
4847 case Intrinsic::amdgcn_add_max_u32:
4848 case Intrinsic::amdgcn_add_min_i32:
4849 case Intrinsic::amdgcn_add_min_u32:
4850 case Intrinsic::amdgcn_pk_add_max_i16:
4851 case Intrinsic::amdgcn_pk_add_max_u16:
4852 case Intrinsic::amdgcn_pk_add_min_i16:
4853 case Intrinsic::amdgcn_pk_add_min_u16:
4855 case Intrinsic::amdgcn_log:
4856 case Intrinsic::amdgcn_exp2:
4857 case Intrinsic::amdgcn_rcp:
4858 case Intrinsic::amdgcn_rsq:
4859 case Intrinsic::amdgcn_sqrt: {
4860 unsigned Size =
MRI.getType(
MI.getOperand(0).getReg()).getSizeInBits();
4866 case Intrinsic::amdgcn_sbfe:
4867 case Intrinsic::amdgcn_ubfe:
4871 case Intrinsic::amdgcn_ds_swizzle:
4872 case Intrinsic::amdgcn_ds_permute:
4873 case Intrinsic::amdgcn_ds_bpermute:
4874 case Intrinsic::amdgcn_update_dpp:
4875 case Intrinsic::amdgcn_mov_dpp8:
4876 case Intrinsic::amdgcn_mov_dpp:
4877 case Intrinsic::amdgcn_strict_wwm:
4878 case Intrinsic::amdgcn_wwm:
4879 case Intrinsic::amdgcn_strict_wqm:
4880 case Intrinsic::amdgcn_wqm:
4881 case Intrinsic::amdgcn_softwqm:
4882 case Intrinsic::amdgcn_set_inactive:
4883 case Intrinsic::amdgcn_set_inactive_chain_arg:
4884 case Intrinsic::amdgcn_permlane64:
4885 case Intrinsic::amdgcn_ds_bpermute_fi_b32:
4887 case Intrinsic::amdgcn_cvt_pkrtz:
4891 case Intrinsic::amdgcn_kernarg_segment_ptr:
4892 case Intrinsic::amdgcn_s_getpc:
4893 case Intrinsic::amdgcn_groupstaticsize:
4894 case Intrinsic::amdgcn_reloc_constant:
4895 case Intrinsic::returnaddress: {
4896 unsigned Size =
MRI.getType(
MI.getOperand(0).getReg()).getSizeInBits();
4897 OpdsMapping[0] = AMDGPU::getValueMapping(AMDGPU::SGPRRegBankID,
Size);
4900 case Intrinsic::amdgcn_wqm_vote: {
4901 unsigned Size =
MRI.getType(
MI.getOperand(0).getReg()).getSizeInBits();
4902 OpdsMapping[0] = OpdsMapping[2]
4903 = AMDGPU::getValueMapping(AMDGPU::VCCRegBankID,
Size);
4906 case Intrinsic::amdgcn_ps_live: {
4907 OpdsMapping[0] = AMDGPU::getValueMapping(AMDGPU::VCCRegBankID, 1);
4910 case Intrinsic::amdgcn_div_scale: {
4911 unsigned Dst0Size =
MRI.getType(
MI.getOperand(0).getReg()).getSizeInBits();
4912 unsigned Dst1Size =
MRI.getType(
MI.getOperand(1).getReg()).getSizeInBits();
4913 OpdsMapping[0] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID, Dst0Size);
4914 OpdsMapping[1] = AMDGPU::getValueMapping(AMDGPU::VCCRegBankID, Dst1Size);
4916 unsigned SrcSize =
MRI.getType(
MI.getOperand(3).getReg()).getSizeInBits();
4917 OpdsMapping[3] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID, SrcSize);
4918 OpdsMapping[4] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID, SrcSize);
4921 case Intrinsic::amdgcn_class: {
4922 Register Src0Reg =
MI.getOperand(2).getReg();
4923 Register Src1Reg =
MI.getOperand(3).getReg();
4924 unsigned Src0Size =
MRI.getType(Src0Reg).getSizeInBits();
4925 unsigned Src1Size =
MRI.getType(Src1Reg).getSizeInBits();
4926 unsigned DstSize =
MRI.getType(
MI.getOperand(0).getReg()).getSizeInBits();
4927 OpdsMapping[0] = AMDGPU::getValueMapping(AMDGPU::VCCRegBankID, DstSize);
4928 OpdsMapping[2] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID, Src0Size);
4929 OpdsMapping[3] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID, Src1Size);
4932 case Intrinsic::amdgcn_icmp:
4933 case Intrinsic::amdgcn_fcmp: {
4934 unsigned DstSize =
MRI.getType(
MI.getOperand(0).getReg()).getSizeInBits();
4936 OpdsMapping[0] = AMDGPU::getValueMapping(AMDGPU::SGPRRegBankID, DstSize);
4937 unsigned OpSize =
MRI.getType(
MI.getOperand(2).getReg()).getSizeInBits();
4938 OpdsMapping[2] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID, OpSize);
4939 OpdsMapping[3] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID, OpSize);
4942 case Intrinsic::amdgcn_readlane: {
4945 unsigned IdxSize =
MRI.getType(IdxReg).getSizeInBits();
4947 OpdsMapping[3] = AMDGPU::getValueMapping(IdxBank, IdxSize);
4950 case Intrinsic::amdgcn_readfirstlane: {
4951 unsigned DstSize =
MRI.getType(
MI.getOperand(0).getReg()).getSizeInBits();
4952 unsigned SrcSize =
MRI.getType(
MI.getOperand(2).getReg()).getSizeInBits();
4953 OpdsMapping[0] = AMDGPU::getValueMapping(AMDGPU::SGPRRegBankID, DstSize);
4954 OpdsMapping[2] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID, SrcSize);
4957 case Intrinsic::amdgcn_writelane: {
4958 unsigned DstSize =
MRI.getType(
MI.getOperand(0).getReg()).getSizeInBits();
4960 unsigned SrcSize =
MRI.getType(SrcReg).getSizeInBits();
4963 unsigned IdxSize =
MRI.getType(IdxReg).getSizeInBits();
4965 OpdsMapping[0] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID, DstSize);
4969 OpdsMapping[2] = AMDGPU::getValueMapping(SrcBank, SrcSize);
4970 OpdsMapping[3] = AMDGPU::getValueMapping(IdxBank, IdxSize);
4971 OpdsMapping[4] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID, SrcSize);
4974 case Intrinsic::amdgcn_if_break: {
4976 OpdsMapping[0] = AMDGPU::getValueMapping(AMDGPU::SGPRRegBankID,
Size);
4977 OpdsMapping[2] = AMDGPU::getValueMapping(AMDGPU::VCCRegBankID, 1);
4978 OpdsMapping[3] = AMDGPU::getValueMapping(AMDGPU::SGPRRegBankID,
Size);
4981 case Intrinsic::amdgcn_permlane16:
4982 case Intrinsic::amdgcn_permlanex16: {
4984 OpdsMapping[0] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID,
Size);
4985 OpdsMapping[2] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID,
Size);
4986 OpdsMapping[3] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID,
Size);
4991 case Intrinsic::amdgcn_permlane_bcast:
4992 case Intrinsic::amdgcn_permlane_up:
4993 case Intrinsic::amdgcn_permlane_down:
4994 case Intrinsic::amdgcn_permlane_xor: {
4996 OpdsMapping[0] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID,
Size);
4997 OpdsMapping[2] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID,
Size);
5002 case Intrinsic::amdgcn_permlane_idx_gen: {
5004 OpdsMapping[0] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID,
Size);
5005 OpdsMapping[2] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID,
Size);
5009 case Intrinsic::amdgcn_permlane16_var:
5010 case Intrinsic::amdgcn_permlanex16_var: {
5012 OpdsMapping[0] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID,
Size);
5013 OpdsMapping[2] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID,
Size);
5014 OpdsMapping[3] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID,
Size);
5015 OpdsMapping[4] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID,
Size);
5018 case Intrinsic::amdgcn_mfma_f32_4x4x1f32:
5019 case Intrinsic::amdgcn_mfma_f32_4x4x4f16:
5020 case Intrinsic::amdgcn_mfma_i32_4x4x4i8:
5021 case Intrinsic::amdgcn_mfma_f32_4x4x2bf16:
5022 case Intrinsic::amdgcn_mfma_f32_16x16x1f32:
5023 case Intrinsic::amdgcn_mfma_f32_16x16x4f32:
5024 case Intrinsic::amdgcn_mfma_f32_16x16x4f16:
5025 case Intrinsic::amdgcn_mfma_f32_16x16x16f16:
5026 case Intrinsic::amdgcn_mfma_i32_16x16x4i8:
5027 case Intrinsic::amdgcn_mfma_i32_16x16x16i8:
5028 case Intrinsic::amdgcn_mfma_f32_16x16x2bf16:
5029 case Intrinsic::amdgcn_mfma_f32_16x16x8bf16:
5030 case Intrinsic::amdgcn_mfma_f32_32x32x1f32:
5031 case Intrinsic::amdgcn_mfma_f32_32x32x2f32:
5032 case Intrinsic::amdgcn_mfma_f32_32x32x4f16:
5033 case Intrinsic::amdgcn_mfma_f32_32x32x8f16:
5034 case Intrinsic::amdgcn_mfma_i32_32x32x4i8:
5035 case Intrinsic::amdgcn_mfma_i32_32x32x8i8:
5036 case Intrinsic::amdgcn_mfma_f32_32x32x2bf16:
5037 case Intrinsic::amdgcn_mfma_f32_32x32x4bf16:
5038 case Intrinsic::amdgcn_mfma_f32_32x32x4bf16_1k:
5039 case Intrinsic::amdgcn_mfma_f32_16x16x4bf16_1k:
5040 case Intrinsic::amdgcn_mfma_f32_4x4x4bf16_1k:
5041 case Intrinsic::amdgcn_mfma_f32_32x32x8bf16_1k:
5042 case Intrinsic::amdgcn_mfma_f32_16x16x16bf16_1k:
5043 case Intrinsic::amdgcn_mfma_f64_16x16x4f64:
5044 case Intrinsic::amdgcn_mfma_f64_4x4x4f64:
5045 case Intrinsic::amdgcn_mfma_i32_16x16x32_i8:
5046 case Intrinsic::amdgcn_mfma_i32_32x32x16_i8:
5047 case Intrinsic::amdgcn_mfma_f32_16x16x8_xf32:
5048 case Intrinsic::amdgcn_mfma_f32_32x32x4_xf32:
5049 case Intrinsic::amdgcn_mfma_f32_16x16x32_bf8_bf8:
5050 case Intrinsic::amdgcn_mfma_f32_16x16x32_bf8_fp8:
5051 case Intrinsic::amdgcn_mfma_f32_16x16x32_fp8_bf8:
5052 case Intrinsic::amdgcn_mfma_f32_16x16x32_fp8_fp8:
5053 case Intrinsic::amdgcn_mfma_f32_32x32x16_bf8_bf8:
5054 case Intrinsic::amdgcn_mfma_f32_32x32x16_bf8_fp8:
5055 case Intrinsic::amdgcn_mfma_f32_32x32x16_fp8_bf8:
5056 case Intrinsic::amdgcn_mfma_f32_32x32x16_fp8_fp8:
5057 case Intrinsic::amdgcn_mfma_f32_16x16x32_f16:
5058 case Intrinsic::amdgcn_mfma_f32_32x32x16_f16:
5059 case Intrinsic::amdgcn_mfma_i32_16x16x64_i8:
5060 case Intrinsic::amdgcn_mfma_i32_32x32x32_i8:
5061 case Intrinsic::amdgcn_mfma_f32_16x16x32_bf16: {
5062 unsigned DstSize =
MRI.getType(
MI.getOperand(0).getReg()).getSizeInBits();
5063 unsigned MinNumRegsRequired = DstSize / 32;
5073 bool UseAGPRForm = !
Subtarget.hasGFX90AInsts() ||
5074 Info->selectAGPRFormMFMA(MinNumRegsRequired);
5086 case Intrinsic::amdgcn_mfma_scale_f32_16x16x128_f8f6f4:
5087 case Intrinsic::amdgcn_mfma_scale_f32_32x32x64_f8f6f4: {
5088 unsigned DstSize =
MRI.getType(
MI.getOperand(0).getReg()).getSizeInBits();
5089 unsigned MinNumRegsRequired = DstSize / 32;
5108 case Intrinsic::amdgcn_smfmac_f32_16x16x32_f16:
5109 case Intrinsic::amdgcn_smfmac_f32_32x32x16_f16:
5110 case Intrinsic::amdgcn_smfmac_f32_16x16x32_bf16:
5111 case Intrinsic::amdgcn_smfmac_f32_32x32x16_bf16:
5112 case Intrinsic::amdgcn_smfmac_i32_16x16x64_i8:
5113 case Intrinsic::amdgcn_smfmac_i32_32x32x32_i8:
5114 case Intrinsic::amdgcn_smfmac_f32_16x16x64_bf8_bf8:
5115 case Intrinsic::amdgcn_smfmac_f32_16x16x64_bf8_fp8:
5116 case Intrinsic::amdgcn_smfmac_f32_16x16x64_fp8_bf8:
5117 case Intrinsic::amdgcn_smfmac_f32_16x16x64_fp8_fp8:
5118 case Intrinsic::amdgcn_smfmac_f32_32x32x32_bf8_bf8:
5119 case Intrinsic::amdgcn_smfmac_f32_32x32x32_bf8_fp8:
5120 case Intrinsic::amdgcn_smfmac_f32_32x32x32_fp8_bf8:
5121 case Intrinsic::amdgcn_smfmac_f32_32x32x32_fp8_fp8:
5122 case Intrinsic::amdgcn_smfmac_f32_16x16x64_f16:
5123 case Intrinsic::amdgcn_smfmac_f32_32x32x32_f16:
5124 case Intrinsic::amdgcn_smfmac_f32_16x16x64_bf16:
5125 case Intrinsic::amdgcn_smfmac_f32_32x32x32_bf16:
5126 case Intrinsic::amdgcn_smfmac_i32_16x16x128_i8:
5127 case Intrinsic::amdgcn_smfmac_i32_32x32x64_i8:
5128 case Intrinsic::amdgcn_smfmac_f32_16x16x128_bf8_bf8:
5129 case Intrinsic::amdgcn_smfmac_f32_16x16x128_bf8_fp8:
5130 case Intrinsic::amdgcn_smfmac_f32_16x16x128_fp8_bf8:
5131 case Intrinsic::amdgcn_smfmac_f32_16x16x128_fp8_fp8:
5132 case Intrinsic::amdgcn_smfmac_f32_32x32x64_bf8_bf8:
5133 case Intrinsic::amdgcn_smfmac_f32_32x32x64_bf8_fp8:
5134 case Intrinsic::amdgcn_smfmac_f32_32x32x64_fp8_bf8:
5135 case Intrinsic::amdgcn_smfmac_f32_32x32x64_fp8_fp8: {
5137 unsigned DstSize =
MRI.getType(DstReg).getSizeInBits();
5138 unsigned MinNumRegsRequired = DstSize / 32;
5154 case Intrinsic::amdgcn_interp_p1:
5155 case Intrinsic::amdgcn_interp_p2:
5156 case Intrinsic::amdgcn_interp_mov:
5157 case Intrinsic::amdgcn_interp_p1_f16:
5158 case Intrinsic::amdgcn_interp_p2_f16:
5159 case Intrinsic::amdgcn_lds_param_load: {
5160 const int M0Idx =
MI.getNumOperands() - 1;
5161 Register M0Reg =
MI.getOperand(M0Idx).getReg();
5163 unsigned DstSize =
MRI.getType(
MI.getOperand(0).getReg()).getSizeInBits();
5165 OpdsMapping[0] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID, DstSize);
5166 for (
int I = 2;
I != M0Idx &&
MI.getOperand(
I).
isReg(); ++
I)
5167 OpdsMapping[
I] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID, 32);
5171 OpdsMapping[M0Idx] = AMDGPU::getValueMapping(M0Bank, 32);
5174 case Intrinsic::amdgcn_interp_inreg_p10:
5175 case Intrinsic::amdgcn_interp_inreg_p2:
5176 case Intrinsic::amdgcn_interp_inreg_p10_f16:
5177 case Intrinsic::amdgcn_interp_inreg_p2_f16:
5178 case Intrinsic::amdgcn_interp_p10_rtz_f16:
5179 case Intrinsic::amdgcn_interp_p2_rtz_f16: {
5180 unsigned DstSize =
MRI.getType(
MI.getOperand(0).getReg()).getSizeInBits();
5181 OpdsMapping[0] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID, DstSize);
5182 OpdsMapping[2] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID, 32);
5183 OpdsMapping[3] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID, 32);
5184 OpdsMapping[4] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID, 32);
5187 case Intrinsic::amdgcn_permlane16_swap:
5188 case Intrinsic::amdgcn_permlane32_swap: {
5189 unsigned DstSize =
MRI.getType(
MI.getOperand(0).getReg()).getSizeInBits();
5190 OpdsMapping[0] = OpdsMapping[1] = OpdsMapping[3] = OpdsMapping[4] =
5191 AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID, DstSize);
5194 case Intrinsic::amdgcn_ballot: {
5195 unsigned DstSize =
MRI.getType(
MI.getOperand(0).getReg()).getSizeInBits();
5196 unsigned SrcSize =
MRI.getType(
MI.getOperand(2).getReg()).getSizeInBits();
5197 OpdsMapping[0] = AMDGPU::getValueMapping(AMDGPU::SGPRRegBankID, DstSize);
5198 OpdsMapping[2] = AMDGPU::getValueMapping(AMDGPU::VCCRegBankID, SrcSize);
5201 case Intrinsic::amdgcn_inverse_ballot: {
5203 Register MaskReg =
MI.getOperand(2).getReg();
5204 unsigned MaskSize =
MRI.getType(MaskReg).getSizeInBits();
5205 unsigned MaskBank =
getRegBankID(MaskReg,
MRI, AMDGPU::SGPRRegBankID);
5206 OpdsMapping[0] = AMDGPU::getValueMapping(AMDGPU::VCCRegBankID, 1);
5207 OpdsMapping[2] = AMDGPU::getValueMapping(MaskBank, MaskSize);
5210 case Intrinsic::amdgcn_bitop3: {
5212 OpdsMapping[0] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID,
Size);
5213 OpdsMapping[2] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID,
Size);
5214 OpdsMapping[3] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID,
Size);
5215 OpdsMapping[4] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID,
Size);
5218 case Intrinsic::amdgcn_s_quadmask:
5219 case Intrinsic::amdgcn_s_wqm: {
5220 Register MaskReg =
MI.getOperand(2).getReg();
5221 unsigned MaskSize =
MRI.getType(MaskReg).getSizeInBits();
5222 unsigned MaskBank =
getRegBankID(MaskReg,
MRI, AMDGPU::SGPRRegBankID);
5223 OpdsMapping[0] = AMDGPU::getValueMapping(AMDGPU::SGPRRegBankID, MaskSize);
5224 OpdsMapping[2] = AMDGPU::getValueMapping(MaskBank, MaskSize);
5227 case Intrinsic::amdgcn_wave_reduce_add:
5228 case Intrinsic::amdgcn_wave_reduce_fadd:
5229 case Intrinsic::amdgcn_wave_reduce_sub:
5230 case Intrinsic::amdgcn_wave_reduce_fsub:
5231 case Intrinsic::amdgcn_wave_reduce_min:
5232 case Intrinsic::amdgcn_wave_reduce_umin:
5233 case Intrinsic::amdgcn_wave_reduce_fmin:
5234 case Intrinsic::amdgcn_wave_reduce_max:
5235 case Intrinsic::amdgcn_wave_reduce_umax:
5236 case Intrinsic::amdgcn_wave_reduce_fmax:
5237 case Intrinsic::amdgcn_wave_reduce_and:
5238 case Intrinsic::amdgcn_wave_reduce_or:
5239 case Intrinsic::amdgcn_wave_reduce_xor: {
5240 unsigned DstSize =
MRI.getType(
MI.getOperand(0).getReg()).getSizeInBits();
5241 OpdsMapping[0] = AMDGPU::getValueMapping(AMDGPU::SGPRRegBankID, DstSize);
5242 unsigned OpSize =
MRI.getType(
MI.getOperand(2).getReg()).getSizeInBits();
5245 OpdsMapping[2] = AMDGPU::getValueMapping(regBankID, OpSize);
5248 case Intrinsic::amdgcn_s_bitreplicate: {
5249 Register MaskReg =
MI.getOperand(2).getReg();
5250 unsigned MaskBank =
getRegBankID(MaskReg,
MRI, AMDGPU::SGPRRegBankID);
5251 OpdsMapping[0] = AMDGPU::getValueMapping(AMDGPU::SGPRRegBankID, 64);
5252 OpdsMapping[2] = AMDGPU::getValueMapping(MaskBank, 32);
5255 case Intrinsic::amdgcn_wave_shuffle: {
5256 unsigned OpSize =
MRI.getType(
MI.getOperand(0).getReg()).getSizeInBits();
5257 OpdsMapping[0] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID, OpSize);
5258 OpdsMapping[2] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID, OpSize);
5259 OpdsMapping[3] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID, OpSize);
5265 case AMDGPU::G_AMDGPU_INTRIN_IMAGE_LOAD:
5266 case AMDGPU::G_AMDGPU_INTRIN_IMAGE_LOAD_D16:
5267 case AMDGPU::G_AMDGPU_INTRIN_IMAGE_LOAD_NORET:
5268 case AMDGPU::G_AMDGPU_INTRIN_IMAGE_STORE:
5269 case AMDGPU::G_AMDGPU_INTRIN_IMAGE_STORE_D16: {
5272 assert(RSrcIntrin &&
"missing RsrcIntrinsic for image intrinsic");
5279 case AMDGPU::G_AMDGPU_BVH_INTERSECT_RAY:
5280 case AMDGPU::G_AMDGPU_BVH8_INTERSECT_RAY:
5281 case AMDGPU::G_AMDGPU_BVH_DUAL_INTERSECT_RAY: {
5283 MI.getOpcode() == AMDGPU::G_AMDGPU_BVH_DUAL_INTERSECT_RAY ||
5284 MI.getOpcode() == AMDGPU::G_AMDGPU_BVH8_INTERSECT_RAY;
5285 unsigned NumMods = IsDualOrBVH8 ? 0 : 1;
5286 unsigned LastRegOpIdx =
MI.getNumExplicitOperands() - 1 - NumMods;
5287 unsigned DstSize =
MRI.getType(
MI.getOperand(0).getReg()).getSizeInBits();
5288 OpdsMapping[0] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID, DstSize);
5290 OpdsMapping[1] = AMDGPU::getValueMapping(
5291 AMDGPU::VGPRRegBankID,
5292 MRI.getType(
MI.getOperand(1).getReg()).getSizeInBits());
5293 OpdsMapping[2] = AMDGPU::getValueMapping(
5294 AMDGPU::VGPRRegBankID,
5295 MRI.getType(
MI.getOperand(2).getReg()).getSizeInBits());
5297 OpdsMapping[LastRegOpIdx] =
5299 if (LastRegOpIdx == 3) {
5301 unsigned Size =
MRI.getType(
MI.getOperand(2).getReg()).getSizeInBits();
5304 OpdsMapping[2] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID,
Size);
5307 unsigned FirstSrcOpIdx = IsDualOrBVH8 ? 4 : 2;
5308 for (
unsigned I = FirstSrcOpIdx;
I < LastRegOpIdx; ++
I) {
5309 unsigned Size =
MRI.getType(
MI.getOperand(
I).getReg()).getSizeInBits();
5310 OpdsMapping[
I] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID,
Size);
5315 case AMDGPU::G_INTRINSIC_W_SIDE_EFFECTS:
5316 case AMDGPU::G_INTRINSIC_CONVERGENT_W_SIDE_EFFECTS: {
5319 case Intrinsic::amdgcn_s_getreg:
5320 case Intrinsic::amdgcn_s_memtime:
5321 case Intrinsic::amdgcn_s_memrealtime:
5322 case Intrinsic::amdgcn_s_get_waveid_in_workgroup:
5323 case Intrinsic::amdgcn_s_sendmsg_rtn: {
5324 unsigned Size =
MRI.getType(
MI.getOperand(0).getReg()).getSizeInBits();
5325 OpdsMapping[0] = AMDGPU::getValueMapping(AMDGPU::SGPRRegBankID,
Size);
5328 case Intrinsic::amdgcn_global_atomic_fmin_num:
5329 case Intrinsic::amdgcn_global_atomic_fmax_num:
5330 case Intrinsic::amdgcn_flat_atomic_fmin_num:
5331 case Intrinsic::amdgcn_flat_atomic_fmax_num:
5332 case Intrinsic::amdgcn_global_atomic_ordered_add_b64:
5333 case Intrinsic::amdgcn_global_load_tr_b64:
5334 case Intrinsic::amdgcn_global_load_tr_b128:
5335 case Intrinsic::amdgcn_global_load_tr4_b64:
5336 case Intrinsic::amdgcn_global_load_tr6_b96:
5337 case Intrinsic::amdgcn_ds_load_tr8_b64:
5338 case Intrinsic::amdgcn_ds_load_tr16_b128:
5339 case Intrinsic::amdgcn_ds_load_tr4_b64:
5340 case Intrinsic::amdgcn_ds_load_tr6_b96:
5341 case Intrinsic::amdgcn_flat_load_monitor_b32:
5342 case Intrinsic::amdgcn_flat_load_monitor_b64:
5343 case Intrinsic::amdgcn_flat_load_monitor_b128:
5344 case Intrinsic::amdgcn_global_load_monitor_b32:
5345 case Intrinsic::amdgcn_global_load_monitor_b64:
5346 case Intrinsic::amdgcn_global_load_monitor_b128:
5347 case Intrinsic::amdgcn_ds_read_tr4_b64:
5348 case Intrinsic::amdgcn_ds_read_tr6_b96:
5349 case Intrinsic::amdgcn_ds_read_tr8_b64:
5350 case Intrinsic::amdgcn_ds_read_tr16_b64:
5351 case Intrinsic::amdgcn_ds_atomic_async_barrier_arrive_b64:
5352 case Intrinsic::amdgcn_ds_atomic_barrier_arrive_rtn_b64:
5354 case Intrinsic::amdgcn_ds_ordered_add:
5355 case Intrinsic::amdgcn_ds_ordered_swap: {
5356 unsigned DstSize =
MRI.getType(
MI.getOperand(0).getReg()).getSizeInBits();
5357 OpdsMapping[0] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID, DstSize);
5359 AMDGPU::SGPRRegBankID);
5360 OpdsMapping[2] = AMDGPU::getValueMapping(M0Bank, 32);
5361 OpdsMapping[3] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID, 32);
5364 case Intrinsic::amdgcn_ds_append:
5365 case Intrinsic::amdgcn_ds_consume: {
5366 unsigned DstSize =
MRI.getType(
MI.getOperand(0).getReg()).getSizeInBits();
5367 OpdsMapping[0] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID, DstSize);
5371 case Intrinsic::amdgcn_exp_compr:
5372 OpdsMapping[3] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID, 32);
5373 OpdsMapping[4] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID, 32);
5375 case Intrinsic::amdgcn_exp:
5377 OpdsMapping[3] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID, 32);
5378 OpdsMapping[4] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID, 32);
5379 OpdsMapping[5] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID, 32);
5380 OpdsMapping[6] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID, 32);
5382 case Intrinsic::amdgcn_exp_row:
5383 OpdsMapping[3] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID, 32);
5384 OpdsMapping[4] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID, 32);
5385 OpdsMapping[5] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID, 32);
5386 OpdsMapping[6] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID, 32);
5389 case Intrinsic::amdgcn_s_sendmsg:
5390 case Intrinsic::amdgcn_s_sendmsghalt: {
5393 AMDGPU::SGPRRegBankID);
5394 OpdsMapping[2] = AMDGPU::getValueMapping(Bank, 32);
5397 case Intrinsic::amdgcn_s_setreg: {
5400 AMDGPU::SGPRRegBankID);
5401 OpdsMapping[2] = AMDGPU::getValueMapping(Bank, 32);
5404 case Intrinsic::amdgcn_s_ttracedata: {
5408 OpdsMapping[1] = AMDGPU::getValueMapping(Bank, 32);
5411 case Intrinsic::amdgcn_end_cf: {
5413 OpdsMapping[1] = AMDGPU::getValueMapping(AMDGPU::SGPRRegBankID,
Size);
5416 case Intrinsic::amdgcn_else: {
5418 OpdsMapping[0] = AMDGPU::getValueMapping(AMDGPU::VCCRegBankID, 1);
5419 OpdsMapping[1] = AMDGPU::getValueMapping(AMDGPU::SGPRRegBankID, WaveSize);
5420 OpdsMapping[3] = AMDGPU::getValueMapping(AMDGPU::SGPRRegBankID, WaveSize);
5423 case Intrinsic::amdgcn_init_whole_wave:
5424 case Intrinsic::amdgcn_live_mask: {
5425 OpdsMapping[0] = AMDGPU::getValueMapping(AMDGPU::VCCRegBankID, 1);
5428 case Intrinsic::amdgcn_wqm_demote:
5429 case Intrinsic::amdgcn_kill: {
5430 OpdsMapping[1] = AMDGPU::getValueMapping(AMDGPU::VCCRegBankID, 1);
5433 case Intrinsic::amdgcn_raw_buffer_load:
5434 case Intrinsic::amdgcn_raw_ptr_buffer_load:
5435 case Intrinsic::amdgcn_raw_atomic_buffer_load:
5436 case Intrinsic::amdgcn_raw_ptr_atomic_buffer_load:
5437 case Intrinsic::amdgcn_raw_tbuffer_load:
5438 case Intrinsic::amdgcn_raw_ptr_tbuffer_load: {
5447 case Intrinsic::amdgcn_raw_buffer_load_lds:
5448 case Intrinsic::amdgcn_raw_ptr_buffer_load_lds: {
5455 case Intrinsic::amdgcn_raw_buffer_store:
5456 case Intrinsic::amdgcn_raw_ptr_buffer_store:
5457 case Intrinsic::amdgcn_raw_buffer_store_format:
5458 case Intrinsic::amdgcn_raw_ptr_buffer_store_format:
5459 case Intrinsic::amdgcn_raw_tbuffer_store:
5460 case Intrinsic::amdgcn_raw_ptr_tbuffer_store: {
5467 case Intrinsic::amdgcn_struct_buffer_load:
5468 case Intrinsic::amdgcn_struct_ptr_buffer_load:
5469 case Intrinsic::amdgcn_struct_tbuffer_load:
5470 case Intrinsic::amdgcn_struct_ptr_tbuffer_load:
5471 case Intrinsic::amdgcn_struct_atomic_buffer_load:
5472 case Intrinsic::amdgcn_struct_ptr_atomic_buffer_load: {
5480 case Intrinsic::amdgcn_struct_buffer_load_lds:
5481 case Intrinsic::amdgcn_struct_ptr_buffer_load_lds: {
5489 case Intrinsic::amdgcn_struct_buffer_store:
5490 case Intrinsic::amdgcn_struct_ptr_buffer_store:
5491 case Intrinsic::amdgcn_struct_tbuffer_store:
5492 case Intrinsic::amdgcn_struct_ptr_tbuffer_store: {
5500 case Intrinsic::amdgcn_init_exec_from_input: {
5502 OpdsMapping[1] = AMDGPU::getValueMapping(AMDGPU::SGPRRegBankID,
Size);
5505 case Intrinsic::amdgcn_ds_gws_init:
5506 case Intrinsic::amdgcn_ds_gws_barrier:
5507 case Intrinsic::amdgcn_ds_gws_sema_br: {
5508 OpdsMapping[1] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID, 32);
5512 AMDGPU::SGPRRegBankID);
5513 OpdsMapping[2] = AMDGPU::getValueMapping(Bank, 32);
5516 case Intrinsic::amdgcn_ds_gws_sema_v:
5517 case Intrinsic::amdgcn_ds_gws_sema_p:
5518 case Intrinsic::amdgcn_ds_gws_sema_release_all: {
5521 AMDGPU::SGPRRegBankID);
5522 OpdsMapping[1] = AMDGPU::getValueMapping(Bank, 32);
5525 case Intrinsic::amdgcn_cluster_load_b32:
5526 case Intrinsic::amdgcn_cluster_load_b64:
5527 case Intrinsic::amdgcn_cluster_load_b128: {
5532 OpdsMapping[4] = AMDGPU::getValueMapping(M0Bank, 32);
5535 case Intrinsic::amdgcn_cluster_load_async_to_lds_b8:
5536 case Intrinsic::amdgcn_cluster_load_async_to_lds_b32:
5537 case Intrinsic::amdgcn_cluster_load_async_to_lds_b64:
5538 case Intrinsic::amdgcn_cluster_load_async_to_lds_b128: {
5543 OpdsMapping[5] = AMDGPU::getValueMapping(M0Bank, 32);
5546 case Intrinsic::amdgcn_global_store_async_from_lds_b8:
5547 case Intrinsic::amdgcn_global_store_async_from_lds_b32:
5548 case Intrinsic::amdgcn_global_store_async_from_lds_b64:
5549 case Intrinsic::amdgcn_global_store_async_from_lds_b128:
5550 case Intrinsic::amdgcn_global_load_async_to_lds_b8:
5551 case Intrinsic::amdgcn_global_load_async_to_lds_b32:
5552 case Intrinsic::amdgcn_global_load_async_to_lds_b64:
5553 case Intrinsic::amdgcn_global_load_async_to_lds_b128:
5554 case Intrinsic::amdgcn_load_to_lds:
5555 case Intrinsic::amdgcn_global_load_lds: {
5560 case Intrinsic::amdgcn_lds_direct_load: {
5561 const int M0Idx =
MI.getNumOperands() - 1;
5562 Register M0Reg =
MI.getOperand(M0Idx).getReg();
5564 unsigned DstSize =
MRI.getType(
MI.getOperand(0).getReg()).getSizeInBits();
5566 OpdsMapping[0] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID, DstSize);
5567 for (
int I = 2;
I != M0Idx &&
MI.getOperand(
I).
isReg(); ++
I)
5568 OpdsMapping[
I] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID, 32);
5572 OpdsMapping[M0Idx] = AMDGPU::getValueMapping(M0Bank, 32);
5575 case Intrinsic::amdgcn_ds_add_gs_reg_rtn:
5576 case Intrinsic::amdgcn_ds_sub_gs_reg_rtn:
5580 case Intrinsic::amdgcn_ds_bvh_stack_rtn:
5581 case Intrinsic::amdgcn_ds_bvh_stack_push4_pop1_rtn:
5582 case Intrinsic::amdgcn_ds_bvh_stack_push8_pop1_rtn:
5583 case Intrinsic::amdgcn_ds_bvh_stack_push8_pop2_rtn: {
5596 case Intrinsic::amdgcn_s_sleep_var:
5599 case Intrinsic::amdgcn_s_barrier_join:
5600 case Intrinsic::amdgcn_s_wakeup_barrier:
5603 case Intrinsic::amdgcn_s_barrier_init:
5604 case Intrinsic::amdgcn_s_barrier_signal_var:
5608 case Intrinsic::amdgcn_s_barrier_signal_isfirst: {
5609 const unsigned ResultSize = 1;
5611 AMDGPU::getValueMapping(AMDGPU::SGPRRegBankID, ResultSize);
5614 case Intrinsic::amdgcn_s_get_barrier_state:
5615 case Intrinsic::amdgcn_s_get_named_barrier_state: {
5620 case Intrinsic::amdgcn_pops_exiting_wave_id:
5622 case Intrinsic::amdgcn_tensor_load_to_lds_d2:
5623 case Intrinsic::amdgcn_tensor_store_from_lds_d2:
5624 case Intrinsic::amdgcn_tensor_load_to_lds:
5625 case Intrinsic::amdgcn_tensor_store_from_lds: {
5628 for (
unsigned I = 1;
I <
MI.getNumOperands(); ++
I) {
5629 if (
MI.getOperand(
I).isReg()) {
5633 OpdsMapping[
I] = AMDGPU::getValueMapping(OpBank,
Size);
5638 case Intrinsic::amdgcn_s_prefetch_data: {
5643 case Intrinsic::amdgcn_flat_prefetch:
5644 case Intrinsic::amdgcn_global_prefetch:
5651 case AMDGPU::G_SELECT: {
5652 unsigned Size =
MRI.getType(
MI.getOperand(0).getReg()).getSizeInBits();
5654 AMDGPU::SGPRRegBankID);
5656 AMDGPU::SGPRRegBankID);
5657 bool SGPRSrcs = Op2Bank == AMDGPU::SGPRRegBankID &&
5658 Op3Bank == AMDGPU::SGPRRegBankID;
5660 unsigned CondBankDefault = SGPRSrcs ?
5661 AMDGPU::SGPRRegBankID : AMDGPU::VCCRegBankID;
5664 if (CondBank == AMDGPU::SGPRRegBankID)
5665 CondBank = SGPRSrcs ? AMDGPU::SGPRRegBankID : AMDGPU::VCCRegBankID;
5666 else if (CondBank == AMDGPU::VGPRRegBankID)
5667 CondBank = AMDGPU::VCCRegBankID;
5669 unsigned Bank = SGPRSrcs && CondBank == AMDGPU::SGPRRegBankID ?
5670 AMDGPU::SGPRRegBankID : AMDGPU::VGPRRegBankID;
5672 assert(CondBank == AMDGPU::VCCRegBankID || CondBank == AMDGPU::SGPRRegBankID);
5676 OpdsMapping[0] = AMDGPU::getValueMappingSGPR64Only(Bank,
Size);
5677 OpdsMapping[1] = AMDGPU::getValueMapping(CondBank, 1);
5678 OpdsMapping[2] = AMDGPU::getValueMappingSGPR64Only(Bank,
Size);
5679 OpdsMapping[3] = AMDGPU::getValueMappingSGPR64Only(Bank,
Size);
5681 OpdsMapping[0] = AMDGPU::getValueMapping(Bank,
Size);
5682 OpdsMapping[1] = AMDGPU::getValueMapping(CondBank, 1);
5683 OpdsMapping[2] = AMDGPU::getValueMapping(Bank,
Size);
5684 OpdsMapping[3] = AMDGPU::getValueMapping(Bank,
Size);
5690 case AMDGPU::G_SI_CALL: {
5691 OpdsMapping[0] = AMDGPU::getValueMapping(AMDGPU::SGPRRegBankID, 64);
5697 for (
unsigned I = 4;
I <
MI.getNumOperands(); ++
I) {
5698 if (
MI.getOperand(
I).isReg()) {
5702 OpdsMapping[
I] = AMDGPU::getValueMapping(OpBank,
Size);
5707 case AMDGPU::G_LOAD:
5708 case AMDGPU::G_ZEXTLOAD:
5709 case AMDGPU::G_SEXTLOAD:
5712 case AMDGPU::G_ATOMICRMW_XCHG:
5713 case AMDGPU::G_ATOMICRMW_ADD:
5714 case AMDGPU::G_ATOMICRMW_SUB:
5715 case AMDGPU::G_ATOMICRMW_AND:
5716 case AMDGPU::G_ATOMICRMW_OR:
5717 case AMDGPU::G_ATOMICRMW_XOR:
5718 case AMDGPU::G_ATOMICRMW_MAX:
5719 case AMDGPU::G_ATOMICRMW_MIN:
5720 case AMDGPU::G_ATOMICRMW_UMAX:
5721 case AMDGPU::G_ATOMICRMW_UMIN:
5722 case AMDGPU::G_ATOMICRMW_FADD:
5723 case AMDGPU::G_ATOMICRMW_FMIN:
5724 case AMDGPU::G_ATOMICRMW_FMAX:
5725 case AMDGPU::G_ATOMICRMW_UINC_WRAP:
5726 case AMDGPU::G_ATOMICRMW_UDEC_WRAP:
5727 case AMDGPU::G_ATOMICRMW_USUB_COND:
5728 case AMDGPU::G_ATOMICRMW_USUB_SAT:
5729 case AMDGPU::G_AMDGPU_ATOMIC_CMPXCHG: {
5735 case AMDGPU::G_ATOMIC_CMPXCHG: {
5742 case AMDGPU::G_BRCOND: {
5744 AMDGPU::SGPRRegBankID);
5745 assert(
MRI.getType(
MI.getOperand(0).getReg()).getSizeInBits() == 1);
5746 if (Bank != AMDGPU::SGPRRegBankID)
5747 Bank = AMDGPU::VCCRegBankID;
5749 OpdsMapping[0] = AMDGPU::getValueMapping(Bank, 1);
5752 case AMDGPU::G_INTRINSIC_FPTRUNC_ROUND:
5754 case AMDGPU::G_PREFETCH:
5757 case AMDGPU::G_AMDGPU_WHOLE_WAVE_FUNC_SETUP:
5758 case AMDGPU::G_AMDGPU_WHOLE_WAVE_FUNC_RETURN:
5759 OpdsMapping[0] = AMDGPU::getValueMapping(AMDGPU::VCCRegBankID, 1);
5765 MI.getNumOperands());