33#define DEBUG_TYPE "gcn-subtarget"
35#define GET_SUBTARGETINFO_TARGET_DESC
36#define GET_SUBTARGETINFO_CTOR
37#define AMDGPUSubtarget GCNSubtarget
38#include "AMDGPUGenSubtargetInfo.inc"
42 "amdgpu-vgpr-index-mode",
43 cl::desc(
"Use GPR indexing mode instead of movrel for vector indexing"),
47 cl::desc(
"Enable the use of AA during codegen."),
52 cl::desc(
"Number of addresses from which to enable MIMG NSA."),
74 FullFS +=
"+flat-for-global,+unaligned-access-mode,+trap-handler,";
76 FullFS +=
"+enable-prt-strict-null,";
79 if (FS.contains_insensitive(
"+wavefrontsize")) {
80 if (!FS.contains_insensitive(
"wavefrontsize16"))
81 FullFS +=
"-wavefrontsize16,";
82 if (!FS.contains_insensitive(
"wavefrontsize32"))
83 FullFS +=
"-wavefrontsize32,";
84 if (!FS.contains_insensitive(
"wavefrontsize64"))
85 FullFS +=
"-wavefrontsize64,";
102 }
else if (!hasFeature(AMDGPU::FeatureWavefrontSize32) &&
103 !hasFeature(AMDGPU::FeatureWavefrontSize64)) {
107 ToggleFeature(AMDGPU::FeatureWavefrontSize32);
122 if (!
hasAddr64() && !FS.contains(
"flat-for-global") && !UseFlatForGlobal) {
123 ToggleFeature(AMDGPU::FeatureUseFlatForGlobal);
124 UseFlatForGlobal =
true;
128 if (!
hasFlat() && !FS.contains(
"flat-for-global") && UseFlatForGlobal) {
129 ToggleFeature(AMDGPU::FeatureUseFlatForGlobal);
130 UseFlatForGlobal =
false;
158 "InstCacheLineSize must be a power of 2");
160 TargetID.setTargetIDFromFeaturesString(FS);
163 <<
TargetID.getXnackSetting() <<
'\n');
165 <<
TargetID.getSramEccSetting() <<
'\n');
172 if (hasFeature(AMDGPU::FeatureWavefrontSize32) &&
173 hasFeature(AMDGPU::FeatureWavefrontSize64)) {
175 F,
"must specify exactly one of wavefrontsize32 and wavefrontsize64"));
195 TSInfo = std::make_unique<AMDGPUSelectionDAGInfo>();
198 InlineAsmLoweringInfo =
200 Legalizer = std::make_unique<AMDGPULegalizerInfo>(*
this, TM);
201 RegBankInfo = std::make_unique<AMDGPURegisterBankInfo>(*
this);
203 std::make_unique<AMDGPUInstructionSelector>(*
this, *RegBankInfo, TM);
215 case AMDGPU::V_LSHLREV_B64_e64:
216 case AMDGPU::V_LSHLREV_B64_gfx10:
217 case AMDGPU::V_LSHLREV_B64_e64_gfx11:
218 case AMDGPU::V_LSHLREV_B64_e32_gfx12:
219 case AMDGPU::V_LSHLREV_B64_e64_gfx12:
220 case AMDGPU::V_LSHL_B64_e64:
221 case AMDGPU::V_LSHRREV_B64_e64:
222 case AMDGPU::V_LSHRREV_B64_gfx10:
223 case AMDGPU::V_LSHRREV_B64_e64_gfx11:
224 case AMDGPU::V_LSHRREV_B64_e64_gfx12:
225 case AMDGPU::V_LSHR_B64_e64:
226 case AMDGPU::V_ASHRREV_I64_e64:
227 case AMDGPU::V_ASHRREV_I64_gfx10:
228 case AMDGPU::V_ASHRREV_I64_e64_gfx11:
229 case AMDGPU::V_ASHRREV_I64_e64_gfx12:
230 case AMDGPU::V_ASHR_I64_e64:
240 case AMDGPU::V_CVT_F16_F32_e32:
241 case AMDGPU::V_CVT_F16_F32_e64:
242 case AMDGPU::V_CVT_F16_U16_e32:
243 case AMDGPU::V_CVT_F16_U16_e64:
244 case AMDGPU::V_CVT_F16_I16_e32:
245 case AMDGPU::V_CVT_F16_I16_e64:
246 case AMDGPU::V_RCP_F16_e64:
247 case AMDGPU::V_RCP_F16_e32:
248 case AMDGPU::V_RSQ_F16_e64:
249 case AMDGPU::V_RSQ_F16_e32:
250 case AMDGPU::V_SQRT_F16_e64:
251 case AMDGPU::V_SQRT_F16_e32:
252 case AMDGPU::V_LOG_F16_e64:
253 case AMDGPU::V_LOG_F16_e32:
254 case AMDGPU::V_EXP_F16_e64:
255 case AMDGPU::V_EXP_F16_e32:
256 case AMDGPU::V_SIN_F16_e64:
257 case AMDGPU::V_SIN_F16_e32:
258 case AMDGPU::V_COS_F16_e64:
259 case AMDGPU::V_COS_F16_e32:
260 case AMDGPU::V_FLOOR_F16_e64:
261 case AMDGPU::V_FLOOR_F16_e32:
262 case AMDGPU::V_CEIL_F16_e64:
263 case AMDGPU::V_CEIL_F16_e32:
264 case AMDGPU::V_TRUNC_F16_e64:
265 case AMDGPU::V_TRUNC_F16_e32:
266 case AMDGPU::V_RNDNE_F16_e64:
267 case AMDGPU::V_RNDNE_F16_e32:
268 case AMDGPU::V_FRACT_F16_e64:
269 case AMDGPU::V_FRACT_F16_e32:
270 case AMDGPU::V_FREXP_MANT_F16_e64:
271 case AMDGPU::V_FREXP_MANT_F16_e32:
272 case AMDGPU::V_FREXP_EXP_I16_F16_e64:
273 case AMDGPU::V_FREXP_EXP_I16_F16_e32:
274 case AMDGPU::V_LDEXP_F16_e64:
275 case AMDGPU::V_LDEXP_F16_e32:
276 case AMDGPU::V_LSHLREV_B16_e64:
277 case AMDGPU::V_LSHLREV_B16_e32:
278 case AMDGPU::V_LSHRREV_B16_e64:
279 case AMDGPU::V_LSHRREV_B16_e32:
280 case AMDGPU::V_ASHRREV_I16_e64:
281 case AMDGPU::V_ASHRREV_I16_e32:
282 case AMDGPU::V_ADD_U16_e64:
283 case AMDGPU::V_ADD_U16_e32:
284 case AMDGPU::V_SUB_U16_e64:
285 case AMDGPU::V_SUB_U16_e32:
286 case AMDGPU::V_SUBREV_U16_e64:
287 case AMDGPU::V_SUBREV_U16_e32:
288 case AMDGPU::V_MUL_LO_U16_e64:
289 case AMDGPU::V_MUL_LO_U16_e32:
290 case AMDGPU::V_ADD_F16_e64:
291 case AMDGPU::V_ADD_F16_e32:
292 case AMDGPU::V_SUB_F16_e64:
293 case AMDGPU::V_SUB_F16_e32:
294 case AMDGPU::V_SUBREV_F16_e64:
295 case AMDGPU::V_SUBREV_F16_e32:
296 case AMDGPU::V_MUL_F16_e64:
297 case AMDGPU::V_MUL_F16_e32:
298 case AMDGPU::V_MAX_F16_e64:
299 case AMDGPU::V_MAX_F16_e32:
300 case AMDGPU::V_MIN_F16_e64:
301 case AMDGPU::V_MIN_F16_e32:
302 case AMDGPU::V_MAX_U16_e64:
303 case AMDGPU::V_MAX_U16_e32:
304 case AMDGPU::V_MIN_U16_e64:
305 case AMDGPU::V_MIN_U16_e32:
306 case AMDGPU::V_MAX_I16_e64:
307 case AMDGPU::V_MAX_I16_e32:
308 case AMDGPU::V_MIN_I16_e64:
309 case AMDGPU::V_MIN_I16_e32:
310 case AMDGPU::V_MAD_F16_e64:
311 case AMDGPU::V_MAD_U16_e64:
312 case AMDGPU::V_MAD_I16_e64:
313 case AMDGPU::V_FMA_F16_e64:
314 case AMDGPU::V_DIV_FIXUP_F16_e64:
317 case AMDGPU::V_MADAK_F16:
318 case AMDGPU::V_MADMK_F16:
319 case AMDGPU::V_MAC_F16_e64:
320 case AMDGPU::V_MAC_F16_e32:
321 case AMDGPU::V_FMAMK_F16:
322 case AMDGPU::V_FMAAK_F16:
323 case AMDGPU::V_FMAC_F16_e64:
324 case AMDGPU::V_FMAC_F16_e32:
329 case AMDGPU::V_MAD_MIXLO_F16:
330 case AMDGPU::V_MAD_MIXHI_F16:
356 if (!enableSIScheduler())
363 Attribute PostRADirectionAttr =
F.getFnAttribute(
"amdgpu-post-ra-direction");
364 if (!PostRADirectionAttr.
isValid())
368 if (PostRADirectionStr ==
"topdown") {
371 }
else if (PostRADirectionStr ==
"bottomup") {
374 }
else if (PostRADirectionStr ==
"bidirectional") {
379 F,
F.getSubprogram(),
"invalid value for postRA direction attribute");
380 F.getContext().diagnose(Diag);
384 const char *DirStr =
"default";
390 DirStr =
"bidirectional";
392 dbgs() <<
"Post-MI-sched direction (" <<
F.getName() <<
"): " << DirStr
401 for (
auto &
MBB : MF) {
403 InstrInfo.fixImplicitOperands(
MI);
409 return InstrInfo.pseudoToMCOpcode(AMDGPU::V_MAD_F16_e64) != -1;
425 unsigned DynamicVGPRBlockSize)
const {
427 DynamicVGPRBlockSize);
435 if (HasFlatScratch || HasArchitectedFlatScratch) {
456 const bool KernelUsesFlatScratch = hasFlatAddressSpace();
460std::pair<unsigned, unsigned>
462 unsigned NumSGPRs,
unsigned NumVGPRs)
const {
474 MaxOcc = std::min(MaxOcc, std::min(SGPROcc, VGPROcc));
475 return {std::min(MinOcc, MaxOcc), MaxOcc};
479 const Function &
F, std::pair<unsigned, unsigned> WavesPerEU,
480 unsigned PreloadedSGPRs,
unsigned ReservedNumSGPRs)
const {
484 unsigned MaxAddressableNumSGPRs =
getMaxNumSGPRs(WavesPerEU.first,
true);
489 F.getFnAttributeAsParsedInteger(
"amdgpu-num-sgpr", MaxNumSGPRs);
491 if (Requested != MaxNumSGPRs) {
493 if (Requested && (Requested <= ReservedNumSGPRs))
503 unsigned InputNumSGPRs = PreloadedSGPRs;
504 if (Requested && Requested < InputNumSGPRs)
505 Requested = InputNumSGPRs;
509 if (Requested && Requested >
getMaxNumSGPRs(WavesPerEU.first,
false))
511 if (WavesPerEU.second && Requested &&
516 MaxNumSGPRs = Requested;
519 if (hasSGPRInitBug())
522 return std::min(MaxNumSGPRs - ReservedNumSGPRs, MaxAddressableNumSGPRs);
535 const unsigned MaxUserSGPRs =
536 USI::getNumUserSGPRForField(USI::PrivateSegmentBufferID) +
537 USI::getNumUserSGPRForField(USI::DispatchPtrID) +
538 USI::getNumUserSGPRForField(USI::QueuePtrID) +
539 USI::getNumUserSGPRForField(USI::KernargSegmentPtrID) +
540 USI::getNumUserSGPRForField(USI::DispatchIdID) +
541 USI::getNumUserSGPRForField(USI::FlatScratchInitID) +
542 USI::getNumUserSGPRForField(USI::ImplicitBufferPtrID);
545 const unsigned MaxSystemSGPRs = 1 +
552 const unsigned SyntheticSGPRs = 1;
554 return MaxUserSGPRs + MaxSystemSGPRs + SyntheticSGPRs;
563 const Function &
F, std::pair<unsigned, unsigned> NumVGPRBounds)
const {
564 const auto [Min, Max] = NumVGPRBounds;
569 unsigned Requested =
F.getFnAttributeAsParsedInteger(
"amdgpu-num-vgpr", Max);
570 if (Requested != Max && hasGFX90AInsts())
574 return std::clamp(Requested, Min, Max);
594std::pair<unsigned, unsigned>
598 unsigned MaxNumVGPRs = MaxVectorRegs;
599 unsigned MaxNumAGPRs = 0;
610 if (hasGFX90AInsts()) {
611 unsigned MinNumAGPRs = 0;
612 const unsigned TotalNumAGPRs = AMDGPU::AGPR_32RegClass.getNumRegs();
614 const std::pair<unsigned, unsigned> DefaultNumAGPR = {~0u, ~0u};
618 std::tie(MinNumAGPRs, MaxNumAGPRs) =
622 if (MinNumAGPRs == DefaultNumAGPR.first) {
624 MinNumAGPRs = MaxNumAGPRs = MaxVectorRegs / 2;
627 MinNumAGPRs =
alignTo(MinNumAGPRs, 4);
629 MinNumAGPRs = std::min(MinNumAGPRs, TotalNumAGPRs);
634 MaxNumAGPRs = std::min(std::max(MinNumAGPRs, MaxNumAGPRs), MaxVectorRegs);
635 MinNumAGPRs = std::min(std::min(MinNumAGPRs, TotalNumAGPRs), MaxNumAGPRs);
637 MaxNumVGPRs = std::min(MaxVectorRegs - MinNumAGPRs, NumArchVGPRs);
638 MaxNumAGPRs = std::min(MaxVectorRegs - MaxNumVGPRs, MaxNumAGPRs);
640 assert(MaxNumVGPRs + MaxNumAGPRs <= MaxVectorRegs &&
641 MaxNumAGPRs <= TotalNumAGPRs && MaxNumVGPRs <= NumArchVGPRs &&
642 "invalid register counts");
643 }
else if (hasMAIInsts()) {
645 MaxNumAGPRs = MaxNumVGPRs = MaxVectorRegs;
648 return std::pair(MaxNumVGPRs, MaxNumAGPRs);
658 AMDGPU::OpName UseName =
659 AMDGPU::getOperandIdxName(UseI.
getOpcode(), UseOpIdx);
661 case AMDGPU::OpName::src0:
662 return InstrInfo.getNamedOperand(UseI, AMDGPU::OpName::src0_modifiers);
663 case AMDGPU::OpName::src1:
664 return InstrInfo.getNamedOperand(UseI, AMDGPU::OpName::src1_modifiers);
665 case AMDGPU::OpName::src2:
666 return InstrInfo.getNamedOperand(UseI, AMDGPU::OpName::src2_modifiers);
679 if (!InstrInfo.isVOP3P(
I) || InstrInfo.isWMMA(
I) || InstrInfo.isSWMMAC(
I))
680 return AMDGPU::NoSubRegister;
685 return AMDGPU::NoSubRegister;
704 if ((!InstrInfo.isVOP3PMix(
I) && (!OpSel || !OpSelHi) &&
705 (OpSel || OpSelHi)) ||
706 (InstrInfo.isVOP3PMix(
I) && !OpSelHi))
707 return AMDGPU::NoSubRegister;
712 if (
unsigned SubRegIdx = OpSel ? AMDGPU::sub1 : AMDGPU::sub0;
713 TRI.getSubClassWithSubReg(RC, SubRegIdx) == RC)
715 if (
unsigned SubRegIdx = OpSel ? AMDGPU::hi16 : AMDGPU::lo16;
716 TRI.getSubClassWithSubReg(RC, SubRegIdx) == RC)
719 return AMDGPU::NoSubRegister;
725 int UseOpIdx)
const {
727 const MachineOperand &DefOp = DefI.
getOperand(DefOpIdx);
728 const MachineOperand &UseOp = UseI.
getOperand(UseOpIdx);
737 unsigned DefSubRegIdx = DefOp.
getSubReg();
738 if (DefReg.
isVirtual() && DefSubRegIdx == AMDGPU::NoSubRegister)
744 if (!
TRI->checkSubRegInterference(DefReg, DefSubRegIdx,
UseReg, UseSubRegIdx))
752 MCRegister DefMCReg =
753 DefSubRegIdx ?
TRI->getSubReg(DefReg, DefSubRegIdx) : DefReg.
asMCReg();
754 MCRegister UseMCReg =
756 return TRI->isSubRegisterEq(DefMCReg, UseMCReg) ? UseMCReg : DefMCReg;
769 if (
Register Reg = getRealSchedDependency(*DefI, DefOpIdx, *UseI, UseOpIdx)) {
782 for (++
I;
I != E &&
I->isBundledWithPred(); ++
I) {
783 if (
I->isMetaInstruction())
785 if (
I->modifiesRegister(Reg,
TRI))
797 for (++
I;
I != E &&
I->isBundledWithPred() && Lat; ++
I) {
798 if (
I->isMetaInstruction())
800 if (
I->readsRegister(Reg,
TRI))
810 Dep.
setLatency(InstrInfo.getSchedModel().computeOperandLatency(
811 DefI, DefOpIdx, UseI, UseOpIdx));
823 "amdgpu-nsa-threshold", -1);
825 return std::max(
Value, 2);
834 const bool IsKernel =
837 if (IsKernel && (!
F.arg_empty() || ST.getImplicitArgNumBytes(
F) != 0))
838 KernargSegmentPtr =
true;
840 bool IsAmdHsaOrMesa = ST.isAmdHsaOrMesa(
F);
841 if (IsAmdHsaOrMesa && !ST.hasFlatScratchEnabled())
842 PrivateSegmentBuffer =
true;
843 else if (ST.isMesaGfxShader(
F))
844 ImplicitBufferPtr =
true;
847 if (!
F.hasFnAttribute(
"amdgpu-no-dispatch-ptr"))
851 if (!
F.hasFnAttribute(
"amdgpu-no-queue-ptr"))
854 if (!
F.hasFnAttribute(
"amdgpu-no-dispatch-id"))
859 (IsAmdHsaOrMesa || ST.hasFlatScratchEnabled()) &&
862 (ST.hasFlatScratchEnabled() ||
864 !
F.hasFnAttribute(
"amdgpu-no-flat-scratch-init"))) &&
865 !ST.hasArchitectedFlatScratch()) {
866 FlatScratchInit = true;
896 NumKernargPreloadSGPRs += NumSGPRs;
897 NumUsedUserSGPRs += NumSGPRs;
assert(UImm &&(UImm !=~static_cast< T >(0)) &&"Invalid immediate!")
static cl::opt< bool > UseAA("aarch64-use-aa", cl::init(true), cl::desc("Enable the use of AA during codegen."))
This file describes how to lower LLVM calls to machine code calls.
This file declares the targeting of the InstructionSelector class for AMDGPU.
This file declares the targeting of the Machinelegalizer class for AMDGPU.
This file declares the targeting of the RegisterBankInfo class for AMDGPU.
The AMDGPU TargetMachine interface definition for hw codegen targets.
static GCRegistry::Add< CoreCLRGC > E("coreclr", "CoreCLR-compatible GC")
static cl::opt< unsigned > NSAThreshold("amdgpu-nsa-threshold", cl::desc("Number of addresses from which to enable MIMG NSA."), cl::init(2), cl::Hidden)
static cl::opt< bool > EnableVGPRIndexMode("amdgpu-vgpr-index-mode", cl::desc("Use GPR indexing mode instead of movrel for vector indexing"), cl::init(false))
static cl::opt< bool > UseAA("amdgpu-use-aa-in-codegen", cl::desc("Enable the use of AA during codegen."), cl::init(true))
static const MachineOperand * getVOP3PSourceModifierFromOpIdx(const MachineInstr &UseI, int UseOpIdx, const SIInstrInfo &InstrInfo)
static unsigned getEffectiveSubRegIdx(const SIRegisterInfo &TRI, const SIInstrInfo &InstrInfo, const MachineInstr &I, const MachineOperand &Op)
AMD GCN specific subclass of TargetSubtarget.
static Register UseReg(const MachineOperand &MO)
This file describes how to lower LLVM inline asm to machine code INLINEASM.
Register const TargetRegisterInfo * TRI
Promote Memory to Register
This file defines the SmallString class.
unsigned FlatOffsetBitWidth
std::pair< unsigned, unsigned > getWavesPerEU(const Function &F) const
std::pair< unsigned, unsigned > getOccupancyWithWorkGroupSizes(uint32_t LDSBytes, const Function &F) const
Subtarget's minimum/maximum occupancy, in number of waves per EU, that can be achieved when the only ...
unsigned getWavefrontSizeLog2() const
AMDGPUSubtarget(const Triple &TT)
unsigned AddressableLocalMemorySize
Functions, function parameters, and return types can have attributes to indicate how they should be t...
LLVM_ABI StringRef getValueAsString() const
Return the attribute's value as a string.
bool isValid() const
Return true if the attribute is any kind of attribute.
Diagnostic information for optimization failures.
Diagnostic information for unsupported feature in backend.
uint64_t getFnAttributeAsParsedInteger(StringRef Kind, uint64_t Default=0) const
For a string attribute Kind, parse attribute as an integer.
InstrItineraryData InstrItins
bool useVGPRIndexMode() const
void mirFileLoaded(MachineFunction &MF) const override
unsigned MaxPrivateElementSize
unsigned getAddressableNumArchVGPRs() const
unsigned getMinNumSGPRs(unsigned WavesPerEU) const
void ParseSubtargetFeatures(StringRef CPU, StringRef TuneCPU, StringRef FS)
unsigned getConstantBusLimit(unsigned Opcode) const
const InstrItineraryData * getInstrItineraryData() const override
void adjustSchedDependency(SUnit *Def, int DefOpIdx, SUnit *Use, int UseOpIdx, SDep &Dep, const TargetSchedModel *SchedModel) const override
void overridePostRASchedPolicy(MachineSchedPolicy &Policy, const SchedRegion &Region) const override
Align getStackAlignment() const
unsigned getMinNumVGPRs(unsigned WavesPerEU, unsigned DynamicVGPRBlockSize) const
bool isDynamicVGPREnabled() const
const SIRegisterInfo * getRegisterInfo() const override
unsigned getBaseMaxNumVGPRs(const Function &F, std::pair< unsigned, unsigned > NumVGPRBounds) const
bool zeroesHigh16BitsOfDest(unsigned Opcode) const
Returns if the result of this instruction with a 16-bit result returned in a 32-bit register implicit...
unsigned getBaseMaxNumSGPRs(const Function &F, std::pair< unsigned, unsigned > WavesPerEU, unsigned PreloadedSGPRs, unsigned ReservedNumSGPRs) const
unsigned getMaxNumPreloadedSGPRs() const
GCNSubtarget & initializeSubtargetDependencies(const Triple &TT, StringRef GPU, StringRef FS)
void overrideSchedPolicy(MachineSchedPolicy &Policy, const SchedRegion &Region) const override
std::pair< unsigned, unsigned > computeOccupancy(const Function &F, unsigned LDSSize=0, unsigned NumSGPRs=0, unsigned NumVGPRs=0) const
Subtarget's minimum/maximum occupancy, in number of waves per EU, that can be achieved when the only ...
unsigned getMaxNumVGPRs(unsigned WavesPerEU, unsigned DynamicVGPRBlockSize) const
const SITargetLowering * getTargetLowering() const override
unsigned getNSAThreshold(const MachineFunction &MF) const
unsigned getReservedNumSGPRs(const MachineFunction &MF) const
bool useAA() const override
unsigned getOccupancyWithNumVGPRs(unsigned VGPRs, unsigned DynamicVGPRBlockSize) const
Return the maximum number of waves per SIMD for kernels using VGPRs VGPRs.
unsigned InstCacheLineSize
unsigned getOccupancyWithNumSGPRs(unsigned SGPRs) const
Return the maximum number of waves per SIMD for kernels using SGPRs SGPRs.
unsigned getMaxWavesPerEU() const
Generation getGeneration() const
GCNSubtarget(const Triple &TT, StringRef GPU, StringRef FS, const GCNTargetMachine &TM)
unsigned getMaxNumSGPRs(unsigned WavesPerEU, bool Addressable) const
std::pair< unsigned, unsigned > getMaxNumVectorRegs(const Function &F) const
Return a pair of maximum numbers of VGPRs and AGPRs that meet the number of waves per execution unit ...
bool isXNACKEnabled() const
unsigned getBaseReservedNumSGPRs(const bool HasFlatScratch) const
unsigned getDynamicVGPRBlockSize() const
void checkSubtargetFeatures(const Function &F) const
Diagnose inconsistent subtarget features before attempting to codegen function F.
const SelectionDAGTargetInfo * getSelectionDAGInfo() const override
AMDGPU::IsaInfo::AMDGPUTargetID TargetID
static unsigned getNumUserSGPRForField(UserSGPRID ID)
bool hasKernargSegmentPtr() const
void allocKernargPreloadSGPRs(unsigned NumSGPRs)
bool hasDispatchID() const
bool hasPrivateSegmentBuffer() const
unsigned getNumFreeUserSGPRs()
bool hasImplicitBufferPtr() const
bool hasPrivateSegmentSize() const
bool hasDispatchPtr() const
GCNUserSGPRUsageInfo(const Function &F, const GCNSubtarget &ST)
bool hasFlatScratchInit() const
This is an important class for using LLVM in a threaded context.
instr_iterator instr_end()
Instructions::const_iterator const_instr_iterator
Function & getFunction()
Return the LLVM function that this machine code represents.
Ty * getInfo()
getInfo - Keep track of various per-function pieces of information for backends that would like to do...
Representation of each machine instruction.
unsigned getOpcode() const
Returns the opcode of this MachineInstr.
const MachineBasicBlock * getParent() const
const MachineOperand & getOperand(unsigned i) const
MachineOperand class - Representation of each machine instruction operand.
unsigned getSubReg() const
Register getReg() const
getReg - Returns the register number.
MachineRegisterInfo - Keep track of information for virtual and physical registers,...
Wrapper class representing virtual and physical registers.
MCRegister asMCReg() const
Utility to check-convert this value to a MCRegister.
constexpr bool isVirtual() const
Return true if the specified register number is in the virtual register namespace.
Kind getKind() const
Returns an enum value representing the kind of the dependence.
@ Data
Regular data dependence (aka true-dependence).
void setLatency(unsigned Lat)
Sets the latency for this edge.
@ Artificial
Arbitrary strong DAG edge (no real dependence).
unsigned getLatency() const
Returns the latency value for this edge, which roughly means the minimum number of cycles that must e...
Register getReg() const
Returns the register associated with this edge.
void setReg(Register Reg)
Assigns the associated register for this edge.
This class keeps track of the SPI_SP_INPUT_ADDR config register, which tells the hardware which inter...
unsigned getNumPreloadedSGPRs() const
std::pair< unsigned, unsigned > getWavesPerEU() const
GCNUserSGPRUsageInfo & getUserSGPRInfo()
Scheduling unit. This is a node in the scheduling DAG.
Targets can subclass this to parameterize the SelectionDAG lowering and instruction selection process...
SmallString - A SmallString is just a SmallVector with methods and accessors that make it work better...
StringRef - Represent a constant reference to a string, i.e.
Information about stack frame layout on the target.
Provide an instruction scheduling machine model to CodeGen passes.
Triple - Helper class for working with autoconf configuration names.
A Use represents the edge between a Value definition and its users.
LLVM Value Representation.
self_iterator getIterator()
@ FIXED_NUM_SGPRS_FOR_INIT_BUG
unsigned getNumWavesPerEUWithNumVGPRs(const MCSubtargetInfo *STI, unsigned NumVGPRs, unsigned DynamicVGPRBlockSize)
unsigned getMaxWavesPerEU(const MCSubtargetInfo *STI)
unsigned getLocalMemorySize(const MCSubtargetInfo *STI)
unsigned getEUsPerCU(const MCSubtargetInfo *STI)
unsigned getOccupancyWithNumSGPRs(unsigned SGPRs, unsigned MaxWaves, AMDGPUSubtarget::Generation Gen)
StringRef getSchedStrategy(const Function &F)
unsigned getMaxNumUserSGPRs(const MCSubtargetInfo &STI)
LLVM_READNONE constexpr bool isEntryFunctionCC(CallingConv::ID CC)
unsigned getDynamicVGPRBlockSize(const Function &F)
std::pair< unsigned, unsigned > getIntegerPairAttribute(const Function &F, StringRef Name, std::pair< unsigned, unsigned > Default, bool OnlyFirstRequired)
LLVM_READNONE constexpr bool isGraphics(CallingConv::ID CC)
unsigned ID
LLVM IR allows to use arbitrary numbers as calling convention identifiers.
@ AMDGPU_KERNEL
Used for AMDGPU code object kernels.
@ SPIR_KERNEL
Used for SPIR kernel functions.
initializer< Ty > init(const Ty &Val)
This is an optimization pass for GlobalISel generic memory operations.
constexpr bool isPowerOf2_32(uint32_t Value)
Return true if the argument is a power of two > 0.
LLVM_ABI raw_ostream & dbgs()
dbgs() - This returns a reference to a raw_ostream for debugging messages.
constexpr uint64_t alignTo(uint64_t Size, Align A)
Returns a multiple of A needed to store Size bytes.
DWARFExpression::Operation Op
This struct is a compact representation of a valid (non-zero power of two) alignment.
Define a generic scheduling policy for targets that don't provide their own MachineSchedStrategy.
bool ShouldTrackLaneMasks
Track LaneMasks to allow reordering of independent subregister writes of the same vreg.
A region of an MBB for scheduling.