37 "amdgpu-mfma-vgpr-form",
38 cl::desc(
"Whether to force use VGPR for Opc and Dest of MFMA. If "
39 "unspecified, default to compiler heuristics"),
53 UserSGPRInfo(
F, *STI), WorkGroupIDX(
false), WorkGroupIDY(
false),
55 PrivateSegmentWaveByteOffset(
false), WorkItemIDX(
false),
57 GITPtrHigh(0xffffffff), HighBitsOf32BitAddress(0),
58 IsWholeWaveFunction(
F.getCallingConv() ==
61 FlatWorkGroupSizes = ST.getFlatWorkGroupSizes(
F);
62 WavesPerEU = ST.getWavesPerEU(
F);
63 MaxNumWorkGroups = ST.getMaxNumWorkGroups(
F);
64 assert(MaxNumWorkGroups.size() == 3);
69 if (DynamicVGPRBlockSize == 0 && ST.isDynamicVGPREnabled())
70 DynamicVGPRBlockSize = ST.getDynamicVGPRBlockSize();
72 Occupancy = ST.computeOccupancy(
F,
getLDSSize()).second;
75 VRegFlags.reserve(1024);
87 if (ST.hasGFX90AInsts()) {
90 auto [MinNumAGPRAttr, MaxNumAGPRAttr] =
93 MinNumAGPRs = MinNumAGPRAttr;
101 FrameOffsetReg = AMDGPU::SGPR33;
102 StackPtrOffsetReg = AMDGPU::SGPR32;
104 if (!ST.hasFlatScratchEnabled()) {
108 ? AMDGPU::SGPR48_SGPR49_SGPR50_SGPR51
109 : ScratchRSrcReg = AMDGPU::SGPR0_SGPR1_SGPR2_SGPR3;
111 ArgInfo.PrivateSegmentBuffer =
115 if (!
F.hasFnAttribute(
"amdgpu-no-implicitarg-ptr") &&
117 ImplicitArgPtr =
true;
119 ImplicitArgPtr =
false;
126 ST.hasArchitectedSGPRs())) {
127 if (IsKernel || !
F.hasFnAttribute(
"amdgpu-no-workgroup-id-x") ||
128 !
F.hasFnAttribute(
"amdgpu-no-cluster-id-x"))
131 if (!
F.hasFnAttribute(
"amdgpu-no-workgroup-id-y") ||
132 !
F.hasFnAttribute(
"amdgpu-no-cluster-id-y"))
135 if (!
F.hasFnAttribute(
"amdgpu-no-workgroup-id-z") ||
136 !
F.hasFnAttribute(
"amdgpu-no-cluster-id-z"))
141 if (IsKernel || !
F.hasFnAttribute(
"amdgpu-no-workitem-id-x"))
144 if (!
F.hasFnAttribute(
"amdgpu-no-workitem-id-y") &&
145 ST.getMaxWorkitemID(
F, 1) != 0)
148 if (!
F.hasFnAttribute(
"amdgpu-no-workitem-id-z") &&
149 ST.getMaxWorkitemID(
F, 2) != 0)
152 if (!IsKernel && !
F.hasFnAttribute(
"amdgpu-no-lds-kernel-id"))
162 if (!ST.hasArchitectedFlatScratch()) {
163 PrivateSegmentWaveByteOffset =
true;
168 ArgInfo.PrivateSegmentWaveByteOffset =
173 Attribute A =
F.getFnAttribute(
"amdgpu-git-ptr-high");
178 A =
F.getFnAttribute(
"amdgpu-32bit-address-high-bits");
179 S =
A.getValueAsString();
183 MaxMemoryClusterDWords =
F.getFnAttributeAsParsedInteger(
189 if (ST.hasMAIInsts() && !ST.hasGFX90AInsts()) {
191 AMDGPU::VGPR_32RegClass.getRegister(ST.getMaxNumVGPRs(
F) - 1);
212 ArgInfo.PrivateSegmentBuffer =
214 getNextUserSGPR(), AMDGPU::sub0, &AMDGPU::SGPR_128RegClass));
216 return ArgInfo.PrivateSegmentBuffer.getRegister();
221 getNextUserSGPR(), AMDGPU::sub0, &AMDGPU::SReg_64RegClass));
223 return ArgInfo.DispatchPtr.getRegister();
228 getNextUserSGPR(), AMDGPU::sub0, &AMDGPU::SReg_64RegClass));
230 return ArgInfo.QueuePtr.getRegister();
234 ArgInfo.KernargSegmentPtr
236 getNextUserSGPR(), AMDGPU::sub0, &AMDGPU::SReg_64RegClass));
238 return ArgInfo.KernargSegmentPtr.getRegister();
243 getNextUserSGPR(), AMDGPU::sub0, &AMDGPU::SReg_64RegClass));
245 return ArgInfo.DispatchID.getRegister();
250 getNextUserSGPR(), AMDGPU::sub0, &AMDGPU::SReg_64RegClass));
252 return ArgInfo.FlatScratchInit.getRegister();
258 return ArgInfo.PrivateSegmentSize.getRegister();
263 getNextUserSGPR(), AMDGPU::sub0, &AMDGPU::SReg_64RegClass));
265 return ArgInfo.ImplicitBufferPtr.getRegister();
271 return ArgInfo.LDSKernelId.getRegister();
276 unsigned AllocSizeDWord,
int KernArgIdx,
int PaddingSGPRs) {
277 auto [It, Inserted] = ArgInfo.PreloadKernArgs.try_emplace(KernArgIdx);
278 assert(Inserted &&
"Preload kernel argument allocated twice.");
279 NumUserSGPRs += PaddingSGPRs;
283 if (!ArgInfo.FirstKernArgPreloadReg)
284 ArgInfo.FirstKernArgPreloadReg = getNextUserSGPR();
286 TRI.getMatchingSuperReg(getNextUserSGPR(), AMDGPU::sub0, RC);
287 auto &Regs = It->second.Regs;
289 (RC == &AMDGPU::SReg_32RegClass || RC == &AMDGPU::SReg_64RegClass)) {
290 Regs.push_back(PreloadReg);
291 NumUserSGPRs += AllocSizeDWord;
293 Regs.reserve(AllocSizeDWord);
294 for (
unsigned I = 0;
I < AllocSizeDWord; ++
I) {
295 Regs.push_back(getNextUserSGPR());
301 UserSGPRInfo.allocKernargPreloadSGPRs(AllocSizeDWord + PaddingSGPRs);
325 WWMSpills.insert(std::make_pair(
335 for (
auto &Reg : WWMSpills) {
337 CalleeSavedRegs.push_back(Reg);
339 ScratchRegs.push_back(Reg);
345 for (
unsigned I = 0; CSRegs[
I]; ++
I) {
346 if (CSRegs[
I] == Reg)
358 for (
unsigned I = 0, E = WWMVGPRs.
size();
I < E; ++
I) {
361 TRI->findUnusedRegister(MRI, &AMDGPU::VGPR_32RegClass, MF);
362 if (!NewReg || NewReg >= Reg)
368 WWMVGPRs[
I] = NewReg;
369 WWMReservedRegs.remove(Reg);
370 WWMReservedRegs.insert(NewReg);
375 auto *RegItr =
llvm::find(SpillPhysVGPRs, Reg);
376 if (RegItr != SpillPhysVGPRs.end()) {
377 unsigned Idx = std::distance(SpillPhysVGPRs.begin(), RegItr);
378 SpillPhysVGPRs[Idx] = NewReg;
383 SavedVGPRs.
reset(Reg);
386 MBB.removeLiveIn(Reg);
387 MBB.sortUniqueLiveIns();
394bool SIMachineFunctionInfo::allocateVirtualVGPRForSGPRSpills(
400 SpillVGPRs.push_back(LaneVGPR);
402 LaneVGPR = SpillVGPRs.back();
405 SGPRSpillsToVirtualVGPRLanes[FI].emplace_back(LaneVGPR, LaneIndex);
409bool SIMachineFunctionInfo::allocatePhysicalVGPRForSGPRSpills(
410 MachineFunction &MF,
int FI,
unsigned LaneIndex,
bool IsPrologEpilog) {
412 const SIRegisterInfo *
TRI =
ST.getRegisterInfo();
419 LaneVGPR =
TRI->findUnusedRegister(MRI, &AMDGPU::VGPR_32RegClass, MF,
421 if (LaneVGPR == AMDGPU::NoRegister) {
424 SGPRSpillsToPhysicalVGPRLanes.erase(FI);
432 for (MachineBasicBlock &
MBB : MF) {
436 SpillPhysVGPRs.push_back(LaneVGPR);
438 LaneVGPR = SpillPhysVGPRs.back();
441 SGPRSpillsToPhysicalVGPRLanes[FI].emplace_back(LaneVGPR, LaneIndex);
447 bool IsPrologEpilog) {
448 std::vector<SIRegisterInfo::SpilledReg> &SpillLanes =
449 SpillToPhysVGPRLane ? SGPRSpillsToPhysicalVGPRLanes[FI]
450 : SGPRSpillsToVirtualVGPRLanes[FI];
453 if (!SpillLanes.empty())
458 unsigned WaveSize = ST.getWavefrontSize();
460 unsigned Size = FrameInfo.getObjectSize(FI);
461 unsigned NumLanes =
Size / 4;
463 if (NumLanes > WaveSize)
466 assert(
Size >= 4 &&
"invalid sgpr spill size");
467 assert(ST.getRegisterInfo()->spillSGPRToVGPR() &&
468 "not spilling SGPRs to VGPRs");
470 unsigned &NumSpillLanes = SpillToPhysVGPRLane ? NumPhysicalVGPRSpillLanes
471 : NumVirtualVGPRSpillLanes;
473 for (
unsigned I = 0;
I < NumLanes; ++
I, ++NumSpillLanes) {
474 unsigned LaneIndex = (NumSpillLanes % WaveSize);
476 bool Allocated = SpillToPhysVGPRLane
477 ? allocatePhysicalVGPRForSGPRSpills(MF, FI, LaneIndex,
479 : allocateVirtualVGPRForSGPRSpills(MF, FI, LaneIndex);
499 assert(ST.hasMAIInsts() && FrameInfo.isSpillSlotObjectIndex(FI));
501 auto &Spill = VGPRToAGPRSpills[FI];
504 if (!Spill.Lanes.empty())
505 return Spill.FullyAllocated;
507 unsigned Size = FrameInfo.getObjectSize(FI);
508 unsigned NumLanes =
Size / 4;
509 Spill.Lanes.resize(NumLanes, AMDGPU::NoRegister);
512 isAGPRtoVGPR ? AMDGPU::VGPR_32RegClass : AMDGPU::AGPR_32RegClass;
515 auto &SpillRegs = isAGPRtoVGPR ? SpillAGPR : SpillVGPR;
517 Spill.FullyAllocated =
true;
532 OtherUsedRegs.
set(Reg);
534 OtherUsedRegs.
set(Reg);
537 for (
int I = NumLanes - 1;
I >= 0; --
I) {
538 NextSpillReg = std::find_if(
539 NextSpillReg, Regs.
end(), [&MRI, &OtherUsedRegs](
MCPhysReg Reg) {
540 return MRI.isAllocatable(Reg) && !MRI.isPhysRegUsed(Reg) &&
544 if (NextSpillReg == Regs.
end()) {
545 Spill.FullyAllocated =
false;
549 OtherUsedRegs.
set(*NextSpillReg);
552 Spill.Lanes[
I] = *NextSpillReg++;
555 return Spill.FullyAllocated;
568 SGPRSpillsToVirtualVGPRLanes.erase(R.first);
573 if (!ResetSGPRSpillStackIDs) {
576 SGPRSpillsToPhysicalVGPRLanes.erase(R.first);
579 bool HaveSGPRToMemory =
false;
581 if (ResetSGPRSpillStackIDs) {
589 HaveSGPRToMemory =
true;
595 for (
auto &R : VGPRToAGPRSpills) {
600 return HaveSGPRToMemory;
610 TRI.getSpillAlign(AMDGPU::SGPR_32RegClass),
false);
614MCPhysReg SIMachineFunctionInfo::getNextUserSGPR()
const {
615 assert(NumSystemSGPRs == 0 &&
"System SGPRs must be added after user SGPRs");
616 return AMDGPU::SGPR0 + NumUserSGPRs;
619MCPhysReg SIMachineFunctionInfo::getNextSystemSGPR()
const {
620 return AMDGPU::SGPR0 + NumUserSGPRs + NumSystemSGPRs;
623void SIMachineFunctionInfo::MRI_NoteNewVirtualRegister(
Register Reg) {
627void SIMachineFunctionInfo::MRI_NoteCloneVirtualRegister(
Register NewReg,
629 VRegFlags.grow(NewReg);
630 VRegFlags[NewReg] = VRegFlags[SrcReg];
636 if (!ST.isAmdPalOS())
639 if (ST.hasMergedShaders()) {
645 GitPtrLo = AMDGPU::SGPR8;
664static std::optional<yaml::SIArgumentInfo>
669 auto convertArg = [&](std::optional<yaml::SIArgument> &
A,
676 if (Arg.isRegister()) {
683 SA.
Mask = Arg.getMask();
703 ArgInfo.PrivateSegmentWaveByteOffset);
712 if (
ArgInfo.FirstKernArgPreloadReg) {
715 "FirstKernArgPreloadReg must be a physical register");
817 SourceRange = YamlMFI.
ScavengeFI->SourceRange;
820 ScavengeFI = *FIOrErr;
822 ScavengeFI = std::nullopt;
828 auto [MinNumAGPR, MaxNumAGPR] =
831 return MinNumAGPR != 0u;
assert(UImm &&(UImm !=~static_cast< T >(0)) &&"Invalid immediate!")
Provides AMDGPU specific target descriptions.
Base class for AMDGPU specific classes of TargetSubtarget.
static GCRegistry::Add< ErlangGC > A("erlang", "erlang-compatible garbage collector")
AMD GCN specific subclass of TargetSubtarget.
Register const TargetRegisterInfo * TRI
Promote Memory to Register
const GCNTargetMachine & getTM(const GCNSubtarget *STI)
static cl::opt< bool, true > MFMAVGPRFormOpt("amdgpu-mfma-vgpr-form", cl::desc("Whether to force use VGPR for Opc and Dest of MFMA. If " "unspecified, default to compiler heuristics"), cl::location(SIMachineFunctionInfo::MFMAVGPRForm), cl::init(true), cl::Hidden)
static std::optional< yaml::SIArgumentInfo > convertArgumentInfo(const AMDGPUFunctionArgInfo &ArgInfo, const TargetRegisterInfo &TRI)
static yaml::StringValue regToString(Register Reg, const TargetRegisterInfo &TRI)
Interface definition for SIRegisterInfo.
AMDGPUMachineFunction(const Function &F, const AMDGPUSubtarget &ST)
uint32_t getLDSSize() const
Align DynLDSAlign
Align for dynamic shared memory if any.
uint32_t LDSSize
Number of bytes in the LDS that are being used.
bool isChainFunction() const
uint64_t ExplicitKernArgSize
bool hasInitWholeWave() const
bool isEntryFunction() const
static ClusterDimsAttr get(const Function &F)
Functions, function parameters, and return types can have attributes to indicate how they should be t...
void resize(unsigned N, bool t=false)
resize - Grow or shrink the bitvector.
void setBitsInMask(const uint32_t *Mask, unsigned MaskWords=~0u)
setBitsInMask - Add '1' bits from Mask to this vector.
Lightweight error class with error context and mandatory checking.
CallingConv::ID getCallingConv() const
getCallingConv()/setCallingConv(CC) - These method get and set the calling convention of this functio...
const SITargetLowering * getTargetLowering() const override
LLVM_ABI void sortUniqueLiveIns()
Sorts and uniques the LiveIns vector.
void addLiveIn(MCRegister PhysReg, LaneBitmask LaneMask=LaneBitmask::getAll())
Adds the specified register as a live in.
The MachineFrameInfo class represents an abstract stack frame until prolog/epilog code is inserted.
LLVM_ABI int CreateStackObject(uint64_t Size, Align Alignment, bool isSpillSlot, const AllocaInst *Alloca=nullptr, uint8_t ID=0)
Create a new statically sized stack object, returning a nonnegative identifier to represent it.
LLVM_ABI int CreateSpillStackObject(uint64_t Size, Align Alignment)
Create a new statically sized stack object that represents a spill slot, returning a nonnegative iden...
void setStackID(int ObjectIdx, uint8_t ID)
bool hasTailCall() const
Returns true if the function contains a tail call.
void RemoveStackObject(int ObjectIdx)
Remove or mark dead a statically sized stack object.
int getObjectIndexEnd() const
Return one past the maximum frame object index.
uint8_t getStackID(int ObjectIdx) const
int getObjectIndexBegin() const
Return the minimum frame object index.
const TargetSubtargetInfo & getSubtarget() const
getSubtarget - Return the subtarget for which this machine code is being compiled.
MachineFrameInfo & getFrameInfo()
getFrameInfo - Return the frame info object for the current function.
MachineRegisterInfo & getRegInfo()
getRegInfo - Return information about the registers currently in use.
Function & getFunction()
Return the LLVM function that this machine code represents.
Ty * cloneInfo(const Ty &Old)
MachineRegisterInfo - Keep track of information for virtual and physical registers,...
LLVM_ABI Register createVirtualRegister(const TargetRegisterClass *RegClass, StringRef Name="")
createVirtualRegister - Create and return a new virtual register in the function with the specified r...
LLVM_ABI const MCPhysReg * getCalleeSavedRegs() const
Returns list of callee saved registers.
void reserveReg(MCRegister PhysReg, const TargetRegisterInfo *TRI)
reserveReg – Mark a register as reserved so checks like isAllocatable will not suggest using it.
LLVM_ABI void replaceRegWith(Register FromReg, Register ToReg)
replaceRegWith - Replace all instances of FromReg with ToReg in the machine function.
This interface provides simple read-only access to a block of memory, and provides simple methods for...
virtual StringRef getBufferIdentifier() const
Return an identifier for this buffer, typically the filename it was read from.
Wrapper class representing virtual and physical registers.
This class keeps track of the SPI_SP_INPUT_ADDR config register, which tells the hardware which inter...
bool initializeBaseYamlFields(const yaml::SIMachineFunctionInfo &YamlMFI, const MachineFunction &MF, PerFunctionMIParsingState &PFS, SMDiagnostic &Error, SMRange &SourceRange)
void shiftWwmVGPRsToLowestRange(MachineFunction &MF, SmallVectorImpl< Register > &WWMVGPRs, BitVector &SavedVGPRs)
Register addPrivateSegmentSize(const SIRegisterInfo &TRI)
void allocateWWMSpill(MachineFunction &MF, Register VGPR, uint64_t Size=4, Align Alignment=Align(4))
Register addDispatchPtr(const SIRegisterInfo &TRI)
Register getLongBranchReservedReg() const
Register addFlatScratchInit(const SIRegisterInfo &TRI)
unsigned getMaxWavesPerEU() const
ArrayRef< Register > getSGPRSpillPhysVGPRs() const
int getScavengeFI(MachineFrameInfo &MFI, const SIRegisterInfo &TRI)
Register addQueuePtr(const SIRegisterInfo &TRI)
SIMachineFunctionInfo(const SIMachineFunctionInfo &MFI)=default
Register getGITPtrLoReg(const MachineFunction &MF) const
bool allocateVGPRSpillToAGPR(MachineFunction &MF, int FI, bool isAGPRtoVGPR)
Reserve AGPRs or VGPRs to support spilling for FrameIndex FI.
void splitWWMSpillRegisters(MachineFunction &MF, SmallVectorImpl< std::pair< Register, int > > &CalleeSavedRegs, SmallVectorImpl< std::pair< Register, int > > &ScratchRegs) const
Register getSGPRForEXECCopy() const
bool mayUseAGPRs(const Function &F) const
bool isCalleeSavedReg(const MCPhysReg *CSRegs, MCPhysReg Reg) const
Register addLDSKernelId()
Register getVGPRForAGPRCopy() const
bool allocateSGPRSpillToVGPRLane(MachineFunction &MF, int FI, bool SpillToPhysVGPRLane=false, bool IsPrologEpilog=false)
Register addKernargSegmentPtr(const SIRegisterInfo &TRI)
Register addDispatchID(const SIRegisterInfo &TRI)
bool removeDeadFrameIndices(MachineFrameInfo &MFI, bool ResetSGPRSpillStackIDs)
If ResetSGPRSpillStackIDs is true, reset the stack ID from sgpr-spill to the default stack.
MachineFunctionInfo * clone(BumpPtrAllocator &Allocator, MachineFunction &DestMF, const DenseMap< MachineBasicBlock *, MachineBasicBlock * > &Src2DstMBB) const override
Make a functionally equivalent copy of this MachineFunctionInfo in MF.
bool checkIndexInPrologEpilogSGPRSpills(int FI) const
Register addPrivateSegmentBuffer(const SIRegisterInfo &TRI)
const ReservedRegSet & getWWMReservedRegs() const
std::optional< int > getOptionalScavengeFI() const
Register addImplicitBufferPtr(const SIRegisterInfo &TRI)
void limitOccupancy(const MachineFunction &MF)
SmallVectorImpl< MCRegister > * addPreloadedKernArg(const SIRegisterInfo &TRI, const TargetRegisterClass *RC, unsigned AllocSizeDWord, int KernArgIdx, int PaddingSGPRs)
void reserveWWMRegister(Register Reg)
static bool isChainScratchRegister(Register VGPR)
Instances of this class encapsulate one diagnostic report, allowing printing to a raw_ostream as a ca...
Represents a location in source code.
Represents a range in source code.
This class consists of common code factored out of the SmallVector class to reduce code duplication b...
typename SuperClass::const_iterator const_iterator
unsigned getMainFileID() const
const MemoryBuffer * getMemoryBuffer(unsigned i) const
StringRef - Represent a constant reference to a string, i.e.
bool consumeInteger(unsigned Radix, T &Result)
Parse the current string as an integer of the specified radix.
constexpr bool empty() const
empty - Check if the string is empty.
const TargetMachine & getTargetMachine() const
ArrayRef< MCPhysReg > getRegisters() const
TargetRegisterInfo base class - We assume that the target defines a static array of TargetRegisterDes...
A raw_ostream that writes to an std::string.
unsigned getInitialPSInputAddr(const Function &F)
unsigned getDynamicVGPRBlockSize(const Function &F)
LLVM_READNONE constexpr bool isChainCC(CallingConv::ID CC)
std::pair< unsigned, unsigned > getIntegerPairAttribute(const Function &F, StringRef Name, std::pair< unsigned, unsigned > Default, bool OnlyFirstRequired)
LLVM_READNONE constexpr bool isGraphics(CallingConv::ID CC)
CallingConv Namespace - This namespace contains an enum with a value for the well-known calling conve...
unsigned ID
LLVM IR allows to use arbitrary numbers as calling convention identifiers.
@ AMDGPU_CS
Used for Mesa/AMDPAL compute shaders.
@ AMDGPU_KERNEL
Used for AMDGPU code object kernels.
@ AMDGPU_Gfx
Used for AMD graphics targets.
@ AMDGPU_HS
Used for Mesa/AMDPAL hull shaders (= tessellation control shaders).
@ AMDGPU_GS
Used for Mesa/AMDPAL geometry shaders.
@ AMDGPU_PS
Used for Mesa/AMDPAL pixel shaders.
@ SPIR_KERNEL
Used for SPIR kernel functions.
initializer< Ty > init(const Ty &Val)
LocationClass< Ty > location(Ty &L)
This is an optimization pass for GlobalISel generic memory operations.
auto find(R &&Range, const T &Val)
Provide wrappers to std::find which take ranges instead of having to pass begin/end explicitly.
iterator_range< early_inc_iterator_impl< detail::IterOfRange< RangeT > > > make_early_inc_range(RangeT &&Range)
Make a range that does early increment to allow mutation of the underlying range without disrupting i...
uint16_t MCPhysReg
An unsigned integer type large enough to represent all physical registers, but not necessarily virtua...
std::string toString(const APInt &I, unsigned Radix, bool Signed, bool formatAsCLiteral=false, bool UpperCase=true, bool InsertSeparators=false)
constexpr unsigned DefaultMemoryClusterDWordsLimit
BumpPtrAllocatorImpl<> BumpPtrAllocator
The standard BumpPtrAllocator which just uses the default template parameters.
LLVM_ABI Printable printReg(Register Reg, const TargetRegisterInfo *TRI=nullptr, unsigned SubIdx=0, const MachineRegisterInfo *MRI=nullptr)
Prints virtual and physical registers with or without a TRI instance.
static const AMDGPUFunctionArgInfo FixedABIFunctionInfo
This struct is a compact representation of a valid (non-zero power of two) alignment.
static ArgDescriptor createRegister(Register Reg, unsigned Mask=~0u)
Helper struct shared between Function Specialization and SCCP Solver.
MachineFunctionInfo - This class can be derived from and used by targets to hold private target-speci...
A serializaable representation of a reference to a stack object or fixed stack object.
This class should be specialized by any type that needs to be converted to/from a YAML mapping.
std::optional< SIArgument > PrivateSegmentWaveByteOffset
std::optional< SIArgument > WorkGroupIDY
std::optional< SIArgument > FlatScratchInit
std::optional< SIArgument > DispatchPtr
std::optional< SIArgument > DispatchID
std::optional< SIArgument > WorkItemIDY
std::optional< SIArgument > WorkGroupIDX
std::optional< SIArgument > ImplicitArgPtr
std::optional< SIArgument > QueuePtr
std::optional< SIArgument > WorkGroupInfo
std::optional< SIArgument > LDSKernelId
std::optional< SIArgument > ImplicitBufferPtr
std::optional< SIArgument > WorkItemIDX
std::optional< SIArgument > KernargSegmentPtr
std::optional< SIArgument > WorkItemIDZ
std::optional< SIArgument > PrivateSegmentSize
std::optional< SIArgument > PrivateSegmentBuffer
std::optional< SIArgument > FirstKernArgPreloadReg
std::optional< SIArgument > WorkGroupIDZ
std::optional< unsigned > Mask
static SIArgument createArgument(bool IsReg)
unsigned MaxMemoryClusterDWords
StringValue SGPRForEXECCopy
SmallVector< StringValue > WWMReservedRegs
uint32_t HighBitsOf32BitAddress
SIMachineFunctionInfo()=default
StringValue FrameOffsetReg
StringValue LongBranchReservedReg
unsigned NumKernargPreloadSGPRs
uint64_t ExplicitKernArgSize
uint16_t NumWaveDispatchSGPRs
void mappingImpl(yaml::IO &YamlIO) override
unsigned DynamicVGPRBlockSize
StringValue VGPRForAGPRCopy
std::optional< SIArgumentInfo > ArgInfo
SmallVector< StringValue, 2 > SpillPhysVGPRS
std::optional< FrameIndex > ScavengeFI
uint16_t NumWaveDispatchVGPRs
unsigned BytesInStackArgArea
unsigned ScratchReservedForDynamicVGPRs
StringValue ScratchRSrcReg
StringValue StackPtrOffsetReg
A wrapper around std::string which contains a source range that's being set during parsing.