| //===-- SIISelLowering.h - SI DAG Lowering Interface ------------*- C++ -*-===// | 
 | // | 
 | // Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. | 
 | // See https://llvm.org/LICENSE.txt for license information. | 
 | // SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception | 
 | // | 
 | //===----------------------------------------------------------------------===// | 
 | // | 
 | /// \file | 
 | /// SI DAG Lowering interface definition | 
 | // | 
 | //===----------------------------------------------------------------------===// | 
 |  | 
 | #ifndef LLVM_LIB_TARGET_AMDGPU_SIISELLOWERING_H | 
 | #define LLVM_LIB_TARGET_AMDGPU_SIISELLOWERING_H | 
 |  | 
 | #include "AMDGPUISelLowering.h" | 
 | #include "AMDGPUArgumentUsageInfo.h" | 
 | #include "llvm/CodeGen/MachineFunction.h" | 
 |  | 
 | namespace llvm { | 
 |  | 
 | class GCNSubtarget; | 
 | class SIMachineFunctionInfo; | 
 | class SIRegisterInfo; | 
 |  | 
 | namespace AMDGPU { | 
 | struct ImageDimIntrinsicInfo; | 
 | } | 
 |  | 
 | class SITargetLowering final : public AMDGPUTargetLowering { | 
 | private: | 
 |   const GCNSubtarget *Subtarget; | 
 |  | 
 | public: | 
 |   MVT getRegisterTypeForCallingConv(LLVMContext &Context, | 
 |                                     CallingConv::ID CC, | 
 |                                     EVT VT) const override; | 
 |   unsigned getNumRegistersForCallingConv(LLVMContext &Context, | 
 |                                          CallingConv::ID CC, | 
 |                                          EVT VT) const override; | 
 |  | 
 |   unsigned getVectorTypeBreakdownForCallingConv( | 
 |     LLVMContext &Context, CallingConv::ID CC, EVT VT, EVT &IntermediateVT, | 
 |     unsigned &NumIntermediates, MVT &RegisterVT) const override; | 
 |  | 
 | private: | 
 |   SDValue lowerKernArgParameterPtr(SelectionDAG &DAG, const SDLoc &SL, | 
 |                                    SDValue Chain, uint64_t Offset) const; | 
 |   SDValue getImplicitArgPtr(SelectionDAG &DAG, const SDLoc &SL) const; | 
 |   SDValue getLDSKernelId(SelectionDAG &DAG, const SDLoc &SL) const; | 
 |   SDValue lowerKernargMemParameter(SelectionDAG &DAG, EVT VT, EVT MemVT, | 
 |                                    const SDLoc &SL, SDValue Chain, | 
 |                                    uint64_t Offset, Align Alignment, | 
 |                                    bool Signed, | 
 |                                    const ISD::InputArg *Arg = nullptr) const; | 
 |   SDValue loadImplicitKernelArgument(SelectionDAG &DAG, MVT VT, const SDLoc &DL, | 
 |                                      Align Alignment, | 
 |                                      ImplicitParameter Param) const; | 
 |  | 
 |   SDValue lowerStackParameter(SelectionDAG &DAG, CCValAssign &VA, | 
 |                               const SDLoc &SL, SDValue Chain, | 
 |                               const ISD::InputArg &Arg) const; | 
 |   SDValue getPreloadedValue(SelectionDAG &DAG, | 
 |                             const SIMachineFunctionInfo &MFI, | 
 |                             EVT VT, | 
 |                             AMDGPUFunctionArgInfo::PreloadedValue) const; | 
 |  | 
 |   SDValue LowerGlobalAddress(AMDGPUMachineFunction *MFI, SDValue Op, | 
 |                              SelectionDAG &DAG) const override; | 
 |   SDValue lowerImplicitZextParam(SelectionDAG &DAG, SDValue Op, | 
 |                                  MVT VT, unsigned Offset) const; | 
 |   SDValue lowerImage(SDValue Op, const AMDGPU::ImageDimIntrinsicInfo *Intr, | 
 |                      SelectionDAG &DAG, bool WithChain) const; | 
 |   SDValue lowerSBuffer(EVT VT, SDLoc DL, SDValue Rsrc, SDValue Offset, | 
 |                        SDValue CachePolicy, SelectionDAG &DAG) const; | 
 |  | 
 |   SDValue lowerRawBufferAtomicIntrin(SDValue Op, SelectionDAG &DAG, | 
 |                                      unsigned NewOpcode) const; | 
 |   SDValue lowerStructBufferAtomicIntrin(SDValue Op, SelectionDAG &DAG, | 
 |                                         unsigned NewOpcode) const; | 
 |  | 
 |   SDValue lowerWorkitemID(SelectionDAG &DAG, SDValue Op, unsigned Dim, | 
 |                           const ArgDescriptor &ArgDesc) const; | 
 |  | 
 |   SDValue LowerINTRINSIC_WO_CHAIN(SDValue Op, SelectionDAG &DAG) const; | 
 |   SDValue LowerINTRINSIC_W_CHAIN(SDValue Op, SelectionDAG &DAG) const; | 
 |   SDValue LowerINTRINSIC_VOID(SDValue Op, SelectionDAG &DAG) const; | 
 |  | 
 |   // The raw.tbuffer and struct.tbuffer intrinsics have two offset args: offset | 
 |   // (the offset that is included in bounds checking and swizzling, to be split | 
 |   // between the instruction's voffset and immoffset fields) and soffset (the | 
 |   // offset that is excluded from bounds checking and swizzling, to go in the | 
 |   // instruction's soffset field).  This function takes the first kind of | 
 |   // offset and figures out how to split it between voffset and immoffset. | 
 |   std::pair<SDValue, SDValue> splitBufferOffsets(SDValue Offset, | 
 |                                                  SelectionDAG &DAG) const; | 
 |  | 
 |   SDValue widenLoad(LoadSDNode *Ld, DAGCombinerInfo &DCI) const; | 
 |   SDValue LowerLOAD(SDValue Op, SelectionDAG &DAG) const; | 
 |   SDValue LowerSELECT(SDValue Op, SelectionDAG &DAG) const; | 
 |   SDValue lowerFastUnsafeFDIV(SDValue Op, SelectionDAG &DAG) const; | 
 |   SDValue lowerFastUnsafeFDIV64(SDValue Op, SelectionDAG &DAG) const; | 
 |   SDValue lowerFDIV_FAST(SDValue Op, SelectionDAG &DAG) const; | 
 |   SDValue LowerFDIV16(SDValue Op, SelectionDAG &DAG) const; | 
 |   SDValue LowerFDIV32(SDValue Op, SelectionDAG &DAG) const; | 
 |   SDValue LowerFDIV64(SDValue Op, SelectionDAG &DAG) const; | 
 |   SDValue LowerFDIV(SDValue Op, SelectionDAG &DAG) const; | 
 |   SDValue LowerFFREXP(SDValue Op, SelectionDAG &DAG) const; | 
 |   SDValue LowerSTORE(SDValue Op, SelectionDAG &DAG) const; | 
 |   SDValue LowerTrig(SDValue Op, SelectionDAG &DAG) const; | 
 |   SDValue lowerFSQRTF64(SDValue Op, SelectionDAG &DAG) const; | 
 |   SDValue LowerATOMIC_CMP_SWAP(SDValue Op, SelectionDAG &DAG) const; | 
 |   SDValue LowerBRCOND(SDValue Op, SelectionDAG &DAG) const; | 
 |   SDValue LowerRETURNADDR(SDValue Op, SelectionDAG &DAG) const; | 
 |   SDValue adjustLoadValueType(unsigned Opcode, MemSDNode *M, | 
 |                               SelectionDAG &DAG, ArrayRef<SDValue> Ops, | 
 |                               bool IsIntrinsic = false) const; | 
 |  | 
 |   SDValue lowerIntrinsicLoad(MemSDNode *M, bool IsFormat, SelectionDAG &DAG, | 
 |                              ArrayRef<SDValue> Ops) const; | 
 |  | 
 |   // Call DAG.getMemIntrinsicNode for a load, but first widen a dwordx3 type to | 
 |   // dwordx4 if on SI. | 
 |   SDValue getMemIntrinsicNode(unsigned Opcode, const SDLoc &DL, SDVTList VTList, | 
 |                               ArrayRef<SDValue> Ops, EVT MemVT, | 
 |                               MachineMemOperand *MMO, SelectionDAG &DAG) const; | 
 |  | 
 |   SDValue handleD16VData(SDValue VData, SelectionDAG &DAG, | 
 |                          bool ImageStore = false) const; | 
 |  | 
 |   /// Converts \p Op, which must be of floating point type, to the | 
 |   /// floating point type \p VT, by either extending or truncating it. | 
 |   SDValue getFPExtOrFPRound(SelectionDAG &DAG, | 
 |                             SDValue Op, | 
 |                             const SDLoc &DL, | 
 |                             EVT VT) const; | 
 |  | 
 |   SDValue convertArgType( | 
 |     SelectionDAG &DAG, EVT VT, EVT MemVT, const SDLoc &SL, SDValue Val, | 
 |     bool Signed, const ISD::InputArg *Arg = nullptr) const; | 
 |  | 
 |   /// Custom lowering for ISD::FP_ROUND for MVT::f16. | 
 |   SDValue lowerFP_ROUND(SDValue Op, SelectionDAG &DAG) const; | 
 |   SDValue lowerFMINNUM_FMAXNUM(SDValue Op, SelectionDAG &DAG) const; | 
 |   SDValue lowerFLDEXP(SDValue Op, SelectionDAG &DAG) const; | 
 |   SDValue lowerXMULO(SDValue Op, SelectionDAG &DAG) const; | 
 |   SDValue lowerXMUL_LOHI(SDValue Op, SelectionDAG &DAG) const; | 
 |  | 
 |   SDValue getSegmentAperture(unsigned AS, const SDLoc &DL, | 
 |                              SelectionDAG &DAG) const; | 
 |  | 
 |   SDValue lowerADDRSPACECAST(SDValue Op, SelectionDAG &DAG) const; | 
 |   SDValue lowerINSERT_SUBVECTOR(SDValue Op, SelectionDAG &DAG) const; | 
 |   SDValue lowerINSERT_VECTOR_ELT(SDValue Op, SelectionDAG &DAG) const; | 
 |   SDValue lowerEXTRACT_VECTOR_ELT(SDValue Op, SelectionDAG &DAG) const; | 
 |   SDValue lowerVECTOR_SHUFFLE(SDValue Op, SelectionDAG &DAG) const; | 
 |   SDValue lowerSCALAR_TO_VECTOR(SDValue Op, SelectionDAG &DAG) const; | 
 |   SDValue lowerBUILD_VECTOR(SDValue Op, SelectionDAG &DAG) const; | 
 |  | 
 |   SDValue lowerTRAP(SDValue Op, SelectionDAG &DAG) const; | 
 |   SDValue lowerTrapEndpgm(SDValue Op, SelectionDAG &DAG) const; | 
 |   SDValue lowerTrapHsaQueuePtr(SDValue Op, SelectionDAG &DAG) const; | 
 |   SDValue lowerTrapHsa(SDValue Op, SelectionDAG &DAG) const; | 
 |   SDValue lowerDEBUGTRAP(SDValue Op, SelectionDAG &DAG) const; | 
 |  | 
 |   SDNode *adjustWritemask(MachineSDNode *&N, SelectionDAG &DAG) const; | 
 |  | 
 |   SDValue performUCharToFloatCombine(SDNode *N, | 
 |                                      DAGCombinerInfo &DCI) const; | 
 |   SDValue performFCopySignCombine(SDNode *N, DAGCombinerInfo &DCI) const; | 
 |  | 
 |   SDValue performSHLPtrCombine(SDNode *N, | 
 |                                unsigned AS, | 
 |                                EVT MemVT, | 
 |                                DAGCombinerInfo &DCI) const; | 
 |  | 
 |   SDValue performMemSDNodeCombine(MemSDNode *N, DAGCombinerInfo &DCI) const; | 
 |  | 
 |   SDValue splitBinaryBitConstantOp(DAGCombinerInfo &DCI, const SDLoc &SL, | 
 |                                    unsigned Opc, SDValue LHS, | 
 |                                    const ConstantSDNode *CRHS) const; | 
 |  | 
 |   SDValue performAndCombine(SDNode *N, DAGCombinerInfo &DCI) const; | 
 |   SDValue performOrCombine(SDNode *N, DAGCombinerInfo &DCI) const; | 
 |   SDValue performXorCombine(SDNode *N, DAGCombinerInfo &DCI) const; | 
 |   SDValue performZeroExtendCombine(SDNode *N, DAGCombinerInfo &DCI) const; | 
 |   SDValue performSignExtendInRegCombine(SDNode *N, DAGCombinerInfo &DCI) const; | 
 |   SDValue performClassCombine(SDNode *N, DAGCombinerInfo &DCI) const; | 
 |   SDValue getCanonicalConstantFP(SelectionDAG &DAG, const SDLoc &SL, EVT VT, | 
 |                                  const APFloat &C) const; | 
 |   SDValue performFCanonicalizeCombine(SDNode *N, DAGCombinerInfo &DCI) const; | 
 |  | 
 |   SDValue performFPMed3ImmCombine(SelectionDAG &DAG, const SDLoc &SL, | 
 |                                   SDValue Op0, SDValue Op1) const; | 
 |   SDValue performIntMed3ImmCombine(SelectionDAG &DAG, const SDLoc &SL, | 
 |                                    SDValue Src, SDValue MinVal, SDValue MaxVal, | 
 |                                    bool Signed) const; | 
 |   SDValue performMinMaxCombine(SDNode *N, DAGCombinerInfo &DCI) const; | 
 |   SDValue performFMed3Combine(SDNode *N, DAGCombinerInfo &DCI) const; | 
 |   SDValue performCvtPkRTZCombine(SDNode *N, DAGCombinerInfo &DCI) const; | 
 |   SDValue performExtractVectorEltCombine(SDNode *N, DAGCombinerInfo &DCI) const; | 
 |   SDValue performInsertVectorEltCombine(SDNode *N, DAGCombinerInfo &DCI) const; | 
 |   SDValue performFPRoundCombine(SDNode *N, DAGCombinerInfo &DCI) const; | 
 |  | 
 |   SDValue reassociateScalarOps(SDNode *N, SelectionDAG &DAG) const; | 
 |   unsigned getFusedOpcode(const SelectionDAG &DAG, | 
 |                           const SDNode *N0, const SDNode *N1) const; | 
 |   SDValue tryFoldToMad64_32(SDNode *N, DAGCombinerInfo &DCI) const; | 
 |   SDValue performAddCombine(SDNode *N, DAGCombinerInfo &DCI) const; | 
 |   SDValue performAddCarrySubCarryCombine(SDNode *N, DAGCombinerInfo &DCI) const; | 
 |   SDValue performSubCombine(SDNode *N, DAGCombinerInfo &DCI) const; | 
 |   SDValue performFAddCombine(SDNode *N, DAGCombinerInfo &DCI) const; | 
 |   SDValue performFSubCombine(SDNode *N, DAGCombinerInfo &DCI) const; | 
 |   SDValue performFMACombine(SDNode *N, DAGCombinerInfo &DCI) const; | 
 |   SDValue performSetCCCombine(SDNode *N, DAGCombinerInfo &DCI) const; | 
 |   SDValue performCvtF32UByteNCombine(SDNode *N, DAGCombinerInfo &DCI) const; | 
 |   SDValue performClampCombine(SDNode *N, DAGCombinerInfo &DCI) const; | 
 |   SDValue performRcpCombine(SDNode *N, DAGCombinerInfo &DCI) const; | 
 |  | 
 |   bool isLegalFlatAddressingMode(const AddrMode &AM) const; | 
 |   bool isLegalMUBUFAddressingMode(const AddrMode &AM) const; | 
 |  | 
 |   unsigned isCFIntrinsic(const SDNode *Intr) const; | 
 |  | 
 | public: | 
 |   /// \returns True if fixup needs to be emitted for given global value \p GV, | 
 |   /// false otherwise. | 
 |   bool shouldEmitFixup(const GlobalValue *GV) const; | 
 |  | 
 |   /// \returns True if GOT relocation needs to be emitted for given global value | 
 |   /// \p GV, false otherwise. | 
 |   bool shouldEmitGOTReloc(const GlobalValue *GV) const; | 
 |  | 
 |   /// \returns True if PC-relative relocation needs to be emitted for given | 
 |   /// global value \p GV, false otherwise. | 
 |   bool shouldEmitPCReloc(const GlobalValue *GV) const; | 
 |  | 
 |   /// \returns true if this should use a literal constant for an LDS address, | 
 |   /// and not emit a relocation for an LDS global. | 
 |   bool shouldUseLDSConstAddress(const GlobalValue *GV) const; | 
 |  | 
 |   /// Check if EXTRACT_VECTOR_ELT/INSERT_VECTOR_ELT (<n x e>, var-idx) should be | 
 |   /// expanded into a set of cmp/select instructions. | 
 |   static bool shouldExpandVectorDynExt(unsigned EltSize, unsigned NumElem, | 
 |                                        bool IsDivergentIdx, | 
 |                                        const GCNSubtarget *Subtarget); | 
 |  | 
 |   bool shouldExpandVectorDynExt(SDNode *N) const; | 
 |  | 
 | private: | 
 |   // Analyze a combined offset from an amdgcn_buffer_ intrinsic and store the | 
 |   // three offsets (voffset, soffset and instoffset) into the SDValue[3] array | 
 |   // pointed to by Offsets. | 
 |   void setBufferOffsets(SDValue CombinedOffset, SelectionDAG &DAG, | 
 |                         SDValue *Offsets, Align Alignment = Align(4)) const; | 
 |  | 
 |   // Convert the i128 that an addrspace(8) pointer is natively represented as | 
 |   // into the v4i32 that all the buffer intrinsics expect to receive. We can't | 
 |   // add register classes for i128 on pain of the promotion logic going haywire, | 
 |   // so this slightly ugly hack is what we've got. If passed a non-pointer | 
 |   // argument (as would be seen in older buffer intrinsics), does nothing. | 
 |   SDValue bufferRsrcPtrToVector(SDValue MaybePointer, SelectionDAG &DAG) const; | 
 |  | 
 |   // Wrap a 64-bit pointer into a v4i32 (which is how all SelectionDAG code | 
 |   // represents ptr addrspace(8)) using the flags specified in the intrinsic. | 
 |   SDValue lowerPointerAsRsrcIntrin(SDNode *Op, SelectionDAG &DAG) const; | 
 |  | 
 |   // Handle 8 bit and 16 bit buffer loads | 
 |   SDValue handleByteShortBufferLoads(SelectionDAG &DAG, EVT LoadVT, SDLoc DL, | 
 |                                      ArrayRef<SDValue> Ops, MemSDNode *M) const; | 
 |  | 
 |   // Handle 8 bit and 16 bit buffer stores | 
 |   SDValue handleByteShortBufferStores(SelectionDAG &DAG, EVT VDataType, | 
 |                                       SDLoc DL, SDValue Ops[], | 
 |                                       MemSDNode *M) const; | 
 |  | 
 | public: | 
 |   SITargetLowering(const TargetMachine &tm, const GCNSubtarget &STI); | 
 |  | 
 |   const GCNSubtarget *getSubtarget() const; | 
 |  | 
 |   bool isFPExtFoldable(const SelectionDAG &DAG, unsigned Opcode, EVT DestVT, | 
 |                        EVT SrcVT) const override; | 
 |  | 
 |   bool isFPExtFoldable(const MachineInstr &MI, unsigned Opcode, LLT DestTy, | 
 |                        LLT SrcTy) const override; | 
 |  | 
 |   bool isShuffleMaskLegal(ArrayRef<int> /*Mask*/, EVT /*VT*/) const override; | 
 |  | 
 |   // While address space 7 should never make it to codegen, it still needs to | 
 |   // have a MVT to prevent some analyses that query this function from breaking, | 
 |   // so, to work around the lack of i160, map it to v5i32. | 
 |   MVT getPointerTy(const DataLayout &DL, unsigned AS) const override; | 
 |   MVT getPointerMemTy(const DataLayout &DL, unsigned AS) const override; | 
 |  | 
 |   bool getTgtMemIntrinsic(IntrinsicInfo &, const CallInst &, | 
 |                           MachineFunction &MF, | 
 |                           unsigned IntrinsicID) const override; | 
 |  | 
 |   bool getAddrModeArguments(IntrinsicInst * /*I*/, | 
 |                             SmallVectorImpl<Value*> &/*Ops*/, | 
 |                             Type *&/*AccessTy*/) const override; | 
 |  | 
 |   bool isLegalGlobalAddressingMode(const AddrMode &AM) const; | 
 |   bool isLegalAddressingMode(const DataLayout &DL, const AddrMode &AM, Type *Ty, | 
 |                              unsigned AS, | 
 |                              Instruction *I = nullptr) const override; | 
 |  | 
 |   bool canMergeStoresTo(unsigned AS, EVT MemVT, | 
 |                         const MachineFunction &MF) const override; | 
 |  | 
 |   bool allowsMisalignedMemoryAccessesImpl( | 
 |       unsigned Size, unsigned AddrSpace, Align Alignment, | 
 |       MachineMemOperand::Flags Flags = MachineMemOperand::MONone, | 
 |       unsigned *IsFast = nullptr) const; | 
 |  | 
 |   bool allowsMisalignedMemoryAccesses( | 
 |       LLT Ty, unsigned AddrSpace, Align Alignment, | 
 |       MachineMemOperand::Flags Flags = MachineMemOperand::MONone, | 
 |       unsigned *IsFast = nullptr) const override { | 
 |     if (IsFast) | 
 |       *IsFast = 0; | 
 |     return allowsMisalignedMemoryAccessesImpl(Ty.getSizeInBits(), AddrSpace, | 
 |                                               Alignment, Flags, IsFast); | 
 |   } | 
 |  | 
 |   bool allowsMisalignedMemoryAccesses( | 
 |       EVT VT, unsigned AS, Align Alignment, | 
 |       MachineMemOperand::Flags Flags = MachineMemOperand::MONone, | 
 |       unsigned *IsFast = nullptr) const override; | 
 |  | 
 |   EVT getOptimalMemOpType(const MemOp &Op, | 
 |                           const AttributeList &FuncAttributes) const override; | 
 |  | 
 |   bool isMemOpUniform(const SDNode *N) const; | 
 |   bool isMemOpHasNoClobberedMemOperand(const SDNode *N) const; | 
 |  | 
 |   static bool isNonGlobalAddrSpace(unsigned AS); | 
 |  | 
 |   bool isFreeAddrSpaceCast(unsigned SrcAS, unsigned DestAS) const override; | 
 |  | 
 |   TargetLoweringBase::LegalizeTypeAction | 
 |   getPreferredVectorAction(MVT VT) const override; | 
 |  | 
 |   bool shouldConvertConstantLoadToIntImm(const APInt &Imm, | 
 |                                         Type *Ty) const override; | 
 |  | 
 |   bool isExtractSubvectorCheap(EVT ResVT, EVT SrcVT, | 
 |                                unsigned Index) const override; | 
 |  | 
 |   bool isTypeDesirableForOp(unsigned Op, EVT VT) const override; | 
 |  | 
 |   bool isOffsetFoldingLegal(const GlobalAddressSDNode *GA) const override; | 
 |  | 
 |   unsigned combineRepeatedFPDivisors() const override { | 
 |     // Combine multiple FDIVs with the same divisor into multiple FMULs by the | 
 |     // reciprocal. | 
 |     return 2; | 
 |   } | 
 |  | 
 |   bool supportSplitCSR(MachineFunction *MF) const override; | 
 |   void initializeSplitCSR(MachineBasicBlock *Entry) const override; | 
 |   void insertCopiesSplitCSR( | 
 |     MachineBasicBlock *Entry, | 
 |     const SmallVectorImpl<MachineBasicBlock *> &Exits) const override; | 
 |  | 
 |   SDValue LowerFormalArguments(SDValue Chain, CallingConv::ID CallConv, | 
 |                                bool isVarArg, | 
 |                                const SmallVectorImpl<ISD::InputArg> &Ins, | 
 |                                const SDLoc &DL, SelectionDAG &DAG, | 
 |                                SmallVectorImpl<SDValue> &InVals) const override; | 
 |  | 
 |   bool CanLowerReturn(CallingConv::ID CallConv, | 
 |                       MachineFunction &MF, bool isVarArg, | 
 |                       const SmallVectorImpl<ISD::OutputArg> &Outs, | 
 |                       LLVMContext &Context) const override; | 
 |  | 
 |   SDValue LowerReturn(SDValue Chain, CallingConv::ID CallConv, bool IsVarArg, | 
 |                       const SmallVectorImpl<ISD::OutputArg> &Outs, | 
 |                       const SmallVectorImpl<SDValue> &OutVals, const SDLoc &DL, | 
 |                       SelectionDAG &DAG) const override; | 
 |  | 
 |   void passSpecialInputs( | 
 |     CallLoweringInfo &CLI, | 
 |     CCState &CCInfo, | 
 |     const SIMachineFunctionInfo &Info, | 
 |     SmallVectorImpl<std::pair<unsigned, SDValue>> &RegsToPass, | 
 |     SmallVectorImpl<SDValue> &MemOpChains, | 
 |     SDValue Chain) const; | 
 |  | 
 |   SDValue LowerCallResult(SDValue Chain, SDValue InGlue, | 
 |                           CallingConv::ID CallConv, bool isVarArg, | 
 |                           const SmallVectorImpl<ISD::InputArg> &Ins, | 
 |                           const SDLoc &DL, SelectionDAG &DAG, | 
 |                           SmallVectorImpl<SDValue> &InVals, bool isThisReturn, | 
 |                           SDValue ThisVal) const; | 
 |  | 
 |   bool mayBeEmittedAsTailCall(const CallInst *) const override; | 
 |  | 
 |   bool isEligibleForTailCallOptimization( | 
 |     SDValue Callee, CallingConv::ID CalleeCC, bool isVarArg, | 
 |     const SmallVectorImpl<ISD::OutputArg> &Outs, | 
 |     const SmallVectorImpl<SDValue> &OutVals, | 
 |     const SmallVectorImpl<ISD::InputArg> &Ins, SelectionDAG &DAG) const; | 
 |  | 
 |   SDValue LowerCall(CallLoweringInfo &CLI, | 
 |                     SmallVectorImpl<SDValue> &InVals) const override; | 
 |  | 
 |   SDValue lowerDYNAMIC_STACKALLOCImpl(SDValue Op, SelectionDAG &DAG) const; | 
 |   SDValue LowerDYNAMIC_STACKALLOC(SDValue Op, SelectionDAG &DAG) const; | 
 |  | 
 |   Register getRegisterByName(const char* RegName, LLT VT, | 
 |                              const MachineFunction &MF) const override; | 
 |  | 
 |   MachineBasicBlock *splitKillBlock(MachineInstr &MI, | 
 |                                     MachineBasicBlock *BB) const; | 
 |  | 
 |   void bundleInstWithWaitcnt(MachineInstr &MI) const; | 
 |   MachineBasicBlock *emitGWSMemViolTestLoop(MachineInstr &MI, | 
 |                                             MachineBasicBlock *BB) const; | 
 |  | 
 |   MachineBasicBlock * | 
 |   EmitInstrWithCustomInserter(MachineInstr &MI, | 
 |                               MachineBasicBlock *BB) const override; | 
 |  | 
 |   bool hasAtomicFaddRtnForTy(SDValue &Op) const; | 
 |   bool enableAggressiveFMAFusion(EVT VT) const override; | 
 |   bool enableAggressiveFMAFusion(LLT Ty) const override; | 
 |   EVT getSetCCResultType(const DataLayout &DL, LLVMContext &Context, | 
 |                          EVT VT) const override; | 
 |   MVT getScalarShiftAmountTy(const DataLayout &, EVT) const override; | 
 |   LLT getPreferredShiftAmountTy(LLT Ty) const override; | 
 |  | 
 |   bool isFMAFasterThanFMulAndFAdd(const MachineFunction &MF, | 
 |                                   EVT VT) const override; | 
 |   bool isFMAFasterThanFMulAndFAdd(const MachineFunction &MF, | 
 |                                   const LLT Ty) const override; | 
 |   bool isFMADLegal(const SelectionDAG &DAG, const SDNode *N) const override; | 
 |   bool isFMADLegal(const MachineInstr &MI, const LLT Ty) const override; | 
 |  | 
 |   SDValue splitUnaryVectorOp(SDValue Op, SelectionDAG &DAG) const; | 
 |   SDValue splitBinaryVectorOp(SDValue Op, SelectionDAG &DAG) const; | 
 |   SDValue splitTernaryVectorOp(SDValue Op, SelectionDAG &DAG) const; | 
 |   SDValue LowerOperation(SDValue Op, SelectionDAG &DAG) const override; | 
 |  | 
 |   void ReplaceNodeResults(SDNode *N, SmallVectorImpl<SDValue> &Results, | 
 |                           SelectionDAG &DAG) const override; | 
 |  | 
 |   SDValue PerformDAGCombine(SDNode *N, DAGCombinerInfo &DCI) const override; | 
 |   SDNode *PostISelFolding(MachineSDNode *N, SelectionDAG &DAG) const override; | 
 |   void AddIMGInit(MachineInstr &MI) const; | 
 |   void AdjustInstrPostInstrSelection(MachineInstr &MI, | 
 |                                      SDNode *Node) const override; | 
 |  | 
 |   SDNode *legalizeTargetIndependentNode(SDNode *Node, SelectionDAG &DAG) const; | 
 |  | 
 |   MachineSDNode *wrapAddr64Rsrc(SelectionDAG &DAG, const SDLoc &DL, | 
 |                                 SDValue Ptr) const; | 
 |   MachineSDNode *buildRSRC(SelectionDAG &DAG, const SDLoc &DL, SDValue Ptr, | 
 |                            uint32_t RsrcDword1, uint64_t RsrcDword2And3) const; | 
 |   std::pair<unsigned, const TargetRegisterClass *> | 
 |   getRegForInlineAsmConstraint(const TargetRegisterInfo *TRI, | 
 |                                StringRef Constraint, MVT VT) const override; | 
 |   ConstraintType getConstraintType(StringRef Constraint) const override; | 
 |   void LowerAsmOperandForConstraint(SDValue Op, | 
 |                                     std::string &Constraint, | 
 |                                     std::vector<SDValue> &Ops, | 
 |                                     SelectionDAG &DAG) const override; | 
 |   bool getAsmOperandConstVal(SDValue Op, uint64_t &Val) const; | 
 |   bool checkAsmConstraintVal(SDValue Op, | 
 |                              const std::string &Constraint, | 
 |                              uint64_t Val) const; | 
 |   bool checkAsmConstraintValA(SDValue Op, | 
 |                               uint64_t Val, | 
 |                               unsigned MaxSize = 64) const; | 
 |   SDValue copyToM0(SelectionDAG &DAG, SDValue Chain, const SDLoc &DL, | 
 |                    SDValue V) const; | 
 |  | 
 |   void finalizeLowering(MachineFunction &MF) const override; | 
 |  | 
 |   void computeKnownBitsForTargetNode(const SDValue Op, KnownBits &Known, | 
 |                                      const APInt &DemandedElts, | 
 |                                      const SelectionDAG &DAG, | 
 |                                      unsigned Depth = 0) const override; | 
 |   void computeKnownBitsForFrameIndex(int FrameIdx, | 
 |                                      KnownBits &Known, | 
 |                                      const MachineFunction &MF) const override; | 
 |   void computeKnownBitsForTargetInstr(GISelKnownBits &Analysis, Register R, | 
 |                                       KnownBits &Known, | 
 |                                       const APInt &DemandedElts, | 
 |                                       const MachineRegisterInfo &MRI, | 
 |                                       unsigned Depth = 0) const override; | 
 |  | 
 |   Align computeKnownAlignForTargetInstr(GISelKnownBits &Analysis, Register R, | 
 |                                         const MachineRegisterInfo &MRI, | 
 |                                         unsigned Depth = 0) const override; | 
 |   bool isSDNodeSourceOfDivergence(const SDNode *N, FunctionLoweringInfo *FLI, | 
 |                                   UniformityInfo *UA) const override; | 
 |  | 
 |   bool hasMemSDNodeUser(SDNode *N) const; | 
 |  | 
 |   bool isReassocProfitable(SelectionDAG &DAG, SDValue N0, | 
 |                            SDValue N1) const override; | 
 |  | 
 |   bool isReassocProfitable(MachineRegisterInfo &MRI, Register N0, | 
 |                            Register N1) const override; | 
 |  | 
 |   bool isCanonicalized(SelectionDAG &DAG, SDValue Op, | 
 |                        unsigned MaxDepth = 5) const; | 
 |   bool isCanonicalized(Register Reg, MachineFunction &MF, | 
 |                        unsigned MaxDepth = 5) const; | 
 |   bool denormalsEnabledForType(const SelectionDAG &DAG, EVT VT) const; | 
 |   bool denormalsEnabledForType(LLT Ty, MachineFunction &MF) const; | 
 |  | 
 |   bool checkForPhysRegDependency(SDNode *Def, SDNode *User, unsigned Op, | 
 |                                  const TargetRegisterInfo *TRI, | 
 |                                  const TargetInstrInfo *TII, unsigned &PhysReg, | 
 |                                  int &Cost) const override; | 
 |  | 
 |   bool isKnownNeverNaNForTargetNode(SDValue Op, | 
 |                                     const SelectionDAG &DAG, | 
 |                                     bool SNaN = false, | 
 |                                     unsigned Depth = 0) const override; | 
 |   AtomicExpansionKind shouldExpandAtomicRMWInIR(AtomicRMWInst *) const override; | 
 |   AtomicExpansionKind shouldExpandAtomicLoadInIR(LoadInst *LI) const override; | 
 |   AtomicExpansionKind shouldExpandAtomicStoreInIR(StoreInst *SI) const override; | 
 |   AtomicExpansionKind | 
 |   shouldExpandAtomicCmpXchgInIR(AtomicCmpXchgInst *AI) const override; | 
 |   void emitExpandAtomicRMW(AtomicRMWInst *AI) const override; | 
 |  | 
 |   LoadInst * | 
 |   lowerIdempotentRMWIntoFencedLoad(AtomicRMWInst *AI) const override; | 
 |  | 
 |   const TargetRegisterClass *getRegClassFor(MVT VT, | 
 |                                             bool isDivergent) const override; | 
 |   bool requiresUniformRegister(MachineFunction &MF, | 
 |                                const Value *V) const override; | 
 |   Align getPrefLoopAlignment(MachineLoop *ML) const override; | 
 |  | 
 |   void allocateHSAUserSGPRs(CCState &CCInfo, | 
 |                             MachineFunction &MF, | 
 |                             const SIRegisterInfo &TRI, | 
 |                             SIMachineFunctionInfo &Info) const; | 
 |  | 
 |   void allocateSystemSGPRs(CCState &CCInfo, | 
 |                            MachineFunction &MF, | 
 |                            SIMachineFunctionInfo &Info, | 
 |                            CallingConv::ID CallConv, | 
 |                            bool IsShader) const; | 
 |  | 
 |   void allocateSpecialEntryInputVGPRs(CCState &CCInfo, | 
 |                                       MachineFunction &MF, | 
 |                                       const SIRegisterInfo &TRI, | 
 |                                       SIMachineFunctionInfo &Info) const; | 
 |   void allocateSpecialInputSGPRs( | 
 |     CCState &CCInfo, | 
 |     MachineFunction &MF, | 
 |     const SIRegisterInfo &TRI, | 
 |     SIMachineFunctionInfo &Info) const; | 
 |  | 
 |   void allocateSpecialInputVGPRs(CCState &CCInfo, | 
 |                                  MachineFunction &MF, | 
 |                                  const SIRegisterInfo &TRI, | 
 |                                  SIMachineFunctionInfo &Info) const; | 
 |   void allocateSpecialInputVGPRsFixed(CCState &CCInfo, | 
 |                                       MachineFunction &MF, | 
 |                                       const SIRegisterInfo &TRI, | 
 |                                       SIMachineFunctionInfo &Info) const; | 
 |  | 
 |   MachineMemOperand::Flags | 
 |   getTargetMMOFlags(const Instruction &I) const override; | 
 | }; | 
 |  | 
 | } // End namespace llvm | 
 |  | 
 | #endif |