|  | //===-- PPCInstrInfo.h - PowerPC Instruction Information --------*- C++ -*-===// | 
|  | // | 
|  | // Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. | 
|  | // See https://llvm.org/LICENSE.txt for license information. | 
|  | // SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception | 
|  | // | 
|  | //===----------------------------------------------------------------------===// | 
|  | // | 
|  | // This file contains the PowerPC implementation of the TargetInstrInfo class. | 
|  | // | 
|  | //===----------------------------------------------------------------------===// | 
|  |  | 
|  | #ifndef LLVM_LIB_TARGET_POWERPC_PPCINSTRINFO_H | 
|  | #define LLVM_LIB_TARGET_POWERPC_PPCINSTRINFO_H | 
|  |  | 
|  | #include "PPCRegisterInfo.h" | 
|  | #include "llvm/CodeGen/TargetInstrInfo.h" | 
|  |  | 
|  | #define GET_INSTRINFO_HEADER | 
|  | #include "PPCGenInstrInfo.inc" | 
|  |  | 
|  | namespace llvm { | 
|  |  | 
|  | /// PPCII - This namespace holds all of the PowerPC target-specific | 
|  | /// per-instruction flags.  These must match the corresponding definitions in | 
|  | /// PPC.td and PPCInstrFormats.td. | 
|  | namespace PPCII { | 
|  | enum { | 
|  | // PPC970 Instruction Flags.  These flags describe the characteristics of the | 
|  | // PowerPC 970 (aka G5) dispatch groups and how they are formed out of | 
|  | // raw machine instructions. | 
|  |  | 
|  | /// PPC970_First - This instruction starts a new dispatch group, so it will | 
|  | /// always be the first one in the group. | 
|  | PPC970_First = 0x1, | 
|  |  | 
|  | /// PPC970_Single - This instruction starts a new dispatch group and | 
|  | /// terminates it, so it will be the sole instruction in the group. | 
|  | PPC970_Single = 0x2, | 
|  |  | 
|  | /// PPC970_Cracked - This instruction is cracked into two pieces, requiring | 
|  | /// two dispatch pipes to be available to issue. | 
|  | PPC970_Cracked = 0x4, | 
|  |  | 
|  | /// PPC970_Mask/Shift - This is a bitmask that selects the pipeline type that | 
|  | /// an instruction is issued to. | 
|  | PPC970_Shift = 3, | 
|  | PPC970_Mask = 0x07 << PPC970_Shift | 
|  | }; | 
|  | enum PPC970_Unit { | 
|  | /// These are the various PPC970 execution unit pipelines.  Each instruction | 
|  | /// is one of these. | 
|  | PPC970_Pseudo = 0 << PPC970_Shift,   // Pseudo instruction | 
|  | PPC970_FXU    = 1 << PPC970_Shift,   // Fixed Point (aka Integer/ALU) Unit | 
|  | PPC970_LSU    = 2 << PPC970_Shift,   // Load Store Unit | 
|  | PPC970_FPU    = 3 << PPC970_Shift,   // Floating Point Unit | 
|  | PPC970_CRU    = 4 << PPC970_Shift,   // Control Register Unit | 
|  | PPC970_VALU   = 5 << PPC970_Shift,   // Vector ALU | 
|  | PPC970_VPERM  = 6 << PPC970_Shift,   // Vector Permute Unit | 
|  | PPC970_BRU    = 7 << PPC970_Shift    // Branch Unit | 
|  | }; | 
|  |  | 
|  | enum { | 
|  | /// Shift count to bypass PPC970 flags | 
|  | NewDef_Shift = 6, | 
|  |  | 
|  | /// This instruction is an X-Form memory operation. | 
|  | XFormMemOp = 0x1 << NewDef_Shift, | 
|  | /// This instruction is prefixed. | 
|  | Prefixed = 0x1 << (NewDef_Shift+1) | 
|  | }; | 
|  | } // end namespace PPCII | 
|  |  | 
|  | // Instructions that have an immediate form might be convertible to that | 
|  | // form if the correct input is a result of a load immediate. In order to | 
|  | // know whether the transformation is special, we might need to know some | 
|  | // of the details of the two forms. | 
|  | struct ImmInstrInfo { | 
|  | // Is the immediate field in the immediate form signed or unsigned? | 
|  | uint64_t SignedImm : 1; | 
|  | // Does the immediate need to be a multiple of some value? | 
|  | uint64_t ImmMustBeMultipleOf : 5; | 
|  | // Is R0/X0 treated specially by the original r+r instruction? | 
|  | // If so, in which operand? | 
|  | uint64_t ZeroIsSpecialOrig : 3; | 
|  | // Is R0/X0 treated specially by the new r+i instruction? | 
|  | // If so, in which operand? | 
|  | uint64_t ZeroIsSpecialNew : 3; | 
|  | // Is the operation commutative? | 
|  | uint64_t IsCommutative : 1; | 
|  | // The operand number to check for add-immediate def. | 
|  | uint64_t OpNoForForwarding : 3; | 
|  | // The operand number for the immediate. | 
|  | uint64_t ImmOpNo : 3; | 
|  | // The opcode of the new instruction. | 
|  | uint64_t ImmOpcode : 16; | 
|  | // The size of the immediate. | 
|  | uint64_t ImmWidth : 5; | 
|  | // The immediate should be truncated to N bits. | 
|  | uint64_t TruncateImmTo : 5; | 
|  | // Is the instruction summing the operand | 
|  | uint64_t IsSummingOperands : 1; | 
|  | }; | 
|  |  | 
|  | // Information required to convert an instruction to just a materialized | 
|  | // immediate. | 
|  | struct LoadImmediateInfo { | 
|  | unsigned Imm : 16; | 
|  | unsigned Is64Bit : 1; | 
|  | unsigned SetCR : 1; | 
|  | }; | 
|  |  | 
|  | // Index into the OpcodesForSpill array. | 
|  | enum SpillOpcodeKey { | 
|  | SOK_Int4Spill, | 
|  | SOK_Int8Spill, | 
|  | SOK_Float8Spill, | 
|  | SOK_Float4Spill, | 
|  | SOK_CRSpill, | 
|  | SOK_CRBitSpill, | 
|  | SOK_VRVectorSpill, | 
|  | SOK_VSXVectorSpill, | 
|  | SOK_VectorFloat8Spill, | 
|  | SOK_VectorFloat4Spill, | 
|  | SOK_SpillToVSR, | 
|  | SOK_PairedVecSpill, | 
|  | SOK_AccumulatorSpill, | 
|  | SOK_UAccumulatorSpill, | 
|  | SOK_SPESpill, | 
|  | SOK_PairedG8Spill, | 
|  | SOK_LastOpcodeSpill // This must be last on the enum. | 
|  | }; | 
|  |  | 
|  | // Define list of load and store spill opcodes. | 
|  | #define NoInstr PPC::INSTRUCTION_LIST_END | 
|  | #define Pwr8LoadOpcodes                                                        \ | 
|  | {                                                                            \ | 
|  | PPC::LWZ, PPC::LD, PPC::LFD, PPC::LFS, PPC::RESTORE_CR,                    \ | 
|  | PPC::RESTORE_CRBIT, PPC::LVX, PPC::LXVD2X, PPC::LXSDX, PPC::LXSSPX,    \ | 
|  | PPC::SPILLTOVSR_LD, NoInstr, NoInstr, NoInstr, PPC::EVLDD,             \ | 
|  | PPC::RESTORE_QUADWORD                                                  \ | 
|  | } | 
|  |  | 
|  | #define Pwr9LoadOpcodes                                                        \ | 
|  | {                                                                            \ | 
|  | PPC::LWZ, PPC::LD, PPC::LFD, PPC::LFS, PPC::RESTORE_CR,                    \ | 
|  | PPC::RESTORE_CRBIT, PPC::LVX, PPC::LXV, PPC::DFLOADf64,                \ | 
|  | PPC::DFLOADf32, PPC::SPILLTOVSR_LD, NoInstr, NoInstr, NoInstr,         \ | 
|  | NoInstr, PPC::RESTORE_QUADWORD                                         \ | 
|  | } | 
|  |  | 
|  | #define Pwr10LoadOpcodes                                                       \ | 
|  | {                                                                            \ | 
|  | PPC::LWZ, PPC::LD, PPC::LFD, PPC::LFS, PPC::RESTORE_CR,                    \ | 
|  | PPC::RESTORE_CRBIT, PPC::LVX, PPC::LXV, PPC::DFLOADf64,                \ | 
|  | PPC::DFLOADf32, PPC::SPILLTOVSR_LD, PPC::LXVP, PPC::RESTORE_ACC,       \ | 
|  | PPC::RESTORE_UACC, NoInstr, PPC::RESTORE_QUADWORD                      \ | 
|  | } | 
|  |  | 
|  | #define Pwr8StoreOpcodes                                                       \ | 
|  | {                                                                            \ | 
|  | PPC::STW, PPC::STD, PPC::STFD, PPC::STFS, PPC::SPILL_CR, PPC::SPILL_CRBIT, \ | 
|  | PPC::STVX, PPC::STXVD2X, PPC::STXSDX, PPC::STXSSPX,                    \ | 
|  | PPC::SPILLTOVSR_ST, NoInstr, NoInstr, NoInstr, PPC::EVSTDD,            \ | 
|  | PPC::SPILL_QUADWORD                                                    \ | 
|  | } | 
|  |  | 
|  | #define Pwr9StoreOpcodes                                                       \ | 
|  | {                                                                            \ | 
|  | PPC::STW, PPC::STD, PPC::STFD, PPC::STFS, PPC::SPILL_CR, PPC::SPILL_CRBIT, \ | 
|  | PPC::STVX, PPC::STXV, PPC::DFSTOREf64, PPC::DFSTOREf32,                \ | 
|  | PPC::SPILLTOVSR_ST, NoInstr, NoInstr, NoInstr, NoInstr,                \ | 
|  | PPC::SPILL_QUADWORD                                                    \ | 
|  | } | 
|  |  | 
|  | #define Pwr10StoreOpcodes                                                      \ | 
|  | {                                                                            \ | 
|  | PPC::STW, PPC::STD, PPC::STFD, PPC::STFS, PPC::SPILL_CR, PPC::SPILL_CRBIT, \ | 
|  | PPC::STVX, PPC::STXV, PPC::DFSTOREf64, PPC::DFSTOREf32,                \ | 
|  | PPC::SPILLTOVSR_ST, PPC::STXVP, PPC::SPILL_ACC, PPC::SPILL_UACC,       \ | 
|  | NoInstr, PPC::SPILL_QUADWORD                                           \ | 
|  | } | 
|  |  | 
|  | // Initialize arrays for load and store spill opcodes on supported subtargets. | 
|  | #define StoreOpcodesForSpill                                                   \ | 
|  | { Pwr8StoreOpcodes, Pwr9StoreOpcodes, Pwr10StoreOpcodes } | 
|  | #define LoadOpcodesForSpill                                                    \ | 
|  | { Pwr8LoadOpcodes, Pwr9LoadOpcodes, Pwr10LoadOpcodes } | 
|  |  | 
|  | class PPCSubtarget; | 
|  | class PPCInstrInfo : public PPCGenInstrInfo { | 
|  | PPCSubtarget &Subtarget; | 
|  | const PPCRegisterInfo RI; | 
|  | const unsigned StoreSpillOpcodesArray[3][SOK_LastOpcodeSpill] = | 
|  | StoreOpcodesForSpill; | 
|  | const unsigned LoadSpillOpcodesArray[3][SOK_LastOpcodeSpill] = | 
|  | LoadOpcodesForSpill; | 
|  |  | 
|  | void StoreRegToStackSlot(MachineFunction &MF, unsigned SrcReg, bool isKill, | 
|  | int FrameIdx, const TargetRegisterClass *RC, | 
|  | SmallVectorImpl<MachineInstr *> &NewMIs) const; | 
|  | void LoadRegFromStackSlot(MachineFunction &MF, const DebugLoc &DL, | 
|  | unsigned DestReg, int FrameIdx, | 
|  | const TargetRegisterClass *RC, | 
|  | SmallVectorImpl<MachineInstr *> &NewMIs) const; | 
|  |  | 
|  | // Replace the instruction with single LI if possible. \p DefMI must be LI or | 
|  | // LI8. | 
|  | bool simplifyToLI(MachineInstr &MI, MachineInstr &DefMI, | 
|  | unsigned OpNoForForwarding, MachineInstr **KilledDef) const; | 
|  | // If the inst is imm-form and its register operand is produced by a ADDI, put | 
|  | // the imm into the inst directly and remove the ADDI if possible. | 
|  | bool transformToNewImmFormFedByAdd(MachineInstr &MI, MachineInstr &DefMI, | 
|  | unsigned OpNoForForwarding) const; | 
|  | // If the inst is x-form and has imm-form and one of its operand is produced | 
|  | // by a LI, put the imm into the inst directly and remove the LI if possible. | 
|  | bool transformToImmFormFedByLI(MachineInstr &MI, const ImmInstrInfo &III, | 
|  | unsigned ConstantOpNo, | 
|  | MachineInstr &DefMI) const; | 
|  | // If the inst is x-form and has imm-form and one of its operand is produced | 
|  | // by an add-immediate, try to transform it when possible. | 
|  | bool transformToImmFormFedByAdd(MachineInstr &MI, const ImmInstrInfo &III, | 
|  | unsigned ConstantOpNo, MachineInstr &DefMI, | 
|  | bool KillDefMI) const; | 
|  | // Try to find that, if the instruction 'MI' contains any operand that | 
|  | // could be forwarded from some inst that feeds it. If yes, return the | 
|  | // Def of that operand. And OpNoForForwarding is the operand index in | 
|  | // the 'MI' for that 'Def'. If we see another use of this Def between | 
|  | // the Def and the MI, SeenIntermediateUse becomes 'true'. | 
|  | MachineInstr *getForwardingDefMI(MachineInstr &MI, | 
|  | unsigned &OpNoForForwarding, | 
|  | bool &SeenIntermediateUse) const; | 
|  |  | 
|  | // Can the user MI have it's source at index \p OpNoForForwarding | 
|  | // forwarded from an add-immediate that feeds it? | 
|  | bool isUseMIElgibleForForwarding(MachineInstr &MI, const ImmInstrInfo &III, | 
|  | unsigned OpNoForForwarding) const; | 
|  | bool isDefMIElgibleForForwarding(MachineInstr &DefMI, | 
|  | const ImmInstrInfo &III, | 
|  | MachineOperand *&ImmMO, | 
|  | MachineOperand *&RegMO) const; | 
|  | bool isImmElgibleForForwarding(const MachineOperand &ImmMO, | 
|  | const MachineInstr &DefMI, | 
|  | const ImmInstrInfo &III, | 
|  | int64_t &Imm, | 
|  | int64_t BaseImm = 0) const; | 
|  | bool isRegElgibleForForwarding(const MachineOperand &RegMO, | 
|  | const MachineInstr &DefMI, | 
|  | const MachineInstr &MI, bool KillDefMI, | 
|  | bool &IsFwdFeederRegKilled) const; | 
|  | unsigned getSpillTarget() const; | 
|  | const unsigned *getStoreOpcodesForSpillArray() const; | 
|  | const unsigned *getLoadOpcodesForSpillArray() const; | 
|  | unsigned getSpillIndex(const TargetRegisterClass *RC) const; | 
|  | int16_t getFMAOpIdxInfo(unsigned Opcode) const; | 
|  | void reassociateFMA(MachineInstr &Root, MachineCombinerPattern Pattern, | 
|  | SmallVectorImpl<MachineInstr *> &InsInstrs, | 
|  | SmallVectorImpl<MachineInstr *> &DelInstrs, | 
|  | DenseMap<unsigned, unsigned> &InstrIdxForVirtReg) const; | 
|  | bool isLoadFromConstantPool(MachineInstr *I) const; | 
|  | Register | 
|  | generateLoadForNewConst(unsigned Idx, MachineInstr *MI, Type *Ty, | 
|  | SmallVectorImpl<MachineInstr *> &InsInstrs) const; | 
|  | const Constant *getConstantFromConstantPool(MachineInstr *I) const; | 
|  | virtual void anchor(); | 
|  |  | 
|  | protected: | 
|  | /// Commutes the operands in the given instruction. | 
|  | /// The commutable operands are specified by their indices OpIdx1 and OpIdx2. | 
|  | /// | 
|  | /// Do not call this method for a non-commutable instruction or for | 
|  | /// non-commutable pair of operand indices OpIdx1 and OpIdx2. | 
|  | /// Even though the instruction is commutable, the method may still | 
|  | /// fail to commute the operands, null pointer is returned in such cases. | 
|  | /// | 
|  | /// For example, we can commute rlwimi instructions, but only if the | 
|  | /// rotate amt is zero.  We also have to munge the immediates a bit. | 
|  | MachineInstr *commuteInstructionImpl(MachineInstr &MI, bool NewMI, | 
|  | unsigned OpIdx1, | 
|  | unsigned OpIdx2) const override; | 
|  |  | 
|  | public: | 
|  | explicit PPCInstrInfo(PPCSubtarget &STI); | 
|  |  | 
|  | /// getRegisterInfo - TargetInstrInfo is a superset of MRegister info.  As | 
|  | /// such, whenever a client has an instance of instruction info, it should | 
|  | /// always be able to get register info as well (through this method). | 
|  | /// | 
|  | const PPCRegisterInfo &getRegisterInfo() const { return RI; } | 
|  |  | 
|  | bool isXFormMemOp(unsigned Opcode) const { | 
|  | return get(Opcode).TSFlags & PPCII::XFormMemOp; | 
|  | } | 
|  | bool isPrefixed(unsigned Opcode) const { | 
|  | return get(Opcode).TSFlags & PPCII::Prefixed; | 
|  | } | 
|  |  | 
|  | /// Check if Opcode corresponds to a call instruction that should be marked | 
|  | /// with the NOTOC relocation. | 
|  | bool isNoTOCCallInstr(unsigned Opcode) const { | 
|  | if (!get(Opcode).isCall()) | 
|  | return false; | 
|  |  | 
|  | switch (Opcode) { | 
|  | default: | 
|  | #ifndef NDEBUG | 
|  | llvm_unreachable("Unknown call opcode"); | 
|  | #endif | 
|  | return false; | 
|  | case PPC::BL8_NOTOC: | 
|  | case PPC::BL8_NOTOC_TLS: | 
|  | case PPC::BL8_NOTOC_RM: | 
|  | return true; | 
|  | #ifndef NDEBUG | 
|  | case PPC::BL8: | 
|  | case PPC::BL: | 
|  | case PPC::BL8_TLS: | 
|  | case PPC::BL_TLS: | 
|  | case PPC::BLA8: | 
|  | case PPC::BLA: | 
|  | case PPC::BCCL: | 
|  | case PPC::BCCLA: | 
|  | case PPC::BCL: | 
|  | case PPC::BCLn: | 
|  | case PPC::BL8_NOP: | 
|  | case PPC::BL_NOP: | 
|  | case PPC::BL8_NOP_TLS: | 
|  | case PPC::BLA8_NOP: | 
|  | case PPC::BCTRL8: | 
|  | case PPC::BCTRL: | 
|  | case PPC::BCCCTRL8: | 
|  | case PPC::BCCCTRL: | 
|  | case PPC::BCCTRL8: | 
|  | case PPC::BCCTRL: | 
|  | case PPC::BCCTRL8n: | 
|  | case PPC::BCCTRLn: | 
|  | case PPC::BL8_RM: | 
|  | case PPC::BLA8_RM: | 
|  | case PPC::BL8_NOP_RM: | 
|  | case PPC::BLA8_NOP_RM: | 
|  | case PPC::BCTRL8_RM: | 
|  | case PPC::BCTRL8_LDinto_toc: | 
|  | case PPC::BCTRL8_LDinto_toc_RM: | 
|  | case PPC::BL8_TLS_: | 
|  | case PPC::TCRETURNdi8: | 
|  | case PPC::TCRETURNai8: | 
|  | case PPC::TCRETURNri8: | 
|  | case PPC::TAILBCTR8: | 
|  | case PPC::TAILB8: | 
|  | case PPC::TAILBA8: | 
|  | case PPC::BCLalways: | 
|  | case PPC::BLRL: | 
|  | case PPC::BCCLRL: | 
|  | case PPC::BCLRL: | 
|  | case PPC::BCLRLn: | 
|  | case PPC::BDZL: | 
|  | case PPC::BDNZL: | 
|  | case PPC::BDZLA: | 
|  | case PPC::BDNZLA: | 
|  | case PPC::BDZLp: | 
|  | case PPC::BDNZLp: | 
|  | case PPC::BDZLAp: | 
|  | case PPC::BDNZLAp: | 
|  | case PPC::BDZLm: | 
|  | case PPC::BDNZLm: | 
|  | case PPC::BDZLAm: | 
|  | case PPC::BDNZLAm: | 
|  | case PPC::BDZLRL: | 
|  | case PPC::BDNZLRL: | 
|  | case PPC::BDZLRLp: | 
|  | case PPC::BDNZLRLp: | 
|  | case PPC::BDZLRLm: | 
|  | case PPC::BDNZLRLm: | 
|  | case PPC::BL_RM: | 
|  | case PPC::BLA_RM: | 
|  | case PPC::BL_NOP_RM: | 
|  | case PPC::BCTRL_RM: | 
|  | case PPC::TCRETURNdi: | 
|  | case PPC::TCRETURNai: | 
|  | case PPC::TCRETURNri: | 
|  | case PPC::BCTRL_LWZinto_toc: | 
|  | case PPC::BCTRL_LWZinto_toc_RM: | 
|  | case PPC::TAILBCTR: | 
|  | case PPC::TAILB: | 
|  | case PPC::TAILBA: | 
|  | return false; | 
|  | #endif | 
|  | } | 
|  | } | 
|  |  | 
|  | static bool isSameClassPhysRegCopy(unsigned Opcode) { | 
|  | unsigned CopyOpcodes[] = {PPC::OR,        PPC::OR8,   PPC::FMR, | 
|  | PPC::VOR,       PPC::XXLOR, PPC::XXLORf, | 
|  | PPC::XSCPSGNDP, PPC::MCRF,  PPC::CROR, | 
|  | PPC::EVOR,      -1U}; | 
|  | for (int i = 0; CopyOpcodes[i] != -1U; i++) | 
|  | if (Opcode == CopyOpcodes[i]) | 
|  | return true; | 
|  | return false; | 
|  | } | 
|  |  | 
|  | ScheduleHazardRecognizer * | 
|  | CreateTargetHazardRecognizer(const TargetSubtargetInfo *STI, | 
|  | const ScheduleDAG *DAG) const override; | 
|  | ScheduleHazardRecognizer * | 
|  | CreateTargetPostRAHazardRecognizer(const InstrItineraryData *II, | 
|  | const ScheduleDAG *DAG) const override; | 
|  |  | 
|  | unsigned getInstrLatency(const InstrItineraryData *ItinData, | 
|  | const MachineInstr &MI, | 
|  | unsigned *PredCost = nullptr) const override; | 
|  |  | 
|  | int getOperandLatency(const InstrItineraryData *ItinData, | 
|  | const MachineInstr &DefMI, unsigned DefIdx, | 
|  | const MachineInstr &UseMI, | 
|  | unsigned UseIdx) const override; | 
|  | int getOperandLatency(const InstrItineraryData *ItinData, | 
|  | SDNode *DefNode, unsigned DefIdx, | 
|  | SDNode *UseNode, unsigned UseIdx) const override { | 
|  | return PPCGenInstrInfo::getOperandLatency(ItinData, DefNode, DefIdx, | 
|  | UseNode, UseIdx); | 
|  | } | 
|  |  | 
|  | bool hasLowDefLatency(const TargetSchedModel &SchedModel, | 
|  | const MachineInstr &DefMI, | 
|  | unsigned DefIdx) const override { | 
|  | // Machine LICM should hoist all instructions in low-register-pressure | 
|  | // situations; none are sufficiently free to justify leaving in a loop | 
|  | // body. | 
|  | return false; | 
|  | } | 
|  |  | 
|  | bool useMachineCombiner() const override { | 
|  | return true; | 
|  | } | 
|  |  | 
|  | /// When getMachineCombinerPatterns() finds patterns, this function generates | 
|  | /// the instructions that could replace the original code sequence | 
|  | void genAlternativeCodeSequence( | 
|  | MachineInstr &Root, MachineCombinerPattern Pattern, | 
|  | SmallVectorImpl<MachineInstr *> &InsInstrs, | 
|  | SmallVectorImpl<MachineInstr *> &DelInstrs, | 
|  | DenseMap<unsigned, unsigned> &InstrIdxForVirtReg) const override; | 
|  |  | 
|  | /// Return true when there is potentially a faster code sequence for a fma | 
|  | /// chain ending in \p Root. All potential patterns are output in the \p | 
|  | /// P array. | 
|  | bool getFMAPatterns(MachineInstr &Root, | 
|  | SmallVectorImpl<MachineCombinerPattern> &P, | 
|  | bool DoRegPressureReduce) const; | 
|  |  | 
|  | /// Return true when there is potentially a faster code sequence | 
|  | /// for an instruction chain ending in <Root>. All potential patterns are | 
|  | /// output in the <Pattern> array. | 
|  | bool getMachineCombinerPatterns(MachineInstr &Root, | 
|  | SmallVectorImpl<MachineCombinerPattern> &P, | 
|  | bool DoRegPressureReduce) const override; | 
|  |  | 
|  | /// On PowerPC, we leverage machine combiner pass to reduce register pressure | 
|  | /// when the register pressure is high for one BB. | 
|  | /// Return true if register pressure for \p MBB is high and ABI is supported | 
|  | /// to reduce register pressure. Otherwise return false. | 
|  | bool | 
|  | shouldReduceRegisterPressure(MachineBasicBlock *MBB, | 
|  | RegisterClassInfo *RegClassInfo) const override; | 
|  |  | 
|  | /// Fixup the placeholders we put in genAlternativeCodeSequence() for | 
|  | /// MachineCombiner. | 
|  | void | 
|  | finalizeInsInstrs(MachineInstr &Root, MachineCombinerPattern &P, | 
|  | SmallVectorImpl<MachineInstr *> &InsInstrs) const override; | 
|  |  | 
|  | bool isAssociativeAndCommutative(const MachineInstr &Inst) const override; | 
|  |  | 
|  | /// On PowerPC, we try to reassociate FMA chain which will increase | 
|  | /// instruction size. Set extension resource length limit to 1 for edge case. | 
|  | /// Resource Length is calculated by scaled resource usage in getCycles(). | 
|  | /// Because of the division in getCycles(), it returns different cycles due to | 
|  | /// legacy scaled resource usage. So new resource length may be same with | 
|  | /// legacy or 1 bigger than legacy. | 
|  | /// We need to execlude the 1 bigger case even the resource length is not | 
|  | /// perserved for more FMA chain reassociations on PowerPC. | 
|  | int getExtendResourceLenLimit() const override { return 1; } | 
|  |  | 
|  | void setSpecialOperandAttr(MachineInstr &OldMI1, MachineInstr &OldMI2, | 
|  | MachineInstr &NewMI1, | 
|  | MachineInstr &NewMI2) const override; | 
|  |  | 
|  | // PowerPC specific version of setSpecialOperandAttr that copies Flags to MI | 
|  | // and clears nuw, nsw, and exact flags. | 
|  | void setSpecialOperandAttr(MachineInstr &MI, uint16_t Flags) const; | 
|  |  | 
|  | bool isCoalescableExtInstr(const MachineInstr &MI, | 
|  | Register &SrcReg, Register &DstReg, | 
|  | unsigned &SubIdx) const override; | 
|  | unsigned isLoadFromStackSlot(const MachineInstr &MI, | 
|  | int &FrameIndex) const override; | 
|  | bool isReallyTriviallyReMaterializable(const MachineInstr &MI) const override; | 
|  | unsigned isStoreToStackSlot(const MachineInstr &MI, | 
|  | int &FrameIndex) const override; | 
|  |  | 
|  | bool findCommutedOpIndices(const MachineInstr &MI, unsigned &SrcOpIdx1, | 
|  | unsigned &SrcOpIdx2) const override; | 
|  |  | 
|  | void insertNoop(MachineBasicBlock &MBB, | 
|  | MachineBasicBlock::iterator MI) const override; | 
|  |  | 
|  |  | 
|  | // Branch analysis. | 
|  | bool analyzeBranch(MachineBasicBlock &MBB, MachineBasicBlock *&TBB, | 
|  | MachineBasicBlock *&FBB, | 
|  | SmallVectorImpl<MachineOperand> &Cond, | 
|  | bool AllowModify) const override; | 
|  | unsigned removeBranch(MachineBasicBlock &MBB, | 
|  | int *BytesRemoved = nullptr) const override; | 
|  | unsigned insertBranch(MachineBasicBlock &MBB, MachineBasicBlock *TBB, | 
|  | MachineBasicBlock *FBB, ArrayRef<MachineOperand> Cond, | 
|  | const DebugLoc &DL, | 
|  | int *BytesAdded = nullptr) const override; | 
|  |  | 
|  | // Select analysis. | 
|  | bool canInsertSelect(const MachineBasicBlock &, ArrayRef<MachineOperand> Cond, | 
|  | Register, Register, Register, int &, int &, | 
|  | int &) const override; | 
|  | void insertSelect(MachineBasicBlock &MBB, MachineBasicBlock::iterator MI, | 
|  | const DebugLoc &DL, Register DstReg, | 
|  | ArrayRef<MachineOperand> Cond, Register TrueReg, | 
|  | Register FalseReg) const override; | 
|  |  | 
|  | void copyPhysReg(MachineBasicBlock &MBB, MachineBasicBlock::iterator I, | 
|  | const DebugLoc &DL, MCRegister DestReg, MCRegister SrcReg, | 
|  | bool KillSrc) const override; | 
|  |  | 
|  | void storeRegToStackSlot(MachineBasicBlock &MBB, | 
|  | MachineBasicBlock::iterator MBBI, | 
|  | Register SrcReg, bool isKill, int FrameIndex, | 
|  | const TargetRegisterClass *RC, | 
|  | const TargetRegisterInfo *TRI) const override; | 
|  |  | 
|  | // Emits a register spill without updating the register class for vector | 
|  | // registers. This ensures that when we spill a vector register the | 
|  | // element order in the register is the same as it was in memory. | 
|  | void storeRegToStackSlotNoUpd(MachineBasicBlock &MBB, | 
|  | MachineBasicBlock::iterator MBBI, | 
|  | unsigned SrcReg, bool isKill, int FrameIndex, | 
|  | const TargetRegisterClass *RC, | 
|  | const TargetRegisterInfo *TRI) const; | 
|  |  | 
|  | void loadRegFromStackSlot(MachineBasicBlock &MBB, | 
|  | MachineBasicBlock::iterator MBBI, | 
|  | Register DestReg, int FrameIndex, | 
|  | const TargetRegisterClass *RC, | 
|  | const TargetRegisterInfo *TRI) const override; | 
|  |  | 
|  | // Emits a register reload without updating the register class for vector | 
|  | // registers. This ensures that when we reload a vector register the | 
|  | // element order in the register is the same as it was in memory. | 
|  | void loadRegFromStackSlotNoUpd(MachineBasicBlock &MBB, | 
|  | MachineBasicBlock::iterator MBBI, | 
|  | unsigned DestReg, int FrameIndex, | 
|  | const TargetRegisterClass *RC, | 
|  | const TargetRegisterInfo *TRI) const; | 
|  |  | 
|  | unsigned getStoreOpcodeForSpill(const TargetRegisterClass *RC) const; | 
|  |  | 
|  | unsigned getLoadOpcodeForSpill(const TargetRegisterClass *RC) const; | 
|  |  | 
|  | bool | 
|  | reverseBranchCondition(SmallVectorImpl<MachineOperand> &Cond) const override; | 
|  |  | 
|  | bool FoldImmediate(MachineInstr &UseMI, MachineInstr &DefMI, Register Reg, | 
|  | MachineRegisterInfo *MRI) const override; | 
|  |  | 
|  | bool onlyFoldImmediate(MachineInstr &UseMI, MachineInstr &DefMI, | 
|  | Register Reg) const; | 
|  |  | 
|  | // If conversion by predication (only supported by some branch instructions). | 
|  | // All of the profitability checks always return true; it is always | 
|  | // profitable to use the predicated branches. | 
|  | bool isProfitableToIfCvt(MachineBasicBlock &MBB, | 
|  | unsigned NumCycles, unsigned ExtraPredCycles, | 
|  | BranchProbability Probability) const override { | 
|  | return true; | 
|  | } | 
|  |  | 
|  | bool isProfitableToIfCvt(MachineBasicBlock &TMBB, | 
|  | unsigned NumT, unsigned ExtraT, | 
|  | MachineBasicBlock &FMBB, | 
|  | unsigned NumF, unsigned ExtraF, | 
|  | BranchProbability Probability) const override; | 
|  |  | 
|  | bool isProfitableToDupForIfCvt(MachineBasicBlock &MBB, unsigned NumCycles, | 
|  | BranchProbability Probability) const override { | 
|  | return true; | 
|  | } | 
|  |  | 
|  | bool isProfitableToUnpredicate(MachineBasicBlock &TMBB, | 
|  | MachineBasicBlock &FMBB) const override { | 
|  | return false; | 
|  | } | 
|  |  | 
|  | // Predication support. | 
|  | bool isPredicated(const MachineInstr &MI) const override; | 
|  |  | 
|  | bool isSchedulingBoundary(const MachineInstr &MI, | 
|  | const MachineBasicBlock *MBB, | 
|  | const MachineFunction &MF) const override; | 
|  |  | 
|  | bool PredicateInstruction(MachineInstr &MI, | 
|  | ArrayRef<MachineOperand> Pred) const override; | 
|  |  | 
|  | bool SubsumesPredicate(ArrayRef<MachineOperand> Pred1, | 
|  | ArrayRef<MachineOperand> Pred2) const override; | 
|  |  | 
|  | bool ClobbersPredicate(MachineInstr &MI, std::vector<MachineOperand> &Pred, | 
|  | bool SkipDead) const override; | 
|  |  | 
|  | // Comparison optimization. | 
|  |  | 
|  | bool analyzeCompare(const MachineInstr &MI, Register &SrcReg, | 
|  | Register &SrcReg2, int64_t &Mask, | 
|  | int64_t &Value) const override; | 
|  |  | 
|  | bool optimizeCompareInstr(MachineInstr &CmpInstr, Register SrcReg, | 
|  | Register SrcReg2, int64_t Mask, int64_t Value, | 
|  | const MachineRegisterInfo *MRI) const override; | 
|  |  | 
|  |  | 
|  | /// Return true if get the base operand, byte offset of an instruction and | 
|  | /// the memory width. Width is the size of memory that is being | 
|  | /// loaded/stored (e.g. 1, 2, 4, 8). | 
|  | bool getMemOperandWithOffsetWidth(const MachineInstr &LdSt, | 
|  | const MachineOperand *&BaseOp, | 
|  | int64_t &Offset, unsigned &Width, | 
|  | const TargetRegisterInfo *TRI) const; | 
|  |  | 
|  | /// Get the base operand and byte offset of an instruction that reads/writes | 
|  | /// memory. | 
|  | bool getMemOperandsWithOffsetWidth( | 
|  | const MachineInstr &LdSt, | 
|  | SmallVectorImpl<const MachineOperand *> &BaseOps, int64_t &Offset, | 
|  | bool &OffsetIsScalable, unsigned &Width, | 
|  | const TargetRegisterInfo *TRI) const override; | 
|  |  | 
|  | /// Returns true if the two given memory operations should be scheduled | 
|  | /// adjacent. | 
|  | bool shouldClusterMemOps(ArrayRef<const MachineOperand *> BaseOps1, | 
|  | ArrayRef<const MachineOperand *> BaseOps2, | 
|  | unsigned NumLoads, unsigned NumBytes) const override; | 
|  |  | 
|  | /// Return true if two MIs access different memory addresses and false | 
|  | /// otherwise | 
|  | bool | 
|  | areMemAccessesTriviallyDisjoint(const MachineInstr &MIa, | 
|  | const MachineInstr &MIb) const override; | 
|  |  | 
|  | /// GetInstSize - Return the number of bytes of code the specified | 
|  | /// instruction may be.  This returns the maximum number of bytes. | 
|  | /// | 
|  | unsigned getInstSizeInBytes(const MachineInstr &MI) const override; | 
|  |  | 
|  | MCInst getNop() const override; | 
|  |  | 
|  | std::pair<unsigned, unsigned> | 
|  | decomposeMachineOperandsTargetFlags(unsigned TF) const override; | 
|  |  | 
|  | ArrayRef<std::pair<unsigned, const char *>> | 
|  | getSerializableDirectMachineOperandTargetFlags() const override; | 
|  |  | 
|  | ArrayRef<std::pair<unsigned, const char *>> | 
|  | getSerializableBitmaskMachineOperandTargetFlags() const override; | 
|  |  | 
|  | // Expand VSX Memory Pseudo instruction to either a VSX or a FP instruction. | 
|  | bool expandVSXMemPseudo(MachineInstr &MI) const; | 
|  |  | 
|  | // Lower pseudo instructions after register allocation. | 
|  | bool expandPostRAPseudo(MachineInstr &MI) const override; | 
|  |  | 
|  | static bool isVFRegister(unsigned Reg) { | 
|  | return Reg >= PPC::VF0 && Reg <= PPC::VF31; | 
|  | } | 
|  | static bool isVRRegister(unsigned Reg) { | 
|  | return Reg >= PPC::V0 && Reg <= PPC::V31; | 
|  | } | 
|  | const TargetRegisterClass *updatedRC(const TargetRegisterClass *RC) const; | 
|  | static int getRecordFormOpcode(unsigned Opcode); | 
|  |  | 
|  | bool isTOCSaveMI(const MachineInstr &MI) const; | 
|  |  | 
|  | bool isSignOrZeroExtended(const MachineInstr &MI, bool SignExt, | 
|  | const unsigned PhiDepth) const; | 
|  |  | 
|  | /// Return true if the output of the instruction is always a sign-extended, | 
|  | /// i.e. 0 to 31-th bits are same as 32-th bit. | 
|  | bool isSignExtended(const MachineInstr &MI, const unsigned depth = 0) const { | 
|  | return isSignOrZeroExtended(MI, true, depth); | 
|  | } | 
|  |  | 
|  | /// Return true if the output of the instruction is always zero-extended, | 
|  | /// i.e. 0 to 31-th bits are all zeros | 
|  | bool isZeroExtended(const MachineInstr &MI, const unsigned depth = 0) const { | 
|  | return isSignOrZeroExtended(MI, false, depth); | 
|  | } | 
|  |  | 
|  | bool convertToImmediateForm(MachineInstr &MI, | 
|  | MachineInstr **KilledDef = nullptr) const; | 
|  | bool foldFrameOffset(MachineInstr &MI) const; | 
|  | bool combineRLWINM(MachineInstr &MI, MachineInstr **ToErase = nullptr) const; | 
|  | bool isADDIInstrEligibleForFolding(MachineInstr &ADDIMI, int64_t &Imm) const; | 
|  | bool isADDInstrEligibleForFolding(MachineInstr &ADDMI) const; | 
|  | bool isImmInstrEligibleForFolding(MachineInstr &MI, unsigned &BaseReg, | 
|  | unsigned &XFormOpcode, | 
|  | int64_t &OffsetOfImmInstr, | 
|  | ImmInstrInfo &III) const; | 
|  | bool isValidToBeChangedReg(MachineInstr *ADDMI, unsigned Index, | 
|  | MachineInstr *&ADDIMI, int64_t &OffsetAddi, | 
|  | int64_t OffsetImm) const; | 
|  |  | 
|  | /// Fixup killed/dead flag for register \p RegNo between instructions [\p | 
|  | /// StartMI, \p EndMI]. Some pre-RA or post-RA transformations may violate | 
|  | /// register killed/dead flags semantics, this function can be called to fix | 
|  | /// up. Before calling this function, | 
|  | /// 1. Ensure that \p RegNo liveness is killed after instruction \p EndMI. | 
|  | /// 2. Ensure that there is no new definition between (\p StartMI, \p EndMI) | 
|  | ///    and possible definition for \p RegNo is \p StartMI or \p EndMI. For | 
|  | ///    pre-RA cases, definition may be \p StartMI through COPY, \p StartMI | 
|  | ///    will be adjust to true definition. | 
|  | /// 3. We can do accurate fixup for the case when all instructions between | 
|  | ///    [\p StartMI, \p EndMI] are in same basic block. | 
|  | /// 4. For the case when \p StartMI and \p EndMI are not in same basic block, | 
|  | ///    we conservatively clear kill flag for all uses of \p RegNo for pre-RA | 
|  | ///    and for post-RA, we give an assertion as without reaching definition | 
|  | ///    analysis post-RA, \p StartMI and \p EndMI are hard to keep right. | 
|  | void fixupIsDeadOrKill(MachineInstr *StartMI, MachineInstr *EndMI, | 
|  | unsigned RegNo) const; | 
|  | void replaceInstrWithLI(MachineInstr &MI, const LoadImmediateInfo &LII) const; | 
|  | void replaceInstrOperandWithImm(MachineInstr &MI, unsigned OpNo, | 
|  | int64_t Imm) const; | 
|  |  | 
|  | bool instrHasImmForm(unsigned Opc, bool IsVFReg, ImmInstrInfo &III, | 
|  | bool PostRA) const; | 
|  |  | 
|  | // In PostRA phase, try to find instruction defines \p Reg before \p MI. | 
|  | // \p SeenIntermediate is set to true if uses between DefMI and \p MI exist. | 
|  | MachineInstr *getDefMIPostRA(unsigned Reg, MachineInstr &MI, | 
|  | bool &SeenIntermediateUse) const; | 
|  |  | 
|  | // Materialize immediate after RA. | 
|  | void materializeImmPostRA(MachineBasicBlock &MBB, | 
|  | MachineBasicBlock::iterator MBBI, | 
|  | const DebugLoc &DL, Register Reg, | 
|  | int64_t Imm) const; | 
|  |  | 
|  | /// getRegNumForOperand - some operands use different numbering schemes | 
|  | /// for the same registers. For example, a VSX instruction may have any of | 
|  | /// vs0-vs63 allocated whereas an Altivec instruction could only have | 
|  | /// vs32-vs63 allocated (numbered as v0-v31). This function returns the actual | 
|  | /// register number needed for the opcode/operand number combination. | 
|  | /// The operand number argument will be useful when we need to extend this | 
|  | /// to instructions that use both Altivec and VSX numbering (for different | 
|  | /// operands). | 
|  | static unsigned getRegNumForOperand(const MCInstrDesc &Desc, unsigned Reg, | 
|  | unsigned OpNo) { | 
|  | int16_t regClass = Desc.OpInfo[OpNo].RegClass; | 
|  | switch (regClass) { | 
|  | // We store F0-F31, VF0-VF31 in MCOperand and it should be F0-F31, | 
|  | // VSX32-VSX63 during encoding/disassembling | 
|  | case PPC::VSSRCRegClassID: | 
|  | case PPC::VSFRCRegClassID: | 
|  | if (isVFRegister(Reg)) | 
|  | return PPC::VSX32 + (Reg - PPC::VF0); | 
|  | break; | 
|  | // We store VSL0-VSL31, V0-V31 in MCOperand and it should be VSL0-VSL31, | 
|  | // VSX32-VSX63 during encoding/disassembling | 
|  | case PPC::VSRCRegClassID: | 
|  | if (isVRRegister(Reg)) | 
|  | return PPC::VSX32 + (Reg - PPC::V0); | 
|  | break; | 
|  | // Other RegClass doesn't need mapping | 
|  | default: | 
|  | break; | 
|  | } | 
|  | return Reg; | 
|  | } | 
|  |  | 
|  | /// Check \p Opcode is BDNZ (Decrement CTR and branch if it is still nonzero). | 
|  | bool isBDNZ(unsigned Opcode) const; | 
|  |  | 
|  | /// Find the hardware loop instruction used to set-up the specified loop. | 
|  | /// On PPC, we have two instructions used to set-up the hardware loop | 
|  | /// (MTCTRloop, MTCTR8loop) with corresponding endloop (BDNZ, BDNZ8) | 
|  | /// instructions to indicate the end of a loop. | 
|  | MachineInstr * | 
|  | findLoopInstr(MachineBasicBlock &PreHeader, | 
|  | SmallPtrSet<MachineBasicBlock *, 8> &Visited) const; | 
|  |  | 
|  | /// Analyze loop L, which must be a single-basic-block loop, and if the | 
|  | /// conditions can be understood enough produce a PipelinerLoopInfo object. | 
|  | std::unique_ptr<TargetInstrInfo::PipelinerLoopInfo> | 
|  | analyzeLoopForPipelining(MachineBasicBlock *LoopBB) const override; | 
|  | }; | 
|  |  | 
|  | } | 
|  |  | 
|  | #endif |