003 File Manager
Current Path:
/usr/src/contrib/llvm-project/llvm/lib/Target/AMDGPU
usr
/
src
/
contrib
/
llvm-project
/
llvm
/
lib
/
Target
/
AMDGPU
/
📁
..
📄
AMDGPU.h
(11.46 KB)
📄
AMDGPU.td
(36.97 KB)
📄
AMDGPUAliasAnalysis.cpp
(5.58 KB)
📄
AMDGPUAliasAnalysis.h
(3.32 KB)
📄
AMDGPUAlwaysInlinePass.cpp
(4.83 KB)
📄
AMDGPUAnnotateKernelFeatures.cpp
(11.94 KB)
📄
AMDGPUAnnotateUniformValues.cpp
(6.13 KB)
📄
AMDGPUArgumentUsageInfo.cpp
(7.66 KB)
📄
AMDGPUArgumentUsageInfo.h
(4.81 KB)
📄
AMDGPUAsmPrinter.cpp
(50.42 KB)
📄
AMDGPUAsmPrinter.h
(5.13 KB)
📄
AMDGPUAtomicOptimizer.cpp
(23.79 KB)
📄
AMDGPUCallLowering.cpp
(28.66 KB)
📄
AMDGPUCallLowering.h
(2.37 KB)
📄
AMDGPUCallingConv.td
(7.33 KB)
📄
AMDGPUCodeGenPrepare.cpp
(46.42 KB)
📄
AMDGPUCombine.td
(2.79 KB)
📄
AMDGPUExportClustering.cpp
(4.52 KB)
📄
AMDGPUExportClustering.h
(533 B)
📄
AMDGPUFeatures.td
(1.81 KB)
📄
AMDGPUFixFunctionBitcasts.cpp
(1.87 KB)
📄
AMDGPUFrameLowering.cpp
(1.98 KB)
📄
AMDGPUFrameLowering.h
(1.39 KB)
📄
AMDGPUGISel.td
(11.57 KB)
📄
AMDGPUGenRegisterBankInfo.def
(5.83 KB)
📄
AMDGPUGlobalISelUtils.cpp
(1.77 KB)
📄
AMDGPUGlobalISelUtils.h
(2.07 KB)
📄
AMDGPUHSAMetadataStreamer.cpp
(31.21 KB)
📄
AMDGPUHSAMetadataStreamer.h
(5.46 KB)
📄
AMDGPUISelDAGToDAG.cpp
(101.59 KB)
📄
AMDGPUISelLowering.cpp
(168.65 KB)
📄
AMDGPUISelLowering.h
(19.23 KB)
📄
AMDGPUInline.cpp
(7.97 KB)
📄
AMDGPUInstrInfo.cpp
(1.71 KB)
📄
AMDGPUInstrInfo.h
(1.66 KB)
📄
AMDGPUInstrInfo.td
(17.18 KB)
📄
AMDGPUInstructionSelector.cpp
(128.53 KB)
📄
AMDGPUInstructionSelector.h
(11.04 KB)
📄
AMDGPUInstructions.td
(25.36 KB)
📄
AMDGPULegalizerInfo.cpp
(149.32 KB)
📄
AMDGPULegalizerInfo.h
(8.49 KB)
📄
AMDGPULibCalls.cpp
(53.89 KB)
📄
AMDGPULibFunc.cpp
(37.85 KB)
📄
AMDGPULibFunc.h
(10.99 KB)
📄
AMDGPULowerIntrinsics.cpp
(4.55 KB)
📄
AMDGPULowerKernelArguments.cpp
(8.89 KB)
📄
AMDGPULowerKernelAttributes.cpp
(7.78 KB)
📄
AMDGPUMCInstLower.cpp
(14.27 KB)
📄
AMDGPUMachineCFGStructurizer.cpp
(101.97 KB)
📄
AMDGPUMachineFunction.cpp
(2.24 KB)
📄
AMDGPUMachineFunction.h
(2.13 KB)
📄
AMDGPUMachineModuleInfo.cpp
(1.34 KB)
📄
AMDGPUMachineModuleInfo.h
(5.46 KB)
📄
AMDGPUMacroFusion.cpp
(2.28 KB)
📄
AMDGPUMacroFusion.h
(679 B)
📄
AMDGPUOpenCLEnqueuedBlockLowering.cpp
(5.31 KB)
📄
AMDGPUPTNote.h
(1.29 KB)
📄
AMDGPUPerfHintAnalysis.cpp
(12.17 KB)
📄
AMDGPUPerfHintAnalysis.h
(1.67 KB)
📄
AMDGPUPostLegalizerCombiner.cpp
(12.02 KB)
📄
AMDGPUPreLegalizerCombiner.cpp
(5.45 KB)
📄
AMDGPUPrintfRuntimeBinding.cpp
(21.7 KB)
📄
AMDGPUPromoteAlloca.cpp
(35.24 KB)
📄
AMDGPUPropagateAttributes.cpp
(11.76 KB)
📄
AMDGPURegBankCombiner.cpp
(5.36 KB)
📄
AMDGPURegisterBankInfo.cpp
(161.67 KB)
📄
AMDGPURegisterBankInfo.h
(7.41 KB)
📄
AMDGPURegisterBanks.td
(921 B)
📄
AMDGPURewriteOutArguments.cpp
(15.82 KB)
📄
AMDGPUSearchableTables.td
(21.04 KB)
📄
AMDGPUSubtarget.cpp
(29.62 KB)
📄
AMDGPUSubtarget.h
(35.82 KB)
📄
AMDGPUTargetMachine.cpp
(42.67 KB)
📄
AMDGPUTargetMachine.h
(4.52 KB)
📄
AMDGPUTargetObjectFile.cpp
(1.54 KB)
📄
AMDGPUTargetObjectFile.h
(1.14 KB)
📄
AMDGPUTargetTransformInfo.cpp
(39.07 KB)
📄
AMDGPUTargetTransformInfo.h
(11.11 KB)
📄
AMDGPUUnifyDivergentExitNodes.cpp
(13.84 KB)
📄
AMDGPUUnifyMetadata.cpp
(4.46 KB)
📄
AMDILCFGStructurizer.cpp
(56.32 KB)
📄
AMDKernelCodeT.h
(32.84 KB)
📁
AsmParser
📄
BUFInstructions.td
(110.75 KB)
📄
CaymanInstructions.td
(7.93 KB)
📄
DSInstructions.td
(52.37 KB)
📁
Disassembler
📄
EvergreenInstructions.td
(28.24 KB)
📄
FLATInstructions.td
(66.93 KB)
📄
GCNDPPCombine.cpp
(19.92 KB)
📄
GCNHazardRecognizer.cpp
(45.3 KB)
📄
GCNHazardRecognizer.h
(3.96 KB)
📄
GCNILPSched.cpp
(11.3 KB)
📄
GCNIterativeScheduler.cpp
(20.62 KB)
📄
GCNIterativeScheduler.h
(4.16 KB)
📄
GCNMinRegStrategy.cpp
(8.47 KB)
📄
GCNNSAReassign.cpp
(10.92 KB)
📄
GCNProcessors.td
(4.84 KB)
📄
GCNRegBankReassign.cpp
(26.68 KB)
📄
GCNRegPressure.cpp
(16.27 KB)
📄
GCNRegPressure.h
(9.15 KB)
📄
GCNSchedStrategy.cpp
(21.67 KB)
📄
GCNSchedStrategy.h
(3.77 KB)
📁
MCTargetDesc
📄
MIMGInstructions.td
(39.85 KB)
📄
R600.td
(1.51 KB)
📄
R600AsmPrinter.cpp
(4.46 KB)
📄
R600AsmPrinter.h
(1.5 KB)
📄
R600ClauseMergePass.cpp
(7.38 KB)
📄
R600ControlFlowFinalizer.cpp
(23.4 KB)
📄
R600Defines.h
(4.25 KB)
📄
R600EmitClauseMarkers.cpp
(12.1 KB)
📄
R600ExpandSpecialInstrs.cpp
(10.11 KB)
📄
R600FrameLowering.cpp
(1.83 KB)
📄
R600FrameLowering.h
(1.25 KB)
📄
R600ISelLowering.cpp
(81.88 KB)
📄
R600ISelLowering.h
(4.8 KB)
📄
R600InstrFormats.td
(11.58 KB)
📄
R600InstrInfo.cpp
(49.47 KB)
📄
R600InstrInfo.h
(13.7 KB)
📄
R600Instructions.td
(55.13 KB)
📄
R600MachineFunctionInfo.cpp
(551 B)
📄
R600MachineFunctionInfo.h
(824 B)
📄
R600MachineScheduler.cpp
(13.57 KB)
📄
R600MachineScheduler.h
(2.53 KB)
📄
R600OpenCLImageTypeLoweringPass.cpp
(11.75 KB)
📄
R600OptimizeVectorRegisters.cpp
(13.4 KB)
📄
R600Packetizer.cpp
(13.4 KB)
📄
R600Processors.td
(4.42 KB)
📄
R600RegisterInfo.cpp
(3.95 KB)
📄
R600RegisterInfo.h
(2 KB)
📄
R600RegisterInfo.td
(9.75 KB)
📄
R600Schedule.td
(1.62 KB)
📄
R700Instructions.td
(783 B)
📄
SIAddIMGInit.cpp
(6.24 KB)
📄
SIAnnotateControlFlow.cpp
(11.18 KB)
📄
SIDefines.h
(20.86 KB)
📄
SIFixSGPRCopies.cpp
(29.46 KB)
📄
SIFixVGPRCopies.cpp
(2 KB)
📄
SIFixupVectorISel.cpp
(8.75 KB)
📄
SIFoldOperands.cpp
(54.56 KB)
📄
SIFormMemoryClauses.cpp
(12.76 KB)
📄
SIFrameLowering.cpp
(48.08 KB)
📄
SIFrameLowering.h
(2.98 KB)
📄
SIISelLowering.cpp
(423.43 KB)
📄
SIISelLowering.h
(22.13 KB)
📄
SIInsertHardClauses.cpp
(7.01 KB)
📄
SIInsertSkips.cpp
(15.29 KB)
📄
SIInsertWaitcnts.cpp
(58.33 KB)
📄
SIInstrFormats.td
(9.44 KB)
📄
SIInstrInfo.cpp
(247.15 KB)
📄
SIInstrInfo.h
(41.24 KB)
📄
SIInstrInfo.td
(90.7 KB)
📄
SIInstructions.td
(77.7 KB)
📄
SILoadStoreOptimizer.cpp
(76.21 KB)
📄
SILowerControlFlow.cpp
(22.66 KB)
📄
SILowerI1Copies.cpp
(27.83 KB)
📄
SILowerSGPRSpills.cpp
(12.68 KB)
📄
SIMachineFunctionInfo.cpp
(20.01 KB)
📄
SIMachineFunctionInfo.h
(26.91 KB)
📄
SIMachineScheduler.cpp
(69.44 KB)
📄
SIMachineScheduler.h
(15.65 KB)
📄
SIMemoryLegalizer.cpp
(45.84 KB)
📄
SIModeRegister.cpp
(17.43 KB)
📄
SIOptimizeExecMasking.cpp
(12.81 KB)
📄
SIOptimizeExecMaskingPreRA.cpp
(11.13 KB)
📄
SIPeepholeSDWA.cpp
(42.84 KB)
📄
SIPostRABundler.cpp
(3.6 KB)
📄
SIPreAllocateWWMRegs.cpp
(6.09 KB)
📄
SIPreEmitPeephole.cpp
(10.51 KB)
📄
SIProgramInfo.h
(2.04 KB)
📄
SIRegisterInfo.cpp
(71.51 KB)
📄
SIRegisterInfo.h
(13.04 KB)
📄
SIRegisterInfo.td
(37.28 KB)
📄
SIRemoveShortExecBranches.cpp
(4.96 KB)
📄
SISchedule.td
(7.58 KB)
📄
SIShrinkInstructions.cpp
(26.86 KB)
📄
SIWholeQuadMode.cpp
(30.22 KB)
📄
SMInstructions.td
(48.14 KB)
📄
SOPInstructions.td
(60.51 KB)
📁
TargetInfo
📁
Utils
📄
VIInstrFormats.td
(645 B)
📄
VOP1Instructions.td
(35.53 KB)
📄
VOP2Instructions.td
(65.04 KB)
📄
VOP3Instructions.td
(53.14 KB)
📄
VOP3PInstructions.td
(26.47 KB)
📄
VOPCInstructions.td
(63.31 KB)
📄
VOPInstructions.td
(23.76 KB)
Editing: SIMemoryLegalizer.cpp
//===- SIMemoryLegalizer.cpp ----------------------------------------------===// // // Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. // See https://llvm.org/LICENSE.txt for license information. // SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception // //===----------------------------------------------------------------------===// // /// \file /// Memory legalizer - implements memory model. More information can be /// found here: /// http://llvm.org/docs/AMDGPUUsage.html#memory-model // //===----------------------------------------------------------------------===// #include "AMDGPU.h" #include "AMDGPUMachineModuleInfo.h" #include "AMDGPUSubtarget.h" #include "SIDefines.h" #include "SIInstrInfo.h" #include "MCTargetDesc/AMDGPUMCTargetDesc.h" #include "Utils/AMDGPUBaseInfo.h" #include "llvm/ADT/BitmaskEnum.h" #include "llvm/ADT/None.h" #include "llvm/ADT/Optional.h" #include "llvm/CodeGen/MachineBasicBlock.h" #include "llvm/CodeGen/MachineFunction.h" #include "llvm/CodeGen/MachineFunctionPass.h" #include "llvm/CodeGen/MachineInstrBuilder.h" #include "llvm/CodeGen/MachineMemOperand.h" #include "llvm/CodeGen/MachineModuleInfo.h" #include "llvm/CodeGen/MachineOperand.h" #include "llvm/IR/DebugLoc.h" #include "llvm/IR/DiagnosticInfo.h" #include "llvm/IR/Function.h" #include "llvm/IR/LLVMContext.h" #include "llvm/MC/MCInstrDesc.h" #include "llvm/Pass.h" #include "llvm/Support/AtomicOrdering.h" #include "llvm/Support/MathExtras.h" #include <cassert> #include <list> using namespace llvm; using namespace llvm::AMDGPU; #define DEBUG_TYPE "si-memory-legalizer" #define PASS_NAME "SI Memory Legalizer" namespace { LLVM_ENABLE_BITMASK_ENUMS_IN_NAMESPACE(); /// Memory operation flags. Can be ORed together. enum class SIMemOp { NONE = 0u, LOAD = 1u << 0, STORE = 1u << 1, LLVM_MARK_AS_BITMASK_ENUM(/* LargestFlag = */ STORE) }; /// Position to insert a new instruction relative to an existing /// instruction. enum class Position { BEFORE, AFTER }; /// The atomic synchronization scopes supported by the AMDGPU target. enum class SIAtomicScope { NONE, SINGLETHREAD, WAVEFRONT, WORKGROUP, AGENT, SYSTEM }; /// The distinct address spaces supported by the AMDGPU target for /// atomic memory operation. Can be ORed toether. enum class SIAtomicAddrSpace { NONE = 0u, GLOBAL = 1u << 0, LDS = 1u << 1, SCRATCH = 1u << 2, GDS = 1u << 3, OTHER = 1u << 4, /// The address spaces that can be accessed by a FLAT instruction. FLAT = GLOBAL | LDS | SCRATCH, /// The address spaces that support atomic instructions. ATOMIC = GLOBAL | LDS | SCRATCH | GDS, /// All address spaces. ALL = GLOBAL | LDS | SCRATCH | GDS | OTHER, LLVM_MARK_AS_BITMASK_ENUM(/* LargestFlag = */ ALL) }; /// Sets named bit \p BitName to "true" if present in instruction \p MI. /// \returns Returns true if \p MI is modified, false otherwise. template <uint16_t BitName> bool enableNamedBit(const MachineBasicBlock::iterator &MI) { int BitIdx = AMDGPU::getNamedOperandIdx(MI->getOpcode(), BitName); if (BitIdx == -1) return false; MachineOperand &Bit = MI->getOperand(BitIdx); if (Bit.getImm() != 0) return false; Bit.setImm(1); return true; } class SIMemOpInfo final { private: friend class SIMemOpAccess; AtomicOrdering Ordering = AtomicOrdering::NotAtomic; AtomicOrdering FailureOrdering = AtomicOrdering::NotAtomic; SIAtomicScope Scope = SIAtomicScope::SYSTEM; SIAtomicAddrSpace OrderingAddrSpace = SIAtomicAddrSpace::NONE; SIAtomicAddrSpace InstrAddrSpace = SIAtomicAddrSpace::NONE; bool IsCrossAddressSpaceOrdering = false; bool IsNonTemporal = false; SIMemOpInfo(AtomicOrdering Ordering = AtomicOrdering::SequentiallyConsistent, SIAtomicScope Scope = SIAtomicScope::SYSTEM, SIAtomicAddrSpace OrderingAddrSpace = SIAtomicAddrSpace::ATOMIC, SIAtomicAddrSpace InstrAddrSpace = SIAtomicAddrSpace::ALL, bool IsCrossAddressSpaceOrdering = true, AtomicOrdering FailureOrdering = AtomicOrdering::SequentiallyConsistent, bool IsNonTemporal = false) : Ordering(Ordering), FailureOrdering(FailureOrdering), Scope(Scope), OrderingAddrSpace(OrderingAddrSpace), InstrAddrSpace(InstrAddrSpace), IsCrossAddressSpaceOrdering(IsCrossAddressSpaceOrdering), IsNonTemporal(IsNonTemporal) { // There is also no cross address space ordering if the ordering // address space is the same as the instruction address space and // only contains a single address space. if ((OrderingAddrSpace == InstrAddrSpace) && isPowerOf2_32(uint32_t(InstrAddrSpace))) this->IsCrossAddressSpaceOrdering = false; } public: /// \returns Atomic synchronization scope of the machine instruction used to /// create this SIMemOpInfo. SIAtomicScope getScope() const { return Scope; } /// \returns Ordering constraint of the machine instruction used to /// create this SIMemOpInfo. AtomicOrdering getOrdering() const { return Ordering; } /// \returns Failure ordering constraint of the machine instruction used to /// create this SIMemOpInfo. AtomicOrdering getFailureOrdering() const { return FailureOrdering; } /// \returns The address spaces be accessed by the machine /// instruction used to create this SiMemOpInfo. SIAtomicAddrSpace getInstrAddrSpace() const { return InstrAddrSpace; } /// \returns The address spaces that must be ordered by the machine /// instruction used to create this SiMemOpInfo. SIAtomicAddrSpace getOrderingAddrSpace() const { return OrderingAddrSpace; } /// \returns Return true iff memory ordering of operations on /// different address spaces is required. bool getIsCrossAddressSpaceOrdering() const { return IsCrossAddressSpaceOrdering; } /// \returns True if memory access of the machine instruction used to /// create this SIMemOpInfo is non-temporal, false otherwise. bool isNonTemporal() const { return IsNonTemporal; } /// \returns True if ordering constraint of the machine instruction used to /// create this SIMemOpInfo is unordered or higher, false otherwise. bool isAtomic() const { return Ordering != AtomicOrdering::NotAtomic; } }; class SIMemOpAccess final { private: AMDGPUMachineModuleInfo *MMI = nullptr; /// Reports unsupported message \p Msg for \p MI to LLVM context. void reportUnsupported(const MachineBasicBlock::iterator &MI, const char *Msg) const; /// Inspects the target synchonization scope \p SSID and determines /// the SI atomic scope it corresponds to, the address spaces it /// covers, and whether the memory ordering applies between address /// spaces. Optional<std::tuple<SIAtomicScope, SIAtomicAddrSpace, bool>> toSIAtomicScope(SyncScope::ID SSID, SIAtomicAddrSpace InstrScope) const; /// \return Return a bit set of the address spaces accessed by \p AS. SIAtomicAddrSpace toSIAtomicAddrSpace(unsigned AS) const; /// \returns Info constructed from \p MI, which has at least machine memory /// operand. Optional<SIMemOpInfo> constructFromMIWithMMO( const MachineBasicBlock::iterator &MI) const; public: /// Construct class to support accessing the machine memory operands /// of instructions in the machine function \p MF. SIMemOpAccess(MachineFunction &MF); /// \returns Load info if \p MI is a load operation, "None" otherwise. Optional<SIMemOpInfo> getLoadInfo( const MachineBasicBlock::iterator &MI) const; /// \returns Store info if \p MI is a store operation, "None" otherwise. Optional<SIMemOpInfo> getStoreInfo( const MachineBasicBlock::iterator &MI) const; /// \returns Atomic fence info if \p MI is an atomic fence operation, /// "None" otherwise. Optional<SIMemOpInfo> getAtomicFenceInfo( const MachineBasicBlock::iterator &MI) const; /// \returns Atomic cmpxchg/rmw info if \p MI is an atomic cmpxchg or /// rmw operation, "None" otherwise. Optional<SIMemOpInfo> getAtomicCmpxchgOrRmwInfo( const MachineBasicBlock::iterator &MI) const; }; class SICacheControl { protected: /// Instruction info. const SIInstrInfo *TII = nullptr; IsaVersion IV; /// Whether to insert cache invalidation instructions. bool InsertCacheInv; SICacheControl(const GCNSubtarget &ST); public: /// Create a cache control for the subtarget \p ST. static std::unique_ptr<SICacheControl> create(const GCNSubtarget &ST); /// Update \p MI memory load instruction to bypass any caches up to /// the \p Scope memory scope for address spaces \p /// AddrSpace. Return true iff the instruction was modified. virtual bool enableLoadCacheBypass(const MachineBasicBlock::iterator &MI, SIAtomicScope Scope, SIAtomicAddrSpace AddrSpace) const = 0; /// Update \p MI memory instruction to indicate it is /// nontemporal. Return true iff the instruction was modified. virtual bool enableNonTemporal(const MachineBasicBlock::iterator &MI) const = 0; /// Inserts any necessary instructions at position \p Pos relative /// to instruction \p MI to ensure any caches associated with /// address spaces \p AddrSpace for memory scopes up to memory scope /// \p Scope are invalidated. Returns true iff any instructions /// inserted. virtual bool insertCacheInvalidate(MachineBasicBlock::iterator &MI, SIAtomicScope Scope, SIAtomicAddrSpace AddrSpace, Position Pos) const = 0; /// Inserts any necessary instructions at position \p Pos relative /// to instruction \p MI to ensure memory instructions of kind \p Op /// associated with address spaces \p AddrSpace have completed as /// observed by other memory instructions executing in memory scope /// \p Scope. \p IsCrossAddrSpaceOrdering indicates if the memory /// ordering is between address spaces. Returns true iff any /// instructions inserted. virtual bool insertWait(MachineBasicBlock::iterator &MI, SIAtomicScope Scope, SIAtomicAddrSpace AddrSpace, SIMemOp Op, bool IsCrossAddrSpaceOrdering, Position Pos) const = 0; /// Virtual destructor to allow derivations to be deleted. virtual ~SICacheControl() = default; }; class SIGfx6CacheControl : public SICacheControl { protected: /// Sets GLC bit to "true" if present in \p MI. Returns true if \p MI /// is modified, false otherwise. bool enableGLCBit(const MachineBasicBlock::iterator &MI) const { return enableNamedBit<AMDGPU::OpName::glc>(MI); } /// Sets SLC bit to "true" if present in \p MI. Returns true if \p MI /// is modified, false otherwise. bool enableSLCBit(const MachineBasicBlock::iterator &MI) const { return enableNamedBit<AMDGPU::OpName::slc>(MI); } public: SIGfx6CacheControl(const GCNSubtarget &ST) : SICacheControl(ST) {}; bool enableLoadCacheBypass(const MachineBasicBlock::iterator &MI, SIAtomicScope Scope, SIAtomicAddrSpace AddrSpace) const override; bool enableNonTemporal(const MachineBasicBlock::iterator &MI) const override; bool insertCacheInvalidate(MachineBasicBlock::iterator &MI, SIAtomicScope Scope, SIAtomicAddrSpace AddrSpace, Position Pos) const override; bool insertWait(MachineBasicBlock::iterator &MI, SIAtomicScope Scope, SIAtomicAddrSpace AddrSpace, SIMemOp Op, bool IsCrossAddrSpaceOrdering, Position Pos) const override; }; class SIGfx7CacheControl : public SIGfx6CacheControl { public: SIGfx7CacheControl(const GCNSubtarget &ST) : SIGfx6CacheControl(ST) {}; bool insertCacheInvalidate(MachineBasicBlock::iterator &MI, SIAtomicScope Scope, SIAtomicAddrSpace AddrSpace, Position Pos) const override; }; class SIGfx10CacheControl : public SIGfx7CacheControl { protected: bool CuMode = false; /// Sets DLC bit to "true" if present in \p MI. Returns true if \p MI /// is modified, false otherwise. bool enableDLCBit(const MachineBasicBlock::iterator &MI) const { return enableNamedBit<AMDGPU::OpName::dlc>(MI); } public: SIGfx10CacheControl(const GCNSubtarget &ST, bool CuMode) : SIGfx7CacheControl(ST), CuMode(CuMode) {}; bool enableLoadCacheBypass(const MachineBasicBlock::iterator &MI, SIAtomicScope Scope, SIAtomicAddrSpace AddrSpace) const override; bool enableNonTemporal(const MachineBasicBlock::iterator &MI) const override; bool insertCacheInvalidate(MachineBasicBlock::iterator &MI, SIAtomicScope Scope, SIAtomicAddrSpace AddrSpace, Position Pos) const override; bool insertWait(MachineBasicBlock::iterator &MI, SIAtomicScope Scope, SIAtomicAddrSpace AddrSpace, SIMemOp Op, bool IsCrossAddrSpaceOrdering, Position Pos) const override; }; class SIMemoryLegalizer final : public MachineFunctionPass { private: /// Cache Control. std::unique_ptr<SICacheControl> CC = nullptr; /// List of atomic pseudo instructions. std::list<MachineBasicBlock::iterator> AtomicPseudoMIs; /// Return true iff instruction \p MI is a atomic instruction that /// returns a result. bool isAtomicRet(const MachineInstr &MI) const { return AMDGPU::getAtomicNoRetOp(MI.getOpcode()) != -1; } /// Removes all processed atomic pseudo instructions from the current /// function. Returns true if current function is modified, false otherwise. bool removeAtomicPseudoMIs(); /// Expands load operation \p MI. Returns true if instructions are /// added/deleted or \p MI is modified, false otherwise. bool expandLoad(const SIMemOpInfo &MOI, MachineBasicBlock::iterator &MI); /// Expands store operation \p MI. Returns true if instructions are /// added/deleted or \p MI is modified, false otherwise. bool expandStore(const SIMemOpInfo &MOI, MachineBasicBlock::iterator &MI); /// Expands atomic fence operation \p MI. Returns true if /// instructions are added/deleted or \p MI is modified, false otherwise. bool expandAtomicFence(const SIMemOpInfo &MOI, MachineBasicBlock::iterator &MI); /// Expands atomic cmpxchg or rmw operation \p MI. Returns true if /// instructions are added/deleted or \p MI is modified, false otherwise. bool expandAtomicCmpxchgOrRmw(const SIMemOpInfo &MOI, MachineBasicBlock::iterator &MI); public: static char ID; SIMemoryLegalizer() : MachineFunctionPass(ID) {} void getAnalysisUsage(AnalysisUsage &AU) const override { AU.setPreservesCFG(); MachineFunctionPass::getAnalysisUsage(AU); } StringRef getPassName() const override { return PASS_NAME; } bool runOnMachineFunction(MachineFunction &MF) override; }; } // end namespace anonymous void SIMemOpAccess::reportUnsupported(const MachineBasicBlock::iterator &MI, const char *Msg) const { const Function &Func = MI->getParent()->getParent()->getFunction(); DiagnosticInfoUnsupported Diag(Func, Msg, MI->getDebugLoc()); Func.getContext().diagnose(Diag); } Optional<std::tuple<SIAtomicScope, SIAtomicAddrSpace, bool>> SIMemOpAccess::toSIAtomicScope(SyncScope::ID SSID, SIAtomicAddrSpace InstrScope) const { if (SSID == SyncScope::System) return std::make_tuple(SIAtomicScope::SYSTEM, SIAtomicAddrSpace::ATOMIC, true); if (SSID == MMI->getAgentSSID()) return std::make_tuple(SIAtomicScope::AGENT, SIAtomicAddrSpace::ATOMIC, true); if (SSID == MMI->getWorkgroupSSID()) return std::make_tuple(SIAtomicScope::WORKGROUP, SIAtomicAddrSpace::ATOMIC, true); if (SSID == MMI->getWavefrontSSID()) return std::make_tuple(SIAtomicScope::WAVEFRONT, SIAtomicAddrSpace::ATOMIC, true); if (SSID == SyncScope::SingleThread) return std::make_tuple(SIAtomicScope::SINGLETHREAD, SIAtomicAddrSpace::ATOMIC, true); if (SSID == MMI->getSystemOneAddressSpaceSSID()) return std::make_tuple(SIAtomicScope::SYSTEM, SIAtomicAddrSpace::ATOMIC & InstrScope, false); if (SSID == MMI->getAgentOneAddressSpaceSSID()) return std::make_tuple(SIAtomicScope::AGENT, SIAtomicAddrSpace::ATOMIC & InstrScope, false); if (SSID == MMI->getWorkgroupOneAddressSpaceSSID()) return std::make_tuple(SIAtomicScope::WORKGROUP, SIAtomicAddrSpace::ATOMIC & InstrScope, false); if (SSID == MMI->getWavefrontOneAddressSpaceSSID()) return std::make_tuple(SIAtomicScope::WAVEFRONT, SIAtomicAddrSpace::ATOMIC & InstrScope, false); if (SSID == MMI->getSingleThreadOneAddressSpaceSSID()) return std::make_tuple(SIAtomicScope::SINGLETHREAD, SIAtomicAddrSpace::ATOMIC & InstrScope, false); return None; } SIAtomicAddrSpace SIMemOpAccess::toSIAtomicAddrSpace(unsigned AS) const { if (AS == AMDGPUAS::FLAT_ADDRESS) return SIAtomicAddrSpace::FLAT; if (AS == AMDGPUAS::GLOBAL_ADDRESS) return SIAtomicAddrSpace::GLOBAL; if (AS == AMDGPUAS::LOCAL_ADDRESS) return SIAtomicAddrSpace::LDS; if (AS == AMDGPUAS::PRIVATE_ADDRESS) return SIAtomicAddrSpace::SCRATCH; if (AS == AMDGPUAS::REGION_ADDRESS) return SIAtomicAddrSpace::GDS; return SIAtomicAddrSpace::OTHER; } SIMemOpAccess::SIMemOpAccess(MachineFunction &MF) { MMI = &MF.getMMI().getObjFileInfo<AMDGPUMachineModuleInfo>(); } Optional<SIMemOpInfo> SIMemOpAccess::constructFromMIWithMMO( const MachineBasicBlock::iterator &MI) const { assert(MI->getNumMemOperands() > 0); SyncScope::ID SSID = SyncScope::SingleThread; AtomicOrdering Ordering = AtomicOrdering::NotAtomic; AtomicOrdering FailureOrdering = AtomicOrdering::NotAtomic; SIAtomicAddrSpace InstrAddrSpace = SIAtomicAddrSpace::NONE; bool IsNonTemporal = true; // Validator should check whether or not MMOs cover the entire set of // locations accessed by the memory instruction. for (const auto &MMO : MI->memoperands()) { IsNonTemporal &= MMO->isNonTemporal(); InstrAddrSpace |= toSIAtomicAddrSpace(MMO->getPointerInfo().getAddrSpace()); AtomicOrdering OpOrdering = MMO->getOrdering(); if (OpOrdering != AtomicOrdering::NotAtomic) { const auto &IsSyncScopeInclusion = MMI->isSyncScopeInclusion(SSID, MMO->getSyncScopeID()); if (!IsSyncScopeInclusion) { reportUnsupported(MI, "Unsupported non-inclusive atomic synchronization scope"); return None; } SSID = IsSyncScopeInclusion.getValue() ? SSID : MMO->getSyncScopeID(); Ordering = isStrongerThan(Ordering, OpOrdering) ? Ordering : MMO->getOrdering(); assert(MMO->getFailureOrdering() != AtomicOrdering::Release && MMO->getFailureOrdering() != AtomicOrdering::AcquireRelease); FailureOrdering = isStrongerThan(FailureOrdering, MMO->getFailureOrdering()) ? FailureOrdering : MMO->getFailureOrdering(); } } SIAtomicScope Scope = SIAtomicScope::NONE; SIAtomicAddrSpace OrderingAddrSpace = SIAtomicAddrSpace::NONE; bool IsCrossAddressSpaceOrdering = false; if (Ordering != AtomicOrdering::NotAtomic) { auto ScopeOrNone = toSIAtomicScope(SSID, InstrAddrSpace); if (!ScopeOrNone) { reportUnsupported(MI, "Unsupported atomic synchronization scope"); return None; } std::tie(Scope, OrderingAddrSpace, IsCrossAddressSpaceOrdering) = ScopeOrNone.getValue(); if ((OrderingAddrSpace == SIAtomicAddrSpace::NONE) || ((OrderingAddrSpace & SIAtomicAddrSpace::ATOMIC) != OrderingAddrSpace)) { reportUnsupported(MI, "Unsupported atomic address space"); return None; } } return SIMemOpInfo(Ordering, Scope, OrderingAddrSpace, InstrAddrSpace, IsCrossAddressSpaceOrdering, FailureOrdering, IsNonTemporal); } Optional<SIMemOpInfo> SIMemOpAccess::getLoadInfo( const MachineBasicBlock::iterator &MI) const { assert(MI->getDesc().TSFlags & SIInstrFlags::maybeAtomic); if (!(MI->mayLoad() && !MI->mayStore())) return None; // Be conservative if there are no memory operands. if (MI->getNumMemOperands() == 0) return SIMemOpInfo(); return constructFromMIWithMMO(MI); } Optional<SIMemOpInfo> SIMemOpAccess::getStoreInfo( const MachineBasicBlock::iterator &MI) const { assert(MI->getDesc().TSFlags & SIInstrFlags::maybeAtomic); if (!(!MI->mayLoad() && MI->mayStore())) return None; // Be conservative if there are no memory operands. if (MI->getNumMemOperands() == 0) return SIMemOpInfo(); return constructFromMIWithMMO(MI); } Optional<SIMemOpInfo> SIMemOpAccess::getAtomicFenceInfo( const MachineBasicBlock::iterator &MI) const { assert(MI->getDesc().TSFlags & SIInstrFlags::maybeAtomic); if (MI->getOpcode() != AMDGPU::ATOMIC_FENCE) return None; AtomicOrdering Ordering = static_cast<AtomicOrdering>(MI->getOperand(0).getImm()); SyncScope::ID SSID = static_cast<SyncScope::ID>(MI->getOperand(1).getImm()); auto ScopeOrNone = toSIAtomicScope(SSID, SIAtomicAddrSpace::ATOMIC); if (!ScopeOrNone) { reportUnsupported(MI, "Unsupported atomic synchronization scope"); return None; } SIAtomicScope Scope = SIAtomicScope::NONE; SIAtomicAddrSpace OrderingAddrSpace = SIAtomicAddrSpace::NONE; bool IsCrossAddressSpaceOrdering = false; std::tie(Scope, OrderingAddrSpace, IsCrossAddressSpaceOrdering) = ScopeOrNone.getValue(); if ((OrderingAddrSpace == SIAtomicAddrSpace::NONE) || ((OrderingAddrSpace & SIAtomicAddrSpace::ATOMIC) != OrderingAddrSpace)) { reportUnsupported(MI, "Unsupported atomic address space"); return None; } return SIMemOpInfo(Ordering, Scope, OrderingAddrSpace, SIAtomicAddrSpace::ATOMIC, IsCrossAddressSpaceOrdering); } Optional<SIMemOpInfo> SIMemOpAccess::getAtomicCmpxchgOrRmwInfo( const MachineBasicBlock::iterator &MI) const { assert(MI->getDesc().TSFlags & SIInstrFlags::maybeAtomic); if (!(MI->mayLoad() && MI->mayStore())) return None; // Be conservative if there are no memory operands. if (MI->getNumMemOperands() == 0) return SIMemOpInfo(); return constructFromMIWithMMO(MI); } SICacheControl::SICacheControl(const GCNSubtarget &ST) { TII = ST.getInstrInfo(); IV = getIsaVersion(ST.getCPU()); InsertCacheInv = !ST.isAmdPalOS(); } /* static */ std::unique_ptr<SICacheControl> SICacheControl::create(const GCNSubtarget &ST) { GCNSubtarget::Generation Generation = ST.getGeneration(); if (Generation <= AMDGPUSubtarget::SOUTHERN_ISLANDS) return std::make_unique<SIGfx6CacheControl>(ST); if (Generation < AMDGPUSubtarget::GFX10) return std::make_unique<SIGfx7CacheControl>(ST); return std::make_unique<SIGfx10CacheControl>(ST, ST.isCuModeEnabled()); } bool SIGfx6CacheControl::enableLoadCacheBypass( const MachineBasicBlock::iterator &MI, SIAtomicScope Scope, SIAtomicAddrSpace AddrSpace) const { assert(MI->mayLoad() && !MI->mayStore()); bool Changed = false; if ((AddrSpace & SIAtomicAddrSpace::GLOBAL) != SIAtomicAddrSpace::NONE) { /// TODO: Do not set glc for rmw atomic operations as they /// implicitly bypass the L1 cache. switch (Scope) { case SIAtomicScope::SYSTEM: case SIAtomicScope::AGENT: Changed |= enableGLCBit(MI); break; case SIAtomicScope::WORKGROUP: case SIAtomicScope::WAVEFRONT: case SIAtomicScope::SINGLETHREAD: // No cache to bypass. break; default: llvm_unreachable("Unsupported synchronization scope"); } } /// The scratch address space does not need the global memory caches /// to be bypassed as all memory operations by the same thread are /// sequentially consistent, and no other thread can access scratch /// memory. /// Other address spaces do not hava a cache. return Changed; } bool SIGfx6CacheControl::enableNonTemporal( const MachineBasicBlock::iterator &MI) const { assert(MI->mayLoad() ^ MI->mayStore()); bool Changed = false; /// TODO: Do not enableGLCBit if rmw atomic. Changed |= enableGLCBit(MI); Changed |= enableSLCBit(MI); return Changed; } bool SIGfx6CacheControl::insertCacheInvalidate(MachineBasicBlock::iterator &MI, SIAtomicScope Scope, SIAtomicAddrSpace AddrSpace, Position Pos) const { if (!InsertCacheInv) return false; bool Changed = false; MachineBasicBlock &MBB = *MI->getParent(); DebugLoc DL = MI->getDebugLoc(); if (Pos == Position::AFTER) ++MI; if ((AddrSpace & SIAtomicAddrSpace::GLOBAL) != SIAtomicAddrSpace::NONE) { switch (Scope) { case SIAtomicScope::SYSTEM: case SIAtomicScope::AGENT: BuildMI(MBB, MI, DL, TII->get(AMDGPU::BUFFER_WBINVL1)); Changed = true; break; case SIAtomicScope::WORKGROUP: case SIAtomicScope::WAVEFRONT: case SIAtomicScope::SINGLETHREAD: // No cache to invalidate. break; default: llvm_unreachable("Unsupported synchronization scope"); } } /// The scratch address space does not need the global memory cache /// to be flushed as all memory operations by the same thread are /// sequentially consistent, and no other thread can access scratch /// memory. /// Other address spaces do not hava a cache. if (Pos == Position::AFTER) --MI; return Changed; } bool SIGfx6CacheControl::insertWait(MachineBasicBlock::iterator &MI, SIAtomicScope Scope, SIAtomicAddrSpace AddrSpace, SIMemOp Op, bool IsCrossAddrSpaceOrdering, Position Pos) const { bool Changed = false; MachineBasicBlock &MBB = *MI->getParent(); DebugLoc DL = MI->getDebugLoc(); if (Pos == Position::AFTER) ++MI; bool VMCnt = false; bool LGKMCnt = false; if ((AddrSpace & SIAtomicAddrSpace::GLOBAL) != SIAtomicAddrSpace::NONE) { switch (Scope) { case SIAtomicScope::SYSTEM: case SIAtomicScope::AGENT: VMCnt |= true; break; case SIAtomicScope::WORKGROUP: case SIAtomicScope::WAVEFRONT: case SIAtomicScope::SINGLETHREAD: // The L1 cache keeps all memory operations in order for // wavefronts in the same work-group. break; default: llvm_unreachable("Unsupported synchronization scope"); } } if ((AddrSpace & SIAtomicAddrSpace::LDS) != SIAtomicAddrSpace::NONE) { switch (Scope) { case SIAtomicScope::SYSTEM: case SIAtomicScope::AGENT: case SIAtomicScope::WORKGROUP: // If no cross address space ordering then an LDS waitcnt is not // needed as LDS operations for all waves are executed in a // total global ordering as observed by all waves. Required if // also synchronizing with global/GDS memory as LDS operations // could be reordered with respect to later global/GDS memory // operations of the same wave. LGKMCnt |= IsCrossAddrSpaceOrdering; break; case SIAtomicScope::WAVEFRONT: case SIAtomicScope::SINGLETHREAD: // The LDS keeps all memory operations in order for // the same wavesfront. break; default: llvm_unreachable("Unsupported synchronization scope"); } } if ((AddrSpace & SIAtomicAddrSpace::GDS) != SIAtomicAddrSpace::NONE) { switch (Scope) { case SIAtomicScope::SYSTEM: case SIAtomicScope::AGENT: // If no cross address space ordering then an GDS waitcnt is not // needed as GDS operations for all waves are executed in a // total global ordering as observed by all waves. Required if // also synchronizing with global/LDS memory as GDS operations // could be reordered with respect to later global/LDS memory // operations of the same wave. LGKMCnt |= IsCrossAddrSpaceOrdering; break; case SIAtomicScope::WORKGROUP: case SIAtomicScope::WAVEFRONT: case SIAtomicScope::SINGLETHREAD: // The GDS keeps all memory operations in order for // the same work-group. break; default: llvm_unreachable("Unsupported synchronization scope"); } } if (VMCnt || LGKMCnt) { unsigned WaitCntImmediate = AMDGPU::encodeWaitcnt(IV, VMCnt ? 0 : getVmcntBitMask(IV), getExpcntBitMask(IV), LGKMCnt ? 0 : getLgkmcntBitMask(IV)); BuildMI(MBB, MI, DL, TII->get(AMDGPU::S_WAITCNT)).addImm(WaitCntImmediate); Changed = true; } if (Pos == Position::AFTER) --MI; return Changed; } bool SIGfx7CacheControl::insertCacheInvalidate(MachineBasicBlock::iterator &MI, SIAtomicScope Scope, SIAtomicAddrSpace AddrSpace, Position Pos) const { if (!InsertCacheInv) return false; bool Changed = false; MachineBasicBlock &MBB = *MI->getParent(); DebugLoc DL = MI->getDebugLoc(); const GCNSubtarget &STM = MBB.getParent()->getSubtarget<GCNSubtarget>(); const unsigned Flush = STM.isAmdPalOS() || STM.isMesa3DOS() ? AMDGPU::BUFFER_WBINVL1 : AMDGPU::BUFFER_WBINVL1_VOL; if (Pos == Position::AFTER) ++MI; if ((AddrSpace & SIAtomicAddrSpace::GLOBAL) != SIAtomicAddrSpace::NONE) { switch (Scope) { case SIAtomicScope::SYSTEM: case SIAtomicScope::AGENT: BuildMI(MBB, MI, DL, TII->get(Flush)); Changed = true; break; case SIAtomicScope::WORKGROUP: case SIAtomicScope::WAVEFRONT: case SIAtomicScope::SINGLETHREAD: // No cache to invalidate. break; default: llvm_unreachable("Unsupported synchronization scope"); } } /// The scratch address space does not need the global memory cache /// to be flushed as all memory operations by the same thread are /// sequentially consistent, and no other thread can access scratch /// memory. /// Other address spaces do not hava a cache. if (Pos == Position::AFTER) --MI; return Changed; } bool SIGfx10CacheControl::enableLoadCacheBypass( const MachineBasicBlock::iterator &MI, SIAtomicScope Scope, SIAtomicAddrSpace AddrSpace) const { assert(MI->mayLoad() && !MI->mayStore()); bool Changed = false; if ((AddrSpace & SIAtomicAddrSpace::GLOBAL) != SIAtomicAddrSpace::NONE) { /// TODO Do not set glc for rmw atomic operations as they /// implicitly bypass the L0/L1 caches. switch (Scope) { case SIAtomicScope::SYSTEM: case SIAtomicScope::AGENT: Changed |= enableGLCBit(MI); Changed |= enableDLCBit(MI); break; case SIAtomicScope::WORKGROUP: // In WGP mode the waves of a work-group can be executing on either CU of // the WGP. Therefore need to bypass the L0 which is per CU. Otherwise in // CU mode and all waves of a work-group are on the same CU, and so the // L0 does not need to be bypassed. if (!CuMode) Changed |= enableGLCBit(MI); break; case SIAtomicScope::WAVEFRONT: case SIAtomicScope::SINGLETHREAD: // No cache to bypass. break; default: llvm_unreachable("Unsupported synchronization scope"); } } /// The scratch address space does not need the global memory caches /// to be bypassed as all memory operations by the same thread are /// sequentially consistent, and no other thread can access scratch /// memory. /// Other address spaces do not hava a cache. return Changed; } bool SIGfx10CacheControl::enableNonTemporal( const MachineBasicBlock::iterator &MI) const { assert(MI->mayLoad() ^ MI->mayStore()); bool Changed = false; Changed |= enableSLCBit(MI); /// TODO for store (non-rmw atomic) instructions also enableGLCBit(MI) return Changed; } bool SIGfx10CacheControl::insertCacheInvalidate(MachineBasicBlock::iterator &MI, SIAtomicScope Scope, SIAtomicAddrSpace AddrSpace, Position Pos) const { if (!InsertCacheInv) return false; bool Changed = false; MachineBasicBlock &MBB = *MI->getParent(); DebugLoc DL = MI->getDebugLoc(); if (Pos == Position::AFTER) ++MI; if ((AddrSpace & SIAtomicAddrSpace::GLOBAL) != SIAtomicAddrSpace::NONE) { switch (Scope) { case SIAtomicScope::SYSTEM: case SIAtomicScope::AGENT: BuildMI(MBB, MI, DL, TII->get(AMDGPU::BUFFER_GL0_INV)); BuildMI(MBB, MI, DL, TII->get(AMDGPU::BUFFER_GL1_INV)); Changed = true; break; case SIAtomicScope::WORKGROUP: // In WGP mode the waves of a work-group can be executing on either CU of // the WGP. Therefore need to invalidate the L0 which is per CU. Otherwise // in CU mode and all waves of a work-group are on the same CU, and so the // L0 does not need to be invalidated. if (!CuMode) { BuildMI(MBB, MI, DL, TII->get(AMDGPU::BUFFER_GL0_INV)); Changed = true; } break; case SIAtomicScope::WAVEFRONT: case SIAtomicScope::SINGLETHREAD: // No cache to invalidate. break; default: llvm_unreachable("Unsupported synchronization scope"); } } /// The scratch address space does not need the global memory cache /// to be flushed as all memory operations by the same thread are /// sequentially consistent, and no other thread can access scratch /// memory. /// Other address spaces do not hava a cache. if (Pos == Position::AFTER) --MI; return Changed; } bool SIGfx10CacheControl::insertWait(MachineBasicBlock::iterator &MI, SIAtomicScope Scope, SIAtomicAddrSpace AddrSpace, SIMemOp Op, bool IsCrossAddrSpaceOrdering, Position Pos) const { bool Changed = false; MachineBasicBlock &MBB = *MI->getParent(); DebugLoc DL = MI->getDebugLoc(); if (Pos == Position::AFTER) ++MI; bool VMCnt = false; bool VSCnt = false; bool LGKMCnt = false; if ((AddrSpace & SIAtomicAddrSpace::GLOBAL) != SIAtomicAddrSpace::NONE) { switch (Scope) { case SIAtomicScope::SYSTEM: case SIAtomicScope::AGENT: if ((Op & SIMemOp::LOAD) != SIMemOp::NONE) VMCnt |= true; if ((Op & SIMemOp::STORE) != SIMemOp::NONE) VSCnt |= true; break; case SIAtomicScope::WORKGROUP: // In WGP mode the waves of a work-group can be executing on either CU of // the WGP. Therefore need to wait for operations to complete to ensure // they are visible to waves in the other CU as the L0 is per CU. // Otherwise in CU mode and all waves of a work-group are on the same CU // which shares the same L0. if (!CuMode) { if ((Op & SIMemOp::LOAD) != SIMemOp::NONE) VMCnt |= true; if ((Op & SIMemOp::STORE) != SIMemOp::NONE) VSCnt |= true; } break; case SIAtomicScope::WAVEFRONT: case SIAtomicScope::SINGLETHREAD: // The L0 cache keeps all memory operations in order for // work-items in the same wavefront. break; default: llvm_unreachable("Unsupported synchronization scope"); } } if ((AddrSpace & SIAtomicAddrSpace::LDS) != SIAtomicAddrSpace::NONE) { switch (Scope) { case SIAtomicScope::SYSTEM: case SIAtomicScope::AGENT: case SIAtomicScope::WORKGROUP: // If no cross address space ordering then an LDS waitcnt is not // needed as LDS operations for all waves are executed in a // total global ordering as observed by all waves. Required if // also synchronizing with global/GDS memory as LDS operations // could be reordered with respect to later global/GDS memory // operations of the same wave. LGKMCnt |= IsCrossAddrSpaceOrdering; break; case SIAtomicScope::WAVEFRONT: case SIAtomicScope::SINGLETHREAD: // The LDS keeps all memory operations in order for // the same wavesfront. break; default: llvm_unreachable("Unsupported synchronization scope"); } } if ((AddrSpace & SIAtomicAddrSpace::GDS) != SIAtomicAddrSpace::NONE) { switch (Scope) { case SIAtomicScope::SYSTEM: case SIAtomicScope::AGENT: // If no cross address space ordering then an GDS waitcnt is not // needed as GDS operations for all waves are executed in a // total global ordering as observed by all waves. Required if // also synchronizing with global/LDS memory as GDS operations // could be reordered with respect to later global/LDS memory // operations of the same wave. LGKMCnt |= IsCrossAddrSpaceOrdering; break; case SIAtomicScope::WORKGROUP: case SIAtomicScope::WAVEFRONT: case SIAtomicScope::SINGLETHREAD: // The GDS keeps all memory operations in order for // the same work-group. break; default: llvm_unreachable("Unsupported synchronization scope"); } } if (VMCnt || LGKMCnt) { unsigned WaitCntImmediate = AMDGPU::encodeWaitcnt(IV, VMCnt ? 0 : getVmcntBitMask(IV), getExpcntBitMask(IV), LGKMCnt ? 0 : getLgkmcntBitMask(IV)); BuildMI(MBB, MI, DL, TII->get(AMDGPU::S_WAITCNT)).addImm(WaitCntImmediate); Changed = true; } if (VSCnt) { BuildMI(MBB, MI, DL, TII->get(AMDGPU::S_WAITCNT_VSCNT)) .addReg(AMDGPU::SGPR_NULL, RegState::Undef) .addImm(0); Changed = true; } if (Pos == Position::AFTER) --MI; return Changed; } bool SIMemoryLegalizer::removeAtomicPseudoMIs() { if (AtomicPseudoMIs.empty()) return false; for (auto &MI : AtomicPseudoMIs) MI->eraseFromParent(); AtomicPseudoMIs.clear(); return true; } bool SIMemoryLegalizer::expandLoad(const SIMemOpInfo &MOI, MachineBasicBlock::iterator &MI) { assert(MI->mayLoad() && !MI->mayStore()); bool Changed = false; if (MOI.isAtomic()) { if (MOI.getOrdering() == AtomicOrdering::Monotonic || MOI.getOrdering() == AtomicOrdering::Acquire || MOI.getOrdering() == AtomicOrdering::SequentiallyConsistent) { Changed |= CC->enableLoadCacheBypass(MI, MOI.getScope(), MOI.getOrderingAddrSpace()); } if (MOI.getOrdering() == AtomicOrdering::SequentiallyConsistent) Changed |= CC->insertWait(MI, MOI.getScope(), MOI.getOrderingAddrSpace(), SIMemOp::LOAD | SIMemOp::STORE, MOI.getIsCrossAddressSpaceOrdering(), Position::BEFORE); if (MOI.getOrdering() == AtomicOrdering::Acquire || MOI.getOrdering() == AtomicOrdering::SequentiallyConsistent) { Changed |= CC->insertWait(MI, MOI.getScope(), MOI.getInstrAddrSpace(), SIMemOp::LOAD, MOI.getIsCrossAddressSpaceOrdering(), Position::AFTER); Changed |= CC->insertCacheInvalidate(MI, MOI.getScope(), MOI.getOrderingAddrSpace(), Position::AFTER); } return Changed; } // Atomic instructions do not have the nontemporal attribute. if (MOI.isNonTemporal()) { Changed |= CC->enableNonTemporal(MI); return Changed; } return Changed; } bool SIMemoryLegalizer::expandStore(const SIMemOpInfo &MOI, MachineBasicBlock::iterator &MI) { assert(!MI->mayLoad() && MI->mayStore()); bool Changed = false; if (MOI.isAtomic()) { if (MOI.getOrdering() == AtomicOrdering::Release || MOI.getOrdering() == AtomicOrdering::SequentiallyConsistent) Changed |= CC->insertWait(MI, MOI.getScope(), MOI.getOrderingAddrSpace(), SIMemOp::LOAD | SIMemOp::STORE, MOI.getIsCrossAddressSpaceOrdering(), Position::BEFORE); return Changed; } // Atomic instructions do not have the nontemporal attribute. if (MOI.isNonTemporal()) { Changed |= CC->enableNonTemporal(MI); return Changed; } return Changed; } bool SIMemoryLegalizer::expandAtomicFence(const SIMemOpInfo &MOI, MachineBasicBlock::iterator &MI) { assert(MI->getOpcode() == AMDGPU::ATOMIC_FENCE); AtomicPseudoMIs.push_back(MI); bool Changed = false; if (MOI.isAtomic()) { if (MOI.getOrdering() == AtomicOrdering::Acquire || MOI.getOrdering() == AtomicOrdering::Release || MOI.getOrdering() == AtomicOrdering::AcquireRelease || MOI.getOrdering() == AtomicOrdering::SequentiallyConsistent) /// TODO: This relies on a barrier always generating a waitcnt /// for LDS to ensure it is not reordered with the completion of /// the proceeding LDS operations. If barrier had a memory /// ordering and memory scope, then library does not need to /// generate a fence. Could add support in this file for /// barrier. SIInsertWaitcnt.cpp could then stop unconditionally /// adding waitcnt before a S_BARRIER. Changed |= CC->insertWait(MI, MOI.getScope(), MOI.getOrderingAddrSpace(), SIMemOp::LOAD | SIMemOp::STORE, MOI.getIsCrossAddressSpaceOrdering(), Position::BEFORE); if (MOI.getOrdering() == AtomicOrdering::Acquire || MOI.getOrdering() == AtomicOrdering::AcquireRelease || MOI.getOrdering() == AtomicOrdering::SequentiallyConsistent) Changed |= CC->insertCacheInvalidate(MI, MOI.getScope(), MOI.getOrderingAddrSpace(), Position::BEFORE); return Changed; } return Changed; } bool SIMemoryLegalizer::expandAtomicCmpxchgOrRmw(const SIMemOpInfo &MOI, MachineBasicBlock::iterator &MI) { assert(MI->mayLoad() && MI->mayStore()); bool Changed = false; if (MOI.isAtomic()) { if (MOI.getOrdering() == AtomicOrdering::Release || MOI.getOrdering() == AtomicOrdering::AcquireRelease || MOI.getOrdering() == AtomicOrdering::SequentiallyConsistent || MOI.getFailureOrdering() == AtomicOrdering::SequentiallyConsistent) Changed |= CC->insertWait(MI, MOI.getScope(), MOI.getOrderingAddrSpace(), SIMemOp::LOAD | SIMemOp::STORE, MOI.getIsCrossAddressSpaceOrdering(), Position::BEFORE); if (MOI.getOrdering() == AtomicOrdering::Acquire || MOI.getOrdering() == AtomicOrdering::AcquireRelease || MOI.getOrdering() == AtomicOrdering::SequentiallyConsistent || MOI.getFailureOrdering() == AtomicOrdering::Acquire || MOI.getFailureOrdering() == AtomicOrdering::SequentiallyConsistent) { Changed |= CC->insertWait(MI, MOI.getScope(), MOI.getOrderingAddrSpace(), isAtomicRet(*MI) ? SIMemOp::LOAD : SIMemOp::STORE, MOI.getIsCrossAddressSpaceOrdering(), Position::AFTER); Changed |= CC->insertCacheInvalidate(MI, MOI.getScope(), MOI.getOrderingAddrSpace(), Position::AFTER); } return Changed; } return Changed; } bool SIMemoryLegalizer::runOnMachineFunction(MachineFunction &MF) { bool Changed = false; SIMemOpAccess MOA(MF); CC = SICacheControl::create(MF.getSubtarget<GCNSubtarget>()); for (auto &MBB : MF) { for (auto MI = MBB.begin(); MI != MBB.end(); ++MI) { if (MI->getOpcode() == TargetOpcode::BUNDLE && MI->mayLoadOrStore()) { MachineBasicBlock::instr_iterator II(MI->getIterator()); for (MachineBasicBlock::instr_iterator I = ++II, E = MBB.instr_end(); I != E && I->isBundledWithPred(); ++I) { I->unbundleFromPred(); for (MachineOperand &MO : I->operands()) if (MO.isReg()) MO.setIsInternalRead(false); } MI->eraseFromParent(); MI = II->getIterator(); } if (!(MI->getDesc().TSFlags & SIInstrFlags::maybeAtomic)) continue; if (const auto &MOI = MOA.getLoadInfo(MI)) Changed |= expandLoad(MOI.getValue(), MI); else if (const auto &MOI = MOA.getStoreInfo(MI)) Changed |= expandStore(MOI.getValue(), MI); else if (const auto &MOI = MOA.getAtomicFenceInfo(MI)) Changed |= expandAtomicFence(MOI.getValue(), MI); else if (const auto &MOI = MOA.getAtomicCmpxchgOrRmwInfo(MI)) Changed |= expandAtomicCmpxchgOrRmw(MOI.getValue(), MI); } } Changed |= removeAtomicPseudoMIs(); return Changed; } INITIALIZE_PASS(SIMemoryLegalizer, DEBUG_TYPE, PASS_NAME, false, false) char SIMemoryLegalizer::ID = 0; char &llvm::SIMemoryLegalizerID = SIMemoryLegalizer::ID; FunctionPass *llvm::createSIMemoryLegalizerPass() { return new SIMemoryLegalizer(); }
Upload File
Create Folder