//===-- AMDGPULowerKernelAttributes.cpp ------------------------------------------===// // // The LLVM Compiler Infrastructure // // This file is distributed under the University of Illinois Open Source // License. See LICENSE.TXT for details. // //===----------------------------------------------------------------------===// // /// \file This pass does attempts to make use of reqd_work_group_size metadata /// to eliminate loads from the dispatch packet and to constant fold OpenCL /// get_local_size-like functions. // //===----------------------------------------------------------------------===// #include "AMDGPU.h" #include "AMDGPUTargetMachine.h" #include "llvm/Analysis/ValueTracking.h" #include "llvm/CodeGen/Passes.h" #include "llvm/CodeGen/TargetPassConfig.h" #include "llvm/IR/Constants.h" #include "llvm/IR/Function.h" #include "llvm/IR/Instructions.h" #include "llvm/IR/PatternMatch.h" #include "llvm/Pass.h" #define DEBUG_TYPE "amdgpu-lower-kernel-attributes" using namespace llvm; namespace { // Field offsets in hsa_kernel_dispatch_packet_t. enum DispatchPackedOffsets { WORKGROUP_SIZE_X = 4, WORKGROUP_SIZE_Y = 6, WORKGROUP_SIZE_Z = 8, GRID_SIZE_X = 12, GRID_SIZE_Y = 16, GRID_SIZE_Z = 20 }; class AMDGPULowerKernelAttributes : public ModulePass { Module *Mod = nullptr; public: static char ID; AMDGPULowerKernelAttributes() : ModulePass(ID) {} bool processUse(CallInst *CI); bool doInitialization(Module &M) override; bool runOnModule(Module &M) override; StringRef getPassName() const override { return "AMDGPU Kernel Attributes"; } void getAnalysisUsage(AnalysisUsage &AU) const override { AU.setPreservesAll(); } }; } // end anonymous namespace bool AMDGPULowerKernelAttributes::doInitialization(Module &M) { Mod = &M; return false; } bool AMDGPULowerKernelAttributes::processUse(CallInst *CI) { Function *F = CI->getParent()->getParent(); auto MD = F->getMetadata("reqd_work_group_size"); const bool HasReqdWorkGroupSize = MD && MD->getNumOperands() == 3; const bool HasUniformWorkGroupSize = F->getFnAttribute("uniform-work-group-size").getValueAsString() == "true"; if (!HasReqdWorkGroupSize && !HasUniformWorkGroupSize) return false; Value *WorkGroupSizeX = nullptr; Value *WorkGroupSizeY = nullptr; Value *WorkGroupSizeZ = nullptr; Value *GridSizeX = nullptr; Value *GridSizeY = nullptr; Value *GridSizeZ = nullptr; const DataLayout &DL = Mod->getDataLayout(); // We expect to see several GEP users, casted to the appropriate type and // loaded. for (User *U : CI->users()) { if (!U->hasOneUse()) continue; int64_t Offset = 0; if (GetPointerBaseWithConstantOffset(U, Offset, DL) != CI) continue; auto *BCI = dyn_cast<BitCastInst>(*U->user_begin()); if (!BCI || !BCI->hasOneUse()) continue; auto *Load = dyn_cast<LoadInst>(*BCI->user_begin()); if (!Load || !Load->isSimple()) continue; unsigned LoadSize = DL.getTypeStoreSize(Load->getType()); // TODO: Handle merged loads. switch (Offset) { case WORKGROUP_SIZE_X: if (LoadSize == 2) WorkGroupSizeX = Load; break; case WORKGROUP_SIZE_Y: if (LoadSize == 2) WorkGroupSizeY = Load; break; case WORKGROUP_SIZE_Z: if (LoadSize == 2) WorkGroupSizeZ = Load; break; case GRID_SIZE_X: if (LoadSize == 4) GridSizeX = Load; break; case GRID_SIZE_Y: if (LoadSize == 4) GridSizeY = Load; break; case GRID_SIZE_Z: if (LoadSize == 4) GridSizeZ = Load; break; default: break; } } // Pattern match the code used to handle partial workgroup dispatches in the // library implementation of get_local_size, so the entire function can be // constant folded with a known group size. // // uint r = grid_size - group_id * group_size; // get_local_size = (r < group_size) ? r : group_size; // // If we have uniform-work-group-size (which is the default in OpenCL 1.2), // the grid_size is required to be a multiple of group_size). In this case: // // grid_size - (group_id * group_size) < group_size // -> // grid_size < group_size + (group_id * group_size) // // (grid_size / group_size) < 1 + group_id // // grid_size / group_size is at least 1, so we can conclude the select // condition is false (except for group_id == 0, where the select result is // the same). bool MadeChange = false; Value *WorkGroupSizes[3] = { WorkGroupSizeX, WorkGroupSizeY, WorkGroupSizeZ }; Value *GridSizes[3] = { GridSizeX, GridSizeY, GridSizeZ }; for (int I = 0; HasUniformWorkGroupSize && I < 3; ++I) { Value *GroupSize = WorkGroupSizes[I]; Value *GridSize = GridSizes[I]; if (!GroupSize || !GridSize) continue; for (User *U : GroupSize->users()) { auto *ZextGroupSize = dyn_cast<ZExtInst>(U); if (!ZextGroupSize) continue; for (User *ZextUser : ZextGroupSize->users()) { auto *SI = dyn_cast<SelectInst>(ZextUser); if (!SI) continue; using namespace llvm::PatternMatch; auto GroupIDIntrin = I == 0 ? m_Intrinsic<Intrinsic::amdgcn_workgroup_id_x>() : (I == 1 ? m_Intrinsic<Intrinsic::amdgcn_workgroup_id_y>() : m_Intrinsic<Intrinsic::amdgcn_workgroup_id_z>()); auto SubExpr = m_Sub(m_Specific(GridSize), m_Mul(GroupIDIntrin, m_Specific(ZextGroupSize))); ICmpInst::Predicate Pred; if (match(SI, m_Select(m_ICmp(Pred, SubExpr, m_Specific(ZextGroupSize)), SubExpr, m_Specific(ZextGroupSize))) && Pred == ICmpInst::ICMP_ULT) { if (HasReqdWorkGroupSize) { ConstantInt *KnownSize = mdconst::extract<ConstantInt>(MD->getOperand(I)); SI->replaceAllUsesWith(ConstantExpr::getIntegerCast(KnownSize, SI->getType(), false)); } else { SI->replaceAllUsesWith(ZextGroupSize); } MadeChange = true; } } } } if (!HasReqdWorkGroupSize) return MadeChange; // Eliminate any other loads we can from the dispatch packet. for (int I = 0; I < 3; ++I) { Value *GroupSize = WorkGroupSizes[I]; if (!GroupSize) continue; ConstantInt *KnownSize = mdconst::extract<ConstantInt>(MD->getOperand(I)); GroupSize->replaceAllUsesWith( ConstantExpr::getIntegerCast(KnownSize, GroupSize->getType(), false)); MadeChange = true; } return MadeChange; } // TODO: Move makeLIDRangeMetadata usage into here. Seem to not get // TargetPassConfig for subtarget. bool AMDGPULowerKernelAttributes::runOnModule(Module &M) { StringRef DispatchPtrName = Intrinsic::getName(Intrinsic::amdgcn_dispatch_ptr); Function *DispatchPtr = Mod->getFunction(DispatchPtrName); if (!DispatchPtr) // Dispatch ptr not used. return false; bool MadeChange = false; SmallPtrSet<Instruction *, 4> HandledUses; for (auto *U : DispatchPtr->users()) { CallInst *CI = cast<CallInst>(U); if (HandledUses.insert(CI).second) { if (processUse(CI)) MadeChange = true; } } return MadeChange; } INITIALIZE_PASS_BEGIN(AMDGPULowerKernelAttributes, DEBUG_TYPE, "AMDGPU IR optimizations", false, false) INITIALIZE_PASS_END(AMDGPULowerKernelAttributes, DEBUG_TYPE, "AMDGPU IR optimizations", false, false) char AMDGPULowerKernelAttributes::ID = 0; ModulePass *llvm::createAMDGPULowerKernelAttributesPass() { return new AMDGPULowerKernelAttributes(); }