//===- FunctionSpecialization.cpp - Function Specialization ---------------===// // // Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. // See https://llvm.org/LICENSE.txt for license information. // SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception // //===----------------------------------------------------------------------===// // // This specialises functions with constant parameters. Constant parameters // like function pointers and constant globals are propagated to the callee by // specializing the function. The main benefit of this pass at the moment is // that indirect calls are transformed into direct calls, which provides inline // opportunities that the inliner would not have been able to achieve. That's // why function specialisation is run before the inliner in the optimisation // pipeline; that is by design. Otherwise, we would only benefit from constant // passing, which is a valid use-case too, but hasn't been explored much in // terms of performance uplifts, cost-model and compile-time impact. // // Current limitations: // - It does not yet handle integer ranges. We do support "literal constants", // but that's off by default under an option. // - The cost-model could be further looked into (it mainly focuses on inlining // benefits), // // Ideas: // - With a function specialization attribute for arguments, we could have // a direct way to steer function specialization, avoiding the cost-model, // and thus control compile-times / code-size. // // Todos: // - Specializing recursive functions relies on running the transformation a // number of times, which is controlled by option // `func-specialization-max-iters`. Thus, increasing this value and the // number of iterations, will linearly increase the number of times recursive // functions get specialized, see also the discussion in // https://reviews.llvm.org/D106426 for details. Perhaps there is a // compile-time friendlier way to control/limit the number of specialisations // for recursive functions. // - Don't transform the function if function specialization does not trigger; // the SCCPSolver may make IR changes. // // References: // - 2021 LLVM Dev Mtg “Introducing function specialisation, and can we enable // it by default?”, https://www.youtube.com/watch?v=zJiCjeXgV5Q // //===----------------------------------------------------------------------===// #include "llvm/ADT/Statistic.h" #include "llvm/Analysis/CodeMetrics.h" #include "llvm/Analysis/InlineCost.h" #include "llvm/Analysis/LoopInfo.h" #include "llvm/Analysis/TargetTransformInfo.h" #include "llvm/Analysis/ValueLattice.h" #include "llvm/Analysis/ValueLatticeUtils.h" #include "llvm/IR/IntrinsicInst.h" #include "llvm/Transforms/Scalar/SCCP.h" #include "llvm/Transforms/Utils/Cloning.h" #include "llvm/Transforms/Utils/SCCPSolver.h" #include "llvm/Transforms/Utils/SizeOpts.h" #include using namespace llvm; #define DEBUG_TYPE "function-specialization" STATISTIC(NumFuncSpecialized, "Number of functions specialized"); static cl::opt ForceFunctionSpecialization( "force-function-specialization", cl::init(false), cl::Hidden, cl::desc("Force function specialization for every call site with a " "constant argument")); static cl::opt FuncSpecializationMaxIters( "func-specialization-max-iters", cl::Hidden, cl::desc("The maximum number of iterations function specialization is run"), cl::init(1)); static cl::opt MaxClonesThreshold( "func-specialization-max-clones", cl::Hidden, cl::desc("The maximum number of clones allowed for a single function " "specialization"), cl::init(3)); static cl::opt SmallFunctionThreshold( "func-specialization-size-threshold", cl::Hidden, cl::desc("Don't specialize functions that have less than this theshold " "number of instructions"), cl::init(100)); static cl::opt AvgLoopIterationCount("func-specialization-avg-iters-cost", cl::Hidden, cl::desc("Average loop iteration count cost"), cl::init(10)); static cl::opt SpecializeOnAddresses( "func-specialization-on-address", cl::init(false), cl::Hidden, cl::desc("Enable function specialization on the address of global values")); // Disabled by default as it can significantly increase compilation times. // Running nikic's compile time tracker on x86 with instruction count as the // metric shows 3-4% regression for SPASS while being neutral for all other // benchmarks of the llvm test suite. // // https://llvm-compile-time-tracker.com // https://github.com/nikic/llvm-compile-time-tracker static cl::opt EnableSpecializationForLiteralConstant( "function-specialization-for-literal-constant", cl::init(false), cl::Hidden, cl::desc("Enable specialization of functions that take a literal constant " "as an argument.")); namespace { // Bookkeeping struct to pass data from the analysis and profitability phase // to the actual transform helper functions. struct SpecializationInfo { SmallVector Args; // Stores the {formal,actual} argument pairs. InstructionCost Gain; // Profitability: Gain = Bonus - Cost. }; } // Anonymous namespace using FuncList = SmallVectorImpl; using CallArgBinding = std::pair; using CallSpecBinding = std::pair; // We are using MapVector because it guarantees deterministic iteration // order across executions. using SpecializationMap = SmallMapVector; // Helper to check if \p LV is either a constant or a constant // range with a single element. This should cover exactly the same cases as the // old ValueLatticeElement::isConstant() and is intended to be used in the // transition to ValueLatticeElement. static bool isConstant(const ValueLatticeElement &LV) { return LV.isConstant() || (LV.isConstantRange() && LV.getConstantRange().isSingleElement()); } // Helper to check if \p LV is either overdefined or a constant int. static bool isOverdefined(const ValueLatticeElement &LV) { return !LV.isUnknownOrUndef() && !isConstant(LV); } static Constant *getPromotableAlloca(AllocaInst *Alloca, CallInst *Call) { Value *StoreValue = nullptr; for (auto *User : Alloca->users()) { // We can't use llvm::isAllocaPromotable() as that would fail because of // the usage in the CallInst, which is what we check here. if (User == Call) continue; if (auto *Bitcast = dyn_cast(User)) { if (!Bitcast->hasOneUse() || *Bitcast->user_begin() != Call) return nullptr; continue; } if (auto *Store = dyn_cast(User)) { // This is a duplicate store, bail out. if (StoreValue || Store->isVolatile()) return nullptr; StoreValue = Store->getValueOperand(); continue; } // Bail if there is any other unknown usage. return nullptr; } return dyn_cast_or_null(StoreValue); } // A constant stack value is an AllocaInst that has a single constant // value stored to it. Return this constant if such an alloca stack value // is a function argument. static Constant *getConstantStackValue(CallInst *Call, Value *Val, SCCPSolver &Solver) { if (!Val) return nullptr; Val = Val->stripPointerCasts(); if (auto *ConstVal = dyn_cast(Val)) return ConstVal; auto *Alloca = dyn_cast(Val); if (!Alloca || !Alloca->getAllocatedType()->isIntegerTy()) return nullptr; return getPromotableAlloca(Alloca, Call); } // To support specializing recursive functions, it is important to propagate // constant arguments because after a first iteration of specialisation, a // reduced example may look like this: // // define internal void @RecursiveFn(i32* arg1) { // %temp = alloca i32, align 4 // store i32 2 i32* %temp, align 4 // call void @RecursiveFn.1(i32* nonnull %temp) // ret void // } // // Before a next iteration, we need to propagate the constant like so // which allows further specialization in next iterations. // // @funcspec.arg = internal constant i32 2 // // define internal void @someFunc(i32* arg1) { // call void @otherFunc(i32* nonnull @funcspec.arg) // ret void // } // static void constantArgPropagation(FuncList &WorkList, Module &M, SCCPSolver &Solver) { // Iterate over the argument tracked functions see if there // are any new constant values for the call instruction via // stack variables. for (auto *F : WorkList) { for (auto *User : F->users()) { auto *Call = dyn_cast(User); if (!Call) continue; bool Changed = false; for (const Use &U : Call->args()) { unsigned Idx = Call->getArgOperandNo(&U); Value *ArgOp = Call->getArgOperand(Idx); Type *ArgOpType = ArgOp->getType(); if (!Call->onlyReadsMemory(Idx) || !ArgOpType->isPointerTy()) continue; auto *ConstVal = getConstantStackValue(Call, ArgOp, Solver); if (!ConstVal) continue; Value *GV = new GlobalVariable(M, ConstVal->getType(), true, GlobalValue::InternalLinkage, ConstVal, "funcspec.arg"); if (ArgOpType != ConstVal->getType()) GV = ConstantExpr::getBitCast(cast(GV), ArgOpType); Call->setArgOperand(Idx, GV); Changed = true; } // Add the changed CallInst to Solver Worklist if (Changed) Solver.visitCall(*Call); } } } // ssa_copy intrinsics are introduced by the SCCP solver. These intrinsics // interfere with the constantArgPropagation optimization. static void removeSSACopy(Function &F) { for (BasicBlock &BB : F) { for (Instruction &Inst : llvm::make_early_inc_range(BB)) { auto *II = dyn_cast(&Inst); if (!II) continue; if (II->getIntrinsicID() != Intrinsic::ssa_copy) continue; Inst.replaceAllUsesWith(II->getOperand(0)); Inst.eraseFromParent(); } } } static void removeSSACopy(Module &M) { for (Function &F : M) removeSSACopy(F); } namespace { class FunctionSpecializer { /// The IPSCCP Solver. SCCPSolver &Solver; /// Analyses used to help determine if a function should be specialized. std::function GetAC; std::function GetTTI; std::function GetTLI; SmallPtrSet SpecializedFuncs; SmallPtrSet FullySpecialized; SmallVector ReplacedWithConstant; DenseMap FunctionMetrics; public: FunctionSpecializer(SCCPSolver &Solver, std::function GetAC, std::function GetTTI, std::function GetTLI) : Solver(Solver), GetAC(GetAC), GetTTI(GetTTI), GetTLI(GetTLI) {} ~FunctionSpecializer() { // Eliminate dead code. removeDeadInstructions(); removeDeadFunctions(); } /// Attempt to specialize functions in the module to enable constant /// propagation across function boundaries. /// /// \returns true if at least one function is specialized. bool specializeFunctions(FuncList &Candidates, FuncList &WorkList) { bool Changed = false; for (auto *F : Candidates) { if (!isCandidateFunction(F)) continue; auto Cost = getSpecializationCost(F); if (!Cost.isValid()) { LLVM_DEBUG( dbgs() << "FnSpecialization: Invalid specialization cost.\n"); continue; } LLVM_DEBUG(dbgs() << "FnSpecialization: Specialization cost for " << F->getName() << " is " << Cost << "\n"); SmallVector Specializations; if (!calculateGains(F, Cost, Specializations)) { LLVM_DEBUG(dbgs() << "FnSpecialization: No possible constants found\n"); continue; } Changed = true; for (auto &Entry : Specializations) specializeFunction(F, Entry.second, WorkList); } updateSpecializedFuncs(Candidates, WorkList); NumFuncSpecialized += NbFunctionsSpecialized; return Changed; } void removeDeadInstructions() { for (auto *I : ReplacedWithConstant) { LLVM_DEBUG(dbgs() << "FnSpecialization: Removing dead instruction " << *I << "\n"); I->eraseFromParent(); } ReplacedWithConstant.clear(); } void removeDeadFunctions() { for (auto *F : FullySpecialized) { LLVM_DEBUG(dbgs() << "FnSpecialization: Removing dead function " << F->getName() << "\n"); F->eraseFromParent(); } FullySpecialized.clear(); } bool tryToReplaceWithConstant(Value *V) { if (!V->getType()->isSingleValueType() || isa(V) || V->user_empty()) return false; const ValueLatticeElement &IV = Solver.getLatticeValueFor(V); if (isOverdefined(IV)) return false; auto *Const = isConstant(IV) ? Solver.getConstant(IV) : UndefValue::get(V->getType()); LLVM_DEBUG(dbgs() << "FnSpecialization: Replacing " << *V << "\nFnSpecialization: with " << *Const << "\n"); // Record uses of V to avoid visiting irrelevant uses of const later. SmallVector UseInsts; for (auto *U : V->users()) if (auto *I = dyn_cast(U)) if (Solver.isBlockExecutable(I->getParent())) UseInsts.push_back(I); V->replaceAllUsesWith(Const); for (auto *I : UseInsts) Solver.visit(I); // Remove the instruction from Block and Solver. if (auto *I = dyn_cast(V)) { if (I->isSafeToRemove()) { ReplacedWithConstant.push_back(I); Solver.removeLatticeValueFor(I); } } return true; } private: // The number of functions specialised, used for collecting statistics and // also in the cost model. unsigned NbFunctionsSpecialized = 0; // Compute the code metrics for function \p F. CodeMetrics &analyzeFunction(Function *F) { auto I = FunctionMetrics.insert({F, CodeMetrics()}); CodeMetrics &Metrics = I.first->second; if (I.second) { // The code metrics were not cached. SmallPtrSet EphValues; CodeMetrics::collectEphemeralValues(F, &(GetAC)(*F), EphValues); for (BasicBlock &BB : *F) Metrics.analyzeBasicBlock(&BB, (GetTTI)(*F), EphValues); LLVM_DEBUG(dbgs() << "FnSpecialization: Code size of function " << F->getName() << " is " << Metrics.NumInsts << " instructions\n"); } return Metrics; } /// Clone the function \p F and remove the ssa_copy intrinsics added by /// the SCCPSolver in the cloned version. Function *cloneCandidateFunction(Function *F, ValueToValueMapTy &Mappings) { Function *Clone = CloneFunction(F, Mappings); removeSSACopy(*Clone); return Clone; } /// This function decides whether it's worthwhile to specialize function /// \p F based on the known constant values its arguments can take on. It /// only discovers potential specialization opportunities without actually /// applying them. /// /// \returns true if any specializations have been found. bool calculateGains(Function *F, InstructionCost Cost, SmallVectorImpl &WorkList) { SpecializationMap Specializations; // Determine if we should specialize the function based on the values the // argument can take on. If specialization is not profitable, we continue // on to the next argument. for (Argument &FormalArg : F->args()) { // Determine if this argument is interesting. If we know the argument can // take on any constant values, they are collected in Constants. SmallVector ActualArgs; if (!isArgumentInteresting(&FormalArg, ActualArgs)) { LLVM_DEBUG(dbgs() << "FnSpecialization: Argument " << FormalArg.getNameOrAsOperand() << " is not interesting\n"); continue; } for (const auto &Entry : ActualArgs) { CallBase *Call = Entry.first; Constant *ActualArg = Entry.second; auto I = Specializations.insert({Call, SpecializationInfo()}); SpecializationInfo &S = I.first->second; if (I.second) S.Gain = ForceFunctionSpecialization ? 1 : 0 - Cost; if (!ForceFunctionSpecialization) S.Gain += getSpecializationBonus(&FormalArg, ActualArg); S.Args.push_back({&FormalArg, ActualArg}); } } // Remove unprofitable specializations. Specializations.remove_if( [](const auto &Entry) { return Entry.second.Gain <= 0; }); // Clear the MapVector and return the underlying vector. WorkList = Specializations.takeVector(); // Sort the candidates in descending order. llvm::stable_sort(WorkList, [](const auto &L, const auto &R) { return L.second.Gain > R.second.Gain; }); // Truncate the worklist to 'MaxClonesThreshold' candidates if necessary. if (WorkList.size() > MaxClonesThreshold) { LLVM_DEBUG(dbgs() << "FnSpecialization: Number of candidates exceed " << "the maximum number of clones threshold.\n" << "FnSpecialization: Truncating worklist to " << MaxClonesThreshold << " candidates.\n"); WorkList.erase(WorkList.begin() + MaxClonesThreshold, WorkList.end()); } LLVM_DEBUG(dbgs() << "FnSpecialization: Specializations for function " << F->getName() << "\n"; for (const auto &Entry : WorkList) { dbgs() << "FnSpecialization: Gain = " << Entry.second.Gain << "\n"; for (const ArgInfo &Arg : Entry.second.Args) dbgs() << "FnSpecialization: FormalArg = " << Arg.Formal->getNameOrAsOperand() << ", ActualArg = " << Arg.Actual->getNameOrAsOperand() << "\n"; }); return !WorkList.empty(); } bool isCandidateFunction(Function *F) { // Do not specialize the cloned function again. if (SpecializedFuncs.contains(F)) return false; // If we're optimizing the function for size, we shouldn't specialize it. if (F->hasOptSize() || shouldOptimizeForSize(F, nullptr, nullptr, PGSOQueryType::IRPass)) return false; // Exit if the function is not executable. There's no point in specializing // a dead function. if (!Solver.isBlockExecutable(&F->getEntryBlock())) return false; // It wastes time to specialize a function which would get inlined finally. if (F->hasFnAttribute(Attribute::AlwaysInline)) return false; LLVM_DEBUG(dbgs() << "FnSpecialization: Try function: " << F->getName() << "\n"); return true; } void specializeFunction(Function *F, SpecializationInfo &S, FuncList &WorkList) { ValueToValueMapTy Mappings; Function *Clone = cloneCandidateFunction(F, Mappings); // Rewrite calls to the function so that they call the clone instead. rewriteCallSites(Clone, S.Args, Mappings); // Initialize the lattice state of the arguments of the function clone, // marking the argument on which we specialized the function constant // with the given value. Solver.markArgInFuncSpecialization(Clone, S.Args); // Mark all the specialized functions WorkList.push_back(Clone); NbFunctionsSpecialized++; // If the function has been completely specialized, the original function // is no longer needed. Mark it unreachable. if (F->getNumUses() == 0 || all_of(F->users(), [F](User *U) { if (auto *CS = dyn_cast(U)) return CS->getFunction() == F; return false; })) { Solver.markFunctionUnreachable(F); FullySpecialized.insert(F); } } /// Compute and return the cost of specializing function \p F. InstructionCost getSpecializationCost(Function *F) { CodeMetrics &Metrics = analyzeFunction(F); // If the code metrics reveal that we shouldn't duplicate the function, we // shouldn't specialize it. Set the specialization cost to Invalid. // Or if the lines of codes implies that this function is easy to get // inlined so that we shouldn't specialize it. if (Metrics.notDuplicatable || !Metrics.NumInsts.isValid() || (!ForceFunctionSpecialization && *Metrics.NumInsts.getValue() < SmallFunctionThreshold)) { InstructionCost C{}; C.setInvalid(); return C; } // Otherwise, set the specialization cost to be the cost of all the // instructions in the function and penalty for specializing more functions. unsigned Penalty = NbFunctionsSpecialized + 1; return Metrics.NumInsts * InlineConstants::InstrCost * Penalty; } InstructionCost getUserBonus(User *U, llvm::TargetTransformInfo &TTI, LoopInfo &LI) { auto *I = dyn_cast_or_null(U); // If not an instruction we do not know how to evaluate. // Keep minimum possible cost for now so that it doesnt affect // specialization. if (!I) return std::numeric_limits::min(); auto Cost = TTI.getUserCost(U, TargetTransformInfo::TCK_SizeAndLatency); // Traverse recursively if there are more uses. // TODO: Any other instructions to be added here? if (I->mayReadFromMemory() || I->isCast()) for (auto *User : I->users()) Cost += getUserBonus(User, TTI, LI); // Increase the cost if it is inside the loop. auto LoopDepth = LI.getLoopDepth(I->getParent()); Cost *= std::pow((double)AvgLoopIterationCount, LoopDepth); return Cost; } /// Compute a bonus for replacing argument \p A with constant \p C. InstructionCost getSpecializationBonus(Argument *A, Constant *C) { Function *F = A->getParent(); DominatorTree DT(*F); LoopInfo LI(DT); auto &TTI = (GetTTI)(*F); LLVM_DEBUG(dbgs() << "FnSpecialization: Analysing bonus for constant: " << C->getNameOrAsOperand() << "\n"); InstructionCost TotalCost = 0; for (auto *U : A->users()) { TotalCost += getUserBonus(U, TTI, LI); LLVM_DEBUG(dbgs() << "FnSpecialization: User cost "; TotalCost.print(dbgs()); dbgs() << " for: " << *U << "\n"); } // The below heuristic is only concerned with exposing inlining // opportunities via indirect call promotion. If the argument is not a // (potentially casted) function pointer, give up. Function *CalledFunction = dyn_cast(C->stripPointerCasts()); if (!CalledFunction) return TotalCost; // Get TTI for the called function (used for the inline cost). auto &CalleeTTI = (GetTTI)(*CalledFunction); // Look at all the call sites whose called value is the argument. // Specializing the function on the argument would allow these indirect // calls to be promoted to direct calls. If the indirect call promotion // would likely enable the called function to be inlined, specializing is a // good idea. int Bonus = 0; for (User *U : A->users()) { if (!isa(U) && !isa(U)) continue; auto *CS = cast(U); if (CS->getCalledOperand() != A) continue; // Get the cost of inlining the called function at this call site. Note // that this is only an estimate. The called function may eventually // change in a way that leads to it not being inlined here, even though // inlining looks profitable now. For example, one of its called // functions may be inlined into it, making the called function too large // to be inlined into this call site. // // We apply a boost for performing indirect call promotion by increasing // the default threshold by the threshold for indirect calls. auto Params = getInlineParams(); Params.DefaultThreshold += InlineConstants::IndirectCallThreshold; InlineCost IC = getInlineCost(*CS, CalledFunction, Params, CalleeTTI, GetAC, GetTLI); // We clamp the bonus for this call to be between zero and the default // threshold. if (IC.isAlways()) Bonus += Params.DefaultThreshold; else if (IC.isVariable() && IC.getCostDelta() > 0) Bonus += IC.getCostDelta(); LLVM_DEBUG(dbgs() << "FnSpecialization: Inlining bonus " << Bonus << " for user " << *U << "\n"); } return TotalCost + Bonus; } /// Determine if we should specialize a function based on the incoming values /// of the given argument. /// /// This function implements the goal-directed heuristic. It determines if /// specializing the function based on the incoming values of argument \p A /// would result in any significant optimization opportunities. If /// optimization opportunities exist, the constant values of \p A on which to /// specialize the function are collected in \p Constants. /// /// \returns true if the function should be specialized on the given /// argument. bool isArgumentInteresting(Argument *A, SmallVectorImpl &Constants) { // For now, don't attempt to specialize functions based on the values of // composite types. if (!A->getType()->isSingleValueType() || A->user_empty()) return false; // If the argument isn't overdefined, there's nothing to do. It should // already be constant. if (!Solver.getLatticeValueFor(A).isOverdefined()) { LLVM_DEBUG(dbgs() << "FnSpecialization: Nothing to do, argument " << A->getNameOrAsOperand() << " is already constant?\n"); return false; } // Collect the constant values that the argument can take on. If the // argument can't take on any constant values, we aren't going to // specialize the function. While it's possible to specialize the function // based on non-constant arguments, there's likely not much benefit to // constant propagation in doing so. // // TODO 1: currently it won't specialize if there are over the threshold of // calls using the same argument, e.g foo(a) x 4 and foo(b) x 1, but it // might be beneficial to take the occurrences into account in the cost // model, so we would need to find the unique constants. // // TODO 2: this currently does not support constants, i.e. integer ranges. // getPossibleConstants(A, Constants); if (Constants.empty()) return false; LLVM_DEBUG(dbgs() << "FnSpecialization: Found interesting argument " << A->getNameOrAsOperand() << "\n"); return true; } /// Collect in \p Constants all the constant values that argument \p A can /// take on. void getPossibleConstants(Argument *A, SmallVectorImpl &Constants) { Function *F = A->getParent(); // SCCP solver does not record an argument that will be constructed on // stack. if (A->hasByValAttr() && !F->onlyReadsMemory()) return; // Iterate over all the call sites of the argument's parent function. for (User *U : F->users()) { if (!isa(U) && !isa(U)) continue; auto &CS = *cast(U); // If the call site has attribute minsize set, that callsite won't be // specialized. if (CS.hasFnAttr(Attribute::MinSize)) continue; // If the parent of the call site will never be executed, we don't need // to worry about the passed value. if (!Solver.isBlockExecutable(CS.getParent())) continue; auto *V = CS.getArgOperand(A->getArgNo()); if (isa(V)) return; // TrackValueOfGlobalVariable only tracks scalar global variables. if (auto *GV = dyn_cast(V)) { // Check if we want to specialize on the address of non-constant // global values. if (!GV->isConstant()) if (!SpecializeOnAddresses) return; if (!GV->getValueType()->isSingleValueType()) return; } if (isa(V) && (Solver.getLatticeValueFor(V).isConstant() || EnableSpecializationForLiteralConstant)) Constants.push_back({&CS, cast(V)}); } } /// Rewrite calls to function \p F to call function \p Clone instead. /// /// This function modifies calls to function \p F as long as the actual /// arguments match those in \p Args. Note that for recursive calls we /// need to compare against the cloned formal arguments. /// /// Callsites that have been marked with the MinSize function attribute won't /// be specialized and rewritten. void rewriteCallSites(Function *Clone, const SmallVectorImpl &Args, ValueToValueMapTy &Mappings) { assert(!Args.empty() && "Specialization without arguments"); Function *F = Args[0].Formal->getParent(); SmallVector CallSitesToRewrite; for (auto *U : F->users()) { if (!isa(U) && !isa(U)) continue; auto &CS = *cast(U); if (!CS.getCalledFunction() || CS.getCalledFunction() != F) continue; CallSitesToRewrite.push_back(&CS); } LLVM_DEBUG(dbgs() << "FnSpecialization: Replacing call sites of " << F->getName() << " with " << Clone->getName() << "\n"); for (auto *CS : CallSitesToRewrite) { LLVM_DEBUG(dbgs() << "FnSpecialization: " << CS->getFunction()->getName() << " ->" << *CS << "\n"); if (/* recursive call */ (CS->getFunction() == Clone && all_of(Args, [CS, &Mappings](const ArgInfo &Arg) { unsigned ArgNo = Arg.Formal->getArgNo(); return CS->getArgOperand(ArgNo) == Mappings[Arg.Formal]; })) || /* normal call */ all_of(Args, [CS](const ArgInfo &Arg) { unsigned ArgNo = Arg.Formal->getArgNo(); return CS->getArgOperand(ArgNo) == Arg.Actual; })) { CS->setCalledFunction(Clone); Solver.markOverdefined(CS); } } } void updateSpecializedFuncs(FuncList &Candidates, FuncList &WorkList) { for (auto *F : WorkList) { SpecializedFuncs.insert(F); // Initialize the state of the newly created functions, marking them // argument-tracked and executable. if (F->hasExactDefinition() && !F->hasFnAttribute(Attribute::Naked)) Solver.addTrackedFunction(F); Solver.addArgumentTrackedFunction(F); Candidates.push_back(F); Solver.markBlockExecutable(&F->front()); // Replace the function arguments for the specialized functions. for (Argument &Arg : F->args()) if (!Arg.use_empty() && tryToReplaceWithConstant(&Arg)) LLVM_DEBUG(dbgs() << "FnSpecialization: Replaced constant argument: " << Arg.getNameOrAsOperand() << "\n"); } } }; } // namespace bool llvm::runFunctionSpecialization( Module &M, const DataLayout &DL, std::function GetTLI, std::function GetTTI, std::function GetAC, function_ref GetAnalysis) { SCCPSolver Solver(DL, GetTLI, M.getContext()); FunctionSpecializer FS(Solver, GetAC, GetTTI, GetTLI); bool Changed = false; // Loop over all functions, marking arguments to those with their addresses // taken or that are external as overdefined. for (Function &F : M) { if (F.isDeclaration()) continue; if (F.hasFnAttribute(Attribute::NoDuplicate)) continue; LLVM_DEBUG(dbgs() << "\nFnSpecialization: Analysing decl: " << F.getName() << "\n"); Solver.addAnalysis(F, GetAnalysis(F)); // Determine if we can track the function's arguments. If so, add the // function to the solver's set of argument-tracked functions. if (canTrackArgumentsInterprocedurally(&F)) { LLVM_DEBUG(dbgs() << "FnSpecialization: Can track arguments\n"); Solver.addArgumentTrackedFunction(&F); continue; } else { LLVM_DEBUG(dbgs() << "FnSpecialization: Can't track arguments!\n" << "FnSpecialization: Doesn't have local linkage, or " << "has its address taken\n"); } // Assume the function is called. Solver.markBlockExecutable(&F.front()); // Assume nothing about the incoming arguments. for (Argument &AI : F.args()) Solver.markOverdefined(&AI); } // Determine if we can track any of the module's global variables. If so, add // the global variables we can track to the solver's set of tracked global // variables. for (GlobalVariable &G : M.globals()) { G.removeDeadConstantUsers(); if (canTrackGlobalVariableInterprocedurally(&G)) Solver.trackValueOfGlobalVariable(&G); } auto &TrackedFuncs = Solver.getArgumentTrackedFunctions(); SmallVector FuncDecls(TrackedFuncs.begin(), TrackedFuncs.end()); // No tracked functions, so nothing to do: don't run the solver and remove // the ssa_copy intrinsics that may have been introduced. if (TrackedFuncs.empty()) { removeSSACopy(M); return false; } // Solve for constants. auto RunSCCPSolver = [&](auto &WorkList) { bool ResolvedUndefs = true; while (ResolvedUndefs) { // Not running the solver unnecessary is checked in regression test // nothing-to-do.ll, so if this debug message is changed, this regression // test needs updating too. LLVM_DEBUG(dbgs() << "FnSpecialization: Running solver\n"); Solver.solve(); LLVM_DEBUG(dbgs() << "FnSpecialization: Resolving undefs\n"); ResolvedUndefs = false; for (Function *F : WorkList) if (Solver.resolvedUndefsIn(*F)) ResolvedUndefs = true; } for (auto *F : WorkList) { for (BasicBlock &BB : *F) { if (!Solver.isBlockExecutable(&BB)) continue; // FIXME: The solver may make changes to the function here, so set // Changed, even if later function specialization does not trigger. for (auto &I : make_early_inc_range(BB)) Changed |= FS.tryToReplaceWithConstant(&I); } } }; #ifndef NDEBUG LLVM_DEBUG(dbgs() << "FnSpecialization: Worklist fn decls:\n"); for (auto *F : FuncDecls) LLVM_DEBUG(dbgs() << "FnSpecialization: *) " << F->getName() << "\n"); #endif // Initially resolve the constants in all the argument tracked functions. RunSCCPSolver(FuncDecls); SmallVector WorkList; unsigned I = 0; while (FuncSpecializationMaxIters != I++ && FS.specializeFunctions(FuncDecls, WorkList)) { LLVM_DEBUG(dbgs() << "FnSpecialization: Finished iteration " << I << "\n"); // Run the solver for the specialized functions. RunSCCPSolver(WorkList); // Replace some unresolved constant arguments. constantArgPropagation(FuncDecls, M, Solver); WorkList.clear(); Changed = true; } LLVM_DEBUG(dbgs() << "FnSpecialization: Number of specializations = " << NumFuncSpecialized << "\n"); // Remove any ssa_copy intrinsics that may have been introduced. removeSSACopy(M); return Changed; }