Book a Demo!
CoCalc Logo Icon
StoreFeaturesDocsShareSupportNewsAboutPoliciesSign UpSign In
freebsd
GitHub Repository: freebsd/freebsd-src
Path: blob/main/contrib/llvm-project/llvm/lib/Transforms/IPO/ModuleInliner.cpp
35266 views
1
//===- ModuleInliner.cpp - Code related to module inliner -----------------===//
2
//
3
// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions.
4
// See https://llvm.org/LICENSE.txt for license information.
5
// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception
6
//
7
//===----------------------------------------------------------------------===//
8
//
9
// This file implements the mechanics required to implement inlining without
10
// missing any calls in the module level. It doesn't need any infromation about
11
// SCC or call graph, which is different from the SCC inliner. The decisions of
12
// which calls are profitable to inline are implemented elsewhere.
13
//
14
//===----------------------------------------------------------------------===//
15
16
#include "llvm/Transforms/IPO/ModuleInliner.h"
17
#include "llvm/ADT/ScopeExit.h"
18
#include "llvm/ADT/SmallVector.h"
19
#include "llvm/ADT/Statistic.h"
20
#include "llvm/Analysis/AliasAnalysis.h"
21
#include "llvm/Analysis/AssumptionCache.h"
22
#include "llvm/Analysis/BlockFrequencyInfo.h"
23
#include "llvm/Analysis/InlineAdvisor.h"
24
#include "llvm/Analysis/InlineCost.h"
25
#include "llvm/Analysis/InlineOrder.h"
26
#include "llvm/Analysis/OptimizationRemarkEmitter.h"
27
#include "llvm/Analysis/ProfileSummaryInfo.h"
28
#include "llvm/Analysis/ReplayInlineAdvisor.h"
29
#include "llvm/Analysis/TargetLibraryInfo.h"
30
#include "llvm/IR/DiagnosticInfo.h"
31
#include "llvm/IR/Function.h"
32
#include "llvm/IR/InstIterator.h"
33
#include "llvm/IR/Instruction.h"
34
#include "llvm/IR/IntrinsicInst.h"
35
#include "llvm/IR/Module.h"
36
#include "llvm/IR/PassManager.h"
37
#include "llvm/Support/CommandLine.h"
38
#include "llvm/Support/Debug.h"
39
#include "llvm/Support/raw_ostream.h"
40
#include "llvm/Transforms/Utils/CallPromotionUtils.h"
41
#include "llvm/Transforms/Utils/Cloning.h"
42
#include <cassert>
43
44
using namespace llvm;
45
46
#define DEBUG_TYPE "module-inline"
47
48
STATISTIC(NumInlined, "Number of functions inlined");
49
STATISTIC(NumDeleted, "Number of functions deleted because all callers found");
50
51
/// Return true if the specified inline history ID
52
/// indicates an inline history that includes the specified function.
53
static bool inlineHistoryIncludes(
54
Function *F, int InlineHistoryID,
55
const SmallVectorImpl<std::pair<Function *, int>> &InlineHistory) {
56
while (InlineHistoryID != -1) {
57
assert(unsigned(InlineHistoryID) < InlineHistory.size() &&
58
"Invalid inline history ID");
59
if (InlineHistory[InlineHistoryID].first == F)
60
return true;
61
InlineHistoryID = InlineHistory[InlineHistoryID].second;
62
}
63
return false;
64
}
65
66
InlineAdvisor &ModuleInlinerPass::getAdvisor(const ModuleAnalysisManager &MAM,
67
FunctionAnalysisManager &FAM,
68
Module &M) {
69
if (OwnedAdvisor)
70
return *OwnedAdvisor;
71
72
auto *IAA = MAM.getCachedResult<InlineAdvisorAnalysis>(M);
73
if (!IAA) {
74
// It should still be possible to run the inliner as a stand-alone module
75
// pass, for test scenarios. In that case, we default to the
76
// DefaultInlineAdvisor, which doesn't need to keep state between module
77
// pass runs. It also uses just the default InlineParams. In this case, we
78
// need to use the provided FAM, which is valid for the duration of the
79
// inliner pass, and thus the lifetime of the owned advisor. The one we
80
// would get from the MAM can be invalidated as a result of the inliner's
81
// activity.
82
OwnedAdvisor = std::make_unique<DefaultInlineAdvisor>(
83
M, FAM, Params, InlineContext{LTOPhase, InlinePass::ModuleInliner});
84
85
return *OwnedAdvisor;
86
}
87
assert(IAA->getAdvisor() &&
88
"Expected a present InlineAdvisorAnalysis also have an "
89
"InlineAdvisor initialized");
90
return *IAA->getAdvisor();
91
}
92
93
static bool isKnownLibFunction(Function &F, TargetLibraryInfo &TLI) {
94
LibFunc LF;
95
96
// Either this is a normal library function or a "vectorizable"
97
// function. Not using the VFDatabase here because this query
98
// is related only to libraries handled via the TLI.
99
return TLI.getLibFunc(F, LF) ||
100
TLI.isKnownVectorFunctionInLibrary(F.getName());
101
}
102
103
PreservedAnalyses ModuleInlinerPass::run(Module &M,
104
ModuleAnalysisManager &MAM) {
105
LLVM_DEBUG(dbgs() << "---- Module Inliner is Running ---- \n");
106
107
auto &IAA = MAM.getResult<InlineAdvisorAnalysis>(M);
108
if (!IAA.tryCreate(Params, Mode, {},
109
InlineContext{LTOPhase, InlinePass::ModuleInliner})) {
110
M.getContext().emitError(
111
"Could not setup Inlining Advisor for the requested "
112
"mode and/or options");
113
return PreservedAnalyses::all();
114
}
115
116
bool Changed = false;
117
118
ProfileSummaryInfo *PSI = MAM.getCachedResult<ProfileSummaryAnalysis>(M);
119
120
FunctionAnalysisManager &FAM =
121
MAM.getResult<FunctionAnalysisManagerModuleProxy>(M).getManager();
122
123
auto GetTLI = [&FAM](Function &F) -> TargetLibraryInfo & {
124
return FAM.getResult<TargetLibraryAnalysis>(F);
125
};
126
127
InlineAdvisor &Advisor = getAdvisor(MAM, FAM, M);
128
Advisor.onPassEntry();
129
130
auto AdvisorOnExit = make_scope_exit([&] { Advisor.onPassExit(); });
131
132
// In the module inliner, a priority-based worklist is used for calls across
133
// the entire Module. With this module inliner, the inline order is not
134
// limited to bottom-up order. More globally scope inline order is enabled.
135
// Also, the inline deferral logic become unnecessary in this module inliner.
136
// It is possible to use other priority heuristics, e.g. profile-based
137
// heuristic.
138
//
139
// TODO: Here is a huge amount duplicate code between the module inliner and
140
// the SCC inliner, which need some refactoring.
141
auto Calls = getInlineOrder(FAM, Params, MAM, M);
142
assert(Calls != nullptr && "Expected an initialized InlineOrder");
143
144
// Populate the initial list of calls in this module.
145
for (Function &F : M) {
146
auto &ORE = FAM.getResult<OptimizationRemarkEmitterAnalysis>(F);
147
for (Instruction &I : instructions(F))
148
if (auto *CB = dyn_cast<CallBase>(&I))
149
if (Function *Callee = CB->getCalledFunction()) {
150
if (!Callee->isDeclaration())
151
Calls->push({CB, -1});
152
else if (!isa<IntrinsicInst>(I)) {
153
using namespace ore;
154
setInlineRemark(*CB, "unavailable definition");
155
ORE.emit([&]() {
156
return OptimizationRemarkMissed(DEBUG_TYPE, "NoDefinition", &I)
157
<< NV("Callee", Callee) << " will not be inlined into "
158
<< NV("Caller", CB->getCaller())
159
<< " because its definition is unavailable"
160
<< setIsVerbose();
161
});
162
}
163
}
164
}
165
if (Calls->empty())
166
return PreservedAnalyses::all();
167
168
// When inlining a callee produces new call sites, we want to keep track of
169
// the fact that they were inlined from the callee. This allows us to avoid
170
// infinite inlining in some obscure cases. To represent this, we use an
171
// index into the InlineHistory vector.
172
SmallVector<std::pair<Function *, int>, 16> InlineHistory;
173
174
// Track the dead functions to delete once finished with inlining calls. We
175
// defer deleting these to make it easier to handle the call graph updates.
176
SmallVector<Function *, 4> DeadFunctions;
177
178
// Loop forward over all of the calls.
179
while (!Calls->empty()) {
180
auto P = Calls->pop();
181
CallBase *CB = P.first;
182
const int InlineHistoryID = P.second;
183
Function &F = *CB->getCaller();
184
Function &Callee = *CB->getCalledFunction();
185
186
LLVM_DEBUG(dbgs() << "Inlining calls in: " << F.getName() << "\n"
187
<< " Function size: " << F.getInstructionCount()
188
<< "\n");
189
(void)F;
190
191
auto GetAssumptionCache = [&](Function &F) -> AssumptionCache & {
192
return FAM.getResult<AssumptionAnalysis>(F);
193
};
194
195
if (InlineHistoryID != -1 &&
196
inlineHistoryIncludes(&Callee, InlineHistoryID, InlineHistory)) {
197
setInlineRemark(*CB, "recursive");
198
continue;
199
}
200
201
auto Advice = Advisor.getAdvice(*CB, /*OnlyMandatory*/ false);
202
// Check whether we want to inline this callsite.
203
if (!Advice->isInliningRecommended()) {
204
Advice->recordUnattemptedInlining();
205
continue;
206
}
207
208
// Setup the data structure used to plumb customization into the
209
// `InlineFunction` routine.
210
InlineFunctionInfo IFI(
211
GetAssumptionCache, PSI,
212
&FAM.getResult<BlockFrequencyAnalysis>(*(CB->getCaller())),
213
&FAM.getResult<BlockFrequencyAnalysis>(Callee));
214
215
InlineResult IR =
216
InlineFunction(*CB, IFI, /*MergeAttributes=*/true,
217
&FAM.getResult<AAManager>(*CB->getCaller()));
218
if (!IR.isSuccess()) {
219
Advice->recordUnsuccessfulInlining(IR);
220
continue;
221
}
222
223
Changed = true;
224
++NumInlined;
225
226
LLVM_DEBUG(dbgs() << " Size after inlining: " << F.getInstructionCount()
227
<< "\n");
228
229
// Add any new callsites to defined functions to the worklist.
230
if (!IFI.InlinedCallSites.empty()) {
231
int NewHistoryID = InlineHistory.size();
232
InlineHistory.push_back({&Callee, InlineHistoryID});
233
234
for (CallBase *ICB : reverse(IFI.InlinedCallSites)) {
235
Function *NewCallee = ICB->getCalledFunction();
236
if (!NewCallee) {
237
// Try to promote an indirect (virtual) call without waiting for
238
// the post-inline cleanup and the next DevirtSCCRepeatedPass
239
// iteration because the next iteration may not happen and we may
240
// miss inlining it.
241
if (tryPromoteCall(*ICB))
242
NewCallee = ICB->getCalledFunction();
243
}
244
if (NewCallee)
245
if (!NewCallee->isDeclaration())
246
Calls->push({ICB, NewHistoryID});
247
}
248
}
249
250
// For local functions, check whether this makes the callee trivially
251
// dead. In that case, we can drop the body of the function eagerly
252
// which may reduce the number of callers of other functions to one,
253
// changing inline cost thresholds.
254
bool CalleeWasDeleted = false;
255
if (Callee.hasLocalLinkage()) {
256
// To check this we also need to nuke any dead constant uses (perhaps
257
// made dead by this operation on other functions).
258
Callee.removeDeadConstantUsers();
259
// if (Callee.use_empty() && !CG.isLibFunction(Callee)) {
260
if (Callee.use_empty() && !isKnownLibFunction(Callee, GetTLI(Callee))) {
261
Calls->erase_if([&](const std::pair<CallBase *, int> &Call) {
262
return Call.first->getCaller() == &Callee;
263
});
264
// Clear the body and queue the function itself for deletion when we
265
// finish inlining.
266
// Note that after this point, it is an error to do anything other
267
// than use the callee's address or delete it.
268
Callee.dropAllReferences();
269
assert(!is_contained(DeadFunctions, &Callee) &&
270
"Cannot put cause a function to become dead twice!");
271
DeadFunctions.push_back(&Callee);
272
CalleeWasDeleted = true;
273
}
274
}
275
if (CalleeWasDeleted)
276
Advice->recordInliningWithCalleeDeleted();
277
else
278
Advice->recordInlining();
279
}
280
281
// Now that we've finished inlining all of the calls across this module,
282
// delete all of the trivially dead functions.
283
//
284
// Note that this walks a pointer set which has non-deterministic order but
285
// that is OK as all we do is delete things and add pointers to unordered
286
// sets.
287
for (Function *DeadF : DeadFunctions) {
288
// Clear out any cached analyses.
289
FAM.clear(*DeadF, DeadF->getName());
290
291
// And delete the actual function from the module.
292
M.getFunctionList().erase(DeadF);
293
294
++NumDeleted;
295
}
296
297
if (!Changed)
298
return PreservedAnalyses::all();
299
300
return PreservedAnalyses::none();
301
}
302
303