xref: /freebsd/contrib/llvm-project/clang/lib/Tooling/Transformer/SourceCode.cpp (revision 0fca6ea1d4eea4c934cfff25ac9ee8ad6fe95583)
1 //===--- SourceCode.cpp - Source code manipulation routines -----*- C++ -*-===//
2 //
3 // Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions.
4 // See https://llvm.org/LICENSE.txt for license information.
5 // SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception
6 //
7 //===----------------------------------------------------------------------===//
8 //
9 //  This file provides functions that simplify extraction of source code.
10 //
11 //===----------------------------------------------------------------------===//
12 #include "clang/Tooling/Transformer/SourceCode.h"
13 #include "clang/AST/ASTContext.h"
14 #include "clang/AST/Attr.h"
15 #include "clang/AST/Comment.h"
16 #include "clang/AST/Decl.h"
17 #include "clang/AST/DeclCXX.h"
18 #include "clang/AST/DeclTemplate.h"
19 #include "clang/AST/Expr.h"
20 #include "clang/Basic/SourceManager.h"
21 #include "clang/Lex/Lexer.h"
22 #include "llvm/Support/Errc.h"
23 #include "llvm/Support/Error.h"
24 #include <set>
25 
26 using namespace clang;
27 
28 using llvm::errc;
29 using llvm::StringError;
30 
getText(CharSourceRange Range,const ASTContext & Context)31 StringRef clang::tooling::getText(CharSourceRange Range,
32                                   const ASTContext &Context) {
33   return Lexer::getSourceText(Range, Context.getSourceManager(),
34                               Context.getLangOpts());
35 }
36 
maybeExtendRange(CharSourceRange Range,tok::TokenKind Next,ASTContext & Context)37 CharSourceRange clang::tooling::maybeExtendRange(CharSourceRange Range,
38                                                  tok::TokenKind Next,
39                                                  ASTContext &Context) {
40   CharSourceRange R = Lexer::getAsCharRange(Range, Context.getSourceManager(),
41                                             Context.getLangOpts());
42   if (R.isInvalid())
43     return Range;
44   Token Tok;
45   bool Err =
46       Lexer::getRawToken(R.getEnd(), Tok, Context.getSourceManager(),
47                          Context.getLangOpts(), /*IgnoreWhiteSpace=*/true);
48   if (Err || !Tok.is(Next))
49     return Range;
50   return CharSourceRange::getTokenRange(Range.getBegin(), Tok.getLocation());
51 }
52 
validateRange(const CharSourceRange & Range,const SourceManager & SM,bool AllowSystemHeaders)53 llvm::Error clang::tooling::validateRange(const CharSourceRange &Range,
54                                           const SourceManager &SM,
55                                           bool AllowSystemHeaders) {
56   if (Range.isInvalid())
57     return llvm::make_error<StringError>(errc::invalid_argument,
58                                          "Invalid range");
59 
60   if (Range.getBegin().isMacroID() || Range.getEnd().isMacroID())
61     return llvm::make_error<StringError>(
62         errc::invalid_argument, "Range starts or ends in a macro expansion");
63 
64   if (!AllowSystemHeaders) {
65     if (SM.isInSystemHeader(Range.getBegin()) ||
66         SM.isInSystemHeader(Range.getEnd()))
67       return llvm::make_error<StringError>(errc::invalid_argument,
68                                            "Range is in system header");
69   }
70 
71   std::pair<FileID, unsigned> BeginInfo = SM.getDecomposedLoc(Range.getBegin());
72   std::pair<FileID, unsigned> EndInfo = SM.getDecomposedLoc(Range.getEnd());
73   if (BeginInfo.first != EndInfo.first)
74     return llvm::make_error<StringError>(
75         errc::invalid_argument, "Range begins and ends in different files");
76 
77   if (BeginInfo.second > EndInfo.second)
78     return llvm::make_error<StringError>(errc::invalid_argument,
79                                          "Range's begin is past its end");
80 
81   return llvm::Error::success();
82 }
83 
validateEditRange(const CharSourceRange & Range,const SourceManager & SM)84 llvm::Error clang::tooling::validateEditRange(const CharSourceRange &Range,
85                                               const SourceManager &SM) {
86   return validateRange(Range, SM, /*AllowSystemHeaders=*/false);
87 }
88 
spelledInMacroDefinition(SourceLocation Loc,const SourceManager & SM)89 static bool spelledInMacroDefinition(SourceLocation Loc,
90                                      const SourceManager &SM) {
91   while (Loc.isMacroID()) {
92     const auto &Expansion = SM.getSLocEntry(SM.getFileID(Loc)).getExpansion();
93     if (Expansion.isMacroArgExpansion()) {
94       // Check the spelling location of the macro arg, in case the arg itself is
95       // in a macro expansion.
96       Loc = Expansion.getSpellingLoc();
97     } else {
98       return true;
99     }
100   }
101   return false;
102 }
103 
104 // Returns the expansion char-range of `Loc` if `Loc` is a split token. For
105 // example, `>>` in nested templates needs the first `>` to be split, otherwise
106 // the `SourceLocation` of the token would lex as `>>` instead of `>`.
107 static std::optional<CharSourceRange>
getExpansionForSplitToken(SourceLocation Loc,const SourceManager & SM,const LangOptions & LangOpts)108 getExpansionForSplitToken(SourceLocation Loc, const SourceManager &SM,
109                           const LangOptions &LangOpts) {
110   if (Loc.isMacroID()) {
111     bool Invalid = false;
112     auto &SLoc = SM.getSLocEntry(SM.getFileID(Loc), &Invalid);
113     if (Invalid)
114       return std::nullopt;
115     if (auto &Expansion = SLoc.getExpansion();
116         !Expansion.isExpansionTokenRange()) {
117       // A char-range expansion is only used where a token-range would be
118       // incorrect, and so identifies this as a split token (and importantly,
119       // not as a macro).
120       return Expansion.getExpansionLocRange();
121     }
122   }
123   return std::nullopt;
124 }
125 
126 // If `Range` covers a split token, returns the expansion range, otherwise
127 // returns `Range`.
getRangeForSplitTokens(CharSourceRange Range,const SourceManager & SM,const LangOptions & LangOpts)128 static CharSourceRange getRangeForSplitTokens(CharSourceRange Range,
129                                               const SourceManager &SM,
130                                               const LangOptions &LangOpts) {
131   if (Range.isTokenRange()) {
132     auto BeginToken = getExpansionForSplitToken(Range.getBegin(), SM, LangOpts);
133     auto EndToken = getExpansionForSplitToken(Range.getEnd(), SM, LangOpts);
134     if (EndToken) {
135       SourceLocation BeginLoc =
136           BeginToken ? BeginToken->getBegin() : Range.getBegin();
137       // We can't use the expansion location with a token-range, because that
138       // will incorrectly lex the end token, so use a char-range that ends at
139       // the split.
140       return CharSourceRange::getCharRange(BeginLoc, EndToken->getEnd());
141     } else if (BeginToken) {
142       // Since the end token is not split, the whole range covers the split, so
143       // the only adjustment we make is to use the expansion location of the
144       // begin token.
145       return CharSourceRange::getTokenRange(BeginToken->getBegin(),
146                                             Range.getEnd());
147     }
148   }
149   return Range;
150 }
151 
getRange(const CharSourceRange & EditRange,const SourceManager & SM,const LangOptions & LangOpts,bool IncludeMacroExpansion)152 static CharSourceRange getRange(const CharSourceRange &EditRange,
153                                 const SourceManager &SM,
154                                 const LangOptions &LangOpts,
155                                 bool IncludeMacroExpansion) {
156   CharSourceRange Range;
157   if (IncludeMacroExpansion) {
158     Range = Lexer::makeFileCharRange(EditRange, SM, LangOpts);
159   } else {
160     auto AdjustedRange = getRangeForSplitTokens(EditRange, SM, LangOpts);
161     if (spelledInMacroDefinition(AdjustedRange.getBegin(), SM) ||
162         spelledInMacroDefinition(AdjustedRange.getEnd(), SM))
163       return {};
164 
165     auto B = SM.getSpellingLoc(AdjustedRange.getBegin());
166     auto E = SM.getSpellingLoc(AdjustedRange.getEnd());
167     if (AdjustedRange.isTokenRange())
168       E = Lexer::getLocForEndOfToken(E, 0, SM, LangOpts);
169     Range = CharSourceRange::getCharRange(B, E);
170   }
171   return Range;
172 }
173 
getFileRangeForEdit(const CharSourceRange & EditRange,const SourceManager & SM,const LangOptions & LangOpts,bool IncludeMacroExpansion)174 std::optional<CharSourceRange> clang::tooling::getFileRangeForEdit(
175     const CharSourceRange &EditRange, const SourceManager &SM,
176     const LangOptions &LangOpts, bool IncludeMacroExpansion) {
177   CharSourceRange Range =
178       getRange(EditRange, SM, LangOpts, IncludeMacroExpansion);
179   bool IsInvalid = llvm::errorToBool(validateEditRange(Range, SM));
180   if (IsInvalid)
181     return std::nullopt;
182   return Range;
183 }
184 
getFileRange(const CharSourceRange & EditRange,const SourceManager & SM,const LangOptions & LangOpts,bool IncludeMacroExpansion)185 std::optional<CharSourceRange> clang::tooling::getFileRange(
186     const CharSourceRange &EditRange, const SourceManager &SM,
187     const LangOptions &LangOpts, bool IncludeMacroExpansion) {
188   CharSourceRange Range =
189       getRange(EditRange, SM, LangOpts, IncludeMacroExpansion);
190   bool IsInvalid =
191       llvm::errorToBool(validateRange(Range, SM, /*AllowSystemHeaders=*/true));
192   if (IsInvalid)
193     return std::nullopt;
194   return Range;
195 }
196 
startsWithNewline(const SourceManager & SM,const Token & Tok)197 static bool startsWithNewline(const SourceManager &SM, const Token &Tok) {
198   return isVerticalWhitespace(SM.getCharacterData(Tok.getLocation())[0]);
199 }
200 
contains(const std::set<tok::TokenKind> & Terminators,const Token & Tok)201 static bool contains(const std::set<tok::TokenKind> &Terminators,
202                      const Token &Tok) {
203   return Terminators.count(Tok.getKind()) > 0;
204 }
205 
206 // Returns the exclusive, *file* end location of the entity whose last token is
207 // at location 'EntityLast'. That is, it returns the location one past the last
208 // relevant character.
209 //
210 // Associated tokens include comments, horizontal whitespace and 'Terminators'
211 // -- optional tokens, which, if any are found, will be included; if
212 // 'Terminators' is empty, we will not include any extra tokens beyond comments
213 // and horizontal whitespace.
214 static SourceLocation
getEntityEndLoc(const SourceManager & SM,SourceLocation EntityLast,const std::set<tok::TokenKind> & Terminators,const LangOptions & LangOpts)215 getEntityEndLoc(const SourceManager &SM, SourceLocation EntityLast,
216                 const std::set<tok::TokenKind> &Terminators,
217                 const LangOptions &LangOpts) {
218   assert(EntityLast.isValid() && "Invalid end location found.");
219 
220   // We remember the last location of a non-horizontal-whitespace token we have
221   // lexed; this is the location up to which we will want to delete.
222   // FIXME: Support using the spelling loc here for cases where we want to
223   // analyze the macro text.
224 
225   CharSourceRange ExpansionRange = SM.getExpansionRange(EntityLast);
226   // FIXME: Should check isTokenRange(), for the (rare) case that
227   // `ExpansionRange` is a character range.
228   std::unique_ptr<Lexer> Lexer = [&]() {
229     bool Invalid = false;
230     auto FileOffset = SM.getDecomposedLoc(ExpansionRange.getEnd());
231     llvm::StringRef File = SM.getBufferData(FileOffset.first, &Invalid);
232     assert(!Invalid && "Cannot get file/offset");
233     return std::make_unique<clang::Lexer>(
234         SM.getLocForStartOfFile(FileOffset.first), LangOpts, File.begin(),
235         File.data() + FileOffset.second, File.end());
236   }();
237 
238   // Tell Lexer to return whitespace as pseudo-tokens (kind is tok::unknown).
239   Lexer->SetKeepWhitespaceMode(true);
240 
241   // Generally, the code we want to include looks like this ([] are optional),
242   // If Terminators is empty:
243   //   [ <comment> ] [ <newline> ]
244   // Otherwise:
245   //   ... <terminator> [ <comment> ] [ <newline> ]
246 
247   Token Tok;
248   bool Terminated = false;
249 
250   // First, lex to the current token (which is the last token of the range that
251   // is definitely associated with the decl). Then, we process the first token
252   // separately from the rest based on conditions that hold specifically for
253   // that first token.
254   //
255   // We do not search for a terminator if none is required or we've already
256   // encountered it. Otherwise, if the original `EntityLast` location was in a
257   // macro expansion, we don't have visibility into the text, so we assume we've
258   // already terminated. However, we note this assumption with
259   // `TerminatedByMacro`, because we'll want to handle it somewhat differently
260   // for the terminators semicolon and comma. These terminators can be safely
261   // associated with the entity when they appear after the macro -- extra
262   // semicolons have no effect on the program and a well-formed program won't
263   // have multiple commas in a row, so we're guaranteed that there is only one.
264   //
265   // FIXME: This handling of macros is more conservative than necessary. When
266   // the end of the expansion coincides with the end of the node, we can still
267   // safely analyze the code. But, it is more complicated, because we need to
268   // start by lexing the spelling loc for the first token and then switch to the
269   // expansion loc.
270   bool TerminatedByMacro = false;
271   Lexer->LexFromRawLexer(Tok);
272   if (Terminators.empty() || contains(Terminators, Tok))
273     Terminated = true;
274   else if (EntityLast.isMacroID()) {
275     Terminated = true;
276     TerminatedByMacro = true;
277   }
278 
279   // We save the most recent candidate for the exclusive end location.
280   SourceLocation End = Tok.getEndLoc();
281 
282   while (!Terminated) {
283     // Lex the next token we want to possibly expand the range with.
284     Lexer->LexFromRawLexer(Tok);
285 
286     switch (Tok.getKind()) {
287     case tok::eof:
288     // Unexpected separators.
289     case tok::l_brace:
290     case tok::r_brace:
291     case tok::comma:
292       return End;
293     // Whitespace pseudo-tokens.
294     case tok::unknown:
295       if (startsWithNewline(SM, Tok))
296         // Include at least until the end of the line.
297         End = Tok.getEndLoc();
298       break;
299     default:
300       if (contains(Terminators, Tok))
301         Terminated = true;
302       End = Tok.getEndLoc();
303       break;
304     }
305   }
306 
307   do {
308     // Lex the next token we want to possibly expand the range with.
309     Lexer->LexFromRawLexer(Tok);
310 
311     switch (Tok.getKind()) {
312     case tok::unknown:
313       if (startsWithNewline(SM, Tok))
314         // We're done, but include this newline.
315         return Tok.getEndLoc();
316       break;
317     case tok::comment:
318       // Include any comments we find on the way.
319       End = Tok.getEndLoc();
320       break;
321     case tok::semi:
322     case tok::comma:
323       if (TerminatedByMacro && contains(Terminators, Tok)) {
324         End = Tok.getEndLoc();
325         // We've found a real terminator.
326         TerminatedByMacro = false;
327         break;
328       }
329       // Found an unrelated token; stop and don't include it.
330       return End;
331     default:
332       // Found an unrelated token; stop and don't include it.
333       return End;
334     }
335   } while (true);
336 }
337 
338 // Returns the expected terminator tokens for the given declaration.
339 //
340 // If we do not know the correct terminator token, returns an empty set.
341 //
342 // There are cases where we have more than one possible terminator (for example,
343 // we find either a comma or a semicolon after a VarDecl).
getTerminators(const Decl & D)344 static std::set<tok::TokenKind> getTerminators(const Decl &D) {
345   if (llvm::isa<RecordDecl>(D) || llvm::isa<UsingDecl>(D))
346     return {tok::semi};
347 
348   if (llvm::isa<FunctionDecl>(D) || llvm::isa<LinkageSpecDecl>(D))
349     return {tok::r_brace, tok::semi};
350 
351   if (llvm::isa<VarDecl>(D) || llvm::isa<FieldDecl>(D))
352     return {tok::comma, tok::semi};
353 
354   return {};
355 }
356 
357 // Starting from `Loc`, skips whitespace up to, and including, a single
358 // newline. Returns the (exclusive) end of any skipped whitespace (that is, the
359 // location immediately after the whitespace).
skipWhitespaceAndNewline(const SourceManager & SM,SourceLocation Loc,const LangOptions & LangOpts)360 static SourceLocation skipWhitespaceAndNewline(const SourceManager &SM,
361                                                SourceLocation Loc,
362                                                const LangOptions &LangOpts) {
363   const char *LocChars = SM.getCharacterData(Loc);
364   int i = 0;
365   while (isHorizontalWhitespace(LocChars[i]))
366     ++i;
367   if (isVerticalWhitespace(LocChars[i]))
368     ++i;
369   return Loc.getLocWithOffset(i);
370 }
371 
372 // Is `Loc` separated from any following decl by something meaningful (e.g. an
373 // empty line, a comment), ignoring horizontal whitespace?  Since this is a
374 // heuristic, we return false when in doubt.  `Loc` cannot be the first location
375 // in the file.
atOrBeforeSeparation(const SourceManager & SM,SourceLocation Loc,const LangOptions & LangOpts)376 static bool atOrBeforeSeparation(const SourceManager &SM, SourceLocation Loc,
377                                  const LangOptions &LangOpts) {
378   // If the preceding character is a newline, we'll check for an empty line as a
379   // separator. However, we can't identify an empty line using tokens, so we
380   // analyse the characters. If we try to use tokens, we'll just end up with a
381   // whitespace token, whose characters we'd have to analyse anyhow.
382   bool Invalid = false;
383   const char *LocChars =
384       SM.getCharacterData(Loc.getLocWithOffset(-1), &Invalid);
385   assert(!Invalid &&
386          "Loc must be a valid character and not the first of the source file.");
387   if (isVerticalWhitespace(LocChars[0])) {
388     for (int i = 1; isWhitespace(LocChars[i]); ++i)
389       if (isVerticalWhitespace(LocChars[i]))
390         return true;
391   }
392   // We didn't find an empty line, so lex the next token, skipping past any
393   // whitespace we just scanned.
394   Token Tok;
395   bool Failed = Lexer::getRawToken(Loc, Tok, SM, LangOpts,
396                                    /*IgnoreWhiteSpace=*/true);
397   if (Failed)
398     // Any text that confuses the lexer seems fair to consider a separation.
399     return true;
400 
401   switch (Tok.getKind()) {
402   case tok::comment:
403   case tok::l_brace:
404   case tok::r_brace:
405   case tok::eof:
406     return true;
407   default:
408     return false;
409   }
410 }
411 
getAssociatedRange(const Decl & Decl,ASTContext & Context)412 CharSourceRange tooling::getAssociatedRange(const Decl &Decl,
413                                             ASTContext &Context) {
414   const SourceManager &SM = Context.getSourceManager();
415   const LangOptions &LangOpts = Context.getLangOpts();
416   CharSourceRange Range = CharSourceRange::getTokenRange(Decl.getSourceRange());
417 
418   // First, expand to the start of the template<> declaration if necessary.
419   if (const auto *Record = llvm::dyn_cast<CXXRecordDecl>(&Decl)) {
420     if (const auto *T = Record->getDescribedClassTemplate())
421       if (SM.isBeforeInTranslationUnit(T->getBeginLoc(), Range.getBegin()))
422         Range.setBegin(T->getBeginLoc());
423   } else if (const auto *F = llvm::dyn_cast<FunctionDecl>(&Decl)) {
424     if (const auto *T = F->getDescribedFunctionTemplate())
425       if (SM.isBeforeInTranslationUnit(T->getBeginLoc(), Range.getBegin()))
426         Range.setBegin(T->getBeginLoc());
427   }
428 
429   // Next, expand the end location past trailing comments to include a potential
430   // newline at the end of the decl's line.
431   Range.setEnd(
432       getEntityEndLoc(SM, Decl.getEndLoc(), getTerminators(Decl), LangOpts));
433   Range.setTokenRange(false);
434 
435   // Expand to include preceeding associated comments. We ignore any comments
436   // that are not preceeding the decl, since we've already skipped trailing
437   // comments with getEntityEndLoc.
438   if (const RawComment *Comment =
439           Decl.getASTContext().getRawCommentForDeclNoCache(&Decl))
440     // Only include a preceding comment if:
441     // * it is *not* separate from the declaration (not including any newline
442     //   that immediately follows the comment),
443     // * the decl *is* separate from any following entity (so, there are no
444     //   other entities the comment could refer to), and
445     // * it is not a IfThisThenThat lint check.
446     if (SM.isBeforeInTranslationUnit(Comment->getBeginLoc(),
447                                      Range.getBegin()) &&
448         !atOrBeforeSeparation(
449             SM, skipWhitespaceAndNewline(SM, Comment->getEndLoc(), LangOpts),
450             LangOpts) &&
451         atOrBeforeSeparation(SM, Range.getEnd(), LangOpts)) {
452       const StringRef CommentText = Comment->getRawText(SM);
453       if (!CommentText.contains("LINT.IfChange") &&
454           !CommentText.contains("LINT.ThenChange"))
455         Range.setBegin(Comment->getBeginLoc());
456     }
457   // Add leading attributes.
458   for (auto *Attr : Decl.attrs()) {
459     if (Attr->getLocation().isInvalid() ||
460         !SM.isBeforeInTranslationUnit(Attr->getLocation(), Range.getBegin()))
461       continue;
462     Range.setBegin(Attr->getLocation());
463 
464     // Extend to the left '[[' or '__attribute((' if we saw the attribute,
465     // unless it is not a valid location.
466     bool Invalid;
467     StringRef Source =
468         SM.getBufferData(SM.getFileID(Range.getBegin()), &Invalid);
469     if (Invalid)
470       continue;
471     llvm::StringRef BeforeAttr =
472         Source.substr(0, SM.getFileOffset(Range.getBegin()));
473     llvm::StringRef BeforeAttrStripped = BeforeAttr.rtrim();
474 
475     for (llvm::StringRef Prefix : {"[[", "__attribute__(("}) {
476       // Handle whitespace between attribute prefix and attribute value.
477       if (BeforeAttrStripped.ends_with(Prefix)) {
478         // Move start to start position of prefix, which is
479         // length(BeforeAttr) - length(BeforeAttrStripped) + length(Prefix)
480         // positions to the left.
481         Range.setBegin(Range.getBegin().getLocWithOffset(static_cast<int>(
482             -BeforeAttr.size() + BeforeAttrStripped.size() - Prefix.size())));
483         break;
484         // If we didn't see '[[' or '__attribute' it's probably coming from a
485         // macro expansion which is already handled by makeFileCharRange(),
486         // below.
487       }
488     }
489   }
490 
491   // Range.getEnd() is already fully un-expanded by getEntityEndLoc. But,
492   // Range.getBegin() may be inside an expansion.
493   return Lexer::makeFileCharRange(Range, SM, LangOpts);
494 }
495