|
123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345 |
- <!DOCTYPE html PUBLIC "-//W3C//DTD HTML 4.01 Transitional//EN" "http://www.w3.org/TR/html4/loose.dtd">
- <html>
- <!-- Created by GNU Texinfo 6.5, http://www.gnu.org/software/texinfo/ -->
- <head>
- <meta http-equiv="Content-Type" content="text/html; charset=utf-8">
- <title>Lexer (The GNU C Preprocessor Internals)</title>
-
- <meta name="description" content="Lexer (The GNU C Preprocessor Internals)">
- <meta name="keywords" content="Lexer (The GNU C Preprocessor Internals)">
- <meta name="resource-type" content="document">
- <meta name="distribution" content="global">
- <meta name="Generator" content="makeinfo">
- <link href="index.html#Top" rel="start" title="Top">
- <link href="Concept-Index.html#Concept-Index" rel="index" title="Concept Index">
- <link href="index.html#SEC_Contents" rel="contents" title="Table of Contents">
- <link href="index.html#Top" rel="up" title="Top">
- <link href="Hash-Nodes.html#Hash-Nodes" rel="next" title="Hash Nodes">
- <link href="Conventions.html#Conventions" rel="prev" title="Conventions">
- <style type="text/css">
- <!--
- a.summary-letter {text-decoration: none}
- blockquote.indentedblock {margin-right: 0em}
- blockquote.smallindentedblock {margin-right: 0em; font-size: smaller}
- blockquote.smallquotation {font-size: smaller}
- div.display {margin-left: 3.2em}
- div.example {margin-left: 3.2em}
- div.lisp {margin-left: 3.2em}
- div.smalldisplay {margin-left: 3.2em}
- div.smallexample {margin-left: 3.2em}
- div.smalllisp {margin-left: 3.2em}
- kbd {font-style: oblique}
- pre.display {font-family: inherit}
- pre.format {font-family: inherit}
- pre.menu-comment {font-family: serif}
- pre.menu-preformatted {font-family: serif}
- pre.smalldisplay {font-family: inherit; font-size: smaller}
- pre.smallexample {font-size: smaller}
- pre.smallformat {font-family: inherit; font-size: smaller}
- pre.smalllisp {font-size: smaller}
- span.nolinebreak {white-space: nowrap}
- span.roman {font-family: initial; font-weight: normal}
- span.sansserif {font-family: sans-serif; font-weight: normal}
- ul.no-bullet {list-style: none}
- -->
- </style>
-
-
- </head>
-
- <body lang="en">
- <a name="Lexer"></a>
- <div class="header">
- <p>
- Next: <a href="Hash-Nodes.html#Hash-Nodes" accesskey="n" rel="next">Hash Nodes</a>, Previous: <a href="Conventions.html#Conventions" accesskey="p" rel="prev">Conventions</a>, Up: <a href="index.html#Top" accesskey="u" rel="up">Top</a> [<a href="index.html#SEC_Contents" title="Table of contents" rel="contents">Contents</a>][<a href="Concept-Index.html#Concept-Index" title="Index" rel="index">Index</a>]</p>
- </div>
- <hr>
- <a name="The-Lexer"></a>
- <h2 class="unnumbered">The Lexer</h2>
- <a name="index-lexer"></a>
- <a name="index-newlines"></a>
- <a name="index-escaped-newlines"></a>
-
- <a name="Overview"></a>
- <h3 class="section">Overview</h3>
- <p>The lexer is contained in the file <samp>lex.c</samp>. It is a hand-coded
- lexer, and not implemented as a state machine. It can understand C, C++
- and Objective-C source code, and has been extended to allow reasonably
- successful preprocessing of assembly language. The lexer does not make
- an initial pass to strip out trigraphs and escaped newlines, but handles
- them as they are encountered in a single pass of the input file. It
- returns preprocessing tokens individually, not a line at a time.
- </p>
- <p>It is mostly transparent to users of the library, since the library’s
- interface for obtaining the next token, <code>cpp_get_token</code>, takes care
- of lexing new tokens, handling directives, and expanding macros as
- necessary. However, the lexer does expose some functionality so that
- clients of the library can easily spell a given token, such as
- <code>cpp_spell_token</code> and <code>cpp_token_len</code>. These functions are
- useful when generating diagnostics, and for emitting the preprocessed
- output.
- </p>
- <a name="Lexing-a-token"></a>
- <h3 class="section">Lexing a token</h3>
- <p>Lexing of an individual token is handled by <code>_cpp_lex_direct</code> and
- its subroutines. In its current form the code is quite complicated,
- with read ahead characters and such-like, since it strives to not step
- back in the character stream in preparation for handling non-ASCII file
- encodings. The current plan is to convert any such files to UTF-8
- before processing them. This complexity is therefore unnecessary and
- will be removed, so I’ll not discuss it further here.
- </p>
- <p>The job of <code>_cpp_lex_direct</code> is simply to lex a token. It is not
- responsible for issues like directive handling, returning lookahead
- tokens directly, multiple-include optimization, or conditional block
- skipping. It necessarily has a minor rôle to play in memory
- management of lexed lines. I discuss these issues in a separate section
- (see <a href="#Lexing-a-line">Lexing a line</a>).
- </p>
- <p>The lexer places the token it lexes into storage pointed to by the
- variable <code>cur_token</code>, and then increments it. This variable is
- important for correct diagnostic positioning. Unless a specific line
- and column are passed to the diagnostic routines, they will examine the
- <code>line</code> and <code>col</code> values of the token just before the location
- that <code>cur_token</code> points to, and use that location to report the
- diagnostic.
- </p>
- <p>The lexer does not consider whitespace to be a token in its own right.
- If whitespace (other than a new line) precedes a token, it sets the
- <code>PREV_WHITE</code> bit in the token’s flags. Each token has its
- <code>line</code> and <code>col</code> variables set to the line and column of the
- first character of the token. This line number is the line number in
- the translation unit, and can be converted to a source (file, line) pair
- using the line map code.
- </p>
- <p>The first token on a logical, i.e. unescaped, line has the flag
- <code>BOL</code> set for beginning-of-line. This flag is intended for
- internal use, both to distinguish a ‘<samp>#</samp>’ that begins a directive
- from one that doesn’t, and to generate a call-back to clients that want
- to be notified about the start of every non-directive line with tokens
- on it. Clients cannot reliably determine this for themselves: the first
- token might be a macro, and the tokens of a macro expansion do not have
- the <code>BOL</code> flag set. The macro expansion may even be empty, and the
- next token on the line certainly won’t have the <code>BOL</code> flag set.
- </p>
- <p>New lines are treated specially; exactly how the lexer handles them is
- context-dependent. The C standard mandates that directives are
- terminated by the first unescaped newline character, even if it appears
- in the middle of a macro expansion. Therefore, if the state variable
- <code>in_directive</code> is set, the lexer returns a <code>CPP_EOF</code> token,
- which is normally used to indicate end-of-file, to indicate
- end-of-directive. In a directive a <code>CPP_EOF</code> token never means
- end-of-file. Conveniently, if the caller was <code>collect_args</code>, it
- already handles <code>CPP_EOF</code> as if it were end-of-file, and reports an
- error about an unterminated macro argument list.
- </p>
- <p>The C standard also specifies that a new line in the middle of the
- arguments to a macro is treated as whitespace. This white space is
- important in case the macro argument is stringized. The state variable
- <code>parsing_args</code> is nonzero when the preprocessor is collecting the
- arguments to a macro call. It is set to 1 when looking for the opening
- parenthesis to a function-like macro, and 2 when collecting the actual
- arguments up to the closing parenthesis, since these two cases need to
- be distinguished sometimes. One such time is here: the lexer sets the
- <code>PREV_WHITE</code> flag of a token if it meets a new line when
- <code>parsing_args</code> is set to 2. It doesn’t set it if it meets a new
- line when <code>parsing_args</code> is 1, since then code like
- </p>
- <div class="smallexample">
- <pre class="smallexample">#define foo() bar
- foo
- baz
- </pre></div>
-
- <p>would be output with an erroneous space before ‘<samp>baz</samp>’:
- </p>
- <div class="smallexample">
- <pre class="smallexample">foo
- baz
- </pre></div>
-
- <p>This is a good example of the subtlety of getting token spacing correct
- in the preprocessor; there are plenty of tests in the testsuite for
- corner cases like this.
- </p>
- <p>The lexer is written to treat each of ‘<samp>\r</samp>’, ‘<samp>\n</samp>’, ‘<samp>\r\n</samp>’
- and ‘<samp>\n\r</samp>’ as a single new line indicator. This allows it to
- transparently preprocess MS-DOS, Macintosh and Unix files without their
- needing to pass through a special filter beforehand.
- </p>
- <p>We also decided to treat a backslash, either ‘<samp>\</samp>’ or the trigraph
- ‘<samp>??/</samp>’, separated from one of the above newline indicators by
- non-comment whitespace only, as intending to escape the newline. It
- tends to be a typing mistake, and cannot reasonably be mistaken for
- anything else in any of the C-family grammars. Since handling it this
- way is not strictly conforming to the ISO standard, the library issues a
- warning wherever it encounters it.
- </p>
- <p>Handling newlines like this is made simpler by doing it in one place
- only. The function <code>handle_newline</code> takes care of all newline
- characters, and <code>skip_escaped_newlines</code> takes care of arbitrarily
- long sequences of escaped newlines, deferring to <code>handle_newline</code>
- to handle the newlines themselves.
- </p>
- <p>The most painful aspect of lexing ISO-standard C and C++ is handling
- trigraphs and backlash-escaped newlines. Trigraphs are processed before
- any interpretation of the meaning of a character is made, and unfortunately
- there is a trigraph representation for a backslash, so it is possible for
- the trigraph ‘<samp>??/</samp>’ to introduce an escaped newline.
- </p>
- <p>Escaped newlines are tedious because theoretically they can occur
- anywhere—between the ‘<samp>+</samp>’ and ‘<samp>=</samp>’ of the ‘<samp>+=</samp>’ token,
- within the characters of an identifier, and even between the ‘<samp>*</samp>’
- and ‘<samp>/</samp>’ that terminates a comment. Moreover, you cannot be sure
- there is just one—there might be an arbitrarily long sequence of them.
- </p>
- <p>So, for example, the routine that lexes a number, <code>parse_number</code>,
- cannot assume that it can scan forwards until the first non-number
- character and be done with it, because this could be the ‘<samp>\</samp>’
- introducing an escaped newline, or the ‘<samp>?</samp>’ introducing the trigraph
- sequence that represents the ‘<samp>\</samp>’ of an escaped newline. If it
- encounters a ‘<samp>?</samp>’ or ‘<samp>\</samp>’, it calls <code>skip_escaped_newlines</code>
- to skip over any potential escaped newlines before checking whether the
- number has been finished.
- </p>
- <p>Similarly code in the main body of <code>_cpp_lex_direct</code> cannot simply
- check for a ‘<samp>=</samp>’ after a ‘<samp>+</samp>’ character to determine whether it
- has a ‘<samp>+=</samp>’ token; it needs to be prepared for an escaped newline of
- some sort. Such cases use the function <code>get_effective_char</code>, which
- returns the first character after any intervening escaped newlines.
- </p>
- <p>The lexer needs to keep track of the correct column position, including
- counting tabs as specified by the <samp>-ftabstop=</samp> option. This
- should be done even within C-style comments; they can appear in the
- middle of a line, and we want to report diagnostics in the correct
- position for text appearing after the end of the comment.
- </p>
- <a name="Invalid-identifiers"></a><p>Some identifiers, such as <code>__VA_ARGS__</code> and poisoned identifiers,
- may be invalid and require a diagnostic. However, if they appear in a
- macro expansion we don’t want to complain with each use of the macro.
- It is therefore best to catch them during the lexing stage, in
- <code>parse_identifier</code>. In both cases, whether a diagnostic is needed
- or not is dependent upon the lexer’s state. For example, we don’t want
- to issue a diagnostic for re-poisoning a poisoned identifier, or for
- using <code>__VA_ARGS__</code> in the expansion of a variable-argument macro.
- Therefore <code>parse_identifier</code> makes use of state flags to determine
- whether a diagnostic is appropriate. Since we change state on a
- per-token basis, and don’t lex whole lines at a time, this is not a
- problem.
- </p>
- <p>Another place where state flags are used to change behavior is whilst
- lexing header names. Normally, a ‘<samp><</samp>’ would be lexed as a single
- token. After a <code>#include</code> directive, though, it should be lexed as
- a single token as far as the nearest ‘<samp>></samp>’ character. Note that we
- don’t allow the terminators of header names to be escaped; the first
- ‘<samp>"</samp>’ or ‘<samp>></samp>’ terminates the header name.
- </p>
- <p>Interpretation of some character sequences depends upon whether we are
- lexing C, C++ or Objective-C, and on the revision of the standard in
- force. For example, ‘<samp>::</samp>’ is a single token in C++, but in C it is
- two separate ‘<samp>:</samp>’ tokens and almost certainly a syntax error. Such
- cases are handled by <code>_cpp_lex_direct</code> based upon command-line
- flags stored in the <code>cpp_options</code> structure.
- </p>
- <p>Once a token has been lexed, it leads an independent existence. The
- spelling of numbers, identifiers and strings is copied to permanent
- storage from the original input buffer, so a token remains valid and
- correct even if its source buffer is freed with <code>_cpp_pop_buffer</code>.
- The storage holding the spellings of such tokens remains until the
- client program calls cpp_destroy, probably at the end of the translation
- unit.
- </p>
- <a name="Lexing-a-line"></a><a name="Lexing-a-line-1"></a>
- <h3 class="section">Lexing a line</h3>
- <a name="index-token-run"></a>
-
- <p>When the preprocessor was changed to return pointers to tokens, one
- feature I wanted was some sort of guarantee regarding how long a
- returned pointer remains valid. This is important to the stand-alone
- preprocessor, the future direction of the C family front ends, and even
- to cpplib itself internally.
- </p>
- <p>Occasionally the preprocessor wants to be able to peek ahead in the
- token stream. For example, after the name of a function-like macro, it
- wants to check the next token to see if it is an opening parenthesis.
- Another example is that, after reading the first few tokens of a
- <code>#pragma</code> directive and not recognizing it as a registered pragma,
- it wants to backtrack and allow the user-defined handler for unknown
- pragmas to access the full <code>#pragma</code> token stream. The stand-alone
- preprocessor wants to be able to test the current token with the
- previous one to see if a space needs to be inserted to preserve their
- separate tokenization upon re-lexing (paste avoidance), so it needs to
- be sure the pointer to the previous token is still valid. The
- recursive-descent C++ parser wants to be able to perform tentative
- parsing arbitrarily far ahead in the token stream, and then to be able
- to jump back to a prior position in that stream if necessary.
- </p>
- <p>The rule I chose, which is fairly natural, is to arrange that the
- preprocessor lex all tokens on a line consecutively into a token buffer,
- which I call a <em>token run</em>, and when meeting an unescaped new line
- (newlines within comments do not count either), to start lexing back at
- the beginning of the run. Note that we do <em>not</em> lex a line of
- tokens at once; if we did that <code>parse_identifier</code> would not have
- state flags available to warn about invalid identifiers (see <a href="#Invalid-identifiers">Invalid identifiers</a>).
- </p>
- <p>In other words, accessing tokens that appeared earlier in the current
- line is valid, but since each logical line overwrites the tokens of the
- previous line, tokens from prior lines are unavailable. In particular,
- since a directive only occupies a single logical line, this means that
- the directive handlers like the <code>#pragma</code> handler can jump around
- in the directive’s tokens if necessary.
- </p>
- <p>Two issues remain: what about tokens that arise from macro expansions,
- and what happens when we have a long line that overflows the token run?
- </p>
- <p>Since we promise clients that we preserve the validity of pointers that
- we have already returned for tokens that appeared earlier in the line,
- we cannot reallocate the run. Instead, on overflow it is expanded by
- chaining a new token run on to the end of the existing one.
- </p>
- <p>The tokens forming a macro’s replacement list are collected by the
- <code>#define</code> handler, and placed in storage that is only freed by
- <code>cpp_destroy</code>. So if a macro is expanded in the line of tokens,
- the pointers to the tokens of its expansion that are returned will always
- remain valid. However, macros are a little trickier than that, since
- they give rise to three sources of fresh tokens. They are the built-in
- macros like <code>__LINE__</code>, and the ‘<samp>#</samp>’ and ‘<samp>##</samp>’ operators
- for stringizing and token pasting. I handled this by allocating
- space for these tokens from the lexer’s token run chain. This means
- they automatically receive the same lifetime guarantees as lexed tokens,
- and we don’t need to concern ourselves with freeing them.
- </p>
- <p>Lexing into a line of tokens solves some of the token memory management
- issues, but not all. The opening parenthesis after a function-like
- macro name might lie on a different line, and the front ends definitely
- want the ability to look ahead past the end of the current line. So
- cpplib only moves back to the start of the token run at the end of a
- line if the variable <code>keep_tokens</code> is zero. Line-buffering is
- quite natural for the preprocessor, and as a result the only time cpplib
- needs to increment this variable is whilst looking for the opening
- parenthesis to, and reading the arguments of, a function-like macro. In
- the near future cpplib will export an interface to increment and
- decrement this variable, so that clients can share full control over the
- lifetime of token pointers too.
- </p>
- <p>The routine <code>_cpp_lex_token</code> handles moving to new token runs,
- calling <code>_cpp_lex_direct</code> to lex new tokens, or returning
- previously-lexed tokens if we stepped back in the token stream. It also
- checks each token for the <code>BOL</code> flag, which might indicate a
- directive that needs to be handled, or require a start-of-line call-back
- to be made. <code>_cpp_lex_token</code> also handles skipping over tokens in
- failed conditional blocks, and invalidates the control macro of the
- multiple-include optimization if a token was successfully lexed outside
- a directive. In other words, its callers do not need to concern
- themselves with such issues.
- </p>
- <hr>
- <div class="header">
- <p>
- Next: <a href="Hash-Nodes.html#Hash-Nodes" accesskey="n" rel="next">Hash Nodes</a>, Previous: <a href="Conventions.html#Conventions" accesskey="p" rel="prev">Conventions</a>, Up: <a href="index.html#Top" accesskey="u" rel="up">Top</a> [<a href="index.html#SEC_Contents" title="Table of contents" rel="contents">Contents</a>][<a href="Concept-Index.html#Concept-Index" title="Index" rel="index">Index</a>]</p>
- </div>
-
-
-
- </body>
- </html>
|