Home | History | Annotate | Download | only in info
      1 This is cppinternals.info, produced by makeinfo version 5.2 from
      2 cppinternals.texi.
      3 
      4 INFO-DIR-SECTION Software development
      5 START-INFO-DIR-ENTRY
      6 * Cpplib: (cppinternals).      Cpplib internals.
      7 END-INFO-DIR-ENTRY
      8 
      9 This file documents the internals of the GNU C Preprocessor.
     10 
     11    Copyright (C) 2000-2013 Free Software Foundation, Inc.
     12 
     13    Permission is granted to make and distribute verbatim copies of this
     14 manual provided the copyright notice and this permission notice are
     15 preserved on all copies.
     16 
     17    Permission is granted to copy and distribute modified versions of
     18 this manual under the conditions for verbatim copying, provided also
     19 that the entire resulting derived work is distributed under the terms of
     20 a permission notice identical to this one.
     21 
     22    Permission is granted to copy and distribute translations of this
     23 manual into another language, under the above conditions for modified
     24 versions.
     25 
     26 
     27 File: cppinternals.info,  Node: Top,  Next: Conventions,  Up: (dir)
     28 
     29 The GNU C Preprocessor Internals
     30 ********************************
     31 
     32 1 Cpplib--the GNU C Preprocessor
     33 ********************************
     34 
     35 The GNU C preprocessor is implemented as a library, "cpplib", so it can
     36 be easily shared between a stand-alone preprocessor, and a preprocessor
     37 integrated with the C, C++ and Objective-C front ends.  It is also
     38 available for use by other programs, though this is not recommended as
     39 its exposed interface has not yet reached a point of reasonable
     40 stability.
     41 
     42    The library has been written to be re-entrant, so that it can be used
     43 to preprocess many files simultaneously if necessary.  It has also been
     44 written with the preprocessing token as the fundamental unit; the
     45 preprocessor in previous versions of GCC would operate on text strings
     46 as the fundamental unit.
     47 
     48    This brief manual documents the internals of cpplib, and explains
     49 some of the tricky issues.  It is intended that, along with the comments
     50 in the source code, a reasonably competent C programmer should be able
     51 to figure out what the code is doing, and why things have been
     52 implemented the way they have.
     53 
     54 * Menu:
     55 
     56 * Conventions::         Conventions used in the code.
     57 * Lexer::               The combined C, C++ and Objective-C Lexer.
     58 * Hash Nodes::          All identifiers are entered into a hash table.
     59 * Macro Expansion::     Macro expansion algorithm.
     60 * Token Spacing::       Spacing and paste avoidance issues.
     61 * Line Numbering::      Tracking location within files.
     62 * Guard Macros::        Optimizing header files with guard macros.
     63 * Files::               File handling.
     64 * Concept Index::       Index.
     65 
     66 
     67 File: cppinternals.info,  Node: Conventions,  Next: Lexer,  Prev: Top,  Up: Top
     68 
     69 Conventions
     70 ***********
     71 
     72 cpplib has two interfaces--one is exposed internally only, and the other
     73 is for both internal and external use.
     74 
     75    The convention is that functions and types that are exposed to
     76 multiple files internally are prefixed with '_cpp_', and are to be found
     77 in the file 'internal.h'.  Functions and types exposed to external
     78 clients are in 'cpplib.h', and prefixed with 'cpp_'.  For historical
     79 reasons this is no longer quite true, but we should strive to stick to
     80 it.
     81 
     82    We are striving to reduce the information exposed in 'cpplib.h' to
     83 the bare minimum necessary, and then to keep it there.  This makes clear
     84 exactly what external clients are entitled to assume, and allows us to
     85 change internals in the future without worrying whether library clients
     86 are perhaps relying on some kind of undocumented implementation-specific
     87 behavior.
     88 
     89 
     90 File: cppinternals.info,  Node: Lexer,  Next: Hash Nodes,  Prev: Conventions,  Up: Top
     91 
     92 The Lexer
     93 *********
     94 
     95 Overview
     96 ========
     97 
     98 The lexer is contained in the file 'lex.c'.  It is a hand-coded lexer,
     99 and not implemented as a state machine.  It can understand C, C++ and
    100 Objective-C source code, and has been extended to allow reasonably
    101 successful preprocessing of assembly language.  The lexer does not make
    102 an initial pass to strip out trigraphs and escaped newlines, but handles
    103 them as they are encountered in a single pass of the input file.  It
    104 returns preprocessing tokens individually, not a line at a time.
    105 
    106    It is mostly transparent to users of the library, since the library's
    107 interface for obtaining the next token, 'cpp_get_token', takes care of
    108 lexing new tokens, handling directives, and expanding macros as
    109 necessary.  However, the lexer does expose some functionality so that
    110 clients of the library can easily spell a given token, such as
    111 'cpp_spell_token' and 'cpp_token_len'.  These functions are useful when
    112 generating diagnostics, and for emitting the preprocessed output.
    113 
    114 Lexing a token
    115 ==============
    116 
    117 Lexing of an individual token is handled by '_cpp_lex_direct' and its
    118 subroutines.  In its current form the code is quite complicated, with
    119 read ahead characters and such-like, since it strives to not step back
    120 in the character stream in preparation for handling non-ASCII file
    121 encodings.  The current plan is to convert any such files to UTF-8
    122 before processing them.  This complexity is therefore unnecessary and
    123 will be removed, so I'll not discuss it further here.
    124 
    125    The job of '_cpp_lex_direct' is simply to lex a token.  It is not
    126 responsible for issues like directive handling, returning lookahead
    127 tokens directly, multiple-include optimization, or conditional block
    128 skipping.  It necessarily has a minor ro^le to play in memory management
    129 of lexed lines.  I discuss these issues in a separate section (*note
    130 Lexing a line::).
    131 
    132    The lexer places the token it lexes into storage pointed to by the
    133 variable 'cur_token', and then increments it.  This variable is
    134 important for correct diagnostic positioning.  Unless a specific line
    135 and column are passed to the diagnostic routines, they will examine the
    136 'line' and 'col' values of the token just before the location that
    137 'cur_token' points to, and use that location to report the diagnostic.
    138 
    139    The lexer does not consider whitespace to be a token in its own
    140 right.  If whitespace (other than a new line) precedes a token, it sets
    141 the 'PREV_WHITE' bit in the token's flags.  Each token has its 'line'
    142 and 'col' variables set to the line and column of the first character of
    143 the token.  This line number is the line number in the translation unit,
    144 and can be converted to a source (file, line) pair using the line map
    145 code.
    146 
    147    The first token on a logical, i.e. unescaped, line has the flag 'BOL'
    148 set for beginning-of-line.  This flag is intended for internal use, both
    149 to distinguish a '#' that begins a directive from one that doesn't, and
    150 to generate a call-back to clients that want to be notified about the
    151 start of every non-directive line with tokens on it.  Clients cannot
    152 reliably determine this for themselves: the first token might be a
    153 macro, and the tokens of a macro expansion do not have the 'BOL' flag
    154 set.  The macro expansion may even be empty, and the next token on the
    155 line certainly won't have the 'BOL' flag set.
    156 
    157    New lines are treated specially; exactly how the lexer handles them
    158 is context-dependent.  The C standard mandates that directives are
    159 terminated by the first unescaped newline character, even if it appears
    160 in the middle of a macro expansion.  Therefore, if the state variable
    161 'in_directive' is set, the lexer returns a 'CPP_EOF' token, which is
    162 normally used to indicate end-of-file, to indicate end-of-directive.  In
    163 a directive a 'CPP_EOF' token never means end-of-file.  Conveniently, if
    164 the caller was 'collect_args', it already handles 'CPP_EOF' as if it
    165 were end-of-file, and reports an error about an unterminated macro
    166 argument list.
    167 
    168    The C standard also specifies that a new line in the middle of the
    169 arguments to a macro is treated as whitespace.  This white space is
    170 important in case the macro argument is stringified.  The state variable
    171 'parsing_args' is nonzero when the preprocessor is collecting the
    172 arguments to a macro call.  It is set to 1 when looking for the opening
    173 parenthesis to a function-like macro, and 2 when collecting the actual
    174 arguments up to the closing parenthesis, since these two cases need to
    175 be distinguished sometimes.  One such time is here: the lexer sets the
    176 'PREV_WHITE' flag of a token if it meets a new line when 'parsing_args'
    177 is set to 2.  It doesn't set it if it meets a new line when
    178 'parsing_args' is 1, since then code like
    179 
    180      #define foo() bar
    181      foo
    182      baz
    183 
    184 would be output with an erroneous space before 'baz':
    185 
    186      foo
    187       baz
    188 
    189    This is a good example of the subtlety of getting token spacing
    190 correct in the preprocessor; there are plenty of tests in the testsuite
    191 for corner cases like this.
    192 
    193    The lexer is written to treat each of '\r', '\n', '\r\n' and '\n\r'
    194 as a single new line indicator.  This allows it to transparently
    195 preprocess MS-DOS, Macintosh and Unix files without their needing to
    196 pass through a special filter beforehand.
    197 
    198    We also decided to treat a backslash, either '\' or the trigraph
    199 '??/', separated from one of the above newline indicators by non-comment
    200 whitespace only, as intending to escape the newline.  It tends to be a
    201 typing mistake, and cannot reasonably be mistaken for anything else in
    202 any of the C-family grammars.  Since handling it this way is not
    203 strictly conforming to the ISO standard, the library issues a warning
    204 wherever it encounters it.
    205 
    206    Handling newlines like this is made simpler by doing it in one place
    207 only.  The function 'handle_newline' takes care of all newline
    208 characters, and 'skip_escaped_newlines' takes care of arbitrarily long
    209 sequences of escaped newlines, deferring to 'handle_newline' to handle
    210 the newlines themselves.
    211 
    212    The most painful aspect of lexing ISO-standard C and C++ is handling
    213 trigraphs and backlash-escaped newlines.  Trigraphs are processed before
    214 any interpretation of the meaning of a character is made, and
    215 unfortunately there is a trigraph representation for a backslash, so it
    216 is possible for the trigraph '??/' to introduce an escaped newline.
    217 
    218    Escaped newlines are tedious because theoretically they can occur
    219 anywhere--between the '+' and '=' of the '+=' token, within the
    220 characters of an identifier, and even between the '*' and '/' that
    221 terminates a comment.  Moreover, you cannot be sure there is just
    222 one--there might be an arbitrarily long sequence of them.
    223 
    224    So, for example, the routine that lexes a number, 'parse_number',
    225 cannot assume that it can scan forwards until the first non-number
    226 character and be done with it, because this could be the '\' introducing
    227 an escaped newline, or the '?' introducing the trigraph sequence that
    228 represents the '\' of an escaped newline.  If it encounters a '?' or
    229 '\', it calls 'skip_escaped_newlines' to skip over any potential escaped
    230 newlines before checking whether the number has been finished.
    231 
    232    Similarly code in the main body of '_cpp_lex_direct' cannot simply
    233 check for a '=' after a '+' character to determine whether it has a '+='
    234 token; it needs to be prepared for an escaped newline of some sort.
    235 Such cases use the function 'get_effective_char', which returns the
    236 first character after any intervening escaped newlines.
    237 
    238    The lexer needs to keep track of the correct column position,
    239 including counting tabs as specified by the '-ftabstop=' option.  This
    240 should be done even within C-style comments; they can appear in the
    241 middle of a line, and we want to report diagnostics in the correct
    242 position for text appearing after the end of the comment.
    243 
    244    Some identifiers, such as '__VA_ARGS__' and poisoned identifiers, may
    245 be invalid and require a diagnostic.  However, if they appear in a macro
    246 expansion we don't want to complain with each use of the macro.  It is
    247 therefore best to catch them during the lexing stage, in
    248 'parse_identifier'.  In both cases, whether a diagnostic is needed or
    249 not is dependent upon the lexer's state.  For example, we don't want to
    250 issue a diagnostic for re-poisoning a poisoned identifier, or for using
    251 '__VA_ARGS__' in the expansion of a variable-argument macro.  Therefore
    252 'parse_identifier' makes use of state flags to determine whether a
    253 diagnostic is appropriate.  Since we change state on a per-token basis,
    254 and don't lex whole lines at a time, this is not a problem.
    255 
    256    Another place where state flags are used to change behavior is whilst
    257 lexing header names.  Normally, a '<' would be lexed as a single token.
    258 After a '#include' directive, though, it should be lexed as a single
    259 token as far as the nearest '>' character.  Note that we don't allow the
    260 terminators of header names to be escaped; the first '"' or '>'
    261 terminates the header name.
    262 
    263    Interpretation of some character sequences depends upon whether we
    264 are lexing C, C++ or Objective-C, and on the revision of the standard in
    265 force.  For example, '::' is a single token in C++, but in C it is two
    266 separate ':' tokens and almost certainly a syntax error.  Such cases are
    267 handled by '_cpp_lex_direct' based upon command-line flags stored in the
    268 'cpp_options' structure.
    269 
    270    Once a token has been lexed, it leads an independent existence.  The
    271 spelling of numbers, identifiers and strings is copied to permanent
    272 storage from the original input buffer, so a token remains valid and
    273 correct even if its source buffer is freed with '_cpp_pop_buffer'.  The
    274 storage holding the spellings of such tokens remains until the client
    275 program calls cpp_destroy, probably at the end of the translation unit.
    276 
    277 Lexing a line
    278 =============
    279 
    280 When the preprocessor was changed to return pointers to tokens, one
    281 feature I wanted was some sort of guarantee regarding how long a
    282 returned pointer remains valid.  This is important to the stand-alone
    283 preprocessor, the future direction of the C family front ends, and even
    284 to cpplib itself internally.
    285 
    286    Occasionally the preprocessor wants to be able to peek ahead in the
    287 token stream.  For example, after the name of a function-like macro, it
    288 wants to check the next token to see if it is an opening parenthesis.
    289 Another example is that, after reading the first few tokens of a
    290 '#pragma' directive and not recognizing it as a registered pragma, it
    291 wants to backtrack and allow the user-defined handler for unknown
    292 pragmas to access the full '#pragma' token stream.  The stand-alone
    293 preprocessor wants to be able to test the current token with the
    294 previous one to see if a space needs to be inserted to preserve their
    295 separate tokenization upon re-lexing (paste avoidance), so it needs to
    296 be sure the pointer to the previous token is still valid.  The
    297 recursive-descent C++ parser wants to be able to perform tentative
    298 parsing arbitrarily far ahead in the token stream, and then to be able
    299 to jump back to a prior position in that stream if necessary.
    300 
    301    The rule I chose, which is fairly natural, is to arrange that the
    302 preprocessor lex all tokens on a line consecutively into a token buffer,
    303 which I call a "token run", and when meeting an unescaped new line
    304 (newlines within comments do not count either), to start lexing back at
    305 the beginning of the run.  Note that we do _not_ lex a line of tokens at
    306 once; if we did that 'parse_identifier' would not have state flags
    307 available to warn about invalid identifiers (*note Invalid
    308 identifiers::).
    309 
    310    In other words, accessing tokens that appeared earlier in the current
    311 line is valid, but since each logical line overwrites the tokens of the
    312 previous line, tokens from prior lines are unavailable.  In particular,
    313 since a directive only occupies a single logical line, this means that
    314 the directive handlers like the '#pragma' handler can jump around in the
    315 directive's tokens if necessary.
    316 
    317    Two issues remain: what about tokens that arise from macro
    318 expansions, and what happens when we have a long line that overflows the
    319 token run?
    320 
    321    Since we promise clients that we preserve the validity of pointers
    322 that we have already returned for tokens that appeared earlier in the
    323 line, we cannot reallocate the run.  Instead, on overflow it is expanded
    324 by chaining a new token run on to the end of the existing one.
    325 
    326    The tokens forming a macro's replacement list are collected by the
    327 '#define' handler, and placed in storage that is only freed by
    328 'cpp_destroy'.  So if a macro is expanded in the line of tokens, the
    329 pointers to the tokens of its expansion that are returned will always
    330 remain valid.  However, macros are a little trickier than that, since
    331 they give rise to three sources of fresh tokens.  They are the built-in
    332 macros like '__LINE__', and the '#' and '##' operators for
    333 stringification and token pasting.  I handled this by allocating space
    334 for these tokens from the lexer's token run chain.  This means they
    335 automatically receive the same lifetime guarantees as lexed tokens, and
    336 we don't need to concern ourselves with freeing them.
    337 
    338    Lexing into a line of tokens solves some of the token memory
    339 management issues, but not all.  The opening parenthesis after a
    340 function-like macro name might lie on a different line, and the front
    341 ends definitely want the ability to look ahead past the end of the
    342 current line.  So cpplib only moves back to the start of the token run
    343 at the end of a line if the variable 'keep_tokens' is zero.
    344 Line-buffering is quite natural for the preprocessor, and as a result
    345 the only time cpplib needs to increment this variable is whilst looking
    346 for the opening parenthesis to, and reading the arguments of, a
    347 function-like macro.  In the near future cpplib will export an interface
    348 to increment and decrement this variable, so that clients can share full
    349 control over the lifetime of token pointers too.
    350 
    351    The routine '_cpp_lex_token' handles moving to new token runs,
    352 calling '_cpp_lex_direct' to lex new tokens, or returning
    353 previously-lexed tokens if we stepped back in the token stream.  It also
    354 checks each token for the 'BOL' flag, which might indicate a directive
    355 that needs to be handled, or require a start-of-line call-back to be
    356 made.  '_cpp_lex_token' also handles skipping over tokens in failed
    357 conditional blocks, and invalidates the control macro of the
    358 multiple-include optimization if a token was successfully lexed outside
    359 a directive.  In other words, its callers do not need to concern
    360 themselves with such issues.
    361 
    362 
    363 File: cppinternals.info,  Node: Hash Nodes,  Next: Macro Expansion,  Prev: Lexer,  Up: Top
    364 
    365 Hash Nodes
    366 **********
    367 
    368 When cpplib encounters an "identifier", it generates a hash code for it
    369 and stores it in the hash table.  By "identifier" we mean tokens with
    370 type 'CPP_NAME'; this includes identifiers in the usual C sense, as well
    371 as keywords, directive names, macro names and so on.  For example, all
    372 of 'pragma', 'int', 'foo' and '__GNUC__' are identifiers and hashed when
    373 lexed.
    374 
    375    Each node in the hash table contain various information about the
    376 identifier it represents.  For example, its length and type.  At any one
    377 time, each identifier falls into exactly one of three categories:
    378 
    379    * Macros
    380 
    381      These have been declared to be macros, either on the command line
    382      or with '#define'.  A few, such as '__TIME__' are built-ins entered
    383      in the hash table during initialization.  The hash node for a
    384      normal macro points to a structure with more information about the
    385      macro, such as whether it is function-like, how many arguments it
    386      takes, and its expansion.  Built-in macros are flagged as special,
    387      and instead contain an enum indicating which of the various
    388      built-in macros it is.
    389 
    390    * Assertions
    391 
    392      Assertions are in a separate namespace to macros.  To enforce this,
    393      cpp actually prepends a '#' character before hashing and entering
    394      it in the hash table.  An assertion's node points to a chain of
    395      answers to that assertion.
    396 
    397    * Void
    398 
    399      Everything else falls into this category--an identifier that is not
    400      currently a macro, or a macro that has since been undefined with
    401      '#undef'.
    402 
    403      When preprocessing C++, this category also includes the named
    404      operators, such as 'xor'.  In expressions these behave like the
    405      operators they represent, but in contexts where the spelling of a
    406      token matters they are spelt differently.  This spelling
    407      distinction is relevant when they are operands of the stringizing
    408      and pasting macro operators '#' and '##'.  Named operator hash
    409      nodes are flagged, both to catch the spelling distinction and to
    410      prevent them from being defined as macros.
    411 
    412    The same identifiers share the same hash node.  Since each identifier
    413 token, after lexing, contains a pointer to its hash node, this is used
    414 to provide rapid lookup of various information.  For example, when
    415 parsing a '#define' statement, CPP flags each argument's identifier hash
    416 node with the index of that argument.  This makes duplicated argument
    417 checking an O(1) operation for each argument.  Similarly, for each
    418 identifier in the macro's expansion, lookup to see if it is an argument,
    419 and which argument it is, is also an O(1) operation.  Further, each
    420 directive name, such as 'endif', has an associated directive enum stored
    421 in its hash node, so that directive lookup is also O(1).
    422 
    423 
    424 File: cppinternals.info,  Node: Macro Expansion,  Next: Token Spacing,  Prev: Hash Nodes,  Up: Top
    425 
    426 Macro Expansion Algorithm
    427 *************************
    428 
    429 Macro expansion is a tricky operation, fraught with nasty corner cases
    430 and situations that render what you thought was a nifty way to optimize
    431 the preprocessor's expansion algorithm wrong in quite subtle ways.
    432 
    433    I strongly recommend you have a good grasp of how the C and C++
    434 standards require macros to be expanded before diving into this section,
    435 let alone the code!.  If you don't have a clear mental picture of how
    436 things like nested macro expansion, stringification and token pasting
    437 are supposed to work, damage to your sanity can quickly result.
    438 
    439 Internal representation of macros
    440 =================================
    441 
    442 The preprocessor stores macro expansions in tokenized form.  This saves
    443 repeated lexing passes during expansion, at the cost of a small increase
    444 in memory consumption on average.  The tokens are stored contiguously in
    445 memory, so a pointer to the first one and a token count is all you need
    446 to get the replacement list of a macro.
    447 
    448    If the macro is a function-like macro the preprocessor also stores
    449 its parameters, in the form of an ordered list of pointers to the hash
    450 table entry of each parameter's identifier.  Further, in the macro's
    451 stored expansion each occurrence of a parameter is replaced with a
    452 special token of type 'CPP_MACRO_ARG'.  Each such token holds the index
    453 of the parameter it represents in the parameter list, which allows rapid
    454 replacement of parameters with their arguments during expansion.
    455 Despite this optimization it is still necessary to store the original
    456 parameters to the macro, both for dumping with e.g., '-dD', and to warn
    457 about non-trivial macro redefinitions when the parameter names have
    458 changed.
    459 
    460 Macro expansion overview
    461 ========================
    462 
    463 The preprocessor maintains a "context stack", implemented as a linked
    464 list of 'cpp_context' structures, which together represent the macro
    465 expansion state at any one time.  The 'struct cpp_reader' member
    466 variable 'context' points to the current top of this stack.  The top
    467 normally holds the unexpanded replacement list of the innermost macro
    468 under expansion, except when cpplib is about to pre-expand an argument,
    469 in which case it holds that argument's unexpanded tokens.
    470 
    471    When there are no macros under expansion, cpplib is in "base
    472 context".  All contexts other than the base context contain a contiguous
    473 list of tokens delimited by a starting and ending token.  When not in
    474 base context, cpplib obtains the next token from the list of the top
    475 context.  If there are no tokens left in the list, it pops that context
    476 off the stack, and subsequent ones if necessary, until an unexhausted
    477 context is found or it returns to base context.  In base context, cpplib
    478 reads tokens directly from the lexer.
    479 
    480    If it encounters an identifier that is both a macro and enabled for
    481 expansion, cpplib prepares to push a new context for that macro on the
    482 stack by calling the routine 'enter_macro_context'.  When this routine
    483 returns, the new context will contain the unexpanded tokens of the
    484 replacement list of that macro.  In the case of function-like macros,
    485 'enter_macro_context' also replaces any parameters in the replacement
    486 list, stored as 'CPP_MACRO_ARG' tokens, with the appropriate macro
    487 argument.  If the standard requires that the parameter be replaced with
    488 its expanded argument, the argument will have been fully macro expanded
    489 first.
    490 
    491    'enter_macro_context' also handles special macros like '__LINE__'.
    492 Although these macros expand to a single token which cannot contain any
    493 further macros, for reasons of token spacing (*note Token Spacing::) and
    494 simplicity of implementation, cpplib handles these special macros by
    495 pushing a context containing just that one token.
    496 
    497    The final thing that 'enter_macro_context' does before returning is
    498 to mark the macro disabled for expansion (except for special macros like
    499 '__TIME__').  The macro is re-enabled when its context is later popped
    500 from the context stack, as described above.  This strict ordering
    501 ensures that a macro is disabled whilst its expansion is being scanned,
    502 but that it is _not_ disabled whilst any arguments to it are being
    503 expanded.
    504 
    505 Scanning the replacement list for macros to expand
    506 ==================================================
    507 
    508 The C standard states that, after any parameters have been replaced with
    509 their possibly-expanded arguments, the replacement list is scanned for
    510 nested macros.  Further, any identifiers in the replacement list that
    511 are not expanded during this scan are never again eligible for expansion
    512 in the future, if the reason they were not expanded is that the macro in
    513 question was disabled.
    514 
    515    Clearly this latter condition can only apply to tokens resulting from
    516 argument pre-expansion.  Other tokens never have an opportunity to be
    517 re-tested for expansion.  It is possible for identifiers that are
    518 function-like macros to not expand initially but to expand during a
    519 later scan.  This occurs when the identifier is the last token of an
    520 argument (and therefore originally followed by a comma or a closing
    521 parenthesis in its macro's argument list), and when it replaces its
    522 parameter in the macro's replacement list, the subsequent token happens
    523 to be an opening parenthesis (itself possibly the first token of an
    524 argument).
    525 
    526    It is important to note that when cpplib reads the last token of a
    527 given context, that context still remains on the stack.  Only when
    528 looking for the _next_ token do we pop it off the stack and drop to a
    529 lower context.  This makes backing up by one token easy, but more
    530 importantly ensures that the macro corresponding to the current context
    531 is still disabled when we are considering the last token of its
    532 replacement list for expansion (or indeed expanding it).  As an example,
    533 which illustrates many of the points above, consider
    534 
    535      #define foo(x) bar x
    536      foo(foo) (2)
    537 
    538 which fully expands to 'bar foo (2)'.  During pre-expansion of the
    539 argument, 'foo' does not expand even though the macro is enabled, since
    540 it has no following parenthesis [pre-expansion of an argument only uses
    541 tokens from that argument; it cannot take tokens from whatever follows
    542 the macro invocation].  This still leaves the argument token 'foo'
    543 eligible for future expansion.  Then, when re-scanning after argument
    544 replacement, the token 'foo' is rejected for expansion, and marked
    545 ineligible for future expansion, since the macro is now disabled.  It is
    546 disabled because the replacement list 'bar foo' of the macro is still on
    547 the context stack.
    548 
    549    If instead the algorithm looked for an opening parenthesis first and
    550 then tested whether the macro were disabled it would be subtly wrong.
    551 In the example above, the replacement list of 'foo' would be popped in
    552 the process of finding the parenthesis, re-enabling 'foo' and expanding
    553 it a second time.
    554 
    555 Looking for a function-like macro's opening parenthesis
    556 =======================================================
    557 
    558 Function-like macros only expand when immediately followed by a
    559 parenthesis.  To do this cpplib needs to temporarily disable macros and
    560 read the next token.  Unfortunately, because of spacing issues (*note
    561 Token Spacing::), there can be fake padding tokens in-between, and if
    562 the next real token is not a parenthesis cpplib needs to be able to back
    563 up that one token as well as retain the information in any intervening
    564 padding tokens.
    565 
    566    Backing up more than one token when macros are involved is not
    567 permitted by cpplib, because in general it might involve issues like
    568 restoring popped contexts onto the context stack, which are too hard.
    569 Instead, searching for the parenthesis is handled by a special function,
    570 'funlike_invocation_p', which remembers padding information as it reads
    571 tokens.  If the next real token is not an opening parenthesis, it backs
    572 up that one token, and then pushes an extra context just containing the
    573 padding information if necessary.
    574 
    575 Marking tokens ineligible for future expansion
    576 ==============================================
    577 
    578 As discussed above, cpplib needs a way of marking tokens as
    579 unexpandable.  Since the tokens cpplib handles are read-only once they
    580 have been lexed, it instead makes a copy of the token and adds the flag
    581 'NO_EXPAND' to the copy.
    582 
    583    For efficiency and to simplify memory management by avoiding having
    584 to remember to free these tokens, they are allocated as temporary tokens
    585 from the lexer's current token run (*note Lexing a line::) using the
    586 function '_cpp_temp_token'.  The tokens are then re-used once the
    587 current line of tokens has been read in.
    588 
    589    This might sound unsafe.  However, tokens runs are not re-used at the
    590 end of a line if it happens to be in the middle of a macro argument
    591 list, and cpplib only wants to back-up more than one lexer token in
    592 situations where no macro expansion is involved, so the optimization is
    593 safe.
    594 
    595 
    596 File: cppinternals.info,  Node: Token Spacing,  Next: Line Numbering,  Prev: Macro Expansion,  Up: Top
    597 
    598 Token Spacing
    599 *************
    600 
    601 First, consider an issue that only concerns the stand-alone
    602 preprocessor: there needs to be a guarantee that re-reading its
    603 preprocessed output results in an identical token stream.  Without
    604 taking special measures, this might not be the case because of macro
    605 substitution.  For example:
    606 
    607      #define PLUS +
    608      #define EMPTY
    609      #define f(x) =x=
    610      +PLUS -EMPTY- PLUS+ f(=)
    611              ==> + + - - + + = = =
    612      _not_
    613              ==> ++ -- ++ ===
    614 
    615    One solution would be to simply insert a space between all adjacent
    616 tokens.  However, we would like to keep space insertion to a minimum,
    617 both for aesthetic reasons and because it causes problems for people who
    618 still try to abuse the preprocessor for things like Fortran source and
    619 Makefiles.
    620 
    621    For now, just notice that when tokens are added (or removed, as shown
    622 by the 'EMPTY' example) from the original lexed token stream, we need to
    623 check for accidental token pasting.  We call this "paste avoidance".
    624 Token addition and removal can only occur because of macro expansion,
    625 but accidental pasting can occur in many places: both before and after
    626 each macro replacement, each argument replacement, and additionally each
    627 token created by the '#' and '##' operators.
    628 
    629    Look at how the preprocessor gets whitespace output correct normally.
    630 The 'cpp_token' structure contains a flags byte, and one of those flags
    631 is 'PREV_WHITE'.  This is flagged by the lexer, and indicates that the
    632 token was preceded by whitespace of some form other than a new line.
    633 The stand-alone preprocessor can use this flag to decide whether to
    634 insert a space between tokens in the output.
    635 
    636    Now consider the result of the following macro expansion:
    637 
    638      #define add(x, y, z) x + y +z;
    639      sum = add (1,2, 3);
    640              ==> sum = 1 + 2 +3;
    641 
    642    The interesting thing here is that the tokens '1' and '2' are output
    643 with a preceding space, and '3' is output without a preceding space, but
    644 when lexed none of these tokens had that property.  Careful
    645 consideration reveals that '1' gets its preceding whitespace from the
    646 space preceding 'add' in the macro invocation, _not_ replacement list.
    647 '2' gets its whitespace from the space preceding the parameter 'y' in
    648 the macro replacement list, and '3' has no preceding space because
    649 parameter 'z' has none in the replacement list.
    650 
    651    Once lexed, tokens are effectively fixed and cannot be altered, since
    652 pointers to them might be held in many places, in particular by
    653 in-progress macro expansions.  So instead of modifying the two tokens
    654 above, the preprocessor inserts a special token, which I call a "padding
    655 token", into the token stream to indicate that spacing of the subsequent
    656 token is special.  The preprocessor inserts padding tokens in front of
    657 every macro expansion and expanded macro argument.  These point to a
    658 "source token" from which the subsequent real token should inherit its
    659 spacing.  In the above example, the source tokens are 'add' in the macro
    660 invocation, and 'y' and 'z' in the macro replacement list, respectively.
    661 
    662    It is quite easy to get multiple padding tokens in a row, for example
    663 if a macro's first replacement token expands straight into another
    664 macro.
    665 
    666      #define foo bar
    667      #define bar baz
    668      [foo]
    669              ==> [baz]
    670 
    671    Here, two padding tokens are generated with sources the 'foo' token
    672 between the brackets, and the 'bar' token from foo's replacement list,
    673 respectively.  Clearly the first padding token is the one to use, so the
    674 output code should contain a rule that the first padding token in a
    675 sequence is the one that matters.
    676 
    677    But what if a macro expansion is left?  Adjusting the above example
    678 slightly:
    679 
    680      #define foo bar
    681      #define bar EMPTY baz
    682      #define EMPTY
    683      [foo] EMPTY;
    684              ==> [ baz] ;
    685 
    686    As shown, now there should be a space before 'baz' and the semicolon
    687 in the output.
    688 
    689    The rules we decided above fail for 'baz': we generate three padding
    690 tokens, one per macro invocation, before the token 'baz'.  We would then
    691 have it take its spacing from the first of these, which carries source
    692 token 'foo' with no leading space.
    693 
    694    It is vital that cpplib get spacing correct in these examples since
    695 any of these macro expansions could be stringified, where spacing
    696 matters.
    697 
    698    So, this demonstrates that not just entering macro and argument
    699 expansions, but leaving them requires special handling too.  I made
    700 cpplib insert a padding token with a 'NULL' source token when leaving
    701 macro expansions, as well as after each replaced argument in a macro's
    702 replacement list.  It also inserts appropriate padding tokens on either
    703 side of tokens created by the '#' and '##' operators.  I expanded the
    704 rule so that, if we see a padding token with a 'NULL' source token,
    705 _and_ that source token has no leading space, then we behave as if we
    706 have seen no padding tokens at all.  A quick check shows this rule will
    707 then get the above example correct as well.
    708 
    709    Now a relationship with paste avoidance is apparent: we have to be
    710 careful about paste avoidance in exactly the same locations we have
    711 padding tokens in order to get white space correct.  This makes
    712 implementation of paste avoidance easy: wherever the stand-alone
    713 preprocessor is fixing up spacing because of padding tokens, and it
    714 turns out that no space is needed, it has to take the extra step to
    715 check that a space is not needed after all to avoid an accidental paste.
    716 The function 'cpp_avoid_paste' advises whether a space is required
    717 between two consecutive tokens.  To avoid excessive spacing, it tries
    718 hard to only require a space if one is likely to be necessary, but for
    719 reasons of efficiency it is slightly conservative and might recommend a
    720 space where one is not strictly needed.
    721 
    722 
    723 File: cppinternals.info,  Node: Line Numbering,  Next: Guard Macros,  Prev: Token Spacing,  Up: Top
    724 
    725 Line numbering
    726 **************
    727 
    728 Just which line number anyway?
    729 ==============================
    730 
    731 There are three reasonable requirements a cpplib client might have for
    732 the line number of a token passed to it:
    733 
    734    * The source line it was lexed on.
    735    * The line it is output on.  This can be different to the line it was
    736      lexed on if, for example, there are intervening escaped newlines or
    737      C-style comments.  For example:
    738 
    739           foo /* A long
    740           comment */ bar \
    741           baz
    742           =>
    743           foo bar baz
    744 
    745    * If the token results from a macro expansion, the line of the macro
    746      name, or possibly the line of the closing parenthesis in the case
    747      of function-like macro expansion.
    748 
    749    The 'cpp_token' structure contains 'line' and 'col' members.  The
    750 lexer fills these in with the line and column of the first character of
    751 the token.  Consequently, but maybe unexpectedly, a token from the
    752 replacement list of a macro expansion carries the location of the token
    753 within the '#define' directive, because cpplib expands a macro by
    754 returning pointers to the tokens in its replacement list.  The current
    755 implementation of cpplib assigns tokens created from built-in macros and
    756 the '#' and '##' operators the location of the most recently lexed
    757 token.  This is a because they are allocated from the lexer's token
    758 runs, and because of the way the diagnostic routines infer the
    759 appropriate location to report.
    760 
    761    The diagnostic routines in cpplib display the location of the most
    762 recently _lexed_ token, unless they are passed a specific line and
    763 column to report.  For diagnostics regarding tokens that arise from
    764 macro expansions, it might also be helpful for the user to see the
    765 original location in the macro definition that the token came from.
    766 Since that is exactly the information each token carries, such an
    767 enhancement could be made relatively easily in future.
    768 
    769    The stand-alone preprocessor faces a similar problem when determining
    770 the correct line to output the token on: the position attached to a
    771 token is fairly useless if the token came from a macro expansion.  All
    772 tokens on a logical line should be output on its first physical line, so
    773 the token's reported location is also wrong if it is part of a physical
    774 line other than the first.
    775 
    776    To solve these issues, cpplib provides a callback that is generated
    777 whenever it lexes a preprocessing token that starts a new logical line
    778 other than a directive.  It passes this token (which may be a 'CPP_EOF'
    779 token indicating the end of the translation unit) to the callback
    780 routine, which can then use the line and column of this token to produce
    781 correct output.
    782 
    783 Representation of line numbers
    784 ==============================
    785 
    786 As mentioned above, cpplib stores with each token the line number that
    787 it was lexed on.  In fact, this number is not the number of the line in
    788 the source file, but instead bears more resemblance to the number of the
    789 line in the translation unit.
    790 
    791    The preprocessor maintains a monotonic increasing line count, which
    792 is incremented at every new line character (and also at the end of any
    793 buffer that does not end in a new line).  Since a line number of zero is
    794 useful to indicate certain special states and conditions, this variable
    795 starts counting from one.
    796 
    797    This variable therefore uniquely enumerates each line in the
    798 translation unit.  With some simple infrastructure, it is straight
    799 forward to map from this to the original source file and line number
    800 pair, saving space whenever line number information needs to be saved.
    801 The code the implements this mapping lies in the files 'line-map.c' and
    802 'line-map.h'.
    803 
    804    Command-line macros and assertions are implemented by pushing a
    805 buffer containing the right hand side of an equivalent '#define' or
    806 '#assert' directive.  Some built-in macros are handled similarly.  Since
    807 these are all processed before the first line of the main input file, it
    808 will typically have an assigned line closer to twenty than to one.
    809 
    810 
    811 File: cppinternals.info,  Node: Guard Macros,  Next: Files,  Prev: Line Numbering,  Up: Top
    812 
    813 The Multiple-Include Optimization
    814 *********************************
    815 
    816 Header files are often of the form
    817 
    818      #ifndef FOO
    819      #define FOO
    820      ...
    821      #endif
    822 
    823 to prevent the compiler from processing them more than once.  The
    824 preprocessor notices such header files, so that if the header file
    825 appears in a subsequent '#include' directive and 'FOO' is defined, then
    826 it is ignored and it doesn't preprocess or even re-open the file a
    827 second time.  This is referred to as the "multiple include
    828 optimization".
    829 
    830    Under what circumstances is such an optimization valid?  If the file
    831 were included a second time, it can only be optimized away if that
    832 inclusion would result in no tokens to return, and no relevant
    833 directives to process.  Therefore the current implementation imposes
    834 requirements and makes some allowances as follows:
    835 
    836   1. There must be no tokens outside the controlling '#if'-'#endif'
    837      pair, but whitespace and comments are permitted.
    838 
    839   2. There must be no directives outside the controlling directive pair,
    840      but the "null directive" (a line containing nothing other than a
    841      single '#' and possibly whitespace) is permitted.
    842 
    843   3. The opening directive must be of the form
    844 
    845           #ifndef FOO
    846 
    847      or
    848 
    849           #if !defined FOO     [equivalently, #if !defined(FOO)]
    850 
    851   4. In the second form above, the tokens forming the '#if' expression
    852      must have come directly from the source file--no macro expansion
    853      must have been involved.  This is because macro definitions can
    854      change, and tracking whether or not a relevant change has been made
    855      is not worth the implementation cost.
    856 
    857   5. There can be no '#else' or '#elif' directives at the outer
    858      conditional block level, because they would probably contain
    859      something of interest to a subsequent pass.
    860 
    861    First, when pushing a new file on the buffer stack,
    862 '_stack_include_file' sets the controlling macro 'mi_cmacro' to 'NULL',
    863 and sets 'mi_valid' to 'true'.  This indicates that the preprocessor has
    864 not yet encountered anything that would invalidate the multiple-include
    865 optimization.  As described in the next few paragraphs, these two
    866 variables having these values effectively indicates top-of-file.
    867 
    868    When about to return a token that is not part of a directive,
    869 '_cpp_lex_token' sets 'mi_valid' to 'false'.  This enforces the
    870 constraint that tokens outside the controlling conditional block
    871 invalidate the optimization.
    872 
    873    The 'do_if', when appropriate, and 'do_ifndef' directive handlers
    874 pass the controlling macro to the function 'push_conditional'.  cpplib
    875 maintains a stack of nested conditional blocks, and after processing
    876 every opening conditional this function pushes an 'if_stack' structure
    877 onto the stack.  In this structure it records the controlling macro for
    878 the block, provided there is one and we're at top-of-file (as described
    879 above).  If an '#elif' or '#else' directive is encountered, the
    880 controlling macro for that block is cleared to 'NULL'.  Otherwise, it
    881 survives until the '#endif' closing the block, upon which 'do_endif'
    882 sets 'mi_valid' to true and stores the controlling macro in 'mi_cmacro'.
    883 
    884    '_cpp_handle_directive' clears 'mi_valid' when processing any
    885 directive other than an opening conditional and the null directive.
    886 With this, and requiring top-of-file to record a controlling macro, and
    887 no '#else' or '#elif' for it to survive and be copied to 'mi_cmacro' by
    888 'do_endif', we have enforced the absence of directives outside the main
    889 conditional block for the optimization to be on.
    890 
    891    Note that whilst we are inside the conditional block, 'mi_valid' is
    892 likely to be reset to 'false', but this does not matter since the
    893 closing '#endif' restores it to 'true' if appropriate.
    894 
    895    Finally, since '_cpp_lex_direct' pops the file off the buffer stack
    896 at 'EOF' without returning a token, if the '#endif' directive was not
    897 followed by any tokens, 'mi_valid' is 'true' and '_cpp_pop_file_buffer'
    898 remembers the controlling macro associated with the file.  Subsequent
    899 calls to 'stack_include_file' result in no buffer being pushed if the
    900 controlling macro is defined, effecting the optimization.
    901 
    902    A quick word on how we handle the
    903 
    904      #if !defined FOO
    905 
    906 case.  '_cpp_parse_expr' and 'parse_defined' take steps to see whether
    907 the three stages '!', 'defined-expression' and 'end-of-directive' occur
    908 in order in a '#if' expression.  If so, they return the guard macro to
    909 'do_if' in the variable 'mi_ind_cmacro', and otherwise set it to 'NULL'.
    910 'enter_macro_context' sets 'mi_valid' to false, so if a macro was
    911 expanded whilst parsing any part of the expression, then the top-of-file
    912 test in 'push_conditional' fails and the optimization is turned off.
    913 
    914 
    915 File: cppinternals.info,  Node: Files,  Next: Concept Index,  Prev: Guard Macros,  Up: Top
    916 
    917 File Handling
    918 *************
    919 
    920 Fairly obviously, the file handling code of cpplib resides in the file
    921 'files.c'.  It takes care of the details of file searching, opening,
    922 reading and caching, for both the main source file and all the headers
    923 it recursively includes.
    924 
    925    The basic strategy is to minimize the number of system calls.  On
    926 many systems, the basic 'open ()' and 'fstat ()' system calls can be
    927 quite expensive.  For every '#include'-d file, we need to try all the
    928 directories in the search path until we find a match.  Some projects,
    929 such as glibc, pass twenty or thirty include paths on the command line,
    930 so this can rapidly become time consuming.
    931 
    932    For a header file we have not encountered before we have little
    933 choice but to do this.  However, it is often the case that the same
    934 headers are repeatedly included, and in these cases we try to avoid
    935 repeating the filesystem queries whilst searching for the correct file.
    936 
    937    For each file we try to open, we store the constructed path in a
    938 splay tree.  This path first undergoes simplification by the function
    939 '_cpp_simplify_pathname'.  For example, '/usr/include/bits/../foo.h' is
    940 simplified to '/usr/include/foo.h' before we enter it in the splay tree
    941 and try to 'open ()' the file.  CPP will then find subsequent uses of
    942 'foo.h', even as '/usr/include/foo.h', in the splay tree and save system
    943 calls.
    944 
    945    Further, it is likely the file contents have also been cached, saving
    946 a 'read ()' system call.  We don't bother caching the contents of header
    947 files that are re-inclusion protected, and whose re-inclusion macro is
    948 defined when we leave the header file for the first time.  If the host
    949 supports it, we try to map suitably large files into memory, rather than
    950 reading them in directly.
    951 
    952    The include paths are internally stored on a null-terminated
    953 singly-linked list, starting with the '"header.h"' directory search
    954 chain, which then links into the '<header.h>' directory chain.
    955 
    956    Files included with the '<foo.h>' syntax start the lookup directly in
    957 the second half of this chain.  However, files included with the
    958 '"foo.h"' syntax start at the beginning of the chain, but with one extra
    959 directory prepended.  This is the directory of the current file; the one
    960 containing the '#include' directive.  Prepending this directory on a
    961 per-file basis is handled by the function 'search_from'.
    962 
    963    Note that a header included with a directory component, such as
    964 '#include "mydir/foo.h"' and opened as '/usr/local/include/mydir/foo.h',
    965 will have the complete path minus the basename 'foo.h' as the current
    966 directory.
    967 
    968    Enough information is stored in the splay tree that CPP can
    969 immediately tell whether it can skip the header file because of the
    970 multiple include optimization, whether the file didn't exist or couldn't
    971 be opened for some reason, or whether the header was flagged not to be
    972 re-used, as it is with the obsolete '#import' directive.
    973 
    974    For the benefit of MS-DOS filesystems with an 8.3 filename
    975 limitation, CPP offers the ability to treat various include file names
    976 as aliases for the real header files with shorter names.  The map from
    977 one to the other is found in a special file called 'header.gcc', stored
    978 in the command line (or system) include directories to which the mapping
    979 applies.  This may be higher up the directory tree than the full path to
    980 the file minus the base name.
    981 
    982 
    983 File: cppinternals.info,  Node: Concept Index,  Prev: Files,  Up: Top
    984 
    985 Concept Index
    986 *************
    987 
    988 [index]
    989 * Menu:
    990 
    991 * assertions:                            Hash Nodes.          (line   6)
    992 * controlling macros:                    Guard Macros.        (line   6)
    993 * escaped newlines:                      Lexer.               (line   5)
    994 * files:                                 Files.               (line   6)
    995 * guard macros:                          Guard Macros.        (line   6)
    996 * hash table:                            Hash Nodes.          (line   6)
    997 * header files:                          Conventions.         (line   6)
    998 * identifiers:                           Hash Nodes.          (line   6)
    999 * interface:                             Conventions.         (line   6)
   1000 * lexer:                                 Lexer.               (line   6)
   1001 * line numbers:                          Line Numbering.      (line   5)
   1002 * macro expansion:                       Macro Expansion.     (line   6)
   1003 * macro representation (internal):       Macro Expansion.     (line  19)
   1004 * macros:                                Hash Nodes.          (line   6)
   1005 * multiple-include optimization:         Guard Macros.        (line   6)
   1006 * named operators:                       Hash Nodes.          (line   6)
   1007 * newlines:                              Lexer.               (line   6)
   1008 * paste avoidance:                       Token Spacing.       (line   6)
   1009 * spacing:                               Token Spacing.       (line   6)
   1010 * token run:                             Lexer.               (line 191)
   1011 * token spacing:                         Token Spacing.       (line   6)
   1012 
   1013 
   1014 
   1015 Tag Table:
   1016 Node: Top905
   1017 Node: Conventions2590
   1018 Node: Lexer3532
   1019 Ref: Invalid identifiers11447
   1020 Ref: Lexing a line13397
   1021 Node: Hash Nodes18170
   1022 Node: Macro Expansion21049
   1023 Node: Token Spacing29997
   1024 Node: Line Numbering35854
   1025 Node: Guard Macros39939
   1026 Node: Files44730
   1027 Node: Concept Index48196
   1028 
   1029 End Tag Table
   1030