3598 lines
104 KiB
C
3598 lines
104 KiB
C
/* CPP Library - lexical analysis.
|
|
Copyright (C) 2000 Free Software Foundation, Inc.
|
|
Contributed by Per Bothner, 1994-95.
|
|
Based on CCCP program by Paul Rubin, June 1986
|
|
Adapted to ANSI C, Richard Stallman, Jan 1987
|
|
Broken out to separate file, Zack Weinberg, Mar 2000
|
|
Single-pass line tokenization by Neil Booth, April 2000
|
|
|
|
This program is free software; you can redistribute it and/or modify it
|
|
under the terms of the GNU General Public License as published by the
|
|
Free Software Foundation; either version 2, or (at your option) any
|
|
later version.
|
|
|
|
This program is distributed in the hope that it will be useful,
|
|
but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
GNU General Public License for more details.
|
|
|
|
You should have received a copy of the GNU General Public License
|
|
along with this program; if not, write to the Free Software
|
|
Foundation, 59 Temple Place - Suite 330, Boston, MA 02111-1307, USA. */
|
|
|
|
/*
|
|
|
|
Cleanups to do:-
|
|
|
|
o -dM and with _cpp_dump_list: too many \n output.
|
|
o Put a printer object in cpp_reader?
|
|
o Check line numbers assigned to all errors.
|
|
o Replace strncmp with memcmp almost everywhere.
|
|
o lex_line's use of cur_token, flags and list->token_used is a bit opaque.
|
|
o Convert do_ functions to return void. Kaveh thinks its OK; and said he'll
|
|
give it a run when we've got some code.
|
|
o Distinguish integers, floats, and 'other' pp-numbers.
|
|
o Store ints and char constants as binary values.
|
|
o New command-line assertion syntax.
|
|
o Work towards functions in cpperror.c taking a message level parameter.
|
|
If we do this, merge the common code of do_warning and do_error.
|
|
o Comment all functions, and describe macro expansion algorithm.
|
|
o Move as much out of header files as possible.
|
|
o Remove single quote pairs `', and some '', from diagnostics.
|
|
o Correct pastability test for CPP_NAME and CPP_NUMBER.
|
|
|
|
*/
|
|
|
|
#include "config.h"
|
|
#include "system.h"
|
|
#include "intl.h"
|
|
#include "cpplib.h"
|
|
#include "cpphash.h"
|
|
#include "symcat.h"
|
|
|
|
static const cpp_token placemarker_token = {0, 0, CPP_PLACEMARKER, 0 UNION_INIT_ZERO};
|
|
static const cpp_token eof_token = {0, 0, CPP_EOF, 0 UNION_INIT_ZERO};
|
|
|
|
/* Flags for cpp_context. */
|
|
#define CONTEXT_PASTEL (1 << 0) /* An argument context on LHS of ##. */
|
|
#define CONTEXT_PASTER (1 << 1) /* An argument context on RHS of ##. */
|
|
#define CONTEXT_RAW (1 << 2) /* If argument tokens already expanded. */
|
|
#define CONTEXT_ARG (1 << 3) /* If an argument context. */
|
|
|
|
typedef struct cpp_context cpp_context;
|
|
struct cpp_context
|
|
{
|
|
union
|
|
{
|
|
const cpp_toklist *list; /* Used for macro contexts only. */
|
|
const cpp_token **arg; /* Used for arg contexts only. */
|
|
} u;
|
|
|
|
/* Pushed token to be returned by next call to get_raw_token. */
|
|
const cpp_token *pushed_token;
|
|
|
|
struct macro_args *args; /* The arguments for a function-like
|
|
macro. NULL otherwise. */
|
|
unsigned short posn; /* Current posn, index into u. */
|
|
unsigned short count; /* No. of tokens in u. */
|
|
unsigned short level;
|
|
unsigned char flags;
|
|
};
|
|
|
|
typedef struct macro_args macro_args;
|
|
struct macro_args
|
|
{
|
|
unsigned int *ends;
|
|
const cpp_token **tokens;
|
|
unsigned int capacity;
|
|
unsigned int used;
|
|
unsigned short level;
|
|
};
|
|
|
|
static const cpp_token *get_raw_token PARAMS ((cpp_reader *));
|
|
static const cpp_token *parse_arg PARAMS ((cpp_reader *, int, unsigned int,
|
|
macro_args *, unsigned int *));
|
|
static int parse_args PARAMS ((cpp_reader *, cpp_hashnode *, macro_args *));
|
|
static void save_token PARAMS ((macro_args *, const cpp_token *));
|
|
static int pop_context PARAMS ((cpp_reader *));
|
|
static int push_macro_context PARAMS ((cpp_reader *, const cpp_token *));
|
|
static void push_arg_context PARAMS ((cpp_reader *, const cpp_token *));
|
|
static void free_macro_args PARAMS ((macro_args *));
|
|
|
|
#define auto_expand_name_space(list) \
|
|
_cpp_expand_name_space ((list), 1 + (list)->name_cap / 2)
|
|
static void safe_fwrite PARAMS ((cpp_reader *, const U_CHAR *,
|
|
size_t, FILE *));
|
|
static void dump_param_spelling PARAMS ((cpp_reader *, const cpp_toklist *,
|
|
unsigned int));
|
|
static void output_line_command PARAMS ((cpp_reader *, cpp_printer *,
|
|
unsigned int));
|
|
|
|
static void process_directive PARAMS ((cpp_reader *, const cpp_token *));
|
|
static unsigned char *trigraph_replace PARAMS ((cpp_reader *, unsigned char *,
|
|
unsigned char *));
|
|
static const unsigned char *backslash_start PARAMS ((cpp_reader *,
|
|
const unsigned char *));
|
|
static int skip_block_comment PARAMS ((cpp_reader *));
|
|
static int skip_line_comment PARAMS ((cpp_reader *));
|
|
static void adjust_column PARAMS ((cpp_reader *, const U_CHAR *));
|
|
static void skip_whitespace PARAMS ((cpp_reader *, int));
|
|
static const U_CHAR *parse_name PARAMS ((cpp_reader *, cpp_token *,
|
|
const U_CHAR *, const U_CHAR *));
|
|
static void parse_number PARAMS ((cpp_reader *, cpp_toklist *, cpp_string *));
|
|
static void parse_string PARAMS ((cpp_reader *, cpp_toklist *, cpp_token *,
|
|
unsigned int));
|
|
static int trigraph_ok PARAMS ((cpp_reader *, const unsigned char *));
|
|
static void save_comment PARAMS ((cpp_toklist *, cpp_token *,
|
|
const unsigned char *,
|
|
unsigned int, unsigned int));
|
|
static void lex_line PARAMS ((cpp_reader *, cpp_toklist *));
|
|
static int lex_next PARAMS ((cpp_reader *, int));
|
|
static int is_macro_disabled PARAMS ((cpp_reader *, const cpp_toklist *,
|
|
const cpp_token *));
|
|
|
|
static cpp_token *stringify_arg PARAMS ((cpp_reader *, const cpp_token *));
|
|
static void expand_context_stack PARAMS ((cpp_reader *));
|
|
static unsigned char * spell_token PARAMS ((cpp_reader *, const cpp_token *,
|
|
unsigned char *));
|
|
static void output_token PARAMS ((cpp_reader *, const cpp_token *,
|
|
const cpp_token *));
|
|
typedef unsigned int (* speller) PARAMS ((unsigned char *, cpp_toklist *,
|
|
cpp_token *));
|
|
static cpp_token *make_string_token PARAMS ((cpp_token *, const U_CHAR *,
|
|
unsigned int));
|
|
static cpp_token *alloc_number_token PARAMS ((cpp_reader *, int number));
|
|
static const cpp_token *special_symbol PARAMS ((cpp_reader *, cpp_hashnode *,
|
|
const cpp_token *));
|
|
static cpp_token *duplicate_token PARAMS ((cpp_reader *, const cpp_token *));
|
|
static const cpp_token *maybe_paste_with_next PARAMS ((cpp_reader *,
|
|
const cpp_token *));
|
|
static enum cpp_ttype can_paste PARAMS ((cpp_reader *, const cpp_token *,
|
|
const cpp_token *, int *));
|
|
static unsigned int prevent_macro_expansion PARAMS ((cpp_reader *));
|
|
static void restore_macro_expansion PARAMS ((cpp_reader *, unsigned int));
|
|
static cpp_token *get_temp_token PARAMS ((cpp_reader *));
|
|
static void release_temp_tokens PARAMS ((cpp_reader *));
|
|
static U_CHAR * quote_string PARAMS ((U_CHAR *, const U_CHAR *, unsigned int));
|
|
static void process_directive PARAMS ((cpp_reader *, const cpp_token *));
|
|
|
|
#define INIT_TOKEN_STR(list, token) \
|
|
do {(token)->val.str.len = 0; \
|
|
(token)->val.str.text = (list)->namebuf + (list)->name_used; \
|
|
} while (0)
|
|
|
|
#define VALID_SIGN(c, prevc) \
|
|
(((c) == '+' || (c) == '-') && \
|
|
((prevc) == 'e' || (prevc) == 'E' \
|
|
|| (((prevc) == 'p' || (prevc) == 'P') && !CPP_OPTION (pfile, c89))))
|
|
|
|
/* Handle LF, CR, CR-LF and LF-CR style newlines. Assumes next
|
|
character, if any, is in buffer. */
|
|
|
|
#define handle_newline(cur, limit, c) \
|
|
do { \
|
|
if ((cur) < (limit) && *(cur) == '\r' + '\n' - c) \
|
|
(cur)++; \
|
|
pfile->buffer->lineno++; \
|
|
pfile->buffer->line_base = (cur); \
|
|
pfile->col_adjust = 0; \
|
|
} while (0)
|
|
|
|
#define IMMED_TOKEN() (!(cur_token->flags & PREV_WHITE))
|
|
#define PREV_TOKEN_TYPE (cur_token[-1].type)
|
|
|
|
#define PUSH_TOKEN(ttype) cur_token++->type = (ttype)
|
|
#define REVISE_TOKEN(ttype) cur_token[-1].type = (ttype)
|
|
#define BACKUP_TOKEN(ttype) (--cur_token)->type = (ttype)
|
|
#define BACKUP_DIGRAPH(ttype) do { \
|
|
BACKUP_TOKEN(ttype); cur_token->flags |= DIGRAPH;} while (0)
|
|
|
|
/* An upper bound on the number of bytes needed to spell a token,
|
|
including preceding whitespace. */
|
|
#define TOKEN_LEN(token) (5 + (TOKEN_SPELL(token) == SPELL_STRING \
|
|
? (token)->val.str.len \
|
|
: (TOKEN_SPELL(token) == SPELL_IDENT \
|
|
? (token)->val.node->length \
|
|
: 0)))
|
|
|
|
#define IS_ARG_CONTEXT(c) ((c)->flags & CONTEXT_ARG)
|
|
#define CURRENT_CONTEXT(pfile) ((pfile)->contexts + (pfile)->cur_context)
|
|
|
|
#define ASSIGN_FLAGS_AND_POS(d, s) \
|
|
do {(d)->flags = (s)->flags & (PREV_WHITE | BOL | PASTE_LEFT); \
|
|
if ((d)->flags & BOL) {(d)->col = (s)->col; (d)->line = (s)->line;} \
|
|
} while (0)
|
|
|
|
/* f is flags, just consisting of PREV_WHITE | BOL. */
|
|
#define MODIFY_FLAGS_AND_POS(d, s, f) \
|
|
do {(d)->flags &= ~(PREV_WHITE | BOL); (d)->flags |= (f); \
|
|
if ((f) & BOL) {(d)->col = (s)->col; (d)->line = (s)->line;} \
|
|
} while (0)
|
|
|
|
#define OP(e, s) { SPELL_OPERATOR, U s },
|
|
#define TK(e, s) { s, U STRINGX (e) },
|
|
|
|
const struct token_spelling
|
|
_cpp_token_spellings [N_TTYPES] = {TTYPE_TABLE };
|
|
|
|
#undef OP
|
|
#undef TK
|
|
|
|
/* The following table is used by trigraph_ok/trigraph_replace. If we
|
|
have designated initializers, it can be constant data; otherwise,
|
|
it is set up at runtime by _cpp_init_input_buffer. */
|
|
|
|
#if (GCC_VERSION >= 2007)
|
|
#define init_trigraph_map() /* nothing */
|
|
#define TRIGRAPH_MAP \
|
|
__extension__ static const U_CHAR trigraph_map[UCHAR_MAX + 1] = {
|
|
#define END };
|
|
#define s(p, v) [p] = v,
|
|
#else
|
|
#define TRIGRAPH_MAP static U_CHAR trigraph_map[UCHAR_MAX + 1] = { 0 }; \
|
|
static void init_trigraph_map PARAMS ((void)) { \
|
|
unsigned char *x = trigraph_map;
|
|
#define END }
|
|
#define s(p, v) x[p] = v;
|
|
#endif
|
|
|
|
TRIGRAPH_MAP
|
|
s('=', '#') s(')', ']') s('!', '|')
|
|
s('(', '[') s('\'', '^') s('>', '}')
|
|
s('/', '\\') s('<', '{') s('-', '~')
|
|
END
|
|
|
|
#undef TRIGRAPH_MAP
|
|
#undef END
|
|
#undef s
|
|
|
|
/* Re-allocates PFILE->token_buffer so it will hold at least N more chars. */
|
|
|
|
void
|
|
_cpp_grow_token_buffer (pfile, n)
|
|
cpp_reader *pfile;
|
|
long n;
|
|
{
|
|
long old_written = CPP_WRITTEN (pfile);
|
|
pfile->token_buffer_size = n + 2 * pfile->token_buffer_size;
|
|
pfile->token_buffer = (U_CHAR *)
|
|
xrealloc(pfile->token_buffer, pfile->token_buffer_size);
|
|
CPP_SET_WRITTEN (pfile, old_written);
|
|
}
|
|
|
|
/* Deal with the annoying semantics of fwrite. */
|
|
static void
|
|
safe_fwrite (pfile, buf, len, fp)
|
|
cpp_reader *pfile;
|
|
const U_CHAR *buf;
|
|
size_t len;
|
|
FILE *fp;
|
|
{
|
|
size_t count;
|
|
|
|
while (len)
|
|
{
|
|
count = fwrite (buf, 1, len, fp);
|
|
if (count == 0)
|
|
goto error;
|
|
len -= count;
|
|
buf += count;
|
|
}
|
|
return;
|
|
|
|
error:
|
|
cpp_notice_from_errno (pfile, CPP_OPTION (pfile, out_fname));
|
|
}
|
|
|
|
/* Notify the compiler proper that the current line number has jumped,
|
|
or the current file name has changed. */
|
|
|
|
static void
|
|
output_line_command (pfile, print, line)
|
|
cpp_reader *pfile;
|
|
cpp_printer *print;
|
|
unsigned int line;
|
|
{
|
|
cpp_buffer *ip = CPP_BUFFER (pfile);
|
|
enum { same = 0, enter, leave, rname } change;
|
|
static const char * const codes[] = { "", " 1", " 2", "" };
|
|
|
|
if (line == 0)
|
|
return;
|
|
|
|
/* End the previous line of text. */
|
|
if (pfile->need_newline)
|
|
putc ('\n', print->outf);
|
|
pfile->need_newline = 0;
|
|
|
|
if (CPP_OPTION (pfile, no_line_commands))
|
|
return;
|
|
|
|
/* If ip is null, we've been called from cpp_finish, and they just
|
|
needed the final flush and trailing newline. */
|
|
if (!ip)
|
|
return;
|
|
|
|
if (pfile->include_depth == print->last_id)
|
|
{
|
|
/* Determine whether the current filename has changed, and if so,
|
|
how. 'nominal_fname' values are unique, so they can be compared
|
|
by comparing pointers. */
|
|
if (ip->nominal_fname == print->last_fname)
|
|
change = same;
|
|
else
|
|
change = rname;
|
|
}
|
|
else
|
|
{
|
|
if (pfile->include_depth > print->last_id)
|
|
change = enter;
|
|
else
|
|
change = leave;
|
|
print->last_id = pfile->include_depth;
|
|
}
|
|
print->last_fname = ip->nominal_fname;
|
|
|
|
/* If the current file has not changed, we can output a few newlines
|
|
instead if we want to increase the line number by a small amount.
|
|
We cannot do this if print->lineno is zero, because that means we
|
|
haven't output any line commands yet. (The very first line
|
|
command output is a `same_file' command.) */
|
|
if (change == same && print->lineno > 0
|
|
&& line >= print->lineno && line < print->lineno + 8)
|
|
{
|
|
while (line > print->lineno)
|
|
{
|
|
putc ('\n', print->outf);
|
|
print->lineno++;
|
|
}
|
|
return;
|
|
}
|
|
|
|
#ifndef NO_IMPLICIT_EXTERN_C
|
|
if (CPP_OPTION (pfile, cplusplus))
|
|
fprintf (print->outf, "# %u \"%s\"%s%s%s\n", line, ip->nominal_fname,
|
|
codes[change],
|
|
ip->inc->sysp ? " 3" : "",
|
|
(ip->inc->sysp == 2) ? " 4" : "");
|
|
else
|
|
#endif
|
|
fprintf (print->outf, "# %u \"%s\"%s%s\n", line, ip->nominal_fname,
|
|
codes[change],
|
|
ip->inc->sysp ? " 3" : "");
|
|
print->lineno = line;
|
|
}
|
|
|
|
/* Write the contents of the token_buffer to the output stream, and
|
|
clear the token_buffer. Also handles generating line commands and
|
|
keeping track of file transitions. */
|
|
|
|
void
|
|
cpp_output_tokens (pfile, print, line)
|
|
cpp_reader *pfile;
|
|
cpp_printer *print;
|
|
unsigned int line;
|
|
{
|
|
if (CPP_WRITTEN (pfile) - print->written)
|
|
{
|
|
safe_fwrite (pfile, pfile->token_buffer,
|
|
CPP_WRITTEN (pfile) - print->written, print->outf);
|
|
pfile->need_newline = 1;
|
|
if (print->lineno)
|
|
print->lineno++;
|
|
|
|
CPP_SET_WRITTEN (pfile, print->written);
|
|
}
|
|
output_line_command (pfile, print, line);
|
|
}
|
|
|
|
/* Scan until CPP_BUFFER (PFILE) is exhausted, discarding output. */
|
|
|
|
void
|
|
cpp_scan_buffer_nooutput (pfile)
|
|
cpp_reader *pfile;
|
|
{
|
|
cpp_buffer *stop = CPP_PREV_BUFFER (CPP_BUFFER (pfile));
|
|
const cpp_token *token;
|
|
|
|
/* In no-output mode, we can ignore everything but directives. */
|
|
for (;;)
|
|
{
|
|
token = _cpp_get_token (pfile);
|
|
|
|
if (token->type == CPP_EOF)
|
|
{
|
|
cpp_pop_buffer (pfile);
|
|
if (CPP_BUFFER (pfile) == stop)
|
|
break;
|
|
}
|
|
|
|
if (token->type == CPP_HASH && token->flags & BOL
|
|
&& pfile->token_list.directive)
|
|
{
|
|
process_directive (pfile, token);
|
|
continue;
|
|
}
|
|
|
|
_cpp_skip_rest_of_line (pfile);
|
|
}
|
|
}
|
|
|
|
/* Scan until CPP_BUFFER (pfile) is exhausted, writing output to PRINT. */
|
|
void
|
|
cpp_scan_buffer (pfile, print)
|
|
cpp_reader *pfile;
|
|
cpp_printer *print;
|
|
{
|
|
cpp_buffer *stop = CPP_PREV_BUFFER (CPP_BUFFER (pfile));
|
|
const cpp_token *token, *prev = 0;
|
|
|
|
for (;;)
|
|
{
|
|
token = _cpp_get_token (pfile);
|
|
if (token->type == CPP_EOF)
|
|
{
|
|
cpp_pop_buffer (pfile);
|
|
|
|
if (CPP_BUFFER (pfile))
|
|
cpp_output_tokens (pfile, print, CPP_BUF_LINE (CPP_BUFFER (pfile)));
|
|
|
|
if (CPP_BUFFER (pfile) == stop)
|
|
return;
|
|
|
|
prev = 0;
|
|
continue;
|
|
}
|
|
|
|
if (token->flags & BOL)
|
|
{
|
|
if (token->type == CPP_HASH && pfile->token_list.directive)
|
|
{
|
|
process_directive (pfile, token);
|
|
continue;
|
|
}
|
|
|
|
cpp_output_tokens (pfile, print, pfile->token_list.line);
|
|
prev = 0;
|
|
}
|
|
|
|
if (token->type != CPP_PLACEMARKER)
|
|
output_token (pfile, token, prev);
|
|
|
|
prev = token;
|
|
}
|
|
}
|
|
|
|
/* Scan a single line of the input into the token_buffer. */
|
|
int
|
|
cpp_scan_line (pfile)
|
|
cpp_reader *pfile;
|
|
{
|
|
const cpp_token *token, *prev = 0;
|
|
|
|
if (pfile->buffer == NULL)
|
|
return 0;
|
|
|
|
do
|
|
{
|
|
token = cpp_get_token (pfile);
|
|
if (token->type == CPP_EOF)
|
|
{
|
|
cpp_pop_buffer (pfile);
|
|
break;
|
|
}
|
|
|
|
/* If the last token on a line results from a macro expansion,
|
|
the check below will fail to stop us from proceeding to the
|
|
next line - so make sure we stick in a newline, at least. */
|
|
if (token->flags & BOL)
|
|
CPP_PUTC (pfile, '\n');
|
|
|
|
output_token (pfile, token, prev);
|
|
prev = token;
|
|
}
|
|
while (pfile->cur_context > 0
|
|
|| pfile->contexts[0].posn < pfile->contexts[0].count);
|
|
return 1;
|
|
}
|
|
|
|
/* Helper routine used by parse_include, which can't see spell_token.
|
|
Reinterpret the current line as an h-char-sequence (< ... >); we are
|
|
looking at the first token after the <. */
|
|
const cpp_token *
|
|
_cpp_glue_header_name (pfile)
|
|
cpp_reader *pfile;
|
|
{
|
|
unsigned int written = CPP_WRITTEN (pfile);
|
|
const cpp_token *t;
|
|
cpp_token *hdr;
|
|
U_CHAR *buf;
|
|
size_t len;
|
|
|
|
for (;;)
|
|
{
|
|
t = _cpp_get_token (pfile);
|
|
if (t->type == CPP_GREATER || t->type == CPP_EOF)
|
|
break;
|
|
|
|
CPP_RESERVE (pfile, TOKEN_LEN (t));
|
|
if (t->flags & PREV_WHITE)
|
|
CPP_PUTC_Q (pfile, ' ');
|
|
pfile->limit = spell_token (pfile, t, pfile->limit);
|
|
}
|
|
|
|
if (t->type == CPP_EOF)
|
|
cpp_error (pfile, "missing terminating > character");
|
|
|
|
len = CPP_WRITTEN (pfile) - written;
|
|
buf = xmalloc (len);
|
|
memcpy (buf, pfile->token_buffer + written, len);
|
|
CPP_SET_WRITTEN (pfile, written);
|
|
|
|
hdr = get_temp_token (pfile);
|
|
hdr->type = CPP_HEADER_NAME;
|
|
hdr->flags = 0;
|
|
hdr->val.str.text = buf;
|
|
hdr->val.str.len = len;
|
|
return hdr;
|
|
}
|
|
|
|
/* Token-buffer helper functions. */
|
|
|
|
/* Expand a token list's string space. It is *vital* that
|
|
list->tokens_used is correct, to get pointer fix-up right. */
|
|
void
|
|
_cpp_expand_name_space (list, len)
|
|
cpp_toklist *list;
|
|
unsigned int len;
|
|
{
|
|
const U_CHAR *old_namebuf;
|
|
|
|
old_namebuf = list->namebuf;
|
|
list->name_cap += len;
|
|
list->namebuf = (unsigned char *) xrealloc (list->namebuf, list->name_cap);
|
|
|
|
/* Fix up token text pointers. */
|
|
if (list->namebuf != old_namebuf)
|
|
{
|
|
unsigned int i;
|
|
|
|
for (i = 0; i < list->tokens_used; i++)
|
|
if (TOKEN_SPELL (&list->tokens[i]) == SPELL_STRING)
|
|
list->tokens[i].val.str.text += (list->namebuf - old_namebuf);
|
|
}
|
|
}
|
|
|
|
/* If there is not enough room for LEN more characters, expand the
|
|
list by just enough to have room for LEN characters. */
|
|
void
|
|
_cpp_reserve_name_space (list, len)
|
|
cpp_toklist *list;
|
|
unsigned int len;
|
|
{
|
|
unsigned int room = list->name_cap - list->name_used;
|
|
|
|
if (room < len)
|
|
_cpp_expand_name_space (list, len - room);
|
|
}
|
|
|
|
/* Expand the number of tokens in a list. */
|
|
void
|
|
_cpp_expand_token_space (list, count)
|
|
cpp_toklist *list;
|
|
unsigned int count;
|
|
{
|
|
unsigned int n;
|
|
|
|
list->tokens_cap += count;
|
|
n = list->tokens_cap;
|
|
if (list->flags & LIST_OFFSET)
|
|
list->tokens--, n++;
|
|
list->tokens = (cpp_token *)
|
|
xrealloc (list->tokens, n * sizeof (cpp_token));
|
|
if (list->flags & LIST_OFFSET)
|
|
list->tokens++; /* Skip the dummy. */
|
|
}
|
|
|
|
/* Initialize a token list. If flags is DUMMY_TOKEN, we allocate
|
|
an extra token in front of the token list, as this allows the lexer
|
|
to always peek at the previous token without worrying about
|
|
underflowing the list, and some initial space. Otherwise, no
|
|
token- or name-space is allocated, and there is no dummy token. */
|
|
void
|
|
_cpp_init_toklist (list, flags)
|
|
cpp_toklist *list;
|
|
int flags;
|
|
{
|
|
if (flags == NO_DUMMY_TOKEN)
|
|
{
|
|
list->tokens_cap = 0;
|
|
list->tokens = 0;
|
|
list->name_cap = 0;
|
|
list->namebuf = 0;
|
|
list->flags = 0;
|
|
}
|
|
else
|
|
{
|
|
/* Initialize token space. Put a dummy token before the start
|
|
that will fail matches. */
|
|
list->tokens_cap = 256; /* 4K's worth. */
|
|
list->tokens = (cpp_token *)
|
|
xmalloc ((list->tokens_cap + 1) * sizeof (cpp_token));
|
|
list->tokens[0].type = CPP_EOF;
|
|
list->tokens++;
|
|
|
|
/* Initialize name space. */
|
|
list->name_cap = 1024;
|
|
list->namebuf = (unsigned char *) xmalloc (list->name_cap);
|
|
list->flags = LIST_OFFSET;
|
|
}
|
|
|
|
_cpp_clear_toklist (list);
|
|
}
|
|
|
|
/* Clear a token list. */
|
|
void
|
|
_cpp_clear_toklist (list)
|
|
cpp_toklist *list;
|
|
{
|
|
list->tokens_used = 0;
|
|
list->name_used = 0;
|
|
list->directive = 0;
|
|
list->paramc = 0;
|
|
list->params_len = 0;
|
|
list->flags &= LIST_OFFSET; /* clear all but that one */
|
|
}
|
|
|
|
/* Free a token list. Does not free the list itself, which may be
|
|
embedded in a larger structure. */
|
|
void
|
|
_cpp_free_toklist (list)
|
|
const cpp_toklist *list;
|
|
{
|
|
if (list->flags & LIST_OFFSET)
|
|
free (list->tokens - 1); /* Backup over dummy token. */
|
|
else
|
|
free (list->tokens);
|
|
free (list->namebuf);
|
|
}
|
|
|
|
/* Compare two tokens. */
|
|
int
|
|
_cpp_equiv_tokens (a, b)
|
|
const cpp_token *a, *b;
|
|
{
|
|
if (a->type == b->type && a->flags == b->flags)
|
|
switch (TOKEN_SPELL (a))
|
|
{
|
|
default: /* Keep compiler happy. */
|
|
case SPELL_OPERATOR:
|
|
return 1;
|
|
case SPELL_CHAR:
|
|
case SPELL_NONE:
|
|
return a->val.aux == b->val.aux; /* arg_no or character. */
|
|
case SPELL_IDENT:
|
|
return a->val.node == b->val.node;
|
|
case SPELL_STRING:
|
|
return (a->val.str.len == b->val.str.len
|
|
&& !memcmp (a->val.str.text, b->val.str.text,
|
|
a->val.str.len));
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
/* Compare two token lists. */
|
|
int
|
|
_cpp_equiv_toklists (a, b)
|
|
const cpp_toklist *a, *b;
|
|
{
|
|
unsigned int i;
|
|
|
|
if (a->tokens_used != b->tokens_used
|
|
|| a->flags != b->flags
|
|
|| a->paramc != b->paramc)
|
|
return 0;
|
|
|
|
for (i = 0; i < a->tokens_used; i++)
|
|
if (! _cpp_equiv_tokens (&a->tokens[i], &b->tokens[i]))
|
|
return 0;
|
|
return 1;
|
|
}
|
|
|
|
/* Utility routine:
|
|
|
|
Compares, the token TOKEN to the NUL-terminated string STRING.
|
|
TOKEN must be a CPP_NAME. Returns 1 for equal, 0 for unequal. */
|
|
|
|
int
|
|
cpp_ideq (token, string)
|
|
const cpp_token *token;
|
|
const char *string;
|
|
{
|
|
if (token->type != CPP_NAME)
|
|
return 0;
|
|
|
|
return !ustrcmp (token->val.node->name, (const U_CHAR *)string);
|
|
}
|
|
|
|
/* Lexing algorithm.
|
|
|
|
The original lexer in cpplib was made up of two passes: a first pass
|
|
that replaced trigraphs and deleted esacped newlines, and a second
|
|
pass that tokenized the result of the first pass. Tokenisation was
|
|
performed by peeking at the next character in the input stream. For
|
|
example, if the input stream contained "!=", the handler for the !
|
|
character would peek at the next character, and if it were a '='
|
|
would skip over it, and return a "!=" token, otherwise it would
|
|
return just the "!" token.
|
|
|
|
To implement a single-pass lexer, this peeking ahead is unworkable.
|
|
An arbitrary number of escaped newlines, and trigraphs (in particular
|
|
??/ which translates to the escape \), could separate the '!' and '='
|
|
in the input stream, yet the next token is still a "!=".
|
|
|
|
Suppose instead that we lex by one logical line at a time, producing
|
|
a token list or stack for each logical line, and when seeing the '!'
|
|
push a CPP_NOT token on the list. Then if the '!' is part of a
|
|
longer token ("!=") we know we must see the remainder of the token by
|
|
the time we reach the end of the logical line. Thus we can have the
|
|
'=' handler look at the previous token (at the end of the list / top
|
|
of the stack) and see if it is a "!" token, and if so, instead of
|
|
pushing a "=" token revise the existing token to be a "!=" token.
|
|
|
|
This works in the presence of escaped newlines, because the '\' would
|
|
have been pushed on the top of the stack as a CPP_BACKSLASH. The
|
|
newline ('\n' or '\r') handler looks at the token at the top of the
|
|
stack to see if it is a CPP_BACKSLASH, and if so discards both.
|
|
Hence the '=' handler would never see any intervening tokens.
|
|
|
|
To make trigraphs work in this context, as in precedence trigraphs
|
|
are highest and converted before anything else, the '?' handler does
|
|
lookahead to see if it is a trigraph, and if so skips the trigraph
|
|
and pushes the token it represents onto the top of the stack. This
|
|
also works in the particular case of a CPP_BACKSLASH trigraph.
|
|
|
|
To the preprocessor, whitespace is only significant to the point of
|
|
knowing whether whitespace precedes a particular token. For example,
|
|
the '=' handler needs to know whether there was whitespace between it
|
|
and a "!" token on the top of the stack, to make the token conversion
|
|
decision correctly. So each token has a PREV_WHITE flag to
|
|
indicate this - the standard permits consecutive whitespace to be
|
|
regarded as a single space. The compiler front ends are not
|
|
interested in whitespace at all; they just require a token stream.
|
|
Another place where whitespace is significant to the preprocessor is
|
|
a #define statment - if there is whitespace between the macro name
|
|
and an initial "(" token the macro is "object-like", otherwise it is
|
|
a function-like macro that takes arguments.
|
|
|
|
However, all is not rosy. Parsing of identifiers, numbers, comments
|
|
and strings becomes trickier because of the possibility of raw
|
|
trigraphs and escaped newlines in the input stream.
|
|
|
|
The trigraphs are three consecutive characters beginning with two
|
|
question marks. A question mark is not valid as part of a number or
|
|
identifier, so parsing of a number or identifier terminates normally
|
|
upon reaching it, returning to the mainloop which handles the
|
|
trigraph just like it would in any other position. Similarly for the
|
|
backslash of a backslash-newline combination. So we just need the
|
|
escaped-newline dropper in the mainloop to check if the token on the
|
|
top of the stack after dropping the escaped newline is a number or
|
|
identifier, and if so to continue the processing it as if nothing had
|
|
happened.
|
|
|
|
For strings, we replace trigraphs whenever we reach a quote or
|
|
newline, because there might be a backslash trigraph escaping them.
|
|
We need to be careful that we start trigraph replacing from where we
|
|
left off previously, because it is possible for a first scan to leave
|
|
"fake" trigraphs that a second scan would pick up as real (e.g. the
|
|
sequence "????/\n=" would find a fake ??= trigraph after removing the
|
|
escaped newline.)
|
|
|
|
For line comments, on reaching a newline we scan the previous
|
|
character(s) to see if it escaped, and continue if it is. Block
|
|
comments ignore everything and just focus on finding the comment
|
|
termination mark. The only difficult thing, and it is surprisingly
|
|
tricky, is checking if an asterisk precedes the final slash since
|
|
they could be separated by escaped newlines. If the preprocessor is
|
|
invoked with the output comments option, we don't bother removing
|
|
escaped newlines and replacing trigraphs for output.
|
|
|
|
Finally, numbers can begin with a period, which is pushed initially
|
|
as a CPP_DOT token in its own right. The digit handler checks if the
|
|
previous token was a CPP_DOT not separated by whitespace, and if so
|
|
pops it off the stack and pushes a period into the number's buffer
|
|
before calling the number parser.
|
|
|
|
*/
|
|
|
|
static const unsigned char *digraph_spellings [] = {U"%:", U"%:%:", U"<:",
|
|
U":>", U"<%", U"%>"};
|
|
|
|
/* Call when a trigraph is encountered. It warns if necessary, and
|
|
returns true if the trigraph should be honoured. END is the third
|
|
character of a trigraph in the input stream. */
|
|
static int
|
|
trigraph_ok (pfile, end)
|
|
cpp_reader *pfile;
|
|
const unsigned char *end;
|
|
{
|
|
int accept = CPP_OPTION (pfile, trigraphs);
|
|
|
|
if (CPP_OPTION (pfile, warn_trigraphs))
|
|
{
|
|
unsigned int col = end - 1 - pfile->buffer->line_base;
|
|
if (accept)
|
|
cpp_warning_with_line (pfile, pfile->buffer->lineno, col,
|
|
"trigraph ??%c converted to %c",
|
|
(int) *end, (int) trigraph_map[*end]);
|
|
else
|
|
cpp_warning_with_line (pfile, pfile->buffer->lineno, col,
|
|
"trigraph ??%c ignored", (int) *end);
|
|
}
|
|
return accept;
|
|
}
|
|
|
|
/* Scan a string for trigraphs, warning or replacing them inline as
|
|
appropriate. When parsing a string, we must call this routine
|
|
before processing a newline character (if trigraphs are enabled),
|
|
since the newline might be escaped by a preceding backslash
|
|
trigraph sequence. Returns a pointer to the end of the name after
|
|
replacement. */
|
|
|
|
static unsigned char *
|
|
trigraph_replace (pfile, src, limit)
|
|
cpp_reader *pfile;
|
|
unsigned char *src;
|
|
unsigned char *limit;
|
|
{
|
|
unsigned char *dest;
|
|
|
|
/* Starting with src[1], find two consecutive '?'. The case of no
|
|
trigraphs is streamlined. */
|
|
|
|
for (src++; src + 1 < limit; src += 2)
|
|
{
|
|
if (src[0] != '?')
|
|
continue;
|
|
|
|
/* Make src point to the 1st (NOT 2nd) of two consecutive '?'s. */
|
|
if (src[-1] == '?')
|
|
src--;
|
|
else if (src + 2 == limit || src[1] != '?')
|
|
continue;
|
|
|
|
/* Check if it really is a trigraph. */
|
|
if (trigraph_map[src[2]] == 0)
|
|
continue;
|
|
|
|
dest = src;
|
|
goto trigraph_found;
|
|
}
|
|
return limit;
|
|
|
|
/* Now we have a trigraph, we need to scan the remaining buffer, and
|
|
copy-shifting its contents left if replacement is enabled. */
|
|
for (; src + 2 < limit; dest++, src++)
|
|
if ((*dest = *src) == '?' && src[1] == '?' && trigraph_map[src[2]])
|
|
{
|
|
trigraph_found:
|
|
src += 2;
|
|
if (trigraph_ok (pfile, pfile->buffer->cur - (limit - src)))
|
|
*dest = trigraph_map[*src];
|
|
}
|
|
|
|
/* Copy remaining (at most 2) characters. */
|
|
while (src < limit)
|
|
*dest++ = *src++;
|
|
return dest;
|
|
}
|
|
|
|
/* If CUR is a backslash or the end of a trigraphed backslash, return
|
|
a pointer to its beginning, otherwise NULL. We don't read beyond
|
|
the buffer start, because there is the start of the comment in the
|
|
buffer. */
|
|
static const unsigned char *
|
|
backslash_start (pfile, cur)
|
|
cpp_reader *pfile;
|
|
const unsigned char *cur;
|
|
{
|
|
if (cur[0] == '\\')
|
|
return cur;
|
|
if (cur[0] == '/' && cur[-1] == '?' && cur[-2] == '?'
|
|
&& trigraph_ok (pfile, cur))
|
|
return cur - 2;
|
|
return 0;
|
|
}
|
|
|
|
/* Skip a C-style block comment. This is probably the trickiest
|
|
handler. We find the end of the comment by seeing if an asterisk
|
|
is before every '/' we encounter. The nasty complication is that a
|
|
previous asterisk may be separated by one or more escaped newlines.
|
|
Returns non-zero if comment terminated by EOF, zero otherwise. */
|
|
static int
|
|
skip_block_comment (pfile)
|
|
cpp_reader *pfile;
|
|
{
|
|
cpp_buffer *buffer = pfile->buffer;
|
|
const unsigned char *char_after_star = 0;
|
|
const unsigned char *cur = buffer->cur;
|
|
|
|
for (; cur < buffer->rlimit; )
|
|
{
|
|
unsigned char c = *cur++;
|
|
|
|
/* People like decorating comments with '*', so check for
|
|
'/' instead for efficiency. */
|
|
if (c == '/')
|
|
{
|
|
/* Don't view / then * then / as finishing the comment. */
|
|
if ((cur[-2] == '*' && cur - 1 > buffer->cur)
|
|
|| cur - 1 == char_after_star)
|
|
{
|
|
buffer->cur = cur;
|
|
return 0;
|
|
}
|
|
|
|
/* Warn about potential nested comments, but not when
|
|
the final character inside the comment is a '/'.
|
|
Don't bother to get it right across escaped newlines. */
|
|
if (CPP_OPTION (pfile, warn_comments) && cur + 1 < buffer->rlimit
|
|
&& cur[0] == '*' && cur[1] != '/')
|
|
{
|
|
buffer->cur = cur;
|
|
cpp_warning (pfile, "'/*' within comment");
|
|
}
|
|
}
|
|
else if (is_vspace (c))
|
|
{
|
|
const unsigned char* bslash = backslash_start (pfile, cur - 2);
|
|
|
|
handle_newline (cur, buffer->rlimit, c);
|
|
/* Work correctly if there is an asterisk before an
|
|
arbirtrarily long sequence of escaped newlines. */
|
|
if (bslash && (bslash[-1] == '*' || bslash == char_after_star))
|
|
char_after_star = cur;
|
|
else
|
|
char_after_star = 0;
|
|
}
|
|
else if (c == '\t')
|
|
adjust_column (pfile, cur - 1);
|
|
}
|
|
|
|
buffer->cur = cur;
|
|
return 1;
|
|
}
|
|
|
|
/* Skip a C++ line comment. Handles escaped newlines. Returns
|
|
non-zero if a multiline comment. */
|
|
static int
|
|
skip_line_comment (pfile)
|
|
cpp_reader *pfile;
|
|
{
|
|
cpp_buffer *buffer = pfile->buffer;
|
|
register const unsigned char *cur = buffer->cur;
|
|
int multiline = 0;
|
|
|
|
for (; cur < buffer->rlimit; )
|
|
{
|
|
unsigned char c = *cur++;
|
|
|
|
if (is_vspace (c))
|
|
{
|
|
/* Check for a (trigaph?) backslash escaping the newline. */
|
|
if (!backslash_start (pfile, cur - 2))
|
|
goto out;
|
|
multiline = 1;
|
|
handle_newline (cur, buffer->rlimit, c);
|
|
}
|
|
}
|
|
cur++;
|
|
|
|
out:
|
|
buffer->cur = cur - 1; /* Leave newline for caller. */
|
|
return multiline;
|
|
}
|
|
|
|
/* TAB points to a \t character. Update col_adjust so we track the
|
|
column correctly. */
|
|
static void
|
|
adjust_column (pfile, tab)
|
|
cpp_reader *pfile;
|
|
const U_CHAR *tab;
|
|
{
|
|
/* Zero-based column. */
|
|
unsigned int col = CPP_BUF_COLUMN (pfile->buffer, tab);
|
|
|
|
/* Round it up to multiple of the tabstop, but subtract 1 since the
|
|
tab itself occupies a character position. */
|
|
pfile->col_adjust += (CPP_OPTION (pfile, tabstop)
|
|
- col % CPP_OPTION (pfile, tabstop)) - 1;
|
|
}
|
|
|
|
/* Skips whitespace, stopping at next non-whitespace character.
|
|
Adjusts pfile->col_adjust to account for tabs. This enables tokens
|
|
to be assigned the correct column. */
|
|
static void
|
|
skip_whitespace (pfile, in_directive)
|
|
cpp_reader *pfile;
|
|
int in_directive;
|
|
{
|
|
cpp_buffer *buffer = pfile->buffer;
|
|
unsigned short warned = 0;
|
|
|
|
/* We only want non-vertical space, i.e. ' ' \t \f \v \0. */
|
|
while (buffer->cur < buffer->rlimit)
|
|
{
|
|
unsigned char c = *buffer->cur;
|
|
|
|
if (!is_nvspace (c))
|
|
break;
|
|
|
|
buffer->cur++;
|
|
/* Horizontal space always OK. */
|
|
if (c == ' ')
|
|
continue;
|
|
else if (c == '\t')
|
|
adjust_column (pfile, buffer->cur - 1);
|
|
/* Must be \f \v or \0. */
|
|
else if (c == '\0')
|
|
{
|
|
if (!warned)
|
|
cpp_warning_with_line (pfile, CPP_BUF_LINE (buffer),
|
|
CPP_BUF_COL (buffer),
|
|
"embedded null character ignored");
|
|
warned = 1;
|
|
}
|
|
else if (in_directive && CPP_PEDANTIC (pfile))
|
|
cpp_pedwarn_with_line (pfile, CPP_BUF_LINE (buffer),
|
|
CPP_BUF_COL (buffer),
|
|
"%s in preprocessing directive",
|
|
c == '\f' ? "form feed" : "vertical tab");
|
|
}
|
|
}
|
|
|
|
/* Parse (append) an identifier. Calculates the hash value of the
|
|
token while parsing, for performance. The algorithm *must* match
|
|
cpp_lookup(). */
|
|
static const U_CHAR *
|
|
parse_name (pfile, tok, cur, rlimit)
|
|
cpp_reader *pfile;
|
|
cpp_token *tok;
|
|
const U_CHAR *cur, *rlimit;
|
|
{
|
|
const U_CHAR *name;
|
|
unsigned int len;
|
|
unsigned int r;
|
|
|
|
name = cur;
|
|
r = 0;
|
|
while (cur < rlimit)
|
|
{
|
|
if (! is_idchar (*cur))
|
|
break;
|
|
/* $ is not a legal identifier character in the standard, but is
|
|
commonly accepted as an extension. Don't warn about it in
|
|
skipped conditional blocks. */
|
|
if (*cur == '$' && CPP_PEDANTIC (pfile) && ! pfile->skipping)
|
|
{
|
|
CPP_BUFFER (pfile)->cur = cur;
|
|
cpp_pedwarn (pfile, "'$' character in identifier");
|
|
}
|
|
|
|
r = HASHSTEP (r, cur);
|
|
cur++;
|
|
}
|
|
len = cur - name;
|
|
|
|
if (tok->val.node == 0)
|
|
tok->val.node = _cpp_lookup_with_hash (pfile, name, len, r);
|
|
else
|
|
{
|
|
unsigned int oldlen = tok->val.node->length;
|
|
U_CHAR *newname = alloca (oldlen + len);
|
|
memcpy (newname, tok->val.node->name, oldlen);
|
|
memcpy (newname + oldlen, name, len);
|
|
tok->val.node = cpp_lookup (pfile, newname, len + oldlen);
|
|
}
|
|
|
|
return cur;
|
|
}
|
|
|
|
/* Parse (append) a number. */
|
|
static void
|
|
parse_number (pfile, list, name)
|
|
cpp_reader *pfile;
|
|
cpp_toklist *list;
|
|
cpp_string *name;
|
|
{
|
|
const unsigned char *name_limit;
|
|
unsigned char *namebuf;
|
|
cpp_buffer *buffer = pfile->buffer;
|
|
register const unsigned char *cur = buffer->cur;
|
|
|
|
expanded:
|
|
name_limit = list->namebuf + list->name_cap;
|
|
namebuf = list->namebuf + list->name_used;
|
|
|
|
for (; cur < buffer->rlimit && namebuf < name_limit; )
|
|
{
|
|
unsigned char c = *namebuf = *cur; /* Copy a single char. */
|
|
|
|
/* Perhaps we should accept '$' here if we accept it for
|
|
identifiers. We know namebuf[-1] is safe, because for c to
|
|
be a sign we must have pushed at least one character. */
|
|
if (!is_numchar (c) && c != '.' && ! VALID_SIGN (c, namebuf[-1]))
|
|
goto out;
|
|
|
|
namebuf++;
|
|
cur++;
|
|
}
|
|
|
|
/* Run out of name space? */
|
|
if (cur < buffer->rlimit)
|
|
{
|
|
list->name_used = namebuf - list->namebuf;
|
|
auto_expand_name_space (list);
|
|
goto expanded;
|
|
}
|
|
|
|
out:
|
|
buffer->cur = cur;
|
|
name->len = namebuf - name->text;
|
|
list->name_used = namebuf - list->namebuf;
|
|
}
|
|
|
|
/* Places a string terminated by an unescaped TERMINATOR into a
|
|
cpp_string, which should be expandable and thus at the top of the
|
|
list's stack. Handles embedded trigraphs, if necessary, and
|
|
escaped newlines.
|
|
|
|
Can be used for character constants (terminator = '\''), string
|
|
constants ('"') and angled headers ('>'). Multi-line strings are
|
|
allowed, except for within directives. */
|
|
|
|
static void
|
|
parse_string (pfile, list, token, terminator)
|
|
cpp_reader *pfile;
|
|
cpp_toklist *list;
|
|
cpp_token *token;
|
|
unsigned int terminator;
|
|
{
|
|
cpp_buffer *buffer = pfile->buffer;
|
|
cpp_string *name = &token->val.str;
|
|
register const unsigned char *cur = buffer->cur;
|
|
const unsigned char *name_limit;
|
|
unsigned char *namebuf;
|
|
unsigned int null_count = 0;
|
|
unsigned int trigraphed = list->name_used;
|
|
|
|
expanded:
|
|
name_limit = list->namebuf + list->name_cap;
|
|
namebuf = list->namebuf + list->name_used;
|
|
|
|
for (; cur < buffer->rlimit && namebuf < name_limit; )
|
|
{
|
|
unsigned int c = *namebuf++ = *cur++; /* Copy a single char. */
|
|
|
|
if (c == '\0')
|
|
null_count++;
|
|
else if (c == terminator || is_vspace (c))
|
|
{
|
|
/* Needed for trigraph_replace and multiline string warning. */
|
|
buffer->cur = cur;
|
|
|
|
/* Scan for trigraphs before checking if backslash-escaped. */
|
|
if ((CPP_OPTION (pfile, trigraphs)
|
|
|| CPP_OPTION (pfile, warn_trigraphs))
|
|
&& namebuf - (list->namebuf + trigraphed) >= 3)
|
|
{
|
|
namebuf = trigraph_replace (pfile, list->namebuf + trigraphed,
|
|
namebuf);
|
|
/* The test above guarantees trigraphed will be positive. */
|
|
trigraphed = namebuf - list->namebuf - 2;
|
|
}
|
|
|
|
namebuf--; /* Drop the newline / terminator from the name. */
|
|
if (is_vspace (c))
|
|
{
|
|
/* Drop a backslash newline, and continue. */
|
|
if (namebuf[-1] == '\\')
|
|
{
|
|
handle_newline (cur, buffer->rlimit, c);
|
|
namebuf--;
|
|
continue;
|
|
}
|
|
|
|
cur--;
|
|
|
|
/* In assembly language, silently terminate strings of
|
|
either variety at end of line. This is a kludge
|
|
around not knowing where comments are. */
|
|
if (CPP_OPTION (pfile, lang_asm))
|
|
goto out;
|
|
|
|
/* Character constants and header names may not extend
|
|
over multiple lines. In Standard C, neither may
|
|
strings. We accept multiline strings as an
|
|
extension. (Even in directives - otherwise, glibc's
|
|
longlong.h breaks.) */
|
|
if (terminator != '"')
|
|
goto unterminated;
|
|
|
|
cur++; /* Move forwards again. */
|
|
|
|
if (pfile->multiline_string_line == 0)
|
|
{
|
|
pfile->multiline_string_line = token->line;
|
|
pfile->multiline_string_column = token->col;
|
|
if (CPP_PEDANTIC (pfile))
|
|
cpp_pedwarn (pfile, "multi-line string constant");
|
|
}
|
|
|
|
*namebuf++ = '\n';
|
|
handle_newline (cur, buffer->rlimit, c);
|
|
}
|
|
else
|
|
{
|
|
unsigned char *temp;
|
|
|
|
/* An odd number of consecutive backslashes represents
|
|
an escaped terminator. */
|
|
temp = namebuf - 1;
|
|
while (temp >= name->text && *temp == '\\')
|
|
temp--;
|
|
|
|
if ((namebuf - temp) & 1)
|
|
goto out;
|
|
namebuf++;
|
|
}
|
|
}
|
|
}
|
|
|
|
/* Run out of name space? */
|
|
if (cur < buffer->rlimit)
|
|
{
|
|
list->name_used = namebuf - list->namebuf;
|
|
auto_expand_name_space (list);
|
|
goto expanded;
|
|
}
|
|
|
|
/* We may not have trigraph-replaced the input for this code path,
|
|
but as the input is in error by being unterminated we don't
|
|
bother. Prevent warnings about no newlines at EOF. */
|
|
if (is_vspace (cur[-1]))
|
|
cur--;
|
|
|
|
unterminated:
|
|
cpp_error (pfile, "missing terminating %c character", (int) terminator);
|
|
|
|
if (terminator == '\"' && pfile->multiline_string_line != list->line
|
|
&& pfile->multiline_string_line != 0)
|
|
{
|
|
cpp_error_with_line (pfile, pfile->multiline_string_line,
|
|
pfile->multiline_string_column,
|
|
"possible start of unterminated string literal");
|
|
pfile->multiline_string_line = 0;
|
|
}
|
|
|
|
out:
|
|
buffer->cur = cur;
|
|
name->len = namebuf - name->text;
|
|
list->name_used = namebuf - list->namebuf;
|
|
|
|
if (null_count > 0)
|
|
cpp_warning (pfile, (null_count > 1 ? "null characters preserved"
|
|
: "null character preserved"));
|
|
}
|
|
|
|
/* The character TYPE helps us distinguish comment types: '*' = C
|
|
style, '/' = C++ style. For code simplicity, the stored comment
|
|
includes the comment start and any terminator. */
|
|
|
|
#define COMMENT_START_LEN 2
|
|
static void
|
|
save_comment (list, token, from, len, type)
|
|
cpp_toklist *list;
|
|
cpp_token *token;
|
|
const unsigned char *from;
|
|
unsigned int len;
|
|
unsigned int type;
|
|
{
|
|
unsigned char *buffer;
|
|
|
|
len += COMMENT_START_LEN;
|
|
|
|
if (list->name_used + len > list->name_cap)
|
|
_cpp_expand_name_space (list, len);
|
|
|
|
INIT_TOKEN_STR (list, token);
|
|
token->type = CPP_COMMENT;
|
|
token->val.str.len = len;
|
|
|
|
buffer = list->namebuf + list->name_used;
|
|
list->name_used += len;
|
|
|
|
/* Copy the comment. */
|
|
if (type == '*')
|
|
{
|
|
*buffer++ = '/';
|
|
*buffer++ = '*';
|
|
}
|
|
else
|
|
{
|
|
*buffer++ = type;
|
|
*buffer++ = type;
|
|
}
|
|
memcpy (buffer, from, len - COMMENT_START_LEN);
|
|
}
|
|
|
|
/*
|
|
* The tokenizer's main loop. Returns a token list, representing a
|
|
* logical line in the input file. On EOF after some tokens have
|
|
* been processed, we return immediately. Then in next call, or if
|
|
* EOF occurred at the beginning of a logical line, a single CPP_EOF
|
|
* token is placed in the list.
|
|
*
|
|
* Implementation relies almost entirely on lookback, rather than
|
|
* looking forwards. This means that tokenization requires just
|
|
* a single pass of the file, even in the presence of trigraphs and
|
|
* escaped newlines, providing significant performance benefits.
|
|
* Trigraph overhead is negligible if they are disabled, and low
|
|
* even when enabled.
|
|
*/
|
|
|
|
#define KNOWN_DIRECTIVE() (list->directive != 0)
|
|
#define MIGHT_BE_DIRECTIVE() \
|
|
(cur_token == &list->tokens[first_token + 1] && cur_token[-1].type == CPP_HASH)
|
|
|
|
static void
|
|
lex_line (pfile, list)
|
|
cpp_reader *pfile;
|
|
cpp_toklist *list;
|
|
{
|
|
cpp_token *cur_token, *token_limit, *first;
|
|
cpp_buffer *buffer = pfile->buffer;
|
|
const unsigned char *cur = buffer->cur;
|
|
unsigned char flags = 0;
|
|
unsigned int first_token = list->tokens_used;
|
|
|
|
if (!(list->flags & LIST_OFFSET))
|
|
(abort) ();
|
|
|
|
list->file = buffer->nominal_fname;
|
|
list->line = CPP_BUF_LINE (buffer);
|
|
pfile->col_adjust = 0;
|
|
pfile->in_lex_line = 1;
|
|
if (cur == buffer->buf)
|
|
list->flags |= BEG_OF_FILE;
|
|
|
|
expanded:
|
|
token_limit = list->tokens + list->tokens_cap;
|
|
cur_token = list->tokens + list->tokens_used;
|
|
|
|
for (; cur < buffer->rlimit && cur_token < token_limit;)
|
|
{
|
|
unsigned char c;
|
|
|
|
/* Optimize non-vertical whitespace skipping; most tokens are
|
|
probably separated by whitespace. (' ' '\t' '\v' '\f' '\0'). */
|
|
c = *cur;
|
|
if (is_nvspace (c))
|
|
{
|
|
buffer->cur = cur;
|
|
skip_whitespace (pfile, (list->tokens[first_token].type == CPP_HASH
|
|
&& cur_token > &list->tokens[first_token]));
|
|
cur = buffer->cur;
|
|
|
|
flags = PREV_WHITE;
|
|
if (cur == buffer->rlimit)
|
|
break;
|
|
c = *cur;
|
|
}
|
|
cur++;
|
|
|
|
/* Initialize current token. CPP_EOF will not be fixed up by
|
|
expand_name_space. */
|
|
list->tokens_used = cur_token - list->tokens + 1;
|
|
cur_token->type = CPP_EOF;
|
|
cur_token->col = CPP_BUF_COLUMN (buffer, cur);
|
|
cur_token->line = CPP_BUF_LINE (buffer);
|
|
cur_token->flags = flags;
|
|
flags = 0;
|
|
|
|
switch (c)
|
|
{
|
|
case '0': case '1': case '2': case '3': case '4':
|
|
case '5': case '6': case '7': case '8': case '9':
|
|
{
|
|
int prev_dot;
|
|
|
|
cur--; /* Backup character. */
|
|
prev_dot = PREV_TOKEN_TYPE == CPP_DOT && IMMED_TOKEN ();
|
|
if (prev_dot)
|
|
cur_token--;
|
|
INIT_TOKEN_STR (list, cur_token);
|
|
/* Prepend an immediately previous CPP_DOT token. */
|
|
if (prev_dot)
|
|
{
|
|
if (list->name_cap == list->name_used)
|
|
auto_expand_name_space (list);
|
|
|
|
cur_token->val.str.len = 1;
|
|
list->namebuf[list->name_used++] = '.';
|
|
}
|
|
|
|
continue_number:
|
|
cur_token->type = CPP_NUMBER; /* Before parse_number. */
|
|
buffer->cur = cur;
|
|
parse_number (pfile, list, &cur_token->val.str);
|
|
cur = buffer->cur;
|
|
}
|
|
/* Check for # 123 form of #line. */
|
|
if (MIGHT_BE_DIRECTIVE ())
|
|
list->directive = _cpp_check_linemarker (pfile, cur_token,
|
|
!(cur_token[-1].flags
|
|
& PREV_WHITE));
|
|
cur_token++;
|
|
break;
|
|
|
|
letter:
|
|
case '_':
|
|
case 'a': case 'b': case 'c': case 'd': case 'e': case 'f':
|
|
case 'g': case 'h': case 'i': case 'j': case 'k': case 'l':
|
|
case 'm': case 'n': case 'o': case 'p': case 'q': case 'r':
|
|
case 's': case 't': case 'u': case 'v': case 'w': case 'x':
|
|
case 'y': case 'z':
|
|
case 'A': case 'B': case 'C': case 'D': case 'E': case 'F':
|
|
case 'G': case 'H': case 'I': case 'J': case 'K': case 'L':
|
|
case 'M': case 'N': case 'O': case 'P': case 'Q': case 'R':
|
|
case 'S': case 'T': case 'U': case 'V': case 'W': case 'X':
|
|
case 'Y': case 'Z':
|
|
cur--; /* Backup character. */
|
|
cur_token->val.node = 0;
|
|
cur_token->type = CPP_NAME; /* Identifier, macro etc. */
|
|
|
|
continue_name:
|
|
cur = parse_name (pfile, cur_token, cur, buffer->rlimit);
|
|
|
|
if (MIGHT_BE_DIRECTIVE ())
|
|
list->directive = _cpp_check_directive (pfile, cur_token,
|
|
!(list->tokens[0].flags
|
|
& PREV_WHITE));
|
|
/* Convert named operators to their proper types. */
|
|
if (cur_token->val.node->type == T_OPERATOR)
|
|
{
|
|
cur_token->flags |= NAMED_OP;
|
|
cur_token->type = cur_token->val.node->value.code;
|
|
}
|
|
|
|
cur_token++;
|
|
break;
|
|
|
|
case '\'':
|
|
case '\"':
|
|
cur_token->type = c == '\'' ? CPP_CHAR : CPP_STRING;
|
|
/* Do we have a wide string? */
|
|
if (cur_token[-1].type == CPP_NAME && IMMED_TOKEN ()
|
|
&& cur_token[-1].val.node == pfile->spec_nodes->n_L)
|
|
BACKUP_TOKEN (c == '\'' ? CPP_WCHAR : CPP_WSTRING);
|
|
|
|
do_parse_string:
|
|
/* Here c is one of ' " or >. */
|
|
INIT_TOKEN_STR (list, cur_token);
|
|
buffer->cur = cur;
|
|
parse_string (pfile, list, cur_token, c);
|
|
cur = buffer->cur;
|
|
cur_token++;
|
|
break;
|
|
|
|
case '/':
|
|
cur_token->type = CPP_DIV;
|
|
if (IMMED_TOKEN ())
|
|
{
|
|
if (PREV_TOKEN_TYPE == CPP_DIV)
|
|
{
|
|
/* We silently allow C++ comments in system headers,
|
|
irrespective of conformance mode, because lots of
|
|
broken systems do that and trying to clean it up
|
|
in fixincludes is a nightmare. */
|
|
if (CPP_IN_SYSTEM_HEADER (pfile))
|
|
goto do_line_comment;
|
|
else if (CPP_OPTION (pfile, cplusplus_comments))
|
|
{
|
|
if (CPP_OPTION (pfile, c89) && CPP_PEDANTIC (pfile)
|
|
&& ! buffer->warned_cplusplus_comments)
|
|
{
|
|
buffer->cur = cur;
|
|
cpp_pedwarn (pfile,
|
|
"C++ style comments are not allowed in ISO C89");
|
|
cpp_pedwarn (pfile,
|
|
"(this will be reported only once per input file)");
|
|
buffer->warned_cplusplus_comments = 1;
|
|
}
|
|
do_line_comment:
|
|
buffer->cur = cur;
|
|
#if 0 /* Leave until new lexer in place. */
|
|
if (cur[-2] != c)
|
|
cpp_warning (pfile,
|
|
"comment start split across lines");
|
|
#endif
|
|
if (skip_line_comment (pfile))
|
|
cpp_warning (pfile, "multi-line comment");
|
|
|
|
/* Back-up to first '-' or '/'. */
|
|
cur_token--;
|
|
if (!CPP_OPTION (pfile, discard_comments)
|
|
&& (!KNOWN_DIRECTIVE()
|
|
|| (list->directive->flags & COMMENTS)))
|
|
save_comment (list, cur_token++, cur,
|
|
buffer->cur - cur, c);
|
|
else
|
|
flags = PREV_WHITE;
|
|
|
|
cur = buffer->cur;
|
|
break;
|
|
}
|
|
}
|
|
}
|
|
cur_token++;
|
|
break;
|
|
|
|
case '*':
|
|
cur_token->type = CPP_MULT;
|
|
if (IMMED_TOKEN ())
|
|
{
|
|
if (PREV_TOKEN_TYPE == CPP_DIV)
|
|
{
|
|
buffer->cur = cur;
|
|
#if 0 /* Leave until new lexer in place. */
|
|
if (cur[-2] != '/')
|
|
cpp_warning (pfile,
|
|
"comment start '/*' split across lines");
|
|
#endif
|
|
if (skip_block_comment (pfile))
|
|
cpp_error_with_line (pfile, list->line, cur_token[-1].col,
|
|
"unterminated comment");
|
|
#if 0 /* Leave until new lexer in place. */
|
|
else if (buffer->cur[-2] != '*')
|
|
cpp_warning (pfile,
|
|
"comment end '*/' split across lines");
|
|
#endif
|
|
/* Back up to opening '/'. */
|
|
cur_token--;
|
|
if (!CPP_OPTION (pfile, discard_comments)
|
|
&& (!KNOWN_DIRECTIVE()
|
|
|| (list->directive->flags & COMMENTS)))
|
|
save_comment (list, cur_token++, cur,
|
|
buffer->cur - cur, c);
|
|
else
|
|
flags = PREV_WHITE;
|
|
|
|
cur = buffer->cur;
|
|
break;
|
|
}
|
|
else if (CPP_OPTION (pfile, cplusplus))
|
|
{
|
|
/* In C++, there are .* and ->* operators. */
|
|
if (PREV_TOKEN_TYPE == CPP_DEREF)
|
|
BACKUP_TOKEN (CPP_DEREF_STAR);
|
|
else if (PREV_TOKEN_TYPE == CPP_DOT)
|
|
BACKUP_TOKEN (CPP_DOT_STAR);
|
|
}
|
|
}
|
|
cur_token++;
|
|
break;
|
|
|
|
case '\n':
|
|
case '\r':
|
|
handle_newline (cur, buffer->rlimit, c);
|
|
if (PREV_TOKEN_TYPE == CPP_BACKSLASH)
|
|
{
|
|
if (IMMED_TOKEN ())
|
|
{
|
|
/* Remove the escaped newline. Then continue to process
|
|
any interrupted name or number. */
|
|
cur_token--;
|
|
/* Backslash-newline may not be immediately followed by
|
|
EOF (C99 5.1.1.2). */
|
|
if (cur >= buffer->rlimit)
|
|
{
|
|
cpp_pedwarn (pfile, "backslash-newline at end of file");
|
|
break;
|
|
}
|
|
if (IMMED_TOKEN ())
|
|
{
|
|
cur_token--;
|
|
if (cur_token->type == CPP_NAME)
|
|
goto continue_name;
|
|
else if (cur_token->type == CPP_NUMBER)
|
|
goto continue_number;
|
|
cur_token++;
|
|
}
|
|
/* Remember whitespace setting. */
|
|
flags = cur_token->flags;
|
|
break;
|
|
}
|
|
else
|
|
{
|
|
buffer->cur = cur;
|
|
cpp_warning (pfile,
|
|
"backslash and newline separated by space");
|
|
}
|
|
}
|
|
else if (MIGHT_BE_DIRECTIVE ())
|
|
{
|
|
/* "Null directive." C99 6.10.7: A preprocessing
|
|
directive of the form # <new-line> has no effect.
|
|
|
|
But it is still a directive, and therefore disappears
|
|
from the output. */
|
|
cur_token--;
|
|
if (cur_token->flags & PREV_WHITE
|
|
&& CPP_WTRADITIONAL (pfile))
|
|
cpp_warning (pfile, "K+R C ignores #\\n with the # indented");
|
|
}
|
|
|
|
/* Skip vertical space until we have at least one token to
|
|
return. */
|
|
if (cur_token != &list->tokens[first_token])
|
|
goto out;
|
|
list->line = CPP_BUF_LINE (buffer);
|
|
break;
|
|
|
|
case '-':
|
|
if (IMMED_TOKEN () && PREV_TOKEN_TYPE == CPP_MINUS)
|
|
REVISE_TOKEN (CPP_MINUS_MINUS);
|
|
else
|
|
PUSH_TOKEN (CPP_MINUS);
|
|
break;
|
|
|
|
make_hash:
|
|
case '#':
|
|
/* The digraph flag checking ensures that ## and %:%:
|
|
are interpreted as CPP_PASTE, but #%: and %:# are not. */
|
|
if (PREV_TOKEN_TYPE == CPP_HASH && IMMED_TOKEN ()
|
|
&& ((cur_token->flags ^ cur_token[-1].flags) & DIGRAPH) == 0)
|
|
REVISE_TOKEN (CPP_PASTE);
|
|
else
|
|
PUSH_TOKEN (CPP_HASH);
|
|
break;
|
|
|
|
case ':':
|
|
cur_token->type = CPP_COLON;
|
|
if (IMMED_TOKEN ())
|
|
{
|
|
if (PREV_TOKEN_TYPE == CPP_COLON
|
|
&& CPP_OPTION (pfile, cplusplus))
|
|
BACKUP_TOKEN (CPP_SCOPE);
|
|
else if (CPP_OPTION (pfile, digraphs))
|
|
{
|
|
/* Digraph: "<:" is a '[' */
|
|
if (PREV_TOKEN_TYPE == CPP_LESS)
|
|
BACKUP_DIGRAPH (CPP_OPEN_SQUARE);
|
|
/* Digraph: "%:" is a '#' */
|
|
else if (PREV_TOKEN_TYPE == CPP_MOD)
|
|
{
|
|
(--cur_token)->flags |= DIGRAPH;
|
|
goto make_hash;
|
|
}
|
|
}
|
|
}
|
|
cur_token++;
|
|
break;
|
|
|
|
case '&':
|
|
if (IMMED_TOKEN () && PREV_TOKEN_TYPE == CPP_AND)
|
|
REVISE_TOKEN (CPP_AND_AND);
|
|
else
|
|
PUSH_TOKEN (CPP_AND);
|
|
break;
|
|
|
|
make_or:
|
|
case '|':
|
|
if (IMMED_TOKEN () && PREV_TOKEN_TYPE == CPP_OR)
|
|
REVISE_TOKEN (CPP_OR_OR);
|
|
else
|
|
PUSH_TOKEN (CPP_OR);
|
|
break;
|
|
|
|
case '+':
|
|
if (IMMED_TOKEN () && PREV_TOKEN_TYPE == CPP_PLUS)
|
|
REVISE_TOKEN (CPP_PLUS_PLUS);
|
|
else
|
|
PUSH_TOKEN (CPP_PLUS);
|
|
break;
|
|
|
|
case '=':
|
|
/* This relies on equidistance of "?=" and "?" tokens. */
|
|
if (IMMED_TOKEN () && PREV_TOKEN_TYPE <= CPP_LAST_EQ)
|
|
REVISE_TOKEN (PREV_TOKEN_TYPE + (CPP_EQ_EQ - CPP_EQ));
|
|
else
|
|
PUSH_TOKEN (CPP_EQ);
|
|
break;
|
|
|
|
case '>':
|
|
cur_token->type = CPP_GREATER;
|
|
if (IMMED_TOKEN ())
|
|
{
|
|
if (PREV_TOKEN_TYPE == CPP_GREATER)
|
|
BACKUP_TOKEN (CPP_RSHIFT);
|
|
else if (PREV_TOKEN_TYPE == CPP_MINUS)
|
|
BACKUP_TOKEN (CPP_DEREF);
|
|
else if (CPP_OPTION (pfile, digraphs))
|
|
{
|
|
/* Digraph: ":>" is a ']' */
|
|
if (PREV_TOKEN_TYPE == CPP_COLON)
|
|
BACKUP_DIGRAPH (CPP_CLOSE_SQUARE);
|
|
/* Digraph: "%>" is a '}' */
|
|
else if (PREV_TOKEN_TYPE == CPP_MOD)
|
|
BACKUP_DIGRAPH (CPP_CLOSE_BRACE);
|
|
}
|
|
}
|
|
cur_token++;
|
|
break;
|
|
|
|
case '<':
|
|
if (IMMED_TOKEN () && PREV_TOKEN_TYPE == CPP_LESS)
|
|
{
|
|
REVISE_TOKEN (CPP_LSHIFT);
|
|
break;
|
|
}
|
|
/* Is this the beginning of a header name? */
|
|
if (KNOWN_DIRECTIVE () && (list->directive->flags & INCL))
|
|
{
|
|
c = '>'; /* Terminator. */
|
|
cur_token->type = CPP_HEADER_NAME;
|
|
goto do_parse_string;
|
|
}
|
|
PUSH_TOKEN (CPP_LESS);
|
|
break;
|
|
|
|
case '%':
|
|
/* Digraph: "<%" is a '{' */
|
|
cur_token->type = CPP_MOD;
|
|
if (IMMED_TOKEN () && PREV_TOKEN_TYPE == CPP_LESS
|
|
&& CPP_OPTION (pfile, digraphs))
|
|
BACKUP_DIGRAPH (CPP_OPEN_BRACE);
|
|
cur_token++;
|
|
break;
|
|
|
|
case '?':
|
|
if (cur + 1 < buffer->rlimit && *cur == '?'
|
|
&& trigraph_map[cur[1]] && trigraph_ok (pfile, cur + 1))
|
|
{
|
|
/* Handle trigraph. */
|
|
cur++;
|
|
switch (*cur++)
|
|
{
|
|
case '(': goto make_open_square;
|
|
case ')': goto make_close_square;
|
|
case '<': goto make_open_brace;
|
|
case '>': goto make_close_brace;
|
|
case '=': goto make_hash;
|
|
case '!': goto make_or;
|
|
case '-': goto make_complement;
|
|
case '/': goto make_backslash;
|
|
case '\'': goto make_xor;
|
|
}
|
|
}
|
|
if (IMMED_TOKEN () && CPP_OPTION (pfile, cplusplus))
|
|
{
|
|
/* GNU C++ defines <? and >? operators. */
|
|
if (PREV_TOKEN_TYPE == CPP_LESS)
|
|
{
|
|
REVISE_TOKEN (CPP_MIN);
|
|
break;
|
|
}
|
|
else if (PREV_TOKEN_TYPE == CPP_GREATER)
|
|
{
|
|
REVISE_TOKEN (CPP_MAX);
|
|
break;
|
|
}
|
|
}
|
|
PUSH_TOKEN (CPP_QUERY);
|
|
break;
|
|
|
|
case '.':
|
|
if (PREV_TOKEN_TYPE == CPP_DOT && cur_token[-2].type == CPP_DOT
|
|
&& IMMED_TOKEN ()
|
|
&& !(cur_token[-1].flags & PREV_WHITE))
|
|
{
|
|
cur_token -= 2;
|
|
PUSH_TOKEN (CPP_ELLIPSIS);
|
|
}
|
|
else
|
|
PUSH_TOKEN (CPP_DOT);
|
|
break;
|
|
|
|
make_complement:
|
|
case '~': PUSH_TOKEN (CPP_COMPL); break;
|
|
make_xor:
|
|
case '^': PUSH_TOKEN (CPP_XOR); break;
|
|
make_open_brace:
|
|
case '{': PUSH_TOKEN (CPP_OPEN_BRACE); break;
|
|
make_close_brace:
|
|
case '}': PUSH_TOKEN (CPP_CLOSE_BRACE); break;
|
|
make_open_square:
|
|
case '[': PUSH_TOKEN (CPP_OPEN_SQUARE); break;
|
|
make_close_square:
|
|
case ']': PUSH_TOKEN (CPP_CLOSE_SQUARE); break;
|
|
make_backslash:
|
|
case '\\': PUSH_TOKEN (CPP_BACKSLASH); break;
|
|
case '!': PUSH_TOKEN (CPP_NOT); break;
|
|
case ',': PUSH_TOKEN (CPP_COMMA); break;
|
|
case ';': PUSH_TOKEN (CPP_SEMICOLON); break;
|
|
case '(': PUSH_TOKEN (CPP_OPEN_PAREN); break;
|
|
case ')': PUSH_TOKEN (CPP_CLOSE_PAREN); break;
|
|
|
|
case '$':
|
|
if (CPP_OPTION (pfile, dollars_in_ident))
|
|
goto letter;
|
|
/* Fall through */
|
|
default:
|
|
cur_token->val.aux = c;
|
|
PUSH_TOKEN (CPP_OTHER);
|
|
break;
|
|
}
|
|
}
|
|
|
|
/* Run out of token space? */
|
|
if (cur_token == token_limit)
|
|
{
|
|
list->tokens_used = cur_token - list->tokens;
|
|
_cpp_expand_token_space (list, 256);
|
|
goto expanded;
|
|
}
|
|
|
|
cur_token->flags = flags;
|
|
if (cur_token == &list->tokens[first_token] && pfile->done_initializing)
|
|
{
|
|
if (cur > buffer->buf && !is_vspace (cur[-1]))
|
|
cpp_pedwarn_with_line (pfile, CPP_BUF_LINE (buffer),
|
|
CPP_BUF_COLUMN (buffer, cur),
|
|
"no newline at end of file");
|
|
cur_token++->type = CPP_EOF;
|
|
}
|
|
|
|
out:
|
|
/* All tokens are allocated, so the memory location is fixed. */
|
|
first = &list->tokens[first_token];
|
|
|
|
/* Don't complain about the null directive, nor directives in
|
|
assembly source: we don't know where the comments are, and # may
|
|
introduce assembler pseudo-ops. Don't complain about invalid
|
|
directives in skipped conditional groups (6.10 p4). */
|
|
if (first->type == CPP_HASH && list->directive == 0 && !pfile->skipping
|
|
&& cur_token > first + 1 && !CPP_OPTION (pfile, lang_asm))
|
|
{
|
|
if (first[1].type == CPP_NAME)
|
|
cpp_error (pfile, "invalid preprocessing directive #%s",
|
|
first[1].val.node->name);
|
|
else
|
|
cpp_error (pfile, "invalid preprocessing directive");
|
|
}
|
|
|
|
/* Put EOF at end of known directives. This covers "directives do
|
|
not extend beyond the end of the line (description 6.10 part 2)". */
|
|
if (KNOWN_DIRECTIVE () || !pfile->done_initializing)
|
|
{
|
|
pfile->first_directive_token = first;
|
|
cur_token++->type = CPP_EOF;
|
|
}
|
|
|
|
first->flags |= BOL;
|
|
if (first_token != 0)
|
|
/* 6.10.3.10: Within the sequence of preprocessing tokens making
|
|
up the invocation of a function-like macro, new line is
|
|
considered a normal white-space character. */
|
|
first->flags |= PREV_WHITE;
|
|
|
|
buffer->cur = cur;
|
|
list->tokens_used = cur_token - list->tokens;
|
|
pfile->in_lex_line = 0;
|
|
}
|
|
|
|
/* Write the spelling of a token TOKEN, with any appropriate
|
|
whitespace before it, to the token_buffer. PREV is the previous
|
|
token, which is used to determine if we need to shove in an extra
|
|
space in order to avoid accidental token paste. */
|
|
static void
|
|
output_token (pfile, token, prev)
|
|
cpp_reader *pfile;
|
|
const cpp_token *token, *prev;
|
|
{
|
|
int dummy;
|
|
|
|
if (token->col && (token->flags & BOL))
|
|
{
|
|
/* Supply enough whitespace to put this token in its original
|
|
column. Don't bother trying to reconstruct tabs; we can't
|
|
get it right in general, and nothing ought to care. (Yes,
|
|
some things do care; the fault lies with them.) */
|
|
unsigned char *buffer;
|
|
unsigned int spaces = token->col - 1;
|
|
|
|
CPP_RESERVE (pfile, token->col);
|
|
buffer = pfile->limit;
|
|
|
|
while (spaces--)
|
|
*buffer++ = ' ';
|
|
pfile->limit = buffer;
|
|
}
|
|
else if (token->flags & PREV_WHITE)
|
|
CPP_PUTC (pfile, ' ');
|
|
else if (prev)
|
|
{
|
|
/* Check for and prevent accidental token pasting. */
|
|
if (can_paste (pfile, prev, token, &dummy) != CPP_EOF)
|
|
CPP_PUTC (pfile, ' ');
|
|
/* can_paste doesn't catch all the accidental pastes.
|
|
Consider a + ++b - if there is not a space between the + and ++, it
|
|
will be misparsed as a++ + b. */
|
|
else if ((prev->type == CPP_PLUS && token->type == CPP_PLUS_PLUS)
|
|
|| (prev->type == CPP_MINUS && token->type == CPP_MINUS_MINUS))
|
|
CPP_PUTC (pfile, ' ');
|
|
}
|
|
|
|
CPP_RESERVE (pfile, TOKEN_LEN (token));
|
|
pfile->limit = spell_token (pfile, token, pfile->limit);
|
|
}
|
|
|
|
/* Write the spelling of a token TOKEN to BUFFER. The buffer must
|
|
already contain the enough space to hold the token's spelling.
|
|
Returns a pointer to the character after the last character
|
|
written. */
|
|
|
|
static unsigned char *
|
|
spell_token (pfile, token, buffer)
|
|
cpp_reader *pfile; /* Would be nice to be rid of this... */
|
|
const cpp_token *token;
|
|
unsigned char *buffer;
|
|
{
|
|
switch (TOKEN_SPELL (token))
|
|
{
|
|
case SPELL_OPERATOR:
|
|
{
|
|
const unsigned char *spelling;
|
|
unsigned char c;
|
|
|
|
if (token->flags & DIGRAPH)
|
|
spelling = digraph_spellings[token->type - CPP_FIRST_DIGRAPH];
|
|
else if (token->flags & NAMED_OP)
|
|
goto spell_ident;
|
|
else
|
|
spelling = TOKEN_NAME (token);
|
|
|
|
while ((c = *spelling++) != '\0')
|
|
*buffer++ = c;
|
|
}
|
|
break;
|
|
|
|
case SPELL_IDENT:
|
|
spell_ident:
|
|
memcpy (buffer, token->val.node->name, token->val.node->length);
|
|
buffer += token->val.node->length;
|
|
break;
|
|
|
|
case SPELL_STRING:
|
|
{
|
|
if (token->type == CPP_WSTRING || token->type == CPP_WCHAR)
|
|
*buffer++ = 'L';
|
|
|
|
if (token->type == CPP_STRING || token->type == CPP_WSTRING)
|
|
*buffer++ = '"';
|
|
if (token->type == CPP_CHAR || token->type == CPP_WCHAR)
|
|
*buffer++ = '\'';
|
|
|
|
memcpy (buffer, token->val.str.text, token->val.str.len);
|
|
buffer += token->val.str.len;
|
|
|
|
if (token->type == CPP_STRING || token->type == CPP_WSTRING)
|
|
*buffer++ = '"';
|
|
if (token->type == CPP_CHAR || token->type == CPP_WCHAR)
|
|
*buffer++ = '\'';
|
|
}
|
|
break;
|
|
|
|
case SPELL_CHAR:
|
|
*buffer++ = token->val.aux;
|
|
break;
|
|
|
|
case SPELL_NONE:
|
|
cpp_ice (pfile, "Unspellable token %s", TOKEN_NAME (token));
|
|
break;
|
|
}
|
|
|
|
return buffer;
|
|
}
|
|
|
|
/* Macro expansion algorithm.
|
|
|
|
Macro expansion is implemented by a single-pass algorithm; there are
|
|
no rescan passes involved. cpp_get_token expands just enough to be
|
|
able to return a token to the caller, a consequence is that when it
|
|
returns the preprocessor can be in a state of mid-expansion. The
|
|
algorithm does not work by fully expanding a macro invocation into
|
|
some kind of token list, and then returning them one by one.
|
|
|
|
Our expansion state is recorded in a context stack. We start out with
|
|
a single context on the stack, let's call it base context. This
|
|
consists of the token list returned by lex_line that forms the next
|
|
logical line in the source file.
|
|
|
|
The current level in the context stack is stored in the cur_context
|
|
member of the cpp_reader structure. The context it references keeps,
|
|
amongst other things, a count of how many tokens form that context and
|
|
our position within those tokens.
|
|
|
|
Fundamentally, calling cpp_get_token will return the next token from
|
|
the current context. If we're at the end of the current context, that
|
|
context is popped from the stack first, unless it is the base context,
|
|
in which case the next logical line is lexed from the source file.
|
|
|
|
However, before returning the token, if it is a CPP_NAME token
|
|
_cpp_get_token checks to see if it is a macro and if it is enabled.
|
|
Each time it encounters a macro name, it calls push_macro_context.
|
|
This function checks that the macro should be expanded (with
|
|
is_macro_enabled), and if so pushes a new macro context on the stack
|
|
which becomes the current context. It then loops back to read the
|
|
first token of the macro context.
|
|
|
|
A macro context basically consists of the token list representing the
|
|
macro's replacement list, which was saved in the hash table by
|
|
save_macro_expansion when its #define statement was parsed. If the
|
|
macro is function-like, it also contains the tokens that form the
|
|
arguments to the macro. I say more about macro arguments below, but
|
|
for now just saying that each argument is a set of pointers to tokens
|
|
is enough.
|
|
|
|
When taking tokens from a macro context, we may get a CPP_MACRO_ARG
|
|
token. This represents an argument passed to the macro, with the
|
|
argument number stored in the token's AUX field. The argument should
|
|
be substituted, this is achieved by pushing an "argument context". An
|
|
argument context is just refers to the tokens forming the argument,
|
|
which are obtained directly from the macro context. The STRINGIFY
|
|
flag on a CPP_MACRO_ARG token indicates that the argument should be
|
|
stringified.
|
|
|
|
Here's a few simple rules the context stack obeys:-
|
|
|
|
1) The lex_line token list is always context zero.
|
|
|
|
2) Context 1, if it exists, must be a macro context.
|
|
|
|
3) An argument context can only appear above a macro context.
|
|
|
|
4) A macro context can appear above the base context, another macro
|
|
context, or an argument context.
|
|
|
|
5) These imply that the minimal level of an argument context is 2.
|
|
|
|
The only tricky thing left is ensuring that macros are enabled and
|
|
disabled correctly. The algorithm controls macro expansion by the
|
|
level of the context a token is taken from in the context stack. If a
|
|
token is taken from a level equal to no_expand_level (a member of
|
|
struct cpp_reader), no expansion is performed.
|
|
|
|
When popping a context off the stack, if no_expand_level equals the
|
|
level of the popped context, it is reduced by one to match the new
|
|
context level, so that expansion is still disabled. It does not
|
|
increase if a context is pushed, though. It starts out life as
|
|
UINT_MAX, which has the effect that initially macro expansion is
|
|
enabled. I explain how this mechanism works below.
|
|
|
|
The standard requires:-
|
|
|
|
1) Arguments to be fully expanded before substitution.
|
|
|
|
2) Stringified arguments to not be expanded, nor the tokens
|
|
immediately surrounding a ## operator.
|
|
|
|
3) Continual rescanning until there are no more macros left to
|
|
replace.
|
|
|
|
4) Once a macro has been expanded in stage 1) or 3), it cannot be
|
|
expanded again during later rescans. This prevents infinite
|
|
recursion.
|
|
|
|
The first thing to observe is that stage 3) is mostly redundant.
|
|
Since a macro is disabled once it has been expanded, how can a rescan
|
|
find an unexpanded macro name? There are only two cases where this is
|
|
possible:-
|
|
|
|
a) If the macro name results from a token paste operation.
|
|
|
|
b) If the macro in question is a function-like macro that hasn't
|
|
already been expanded because previously there was not the required
|
|
'(' token immediately following it. This is only possible when an
|
|
argument is substituted, and after substitution the last token of
|
|
the argument can bind with a parenthesis appearing in the tokens
|
|
following the substitution. Note that if the '(' appears within the
|
|
argument, the ')' must too, as expanding macro arguments cannot
|
|
"suck in" tokens outside the argument.
|
|
|
|
So we tackle this as follows. When parsing the macro invocation for
|
|
arguments, we record the tokens forming each argument as a list of
|
|
pointers to those tokens. We do not expand any tokens that are "raw",
|
|
i.e. directly from the macro invocation, but other tokens that come
|
|
from (nested) argument substitution are fully expanded.
|
|
|
|
This is achieved by setting the no_expand_level to that of the macro
|
|
invocation. A CPP_MACRO_ARG token never appears in the list of tokens
|
|
forming an argument, because parse_args (indirectly) calls
|
|
get_raw_token which automatically pushes argument contexts and traces
|
|
into them. Since these contexts are at a higher level than the
|
|
no_expand_level, they get fully macro expanded.
|
|
|
|
"Raw" and non-raw tokens are separated in arguments by null pointers,
|
|
with the policy that the initial state of an argument is raw. If the
|
|
first token is not raw, it should be preceded by a null pointer. When
|
|
tracing through the tokens of an argument context, each time
|
|
get_raw_token encounters a null pointer, it toggles the flag
|
|
CONTEXT_RAW.
|
|
|
|
This flag, when set, indicates to is_macro_disabled that we are
|
|
reading raw tokens which should be macro-expanded. Similarly, if
|
|
clear, is_macro_disabled suppresses re-expansion.
|
|
|
|
It's probably time for an example.
|
|
|
|
#define hash #
|
|
#define str(x) #x
|
|
#define xstr(y) str(y hash)
|
|
str(hash) // "hash"
|
|
xstr(hash) // "# hash"
|
|
|
|
In the invocation of str, parse_args turns off macro expansion and so
|
|
parses the argument as <hash>. This is the only token (pointer)
|
|
passed as the argument to str. Since <hash> is raw there is no need
|
|
for an initial null pointer. stringify_arg is called from
|
|
get_raw_token when tracing through the expansion of str, since the
|
|
argument has the STRINGIFY flag set. stringify_arg turns off
|
|
macro_expansion by setting the no_expand_level to that of the argument
|
|
context. Thus it gets the token <hash> and stringifies it to "hash"
|
|
correctly.
|
|
|
|
Similary xstr is passed <hash>. However, when parse_args is parsing
|
|
the invocation of str() in xstr's expansion, get_raw_token encounters
|
|
a CPP_MACRO_ARG token for y. Transparently to parse_args, it pushes
|
|
an argument context, and enters the tokens of the argument,
|
|
i.e. <hash>. This is at a higher context level than parse_args
|
|
disabled, and so is_macro_disabled permits expansion of it and a macro
|
|
context is pushed on top of the argument context. This contains the
|
|
<#> token, and the end result is that <hash> is macro expanded.
|
|
However, after popping off the argument context, the <hash> of xstr's
|
|
expansion does not get macro expanded because we're back at the
|
|
no_expand_level. The end result is that the argument passed to str is
|
|
<NULL> <#> <NULL> <hash>. Note the nulls - policy is we start off
|
|
raw, <#> is not raw, but then <hash> is.
|
|
|
|
*/
|
|
|
|
|
|
/* Free the storage allocated for macro arguments. */
|
|
static void
|
|
free_macro_args (args)
|
|
macro_args *args;
|
|
{
|
|
if (args->tokens)
|
|
free ((PTR) args->tokens);
|
|
free (args->ends);
|
|
free (args);
|
|
}
|
|
|
|
/* Determines if a macro has been already used (and is therefore
|
|
disabled). */
|
|
static int
|
|
is_macro_disabled (pfile, expansion, token)
|
|
cpp_reader *pfile;
|
|
const cpp_toklist *expansion;
|
|
const cpp_token *token;
|
|
{
|
|
cpp_context *context = CURRENT_CONTEXT (pfile);
|
|
|
|
/* Don't expand anything if this file has already been preprocessed. */
|
|
if (CPP_OPTION (pfile, preprocessed))
|
|
return 1;
|
|
|
|
/* Arguments on either side of ## are inserted in place without
|
|
macro expansion (6.10.3.3.2). Conceptually, any macro expansion
|
|
occurs during a later rescan pass. The effect is that we expand
|
|
iff we would as part of the macro's expansion list, so we should
|
|
drop to the macro's context. */
|
|
if (IS_ARG_CONTEXT (context))
|
|
{
|
|
if (token->flags & PASTED)
|
|
context--;
|
|
else if (!(context->flags & CONTEXT_RAW))
|
|
return 1;
|
|
else if (context->flags & (CONTEXT_PASTEL | CONTEXT_PASTER))
|
|
context--;
|
|
}
|
|
|
|
/* Have we already used this macro? */
|
|
while (context->level > 0)
|
|
{
|
|
if (!IS_ARG_CONTEXT (context) && context->u.list == expansion)
|
|
return 1;
|
|
/* Raw argument tokens are judged based on the token list they
|
|
came from. */
|
|
if (context->flags & CONTEXT_RAW)
|
|
context = pfile->contexts + context->level;
|
|
else
|
|
context--;
|
|
}
|
|
|
|
/* Function-like macros may be disabled if the '(' is not in the
|
|
current context. We check this without disrupting the context
|
|
stack. */
|
|
if (expansion->paramc >= 0)
|
|
{
|
|
const cpp_token *next;
|
|
unsigned int prev_nme;
|
|
|
|
context = CURRENT_CONTEXT (pfile);
|
|
/* Drop down any contexts we're at the end of: the '(' may
|
|
appear in lower macro expansions, or in the rest of the file. */
|
|
while (context->posn == context->count && context > pfile->contexts)
|
|
{
|
|
context--;
|
|
/* If we matched, we are disabled, as we appear in the
|
|
expansion of each macro we meet. */
|
|
if (!IS_ARG_CONTEXT (context) && context->u.list == expansion)
|
|
return 1;
|
|
}
|
|
|
|
prev_nme = pfile->no_expand_level;
|
|
pfile->no_expand_level = context - pfile->contexts;
|
|
next = _cpp_get_token (pfile);
|
|
restore_macro_expansion (pfile, prev_nme);
|
|
if (next->type != CPP_OPEN_PAREN)
|
|
{
|
|
_cpp_push_token (pfile, next);
|
|
if (CPP_WTRADITIONAL (pfile))
|
|
cpp_warning (pfile,
|
|
"function macro %s must be used with arguments in traditional C",
|
|
token->val.node->name);
|
|
return 1;
|
|
}
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
/* Add a token to the set of tokens forming the arguments to the macro
|
|
being parsed in parse_args. */
|
|
static void
|
|
save_token (args, token)
|
|
macro_args *args;
|
|
const cpp_token *token;
|
|
{
|
|
if (args->used == args->capacity)
|
|
{
|
|
args->capacity += args->capacity + 100;
|
|
args->tokens = (const cpp_token **)
|
|
xrealloc ((PTR) args->tokens,
|
|
args->capacity * sizeof (const cpp_token *));
|
|
}
|
|
args->tokens[args->used++] = token;
|
|
}
|
|
|
|
/* Take and save raw tokens until we finish one argument. Empty
|
|
arguments are saved as a single CPP_PLACEMARKER token. */
|
|
static const cpp_token *
|
|
parse_arg (pfile, var_args, paren_context, args, pcount)
|
|
cpp_reader *pfile;
|
|
int var_args;
|
|
unsigned int paren_context;
|
|
macro_args *args;
|
|
unsigned int *pcount;
|
|
{
|
|
const cpp_token *token;
|
|
unsigned int paren = 0, count = 0;
|
|
int raw, was_raw = 1;
|
|
|
|
for (count = 0;; count++)
|
|
{
|
|
token = _cpp_get_token (pfile);
|
|
|
|
switch (token->type)
|
|
{
|
|
default:
|
|
break;
|
|
|
|
case CPP_OPEN_PAREN:
|
|
paren++;
|
|
break;
|
|
|
|
case CPP_CLOSE_PAREN:
|
|
if (paren-- != 0)
|
|
break;
|
|
goto out;
|
|
|
|
case CPP_COMMA:
|
|
/* Commas are not terminators within parantheses or var_args. */
|
|
if (paren || var_args)
|
|
break;
|
|
goto out;
|
|
|
|
case CPP_EOF: /* Error reported by caller. */
|
|
goto out;
|
|
}
|
|
|
|
raw = pfile->cur_context <= paren_context;
|
|
if (raw != was_raw)
|
|
{
|
|
was_raw = raw;
|
|
save_token (args, 0);
|
|
count++;
|
|
}
|
|
save_token (args, token);
|
|
}
|
|
|
|
out:
|
|
if (count == 0)
|
|
{
|
|
/* Duplicate the placemarker. Then we can set its flags and
|
|
position and safely be using more than one. */
|
|
save_token (args, duplicate_token (pfile, &placemarker_token));
|
|
count++;
|
|
}
|
|
|
|
*pcount = count;
|
|
return token;
|
|
}
|
|
|
|
/* This macro returns true if the argument starting at offset O of arglist
|
|
A is empty - that is, it's either a single PLACEMARKER token, or a null
|
|
pointer followed by a PLACEMARKER. */
|
|
|
|
#define empty_argument(A, O) \
|
|
((A)->tokens[O] ? (A)->tokens[O]->type == CPP_PLACEMARKER \
|
|
: (A)->tokens[(O)+1]->type == CPP_PLACEMARKER)
|
|
|
|
/* Parse the arguments making up a macro invocation. Nested arguments
|
|
are automatically macro expanded, but immediate macros are not
|
|
expanded; this enables e.g. operator # to work correctly. Returns
|
|
non-zero on error. */
|
|
static int
|
|
parse_args (pfile, hp, args)
|
|
cpp_reader *pfile;
|
|
cpp_hashnode *hp;
|
|
macro_args *args;
|
|
{
|
|
const cpp_token *token;
|
|
const cpp_toklist *macro;
|
|
unsigned int total = 0;
|
|
unsigned int paren_context = pfile->cur_context;
|
|
int argc = 0;
|
|
|
|
macro = hp->value.expansion;
|
|
do
|
|
{
|
|
unsigned int count;
|
|
|
|
token = parse_arg (pfile, (argc + 1 == macro->paramc
|
|
&& (macro->flags & VAR_ARGS)),
|
|
paren_context, args, &count);
|
|
if (argc < macro->paramc)
|
|
{
|
|
total += count;
|
|
args->ends[argc] = total;
|
|
}
|
|
argc++;
|
|
}
|
|
while (token->type != CPP_CLOSE_PAREN && token->type != CPP_EOF);
|
|
|
|
if (token->type == CPP_EOF)
|
|
{
|
|
cpp_error(pfile, "unterminated argument list for macro \"%s\"", hp->name);
|
|
return 1;
|
|
}
|
|
else if (argc < macro->paramc)
|
|
{
|
|
/* A rest argument is allowed to not appear in the invocation at all.
|
|
e.g. #define debug(format, args...) ...
|
|
debug("string");
|
|
This is exactly the same as if the rest argument had received no
|
|
tokens - debug("string",); This extension is deprecated. */
|
|
|
|
if (argc + 1 == macro->paramc && (macro->flags & VAR_ARGS))
|
|
{
|
|
/* Duplicate the placemarker. Then we can set its flags and
|
|
position and safely be using more than one. */
|
|
cpp_token *pm = duplicate_token (pfile, &placemarker_token);
|
|
pm->flags = VOID_REST;
|
|
save_token (args, pm);
|
|
args->ends[argc] = total + 1;
|
|
|
|
if (CPP_OPTION (pfile, c99) && CPP_PEDANTIC (pfile))
|
|
cpp_pedwarn (pfile, "ISO C99 requires rest arguments to be used");
|
|
|
|
return 0;
|
|
}
|
|
else
|
|
{
|
|
cpp_error (pfile, "not enough arguments for macro \"%s\"", hp->name);
|
|
return 1;
|
|
}
|
|
}
|
|
/* An empty argument to an empty function-like macro is fine. */
|
|
else if (argc > macro->paramc
|
|
&& !(macro->paramc == 0 && argc == 1 && empty_argument (args, 0)))
|
|
{
|
|
cpp_error (pfile, "too many arguments for macro \"%s\"", hp->name);
|
|
return 1;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
/* Adds backslashes before all backslashes and double quotes appearing
|
|
in strings. Non-printable characters are converted to octal. */
|
|
static U_CHAR *
|
|
quote_string (dest, src, len)
|
|
U_CHAR *dest;
|
|
const U_CHAR *src;
|
|
unsigned int len;
|
|
{
|
|
while (len--)
|
|
{
|
|
U_CHAR c = *src++;
|
|
|
|
if (c == '\\' || c == '"')
|
|
{
|
|
*dest++ = '\\';
|
|
*dest++ = c;
|
|
}
|
|
else
|
|
{
|
|
if (ISPRINT (c))
|
|
*dest++ = c;
|
|
else
|
|
{
|
|
sprintf ((char *) dest, "\\%03o", c);
|
|
dest += 4;
|
|
}
|
|
}
|
|
}
|
|
|
|
return dest;
|
|
}
|
|
|
|
/* Allocates a buffer to hold a token's TEXT, and converts TOKEN to a
|
|
CPP_STRING token containing TEXT in quoted form. */
|
|
static cpp_token *
|
|
make_string_token (token, text, len)
|
|
cpp_token *token;
|
|
const U_CHAR *text;
|
|
unsigned int len;
|
|
{
|
|
U_CHAR *buf;
|
|
|
|
buf = (U_CHAR *) xmalloc (len * 4);
|
|
token->type = CPP_STRING;
|
|
token->flags = 0;
|
|
token->val.str.text = buf;
|
|
token->val.str.len = quote_string (buf, text, len) - buf;
|
|
return token;
|
|
}
|
|
|
|
/* Allocates and converts a temporary token to a CPP_NUMBER token,
|
|
evaluating to NUMBER. */
|
|
static cpp_token *
|
|
alloc_number_token (pfile, number)
|
|
cpp_reader *pfile;
|
|
int number;
|
|
{
|
|
cpp_token *result;
|
|
char *buf;
|
|
|
|
result = get_temp_token (pfile);
|
|
buf = xmalloc (20);
|
|
sprintf (buf, "%d", number);
|
|
|
|
result->type = CPP_NUMBER;
|
|
result->flags = 0;
|
|
result->val.str.text = (U_CHAR *) buf;
|
|
result->val.str.len = strlen (buf);
|
|
return result;
|
|
}
|
|
|
|
/* Returns a temporary token from the temporary token store of PFILE. */
|
|
static cpp_token *
|
|
get_temp_token (pfile)
|
|
cpp_reader *pfile;
|
|
{
|
|
if (pfile->temp_used == pfile->temp_alloced)
|
|
{
|
|
if (pfile->temp_used == pfile->temp_cap)
|
|
{
|
|
pfile->temp_cap += pfile->temp_cap + 20;
|
|
pfile->temp_tokens = (cpp_token **) xrealloc
|
|
(pfile->temp_tokens, pfile->temp_cap * sizeof (cpp_token *));
|
|
}
|
|
pfile->temp_tokens[pfile->temp_alloced++] = (cpp_token *) xmalloc
|
|
(sizeof (cpp_token));
|
|
}
|
|
|
|
return pfile->temp_tokens[pfile->temp_used++];
|
|
}
|
|
|
|
/* Release (not free) for re-use the temporary tokens of PFILE. */
|
|
static void
|
|
release_temp_tokens (pfile)
|
|
cpp_reader *pfile;
|
|
{
|
|
while (pfile->temp_used)
|
|
{
|
|
cpp_token *token = pfile->temp_tokens[--pfile->temp_used];
|
|
|
|
if (TOKEN_SPELL (token) == SPELL_STRING)
|
|
{
|
|
free ((char *) token->val.str.text);
|
|
token->val.str.text = 0;
|
|
}
|
|
}
|
|
}
|
|
|
|
/* Free all of PFILE's dynamically-allocated temporary tokens. */
|
|
void
|
|
_cpp_free_temp_tokens (pfile)
|
|
cpp_reader *pfile;
|
|
{
|
|
if (pfile->temp_tokens)
|
|
{
|
|
/* It is possible, though unlikely (looking for '(' of a funlike
|
|
macro into EOF), that we haven't released the tokens yet. */
|
|
release_temp_tokens (pfile);
|
|
while (pfile->temp_alloced)
|
|
free (pfile->temp_tokens[--pfile->temp_alloced]);
|
|
free (pfile->temp_tokens);
|
|
}
|
|
|
|
if (pfile->date)
|
|
{
|
|
free ((char *) pfile->date->val.str.text);
|
|
free (pfile->date);
|
|
free ((char *) pfile->time->val.str.text);
|
|
free (pfile->time);
|
|
}
|
|
}
|
|
|
|
/* Copy TOKEN into a temporary token from PFILE's store. */
|
|
static cpp_token *
|
|
duplicate_token (pfile, token)
|
|
cpp_reader *pfile;
|
|
const cpp_token *token;
|
|
{
|
|
cpp_token *result = get_temp_token (pfile);
|
|
|
|
*result = *token;
|
|
if (TOKEN_SPELL (token) == SPELL_STRING)
|
|
{
|
|
U_CHAR *buff = (U_CHAR *) xmalloc (token->val.str.len);
|
|
memcpy (buff, token->val.str.text, token->val.str.len);
|
|
result->val.str.text = buff;
|
|
}
|
|
return result;
|
|
}
|
|
|
|
/* Determine whether two tokens can be pasted together, and if so,
|
|
what the resulting token is. Returns CPP_EOF if the tokens cannot
|
|
be pasted, or the appropriate type for the merged token if they
|
|
can. */
|
|
static enum cpp_ttype
|
|
can_paste (pfile, token1, token2, digraph)
|
|
cpp_reader * pfile;
|
|
const cpp_token *token1, *token2;
|
|
int* digraph;
|
|
{
|
|
enum cpp_ttype a = token1->type, b = token2->type;
|
|
int cxx = CPP_OPTION (pfile, cplusplus);
|
|
|
|
/* Treat named operators as if they were ordinary NAMEs. */
|
|
if (token1->flags & NAMED_OP)
|
|
a = CPP_NAME;
|
|
if (token2->flags & NAMED_OP)
|
|
b = CPP_NAME;
|
|
|
|
if (a <= CPP_LAST_EQ && b == CPP_EQ)
|
|
return a + (CPP_EQ_EQ - CPP_EQ);
|
|
|
|
switch (a)
|
|
{
|
|
case CPP_GREATER:
|
|
if (b == a) return CPP_RSHIFT;
|
|
if (b == CPP_QUERY && cxx) return CPP_MAX;
|
|
if (b == CPP_GREATER_EQ) return CPP_RSHIFT_EQ;
|
|
break;
|
|
case CPP_LESS:
|
|
if (b == a) return CPP_LSHIFT;
|
|
if (b == CPP_QUERY && cxx) return CPP_MIN;
|
|
if (b == CPP_LESS_EQ) return CPP_LSHIFT_EQ;
|
|
if (CPP_OPTION (pfile, digraphs))
|
|
{
|
|
if (b == CPP_COLON)
|
|
{*digraph = 1; return CPP_OPEN_SQUARE;} /* <: digraph */
|
|
if (b == CPP_MOD)
|
|
{*digraph = 1; return CPP_OPEN_BRACE;} /* <% digraph */
|
|
}
|
|
break;
|
|
|
|
case CPP_PLUS: if (b == a) return CPP_PLUS_PLUS; break;
|
|
case CPP_AND: if (b == a) return CPP_AND_AND; break;
|
|
case CPP_OR: if (b == a) return CPP_OR_OR; break;
|
|
|
|
case CPP_MINUS:
|
|
if (b == a) return CPP_MINUS_MINUS;
|
|
if (b == CPP_GREATER) return CPP_DEREF;
|
|
break;
|
|
case CPP_COLON:
|
|
if (b == a && cxx) return CPP_SCOPE;
|
|
if (b == CPP_GREATER && CPP_OPTION (pfile, digraphs))
|
|
{*digraph = 1; return CPP_CLOSE_SQUARE;} /* :> digraph */
|
|
break;
|
|
|
|
case CPP_MOD:
|
|
if (CPP_OPTION (pfile, digraphs))
|
|
{
|
|
if (b == CPP_GREATER)
|
|
{*digraph = 1; return CPP_CLOSE_BRACE;} /* %> digraph */
|
|
if (b == CPP_COLON)
|
|
{*digraph = 1; return CPP_HASH;} /* %: digraph */
|
|
}
|
|
break;
|
|
case CPP_DEREF:
|
|
if (b == CPP_MULT && cxx) return CPP_DEREF_STAR;
|
|
break;
|
|
case CPP_DOT:
|
|
if (b == CPP_MULT && cxx) return CPP_DOT_STAR;
|
|
if (b == CPP_NUMBER) return CPP_NUMBER;
|
|
break;
|
|
|
|
case CPP_HASH:
|
|
if (b == a && (token1->flags & DIGRAPH) == (token2->flags & DIGRAPH))
|
|
/* %:%: digraph */
|
|
{*digraph = (token1->flags & DIGRAPH); return CPP_PASTE;}
|
|
break;
|
|
|
|
case CPP_NAME:
|
|
if (b == CPP_NAME) return CPP_NAME;
|
|
if (b == CPP_NUMBER
|
|
&& is_numstart(token2->val.str.text[0])) return CPP_NAME;
|
|
if (b == CPP_CHAR
|
|
&& token1->val.node == pfile->spec_nodes->n_L) return CPP_WCHAR;
|
|
if (b == CPP_STRING
|
|
&& token1->val.node == pfile->spec_nodes->n_L) return CPP_WSTRING;
|
|
break;
|
|
|
|
case CPP_NUMBER:
|
|
if (b == CPP_NUMBER) return CPP_NUMBER;
|
|
if (b == CPP_NAME) return CPP_NUMBER;
|
|
if (b == CPP_DOT) return CPP_NUMBER;
|
|
/* Numbers cannot have length zero, so this is safe. */
|
|
if ((b == CPP_PLUS || b == CPP_MINUS)
|
|
&& VALID_SIGN ('+', token1->val.str.text[token1->val.str.len - 1]))
|
|
return CPP_NUMBER;
|
|
break;
|
|
|
|
default:
|
|
break;
|
|
}
|
|
|
|
return CPP_EOF;
|
|
}
|
|
|
|
/* Check if TOKEN is to be ##-pasted with the token after it. */
|
|
static const cpp_token *
|
|
maybe_paste_with_next (pfile, token)
|
|
cpp_reader *pfile;
|
|
const cpp_token *token;
|
|
{
|
|
cpp_token *pasted;
|
|
const cpp_token *second;
|
|
cpp_context *context = CURRENT_CONTEXT (pfile);
|
|
|
|
/* Is this token on the LHS of ## ? */
|
|
|
|
while ((token->flags & PASTE_LEFT)
|
|
|| ((context->flags & CONTEXT_PASTEL)
|
|
&& context->posn == context->count))
|
|
{
|
|
/* Suppress macro expansion for next token, but don't conflict
|
|
with the other method of suppression. If it is an argument,
|
|
macro expansion within the argument will still occur. */
|
|
pfile->paste_level = pfile->cur_context;
|
|
second = _cpp_get_token (pfile);
|
|
pfile->paste_level = 0;
|
|
|
|
/* Ignore placemarker argument tokens (cannot be from an empty
|
|
macro since macros are not expanded). */
|
|
if (token->type == CPP_PLACEMARKER)
|
|
pasted = duplicate_token (pfile, second);
|
|
else if (second->type == CPP_PLACEMARKER)
|
|
{
|
|
/* GCC has special extended semantics for a ## b where b is
|
|
a varargs parameter: a disappears if b was given no actual
|
|
arguments (not merely if b is an empty argument). */
|
|
if (second->flags & VOID_REST)
|
|
pasted = duplicate_token (pfile, second);
|
|
else
|
|
pasted = duplicate_token (pfile, token);
|
|
}
|
|
else
|
|
{
|
|
int digraph = 0;
|
|
enum cpp_ttype type = can_paste (pfile, token, second, &digraph);
|
|
|
|
if (type == CPP_EOF)
|
|
{
|
|
if (CPP_OPTION (pfile, warn_paste))
|
|
cpp_warning (pfile,
|
|
"pasting would not give a valid preprocessing token");
|
|
_cpp_push_token (pfile, second);
|
|
return token;
|
|
}
|
|
|
|
if (type == CPP_NAME || type == CPP_NUMBER)
|
|
{
|
|
/* Join spellings. */
|
|
U_CHAR *buf, *end;
|
|
|
|
pasted = get_temp_token (pfile);
|
|
buf = (U_CHAR *) alloca (TOKEN_LEN (token) + TOKEN_LEN (second));
|
|
end = spell_token (pfile, token, buf);
|
|
end = spell_token (pfile, second, end);
|
|
*end = '\0';
|
|
|
|
if (type == CPP_NAME)
|
|
pasted->val.node = cpp_lookup (pfile, buf, end - buf);
|
|
else
|
|
{
|
|
pasted->val.str.text = uxstrdup (buf);
|
|
pasted->val.str.len = end - buf;
|
|
}
|
|
}
|
|
else if (type == CPP_WCHAR || type == CPP_WSTRING)
|
|
pasted = duplicate_token (pfile, second);
|
|
else
|
|
{
|
|
pasted = get_temp_token (pfile);
|
|
pasted->val.integer = 0;
|
|
}
|
|
|
|
pasted->type = type;
|
|
pasted->flags = digraph ? DIGRAPH : 0;
|
|
|
|
if (type == CPP_NAME && pasted->val.node->type == T_OPERATOR)
|
|
{
|
|
pasted->type = pasted->val.node->value.code;
|
|
pasted->flags |= NAMED_OP;
|
|
}
|
|
}
|
|
|
|
/* The pasted token gets the whitespace flags and position of the
|
|
first token, the PASTE_LEFT flag of the second token, plus the
|
|
PASTED flag to indicate it is the result of a paste. However, we
|
|
want to preserve the DIGRAPH flag. */
|
|
pasted->flags &= ~(PREV_WHITE | BOL | PASTE_LEFT);
|
|
pasted->flags |= ((token->flags & (PREV_WHITE | BOL))
|
|
| (second->flags & PASTE_LEFT) | PASTED);
|
|
pasted->col = token->col;
|
|
pasted->line = token->line;
|
|
|
|
/* See if there is another token to be pasted onto the one we just
|
|
constructed. */
|
|
token = pasted;
|
|
context = CURRENT_CONTEXT (pfile);
|
|
/* and loop */
|
|
}
|
|
return token;
|
|
}
|
|
|
|
/* Convert a token sequence to a single string token according to the
|
|
rules of the ISO C #-operator. */
|
|
#define INIT_SIZE 200
|
|
static cpp_token *
|
|
stringify_arg (pfile, token)
|
|
cpp_reader *pfile;
|
|
const cpp_token *token;
|
|
{
|
|
cpp_token *result;
|
|
unsigned char *main_buf;
|
|
unsigned int prev_value, backslash_count = 0;
|
|
unsigned int buf_used = 0, whitespace = 0, buf_cap = INIT_SIZE;
|
|
|
|
push_arg_context (pfile, token);
|
|
prev_value = prevent_macro_expansion (pfile);
|
|
main_buf = (unsigned char *) xmalloc (buf_cap);
|
|
|
|
result = get_temp_token (pfile);
|
|
ASSIGN_FLAGS_AND_POS (result, token);
|
|
|
|
for (; (token = _cpp_get_token (pfile))->type != CPP_EOF; )
|
|
{
|
|
int escape;
|
|
unsigned char *buf;
|
|
unsigned int len = TOKEN_LEN (token);
|
|
|
|
if (token->type == CPP_PLACEMARKER)
|
|
continue;
|
|
|
|
escape = (token->type == CPP_STRING || token->type == CPP_WSTRING
|
|
|| token->type == CPP_CHAR || token->type == CPP_WCHAR);
|
|
if (escape)
|
|
len *= 4 + 1;
|
|
|
|
if (buf_used + len > buf_cap)
|
|
{
|
|
buf_cap = buf_used + len + INIT_SIZE;
|
|
main_buf = xrealloc (main_buf, buf_cap);
|
|
}
|
|
|
|
if (whitespace && (token->flags & PREV_WHITE))
|
|
main_buf[buf_used++] = ' ';
|
|
|
|
if (escape)
|
|
buf = (unsigned char *) xmalloc (len);
|
|
else
|
|
buf = main_buf + buf_used;
|
|
|
|
len = spell_token (pfile, token, buf) - buf;
|
|
if (escape)
|
|
{
|
|
buf_used = quote_string (&main_buf[buf_used], buf, len) - main_buf;
|
|
free (buf);
|
|
}
|
|
else
|
|
buf_used += len;
|
|
|
|
whitespace = 1;
|
|
if (token->type == CPP_BACKSLASH)
|
|
backslash_count++;
|
|
else
|
|
backslash_count = 0;
|
|
}
|
|
|
|
/* Ignore the final \ of invalid string literals. */
|
|
if (backslash_count & 1)
|
|
{
|
|
cpp_warning (pfile, "invalid string literal, ignoring final '\\'");
|
|
buf_used--;
|
|
}
|
|
|
|
result->type = CPP_STRING;
|
|
result->val.str.text = main_buf;
|
|
result->val.str.len = buf_used;
|
|
restore_macro_expansion (pfile, prev_value);
|
|
return result;
|
|
}
|
|
|
|
/* Allocate more room on the context stack of PFILE. */
|
|
static void
|
|
expand_context_stack (pfile)
|
|
cpp_reader *pfile;
|
|
{
|
|
pfile->context_cap += pfile->context_cap + 20;
|
|
pfile->contexts = (cpp_context *)
|
|
xrealloc (pfile->contexts, pfile->context_cap * sizeof (cpp_context));
|
|
}
|
|
|
|
/* Push the context of macro NODE onto the context stack. TOKEN is
|
|
the CPP_NAME token invoking the macro. */
|
|
static int
|
|
push_macro_context (pfile, token)
|
|
cpp_reader *pfile;
|
|
const cpp_token *token;
|
|
{
|
|
unsigned char orig_flags;
|
|
macro_args *args;
|
|
cpp_context *context;
|
|
cpp_hashnode *node = token->val.node;
|
|
|
|
/* Token's flags may change when parsing args containing a nested
|
|
invocation of this macro. */
|
|
orig_flags = token->flags & (PREV_WHITE | BOL);
|
|
args = 0;
|
|
if (node->value.expansion->paramc >= 0)
|
|
{
|
|
unsigned int error, prev_nme;
|
|
|
|
/* Allocate room for the argument contexts, and parse them. */
|
|
args = (macro_args *) xmalloc (sizeof (macro_args));
|
|
args->ends = (unsigned int *)
|
|
xmalloc (node->value.expansion->paramc * sizeof (unsigned int));
|
|
args->tokens = 0;
|
|
args->capacity = 0;
|
|
args->used = 0;
|
|
args->level = pfile->cur_context;
|
|
|
|
prev_nme = prevent_macro_expansion (pfile);
|
|
pfile->args = args;
|
|
error = parse_args (pfile, node, args);
|
|
pfile->args = 0;
|
|
restore_macro_expansion (pfile, prev_nme);
|
|
if (error)
|
|
{
|
|
free_macro_args (args);
|
|
return 1;
|
|
}
|
|
}
|
|
|
|
/* Now push its context. */
|
|
pfile->cur_context++;
|
|
if (pfile->cur_context == pfile->context_cap)
|
|
expand_context_stack (pfile);
|
|
|
|
context = CURRENT_CONTEXT (pfile);
|
|
context->u.list = node->value.expansion;
|
|
context->args = args;
|
|
context->posn = 0;
|
|
context->count = context->u.list->tokens_used;
|
|
context->level = pfile->cur_context;
|
|
context->flags = 0;
|
|
context->pushed_token = 0;
|
|
|
|
/* Set the flags of the first token. We know there must
|
|
be one, empty macros are a single placemarker token. */
|
|
MODIFY_FLAGS_AND_POS (&context->u.list->tokens[0], token, orig_flags);
|
|
|
|
return 0;
|
|
}
|
|
|
|
/* Push an argument to the current macro onto the context stack.
|
|
TOKEN is the MACRO_ARG token representing the argument expansion. */
|
|
static void
|
|
push_arg_context (pfile, token)
|
|
cpp_reader *pfile;
|
|
const cpp_token *token;
|
|
{
|
|
cpp_context *context;
|
|
macro_args *args;
|
|
|
|
pfile->cur_context++;
|
|
if (pfile->cur_context == pfile->context_cap)
|
|
expand_context_stack (pfile);
|
|
|
|
context = CURRENT_CONTEXT (pfile);
|
|
args = context[-1].args;
|
|
|
|
context->count = token->val.aux ? args->ends[token->val.aux - 1]: 0;
|
|
context->u.arg = args->tokens + context->count;
|
|
context->count = args->ends[token->val.aux] - context->count;
|
|
context->args = 0;
|
|
context->posn = 0;
|
|
context->level = args->level;
|
|
context->flags = CONTEXT_ARG | CONTEXT_RAW;
|
|
context->pushed_token = 0;
|
|
|
|
/* Set the flags of the first token. There is one. */
|
|
{
|
|
const cpp_token *first = context->u.arg[0];
|
|
if (!first)
|
|
first = context->u.arg[1];
|
|
|
|
MODIFY_FLAGS_AND_POS ((cpp_token *) first, token,
|
|
token->flags & (PREV_WHITE | BOL));
|
|
}
|
|
|
|
if (token->flags & PASTE_LEFT)
|
|
context->flags |= CONTEXT_PASTEL;
|
|
if (pfile->paste_level)
|
|
context->flags |= CONTEXT_PASTER;
|
|
}
|
|
|
|
/* "Unget" a token. It is effectively inserted in the token queue and
|
|
will be returned by the next call to get_raw_token. */
|
|
void
|
|
_cpp_push_token (pfile, token)
|
|
cpp_reader *pfile;
|
|
const cpp_token *token;
|
|
{
|
|
cpp_context *context = CURRENT_CONTEXT (pfile);
|
|
|
|
if (context->posn > 0)
|
|
{
|
|
const cpp_token *prev;
|
|
if (IS_ARG_CONTEXT (context))
|
|
prev = context->u.arg[context->posn - 1];
|
|
else
|
|
prev = &context->u.list->tokens[context->posn - 1];
|
|
|
|
if (prev == token)
|
|
{
|
|
context->posn--;
|
|
return;
|
|
}
|
|
}
|
|
|
|
if (context->pushed_token)
|
|
cpp_ice (pfile, "two tokens pushed in a row");
|
|
if (token->type != CPP_EOF)
|
|
context->pushed_token = token;
|
|
/* Don't push back a directive's CPP_EOF, step back instead. */
|
|
else if (pfile->cur_context == 0)
|
|
pfile->contexts[0].posn--;
|
|
}
|
|
|
|
/* Handle a preprocessing directive. TOKEN is the CPP_HASH token
|
|
introducing the directive. */
|
|
static void
|
|
process_directive (pfile, token)
|
|
cpp_reader *pfile;
|
|
const cpp_token *token;
|
|
{
|
|
const struct directive *d = pfile->token_list.directive;
|
|
int prev_nme = 0;
|
|
|
|
/* Skip over the directive name. */
|
|
if (token[1].type == CPP_NAME)
|
|
_cpp_get_raw_token (pfile);
|
|
else if (token[1].type != CPP_NUMBER)
|
|
cpp_ice (pfile, "directive begins with %s?!", TOKEN_NAME (token));
|
|
|
|
/* Flush pending tokens at this point, in case the directive produces
|
|
output. XXX Directive output won't be visible to a direct caller of
|
|
cpp_get_token. */
|
|
if (pfile->printer && CPP_WRITTEN (pfile) - pfile->printer->written)
|
|
cpp_output_tokens (pfile, pfile->printer, pfile->token_list.line);
|
|
|
|
if (! (d->flags & EXPAND))
|
|
prev_nme = prevent_macro_expansion (pfile);
|
|
(void) (*d->handler) (pfile);
|
|
if (! (d->flags & EXPAND))
|
|
restore_macro_expansion (pfile, prev_nme);
|
|
_cpp_skip_rest_of_line (pfile);
|
|
}
|
|
|
|
/* The external interface to return the next token. All macro
|
|
expansion and directive processing is handled internally, the
|
|
caller only ever sees the output after preprocessing. */
|
|
const cpp_token *
|
|
cpp_get_token (pfile)
|
|
cpp_reader *pfile;
|
|
{
|
|
const cpp_token *token;
|
|
/* Loop till we hit a non-directive, non-placemarker token. */
|
|
for (;;)
|
|
{
|
|
token = _cpp_get_token (pfile);
|
|
|
|
if (token->type == CPP_PLACEMARKER)
|
|
continue;
|
|
|
|
if (token->type == CPP_HASH && token->flags & BOL
|
|
&& pfile->token_list.directive)
|
|
{
|
|
process_directive (pfile, token);
|
|
continue;
|
|
}
|
|
|
|
return token;
|
|
}
|
|
}
|
|
|
|
/* The internal interface to return the next token. There are two
|
|
differences between the internal and external interfaces: the
|
|
internal interface may return a PLACEMARKER token, and it does not
|
|
process directives. */
|
|
const cpp_token *
|
|
_cpp_get_token (pfile)
|
|
cpp_reader *pfile;
|
|
{
|
|
const cpp_token *token, *old_token;
|
|
cpp_hashnode *node;
|
|
|
|
/* Loop until we hit a non-macro token. */
|
|
for (;;)
|
|
{
|
|
token = get_raw_token (pfile);
|
|
|
|
/* Short circuit EOF. */
|
|
if (token->type == CPP_EOF)
|
|
return token;
|
|
|
|
/* If we are skipping... */
|
|
if (pfile->skipping)
|
|
{
|
|
/* we still have to process directives, */
|
|
if (pfile->token_list.directive)
|
|
return token;
|
|
|
|
/* but everything else is ignored. */
|
|
_cpp_skip_rest_of_line (pfile);
|
|
continue;
|
|
}
|
|
|
|
/* If there's a potential control macro and we get here, then that
|
|
#ifndef didn't cover the entire file and its argument shouldn't
|
|
be taken as a control macro. */
|
|
pfile->potential_control_macro = 0;
|
|
|
|
old_token = token;
|
|
|
|
/* See if there's a token to paste with this one. */
|
|
if (!pfile->paste_level)
|
|
token = maybe_paste_with_next (pfile, token);
|
|
|
|
/* If it isn't a macro, return it now. */
|
|
if (token->type != CPP_NAME || token->val.node->type == T_VOID)
|
|
return token;
|
|
|
|
/* Is macro expansion disabled in general, or are we in the
|
|
middle of a token paste, or was this token just pasted?
|
|
(Note we don't check token->flags & PASTED, because that
|
|
counts tokens that were pasted at some point in the past,
|
|
we're only interested in tokens that were pasted by this call
|
|
to maybe_paste_with_next.) */
|
|
if (pfile->no_expand_level == pfile->cur_context
|
|
|| pfile->paste_level
|
|
|| (token != old_token
|
|
&& pfile->no_expand_level + 1 == pfile->cur_context))
|
|
return token;
|
|
|
|
node = token->val.node;
|
|
if (node->type != T_MACRO)
|
|
return special_symbol (pfile, node, token);
|
|
|
|
if (is_macro_disabled (pfile, node->value.expansion, token))
|
|
return token;
|
|
|
|
if (pfile->cur_context > CPP_STACK_MAX)
|
|
{
|
|
cpp_error (pfile, "macros nested too deep invoking '%s'", node->name);
|
|
return token;
|
|
}
|
|
|
|
if (push_macro_context (pfile, token))
|
|
return token;
|
|
/* else loop */
|
|
}
|
|
}
|
|
|
|
/* Returns the next raw token, i.e. without performing macro
|
|
expansion. Argument contexts are automatically entered. */
|
|
static const cpp_token *
|
|
get_raw_token (pfile)
|
|
cpp_reader *pfile;
|
|
{
|
|
const cpp_token *result;
|
|
cpp_context *context;
|
|
|
|
for (;;)
|
|
{
|
|
context = CURRENT_CONTEXT (pfile);
|
|
if (context->pushed_token)
|
|
{
|
|
result = context->pushed_token;
|
|
context->pushed_token = 0;
|
|
return result; /* Cannot be a CPP_MACRO_ARG */
|
|
}
|
|
else if (context->posn == context->count)
|
|
{
|
|
if (pop_context (pfile))
|
|
return &eof_token;
|
|
continue;
|
|
}
|
|
else if (IS_ARG_CONTEXT (context))
|
|
{
|
|
result = context->u.arg[context->posn++];
|
|
if (result == 0)
|
|
{
|
|
context->flags ^= CONTEXT_RAW;
|
|
result = context->u.arg[context->posn++];
|
|
}
|
|
return result; /* Cannot be a CPP_MACRO_ARG */
|
|
}
|
|
|
|
result = &context->u.list->tokens[context->posn++];
|
|
|
|
if (result->type != CPP_MACRO_ARG)
|
|
return result;
|
|
|
|
if (result->flags & STRINGIFY_ARG)
|
|
return stringify_arg (pfile, result);
|
|
|
|
push_arg_context (pfile, result);
|
|
}
|
|
}
|
|
|
|
/* Internal interface to get the token without macro expanding. */
|
|
const cpp_token *
|
|
_cpp_get_raw_token (pfile)
|
|
cpp_reader *pfile;
|
|
{
|
|
int prev_nme = prevent_macro_expansion (pfile);
|
|
const cpp_token *result = _cpp_get_token (pfile);
|
|
restore_macro_expansion (pfile, prev_nme);
|
|
return result;
|
|
}
|
|
|
|
/* A thin wrapper to lex_line. CLEAR is non-zero if the current token
|
|
list should be overwritten, or zero if we need to append
|
|
(typically, if we are within the arguments to a macro, or looking
|
|
for the '(' to start a function-like macro invocation). */
|
|
static int
|
|
lex_next (pfile, clear)
|
|
cpp_reader *pfile;
|
|
int clear;
|
|
{
|
|
cpp_toklist *list = &pfile->token_list;
|
|
const cpp_token *old_list = list->tokens;
|
|
unsigned int old_used = list->tokens_used;
|
|
|
|
if (clear)
|
|
{
|
|
/* Release all temporary tokens. */
|
|
_cpp_clear_toklist (list);
|
|
pfile->contexts[0].posn = 0;
|
|
if (pfile->temp_used)
|
|
release_temp_tokens (pfile);
|
|
}
|
|
lex_line (pfile, list);
|
|
pfile->contexts[0].count = list->tokens_used;
|
|
|
|
if (!clear && pfile->args)
|
|
{
|
|
/* Fix up argument token pointers. */
|
|
if (old_list != list->tokens)
|
|
{
|
|
unsigned int i;
|
|
|
|
for (i = 0; i < pfile->args->used; i++)
|
|
{
|
|
const cpp_token *token = pfile->args->tokens[i];
|
|
if (token >= old_list && token < old_list + old_used)
|
|
pfile->args->tokens[i] = (const cpp_token *)
|
|
((char *) token + ((char *) list->tokens - (char *) old_list));
|
|
}
|
|
}
|
|
|
|
/* 6.10.3 paragraph 11: If there are sequences of preprocessing
|
|
tokens within the list of arguments that would otherwise act as
|
|
preprocessing directives, the behavior is undefined.
|
|
|
|
This implementation will report a hard error and treat the
|
|
'sequence of preprocessing tokens' as part of the macro argument,
|
|
not a directive.
|
|
|
|
Note if pfile->args == 0, we're OK since we're only inside a
|
|
macro argument after a '('. */
|
|
if (list->directive)
|
|
{
|
|
cpp_error_with_line (pfile, list->tokens[old_used].line,
|
|
list->tokens[old_used].col,
|
|
"#%s may not be used inside a macro argument",
|
|
list->directive->name);
|
|
return 1;
|
|
}
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
/* Pops a context off the context stack. If we're at the bottom, lexes
|
|
the next logical line. Returns EOF if we're at the end of the
|
|
argument list to the # operator, or if it is illegal to "overflow"
|
|
into the rest of the file (e.g. 6.10.3.1.1). */
|
|
static int
|
|
pop_context (pfile)
|
|
cpp_reader *pfile;
|
|
{
|
|
cpp_context *context;
|
|
|
|
if (pfile->cur_context == 0)
|
|
{
|
|
/* If we are currently processing a directive, do not advance. 6.10
|
|
paragraph 2: A new-line character ends the directive even if it
|
|
occurs within what would otherwise be an invocation of a
|
|
function-like macro. */
|
|
if (pfile->token_list.directive)
|
|
return 1;
|
|
|
|
return lex_next (pfile, pfile->no_expand_level == UINT_MAX);
|
|
}
|
|
|
|
/* Argument contexts, when parsing args or handling # operator
|
|
return CPP_EOF at the end. */
|
|
context = CURRENT_CONTEXT (pfile);
|
|
if (IS_ARG_CONTEXT (context) && pfile->cur_context == pfile->no_expand_level)
|
|
return 1;
|
|
|
|
/* Free resources when leaving macro contexts. */
|
|
if (context->args)
|
|
free_macro_args (context->args);
|
|
|
|
if (pfile->cur_context == pfile->no_expand_level)
|
|
pfile->no_expand_level--;
|
|
pfile->cur_context--;
|
|
|
|
return 0;
|
|
}
|
|
|
|
/* Turn off macro expansion at the current context level. */
|
|
static unsigned int
|
|
prevent_macro_expansion (pfile)
|
|
cpp_reader *pfile;
|
|
{
|
|
unsigned int prev_value = pfile->no_expand_level;
|
|
pfile->no_expand_level = pfile->cur_context;
|
|
return prev_value;
|
|
}
|
|
|
|
/* Restore macro expansion to its previous state. */
|
|
static void
|
|
restore_macro_expansion (pfile, prev_value)
|
|
cpp_reader *pfile;
|
|
unsigned int prev_value;
|
|
{
|
|
pfile->no_expand_level = prev_value;
|
|
}
|
|
|
|
/* Used by cpperror.c to obtain the correct line and column to report
|
|
in a diagnostic. */
|
|
unsigned int
|
|
_cpp_get_line (pfile, pcol)
|
|
cpp_reader *pfile;
|
|
unsigned int *pcol;
|
|
{
|
|
unsigned int index;
|
|
const cpp_token *cur_token;
|
|
|
|
if (pfile->in_lex_line)
|
|
index = pfile->token_list.tokens_used;
|
|
else
|
|
index = pfile->contexts[0].posn;
|
|
|
|
if (index == 0)
|
|
{
|
|
if (pcol)
|
|
*pcol = 0;
|
|
return 0;
|
|
}
|
|
|
|
cur_token = &pfile->token_list.tokens[index - 1];
|
|
if (pcol)
|
|
*pcol = cur_token->col;
|
|
return cur_token->line;
|
|
}
|
|
|
|
#define DSC(str) (const U_CHAR *)str, sizeof str - 1
|
|
static const char * const monthnames[] =
|
|
{
|
|
"Jan", "Feb", "Mar", "Apr", "May", "Jun",
|
|
"Jul", "Aug", "Sep", "Oct", "Nov", "Dec",
|
|
};
|
|
|
|
/* Handle builtin macros like __FILE__. */
|
|
static const cpp_token *
|
|
special_symbol (pfile, node, token)
|
|
cpp_reader *pfile;
|
|
cpp_hashnode *node;
|
|
const cpp_token *token;
|
|
{
|
|
cpp_token *result;
|
|
cpp_buffer *ip;
|
|
|
|
switch (node->type)
|
|
{
|
|
case T_FILE:
|
|
case T_BASE_FILE:
|
|
{
|
|
const char *file;
|
|
|
|
ip = CPP_BUFFER (pfile);
|
|
if (ip == 0)
|
|
file = "";
|
|
else
|
|
{
|
|
if (node->type == T_BASE_FILE)
|
|
while (CPP_PREV_BUFFER (ip) != NULL)
|
|
ip = CPP_PREV_BUFFER (ip);
|
|
|
|
file = ip->nominal_fname;
|
|
}
|
|
result = make_string_token (get_temp_token (pfile), (U_CHAR *) file,
|
|
strlen (file));
|
|
}
|
|
break;
|
|
|
|
case T_INCLUDE_LEVEL:
|
|
/* pfile->include_depth counts the primary source as level 1,
|
|
but historically __INCLUDE_DEPTH__ has called the primary
|
|
source level 0. */
|
|
result = alloc_number_token (pfile, pfile->include_depth - 1);
|
|
break;
|
|
|
|
case T_SPECLINE:
|
|
/* If __LINE__ is embedded in a macro, it must expand to the
|
|
line of the macro's invocation, not its definition.
|
|
Otherwise things like assert() will not work properly. */
|
|
result = alloc_number_token (pfile, _cpp_get_line (pfile, NULL));
|
|
break;
|
|
|
|
case T_STDC:
|
|
{
|
|
int stdc = 1;
|
|
|
|
#ifdef STDC_0_IN_SYSTEM_HEADERS
|
|
if (CPP_IN_SYSTEM_HEADER (pfile)
|
|
&& pfile->spec_nodes->n__STRICT_ANSI__->type == T_VOID)
|
|
stdc = 0;
|
|
#endif
|
|
result = alloc_number_token (pfile, stdc);
|
|
}
|
|
break;
|
|
|
|
case T_DATE:
|
|
case T_TIME:
|
|
if (pfile->date == 0)
|
|
{
|
|
/* Allocate __DATE__ and __TIME__ from permanent storage,
|
|
and save them in pfile so we don't have to do this again.
|
|
We don't generate these strings at init time because
|
|
time() and localtime() are very slow on some systems. */
|
|
time_t tt = time (NULL);
|
|
struct tm *tb = localtime (&tt);
|
|
|
|
pfile->date = make_string_token
|
|
((cpp_token *) xmalloc (sizeof (cpp_token)), DSC("Oct 11 1347"));
|
|
pfile->time = make_string_token
|
|
((cpp_token *) xmalloc (sizeof (cpp_token)), DSC("12:34:56"));
|
|
|
|
sprintf ((char *) pfile->date->val.str.text, "%s %2d %4d",
|
|
monthnames[tb->tm_mon], tb->tm_mday, tb->tm_year + 1900);
|
|
sprintf ((char *) pfile->time->val.str.text, "%02d:%02d:%02d",
|
|
tb->tm_hour, tb->tm_min, tb->tm_sec);
|
|
}
|
|
result = node->type == T_DATE ? pfile->date: pfile->time;
|
|
break;
|
|
|
|
case T_POISON:
|
|
cpp_error (pfile, "attempt to use poisoned \"%s\"", node->name);
|
|
return token;
|
|
|
|
default:
|
|
cpp_ice (pfile, "invalid special hash type");
|
|
return token;
|
|
}
|
|
|
|
ASSIGN_FLAGS_AND_POS (result, token);
|
|
return result;
|
|
}
|
|
#undef DSC
|
|
|
|
/* Dump the original user's spelling of argument index ARG_NO to the
|
|
macro whose expansion is LIST. */
|
|
static void
|
|
dump_param_spelling (pfile, list, arg_no)
|
|
cpp_reader *pfile;
|
|
const cpp_toklist *list;
|
|
unsigned int arg_no;
|
|
{
|
|
const U_CHAR *param = list->namebuf;
|
|
|
|
while (arg_no--)
|
|
param += ustrlen (param) + 1;
|
|
CPP_PUTS (pfile, param, ustrlen (param));
|
|
}
|
|
|
|
/* Dump a token list to the output. */
|
|
void
|
|
_cpp_dump_list (pfile, list, token, flush)
|
|
cpp_reader *pfile;
|
|
const cpp_toklist *list;
|
|
const cpp_token *token;
|
|
int flush;
|
|
{
|
|
const cpp_token *limit = list->tokens + list->tokens_used;
|
|
const cpp_token *prev = 0;
|
|
|
|
/* Avoid the CPP_EOF. */
|
|
if (list->directive)
|
|
limit--;
|
|
|
|
while (token < limit)
|
|
{
|
|
if (token->type == CPP_MACRO_ARG)
|
|
{
|
|
if (token->flags & PREV_WHITE)
|
|
CPP_PUTC (pfile, ' ');
|
|
if (token->flags & STRINGIFY_ARG)
|
|
CPP_PUTC (pfile, '#');
|
|
dump_param_spelling (pfile, list, token->val.aux);
|
|
}
|
|
else if (token->type != CPP_PLACEMARKER)
|
|
output_token (pfile, token, prev);
|
|
if (token->flags & PASTE_LEFT)
|
|
CPP_PUTS (pfile, " ##", 3);
|
|
prev = token;
|
|
token++;
|
|
}
|
|
|
|
if (flush && pfile->printer)
|
|
cpp_output_tokens (pfile, pfile->printer, pfile->token_list.line);
|
|
}
|
|
|
|
/* Allocate pfile->input_buffer, and initialize trigraph_map[]
|
|
if it hasn't happened already. */
|
|
|
|
void
|
|
_cpp_init_input_buffer (pfile)
|
|
cpp_reader *pfile;
|
|
{
|
|
cpp_context *base;
|
|
|
|
init_trigraph_map ();
|
|
_cpp_init_toklist (&pfile->token_list, DUMMY_TOKEN);
|
|
pfile->no_expand_level = UINT_MAX;
|
|
pfile->context_cap = 20;
|
|
pfile->cur_context = 0;
|
|
|
|
pfile->contexts = (cpp_context *)
|
|
xmalloc (pfile->context_cap * sizeof (cpp_context));
|
|
|
|
/* Clear the base context. */
|
|
base = &pfile->contexts[0];
|
|
base->u.list = &pfile->token_list;
|
|
base->posn = 0;
|
|
base->count = 0;
|
|
base->args = 0;
|
|
base->level = 0;
|
|
base->flags = 0;
|
|
base->pushed_token = 0;
|
|
}
|
|
|
|
/* Moves to the end of the directive line, popping contexts as
|
|
necessary. */
|
|
void
|
|
_cpp_skip_rest_of_line (pfile)
|
|
cpp_reader *pfile;
|
|
{
|
|
/* Discard all stacked contexts. */
|
|
int i;
|
|
for (i = pfile->cur_context; i > 0; i--)
|
|
if (pfile->contexts[i].args)
|
|
free_macro_args (pfile->contexts[i].args);
|
|
|
|
if (pfile->no_expand_level <= pfile->cur_context)
|
|
pfile->no_expand_level = 0;
|
|
pfile->cur_context = 0;
|
|
|
|
/* Clear the base context, and clear the directive pointer so that
|
|
get_raw_token will advance to the next line. */
|
|
pfile->contexts[0].count = 0;
|
|
pfile->contexts[0].posn = 0;
|
|
pfile->token_list.directive = 0;
|
|
}
|
|
|
|
/* Directive handler wrapper used by the command line option
|
|
processor. */
|
|
void
|
|
_cpp_run_directive (pfile, dir, buf, count)
|
|
cpp_reader *pfile;
|
|
const struct directive *dir;
|
|
const char *buf;
|
|
size_t count;
|
|
{
|
|
if (cpp_push_buffer (pfile, (const U_CHAR *)buf, count) != NULL)
|
|
{
|
|
unsigned int prev_lvl = 0;
|
|
|
|
/* Scan the line now, else prevent_macro_expansion won't work. */
|
|
lex_next (pfile, 1);
|
|
if (! (dir->flags & EXPAND))
|
|
prev_lvl = prevent_macro_expansion (pfile);
|
|
|
|
(void) (*dir->handler) (pfile);
|
|
|
|
if (! (dir->flags & EXPAND))
|
|
restore_macro_expansion (pfile, prev_lvl);
|
|
|
|
_cpp_skip_rest_of_line (pfile);
|
|
cpp_pop_buffer (pfile);
|
|
}
|
|
}
|