optimize.c (struct inline_data): Moved to ../tree-inline.c.
* optimize.c (struct inline_data): Moved to ../tree-inline.c. (INSNS_PER_STMT): Likewise. (remap_decl, remap_block, copy_scopy_stmt, copy_body_r): Likewise. (copy_body, initialize_inlined_parameters): Likewise. (declare_return_variable, inlinable_function_p): Likewise. (expand_call_inline, expand_calls_inline): Likewise. (optimize_inline_calls, clone_body): Likewise. * tree.c (walk_tree): Moved to ../tree-inline.c. (walk_tree_without_duplicates): Likewise. (copy_tree_r, remap_save_expr): Likewise. From-SVN: r46022
This commit is contained in:
parent
588d3ade13
commit
bc4c7159d8
@ -1,3 +1,16 @@
|
||||
2001-10-05 Alexandre Oliva <aoliva@redhat.com>
|
||||
|
||||
* optimize.c (struct inline_data): Moved to ../tree-inline.c.
|
||||
(INSNS_PER_STMT): Likewise.
|
||||
(remap_decl, remap_block, copy_scopy_stmt, copy_body_r): Likewise.
|
||||
(copy_body, initialize_inlined_parameters): Likewise.
|
||||
(declare_return_variable, inlinable_function_p): Likewise.
|
||||
(expand_call_inline, expand_calls_inline): Likewise.
|
||||
(optimize_inline_calls, clone_body): Likewise.
|
||||
* tree.c (walk_tree): Moved to ../tree-inline.c.
|
||||
(walk_tree_without_duplicates): Likewise.
|
||||
(copy_tree_r, remap_save_expr): Likewise.
|
||||
|
||||
2001-10-04 Alexandre Oliva <aoliva@redhat.com>
|
||||
|
||||
* Make-lang.in (cp/decl.o, cp/tree.o): Depend on tree-inline.h.
|
||||
|
@ -35,949 +35,12 @@ Software Foundation, 59 Temple Place - Suite 330, Boston, MA
|
||||
#include "debug.h"
|
||||
#include "tree-inline.h"
|
||||
|
||||
/* To Do:
|
||||
|
||||
o In order to make inlining-on-trees work, we pessimized
|
||||
function-local static constants. In particular, they are now
|
||||
always output, even when not addressed. Fix this by treating
|
||||
function-local static constants just like global static
|
||||
constants; the back-end already knows not to output them if they
|
||||
are not needed.
|
||||
|
||||
o Provide heuristics to clamp inlining of recursive template
|
||||
calls? */
|
||||
|
||||
/* Data required for function inlining. */
|
||||
|
||||
typedef struct inline_data
|
||||
{
|
||||
/* A stack of the functions we are inlining. For example, if we are
|
||||
compiling `f', which calls `g', which calls `h', and we are
|
||||
inlining the body of `h', the stack will contain, `h', followed
|
||||
by `g', followed by `f'. The first few elements of the stack may
|
||||
contain other functions that we know we should not recurse into,
|
||||
even though they are not directly being inlined. */
|
||||
varray_type fns;
|
||||
/* The index of the first element of FNS that really represents an
|
||||
inlined function. */
|
||||
unsigned first_inlined_fn;
|
||||
/* The label to jump to when a return statement is encountered. If
|
||||
this value is NULL, then return statements will simply be
|
||||
remapped as return statements, rather than as jumps. */
|
||||
tree ret_label;
|
||||
/* The map from local declarations in the inlined function to
|
||||
equivalents in the function into which it is being inlined. */
|
||||
splay_tree decl_map;
|
||||
/* Nonzero if we are currently within the cleanup for a
|
||||
TARGET_EXPR. */
|
||||
int in_target_cleanup_p;
|
||||
/* A stack of the TARGET_EXPRs that we are currently processing. */
|
||||
varray_type target_exprs;
|
||||
/* A list of the functions current function has inlined. */
|
||||
varray_type inlined_fns;
|
||||
/* The approximate number of statements we have inlined in the
|
||||
current call stack. */
|
||||
int inlined_stmts;
|
||||
/* We use the same mechanism to build clones that we do to perform
|
||||
inlining. However, there are a few places where we need to
|
||||
distinguish between those two situations. This flag is true if
|
||||
we are cloning, rather than inlining. */
|
||||
bool cloning_p;
|
||||
/* Hash table used to prevent walk_tree from visiting the same node
|
||||
umpteen million times. */
|
||||
htab_t tree_pruner;
|
||||
} inline_data;
|
||||
|
||||
/* Prototypes. */
|
||||
|
||||
static tree initialize_inlined_parameters PARAMS ((inline_data *, tree, tree));
|
||||
static tree declare_return_variable PARAMS ((inline_data *, tree *));
|
||||
static tree copy_body_r PARAMS ((tree *, int *, void *));
|
||||
static tree copy_body PARAMS ((inline_data *));
|
||||
static tree expand_call_inline PARAMS ((tree *, int *, void *));
|
||||
static void expand_calls_inline PARAMS ((tree *, inline_data *));
|
||||
static int inlinable_function_p PARAMS ((tree, inline_data *));
|
||||
static tree remap_decl PARAMS ((tree, inline_data *));
|
||||
static void remap_block PARAMS ((tree, tree, inline_data *));
|
||||
static void copy_scope_stmt PARAMS ((tree *, int *, inline_data *));
|
||||
static tree calls_setjmp_r PARAMS ((tree *, int *, void *));
|
||||
static void update_cloned_parm PARAMS ((tree, tree));
|
||||
static void dump_function PARAMS ((enum tree_dump_index, tree));
|
||||
|
||||
/* The approximate number of instructions per statement. This number
|
||||
need not be particularly accurate; it is used only to make
|
||||
decisions about when a function is too big to inline. */
|
||||
#define INSNS_PER_STMT (10)
|
||||
|
||||
/* Remap DECL during the copying of the BLOCK tree for the function. */
|
||||
|
||||
static tree
|
||||
remap_decl (decl, id)
|
||||
tree decl;
|
||||
inline_data *id;
|
||||
{
|
||||
splay_tree_node n;
|
||||
tree fn;
|
||||
|
||||
/* We only remap local variables in the current function. */
|
||||
fn = VARRAY_TOP_TREE (id->fns);
|
||||
if (! LANG_AUTO_VAR_IN_FN_P (decl, fn))
|
||||
return NULL_TREE;
|
||||
|
||||
/* See if we have remapped this declaration. */
|
||||
n = splay_tree_lookup (id->decl_map, (splay_tree_key) decl);
|
||||
/* If we didn't already have an equivalent for this declaration,
|
||||
create one now. */
|
||||
if (!n)
|
||||
{
|
||||
tree t;
|
||||
|
||||
/* Make a copy of the variable or label. */
|
||||
t = copy_decl_for_inlining (decl, fn,
|
||||
VARRAY_TREE (id->fns, 0));
|
||||
|
||||
/* The decl T could be a dynamic array or other variable size type,
|
||||
in which case some fields need to be remapped because they may
|
||||
contain SAVE_EXPRs. */
|
||||
walk_tree (&DECL_SIZE (t), copy_body_r, id, NULL);
|
||||
walk_tree (&DECL_SIZE_UNIT (t), copy_body_r, id, NULL);
|
||||
if (TREE_TYPE (t) && TREE_CODE (TREE_TYPE (t)) == ARRAY_TYPE
|
||||
&& TYPE_DOMAIN (TREE_TYPE (t)))
|
||||
{
|
||||
TREE_TYPE (t) = copy_node (TREE_TYPE (t));
|
||||
TYPE_DOMAIN (TREE_TYPE (t))
|
||||
= copy_node (TYPE_DOMAIN (TREE_TYPE (t)));
|
||||
walk_tree (&TYPE_MAX_VALUE (TYPE_DOMAIN (TREE_TYPE (t))),
|
||||
copy_body_r, id, NULL);
|
||||
}
|
||||
|
||||
if (! DECL_NAME (t) && TREE_TYPE (t)
|
||||
&& LANG_ANON_AGGR_TYPE_P (TREE_TYPE (t)))
|
||||
{
|
||||
/* For a VAR_DECL of anonymous type, we must also copy the
|
||||
member VAR_DECLS here and rechain the
|
||||
DECL_ANON_UNION_ELEMS. */
|
||||
tree members = NULL;
|
||||
tree src;
|
||||
|
||||
for (src = DECL_ANON_UNION_ELEMS (t); src;
|
||||
src = TREE_CHAIN (src))
|
||||
{
|
||||
tree member = remap_decl (TREE_VALUE (src), id);
|
||||
|
||||
if (TREE_PURPOSE (src))
|
||||
abort ();
|
||||
members = tree_cons (NULL, member, members);
|
||||
}
|
||||
DECL_ANON_UNION_ELEMS (t) = nreverse (members);
|
||||
}
|
||||
|
||||
/* Remember it, so that if we encounter this local entity
|
||||
again we can reuse this copy. */
|
||||
n = splay_tree_insert (id->decl_map,
|
||||
(splay_tree_key) decl,
|
||||
(splay_tree_value) t);
|
||||
}
|
||||
|
||||
return (tree) n->value;
|
||||
}
|
||||
|
||||
/* Copy the SCOPE_STMT_BLOCK associated with SCOPE_STMT to contain
|
||||
remapped versions of the variables therein. And hook the new block
|
||||
into the block-tree. If non-NULL, the DECLS are declarations to
|
||||
add to use instead of the BLOCK_VARS in the old block. */
|
||||
|
||||
static void
|
||||
remap_block (scope_stmt, decls, id)
|
||||
tree scope_stmt;
|
||||
tree decls;
|
||||
inline_data *id;
|
||||
{
|
||||
/* We cannot do this in the cleanup for a TARGET_EXPR since we do
|
||||
not know whether or not expand_expr will actually write out the
|
||||
code we put there. If it does not, then we'll have more BLOCKs
|
||||
than block-notes, and things will go awry. At some point, we
|
||||
should make the back-end handle BLOCK notes in a tidier way,
|
||||
without requiring a strict correspondence to the block-tree; then
|
||||
this check can go. */
|
||||
if (id->in_target_cleanup_p)
|
||||
{
|
||||
SCOPE_STMT_BLOCK (scope_stmt) = NULL_TREE;
|
||||
return;
|
||||
}
|
||||
|
||||
/* If this is the beginning of a scope, remap the associated BLOCK. */
|
||||
if (SCOPE_BEGIN_P (scope_stmt) && SCOPE_STMT_BLOCK (scope_stmt))
|
||||
{
|
||||
tree old_block;
|
||||
tree new_block;
|
||||
tree old_var;
|
||||
tree fn;
|
||||
|
||||
/* Make the new block. */
|
||||
old_block = SCOPE_STMT_BLOCK (scope_stmt);
|
||||
new_block = make_node (BLOCK);
|
||||
TREE_USED (new_block) = TREE_USED (old_block);
|
||||
BLOCK_ABSTRACT_ORIGIN (new_block) = old_block;
|
||||
SCOPE_STMT_BLOCK (scope_stmt) = new_block;
|
||||
|
||||
/* Remap its variables. */
|
||||
for (old_var = decls ? decls : BLOCK_VARS (old_block);
|
||||
old_var;
|
||||
old_var = TREE_CHAIN (old_var))
|
||||
{
|
||||
tree new_var;
|
||||
|
||||
/* Remap the variable. */
|
||||
new_var = remap_decl (old_var, id);
|
||||
/* If we didn't remap this variable, so we can't mess with
|
||||
its TREE_CHAIN. If we remapped this variable to
|
||||
something other than a declaration (say, if we mapped it
|
||||
to a constant), then we must similarly omit any mention
|
||||
of it here. */
|
||||
if (!new_var || !DECL_P (new_var))
|
||||
;
|
||||
else
|
||||
{
|
||||
TREE_CHAIN (new_var) = BLOCK_VARS (new_block);
|
||||
BLOCK_VARS (new_block) = new_var;
|
||||
}
|
||||
}
|
||||
/* We put the BLOCK_VARS in reverse order; fix that now. */
|
||||
BLOCK_VARS (new_block) = nreverse (BLOCK_VARS (new_block));
|
||||
fn = VARRAY_TREE (id->fns, 0);
|
||||
if (id->cloning_p)
|
||||
/* We're building a clone; DECL_INITIAL is still
|
||||
error_mark_node, and current_binding_level is the parm
|
||||
binding level. */
|
||||
insert_block (new_block);
|
||||
else
|
||||
{
|
||||
/* Attach this new block after the DECL_INITIAL block for the
|
||||
function into which this block is being inlined. In
|
||||
rest_of_compilation we will straighten out the BLOCK tree. */
|
||||
tree *first_block;
|
||||
if (DECL_INITIAL (fn))
|
||||
first_block = &BLOCK_CHAIN (DECL_INITIAL (fn));
|
||||
else
|
||||
first_block = &DECL_INITIAL (fn);
|
||||
BLOCK_CHAIN (new_block) = *first_block;
|
||||
*first_block = new_block;
|
||||
}
|
||||
/* Remember the remapped block. */
|
||||
splay_tree_insert (id->decl_map,
|
||||
(splay_tree_key) old_block,
|
||||
(splay_tree_value) new_block);
|
||||
}
|
||||
/* If this is the end of a scope, set the SCOPE_STMT_BLOCK to be the
|
||||
remapped block. */
|
||||
else if (SCOPE_END_P (scope_stmt) && SCOPE_STMT_BLOCK (scope_stmt))
|
||||
{
|
||||
splay_tree_node n;
|
||||
|
||||
/* Find this block in the table of remapped things. */
|
||||
n = splay_tree_lookup (id->decl_map,
|
||||
(splay_tree_key) SCOPE_STMT_BLOCK (scope_stmt));
|
||||
if (! n)
|
||||
abort ();
|
||||
SCOPE_STMT_BLOCK (scope_stmt) = (tree) n->value;
|
||||
}
|
||||
}
|
||||
|
||||
/* Copy the SCOPE_STMT pointed to by TP. */
|
||||
|
||||
static void
|
||||
copy_scope_stmt (tp, walk_subtrees, id)
|
||||
tree *tp;
|
||||
int *walk_subtrees;
|
||||
inline_data *id;
|
||||
{
|
||||
tree block;
|
||||
|
||||
/* Remember whether or not this statement was nullified. When
|
||||
making a copy, copy_tree_r always sets SCOPE_NULLIFIED_P (and
|
||||
doesn't copy the SCOPE_STMT_BLOCK) to free callers from having to
|
||||
deal with copying BLOCKs if they do not wish to do so. */
|
||||
block = SCOPE_STMT_BLOCK (*tp);
|
||||
/* Copy (and replace) the statement. */
|
||||
copy_tree_r (tp, walk_subtrees, NULL);
|
||||
/* Restore the SCOPE_STMT_BLOCK. */
|
||||
SCOPE_STMT_BLOCK (*tp) = block;
|
||||
|
||||
/* Remap the associated block. */
|
||||
remap_block (*tp, NULL_TREE, id);
|
||||
}
|
||||
|
||||
/* Called from copy_body via walk_tree. DATA is really an
|
||||
`inline_data *'. */
|
||||
|
||||
static tree
|
||||
copy_body_r (tp, walk_subtrees, data)
|
||||
tree *tp;
|
||||
int *walk_subtrees;
|
||||
void *data;
|
||||
{
|
||||
inline_data* id;
|
||||
tree fn;
|
||||
|
||||
/* Set up. */
|
||||
id = (inline_data *) data;
|
||||
fn = VARRAY_TOP_TREE (id->fns);
|
||||
|
||||
#if 0
|
||||
/* All automatic variables should have a DECL_CONTEXT indicating
|
||||
what function they come from. */
|
||||
if ((TREE_CODE (*tp) == VAR_DECL || TREE_CODE (*tp) == LABEL_DECL)
|
||||
&& DECL_NAMESPACE_SCOPE_P (*tp))
|
||||
if (! DECL_EXTERNAL (*tp) && ! TREE_STATIC (*tp))
|
||||
abort ();
|
||||
#endif
|
||||
|
||||
/* If this is a RETURN_STMT, change it into an EXPR_STMT and a
|
||||
GOTO_STMT with the RET_LABEL as its target. */
|
||||
if (TREE_CODE (*tp) == RETURN_STMT && id->ret_label)
|
||||
{
|
||||
tree return_stmt = *tp;
|
||||
tree goto_stmt;
|
||||
|
||||
/* Build the GOTO_STMT. */
|
||||
goto_stmt = build_stmt (GOTO_STMT, id->ret_label);
|
||||
TREE_CHAIN (goto_stmt) = TREE_CHAIN (return_stmt);
|
||||
|
||||
/* If we're returning something, just turn that into an
|
||||
assignment into the equivalent of the original
|
||||
RESULT_DECL. */
|
||||
if (RETURN_EXPR (return_stmt))
|
||||
{
|
||||
*tp = build_stmt (EXPR_STMT,
|
||||
RETURN_EXPR (return_stmt));
|
||||
STMT_IS_FULL_EXPR_P (*tp) = 1;
|
||||
/* And then jump to the end of the function. */
|
||||
TREE_CHAIN (*tp) = goto_stmt;
|
||||
}
|
||||
/* If we're not returning anything just do the jump. */
|
||||
else
|
||||
*tp = goto_stmt;
|
||||
}
|
||||
/* Local variables and labels need to be replaced by equivalent
|
||||
variables. We don't want to copy static variables; there's only
|
||||
one of those, no matter how many times we inline the containing
|
||||
function. */
|
||||
else if (LANG_AUTO_VAR_IN_FN_P (*tp, fn))
|
||||
{
|
||||
tree new_decl;
|
||||
|
||||
/* Remap the declaration. */
|
||||
new_decl = remap_decl (*tp, id);
|
||||
if (! new_decl)
|
||||
abort ();
|
||||
/* Replace this variable with the copy. */
|
||||
STRIP_TYPE_NOPS (new_decl);
|
||||
*tp = new_decl;
|
||||
}
|
||||
#if 0
|
||||
else if (nonstatic_local_decl_p (*tp)
|
||||
&& DECL_CONTEXT (*tp) != VARRAY_TREE (id->fns, 0))
|
||||
abort ();
|
||||
#endif
|
||||
else if (TREE_CODE (*tp) == SAVE_EXPR)
|
||||
remap_save_expr (tp, id->decl_map, VARRAY_TREE (id->fns, 0),
|
||||
walk_subtrees);
|
||||
else if (TREE_CODE (*tp) == UNSAVE_EXPR)
|
||||
/* UNSAVE_EXPRs should not be generated until expansion time. */
|
||||
abort ();
|
||||
/* For a SCOPE_STMT, we must copy the associated block so that we
|
||||
can write out debugging information for the inlined variables. */
|
||||
else if (TREE_CODE (*tp) == SCOPE_STMT && !id->in_target_cleanup_p)
|
||||
copy_scope_stmt (tp, walk_subtrees, id);
|
||||
/* Otherwise, just copy the node. Note that copy_tree_r already
|
||||
knows not to copy VAR_DECLs, etc., so this is safe. */
|
||||
else
|
||||
{
|
||||
copy_tree_r (tp, walk_subtrees, NULL);
|
||||
|
||||
/* The copied TARGET_EXPR has never been expanded, even if the
|
||||
original node was expanded already. */
|
||||
if (TREE_CODE (*tp) == TARGET_EXPR && TREE_OPERAND (*tp, 3))
|
||||
{
|
||||
TREE_OPERAND (*tp, 1) = TREE_OPERAND (*tp, 3);
|
||||
TREE_OPERAND (*tp, 3) = NULL_TREE;
|
||||
}
|
||||
else if (TREE_CODE (*tp) == MODIFY_EXPR
|
||||
&& TREE_OPERAND (*tp, 0) == TREE_OPERAND (*tp, 1)
|
||||
&& LANG_AUTO_VAR_IN_FN_P (TREE_OPERAND (*tp, 0), fn))
|
||||
{
|
||||
/* Some assignments VAR = VAR; don't generate any rtl code
|
||||
and thus don't count as variable modification. Avoid
|
||||
keeping bogosities like 0 = 0. */
|
||||
tree decl = TREE_OPERAND (*tp, 0), value;
|
||||
splay_tree_node n;
|
||||
|
||||
n = splay_tree_lookup (id->decl_map, (splay_tree_key) decl);
|
||||
if (n)
|
||||
{
|
||||
value = (tree) n->value;
|
||||
STRIP_TYPE_NOPS (value);
|
||||
if (TREE_CONSTANT (value) || TREE_READONLY_DECL_P (value))
|
||||
*tp = value;
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
/* Keep iterating. */
|
||||
return NULL_TREE;
|
||||
}
|
||||
|
||||
/* Make a copy of the body of FN so that it can be inserted inline in
|
||||
another function. */
|
||||
|
||||
static tree
|
||||
copy_body (id)
|
||||
inline_data *id;
|
||||
{
|
||||
tree body;
|
||||
|
||||
body = DECL_SAVED_TREE (VARRAY_TOP_TREE (id->fns));
|
||||
walk_tree (&body, copy_body_r, id, NULL);
|
||||
|
||||
return body;
|
||||
}
|
||||
|
||||
/* Generate code to initialize the parameters of the function at the
|
||||
top of the stack in ID from the ARGS (presented as a TREE_LIST). */
|
||||
|
||||
static tree
|
||||
initialize_inlined_parameters (id, args, fn)
|
||||
inline_data *id;
|
||||
tree args;
|
||||
tree fn;
|
||||
{
|
||||
tree init_stmts;
|
||||
tree parms;
|
||||
tree a;
|
||||
tree p;
|
||||
|
||||
/* Figure out what the parameters are. */
|
||||
parms = DECL_ARGUMENTS (fn);
|
||||
|
||||
/* Start with no initializations whatsoever. */
|
||||
init_stmts = NULL_TREE;
|
||||
|
||||
/* Loop through the parameter declarations, replacing each with an
|
||||
equivalent VAR_DECL, appropriately initialized. */
|
||||
for (p = parms, a = args; p; a = TREE_CHAIN (a), p = TREE_CHAIN (p))
|
||||
{
|
||||
tree init_stmt;
|
||||
tree var;
|
||||
tree value;
|
||||
|
||||
/* Find the initializer. */
|
||||
value = TREE_VALUE (a);
|
||||
/* If the parameter is never assigned to, we may not need to
|
||||
create a new variable here at all. Instead, we may be able
|
||||
to just use the argument value. */
|
||||
if (TREE_READONLY (p)
|
||||
&& !TREE_ADDRESSABLE (p)
|
||||
&& !TREE_SIDE_EFFECTS (value))
|
||||
{
|
||||
/* Simplify the value, if possible. */
|
||||
value = fold (decl_constant_value (value));
|
||||
|
||||
/* We can't risk substituting complex expressions. They
|
||||
might contain variables that will be assigned to later.
|
||||
Theoretically, we could check the expression to see if
|
||||
all of the variables that determine its value are
|
||||
read-only, but we don't bother. */
|
||||
if (TREE_CONSTANT (value) || TREE_READONLY_DECL_P (value))
|
||||
{
|
||||
/* If this is a declaration, wrap it a NOP_EXPR so that
|
||||
we don't try to put the VALUE on the list of
|
||||
BLOCK_VARS. */
|
||||
if (DECL_P (value))
|
||||
value = build1 (NOP_EXPR, TREE_TYPE (value), value);
|
||||
|
||||
splay_tree_insert (id->decl_map,
|
||||
(splay_tree_key) p,
|
||||
(splay_tree_value) value);
|
||||
continue;
|
||||
}
|
||||
}
|
||||
|
||||
/* Make an equivalent VAR_DECL. */
|
||||
var = copy_decl_for_inlining (p, fn, VARRAY_TREE (id->fns, 0));
|
||||
/* Register the VAR_DECL as the equivalent for the PARM_DECL;
|
||||
that way, when the PARM_DECL is encountered, it will be
|
||||
automatically replaced by the VAR_DECL. */
|
||||
splay_tree_insert (id->decl_map,
|
||||
(splay_tree_key) p,
|
||||
(splay_tree_value) var);
|
||||
|
||||
/* Declare this new variable. */
|
||||
init_stmt = build_stmt (DECL_STMT, var);
|
||||
TREE_CHAIN (init_stmt) = init_stmts;
|
||||
init_stmts = init_stmt;
|
||||
|
||||
/* Initialize this VAR_DECL from the equivalent argument. If
|
||||
the argument is an object, created via a constructor or copy,
|
||||
this will not result in an extra copy: the TARGET_EXPR
|
||||
representing the argument will be bound to VAR, and the
|
||||
object will be constructed in VAR. */
|
||||
if (! TYPE_NEEDS_CONSTRUCTING (TREE_TYPE (p)))
|
||||
DECL_INITIAL (var) = value;
|
||||
else
|
||||
{
|
||||
/* Even if P was TREE_READONLY, the new VAR should not be.
|
||||
In the original code, we would have constructed a
|
||||
temporary, and then the function body would have never
|
||||
changed the value of P. However, now, we will be
|
||||
constructing VAR directly. The constructor body may
|
||||
change its value multiple times as it is being
|
||||
constructed. Therefore, it must not be TREE_READONLY;
|
||||
the back-end assumes that TREE_READONLY variable is
|
||||
assigned to only once. */
|
||||
TREE_READONLY (var) = 0;
|
||||
|
||||
/* Build a run-time initialization. */
|
||||
init_stmt = build_stmt (EXPR_STMT,
|
||||
build (INIT_EXPR, TREE_TYPE (p),
|
||||
var, value));
|
||||
/* Add this initialization to the list. Note that we want the
|
||||
declaration *after* the initialization because we are going
|
||||
to reverse all the initialization statements below. */
|
||||
TREE_CHAIN (init_stmt) = init_stmts;
|
||||
init_stmts = init_stmt;
|
||||
}
|
||||
}
|
||||
|
||||
/* The initialization statements have been built up in reverse
|
||||
order. Straighten them out now. */
|
||||
return nreverse (init_stmts);
|
||||
}
|
||||
|
||||
/* Declare a return variable to replace the RESULT_DECL for the
|
||||
function we are calling. An appropriate DECL_STMT is returned.
|
||||
The USE_STMT is filled in to contain a use of the declaration to
|
||||
indicate the return value of the function. */
|
||||
|
||||
static tree
|
||||
declare_return_variable (id, use_stmt)
|
||||
struct inline_data *id;
|
||||
tree *use_stmt;
|
||||
{
|
||||
tree fn = VARRAY_TOP_TREE (id->fns);
|
||||
tree result = DECL_RESULT (fn);
|
||||
tree var;
|
||||
int need_return_decl = 1;
|
||||
|
||||
/* We don't need to do anything for functions that don't return
|
||||
anything. */
|
||||
if (!result || VOID_TYPE_P (TREE_TYPE (result)))
|
||||
{
|
||||
*use_stmt = NULL_TREE;
|
||||
return NULL_TREE;
|
||||
}
|
||||
|
||||
var = LANG_COPY_RES_DECL_FOR_INLINING (result, fn, VARRAY_TREE (id->fns, 0),
|
||||
id->decl_map, &need_return_decl,
|
||||
&id->target_exprs);
|
||||
|
||||
/* Register the VAR_DECL as the equivalent for the RESULT_DECL; that
|
||||
way, when the RESULT_DECL is encountered, it will be
|
||||
automatically replaced by the VAR_DECL. */
|
||||
splay_tree_insert (id->decl_map,
|
||||
(splay_tree_key) result,
|
||||
(splay_tree_value) var);
|
||||
|
||||
/* Build the USE_STMT. */
|
||||
*use_stmt = build_stmt (EXPR_STMT, var);
|
||||
|
||||
/* Build the declaration statement if FN does not return an
|
||||
aggregate. */
|
||||
if (need_return_decl)
|
||||
return build_stmt (DECL_STMT, var);
|
||||
/* If FN does return an aggregate, there's no need to declare the
|
||||
return variable; we're using a variable in our caller's frame. */
|
||||
else
|
||||
return NULL_TREE;
|
||||
}
|
||||
|
||||
/* Returns non-zero if FN is a function that can be inlined. */
|
||||
|
||||
static int
|
||||
inlinable_function_p (fn, id)
|
||||
tree fn;
|
||||
inline_data *id;
|
||||
{
|
||||
int inlinable;
|
||||
|
||||
/* If we've already decided this function shouldn't be inlined,
|
||||
there's no need to check again. */
|
||||
if (DECL_UNINLINABLE (fn))
|
||||
return 0;
|
||||
|
||||
/* Assume it is not inlinable. */
|
||||
inlinable = 0;
|
||||
|
||||
/* If we're not inlining things, then nothing is inlinable. */
|
||||
if (!flag_inline_trees)
|
||||
;
|
||||
/* If the function was not declared `inline', then we don't inline
|
||||
it. */
|
||||
else if (!DECL_INLINE (fn))
|
||||
;
|
||||
/* We can't inline functions that are too big. Only allow a single
|
||||
function to eat up half of our budget. Make special allowance
|
||||
for extern inline functions, though. */
|
||||
else if (! LANG_DISREGARD_INLINE_LIMITS (fn)
|
||||
&& DECL_NUM_STMTS (fn) * INSNS_PER_STMT > MAX_INLINE_INSNS / 2)
|
||||
;
|
||||
/* All is well. We can inline this function. Traditionally, GCC
|
||||
has refused to inline functions using alloca, or functions whose
|
||||
values are returned in a PARALLEL, and a few other such obscure
|
||||
conditions. We are not equally constrained at the tree level. */
|
||||
else
|
||||
inlinable = 1;
|
||||
|
||||
/* Squirrel away the result so that we don't have to check again. */
|
||||
DECL_UNINLINABLE (fn) = !inlinable;
|
||||
|
||||
/* Even if this function is not itself too big to inline, it might
|
||||
be that we've done so much inlining already that we don't want to
|
||||
risk too much inlining any more and thus halve the acceptable
|
||||
size. */
|
||||
if (! LANG_DISREGARD_INLINE_LIMITS (fn)
|
||||
&& ((DECL_NUM_STMTS (fn) + id->inlined_stmts) * INSNS_PER_STMT
|
||||
> MAX_INLINE_INSNS)
|
||||
&& DECL_NUM_STMTS (fn) * INSNS_PER_STMT > MAX_INLINE_INSNS / 4)
|
||||
inlinable = 0;
|
||||
|
||||
if (inlinable && LANG_CANNOT_INLINE_TREE_FN (&fn))
|
||||
inlinable = 0;
|
||||
|
||||
/* If we don't have the function body available, we can't inline
|
||||
it. */
|
||||
if (!DECL_SAVED_TREE (fn))
|
||||
inlinable = 0;
|
||||
|
||||
/* Check again, language hooks may have modified it. */
|
||||
if (! inlinable || DECL_UNINLINABLE (fn))
|
||||
return 0;
|
||||
|
||||
/* Don't do recursive inlining, either. We don't record this in
|
||||
DECL_UNINLINABLE; we may be able to inline this function later. */
|
||||
if (inlinable)
|
||||
{
|
||||
size_t i;
|
||||
|
||||
for (i = 0; i < VARRAY_ACTIVE_SIZE (id->fns); ++i)
|
||||
if (VARRAY_TREE (id->fns, i) == fn)
|
||||
return 0;
|
||||
|
||||
if (inlinable && DECL_INLINED_FNS (fn))
|
||||
{
|
||||
int j;
|
||||
tree inlined_fns = DECL_INLINED_FNS (fn);
|
||||
|
||||
for (j = 0; j < TREE_VEC_LENGTH (inlined_fns); ++j)
|
||||
if (TREE_VEC_ELT (inlined_fns, j) == VARRAY_TREE (id->fns, 0))
|
||||
return 0;
|
||||
}
|
||||
}
|
||||
|
||||
/* Return the result. */
|
||||
return inlinable;
|
||||
}
|
||||
|
||||
/* If *TP is a CALL_EXPR, replace it with its inline expansion. */
|
||||
|
||||
static tree
|
||||
expand_call_inline (tp, walk_subtrees, data)
|
||||
tree *tp;
|
||||
int *walk_subtrees;
|
||||
void *data;
|
||||
{
|
||||
inline_data *id;
|
||||
tree t;
|
||||
tree expr;
|
||||
tree chain;
|
||||
tree fn;
|
||||
tree scope_stmt;
|
||||
tree use_stmt;
|
||||
tree arg_inits;
|
||||
tree *inlined_body;
|
||||
splay_tree st;
|
||||
|
||||
/* See what we've got. */
|
||||
id = (inline_data *) data;
|
||||
t = *tp;
|
||||
|
||||
/* Recurse, but letting recursive invocations know that we are
|
||||
inside the body of a TARGET_EXPR. */
|
||||
if (TREE_CODE (*tp) == TARGET_EXPR)
|
||||
{
|
||||
int i, len = first_rtl_op (TARGET_EXPR);
|
||||
|
||||
/* We're walking our own subtrees. */
|
||||
*walk_subtrees = 0;
|
||||
|
||||
/* Push *TP on the stack of pending TARGET_EXPRs. */
|
||||
VARRAY_PUSH_TREE (id->target_exprs, *tp);
|
||||
|
||||
/* Actually walk over them. This loop is the body of
|
||||
walk_trees, omitting the case where the TARGET_EXPR
|
||||
itself is handled. */
|
||||
for (i = 0; i < len; ++i)
|
||||
{
|
||||
if (i == 2)
|
||||
++id->in_target_cleanup_p;
|
||||
walk_tree (&TREE_OPERAND (*tp, i), expand_call_inline, data,
|
||||
id->tree_pruner);
|
||||
if (i == 2)
|
||||
--id->in_target_cleanup_p;
|
||||
}
|
||||
|
||||
/* We're done with this TARGET_EXPR now. */
|
||||
VARRAY_POP (id->target_exprs);
|
||||
|
||||
return NULL_TREE;
|
||||
}
|
||||
|
||||
if (TYPE_P (t))
|
||||
/* Because types were not copied in copy_body, CALL_EXPRs beneath
|
||||
them should not be expanded. This can happen if the type is a
|
||||
dynamic array type, for example. */
|
||||
*walk_subtrees = 0;
|
||||
|
||||
/* From here on, we're only interested in CALL_EXPRs. */
|
||||
if (TREE_CODE (t) != CALL_EXPR)
|
||||
return NULL_TREE;
|
||||
|
||||
/* First, see if we can figure out what function is being called.
|
||||
If we cannot, then there is no hope of inlining the function. */
|
||||
fn = get_callee_fndecl (t);
|
||||
if (!fn)
|
||||
return NULL_TREE;
|
||||
|
||||
/* Don't try to inline functions that are not well-suited to
|
||||
inlining. */
|
||||
if (!inlinable_function_p (fn, id))
|
||||
return NULL_TREE;
|
||||
|
||||
/* Set the current filename and line number to the function we are
|
||||
inlining so that when we create new _STMT nodes here they get
|
||||
line numbers corresponding to the function we are calling. We
|
||||
wrap the whole inlined body in an EXPR_WITH_FILE_AND_LINE as well
|
||||
because individual statements don't record the filename. */
|
||||
push_srcloc (fn->decl.filename, fn->decl.linenum);
|
||||
|
||||
/* Build a statement-expression containing code to initialize the
|
||||
arguments, the actual inline expansion of the body, and a label
|
||||
for the return statements within the function to jump to. The
|
||||
type of the statement expression is the return type of the
|
||||
function call. */
|
||||
expr = build1 (STMT_EXPR, TREE_TYPE (TREE_TYPE (fn)), NULL_TREE);
|
||||
|
||||
/* Local declarations will be replaced by their equivalents in this
|
||||
map. */
|
||||
st = id->decl_map;
|
||||
id->decl_map = splay_tree_new (splay_tree_compare_pointers,
|
||||
NULL, NULL);
|
||||
|
||||
/* Initialize the parameters. */
|
||||
arg_inits = initialize_inlined_parameters (id, TREE_OPERAND (t, 1), fn);
|
||||
/* Expand any inlined calls in the initializers. Do this before we
|
||||
push FN on the stack of functions we are inlining; we want to
|
||||
inline calls to FN that appear in the initializers for the
|
||||
parameters. */
|
||||
expand_calls_inline (&arg_inits, id);
|
||||
/* And add them to the tree. */
|
||||
STMT_EXPR_STMT (expr) = chainon (STMT_EXPR_STMT (expr), arg_inits);
|
||||
|
||||
/* Record the function we are about to inline so that we can avoid
|
||||
recursing into it. */
|
||||
VARRAY_PUSH_TREE (id->fns, fn);
|
||||
|
||||
/* Record the function we are about to inline if optimize_function
|
||||
has not been called on it yet and we don't have it in the list. */
|
||||
if (! DECL_INLINED_FNS (fn))
|
||||
{
|
||||
int i;
|
||||
|
||||
for (i = VARRAY_ACTIVE_SIZE (id->inlined_fns) - 1; i >= 0; i--)
|
||||
if (VARRAY_TREE (id->inlined_fns, i) == fn)
|
||||
break;
|
||||
if (i < 0)
|
||||
VARRAY_PUSH_TREE (id->inlined_fns, fn);
|
||||
}
|
||||
|
||||
/* Return statements in the function body will be replaced by jumps
|
||||
to the RET_LABEL. */
|
||||
id->ret_label = build_decl (LABEL_DECL, NULL_TREE, NULL_TREE);
|
||||
DECL_CONTEXT (id->ret_label) = VARRAY_TREE (id->fns, 0);
|
||||
|
||||
/* Create a block to put the parameters in. We have to do this
|
||||
after the parameters have been remapped because remapping
|
||||
parameters is different from remapping ordinary variables. */
|
||||
scope_stmt = build_stmt (SCOPE_STMT, DECL_INITIAL (fn));
|
||||
SCOPE_BEGIN_P (scope_stmt) = 1;
|
||||
SCOPE_NO_CLEANUPS_P (scope_stmt) = 1;
|
||||
remap_block (scope_stmt, DECL_ARGUMENTS (fn), id);
|
||||
TREE_CHAIN (scope_stmt) = STMT_EXPR_STMT (expr);
|
||||
STMT_EXPR_STMT (expr) = scope_stmt;
|
||||
|
||||
/* Tell the debugging backends that this block represents the
|
||||
outermost scope of the inlined function. */
|
||||
if (SCOPE_STMT_BLOCK (scope_stmt))
|
||||
BLOCK_ABSTRACT_ORIGIN (SCOPE_STMT_BLOCK (scope_stmt)) = DECL_ORIGIN (fn);
|
||||
|
||||
/* Declare the return variable for the function. */
|
||||
STMT_EXPR_STMT (expr)
|
||||
= chainon (STMT_EXPR_STMT (expr),
|
||||
declare_return_variable (id, &use_stmt));
|
||||
|
||||
/* After we've initialized the parameters, we insert the body of the
|
||||
function itself. */
|
||||
inlined_body = &STMT_EXPR_STMT (expr);
|
||||
while (*inlined_body)
|
||||
inlined_body = &TREE_CHAIN (*inlined_body);
|
||||
*inlined_body = copy_body (id);
|
||||
|
||||
/* Close the block for the parameters. */
|
||||
scope_stmt = build_stmt (SCOPE_STMT, DECL_INITIAL (fn));
|
||||
SCOPE_NO_CLEANUPS_P (scope_stmt) = 1;
|
||||
if (! DECL_INITIAL (fn)
|
||||
|| TREE_CODE (DECL_INITIAL (fn)) != BLOCK)
|
||||
abort ();
|
||||
remap_block (scope_stmt, NULL_TREE, id);
|
||||
STMT_EXPR_STMT (expr)
|
||||
= chainon (STMT_EXPR_STMT (expr), scope_stmt);
|
||||
|
||||
/* After the body of the function comes the RET_LABEL. This must come
|
||||
before we evaluate the returned value below, because that evalulation
|
||||
may cause RTL to be generated. */
|
||||
STMT_EXPR_STMT (expr)
|
||||
= chainon (STMT_EXPR_STMT (expr),
|
||||
build_stmt (LABEL_STMT, id->ret_label));
|
||||
|
||||
/* Finally, mention the returned value so that the value of the
|
||||
statement-expression is the returned value of the function. */
|
||||
STMT_EXPR_STMT (expr) = chainon (STMT_EXPR_STMT (expr), use_stmt);
|
||||
|
||||
/* Clean up. */
|
||||
splay_tree_delete (id->decl_map);
|
||||
id->decl_map = st;
|
||||
|
||||
/* The new expression has side-effects if the old one did. */
|
||||
TREE_SIDE_EFFECTS (expr) = TREE_SIDE_EFFECTS (t);
|
||||
|
||||
/* Replace the call by the inlined body. Wrap it in an
|
||||
EXPR_WITH_FILE_LOCATION so that we'll get debugging line notes
|
||||
pointing to the right place. */
|
||||
chain = TREE_CHAIN (*tp);
|
||||
*tp = build_expr_wfl (expr, DECL_SOURCE_FILE (fn), DECL_SOURCE_LINE (fn),
|
||||
/*col=*/0);
|
||||
EXPR_WFL_EMIT_LINE_NOTE (*tp) = 1;
|
||||
TREE_CHAIN (*tp) = chain;
|
||||
pop_srcloc ();
|
||||
|
||||
/* If the value of the new expression is ignored, that's OK. We
|
||||
don't warn about this for CALL_EXPRs, so we shouldn't warn about
|
||||
the equivalent inlined version either. */
|
||||
TREE_USED (*tp) = 1;
|
||||
|
||||
/* Our function now has more statements than it did before. */
|
||||
DECL_NUM_STMTS (VARRAY_TREE (id->fns, 0)) += DECL_NUM_STMTS (fn);
|
||||
id->inlined_stmts += DECL_NUM_STMTS (fn);
|
||||
|
||||
/* Recurse into the body of the just inlined function. */
|
||||
expand_calls_inline (inlined_body, id);
|
||||
VARRAY_POP (id->fns);
|
||||
|
||||
/* If we've returned to the top level, clear out the record of how
|
||||
much inlining has been done. */
|
||||
if (VARRAY_ACTIVE_SIZE (id->fns) == id->first_inlined_fn)
|
||||
id->inlined_stmts = 0;
|
||||
|
||||
/* Don't walk into subtrees. We've already handled them above. */
|
||||
*walk_subtrees = 0;
|
||||
|
||||
/* Keep iterating. */
|
||||
return NULL_TREE;
|
||||
}
|
||||
|
||||
/* Walk over the entire tree *TP, replacing CALL_EXPRs with inline
|
||||
expansions as appropriate. */
|
||||
|
||||
static void
|
||||
expand_calls_inline (tp, id)
|
||||
tree *tp;
|
||||
inline_data *id;
|
||||
{
|
||||
/* Search through *TP, replacing all calls to inline functions by
|
||||
appropriate equivalents. Use walk_tree in no-duplicates mode
|
||||
to avoid exponential time complexity. (We can't just use
|
||||
walk_tree_without_duplicates, because of the special TARGET_EXPR
|
||||
handling in expand_calls. The hash table is set up in
|
||||
optimize_function. */
|
||||
walk_tree (tp, expand_call_inline, id, id->tree_pruner);
|
||||
}
|
||||
|
||||
/* Expand calls to inline functions in the body of FN. */
|
||||
|
||||
void
|
||||
optimize_inline_calls (fn)
|
||||
tree fn;
|
||||
{
|
||||
inline_data id;
|
||||
tree prev_fn;
|
||||
|
||||
/* Clear out ID. */
|
||||
memset (&id, 0, sizeof (id));
|
||||
|
||||
/* Don't allow recursion into FN. */
|
||||
VARRAY_TREE_INIT (id.fns, 32, "fns");
|
||||
VARRAY_PUSH_TREE (id.fns, fn);
|
||||
/* Or any functions that aren't finished yet. */
|
||||
prev_fn = NULL_TREE;
|
||||
if (current_function_decl)
|
||||
{
|
||||
VARRAY_PUSH_TREE (id.fns, current_function_decl);
|
||||
prev_fn = current_function_decl;
|
||||
}
|
||||
|
||||
prev_fn = LANG_ADD_PENDING_FN_DECLS (&id.fns, prev_fn);
|
||||
|
||||
/* Create the stack of TARGET_EXPRs. */
|
||||
VARRAY_TREE_INIT (id.target_exprs, 32, "target_exprs");
|
||||
|
||||
/* Create the list of functions this call will inline. */
|
||||
VARRAY_TREE_INIT (id.inlined_fns, 32, "inlined_fns");
|
||||
|
||||
/* Keep track of the low-water mark, i.e., the point where the first
|
||||
real inlining is represented in ID.FNS. */
|
||||
id.first_inlined_fn = VARRAY_ACTIVE_SIZE (id.fns);
|
||||
|
||||
/* Replace all calls to inline functions with the bodies of those
|
||||
functions. */
|
||||
id.tree_pruner = htab_create (37, htab_hash_pointer,
|
||||
htab_eq_pointer, NULL);
|
||||
expand_calls_inline (&DECL_SAVED_TREE (fn), &id);
|
||||
|
||||
/* Clean up. */
|
||||
htab_delete (id.tree_pruner);
|
||||
VARRAY_FREE (id.fns);
|
||||
VARRAY_FREE (id.target_exprs);
|
||||
if (DECL_LANG_SPECIFIC (fn))
|
||||
{
|
||||
tree ifn = make_tree_vec (VARRAY_ACTIVE_SIZE (id.inlined_fns));
|
||||
|
||||
memcpy (&TREE_VEC_ELT (ifn, 0), &VARRAY_TREE (id.inlined_fns, 0),
|
||||
VARRAY_ACTIVE_SIZE (id.inlined_fns) * sizeof (tree));
|
||||
DECL_INLINED_FNS (fn) = ifn;
|
||||
}
|
||||
VARRAY_FREE (id.inlined_fns);
|
||||
}
|
||||
|
||||
/* Optimize the body of FN. */
|
||||
|
||||
void
|
||||
@ -1070,38 +133,6 @@ update_cloned_parm (parm, cloned_parm)
|
||||
DECL_SOURCE_LINE (cloned_parm) = DECL_SOURCE_LINE (parm);
|
||||
}
|
||||
|
||||
/* FN is a function that has a complete body, and CLONE is a function
|
||||
whose body is to be set to a copy of FN, mapping argument
|
||||
declarations according to the ARG_MAP splay_tree. */
|
||||
|
||||
void
|
||||
clone_body (clone, fn, arg_map)
|
||||
tree clone, fn;
|
||||
void *arg_map;
|
||||
{
|
||||
inline_data id;
|
||||
|
||||
/* Clone the body, as if we were making an inline call. But, remap
|
||||
the parameters in the callee to the parameters of caller. If
|
||||
there's an in-charge parameter, map it to an appropriate
|
||||
constant. */
|
||||
memset (&id, 0, sizeof (id));
|
||||
VARRAY_TREE_INIT (id.fns, 2, "fns");
|
||||
VARRAY_PUSH_TREE (id.fns, clone);
|
||||
VARRAY_PUSH_TREE (id.fns, fn);
|
||||
id.decl_map = (splay_tree)arg_map;
|
||||
|
||||
/* Cloning is treated slightly differently from inlining. Set
|
||||
CLONING_P so that it's clear which operation we're performing. */
|
||||
id.cloning_p = true;
|
||||
|
||||
/* Actually copy the body. */
|
||||
TREE_CHAIN (DECL_SAVED_TREE (clone)) = copy_body (&id);
|
||||
|
||||
/* Clean up. */
|
||||
VARRAY_FREE (id.fns);
|
||||
}
|
||||
|
||||
/* FN is a function that has a complete body. Clone the body as
|
||||
necessary. Returns non-zero if there's no longer any need to
|
||||
process the main body. */
|
||||
|
322
gcc/cp/tree.c
322
gcc/cp/tree.c
@ -1153,241 +1153,6 @@ bind_template_template_parm (t, newargs)
|
||||
return t2;
|
||||
}
|
||||
|
||||
/* Apply FUNC to all the sub-trees of TP in a pre-order traversal.
|
||||
FUNC is called with the DATA and the address of each sub-tree. If
|
||||
FUNC returns a non-NULL value, the traversal is aborted, and the
|
||||
value returned by FUNC is returned. If HTAB is non-NULL it is used
|
||||
to record the nodes visited, and to avoid visiting a node more than
|
||||
once. */
|
||||
|
||||
tree
|
||||
walk_tree (tp, func, data, htab_)
|
||||
tree *tp;
|
||||
walk_tree_fn func;
|
||||
void *data;
|
||||
void *htab_;
|
||||
{
|
||||
htab_t htab = (htab_t) htab_;
|
||||
enum tree_code code;
|
||||
int walk_subtrees;
|
||||
tree result;
|
||||
|
||||
#define WALK_SUBTREE(NODE) \
|
||||
do \
|
||||
{ \
|
||||
result = walk_tree (&(NODE), func, data, htab); \
|
||||
if (result) \
|
||||
return result; \
|
||||
} \
|
||||
while (0)
|
||||
|
||||
/* Skip empty subtrees. */
|
||||
if (!*tp)
|
||||
return NULL_TREE;
|
||||
|
||||
if (htab)
|
||||
{
|
||||
void **slot;
|
||||
|
||||
/* Don't walk the same tree twice, if the user has requested
|
||||
that we avoid doing so. */
|
||||
if (htab_find (htab, *tp))
|
||||
return NULL_TREE;
|
||||
/* If we haven't already seen this node, add it to the table. */
|
||||
slot = htab_find_slot (htab, *tp, INSERT);
|
||||
*slot = *tp;
|
||||
}
|
||||
|
||||
/* Call the function. */
|
||||
walk_subtrees = 1;
|
||||
result = (*func) (tp, &walk_subtrees, data);
|
||||
|
||||
/* If we found something, return it. */
|
||||
if (result)
|
||||
return result;
|
||||
|
||||
code = TREE_CODE (*tp);
|
||||
|
||||
/* Even if we didn't, FUNC may have decided that there was nothing
|
||||
interesting below this point in the tree. */
|
||||
if (!walk_subtrees)
|
||||
{
|
||||
if (statement_code_p (code) || code == TREE_LIST
|
||||
|| LANG_TREE_CHAIN_MATTERS_P (*tp))
|
||||
/* But we still need to check our siblings. */
|
||||
return walk_tree (&TREE_CHAIN (*tp), func, data, htab);
|
||||
else
|
||||
return NULL_TREE;
|
||||
}
|
||||
|
||||
/* Handle common cases up front. */
|
||||
if (IS_EXPR_CODE_CLASS (TREE_CODE_CLASS (code))
|
||||
|| TREE_CODE_CLASS (code) == 'r'
|
||||
|| TREE_CODE_CLASS (code) == 's')
|
||||
{
|
||||
int i, len;
|
||||
|
||||
/* Set lineno here so we get the right instantiation context
|
||||
if we call instantiate_decl from inlinable_function_p. */
|
||||
if (statement_code_p (code) && !STMT_LINENO_FOR_FN_P (*tp))
|
||||
lineno = STMT_LINENO (*tp);
|
||||
|
||||
/* Walk over all the sub-trees of this operand. */
|
||||
len = first_rtl_op (code);
|
||||
/* TARGET_EXPRs are peculiar: operands 1 and 3 can be the same.
|
||||
But, we only want to walk once. */
|
||||
if (code == TARGET_EXPR
|
||||
&& TREE_OPERAND (*tp, 3) == TREE_OPERAND (*tp, 1))
|
||||
--len;
|
||||
/* Go through the subtrees. We need to do this in forward order so
|
||||
that the scope of a FOR_EXPR is handled properly. */
|
||||
for (i = 0; i < len; ++i)
|
||||
WALK_SUBTREE (TREE_OPERAND (*tp, i));
|
||||
|
||||
/* For statements, we also walk the chain so that we cover the
|
||||
entire statement tree. */
|
||||
if (statement_code_p (code))
|
||||
{
|
||||
if (code == DECL_STMT
|
||||
&& DECL_STMT_DECL (*tp)
|
||||
&& DECL_P (DECL_STMT_DECL (*tp)))
|
||||
{
|
||||
/* Walk the DECL_INITIAL and DECL_SIZE. We don't want to walk
|
||||
into declarations that are just mentioned, rather than
|
||||
declared; they don't really belong to this part of the tree.
|
||||
And, we can see cycles: the initializer for a declaration can
|
||||
refer to the declaration itself. */
|
||||
WALK_SUBTREE (DECL_INITIAL (DECL_STMT_DECL (*tp)));
|
||||
WALK_SUBTREE (DECL_SIZE (DECL_STMT_DECL (*tp)));
|
||||
WALK_SUBTREE (DECL_SIZE_UNIT (DECL_STMT_DECL (*tp)));
|
||||
}
|
||||
|
||||
/* This can be tail-recursion optimized if we write it this way. */
|
||||
return walk_tree (&TREE_CHAIN (*tp), func, data, htab);
|
||||
}
|
||||
|
||||
/* We didn't find what we were looking for. */
|
||||
return NULL_TREE;
|
||||
}
|
||||
else if (TREE_CODE_CLASS (code) == 'd')
|
||||
{
|
||||
WALK_SUBTREE (TREE_TYPE (*tp));
|
||||
|
||||
/* We didn't find what we were looking for. */
|
||||
return NULL_TREE;
|
||||
}
|
||||
|
||||
result = LANG_WALK_SUBTREES (tp, &walk_subtrees, func, data, htab);
|
||||
if (result || ! walk_subtrees)
|
||||
return result;
|
||||
|
||||
/* Not one of the easy cases. We must explicitly go through the
|
||||
children. */
|
||||
switch (code)
|
||||
{
|
||||
case ERROR_MARK:
|
||||
case IDENTIFIER_NODE:
|
||||
case INTEGER_CST:
|
||||
case REAL_CST:
|
||||
case STRING_CST:
|
||||
case REAL_TYPE:
|
||||
case COMPLEX_TYPE:
|
||||
case VECTOR_TYPE:
|
||||
case VOID_TYPE:
|
||||
case BOOLEAN_TYPE:
|
||||
case UNION_TYPE:
|
||||
case ENUMERAL_TYPE:
|
||||
case BLOCK:
|
||||
case RECORD_TYPE:
|
||||
/* None of thse have subtrees other than those already walked
|
||||
above. */
|
||||
break;
|
||||
|
||||
case POINTER_TYPE:
|
||||
case REFERENCE_TYPE:
|
||||
WALK_SUBTREE (TREE_TYPE (*tp));
|
||||
break;
|
||||
|
||||
case TREE_LIST:
|
||||
WALK_SUBTREE (TREE_VALUE (*tp));
|
||||
WALK_SUBTREE (TREE_CHAIN (*tp));
|
||||
break;
|
||||
|
||||
case TREE_VEC:
|
||||
{
|
||||
int len = TREE_VEC_LENGTH (*tp);
|
||||
while (len--)
|
||||
WALK_SUBTREE (TREE_VEC_ELT (*tp, len));
|
||||
}
|
||||
break;
|
||||
|
||||
case COMPLEX_CST:
|
||||
WALK_SUBTREE (TREE_REALPART (*tp));
|
||||
WALK_SUBTREE (TREE_IMAGPART (*tp));
|
||||
break;
|
||||
|
||||
case CONSTRUCTOR:
|
||||
WALK_SUBTREE (CONSTRUCTOR_ELTS (*tp));
|
||||
break;
|
||||
|
||||
case METHOD_TYPE:
|
||||
WALK_SUBTREE (TYPE_METHOD_BASETYPE (*tp));
|
||||
/* Fall through. */
|
||||
|
||||
case FUNCTION_TYPE:
|
||||
WALK_SUBTREE (TREE_TYPE (*tp));
|
||||
{
|
||||
tree arg = TYPE_ARG_TYPES (*tp);
|
||||
|
||||
/* We never want to walk into default arguments. */
|
||||
for (; arg; arg = TREE_CHAIN (arg))
|
||||
WALK_SUBTREE (TREE_VALUE (arg));
|
||||
}
|
||||
break;
|
||||
|
||||
case ARRAY_TYPE:
|
||||
WALK_SUBTREE (TREE_TYPE (*tp));
|
||||
WALK_SUBTREE (TYPE_DOMAIN (*tp));
|
||||
break;
|
||||
|
||||
case INTEGER_TYPE:
|
||||
WALK_SUBTREE (TYPE_MIN_VALUE (*tp));
|
||||
WALK_SUBTREE (TYPE_MAX_VALUE (*tp));
|
||||
break;
|
||||
|
||||
case OFFSET_TYPE:
|
||||
WALK_SUBTREE (TREE_TYPE (*tp));
|
||||
WALK_SUBTREE (TYPE_OFFSET_BASETYPE (*tp));
|
||||
break;
|
||||
|
||||
default:
|
||||
abort ();
|
||||
}
|
||||
|
||||
/* We didn't find what we were looking for. */
|
||||
return NULL_TREE;
|
||||
|
||||
#undef WALK_SUBTREE
|
||||
}
|
||||
|
||||
/* Like walk_tree, but does not walk duplicate nodes more than
|
||||
once. */
|
||||
|
||||
tree
|
||||
walk_tree_without_duplicates (tp, func, data)
|
||||
tree *tp;
|
||||
walk_tree_fn func;
|
||||
void *data;
|
||||
{
|
||||
tree result;
|
||||
htab_t htab;
|
||||
|
||||
htab = htab_create (37, htab_hash_pointer, htab_eq_pointer, NULL);
|
||||
result = walk_tree (tp, func, data, htab);
|
||||
htab_delete (htab);
|
||||
return result;
|
||||
}
|
||||
|
||||
/* Called from count_trees via walk_tree. */
|
||||
|
||||
static tree
|
||||
@ -1512,51 +1277,6 @@ no_linkage_check (t)
|
||||
return NULL_TREE;
|
||||
}
|
||||
|
||||
/* Passed to walk_tree. Copies the node pointed to, if appropriate. */
|
||||
|
||||
tree
|
||||
copy_tree_r (tp, walk_subtrees, data)
|
||||
tree *tp;
|
||||
int *walk_subtrees;
|
||||
void *data ATTRIBUTE_UNUSED;
|
||||
{
|
||||
enum tree_code code = TREE_CODE (*tp);
|
||||
|
||||
/* We make copies of most nodes. */
|
||||
if (IS_EXPR_CODE_CLASS (TREE_CODE_CLASS (code))
|
||||
|| TREE_CODE_CLASS (code) == 'r'
|
||||
|| TREE_CODE_CLASS (code) == 'c'
|
||||
|| TREE_CODE_CLASS (code) == 's'
|
||||
|| code == TREE_LIST
|
||||
|| code == TREE_VEC
|
||||
|| LANG_TREE_CHAIN_MATTERS_P (*tp))
|
||||
{
|
||||
/* Because the chain gets clobbered when we make a copy, we save it
|
||||
here. */
|
||||
tree chain = TREE_CHAIN (*tp);
|
||||
|
||||
/* Copy the node. */
|
||||
*tp = copy_node (*tp);
|
||||
|
||||
/* Now, restore the chain, if appropriate. That will cause
|
||||
walk_tree to walk into the chain as well. */
|
||||
if (code == PARM_DECL || code == TREE_LIST
|
||||
|| LANG_TREE_CHAIN_MATTERS_P (*tp)
|
||||
|| statement_code_p (code))
|
||||
TREE_CHAIN (*tp) = chain;
|
||||
|
||||
/* For now, we don't update BLOCKs when we make copies. So, we
|
||||
have to nullify all scope-statements. */
|
||||
if (TREE_CODE (*tp) == SCOPE_STMT)
|
||||
SCOPE_STMT_BLOCK (*tp) = NULL_TREE;
|
||||
}
|
||||
else if (TREE_CODE_CLASS (code) == 't')
|
||||
/* There's no need to copy types, or anything beneath them. */
|
||||
*walk_subtrees = 0;
|
||||
|
||||
return NULL_TREE;
|
||||
}
|
||||
|
||||
#ifdef GATHER_STATISTICS
|
||||
extern int depth_reached;
|
||||
#endif
|
||||
@ -2565,48 +2285,6 @@ init_tree ()
|
||||
mark_tree_hashtable);
|
||||
}
|
||||
|
||||
/* The SAVE_EXPR pointed to by TP is being copied. If ST contains
|
||||
information indicating to what new SAVE_EXPR this one should be
|
||||
mapped, use that one. Otherwise, create a new node and enter it in
|
||||
ST. FN is the function into which the copy will be placed. */
|
||||
|
||||
void
|
||||
remap_save_expr (tp, st_, fn, walk_subtrees)
|
||||
tree *tp;
|
||||
void *st_;
|
||||
tree fn;
|
||||
int *walk_subtrees;
|
||||
{
|
||||
splay_tree st = (splay_tree) st_;
|
||||
splay_tree_node n;
|
||||
|
||||
/* See if we already encountered this SAVE_EXPR. */
|
||||
n = splay_tree_lookup (st, (splay_tree_key) *tp);
|
||||
|
||||
/* If we didn't already remap this SAVE_EXPR, do so now. */
|
||||
if (!n)
|
||||
{
|
||||
tree t = copy_node (*tp);
|
||||
|
||||
/* The SAVE_EXPR is now part of the function into which we
|
||||
are inlining this body. */
|
||||
SAVE_EXPR_CONTEXT (t) = fn;
|
||||
/* And we haven't evaluated it yet. */
|
||||
SAVE_EXPR_RTL (t) = NULL_RTX;
|
||||
/* Remember this SAVE_EXPR. */
|
||||
n = splay_tree_insert (st,
|
||||
(splay_tree_key) *tp,
|
||||
(splay_tree_value) t);
|
||||
}
|
||||
else
|
||||
/* We've already walked into this SAVE_EXPR, so we needn't do it
|
||||
again. */
|
||||
*walk_subtrees = 0;
|
||||
|
||||
/* Replace this SAVE_EXPR with the copy. */
|
||||
*tp = (tree) n->value;
|
||||
}
|
||||
|
||||
/* Called via walk_tree. If *TP points to a DECL_STMT for a local
|
||||
declaration, copies the declaration and enters it in the splay_tree
|
||||
pointed to by DATA (which is really a `splay_tree *'). */
|
||||
|
Loading…
Reference in New Issue
Block a user