gcc/gcc/cp/optimize.c

776 lines
24 KiB
C

/* Perform optimizations on tree structure.
Copyright (C) 1998, 1999 Free Software Foundation, Inc.
Written by Mark Michell (mark@codesourcery.com).
This file is part of GNU CC.
GNU CC is free software; you can redistribute it and/or modify it
under the terms of the GNU General Public License as published by
the Free Software Foundation; either version 2, or (at your option)
any later version.
GNU CC is distributed in the hope that it will be useful, but
WITHOUT ANY WARRANTY; without even the implied warranty of
MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
General Public License for more details.
You should have received a copy of the GNU General Public License
along with GNU CC; see the file COPYING. If not, write to the Free
Software Foundation, 59 Temple Place - Suite 330, Boston, MA
02111-1307, USA. */
#include "config.h"
#include "system.h"
#include "tree.h"
#include "cp-tree.h"
#include "rtl.h"
#include "insn-config.h"
#include "integrate.h"
#include "varray.h"
/* To Do:
o In order to make inlining-on-trees work, we pessimized
function-local static constants. In particular, they are now
always output, even when not addressed. Fix this by treating
function-local static constants just like global static
constants; the back-end already knows not to output them if they
are not needed.
o Provide heuristics to clamp inlining of recursive template
calls?
o It looks like the return label is not being placed in the optimal
place. Shouldn't it come before the returned value? */
/* Data required for function inlining. */
typedef struct inline_data
{
/* A stack of the functions we are inlining. For example, if we are
compiling `f', which calls `g', which calls `h', and we are
inlining the body of `h', the stack will contain, `h', followed
by `g', followed by `f'. */
varray_type fns;
/* The last SCOPE_STMT we have encountered. */
tree scope_stmt;
/* The label to jump to when a return statement is encountered. */
tree ret_label;
/* The map from local declarations in the inlined function to
equivalents in the function into which it is being inlined. */
splay_tree decl_map;
/* Nonzero if we are currently within the cleanup for a
TARGET_EXPR. */
int in_target_cleanup_p;
} inline_data;
/* Prototypes. */
static tree initialize_inlined_parameters PROTO((inline_data *, tree, tree));
static tree declare_return_variable PROTO((inline_data *, tree *));
static tree copy_body_r PROTO((tree *, int *, void *));
static tree copy_body PROTO((inline_data *));
static tree expand_call_inline PROTO((tree *, int *, void *));
static void expand_calls_inline PROTO((tree *, inline_data *));
static int inlinable_function_p PROTO((tree, inline_data *));
static tree remap_decl PROTO((tree, inline_data *));
static void remap_block PROTO((tree, tree, inline_data *));
static void copy_scope_stmt PROTO((tree *, int *, inline_data *));
static tree calls_setjmp_r PROTO((tree *, int *, void *));
/* Remap DECL during the copying of the BLOCK tree for the function.
DATA is really an `inline_data *'. */
static tree
remap_decl (decl, id)
tree decl;
inline_data *id;
{
splay_tree_node n;
tree fn;
/* We only remap local variables in the current function. */
fn = VARRAY_TOP_TREE (id->fns);
if (!nonstatic_local_decl_p (decl) || DECL_CONTEXT (decl) != fn)
return NULL_TREE;
/* See if we have remapped this declaration. */
n = splay_tree_lookup (id->decl_map, (splay_tree_key) decl);
/* If we didn't already have an equivalent for this declaration,
create one now. */
if (!n)
{
tree t;
/* Make a copy of the variable or label. */
t = copy_decl_for_inlining (decl, fn,
VARRAY_TREE (id->fns, 0));
/* Remember it, so that if we encounter this local entity
again we can reuse this copy. */
n = splay_tree_insert (id->decl_map,
(splay_tree_key) decl,
(splay_tree_value) t);
}
return (tree) n->value;
}
/* Copy the SCOPE_STMT_BLOCK associated with SCOPE_STMT to contain
remapped versions of the variables therein. And hook the new block
into the block-tree. If non-NULL, the DECLS are declarations to
add to use instead of the BLOCK_VARS in the old block. */
static void
remap_block (scope_stmt, decls, id)
tree scope_stmt;
tree decls;
inline_data *id;
{
/* We cannot do this in the cleanup for a TARGET_EXPR since we do
not know whether or not expand_expr will actually write out the
code we put there. If it does not, then we'll have more BLOCKs
than block-notes, and things will go awry. At some point, we
should make the back-end handle BLOCK notes in a tidier way,
without requiring a strict correspondence to the block-tree; then
this check can go. */
if (id->in_target_cleanup_p)
{
SCOPE_STMT_BLOCK (scope_stmt) = NULL_TREE;
return;
}
/* If this is the beginning of a scope, remap the associated BLOCK. */
if (SCOPE_BEGIN_P (scope_stmt) && SCOPE_STMT_BLOCK (scope_stmt))
{
tree old_block;
tree new_block;
tree old_var;
/* Make the new block. */
old_block = SCOPE_STMT_BLOCK (scope_stmt);
new_block = make_node (BLOCK);
TREE_USED (new_block) = TREE_USED (old_block);
BLOCK_ABSTRACT_ORIGIN (new_block) = old_block;
SCOPE_STMT_BLOCK (scope_stmt) = new_block;
/* Remap its variables. */
for (old_var = decls ? decls : BLOCK_VARS (old_block);
old_var;
old_var = TREE_CHAIN (old_var))
{
tree new_var;
/* Remap the variable. */
new_var = remap_decl (old_var, id);
if (!new_var)
/* We didn't remap this variable, so we can't mess with
its TREE_CHAIN. */
;
else
{
TREE_CHAIN (new_var) = BLOCK_VARS (new_block);
BLOCK_VARS (new_block) = new_var;
}
}
/* We put the BLOCK_VARS in reverse order; fix that now. */
BLOCK_VARS (new_block) = nreverse (BLOCK_VARS (new_block));
/* Graft the new block into the tree. */
insert_block_after_note (new_block,
SCOPE_STMT_BLOCK (id->scope_stmt),
SCOPE_BEGIN_P (id->scope_stmt));
/* Remember that this is now the last scope statement with
an associated block. */
id->scope_stmt = scope_stmt;
/* Remember the remapped block. */
splay_tree_insert (id->decl_map,
(splay_tree_key) old_block,
(splay_tree_value) new_block);
}
/* If this is the end of a scope, set the SCOPE_STMT_BLOCK to be the
remapped block. */
else if (SCOPE_END_P (scope_stmt) && SCOPE_STMT_BLOCK (scope_stmt))
{
splay_tree_node n;
/* Find this block in the table of remapped things. */
n = splay_tree_lookup (id->decl_map,
(splay_tree_key) SCOPE_STMT_BLOCK (scope_stmt));
my_friendly_assert (n != NULL, 19991203);
SCOPE_STMT_BLOCK (scope_stmt) = (tree) n->value;
/* Remember that this is now the last scope statement with an
associated block. */
id->scope_stmt = scope_stmt;
}
}
/* Copy the SCOPE_STMT pointed to by TP. */
static void
copy_scope_stmt (tp, walk_subtrees, id)
tree *tp;
int *walk_subtrees;
inline_data *id;
{
tree block;
/* Remember whether or not this statement was nullified. When
making a copy, copy_tree_r always sets SCOPE_NULLIFIED_P (and
doesn't copy the SCOPE_STMT_BLOCK) to free callers from having to
deal with copying BLOCKs if they do not wish to do so. */
block = SCOPE_STMT_BLOCK (*tp);
/* Copy (and replace) the statement. */
copy_tree_r (tp, walk_subtrees, NULL);
/* Restore the SCOPE_STMT_BLOCK. */
SCOPE_STMT_BLOCK (*tp) = block;
/* Remap the associated block. */
remap_block (*tp, NULL_TREE, id);
}
/* Called from copy_body via walk_tree. DATA is really an
`inline_data *'. */
static tree
copy_body_r (tp, walk_subtrees, data)
tree *tp;
int *walk_subtrees;
void *data;
{
inline_data* id;
tree fn;
/* Set up. */
id = (inline_data *) data;
fn = VARRAY_TOP_TREE (id->fns);
/* All automatic variables should have a DECL_CONTEXT indicating
what function they come from. */
if ((TREE_CODE (*tp) == VAR_DECL || TREE_CODE (*tp) == LABEL_DECL)
&& DECL_NAMESPACE_SCOPE_P (*tp))
my_friendly_assert (DECL_EXTERNAL (*tp) || TREE_STATIC (*tp),
19991113);
/* If this is a RETURN_STMT, change it into an EXPR_STMT and a
GOTO_STMT with the RET_LABEL as its target. */
if (TREE_CODE (*tp) == RETURN_STMT)
{
tree return_stmt = *tp;
tree goto_stmt;
/* Build the GOTO_STMT. */
goto_stmt = build_min_nt (GOTO_STMT, id->ret_label);
TREE_CHAIN (goto_stmt) = TREE_CHAIN (return_stmt);
/* If we're returning something, just turn that into an
assignment into the equivalent of the original
RESULT_DECL. */
if (RETURN_EXPR (return_stmt))
{
*tp = build_min_nt (EXPR_STMT,
RETURN_EXPR (return_stmt));
/* And then jump to the end of the function. */
TREE_CHAIN (*tp) = goto_stmt;
}
/* If we're not returning anything just do the jump. */
else
*tp = goto_stmt;
}
/* Local variables and labels need to be replaced by equivalent
variables. We don't want to copy static variables; there's only
one of those, no matter how many times we inline the containing
function. */
else if (nonstatic_local_decl_p (*tp) && DECL_CONTEXT (*tp) == fn)
{
tree new_decl;
/* Remap the declaration. */
new_decl = remap_decl (*tp, id);
my_friendly_assert (new_decl != NULL_TREE, 19991203);
/* Replace this variable with the copy. */
*tp = new_decl;
}
else if (TREE_CODE (*tp) == SAVE_EXPR)
remap_save_expr (tp, id->decl_map, VARRAY_TREE (id->fns, 0),
walk_subtrees);
else if (TREE_CODE (*tp) == UNSAVE_EXPR)
my_friendly_abort (19991113);
/* For a SCOPE_STMT, we must copy the associated block so that we
can write out debugging information for the inlined variables. */
else if (TREE_CODE (*tp) == SCOPE_STMT && !id->in_target_cleanup_p)
copy_scope_stmt (tp, walk_subtrees, id);
/* Otherwise, just copy the node. Note that copy_tree_r already
knows not to copy VAR_DECLs, etc., so this is safe. */
else
{
copy_tree_r (tp, walk_subtrees, NULL);
/* The copied TARGET_EXPR has never been expanded, even if the
original node was expanded already. */
if (TREE_CODE (*tp) == TARGET_EXPR && TREE_OPERAND (*tp, 3))
TREE_OPERAND (*tp, 1) = TREE_OPERAND (*tp, 3);
/* Similarly, if we're copying a CALL_EXPR, the RTL for the
result is no longer valid. */
else if (TREE_CODE (*tp) == CALL_EXPR)
CALL_EXPR_RTL (*tp) = NULL_RTX;
}
/* Keep iterating. */
return NULL_TREE;
}
/* Make a copy of the body of FN so that it can be inserted inline in
another function. */
static tree
copy_body (id)
inline_data *id;
{
tree body;
body = DECL_SAVED_TREE (VARRAY_TOP_TREE (id->fns));
walk_tree (&body, copy_body_r, id);
return body;
}
/* Generate code to initialize the parameters of the function at the
top of the stack in ID from the ARGS (presented as a TREE_LIST). */
static tree
initialize_inlined_parameters (id, args, fn)
inline_data *id;
tree args;
tree fn;
{
tree init_stmts;
tree parms;
tree a;
tree p;
/* Figure out what the parameters are. */
parms = DECL_ARGUMENTS (fn);
/* Start with no initializations whatsoever. */
init_stmts = NULL_TREE;
/* Loop through the parameter declarations, replacing each with an
equivalent VAR_DECL, appropriately initialized. */
for (p = parms, a = args; p; a = TREE_CHAIN (a), p = TREE_CHAIN (p))
{
tree init_stmt;
tree var;
/* Make an equivalent VAR_DECL. */
var = copy_decl_for_inlining (p, fn, VARRAY_TREE (id->fns, 0));
/* Register the VAR_DECL as the equivalent for the PARM_DECL;
that way, when the PARM_DECL is encountered, it will be
automatically replaced by the VAR_DECL. */
splay_tree_insert (id->decl_map,
(splay_tree_key) p,
(splay_tree_value) var);
/* Initialize this VAR_DECL from the equivalent argument. If
the argument is an object, created via a constructor or copy,
this will not result in an extra copy: the TARGET_EXPR
representing the argument will be bound to VAR, and the
object will be constructed in VAR. */
init_stmt = build_min_nt (EXPR_STMT,
build (INIT_EXPR, TREE_TYPE (p),
var, TREE_VALUE (a)));
/* Declare this new variable. Note that we do this *after* the
initialization because we are going to reverse all the
initialization statements below. */
TREE_CHAIN (init_stmt) = build_min_nt (DECL_STMT, var);
/* Add this initialization to the list. */
TREE_CHAIN (TREE_CHAIN (init_stmt)) = init_stmts;
init_stmts = init_stmt;
}
/* The initialization statements have been built up in reverse
order. Straighten them out now. */
return nreverse (init_stmts);
}
/* Declare a return variable to replace the RESULT_DECL for the
function we are calling. An appropriate DECL_STMT is returned.
The USE_STMT is filled in to contain a use of the declaration to
indicate the return value of the function. */
static tree
declare_return_variable (id, use_stmt)
struct inline_data *id;
tree *use_stmt;
{
tree fn = VARRAY_TOP_TREE (id->fns);
tree result = DECL_RESULT (fn);
tree var;
/* We don't need to do anything for functions that don't return
anything. */
if (!result || same_type_p (TYPE_MAIN_VARIANT (TREE_TYPE (result)),
void_type_node))
{
*use_stmt = NULL_TREE;
return NULL_TREE;
}
/* Make an appropriate copy. */
var = copy_decl_for_inlining (result, fn, VARRAY_TREE (id->fns, 0));
/* Register the VAR_DECL as the equivalent for the RESULT_DECL; that
way, when the RESULT_DECL is encountered, it will be
automatically replaced by the VAR_DECL. */
splay_tree_insert (id->decl_map,
(splay_tree_key) result,
(splay_tree_value) var);
/* Build the USE_STMT. */
*use_stmt = build_min_nt (EXPR_STMT, var);
/* Build the declaration statement. */
return build_min_nt (DECL_STMT, var);
}
/* Returns non-zero if FN is a function that can be inlined. */
static int
inlinable_function_p (fn, id)
tree fn;
inline_data *id;
{
int inlinable;
/* If we've already decided this function shouldn't be inlined,
there's no need to check again. */
if (DECL_UNINLINABLE (fn))
return 0;
/* Assume it is not inlinable. */
inlinable = 0;
/* If we're not inlining things, then nothing is inlinable. */
if (!flag_inline_trees)
;
/* If the function was not declared `inline', then we don't inline
it. */
else if (!DECL_INLINE (fn))
;
/* If we don't have the function body available, we can't inline
it. */
else if (!DECL_SAVED_TREE (fn))
;
/* We can't inline varargs functions. */
else if (varargs_function_p (fn))
;
/* All is well. We can inline this function. Traditionally, GCC
has refused to inline functions using setjmp or alloca, or
functions whose values are returned in a PARALLEL, and a few
other such obscure conditions. We are not equally constrained at
the tree level. */
else
inlinable = 1;
/* Squirrel away the result so that we don't have to check again. */
DECL_UNINLINABLE (fn) = !inlinable;
/* Don't do recursive inlining, either. We don't record this in
DECL_UNLINABLE; we may be able to inline this function later. */
if (inlinable)
{
size_t i;
for (i = 0; i < id->fns->elements_used; ++i)
if (VARRAY_TREE (id->fns, i) == fn)
inlinable = 0;
}
/* We can inline a template instantiation only if it's fully
instantiated. */
if (inlinable
&& DECL_TEMPLATE_INFO (fn)
&& TI_PENDING_TEMPLATE_FLAG (DECL_TEMPLATE_INFO (fn)))
{
fn = instantiate_decl (fn);
inlinable = !TI_PENDING_TEMPLATE_FLAG (DECL_TEMPLATE_INFO (fn));
}
/* Return the result. */
return inlinable;
}
/* If *TP is a CALL_EXPR, replace it with its inline expansion. */
static tree
expand_call_inline (tp, walk_subtrees, data)
tree *tp;
int *walk_subtrees;
void *data;
{
inline_data *id;
tree t;
tree expr;
tree chain;
tree fn;
tree scope_stmt;
tree use_stmt;
tree arg_inits;
splay_tree st;
/* See what we've got. */
id = (inline_data *) data;
t = *tp;
/* Keep track of the last SCOPE_STMT we've seen. */
if (TREE_CODE (t) == SCOPE_STMT)
{
if (SCOPE_STMT_BLOCK (t) && !id->in_target_cleanup_p)
id->scope_stmt = t;
return NULL_TREE;
}
/* Recurse, but letting recursive invocations know that we are
inside the body of a TARGET_EXPR. */
if (TREE_CODE (*tp) == TARGET_EXPR)
{
int i, len = first_rtl_op (TARGET_EXPR);
/* We're walking our own subtrees. */
*walk_subtrees = 0;
/* Actually walk over them. This loop is the body of
walk_trees, omitting the case where the TARGET_EXPR
itself is handled. */
for (i = 0; i < len; ++i)
{
if (i == 2)
++id->in_target_cleanup_p;
walk_tree (&TREE_OPERAND (*tp, i), expand_call_inline, data);
if (i == 2)
--id->in_target_cleanup_p;
}
return NULL_TREE;
}
/* From here on, we're only interested in CALL_EXPRs. */
if (TREE_CODE (t) != CALL_EXPR)
return NULL_TREE;
/* First, see if we can figure out what function is being called.
If we cannot, then there is no hope of inlining the function. */
fn = get_callee_fndecl (t);
if (!fn)
return NULL_TREE;
/* Don't try to inline functions that are not well-suited to
inlining. */
if (!inlinable_function_p (fn, id))
return NULL_TREE;
/* Build a statement-expression containing code to initialize the
arguments, the actual inline expansion of the body, and a label
for the return statements within the function to jump to. The
type of the statement expression is the return type of the
function call. */
expr = build_min (STMT_EXPR, TREE_TYPE (TREE_TYPE (fn)), NULL_TREE);
/* Local declarations will be replaced by their equivalents in this
map. */
st = id->decl_map;
id->decl_map = splay_tree_new (splay_tree_compare_pointers,
NULL, NULL);
/* Initialize the parameters. */
arg_inits = initialize_inlined_parameters (id, TREE_OPERAND (t, 1), fn);
/* Expand any inlined calls in the initializers. Do this before we
push FN on the stack of functions we are inlining; we want to
inline calls to FN that appear in the initializers for the
parameters. */
expand_calls_inline (&arg_inits, id);
/* And add them to the tree. */
STMT_EXPR_STMT (expr) = chainon (STMT_EXPR_STMT (expr), arg_inits);
/* Record the function we are about to inline so that we can avoid
recursing into it. */
VARRAY_PUSH_TREE (id->fns, fn);
/* Return statements in the function body will be replaced by jumps
to the RET_LABEL. */
id->ret_label = build_decl (LABEL_DECL, NULL_TREE, NULL_TREE);
DECL_CONTEXT (id->ret_label) = VARRAY_TREE (id->fns, 0);
/* Create a block to put the parameters in. We have to do this
after the parameters have been remapped because remapping
parameters is different from remapping ordinary variables. */
scope_stmt = build_min_nt (SCOPE_STMT, DECL_INITIAL (fn));
SCOPE_BEGIN_P (scope_stmt) = 1;
SCOPE_NO_CLEANUPS_P (scope_stmt) = 1;
remap_block (scope_stmt, DECL_ARGUMENTS (fn), id);
TREE_CHAIN (scope_stmt) = STMT_EXPR_STMT (expr);
STMT_EXPR_STMT (expr) = scope_stmt;
id->scope_stmt = scope_stmt;
/* Tell the debugging backends that this block represents the
outermost scope of the inlined function. FIXME what to do for
inlines in cleanups? */
if (SCOPE_STMT_BLOCK (scope_stmt))
BLOCK_ABSTRACT_ORIGIN (SCOPE_STMT_BLOCK (scope_stmt)) = DECL_ORIGIN (fn);
/* Declare the return variable for the function. */
STMT_EXPR_STMT (expr)
= chainon (STMT_EXPR_STMT (expr),
declare_return_variable (id, &use_stmt));
/* After we've initialized the parameters, we insert the body of the
function itself. */
STMT_EXPR_STMT (expr)
= chainon (STMT_EXPR_STMT (expr), copy_body (id));
/* Close the block for the parameters. */
scope_stmt = build_min_nt (SCOPE_STMT, DECL_INITIAL (fn));
SCOPE_NO_CLEANUPS_P (scope_stmt) = 1;
my_friendly_assert (DECL_INITIAL (fn)
&& TREE_CODE (DECL_INITIAL (fn)) == BLOCK,
19991203);
remap_block (scope_stmt, NULL_TREE, id);
STMT_EXPR_STMT (expr)
= chainon (STMT_EXPR_STMT (expr), scope_stmt);
/* Finally, mention the returned value so that the value of the
statement-expression is the returned value of the function. */
STMT_EXPR_STMT (expr) = chainon (STMT_EXPR_STMT (expr), use_stmt);
/* Clean up. */
splay_tree_delete (id->decl_map);
id->decl_map = st;
/* After the body of the function comes the RET_LABEL. */
STMT_EXPR_STMT (expr)
= chainon (STMT_EXPR_STMT (expr),
build_min_nt (LABEL_STMT, id->ret_label));
/* The new expression has side-effects if the old one did. */
TREE_SIDE_EFFECTS (expr) = TREE_SIDE_EFFECTS (t);
/* Replace the call by the inlined body. Wrap it in an
EXPR_WITH_FILE_LOCATION so that we'll get debugging line notes
pointing to the right place. */
chain = TREE_CHAIN (*tp);
*tp = build_expr_wfl (expr, DECL_SOURCE_FILE (fn), DECL_SOURCE_LINE (fn),
/*col=*/0);
EXPR_WFL_EMIT_LINE_NOTE (*tp) = 1;
TREE_CHAIN (*tp) = chain;
/* If the value of the new expression is ignored, that's OK. We
don't warn about this for CALL_EXPRs, so we shouldn't warn about
the equivalent inlined version either. */
TREE_USED (*tp) = 1;
/* Recurse into the body of the just inlined function. */
expand_calls_inline (tp, id);
VARRAY_POP (id->fns);
/* Don't walk into subtrees. We've already handled them above. */
*walk_subtrees = 0;
/* Keep iterating. */
return NULL_TREE;
}
/* Walk over the entire tree *TP, replacing CALL_EXPRs with inline
expansions as appropriate. */
static void
expand_calls_inline (tp, id)
tree *tp;
inline_data *id;
{
/* Search through *TP, replacing all calls to inline functions by
appropriate equivalents. */
walk_tree (tp, expand_call_inline, id);
}
/* Optimize the body of FN. */
void
optimize_function (fn)
tree fn;
{
/* Expand calls to inline functions. */
if (flag_inline_trees)
{
inline_data id;
tree prev_fn;
struct saved_scope *s;
/* Clear out ID. */
memset (&id, 0, sizeof (id));
/* Don't allow recursion into FN. */
VARRAY_TREE_INIT (id.fns, 32, "fns");
VARRAY_PUSH_TREE (id.fns, fn);
/* Or any functions that aren't finished yet. */
prev_fn = NULL_TREE;
if (current_function_decl)
{
VARRAY_PUSH_TREE (id.fns, current_function_decl);
prev_fn = current_function_decl;
}
for (s = scope_chain; s; s = s->prev)
if (s->function_decl && s->function_decl != prev_fn)
{
VARRAY_PUSH_TREE (id.fns, s->function_decl);
prev_fn = s->function_decl;
}
/* Initialize id->scope_stmt with a fake SCOPE_STMT for the outermost
block of the function (i.e. the BLOCK with __FUNCTION__ et al). */
id.scope_stmt = build_min_nt (SCOPE_STMT,
BLOCK_SUBBLOCKS (DECL_INITIAL (fn)));
SCOPE_BEGIN_P (id.scope_stmt) = 1;
/* Replace all calls to inline functions with the bodies of those
functions. */
expand_calls_inline (&DECL_SAVED_TREE (fn), &id);
/* Clean up. */
VARRAY_FREE (id.fns);
}
}
/* Called from calls_setjmp_p via walk_tree. */
static tree
calls_setjmp_r (tp, walk_subtrees, data)
tree *tp;
int *walk_subtrees ATTRIBUTE_UNUSED;
void *data ATTRIBUTE_UNUSED;
{
int setjmp_p;
int longjmp_p;
int malloc_p;
int alloca_p;
/* We're only interested in FUNCTION_DECLS. */
if (TREE_CODE (*tp) != FUNCTION_DECL)
return NULL_TREE;
special_function_p (*tp, &setjmp_p, &longjmp_p, &malloc_p, &alloca_p);
return setjmp_p ? *tp : NULL_TREE;
}
/* Returns non-zero if FN calls `setjmp' or some other function that
can return more than once. This function is conservative; it may
occasionally return a non-zero value even when FN does not actually
call `setjmp'. */
int
calls_setjmp_p (fn)
tree fn;
{
return (walk_tree (&DECL_SAVED_TREE (fn), calls_setjmp_r, NULL)
!= NULL_TREE);
}