cbe34bb5ed
From-SVN: r243994
431 lines
12 KiB
C
431 lines
12 KiB
C
/* Miscellaneous utilities for tree streaming. Things that are used
|
|
in both input and output are here.
|
|
|
|
Copyright (C) 2011-2017 Free Software Foundation, Inc.
|
|
Contributed by Diego Novillo <dnovillo@google.com>
|
|
|
|
This file is part of GCC.
|
|
|
|
GCC is free software; you can redistribute it and/or modify it under
|
|
the terms of the GNU General Public License as published by the Free
|
|
Software Foundation; either version 3, or (at your option) any later
|
|
version.
|
|
|
|
GCC is distributed in the hope that it will be useful, but WITHOUT ANY
|
|
WARRANTY; without even the implied warranty of MERCHANTABILITY or
|
|
FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License
|
|
for more details.
|
|
|
|
You should have received a copy of the GNU General Public License
|
|
along with GCC; see the file COPYING3. If not see
|
|
<http://www.gnu.org/licenses/>. */
|
|
|
|
#include "config.h"
|
|
#include "system.h"
|
|
#include "coretypes.h"
|
|
#include "backend.h"
|
|
#include "tree.h"
|
|
#include "gimple.h"
|
|
#include "tree-streamer.h"
|
|
#include "cgraph.h"
|
|
|
|
/* Table indexed by machine_mode, used for 2 different purposes.
|
|
During streaming out we record there non-zero value for all modes
|
|
that were streamed out.
|
|
During streaming in, we translate the on the disk mode using this
|
|
table. For normal LTO it is set to identity, for ACCEL_COMPILER
|
|
depending on the mode_table content. */
|
|
unsigned char streamer_mode_table[1 << 8];
|
|
|
|
/* Check that all the TS_* structures handled by the streamer_write_* and
|
|
streamer_read_* routines are exactly ALL the structures defined in
|
|
treestruct.def. */
|
|
|
|
void
|
|
streamer_check_handled_ts_structures (void)
|
|
{
|
|
bool handled_p[LAST_TS_ENUM];
|
|
unsigned i;
|
|
|
|
memset (&handled_p, 0, sizeof (handled_p));
|
|
|
|
/* These are the TS_* structures that are either handled or
|
|
explicitly ignored by the streamer routines. */
|
|
handled_p[TS_BASE] = true;
|
|
handled_p[TS_TYPED] = true;
|
|
handled_p[TS_COMMON] = true;
|
|
handled_p[TS_INT_CST] = true;
|
|
handled_p[TS_REAL_CST] = true;
|
|
handled_p[TS_FIXED_CST] = true;
|
|
handled_p[TS_VECTOR] = true;
|
|
handled_p[TS_STRING] = true;
|
|
handled_p[TS_COMPLEX] = true;
|
|
handled_p[TS_IDENTIFIER] = true;
|
|
handled_p[TS_DECL_MINIMAL] = true;
|
|
handled_p[TS_DECL_COMMON] = true;
|
|
handled_p[TS_DECL_WRTL] = true;
|
|
handled_p[TS_DECL_NON_COMMON] = true;
|
|
handled_p[TS_DECL_WITH_VIS] = true;
|
|
handled_p[TS_FIELD_DECL] = true;
|
|
handled_p[TS_VAR_DECL] = true;
|
|
handled_p[TS_PARM_DECL] = true;
|
|
handled_p[TS_LABEL_DECL] = true;
|
|
handled_p[TS_RESULT_DECL] = true;
|
|
handled_p[TS_CONST_DECL] = true;
|
|
handled_p[TS_TYPE_DECL] = true;
|
|
handled_p[TS_FUNCTION_DECL] = true;
|
|
handled_p[TS_TYPE_COMMON] = true;
|
|
handled_p[TS_TYPE_WITH_LANG_SPECIFIC] = true;
|
|
handled_p[TS_TYPE_NON_COMMON] = true;
|
|
handled_p[TS_LIST] = true;
|
|
handled_p[TS_VEC] = true;
|
|
handled_p[TS_EXP] = true;
|
|
handled_p[TS_SSA_NAME] = true;
|
|
handled_p[TS_BLOCK] = true;
|
|
handled_p[TS_BINFO] = true;
|
|
handled_p[TS_STATEMENT_LIST] = true;
|
|
handled_p[TS_CONSTRUCTOR] = true;
|
|
handled_p[TS_OMP_CLAUSE] = true;
|
|
handled_p[TS_OPTIMIZATION] = true;
|
|
handled_p[TS_TARGET_OPTION] = true;
|
|
handled_p[TS_TRANSLATION_UNIT_DECL] = true;
|
|
|
|
/* Anything not marked above will trigger the following assertion.
|
|
If this assertion triggers, it means that there is a new TS_*
|
|
structure that should be handled by the streamer. */
|
|
for (i = 0; i < LAST_TS_ENUM; i++)
|
|
gcc_assert (handled_p[i]);
|
|
}
|
|
|
|
|
|
/* Helper for streamer_tree_cache_insert_1. Add T to CACHE->NODES at
|
|
slot IX. */
|
|
|
|
static void
|
|
streamer_tree_cache_add_to_node_array (struct streamer_tree_cache_d *cache,
|
|
unsigned ix, tree t, hashval_t hash)
|
|
{
|
|
/* We're either replacing an old element or appending consecutively. */
|
|
if (cache->nodes.exists ())
|
|
{
|
|
if (cache->nodes.length () == ix)
|
|
cache->nodes.safe_push (t);
|
|
else
|
|
cache->nodes[ix] = t;
|
|
}
|
|
if (cache->hashes.exists ())
|
|
{
|
|
if (cache->hashes.length () == ix)
|
|
cache->hashes.safe_push (hash);
|
|
else
|
|
cache->hashes[ix] = hash;
|
|
}
|
|
}
|
|
|
|
|
|
/* Helper for streamer_tree_cache_insert and streamer_tree_cache_insert_at.
|
|
CACHE, T, and IX_P are as in streamer_tree_cache_insert.
|
|
|
|
If INSERT_AT_NEXT_SLOT_P is true, T is inserted at the next available
|
|
slot in the cache. Otherwise, T is inserted at the position indicated
|
|
in *IX_P.
|
|
|
|
If T already existed in CACHE, return true. Otherwise,
|
|
return false. */
|
|
|
|
static bool
|
|
streamer_tree_cache_insert_1 (struct streamer_tree_cache_d *cache,
|
|
tree t, hashval_t hash, unsigned *ix_p,
|
|
bool insert_at_next_slot_p)
|
|
{
|
|
bool existed_p;
|
|
|
|
gcc_assert (t);
|
|
|
|
unsigned int &ix = cache->node_map->get_or_insert (t, &existed_p);
|
|
if (!existed_p)
|
|
{
|
|
/* Determine the next slot to use in the cache. */
|
|
if (insert_at_next_slot_p)
|
|
ix = cache->next_idx++;
|
|
else
|
|
ix = *ix_p;
|
|
|
|
streamer_tree_cache_add_to_node_array (cache, ix, t, hash);
|
|
}
|
|
else
|
|
{
|
|
if (!insert_at_next_slot_p && ix != *ix_p)
|
|
{
|
|
/* If the caller wants to insert T at a specific slot
|
|
location, and ENTRY->TO does not match *IX_P, add T to
|
|
the requested location slot. */
|
|
ix = *ix_p;
|
|
streamer_tree_cache_add_to_node_array (cache, ix, t, hash);
|
|
}
|
|
}
|
|
|
|
if (ix_p)
|
|
*ix_p = ix;
|
|
|
|
return existed_p;
|
|
}
|
|
|
|
|
|
/* Insert tree node T in CACHE. If T already existed in the cache
|
|
return true. Otherwise, return false.
|
|
|
|
If IX_P is non-null, update it with the index into the cache where
|
|
T has been stored. */
|
|
|
|
bool
|
|
streamer_tree_cache_insert (struct streamer_tree_cache_d *cache, tree t,
|
|
hashval_t hash, unsigned *ix_p)
|
|
{
|
|
return streamer_tree_cache_insert_1 (cache, t, hash, ix_p, true);
|
|
}
|
|
|
|
|
|
/* Replace the tree node with T in CACHE at slot IX. */
|
|
|
|
void
|
|
streamer_tree_cache_replace_tree (struct streamer_tree_cache_d *cache,
|
|
tree t, unsigned ix)
|
|
{
|
|
hashval_t hash = 0;
|
|
if (cache->hashes.exists ())
|
|
hash = streamer_tree_cache_get_hash (cache, ix);
|
|
if (!cache->node_map)
|
|
streamer_tree_cache_add_to_node_array (cache, ix, t, hash);
|
|
else
|
|
streamer_tree_cache_insert_1 (cache, t, hash, &ix, false);
|
|
}
|
|
|
|
|
|
/* Appends tree node T to CACHE, even if T already existed in it. */
|
|
|
|
void
|
|
streamer_tree_cache_append (struct streamer_tree_cache_d *cache,
|
|
tree t, hashval_t hash)
|
|
{
|
|
unsigned ix = cache->next_idx++;
|
|
if (!cache->node_map)
|
|
streamer_tree_cache_add_to_node_array (cache, ix, t, hash);
|
|
else
|
|
streamer_tree_cache_insert_1 (cache, t, hash, &ix, false);
|
|
}
|
|
|
|
/* Return true if tree node T exists in CACHE, otherwise false. If IX_P is
|
|
not NULL, write to *IX_P the index into the cache where T is stored
|
|
((unsigned)-1 if T is not found). */
|
|
|
|
bool
|
|
streamer_tree_cache_lookup (struct streamer_tree_cache_d *cache, tree t,
|
|
unsigned *ix_p)
|
|
{
|
|
unsigned *slot;
|
|
bool retval;
|
|
unsigned ix;
|
|
|
|
gcc_assert (t);
|
|
|
|
slot = cache->node_map->get (t);
|
|
if (slot == NULL)
|
|
{
|
|
retval = false;
|
|
ix = -1;
|
|
}
|
|
else
|
|
{
|
|
retval = true;
|
|
ix = *slot;
|
|
}
|
|
|
|
if (ix_p)
|
|
*ix_p = ix;
|
|
|
|
return retval;
|
|
}
|
|
|
|
|
|
/* Verify that NODE is in CACHE. */
|
|
|
|
static void
|
|
verify_common_node_recorded (struct streamer_tree_cache_d *cache, tree node)
|
|
{
|
|
/* Restrict this to flag_checking only because in general violating it is
|
|
harmless plus we never know what happens on all targets/frontend/flag(!)
|
|
combinations. */
|
|
if (!flag_checking)
|
|
return;
|
|
|
|
if (cache->node_map)
|
|
gcc_assert (streamer_tree_cache_lookup (cache, node, NULL));
|
|
else
|
|
{
|
|
bool found = false;
|
|
gcc_assert (cache->nodes.exists ());
|
|
/* Linear search... */
|
|
for (unsigned i = 0; !found && i < cache->nodes.length (); ++i)
|
|
if (cache->nodes[i] == node)
|
|
found = true;
|
|
gcc_assert (found);
|
|
}
|
|
}
|
|
|
|
|
|
/* Record NODE in CACHE. */
|
|
|
|
static void
|
|
record_common_node (struct streamer_tree_cache_d *cache, tree node)
|
|
{
|
|
/* If we recursively end up at nodes we do not want to preload simply don't.
|
|
??? We'd want to verify that this doesn't happen, or alternatively
|
|
do not recurse at all. */
|
|
if (node == char_type_node)
|
|
return;
|
|
|
|
gcc_checking_assert (node != boolean_type_node
|
|
&& node != boolean_true_node
|
|
&& node != boolean_false_node);
|
|
|
|
/* We have to make sure to fill exactly the same number of
|
|
elements for all frontends. That can include NULL trees.
|
|
As our hash table can't deal with zero entries we'll simply stream
|
|
a random other tree. A NULL tree never will be looked up so it
|
|
doesn't matter which tree we replace it with, just to be sure
|
|
use error_mark_node. */
|
|
if (!node)
|
|
node = error_mark_node;
|
|
|
|
/* ??? FIXME, devise a better hash value. But the hash needs to be equal
|
|
for all frontend and lto1 invocations. So just use the position
|
|
in the cache as hash value. */
|
|
streamer_tree_cache_append (cache, node, cache->nodes.length ());
|
|
|
|
switch (TREE_CODE (node))
|
|
{
|
|
case ERROR_MARK:
|
|
case FIELD_DECL:
|
|
case FIXED_POINT_TYPE:
|
|
case IDENTIFIER_NODE:
|
|
case INTEGER_CST:
|
|
case INTEGER_TYPE:
|
|
case POINTER_BOUNDS_TYPE:
|
|
case REAL_TYPE:
|
|
case TREE_LIST:
|
|
case VOID_CST:
|
|
case VOID_TYPE:
|
|
/* No recursive trees. */
|
|
break;
|
|
case ARRAY_TYPE:
|
|
case POINTER_TYPE:
|
|
case REFERENCE_TYPE:
|
|
record_common_node (cache, TREE_TYPE (node));
|
|
break;
|
|
case COMPLEX_TYPE:
|
|
/* Verify that a complex type's component type (node_type) has been
|
|
handled already (and we thus don't need to recurse here). */
|
|
verify_common_node_recorded (cache, TREE_TYPE (node));
|
|
break;
|
|
case RECORD_TYPE:
|
|
/* The FIELD_DECLs of structures should be shared, so that every
|
|
COMPONENT_REF uses the same tree node when referencing a field.
|
|
Pointer equality between FIELD_DECLs is used by the alias
|
|
machinery to compute overlapping component references (see
|
|
nonoverlapping_component_refs_p and
|
|
nonoverlapping_component_refs_of_decl_p). */
|
|
for (tree f = TYPE_FIELDS (node); f; f = TREE_CHAIN (f))
|
|
record_common_node (cache, f);
|
|
break;
|
|
default:
|
|
/* Unexpected tree code. */
|
|
gcc_unreachable ();
|
|
}
|
|
}
|
|
|
|
|
|
/* Preload common nodes into CACHE and make sure they are merged
|
|
properly according to the gimple type table. */
|
|
|
|
static void
|
|
preload_common_nodes (struct streamer_tree_cache_d *cache)
|
|
{
|
|
unsigned i;
|
|
|
|
for (i = 0; i < itk_none; i++)
|
|
/* Skip itk_char. char_type_node is dependent on -f[un]signed-char. */
|
|
if (i != itk_char)
|
|
record_common_node (cache, integer_types[i]);
|
|
|
|
for (i = 0; i < stk_type_kind_last; i++)
|
|
record_common_node (cache, sizetype_tab[i]);
|
|
|
|
for (i = 0; i < TI_MAX; i++)
|
|
/* Skip boolean type and constants, they are frontend dependent. */
|
|
if (i != TI_BOOLEAN_TYPE
|
|
&& i != TI_BOOLEAN_FALSE
|
|
&& i != TI_BOOLEAN_TRUE
|
|
/* MAIN_IDENTIFIER is not always initialized by Fortran FE. */
|
|
&& i != TI_MAIN_IDENTIFIER
|
|
/* PID_TYPE is initialized only by C family front-ends. */
|
|
&& i != TI_PID_TYPE
|
|
/* Skip optimization and target option nodes; they depend on flags. */
|
|
&& i != TI_OPTIMIZATION_DEFAULT
|
|
&& i != TI_OPTIMIZATION_CURRENT
|
|
&& i != TI_TARGET_OPTION_DEFAULT
|
|
&& i != TI_TARGET_OPTION_CURRENT
|
|
&& i != TI_CURRENT_TARGET_PRAGMA
|
|
&& i != TI_CURRENT_OPTIMIZE_PRAGMA
|
|
/* Skip va_list* related nodes if offloading. For native LTO
|
|
we want them to be merged for the stdarg pass, for offloading
|
|
they might not be identical between host and offloading target. */
|
|
&& (!lto_stream_offload_p
|
|
|| (i != TI_VA_LIST_TYPE
|
|
&& i != TI_VA_LIST_GPR_COUNTER_FIELD
|
|
&& i != TI_VA_LIST_FPR_COUNTER_FIELD)))
|
|
record_common_node (cache, global_trees[i]);
|
|
}
|
|
|
|
|
|
/* Create a cache of pickled nodes. */
|
|
|
|
struct streamer_tree_cache_d *
|
|
streamer_tree_cache_create (bool with_hashes, bool with_map, bool with_vec)
|
|
{
|
|
struct streamer_tree_cache_d *cache;
|
|
|
|
cache = XCNEW (struct streamer_tree_cache_d);
|
|
|
|
if (with_map)
|
|
cache->node_map = new hash_map<tree, unsigned> (251);
|
|
cache->next_idx = 0;
|
|
if (with_vec)
|
|
cache->nodes.create (165);
|
|
if (with_hashes)
|
|
cache->hashes.create (165);
|
|
|
|
/* Load all the well-known tree nodes that are always created by
|
|
the compiler on startup. This prevents writing them out
|
|
unnecessarily. */
|
|
preload_common_nodes (cache);
|
|
|
|
return cache;
|
|
}
|
|
|
|
|
|
/* Delete the streamer cache C. */
|
|
|
|
void
|
|
streamer_tree_cache_delete (struct streamer_tree_cache_d *c)
|
|
{
|
|
if (c == NULL)
|
|
return;
|
|
|
|
delete c->node_map;
|
|
c->node_map = NULL;
|
|
c->nodes.release ();
|
|
c->hashes.release ();
|
|
free (c);
|
|
}
|