e287a2a11d
As discussed in the PR, the problem here is that the routines changed in this patch sign extend the difference of index and low_bound from the precision of the index, so e.g. when index is unsigned int and contains value -2U, we treat it as index -2 rather than 0x00000000fffffffeU on 64-bit arches. On the other hand, get_inner_reference which is used during expansion, does: if (! integer_zerop (low_bound)) index = fold_build2 (MINUS_EXPR, TREE_TYPE (index), index, low_bound); offset = size_binop (PLUS_EXPR, offset, size_binop (MULT_EXPR, fold_convert (sizetype, index), unit_size)); which effectively requires that either low_bound is constant 0 and then index in ARRAY_REFs can be arbitrary type which is then sign or zero extended to sizetype, or low_bound is something else and then index and low_bound must have compatible types and it is still converted afterwards to sizetype and from there then a few lines later: expr.c- if (poly_int_tree_p (offset)) expr.c- { expr.c: poly_offset_int tem = wi::sext (wi::to_poly_offset (offset), expr.c- TYPE_PRECISION (sizetype)); The following patch makes those routines match what get_inner_reference is doing. 2021-01-22 Jakub Jelinek <jakub@redhat.com> PR tree-optimization/98255 * tree-dfa.c (get_ref_base_and_extent): For ARRAY_REFs, sign extend index - low_bound from sizetype's precision rather than index precision. (get_addr_base_and_unit_offset_1): Likewise. * tree-ssa-sccvn.c (ao_ref_init_from_vn_reference): Likewise. * gimple-fold.c (fold_const_aggregate_ref_1): Likewise. * gcc.dg/pr98255.c: New test.
1034 lines
29 KiB
C
1034 lines
29 KiB
C
/* Data flow functions for trees.
|
|
Copyright (C) 2001-2021 Free Software Foundation, Inc.
|
|
Contributed by Diego Novillo <dnovillo@redhat.com>
|
|
|
|
This file is part of GCC.
|
|
|
|
GCC is free software; you can redistribute it and/or modify
|
|
it under the terms of the GNU General Public License as published by
|
|
the Free Software Foundation; either version 3, or (at your option)
|
|
any later version.
|
|
|
|
GCC is distributed in the hope that it will be useful,
|
|
but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
GNU General Public License for more details.
|
|
|
|
You should have received a copy of the GNU General Public License
|
|
along with GCC; see the file COPYING3. If not see
|
|
<http://www.gnu.org/licenses/>. */
|
|
|
|
#include "config.h"
|
|
#include "system.h"
|
|
#include "coretypes.h"
|
|
#include "backend.h"
|
|
#include "rtl.h"
|
|
#include "tree.h"
|
|
#include "gimple.h"
|
|
#include "tree-pass.h"
|
|
#include "ssa.h"
|
|
#include "tree-pretty-print.h"
|
|
#include "fold-const.h"
|
|
#include "stor-layout.h"
|
|
#include "langhooks.h"
|
|
#include "gimple-iterator.h"
|
|
#include "gimple-walk.h"
|
|
#include "tree-dfa.h"
|
|
|
|
/* Build and maintain data flow information for trees. */
|
|
|
|
/* Counters used to display DFA and SSA statistics. */
|
|
struct dfa_stats_d
|
|
{
|
|
long num_defs;
|
|
long num_uses;
|
|
long num_phis;
|
|
long num_phi_args;
|
|
size_t max_num_phi_args;
|
|
long num_vdefs;
|
|
long num_vuses;
|
|
};
|
|
|
|
|
|
/* Local functions. */
|
|
static void collect_dfa_stats (struct dfa_stats_d *);
|
|
|
|
|
|
/*---------------------------------------------------------------------------
|
|
Dataflow analysis (DFA) routines
|
|
---------------------------------------------------------------------------*/
|
|
|
|
/* Renumber all of the gimple stmt uids. */
|
|
|
|
void
|
|
renumber_gimple_stmt_uids (struct function *fun)
|
|
{
|
|
basic_block bb;
|
|
|
|
set_gimple_stmt_max_uid (fun, 0);
|
|
FOR_ALL_BB_FN (bb, fun)
|
|
{
|
|
gimple_stmt_iterator bsi;
|
|
for (bsi = gsi_start_phis (bb); !gsi_end_p (bsi); gsi_next (&bsi))
|
|
{
|
|
gimple *stmt = gsi_stmt (bsi);
|
|
gimple_set_uid (stmt, inc_gimple_stmt_max_uid (fun));
|
|
}
|
|
for (bsi = gsi_start_bb (bb); !gsi_end_p (bsi); gsi_next (&bsi))
|
|
{
|
|
gimple *stmt = gsi_stmt (bsi);
|
|
gimple_set_uid (stmt, inc_gimple_stmt_max_uid (fun));
|
|
}
|
|
}
|
|
}
|
|
|
|
/* Like renumber_gimple_stmt_uids, but only do work on the basic blocks
|
|
in BLOCKS, of which there are N_BLOCKS. Also renumbers PHIs. */
|
|
|
|
void
|
|
renumber_gimple_stmt_uids_in_blocks (basic_block *blocks, int n_blocks)
|
|
{
|
|
int i;
|
|
|
|
set_gimple_stmt_max_uid (cfun, 0);
|
|
for (i = 0; i < n_blocks; i++)
|
|
{
|
|
basic_block bb = blocks[i];
|
|
gimple_stmt_iterator bsi;
|
|
for (bsi = gsi_start_phis (bb); !gsi_end_p (bsi); gsi_next (&bsi))
|
|
{
|
|
gimple *stmt = gsi_stmt (bsi);
|
|
gimple_set_uid (stmt, inc_gimple_stmt_max_uid (cfun));
|
|
}
|
|
for (bsi = gsi_start_bb (bb); !gsi_end_p (bsi); gsi_next (&bsi))
|
|
{
|
|
gimple *stmt = gsi_stmt (bsi);
|
|
gimple_set_uid (stmt, inc_gimple_stmt_max_uid (cfun));
|
|
}
|
|
}
|
|
}
|
|
|
|
|
|
|
|
/*---------------------------------------------------------------------------
|
|
Debugging functions
|
|
---------------------------------------------------------------------------*/
|
|
|
|
/* Dump variable VAR and its may-aliases to FILE. */
|
|
|
|
void
|
|
dump_variable (FILE *file, tree var)
|
|
{
|
|
if (TREE_CODE (var) == SSA_NAME)
|
|
{
|
|
if (POINTER_TYPE_P (TREE_TYPE (var)))
|
|
dump_points_to_info_for (file, var);
|
|
var = SSA_NAME_VAR (var);
|
|
}
|
|
|
|
if (var == NULL_TREE)
|
|
{
|
|
fprintf (file, "<nil>");
|
|
return;
|
|
}
|
|
|
|
print_generic_expr (file, var, dump_flags);
|
|
|
|
fprintf (file, ", UID D.%u", (unsigned) DECL_UID (var));
|
|
if (DECL_PT_UID (var) != DECL_UID (var))
|
|
fprintf (file, ", PT-UID D.%u", (unsigned) DECL_PT_UID (var));
|
|
|
|
fprintf (file, ", ");
|
|
print_generic_expr (file, TREE_TYPE (var), dump_flags);
|
|
|
|
if (TREE_ADDRESSABLE (var))
|
|
fprintf (file, ", is addressable");
|
|
|
|
if (is_global_var (var))
|
|
fprintf (file, ", is global");
|
|
|
|
if (TREE_THIS_VOLATILE (var))
|
|
fprintf (file, ", is volatile");
|
|
|
|
if (cfun && ssa_default_def (cfun, var))
|
|
{
|
|
fprintf (file, ", default def: ");
|
|
print_generic_expr (file, ssa_default_def (cfun, var), dump_flags);
|
|
}
|
|
|
|
if (DECL_INITIAL (var))
|
|
{
|
|
fprintf (file, ", initial: ");
|
|
print_generic_expr (file, DECL_INITIAL (var), dump_flags);
|
|
}
|
|
|
|
fprintf (file, "\n");
|
|
}
|
|
|
|
|
|
/* Dump variable VAR and its may-aliases to stderr. */
|
|
|
|
DEBUG_FUNCTION void
|
|
debug_variable (tree var)
|
|
{
|
|
dump_variable (stderr, var);
|
|
}
|
|
|
|
|
|
/* Dump various DFA statistics to FILE. */
|
|
|
|
void
|
|
dump_dfa_stats (FILE *file)
|
|
{
|
|
struct dfa_stats_d dfa_stats;
|
|
|
|
unsigned long size, total = 0;
|
|
const char * const fmt_str = "%-30s%-13s%12s\n";
|
|
const char * const fmt_str_1 = "%-30s%13lu" PRsa (11) "\n";
|
|
const char * const fmt_str_3 = "%-43s" PRsa (11) "\n";
|
|
const char *funcname
|
|
= lang_hooks.decl_printable_name (current_function_decl, 2);
|
|
|
|
collect_dfa_stats (&dfa_stats);
|
|
|
|
fprintf (file, "\nDFA Statistics for %s\n\n", funcname);
|
|
|
|
fprintf (file, "---------------------------------------------------------\n");
|
|
fprintf (file, fmt_str, "", " Number of ", "Memory");
|
|
fprintf (file, fmt_str, "", " instances ", "used ");
|
|
fprintf (file, "---------------------------------------------------------\n");
|
|
|
|
size = dfa_stats.num_uses * sizeof (tree *);
|
|
total += size;
|
|
fprintf (file, fmt_str_1, "USE operands", dfa_stats.num_uses,
|
|
SIZE_AMOUNT (size));
|
|
|
|
size = dfa_stats.num_defs * sizeof (tree *);
|
|
total += size;
|
|
fprintf (file, fmt_str_1, "DEF operands", dfa_stats.num_defs,
|
|
SIZE_AMOUNT (size));
|
|
|
|
size = dfa_stats.num_vuses * sizeof (tree *);
|
|
total += size;
|
|
fprintf (file, fmt_str_1, "VUSE operands", dfa_stats.num_vuses,
|
|
SIZE_AMOUNT (size));
|
|
|
|
size = dfa_stats.num_vdefs * sizeof (tree *);
|
|
total += size;
|
|
fprintf (file, fmt_str_1, "VDEF operands", dfa_stats.num_vdefs,
|
|
SIZE_AMOUNT (size));
|
|
|
|
size = dfa_stats.num_phis * sizeof (struct gphi);
|
|
total += size;
|
|
fprintf (file, fmt_str_1, "PHI nodes", dfa_stats.num_phis,
|
|
SIZE_AMOUNT (size));
|
|
|
|
size = dfa_stats.num_phi_args * sizeof (struct phi_arg_d);
|
|
total += size;
|
|
fprintf (file, fmt_str_1, "PHI arguments", dfa_stats.num_phi_args,
|
|
SIZE_AMOUNT (size));
|
|
|
|
fprintf (file, "---------------------------------------------------------\n");
|
|
fprintf (file, fmt_str_3, "Total memory used by DFA/SSA data",
|
|
SIZE_AMOUNT (total));
|
|
fprintf (file, "---------------------------------------------------------\n");
|
|
fprintf (file, "\n");
|
|
|
|
if (dfa_stats.num_phis)
|
|
fprintf (file, "Average number of arguments per PHI node: %.1f (max: %ld)\n",
|
|
(float) dfa_stats.num_phi_args / (float) dfa_stats.num_phis,
|
|
(long) dfa_stats.max_num_phi_args);
|
|
|
|
fprintf (file, "\n");
|
|
}
|
|
|
|
|
|
/* Dump DFA statistics on stderr. */
|
|
|
|
DEBUG_FUNCTION void
|
|
debug_dfa_stats (void)
|
|
{
|
|
dump_dfa_stats (stderr);
|
|
}
|
|
|
|
|
|
/* Collect DFA statistics and store them in the structure pointed to by
|
|
DFA_STATS_P. */
|
|
|
|
static void
|
|
collect_dfa_stats (struct dfa_stats_d *dfa_stats_p ATTRIBUTE_UNUSED)
|
|
{
|
|
basic_block bb;
|
|
|
|
gcc_assert (dfa_stats_p);
|
|
|
|
memset ((void *)dfa_stats_p, 0, sizeof (struct dfa_stats_d));
|
|
|
|
/* Walk all the statements in the function counting references. */
|
|
FOR_EACH_BB_FN (bb, cfun)
|
|
{
|
|
for (gphi_iterator si = gsi_start_phis (bb); !gsi_end_p (si);
|
|
gsi_next (&si))
|
|
{
|
|
gphi *phi = si.phi ();
|
|
dfa_stats_p->num_phis++;
|
|
dfa_stats_p->num_phi_args += gimple_phi_num_args (phi);
|
|
if (gimple_phi_num_args (phi) > dfa_stats_p->max_num_phi_args)
|
|
dfa_stats_p->max_num_phi_args = gimple_phi_num_args (phi);
|
|
}
|
|
|
|
for (gimple_stmt_iterator si = gsi_start_bb (bb); !gsi_end_p (si);
|
|
gsi_next (&si))
|
|
{
|
|
gimple *stmt = gsi_stmt (si);
|
|
dfa_stats_p->num_defs += NUM_SSA_OPERANDS (stmt, SSA_OP_DEF);
|
|
dfa_stats_p->num_uses += NUM_SSA_OPERANDS (stmt, SSA_OP_USE);
|
|
dfa_stats_p->num_vdefs += gimple_vdef (stmt) ? 1 : 0;
|
|
dfa_stats_p->num_vuses += gimple_vuse (stmt) ? 1 : 0;
|
|
}
|
|
}
|
|
}
|
|
|
|
|
|
/*---------------------------------------------------------------------------
|
|
Miscellaneous helpers
|
|
---------------------------------------------------------------------------*/
|
|
|
|
/* Lookup VAR UID in the default_defs hashtable and return the associated
|
|
variable. */
|
|
|
|
tree
|
|
ssa_default_def (struct function *fn, tree var)
|
|
{
|
|
struct tree_decl_minimal ind;
|
|
struct tree_ssa_name in;
|
|
gcc_assert (VAR_P (var)
|
|
|| TREE_CODE (var) == PARM_DECL
|
|
|| TREE_CODE (var) == RESULT_DECL);
|
|
|
|
/* Always NULL_TREE for rtl function dumps. */
|
|
if (!fn->gimple_df)
|
|
return NULL_TREE;
|
|
|
|
in.var = (tree)&ind;
|
|
ind.uid = DECL_UID (var);
|
|
return DEFAULT_DEFS (fn)->find_with_hash ((tree)&in, DECL_UID (var));
|
|
}
|
|
|
|
/* Insert the pair VAR's UID, DEF into the default_defs hashtable
|
|
of function FN. */
|
|
|
|
void
|
|
set_ssa_default_def (struct function *fn, tree var, tree def)
|
|
{
|
|
struct tree_decl_minimal ind;
|
|
struct tree_ssa_name in;
|
|
|
|
gcc_assert (VAR_P (var)
|
|
|| TREE_CODE (var) == PARM_DECL
|
|
|| TREE_CODE (var) == RESULT_DECL);
|
|
in.var = (tree)&ind;
|
|
ind.uid = DECL_UID (var);
|
|
if (!def)
|
|
{
|
|
tree *loc = DEFAULT_DEFS (fn)->find_slot_with_hash ((tree)&in,
|
|
DECL_UID (var),
|
|
NO_INSERT);
|
|
if (loc)
|
|
{
|
|
SSA_NAME_IS_DEFAULT_DEF (*(tree *)loc) = false;
|
|
DEFAULT_DEFS (fn)->clear_slot (loc);
|
|
}
|
|
return;
|
|
}
|
|
gcc_assert (TREE_CODE (def) == SSA_NAME && SSA_NAME_VAR (def) == var);
|
|
tree *loc = DEFAULT_DEFS (fn)->find_slot_with_hash ((tree)&in,
|
|
DECL_UID (var), INSERT);
|
|
|
|
/* Default definition might be changed by tail call optimization. */
|
|
if (*loc)
|
|
SSA_NAME_IS_DEFAULT_DEF (*loc) = false;
|
|
|
|
/* Mark DEF as the default definition for VAR. */
|
|
*loc = def;
|
|
SSA_NAME_IS_DEFAULT_DEF (def) = true;
|
|
}
|
|
|
|
/* Retrieve or create a default definition for VAR. */
|
|
|
|
tree
|
|
get_or_create_ssa_default_def (struct function *fn, tree var)
|
|
{
|
|
tree ddef = ssa_default_def (fn, var);
|
|
if (ddef == NULL_TREE)
|
|
{
|
|
ddef = make_ssa_name_fn (fn, var, gimple_build_nop ());
|
|
set_ssa_default_def (fn, var, ddef);
|
|
}
|
|
return ddef;
|
|
}
|
|
|
|
|
|
/* If EXP is a handled component reference for a structure, return the
|
|
base variable. The access range is delimited by bit positions *POFFSET and
|
|
*POFFSET + *PMAX_SIZE. The access size is *PSIZE bits. If either
|
|
*PSIZE or *PMAX_SIZE is -1, they could not be determined. If *PSIZE
|
|
and *PMAX_SIZE are equal, the access is non-variable. If *PREVERSE is
|
|
true, the storage order of the reference is reversed. */
|
|
|
|
tree
|
|
get_ref_base_and_extent (tree exp, poly_int64_pod *poffset,
|
|
poly_int64_pod *psize,
|
|
poly_int64_pod *pmax_size,
|
|
bool *preverse)
|
|
{
|
|
poly_offset_int bitsize = -1;
|
|
poly_offset_int maxsize;
|
|
tree size_tree = NULL_TREE;
|
|
poly_offset_int bit_offset = 0;
|
|
bool seen_variable_array_ref = false;
|
|
|
|
/* First get the final access size and the storage order from just the
|
|
outermost expression. */
|
|
if (TREE_CODE (exp) == COMPONENT_REF)
|
|
size_tree = DECL_SIZE (TREE_OPERAND (exp, 1));
|
|
else if (TREE_CODE (exp) == BIT_FIELD_REF)
|
|
size_tree = TREE_OPERAND (exp, 1);
|
|
else if (!VOID_TYPE_P (TREE_TYPE (exp)))
|
|
{
|
|
machine_mode mode = TYPE_MODE (TREE_TYPE (exp));
|
|
if (mode == BLKmode)
|
|
size_tree = TYPE_SIZE (TREE_TYPE (exp));
|
|
else
|
|
bitsize = GET_MODE_BITSIZE (mode);
|
|
}
|
|
if (size_tree != NULL_TREE
|
|
&& poly_int_tree_p (size_tree))
|
|
bitsize = wi::to_poly_offset (size_tree);
|
|
|
|
*preverse = reverse_storage_order_for_component_p (exp);
|
|
|
|
/* Initially, maxsize is the same as the accessed element size.
|
|
In the following it will only grow (or become -1). */
|
|
maxsize = bitsize;
|
|
|
|
/* Compute cumulative bit-offset for nested component-refs and array-refs,
|
|
and find the ultimate containing object. */
|
|
while (1)
|
|
{
|
|
switch (TREE_CODE (exp))
|
|
{
|
|
case BIT_FIELD_REF:
|
|
bit_offset += wi::to_poly_offset (TREE_OPERAND (exp, 2));
|
|
break;
|
|
|
|
case COMPONENT_REF:
|
|
{
|
|
tree field = TREE_OPERAND (exp, 1);
|
|
tree this_offset = component_ref_field_offset (exp);
|
|
|
|
if (this_offset && poly_int_tree_p (this_offset))
|
|
{
|
|
poly_offset_int woffset = (wi::to_poly_offset (this_offset)
|
|
<< LOG2_BITS_PER_UNIT);
|
|
woffset += wi::to_offset (DECL_FIELD_BIT_OFFSET (field));
|
|
bit_offset += woffset;
|
|
|
|
/* If we had seen a variable array ref already and we just
|
|
referenced the last field of a struct or a union member
|
|
then we have to adjust maxsize by the padding at the end
|
|
of our field. */
|
|
if (seen_variable_array_ref)
|
|
{
|
|
tree stype = TREE_TYPE (TREE_OPERAND (exp, 0));
|
|
tree next = DECL_CHAIN (field);
|
|
while (next && TREE_CODE (next) != FIELD_DECL)
|
|
next = DECL_CHAIN (next);
|
|
if (!next
|
|
|| TREE_CODE (stype) != RECORD_TYPE)
|
|
{
|
|
tree fsize = DECL_SIZE_UNIT (field);
|
|
tree ssize = TYPE_SIZE_UNIT (stype);
|
|
if (fsize == NULL
|
|
|| !poly_int_tree_p (fsize)
|
|
|| ssize == NULL
|
|
|| !poly_int_tree_p (ssize))
|
|
maxsize = -1;
|
|
else if (known_size_p (maxsize))
|
|
{
|
|
poly_offset_int tem
|
|
= (wi::to_poly_offset (ssize)
|
|
- wi::to_poly_offset (fsize));
|
|
tem <<= LOG2_BITS_PER_UNIT;
|
|
tem -= woffset;
|
|
maxsize += tem;
|
|
}
|
|
}
|
|
/* An component ref with an adjacent field up in the
|
|
structure hierarchy constrains the size of any variable
|
|
array ref lower in the access hierarchy. */
|
|
else
|
|
seen_variable_array_ref = false;
|
|
}
|
|
}
|
|
else
|
|
{
|
|
tree csize = TYPE_SIZE (TREE_TYPE (TREE_OPERAND (exp, 0)));
|
|
/* We need to adjust maxsize to the whole structure bitsize.
|
|
But we can subtract any constant offset seen so far,
|
|
because that would get us out of the structure otherwise. */
|
|
if (known_size_p (maxsize)
|
|
&& csize
|
|
&& poly_int_tree_p (csize))
|
|
maxsize = wi::to_poly_offset (csize) - bit_offset;
|
|
else
|
|
maxsize = -1;
|
|
}
|
|
}
|
|
break;
|
|
|
|
case ARRAY_REF:
|
|
case ARRAY_RANGE_REF:
|
|
{
|
|
tree index = TREE_OPERAND (exp, 1);
|
|
tree low_bound, unit_size;
|
|
|
|
/* If the resulting bit-offset is constant, track it. */
|
|
if (poly_int_tree_p (index)
|
|
&& (low_bound = array_ref_low_bound (exp),
|
|
poly_int_tree_p (low_bound))
|
|
&& (unit_size = array_ref_element_size (exp),
|
|
TREE_CODE (unit_size) == INTEGER_CST))
|
|
{
|
|
poly_offset_int woffset
|
|
= wi::sext (wi::to_poly_offset (index)
|
|
- wi::to_poly_offset (low_bound),
|
|
TYPE_PRECISION (sizetype));
|
|
woffset *= wi::to_offset (unit_size);
|
|
woffset <<= LOG2_BITS_PER_UNIT;
|
|
bit_offset += woffset;
|
|
|
|
/* An array ref with a constant index up in the structure
|
|
hierarchy will constrain the size of any variable array ref
|
|
lower in the access hierarchy. */
|
|
seen_variable_array_ref = false;
|
|
}
|
|
else
|
|
{
|
|
tree asize = TYPE_SIZE (TREE_TYPE (TREE_OPERAND (exp, 0)));
|
|
/* We need to adjust maxsize to the whole array bitsize.
|
|
But we can subtract any constant offset seen so far,
|
|
because that would get us outside of the array otherwise. */
|
|
if (known_size_p (maxsize)
|
|
&& asize
|
|
&& poly_int_tree_p (asize))
|
|
maxsize = wi::to_poly_offset (asize) - bit_offset;
|
|
else
|
|
maxsize = -1;
|
|
|
|
/* Remember that we have seen an array ref with a variable
|
|
index. */
|
|
seen_variable_array_ref = true;
|
|
|
|
wide_int min, max;
|
|
if (TREE_CODE (index) == SSA_NAME
|
|
&& (low_bound = array_ref_low_bound (exp),
|
|
poly_int_tree_p (low_bound))
|
|
&& (unit_size = array_ref_element_size (exp),
|
|
TREE_CODE (unit_size) == INTEGER_CST)
|
|
&& get_range_info (index, &min, &max) == VR_RANGE)
|
|
{
|
|
poly_offset_int lbound = wi::to_poly_offset (low_bound);
|
|
/* Try to constrain maxsize with range information. */
|
|
offset_int omax
|
|
= offset_int::from (max, TYPE_SIGN (TREE_TYPE (index)));
|
|
if (known_lt (lbound, omax))
|
|
{
|
|
poly_offset_int rmaxsize;
|
|
rmaxsize = (omax - lbound + 1)
|
|
* wi::to_offset (unit_size) << LOG2_BITS_PER_UNIT;
|
|
if (!known_size_p (maxsize)
|
|
|| known_lt (rmaxsize, maxsize))
|
|
{
|
|
/* If we know an upper bound below the declared
|
|
one this is no longer variable. */
|
|
if (known_size_p (maxsize))
|
|
seen_variable_array_ref = false;
|
|
maxsize = rmaxsize;
|
|
}
|
|
}
|
|
/* Try to adjust bit_offset with range information. */
|
|
offset_int omin
|
|
= offset_int::from (min, TYPE_SIGN (TREE_TYPE (index)));
|
|
if (known_le (lbound, omin))
|
|
{
|
|
poly_offset_int woffset
|
|
= wi::sext (omin - lbound,
|
|
TYPE_PRECISION (sizetype));
|
|
woffset *= wi::to_offset (unit_size);
|
|
woffset <<= LOG2_BITS_PER_UNIT;
|
|
bit_offset += woffset;
|
|
if (known_size_p (maxsize))
|
|
maxsize -= woffset;
|
|
}
|
|
}
|
|
}
|
|
}
|
|
break;
|
|
|
|
case REALPART_EXPR:
|
|
break;
|
|
|
|
case IMAGPART_EXPR:
|
|
bit_offset += bitsize;
|
|
break;
|
|
|
|
case VIEW_CONVERT_EXPR:
|
|
break;
|
|
|
|
case TARGET_MEM_REF:
|
|
/* Via the variable index or index2 we can reach the
|
|
whole object. Still hand back the decl here. */
|
|
if (TREE_CODE (TMR_BASE (exp)) == ADDR_EXPR
|
|
&& (TMR_INDEX (exp) || TMR_INDEX2 (exp)))
|
|
{
|
|
exp = TREE_OPERAND (TMR_BASE (exp), 0);
|
|
bit_offset = 0;
|
|
maxsize = -1;
|
|
goto done;
|
|
}
|
|
/* Fallthru. */
|
|
case MEM_REF:
|
|
/* We need to deal with variable arrays ending structures such as
|
|
struct { int length; int a[1]; } x; x.a[d]
|
|
struct { struct { int a; int b; } a[1]; } x; x.a[d].a
|
|
struct { struct { int a[1]; } a[1]; } x; x.a[0][d], x.a[d][0]
|
|
struct { int len; union { int a[1]; struct X x; } u; } x; x.u.a[d]
|
|
where we do not know maxsize for variable index accesses to
|
|
the array. The simplest way to conservatively deal with this
|
|
is to punt in the case that offset + maxsize reaches the
|
|
base type boundary. This needs to include possible trailing
|
|
padding that is there for alignment purposes. */
|
|
if (seen_variable_array_ref
|
|
&& known_size_p (maxsize)
|
|
&& (TYPE_SIZE (TREE_TYPE (exp)) == NULL_TREE
|
|
|| !poly_int_tree_p (TYPE_SIZE (TREE_TYPE (exp)))
|
|
|| (maybe_eq
|
|
(bit_offset + maxsize,
|
|
wi::to_poly_offset (TYPE_SIZE (TREE_TYPE (exp)))))))
|
|
maxsize = -1;
|
|
|
|
/* Hand back the decl for MEM[&decl, off]. */
|
|
if (TREE_CODE (TREE_OPERAND (exp, 0)) == ADDR_EXPR)
|
|
{
|
|
if (integer_zerop (TREE_OPERAND (exp, 1)))
|
|
exp = TREE_OPERAND (TREE_OPERAND (exp, 0), 0);
|
|
else
|
|
{
|
|
poly_offset_int off = mem_ref_offset (exp);
|
|
off <<= LOG2_BITS_PER_UNIT;
|
|
off += bit_offset;
|
|
poly_int64 off_hwi;
|
|
if (off.to_shwi (&off_hwi))
|
|
{
|
|
bit_offset = off_hwi;
|
|
exp = TREE_OPERAND (TREE_OPERAND (exp, 0), 0);
|
|
}
|
|
}
|
|
}
|
|
goto done;
|
|
|
|
default:
|
|
goto done;
|
|
}
|
|
|
|
exp = TREE_OPERAND (exp, 0);
|
|
}
|
|
|
|
done:
|
|
if (!bitsize.to_shwi (psize) || maybe_lt (*psize, 0))
|
|
{
|
|
*poffset = 0;
|
|
*psize = -1;
|
|
*pmax_size = -1;
|
|
|
|
return exp;
|
|
}
|
|
|
|
/* ??? Due to negative offsets in ARRAY_REF we can end up with
|
|
negative bit_offset here. We might want to store a zero offset
|
|
in this case. */
|
|
if (!bit_offset.to_shwi (poffset))
|
|
{
|
|
*poffset = 0;
|
|
*pmax_size = -1;
|
|
|
|
return exp;
|
|
}
|
|
|
|
/* In case of a decl or constant base object we can do better. */
|
|
|
|
if (DECL_P (exp))
|
|
{
|
|
if (VAR_P (exp)
|
|
&& ((flag_unconstrained_commons && DECL_COMMON (exp))
|
|
|| (DECL_EXTERNAL (exp) && seen_variable_array_ref)))
|
|
{
|
|
tree sz_tree = TYPE_SIZE (TREE_TYPE (exp));
|
|
/* If size is unknown, or we have read to the end, assume there
|
|
may be more to the structure than we are told. */
|
|
if (TREE_CODE (TREE_TYPE (exp)) == ARRAY_TYPE
|
|
|| (seen_variable_array_ref
|
|
&& (sz_tree == NULL_TREE
|
|
|| !poly_int_tree_p (sz_tree)
|
|
|| maybe_eq (bit_offset + maxsize,
|
|
wi::to_poly_offset (sz_tree)))))
|
|
maxsize = -1;
|
|
}
|
|
/* If maxsize is unknown adjust it according to the size of the
|
|
base decl. */
|
|
else if (!known_size_p (maxsize)
|
|
&& DECL_SIZE (exp)
|
|
&& poly_int_tree_p (DECL_SIZE (exp)))
|
|
maxsize = wi::to_poly_offset (DECL_SIZE (exp)) - bit_offset;
|
|
}
|
|
else if (CONSTANT_CLASS_P (exp))
|
|
{
|
|
/* If maxsize is unknown adjust it according to the size of the
|
|
base type constant. */
|
|
if (!known_size_p (maxsize)
|
|
&& TYPE_SIZE (TREE_TYPE (exp))
|
|
&& poly_int_tree_p (TYPE_SIZE (TREE_TYPE (exp))))
|
|
maxsize = (wi::to_poly_offset (TYPE_SIZE (TREE_TYPE (exp)))
|
|
- bit_offset);
|
|
}
|
|
|
|
if (!maxsize.to_shwi (pmax_size)
|
|
|| maybe_lt (*pmax_size, 0)
|
|
|| !endpoint_representable_p (*poffset, *pmax_size))
|
|
*pmax_size = -1;
|
|
|
|
/* Punt if *POFFSET + *PSIZE overflows in HOST_WIDE_INT, the callers don't
|
|
check for such overflows individually and assume it works. */
|
|
if (!endpoint_representable_p (*poffset, *psize))
|
|
{
|
|
*poffset = 0;
|
|
*psize = -1;
|
|
*pmax_size = -1;
|
|
|
|
return exp;
|
|
}
|
|
|
|
return exp;
|
|
}
|
|
|
|
/* Like get_ref_base_and_extent, but for cases in which we only care
|
|
about constant-width accesses at constant offsets. Return null
|
|
if the access is anything else. */
|
|
|
|
tree
|
|
get_ref_base_and_extent_hwi (tree exp, HOST_WIDE_INT *poffset,
|
|
HOST_WIDE_INT *psize, bool *preverse)
|
|
{
|
|
poly_int64 offset, size, max_size;
|
|
HOST_WIDE_INT const_offset, const_size;
|
|
bool reverse;
|
|
tree decl = get_ref_base_and_extent (exp, &offset, &size, &max_size,
|
|
&reverse);
|
|
if (!offset.is_constant (&const_offset)
|
|
|| !size.is_constant (&const_size)
|
|
|| const_offset < 0
|
|
|| !known_size_p (max_size)
|
|
|| maybe_ne (max_size, const_size))
|
|
return NULL_TREE;
|
|
|
|
*poffset = const_offset;
|
|
*psize = const_size;
|
|
*preverse = reverse;
|
|
return decl;
|
|
}
|
|
|
|
/* Returns the base object and a constant BITS_PER_UNIT offset in *POFFSET that
|
|
denotes the starting address of the memory access EXP.
|
|
Returns NULL_TREE if the offset is not constant or any component
|
|
is not BITS_PER_UNIT-aligned.
|
|
VALUEIZE if non-NULL is used to valueize SSA names. It should return
|
|
its argument or a constant if the argument is known to be constant. */
|
|
|
|
tree
|
|
get_addr_base_and_unit_offset_1 (tree exp, poly_int64_pod *poffset,
|
|
tree (*valueize) (tree))
|
|
{
|
|
poly_int64 byte_offset = 0;
|
|
|
|
/* Compute cumulative byte-offset for nested component-refs and array-refs,
|
|
and find the ultimate containing object. */
|
|
while (1)
|
|
{
|
|
switch (TREE_CODE (exp))
|
|
{
|
|
case BIT_FIELD_REF:
|
|
{
|
|
poly_int64 this_byte_offset;
|
|
poly_uint64 this_bit_offset;
|
|
if (!poly_int_tree_p (TREE_OPERAND (exp, 2), &this_bit_offset)
|
|
|| !multiple_p (this_bit_offset, BITS_PER_UNIT,
|
|
&this_byte_offset))
|
|
return NULL_TREE;
|
|
byte_offset += this_byte_offset;
|
|
}
|
|
break;
|
|
|
|
case COMPONENT_REF:
|
|
{
|
|
tree field = TREE_OPERAND (exp, 1);
|
|
tree this_offset = component_ref_field_offset (exp);
|
|
poly_int64 hthis_offset;
|
|
|
|
if (!this_offset
|
|
|| !poly_int_tree_p (this_offset, &hthis_offset)
|
|
|| (TREE_INT_CST_LOW (DECL_FIELD_BIT_OFFSET (field))
|
|
% BITS_PER_UNIT))
|
|
return NULL_TREE;
|
|
|
|
hthis_offset += (TREE_INT_CST_LOW (DECL_FIELD_BIT_OFFSET (field))
|
|
/ BITS_PER_UNIT);
|
|
byte_offset += hthis_offset;
|
|
}
|
|
break;
|
|
|
|
case ARRAY_REF:
|
|
case ARRAY_RANGE_REF:
|
|
{
|
|
tree index = TREE_OPERAND (exp, 1);
|
|
tree low_bound, unit_size;
|
|
|
|
if (valueize
|
|
&& TREE_CODE (index) == SSA_NAME)
|
|
index = (*valueize) (index);
|
|
if (!poly_int_tree_p (index))
|
|
return NULL_TREE;
|
|
low_bound = array_ref_low_bound (exp);
|
|
if (valueize
|
|
&& TREE_CODE (low_bound) == SSA_NAME)
|
|
low_bound = (*valueize) (low_bound);
|
|
if (!poly_int_tree_p (low_bound))
|
|
return NULL_TREE;
|
|
unit_size = array_ref_element_size (exp);
|
|
if (TREE_CODE (unit_size) != INTEGER_CST)
|
|
return NULL_TREE;
|
|
|
|
/* If the resulting bit-offset is constant, track it. */
|
|
poly_offset_int woffset
|
|
= wi::sext (wi::to_poly_offset (index)
|
|
- wi::to_poly_offset (low_bound),
|
|
TYPE_PRECISION (sizetype));
|
|
woffset *= wi::to_offset (unit_size);
|
|
byte_offset += woffset.force_shwi ();
|
|
}
|
|
break;
|
|
|
|
case REALPART_EXPR:
|
|
break;
|
|
|
|
case IMAGPART_EXPR:
|
|
byte_offset += TREE_INT_CST_LOW (TYPE_SIZE_UNIT (TREE_TYPE (exp)));
|
|
break;
|
|
|
|
case VIEW_CONVERT_EXPR:
|
|
break;
|
|
|
|
case MEM_REF:
|
|
{
|
|
tree base = TREE_OPERAND (exp, 0);
|
|
if (valueize
|
|
&& TREE_CODE (base) == SSA_NAME)
|
|
base = (*valueize) (base);
|
|
|
|
/* Hand back the decl for MEM[&decl, off]. */
|
|
if (TREE_CODE (base) == ADDR_EXPR)
|
|
{
|
|
if (!integer_zerop (TREE_OPERAND (exp, 1)))
|
|
{
|
|
poly_offset_int off = mem_ref_offset (exp);
|
|
byte_offset += off.force_shwi ();
|
|
}
|
|
exp = TREE_OPERAND (base, 0);
|
|
}
|
|
goto done;
|
|
}
|
|
|
|
case TARGET_MEM_REF:
|
|
{
|
|
tree base = TREE_OPERAND (exp, 0);
|
|
if (valueize
|
|
&& TREE_CODE (base) == SSA_NAME)
|
|
base = (*valueize) (base);
|
|
|
|
/* Hand back the decl for MEM[&decl, off]. */
|
|
if (TREE_CODE (base) == ADDR_EXPR)
|
|
{
|
|
if (TMR_INDEX (exp) || TMR_INDEX2 (exp))
|
|
return NULL_TREE;
|
|
if (!integer_zerop (TMR_OFFSET (exp)))
|
|
{
|
|
poly_offset_int off = mem_ref_offset (exp);
|
|
byte_offset += off.force_shwi ();
|
|
}
|
|
exp = TREE_OPERAND (base, 0);
|
|
}
|
|
goto done;
|
|
}
|
|
|
|
default:
|
|
goto done;
|
|
}
|
|
|
|
exp = TREE_OPERAND (exp, 0);
|
|
}
|
|
done:
|
|
|
|
*poffset = byte_offset;
|
|
return exp;
|
|
}
|
|
|
|
/* Returns the base object and a constant BITS_PER_UNIT offset in *POFFSET that
|
|
denotes the starting address of the memory access EXP.
|
|
Returns NULL_TREE if the offset is not constant or any component
|
|
is not BITS_PER_UNIT-aligned. */
|
|
|
|
tree
|
|
get_addr_base_and_unit_offset (tree exp, poly_int64_pod *poffset)
|
|
{
|
|
return get_addr_base_and_unit_offset_1 (exp, poffset, NULL);
|
|
}
|
|
|
|
/* Returns true if STMT references an SSA_NAME that has
|
|
SSA_NAME_OCCURS_IN_ABNORMAL_PHI set, otherwise false. */
|
|
|
|
bool
|
|
stmt_references_abnormal_ssa_name (gimple *stmt)
|
|
{
|
|
ssa_op_iter oi;
|
|
use_operand_p use_p;
|
|
|
|
FOR_EACH_SSA_USE_OPERAND (use_p, stmt, oi, SSA_OP_USE)
|
|
{
|
|
if (SSA_NAME_OCCURS_IN_ABNORMAL_PHI (USE_FROM_PTR (use_p)))
|
|
return true;
|
|
}
|
|
|
|
return false;
|
|
}
|
|
|
|
/* If STMT takes any abnormal PHI values as input, replace them with
|
|
local copies. */
|
|
|
|
void
|
|
replace_abnormal_ssa_names (gimple *stmt)
|
|
{
|
|
ssa_op_iter oi;
|
|
use_operand_p use_p;
|
|
|
|
FOR_EACH_SSA_USE_OPERAND (use_p, stmt, oi, SSA_OP_USE)
|
|
{
|
|
tree op = USE_FROM_PTR (use_p);
|
|
if (TREE_CODE (op) == SSA_NAME && SSA_NAME_OCCURS_IN_ABNORMAL_PHI (op))
|
|
{
|
|
gimple_stmt_iterator gsi = gsi_for_stmt (stmt);
|
|
tree new_name = make_ssa_name (TREE_TYPE (op));
|
|
gassign *assign = gimple_build_assign (new_name, op);
|
|
gsi_insert_before (&gsi, assign, GSI_SAME_STMT);
|
|
SET_USE (use_p, new_name);
|
|
}
|
|
}
|
|
}
|
|
|
|
/* Pair of tree and a sorting index, for dump_enumerated_decls. */
|
|
struct GTY(()) numbered_tree
|
|
{
|
|
tree t;
|
|
int num;
|
|
};
|
|
|
|
|
|
/* Compare two declarations references by their DECL_UID / sequence number.
|
|
Called via qsort. */
|
|
|
|
static int
|
|
compare_decls_by_uid (const void *pa, const void *pb)
|
|
{
|
|
const numbered_tree *nt_a = ((const numbered_tree *)pa);
|
|
const numbered_tree *nt_b = ((const numbered_tree *)pb);
|
|
|
|
if (DECL_UID (nt_a->t) != DECL_UID (nt_b->t))
|
|
return DECL_UID (nt_a->t) - DECL_UID (nt_b->t);
|
|
return nt_a->num - nt_b->num;
|
|
}
|
|
|
|
/* Called via walk_gimple_stmt / walk_gimple_op by dump_enumerated_decls. */
|
|
static tree
|
|
dump_enumerated_decls_push (tree *tp, int *walk_subtrees, void *data)
|
|
{
|
|
struct walk_stmt_info *wi = (struct walk_stmt_info *) data;
|
|
vec<numbered_tree> *list = (vec<numbered_tree> *) wi->info;
|
|
numbered_tree nt;
|
|
|
|
if (!DECL_P (*tp))
|
|
return NULL_TREE;
|
|
nt.t = *tp;
|
|
nt.num = list->length ();
|
|
list->safe_push (nt);
|
|
*walk_subtrees = 0;
|
|
return NULL_TREE;
|
|
}
|
|
|
|
/* Find all the declarations used by the current function, sort them by uid,
|
|
and emit the sorted list. Each declaration is tagged with a sequence
|
|
number indicating when it was found during statement / tree walking,
|
|
so that TDF_NOUID comparisons of anonymous declarations are still
|
|
meaningful. Where a declaration was encountered more than once, we
|
|
emit only the sequence number of the first encounter.
|
|
FILE is the dump file where to output the list and FLAGS is as in
|
|
print_generic_expr. */
|
|
void
|
|
dump_enumerated_decls (FILE *file, dump_flags_t flags)
|
|
{
|
|
if (!cfun->cfg)
|
|
return;
|
|
|
|
basic_block bb;
|
|
struct walk_stmt_info wi;
|
|
auto_vec<numbered_tree, 40> decl_list;
|
|
|
|
memset (&wi, '\0', sizeof (wi));
|
|
wi.info = (void *) &decl_list;
|
|
FOR_EACH_BB_FN (bb, cfun)
|
|
{
|
|
gimple_stmt_iterator gsi;
|
|
|
|
for (gsi = gsi_start_bb (bb); !gsi_end_p (gsi); gsi_next (&gsi))
|
|
if (!is_gimple_debug (gsi_stmt (gsi)))
|
|
walk_gimple_stmt (&gsi, NULL, dump_enumerated_decls_push, &wi);
|
|
}
|
|
decl_list.qsort (compare_decls_by_uid);
|
|
if (decl_list.length ())
|
|
{
|
|
unsigned ix;
|
|
numbered_tree *ntp;
|
|
tree last = NULL_TREE;
|
|
|
|
fprintf (file, "Declarations used by %s, sorted by DECL_UID:\n",
|
|
current_function_name ());
|
|
FOR_EACH_VEC_ELT (decl_list, ix, ntp)
|
|
{
|
|
if (ntp->t == last)
|
|
continue;
|
|
fprintf (file, "%d: ", ntp->num);
|
|
print_generic_decl (file, ntp->t, flags);
|
|
fprintf (file, "\n");
|
|
last = ntp->t;
|
|
}
|
|
}
|
|
}
|