7ee2468b92
gcc/ * system.h (dump_file): Do not define. * tree-pass.h: Include dumpfile.h, which is a new file containing... (enum tree_dump_index, TDF_*, get_dump_file_name, dump_enabled_p, dump_initialized_p, dump_begin, dump_end, dump_node, dump_switch_p, dump_flag_name, dump_file, dump_flags, dump_file_name, get_dump_file_info, struct dump_file_info): all of this, moved to... * dumpfile.h: Here, new file. * tree-dump.h: Include dumpfile.h, but not tree-pass.h. (dump_stmt): Remove prototype for C++ specific function. (dump_enumerated_decls): Move prototype from here... * tree-flow.h (dump_enumerated_decls): ... to here. (get_ref_base_and_extent) Move prototype from here ... * tree.h (get_ref_base_and_extent) ... to here. * tree-ssa-live.c: Do not inclde tree-pretty-print.h, because gimple-pretty-print.h is enough. Do not include tree-dump.h, include timevar.h and dumpfile.h instead. (struct numbered_tree_d, compare_decls_by_uid, dump_enumerated_decls_push, dump_enumerated_decls): Move from here ... * tree-dfa.c:(struct numbered_tree_d, compare_decls_by_uid, dump_enumerated_decls_push, dump_enumerated_decls):... to here. Do not include timevar.h. * tree.c: Do not include timevar.h. * tree-cfg.c: Do not include langhooks.h, tree-pretty-print.h, and timevar.h. (dump_cfg_stats): Use current_function_name. (gimple_cfg2vcg): Likewise. (dump_function_to_file): Likewise. * df-scan.c: Do not include tree-pass.h and timevar.h. Include dumpfile.h. (df_entry_block_bitmap_verify, df_exit_block_bitmap_verify): Do not use print_current_pass. * df-problems.c: Include dumpfile.h. Always define REG_DEAD_DEBUGGING, avoid #ifdef code, because it leads to errors in the code not selected. (df_note_compute): Do not print_rtl_with_bb here. Fix compilation bug if REG_DEAD_DEBUGGING is not 0, get_insns is not available here. * lcm.c: Include dumpfile.h. Remove obsolete include of insn-attr.h. * dojump.c (do_compare_rtx_and_jump): Remove failure printing for missing probability notes. * stmt.c: Include dumpfile.h. (emit_case_decision_tree): Re-enable printing expand details only if TDF_DETAILS. * alias.c, auto-inc-dec.c, bb-reorder.c, caller-save.c, cfg.c, cfgcleanup.c, cfgexpand.c, cfgloop.c, cfgloopmanip.c, cgraph.c, cgraphclones.c, cgraphunit.c, combine.c, combine-stack-adj.c, coverage.c, cprop.c, cse.c, cselib.c, dbgcnt.c, dce.c, df-core.c, dse.c, dwarf2out.c, emit-rtl.c, except.c, expr.c, final.c, function.c, fwprop.c, gcse.c, gimple-fold.c, gimple-pretty-print.c, gimple-ssa-strength-reduction.c, gimplify.c, graphite-blocking.c, graphite-clast-to-gimple.c, graphite-dependences.c, graphite-interchange.c, graphite-optimize-isl.c, graphite-poly.c, graphite-sese-to-poly.c, haifa-sched.c, hw-doloop.c, ifcvt.c, ipa.c, ipa-cp.c, ipa-inline-analysis.c, ipa-inline.c, ipa-inline-transform.c, ipa-prop.c, ipa-pure-const.c, ipa-reference.c, ipa-split.c, ipa-utils.c, ira.c, ira-emit.c, jump.c, loop-doloop.c, loop-init.c, loop-invariant.c, loop-iv.c, loop-unroll.c, loop-unswitch.c, lower-subreg.c, lto-section-out.c, lto-streamer-in.c, matrix-reorg.c, mcf.c, mode-switching.c, modulo-sched.c, omega.c, omp-low.c, passes.c, plugin.c, postreload.c, postreload-gcse.c, predict.c, print-rtl.c, print-tree.c, profile.c, recog.c, ree.c, regcprop.c, reginfo.c, regmove.c, regrename.c, reg-stack.c, reload1.c, reorg.c, sched-rgn.c, sched-vis.c, sel-sched.c, sel-sched-ir.c, store-motion.c, tracer.c, trans-mem.c, tree-affine.c, tree-call-cdce.c, tree-cfgcleanup.c, tree-chrec.c, tree-data-ref.c, tree-diagnostic.c, tree-dump.c, tree-eh.c, tree-flow-inline.h, tree-if-conv.c, tree-into-ssa.c, tree-mudflap.c, tree-nrv.c, tree-object-size.c, tree-optimize.c, tree-outof-ssa.c, tree-predcom.c, tree-pretty-print.c, tree-profile.c, tree-scalar-evolution.c, tree-sra.c, tree-ssa-address.c, tree-ssa-alias.c, tree-ssa.c, tree-ssa-ccp.c, tree-ssa-coalesce.c, tree-ssa-copy.c, tree-ssa-copyrename.c,, tree-ssa-dce.c, tree-ssa-dom.c, tree-ssa-dse.c, tree-ssa-forwprop.c, tree-ssa-ifcombine.c, tree-ssa-loop.c, tree-ssa-loop-ch.c, tree-ssa-loop-im.c, tree-ssa-loop-ivcanon.c, tree-ssa-loop-ivopts.c, tree-ssa-loop-manip.c, tree-ssa-loop-niter.c, tree-ssa-loop-prefetch.c, tree-ssa-loop-unswitch.c, tree-ssa-math-opts.c, tree-ssa-operands.c, tree-ssa-phiopt.c, tree-ssa-phiprop.c, tree-ssa-pre.c, tree-ssa-propagate.c, tree-ssa-reassoc.c, tree-ssa-sccvn.c, tree-ssa-sink.c, tree-ssa-structalias.c, tree-ssa-tail-merge.c, tree-ssa-ter.c, tree-ssa-threadedge.c, tree-ssa-threadupdate.c, tree-ssa-uncprop.c, tree-ssa-uninit.c, tree-switch-conversion.c, tree-tailcall.c, tree-vect-data-refs.c, tree-vect-loop.c, tree-vect-loop-manip.c, tree-vectorizer.c, tree-vect-patterns.c, tree-vect-slp.c, tree-vect-stmts.c, tree-vrp.c, value-prof.c, var-tracking.c, web.c: Include tree-pass.h only if needed. If tree-pass.h is included, do not include timevar.h and dumpfile.h. If tree-pass.h is not included but dump_file, or dump_flags, or the TDF_* flags are used, include dumpfile.h. If gimple-pretty-print.h is included, don't include tree-pretty-print.h. Remove assorted unnecessary includes. * config/mn10300/mn10300.c, config/c6x/c6x.c, config/ia64/ia64.c, config/arm/arm.c, config/bfin/bfin.c, config/frv/frv.c, config/spu/spu.c, config/mep/mep.c, config/i386/i386.c: Include dumpfile.h. * config/rl78/rl78.c: Include dumpfile.h instead of tree-pass.h. * arm/t-arm, avr/t-avr, i386/t-i386, ia64/t-ia64, mep/t-mep, spu/t-spu-elf: Fix dependencies. c-family/ * c-gimplify.c: Include dumpfile.h instead of tree-dump.h. * c-ada-spec.c: Likewise. * c-dump.c (dump_stmt): Move to cp/dump.c, the only user. c/ * c-decl.c: Include dumpfile.h instead of tree-dump.h. * Make-lang.in: Fix dependencies. cp/ * dump.c (dump_stmt): Moved here from c-dump.c. * optimize.c: Include dumpfile.h instead of tree-dump.h. * class.c: Likewise. * decl2.c: Likewise. * Make-lang.in: Fix dependencies. fortran/ * f95-lang.c: Include dumpfile.h instead of tree-dump.h. * Make-lang.in: Fix dependencies. java/ * java-gimplify.c Include dumpfile.h instead of tree-dump.h * Make-lang.in: Fix dependencies. lto/ * lto.c: Do not include timevar.h. * Make-lang.in: Fix dependencies. ada/ * gcc-interface/utils.c: Include timevar.h. * Make-lang.in: Fix dependencies. From-SVN: r189519
916 lines
23 KiB
C
916 lines
23 KiB
C
/* Operations with affine combinations of trees.
|
|
Copyright (C) 2005, 2007, 2008, 2010 Free Software Foundation, Inc.
|
|
|
|
This file is part of GCC.
|
|
|
|
GCC is free software; you can redistribute it and/or modify it
|
|
under the terms of the GNU General Public License as published by the
|
|
Free Software Foundation; either version 3, or (at your option) any
|
|
later version.
|
|
|
|
GCC is distributed in the hope that it will be useful, but WITHOUT
|
|
ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
|
|
FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License
|
|
for more details.
|
|
|
|
You should have received a copy of the GNU General Public License
|
|
along with GCC; see the file COPYING3. If not see
|
|
<http://www.gnu.org/licenses/>. */
|
|
|
|
#include "config.h"
|
|
#include "system.h"
|
|
#include "coretypes.h"
|
|
#include "tree.h"
|
|
#include "tree-pretty-print.h"
|
|
#include "pointer-set.h"
|
|
#include "tree-affine.h"
|
|
#include "gimple.h"
|
|
#include "flags.h"
|
|
#include "dumpfile.h"
|
|
|
|
/* Extends CST as appropriate for the affine combinations COMB. */
|
|
|
|
double_int
|
|
double_int_ext_for_comb (double_int cst, aff_tree *comb)
|
|
{
|
|
return double_int_sext (cst, TYPE_PRECISION (comb->type));
|
|
}
|
|
|
|
/* Initializes affine combination COMB so that its value is zero in TYPE. */
|
|
|
|
static void
|
|
aff_combination_zero (aff_tree *comb, tree type)
|
|
{
|
|
comb->type = type;
|
|
comb->offset = double_int_zero;
|
|
comb->n = 0;
|
|
comb->rest = NULL_TREE;
|
|
}
|
|
|
|
/* Sets COMB to CST. */
|
|
|
|
void
|
|
aff_combination_const (aff_tree *comb, tree type, double_int cst)
|
|
{
|
|
aff_combination_zero (comb, type);
|
|
comb->offset = double_int_ext_for_comb (cst, comb);
|
|
}
|
|
|
|
/* Sets COMB to single element ELT. */
|
|
|
|
void
|
|
aff_combination_elt (aff_tree *comb, tree type, tree elt)
|
|
{
|
|
aff_combination_zero (comb, type);
|
|
|
|
comb->n = 1;
|
|
comb->elts[0].val = elt;
|
|
comb->elts[0].coef = double_int_one;
|
|
}
|
|
|
|
/* Scales COMB by SCALE. */
|
|
|
|
void
|
|
aff_combination_scale (aff_tree *comb, double_int scale)
|
|
{
|
|
unsigned i, j;
|
|
|
|
scale = double_int_ext_for_comb (scale, comb);
|
|
if (double_int_one_p (scale))
|
|
return;
|
|
|
|
if (double_int_zero_p (scale))
|
|
{
|
|
aff_combination_zero (comb, comb->type);
|
|
return;
|
|
}
|
|
|
|
comb->offset
|
|
= double_int_ext_for_comb (double_int_mul (scale, comb->offset), comb);
|
|
for (i = 0, j = 0; i < comb->n; i++)
|
|
{
|
|
double_int new_coef;
|
|
|
|
new_coef
|
|
= double_int_ext_for_comb (double_int_mul (scale, comb->elts[i].coef),
|
|
comb);
|
|
/* A coefficient may become zero due to overflow. Remove the zero
|
|
elements. */
|
|
if (double_int_zero_p (new_coef))
|
|
continue;
|
|
comb->elts[j].coef = new_coef;
|
|
comb->elts[j].val = comb->elts[i].val;
|
|
j++;
|
|
}
|
|
comb->n = j;
|
|
|
|
if (comb->rest)
|
|
{
|
|
tree type = comb->type;
|
|
if (POINTER_TYPE_P (type))
|
|
type = sizetype;
|
|
if (comb->n < MAX_AFF_ELTS)
|
|
{
|
|
comb->elts[comb->n].coef = scale;
|
|
comb->elts[comb->n].val = comb->rest;
|
|
comb->rest = NULL_TREE;
|
|
comb->n++;
|
|
}
|
|
else
|
|
comb->rest = fold_build2 (MULT_EXPR, type, comb->rest,
|
|
double_int_to_tree (type, scale));
|
|
}
|
|
}
|
|
|
|
/* Adds ELT * SCALE to COMB. */
|
|
|
|
void
|
|
aff_combination_add_elt (aff_tree *comb, tree elt, double_int scale)
|
|
{
|
|
unsigned i;
|
|
tree type;
|
|
|
|
scale = double_int_ext_for_comb (scale, comb);
|
|
if (double_int_zero_p (scale))
|
|
return;
|
|
|
|
for (i = 0; i < comb->n; i++)
|
|
if (operand_equal_p (comb->elts[i].val, elt, 0))
|
|
{
|
|
double_int new_coef;
|
|
|
|
new_coef = double_int_add (comb->elts[i].coef, scale);
|
|
new_coef = double_int_ext_for_comb (new_coef, comb);
|
|
if (!double_int_zero_p (new_coef))
|
|
{
|
|
comb->elts[i].coef = new_coef;
|
|
return;
|
|
}
|
|
|
|
comb->n--;
|
|
comb->elts[i] = comb->elts[comb->n];
|
|
|
|
if (comb->rest)
|
|
{
|
|
gcc_assert (comb->n == MAX_AFF_ELTS - 1);
|
|
comb->elts[comb->n].coef = double_int_one;
|
|
comb->elts[comb->n].val = comb->rest;
|
|
comb->rest = NULL_TREE;
|
|
comb->n++;
|
|
}
|
|
return;
|
|
}
|
|
if (comb->n < MAX_AFF_ELTS)
|
|
{
|
|
comb->elts[comb->n].coef = scale;
|
|
comb->elts[comb->n].val = elt;
|
|
comb->n++;
|
|
return;
|
|
}
|
|
|
|
type = comb->type;
|
|
if (POINTER_TYPE_P (type))
|
|
type = sizetype;
|
|
|
|
if (double_int_one_p (scale))
|
|
elt = fold_convert (type, elt);
|
|
else
|
|
elt = fold_build2 (MULT_EXPR, type,
|
|
fold_convert (type, elt),
|
|
double_int_to_tree (type, scale));
|
|
|
|
if (comb->rest)
|
|
comb->rest = fold_build2 (PLUS_EXPR, type, comb->rest,
|
|
elt);
|
|
else
|
|
comb->rest = elt;
|
|
}
|
|
|
|
/* Adds CST to C. */
|
|
|
|
static void
|
|
aff_combination_add_cst (aff_tree *c, double_int cst)
|
|
{
|
|
c->offset = double_int_ext_for_comb (double_int_add (c->offset, cst), c);
|
|
}
|
|
|
|
/* Adds COMB2 to COMB1. */
|
|
|
|
void
|
|
aff_combination_add (aff_tree *comb1, aff_tree *comb2)
|
|
{
|
|
unsigned i;
|
|
|
|
aff_combination_add_cst (comb1, comb2->offset);
|
|
for (i = 0; i < comb2->n; i++)
|
|
aff_combination_add_elt (comb1, comb2->elts[i].val, comb2->elts[i].coef);
|
|
if (comb2->rest)
|
|
aff_combination_add_elt (comb1, comb2->rest, double_int_one);
|
|
}
|
|
|
|
/* Converts affine combination COMB to TYPE. */
|
|
|
|
void
|
|
aff_combination_convert (aff_tree *comb, tree type)
|
|
{
|
|
unsigned i, j;
|
|
tree comb_type = comb->type;
|
|
|
|
if (TYPE_PRECISION (type) > TYPE_PRECISION (comb_type))
|
|
{
|
|
tree val = fold_convert (type, aff_combination_to_tree (comb));
|
|
tree_to_aff_combination (val, type, comb);
|
|
return;
|
|
}
|
|
|
|
comb->type = type;
|
|
if (comb->rest && !POINTER_TYPE_P (type))
|
|
comb->rest = fold_convert (type, comb->rest);
|
|
|
|
if (TYPE_PRECISION (type) == TYPE_PRECISION (comb_type))
|
|
return;
|
|
|
|
comb->offset = double_int_ext_for_comb (comb->offset, comb);
|
|
for (i = j = 0; i < comb->n; i++)
|
|
{
|
|
double_int new_coef = double_int_ext_for_comb (comb->elts[i].coef, comb);
|
|
if (double_int_zero_p (new_coef))
|
|
continue;
|
|
comb->elts[j].coef = new_coef;
|
|
comb->elts[j].val = fold_convert (type, comb->elts[i].val);
|
|
j++;
|
|
}
|
|
|
|
comb->n = j;
|
|
if (comb->n < MAX_AFF_ELTS && comb->rest)
|
|
{
|
|
comb->elts[comb->n].coef = double_int_one;
|
|
comb->elts[comb->n].val = comb->rest;
|
|
comb->rest = NULL_TREE;
|
|
comb->n++;
|
|
}
|
|
}
|
|
|
|
/* Splits EXPR into an affine combination of parts. */
|
|
|
|
void
|
|
tree_to_aff_combination (tree expr, tree type, aff_tree *comb)
|
|
{
|
|
aff_tree tmp;
|
|
enum tree_code code;
|
|
tree cst, core, toffset;
|
|
HOST_WIDE_INT bitpos, bitsize;
|
|
enum machine_mode mode;
|
|
int unsignedp, volatilep;
|
|
|
|
STRIP_NOPS (expr);
|
|
|
|
code = TREE_CODE (expr);
|
|
switch (code)
|
|
{
|
|
case INTEGER_CST:
|
|
aff_combination_const (comb, type, tree_to_double_int (expr));
|
|
return;
|
|
|
|
case POINTER_PLUS_EXPR:
|
|
tree_to_aff_combination (TREE_OPERAND (expr, 0), type, comb);
|
|
tree_to_aff_combination (TREE_OPERAND (expr, 1), sizetype, &tmp);
|
|
aff_combination_add (comb, &tmp);
|
|
return;
|
|
|
|
case PLUS_EXPR:
|
|
case MINUS_EXPR:
|
|
tree_to_aff_combination (TREE_OPERAND (expr, 0), type, comb);
|
|
tree_to_aff_combination (TREE_OPERAND (expr, 1), type, &tmp);
|
|
if (code == MINUS_EXPR)
|
|
aff_combination_scale (&tmp, double_int_minus_one);
|
|
aff_combination_add (comb, &tmp);
|
|
return;
|
|
|
|
case MULT_EXPR:
|
|
cst = TREE_OPERAND (expr, 1);
|
|
if (TREE_CODE (cst) != INTEGER_CST)
|
|
break;
|
|
tree_to_aff_combination (TREE_OPERAND (expr, 0), type, comb);
|
|
aff_combination_scale (comb, tree_to_double_int (cst));
|
|
return;
|
|
|
|
case NEGATE_EXPR:
|
|
tree_to_aff_combination (TREE_OPERAND (expr, 0), type, comb);
|
|
aff_combination_scale (comb, double_int_minus_one);
|
|
return;
|
|
|
|
case BIT_NOT_EXPR:
|
|
/* ~x = -x - 1 */
|
|
tree_to_aff_combination (TREE_OPERAND (expr, 0), type, comb);
|
|
aff_combination_scale (comb, double_int_minus_one);
|
|
aff_combination_add_cst (comb, double_int_minus_one);
|
|
return;
|
|
|
|
case ADDR_EXPR:
|
|
/* Handle &MEM[ptr + CST] which is equivalent to POINTER_PLUS_EXPR. */
|
|
if (TREE_CODE (TREE_OPERAND (expr, 0)) == MEM_REF)
|
|
{
|
|
expr = TREE_OPERAND (expr, 0);
|
|
tree_to_aff_combination (TREE_OPERAND (expr, 0), type, comb);
|
|
tree_to_aff_combination (TREE_OPERAND (expr, 1), sizetype, &tmp);
|
|
aff_combination_add (comb, &tmp);
|
|
return;
|
|
}
|
|
core = get_inner_reference (TREE_OPERAND (expr, 0), &bitsize, &bitpos,
|
|
&toffset, &mode, &unsignedp, &volatilep,
|
|
false);
|
|
if (bitpos % BITS_PER_UNIT != 0)
|
|
break;
|
|
aff_combination_const (comb, type,
|
|
uhwi_to_double_int (bitpos / BITS_PER_UNIT));
|
|
core = build_fold_addr_expr (core);
|
|
if (TREE_CODE (core) == ADDR_EXPR)
|
|
aff_combination_add_elt (comb, core, double_int_one);
|
|
else
|
|
{
|
|
tree_to_aff_combination (core, type, &tmp);
|
|
aff_combination_add (comb, &tmp);
|
|
}
|
|
if (toffset)
|
|
{
|
|
tree_to_aff_combination (toffset, type, &tmp);
|
|
aff_combination_add (comb, &tmp);
|
|
}
|
|
return;
|
|
|
|
case MEM_REF:
|
|
if (TREE_CODE (TREE_OPERAND (expr, 0)) == ADDR_EXPR)
|
|
tree_to_aff_combination (TREE_OPERAND (TREE_OPERAND (expr, 0), 0),
|
|
type, comb);
|
|
else if (integer_zerop (TREE_OPERAND (expr, 1)))
|
|
{
|
|
aff_combination_elt (comb, type, expr);
|
|
return;
|
|
}
|
|
else
|
|
aff_combination_elt (comb, type,
|
|
build2 (MEM_REF, TREE_TYPE (expr),
|
|
TREE_OPERAND (expr, 0),
|
|
build_int_cst
|
|
(TREE_TYPE (TREE_OPERAND (expr, 1)), 0)));
|
|
tree_to_aff_combination (TREE_OPERAND (expr, 1), sizetype, &tmp);
|
|
aff_combination_add (comb, &tmp);
|
|
return;
|
|
|
|
default:
|
|
break;
|
|
}
|
|
|
|
aff_combination_elt (comb, type, expr);
|
|
}
|
|
|
|
/* Creates EXPR + ELT * SCALE in TYPE. EXPR is taken from affine
|
|
combination COMB. */
|
|
|
|
static tree
|
|
add_elt_to_tree (tree expr, tree type, tree elt, double_int scale,
|
|
aff_tree *comb)
|
|
{
|
|
enum tree_code code;
|
|
tree type1 = type;
|
|
if (POINTER_TYPE_P (type))
|
|
type1 = sizetype;
|
|
|
|
scale = double_int_ext_for_comb (scale, comb);
|
|
elt = fold_convert (type1, elt);
|
|
|
|
if (double_int_one_p (scale))
|
|
{
|
|
if (!expr)
|
|
return fold_convert (type, elt);
|
|
|
|
if (POINTER_TYPE_P (type))
|
|
return fold_build_pointer_plus (expr, elt);
|
|
return fold_build2 (PLUS_EXPR, type, expr, elt);
|
|
}
|
|
|
|
if (double_int_minus_one_p (scale))
|
|
{
|
|
if (!expr)
|
|
return fold_convert (type, fold_build1 (NEGATE_EXPR, type1, elt));
|
|
|
|
if (POINTER_TYPE_P (type))
|
|
{
|
|
elt = fold_build1 (NEGATE_EXPR, type1, elt);
|
|
return fold_build_pointer_plus (expr, elt);
|
|
}
|
|
return fold_build2 (MINUS_EXPR, type, expr, elt);
|
|
}
|
|
|
|
if (!expr)
|
|
return fold_convert (type,
|
|
fold_build2 (MULT_EXPR, type1, elt,
|
|
double_int_to_tree (type1, scale)));
|
|
|
|
if (double_int_negative_p (scale))
|
|
{
|
|
code = MINUS_EXPR;
|
|
scale = double_int_neg (scale);
|
|
}
|
|
else
|
|
code = PLUS_EXPR;
|
|
|
|
elt = fold_build2 (MULT_EXPR, type1, elt,
|
|
double_int_to_tree (type1, scale));
|
|
if (POINTER_TYPE_P (type))
|
|
{
|
|
if (code == MINUS_EXPR)
|
|
elt = fold_build1 (NEGATE_EXPR, type1, elt);
|
|
return fold_build_pointer_plus (expr, elt);
|
|
}
|
|
return fold_build2 (code, type, expr, elt);
|
|
}
|
|
|
|
/* Makes tree from the affine combination COMB. */
|
|
|
|
tree
|
|
aff_combination_to_tree (aff_tree *comb)
|
|
{
|
|
tree type = comb->type;
|
|
tree expr = NULL_TREE;
|
|
unsigned i;
|
|
double_int off, sgn;
|
|
tree type1 = type;
|
|
if (POINTER_TYPE_P (type))
|
|
type1 = sizetype;
|
|
|
|
gcc_assert (comb->n == MAX_AFF_ELTS || comb->rest == NULL_TREE);
|
|
|
|
for (i = 0; i < comb->n; i++)
|
|
expr = add_elt_to_tree (expr, type, comb->elts[i].val, comb->elts[i].coef,
|
|
comb);
|
|
|
|
if (comb->rest)
|
|
expr = add_elt_to_tree (expr, type, comb->rest, double_int_one, comb);
|
|
|
|
/* Ensure that we get x - 1, not x + (-1) or x + 0xff..f if x is
|
|
unsigned. */
|
|
if (double_int_negative_p (comb->offset))
|
|
{
|
|
off = double_int_neg (comb->offset);
|
|
sgn = double_int_minus_one;
|
|
}
|
|
else
|
|
{
|
|
off = comb->offset;
|
|
sgn = double_int_one;
|
|
}
|
|
return add_elt_to_tree (expr, type, double_int_to_tree (type1, off), sgn,
|
|
comb);
|
|
}
|
|
|
|
/* Copies the tree elements of COMB to ensure that they are not shared. */
|
|
|
|
void
|
|
unshare_aff_combination (aff_tree *comb)
|
|
{
|
|
unsigned i;
|
|
|
|
for (i = 0; i < comb->n; i++)
|
|
comb->elts[i].val = unshare_expr (comb->elts[i].val);
|
|
if (comb->rest)
|
|
comb->rest = unshare_expr (comb->rest);
|
|
}
|
|
|
|
/* Remove M-th element from COMB. */
|
|
|
|
void
|
|
aff_combination_remove_elt (aff_tree *comb, unsigned m)
|
|
{
|
|
comb->n--;
|
|
if (m <= comb->n)
|
|
comb->elts[m] = comb->elts[comb->n];
|
|
if (comb->rest)
|
|
{
|
|
comb->elts[comb->n].coef = double_int_one;
|
|
comb->elts[comb->n].val = comb->rest;
|
|
comb->rest = NULL_TREE;
|
|
comb->n++;
|
|
}
|
|
}
|
|
|
|
/* Adds C * COEF * VAL to R. VAL may be NULL, in that case only
|
|
C * COEF is added to R. */
|
|
|
|
|
|
static void
|
|
aff_combination_add_product (aff_tree *c, double_int coef, tree val,
|
|
aff_tree *r)
|
|
{
|
|
unsigned i;
|
|
tree aval, type;
|
|
|
|
for (i = 0; i < c->n; i++)
|
|
{
|
|
aval = c->elts[i].val;
|
|
if (val)
|
|
{
|
|
type = TREE_TYPE (aval);
|
|
aval = fold_build2 (MULT_EXPR, type, aval,
|
|
fold_convert (type, val));
|
|
}
|
|
|
|
aff_combination_add_elt (r, aval,
|
|
double_int_mul (coef, c->elts[i].coef));
|
|
}
|
|
|
|
if (c->rest)
|
|
{
|
|
aval = c->rest;
|
|
if (val)
|
|
{
|
|
type = TREE_TYPE (aval);
|
|
aval = fold_build2 (MULT_EXPR, type, aval,
|
|
fold_convert (type, val));
|
|
}
|
|
|
|
aff_combination_add_elt (r, aval, coef);
|
|
}
|
|
|
|
if (val)
|
|
aff_combination_add_elt (r, val,
|
|
double_int_mul (coef, c->offset));
|
|
else
|
|
aff_combination_add_cst (r, double_int_mul (coef, c->offset));
|
|
}
|
|
|
|
/* Multiplies C1 by C2, storing the result to R */
|
|
|
|
void
|
|
aff_combination_mult (aff_tree *c1, aff_tree *c2, aff_tree *r)
|
|
{
|
|
unsigned i;
|
|
gcc_assert (TYPE_PRECISION (c1->type) == TYPE_PRECISION (c2->type));
|
|
|
|
aff_combination_zero (r, c1->type);
|
|
|
|
for (i = 0; i < c2->n; i++)
|
|
aff_combination_add_product (c1, c2->elts[i].coef, c2->elts[i].val, r);
|
|
if (c2->rest)
|
|
aff_combination_add_product (c1, double_int_one, c2->rest, r);
|
|
aff_combination_add_product (c1, c2->offset, NULL, r);
|
|
}
|
|
|
|
/* Returns the element of COMB whose value is VAL, or NULL if no such
|
|
element exists. If IDX is not NULL, it is set to the index of VAL in
|
|
COMB. */
|
|
|
|
static struct aff_comb_elt *
|
|
aff_combination_find_elt (aff_tree *comb, tree val, unsigned *idx)
|
|
{
|
|
unsigned i;
|
|
|
|
for (i = 0; i < comb->n; i++)
|
|
if (operand_equal_p (comb->elts[i].val, val, 0))
|
|
{
|
|
if (idx)
|
|
*idx = i;
|
|
|
|
return &comb->elts[i];
|
|
}
|
|
|
|
return NULL;
|
|
}
|
|
|
|
/* Element of the cache that maps ssa name NAME to its expanded form
|
|
as an affine expression EXPANSION. */
|
|
|
|
struct name_expansion
|
|
{
|
|
aff_tree expansion;
|
|
|
|
/* True if the expansion for the name is just being generated. */
|
|
unsigned in_progress : 1;
|
|
};
|
|
|
|
/* Expands SSA names in COMB recursively. CACHE is used to cache the
|
|
results. */
|
|
|
|
void
|
|
aff_combination_expand (aff_tree *comb ATTRIBUTE_UNUSED,
|
|
struct pointer_map_t **cache ATTRIBUTE_UNUSED)
|
|
{
|
|
unsigned i;
|
|
aff_tree to_add, current, curre;
|
|
tree e, rhs;
|
|
gimple def;
|
|
double_int scale;
|
|
void **slot;
|
|
struct name_expansion *exp;
|
|
|
|
aff_combination_zero (&to_add, comb->type);
|
|
for (i = 0; i < comb->n; i++)
|
|
{
|
|
tree type, name;
|
|
enum tree_code code;
|
|
|
|
e = comb->elts[i].val;
|
|
type = TREE_TYPE (e);
|
|
name = e;
|
|
/* Look through some conversions. */
|
|
if (TREE_CODE (e) == NOP_EXPR
|
|
&& (TYPE_PRECISION (type)
|
|
>= TYPE_PRECISION (TREE_TYPE (TREE_OPERAND (e, 0)))))
|
|
name = TREE_OPERAND (e, 0);
|
|
if (TREE_CODE (name) != SSA_NAME)
|
|
continue;
|
|
def = SSA_NAME_DEF_STMT (name);
|
|
if (!is_gimple_assign (def) || gimple_assign_lhs (def) != name)
|
|
continue;
|
|
|
|
code = gimple_assign_rhs_code (def);
|
|
if (code != SSA_NAME
|
|
&& !IS_EXPR_CODE_CLASS (TREE_CODE_CLASS (code))
|
|
&& (get_gimple_rhs_class (code) != GIMPLE_SINGLE_RHS
|
|
|| !is_gimple_min_invariant (gimple_assign_rhs1 (def))))
|
|
continue;
|
|
|
|
/* We do not know whether the reference retains its value at the
|
|
place where the expansion is used. */
|
|
if (TREE_CODE_CLASS (code) == tcc_reference)
|
|
continue;
|
|
|
|
if (!*cache)
|
|
*cache = pointer_map_create ();
|
|
slot = pointer_map_insert (*cache, e);
|
|
exp = (struct name_expansion *) *slot;
|
|
|
|
if (!exp)
|
|
{
|
|
exp = XNEW (struct name_expansion);
|
|
exp->in_progress = 1;
|
|
*slot = exp;
|
|
/* In principle this is a generally valid folding, but
|
|
it is not unconditionally an optimization, so do it
|
|
here and not in fold_unary. */
|
|
/* Convert (T1)(X *+- CST) into (T1)X *+- (T1)CST if T1 is wider
|
|
than the type of X and overflow for the type of X is
|
|
undefined. */
|
|
if (e != name
|
|
&& INTEGRAL_TYPE_P (type)
|
|
&& INTEGRAL_TYPE_P (TREE_TYPE (name))
|
|
&& TYPE_OVERFLOW_UNDEFINED (TREE_TYPE (name))
|
|
&& TYPE_PRECISION (type) > TYPE_PRECISION (TREE_TYPE (name))
|
|
&& (code == PLUS_EXPR || code == MINUS_EXPR || code == MULT_EXPR)
|
|
&& TREE_CODE (gimple_assign_rhs2 (def)) == INTEGER_CST)
|
|
rhs = fold_build2 (code, type,
|
|
fold_convert (type, gimple_assign_rhs1 (def)),
|
|
fold_convert (type, gimple_assign_rhs2 (def)));
|
|
else
|
|
{
|
|
rhs = gimple_assign_rhs_to_tree (def);
|
|
if (e != name)
|
|
rhs = fold_convert (type, rhs);
|
|
}
|
|
tree_to_aff_combination_expand (rhs, comb->type, ¤t, cache);
|
|
exp->expansion = current;
|
|
exp->in_progress = 0;
|
|
}
|
|
else
|
|
{
|
|
/* Since we follow the definitions in the SSA form, we should not
|
|
enter a cycle unless we pass through a phi node. */
|
|
gcc_assert (!exp->in_progress);
|
|
current = exp->expansion;
|
|
}
|
|
|
|
/* Accumulate the new terms to TO_ADD, so that we do not modify
|
|
COMB while traversing it; include the term -coef * E, to remove
|
|
it from COMB. */
|
|
scale = comb->elts[i].coef;
|
|
aff_combination_zero (&curre, comb->type);
|
|
aff_combination_add_elt (&curre, e, double_int_neg (scale));
|
|
aff_combination_scale (¤t, scale);
|
|
aff_combination_add (&to_add, ¤t);
|
|
aff_combination_add (&to_add, &curre);
|
|
}
|
|
aff_combination_add (comb, &to_add);
|
|
}
|
|
|
|
/* Similar to tree_to_aff_combination, but follows SSA name definitions
|
|
and expands them recursively. CACHE is used to cache the expansions
|
|
of the ssa names, to avoid exponential time complexity for cases
|
|
like
|
|
|
|
a1 = a0 + a0;
|
|
a2 = a1 + a1;
|
|
a3 = a2 + a2;
|
|
... */
|
|
|
|
void
|
|
tree_to_aff_combination_expand (tree expr, tree type, aff_tree *comb,
|
|
struct pointer_map_t **cache)
|
|
{
|
|
tree_to_aff_combination (expr, type, comb);
|
|
aff_combination_expand (comb, cache);
|
|
}
|
|
|
|
/* Frees memory occupied by struct name_expansion in *VALUE. Callback for
|
|
pointer_map_traverse. */
|
|
|
|
static bool
|
|
free_name_expansion (const void *key ATTRIBUTE_UNUSED, void **value,
|
|
void *data ATTRIBUTE_UNUSED)
|
|
{
|
|
struct name_expansion *const exp = (struct name_expansion *) *value;
|
|
|
|
free (exp);
|
|
return true;
|
|
}
|
|
|
|
/* Frees memory allocated for the CACHE used by
|
|
tree_to_aff_combination_expand. */
|
|
|
|
void
|
|
free_affine_expand_cache (struct pointer_map_t **cache)
|
|
{
|
|
if (!*cache)
|
|
return;
|
|
|
|
pointer_map_traverse (*cache, free_name_expansion, NULL);
|
|
pointer_map_destroy (*cache);
|
|
*cache = NULL;
|
|
}
|
|
|
|
/* If VAL != CST * DIV for any constant CST, returns false.
|
|
Otherwise, if VAL != 0 (and hence CST != 0), and *MULT_SET is true,
|
|
additionally compares CST and MULT, and if they are different,
|
|
returns false. Finally, if neither of these two cases occur,
|
|
true is returned, and if CST != 0, CST is stored to MULT and
|
|
MULT_SET is set to true. */
|
|
|
|
static bool
|
|
double_int_constant_multiple_p (double_int val, double_int div,
|
|
bool *mult_set, double_int *mult)
|
|
{
|
|
double_int rem, cst;
|
|
|
|
if (double_int_zero_p (val))
|
|
return true;
|
|
|
|
if (double_int_zero_p (div))
|
|
return false;
|
|
|
|
cst = double_int_sdivmod (val, div, FLOOR_DIV_EXPR, &rem);
|
|
if (!double_int_zero_p (rem))
|
|
return false;
|
|
|
|
if (*mult_set && !double_int_equal_p (*mult, cst))
|
|
return false;
|
|
|
|
*mult_set = true;
|
|
*mult = cst;
|
|
return true;
|
|
}
|
|
|
|
/* Returns true if VAL = X * DIV for some constant X. If this is the case,
|
|
X is stored to MULT. */
|
|
|
|
bool
|
|
aff_combination_constant_multiple_p (aff_tree *val, aff_tree *div,
|
|
double_int *mult)
|
|
{
|
|
bool mult_set = false;
|
|
unsigned i;
|
|
|
|
if (val->n == 0 && double_int_zero_p (val->offset))
|
|
{
|
|
*mult = double_int_zero;
|
|
return true;
|
|
}
|
|
if (val->n != div->n)
|
|
return false;
|
|
|
|
if (val->rest || div->rest)
|
|
return false;
|
|
|
|
if (!double_int_constant_multiple_p (val->offset, div->offset,
|
|
&mult_set, mult))
|
|
return false;
|
|
|
|
for (i = 0; i < div->n; i++)
|
|
{
|
|
struct aff_comb_elt *elt
|
|
= aff_combination_find_elt (val, div->elts[i].val, NULL);
|
|
if (!elt)
|
|
return false;
|
|
if (!double_int_constant_multiple_p (elt->coef, div->elts[i].coef,
|
|
&mult_set, mult))
|
|
return false;
|
|
}
|
|
|
|
gcc_assert (mult_set);
|
|
return true;
|
|
}
|
|
|
|
/* Prints the affine VAL to the FILE. */
|
|
|
|
static void
|
|
print_aff (FILE *file, aff_tree *val)
|
|
{
|
|
unsigned i;
|
|
bool uns = TYPE_UNSIGNED (val->type);
|
|
if (POINTER_TYPE_P (val->type))
|
|
uns = false;
|
|
fprintf (file, "{\n type = ");
|
|
print_generic_expr (file, val->type, TDF_VOPS|TDF_MEMSYMS);
|
|
fprintf (file, "\n offset = ");
|
|
dump_double_int (file, val->offset, uns);
|
|
if (val->n > 0)
|
|
{
|
|
fprintf (file, "\n elements = {\n");
|
|
for (i = 0; i < val->n; i++)
|
|
{
|
|
fprintf (file, " [%d] = ", i);
|
|
print_generic_expr (file, val->elts[i].val, TDF_VOPS|TDF_MEMSYMS);
|
|
|
|
fprintf (file, " * ");
|
|
dump_double_int (file, val->elts[i].coef, uns);
|
|
if (i != val->n - 1)
|
|
fprintf (file, ", \n");
|
|
}
|
|
fprintf (file, "\n }");
|
|
}
|
|
if (val->rest)
|
|
{
|
|
fprintf (file, "\n rest = ");
|
|
print_generic_expr (file, val->rest, TDF_VOPS|TDF_MEMSYMS);
|
|
}
|
|
fprintf (file, "\n}");
|
|
}
|
|
|
|
/* Prints the affine VAL to the standard error, used for debugging. */
|
|
|
|
DEBUG_FUNCTION void
|
|
debug_aff (aff_tree *val)
|
|
{
|
|
print_aff (stderr, val);
|
|
fprintf (stderr, "\n");
|
|
}
|
|
|
|
/* Returns address of the reference REF in ADDR. The size of the accessed
|
|
location is stored to SIZE. */
|
|
|
|
void
|
|
get_inner_reference_aff (tree ref, aff_tree *addr, double_int *size)
|
|
{
|
|
HOST_WIDE_INT bitsize, bitpos;
|
|
tree toff;
|
|
enum machine_mode mode;
|
|
int uns, vol;
|
|
aff_tree tmp;
|
|
tree base = get_inner_reference (ref, &bitsize, &bitpos, &toff, &mode,
|
|
&uns, &vol, false);
|
|
tree base_addr = build_fold_addr_expr (base);
|
|
|
|
/* ADDR = &BASE + TOFF + BITPOS / BITS_PER_UNIT. */
|
|
|
|
tree_to_aff_combination (base_addr, sizetype, addr);
|
|
|
|
if (toff)
|
|
{
|
|
tree_to_aff_combination (toff, sizetype, &tmp);
|
|
aff_combination_add (addr, &tmp);
|
|
}
|
|
|
|
aff_combination_const (&tmp, sizetype,
|
|
shwi_to_double_int (bitpos / BITS_PER_UNIT));
|
|
aff_combination_add (addr, &tmp);
|
|
|
|
*size = shwi_to_double_int ((bitsize + BITS_PER_UNIT - 1) / BITS_PER_UNIT);
|
|
}
|
|
|
|
/* Returns true if a region of size SIZE1 at position 0 and a region of
|
|
size SIZE2 at position DIFF cannot overlap. */
|
|
|
|
bool
|
|
aff_comb_cannot_overlap_p (aff_tree *diff, double_int size1, double_int size2)
|
|
{
|
|
double_int d, bound;
|
|
|
|
/* Unless the difference is a constant, we fail. */
|
|
if (diff->n != 0)
|
|
return false;
|
|
|
|
d = diff->offset;
|
|
if (double_int_negative_p (d))
|
|
{
|
|
/* The second object is before the first one, we succeed if the last
|
|
element of the second object is before the start of the first one. */
|
|
bound = double_int_add (d, double_int_add (size2, double_int_minus_one));
|
|
return double_int_negative_p (bound);
|
|
}
|
|
else
|
|
{
|
|
/* We succeed if the second object starts after the first one ends. */
|
|
return double_int_scmp (size1, d) <= 0;
|
|
}
|
|
}
|
|
|