gcc/gcc/gimple-match-head.c
Aldy Hernandez 45f4e2b01b Convert remaining passes to get_range_query.
This patch converts the remaining users of get_range_info and
get_ptr_nonnull to the get_range_query API.

No effort was made to move passes away from VR_ANTI_RANGE, or any other
use of deprecated methods.  This was a straight up conversion to the new
API, nothing else.

gcc/ChangeLog:

	* builtins.c (check_nul_terminated_array): Convert to get_range_query.
	(expand_builtin_strnlen): Same.
	(determine_block_size): Same.
	* fold-const.c (expr_not_equal_to): Same.
	* gimple-fold.c (size_must_be_zero_p): Same.
	* gimple-match-head.c: Include gimple-range.h.
	* gimple-pretty-print.c (dump_ssaname_info): Convert to get_range_query.
	* gimple-ssa-warn-restrict.c
	(builtin_memref::extend_offset_range): Same.
	* graphite-sese-to-poly.c (add_param_constraints): Same.
	* internal-fn.c (get_min_precision): Same.
	* ipa-fnsummary.c (set_switch_stmt_execution_predicate): Same.
	* ipa-prop.c (ipa_compute_jump_functions_for_edge): Same.
	* match.pd: Same.
	* tree-data-ref.c (split_constant_offset): Same.
	(dr_step_indicator): Same.
	* tree-dfa.c (get_ref_base_and_extent): Same.
	* tree-scalar-evolution.c (iv_can_overflow_p): Same.
	* tree-ssa-loop-niter.c (refine_value_range_using_guard): Same.
	(determine_value_range): Same.
	(record_nonwrapping_iv): Same.
	(infer_loop_bounds_from_signedness): Same.
	(scev_var_range_cant_overflow): Same.
	* tree-ssa-phiopt.c (two_value_replacement): Same.
	* tree-ssa-pre.c (insert_into_preds_of_block): Same.
	* tree-ssa-reassoc.c (optimize_range_tests_to_bit_test): Same.
	* tree-ssa-strlen.c (handle_builtin_stxncpy_strncat): Same.
	(get_range): Same.
	(dump_strlen_info): Same.
	(set_strlen_range): Same.
	(maybe_diag_stxncpy_trunc): Same.
	(get_len_or_size): Same.
	(handle_integral_assign): Same.
	* tree-ssa-structalias.c (find_what_p_points_to): Same.
	* tree-ssa-uninit.c (find_var_cmp_const): Same.
	* tree-switch-conversion.c (bit_test_cluster::emit): Same.
	* tree-vect-patterns.c (vect_get_range_info): Same.
	(vect_recog_divmod_pattern): Same.
	* tree-vrp.c (intersect_range_with_nonzero_bits): Same.
	(register_edge_assert_for_2): Same.
	(determine_value_range_1): Same.
	* tree.c (get_range_pos_neg): Same.
	* vr-values.c (vr_values::get_lattice_entry): Same.
	(vr_values::update_value_range): Same.
	(simplify_conversion_using_ranges): Same.
2021-05-26 21:31:27 +02:00

1263 lines
36 KiB
C

/* Preamble and helpers for the autogenerated gimple-match.c file.
Copyright (C) 2014-2021 Free Software Foundation, Inc.
This file is part of GCC.
GCC is free software; you can redistribute it and/or modify it under
the terms of the GNU General Public License as published by the Free
Software Foundation; either version 3, or (at your option) any later
version.
GCC is distributed in the hope that it will be useful, but WITHOUT ANY
WARRANTY; without even the implied warranty of MERCHANTABILITY or
FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License
for more details.
You should have received a copy of the GNU General Public License
along with GCC; see the file COPYING3. If not see
<http://www.gnu.org/licenses/>. */
#include "config.h"
#include "system.h"
#include "coretypes.h"
#include "backend.h"
#include "target.h"
#include "rtl.h"
#include "tree.h"
#include "gimple.h"
#include "ssa.h"
#include "cgraph.h"
#include "vec-perm-indices.h"
#include "fold-const.h"
#include "fold-const-call.h"
#include "stor-layout.h"
#include "gimple-fold.h"
#include "calls.h"
#include "tree-dfa.h"
#include "builtins.h"
#include "gimple-match.h"
#include "tree-pass.h"
#include "internal-fn.h"
#include "case-cfn-macros.h"
#include "gimplify.h"
#include "optabs-tree.h"
#include "tree-eh.h"
#include "dbgcnt.h"
#include "tm.h"
#include "gimple-range.h"
/* Forward declarations of the private auto-generated matchers.
They expect valueized operands in canonical order and do not
perform simplification of all-constant operands. */
static bool gimple_simplify (gimple_match_op *, gimple_seq *, tree (*)(tree),
code_helper, tree, tree);
static bool gimple_simplify (gimple_match_op *, gimple_seq *, tree (*)(tree),
code_helper, tree, tree, tree);
static bool gimple_simplify (gimple_match_op *, gimple_seq *, tree (*)(tree),
code_helper, tree, tree, tree, tree);
static bool gimple_simplify (gimple_match_op *, gimple_seq *, tree (*)(tree),
code_helper, tree, tree, tree, tree, tree);
static bool gimple_simplify (gimple_match_op *, gimple_seq *, tree (*)(tree),
code_helper, tree, tree, tree, tree, tree, tree);
static bool gimple_resimplify1 (gimple_seq *, gimple_match_op *,
tree (*)(tree));
static bool gimple_resimplify2 (gimple_seq *, gimple_match_op *,
tree (*)(tree));
static bool gimple_resimplify3 (gimple_seq *, gimple_match_op *,
tree (*)(tree));
static bool gimple_resimplify4 (gimple_seq *, gimple_match_op *,
tree (*)(tree));
static bool gimple_resimplify5 (gimple_seq *, gimple_match_op *,
tree (*)(tree));
const unsigned int gimple_match_op::MAX_NUM_OPS;
/* Return whether T is a constant that we'll dispatch to fold to
evaluate fully constant expressions. */
static inline bool
constant_for_folding (tree t)
{
return (CONSTANT_CLASS_P (t)
/* The following is only interesting to string builtins. */
|| (TREE_CODE (t) == ADDR_EXPR
&& TREE_CODE (TREE_OPERAND (t, 0)) == STRING_CST));
}
/* Try to convert conditional operation ORIG_OP into an IFN_COND_*
operation. Return true on success, storing the new operation in NEW_OP. */
static bool
convert_conditional_op (gimple_match_op *orig_op,
gimple_match_op *new_op)
{
internal_fn ifn;
if (orig_op->code.is_tree_code ())
ifn = get_conditional_internal_fn ((tree_code) orig_op->code);
else
{
combined_fn cfn = orig_op->code;
if (!internal_fn_p (cfn))
return false;
ifn = get_conditional_internal_fn (as_internal_fn (cfn));
}
if (ifn == IFN_LAST)
return false;
unsigned int num_ops = orig_op->num_ops;
new_op->set_op (as_combined_fn (ifn), orig_op->type, num_ops + 2);
new_op->ops[0] = orig_op->cond.cond;
for (unsigned int i = 0; i < num_ops; ++i)
new_op->ops[i + 1] = orig_op->ops[i];
tree else_value = orig_op->cond.else_value;
if (!else_value)
else_value = targetm.preferred_else_value (ifn, orig_op->type,
num_ops, orig_op->ops);
new_op->ops[num_ops + 1] = else_value;
return true;
}
/* RES_OP is the result of a simplification. If it is conditional,
try to replace it with the equivalent UNCOND form, such as an
IFN_COND_* call or a VEC_COND_EXPR. Also try to resimplify the
result of the replacement if appropriate, adding any new statements to
SEQ and using VALUEIZE as the valueization function. Return true if
this resimplification occurred and resulted in at least one change. */
static bool
maybe_resimplify_conditional_op (gimple_seq *seq, gimple_match_op *res_op,
tree (*valueize) (tree))
{
if (!res_op->cond.cond)
return false;
if (!res_op->cond.else_value
&& res_op->code.is_tree_code ())
{
/* The "else" value doesn't matter. If the "then" value is a
gimple value, just use it unconditionally. This isn't a
simplification in itself, since there was no operation to
build in the first place. */
if (gimple_simplified_result_is_gimple_val (res_op))
{
res_op->cond.cond = NULL_TREE;
return false;
}
/* Likewise if the operation would not trap. */
bool honor_trapv = (INTEGRAL_TYPE_P (res_op->type)
&& TYPE_OVERFLOW_TRAPS (res_op->type));
tree_code op_code = (tree_code) res_op->code;
bool op_could_trap;
/* COND_EXPR will trap if, and only if, the condition
traps and hence we have to check this. For all other operations, we
don't need to consider the operands. */
if (op_code == COND_EXPR)
op_could_trap = generic_expr_could_trap_p (res_op->ops[0]);
else
op_could_trap = operation_could_trap_p ((tree_code) res_op->code,
FLOAT_TYPE_P (res_op->type),
honor_trapv,
res_op->op_or_null (1));
if (!op_could_trap)
{
res_op->cond.cond = NULL_TREE;
return false;
}
}
/* If the "then" value is a gimple value and the "else" value matters,
create a VEC_COND_EXPR between them, then see if it can be further
simplified. */
gimple_match_op new_op;
if (res_op->cond.else_value
&& VECTOR_TYPE_P (res_op->type)
&& gimple_simplified_result_is_gimple_val (res_op))
{
new_op.set_op (VEC_COND_EXPR, res_op->type,
res_op->cond.cond, res_op->ops[0],
res_op->cond.else_value);
*res_op = new_op;
return gimple_resimplify3 (seq, res_op, valueize);
}
/* Otherwise try rewriting the operation as an IFN_COND_* call.
Again, this isn't a simplification in itself, since it's what
RES_OP already described. */
if (convert_conditional_op (res_op, &new_op))
*res_op = new_op;
return false;
}
/* Helper that matches and simplifies the toplevel result from
a gimple_simplify run (where we don't want to build
a stmt in case it's used in in-place folding). Replaces
RES_OP with a simplified and/or canonicalized result and
returns whether any change was made. */
static bool
gimple_resimplify1 (gimple_seq *seq, gimple_match_op *res_op,
tree (*valueize)(tree))
{
if (constant_for_folding (res_op->ops[0]))
{
tree tem = NULL_TREE;
if (res_op->code.is_tree_code ())
{
tree_code code = res_op->code;
if (IS_EXPR_CODE_CLASS (TREE_CODE_CLASS (code))
&& TREE_CODE_LENGTH (code) == 1)
tem = const_unop (res_op->code, res_op->type, res_op->ops[0]);
}
else
tem = fold_const_call (combined_fn (res_op->code), res_op->type,
res_op->ops[0]);
if (tem != NULL_TREE
&& CONSTANT_CLASS_P (tem))
{
if (TREE_OVERFLOW_P (tem))
tem = drop_tree_overflow (tem);
res_op->set_value (tem);
maybe_resimplify_conditional_op (seq, res_op, valueize);
return true;
}
}
/* Limit recursion, there are cases like PR80887 and others, for
example when value-numbering presents us with unfolded expressions
that we are really not prepared to handle without eventual
oscillation like ((_50 + 0) + 8) where _50 gets mapped to _50
itself as available expression. */
static unsigned depth;
if (depth > 10)
{
if (dump_file && (dump_flags & TDF_FOLDING))
fprintf (dump_file, "Aborting expression simplification due to "
"deep recursion\n");
return false;
}
++depth;
gimple_match_op res_op2 (*res_op);
if (gimple_simplify (&res_op2, seq, valueize,
res_op->code, res_op->type, res_op->ops[0]))
{
--depth;
*res_op = res_op2;
return true;
}
--depth;
if (maybe_resimplify_conditional_op (seq, res_op, valueize))
return true;
return false;
}
/* Helper that matches and simplifies the toplevel result from
a gimple_simplify run (where we don't want to build
a stmt in case it's used in in-place folding). Replaces
RES_OP with a simplified and/or canonicalized result and
returns whether any change was made. */
static bool
gimple_resimplify2 (gimple_seq *seq, gimple_match_op *res_op,
tree (*valueize)(tree))
{
if (constant_for_folding (res_op->ops[0])
&& constant_for_folding (res_op->ops[1]))
{
tree tem = NULL_TREE;
if (res_op->code.is_tree_code ())
{
tree_code code = res_op->code;
if (IS_EXPR_CODE_CLASS (TREE_CODE_CLASS (code))
&& TREE_CODE_LENGTH (code) == 2)
tem = const_binop (res_op->code, res_op->type,
res_op->ops[0], res_op->ops[1]);
}
else
tem = fold_const_call (combined_fn (res_op->code), res_op->type,
res_op->ops[0], res_op->ops[1]);
if (tem != NULL_TREE
&& CONSTANT_CLASS_P (tem))
{
if (TREE_OVERFLOW_P (tem))
tem = drop_tree_overflow (tem);
res_op->set_value (tem);
maybe_resimplify_conditional_op (seq, res_op, valueize);
return true;
}
}
/* Canonicalize operand order. */
bool canonicalized = false;
if (res_op->code.is_tree_code ()
&& (TREE_CODE_CLASS ((enum tree_code) res_op->code) == tcc_comparison
|| commutative_tree_code (res_op->code))
&& tree_swap_operands_p (res_op->ops[0], res_op->ops[1]))
{
std::swap (res_op->ops[0], res_op->ops[1]);
if (TREE_CODE_CLASS ((enum tree_code) res_op->code) == tcc_comparison)
res_op->code = swap_tree_comparison (res_op->code);
canonicalized = true;
}
/* Limit recursion, see gimple_resimplify1. */
static unsigned depth;
if (depth > 10)
{
if (dump_file && (dump_flags & TDF_FOLDING))
fprintf (dump_file, "Aborting expression simplification due to "
"deep recursion\n");
return false;
}
++depth;
gimple_match_op res_op2 (*res_op);
if (gimple_simplify (&res_op2, seq, valueize,
res_op->code, res_op->type,
res_op->ops[0], res_op->ops[1]))
{
--depth;
*res_op = res_op2;
return true;
}
--depth;
if (maybe_resimplify_conditional_op (seq, res_op, valueize))
return true;
return canonicalized;
}
/* Helper that matches and simplifies the toplevel result from
a gimple_simplify run (where we don't want to build
a stmt in case it's used in in-place folding). Replaces
RES_OP with a simplified and/or canonicalized result and
returns whether any change was made. */
static bool
gimple_resimplify3 (gimple_seq *seq, gimple_match_op *res_op,
tree (*valueize)(tree))
{
if (constant_for_folding (res_op->ops[0])
&& constant_for_folding (res_op->ops[1])
&& constant_for_folding (res_op->ops[2]))
{
tree tem = NULL_TREE;
if (res_op->code.is_tree_code ())
{
tree_code code = res_op->code;
if (IS_EXPR_CODE_CLASS (TREE_CODE_CLASS (code))
&& TREE_CODE_LENGTH (code) == 3)
tem = fold_ternary/*_to_constant*/ (res_op->code, res_op->type,
res_op->ops[0], res_op->ops[1],
res_op->ops[2]);
}
else
tem = fold_const_call (combined_fn (res_op->code), res_op->type,
res_op->ops[0], res_op->ops[1], res_op->ops[2]);
if (tem != NULL_TREE
&& CONSTANT_CLASS_P (tem))
{
if (TREE_OVERFLOW_P (tem))
tem = drop_tree_overflow (tem);
res_op->set_value (tem);
maybe_resimplify_conditional_op (seq, res_op, valueize);
return true;
}
}
/* Canonicalize operand order. */
bool canonicalized = false;
if (res_op->code.is_tree_code ()
&& commutative_ternary_tree_code (res_op->code)
&& tree_swap_operands_p (res_op->ops[0], res_op->ops[1]))
{
std::swap (res_op->ops[0], res_op->ops[1]);
canonicalized = true;
}
/* Limit recursion, see gimple_resimplify1. */
static unsigned depth;
if (depth > 10)
{
if (dump_file && (dump_flags & TDF_FOLDING))
fprintf (dump_file, "Aborting expression simplification due to "
"deep recursion\n");
return false;
}
++depth;
gimple_match_op res_op2 (*res_op);
if (gimple_simplify (&res_op2, seq, valueize,
res_op->code, res_op->type,
res_op->ops[0], res_op->ops[1], res_op->ops[2]))
{
--depth;
*res_op = res_op2;
return true;
}
--depth;
if (maybe_resimplify_conditional_op (seq, res_op, valueize))
return true;
return canonicalized;
}
/* Helper that matches and simplifies the toplevel result from
a gimple_simplify run (where we don't want to build
a stmt in case it's used in in-place folding). Replaces
RES_OP with a simplified and/or canonicalized result and
returns whether any change was made. */
static bool
gimple_resimplify4 (gimple_seq *seq, gimple_match_op *res_op,
tree (*valueize)(tree))
{
/* No constant folding is defined for four-operand functions. */
/* Limit recursion, see gimple_resimplify1. */
static unsigned depth;
if (depth > 10)
{
if (dump_file && (dump_flags & TDF_FOLDING))
fprintf (dump_file, "Aborting expression simplification due to "
"deep recursion\n");
return false;
}
++depth;
gimple_match_op res_op2 (*res_op);
if (gimple_simplify (&res_op2, seq, valueize,
res_op->code, res_op->type,
res_op->ops[0], res_op->ops[1], res_op->ops[2],
res_op->ops[3]))
{
--depth;
*res_op = res_op2;
return true;
}
--depth;
if (maybe_resimplify_conditional_op (seq, res_op, valueize))
return true;
return false;
}
/* Helper that matches and simplifies the toplevel result from
a gimple_simplify run (where we don't want to build
a stmt in case it's used in in-place folding). Replaces
RES_OP with a simplified and/or canonicalized result and
returns whether any change was made. */
static bool
gimple_resimplify5 (gimple_seq *seq, gimple_match_op *res_op,
tree (*valueize)(tree))
{
/* No constant folding is defined for five-operand functions. */
gimple_match_op res_op2 (*res_op);
if (gimple_simplify (&res_op2, seq, valueize,
res_op->code, res_op->type,
res_op->ops[0], res_op->ops[1], res_op->ops[2],
res_op->ops[3], res_op->ops[4]))
{
*res_op = res_op2;
return true;
}
if (maybe_resimplify_conditional_op (seq, res_op, valueize))
return true;
return false;
}
/* Match and simplify the toplevel valueized operation THIS.
Replaces THIS with a simplified and/or canonicalized result and
returns whether any change was made. */
bool
gimple_match_op::resimplify (gimple_seq *seq, tree (*valueize)(tree))
{
switch (num_ops)
{
case 1:
return gimple_resimplify1 (seq, this, valueize);
case 2:
return gimple_resimplify2 (seq, this, valueize);
case 3:
return gimple_resimplify3 (seq, this, valueize);
case 4:
return gimple_resimplify4 (seq, this, valueize);
case 5:
return gimple_resimplify5 (seq, this, valueize);
default:
gcc_unreachable ();
}
}
/* If in GIMPLE the operation described by RES_OP should be single-rhs,
build a GENERIC tree for that expression and update RES_OP accordingly. */
void
maybe_build_generic_op (gimple_match_op *res_op)
{
tree_code code = (tree_code) res_op->code;
tree val;
switch (code)
{
case REALPART_EXPR:
case IMAGPART_EXPR:
case VIEW_CONVERT_EXPR:
val = build1 (code, res_op->type, res_op->ops[0]);
res_op->set_value (val);
break;
case BIT_FIELD_REF:
val = build3 (code, res_op->type, res_op->ops[0], res_op->ops[1],
res_op->ops[2]);
REF_REVERSE_STORAGE_ORDER (val) = res_op->reverse;
res_op->set_value (val);
break;
default:;
}
}
tree (*mprts_hook) (gimple_match_op *);
/* Try to build RES_OP, which is known to be a call to FN. Return null
if the target doesn't support the function. */
static gcall *
build_call_internal (internal_fn fn, gimple_match_op *res_op)
{
if (direct_internal_fn_p (fn))
{
tree_pair types = direct_internal_fn_types (fn, res_op->type,
res_op->ops);
if (!direct_internal_fn_supported_p (fn, types, OPTIMIZE_FOR_BOTH))
return NULL;
}
return gimple_build_call_internal (fn, res_op->num_ops,
res_op->op_or_null (0),
res_op->op_or_null (1),
res_op->op_or_null (2),
res_op->op_or_null (3),
res_op->op_or_null (4));
}
/* Push the exploded expression described by RES_OP as a statement to
SEQ if necessary and return a gimple value denoting the value of the
expression. If RES is not NULL then the result will be always RES
and even gimple values are pushed to SEQ. */
tree
maybe_push_res_to_seq (gimple_match_op *res_op, gimple_seq *seq, tree res)
{
tree *ops = res_op->ops;
unsigned num_ops = res_op->num_ops;
/* The caller should have converted conditional operations into an UNCOND
form and resimplified as appropriate. The conditional form only
survives this far if that conversion failed. */
if (res_op->cond.cond)
return NULL_TREE;
if (res_op->code.is_tree_code ())
{
if (!res
&& gimple_simplified_result_is_gimple_val (res_op))
return ops[0];
if (mprts_hook)
{
tree tem = mprts_hook (res_op);
if (tem)
return tem;
}
}
if (!seq)
return NULL_TREE;
/* Play safe and do not allow abnormals to be mentioned in
newly created statements. */
for (unsigned int i = 0; i < num_ops; ++i)
if (TREE_CODE (ops[i]) == SSA_NAME
&& SSA_NAME_OCCURS_IN_ABNORMAL_PHI (ops[i]))
return NULL_TREE;
if (num_ops > 0 && COMPARISON_CLASS_P (ops[0]))
for (unsigned int i = 0; i < 2; ++i)
if (TREE_CODE (TREE_OPERAND (ops[0], i)) == SSA_NAME
&& SSA_NAME_OCCURS_IN_ABNORMAL_PHI (TREE_OPERAND (ops[0], i)))
return NULL_TREE;
if (res_op->code.is_tree_code ())
{
if (!res)
{
if (gimple_in_ssa_p (cfun))
res = make_ssa_name (res_op->type);
else
res = create_tmp_reg (res_op->type);
}
maybe_build_generic_op (res_op);
gimple *new_stmt = gimple_build_assign (res, res_op->code,
res_op->op_or_null (0),
res_op->op_or_null (1),
res_op->op_or_null (2));
gimple_seq_add_stmt_without_update (seq, new_stmt);
return res;
}
else
{
gcc_assert (num_ops != 0);
combined_fn fn = res_op->code;
gcall *new_stmt = NULL;
if (internal_fn_p (fn))
{
/* Generate the given function if we can. */
internal_fn ifn = as_internal_fn (fn);
new_stmt = build_call_internal (ifn, res_op);
if (!new_stmt)
return NULL_TREE;
}
else
{
/* Find the function we want to call. */
tree decl = builtin_decl_implicit (as_builtin_fn (fn));
if (!decl)
return NULL;
/* We can't and should not emit calls to non-const functions. */
if (!(flags_from_decl_or_type (decl) & ECF_CONST))
return NULL;
new_stmt = gimple_build_call (decl, num_ops,
res_op->op_or_null (0),
res_op->op_or_null (1),
res_op->op_or_null (2),
res_op->op_or_null (3),
res_op->op_or_null (4));
}
if (!res)
{
if (gimple_in_ssa_p (cfun))
res = make_ssa_name (res_op->type);
else
res = create_tmp_reg (res_op->type);
}
gimple_call_set_lhs (new_stmt, res);
gimple_seq_add_stmt_without_update (seq, new_stmt);
return res;
}
}
/* Public API overloads follow for operation being tree_code or
built_in_function and for one to three operands or arguments.
They return NULL_TREE if nothing could be simplified or
the resulting simplified value with parts pushed to SEQ.
If SEQ is NULL then if the simplification needs to create
new stmts it will fail. If VALUEIZE is non-NULL then all
SSA names will be valueized using that hook prior to
applying simplifications. */
/* Unary ops. */
tree
gimple_simplify (enum tree_code code, tree type,
tree op0,
gimple_seq *seq, tree (*valueize)(tree))
{
if (constant_for_folding (op0))
{
tree res = const_unop (code, type, op0);
if (res != NULL_TREE
&& CONSTANT_CLASS_P (res))
return res;
}
gimple_match_op res_op;
if (!gimple_simplify (&res_op, seq, valueize, code, type, op0))
return NULL_TREE;
return maybe_push_res_to_seq (&res_op, seq);
}
/* Binary ops. */
tree
gimple_simplify (enum tree_code code, tree type,
tree op0, tree op1,
gimple_seq *seq, tree (*valueize)(tree))
{
if (constant_for_folding (op0) && constant_for_folding (op1))
{
tree res = const_binop (code, type, op0, op1);
if (res != NULL_TREE
&& CONSTANT_CLASS_P (res))
return res;
}
/* Canonicalize operand order both for matching and fallback stmt
generation. */
if ((commutative_tree_code (code)
|| TREE_CODE_CLASS (code) == tcc_comparison)
&& tree_swap_operands_p (op0, op1))
{
std::swap (op0, op1);
if (TREE_CODE_CLASS (code) == tcc_comparison)
code = swap_tree_comparison (code);
}
gimple_match_op res_op;
if (!gimple_simplify (&res_op, seq, valueize, code, type, op0, op1))
return NULL_TREE;
return maybe_push_res_to_seq (&res_op, seq);
}
/* Ternary ops. */
tree
gimple_simplify (enum tree_code code, tree type,
tree op0, tree op1, tree op2,
gimple_seq *seq, tree (*valueize)(tree))
{
if (constant_for_folding (op0) && constant_for_folding (op1)
&& constant_for_folding (op2))
{
tree res = fold_ternary/*_to_constant */ (code, type, op0, op1, op2);
if (res != NULL_TREE
&& CONSTANT_CLASS_P (res))
return res;
}
/* Canonicalize operand order both for matching and fallback stmt
generation. */
if (commutative_ternary_tree_code (code)
&& tree_swap_operands_p (op0, op1))
std::swap (op0, op1);
gimple_match_op res_op;
if (!gimple_simplify (&res_op, seq, valueize, code, type, op0, op1, op2))
return NULL_TREE;
return maybe_push_res_to_seq (&res_op, seq);
}
/* Builtin or internal function with one argument. */
tree
gimple_simplify (combined_fn fn, tree type,
tree arg0,
gimple_seq *seq, tree (*valueize)(tree))
{
if (constant_for_folding (arg0))
{
tree res = fold_const_call (fn, type, arg0);
if (res && CONSTANT_CLASS_P (res))
return res;
}
gimple_match_op res_op;
if (!gimple_simplify (&res_op, seq, valueize, fn, type, arg0))
return NULL_TREE;
return maybe_push_res_to_seq (&res_op, seq);
}
/* Builtin or internal function with two arguments. */
tree
gimple_simplify (combined_fn fn, tree type,
tree arg0, tree arg1,
gimple_seq *seq, tree (*valueize)(tree))
{
if (constant_for_folding (arg0)
&& constant_for_folding (arg1))
{
tree res = fold_const_call (fn, type, arg0, arg1);
if (res && CONSTANT_CLASS_P (res))
return res;
}
gimple_match_op res_op;
if (!gimple_simplify (&res_op, seq, valueize, fn, type, arg0, arg1))
return NULL_TREE;
return maybe_push_res_to_seq (&res_op, seq);
}
/* Builtin or internal function with three arguments. */
tree
gimple_simplify (combined_fn fn, tree type,
tree arg0, tree arg1, tree arg2,
gimple_seq *seq, tree (*valueize)(tree))
{
if (constant_for_folding (arg0)
&& constant_for_folding (arg1)
&& constant_for_folding (arg2))
{
tree res = fold_const_call (fn, type, arg0, arg1, arg2);
if (res && CONSTANT_CLASS_P (res))
return res;
}
gimple_match_op res_op;
if (!gimple_simplify (&res_op, seq, valueize, fn, type, arg0, arg1, arg2))
return NULL_TREE;
return maybe_push_res_to_seq (&res_op, seq);
}
/* Helper for gimple_simplify valueizing OP using VALUEIZE and setting
VALUEIZED to true if valueization changed OP. */
static inline tree
do_valueize (tree op, tree (*valueize)(tree), bool &valueized)
{
if (valueize && TREE_CODE (op) == SSA_NAME)
{
tree tem = valueize (op);
if (tem && tem != op)
{
op = tem;
valueized = true;
}
}
return op;
}
/* If RES_OP is a call to a conditional internal function, try simplifying
the associated unconditional operation and using the result to build
a new conditional operation. For example, if RES_OP is:
IFN_COND_ADD (COND, A, B, ELSE)
try simplifying (plus A B) and using the result to build a replacement
for the whole IFN_COND_ADD.
Return true if this approach led to a simplification, otherwise leave
RES_OP unchanged (and so suitable for other simplifications). When
returning true, add any new statements to SEQ and use VALUEIZE as the
valueization function.
RES_OP is known to be a call to IFN. */
static bool
try_conditional_simplification (internal_fn ifn, gimple_match_op *res_op,
gimple_seq *seq, tree (*valueize) (tree))
{
code_helper op;
tree_code code = conditional_internal_fn_code (ifn);
if (code != ERROR_MARK)
op = code;
else
{
ifn = get_unconditional_internal_fn (ifn);
if (ifn == IFN_LAST)
return false;
op = as_combined_fn (ifn);
}
unsigned int num_ops = res_op->num_ops;
gimple_match_op cond_op (gimple_match_cond (res_op->ops[0],
res_op->ops[num_ops - 1]),
op, res_op->type, num_ops - 2);
memcpy (cond_op.ops, res_op->ops + 1, (num_ops - 1) * sizeof *cond_op.ops);
switch (num_ops - 2)
{
case 2:
if (!gimple_resimplify2 (seq, &cond_op, valueize))
return false;
break;
case 3:
if (!gimple_resimplify3 (seq, &cond_op, valueize))
return false;
break;
default:
gcc_unreachable ();
}
*res_op = cond_op;
maybe_resimplify_conditional_op (seq, res_op, valueize);
return true;
}
/* The main STMT based simplification entry. It is used by the fold_stmt
and the fold_stmt_to_constant APIs. */
bool
gimple_simplify (gimple *stmt, gimple_match_op *res_op, gimple_seq *seq,
tree (*valueize)(tree), tree (*top_valueize)(tree))
{
switch (gimple_code (stmt))
{
case GIMPLE_ASSIGN:
{
enum tree_code code = gimple_assign_rhs_code (stmt);
tree type = TREE_TYPE (gimple_assign_lhs (stmt));
switch (gimple_assign_rhs_class (stmt))
{
case GIMPLE_SINGLE_RHS:
if (code == REALPART_EXPR
|| code == IMAGPART_EXPR
|| code == VIEW_CONVERT_EXPR)
{
tree op0 = TREE_OPERAND (gimple_assign_rhs1 (stmt), 0);
bool valueized = false;
op0 = do_valueize (op0, top_valueize, valueized);
res_op->set_op (code, type, op0);
return (gimple_resimplify1 (seq, res_op, valueize)
|| valueized);
}
else if (code == BIT_FIELD_REF)
{
tree rhs1 = gimple_assign_rhs1 (stmt);
tree op0 = TREE_OPERAND (rhs1, 0);
bool valueized = false;
op0 = do_valueize (op0, top_valueize, valueized);
res_op->set_op (code, type, op0,
TREE_OPERAND (rhs1, 1),
TREE_OPERAND (rhs1, 2),
REF_REVERSE_STORAGE_ORDER (rhs1));
if (res_op->reverse)
return valueized;
return (gimple_resimplify3 (seq, res_op, valueize)
|| valueized);
}
else if (code == SSA_NAME
&& top_valueize)
{
tree op0 = gimple_assign_rhs1 (stmt);
tree valueized = top_valueize (op0);
if (!valueized || op0 == valueized)
return false;
res_op->set_op (TREE_CODE (op0), type, valueized);
return true;
}
break;
case GIMPLE_UNARY_RHS:
{
tree rhs1 = gimple_assign_rhs1 (stmt);
bool valueized = false;
rhs1 = do_valueize (rhs1, top_valueize, valueized);
res_op->set_op (code, type, rhs1);
return (gimple_resimplify1 (seq, res_op, valueize)
|| valueized);
}
case GIMPLE_BINARY_RHS:
{
tree rhs1 = gimple_assign_rhs1 (stmt);
tree rhs2 = gimple_assign_rhs2 (stmt);
bool valueized = false;
rhs1 = do_valueize (rhs1, top_valueize, valueized);
rhs2 = do_valueize (rhs2, top_valueize, valueized);
res_op->set_op (code, type, rhs1, rhs2);
return (gimple_resimplify2 (seq, res_op, valueize)
|| valueized);
}
case GIMPLE_TERNARY_RHS:
{
bool valueized = false;
tree rhs1 = gimple_assign_rhs1 (stmt);
/* If this is a COND_EXPR first try to simplify an
embedded GENERIC condition. */
if (code == COND_EXPR)
{
if (COMPARISON_CLASS_P (rhs1))
{
tree lhs = TREE_OPERAND (rhs1, 0);
tree rhs = TREE_OPERAND (rhs1, 1);
lhs = do_valueize (lhs, top_valueize, valueized);
rhs = do_valueize (rhs, top_valueize, valueized);
gimple_match_op res_op2 (res_op->cond, TREE_CODE (rhs1),
TREE_TYPE (rhs1), lhs, rhs);
if ((gimple_resimplify2 (seq, &res_op2, valueize)
|| valueized)
&& res_op2.code.is_tree_code ())
{
valueized = true;
if (TREE_CODE_CLASS ((enum tree_code) res_op2.code)
== tcc_comparison)
rhs1 = build2 (res_op2.code, TREE_TYPE (rhs1),
res_op2.ops[0], res_op2.ops[1]);
else if (res_op2.code == SSA_NAME
|| res_op2.code == INTEGER_CST
|| res_op2.code == VECTOR_CST)
rhs1 = res_op2.ops[0];
else
valueized = false;
}
}
}
tree rhs2 = gimple_assign_rhs2 (stmt);
tree rhs3 = gimple_assign_rhs3 (stmt);
rhs1 = do_valueize (rhs1, top_valueize, valueized);
rhs2 = do_valueize (rhs2, top_valueize, valueized);
rhs3 = do_valueize (rhs3, top_valueize, valueized);
res_op->set_op (code, type, rhs1, rhs2, rhs3);
return (gimple_resimplify3 (seq, res_op, valueize)
|| valueized);
}
default:
gcc_unreachable ();
}
break;
}
case GIMPLE_CALL:
/* ??? This way we can't simplify calls with side-effects. */
if (gimple_call_lhs (stmt) != NULL_TREE
&& gimple_call_num_args (stmt) >= 1
&& gimple_call_num_args (stmt) <= 5)
{
bool valueized = false;
combined_fn cfn;
if (gimple_call_internal_p (stmt))
cfn = as_combined_fn (gimple_call_internal_fn (stmt));
else
{
tree fn = gimple_call_fn (stmt);
if (!fn)
return false;
fn = do_valueize (fn, top_valueize, valueized);
if (TREE_CODE (fn) != ADDR_EXPR
|| TREE_CODE (TREE_OPERAND (fn, 0)) != FUNCTION_DECL)
return false;
tree decl = TREE_OPERAND (fn, 0);
if (DECL_BUILT_IN_CLASS (decl) != BUILT_IN_NORMAL
|| !gimple_builtin_call_types_compatible_p (stmt, decl))
return false;
cfn = as_combined_fn (DECL_FUNCTION_CODE (decl));
}
unsigned int num_args = gimple_call_num_args (stmt);
res_op->set_op (cfn, TREE_TYPE (gimple_call_lhs (stmt)), num_args);
for (unsigned i = 0; i < num_args; ++i)
{
tree arg = gimple_call_arg (stmt, i);
res_op->ops[i] = do_valueize (arg, top_valueize, valueized);
}
if (internal_fn_p (cfn)
&& try_conditional_simplification (as_internal_fn (cfn),
res_op, seq, valueize))
return true;
switch (num_args)
{
case 1:
return (gimple_resimplify1 (seq, res_op, valueize)
|| valueized);
case 2:
return (gimple_resimplify2 (seq, res_op, valueize)
|| valueized);
case 3:
return (gimple_resimplify3 (seq, res_op, valueize)
|| valueized);
case 4:
return (gimple_resimplify4 (seq, res_op, valueize)
|| valueized);
case 5:
return (gimple_resimplify5 (seq, res_op, valueize)
|| valueized);
default:
gcc_unreachable ();
}
}
break;
case GIMPLE_COND:
{
tree lhs = gimple_cond_lhs (stmt);
tree rhs = gimple_cond_rhs (stmt);
bool valueized = false;
lhs = do_valueize (lhs, top_valueize, valueized);
rhs = do_valueize (rhs, top_valueize, valueized);
res_op->set_op (gimple_cond_code (stmt), boolean_type_node, lhs, rhs);
return (gimple_resimplify2 (seq, res_op, valueize)
|| valueized);
}
default:
break;
}
return false;
}
/* Helper for the autogenerated code, valueize OP. */
inline tree
do_valueize (tree (*valueize)(tree), tree op)
{
if (valueize && TREE_CODE (op) == SSA_NAME)
{
tree tem = valueize (op);
if (tem)
return tem;
}
return op;
}
/* Helper for the autogenerated code, get at the definition of NAME when
VALUEIZE allows that. */
inline gimple *
get_def (tree (*valueize)(tree), tree name)
{
if (valueize && ! valueize (name))
return NULL;
return SSA_NAME_DEF_STMT (name);
}
/* Routine to determine if the types T1 and T2 are effectively
the same for GIMPLE. If T1 or T2 is not a type, the test
applies to their TREE_TYPE. */
static inline bool
types_match (tree t1, tree t2)
{
if (!TYPE_P (t1))
t1 = TREE_TYPE (t1);
if (!TYPE_P (t2))
t2 = TREE_TYPE (t2);
return types_compatible_p (t1, t2);
}
/* Return if T has a single use. For GIMPLE, we also allow any
non-SSA_NAME (ie constants) and zero uses to cope with uses
that aren't linked up yet. */
static inline bool
single_use (tree t)
{
return TREE_CODE (t) != SSA_NAME || has_zero_uses (t) || has_single_use (t);
}
/* Return true if math operations should be canonicalized,
e.g. sqrt(sqrt(x)) -> pow(x, 0.25). */
static inline bool
canonicalize_math_p ()
{
return !cfun || (cfun->curr_properties & PROP_gimple_opt_math) == 0;
}
/* Return true if math operations that are beneficial only after
vectorization should be canonicalized. */
static inline bool
canonicalize_math_after_vectorization_p ()
{
return !cfun || (cfun->curr_properties & PROP_gimple_lvec) != 0;
}
/* Return true if we can still perform transformations that may introduce
vector operations that are not supported by the target. Vector lowering
normally handles those, but after that pass, it becomes unsafe. */
static inline bool
optimize_vectors_before_lowering_p ()
{
return !cfun || (cfun->curr_properties & PROP_gimple_lvec) == 0;
}
/* Return true if pow(cst, x) should be optimized into exp(log(cst) * x).
As a workaround for SPEC CPU2017 628.pop2_s, don't do it if arg0
is an exact integer, arg1 = phi_res +/- cst1 and phi_res = PHI <cst2, ...>
where cst2 +/- cst1 is an exact integer, because then pow (arg0, arg1)
will likely be exact, while exp (log (arg0) * arg1) might be not.
Also don't do it if arg1 is phi_res above and cst2 is an exact integer. */
static bool
optimize_pow_to_exp (tree arg0, tree arg1)
{
gcc_assert (TREE_CODE (arg0) == REAL_CST);
if (!real_isinteger (TREE_REAL_CST_PTR (arg0), TYPE_MODE (TREE_TYPE (arg0))))
return true;
if (TREE_CODE (arg1) != SSA_NAME)
return true;
gimple *def = SSA_NAME_DEF_STMT (arg1);
gphi *phi = dyn_cast <gphi *> (def);
tree cst1 = NULL_TREE;
enum tree_code code = ERROR_MARK;
if (!phi)
{
if (!is_gimple_assign (def))
return true;
code = gimple_assign_rhs_code (def);
switch (code)
{
case PLUS_EXPR:
case MINUS_EXPR:
break;
default:
return true;
}
if (TREE_CODE (gimple_assign_rhs1 (def)) != SSA_NAME
|| TREE_CODE (gimple_assign_rhs2 (def)) != REAL_CST)
return true;
cst1 = gimple_assign_rhs2 (def);
phi = dyn_cast <gphi *> (SSA_NAME_DEF_STMT (gimple_assign_rhs1 (def)));
if (!phi)
return true;
}
tree cst2 = NULL_TREE;
int n = gimple_phi_num_args (phi);
for (int i = 0; i < n; i++)
{
tree arg = PHI_ARG_DEF (phi, i);
if (TREE_CODE (arg) != REAL_CST)
continue;
else if (cst2 == NULL_TREE)
cst2 = arg;
else if (!operand_equal_p (cst2, arg, 0))
return true;
}
if (cst1 && cst2)
cst2 = const_binop (code, TREE_TYPE (cst2), cst2, cst1);
if (cst2
&& TREE_CODE (cst2) == REAL_CST
&& real_isinteger (TREE_REAL_CST_PTR (cst2),
TYPE_MODE (TREE_TYPE (cst2))))
return false;
return true;
}
/* Return true if a division INNER_DIV / DIVISOR where INNER_DIV
is another division can be optimized. Don't optimize if INNER_DIV
is used in a TRUNC_MOD_EXPR with DIVISOR as second operand. */
static bool
optimize_successive_divisions_p (tree divisor, tree inner_div)
{
if (!gimple_in_ssa_p (cfun))
return false;
imm_use_iterator imm_iter;
use_operand_p use_p;
FOR_EACH_IMM_USE_FAST (use_p, imm_iter, inner_div)
{
gimple *use_stmt = USE_STMT (use_p);
if (!is_gimple_assign (use_stmt)
|| gimple_assign_rhs_code (use_stmt) != TRUNC_MOD_EXPR
|| !operand_equal_p (gimple_assign_rhs2 (use_stmt), divisor, 0))
continue;
return false;
}
return true;
}