b2b4005150
2016-01-19 Martin Jambor <mjambor@suse.cz> Martin Liska <mliska@suse.cz> Michael Matz <matz@suse.de> libgomp/ * plugin/Makefrag.am: Add HSA plugin requirements. * plugin/configfrag.ac (HSA_RUNTIME_INCLUDE): New variable. (HSA_RUNTIME_LIB): Likewise. (HSA_RUNTIME_CPPFLAGS): Likewise. (HSA_RUNTIME_INCLUDE): New substitution. (HSA_RUNTIME_LIB): Likewise. (HSA_RUNTIME_LDFLAGS): Likewise. (hsa-runtime): New configure option. (hsa-runtime-include): Likewise. (hsa-runtime-lib): Likewise. (PLUGIN_HSA): New substitution variable. Fill HSA_RUNTIME_INCLUDE and HSA_RUNTIME_LIB according to the new configure options. (PLUGIN_HSA_CPPFLAGS): Likewise. (PLUGIN_HSA_LDFLAGS): Likewise. (PLUGIN_HSA_LIBS): Likewise. Check that we have access to HSA run-time. * libgomp-plugin.h (offload_target_type): New element OFFLOAD_TARGET_TYPE_HSA. * libgomp.h (gomp_target_task): New fields firstprivate_copies and args. (bool gomp_create_target_task): Updated. (gomp_device_descr): Extra parameter of run_func and async_run_func, new field can_run_func. * libgomp_g.h (GOMP_target_ext): Update prototype. * oacc-host.c (host_run): Added a new parameter args. * target.c (calculate_firstprivate_requirements): New function. (copy_firstprivate_data): Likewise. (gomp_target_fallback_firstprivate): Use them. (gomp_target_unshare_firstprivate): New function. (gomp_get_target_fn_addr): Allow returning NULL for shared memory devices. (GOMP_target): Do host fallback for all shared memory devices. Do not pass any args to plugins. (GOMP_target_ext): Introduce device-specific argument parameter args. Allow host fallback if device shares memory. Do not remap data if device has shared memory. (gomp_target_task_fn): Likewise. Also treat shared memory devices like host fallback for mappings. (GOMP_target_data): Treat shared memory devices like host fallback. (GOMP_target_data_ext): Likewise. (GOMP_target_update): Likewise. (GOMP_target_update_ext): Likewise. Also pass NULL as args to gomp_create_target_task. (GOMP_target_enter_exit_data): Likewise. (omp_target_alloc): Treat shared memory devices like host fallback. (omp_target_free): Likewise. (omp_target_is_present): Likewise. (omp_target_memcpy): Likewise. (omp_target_memcpy_rect): Likewise. (omp_target_associate_ptr): Likewise. (gomp_load_plugin_for_device): Also load can_run. * task.c (GOMP_PLUGIN_target_task_completion): Free firstprivate_copies. (gomp_create_target_task): Accept new argument args and store it to ttask. * plugin/plugin-hsa.c: New file. gcc/ * Makefile.in (OBJS): Add new source files. (GTFILES): Add hsa.c. * common.opt (disable_hsa): New variable. (-Whsa): New warning. * config.in (ENABLE_HSA): New. * configure.ac: Treat hsa differently from other accelerators. (OFFLOAD_TARGETS): Define ENABLE_OFFLOADING according to $enable_offloading. (ENABLE_HSA): Define ENABLE_HSA according to $enable_hsa. * doc/install.texi (Configuration): Document --with-hsa-runtime, --with-hsa-runtime-include, --with-hsa-runtime-lib and --with-hsa-kmt-lib. * doc/invoke.texi (-Whsa): Document. (hsa-gen-debug-stores): Likewise. * lto-wrapper.c (compile_images_for_offload_targets): Do not attempt to invoke offload compiler for hsa acclerator. * opts.c (common_handle_option): Determine whether HSA offloading should be performed. * params.def (PARAM_HSA_GEN_DEBUG_STORES): New parameter. * builtin-types.def (BT_FN_VOID_UINT_PTR_INT_PTR): New. (BT_FN_VOID_INT_OMPFN_SIZE_PTR_PTR_PTR_UINT_PTR_INT_INT): Removed. (BT_FN_VOID_INT_OMPFN_SIZE_PTR_PTR_PTR_UINT_PTR_PTR): New. * gimple-low.c (lower_stmt): Also handle GIMPLE_OMP_GRID_BODY. * gimple-pretty-print.c (dump_gimple_omp_for): Also handle GF_OMP_FOR_KIND_GRID_LOOP. (dump_gimple_omp_block): Also handle GIMPLE_OMP_GRID_BODY. (pp_gimple_stmt_1): Likewise. * gimple-walk.c (walk_gimple_stmt): Likewise. * gimple.c (gimple_build_omp_grid_body): New function. (gimple_copy): Also handle GIMPLE_OMP_GRID_BODY. * gimple.def (GIMPLE_OMP_GRID_BODY): New. * gimple.h (enum gf_mask): Added GF_OMP_PARALLEL_GRID_PHONY, GF_OMP_FOR_KIND_GRID_LOOP, GF_OMP_FOR_GRID_PHONY and GF_OMP_TEAMS_GRID_PHONY. (gimple_statement_omp_single_layout): Updated comments. (gimple_build_omp_grid_body): New function. (gimple_has_substatements): Also handle GIMPLE_OMP_GRID_BODY. (gimple_omp_for_grid_phony): New function. (gimple_omp_for_set_grid_phony): Likewise. (gimple_omp_parallel_grid_phony): Likewise. (gimple_omp_parallel_set_grid_phony): Likewise. (gimple_omp_teams_grid_phony): Likewise. (gimple_omp_teams_set_grid_phony): Likewise. (gimple_return_set_retbnd): Also handle GIMPLE_OMP_GRID_BODY. * omp-builtins.def (BUILT_IN_GOMP_OFFLOAD_REGISTER): New. (BUILT_IN_GOMP_OFFLOAD_UNREGISTER): Likewise. (BUILT_IN_GOMP_TARGET): Updated type. * omp-low.c: Include symbol-summary.h, hsa.h and params.h. (adjust_for_condition): New function. (get_omp_for_step_from_incr): Likewise. (extract_omp_for_data): Moved parts to adjust_for_condition and get_omp_for_step_from_incr. (build_outer_var_ref): Handle GIMPLE_OMP_GRID_BODY. (fixup_child_record_type): Bail out if receiver_decl is NULL. (scan_sharing_clauses): Handle OMP_CLAUSE__GRIDDIM_. (scan_omp_parallel): Do not create child functions for phony constructs. (check_omp_nesting_restrictions): Handle GIMPLE_OMP_GRID_BODY. (scan_omp_1_op): Checking assert we are not remapping to ERROR_MARK. Also also handle GIMPLE_OMP_GRID_BODY. (parallel_needs_hsa_kernel_p): New function. (expand_parallel_call): Register apprpriate parallel child functions as HSA kernels. (grid_launch_attributes_trees): New type. (grid_attr_trees): New variable. (grid_create_kernel_launch_attr_types): New function. (grid_insert_store_range_dim): Likewise. (grid_get_kernel_launch_attributes): Likewise. (get_target_argument_identifier_1): Likewise. (get_target_argument_identifier): Likewise. (get_target_argument_value): Likewise. (push_target_argument_according_to_value): Likewise. (get_target_arguments): Likewise. (expand_omp_target): Call get_target_arguments instead of looking up for teams and thread limit. (grid_expand_omp_for_loop): New function. (grid_arg_decl_map): New type. (grid_remap_kernel_arg_accesses): New function. (grid_expand_target_kernel_body): New function. (expand_omp): Call it. (lower_omp_for): Do not emit phony constructs. (lower_omp_taskreg): Do not emit phony constructs but create for them a temporary variable receiver_decl. (lower_omp_taskreg): Do not emit phony constructs. (lower_omp_teams): Likewise. (lower_omp_grid_body): New function. (lower_omp_1): Call it. (grid_reg_assignment_to_local_var_p): New function. (grid_seq_only_contains_local_assignments): Likewise. (grid_find_single_omp_among_assignments_1): Likewise. (grid_find_single_omp_among_assignments): Likewise. (grid_find_ungridifiable_statement): Likewise. (grid_target_follows_gridifiable_pattern): Likewise. (grid_remap_prebody_decls): Likewise. (grid_copy_leading_local_assignments): Likewise. (grid_process_kernel_body_copy): Likewise. (grid_attempt_target_gridification): Likewise. (grid_gridify_all_targets_stmt): Likewise. (grid_gridify_all_targets): Likewise. (execute_lower_omp): Call grid_gridify_all_targets. (make_gimple_omp_edges): Handle GIMPLE_OMP_GRID_BODY. * tree-core.h (omp_clause_code): Added OMP_CLAUSE__GRIDDIM_. (tree_omp_clause): Added union field dimension. * tree-pretty-print.c (dump_omp_clause): Handle OMP_CLAUSE__GRIDDIM_. * tree.c (omp_clause_num_ops): Added number of arguments of OMP_CLAUSE__GRIDDIM_. (omp_clause_code_name): Added name of OMP_CLAUSE__GRIDDIM_. (walk_tree_1): Handle OMP_CLAUSE__GRIDDIM_. * tree.h (OMP_CLAUSE_GRIDDIM_DIMENSION): New. (OMP_CLAUSE_SET_GRIDDIM_DIMENSION): Likewise. (OMP_CLAUSE_GRIDDIM_SIZE): Likewise. (OMP_CLAUSE_GRIDDIM_GROUP): Likewise. * passes.def: Schedule pass_ipa_hsa and pass_gen_hsail. * tree-pass.h (make_pass_gen_hsail): Declare. (make_pass_ipa_hsa): Likewise. * ipa-hsa.c: New file. * lto-section-in.c (lto_section_name): Add hsa section name. * lto-streamer.h (lto_section_type): Add hsa section. * timevar.def (TV_IPA_HSA): New. * hsa-brig-format.h: New file. * hsa-brig.c: New file. * hsa-dump.c: Likewise. * hsa-gen.c: Likewise. * hsa.c: Likewise. * hsa.h: Likewise. * toplev.c (compile_file): Call hsa_output_brig. * hsa-regalloc.c: New file. gcc/fortran/ * types.def (BT_FN_VOID_UINT_PTR_INT_PTR): New. (BT_FN_VOID_INT_OMPFN_SIZE_PTR_PTR_PTR_UINT_PTR_INT_INT): Removed. (BT_FN_VOID_INT_OMPFN_SIZE_PTR_PTR_PTR_UINT_PTR_PTR): New. gcc/lto/ * lto-partition.c: Include "hsa.h" (add_symbol_to_partition_1): Put hsa implementations into the same partition as host implementations. liboffloadmic/ * plugin/libgomp-plugin-intelmic.cpp (GOMP_OFFLOAD_async_run): New unused parameter. (GOMP_OFFLOAD_run): Likewise. include/ * gomp-constants.h (GOMP_DEVICE_HSA): New macro. (GOMP_VERSION_HSA): Likewise. (GOMP_TARGET_ARG_DEVICE_MASK): Likewise. (GOMP_TARGET_ARG_DEVICE_ALL): Likewise. (GOMP_TARGET_ARG_SUBSEQUENT_PARAM): Likewise. (GOMP_TARGET_ARG_ID_MASK): Likewise. (GOMP_TARGET_ARG_NUM_TEAMS): Likewise. (GOMP_TARGET_ARG_THREAD_LIMIT): Likewise. (GOMP_TARGET_ARG_VALUE_SHIFT): Likewise. (GOMP_TARGET_ARG_HSA_KERNEL_ATTRIBUTES): Likewise. From-SVN: r232549
939 lines
25 KiB
C
939 lines
25 KiB
C
/* Gimple walk support.
|
|
|
|
Copyright (C) 2007-2016 Free Software Foundation, Inc.
|
|
Contributed by Aldy Hernandez <aldyh@redhat.com>
|
|
|
|
This file is part of GCC.
|
|
|
|
GCC is free software; you can redistribute it and/or modify it under
|
|
the terms of the GNU General Public License as published by the Free
|
|
Software Foundation; either version 3, or (at your option) any later
|
|
version.
|
|
|
|
GCC is distributed in the hope that it will be useful, but WITHOUT ANY
|
|
WARRANTY; without even the implied warranty of MERCHANTABILITY or
|
|
FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License
|
|
for more details.
|
|
|
|
You should have received a copy of the GNU General Public License
|
|
along with GCC; see the file COPYING3. If not see
|
|
<http://www.gnu.org/licenses/>. */
|
|
|
|
#include "config.h"
|
|
#include "system.h"
|
|
#include "coretypes.h"
|
|
#include "backend.h"
|
|
#include "tree.h"
|
|
#include "gimple.h"
|
|
#include "gimple-iterator.h"
|
|
#include "gimple-walk.h"
|
|
#include "stmt.h"
|
|
|
|
/* Walk all the statements in the sequence *PSEQ calling walk_gimple_stmt
|
|
on each one. WI is as in walk_gimple_stmt.
|
|
|
|
If walk_gimple_stmt returns non-NULL, the walk is stopped, and the
|
|
value is stored in WI->CALLBACK_RESULT. Also, the statement that
|
|
produced the value is returned if this statement has not been
|
|
removed by a callback (wi->removed_stmt). If the statement has
|
|
been removed, NULL is returned.
|
|
|
|
Otherwise, all the statements are walked and NULL returned. */
|
|
|
|
gimple *
|
|
walk_gimple_seq_mod (gimple_seq *pseq, walk_stmt_fn callback_stmt,
|
|
walk_tree_fn callback_op, struct walk_stmt_info *wi)
|
|
{
|
|
gimple_stmt_iterator gsi;
|
|
|
|
for (gsi = gsi_start (*pseq); !gsi_end_p (gsi); )
|
|
{
|
|
tree ret = walk_gimple_stmt (&gsi, callback_stmt, callback_op, wi);
|
|
if (ret)
|
|
{
|
|
/* If CALLBACK_STMT or CALLBACK_OP return a value, WI must exist
|
|
to hold it. */
|
|
gcc_assert (wi);
|
|
wi->callback_result = ret;
|
|
|
|
return wi->removed_stmt ? NULL : gsi_stmt (gsi);
|
|
}
|
|
|
|
if (!wi->removed_stmt)
|
|
gsi_next (&gsi);
|
|
}
|
|
|
|
if (wi)
|
|
wi->callback_result = NULL_TREE;
|
|
|
|
return NULL;
|
|
}
|
|
|
|
|
|
/* Like walk_gimple_seq_mod, but ensure that the head of SEQ isn't
|
|
changed by the callbacks. */
|
|
|
|
gimple *
|
|
walk_gimple_seq (gimple_seq seq, walk_stmt_fn callback_stmt,
|
|
walk_tree_fn callback_op, struct walk_stmt_info *wi)
|
|
{
|
|
gimple_seq seq2 = seq;
|
|
gimple *ret = walk_gimple_seq_mod (&seq2, callback_stmt, callback_op, wi);
|
|
gcc_assert (seq2 == seq);
|
|
return ret;
|
|
}
|
|
|
|
|
|
/* Helper function for walk_gimple_stmt. Walk operands of a GIMPLE_ASM. */
|
|
|
|
static tree
|
|
walk_gimple_asm (gasm *stmt, walk_tree_fn callback_op,
|
|
struct walk_stmt_info *wi)
|
|
{
|
|
tree ret, op;
|
|
unsigned noutputs;
|
|
const char **oconstraints;
|
|
unsigned i, n;
|
|
const char *constraint;
|
|
bool allows_mem, allows_reg, is_inout;
|
|
|
|
noutputs = gimple_asm_noutputs (stmt);
|
|
oconstraints = (const char **) alloca ((noutputs) * sizeof (const char *));
|
|
|
|
if (wi)
|
|
wi->is_lhs = true;
|
|
|
|
for (i = 0; i < noutputs; i++)
|
|
{
|
|
op = gimple_asm_output_op (stmt, i);
|
|
constraint = TREE_STRING_POINTER (TREE_VALUE (TREE_PURPOSE (op)));
|
|
oconstraints[i] = constraint;
|
|
if (wi)
|
|
{
|
|
if (parse_output_constraint (&constraint, i, 0, 0, &allows_mem,
|
|
&allows_reg, &is_inout))
|
|
wi->val_only = (allows_reg || !allows_mem);
|
|
}
|
|
ret = walk_tree (&TREE_VALUE (op), callback_op, wi, NULL);
|
|
if (ret)
|
|
return ret;
|
|
}
|
|
|
|
n = gimple_asm_ninputs (stmt);
|
|
for (i = 0; i < n; i++)
|
|
{
|
|
op = gimple_asm_input_op (stmt, i);
|
|
constraint = TREE_STRING_POINTER (TREE_VALUE (TREE_PURPOSE (op)));
|
|
|
|
if (wi)
|
|
{
|
|
if (parse_input_constraint (&constraint, 0, 0, noutputs, 0,
|
|
oconstraints, &allows_mem, &allows_reg))
|
|
{
|
|
wi->val_only = (allows_reg || !allows_mem);
|
|
/* Although input "m" is not really a LHS, we need a lvalue. */
|
|
wi->is_lhs = !wi->val_only;
|
|
}
|
|
}
|
|
ret = walk_tree (&TREE_VALUE (op), callback_op, wi, NULL);
|
|
if (ret)
|
|
return ret;
|
|
}
|
|
|
|
if (wi)
|
|
{
|
|
wi->is_lhs = false;
|
|
wi->val_only = true;
|
|
}
|
|
|
|
n = gimple_asm_nlabels (stmt);
|
|
for (i = 0; i < n; i++)
|
|
{
|
|
op = gimple_asm_label_op (stmt, i);
|
|
ret = walk_tree (&TREE_VALUE (op), callback_op, wi, NULL);
|
|
if (ret)
|
|
return ret;
|
|
}
|
|
|
|
return NULL_TREE;
|
|
}
|
|
|
|
|
|
/* Helper function of WALK_GIMPLE_STMT. Walk every tree operand in
|
|
STMT. CALLBACK_OP and WI are as in WALK_GIMPLE_STMT.
|
|
|
|
CALLBACK_OP is called on each operand of STMT via walk_tree.
|
|
Additional parameters to walk_tree must be stored in WI. For each operand
|
|
OP, walk_tree is called as:
|
|
|
|
walk_tree (&OP, CALLBACK_OP, WI, WI->PSET)
|
|
|
|
If CALLBACK_OP returns non-NULL for an operand, the remaining
|
|
operands are not scanned.
|
|
|
|
The return value is that returned by the last call to walk_tree, or
|
|
NULL_TREE if no CALLBACK_OP is specified. */
|
|
|
|
tree
|
|
walk_gimple_op (gimple *stmt, walk_tree_fn callback_op,
|
|
struct walk_stmt_info *wi)
|
|
{
|
|
hash_set<tree> *pset = (wi) ? wi->pset : NULL;
|
|
unsigned i;
|
|
tree ret = NULL_TREE;
|
|
|
|
switch (gimple_code (stmt))
|
|
{
|
|
case GIMPLE_ASSIGN:
|
|
/* Walk the RHS operands. If the LHS is of a non-renamable type or
|
|
is a register variable, we may use a COMPONENT_REF on the RHS. */
|
|
if (wi)
|
|
{
|
|
tree lhs = gimple_assign_lhs (stmt);
|
|
wi->val_only
|
|
= (is_gimple_reg_type (TREE_TYPE (lhs)) && !is_gimple_reg (lhs))
|
|
|| gimple_assign_rhs_class (stmt) != GIMPLE_SINGLE_RHS;
|
|
}
|
|
|
|
for (i = 1; i < gimple_num_ops (stmt); i++)
|
|
{
|
|
ret = walk_tree (gimple_op_ptr (stmt, i), callback_op, wi,
|
|
pset);
|
|
if (ret)
|
|
return ret;
|
|
}
|
|
|
|
/* Walk the LHS. If the RHS is appropriate for a memory, we
|
|
may use a COMPONENT_REF on the LHS. */
|
|
if (wi)
|
|
{
|
|
/* If the RHS is of a non-renamable type or is a register variable,
|
|
we may use a COMPONENT_REF on the LHS. */
|
|
tree rhs1 = gimple_assign_rhs1 (stmt);
|
|
wi->val_only
|
|
= (is_gimple_reg_type (TREE_TYPE (rhs1)) && !is_gimple_reg (rhs1))
|
|
|| gimple_assign_rhs_class (stmt) != GIMPLE_SINGLE_RHS;
|
|
wi->is_lhs = true;
|
|
}
|
|
|
|
ret = walk_tree (gimple_op_ptr (stmt, 0), callback_op, wi, pset);
|
|
if (ret)
|
|
return ret;
|
|
|
|
if (wi)
|
|
{
|
|
wi->val_only = true;
|
|
wi->is_lhs = false;
|
|
}
|
|
break;
|
|
|
|
case GIMPLE_CALL:
|
|
if (wi)
|
|
{
|
|
wi->is_lhs = false;
|
|
wi->val_only = true;
|
|
}
|
|
|
|
ret = walk_tree (gimple_call_chain_ptr (as_a <gcall *> (stmt)),
|
|
callback_op, wi, pset);
|
|
if (ret)
|
|
return ret;
|
|
|
|
ret = walk_tree (gimple_call_fn_ptr (stmt), callback_op, wi, pset);
|
|
if (ret)
|
|
return ret;
|
|
|
|
for (i = 0; i < gimple_call_num_args (stmt); i++)
|
|
{
|
|
if (wi)
|
|
wi->val_only
|
|
= is_gimple_reg_type (TREE_TYPE (gimple_call_arg (stmt, i)));
|
|
ret = walk_tree (gimple_call_arg_ptr (stmt, i), callback_op, wi,
|
|
pset);
|
|
if (ret)
|
|
return ret;
|
|
}
|
|
|
|
if (gimple_call_lhs (stmt))
|
|
{
|
|
if (wi)
|
|
{
|
|
wi->is_lhs = true;
|
|
wi->val_only
|
|
= is_gimple_reg_type (TREE_TYPE (gimple_call_lhs (stmt)));
|
|
}
|
|
|
|
ret = walk_tree (gimple_call_lhs_ptr (stmt), callback_op, wi, pset);
|
|
if (ret)
|
|
return ret;
|
|
}
|
|
|
|
if (wi)
|
|
{
|
|
wi->is_lhs = false;
|
|
wi->val_only = true;
|
|
}
|
|
break;
|
|
|
|
case GIMPLE_CATCH:
|
|
ret = walk_tree (gimple_catch_types_ptr (as_a <gcatch *> (stmt)),
|
|
callback_op, wi, pset);
|
|
if (ret)
|
|
return ret;
|
|
break;
|
|
|
|
case GIMPLE_EH_FILTER:
|
|
ret = walk_tree (gimple_eh_filter_types_ptr (stmt), callback_op, wi,
|
|
pset);
|
|
if (ret)
|
|
return ret;
|
|
break;
|
|
|
|
case GIMPLE_ASM:
|
|
ret = walk_gimple_asm (as_a <gasm *> (stmt), callback_op, wi);
|
|
if (ret)
|
|
return ret;
|
|
break;
|
|
|
|
case GIMPLE_OMP_CONTINUE:
|
|
{
|
|
gomp_continue *cont_stmt = as_a <gomp_continue *> (stmt);
|
|
ret = walk_tree (gimple_omp_continue_control_def_ptr (cont_stmt),
|
|
callback_op, wi, pset);
|
|
if (ret)
|
|
return ret;
|
|
|
|
ret = walk_tree (gimple_omp_continue_control_use_ptr (cont_stmt),
|
|
callback_op, wi, pset);
|
|
if (ret)
|
|
return ret;
|
|
}
|
|
break;
|
|
|
|
case GIMPLE_OMP_CRITICAL:
|
|
{
|
|
gomp_critical *omp_stmt = as_a <gomp_critical *> (stmt);
|
|
ret = walk_tree (gimple_omp_critical_name_ptr (omp_stmt),
|
|
callback_op, wi, pset);
|
|
if (ret)
|
|
return ret;
|
|
ret = walk_tree (gimple_omp_critical_clauses_ptr (omp_stmt),
|
|
callback_op, wi, pset);
|
|
if (ret)
|
|
return ret;
|
|
}
|
|
break;
|
|
|
|
case GIMPLE_OMP_ORDERED:
|
|
{
|
|
gomp_ordered *omp_stmt = as_a <gomp_ordered *> (stmt);
|
|
ret = walk_tree (gimple_omp_ordered_clauses_ptr (omp_stmt),
|
|
callback_op, wi, pset);
|
|
if (ret)
|
|
return ret;
|
|
}
|
|
break;
|
|
|
|
case GIMPLE_OMP_FOR:
|
|
ret = walk_tree (gimple_omp_for_clauses_ptr (stmt), callback_op, wi,
|
|
pset);
|
|
if (ret)
|
|
return ret;
|
|
for (i = 0; i < gimple_omp_for_collapse (stmt); i++)
|
|
{
|
|
ret = walk_tree (gimple_omp_for_index_ptr (stmt, i), callback_op,
|
|
wi, pset);
|
|
if (ret)
|
|
return ret;
|
|
ret = walk_tree (gimple_omp_for_initial_ptr (stmt, i), callback_op,
|
|
wi, pset);
|
|
if (ret)
|
|
return ret;
|
|
ret = walk_tree (gimple_omp_for_final_ptr (stmt, i), callback_op,
|
|
wi, pset);
|
|
if (ret)
|
|
return ret;
|
|
ret = walk_tree (gimple_omp_for_incr_ptr (stmt, i), callback_op,
|
|
wi, pset);
|
|
if (ret)
|
|
return ret;
|
|
}
|
|
break;
|
|
|
|
case GIMPLE_OMP_PARALLEL:
|
|
{
|
|
gomp_parallel *omp_par_stmt = as_a <gomp_parallel *> (stmt);
|
|
ret = walk_tree (gimple_omp_parallel_clauses_ptr (omp_par_stmt),
|
|
callback_op, wi, pset);
|
|
if (ret)
|
|
return ret;
|
|
ret = walk_tree (gimple_omp_parallel_child_fn_ptr (omp_par_stmt),
|
|
callback_op, wi, pset);
|
|
if (ret)
|
|
return ret;
|
|
ret = walk_tree (gimple_omp_parallel_data_arg_ptr (omp_par_stmt),
|
|
callback_op, wi, pset);
|
|
if (ret)
|
|
return ret;
|
|
}
|
|
break;
|
|
|
|
case GIMPLE_OMP_TASK:
|
|
ret = walk_tree (gimple_omp_task_clauses_ptr (stmt), callback_op,
|
|
wi, pset);
|
|
if (ret)
|
|
return ret;
|
|
ret = walk_tree (gimple_omp_task_child_fn_ptr (stmt), callback_op,
|
|
wi, pset);
|
|
if (ret)
|
|
return ret;
|
|
ret = walk_tree (gimple_omp_task_data_arg_ptr (stmt), callback_op,
|
|
wi, pset);
|
|
if (ret)
|
|
return ret;
|
|
ret = walk_tree (gimple_omp_task_copy_fn_ptr (stmt), callback_op,
|
|
wi, pset);
|
|
if (ret)
|
|
return ret;
|
|
ret = walk_tree (gimple_omp_task_arg_size_ptr (stmt), callback_op,
|
|
wi, pset);
|
|
if (ret)
|
|
return ret;
|
|
ret = walk_tree (gimple_omp_task_arg_align_ptr (stmt), callback_op,
|
|
wi, pset);
|
|
if (ret)
|
|
return ret;
|
|
break;
|
|
|
|
case GIMPLE_OMP_SECTIONS:
|
|
ret = walk_tree (gimple_omp_sections_clauses_ptr (stmt), callback_op,
|
|
wi, pset);
|
|
if (ret)
|
|
return ret;
|
|
ret = walk_tree (gimple_omp_sections_control_ptr (stmt), callback_op,
|
|
wi, pset);
|
|
if (ret)
|
|
return ret;
|
|
|
|
break;
|
|
|
|
case GIMPLE_OMP_SINGLE:
|
|
ret = walk_tree (gimple_omp_single_clauses_ptr (stmt), callback_op, wi,
|
|
pset);
|
|
if (ret)
|
|
return ret;
|
|
break;
|
|
|
|
case GIMPLE_OMP_TARGET:
|
|
{
|
|
gomp_target *omp_stmt = as_a <gomp_target *> (stmt);
|
|
ret = walk_tree (gimple_omp_target_clauses_ptr (omp_stmt),
|
|
callback_op, wi, pset);
|
|
if (ret)
|
|
return ret;
|
|
ret = walk_tree (gimple_omp_target_child_fn_ptr (omp_stmt),
|
|
callback_op, wi, pset);
|
|
if (ret)
|
|
return ret;
|
|
ret = walk_tree (gimple_omp_target_data_arg_ptr (omp_stmt),
|
|
callback_op, wi, pset);
|
|
if (ret)
|
|
return ret;
|
|
}
|
|
break;
|
|
|
|
case GIMPLE_OMP_TEAMS:
|
|
ret = walk_tree (gimple_omp_teams_clauses_ptr (stmt), callback_op, wi,
|
|
pset);
|
|
if (ret)
|
|
return ret;
|
|
break;
|
|
|
|
case GIMPLE_OMP_ATOMIC_LOAD:
|
|
{
|
|
gomp_atomic_load *omp_stmt = as_a <gomp_atomic_load *> (stmt);
|
|
ret = walk_tree (gimple_omp_atomic_load_lhs_ptr (omp_stmt),
|
|
callback_op, wi, pset);
|
|
if (ret)
|
|
return ret;
|
|
ret = walk_tree (gimple_omp_atomic_load_rhs_ptr (omp_stmt),
|
|
callback_op, wi, pset);
|
|
if (ret)
|
|
return ret;
|
|
}
|
|
break;
|
|
|
|
case GIMPLE_OMP_ATOMIC_STORE:
|
|
{
|
|
gomp_atomic_store *omp_stmt = as_a <gomp_atomic_store *> (stmt);
|
|
ret = walk_tree (gimple_omp_atomic_store_val_ptr (omp_stmt),
|
|
callback_op, wi, pset);
|
|
if (ret)
|
|
return ret;
|
|
}
|
|
break;
|
|
|
|
case GIMPLE_TRANSACTION:
|
|
{
|
|
gtransaction *txn = as_a <gtransaction *> (stmt);
|
|
|
|
ret = walk_tree (gimple_transaction_label_norm_ptr (txn),
|
|
callback_op, wi, pset);
|
|
if (ret)
|
|
return ret;
|
|
ret = walk_tree (gimple_transaction_label_uninst_ptr (txn),
|
|
callback_op, wi, pset);
|
|
if (ret)
|
|
return ret;
|
|
ret = walk_tree (gimple_transaction_label_over_ptr (txn),
|
|
callback_op, wi, pset);
|
|
if (ret)
|
|
return ret;
|
|
}
|
|
break;
|
|
|
|
case GIMPLE_OMP_RETURN:
|
|
ret = walk_tree (gimple_omp_return_lhs_ptr (stmt), callback_op, wi,
|
|
pset);
|
|
if (ret)
|
|
return ret;
|
|
break;
|
|
|
|
/* Tuples that do not have operands. */
|
|
case GIMPLE_NOP:
|
|
case GIMPLE_RESX:
|
|
case GIMPLE_PREDICT:
|
|
break;
|
|
|
|
default:
|
|
{
|
|
enum gimple_statement_structure_enum gss;
|
|
gss = gimple_statement_structure (stmt);
|
|
if (gss == GSS_WITH_OPS || gss == GSS_WITH_MEM_OPS)
|
|
for (i = 0; i < gimple_num_ops (stmt); i++)
|
|
{
|
|
ret = walk_tree (gimple_op_ptr (stmt, i), callback_op, wi, pset);
|
|
if (ret)
|
|
return ret;
|
|
}
|
|
}
|
|
break;
|
|
}
|
|
|
|
return NULL_TREE;
|
|
}
|
|
|
|
|
|
/* Walk the current statement in GSI (optionally using traversal state
|
|
stored in WI). If WI is NULL, no state is kept during traversal.
|
|
The callback CALLBACK_STMT is called. If CALLBACK_STMT indicates
|
|
that it has handled all the operands of the statement, its return
|
|
value is returned. Otherwise, the return value from CALLBACK_STMT
|
|
is discarded and its operands are scanned.
|
|
|
|
If CALLBACK_STMT is NULL or it didn't handle the operands,
|
|
CALLBACK_OP is called on each operand of the statement via
|
|
walk_gimple_op. If walk_gimple_op returns non-NULL for any
|
|
operand, the remaining operands are not scanned. In this case, the
|
|
return value from CALLBACK_OP is returned.
|
|
|
|
In any other case, NULL_TREE is returned. */
|
|
|
|
tree
|
|
walk_gimple_stmt (gimple_stmt_iterator *gsi, walk_stmt_fn callback_stmt,
|
|
walk_tree_fn callback_op, struct walk_stmt_info *wi)
|
|
{
|
|
gimple *ret;
|
|
tree tree_ret;
|
|
gimple *stmt = gsi_stmt (*gsi);
|
|
|
|
if (wi)
|
|
{
|
|
wi->gsi = *gsi;
|
|
wi->removed_stmt = false;
|
|
|
|
if (wi->want_locations && gimple_has_location (stmt))
|
|
input_location = gimple_location (stmt);
|
|
}
|
|
|
|
ret = NULL;
|
|
|
|
/* Invoke the statement callback. Return if the callback handled
|
|
all of STMT operands by itself. */
|
|
if (callback_stmt)
|
|
{
|
|
bool handled_ops = false;
|
|
tree_ret = callback_stmt (gsi, &handled_ops, wi);
|
|
if (handled_ops)
|
|
return tree_ret;
|
|
|
|
/* If CALLBACK_STMT did not handle operands, it should not have
|
|
a value to return. */
|
|
gcc_assert (tree_ret == NULL);
|
|
|
|
if (wi && wi->removed_stmt)
|
|
return NULL;
|
|
|
|
/* Re-read stmt in case the callback changed it. */
|
|
stmt = gsi_stmt (*gsi);
|
|
}
|
|
|
|
/* If CALLBACK_OP is defined, invoke it on every operand of STMT. */
|
|
if (callback_op)
|
|
{
|
|
tree_ret = walk_gimple_op (stmt, callback_op, wi);
|
|
if (tree_ret)
|
|
return tree_ret;
|
|
}
|
|
|
|
/* If STMT can have statements inside (e.g. GIMPLE_BIND), walk them. */
|
|
switch (gimple_code (stmt))
|
|
{
|
|
case GIMPLE_BIND:
|
|
ret = walk_gimple_seq_mod (gimple_bind_body_ptr (as_a <gbind *> (stmt)),
|
|
callback_stmt, callback_op, wi);
|
|
if (ret)
|
|
return wi->callback_result;
|
|
break;
|
|
|
|
case GIMPLE_CATCH:
|
|
ret = walk_gimple_seq_mod (gimple_catch_handler_ptr (
|
|
as_a <gcatch *> (stmt)),
|
|
callback_stmt, callback_op, wi);
|
|
if (ret)
|
|
return wi->callback_result;
|
|
break;
|
|
|
|
case GIMPLE_EH_FILTER:
|
|
ret = walk_gimple_seq_mod (gimple_eh_filter_failure_ptr (stmt), callback_stmt,
|
|
callback_op, wi);
|
|
if (ret)
|
|
return wi->callback_result;
|
|
break;
|
|
|
|
case GIMPLE_EH_ELSE:
|
|
{
|
|
geh_else *eh_else_stmt = as_a <geh_else *> (stmt);
|
|
ret = walk_gimple_seq_mod (gimple_eh_else_n_body_ptr (eh_else_stmt),
|
|
callback_stmt, callback_op, wi);
|
|
if (ret)
|
|
return wi->callback_result;
|
|
ret = walk_gimple_seq_mod (gimple_eh_else_e_body_ptr (eh_else_stmt),
|
|
callback_stmt, callback_op, wi);
|
|
if (ret)
|
|
return wi->callback_result;
|
|
}
|
|
break;
|
|
|
|
case GIMPLE_TRY:
|
|
ret = walk_gimple_seq_mod (gimple_try_eval_ptr (stmt), callback_stmt, callback_op,
|
|
wi);
|
|
if (ret)
|
|
return wi->callback_result;
|
|
|
|
ret = walk_gimple_seq_mod (gimple_try_cleanup_ptr (stmt), callback_stmt,
|
|
callback_op, wi);
|
|
if (ret)
|
|
return wi->callback_result;
|
|
break;
|
|
|
|
case GIMPLE_OMP_FOR:
|
|
ret = walk_gimple_seq_mod (gimple_omp_for_pre_body_ptr (stmt), callback_stmt,
|
|
callback_op, wi);
|
|
if (ret)
|
|
return wi->callback_result;
|
|
|
|
/* FALL THROUGH. */
|
|
case GIMPLE_OMP_CRITICAL:
|
|
case GIMPLE_OMP_MASTER:
|
|
case GIMPLE_OMP_TASKGROUP:
|
|
case GIMPLE_OMP_ORDERED:
|
|
case GIMPLE_OMP_SECTION:
|
|
case GIMPLE_OMP_PARALLEL:
|
|
case GIMPLE_OMP_TASK:
|
|
case GIMPLE_OMP_SECTIONS:
|
|
case GIMPLE_OMP_SINGLE:
|
|
case GIMPLE_OMP_TARGET:
|
|
case GIMPLE_OMP_TEAMS:
|
|
case GIMPLE_OMP_GRID_BODY:
|
|
ret = walk_gimple_seq_mod (gimple_omp_body_ptr (stmt), callback_stmt,
|
|
callback_op, wi);
|
|
if (ret)
|
|
return wi->callback_result;
|
|
break;
|
|
|
|
case GIMPLE_WITH_CLEANUP_EXPR:
|
|
ret = walk_gimple_seq_mod (gimple_wce_cleanup_ptr (stmt), callback_stmt,
|
|
callback_op, wi);
|
|
if (ret)
|
|
return wi->callback_result;
|
|
break;
|
|
|
|
case GIMPLE_TRANSACTION:
|
|
ret = walk_gimple_seq_mod (gimple_transaction_body_ptr (
|
|
as_a <gtransaction *> (stmt)),
|
|
callback_stmt, callback_op, wi);
|
|
if (ret)
|
|
return wi->callback_result;
|
|
break;
|
|
|
|
default:
|
|
gcc_assert (!gimple_has_substatements (stmt));
|
|
break;
|
|
}
|
|
|
|
return NULL;
|
|
}
|
|
|
|
/* From a tree operand OP return the base of a load or store operation
|
|
or NULL_TREE if OP is not a load or a store. */
|
|
|
|
static tree
|
|
get_base_loadstore (tree op)
|
|
{
|
|
while (handled_component_p (op))
|
|
op = TREE_OPERAND (op, 0);
|
|
if (DECL_P (op)
|
|
|| INDIRECT_REF_P (op)
|
|
|| TREE_CODE (op) == MEM_REF
|
|
|| TREE_CODE (op) == TARGET_MEM_REF)
|
|
return op;
|
|
return NULL_TREE;
|
|
}
|
|
|
|
|
|
/* For the statement STMT call the callbacks VISIT_LOAD, VISIT_STORE and
|
|
VISIT_ADDR if non-NULL on loads, store and address-taken operands
|
|
passing the STMT, the base of the operand, the operand itself containing
|
|
the base and DATA to it. The base will be either a decl, an indirect
|
|
reference (including TARGET_MEM_REF) or the argument of an address
|
|
expression.
|
|
Returns the results of these callbacks or'ed. */
|
|
|
|
bool
|
|
walk_stmt_load_store_addr_ops (gimple *stmt, void *data,
|
|
walk_stmt_load_store_addr_fn visit_load,
|
|
walk_stmt_load_store_addr_fn visit_store,
|
|
walk_stmt_load_store_addr_fn visit_addr)
|
|
{
|
|
bool ret = false;
|
|
unsigned i;
|
|
if (gimple_assign_single_p (stmt))
|
|
{
|
|
tree lhs, rhs, arg;
|
|
if (visit_store)
|
|
{
|
|
arg = gimple_assign_lhs (stmt);
|
|
lhs = get_base_loadstore (arg);
|
|
if (lhs)
|
|
ret |= visit_store (stmt, lhs, arg, data);
|
|
}
|
|
arg = gimple_assign_rhs1 (stmt);
|
|
rhs = arg;
|
|
while (handled_component_p (rhs))
|
|
rhs = TREE_OPERAND (rhs, 0);
|
|
if (visit_addr)
|
|
{
|
|
if (TREE_CODE (rhs) == ADDR_EXPR)
|
|
ret |= visit_addr (stmt, TREE_OPERAND (rhs, 0), arg, data);
|
|
else if (TREE_CODE (rhs) == TARGET_MEM_REF
|
|
&& TREE_CODE (TMR_BASE (rhs)) == ADDR_EXPR)
|
|
ret |= visit_addr (stmt, TREE_OPERAND (TMR_BASE (rhs), 0), arg,
|
|
data);
|
|
else if (TREE_CODE (rhs) == OBJ_TYPE_REF
|
|
&& TREE_CODE (OBJ_TYPE_REF_OBJECT (rhs)) == ADDR_EXPR)
|
|
ret |= visit_addr (stmt, TREE_OPERAND (OBJ_TYPE_REF_OBJECT (rhs),
|
|
0), arg, data);
|
|
else if (TREE_CODE (rhs) == CONSTRUCTOR)
|
|
{
|
|
unsigned int ix;
|
|
tree val;
|
|
|
|
FOR_EACH_CONSTRUCTOR_VALUE (CONSTRUCTOR_ELTS (rhs), ix, val)
|
|
if (TREE_CODE (val) == ADDR_EXPR)
|
|
ret |= visit_addr (stmt, TREE_OPERAND (val, 0), arg, data);
|
|
else if (TREE_CODE (val) == OBJ_TYPE_REF
|
|
&& TREE_CODE (OBJ_TYPE_REF_OBJECT (val)) == ADDR_EXPR)
|
|
ret |= visit_addr (stmt,
|
|
TREE_OPERAND (OBJ_TYPE_REF_OBJECT (val),
|
|
0), arg, data);
|
|
}
|
|
lhs = gimple_assign_lhs (stmt);
|
|
if (TREE_CODE (lhs) == TARGET_MEM_REF
|
|
&& TREE_CODE (TMR_BASE (lhs)) == ADDR_EXPR)
|
|
ret |= visit_addr (stmt, TREE_OPERAND (TMR_BASE (lhs), 0), lhs, data);
|
|
}
|
|
if (visit_load)
|
|
{
|
|
rhs = get_base_loadstore (rhs);
|
|
if (rhs)
|
|
ret |= visit_load (stmt, rhs, arg, data);
|
|
}
|
|
}
|
|
else if (visit_addr
|
|
&& (is_gimple_assign (stmt)
|
|
|| gimple_code (stmt) == GIMPLE_COND))
|
|
{
|
|
for (i = 0; i < gimple_num_ops (stmt); ++i)
|
|
{
|
|
tree op = gimple_op (stmt, i);
|
|
if (op == NULL_TREE)
|
|
;
|
|
else if (TREE_CODE (op) == ADDR_EXPR)
|
|
ret |= visit_addr (stmt, TREE_OPERAND (op, 0), op, data);
|
|
/* COND_EXPR and VCOND_EXPR rhs1 argument is a comparison
|
|
tree with two operands. */
|
|
else if (i == 1 && COMPARISON_CLASS_P (op))
|
|
{
|
|
if (TREE_CODE (TREE_OPERAND (op, 0)) == ADDR_EXPR)
|
|
ret |= visit_addr (stmt, TREE_OPERAND (TREE_OPERAND (op, 0),
|
|
0), op, data);
|
|
if (TREE_CODE (TREE_OPERAND (op, 1)) == ADDR_EXPR)
|
|
ret |= visit_addr (stmt, TREE_OPERAND (TREE_OPERAND (op, 1),
|
|
0), op, data);
|
|
}
|
|
}
|
|
}
|
|
else if (gcall *call_stmt = dyn_cast <gcall *> (stmt))
|
|
{
|
|
if (visit_store)
|
|
{
|
|
tree arg = gimple_call_lhs (call_stmt);
|
|
if (arg)
|
|
{
|
|
tree lhs = get_base_loadstore (arg);
|
|
if (lhs)
|
|
ret |= visit_store (stmt, lhs, arg, data);
|
|
}
|
|
}
|
|
if (visit_load || visit_addr)
|
|
for (i = 0; i < gimple_call_num_args (call_stmt); ++i)
|
|
{
|
|
tree arg = gimple_call_arg (call_stmt, i);
|
|
if (visit_addr
|
|
&& TREE_CODE (arg) == ADDR_EXPR)
|
|
ret |= visit_addr (stmt, TREE_OPERAND (arg, 0), arg, data);
|
|
else if (visit_load)
|
|
{
|
|
tree rhs = get_base_loadstore (arg);
|
|
if (rhs)
|
|
ret |= visit_load (stmt, rhs, arg, data);
|
|
}
|
|
}
|
|
if (visit_addr
|
|
&& gimple_call_chain (call_stmt)
|
|
&& TREE_CODE (gimple_call_chain (call_stmt)) == ADDR_EXPR)
|
|
ret |= visit_addr (stmt, TREE_OPERAND (gimple_call_chain (call_stmt), 0),
|
|
gimple_call_chain (call_stmt), data);
|
|
if (visit_addr
|
|
&& gimple_call_return_slot_opt_p (call_stmt)
|
|
&& gimple_call_lhs (call_stmt) != NULL_TREE
|
|
&& TREE_ADDRESSABLE (TREE_TYPE (gimple_call_lhs (call_stmt))))
|
|
ret |= visit_addr (stmt, gimple_call_lhs (call_stmt),
|
|
gimple_call_lhs (call_stmt), data);
|
|
}
|
|
else if (gasm *asm_stmt = dyn_cast <gasm *> (stmt))
|
|
{
|
|
unsigned noutputs;
|
|
const char *constraint;
|
|
const char **oconstraints;
|
|
bool allows_mem, allows_reg, is_inout;
|
|
noutputs = gimple_asm_noutputs (asm_stmt);
|
|
oconstraints = XALLOCAVEC (const char *, noutputs);
|
|
if (visit_store || visit_addr)
|
|
for (i = 0; i < gimple_asm_noutputs (asm_stmt); ++i)
|
|
{
|
|
tree link = gimple_asm_output_op (asm_stmt, i);
|
|
tree op = get_base_loadstore (TREE_VALUE (link));
|
|
if (op && visit_store)
|
|
ret |= visit_store (stmt, op, TREE_VALUE (link), data);
|
|
if (visit_addr)
|
|
{
|
|
constraint = TREE_STRING_POINTER
|
|
(TREE_VALUE (TREE_PURPOSE (link)));
|
|
oconstraints[i] = constraint;
|
|
parse_output_constraint (&constraint, i, 0, 0, &allows_mem,
|
|
&allows_reg, &is_inout);
|
|
if (op && !allows_reg && allows_mem)
|
|
ret |= visit_addr (stmt, op, TREE_VALUE (link), data);
|
|
}
|
|
}
|
|
if (visit_load || visit_addr)
|
|
for (i = 0; i < gimple_asm_ninputs (asm_stmt); ++i)
|
|
{
|
|
tree link = gimple_asm_input_op (asm_stmt, i);
|
|
tree op = TREE_VALUE (link);
|
|
if (visit_addr
|
|
&& TREE_CODE (op) == ADDR_EXPR)
|
|
ret |= visit_addr (stmt, TREE_OPERAND (op, 0), op, data);
|
|
else if (visit_load || visit_addr)
|
|
{
|
|
op = get_base_loadstore (op);
|
|
if (op)
|
|
{
|
|
if (visit_load)
|
|
ret |= visit_load (stmt, op, TREE_VALUE (link), data);
|
|
if (visit_addr)
|
|
{
|
|
constraint = TREE_STRING_POINTER
|
|
(TREE_VALUE (TREE_PURPOSE (link)));
|
|
parse_input_constraint (&constraint, 0, 0, noutputs,
|
|
0, oconstraints,
|
|
&allows_mem, &allows_reg);
|
|
if (!allows_reg && allows_mem)
|
|
ret |= visit_addr (stmt, op, TREE_VALUE (link),
|
|
data);
|
|
}
|
|
}
|
|
}
|
|
}
|
|
}
|
|
else if (greturn *return_stmt = dyn_cast <greturn *> (stmt))
|
|
{
|
|
tree op = gimple_return_retval (return_stmt);
|
|
if (op)
|
|
{
|
|
if (visit_addr
|
|
&& TREE_CODE (op) == ADDR_EXPR)
|
|
ret |= visit_addr (stmt, TREE_OPERAND (op, 0), op, data);
|
|
else if (visit_load)
|
|
{
|
|
tree base = get_base_loadstore (op);
|
|
if (base)
|
|
ret |= visit_load (stmt, base, op, data);
|
|
}
|
|
}
|
|
}
|
|
else if (visit_addr
|
|
&& gimple_code (stmt) == GIMPLE_PHI)
|
|
{
|
|
for (i = 0; i < gimple_phi_num_args (stmt); ++i)
|
|
{
|
|
tree op = gimple_phi_arg_def (stmt, i);
|
|
if (TREE_CODE (op) == ADDR_EXPR)
|
|
ret |= visit_addr (stmt, TREE_OPERAND (op, 0), op, data);
|
|
}
|
|
}
|
|
else if (visit_addr
|
|
&& gimple_code (stmt) == GIMPLE_GOTO)
|
|
{
|
|
tree op = gimple_goto_dest (stmt);
|
|
if (TREE_CODE (op) == ADDR_EXPR)
|
|
ret |= visit_addr (stmt, TREE_OPERAND (op, 0), op, data);
|
|
}
|
|
|
|
return ret;
|
|
}
|
|
|
|
/* Like walk_stmt_load_store_addr_ops but with NULL visit_addr. IPA-CP
|
|
should make a faster clone for this case. */
|
|
|
|
bool
|
|
walk_stmt_load_store_ops (gimple *stmt, void *data,
|
|
walk_stmt_load_store_addr_fn visit_load,
|
|
walk_stmt_load_store_addr_fn visit_store)
|
|
{
|
|
return walk_stmt_load_store_addr_ops (stmt, data,
|
|
visit_load, visit_store, NULL);
|
|
}
|