b5b8b0ac64
gcc/ChangeLog: * doc/invoke.texi (-fvar-tracking-assignments): New. (-fvar-tracking-assignments-toggle): New. (-fdump-final-insns=file): Mark filename as optional. (--param min-nondebug-insn-uid): New. (-gdwarf-@{version}): Mention version 4. * opts.c (common_handle_option): Accept it. * tree-vrp.c (find_assert_locations_1): Skip debug stmts. * regrename.c (regrename_optimize): Drop last. Don't count debug insns as uses. Don't reject change because of debug insn. (do_replace): Reject DEBUG_INSN as chain starter. Take base_regno from the chain starter, and check for inexact matches in DEBUG_INSNS. (scan_rtx_reg): Accept inexact matches in DEBUG_INSNs. (build_def_use): Simplify and fix the marking of DEBUG_INSNs. * sched-ebb.c (schedule_ebbs): Skip boundary debug insns. * fwprop.c (forward_propagate_and_simplify): ...into debug insns. * doc/gimple.texi (is_gimple_debug): New. (gimple_debug_bind_p): New. (is_gimple_call, gimple_assign_cast_p): End sentence with period. * doc/install.texi (bootstrap-debug): More details. (bootstrap-debug-big, bootstrap-debug-lean): Document. (bootstrap-debug-lib): More details. (bootstrap-debug-ckovw): Update. (bootstrap-time): New. * tree-into-ssa.c (mark_def_sites): Skip debug stmts. (insert_phi_nodes_for): Insert debug stmts. (rewrite_stmt): Take iterator. Insert debug stmts. (rewrite_enter_block): Adjust. (maybe_replace_use_in_debug_stmt): New. (rewrite_update_stmt): Use it. (mark_use_interesting): Return early for debug stmts. * tree-ssa-loop-im.c (rewrite_bittest): Propagate DEFs into debug stmts before replacing stmt. (move_computations_stmt): Likewise. * ira-conflicts.c (add_copies): Skip debug insns. * regstat.c (regstat_init_n_sets_and_refs): Discount debug insns. (regstat_bb_compute_ri): Skip debug insns. * tree-ssa-threadupdate.c (redirection_block_p): Skip debug stmts. * tree-ssa-loop-manip.c (find_uses_to_rename_stmt, check_loop_closed_ssa_stmt): Skip debug stmts. * tree-tailcall.c (find_tail_calls): Likewise. * tree-ssa-loop-ch.c (should_duplicate_loop_header_p): Likewise. * tree.h (MAY_HAVE_DEBUG_STMTS): New. (build_var_debug_value_stat): Declare. (build_var_debug_value): Define. (target_for_debug_bind): Declare. * reload.c (find_equiv_reg): Skip debug insns. * rtlanal.c (reg_used_between_p): Skip debug insns. (side_effects_p): Likewise. (canonicalize_condition): Likewise. * ddg.c (create_ddg_dep_from_intra_loop_link): Check that non-debug insns never depend on debug insns. (create_ddg_dep_no_link): Likewise. (add_cross_iteration_register_deps): Use ANTI_DEP for debug insns. Don't add inter-loop dependencies for debug insns. (build_intra_loop_deps): Likewise. (create_ddg): Count debug insns. * ddg.h (struct ddg::num_debug): New. (num_backargs): Pair up with previous int field. * diagnostic.c (diagnostic_report_diagnostic): Skip notes on -fcompare-debug-second. * final.c (get_attr_length_1): Skip debug insns. (rest_of_clean-state): Don't dump CFA_RESTORE_STATE. * gcc.c (invoke_as): Call compare-debug-dump-opt. (driver_self_specs): Map -fdump-final-insns to -fdump-final-insns=.. (get_local_tick): New. (compare_debug_dump_opt_spec_function): Test for . argument and compute output name. Compute temp output spec without flag name. Compute -frandom-seed. (OPT): Undef after use. * cfgloopanal.c (num_loop_insns): Skip debug insns. (average_num_loop_insns): Likewise. * params.h (MIN_NONDEBUG_INSN_UID): New. * gimple.def (GIMPLE_DEBUG): New. * ipa-reference.c (scan_stmt_for_static_refs): Skip debug stmts. * auto-inc-dec.c (merge_in_block): Skip debug insns. (merge_in_block): Fix whitespace. * toplev.c (flag_var_tracking): Update comment. (flag_var_tracking_assignments): New. (flag_var_tracking_assignments_toggle): New. (process_options): Don't open final insns dump file if we're not going to write to it. Compute defaults for var_tracking. * df-scan.c (df_insn_rescan_debug_internal): New. (df_uses_record): Handle debug insns. * haifa-sched.c (ready): Initialize n_debug. (contributes_to_priority): Skip debug insns. (dep_list_size): New. (priority): Use it. (rank_for_schedule): Likewise. Schedule debug insns as soon as they're ready. Disregard previous debug insns to make decisions. (queue_insn): Never queue debug insns. (ready_add, ready_remove_first, ready_remove): Count debug insns. (schedule_insn): Don't reject debug insns because of issue rate. (get_ebb_head_tail, no_real_insns_p): Skip boundary debug insns. (queue_to_ready): Skip and discount debug insns. (choose_ready): Let debug insns through. (schedule_block): Check boundary debug insns. Discount debug insns, schedule them early. Adjust whitespace. (set_priorities): Check for boundary debug insns. (add_jump_dependencies): Use dep_list_size. (prev_non_location_insn): New. (check_cfg): Use it. * tree-ssa-loop-ivopts.c (find-interesting_users): Skip debug stmts. (remove_unused_ivs): Reset debug stmts. * modulo-sched.c (const_iteration_count): Skip debug insns. (res_MII): Discount debug insns. (loop_single_full_bb_p): Skip debug insns. (sms_schedule): Likewise. (sms_schedule_by_order): Likewise. (ps_has_conflicts): Likewise. * caller-save.c (refmarker_fn): New. (save_call_clobbered_regs): Replace regs with saved mem in debug insns. (mark_referenced_regs): Take pointer, mark and arg. Adjust. Call refmarker_fn mark for hardregnos. (mark_reg_as_referenced): New. (replace_reg_with_saved_mem): New. * ipa-pure-const.c (check_stmt): Skip debug stmts. * cse.c (cse_insn): Canonicalize debug insns. Skip them when searching back. (cse_extended_basic_block): Skip debug insns. (count_reg_usage): Likewise. (is_dead_reg): New, split out of... (set_live_p): ... here. (insn_live_p): Use it for debug insns. * tree-stdarg.c (check_all_va_list_escapes): Skip debug stmts. (execute_optimize_stdarg): Likewise. * tree-ssa-dom.c (propagate_rhs_into_lhs): Likewise. * tree-ssa-propagate.c (substitute_and_fold): Don't regard changes in debug stmts as changes. * sel-sched.c (moving_insn_creates_bookkeeping_block_p): New. (moveup_expr): Don't move across debug insns. Don't move debug insn if it would create a bookkeeping block. (moveup_expr_cached): Don't use cache for debug insns that are heads of blocks. (compute_av_set_inside_bb): Skip debug insns. (sel_rank_for_schedule): Schedule debug insns first. Remove dead code. (block_valid_for_bookkeeping_p); Support lax searches. (create_block_for_bookkeeping): Adjust block numbers when encountering debug-only blocks. (find_place_for_bookkeeping): Deal with debug-only blocks. (generate_bookkeeping_insn): Accept no place to insert. (remove_temp_moveop_nops): New argument full_tidying. (prepare_place_to_insert): Deal with debug insns. (advance_state_on_fence): Debug insns don't start cycles. (update_boundaries): Take fence as argument. Deal with debug insns. (schedule_expr_on_boundary): No full_tidying on debug insns. (fill_insns): Deal with debug insns. (track_scheduled_insns_and_blocks): Don't count debug insns. (need_nop_to_preserve_insn_bb): New, split out of... (remove_insn_from_stream): ... this. (fur_orig_expr_not_found): Skip debug insns. * rtl.def (VALUE): Move up. (DEBUG_INSN): New. * tree-ssa-sink.c (all_immediate_uses_same_place): Skip debug stmts. (nearest_common_dominator_of_uses): Take debug_stmts argument. Set it if debug stmts are found. (statement_sink_location): Skip debug stmts. Propagate moving defs into debug stmts. * ifcvt.c (first_active_insn): Skip debug insns. (last_active_insns): Likewise. (cond_exec_process_insns): Likewise. (noce_process_if_block): Likewise. (check_cond_move_block): Likewise. (cond_move_convert_if_block): Likewise. (block_jumps_and_fallthru_p): Likewise. (dead_or_predicable): Likewise. * dwarf2out.c (debug_str_hash_forced): New. (find_AT_string): Add comment. (gen_label_for_indirect_string): New. (get_debug_string_label): New. (AT_string_form): Use it. (mem_loc_descriptor): Handle non-TLS symbols. Handle MINUS , DIV, MOD, AND, IOR, XOR, NOT, ABS, NEG, and CONST_STRING. Accept but discard COMPARE, IF_THEN_ELSE, ROTATE, ROTATERT, TRUNCATE and several operations that cannot be represented with DWARF opcodes. (loc_descriptor): Ignore SIGN_EXTEND and ZERO_EXTEND. Require dwarf_version 4 for DW_OP_implicit_value and DW_OP_stack_value. (dwarf2out_var_location): Take during-call mark into account. (output_indirect_string): Update comment. Output if there are label and references. (prune_indirect_string): New. (prune_unused_types): Call it if debug_str_hash_forced. More in dwarf2out.c, from Jakub Jelinek <jakub@redhat.com>: (dw_long_long_const): Remove. (struct dw_val_struct): Change val_long_long type to rtx. (print_die, attr_checksum, same_dw_val_p, loc_descriptor): Adjust for val_long_long change to CONST_DOUBLE rtx from a long hi/lo pair. (output_die): Likewise. Use HOST_BITS_PER_WIDE_INT size of each component instead of HOST_BITS_PER_LONG. (output_loc_operands): Likewise. For const8* assert HOST_BITS_PER_WIDE_INT rather than HOST_BITS_PER_LONG is >= 64. (output_loc_operands_raw): For const8* assert HOST_BITS_PER_WIDE_INT rather than HOST_BITS_PER_LONG is >= 64. (add_AT_long_long): Remove val_hi and val_lo arguments, add val_const_double. (size_of_die): Use HOST_BITS_PER_WIDE_INT size multiplier instead of HOST_BITS_PER_LONG for dw_val_class_long_long. (add_const_value_attribute): Adjust add_AT_long_long caller. Don't handle TLS SYMBOL_REFs. If CONST wraps a constant, tail recurse. (dwarf_stack_op_name): Handle DW_OP_implicit_value and DW_OP_stack_value. (size_of_loc_descr, output_loc_operands, output_loc_operands_raw): Handle DW_OP_implicit_value. (extract_int): Move prototype earlier. (mem_loc_descriptor): For SUBREG punt if inner mode size is wider than DWARF2_ADDR_SIZE. Handle SIGN_EXTEND and ZERO_EXTEND by DW_OP_shl and DW_OP_shr{a,}. Handle EQ, NE, GT, GE, LT, LE, GTU, GEU, LTU, LEU, SMIN, SMAX, UMIN, UMAX, SIGN_EXTRACT, ZERO_EXTRACT. (loc_descriptor): Compare mode size with DWARF2_ADDR_SIZE instead of Pmode size. (loc_descriptor): Add MODE argument. Handle CONST_INT, CONST_DOUBLE, CONST_VECTOR, CONST, LABEL_REF and SYMBOL_REF if mode != VOIDmode, attempt to handle other expressions. Don't handle TLS SYMBOL_REFs. (concat_loc_descriptor, concatn_loc_descriptor, loc_descriptor_from_tree_1): Adjust loc_descriptor callers. (add_location_or_const_value_attribute): Likewise. For single location loc_lists attempt to use add_const_value_attribute for constant decls. Add DW_AT_const_value even if NOTE_VAR_LOCATION is VAR_LOCATION with CONSTANT_P or CONST_STRING in its expression. * cfgbuild.c (inside_basic_block_p): Handle debug insns. (control_flow_insn_p): Likewise. * tree-parloops.c (eliminate_local_variables_stmt): Handle debug stmt. (separate_decls_in_region_debug_bind): New. (separate_decls_in_region): Process debug bind stmts afterwards. * recog.c (verify_changes): Handle debug insns. (extract_insn): Likewise. (peephole2_optimize): Skip debug insns. * dse.c (scan_insn): Skip debug insns. * sel-sched-ir.c (return_nop_to_pool): Take full_tidying argument. Pass it on. (setup_id_for_insn): Handle debug insns. (maybe_tidy_empty_bb): Adjust whitespace. (tidy_control_flow): Skip debug insns. (sel_remove_insn): Adjust for debug insns. (sel_estimate_number_of_insns): Skip debug insns. (create_insn_rtx_from_pattern): Handle debug insns. (create_copy_of_insn_rtx): Likewise. * sel-sched-.h (sel_bb_end): Declare. (sel_bb_empty_or_nop_p): New. (get_all_loop_exits): Use it. (_eligible_successor_edge_p): Likewise. (return_nop_to_pool): Adjust. * tree-eh.c (tre_empty_eh_handler_p): Skip debug stmts. * ira-lives.c (process_bb_node_lives): Skip debug insns. * gimple-pretty-print.c (dump_gimple_debug): New. (dump_gimple_stmt): Use it. (dump_bb_header): Skip gimple debug stmts. * regmove.c (optimize_reg_copy_1): Discount debug insns. (fixup_match_2): Likewise. (regmove_backward_pass): Likewise. Simplify combined replacement. Handle debug insns. * function.c (instantiate_virtual_regs): Handle debug insns. * function.h (struct emit_status): Add x_cur_debug_insn_uid. * print-rtl.h: Include cselib.h. (print_rtx): Print VALUEs. Split out and recurse for VAR_LOCATIONs. * df.h (df_inns_rescan_debug_internal): Declare. * gcse.c (alloc_hash_table): Estimate n_insns. (cprop_insn): Don't regard debug insns as changes. (bypass_conditional_jumps): Skip debug insns. (one_pre_gcse_pass): Adjust. (one_code_hoisting_pass): Likewise. (compute_ld_motion_mems): Skip debug insns. (one_cprop_pass): Adjust. * tree-if-conv.c (tree_if_convert_stmt): Reset debug stmts. (if_convertible_stmt_p): Handle debug stmts. * init-regs.c (initialize_uninitialized_regs): Skip debug insns. * tree-vect-loop.c (vect_is_simple_reduction): Skip debug stmts. * ira-build.c (create_bb_allocnos): Skip debug insns. * tree-flow-inline.h (has_zero_uses): Discount debug stmts. (has_single_use): Likewise. (single_imm_use): Likewise. (num_imm_uses): Likewise. * tree-ssa-phiopt.c (empty_block_p): Skip debug stmts. * tree-ssa-coalesce.c (build_ssa_conflict_graph): Skip debug stmts. (create_outofssa_var_map): Likewise. * lower-subreg.c (adjust_decomposed_uses): New. (resolve_debug): New. (decompose_multiword_subregs): Use it. * tree-dfa.c (find_referenced_vars): Skip debug stmts. * emit-rtl.c: Include params.h. (cur_debug_insn_uid): Define. (set_new_first_and_last_insn): Set cur_debug_insn_uid too. (copy_rtx_if_shared_1): Handle debug insns. (reset_used_flags): Likewise. (set_used_flags): LIkewise. (get_max_insn_count): New. (next_nondebug_insn): New. (prev_nondebug_insn): New. (make_debug_insn_raw): New. (emit_insn_before_noloc): Handle debug insns. (emit_jump_insn_before_noloc): Likewise. (emit_call_insn_before_noloc): Likewise. (emit_debug_insn_before_noloc): New. (emit_insn_after_noloc): Handle debug insns. (emit_jump_insn_after_noloc): Likewise. (emit_call_insn_after_noloc): Likewise. (emit_debug_insn_after_noloc): Likewise. (emit_insn_after): Take loc from earlier non-debug insn. (emit_jump_insn_after): Likewise. (emit_call_insn_after): Likewise. (emit_debug_insn_after_setloc): New. (emit_debug_insn_after): New. (emit_insn_before): Take loc from later non-debug insn. (emit_jump_insn_before): Likewise. (emit_call_insn_before): Likewise. (emit_debug_insn_before_setloc): New. (emit_debug_insn_before): New. (emit_insn): Handle debug insns. (emit_debug_insn): New. (emit_jump_insn): Handle debug insns. (emit_call_insn): Likewise. (emit): Likewise. (init_emit): Take min-nondebug-insn-uid into account. Initialize cur_debug_insn_uid. (emit_copy_of_insn_after): Handle debug insns. * cfgexpand.c (gimple_assign_rhs_to_tree): Do not overwrite location of single rhs in place. (maybe_dump_rtl_for_gimple_stmt): Dump lineno. (floor_sdiv_adjust): New. (cell_sdiv_adjust): New. (cell_udiv_adjust): New. (round_sdiv_adjust): New. (round_udiv_adjust): New. (wrap_constant): Moved from cselib. (unwrap_constant): New. (expand_debug_expr): New. (expand_debug_locations): New. (expand_gimple_basic_block): Drop hiding redeclaration. Expand debug bind stmts. (gimple_expand_cfg): Expand debug locations. * cselib.c: Include tree-pass.h. (struct expand_value_data): New. (cselib_record_sets_hook): New. (PRESERVED_VALUE_P, LONG_TERM_PRESERVED_VALUE_P): New. (cselib_clear_table): Move, and implemnet in terms of... (cselib_reset_table_with_next_value): ... this. (cselib_get_next_unknown_value): New. (discard_useless_locs): Don't discard preserved values. (cselib_preserve_value): New. (cselib_preserved_value_p): New. (cselib_preserve_definitely): New. (cselib_clear_preserve): New. (cselib_preserve_only_values): New. (new_cselib_val): Take rtx argument. Dump it in details. (cselib_lookup_mem): Adjust. (expand_loc): Take regs_active in struct. Adjust. Silence dumps unless details are requested. (cselib_expand_value_rtx_cb): New. (cselib_expand_value_rtx): Rename and reimplment in terms of... (cselib_expand_value_rtx_1): ... this. Adjust. Silence dumps without details. Copy more subregs. Try to resolve values using a callback. Wrap constants. (cselib_subst_to_values): Adjust. (cselib_log_lookup): New. (cselib_lookup): Call it. (cselib_invalidate_regno): Don't count preserved values as useless. (cselib_invalidate_mem): Likewise. (cselib_record_set): Likewise. (struct set): Renamed to cselib_set, moved to cselib.h. (cselib_record_sets): Adjust. Call hook. (cselib_process_insn): Reset table when it would be cleared. (dump_cselib_val): New. (dump_cselib_table): New. * tree-cfgcleanup.c (tree_forwarded_block_p): Skip debug stmts. (remove_forwarder_block): Support moving debug stmts. * cselib.h (cselib_record_sets_hook): Declare. (cselib_expand_callback): New type. (cselib_expand_value_rtx_cb): Declare. (cselib_reset_table_with_next_value): Declare. (cselib_get_next_unknown_value): Declare. (cselib_preserve_value): Declare. (cselib_preserved_value_p): Declare. (cselib_preserve_only_values): Declare. (dump_cselib_table): Declare. * cfgcleanup.c (flow_find_cross_jump): Skip debug insns. (try_crossjump_to_edge): Likewise. (delete_unreachable_blocks): Remove dominant GIMPLE blocks after dominated blocks when debug stmts are present. * simplify-rtx.c (delegitimize_mem_from_attrs): New. * tree-ssa-live.c (remove_unused_locals): Skip debug stmts. (set_var_live_on_entry): Likewise. * loop-invariant.c (find_invariants_bb): Skip debug insns. * cfglayout.c (curr_location, last_location): Make static. (set_curr_insn_source_location): Don't avoid bouncing. (get_curr_insn_source_location): New. (get_curr_insn_block): New. (duplicate_insn_chain): Handle debug insns. * tree-ssa-forwprop.c (forward_propagate_addr_expr): Propagate into debug stmts. * common.opt (fcompare-debug): Move to sort order. (fdump-unnumbered-links): Likewise. (fvar-tracking-assignments): New. (fvar-tracking-assignments-toggle): New. * tree-ssa-dce.c (mark_stmt_necessary): Don't mark blocks because of debug stmts. (mark_stmt_if_obviously_necessary): Mark debug stmts. (eliminate_unnecessary_stmts): Walk dominated blocks before dominators. * tree-ssa-ter.c (find_replaceable_in_bb): Skip debug stmts. * ira.c (memref_used_between_p): Skip debug insns. (update_equiv_regs): Likewise. * sched-deps.c (sd_lists_size): Accept empty list. (sd_init_insn): Mark debug insns. (sd_finish_insn): Unmark them. (sd_add_dep): Reject non-debug deps on debug insns. (fixup_sched_groups): Give debug insns group treatment. Skip debug insns. (sched_analyze_reg): Don't mark debug insns for sched before call. (sched_analyze_2): Handle debug insns. (sched_analyze_insn): Compute next non-debug insn. Handle debug insns. (deps_analyze_insn): Handle debug insns. (deps_start_bb): Skip debug insns. (init_deps): Initialize last_debug_insn. * tree-ssa.c (target_for_debug_bind): New. (find_released_ssa_name): New. (propagate_var_def_into_debug_stmts): New. (propagate_defs_into_debug_stmts): New. (verify_ssa): Skip debug bind stmts without values. (warn_uninialized_vars): Skip debug stmts. * target-def.h (TARGET_DELEGITIMIZE_ADDRESS): Set default. * rtl.c (rtx_equal_p_cb): Handle VALUEs. (rtx_equal_p): Likewise. * ira-costs.c (scan_one_insn): Skip debug insns. (process_bb_node_for_hard_reg_moves): Likewise. * rtl.h (DEBUG_INSN_P): New. (NONDEBUG_INSN_P): New. (MAY_HAVE_DEBUG_INSNS): New. (INSN_P): Accept debug insns. (RTX_FRAME_RELATED_P): Likewise. (INSN_DELETED_P): Likewise (PAT_VAR_LOCATION_DECL): New. (PAT_VAR_LOCATION_LOC): New. (PAT_VAR_OCATION_STATUS): New. (NOTE_VAR_LOCATION_DECL): Reimplement. (NOTE_VAR_LOCATION_LOC): Likewise. (NOTE_VAR_LOCATION_STATUS): Likewise. (INSN_VAR_LOCATION): New. (INSN_VAR_LOCATION_DECL): New. (INSN_VAR_LOCATION_LOC): New. (INSN_VAR_LOCATION_STATUS): New. (gen_rtx_UNKNOWN_VAR_LOC): New. (VAR_LOC_UNKNOWN_P): New. (NOTE_DURING_CALL_P): New. (SCHED_GROUP_P): Accept debug insns. (emit_debug_insn_before): Declare. (emit_debug_insn_before_noloc): Declare. (emit_debug_insn_beore_setloc): Declare. (emit_debug_insn_after): Declare. (emit_debug_insn_after_noloc): Declare. (emit_debug_insn_after_setloc): Declare. (emit_debug_insn): Declare. (make_debug_insn_raw): Declare. (prev_nondebug_insn): Declare. (next_nondebug_insn): Declare. (delegitimize_mem_from_attrs): Declare. (get_max_insn_count): Declare. (wrap_constant): Declare. (unwrap_constant): Declare. (get_curr_insn_source_location): Declare. (get_curr_insn_block): Declare. * tree-inline.c (insert_debug_decl_map): New. (processing_debug_stmt): New. (remap_decl): Don't create new mappings in debug stmts. (remap_gimple_op_r): Don't add references in debug stmts. (copy_tree_body_r): Likewise. (remap_gimple_stmt): Handle debug bind stmts. (copy_bb): Skip debug stmts. (copy_edges_for_bb): Likewise. (copy_debug_stmt): New. (copy_debug_stmts): New. (copy_body): Copy debug stmts at the end. (insert_init_debug_bind): New. (insert_init_stmt): Take id. Skip and emit debug stmts. (setup_one_parameter): Remap variable earlier, register debug mapping. (estimate_num_insns): Skip debug stmts. (expand_call_inline): Preserve debug_map. (optimize_inline_calls): Check for no debug_stmts left-overs. (unsave_expr_now): Preserve debug_map. (copy_gimple_seq_and_replace_locals): Likewise. (tree_function_versioning): Check for no debug_stmts left-overs. Init and destroy debug_map as needed. Split edges unconditionally. (build_duplicate_type): Init and destroy debug_map as needed. * tree-inline.h: Include gimple.h instead of pointer-set.h. (struct copy_body_data): Add debug_stmts and debug_map. * sched-int.h (struct ready_list): Add n_debug. (struct deps): Add last_debug_insn. (DEBUG_INSN_SCHED_P): New. (BOUNDARY_DEBUG_INSN_P): New. (SCHEDULE_DEBUG_INSN_P): New. (sd_iterator_cond): Accept empty list. * combine.c (create_log_links): Skip debug insns. (combine_instructions): Likewise. (cleanup_auto_inc_dec): New. From Jakub Jelinek: Make sure the return value is always unshared. (struct rtx_subst_pair): New. (auto_adjust_pair): New. (propagate_for_debug_subst): New. (propagate_for_debug): New. (try_combine): Skip debug insns. Propagate removed defs into debug insns. (next_nonnote_nondebug_insn): New. (distribute_notes): Use it. Skip debug insns. (distribute_links): Skip debug insns. * tree-outof-ssa.c (set_location_for_edge): Likewise. * resource.c (mark_target_live_regs): Likewise. * var-tracking.c: Include cselib.h and target.h. (enum micro_operation_type): Add MO_VAL_USE, MO_VAL_LOC, and MO_VAL_SET. (micro_operation_type_name): New. (enum emit_note_where): Add EMIT_NOTE_AFTER_CALL_INSN. (struct micro_operation_def): Update comments. (decl_or_value): New type. Use instead of decls. (struct emit_note_data_def): Add vars. (struct attrs_def): Use decl_or_value. (struct variable_tracking_info_def): Add permp, flooded. (struct location_chain_def): Update comment. (struct variable_part_def): Use decl_or_value. (struct variable_def): Make var_part a variable length array. (valvar_pool): New. (scratch_regs): New. (cselib_hook_called): New. (dv_is_decl_p): New. (dv_is_value_p): New. (dv_as_decl): New. (dv_as_value): New. (dv_as_opaque): New. (dv_onepart_p): New. (dv_pool): New. (IS_DECL_CODE): New. (check_value_is_not_decl): New. (dv_from_decl): New. (dv_from_value): New. (dv_htab_hash): New. (variable_htab_hash): Use it. (variable_htab_eq): Support values. (variable_htab_free): Free from the right pool. (attrs_list_member, attrs_list_insert): Use decl_or_value. (attrs_list_union): Adjust. (attrs_list_mpdv_union): New. (tie_break_pointers): New. (canon_value_cmp): New. (unshare_variable): Return possibly-modified slot. (vars_copy_1): Adjust. (var_reg_decl_set): Adjust. Split out of... (var_reg_set): ... this. (get_init_value): Adjust. (var_reg_delete_and_set): Adjust. (var_reg_delete): Adjust. (var_regno_delete): Adjust. (var_mem_decl_set): Split out of... (var_mem_set): ... this. (var_mem_delete_and_set): Adjust. (var_mem_delete): Adjust. (val_store): New. (val_reset): New. (val_resolve): New. (variable_union): Adjust. Speed up merge of 1-part vars. (variable_canonicalize): Use unshared slot. (VALUED_RECURSED_INTO): New. (find_loc_in_1pdv): New. (struct dfset_merge): New. (insert_into_intersection): New. (intersect_loc_chains): New. (loc_cmp): New. (canonicalize_loc_order_check): New. (canonicalize_values_mark): New. (canonicalize_values_star): New. (variable_merge_over_cur): New. (variable_merge_over_src): New. (dataflow_set_merge): New. (dataflow_set_equiv_regs): New. (remove_duplicate_values): New. (struct dfset_post_merge): New. (variable_post_merge_new_vals): New. (variable_post_merge_perm_vals): New. (dataflow_post_merge_adjust): New. (find_mem_expr_in_1pdv): New. (dataflow_set_preserve_mem_locs): New. (dataflow_set_remove_mem_locs): New. (dataflow_set_clear_at_call): New. (onepart_variable_different_p): New. (variable_different_p): Use it. (dataflow_set_different_1): Adjust. Make detailed dump more verbose. (track_expr_p): Add need_rtl parameter. Don't generate rtl if not needed. (track_loc_p): Pass it true. (struct count_use_info): New. (find_use_val): New. (replace_expr_with_values): New. (log_op_type): New. (use_type): New, partially split out of... (count_uses): ... this. Count new micro-ops. (count_uses_1): Adjust. (count_stores): Adjust. (count_with_sets): New. (VAL_NEEDS_RESOLUTION): New. (VAL_HOLDS_TRACK_EXPR): New. (VAL_EXPR_IS_COPIED): New. (VAL_EXPR_IS_CLOBBERED): New. (add_uses): Adjust. Generate new micro-ops. (add_uses_1): Adjust. (add_stores): Generate new micro-ops. (add_with_sets): New. (find_src_status): Adjust. (find_src_set_src): Adjust. (compute_bb_dataflow): Use dataflow_set_clear_at_call. Handle new micro-ops. Canonicalize value equivalances. (vt_find_locations): Compute total size of hash tables for dumping. Perform merge for var-tracking-assignments. Don't disregard single-block loops. (dump_attrs_list): Handle decl_or_value. (dump_variable): Take variable. Deal with decl_or_value. (dump_variable_slot): New. (dump_vars): Use it. (dump_dataflow_sets): Adjust. (set_slot_part): New, extended to support one-part variables after splitting out of... (set_variable_part): ... this. (clobber_slot_part): New, split out of... (clobber_variable_part): ... this. (delete_slot_part): New, split out of... (delete_variable_part): .... this. (check_wrap_constant): New. (vt_expand_loc_callback): New. (vt_expand_loc): New. (emit_note_insn_var_location): Adjust. Handle values. Handle EMIT_NOTE_AFTER_CALL_INSN. (emit_notes_for_differences_1): Adjust. Handle values. (emit_notes_for_differences_2): Likewise. (emit_notes_for_differences): Adjust. (emit_notes_in_bb): Take pointer to set. Emit AFTER_CALL_INSN notes. Adjust. Handle new micro-ops. (vt_add_function_parameters): Adjust. Create and bind values. (vt_initialize): Adjust. Initialize scratch_regs and valvar_pool, flooded and perm.. Initialize and use cselib. Log operations. Move some code to count_with_sets and add_with_sets. (delete_debug_insns): New. (vt_debug_insns_local): New. (vt_finalize): Release permp, valvar_pool, scratch_regs. Finish cselib. (var_tracking_main): If var-tracking-assignments is enabled but var-tracking isn't, delete debug insns and leave. Likewise if we exceed limits or fail the stack adjustments tests, and after all var-tracking processing. More in var-tracking, from Jakub Jelinek <jakub@redhat.com>: (dataflow_set): Add traversed_vars. (value_chain, const_value_chain): New typedefs. (value_chain_pool, value_chains): New variables. (value_chain_htab_hash, value_chain_htab_eq, add_value_chain, add_value_chains, add_cselib_value_chains, remove_value_chain, remove_value_chains, remove_cselib_value_chains): New functions. (shared_hash_find_slot_unshare_1, shared_hash_find_slot_1, shared_hash_find_slot_noinsert_1, shared_hash_find_1): New static inlines. (shared_hash_find_slot_unshare, shared_hash_find_slot, shared_hash_find_slot_noinsert, shared_hash_find): Update. (dst_can_be_shared): New variable. (unshare_variable): Unshare set->vars if shared, use shared_hash_*. Clear dst_can_be_shared. If set->traversed_vars is non-NULL and different from set->vars, look up slot again instead of using the passed in slot. (dataflow_set_init): Initialize traversed_vars. (variable_union): Use shared_hash_*. Use initially NO_INSERT lookup if set->vars is shared. Don't keep slot cleared before calling unshare_variable. Unshare set->vars if needed. Adjust unshare_variable callers. Clear dst_can_be_shared if needed. Even ->refcount == 1 vars must be unshared if set->vars is shared and var needs to be modified. (dataflow_set_union): Set traversed_vars during canonicalization. (VALUE_CHANGED, DECL_CHANGED): Define. (set_dv_changed, dv_changed_p): New static inlines. (track_expr_p): Clear DECL_CHANGED. (dump_dataflow_sets): Set it. (variable_was_changed): Call set_dv_changed. (emit_note_insn_var_location): Likewise. (changed_variables_stack): New variable. (check_changed_vars_1, check_changed_vars_2): New functions. (emit_notes_for_changes): Do nothing if changed_variables is empty. Traverse changed_variables with check_changed_vars_1, call check_changed_vars_2 on each changed_variables_stack entry. (emit_notes_in_bb): Add SET argument. Just clear it at the beginning, use it instead of local &set, don't destroy it at the end. (vt_emit_notes): Call dataflow_set_clear early on all VTI(bb)->out sets, never use them, instead use emit_notes_in_bb computed set, dataflow_set_clear also VTI(bb)->in when we are done with the basic block. Initialize changed_variables_stack, free it afterwards. If ENABLE_CHECKING verify that after noting differences to an empty set value_chains hash table is empty. (vt_initialize): Initialize value_chains and value_chain_pool. (vt_finalize): Delete value_chains htab, free value_chain_pool. (variable_tracking_main): Call dump_dataflow_sets before calling vt_emit_notes, not after it. * tree-flow.h (propagate_defs_into_debug_stmts): Declare. (propagate_var_def_into_debug_stmts): Declare. * df-problems.c (df_lr_bb_local_compute): Skip debug insns. (df_set_note): Reject debug insns. (df_whole_mw_reg_dead_p): Take added_notes_p argument. Don't add notes to debug insns. (df_note_bb_compute): Adjust. Likewise. (df_simulate_uses): Skip debug insns. (df_simulate_initialize_backwards): Likewise. * reg-stack.c (subst_stack_regs_in_debug_insn): New. (subst_stack_regs_pat): Reject debug insns. (convert_regs_1): Handle debug insns. * Makefile.in (TREE_INLINE_H): Take pointer-set.h from GIMPLE_H. (print-rtl.o): Depend on cselib.h. (cselib.o): Depend on TREE_PASS_H. (var-tracking.o): Depend on cselib.h and TARGET_H. * sched-rgn.c (rgn_estimate_number_of_insns): Discount debug insns. (init_ready_list): Skip boundary debug insns. (add_branch_dependences): Skip debug insns. (free_block_dependencies): Check for blocks with only debug insns. (compute_priorities): Likewise. * gimple.c (gss_for_code): Handle GIMPLE_DEBUG. (gimple_build_with_ops_stat): Take subcode as unsigned. Adjust all callers. (gimple_build_debug_bind_stat): New. (empty_body_p): Skip debug stmts. (gimple_has_side_effects): Likewise. (gimple_rhs_has_side_effects): Likewise. * gimple.h (enum gimple_debug_subcode, GIMPLE_DEBUG_BIND): New. (gimple_build_debug_bind_stat): Declare. (gimple_build_debug_bind): Define. (is_gimple_debug): New. (gimple_debug_bind_p): New. (gimple_debug_bind_get_var): New. (gimple_debug_bind_get_value): New. (gimple_debug_bind_get_value_ptr): New. (gimple_debug_bind_set_var): New. (gimple_debug_bind_set_value): New. (GIMPLE_DEBUG_BIND_NOVALUE): New internal temporary macro. (gimple_debug_bind_reset_value): New. (gimple_debug_bind_has_value_p): New. (gsi_next_nondebug): New. (gsi_prev_nondebug): New. (gsi_start_nondebug_bb): New. (gsi_last_nondebug_bb): New. * sched-vis.c (print_pattern): Handle VAR_LOCATION. (print_insn): Handle DEBUG_INSN. * tree-cfg.c (remove_bb): Walk stmts backwards. Let loc of first insn prevail. (first_stmt): Skip debug stmts. (first_non_label_stmt): Likewise. (last_stmt): Likewise. (has_zero_uses_1): New. (single_imm_use_1): New. (verify_gimple_debug): New. (verify_types_in_gimple_stmt): Handle debug stmts. (verify_stmt): Likewise. (debug_loop_num): Skip debug stmts. (remove_edge_and_dominated_blocks): Remove dominators last. * tree-ssa-reasssoc.c (rewrite_expr_tree): Propagate into debug stmts. (linearize_expr): Likewise. * config/i386/i386.c (ix86_delegitimize_address): Call default implementation. * config/ia64/ia64.c (ia64_safe_itanium_class): Handle debug insns. (group_barrier_needed): Skip debug insns. (emit_insn_group_barriers): Likewise. (emit_all_insn_group_barriers): Likewise. (ia64_variable_issue): Handle debug insns. (ia64_dfa_new_cycle): Likewise. (final_emit_insn_group_barriers): Skip debug insns. (ia64_dwarf2out_def_steady_cfa): Take frame argument. Don't def cfa without frame. (process_set): Likewise. (process_for_unwind_directive): Pass frame on. * config/rs6000/rs6000.c (TARGET_DELEGITIMIZE_ADDRESS): Define. (rs6000_delegitimize_address): New. (rs6000_debug_adjust_cost): Handle debug insns. (is_microcoded_insn): Likewise. (is_cracked_insn): Likewise. (is_nonpipeline_insn): Likewise. (insn_must_be_first_in_group): Likewise. (insn_must_be_last_in_group): Likewise. (force_new_group): Likewise. * cfgrtl.c (rtl_split_block): Emit INSN_DELETED note if block contains only debug insns. (rtl_merge_blocks): Skip debug insns. (purge_dead_edges): Likewise. (rtl_block_ends_with_call_p): Skip debug insns. * dce.c (deletable_insn_p): Handle VAR_LOCATION. (mark_reg_dependencies): Skip debug insns. * params.def (PARAM_MIN_NONDEBUG_INSN_UID): New. * tree-ssanames.c (release_ssa_name): Propagate def into debug stmts. * tree-ssa-threadedge.c (record_temporary_equivalences_from_stmts): Skip debug stmts. * regcprop.c (replace_oldest_value_addr): Skip debug insns. (replace_oldest_value_mem): Use ALL_REGS for debug insns. (copyprop_hardreg_forward_1): Handle debug insns. * reload1.c (reload): Skip debug insns. Replace unassigned pseudos in debug insns with their equivalences. (eliminate_regs_in_insn): Skip debug insns. (emit_input_reload_insns): Skip debug insns at first, adjust them later. * tree-ssa-operands.c (add_virtual_operand): Reject debug stmts. (get_indirect_ref_operands): Pass opf_no_vops on. (get_expr_operands): Likewise. Skip debug stmts. (parse_ssa_operands): Scan debug insns with opf_no_vops. gcc/testsuite/ChangeLog: * gcc.dg/guality/guality.c: New. * gcc.dg/guality/guality.h: New. * gcc.dg/guality/guality.exp: New. * gcc.dg/guality/example.c: New. * lib/gcc-dg.exp (cleanup-dump): Remove .gk files. (cleanup-saved-temps): Likewise, .gkd files too. gcc/cp/ChangeLog: * cp-tree.h (TFF_NO_OMIT_DEFAULT_TEMPLATE_ARGUMENTS): New. * cp-lang.c (cxx_dwarf_name): Pass it. * error.c (count_non_default_template_args): Take flags as argument. Adjust all callers. Skip counting of default arguments if the new flag is given. ChangeLog: * Makefile.tpl (BUILD_CONFIG): Default to bootstrap-debug. * Makefile.in: Rebuilt. contrib/ChangeLog: * compare-debug: Look for .gkd files and compare them. config/ChangeLog: * bootstrap-debug.mk: Add comments. * bootstrap-debug-big.mk: New. * bootstrap-debug-lean.mk: New. * bootstrap-debug-ckovw.mk: Add comments. * bootstrap-debug-lib.mk: Drop CFLAGS for stages. Use -g0 for TFLAGS in stage1. Drop -fvar-tracking-assignments-toggle. From-SVN: r151312
2906 lines
87 KiB
C
2906 lines
87 KiB
C
/* Swing Modulo Scheduling implementation.
|
||
Copyright (C) 2004, 2005, 2006, 2007, 2008
|
||
Free Software Foundation, Inc.
|
||
Contributed by Ayal Zaks and Mustafa Hagog <zaks,mustafa@il.ibm.com>
|
||
|
||
This file is part of GCC.
|
||
|
||
GCC is free software; you can redistribute it and/or modify it under
|
||
the terms of the GNU General Public License as published by the Free
|
||
Software Foundation; either version 3, or (at your option) any later
|
||
version.
|
||
|
||
GCC is distributed in the hope that it will be useful, but WITHOUT ANY
|
||
WARRANTY; without even the implied warranty of MERCHANTABILITY or
|
||
FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License
|
||
for more details.
|
||
|
||
You should have received a copy of the GNU General Public License
|
||
along with GCC; see the file COPYING3. If not see
|
||
<http://www.gnu.org/licenses/>. */
|
||
|
||
|
||
#include "config.h"
|
||
#include "system.h"
|
||
#include "coretypes.h"
|
||
#include "tm.h"
|
||
#include "toplev.h"
|
||
#include "rtl.h"
|
||
#include "tm_p.h"
|
||
#include "hard-reg-set.h"
|
||
#include "regs.h"
|
||
#include "function.h"
|
||
#include "flags.h"
|
||
#include "insn-config.h"
|
||
#include "insn-attr.h"
|
||
#include "except.h"
|
||
#include "toplev.h"
|
||
#include "recog.h"
|
||
#include "sched-int.h"
|
||
#include "target.h"
|
||
#include "cfglayout.h"
|
||
#include "cfgloop.h"
|
||
#include "cfghooks.h"
|
||
#include "expr.h"
|
||
#include "params.h"
|
||
#include "gcov-io.h"
|
||
#include "ddg.h"
|
||
#include "timevar.h"
|
||
#include "tree-pass.h"
|
||
#include "dbgcnt.h"
|
||
|
||
#ifdef INSN_SCHEDULING
|
||
|
||
/* This file contains the implementation of the Swing Modulo Scheduler,
|
||
described in the following references:
|
||
[1] J. Llosa, A. Gonzalez, E. Ayguade, M. Valero., and J. Eckhardt.
|
||
Lifetime--sensitive modulo scheduling in a production environment.
|
||
IEEE Trans. on Comps., 50(3), March 2001
|
||
[2] J. Llosa, A. Gonzalez, E. Ayguade, and M. Valero.
|
||
Swing Modulo Scheduling: A Lifetime Sensitive Approach.
|
||
PACT '96 , pages 80-87, October 1996 (Boston - Massachusetts - USA).
|
||
|
||
The basic structure is:
|
||
1. Build a data-dependence graph (DDG) for each loop.
|
||
2. Use the DDG to order the insns of a loop (not in topological order
|
||
necessarily, but rather) trying to place each insn after all its
|
||
predecessors _or_ after all its successors.
|
||
3. Compute MII: a lower bound on the number of cycles to schedule the loop.
|
||
4. Use the ordering to perform list-scheduling of the loop:
|
||
1. Set II = MII. We will try to schedule the loop within II cycles.
|
||
2. Try to schedule the insns one by one according to the ordering.
|
||
For each insn compute an interval of cycles by considering already-
|
||
scheduled preds and succs (and associated latencies); try to place
|
||
the insn in the cycles of this window checking for potential
|
||
resource conflicts (using the DFA interface).
|
||
Note: this is different from the cycle-scheduling of schedule_insns;
|
||
here the insns are not scheduled monotonically top-down (nor bottom-
|
||
up).
|
||
3. If failed in scheduling all insns - bump II++ and try again, unless
|
||
II reaches an upper bound MaxII, in which case report failure.
|
||
5. If we succeeded in scheduling the loop within II cycles, we now
|
||
generate prolog and epilog, decrease the counter of the loop, and
|
||
perform modulo variable expansion for live ranges that span more than
|
||
II cycles (i.e. use register copies to prevent a def from overwriting
|
||
itself before reaching the use).
|
||
|
||
SMS works with countable loops (1) whose control part can be easily
|
||
decoupled from the rest of the loop and (2) whose loop count can
|
||
be easily adjusted. This is because we peel a constant number of
|
||
iterations into a prologue and epilogue for which we want to avoid
|
||
emitting the control part, and a kernel which is to iterate that
|
||
constant number of iterations less than the original loop. So the
|
||
control part should be a set of insns clearly identified and having
|
||
its own iv, not otherwise used in the loop (at-least for now), which
|
||
initializes a register before the loop to the number of iterations.
|
||
Currently SMS relies on the do-loop pattern to recognize such loops,
|
||
where (1) the control part comprises of all insns defining and/or
|
||
using a certain 'count' register and (2) the loop count can be
|
||
adjusted by modifying this register prior to the loop.
|
||
TODO: Rely on cfgloop analysis instead. */
|
||
|
||
/* This page defines partial-schedule structures and functions for
|
||
modulo scheduling. */
|
||
|
||
typedef struct partial_schedule *partial_schedule_ptr;
|
||
typedef struct ps_insn *ps_insn_ptr;
|
||
|
||
/* The minimum (absolute) cycle that a node of ps was scheduled in. */
|
||
#define PS_MIN_CYCLE(ps) (((partial_schedule_ptr)(ps))->min_cycle)
|
||
|
||
/* The maximum (absolute) cycle that a node of ps was scheduled in. */
|
||
#define PS_MAX_CYCLE(ps) (((partial_schedule_ptr)(ps))->max_cycle)
|
||
|
||
/* Perform signed modulo, always returning a non-negative value. */
|
||
#define SMODULO(x,y) ((x) % (y) < 0 ? ((x) % (y) + (y)) : (x) % (y))
|
||
|
||
/* The number of different iterations the nodes in ps span, assuming
|
||
the stage boundaries are placed efficiently. */
|
||
#define PS_STAGE_COUNT(ps) ((PS_MAX_CYCLE (ps) - PS_MIN_CYCLE (ps) \
|
||
+ 1 + (ps)->ii - 1) / (ps)->ii)
|
||
|
||
/* A single instruction in the partial schedule. */
|
||
struct ps_insn
|
||
{
|
||
/* The corresponding DDG_NODE. */
|
||
ddg_node_ptr node;
|
||
|
||
/* The (absolute) cycle in which the PS instruction is scheduled.
|
||
Same as SCHED_TIME (node). */
|
||
int cycle;
|
||
|
||
/* The next/prev PS_INSN in the same row. */
|
||
ps_insn_ptr next_in_row,
|
||
prev_in_row;
|
||
|
||
/* The number of nodes in the same row that come after this node. */
|
||
int row_rest_count;
|
||
};
|
||
|
||
/* Holds the partial schedule as an array of II rows. Each entry of the
|
||
array points to a linked list of PS_INSNs, which represents the
|
||
instructions that are scheduled for that row. */
|
||
struct partial_schedule
|
||
{
|
||
int ii; /* Number of rows in the partial schedule. */
|
||
int history; /* Threshold for conflict checking using DFA. */
|
||
|
||
/* rows[i] points to linked list of insns scheduled in row i (0<=i<ii). */
|
||
ps_insn_ptr *rows;
|
||
|
||
/* The earliest absolute cycle of an insn in the partial schedule. */
|
||
int min_cycle;
|
||
|
||
/* The latest absolute cycle of an insn in the partial schedule. */
|
||
int max_cycle;
|
||
|
||
ddg_ptr g; /* The DDG of the insns in the partial schedule. */
|
||
};
|
||
|
||
/* We use this to record all the register replacements we do in
|
||
the kernel so we can undo SMS if it is not profitable. */
|
||
struct undo_replace_buff_elem
|
||
{
|
||
rtx insn;
|
||
rtx orig_reg;
|
||
rtx new_reg;
|
||
struct undo_replace_buff_elem *next;
|
||
};
|
||
|
||
|
||
|
||
static partial_schedule_ptr create_partial_schedule (int ii, ddg_ptr, int history);
|
||
static void free_partial_schedule (partial_schedule_ptr);
|
||
static void reset_partial_schedule (partial_schedule_ptr, int new_ii);
|
||
void print_partial_schedule (partial_schedule_ptr, FILE *);
|
||
static void verify_partial_schedule (partial_schedule_ptr, sbitmap);
|
||
static ps_insn_ptr ps_add_node_check_conflicts (partial_schedule_ptr,
|
||
ddg_node_ptr node, int cycle,
|
||
sbitmap must_precede,
|
||
sbitmap must_follow);
|
||
static void rotate_partial_schedule (partial_schedule_ptr, int);
|
||
void set_row_column_for_ps (partial_schedule_ptr);
|
||
static void ps_insert_empty_row (partial_schedule_ptr, int, sbitmap);
|
||
static int compute_split_row (sbitmap, int, int, int, ddg_node_ptr);
|
||
|
||
|
||
/* This page defines constants and structures for the modulo scheduling
|
||
driver. */
|
||
|
||
static int sms_order_nodes (ddg_ptr, int, int *, int *);
|
||
static void set_node_sched_params (ddg_ptr);
|
||
static partial_schedule_ptr sms_schedule_by_order (ddg_ptr, int, int, int *);
|
||
static void permute_partial_schedule (partial_schedule_ptr, rtx);
|
||
static void generate_prolog_epilog (partial_schedule_ptr, struct loop *,
|
||
rtx, rtx);
|
||
static void duplicate_insns_of_cycles (partial_schedule_ptr,
|
||
int, int, int, rtx);
|
||
|
||
#define SCHED_ASAP(x) (((node_sched_params_ptr)(x)->aux.info)->asap)
|
||
#define SCHED_TIME(x) (((node_sched_params_ptr)(x)->aux.info)->time)
|
||
#define SCHED_FIRST_REG_MOVE(x) \
|
||
(((node_sched_params_ptr)(x)->aux.info)->first_reg_move)
|
||
#define SCHED_NREG_MOVES(x) \
|
||
(((node_sched_params_ptr)(x)->aux.info)->nreg_moves)
|
||
#define SCHED_ROW(x) (((node_sched_params_ptr)(x)->aux.info)->row)
|
||
#define SCHED_STAGE(x) (((node_sched_params_ptr)(x)->aux.info)->stage)
|
||
#define SCHED_COLUMN(x) (((node_sched_params_ptr)(x)->aux.info)->column)
|
||
|
||
/* The scheduling parameters held for each node. */
|
||
typedef struct node_sched_params
|
||
{
|
||
int asap; /* A lower-bound on the absolute scheduling cycle. */
|
||
int time; /* The absolute scheduling cycle (time >= asap). */
|
||
|
||
/* The following field (first_reg_move) is a pointer to the first
|
||
register-move instruction added to handle the modulo-variable-expansion
|
||
of the register defined by this node. This register-move copies the
|
||
original register defined by the node. */
|
||
rtx first_reg_move;
|
||
|
||
/* The number of register-move instructions added, immediately preceding
|
||
first_reg_move. */
|
||
int nreg_moves;
|
||
|
||
int row; /* Holds time % ii. */
|
||
int stage; /* Holds time / ii. */
|
||
|
||
/* The column of a node inside the ps. If nodes u, v are on the same row,
|
||
u will precede v if column (u) < column (v). */
|
||
int column;
|
||
} *node_sched_params_ptr;
|
||
|
||
|
||
/* The following three functions are copied from the current scheduler
|
||
code in order to use sched_analyze() for computing the dependencies.
|
||
They are used when initializing the sched_info structure. */
|
||
static const char *
|
||
sms_print_insn (const_rtx insn, int aligned ATTRIBUTE_UNUSED)
|
||
{
|
||
static char tmp[80];
|
||
|
||
sprintf (tmp, "i%4d", INSN_UID (insn));
|
||
return tmp;
|
||
}
|
||
|
||
static void
|
||
compute_jump_reg_dependencies (rtx insn ATTRIBUTE_UNUSED,
|
||
regset cond_exec ATTRIBUTE_UNUSED,
|
||
regset used ATTRIBUTE_UNUSED,
|
||
regset set ATTRIBUTE_UNUSED)
|
||
{
|
||
}
|
||
|
||
static struct common_sched_info_def sms_common_sched_info;
|
||
|
||
static struct sched_deps_info_def sms_sched_deps_info =
|
||
{
|
||
compute_jump_reg_dependencies,
|
||
NULL, NULL, NULL, NULL, NULL, NULL, NULL, NULL, NULL, NULL,
|
||
NULL,
|
||
0, 0, 0
|
||
};
|
||
|
||
static struct haifa_sched_info sms_sched_info =
|
||
{
|
||
NULL,
|
||
NULL,
|
||
NULL,
|
||
NULL,
|
||
NULL,
|
||
sms_print_insn,
|
||
NULL,
|
||
NULL, /* insn_finishes_block_p */
|
||
NULL, NULL,
|
||
NULL, NULL,
|
||
0, 0,
|
||
|
||
NULL, NULL, NULL,
|
||
0
|
||
};
|
||
|
||
/* Given HEAD and TAIL which are the first and last insns in a loop;
|
||
return the register which controls the loop. Return zero if it has
|
||
more than one occurrence in the loop besides the control part or the
|
||
do-loop pattern is not of the form we expect. */
|
||
static rtx
|
||
doloop_register_get (rtx head ATTRIBUTE_UNUSED, rtx tail ATTRIBUTE_UNUSED)
|
||
{
|
||
#ifdef HAVE_doloop_end
|
||
rtx reg, condition, insn, first_insn_not_to_check;
|
||
|
||
if (!JUMP_P (tail))
|
||
return NULL_RTX;
|
||
|
||
/* TODO: Free SMS's dependence on doloop_condition_get. */
|
||
condition = doloop_condition_get (tail);
|
||
if (! condition)
|
||
return NULL_RTX;
|
||
|
||
if (REG_P (XEXP (condition, 0)))
|
||
reg = XEXP (condition, 0);
|
||
else if (GET_CODE (XEXP (condition, 0)) == PLUS
|
||
&& REG_P (XEXP (XEXP (condition, 0), 0)))
|
||
reg = XEXP (XEXP (condition, 0), 0);
|
||
else
|
||
gcc_unreachable ();
|
||
|
||
/* Check that the COUNT_REG has no other occurrences in the loop
|
||
until the decrement. We assume the control part consists of
|
||
either a single (parallel) branch-on-count or a (non-parallel)
|
||
branch immediately preceded by a single (decrement) insn. */
|
||
first_insn_not_to_check = (GET_CODE (PATTERN (tail)) == PARALLEL ? tail
|
||
: PREV_INSN (tail));
|
||
|
||
for (insn = head; insn != first_insn_not_to_check; insn = NEXT_INSN (insn))
|
||
if (reg_mentioned_p (reg, insn))
|
||
{
|
||
if (dump_file)
|
||
{
|
||
fprintf (dump_file, "SMS count_reg found ");
|
||
print_rtl_single (dump_file, reg);
|
||
fprintf (dump_file, " outside control in insn:\n");
|
||
print_rtl_single (dump_file, insn);
|
||
}
|
||
|
||
return NULL_RTX;
|
||
}
|
||
|
||
return reg;
|
||
#else
|
||
return NULL_RTX;
|
||
#endif
|
||
}
|
||
|
||
/* Check if COUNT_REG is set to a constant in the PRE_HEADER block, so
|
||
that the number of iterations is a compile-time constant. If so,
|
||
return the rtx that sets COUNT_REG to a constant, and set COUNT to
|
||
this constant. Otherwise return 0. */
|
||
static rtx
|
||
const_iteration_count (rtx count_reg, basic_block pre_header,
|
||
HOST_WIDEST_INT * count)
|
||
{
|
||
rtx insn;
|
||
rtx head, tail;
|
||
|
||
if (! pre_header)
|
||
return NULL_RTX;
|
||
|
||
get_ebb_head_tail (pre_header, pre_header, &head, &tail);
|
||
|
||
for (insn = tail; insn != PREV_INSN (head); insn = PREV_INSN (insn))
|
||
if (NONDEBUG_INSN_P (insn) && single_set (insn) &&
|
||
rtx_equal_p (count_reg, SET_DEST (single_set (insn))))
|
||
{
|
||
rtx pat = single_set (insn);
|
||
|
||
if (CONST_INT_P (SET_SRC (pat)))
|
||
{
|
||
*count = INTVAL (SET_SRC (pat));
|
||
return insn;
|
||
}
|
||
|
||
return NULL_RTX;
|
||
}
|
||
|
||
return NULL_RTX;
|
||
}
|
||
|
||
/* A very simple resource-based lower bound on the initiation interval.
|
||
??? Improve the accuracy of this bound by considering the
|
||
utilization of various units. */
|
||
static int
|
||
res_MII (ddg_ptr g)
|
||
{
|
||
if (targetm.sched.sms_res_mii)
|
||
return targetm.sched.sms_res_mii (g);
|
||
|
||
return ((g->num_nodes - g->num_debug) / issue_rate);
|
||
}
|
||
|
||
|
||
/* Points to the array that contains the sched data for each node. */
|
||
static node_sched_params_ptr node_sched_params;
|
||
|
||
/* Allocate sched_params for each node and initialize it. Assumes that
|
||
the aux field of each node contain the asap bound (computed earlier),
|
||
and copies it into the sched_params field. */
|
||
static void
|
||
set_node_sched_params (ddg_ptr g)
|
||
{
|
||
int i;
|
||
|
||
/* Allocate for each node in the DDG a place to hold the "sched_data". */
|
||
/* Initialize ASAP/ALAP/HIGHT to zero. */
|
||
node_sched_params = (node_sched_params_ptr)
|
||
xcalloc (g->num_nodes,
|
||
sizeof (struct node_sched_params));
|
||
|
||
/* Set the pointer of the general data of the node to point to the
|
||
appropriate sched_params structure. */
|
||
for (i = 0; i < g->num_nodes; i++)
|
||
{
|
||
/* Watch out for aliasing problems? */
|
||
node_sched_params[i].asap = g->nodes[i].aux.count;
|
||
g->nodes[i].aux.info = &node_sched_params[i];
|
||
}
|
||
}
|
||
|
||
static void
|
||
print_node_sched_params (FILE *file, int num_nodes, ddg_ptr g)
|
||
{
|
||
int i;
|
||
|
||
if (! file)
|
||
return;
|
||
for (i = 0; i < num_nodes; i++)
|
||
{
|
||
node_sched_params_ptr nsp = &node_sched_params[i];
|
||
rtx reg_move = nsp->first_reg_move;
|
||
int j;
|
||
|
||
fprintf (file, "Node = %d; INSN = %d\n", i,
|
||
(INSN_UID (g->nodes[i].insn)));
|
||
fprintf (file, " asap = %d:\n", nsp->asap);
|
||
fprintf (file, " time = %d:\n", nsp->time);
|
||
fprintf (file, " nreg_moves = %d:\n", nsp->nreg_moves);
|
||
for (j = 0; j < nsp->nreg_moves; j++)
|
||
{
|
||
fprintf (file, " reg_move = ");
|
||
print_rtl_single (file, reg_move);
|
||
reg_move = PREV_INSN (reg_move);
|
||
}
|
||
}
|
||
}
|
||
|
||
/*
|
||
Breaking intra-loop register anti-dependences:
|
||
Each intra-loop register anti-dependence implies a cross-iteration true
|
||
dependence of distance 1. Therefore, we can remove such false dependencies
|
||
and figure out if the partial schedule broke them by checking if (for a
|
||
true-dependence of distance 1): SCHED_TIME (def) < SCHED_TIME (use) and
|
||
if so generate a register move. The number of such moves is equal to:
|
||
SCHED_TIME (use) - SCHED_TIME (def) { 0 broken
|
||
nreg_moves = ----------------------------------- + 1 - { dependence.
|
||
ii { 1 if not.
|
||
*/
|
||
static struct undo_replace_buff_elem *
|
||
generate_reg_moves (partial_schedule_ptr ps, bool rescan)
|
||
{
|
||
ddg_ptr g = ps->g;
|
||
int ii = ps->ii;
|
||
int i;
|
||
struct undo_replace_buff_elem *reg_move_replaces = NULL;
|
||
|
||
for (i = 0; i < g->num_nodes; i++)
|
||
{
|
||
ddg_node_ptr u = &g->nodes[i];
|
||
ddg_edge_ptr e;
|
||
int nreg_moves = 0, i_reg_move;
|
||
sbitmap *uses_of_defs;
|
||
rtx last_reg_move;
|
||
rtx prev_reg, old_reg;
|
||
|
||
/* Compute the number of reg_moves needed for u, by looking at life
|
||
ranges started at u (excluding self-loops). */
|
||
for (e = u->out; e; e = e->next_out)
|
||
if (e->type == TRUE_DEP && e->dest != e->src)
|
||
{
|
||
int nreg_moves4e = (SCHED_TIME (e->dest) - SCHED_TIME (e->src)) / ii;
|
||
|
||
if (e->distance == 1)
|
||
nreg_moves4e = (SCHED_TIME (e->dest) - SCHED_TIME (e->src) + ii) / ii;
|
||
|
||
/* If dest precedes src in the schedule of the kernel, then dest
|
||
will read before src writes and we can save one reg_copy. */
|
||
if (SCHED_ROW (e->dest) == SCHED_ROW (e->src)
|
||
&& SCHED_COLUMN (e->dest) < SCHED_COLUMN (e->src))
|
||
nreg_moves4e--;
|
||
|
||
nreg_moves = MAX (nreg_moves, nreg_moves4e);
|
||
}
|
||
|
||
if (nreg_moves == 0)
|
||
continue;
|
||
|
||
/* Every use of the register defined by node may require a different
|
||
copy of this register, depending on the time the use is scheduled.
|
||
Set a bitmap vector, telling which nodes use each copy of this
|
||
register. */
|
||
uses_of_defs = sbitmap_vector_alloc (nreg_moves, g->num_nodes);
|
||
sbitmap_vector_zero (uses_of_defs, nreg_moves);
|
||
for (e = u->out; e; e = e->next_out)
|
||
if (e->type == TRUE_DEP && e->dest != e->src)
|
||
{
|
||
int dest_copy = (SCHED_TIME (e->dest) - SCHED_TIME (e->src)) / ii;
|
||
|
||
if (e->distance == 1)
|
||
dest_copy = (SCHED_TIME (e->dest) - SCHED_TIME (e->src) + ii) / ii;
|
||
|
||
if (SCHED_ROW (e->dest) == SCHED_ROW (e->src)
|
||
&& SCHED_COLUMN (e->dest) < SCHED_COLUMN (e->src))
|
||
dest_copy--;
|
||
|
||
if (dest_copy)
|
||
SET_BIT (uses_of_defs[dest_copy - 1], e->dest->cuid);
|
||
}
|
||
|
||
/* Now generate the reg_moves, attaching relevant uses to them. */
|
||
SCHED_NREG_MOVES (u) = nreg_moves;
|
||
old_reg = prev_reg = copy_rtx (SET_DEST (single_set (u->insn)));
|
||
/* Insert the reg-moves right before the notes which precede
|
||
the insn they relates to. */
|
||
last_reg_move = u->first_note;
|
||
|
||
for (i_reg_move = 0; i_reg_move < nreg_moves; i_reg_move++)
|
||
{
|
||
unsigned int i_use = 0;
|
||
rtx new_reg = gen_reg_rtx (GET_MODE (prev_reg));
|
||
rtx reg_move = gen_move_insn (new_reg, prev_reg);
|
||
sbitmap_iterator sbi;
|
||
|
||
add_insn_before (reg_move, last_reg_move, NULL);
|
||
last_reg_move = reg_move;
|
||
|
||
if (!SCHED_FIRST_REG_MOVE (u))
|
||
SCHED_FIRST_REG_MOVE (u) = reg_move;
|
||
|
||
EXECUTE_IF_SET_IN_SBITMAP (uses_of_defs[i_reg_move], 0, i_use, sbi)
|
||
{
|
||
struct undo_replace_buff_elem *rep;
|
||
|
||
rep = (struct undo_replace_buff_elem *)
|
||
xcalloc (1, sizeof (struct undo_replace_buff_elem));
|
||
rep->insn = g->nodes[i_use].insn;
|
||
rep->orig_reg = old_reg;
|
||
rep->new_reg = new_reg;
|
||
|
||
if (! reg_move_replaces)
|
||
reg_move_replaces = rep;
|
||
else
|
||
{
|
||
rep->next = reg_move_replaces;
|
||
reg_move_replaces = rep;
|
||
}
|
||
|
||
replace_rtx (g->nodes[i_use].insn, old_reg, new_reg);
|
||
if (rescan)
|
||
df_insn_rescan (g->nodes[i_use].insn);
|
||
}
|
||
|
||
prev_reg = new_reg;
|
||
}
|
||
sbitmap_vector_free (uses_of_defs);
|
||
}
|
||
return reg_move_replaces;
|
||
}
|
||
|
||
/* Free memory allocated for the undo buffer. */
|
||
static void
|
||
free_undo_replace_buff (struct undo_replace_buff_elem *reg_move_replaces)
|
||
{
|
||
|
||
while (reg_move_replaces)
|
||
{
|
||
struct undo_replace_buff_elem *rep = reg_move_replaces;
|
||
|
||
reg_move_replaces = reg_move_replaces->next;
|
||
free (rep);
|
||
}
|
||
}
|
||
|
||
/* Bump the SCHED_TIMEs of all nodes to start from zero. Set the values
|
||
of SCHED_ROW and SCHED_STAGE. */
|
||
static void
|
||
normalize_sched_times (partial_schedule_ptr ps)
|
||
{
|
||
int row;
|
||
int amount = PS_MIN_CYCLE (ps);
|
||
int ii = ps->ii;
|
||
ps_insn_ptr crr_insn;
|
||
|
||
for (row = 0; row < ii; row++)
|
||
for (crr_insn = ps->rows[row]; crr_insn; crr_insn = crr_insn->next_in_row)
|
||
{
|
||
ddg_node_ptr u = crr_insn->node;
|
||
int normalized_time = SCHED_TIME (u) - amount;
|
||
|
||
if (dump_file)
|
||
fprintf (dump_file, "crr_insn->node=%d, crr_insn->cycle=%d,\
|
||
min_cycle=%d\n", crr_insn->node->cuid, SCHED_TIME
|
||
(u), ps->min_cycle);
|
||
gcc_assert (SCHED_TIME (u) >= ps->min_cycle);
|
||
gcc_assert (SCHED_TIME (u) <= ps->max_cycle);
|
||
SCHED_TIME (u) = normalized_time;
|
||
SCHED_ROW (u) = normalized_time % ii;
|
||
SCHED_STAGE (u) = normalized_time / ii;
|
||
}
|
||
}
|
||
|
||
/* Set SCHED_COLUMN of each node according to its position in PS. */
|
||
static void
|
||
set_columns_for_ps (partial_schedule_ptr ps)
|
||
{
|
||
int row;
|
||
|
||
for (row = 0; row < ps->ii; row++)
|
||
{
|
||
ps_insn_ptr cur_insn = ps->rows[row];
|
||
int column = 0;
|
||
|
||
for (; cur_insn; cur_insn = cur_insn->next_in_row)
|
||
SCHED_COLUMN (cur_insn->node) = column++;
|
||
}
|
||
}
|
||
|
||
/* Permute the insns according to their order in PS, from row 0 to
|
||
row ii-1, and position them right before LAST. This schedules
|
||
the insns of the loop kernel. */
|
||
static void
|
||
permute_partial_schedule (partial_schedule_ptr ps, rtx last)
|
||
{
|
||
int ii = ps->ii;
|
||
int row;
|
||
ps_insn_ptr ps_ij;
|
||
|
||
for (row = 0; row < ii ; row++)
|
||
for (ps_ij = ps->rows[row]; ps_ij; ps_ij = ps_ij->next_in_row)
|
||
if (PREV_INSN (last) != ps_ij->node->insn)
|
||
reorder_insns_nobb (ps_ij->node->first_note, ps_ij->node->insn,
|
||
PREV_INSN (last));
|
||
}
|
||
|
||
static void
|
||
duplicate_insns_of_cycles (partial_schedule_ptr ps, int from_stage,
|
||
int to_stage, int for_prolog, rtx count_reg)
|
||
{
|
||
int row;
|
||
ps_insn_ptr ps_ij;
|
||
|
||
for (row = 0; row < ps->ii; row++)
|
||
for (ps_ij = ps->rows[row]; ps_ij; ps_ij = ps_ij->next_in_row)
|
||
{
|
||
ddg_node_ptr u_node = ps_ij->node;
|
||
int j, i_reg_moves;
|
||
rtx reg_move = NULL_RTX;
|
||
|
||
/* Do not duplicate any insn which refers to count_reg as it
|
||
belongs to the control part.
|
||
TODO: This should be done by analyzing the control part of
|
||
the loop. */
|
||
if (reg_mentioned_p (count_reg, u_node->insn))
|
||
continue;
|
||
|
||
if (for_prolog)
|
||
{
|
||
/* SCHED_STAGE (u_node) >= from_stage == 0. Generate increasing
|
||
number of reg_moves starting with the second occurrence of
|
||
u_node, which is generated if its SCHED_STAGE <= to_stage. */
|
||
i_reg_moves = to_stage - SCHED_STAGE (u_node) + 1;
|
||
i_reg_moves = MAX (i_reg_moves, 0);
|
||
i_reg_moves = MIN (i_reg_moves, SCHED_NREG_MOVES (u_node));
|
||
|
||
/* The reg_moves start from the *first* reg_move backwards. */
|
||
if (i_reg_moves)
|
||
{
|
||
reg_move = SCHED_FIRST_REG_MOVE (u_node);
|
||
for (j = 1; j < i_reg_moves; j++)
|
||
reg_move = PREV_INSN (reg_move);
|
||
}
|
||
}
|
||
else /* It's for the epilog. */
|
||
{
|
||
/* SCHED_STAGE (u_node) <= to_stage. Generate all reg_moves,
|
||
starting to decrease one stage after u_node no longer occurs;
|
||
that is, generate all reg_moves until
|
||
SCHED_STAGE (u_node) == from_stage - 1. */
|
||
i_reg_moves = SCHED_NREG_MOVES (u_node)
|
||
- (from_stage - SCHED_STAGE (u_node) - 1);
|
||
i_reg_moves = MAX (i_reg_moves, 0);
|
||
i_reg_moves = MIN (i_reg_moves, SCHED_NREG_MOVES (u_node));
|
||
|
||
/* The reg_moves start from the *last* reg_move forwards. */
|
||
if (i_reg_moves)
|
||
{
|
||
reg_move = SCHED_FIRST_REG_MOVE (u_node);
|
||
for (j = 1; j < SCHED_NREG_MOVES (u_node); j++)
|
||
reg_move = PREV_INSN (reg_move);
|
||
}
|
||
}
|
||
|
||
for (j = 0; j < i_reg_moves; j++, reg_move = NEXT_INSN (reg_move))
|
||
emit_insn (copy_rtx (PATTERN (reg_move)));
|
||
if (SCHED_STAGE (u_node) >= from_stage
|
||
&& SCHED_STAGE (u_node) <= to_stage)
|
||
duplicate_insn_chain (u_node->first_note, u_node->insn);
|
||
}
|
||
}
|
||
|
||
|
||
/* Generate the instructions (including reg_moves) for prolog & epilog. */
|
||
static void
|
||
generate_prolog_epilog (partial_schedule_ptr ps, struct loop *loop,
|
||
rtx count_reg, rtx count_init)
|
||
{
|
||
int i;
|
||
int last_stage = PS_STAGE_COUNT (ps) - 1;
|
||
edge e;
|
||
|
||
/* Generate the prolog, inserting its insns on the loop-entry edge. */
|
||
start_sequence ();
|
||
|
||
if (!count_init)
|
||
{
|
||
/* Generate instructions at the beginning of the prolog to
|
||
adjust the loop count by STAGE_COUNT. If loop count is constant
|
||
(count_init), this constant is adjusted by STAGE_COUNT in
|
||
generate_prolog_epilog function. */
|
||
rtx sub_reg = NULL_RTX;
|
||
|
||
sub_reg = expand_simple_binop (GET_MODE (count_reg), MINUS,
|
||
count_reg, GEN_INT (last_stage),
|
||
count_reg, 1, OPTAB_DIRECT);
|
||
gcc_assert (REG_P (sub_reg));
|
||
if (REGNO (sub_reg) != REGNO (count_reg))
|
||
emit_move_insn (count_reg, sub_reg);
|
||
}
|
||
|
||
for (i = 0; i < last_stage; i++)
|
||
duplicate_insns_of_cycles (ps, 0, i, 1, count_reg);
|
||
|
||
/* Put the prolog on the entry edge. */
|
||
e = loop_preheader_edge (loop);
|
||
split_edge_and_insert (e, get_insns ());
|
||
|
||
end_sequence ();
|
||
|
||
/* Generate the epilog, inserting its insns on the loop-exit edge. */
|
||
start_sequence ();
|
||
|
||
for (i = 0; i < last_stage; i++)
|
||
duplicate_insns_of_cycles (ps, i + 1, last_stage, 0, count_reg);
|
||
|
||
/* Put the epilogue on the exit edge. */
|
||
gcc_assert (single_exit (loop));
|
||
e = single_exit (loop);
|
||
split_edge_and_insert (e, get_insns ());
|
||
end_sequence ();
|
||
}
|
||
|
||
/* Return true if all the BBs of the loop are empty except the
|
||
loop header. */
|
||
static bool
|
||
loop_single_full_bb_p (struct loop *loop)
|
||
{
|
||
unsigned i;
|
||
basic_block *bbs = get_loop_body (loop);
|
||
|
||
for (i = 0; i < loop->num_nodes ; i++)
|
||
{
|
||
rtx head, tail;
|
||
bool empty_bb = true;
|
||
|
||
if (bbs[i] == loop->header)
|
||
continue;
|
||
|
||
/* Make sure that basic blocks other than the header
|
||
have only notes labels or jumps. */
|
||
get_ebb_head_tail (bbs[i], bbs[i], &head, &tail);
|
||
for (; head != NEXT_INSN (tail); head = NEXT_INSN (head))
|
||
{
|
||
if (NOTE_P (head) || LABEL_P (head)
|
||
|| (INSN_P (head) && (DEBUG_INSN_P (head) || JUMP_P (head))))
|
||
continue;
|
||
empty_bb = false;
|
||
break;
|
||
}
|
||
|
||
if (! empty_bb)
|
||
{
|
||
free (bbs);
|
||
return false;
|
||
}
|
||
}
|
||
free (bbs);
|
||
return true;
|
||
}
|
||
|
||
/* A simple loop from SMS point of view; it is a loop that is composed of
|
||
either a single basic block or two BBs - a header and a latch. */
|
||
#define SIMPLE_SMS_LOOP_P(loop) ((loop->num_nodes < 3 ) \
|
||
&& (EDGE_COUNT (loop->latch->preds) == 1) \
|
||
&& (EDGE_COUNT (loop->latch->succs) == 1))
|
||
|
||
/* Return true if the loop is in its canonical form and false if not.
|
||
i.e. SIMPLE_SMS_LOOP_P and have one preheader block, and single exit. */
|
||
static bool
|
||
loop_canon_p (struct loop *loop)
|
||
{
|
||
|
||
if (loop->inner || !loop_outer (loop))
|
||
{
|
||
if (dump_file)
|
||
fprintf (dump_file, "SMS loop inner or !loop_outer\n");
|
||
return false;
|
||
}
|
||
|
||
if (!single_exit (loop))
|
||
{
|
||
if (dump_file)
|
||
{
|
||
rtx insn = BB_END (loop->header);
|
||
|
||
fprintf (dump_file, "SMS loop many exits ");
|
||
fprintf (dump_file, " %s %d (file, line)\n",
|
||
insn_file (insn), insn_line (insn));
|
||
}
|
||
return false;
|
||
}
|
||
|
||
if (! SIMPLE_SMS_LOOP_P (loop) && ! loop_single_full_bb_p (loop))
|
||
{
|
||
if (dump_file)
|
||
{
|
||
rtx insn = BB_END (loop->header);
|
||
|
||
fprintf (dump_file, "SMS loop many BBs. ");
|
||
fprintf (dump_file, " %s %d (file, line)\n",
|
||
insn_file (insn), insn_line (insn));
|
||
}
|
||
return false;
|
||
}
|
||
|
||
return true;
|
||
}
|
||
|
||
/* If there are more than one entry for the loop,
|
||
make it one by splitting the first entry edge and
|
||
redirecting the others to the new BB. */
|
||
static void
|
||
canon_loop (struct loop *loop)
|
||
{
|
||
edge e;
|
||
edge_iterator i;
|
||
|
||
/* Avoid annoying special cases of edges going to exit
|
||
block. */
|
||
FOR_EACH_EDGE (e, i, EXIT_BLOCK_PTR->preds)
|
||
if ((e->flags & EDGE_FALLTHRU) && (EDGE_COUNT (e->src->succs) > 1))
|
||
split_edge (e);
|
||
|
||
if (loop->latch == loop->header
|
||
|| EDGE_COUNT (loop->latch->succs) > 1)
|
||
{
|
||
FOR_EACH_EDGE (e, i, loop->header->preds)
|
||
if (e->src == loop->latch)
|
||
break;
|
||
split_edge (e);
|
||
}
|
||
}
|
||
|
||
/* Setup infos. */
|
||
static void
|
||
setup_sched_infos (void)
|
||
{
|
||
memcpy (&sms_common_sched_info, &haifa_common_sched_info,
|
||
sizeof (sms_common_sched_info));
|
||
sms_common_sched_info.sched_pass_id = SCHED_SMS_PASS;
|
||
common_sched_info = &sms_common_sched_info;
|
||
|
||
sched_deps_info = &sms_sched_deps_info;
|
||
current_sched_info = &sms_sched_info;
|
||
}
|
||
|
||
/* Probability in % that the sms-ed loop rolls enough so that optimized
|
||
version may be entered. Just a guess. */
|
||
#define PROB_SMS_ENOUGH_ITERATIONS 80
|
||
|
||
/* Used to calculate the upper bound of ii. */
|
||
#define MAXII_FACTOR 2
|
||
|
||
/* Main entry point, perform SMS scheduling on the loops of the function
|
||
that consist of single basic blocks. */
|
||
static void
|
||
sms_schedule (void)
|
||
{
|
||
rtx insn;
|
||
ddg_ptr *g_arr, g;
|
||
int * node_order;
|
||
int maxii, max_asap;
|
||
loop_iterator li;
|
||
partial_schedule_ptr ps;
|
||
basic_block bb = NULL;
|
||
struct loop *loop;
|
||
basic_block condition_bb = NULL;
|
||
edge latch_edge;
|
||
gcov_type trip_count = 0;
|
||
|
||
loop_optimizer_init (LOOPS_HAVE_PREHEADERS
|
||
| LOOPS_HAVE_RECORDED_EXITS);
|
||
if (number_of_loops () <= 1)
|
||
{
|
||
loop_optimizer_finalize ();
|
||
return; /* There are no loops to schedule. */
|
||
}
|
||
|
||
/* Initialize issue_rate. */
|
||
if (targetm.sched.issue_rate)
|
||
{
|
||
int temp = reload_completed;
|
||
|
||
reload_completed = 1;
|
||
issue_rate = targetm.sched.issue_rate ();
|
||
reload_completed = temp;
|
||
}
|
||
else
|
||
issue_rate = 1;
|
||
|
||
/* Initialize the scheduler. */
|
||
setup_sched_infos ();
|
||
haifa_sched_init ();
|
||
|
||
/* Allocate memory to hold the DDG array one entry for each loop.
|
||
We use loop->num as index into this array. */
|
||
g_arr = XCNEWVEC (ddg_ptr, number_of_loops ());
|
||
|
||
if (dump_file)
|
||
{
|
||
fprintf (dump_file, "\n\nSMS analysis phase\n");
|
||
fprintf (dump_file, "===================\n\n");
|
||
}
|
||
|
||
/* Build DDGs for all the relevant loops and hold them in G_ARR
|
||
indexed by the loop index. */
|
||
FOR_EACH_LOOP (li, loop, 0)
|
||
{
|
||
rtx head, tail;
|
||
rtx count_reg;
|
||
|
||
/* For debugging. */
|
||
if (dbg_cnt (sms_sched_loop) == false)
|
||
{
|
||
if (dump_file)
|
||
fprintf (dump_file, "SMS reached max limit... \n");
|
||
|
||
break;
|
||
}
|
||
|
||
if (dump_file)
|
||
{
|
||
rtx insn = BB_END (loop->header);
|
||
|
||
fprintf (dump_file, "SMS loop num: %d, file: %s, line: %d\n",
|
||
loop->num, insn_file (insn), insn_line (insn));
|
||
|
||
}
|
||
|
||
if (! loop_canon_p (loop))
|
||
continue;
|
||
|
||
if (! loop_single_full_bb_p (loop))
|
||
{
|
||
if (dump_file)
|
||
fprintf (dump_file, "SMS not loop_single_full_bb_p\n");
|
||
continue;
|
||
}
|
||
|
||
bb = loop->header;
|
||
|
||
get_ebb_head_tail (bb, bb, &head, &tail);
|
||
latch_edge = loop_latch_edge (loop);
|
||
gcc_assert (single_exit (loop));
|
||
if (single_exit (loop)->count)
|
||
trip_count = latch_edge->count / single_exit (loop)->count;
|
||
|
||
/* Perform SMS only on loops that their average count is above threshold. */
|
||
|
||
if ( latch_edge->count
|
||
&& (latch_edge->count < single_exit (loop)->count * SMS_LOOP_AVERAGE_COUNT_THRESHOLD))
|
||
{
|
||
if (dump_file)
|
||
{
|
||
fprintf (dump_file, " %s %d (file, line)\n",
|
||
insn_file (tail), insn_line (tail));
|
||
fprintf (dump_file, "SMS single-bb-loop\n");
|
||
if (profile_info && flag_branch_probabilities)
|
||
{
|
||
fprintf (dump_file, "SMS loop-count ");
|
||
fprintf (dump_file, HOST_WIDEST_INT_PRINT_DEC,
|
||
(HOST_WIDEST_INT) bb->count);
|
||
fprintf (dump_file, "\n");
|
||
fprintf (dump_file, "SMS trip-count ");
|
||
fprintf (dump_file, HOST_WIDEST_INT_PRINT_DEC,
|
||
(HOST_WIDEST_INT) trip_count);
|
||
fprintf (dump_file, "\n");
|
||
fprintf (dump_file, "SMS profile-sum-max ");
|
||
fprintf (dump_file, HOST_WIDEST_INT_PRINT_DEC,
|
||
(HOST_WIDEST_INT) profile_info->sum_max);
|
||
fprintf (dump_file, "\n");
|
||
}
|
||
}
|
||
continue;
|
||
}
|
||
|
||
/* Make sure this is a doloop. */
|
||
if ( !(count_reg = doloop_register_get (head, tail)))
|
||
{
|
||
if (dump_file)
|
||
fprintf (dump_file, "SMS doloop_register_get failed\n");
|
||
continue;
|
||
}
|
||
|
||
/* Don't handle BBs with calls or barriers, or !single_set insns,
|
||
or auto-increment insns (to avoid creating invalid reg-moves
|
||
for the auto-increment insns).
|
||
??? Should handle auto-increment insns.
|
||
??? Should handle insns defining subregs. */
|
||
for (insn = head; insn != NEXT_INSN (tail); insn = NEXT_INSN (insn))
|
||
{
|
||
rtx set;
|
||
|
||
if (CALL_P (insn)
|
||
|| BARRIER_P (insn)
|
||
|| (NONDEBUG_INSN_P (insn) && !JUMP_P (insn)
|
||
&& !single_set (insn) && GET_CODE (PATTERN (insn)) != USE)
|
||
|| (FIND_REG_INC_NOTE (insn, NULL_RTX) != 0)
|
||
|| (INSN_P (insn) && (set = single_set (insn))
|
||
&& GET_CODE (SET_DEST (set)) == SUBREG))
|
||
break;
|
||
}
|
||
|
||
if (insn != NEXT_INSN (tail))
|
||
{
|
||
if (dump_file)
|
||
{
|
||
if (CALL_P (insn))
|
||
fprintf (dump_file, "SMS loop-with-call\n");
|
||
else if (BARRIER_P (insn))
|
||
fprintf (dump_file, "SMS loop-with-barrier\n");
|
||
else if (FIND_REG_INC_NOTE (insn, NULL_RTX) != 0)
|
||
fprintf (dump_file, "SMS reg inc\n");
|
||
else if ((NONDEBUG_INSN_P (insn) && !JUMP_P (insn)
|
||
&& !single_set (insn) && GET_CODE (PATTERN (insn)) != USE))
|
||
fprintf (dump_file, "SMS loop-with-not-single-set\n");
|
||
else
|
||
fprintf (dump_file, "SMS loop with subreg in lhs\n");
|
||
print_rtl_single (dump_file, insn);
|
||
}
|
||
|
||
continue;
|
||
}
|
||
|
||
if (! (g = create_ddg (bb, 0)))
|
||
{
|
||
if (dump_file)
|
||
fprintf (dump_file, "SMS create_ddg failed\n");
|
||
continue;
|
||
}
|
||
|
||
g_arr[loop->num] = g;
|
||
if (dump_file)
|
||
fprintf (dump_file, "...OK\n");
|
||
|
||
}
|
||
if (dump_file)
|
||
{
|
||
fprintf (dump_file, "\nSMS transformation phase\n");
|
||
fprintf (dump_file, "=========================\n\n");
|
||
}
|
||
|
||
/* We don't want to perform SMS on new loops - created by versioning. */
|
||
FOR_EACH_LOOP (li, loop, 0)
|
||
{
|
||
rtx head, tail;
|
||
rtx count_reg, count_init;
|
||
int mii, rec_mii;
|
||
unsigned stage_count = 0;
|
||
HOST_WIDEST_INT loop_count = 0;
|
||
|
||
if (! (g = g_arr[loop->num]))
|
||
continue;
|
||
|
||
if (dump_file)
|
||
{
|
||
rtx insn = BB_END (loop->header);
|
||
|
||
fprintf (dump_file, "SMS loop num: %d, file: %s, line: %d\n",
|
||
loop->num, insn_file (insn), insn_line (insn));
|
||
|
||
print_ddg (dump_file, g);
|
||
}
|
||
|
||
get_ebb_head_tail (loop->header, loop->header, &head, &tail);
|
||
|
||
latch_edge = loop_latch_edge (loop);
|
||
gcc_assert (single_exit (loop));
|
||
if (single_exit (loop)->count)
|
||
trip_count = latch_edge->count / single_exit (loop)->count;
|
||
|
||
if (dump_file)
|
||
{
|
||
fprintf (dump_file, " %s %d (file, line)\n",
|
||
insn_file (tail), insn_line (tail));
|
||
fprintf (dump_file, "SMS single-bb-loop\n");
|
||
if (profile_info && flag_branch_probabilities)
|
||
{
|
||
fprintf (dump_file, "SMS loop-count ");
|
||
fprintf (dump_file, HOST_WIDEST_INT_PRINT_DEC,
|
||
(HOST_WIDEST_INT) bb->count);
|
||
fprintf (dump_file, "\n");
|
||
fprintf (dump_file, "SMS profile-sum-max ");
|
||
fprintf (dump_file, HOST_WIDEST_INT_PRINT_DEC,
|
||
(HOST_WIDEST_INT) profile_info->sum_max);
|
||
fprintf (dump_file, "\n");
|
||
}
|
||
fprintf (dump_file, "SMS doloop\n");
|
||
fprintf (dump_file, "SMS built-ddg %d\n", g->num_nodes);
|
||
fprintf (dump_file, "SMS num-loads %d\n", g->num_loads);
|
||
fprintf (dump_file, "SMS num-stores %d\n", g->num_stores);
|
||
}
|
||
|
||
|
||
/* In case of th loop have doloop register it gets special
|
||
handling. */
|
||
count_init = NULL_RTX;
|
||
if ((count_reg = doloop_register_get (head, tail)))
|
||
{
|
||
basic_block pre_header;
|
||
|
||
pre_header = loop_preheader_edge (loop)->src;
|
||
count_init = const_iteration_count (count_reg, pre_header,
|
||
&loop_count);
|
||
}
|
||
gcc_assert (count_reg);
|
||
|
||
if (dump_file && count_init)
|
||
{
|
||
fprintf (dump_file, "SMS const-doloop ");
|
||
fprintf (dump_file, HOST_WIDEST_INT_PRINT_DEC,
|
||
loop_count);
|
||
fprintf (dump_file, "\n");
|
||
}
|
||
|
||
node_order = XNEWVEC (int, g->num_nodes);
|
||
|
||
mii = 1; /* Need to pass some estimate of mii. */
|
||
rec_mii = sms_order_nodes (g, mii, node_order, &max_asap);
|
||
mii = MAX (res_MII (g), rec_mii);
|
||
maxii = MAX (max_asap, MAXII_FACTOR * mii);
|
||
|
||
if (dump_file)
|
||
fprintf (dump_file, "SMS iis %d %d %d (rec_mii, mii, maxii)\n",
|
||
rec_mii, mii, maxii);
|
||
|
||
/* After sms_order_nodes and before sms_schedule_by_order, to copy over
|
||
ASAP. */
|
||
set_node_sched_params (g);
|
||
|
||
ps = sms_schedule_by_order (g, mii, maxii, node_order);
|
||
|
||
if (ps){
|
||
stage_count = PS_STAGE_COUNT (ps);
|
||
gcc_assert(stage_count >= 1);
|
||
}
|
||
|
||
/* Stage count of 1 means that there is no interleaving between
|
||
iterations, let the scheduling passes do the job. */
|
||
if (stage_count <= 1
|
||
|| (count_init && (loop_count <= stage_count))
|
||
|| (flag_branch_probabilities && (trip_count <= stage_count)))
|
||
{
|
||
if (dump_file)
|
||
{
|
||
fprintf (dump_file, "SMS failed... \n");
|
||
fprintf (dump_file, "SMS sched-failed (stage-count=%d, loop-count=", stage_count);
|
||
fprintf (dump_file, HOST_WIDEST_INT_PRINT_DEC, loop_count);
|
||
fprintf (dump_file, ", trip-count=");
|
||
fprintf (dump_file, HOST_WIDEST_INT_PRINT_DEC, trip_count);
|
||
fprintf (dump_file, ")\n");
|
||
}
|
||
continue;
|
||
}
|
||
else
|
||
{
|
||
struct undo_replace_buff_elem *reg_move_replaces;
|
||
|
||
if (dump_file)
|
||
{
|
||
fprintf (dump_file,
|
||
"SMS succeeded %d %d (with ii, sc)\n", ps->ii,
|
||
stage_count);
|
||
print_partial_schedule (ps, dump_file);
|
||
fprintf (dump_file,
|
||
"SMS Branch (%d) will later be scheduled at cycle %d.\n",
|
||
g->closing_branch->cuid, PS_MIN_CYCLE (ps) - 1);
|
||
}
|
||
|
||
/* Set the stage boundaries. If the DDG is built with closing_branch_deps,
|
||
the closing_branch was scheduled and should appear in the last (ii-1)
|
||
row. Otherwise, we are free to schedule the branch, and we let nodes
|
||
that were scheduled at the first PS_MIN_CYCLE cycle appear in the first
|
||
row; this should reduce stage_count to minimum.
|
||
TODO: Revisit the issue of scheduling the insns of the
|
||
control part relative to the branch when the control part
|
||
has more than one insn. */
|
||
normalize_sched_times (ps);
|
||
rotate_partial_schedule (ps, PS_MIN_CYCLE (ps));
|
||
set_columns_for_ps (ps);
|
||
|
||
canon_loop (loop);
|
||
|
||
/* case the BCT count is not known , Do loop-versioning */
|
||
if (count_reg && ! count_init)
|
||
{
|
||
rtx comp_rtx = gen_rtx_fmt_ee (GT, VOIDmode, count_reg,
|
||
GEN_INT(stage_count));
|
||
unsigned prob = (PROB_SMS_ENOUGH_ITERATIONS
|
||
* REG_BR_PROB_BASE) / 100;
|
||
|
||
loop_version (loop, comp_rtx, &condition_bb,
|
||
prob, prob, REG_BR_PROB_BASE - prob,
|
||
true);
|
||
}
|
||
|
||
/* Set new iteration count of loop kernel. */
|
||
if (count_reg && count_init)
|
||
SET_SRC (single_set (count_init)) = GEN_INT (loop_count
|
||
- stage_count + 1);
|
||
|
||
/* Now apply the scheduled kernel to the RTL of the loop. */
|
||
permute_partial_schedule (ps, g->closing_branch->first_note);
|
||
|
||
/* Mark this loop as software pipelined so the later
|
||
scheduling passes doesn't touch it. */
|
||
if (! flag_resched_modulo_sched)
|
||
g->bb->flags |= BB_DISABLE_SCHEDULE;
|
||
/* The life-info is not valid any more. */
|
||
df_set_bb_dirty (g->bb);
|
||
|
||
reg_move_replaces = generate_reg_moves (ps, true);
|
||
if (dump_file)
|
||
print_node_sched_params (dump_file, g->num_nodes, g);
|
||
/* Generate prolog and epilog. */
|
||
generate_prolog_epilog (ps, loop, count_reg, count_init);
|
||
|
||
free_undo_replace_buff (reg_move_replaces);
|
||
}
|
||
|
||
free_partial_schedule (ps);
|
||
free (node_sched_params);
|
||
free (node_order);
|
||
free_ddg (g);
|
||
}
|
||
|
||
free (g_arr);
|
||
|
||
/* Release scheduler data, needed until now because of DFA. */
|
||
haifa_sched_finish ();
|
||
loop_optimizer_finalize ();
|
||
}
|
||
|
||
/* The SMS scheduling algorithm itself
|
||
-----------------------------------
|
||
Input: 'O' an ordered list of insns of a loop.
|
||
Output: A scheduling of the loop - kernel, prolog, and epilogue.
|
||
|
||
'Q' is the empty Set
|
||
'PS' is the partial schedule; it holds the currently scheduled nodes with
|
||
their cycle/slot.
|
||
'PSP' previously scheduled predecessors.
|
||
'PSS' previously scheduled successors.
|
||
't(u)' the cycle where u is scheduled.
|
||
'l(u)' is the latency of u.
|
||
'd(v,u)' is the dependence distance from v to u.
|
||
'ASAP(u)' the earliest time at which u could be scheduled as computed in
|
||
the node ordering phase.
|
||
'check_hardware_resources_conflicts(u, PS, c)'
|
||
run a trace around cycle/slot through DFA model
|
||
to check resource conflicts involving instruction u
|
||
at cycle c given the partial schedule PS.
|
||
'add_to_partial_schedule_at_time(u, PS, c)'
|
||
Add the node/instruction u to the partial schedule
|
||
PS at time c.
|
||
'calculate_register_pressure(PS)'
|
||
Given a schedule of instructions, calculate the register
|
||
pressure it implies. One implementation could be the
|
||
maximum number of overlapping live ranges.
|
||
'maxRP' The maximum allowed register pressure, it is usually derived from the number
|
||
registers available in the hardware.
|
||
|
||
1. II = MII.
|
||
2. PS = empty list
|
||
3. for each node u in O in pre-computed order
|
||
4. if (PSP(u) != Q && PSS(u) == Q) then
|
||
5. Early_start(u) = max ( t(v) + l(v) - d(v,u)*II ) over all every v in PSP(u).
|
||
6. start = Early_start; end = Early_start + II - 1; step = 1
|
||
11. else if (PSP(u) == Q && PSS(u) != Q) then
|
||
12. Late_start(u) = min ( t(v) - l(v) + d(v,u)*II ) over all every v in PSS(u).
|
||
13. start = Late_start; end = Late_start - II + 1; step = -1
|
||
14. else if (PSP(u) != Q && PSS(u) != Q) then
|
||
15. Early_start(u) = max ( t(v) + l(v) - d(v,u)*II ) over all every v in PSP(u).
|
||
16. Late_start(u) = min ( t(v) - l(v) + d(v,u)*II ) over all every v in PSS(u).
|
||
17. start = Early_start;
|
||
18. end = min(Early_start + II - 1 , Late_start);
|
||
19. step = 1
|
||
20. else "if (PSP(u) == Q && PSS(u) == Q)"
|
||
21. start = ASAP(u); end = start + II - 1; step = 1
|
||
22. endif
|
||
|
||
23. success = false
|
||
24. for (c = start ; c != end ; c += step)
|
||
25. if check_hardware_resources_conflicts(u, PS, c) then
|
||
26. add_to_partial_schedule_at_time(u, PS, c)
|
||
27. success = true
|
||
28. break
|
||
29. endif
|
||
30. endfor
|
||
31. if (success == false) then
|
||
32. II = II + 1
|
||
33. if (II > maxII) then
|
||
34. finish - failed to schedule
|
||
35. endif
|
||
36. goto 2.
|
||
37. endif
|
||
38. endfor
|
||
39. if (calculate_register_pressure(PS) > maxRP) then
|
||
40. goto 32.
|
||
41. endif
|
||
42. compute epilogue & prologue
|
||
43. finish - succeeded to schedule
|
||
*/
|
||
|
||
/* A limit on the number of cycles that resource conflicts can span. ??? Should
|
||
be provided by DFA, and be dependent on the type of insn scheduled. Currently
|
||
set to 0 to save compile time. */
|
||
#define DFA_HISTORY SMS_DFA_HISTORY
|
||
|
||
/* A threshold for the number of repeated unsuccessful attempts to insert
|
||
an empty row, before we flush the partial schedule and start over. */
|
||
#define MAX_SPLIT_NUM 10
|
||
/* Given the partial schedule PS, this function calculates and returns the
|
||
cycles in which we can schedule the node with the given index I.
|
||
NOTE: Here we do the backtracking in SMS, in some special cases. We have
|
||
noticed that there are several cases in which we fail to SMS the loop
|
||
because the sched window of a node is empty due to tight data-deps. In
|
||
such cases we want to unschedule some of the predecessors/successors
|
||
until we get non-empty scheduling window. It returns -1 if the
|
||
scheduling window is empty and zero otherwise. */
|
||
|
||
static int
|
||
get_sched_window (partial_schedule_ptr ps, int *nodes_order, int i,
|
||
sbitmap sched_nodes, int ii, int *start_p, int *step_p, int *end_p)
|
||
{
|
||
int start, step, end;
|
||
ddg_edge_ptr e;
|
||
int u = nodes_order [i];
|
||
ddg_node_ptr u_node = &ps->g->nodes[u];
|
||
sbitmap psp = sbitmap_alloc (ps->g->num_nodes);
|
||
sbitmap pss = sbitmap_alloc (ps->g->num_nodes);
|
||
sbitmap u_node_preds = NODE_PREDECESSORS (u_node);
|
||
sbitmap u_node_succs = NODE_SUCCESSORS (u_node);
|
||
int psp_not_empty;
|
||
int pss_not_empty;
|
||
|
||
/* 1. compute sched window for u (start, end, step). */
|
||
sbitmap_zero (psp);
|
||
sbitmap_zero (pss);
|
||
psp_not_empty = sbitmap_a_and_b_cg (psp, u_node_preds, sched_nodes);
|
||
pss_not_empty = sbitmap_a_and_b_cg (pss, u_node_succs, sched_nodes);
|
||
|
||
if (psp_not_empty && !pss_not_empty)
|
||
{
|
||
int early_start = INT_MIN;
|
||
|
||
end = INT_MAX;
|
||
for (e = u_node->in; e != 0; e = e->next_in)
|
||
{
|
||
ddg_node_ptr v_node = e->src;
|
||
|
||
if (dump_file)
|
||
{
|
||
fprintf (dump_file, "\nProcessing edge: ");
|
||
print_ddg_edge (dump_file, e);
|
||
fprintf (dump_file,
|
||
"\nScheduling %d (%d) in psp_not_empty,"
|
||
" checking p %d (%d): ", u_node->cuid,
|
||
INSN_UID (u_node->insn), v_node->cuid, INSN_UID
|
||
(v_node->insn));
|
||
}
|
||
|
||
if (TEST_BIT (sched_nodes, v_node->cuid))
|
||
{
|
||
int p_st = SCHED_TIME (v_node);
|
||
|
||
early_start =
|
||
MAX (early_start, p_st + e->latency - (e->distance * ii));
|
||
|
||
if (dump_file)
|
||
fprintf (dump_file,
|
||
"pred st = %d; early_start = %d; latency: %d",
|
||
p_st, early_start, e->latency);
|
||
|
||
if (e->data_type == MEM_DEP)
|
||
end = MIN (end, SCHED_TIME (v_node) + ii - 1);
|
||
}
|
||
else if (dump_file)
|
||
fprintf (dump_file, "the node is not scheduled\n");
|
||
}
|
||
start = early_start;
|
||
end = MIN (end, early_start + ii);
|
||
/* Schedule the node close to it's predecessors. */
|
||
step = 1;
|
||
|
||
if (dump_file)
|
||
fprintf (dump_file,
|
||
"\nScheduling %d (%d) in a window (%d..%d) with step %d\n",
|
||
u_node->cuid, INSN_UID (u_node->insn), start, end, step);
|
||
}
|
||
|
||
else if (!psp_not_empty && pss_not_empty)
|
||
{
|
||
int late_start = INT_MAX;
|
||
|
||
end = INT_MIN;
|
||
for (e = u_node->out; e != 0; e = e->next_out)
|
||
{
|
||
ddg_node_ptr v_node = e->dest;
|
||
|
||
if (dump_file)
|
||
{
|
||
fprintf (dump_file, "\nProcessing edge:");
|
||
print_ddg_edge (dump_file, e);
|
||
fprintf (dump_file,
|
||
"\nScheduling %d (%d) in pss_not_empty,"
|
||
" checking s %d (%d): ", u_node->cuid,
|
||
INSN_UID (u_node->insn), v_node->cuid, INSN_UID
|
||
(v_node->insn));
|
||
}
|
||
|
||
if (TEST_BIT (sched_nodes, v_node->cuid))
|
||
{
|
||
int s_st = SCHED_TIME (v_node);
|
||
|
||
late_start = MIN (late_start,
|
||
s_st - e->latency + (e->distance * ii));
|
||
|
||
if (dump_file)
|
||
fprintf (dump_file,
|
||
"succ st = %d; late_start = %d; latency = %d",
|
||
s_st, late_start, e->latency);
|
||
|
||
if (e->data_type == MEM_DEP)
|
||
end = MAX (end, SCHED_TIME (v_node) - ii + 1);
|
||
if (dump_file)
|
||
fprintf (dump_file, "end = %d\n", end);
|
||
|
||
}
|
||
else if (dump_file)
|
||
fprintf (dump_file, "the node is not scheduled\n");
|
||
|
||
}
|
||
start = late_start;
|
||
end = MAX (end, late_start - ii);
|
||
/* Schedule the node close to it's successors. */
|
||
step = -1;
|
||
|
||
if (dump_file)
|
||
fprintf (dump_file,
|
||
"\nScheduling %d (%d) in a window (%d..%d) with step %d\n",
|
||
u_node->cuid, INSN_UID (u_node->insn), start, end, step);
|
||
|
||
}
|
||
|
||
else if (psp_not_empty && pss_not_empty)
|
||
{
|
||
int early_start = INT_MIN;
|
||
int late_start = INT_MAX;
|
||
int count_preds = 0;
|
||
int count_succs = 0;
|
||
|
||
start = INT_MIN;
|
||
end = INT_MAX;
|
||
for (e = u_node->in; e != 0; e = e->next_in)
|
||
{
|
||
ddg_node_ptr v_node = e->src;
|
||
|
||
if (dump_file)
|
||
{
|
||
fprintf (dump_file, "\nProcessing edge:");
|
||
print_ddg_edge (dump_file, e);
|
||
fprintf (dump_file,
|
||
"\nScheduling %d (%d) in psp_pss_not_empty,"
|
||
" checking p %d (%d): ", u_node->cuid, INSN_UID
|
||
(u_node->insn), v_node->cuid, INSN_UID
|
||
(v_node->insn));
|
||
}
|
||
|
||
if (TEST_BIT (sched_nodes, v_node->cuid))
|
||
{
|
||
int p_st = SCHED_TIME (v_node);
|
||
|
||
early_start = MAX (early_start,
|
||
p_st + e->latency
|
||
- (e->distance * ii));
|
||
|
||
if (dump_file)
|
||
fprintf (dump_file,
|
||
"pred st = %d; early_start = %d; latency = %d",
|
||
p_st, early_start, e->latency);
|
||
|
||
if (e->type == TRUE_DEP && e->data_type == REG_DEP)
|
||
count_preds++;
|
||
|
||
if (e->data_type == MEM_DEP)
|
||
end = MIN (end, SCHED_TIME (v_node) + ii - 1);
|
||
}
|
||
else if (dump_file)
|
||
fprintf (dump_file, "the node is not scheduled\n");
|
||
|
||
}
|
||
for (e = u_node->out; e != 0; e = e->next_out)
|
||
{
|
||
ddg_node_ptr v_node = e->dest;
|
||
|
||
if (dump_file)
|
||
{
|
||
fprintf (dump_file, "\nProcessing edge:");
|
||
print_ddg_edge (dump_file, e);
|
||
fprintf (dump_file,
|
||
"\nScheduling %d (%d) in psp_pss_not_empty,"
|
||
" checking s %d (%d): ", u_node->cuid, INSN_UID
|
||
(u_node->insn), v_node->cuid, INSN_UID
|
||
(v_node->insn));
|
||
}
|
||
|
||
if (TEST_BIT (sched_nodes, v_node->cuid))
|
||
{
|
||
int s_st = SCHED_TIME (v_node);
|
||
|
||
late_start = MIN (late_start,
|
||
s_st - e->latency
|
||
+ (e->distance * ii));
|
||
|
||
if (dump_file)
|
||
fprintf (dump_file,
|
||
"succ st = %d; late_start = %d; latency = %d",
|
||
s_st, late_start, e->latency);
|
||
|
||
if (e->type == TRUE_DEP && e->data_type == REG_DEP)
|
||
count_succs++;
|
||
|
||
if (e->data_type == MEM_DEP)
|
||
start = MAX (start, SCHED_TIME (v_node) - ii + 1);
|
||
}
|
||
else if (dump_file)
|
||
fprintf (dump_file, "the node is not scheduled\n");
|
||
|
||
}
|
||
start = MAX (start, early_start);
|
||
end = MIN (end, MIN (early_start + ii, late_start + 1));
|
||
step = 1;
|
||
/* If there are more successors than predecessors schedule the
|
||
node close to it's successors. */
|
||
if (count_succs >= count_preds)
|
||
{
|
||
int old_start = start;
|
||
|
||
start = end - 1;
|
||
end = old_start - 1;
|
||
step = -1;
|
||
}
|
||
}
|
||
else /* psp is empty && pss is empty. */
|
||
{
|
||
start = SCHED_ASAP (u_node);
|
||
end = start + ii;
|
||
step = 1;
|
||
}
|
||
|
||
*start_p = start;
|
||
*step_p = step;
|
||
*end_p = end;
|
||
sbitmap_free (psp);
|
||
sbitmap_free (pss);
|
||
|
||
if ((start >= end && step == 1) || (start <= end && step == -1))
|
||
{
|
||
if (dump_file)
|
||
fprintf (dump_file, "\nEmpty window: start=%d, end=%d, step=%d\n",
|
||
start, end, step);
|
||
return -1;
|
||
}
|
||
|
||
return 0;
|
||
}
|
||
|
||
/* Calculate MUST_PRECEDE/MUST_FOLLOW bitmaps of U_NODE; which is the
|
||
node currently been scheduled. At the end of the calculation
|
||
MUST_PRECEDE/MUST_FOLLOW contains all predecessors/successors of
|
||
U_NODE which are (1) already scheduled in the first/last row of
|
||
U_NODE's scheduling window, (2) whose dependence inequality with U
|
||
becomes an equality when U is scheduled in this same row, and (3)
|
||
whose dependence latency is zero.
|
||
|
||
The first and last rows are calculated using the following parameters:
|
||
START/END rows - The cycles that begins/ends the traversal on the window;
|
||
searching for an empty cycle to schedule U_NODE.
|
||
STEP - The direction in which we traverse the window.
|
||
II - The initiation interval. */
|
||
|
||
static void
|
||
calculate_must_precede_follow (ddg_node_ptr u_node, int start, int end,
|
||
int step, int ii, sbitmap sched_nodes,
|
||
sbitmap must_precede, sbitmap must_follow)
|
||
{
|
||
ddg_edge_ptr e;
|
||
int first_cycle_in_window, last_cycle_in_window;
|
||
|
||
gcc_assert (must_precede && must_follow);
|
||
|
||
/* Consider the following scheduling window:
|
||
{first_cycle_in_window, first_cycle_in_window+1, ...,
|
||
last_cycle_in_window}. If step is 1 then the following will be
|
||
the order we traverse the window: {start=first_cycle_in_window,
|
||
first_cycle_in_window+1, ..., end=last_cycle_in_window+1},
|
||
or {start=last_cycle_in_window, last_cycle_in_window-1, ...,
|
||
end=first_cycle_in_window-1} if step is -1. */
|
||
first_cycle_in_window = (step == 1) ? start : end - step;
|
||
last_cycle_in_window = (step == 1) ? end - step : start;
|
||
|
||
sbitmap_zero (must_precede);
|
||
sbitmap_zero (must_follow);
|
||
|
||
if (dump_file)
|
||
fprintf (dump_file, "\nmust_precede: ");
|
||
|
||
/* Instead of checking if:
|
||
(SMODULO (SCHED_TIME (e->src), ii) == first_row_in_window)
|
||
&& ((SCHED_TIME (e->src) + e->latency - (e->distance * ii)) ==
|
||
first_cycle_in_window)
|
||
&& e->latency == 0
|
||
we use the fact that latency is non-negative:
|
||
SCHED_TIME (e->src) - (e->distance * ii) <=
|
||
SCHED_TIME (e->src) + e->latency - (e->distance * ii)) <=
|
||
first_cycle_in_window
|
||
and check only if
|
||
SCHED_TIME (e->src) - (e->distance * ii) == first_cycle_in_window */
|
||
for (e = u_node->in; e != 0; e = e->next_in)
|
||
if (TEST_BIT (sched_nodes, e->src->cuid)
|
||
&& ((SCHED_TIME (e->src) - (e->distance * ii)) ==
|
||
first_cycle_in_window))
|
||
{
|
||
if (dump_file)
|
||
fprintf (dump_file, "%d ", e->src->cuid);
|
||
|
||
SET_BIT (must_precede, e->src->cuid);
|
||
}
|
||
|
||
if (dump_file)
|
||
fprintf (dump_file, "\nmust_follow: ");
|
||
|
||
/* Instead of checking if:
|
||
(SMODULO (SCHED_TIME (e->dest), ii) == last_row_in_window)
|
||
&& ((SCHED_TIME (e->dest) - e->latency + (e->distance * ii)) ==
|
||
last_cycle_in_window)
|
||
&& e->latency == 0
|
||
we use the fact that latency is non-negative:
|
||
SCHED_TIME (e->dest) + (e->distance * ii) >=
|
||
SCHED_TIME (e->dest) - e->latency + (e->distance * ii)) >=
|
||
last_cycle_in_window
|
||
and check only if
|
||
SCHED_TIME (e->dest) + (e->distance * ii) == last_cycle_in_window */
|
||
for (e = u_node->out; e != 0; e = e->next_out)
|
||
if (TEST_BIT (sched_nodes, e->dest->cuid)
|
||
&& ((SCHED_TIME (e->dest) + (e->distance * ii)) ==
|
||
last_cycle_in_window))
|
||
{
|
||
if (dump_file)
|
||
fprintf (dump_file, "%d ", e->dest->cuid);
|
||
|
||
SET_BIT (must_follow, e->dest->cuid);
|
||
}
|
||
|
||
if (dump_file)
|
||
fprintf (dump_file, "\n");
|
||
}
|
||
|
||
/* Return 1 if U_NODE can be scheduled in CYCLE. Use the following
|
||
parameters to decide if that's possible:
|
||
PS - The partial schedule.
|
||
U - The serial number of U_NODE.
|
||
NUM_SPLITS - The number of row splits made so far.
|
||
MUST_PRECEDE - The nodes that must precede U_NODE. (only valid at
|
||
the first row of the scheduling window)
|
||
MUST_FOLLOW - The nodes that must follow U_NODE. (only valid at the
|
||
last row of the scheduling window) */
|
||
|
||
static bool
|
||
try_scheduling_node_in_cycle (partial_schedule_ptr ps, ddg_node_ptr u_node,
|
||
int u, int cycle, sbitmap sched_nodes,
|
||
int *num_splits, sbitmap must_precede,
|
||
sbitmap must_follow)
|
||
{
|
||
ps_insn_ptr psi;
|
||
bool success = 0;
|
||
|
||
verify_partial_schedule (ps, sched_nodes);
|
||
psi = ps_add_node_check_conflicts (ps, u_node, cycle,
|
||
must_precede, must_follow);
|
||
if (psi)
|
||
{
|
||
SCHED_TIME (u_node) = cycle;
|
||
SET_BIT (sched_nodes, u);
|
||
success = 1;
|
||
*num_splits = 0;
|
||
if (dump_file)
|
||
fprintf (dump_file, "Scheduled w/o split in %d\n", cycle);
|
||
|
||
}
|
||
|
||
return success;
|
||
}
|
||
|
||
/* This function implements the scheduling algorithm for SMS according to the
|
||
above algorithm. */
|
||
static partial_schedule_ptr
|
||
sms_schedule_by_order (ddg_ptr g, int mii, int maxii, int *nodes_order)
|
||
{
|
||
int ii = mii;
|
||
int i, c, success, num_splits = 0;
|
||
int flush_and_start_over = true;
|
||
int num_nodes = g->num_nodes;
|
||
int start, end, step; /* Place together into one struct? */
|
||
sbitmap sched_nodes = sbitmap_alloc (num_nodes);
|
||
sbitmap must_precede = sbitmap_alloc (num_nodes);
|
||
sbitmap must_follow = sbitmap_alloc (num_nodes);
|
||
sbitmap tobe_scheduled = sbitmap_alloc (num_nodes);
|
||
|
||
partial_schedule_ptr ps = create_partial_schedule (ii, g, DFA_HISTORY);
|
||
|
||
sbitmap_ones (tobe_scheduled);
|
||
sbitmap_zero (sched_nodes);
|
||
|
||
while (flush_and_start_over && (ii < maxii))
|
||
{
|
||
|
||
if (dump_file)
|
||
fprintf (dump_file, "Starting with ii=%d\n", ii);
|
||
flush_and_start_over = false;
|
||
sbitmap_zero (sched_nodes);
|
||
|
||
for (i = 0; i < num_nodes; i++)
|
||
{
|
||
int u = nodes_order[i];
|
||
ddg_node_ptr u_node = &ps->g->nodes[u];
|
||
rtx insn = u_node->insn;
|
||
|
||
if (!NONDEBUG_INSN_P (insn))
|
||
{
|
||
RESET_BIT (tobe_scheduled, u);
|
||
continue;
|
||
}
|
||
|
||
if (JUMP_P (insn)) /* Closing branch handled later. */
|
||
{
|
||
RESET_BIT (tobe_scheduled, u);
|
||
continue;
|
||
}
|
||
|
||
if (TEST_BIT (sched_nodes, u))
|
||
continue;
|
||
|
||
/* Try to get non-empty scheduling window. */
|
||
success = 0;
|
||
if (get_sched_window (ps, nodes_order, i, sched_nodes, ii, &start,
|
||
&step, &end) == 0)
|
||
{
|
||
if (dump_file)
|
||
fprintf (dump_file, "\nTrying to schedule node %d \
|
||
INSN = %d in (%d .. %d) step %d\n", u, (INSN_UID
|
||
(g->nodes[u].insn)), start, end, step);
|
||
|
||
gcc_assert ((step > 0 && start < end)
|
||
|| (step < 0 && start > end));
|
||
|
||
calculate_must_precede_follow (u_node, start, end, step, ii,
|
||
sched_nodes, must_precede,
|
||
must_follow);
|
||
|
||
for (c = start; c != end; c += step)
|
||
{
|
||
sbitmap tmp_precede = NULL;
|
||
sbitmap tmp_follow = NULL;
|
||
|
||
if (c == start)
|
||
{
|
||
if (step == 1)
|
||
tmp_precede = must_precede;
|
||
else /* step == -1. */
|
||
tmp_follow = must_follow;
|
||
}
|
||
if (c == end - step)
|
||
{
|
||
if (step == 1)
|
||
tmp_follow = must_follow;
|
||
else /* step == -1. */
|
||
tmp_precede = must_precede;
|
||
}
|
||
|
||
success =
|
||
try_scheduling_node_in_cycle (ps, u_node, u, c,
|
||
sched_nodes,
|
||
&num_splits, tmp_precede,
|
||
tmp_follow);
|
||
if (success)
|
||
break;
|
||
}
|
||
|
||
verify_partial_schedule (ps, sched_nodes);
|
||
}
|
||
if (!success)
|
||
{
|
||
int split_row;
|
||
|
||
if (ii++ == maxii)
|
||
break;
|
||
|
||
if (num_splits >= MAX_SPLIT_NUM)
|
||
{
|
||
num_splits = 0;
|
||
flush_and_start_over = true;
|
||
verify_partial_schedule (ps, sched_nodes);
|
||
reset_partial_schedule (ps, ii);
|
||
verify_partial_schedule (ps, sched_nodes);
|
||
break;
|
||
}
|
||
|
||
num_splits++;
|
||
/* The scheduling window is exclusive of 'end'
|
||
whereas compute_split_window() expects an inclusive,
|
||
ordered range. */
|
||
if (step == 1)
|
||
split_row = compute_split_row (sched_nodes, start, end - 1,
|
||
ps->ii, u_node);
|
||
else
|
||
split_row = compute_split_row (sched_nodes, end + 1, start,
|
||
ps->ii, u_node);
|
||
|
||
ps_insert_empty_row (ps, split_row, sched_nodes);
|
||
i--; /* Go back and retry node i. */
|
||
|
||
if (dump_file)
|
||
fprintf (dump_file, "num_splits=%d\n", num_splits);
|
||
}
|
||
|
||
/* ??? If (success), check register pressure estimates. */
|
||
} /* Continue with next node. */
|
||
} /* While flush_and_start_over. */
|
||
if (ii >= maxii)
|
||
{
|
||
free_partial_schedule (ps);
|
||
ps = NULL;
|
||
}
|
||
else
|
||
gcc_assert (sbitmap_equal (tobe_scheduled, sched_nodes));
|
||
|
||
sbitmap_free (sched_nodes);
|
||
sbitmap_free (must_precede);
|
||
sbitmap_free (must_follow);
|
||
sbitmap_free (tobe_scheduled);
|
||
|
||
return ps;
|
||
}
|
||
|
||
/* This function inserts a new empty row into PS at the position
|
||
according to SPLITROW, keeping all already scheduled instructions
|
||
intact and updating their SCHED_TIME and cycle accordingly. */
|
||
static void
|
||
ps_insert_empty_row (partial_schedule_ptr ps, int split_row,
|
||
sbitmap sched_nodes)
|
||
{
|
||
ps_insn_ptr crr_insn;
|
||
ps_insn_ptr *rows_new;
|
||
int ii = ps->ii;
|
||
int new_ii = ii + 1;
|
||
int row;
|
||
|
||
verify_partial_schedule (ps, sched_nodes);
|
||
|
||
/* We normalize sched_time and rotate ps to have only non-negative sched
|
||
times, for simplicity of updating cycles after inserting new row. */
|
||
split_row -= ps->min_cycle;
|
||
split_row = SMODULO (split_row, ii);
|
||
if (dump_file)
|
||
fprintf (dump_file, "split_row=%d\n", split_row);
|
||
|
||
normalize_sched_times (ps);
|
||
rotate_partial_schedule (ps, ps->min_cycle);
|
||
|
||
rows_new = (ps_insn_ptr *) xcalloc (new_ii, sizeof (ps_insn_ptr));
|
||
for (row = 0; row < split_row; row++)
|
||
{
|
||
rows_new[row] = ps->rows[row];
|
||
ps->rows[row] = NULL;
|
||
for (crr_insn = rows_new[row];
|
||
crr_insn; crr_insn = crr_insn->next_in_row)
|
||
{
|
||
ddg_node_ptr u = crr_insn->node;
|
||
int new_time = SCHED_TIME (u) + (SCHED_TIME (u) / ii);
|
||
|
||
SCHED_TIME (u) = new_time;
|
||
crr_insn->cycle = new_time;
|
||
SCHED_ROW (u) = new_time % new_ii;
|
||
SCHED_STAGE (u) = new_time / new_ii;
|
||
}
|
||
|
||
}
|
||
|
||
rows_new[split_row] = NULL;
|
||
|
||
for (row = split_row; row < ii; row++)
|
||
{
|
||
rows_new[row + 1] = ps->rows[row];
|
||
ps->rows[row] = NULL;
|
||
for (crr_insn = rows_new[row + 1];
|
||
crr_insn; crr_insn = crr_insn->next_in_row)
|
||
{
|
||
ddg_node_ptr u = crr_insn->node;
|
||
int new_time = SCHED_TIME (u) + (SCHED_TIME (u) / ii) + 1;
|
||
|
||
SCHED_TIME (u) = new_time;
|
||
crr_insn->cycle = new_time;
|
||
SCHED_ROW (u) = new_time % new_ii;
|
||
SCHED_STAGE (u) = new_time / new_ii;
|
||
}
|
||
}
|
||
|
||
/* Updating ps. */
|
||
ps->min_cycle = ps->min_cycle + ps->min_cycle / ii
|
||
+ (SMODULO (ps->min_cycle, ii) >= split_row ? 1 : 0);
|
||
ps->max_cycle = ps->max_cycle + ps->max_cycle / ii
|
||
+ (SMODULO (ps->max_cycle, ii) >= split_row ? 1 : 0);
|
||
free (ps->rows);
|
||
ps->rows = rows_new;
|
||
ps->ii = new_ii;
|
||
gcc_assert (ps->min_cycle >= 0);
|
||
|
||
verify_partial_schedule (ps, sched_nodes);
|
||
|
||
if (dump_file)
|
||
fprintf (dump_file, "min_cycle=%d, max_cycle=%d\n", ps->min_cycle,
|
||
ps->max_cycle);
|
||
}
|
||
|
||
/* Given U_NODE which is the node that failed to be scheduled; LOW and
|
||
UP which are the boundaries of it's scheduling window; compute using
|
||
SCHED_NODES and II a row in the partial schedule that can be split
|
||
which will separate a critical predecessor from a critical successor
|
||
thereby expanding the window, and return it. */
|
||
static int
|
||
compute_split_row (sbitmap sched_nodes, int low, int up, int ii,
|
||
ddg_node_ptr u_node)
|
||
{
|
||
ddg_edge_ptr e;
|
||
int lower = INT_MIN, upper = INT_MAX;
|
||
ddg_node_ptr crit_pred = NULL;
|
||
ddg_node_ptr crit_succ = NULL;
|
||
int crit_cycle;
|
||
|
||
for (e = u_node->in; e != 0; e = e->next_in)
|
||
{
|
||
ddg_node_ptr v_node = e->src;
|
||
|
||
if (TEST_BIT (sched_nodes, v_node->cuid)
|
||
&& (low == SCHED_TIME (v_node) + e->latency - (e->distance * ii)))
|
||
if (SCHED_TIME (v_node) > lower)
|
||
{
|
||
crit_pred = v_node;
|
||
lower = SCHED_TIME (v_node);
|
||
}
|
||
}
|
||
|
||
if (crit_pred != NULL)
|
||
{
|
||
crit_cycle = SCHED_TIME (crit_pred) + 1;
|
||
return SMODULO (crit_cycle, ii);
|
||
}
|
||
|
||
for (e = u_node->out; e != 0; e = e->next_out)
|
||
{
|
||
ddg_node_ptr v_node = e->dest;
|
||
if (TEST_BIT (sched_nodes, v_node->cuid)
|
||
&& (up == SCHED_TIME (v_node) - e->latency + (e->distance * ii)))
|
||
if (SCHED_TIME (v_node) < upper)
|
||
{
|
||
crit_succ = v_node;
|
||
upper = SCHED_TIME (v_node);
|
||
}
|
||
}
|
||
|
||
if (crit_succ != NULL)
|
||
{
|
||
crit_cycle = SCHED_TIME (crit_succ);
|
||
return SMODULO (crit_cycle, ii);
|
||
}
|
||
|
||
if (dump_file)
|
||
fprintf (dump_file, "Both crit_pred and crit_succ are NULL\n");
|
||
|
||
return SMODULO ((low + up + 1) / 2, ii);
|
||
}
|
||
|
||
static void
|
||
verify_partial_schedule (partial_schedule_ptr ps, sbitmap sched_nodes)
|
||
{
|
||
int row;
|
||
ps_insn_ptr crr_insn;
|
||
|
||
for (row = 0; row < ps->ii; row++)
|
||
for (crr_insn = ps->rows[row]; crr_insn; crr_insn = crr_insn->next_in_row)
|
||
{
|
||
ddg_node_ptr u = crr_insn->node;
|
||
|
||
gcc_assert (TEST_BIT (sched_nodes, u->cuid));
|
||
/* ??? Test also that all nodes of sched_nodes are in ps, perhaps by
|
||
popcount (sched_nodes) == number of insns in ps. */
|
||
gcc_assert (SCHED_TIME (u) >= ps->min_cycle);
|
||
gcc_assert (SCHED_TIME (u) <= ps->max_cycle);
|
||
}
|
||
}
|
||
|
||
|
||
/* This page implements the algorithm for ordering the nodes of a DDG
|
||
for modulo scheduling, activated through the
|
||
"int sms_order_nodes (ddg_ptr, int mii, int * result)" API. */
|
||
|
||
#define ORDER_PARAMS(x) ((struct node_order_params *) (x)->aux.info)
|
||
#define ASAP(x) (ORDER_PARAMS ((x))->asap)
|
||
#define ALAP(x) (ORDER_PARAMS ((x))->alap)
|
||
#define HEIGHT(x) (ORDER_PARAMS ((x))->height)
|
||
#define MOB(x) (ALAP ((x)) - ASAP ((x)))
|
||
#define DEPTH(x) (ASAP ((x)))
|
||
|
||
typedef struct node_order_params * nopa;
|
||
|
||
static void order_nodes_of_sccs (ddg_all_sccs_ptr, int * result);
|
||
static int order_nodes_in_scc (ddg_ptr, sbitmap, sbitmap, int*, int);
|
||
static nopa calculate_order_params (ddg_ptr, int, int *);
|
||
static int find_max_asap (ddg_ptr, sbitmap);
|
||
static int find_max_hv_min_mob (ddg_ptr, sbitmap);
|
||
static int find_max_dv_min_mob (ddg_ptr, sbitmap);
|
||
|
||
enum sms_direction {BOTTOMUP, TOPDOWN};
|
||
|
||
struct node_order_params
|
||
{
|
||
int asap;
|
||
int alap;
|
||
int height;
|
||
};
|
||
|
||
/* Check if NODE_ORDER contains a permutation of 0 .. NUM_NODES-1. */
|
||
static void
|
||
check_nodes_order (int *node_order, int num_nodes)
|
||
{
|
||
int i;
|
||
sbitmap tmp = sbitmap_alloc (num_nodes);
|
||
|
||
sbitmap_zero (tmp);
|
||
|
||
if (dump_file)
|
||
fprintf (dump_file, "SMS final nodes order: \n");
|
||
|
||
for (i = 0; i < num_nodes; i++)
|
||
{
|
||
int u = node_order[i];
|
||
|
||
if (dump_file)
|
||
fprintf (dump_file, "%d ", u);
|
||
gcc_assert (u < num_nodes && u >= 0 && !TEST_BIT (tmp, u));
|
||
|
||
SET_BIT (tmp, u);
|
||
}
|
||
|
||
if (dump_file)
|
||
fprintf (dump_file, "\n");
|
||
|
||
sbitmap_free (tmp);
|
||
}
|
||
|
||
/* Order the nodes of G for scheduling and pass the result in
|
||
NODE_ORDER. Also set aux.count of each node to ASAP.
|
||
Put maximal ASAP to PMAX_ASAP. Return the recMII for the given DDG. */
|
||
static int
|
||
sms_order_nodes (ddg_ptr g, int mii, int * node_order, int *pmax_asap)
|
||
{
|
||
int i;
|
||
int rec_mii = 0;
|
||
ddg_all_sccs_ptr sccs = create_ddg_all_sccs (g);
|
||
|
||
nopa nops = calculate_order_params (g, mii, pmax_asap);
|
||
|
||
if (dump_file)
|
||
print_sccs (dump_file, sccs, g);
|
||
|
||
order_nodes_of_sccs (sccs, node_order);
|
||
|
||
if (sccs->num_sccs > 0)
|
||
/* First SCC has the largest recurrence_length. */
|
||
rec_mii = sccs->sccs[0]->recurrence_length;
|
||
|
||
/* Save ASAP before destroying node_order_params. */
|
||
for (i = 0; i < g->num_nodes; i++)
|
||
{
|
||
ddg_node_ptr v = &g->nodes[i];
|
||
v->aux.count = ASAP (v);
|
||
}
|
||
|
||
free (nops);
|
||
free_ddg_all_sccs (sccs);
|
||
check_nodes_order (node_order, g->num_nodes);
|
||
|
||
return rec_mii;
|
||
}
|
||
|
||
static void
|
||
order_nodes_of_sccs (ddg_all_sccs_ptr all_sccs, int * node_order)
|
||
{
|
||
int i, pos = 0;
|
||
ddg_ptr g = all_sccs->ddg;
|
||
int num_nodes = g->num_nodes;
|
||
sbitmap prev_sccs = sbitmap_alloc (num_nodes);
|
||
sbitmap on_path = sbitmap_alloc (num_nodes);
|
||
sbitmap tmp = sbitmap_alloc (num_nodes);
|
||
sbitmap ones = sbitmap_alloc (num_nodes);
|
||
|
||
sbitmap_zero (prev_sccs);
|
||
sbitmap_ones (ones);
|
||
|
||
/* Perform the node ordering starting from the SCC with the highest recMII.
|
||
For each SCC order the nodes according to their ASAP/ALAP/HEIGHT etc. */
|
||
for (i = 0; i < all_sccs->num_sccs; i++)
|
||
{
|
||
ddg_scc_ptr scc = all_sccs->sccs[i];
|
||
|
||
/* Add nodes on paths from previous SCCs to the current SCC. */
|
||
find_nodes_on_paths (on_path, g, prev_sccs, scc->nodes);
|
||
sbitmap_a_or_b (tmp, scc->nodes, on_path);
|
||
|
||
/* Add nodes on paths from the current SCC to previous SCCs. */
|
||
find_nodes_on_paths (on_path, g, scc->nodes, prev_sccs);
|
||
sbitmap_a_or_b (tmp, tmp, on_path);
|
||
|
||
/* Remove nodes of previous SCCs from current extended SCC. */
|
||
sbitmap_difference (tmp, tmp, prev_sccs);
|
||
|
||
pos = order_nodes_in_scc (g, prev_sccs, tmp, node_order, pos);
|
||
/* Above call to order_nodes_in_scc updated prev_sccs |= tmp. */
|
||
}
|
||
|
||
/* Handle the remaining nodes that do not belong to any scc. Each call
|
||
to order_nodes_in_scc handles a single connected component. */
|
||
while (pos < g->num_nodes)
|
||
{
|
||
sbitmap_difference (tmp, ones, prev_sccs);
|
||
pos = order_nodes_in_scc (g, prev_sccs, tmp, node_order, pos);
|
||
}
|
||
sbitmap_free (prev_sccs);
|
||
sbitmap_free (on_path);
|
||
sbitmap_free (tmp);
|
||
sbitmap_free (ones);
|
||
}
|
||
|
||
/* MII is needed if we consider backarcs (that do not close recursive cycles). */
|
||
static struct node_order_params *
|
||
calculate_order_params (ddg_ptr g, int mii ATTRIBUTE_UNUSED, int *pmax_asap)
|
||
{
|
||
int u;
|
||
int max_asap;
|
||
int num_nodes = g->num_nodes;
|
||
ddg_edge_ptr e;
|
||
/* Allocate a place to hold ordering params for each node in the DDG. */
|
||
nopa node_order_params_arr;
|
||
|
||
/* Initialize of ASAP/ALAP/HEIGHT to zero. */
|
||
node_order_params_arr = (nopa) xcalloc (num_nodes,
|
||
sizeof (struct node_order_params));
|
||
|
||
/* Set the aux pointer of each node to point to its order_params structure. */
|
||
for (u = 0; u < num_nodes; u++)
|
||
g->nodes[u].aux.info = &node_order_params_arr[u];
|
||
|
||
/* Disregarding a backarc from each recursive cycle to obtain a DAG,
|
||
calculate ASAP, ALAP, mobility, distance, and height for each node
|
||
in the dependence (direct acyclic) graph. */
|
||
|
||
/* We assume that the nodes in the array are in topological order. */
|
||
|
||
max_asap = 0;
|
||
for (u = 0; u < num_nodes; u++)
|
||
{
|
||
ddg_node_ptr u_node = &g->nodes[u];
|
||
|
||
ASAP (u_node) = 0;
|
||
for (e = u_node->in; e; e = e->next_in)
|
||
if (e->distance == 0)
|
||
ASAP (u_node) = MAX (ASAP (u_node),
|
||
ASAP (e->src) + e->latency);
|
||
max_asap = MAX (max_asap, ASAP (u_node));
|
||
}
|
||
|
||
for (u = num_nodes - 1; u > -1; u--)
|
||
{
|
||
ddg_node_ptr u_node = &g->nodes[u];
|
||
|
||
ALAP (u_node) = max_asap;
|
||
HEIGHT (u_node) = 0;
|
||
for (e = u_node->out; e; e = e->next_out)
|
||
if (e->distance == 0)
|
||
{
|
||
ALAP (u_node) = MIN (ALAP (u_node),
|
||
ALAP (e->dest) - e->latency);
|
||
HEIGHT (u_node) = MAX (HEIGHT (u_node),
|
||
HEIGHT (e->dest) + e->latency);
|
||
}
|
||
}
|
||
if (dump_file)
|
||
{
|
||
fprintf (dump_file, "\nOrder params\n");
|
||
for (u = 0; u < num_nodes; u++)
|
||
{
|
||
ddg_node_ptr u_node = &g->nodes[u];
|
||
|
||
fprintf (dump_file, "node %d, ASAP: %d, ALAP: %d, HEIGHT: %d\n", u,
|
||
ASAP (u_node), ALAP (u_node), HEIGHT (u_node));
|
||
}
|
||
}
|
||
|
||
*pmax_asap = max_asap;
|
||
return node_order_params_arr;
|
||
}
|
||
|
||
static int
|
||
find_max_asap (ddg_ptr g, sbitmap nodes)
|
||
{
|
||
unsigned int u = 0;
|
||
int max_asap = -1;
|
||
int result = -1;
|
||
sbitmap_iterator sbi;
|
||
|
||
EXECUTE_IF_SET_IN_SBITMAP (nodes, 0, u, sbi)
|
||
{
|
||
ddg_node_ptr u_node = &g->nodes[u];
|
||
|
||
if (max_asap < ASAP (u_node))
|
||
{
|
||
max_asap = ASAP (u_node);
|
||
result = u;
|
||
}
|
||
}
|
||
return result;
|
||
}
|
||
|
||
static int
|
||
find_max_hv_min_mob (ddg_ptr g, sbitmap nodes)
|
||
{
|
||
unsigned int u = 0;
|
||
int max_hv = -1;
|
||
int min_mob = INT_MAX;
|
||
int result = -1;
|
||
sbitmap_iterator sbi;
|
||
|
||
EXECUTE_IF_SET_IN_SBITMAP (nodes, 0, u, sbi)
|
||
{
|
||
ddg_node_ptr u_node = &g->nodes[u];
|
||
|
||
if (max_hv < HEIGHT (u_node))
|
||
{
|
||
max_hv = HEIGHT (u_node);
|
||
min_mob = MOB (u_node);
|
||
result = u;
|
||
}
|
||
else if ((max_hv == HEIGHT (u_node))
|
||
&& (min_mob > MOB (u_node)))
|
||
{
|
||
min_mob = MOB (u_node);
|
||
result = u;
|
||
}
|
||
}
|
||
return result;
|
||
}
|
||
|
||
static int
|
||
find_max_dv_min_mob (ddg_ptr g, sbitmap nodes)
|
||
{
|
||
unsigned int u = 0;
|
||
int max_dv = -1;
|
||
int min_mob = INT_MAX;
|
||
int result = -1;
|
||
sbitmap_iterator sbi;
|
||
|
||
EXECUTE_IF_SET_IN_SBITMAP (nodes, 0, u, sbi)
|
||
{
|
||
ddg_node_ptr u_node = &g->nodes[u];
|
||
|
||
if (max_dv < DEPTH (u_node))
|
||
{
|
||
max_dv = DEPTH (u_node);
|
||
min_mob = MOB (u_node);
|
||
result = u;
|
||
}
|
||
else if ((max_dv == DEPTH (u_node))
|
||
&& (min_mob > MOB (u_node)))
|
||
{
|
||
min_mob = MOB (u_node);
|
||
result = u;
|
||
}
|
||
}
|
||
return result;
|
||
}
|
||
|
||
/* Places the nodes of SCC into the NODE_ORDER array starting
|
||
at position POS, according to the SMS ordering algorithm.
|
||
NODES_ORDERED (in&out parameter) holds the bitset of all nodes in
|
||
the NODE_ORDER array, starting from position zero. */
|
||
static int
|
||
order_nodes_in_scc (ddg_ptr g, sbitmap nodes_ordered, sbitmap scc,
|
||
int * node_order, int pos)
|
||
{
|
||
enum sms_direction dir;
|
||
int num_nodes = g->num_nodes;
|
||
sbitmap workset = sbitmap_alloc (num_nodes);
|
||
sbitmap tmp = sbitmap_alloc (num_nodes);
|
||
sbitmap zero_bitmap = sbitmap_alloc (num_nodes);
|
||
sbitmap predecessors = sbitmap_alloc (num_nodes);
|
||
sbitmap successors = sbitmap_alloc (num_nodes);
|
||
|
||
sbitmap_zero (predecessors);
|
||
find_predecessors (predecessors, g, nodes_ordered);
|
||
|
||
sbitmap_zero (successors);
|
||
find_successors (successors, g, nodes_ordered);
|
||
|
||
sbitmap_zero (tmp);
|
||
if (sbitmap_a_and_b_cg (tmp, predecessors, scc))
|
||
{
|
||
sbitmap_copy (workset, tmp);
|
||
dir = BOTTOMUP;
|
||
}
|
||
else if (sbitmap_a_and_b_cg (tmp, successors, scc))
|
||
{
|
||
sbitmap_copy (workset, tmp);
|
||
dir = TOPDOWN;
|
||
}
|
||
else
|
||
{
|
||
int u;
|
||
|
||
sbitmap_zero (workset);
|
||
if ((u = find_max_asap (g, scc)) >= 0)
|
||
SET_BIT (workset, u);
|
||
dir = BOTTOMUP;
|
||
}
|
||
|
||
sbitmap_zero (zero_bitmap);
|
||
while (!sbitmap_equal (workset, zero_bitmap))
|
||
{
|
||
int v;
|
||
ddg_node_ptr v_node;
|
||
sbitmap v_node_preds;
|
||
sbitmap v_node_succs;
|
||
|
||
if (dir == TOPDOWN)
|
||
{
|
||
while (!sbitmap_equal (workset, zero_bitmap))
|
||
{
|
||
v = find_max_hv_min_mob (g, workset);
|
||
v_node = &g->nodes[v];
|
||
node_order[pos++] = v;
|
||
v_node_succs = NODE_SUCCESSORS (v_node);
|
||
sbitmap_a_and_b (tmp, v_node_succs, scc);
|
||
|
||
/* Don't consider the already ordered successors again. */
|
||
sbitmap_difference (tmp, tmp, nodes_ordered);
|
||
sbitmap_a_or_b (workset, workset, tmp);
|
||
RESET_BIT (workset, v);
|
||
SET_BIT (nodes_ordered, v);
|
||
}
|
||
dir = BOTTOMUP;
|
||
sbitmap_zero (predecessors);
|
||
find_predecessors (predecessors, g, nodes_ordered);
|
||
sbitmap_a_and_b (workset, predecessors, scc);
|
||
}
|
||
else
|
||
{
|
||
while (!sbitmap_equal (workset, zero_bitmap))
|
||
{
|
||
v = find_max_dv_min_mob (g, workset);
|
||
v_node = &g->nodes[v];
|
||
node_order[pos++] = v;
|
||
v_node_preds = NODE_PREDECESSORS (v_node);
|
||
sbitmap_a_and_b (tmp, v_node_preds, scc);
|
||
|
||
/* Don't consider the already ordered predecessors again. */
|
||
sbitmap_difference (tmp, tmp, nodes_ordered);
|
||
sbitmap_a_or_b (workset, workset, tmp);
|
||
RESET_BIT (workset, v);
|
||
SET_BIT (nodes_ordered, v);
|
||
}
|
||
dir = TOPDOWN;
|
||
sbitmap_zero (successors);
|
||
find_successors (successors, g, nodes_ordered);
|
||
sbitmap_a_and_b (workset, successors, scc);
|
||
}
|
||
}
|
||
sbitmap_free (tmp);
|
||
sbitmap_free (workset);
|
||
sbitmap_free (zero_bitmap);
|
||
sbitmap_free (predecessors);
|
||
sbitmap_free (successors);
|
||
return pos;
|
||
}
|
||
|
||
|
||
/* This page contains functions for manipulating partial-schedules during
|
||
modulo scheduling. */
|
||
|
||
/* Create a partial schedule and allocate a memory to hold II rows. */
|
||
|
||
static partial_schedule_ptr
|
||
create_partial_schedule (int ii, ddg_ptr g, int history)
|
||
{
|
||
partial_schedule_ptr ps = XNEW (struct partial_schedule);
|
||
ps->rows = (ps_insn_ptr *) xcalloc (ii, sizeof (ps_insn_ptr));
|
||
ps->ii = ii;
|
||
ps->history = history;
|
||
ps->min_cycle = INT_MAX;
|
||
ps->max_cycle = INT_MIN;
|
||
ps->g = g;
|
||
|
||
return ps;
|
||
}
|
||
|
||
/* Free the PS_INSNs in rows array of the given partial schedule.
|
||
??? Consider caching the PS_INSN's. */
|
||
static void
|
||
free_ps_insns (partial_schedule_ptr ps)
|
||
{
|
||
int i;
|
||
|
||
for (i = 0; i < ps->ii; i++)
|
||
{
|
||
while (ps->rows[i])
|
||
{
|
||
ps_insn_ptr ps_insn = ps->rows[i]->next_in_row;
|
||
|
||
free (ps->rows[i]);
|
||
ps->rows[i] = ps_insn;
|
||
}
|
||
ps->rows[i] = NULL;
|
||
}
|
||
}
|
||
|
||
/* Free all the memory allocated to the partial schedule. */
|
||
|
||
static void
|
||
free_partial_schedule (partial_schedule_ptr ps)
|
||
{
|
||
if (!ps)
|
||
return;
|
||
free_ps_insns (ps);
|
||
free (ps->rows);
|
||
free (ps);
|
||
}
|
||
|
||
/* Clear the rows array with its PS_INSNs, and create a new one with
|
||
NEW_II rows. */
|
||
|
||
static void
|
||
reset_partial_schedule (partial_schedule_ptr ps, int new_ii)
|
||
{
|
||
if (!ps)
|
||
return;
|
||
free_ps_insns (ps);
|
||
if (new_ii == ps->ii)
|
||
return;
|
||
ps->rows = (ps_insn_ptr *) xrealloc (ps->rows, new_ii
|
||
* sizeof (ps_insn_ptr));
|
||
memset (ps->rows, 0, new_ii * sizeof (ps_insn_ptr));
|
||
ps->ii = new_ii;
|
||
ps->min_cycle = INT_MAX;
|
||
ps->max_cycle = INT_MIN;
|
||
}
|
||
|
||
/* Prints the partial schedule as an ii rows array, for each rows
|
||
print the ids of the insns in it. */
|
||
void
|
||
print_partial_schedule (partial_schedule_ptr ps, FILE *dump)
|
||
{
|
||
int i;
|
||
|
||
for (i = 0; i < ps->ii; i++)
|
||
{
|
||
ps_insn_ptr ps_i = ps->rows[i];
|
||
|
||
fprintf (dump, "\n[ROW %d ]: ", i);
|
||
while (ps_i)
|
||
{
|
||
fprintf (dump, "%d, ",
|
||
INSN_UID (ps_i->node->insn));
|
||
ps_i = ps_i->next_in_row;
|
||
}
|
||
}
|
||
}
|
||
|
||
/* Creates an object of PS_INSN and initializes it to the given parameters. */
|
||
static ps_insn_ptr
|
||
create_ps_insn (ddg_node_ptr node, int rest_count, int cycle)
|
||
{
|
||
ps_insn_ptr ps_i = XNEW (struct ps_insn);
|
||
|
||
ps_i->node = node;
|
||
ps_i->next_in_row = NULL;
|
||
ps_i->prev_in_row = NULL;
|
||
ps_i->row_rest_count = rest_count;
|
||
ps_i->cycle = cycle;
|
||
|
||
return ps_i;
|
||
}
|
||
|
||
|
||
/* Removes the given PS_INSN from the partial schedule. Returns false if the
|
||
node is not found in the partial schedule, else returns true. */
|
||
static bool
|
||
remove_node_from_ps (partial_schedule_ptr ps, ps_insn_ptr ps_i)
|
||
{
|
||
int row;
|
||
|
||
if (!ps || !ps_i)
|
||
return false;
|
||
|
||
row = SMODULO (ps_i->cycle, ps->ii);
|
||
if (! ps_i->prev_in_row)
|
||
{
|
||
if (ps_i != ps->rows[row])
|
||
return false;
|
||
|
||
ps->rows[row] = ps_i->next_in_row;
|
||
if (ps->rows[row])
|
||
ps->rows[row]->prev_in_row = NULL;
|
||
}
|
||
else
|
||
{
|
||
ps_i->prev_in_row->next_in_row = ps_i->next_in_row;
|
||
if (ps_i->next_in_row)
|
||
ps_i->next_in_row->prev_in_row = ps_i->prev_in_row;
|
||
}
|
||
free (ps_i);
|
||
return true;
|
||
}
|
||
|
||
/* Unlike what literature describes for modulo scheduling (which focuses
|
||
on VLIW machines) the order of the instructions inside a cycle is
|
||
important. Given the bitmaps MUST_FOLLOW and MUST_PRECEDE we know
|
||
where the current instruction should go relative to the already
|
||
scheduled instructions in the given cycle. Go over these
|
||
instructions and find the first possible column to put it in. */
|
||
static bool
|
||
ps_insn_find_column (partial_schedule_ptr ps, ps_insn_ptr ps_i,
|
||
sbitmap must_precede, sbitmap must_follow)
|
||
{
|
||
ps_insn_ptr next_ps_i;
|
||
ps_insn_ptr first_must_follow = NULL;
|
||
ps_insn_ptr last_must_precede = NULL;
|
||
int row;
|
||
|
||
if (! ps_i)
|
||
return false;
|
||
|
||
row = SMODULO (ps_i->cycle, ps->ii);
|
||
|
||
/* Find the first must follow and the last must precede
|
||
and insert the node immediately after the must precede
|
||
but make sure that it there is no must follow after it. */
|
||
for (next_ps_i = ps->rows[row];
|
||
next_ps_i;
|
||
next_ps_i = next_ps_i->next_in_row)
|
||
{
|
||
if (must_follow && TEST_BIT (must_follow, next_ps_i->node->cuid)
|
||
&& ! first_must_follow)
|
||
first_must_follow = next_ps_i;
|
||
if (must_precede && TEST_BIT (must_precede, next_ps_i->node->cuid))
|
||
{
|
||
/* If we have already met a node that must follow, then
|
||
there is no possible column. */
|
||
if (first_must_follow)
|
||
return false;
|
||
else
|
||
last_must_precede = next_ps_i;
|
||
}
|
||
}
|
||
|
||
/* Now insert the node after INSERT_AFTER_PSI. */
|
||
|
||
if (! last_must_precede)
|
||
{
|
||
ps_i->next_in_row = ps->rows[row];
|
||
ps_i->prev_in_row = NULL;
|
||
if (ps_i->next_in_row)
|
||
ps_i->next_in_row->prev_in_row = ps_i;
|
||
ps->rows[row] = ps_i;
|
||
}
|
||
else
|
||
{
|
||
ps_i->next_in_row = last_must_precede->next_in_row;
|
||
last_must_precede->next_in_row = ps_i;
|
||
ps_i->prev_in_row = last_must_precede;
|
||
if (ps_i->next_in_row)
|
||
ps_i->next_in_row->prev_in_row = ps_i;
|
||
}
|
||
|
||
return true;
|
||
}
|
||
|
||
/* Advances the PS_INSN one column in its current row; returns false
|
||
in failure and true in success. Bit N is set in MUST_FOLLOW if
|
||
the node with cuid N must be come after the node pointed to by
|
||
PS_I when scheduled in the same cycle. */
|
||
static int
|
||
ps_insn_advance_column (partial_schedule_ptr ps, ps_insn_ptr ps_i,
|
||
sbitmap must_follow)
|
||
{
|
||
ps_insn_ptr prev, next;
|
||
int row;
|
||
ddg_node_ptr next_node;
|
||
|
||
if (!ps || !ps_i)
|
||
return false;
|
||
|
||
row = SMODULO (ps_i->cycle, ps->ii);
|
||
|
||
if (! ps_i->next_in_row)
|
||
return false;
|
||
|
||
next_node = ps_i->next_in_row->node;
|
||
|
||
/* Check if next_in_row is dependent on ps_i, both having same sched
|
||
times (typically ANTI_DEP). If so, ps_i cannot skip over it. */
|
||
if (must_follow && TEST_BIT (must_follow, next_node->cuid))
|
||
return false;
|
||
|
||
/* Advance PS_I over its next_in_row in the doubly linked list. */
|
||
prev = ps_i->prev_in_row;
|
||
next = ps_i->next_in_row;
|
||
|
||
if (ps_i == ps->rows[row])
|
||
ps->rows[row] = next;
|
||
|
||
ps_i->next_in_row = next->next_in_row;
|
||
|
||
if (next->next_in_row)
|
||
next->next_in_row->prev_in_row = ps_i;
|
||
|
||
next->next_in_row = ps_i;
|
||
ps_i->prev_in_row = next;
|
||
|
||
next->prev_in_row = prev;
|
||
if (prev)
|
||
prev->next_in_row = next;
|
||
|
||
return true;
|
||
}
|
||
|
||
/* Inserts a DDG_NODE to the given partial schedule at the given cycle.
|
||
Returns 0 if this is not possible and a PS_INSN otherwise. Bit N is
|
||
set in MUST_PRECEDE/MUST_FOLLOW if the node with cuid N must be come
|
||
before/after (respectively) the node pointed to by PS_I when scheduled
|
||
in the same cycle. */
|
||
static ps_insn_ptr
|
||
add_node_to_ps (partial_schedule_ptr ps, ddg_node_ptr node, int cycle,
|
||
sbitmap must_precede, sbitmap must_follow)
|
||
{
|
||
ps_insn_ptr ps_i;
|
||
int rest_count = 1;
|
||
int row = SMODULO (cycle, ps->ii);
|
||
|
||
if (ps->rows[row]
|
||
&& ps->rows[row]->row_rest_count >= issue_rate)
|
||
return NULL;
|
||
|
||
if (ps->rows[row])
|
||
rest_count += ps->rows[row]->row_rest_count;
|
||
|
||
ps_i = create_ps_insn (node, rest_count, cycle);
|
||
|
||
/* Finds and inserts PS_I according to MUST_FOLLOW and
|
||
MUST_PRECEDE. */
|
||
if (! ps_insn_find_column (ps, ps_i, must_precede, must_follow))
|
||
{
|
||
free (ps_i);
|
||
return NULL;
|
||
}
|
||
|
||
return ps_i;
|
||
}
|
||
|
||
/* Advance time one cycle. Assumes DFA is being used. */
|
||
static void
|
||
advance_one_cycle (void)
|
||
{
|
||
if (targetm.sched.dfa_pre_cycle_insn)
|
||
state_transition (curr_state,
|
||
targetm.sched.dfa_pre_cycle_insn ());
|
||
|
||
state_transition (curr_state, NULL);
|
||
|
||
if (targetm.sched.dfa_post_cycle_insn)
|
||
state_transition (curr_state,
|
||
targetm.sched.dfa_post_cycle_insn ());
|
||
}
|
||
|
||
|
||
|
||
/* Checks if PS has resource conflicts according to DFA, starting from
|
||
FROM cycle to TO cycle; returns true if there are conflicts and false
|
||
if there are no conflicts. Assumes DFA is being used. */
|
||
static int
|
||
ps_has_conflicts (partial_schedule_ptr ps, int from, int to)
|
||
{
|
||
int cycle;
|
||
|
||
state_reset (curr_state);
|
||
|
||
for (cycle = from; cycle <= to; cycle++)
|
||
{
|
||
ps_insn_ptr crr_insn;
|
||
/* Holds the remaining issue slots in the current row. */
|
||
int can_issue_more = issue_rate;
|
||
|
||
/* Walk through the DFA for the current row. */
|
||
for (crr_insn = ps->rows[SMODULO (cycle, ps->ii)];
|
||
crr_insn;
|
||
crr_insn = crr_insn->next_in_row)
|
||
{
|
||
rtx insn = crr_insn->node->insn;
|
||
|
||
if (!NONDEBUG_INSN_P (insn))
|
||
continue;
|
||
|
||
/* Check if there is room for the current insn. */
|
||
if (!can_issue_more || state_dead_lock_p (curr_state))
|
||
return true;
|
||
|
||
/* Update the DFA state and return with failure if the DFA found
|
||
resource conflicts. */
|
||
if (state_transition (curr_state, insn) >= 0)
|
||
return true;
|
||
|
||
if (targetm.sched.variable_issue)
|
||
can_issue_more =
|
||
targetm.sched.variable_issue (sched_dump, sched_verbose,
|
||
insn, can_issue_more);
|
||
/* A naked CLOBBER or USE generates no instruction, so don't
|
||
let them consume issue slots. */
|
||
else if (GET_CODE (PATTERN (insn)) != USE
|
||
&& GET_CODE (PATTERN (insn)) != CLOBBER)
|
||
can_issue_more--;
|
||
}
|
||
|
||
/* Advance the DFA to the next cycle. */
|
||
advance_one_cycle ();
|
||
}
|
||
return false;
|
||
}
|
||
|
||
/* Checks if the given node causes resource conflicts when added to PS at
|
||
cycle C. If not the node is added to PS and returned; otherwise zero
|
||
is returned. Bit N is set in MUST_PRECEDE/MUST_FOLLOW if the node with
|
||
cuid N must be come before/after (respectively) the node pointed to by
|
||
PS_I when scheduled in the same cycle. */
|
||
ps_insn_ptr
|
||
ps_add_node_check_conflicts (partial_schedule_ptr ps, ddg_node_ptr n,
|
||
int c, sbitmap must_precede,
|
||
sbitmap must_follow)
|
||
{
|
||
int has_conflicts = 0;
|
||
ps_insn_ptr ps_i;
|
||
|
||
/* First add the node to the PS, if this succeeds check for
|
||
conflicts, trying different issue slots in the same row. */
|
||
if (! (ps_i = add_node_to_ps (ps, n, c, must_precede, must_follow)))
|
||
return NULL; /* Failed to insert the node at the given cycle. */
|
||
|
||
has_conflicts = ps_has_conflicts (ps, c, c)
|
||
|| (ps->history > 0
|
||
&& ps_has_conflicts (ps,
|
||
c - ps->history,
|
||
c + ps->history));
|
||
|
||
/* Try different issue slots to find one that the given node can be
|
||
scheduled in without conflicts. */
|
||
while (has_conflicts)
|
||
{
|
||
if (! ps_insn_advance_column (ps, ps_i, must_follow))
|
||
break;
|
||
has_conflicts = ps_has_conflicts (ps, c, c)
|
||
|| (ps->history > 0
|
||
&& ps_has_conflicts (ps,
|
||
c - ps->history,
|
||
c + ps->history));
|
||
}
|
||
|
||
if (has_conflicts)
|
||
{
|
||
remove_node_from_ps (ps, ps_i);
|
||
return NULL;
|
||
}
|
||
|
||
ps->min_cycle = MIN (ps->min_cycle, c);
|
||
ps->max_cycle = MAX (ps->max_cycle, c);
|
||
return ps_i;
|
||
}
|
||
|
||
/* Rotate the rows of PS such that insns scheduled at time
|
||
START_CYCLE will appear in row 0. Updates max/min_cycles. */
|
||
void
|
||
rotate_partial_schedule (partial_schedule_ptr ps, int start_cycle)
|
||
{
|
||
int i, row, backward_rotates;
|
||
int last_row = ps->ii - 1;
|
||
|
||
if (start_cycle == 0)
|
||
return;
|
||
|
||
backward_rotates = SMODULO (start_cycle, ps->ii);
|
||
|
||
/* Revisit later and optimize this into a single loop. */
|
||
for (i = 0; i < backward_rotates; i++)
|
||
{
|
||
ps_insn_ptr first_row = ps->rows[0];
|
||
|
||
for (row = 0; row < last_row; row++)
|
||
ps->rows[row] = ps->rows[row+1];
|
||
|
||
ps->rows[last_row] = first_row;
|
||
}
|
||
|
||
ps->max_cycle -= start_cycle;
|
||
ps->min_cycle -= start_cycle;
|
||
}
|
||
|
||
#endif /* INSN_SCHEDULING */
|
||
|
||
static bool
|
||
gate_handle_sms (void)
|
||
{
|
||
return (optimize > 0 && flag_modulo_sched);
|
||
}
|
||
|
||
|
||
/* Run instruction scheduler. */
|
||
/* Perform SMS module scheduling. */
|
||
static unsigned int
|
||
rest_of_handle_sms (void)
|
||
{
|
||
#ifdef INSN_SCHEDULING
|
||
basic_block bb;
|
||
|
||
/* Collect loop information to be used in SMS. */
|
||
cfg_layout_initialize (0);
|
||
sms_schedule ();
|
||
|
||
/* Update the life information, because we add pseudos. */
|
||
max_regno = max_reg_num ();
|
||
|
||
/* Finalize layout changes. */
|
||
FOR_EACH_BB (bb)
|
||
if (bb->next_bb != EXIT_BLOCK_PTR)
|
||
bb->aux = bb->next_bb;
|
||
free_dominance_info (CDI_DOMINATORS);
|
||
cfg_layout_finalize ();
|
||
#endif /* INSN_SCHEDULING */
|
||
return 0;
|
||
}
|
||
|
||
struct rtl_opt_pass pass_sms =
|
||
{
|
||
{
|
||
RTL_PASS,
|
||
"sms", /* name */
|
||
gate_handle_sms, /* gate */
|
||
rest_of_handle_sms, /* execute */
|
||
NULL, /* sub */
|
||
NULL, /* next */
|
||
0, /* static_pass_number */
|
||
TV_SMS, /* tv_id */
|
||
0, /* properties_required */
|
||
0, /* properties_provided */
|
||
0, /* properties_destroyed */
|
||
TODO_dump_func, /* todo_flags_start */
|
||
TODO_df_finish | TODO_verify_rtl_sharing |
|
||
TODO_dump_func |
|
||
TODO_ggc_collect /* todo_flags_finish */
|
||
}
|
||
};
|
||
|