cgraphunit.c, [...]: Fix comment typos.
* cgraphunit.c, config/arm/arm.c, config/m68k/m68k.c, ipa-inline.c, tree-profile.c, tree-ssa-live.c, tree-ssa-math-opts.c, tree-ssanames.c, tree-vect-analyze.c, value-prof.c: Fix comment typos. From-SVN: r121374
This commit is contained in:
parent
65de695fa5
commit
88512ba0b8
|
@ -1,3 +1,10 @@
|
||||||
|
2007-01-31 Kazu Hirata <kazu@codesourcery.com>
|
||||||
|
|
||||||
|
* cgraphunit.c, config/arm/arm.c, config/m68k/m68k.c,
|
||||||
|
ipa-inline.c, tree-profile.c, tree-ssa-live.c,
|
||||||
|
tree-ssa-math-opts.c, tree-ssanames.c, tree-vect-analyze.c,
|
||||||
|
value-prof.c: Fix comment typos.
|
||||||
|
|
||||||
2007-01-30 Manuel Lopez-Ibanez <manu@gcc.gnu.org>
|
2007-01-30 Manuel Lopez-Ibanez <manu@gcc.gnu.org>
|
||||||
|
|
||||||
PR c++/24745
|
PR c++/24745
|
||||||
|
|
|
@ -783,7 +783,7 @@ process_function_and_variable_attributes (struct cgraph_node *first,
|
||||||
/* Process CGRAPH_NODES_NEEDED queue, analyze each function (and transitively
|
/* Process CGRAPH_NODES_NEEDED queue, analyze each function (and transitively
|
||||||
each reachable functions) and build cgraph.
|
each reachable functions) and build cgraph.
|
||||||
The function can be called multiple times after inserting new nodes
|
The function can be called multiple times after inserting new nodes
|
||||||
into beggining of queue. Just the new part of queue is re-scanned then. */
|
into beginning of queue. Just the new part of queue is re-scanned then. */
|
||||||
|
|
||||||
static void
|
static void
|
||||||
cgraph_analyze_functions (void)
|
cgraph_analyze_functions (void)
|
||||||
|
|
|
@ -7710,7 +7710,7 @@ get_jump_table_size (rtx insn)
|
||||||
switch (modesize)
|
switch (modesize)
|
||||||
{
|
{
|
||||||
case 1:
|
case 1:
|
||||||
/* Round up size of TBB table to a haflword boundary. */
|
/* Round up size of TBB table to a halfword boundary. */
|
||||||
size = (size + 1) & ~(HOST_WIDE_INT)1;
|
size = (size + 1) & ~(HOST_WIDE_INT)1;
|
||||||
break;
|
break;
|
||||||
case 2:
|
case 2:
|
||||||
|
|
|
@ -1957,7 +1957,7 @@ m68k_rtx_costs (rtx x, int code, int outer_code, int *total)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
/* Return an instruction to move CONST_INT OPERANDS[1] into data regsiter
|
/* Return an instruction to move CONST_INT OPERANDS[1] into data register
|
||||||
OPERANDS[0]. */
|
OPERANDS[0]. */
|
||||||
|
|
||||||
static const char *
|
static const char *
|
||||||
|
@ -2838,7 +2838,7 @@ notice_update_cc (rtx exp, rtx insn)
|
||||||
codes. Normal moves _do_ set the condition codes, but not in
|
codes. Normal moves _do_ set the condition codes, but not in
|
||||||
a way that is appropriate for comparison with 0, because -0.0
|
a way that is appropriate for comparison with 0, because -0.0
|
||||||
would be treated as a negative nonzero number. Note that it
|
would be treated as a negative nonzero number. Note that it
|
||||||
isn't appropriate to conditionalize this restiction on
|
isn't appropriate to conditionalize this restriction on
|
||||||
HONOR_SIGNED_ZEROS because that macro merely indicates whether
|
HONOR_SIGNED_ZEROS because that macro merely indicates whether
|
||||||
we care about the difference between -0.0 and +0.0. */
|
we care about the difference between -0.0 and +0.0. */
|
||||||
else if (!FP_REG_P (SET_DEST (exp))
|
else if (!FP_REG_P (SET_DEST (exp))
|
||||||
|
|
|
@ -91,11 +91,11 @@ Software Foundation, 51 Franklin Street, Fifth Floor, Boston, MA
|
||||||
maintained by pass manager). The functions after inlining are early
|
maintained by pass manager). The functions after inlining are early
|
||||||
optimized so the early inliner sees unoptimized function itself, but
|
optimized so the early inliner sees unoptimized function itself, but
|
||||||
all considered callees are already optimized allowing it to unfold
|
all considered callees are already optimized allowing it to unfold
|
||||||
abstraction penalty on C++ effectivly and cheaply.
|
abstraction penalty on C++ effectively and cheaply.
|
||||||
|
|
||||||
pass_ipa_early_inlining
|
pass_ipa_early_inlining
|
||||||
|
|
||||||
With profiling, the early inlining is also neccesary to reduce
|
With profiling, the early inlining is also necessary to reduce
|
||||||
instrumentation costs on program with high abstraction penalty (doing
|
instrumentation costs on program with high abstraction penalty (doing
|
||||||
many redundant calls). This can't happen in parallel with early
|
many redundant calls). This can't happen in parallel with early
|
||||||
optimization and profile instrumentation, because we would end up
|
optimization and profile instrumentation, because we would end up
|
||||||
|
@ -751,7 +751,7 @@ cgraph_set_inline_failed (struct cgraph_node *node, const char *reason)
|
||||||
e->inline_failed = reason;
|
e->inline_failed = reason;
|
||||||
}
|
}
|
||||||
|
|
||||||
/* Given whole compilation unit esitmate of INSNS, compute how large we can
|
/* Given whole compilation unit estimate of INSNS, compute how large we can
|
||||||
allow the unit to grow. */
|
allow the unit to grow. */
|
||||||
static int
|
static int
|
||||||
compute_max_insns (int insns)
|
compute_max_insns (int insns)
|
||||||
|
@ -1043,7 +1043,7 @@ cgraph_decide_inlining (void)
|
||||||
e->caller->global.insns);
|
e->caller->global.insns);
|
||||||
}
|
}
|
||||||
/* Inlining self recursive function might introduce new calls to
|
/* Inlining self recursive function might introduce new calls to
|
||||||
thsemselves we didn't see in the loop above. Fill in the proper
|
themselves we didn't see in the loop above. Fill in the proper
|
||||||
reason why inline failed. */
|
reason why inline failed. */
|
||||||
for (e = node->callers; e; e = e->next_caller)
|
for (e = node->callers; e; e = e->next_caller)
|
||||||
if (e->inline_failed)
|
if (e->inline_failed)
|
||||||
|
@ -1126,7 +1126,7 @@ cgraph_decide_inlining (void)
|
||||||
recursive inlining, but as an special case, we want to try harder inline
|
recursive inlining, but as an special case, we want to try harder inline
|
||||||
ALWAYS_INLINE functions: consider callgraph a->b->c->b, with a being
|
ALWAYS_INLINE functions: consider callgraph a->b->c->b, with a being
|
||||||
flatten, b being always inline. Flattening 'a' will collapse
|
flatten, b being always inline. Flattening 'a' will collapse
|
||||||
a->b->c before hitting cycle. To accomondate always inline, we however
|
a->b->c before hitting cycle. To accommodate always inline, we however
|
||||||
need to inline a->b->c->b.
|
need to inline a->b->c->b.
|
||||||
|
|
||||||
So after hitting cycle first time, we switch into ALWAYS_INLINE mode and
|
So after hitting cycle first time, we switch into ALWAYS_INLINE mode and
|
||||||
|
@ -1145,7 +1145,7 @@ try_inline (struct cgraph_edge *e, enum inlining_mode mode, int depth)
|
||||||
mode yet. and the function in question is always_inline. */
|
mode yet. and the function in question is always_inline. */
|
||||||
if (always_inline && mode != INLINE_ALWAYS_INLINE)
|
if (always_inline && mode != INLINE_ALWAYS_INLINE)
|
||||||
mode = INLINE_ALWAYS_INLINE;
|
mode = INLINE_ALWAYS_INLINE;
|
||||||
/* Otheriwse it is time to give up. */
|
/* Otherwise it is time to give up. */
|
||||||
else
|
else
|
||||||
{
|
{
|
||||||
if (dump_file)
|
if (dump_file)
|
||||||
|
|
|
@ -64,7 +64,7 @@ static GTY(()) tree ptr_void;
|
||||||
|
|
||||||
/* Add code:
|
/* Add code:
|
||||||
static gcov* __gcov_indirect_call_counters; // pointer to actual counter
|
static gcov* __gcov_indirect_call_counters; // pointer to actual counter
|
||||||
static void* __gcov_indirect_call_callee; // actual callie addres
|
static void* __gcov_indirect_call_callee; // actual callee address
|
||||||
*/
|
*/
|
||||||
static void
|
static void
|
||||||
tree_init_ic_make_global_vars (void)
|
tree_init_ic_make_global_vars (void)
|
||||||
|
@ -269,7 +269,7 @@ tree_gen_one_value_profiler (histogram_value value, unsigned tag, unsigned base)
|
||||||
|
|
||||||
/* Output instructions as GIMPLE trees for code to find the most
|
/* Output instructions as GIMPLE trees for code to find the most
|
||||||
common called function in indirect call.
|
common called function in indirect call.
|
||||||
VALUE is the call expression whose indirect callie is profiled.
|
VALUE is the call expression whose indirect callee is profiled.
|
||||||
TAG is the tag of the section for counters, BASE is offset of the
|
TAG is the tag of the section for counters, BASE is offset of the
|
||||||
counter position. */
|
counter position. */
|
||||||
|
|
||||||
|
@ -308,7 +308,7 @@ tree_gen_ic_profiler (histogram_value value, unsigned tag, unsigned base)
|
||||||
|
|
||||||
/* Output instructions as GIMPLE trees for code to find the most
|
/* Output instructions as GIMPLE trees for code to find the most
|
||||||
common called function in indirect call. Insert instructions at the
|
common called function in indirect call. Insert instructions at the
|
||||||
begining of every possible called function.
|
beginning of every possible called function.
|
||||||
*/
|
*/
|
||||||
|
|
||||||
static void
|
static void
|
||||||
|
|
|
@ -504,7 +504,7 @@ remove_unused_locals (void)
|
||||||
|
|
||||||
/* Remove unused variables from REFERENCED_VARs. As an special exception
|
/* Remove unused variables from REFERENCED_VARs. As an special exception
|
||||||
keep the variables that are believed to be aliased. Those can't be
|
keep the variables that are believed to be aliased. Those can't be
|
||||||
easilly removed from the alias sets and and operand caches.
|
easily removed from the alias sets and and operand caches.
|
||||||
They will be removed shortly after next may_alias pass is performed. */
|
They will be removed shortly after next may_alias pass is performed. */
|
||||||
FOR_EACH_REFERENCED_VAR (t, rvi)
|
FOR_EACH_REFERENCED_VAR (t, rvi)
|
||||||
if (!is_global_var (t)
|
if (!is_global_var (t)
|
||||||
|
|
|
@ -521,9 +521,9 @@ struct tree_opt_pass pass_cse_reciprocals =
|
||||||
0 /* letter */
|
0 /* letter */
|
||||||
};
|
};
|
||||||
|
|
||||||
/* Records an occurance at statement USE_STMT in the vector of trees
|
/* Records an occurrence at statement USE_STMT in the vector of trees
|
||||||
STMTS if it is dominated by *TOP_BB or dominates it or this basic block
|
STMTS if it is dominated by *TOP_BB or dominates it or this basic block
|
||||||
is not yet initialized. Returns true if the occurance was pushed on
|
is not yet initialized. Returns true if the occurrence was pushed on
|
||||||
the vector. Adjusts *TOP_BB to be the basic block dominating all
|
the vector. Adjusts *TOP_BB to be the basic block dominating all
|
||||||
statements in the vector. */
|
statements in the vector. */
|
||||||
|
|
||||||
|
|
|
@ -318,7 +318,7 @@ release_dead_ssa_names (void)
|
||||||
referenced_var_iterator rvi;
|
referenced_var_iterator rvi;
|
||||||
|
|
||||||
/* Current defs point to various dead SSA names that in turn points to dead
|
/* Current defs point to various dead SSA names that in turn points to dead
|
||||||
statements so bunch of dead memory is holded from releasing. */
|
statements so bunch of dead memory is held from releasing. */
|
||||||
FOR_EACH_REFERENCED_VAR (t, rvi)
|
FOR_EACH_REFERENCED_VAR (t, rvi)
|
||||||
set_current_def (t, NULL);
|
set_current_def (t, NULL);
|
||||||
/* Now release the freelist. */
|
/* Now release the freelist. */
|
||||||
|
@ -328,7 +328,7 @@ release_dead_ssa_names (void)
|
||||||
/* Dangling pointers might make GGC to still see dead SSA names, so it is
|
/* Dangling pointers might make GGC to still see dead SSA names, so it is
|
||||||
important to unlink the list and avoid GGC from seeing all subsequent
|
important to unlink the list and avoid GGC from seeing all subsequent
|
||||||
SSA names. In longer run we want to have all dangling pointers here
|
SSA names. In longer run we want to have all dangling pointers here
|
||||||
removed (since they usually go trhough dead statements that consume
|
removed (since they usually go through dead statements that consume
|
||||||
considerable amounts of memory). */
|
considerable amounts of memory). */
|
||||||
TREE_CHAIN (t) = NULL_TREE;
|
TREE_CHAIN (t) = NULL_TREE;
|
||||||
n++;
|
n++;
|
||||||
|
|
|
@ -164,7 +164,7 @@ vect_determine_vectorization_factor (loop_vec_info loop_vinfo)
|
||||||
arguments (e.g. demotion, promotion), vectype will be reset
|
arguments (e.g. demotion, promotion), vectype will be reset
|
||||||
appropriately (later). Note that we have to visit the smallest
|
appropriately (later). Note that we have to visit the smallest
|
||||||
datatype in this function, because that determines the VF.
|
datatype in this function, because that determines the VF.
|
||||||
If the samallest datatype in the loop is present only as the
|
If the smallest datatype in the loop is present only as the
|
||||||
rhs of a promotion operation - we'd miss it here.
|
rhs of a promotion operation - we'd miss it here.
|
||||||
However, in such a case, that a variable of this datatype
|
However, in such a case, that a variable of this datatype
|
||||||
does not appear in the lhs anywhere in the loop, it shouldn't
|
does not appear in the lhs anywhere in the loop, it shouldn't
|
||||||
|
@ -1752,7 +1752,7 @@ vect_analyze_data_ref_access (struct data_reference *dr)
|
||||||
return false;
|
return false;
|
||||||
}
|
}
|
||||||
|
|
||||||
/* Check that there is no load-store dependecies for this loads
|
/* Check that there is no load-store dependencies for this loads
|
||||||
to prevent a case of load-store-load to the same location. */
|
to prevent a case of load-store-load to the same location. */
|
||||||
if (DR_GROUP_READ_WRITE_DEPENDENCE (vinfo_for_stmt (next))
|
if (DR_GROUP_READ_WRITE_DEPENDENCE (vinfo_for_stmt (next))
|
||||||
|| DR_GROUP_READ_WRITE_DEPENDENCE (vinfo_for_stmt (prev)))
|
|| DR_GROUP_READ_WRITE_DEPENDENCE (vinfo_for_stmt (prev)))
|
||||||
|
|
|
@ -63,7 +63,7 @@ static struct value_prof_hooks *value_prof_hooks;
|
||||||
|
|
||||||
3) Indirect/virtual call specialization. If we can determine most
|
3) Indirect/virtual call specialization. If we can determine most
|
||||||
common function callee in indirect/virtual call. We can use this
|
common function callee in indirect/virtual call. We can use this
|
||||||
information to improve code effectivity (espetialy info for
|
information to improve code effectiveness (especially info for
|
||||||
inliner).
|
inliner).
|
||||||
|
|
||||||
Every such optimization should add its requirements for profiled values to
|
Every such optimization should add its requirements for profiled values to
|
||||||
|
|
Loading…
Reference in New Issue