diff --git a/gcc/ChangeLog b/gcc/ChangeLog index a58b1dfa0ea..64e97e50c5b 100644 --- a/gcc/ChangeLog +++ b/gcc/ChangeLog @@ -1,3 +1,10 @@ +2007-01-31 Kazu Hirata + + * cgraphunit.c, config/arm/arm.c, config/m68k/m68k.c, + ipa-inline.c, tree-profile.c, tree-ssa-live.c, + tree-ssa-math-opts.c, tree-ssanames.c, tree-vect-analyze.c, + value-prof.c: Fix comment typos. + 2007-01-30 Manuel Lopez-Ibanez PR c++/24745 diff --git a/gcc/cgraphunit.c b/gcc/cgraphunit.c index c4cf9e92580..892e952e238 100644 --- a/gcc/cgraphunit.c +++ b/gcc/cgraphunit.c @@ -783,7 +783,7 @@ process_function_and_variable_attributes (struct cgraph_node *first, /* Process CGRAPH_NODES_NEEDED queue, analyze each function (and transitively each reachable functions) and build cgraph. The function can be called multiple times after inserting new nodes - into beggining of queue. Just the new part of queue is re-scanned then. */ + into beginning of queue. Just the new part of queue is re-scanned then. */ static void cgraph_analyze_functions (void) diff --git a/gcc/config/arm/arm.c b/gcc/config/arm/arm.c index 20b8fd75f7b..c342f2fdac2 100644 --- a/gcc/config/arm/arm.c +++ b/gcc/config/arm/arm.c @@ -7710,7 +7710,7 @@ get_jump_table_size (rtx insn) switch (modesize) { case 1: - /* Round up size of TBB table to a haflword boundary. */ + /* Round up size of TBB table to a halfword boundary. */ size = (size + 1) & ~(HOST_WIDE_INT)1; break; case 2: diff --git a/gcc/config/m68k/m68k.c b/gcc/config/m68k/m68k.c index 78aae6dc0b5..edf1dc61fe4 100644 --- a/gcc/config/m68k/m68k.c +++ b/gcc/config/m68k/m68k.c @@ -1957,7 +1957,7 @@ m68k_rtx_costs (rtx x, int code, int outer_code, int *total) } } -/* Return an instruction to move CONST_INT OPERANDS[1] into data regsiter +/* Return an instruction to move CONST_INT OPERANDS[1] into data register OPERANDS[0]. */ static const char * @@ -2838,7 +2838,7 @@ notice_update_cc (rtx exp, rtx insn) codes. Normal moves _do_ set the condition codes, but not in a way that is appropriate for comparison with 0, because -0.0 would be treated as a negative nonzero number. Note that it - isn't appropriate to conditionalize this restiction on + isn't appropriate to conditionalize this restriction on HONOR_SIGNED_ZEROS because that macro merely indicates whether we care about the difference between -0.0 and +0.0. */ else if (!FP_REG_P (SET_DEST (exp)) diff --git a/gcc/ipa-inline.c b/gcc/ipa-inline.c index 7640fb79501..8b63b51c7f5 100644 --- a/gcc/ipa-inline.c +++ b/gcc/ipa-inline.c @@ -91,11 +91,11 @@ Software Foundation, 51 Franklin Street, Fifth Floor, Boston, MA maintained by pass manager). The functions after inlining are early optimized so the early inliner sees unoptimized function itself, but all considered callees are already optimized allowing it to unfold - abstraction penalty on C++ effectivly and cheaply. + abstraction penalty on C++ effectively and cheaply. pass_ipa_early_inlining - With profiling, the early inlining is also neccesary to reduce + With profiling, the early inlining is also necessary to reduce instrumentation costs on program with high abstraction penalty (doing many redundant calls). This can't happen in parallel with early optimization and profile instrumentation, because we would end up @@ -751,7 +751,7 @@ cgraph_set_inline_failed (struct cgraph_node *node, const char *reason) e->inline_failed = reason; } -/* Given whole compilation unit esitmate of INSNS, compute how large we can +/* Given whole compilation unit estimate of INSNS, compute how large we can allow the unit to grow. */ static int compute_max_insns (int insns) @@ -1043,7 +1043,7 @@ cgraph_decide_inlining (void) e->caller->global.insns); } /* Inlining self recursive function might introduce new calls to - thsemselves we didn't see in the loop above. Fill in the proper + themselves we didn't see in the loop above. Fill in the proper reason why inline failed. */ for (e = node->callers; e; e = e->next_caller) if (e->inline_failed) @@ -1126,7 +1126,7 @@ cgraph_decide_inlining (void) recursive inlining, but as an special case, we want to try harder inline ALWAYS_INLINE functions: consider callgraph a->b->c->b, with a being flatten, b being always inline. Flattening 'a' will collapse - a->b->c before hitting cycle. To accomondate always inline, we however + a->b->c before hitting cycle. To accommodate always inline, we however need to inline a->b->c->b. So after hitting cycle first time, we switch into ALWAYS_INLINE mode and @@ -1145,7 +1145,7 @@ try_inline (struct cgraph_edge *e, enum inlining_mode mode, int depth) mode yet. and the function in question is always_inline. */ if (always_inline && mode != INLINE_ALWAYS_INLINE) mode = INLINE_ALWAYS_INLINE; - /* Otheriwse it is time to give up. */ + /* Otherwise it is time to give up. */ else { if (dump_file) diff --git a/gcc/tree-profile.c b/gcc/tree-profile.c index b5088f07d32..b8196658385 100644 --- a/gcc/tree-profile.c +++ b/gcc/tree-profile.c @@ -64,7 +64,7 @@ static GTY(()) tree ptr_void; /* Add code: static gcov* __gcov_indirect_call_counters; // pointer to actual counter - static void* __gcov_indirect_call_callee; // actual callie addres + static void* __gcov_indirect_call_callee; // actual callee address */ static void tree_init_ic_make_global_vars (void) @@ -269,7 +269,7 @@ tree_gen_one_value_profiler (histogram_value value, unsigned tag, unsigned base) /* Output instructions as GIMPLE trees for code to find the most common called function in indirect call. - VALUE is the call expression whose indirect callie is profiled. + VALUE is the call expression whose indirect callee is profiled. TAG is the tag of the section for counters, BASE is offset of the counter position. */ @@ -308,7 +308,7 @@ tree_gen_ic_profiler (histogram_value value, unsigned tag, unsigned base) /* Output instructions as GIMPLE trees for code to find the most common called function in indirect call. Insert instructions at the - begining of every possible called function. + beginning of every possible called function. */ static void diff --git a/gcc/tree-ssa-live.c b/gcc/tree-ssa-live.c index cc4972f54b2..0113b1b36e7 100644 --- a/gcc/tree-ssa-live.c +++ b/gcc/tree-ssa-live.c @@ -504,7 +504,7 @@ remove_unused_locals (void) /* Remove unused variables from REFERENCED_VARs. As an special exception keep the variables that are believed to be aliased. Those can't be - easilly removed from the alias sets and and operand caches. + easily removed from the alias sets and and operand caches. They will be removed shortly after next may_alias pass is performed. */ FOR_EACH_REFERENCED_VAR (t, rvi) if (!is_global_var (t) diff --git a/gcc/tree-ssa-math-opts.c b/gcc/tree-ssa-math-opts.c index ae7a40bf8ca..bb51c3186e6 100644 --- a/gcc/tree-ssa-math-opts.c +++ b/gcc/tree-ssa-math-opts.c @@ -521,9 +521,9 @@ struct tree_opt_pass pass_cse_reciprocals = 0 /* letter */ }; -/* Records an occurance at statement USE_STMT in the vector of trees +/* Records an occurrence at statement USE_STMT in the vector of trees STMTS if it is dominated by *TOP_BB or dominates it or this basic block - is not yet initialized. Returns true if the occurance was pushed on + is not yet initialized. Returns true if the occurrence was pushed on the vector. Adjusts *TOP_BB to be the basic block dominating all statements in the vector. */ diff --git a/gcc/tree-ssanames.c b/gcc/tree-ssanames.c index 07b83f831af..9c61a3f1c38 100644 --- a/gcc/tree-ssanames.c +++ b/gcc/tree-ssanames.c @@ -318,7 +318,7 @@ release_dead_ssa_names (void) referenced_var_iterator rvi; /* Current defs point to various dead SSA names that in turn points to dead - statements so bunch of dead memory is holded from releasing. */ + statements so bunch of dead memory is held from releasing. */ FOR_EACH_REFERENCED_VAR (t, rvi) set_current_def (t, NULL); /* Now release the freelist. */ @@ -328,7 +328,7 @@ release_dead_ssa_names (void) /* Dangling pointers might make GGC to still see dead SSA names, so it is important to unlink the list and avoid GGC from seeing all subsequent SSA names. In longer run we want to have all dangling pointers here - removed (since they usually go trhough dead statements that consume + removed (since they usually go through dead statements that consume considerable amounts of memory). */ TREE_CHAIN (t) = NULL_TREE; n++; diff --git a/gcc/tree-vect-analyze.c b/gcc/tree-vect-analyze.c index c53c34e2c51..88c937c3fa3 100644 --- a/gcc/tree-vect-analyze.c +++ b/gcc/tree-vect-analyze.c @@ -164,7 +164,7 @@ vect_determine_vectorization_factor (loop_vec_info loop_vinfo) arguments (e.g. demotion, promotion), vectype will be reset appropriately (later). Note that we have to visit the smallest datatype in this function, because that determines the VF. - If the samallest datatype in the loop is present only as the + If the smallest datatype in the loop is present only as the rhs of a promotion operation - we'd miss it here. However, in such a case, that a variable of this datatype does not appear in the lhs anywhere in the loop, it shouldn't @@ -1752,7 +1752,7 @@ vect_analyze_data_ref_access (struct data_reference *dr) return false; } - /* Check that there is no load-store dependecies for this loads + /* Check that there is no load-store dependencies for this loads to prevent a case of load-store-load to the same location. */ if (DR_GROUP_READ_WRITE_DEPENDENCE (vinfo_for_stmt (next)) || DR_GROUP_READ_WRITE_DEPENDENCE (vinfo_for_stmt (prev))) diff --git a/gcc/value-prof.c b/gcc/value-prof.c index 869e1693ac4..0d5f0411e1d 100644 --- a/gcc/value-prof.c +++ b/gcc/value-prof.c @@ -63,7 +63,7 @@ static struct value_prof_hooks *value_prof_hooks; 3) Indirect/virtual call specialization. If we can determine most common function callee in indirect/virtual call. We can use this - information to improve code effectivity (espetialy info for + information to improve code effectiveness (especially info for inliner). Every such optimization should add its requirements for profiled values to