aboutsummaryrefslogtreecommitdiff
path: root/gcc
diff options
context:
space:
mode:
authorKazu Hirata <kazu@codesourcery.com>2006-03-04 23:05:24 +0000
committerKazu Hirata <kazu@gcc.gnu.org>2006-03-04 23:05:24 +0000
commitc0220ea4c509e3c14cef20c864545340aa4b98b4 (patch)
tree4350c209f16fe5635562cd7e6a9850c631419372 /gcc
parent8156c8b0b9a1880ea8501385d035dbf9ae990306 (diff)
downloadgcc-c0220ea4c509e3c14cef20c864545340aa4b98b4.zip
gcc-c0220ea4c509e3c14cef20c864545340aa4b98b4.tar.gz
gcc-c0220ea4c509e3c14cef20c864545340aa4b98b4.tar.bz2
builtins.c, [...]: Fix comment typos.
* builtins.c, c-pragma.h, c-typeck.c, cgraph.c, cgraphunit.c, combine.c, common.opt, config/dfp-bit.c, config/i386/i386.c, config/m68k/m68k.c, config/m68k/m68k.md, config/mt/mt.c, config/mt/mt.h, config/s390/s390.md, df-core.c, df-problems.c, df-scan.c, df.h, diagnostic.c, expr.c, function.h, gimplify.c, loop-invariant.c, omp-low.c, opts.c, passes.c, rtl-factoring.c, rtlanal.c, struct-equiv.c, tree-cfgcleanup.c, tree-ssa-loop-niter.c, tree-ssa-loop-prefetch.c, tree-ssa-structalias.c, tree-ssa-threadedge.c, tree-ssa-threadupdate.c, tree-vect-patterns.c, tree-vect-transform.c, tree-vectorizer.h, tree-vrp.c, unwind-dw2.c: Fix comment typos. Follow spelling conventions. From-SVN: r111721
Diffstat (limited to 'gcc')
-rw-r--r--gcc/ChangeLog15
-rw-r--r--gcc/builtins.c2
-rw-r--r--gcc/c-pragma.h2
-rw-r--r--gcc/c-typeck.c2
-rw-r--r--gcc/cgraph.c2
-rw-r--r--gcc/cgraphunit.c2
-rw-r--r--gcc/combine.c2
-rw-r--r--gcc/common.opt6
-rw-r--r--gcc/config/dfp-bit.c4
-rw-r--r--gcc/config/i386/i386.c4
-rw-r--r--gcc/config/m68k/m68k.c4
-rw-r--r--gcc/config/m68k/m68k.md2
-rw-r--r--gcc/config/mt/mt.c8
-rw-r--r--gcc/config/mt/mt.h2
-rw-r--r--gcc/config/s390/s390.md4
-rw-r--r--gcc/df-core.c8
-rw-r--r--gcc/df-problems.c4
-rw-r--r--gcc/df-scan.c6
-rw-r--r--gcc/df.h2
-rw-r--r--gcc/diagnostic.c2
-rw-r--r--gcc/expr.c2
-rw-r--r--gcc/function.h2
-rw-r--r--gcc/gimplify.c6
-rw-r--r--gcc/loop-invariant.c2
-rw-r--r--gcc/omp-low.c4
-rw-r--r--gcc/opts.c2
-rw-r--r--gcc/passes.c2
-rw-r--r--gcc/rtl-factoring.c10
-rw-r--r--gcc/rtlanal.c2
-rw-r--r--gcc/struct-equiv.c2
-rw-r--r--gcc/tree-cfgcleanup.c2
-rw-r--r--gcc/tree-ssa-loop-niter.c2
-rw-r--r--gcc/tree-ssa-loop-prefetch.c4
-rw-r--r--gcc/tree-ssa-structalias.c6
-rw-r--r--gcc/tree-ssa-threadedge.c2
-rw-r--r--gcc/tree-ssa-threadupdate.c2
-rw-r--r--gcc/tree-vect-patterns.c8
-rw-r--r--gcc/tree-vect-transform.c4
-rw-r--r--gcc/tree-vectorizer.h2
-rw-r--r--gcc/tree-vrp.c2
-rw-r--r--gcc/unwind-dw2.c2
41 files changed, 84 insertions, 69 deletions
diff --git a/gcc/ChangeLog b/gcc/ChangeLog
index 3474253..e6fa683 100644
--- a/gcc/ChangeLog
+++ b/gcc/ChangeLog
@@ -1,5 +1,20 @@
2006-03-04 Kazu Hirata <kazu@codesourcery.com>
+ * builtins.c, c-pragma.h, c-typeck.c, cgraph.c, cgraphunit.c,
+ combine.c, common.opt, config/dfp-bit.c, config/i386/i386.c,
+ config/m68k/m68k.c, config/m68k/m68k.md, config/mt/mt.c,
+ config/mt/mt.h, config/s390/s390.md, df-core.c, df-problems.c,
+ df-scan.c, df.h, diagnostic.c, expr.c, function.h, gimplify.c,
+ loop-invariant.c, omp-low.c, opts.c, passes.c,
+ rtl-factoring.c, rtlanal.c, struct-equiv.c, tree-cfgcleanup.c,
+ tree-ssa-loop-niter.c, tree-ssa-loop-prefetch.c,
+ tree-ssa-structalias.c, tree-ssa-threadedge.c,
+ tree-ssa-threadupdate.c, tree-vect-patterns.c,
+ tree-vect-transform.c, tree-vectorizer.h, tree-vrp.c,
+ unwind-dw2.c: Fix comment typos. Follow spelling conventions.
+
+2006-03-04 Kazu Hirata <kazu@codesourcery.com>
+
* dwarf2out.c (dwarf2out_add_library_unit_info): Remove.
* dwarf2out.h: Remove the prototype for
dwarf2out_add_library_unit_info.
diff --git a/gcc/builtins.c b/gcc/builtins.c
index b9755bf..f6095e8 100644
--- a/gcc/builtins.c
+++ b/gcc/builtins.c
@@ -498,7 +498,7 @@ expand_builtin_return_addr (enum built_in_function fndecl_code, int count)
/* For a zero count, we don't care what frame address we return, so frame
pointer elimination is OK, and using the soft frame pointer is OK.
- For a non-zero count, we require a stable offset from the current frame
+ For a nonzero count, we require a stable offset from the current frame
pointer to the previous one, so we must use the hard frame pointer, and
we must disable frame pointer elimination. */
if (count == 0)
diff --git a/gcc/c-pragma.h b/gcc/c-pragma.h
index 28ef4e8..5868f07 100644
--- a/gcc/c-pragma.h
+++ b/gcc/c-pragma.h
@@ -25,7 +25,7 @@ Software Foundation, 51 Franklin Street, Fifth Floor, Boston, MA
#include <cpplib.h> /* For enum cpp_ttype. */
/* Pragma identifiers built in to the front end parsers. Identifiers
- for anciliary handlers will follow these. */
+ for ancillary handlers will follow these. */
typedef enum pragma_kind {
PRAGMA_NONE = 0,
diff --git a/gcc/c-typeck.c b/gcc/c-typeck.c
index ef51895..98eeacd 100644
--- a/gcc/c-typeck.c
+++ b/gcc/c-typeck.c
@@ -8456,7 +8456,7 @@ c_expr_to_decl (tree expr, bool *tc ATTRIBUTE_UNUSED,
}
-/* Like c_begin_compound_stmt, except force the retension of the BLOCK. */
+/* Like c_begin_compound_stmt, except force the retention of the BLOCK. */
tree
c_begin_omp_parallel (void)
diff --git a/gcc/cgraph.c b/gcc/cgraph.c
index a05f76d..6a8ac08 100644
--- a/gcc/cgraph.c
+++ b/gcc/cgraph.c
@@ -114,7 +114,7 @@ struct cgraph_node *cgraph_nodes;
struct cgraph_node *cgraph_nodes_queue;
/* Queue of cgraph nodes scheduled to be expanded. This is a
- secondary queue used during optimization to accomodate passes that
+ secondary queue used during optimization to accommodate passes that
may generate new functions that need to be optimized and expanded. */
struct cgraph_node *cgraph_expand_queue;
diff --git a/gcc/cgraphunit.c b/gcc/cgraphunit.c
index f6540c0..80f26fa 100644
--- a/gcc/cgraphunit.c
+++ b/gcc/cgraphunit.c
@@ -1692,7 +1692,7 @@ save_inline_function_body (struct cgraph_node *node)
cgraph_lower_function (node);
/* In non-unit-at-a-time we construct full fledged clone we never output to
- assembly file. This clone is pointed out by inline_decl of orginal function
+ assembly file. This clone is pointed out by inline_decl of original function
and inlining infrastructure knows how to deal with this. */
if (!flag_unit_at_a_time)
{
diff --git a/gcc/combine.c b/gcc/combine.c
index b7199ad..74b01cb 100644
--- a/gcc/combine.c
+++ b/gcc/combine.c
@@ -11123,7 +11123,7 @@ record_truncated_value (rtx x)
x = SUBREG_REG (x);
}
- /* ??? For hard-regs we now record everthing. We might be able to
+ /* ??? For hard-regs we now record everything. We might be able to
optimize this using last_set_mode. */
else if (REG_P (x) && REGNO (x) < FIRST_PSEUDO_REGISTER)
truncated_mode = GET_MODE (x);
diff --git a/gcc/common.opt b/gcc/common.opt
index 02ade79..5b00a7b 100644
--- a/gcc/common.opt
+++ b/gcc/common.opt
@@ -557,7 +557,7 @@ Give external symbols a leading underscore
floop-optimize
Common
-Does nothing. Preserved for backward compatability.
+Does nothing. Preserved for backward compatibility.
fmath-errno
Common Report Var(flag_errno_math) Init(1)
@@ -723,7 +723,7 @@ Add a common subexpression elimination pass after loop optimizations
frerun-loop-opt
Common
-Does nothing. Preserved for backward compatability.
+Does nothing. Preserved for backward compatibility.
frounding-math
Common Report Var(flag_rounding_math)
@@ -842,7 +842,7 @@ Use a stack protection method for every function
fstrength-reduce
Common
-Does nothing. Preserved for backward compatability.
+Does nothing. Preserved for backward compatibility.
; Nonzero if we should do (language-dependent) alias analysis.
; Typically, this analysis will assume that expressions of certain
diff --git a/gcc/config/dfp-bit.c b/gcc/config/dfp-bit.c
index 835586b..271bfb5 100644
--- a/gcc/config/dfp-bit.c
+++ b/gcc/config/dfp-bit.c
@@ -415,7 +415,7 @@ DFP_TO_INT (DFP_C_TYPE x)
decNumberFromString (&qval, (char *) "1.0", &context);
/* Force the exponent to zero. */
decNumberQuantize (&n1, &n2, &qval, &context);
- /* This is based on text in N1107 secton 5.1; it might turn out to be
+ /* This is based on text in N1107 section 5.1; it might turn out to be
undefined behavior instead. */
if (context.status & DEC_Invalid_operation)
{
@@ -510,7 +510,7 @@ BFP_TO_DFP (BFP_TYPE x)
/* Use a C library function to write the floating point value to a string. */
#ifdef BFP_VIA_TYPE
- /* FIXME: Is threre a better way to output an XFmode variable in C? */
+ /* FIXME: Is there a better way to output an XFmode variable in C? */
sprintf (buf, BFP_FMT, (BFP_VIA_TYPE) x);
#else
sprintf (buf, BFP_FMT, x);
diff --git a/gcc/config/i386/i386.c b/gcc/config/i386/i386.c
index 262c006..7758f43 100644
--- a/gcc/config/i386/i386.c
+++ b/gcc/config/i386/i386.c
@@ -742,7 +742,7 @@ const int x86_use_sahf = m_PPRO | m_K6 | m_PENT4 | m_NOCONA | m_GENERIC32; /*m_G
/* We probably ought to watch for partial register stalls on Generic32
compilation setting as well. However in current implementation the
partial register stalls are not eliminated very well - they can
- be introduced via subregs synthetized by combine and can happen
+ be introduced via subregs synthesized by combine and can happen
in caller/callee saving sequences.
Because this option pays back little on PPro based chips and is in conflict
with partial reg. dependencies used by Athlon/P4 based chips, it is better
@@ -777,7 +777,7 @@ const int x86_prologue_using_move = m_ATHLON_K8 | m_PPRO | m_GENERIC;
const int x86_epilogue_using_move = m_ATHLON_K8 | m_PPRO | m_GENERIC;
const int x86_shift1 = ~m_486;
const int x86_arch_always_fancy_math_387 = m_PENT | m_PPRO | m_ATHLON_K8 | m_PENT4 | m_NOCONA | m_GENERIC;
-/* In Generic model we have an confict here in between PPro/Pentium4 based chips
+/* In Generic model we have an conflict here in between PPro/Pentium4 based chips
that thread 128bit SSE registers as single units versus K8 based chips that
divide SSE registers to two 64bit halves.
x86_sse_partial_reg_dependency promote all store destinations to be 128bit
diff --git a/gcc/config/m68k/m68k.c b/gcc/config/m68k/m68k.c
index 1e753df..7b3a8ed 100644
--- a/gcc/config/m68k/m68k.c
+++ b/gcc/config/m68k/m68k.c
@@ -2342,7 +2342,7 @@ fp_reg_operand (rtx op, enum machine_mode mode ATTRIBUTE_UNUSED)
normally.
Note SCRATCH_REG may not be in the proper mode depending on how it
- will be used. This routine is resposible for creating a new copy
+ will be used. This routine is responsible for creating a new copy
of SCRATCH_REG in the proper mode. */
int
@@ -3647,7 +3647,7 @@ m68k_regno_mode_ok (int regno, enum machine_mode mode)
/* Return floating point values in a 68881 register. This makes 68881 code
a little bit faster. It also makes -msoft-float code incompatible with
hard-float code, so people have to be careful not to mix the two.
- For ColdFire it was decided the ABI incopmatibility is undesirable.
+ For ColdFire it was decided the ABI incompatibility is undesirable.
If there is need for a hard-float ABI it is probably worth doing it
properly and also passing function arguments in FP registers. */
rtx
diff --git a/gcc/config/m68k/m68k.md b/gcc/config/m68k/m68k.md
index 4c24756..6bbeff2 100644
--- a/gcc/config/m68k/m68k.md
+++ b/gcc/config/m68k/m68k.md
@@ -138,7 +138,7 @@
(define_mode_attr round [(SF "%$") (DF "%&") (XF "")])
;; Mnemonic infix to round result for mul or div instruction
(define_mode_attr round_mul [(SF "sgl") (DF "%&") (XF "")])
-;; Suffix specifiying source operand format
+;; Suffix specifying source operand format
(define_mode_attr prec [(SF "s") (DF "d") (XF "x")])
;; Allowable D registers
(define_mode_attr dreg [(SF "d") (DF "") (XF "")])
diff --git a/gcc/config/mt/mt.c b/gcc/config/mt/mt.c
index 59529ac..44c494e 100644
--- a/gcc/config/mt/mt.c
+++ b/gcc/config/mt/mt.c
@@ -683,7 +683,7 @@ mt_legitimate_simple_address_p (enum machine_mode mode ATTRIBUTE_UNUSED,
}
-/* Helper function of GO_IF_LEGITIMATE_ADDRESS. Return non-zero if
+/* Helper function of GO_IF_LEGITIMATE_ADDRESS. Return nonzero if
XINSN is a legitimate address on MT. */
int
mt_legitimate_address_p (enum machine_mode mode, rtx xinsn, int strict)
@@ -1648,9 +1648,9 @@ void mt_add_loop (void)
}
-/* Maxium loop nesting depth. */
+/* Maximum loop nesting depth. */
#define MAX_LOOP_DEPTH 4
-/* Maxium size of a loop (allows some headroom for delayed branch slot
+/* Maximum size of a loop (allows some headroom for delayed branch slot
filling. */
#define MAX_LOOP_LENGTH (200 * 4)
@@ -1850,7 +1850,7 @@ mt_scan_loop (loop_info loop, rtx reg, rtx dbnz)
loop iterations. It can be nested with an automatically maintained
stack of counter and end address registers. It's an ideal
candidate for doloop. Unfortunately, gcc presumes that loops
- always end with an explicit instriction, and the doloop_begin
+ always end with an explicit instruction, and the doloop_begin
instruction is not a flow control instruction so it can be
scheduled earlier than just before the start of the loop. To make
matters worse, the optimization pipeline can duplicate loop exit
diff --git a/gcc/config/mt/mt.h b/gcc/config/mt/mt.h
index 9764d7f..90d98ac 100644
--- a/gcc/config/mt/mt.h
+++ b/gcc/config/mt/mt.h
@@ -512,7 +512,7 @@ extern struct mt_frame_info current_frame_info;
{FRAME_POINTER_REGNUM, STACK_POINTER_REGNUM} \
}
-/* A C expression that returns non-zero if the compiler is allowed to try to
+/* A C expression that returns nonzero if the compiler is allowed to try to
replace register number FROM with register number TO. */
#define CAN_ELIMINATE(FROM, TO) \
((FROM) == ARG_POINTER_REGNUM && (TO) == STACK_POINTER_REGNUM \
diff --git a/gcc/config/s390/s390.md b/gcc/config/s390/s390.md
index 0f3e562..c03f880 100644
--- a/gcc/config/s390/s390.md
+++ b/gcc/config/s390/s390.md
@@ -454,7 +454,7 @@
; Load-and-Test instructions
;
-; tst(di|si) intruction pattern(s).
+; tst(di|si) instruction pattern(s).
(define_insn "*tstdi_sign"
[(set (reg CC_REGNUM)
@@ -542,7 +542,7 @@
"lt<g>r\t%0,%0"
[(set_attr "op_type" "RR<E>")])
-; tst(hi|qi) intruction pattern(s).
+; tst(hi|qi) instruction pattern(s).
(define_insn "*tst<mode>CCT"
[(set (reg CC_REGNUM)
diff --git a/gcc/df-core.c b/gcc/df-core.c
index e9da8b6..06b1085 100644
--- a/gcc/df-core.c
+++ b/gcc/df-core.c
@@ -80,12 +80,12 @@ calls to add a problem for a given instance of df must occur before
the first call to DF_RESCAN_BLOCKS or DF_ANALYZE.
For all of the problems defined in df-problems.c, there are
-convienence functions named DF_*_ADD_PROBLEM.
+convenience functions named DF_*_ADD_PROBLEM.
Problems can be dependent on other problems. For instance, solving
def-use or use-def chains is dependant on solving reaching
-definitions. As long as these dependancies are listed in the problem
+definitions. As long as these dependencies are listed in the problem
definition, the order of adding the problems is not material.
Otherwise, the problems will be solved in the order of calls to
df_add_problem. Note that it is not necessary to have a problem. In
@@ -100,7 +100,7 @@ to analyze the entire function and no call to df_set_blocks is made.
When a subset is given, the analysis behaves as if the function only
contains those blocks and any edges that occur directly between the
blocks in the set. Care should be taken to call df_set_blocks right
-before the call to analyze in order to eliminate the possiblity that
+before the call to analyze in order to eliminate the possibility that
optimizations that reorder blocks invalidate the bitvector.
@@ -220,7 +220,7 @@ There are 4 ways to obtain access to refs:
register and are put there to keep the code from forgetting about
them.
- Artifical defs occur at the end of the entry block. These arise
+ Artificial defs occur at the end of the entry block. These arise
from registers that are live at entry to the function.
2) All of the uses and defs associated with each pseudo or hard
diff --git a/gcc/df-problems.c b/gcc/df-problems.c
index 257baad..a0422f8 100644
--- a/gcc/df-problems.c
+++ b/gcc/df-problems.c
@@ -184,7 +184,7 @@ df_get_live_out (struct df *df, basic_block bb)
----------------------------------------------------------------------------*/
/* Generic versions to get the void* version of the block info. Only
- used inside the problem instace vectors. */
+ used inside the problem instance vectors. */
/* Grow the bb_info array. */
@@ -2781,7 +2781,7 @@ df_chain_bb_reset (struct dataflow *dflow, unsigned int bb_index)
}
}
- /* Get rid of any chains in artifical uses or defs. */
+ /* Get rid of any chains in artificial uses or defs. */
if (problem_data->flags & DF_DU_CHAIN)
{
struct df_ref *def;
diff --git a/gcc/df-scan.c b/gcc/df-scan.c
index 934f98d..1ed153d 100644
--- a/gcc/df-scan.c
+++ b/gcc/df-scan.c
@@ -549,7 +549,7 @@ df_ref_create (struct df *df, rtx reg, rtx *loc, rtx insn,
----------------------------------------------------------------------------*/
-/* Get the artifical uses for a basic block. */
+/* Get the artificial uses for a basic block. */
struct df_ref *
df_get_artificial_defs (struct df *df, unsigned int bb_index)
@@ -559,7 +559,7 @@ df_get_artificial_defs (struct df *df, unsigned int bb_index)
}
-/* Get the artifical uses for a basic block. */
+/* Get the artificial uses for a basic block. */
struct df_ref *
df_get_artificial_uses (struct df *df, unsigned int bb_index)
@@ -797,7 +797,7 @@ df_bb_refs_delete (struct dataflow *dflow, int bb_index)
}
}
- /* Get rid of any artifical uses or defs. */
+ /* Get rid of any artificial uses or defs. */
if (bb_info)
{
def = bb_info->artificial_defs;
diff --git a/gcc/df.h b/gcc/df.h
index d6a76d8..720bbfa 100644
--- a/gcc/df.h
+++ b/gcc/df.h
@@ -271,7 +271,7 @@ struct df_ref_info
unsigned int bitmap_size; /* Number of refs seen. */
/* True if refs table is organized so that every reference for a
- pseudo is contigious. */
+ pseudo is contiguous. */
bool refs_organized;
/* True if the next refs should be added immediately or false to
defer to later to reorganize the table. */
diff --git a/gcc/diagnostic.c b/gcc/diagnostic.c
index fb81bbd..b671b1a 100644
--- a/gcc/diagnostic.c
+++ b/gcc/diagnostic.c
@@ -383,7 +383,7 @@ diagnostic_report_diagnostic (diagnostic_context *context,
option. */
if (context->classify_diagnostic[diagnostic->option_index] != DK_UNSPECIFIED)
diagnostic->kind = context->classify_diagnostic[diagnostic->option_index];
- /* This allows for future extenions, like temporarily disabling
+ /* This allows for future extensions, like temporarily disabling
warnings for ranges of source code. */
if (diagnostic->kind == DK_IGNORED)
return;
diff --git a/gcc/expr.c b/gcc/expr.c
index 0107f2a..e3cb784 100644
--- a/gcc/expr.c
+++ b/gcc/expr.c
@@ -6282,7 +6282,7 @@ expand_operands (tree exp0, tree exp1, rtx target, rtx *op0, rtx *op1,
}
-/* Return a MEM that constains constant EXP. DEFER is as for
+/* Return a MEM that contains constant EXP. DEFER is as for
output_constant_def and MODIFIER is as for expand_expr. */
static rtx
diff --git a/gcc/function.h b/gcc/function.h
index 727879e..a904e0f 100644
--- a/gcc/function.h
+++ b/gcc/function.h
@@ -386,7 +386,7 @@ struct function GTY(())
unsigned int calls_alloca : 1;
/* Nonzero if function being compiled called builtin_return_addr or
- builtin_frame_address with non-zero count. */
+ builtin_frame_address with nonzero count. */
unsigned int accesses_prior_frames : 1;
/* Nonzero if the function calls __builtin_eh_return. */
diff --git a/gcc/gimplify.c b/gcc/gimplify.c
index 10373c2..4b58201 100644
--- a/gcc/gimplify.c
+++ b/gcc/gimplify.c
@@ -404,8 +404,8 @@ find_single_pointer_decl (tree t)
if (walk_tree (&t, find_single_pointer_decl_1, &decl, NULL))
{
- /* find_single_pointer_decl_1 returns a non-zero value, causing
- walk_tree to return a non-zero value, to indicate that it
+ /* find_single_pointer_decl_1 returns a nonzero value, causing
+ walk_tree to return a nonzero value, to indicate that it
found more than one pointer DECL. */
return NULL_TREE;
}
@@ -4884,7 +4884,7 @@ gimplify_omp_atomic_fetch_op (tree *expr_p, tree addr, tree rhs, int index)
}
/* A subroutine of gimplify_omp_atomic_pipeline. Walk *EXPR_P and replace
- appearences of *LHS_ADDR with LHS_VAR. If an expression does not involve
+ appearances of *LHS_ADDR with LHS_VAR. If an expression does not involve
the lhs, evaluate it into a temporary. Return 1 if the lhs appeared as
a subexpression, 0 if it did not, or -1 if an error was encountered. */
diff --git a/gcc/loop-invariant.c b/gcc/loop-invariant.c
index 553e9fe..2185d18 100644
--- a/gcc/loop-invariant.c
+++ b/gcc/loop-invariant.c
@@ -936,7 +936,7 @@ get_inv_cost (struct invariant *inv, int *comp_cost, unsigned *regs_needed)
{
/* Hoisting constant pool constants into stack regs may cost more than
just single register. On x87, the balance is affected both by the
- small number of FP registers, and by its register stack organisation,
+ small number of FP registers, and by its register stack organization,
that forces us to add compensation code in and around the loop to
shuffle the operands to the top of stack before use, and pop them
from the stack after the loop finishes.
diff --git a/gcc/omp-low.c b/gcc/omp-low.c
index 74be1e8..0a9c96f 100644
--- a/gcc/omp-low.c
+++ b/gcc/omp-low.c
@@ -47,7 +47,7 @@ Software Foundation, 51 Franklin Street, Fifth Floor, Boston, MA
phases. The first phase scans the function looking for OMP statements
and then for variables that must be replaced to satisfy data sharing
clauses. The second phase expands code for the constructs, as well as
- re-gimplifing things when variables have been replaced with complex
+ re-gimplifying things when variables have been replaced with complex
expressions.
Final code generation is done by pass_expand_omp. The flowgraph is
@@ -2042,7 +2042,7 @@ expand_parallel_call (struct omp_region *region, basic_block bb, tree ws_args)
clauses = OMP_PARALLEL_CLAUSES (region->entry);
push_gimplify_context ();
- /* Determine what flavour of GOMP_parallel_start we will be
+ /* Determine what flavor of GOMP_parallel_start we will be
emitting. */
start_ix = BUILT_IN_GOMP_PARALLEL_START;
if (is_combined_parallel (region))
diff --git a/gcc/opts.c b/gcc/opts.c
index d804761..00b9716 100644
--- a/gcc/opts.c
+++ b/gcc/opts.c
@@ -1085,7 +1085,7 @@ common_handle_option (size_t scode, const char *arg, int value,
case OPT_floop_optimize:
case OPT_frerun_loop_opt:
case OPT_fstrength_reduce:
- /* These are no-ops, preserved for backward compatability. */
+ /* These are no-ops, preserved for backward compatibility. */
break;
default:
diff --git a/gcc/passes.c b/gcc/passes.c
index c3702ef..9732b25 100644
--- a/gcc/passes.c
+++ b/gcc/passes.c
@@ -364,7 +364,7 @@ register_dump_files_1 (struct tree_opt_pass *pass, bool ipa, int properties)
/* Register the dump files for the pipeline starting at PASS. IPA is
true if the pass is inter-procedural, and PROPERTIES reflects the
- properties that are guarenteed to be available at the beginning of
+ properties that are guaranteed to be available at the beginning of
the pipeline. */
static void
diff --git a/gcc/rtl-factoring.c b/gcc/rtl-factoring.c
index 0cd1aaa..021a802 100644
--- a/gcc/rtl-factoring.c
+++ b/gcc/rtl-factoring.c
@@ -125,7 +125,7 @@ Software Foundation, 51 Franklin Street, Fifth Floor, Boston, MA
TODO:
- Use REG_ALLOC_ORDER when choosing link register.
- Handle JUMP_INSNs. Also handle volatile function calls (handle them
- simmilar to unconditional jumps.)
+ similar to unconditional jumps.)
- Test command line option -fpic.
*/
@@ -227,7 +227,7 @@ typedef struct seq_block_def
struct seq_block_def *next_seq_block;
} *seq_block;
-/* Contains same sequence candidates for futher searching. */
+/* Contains same sequence candidates for further searching. */
typedef struct hash_bucket_def
{
/* The hash value of the group. */
@@ -888,7 +888,7 @@ determine_seq_blocks (void)
}
/* Ensure that SB contains a seq_block with the appropriate length.
- Insert a new seq_block if neccessary. */
+ Insert a new seq_block if necessary. */
if (!seq_blocks || ((*mseq)->abstracted_length < seq_blocks->length))
{
sb = (seq_block) xmalloc (sizeof (struct seq_block_def));
@@ -979,7 +979,7 @@ split_blocks_after_seqs (void)
}
}
-/* Splits the best pattern sequence accoring to SEQ_BLOCKS. Emits pseudo-call
+/* Splits the best pattern sequence according to SEQ_BLOCKS. Emits pseudo-call
and -return insns before and after the sequence. */
static void
@@ -1373,7 +1373,7 @@ rtl_seqabstr (void)
/* Iterate until there are no sequences to abstract. */
for (iter = 1;; iter++)
{
- /* Recompute gain for sequences if neccessary and select sequence with
+ /* Recompute gain for sequences if necessary and select sequence with
biggest gain. */
recompute_gain ();
if (!pattern_seqs)
diff --git a/gcc/rtlanal.c b/gcc/rtlanal.c
index dc3b5ad..d740ed5 100644
--- a/gcc/rtlanal.c
+++ b/gcc/rtlanal.c
@@ -2196,7 +2196,7 @@ may_trap_after_code_motion_p (rtx x)
return may_trap_p_1 (x, MTP_AFTER_MOVE);
}
-/* Same as above, but additionally return non-zero if evaluating rtx X might
+/* Same as above, but additionally return nonzero if evaluating rtx X might
cause a fault. We define a fault for the purpose of this function as a
erroneous execution condition that cannot be encountered during the normal
execution of a valid program; the typical example is an unaligned memory
diff --git a/gcc/struct-equiv.c b/gcc/struct-equiv.c
index 1f11ff8..265cfd0 100644
--- a/gcc/struct-equiv.c
+++ b/gcc/struct-equiv.c
@@ -483,7 +483,7 @@ rtx_equiv_p (rtx *xp, rtx y, int rvalue, struct equiv_info *info)
pair, use the old one. If the width is the same, use the
old one if the modes match, but the new if they don't.
We don't want to get too fancy with subreg_regno_offset
- here, so we just test two straightforwad cases each. */
+ here, so we just test two straightforward cases each. */
if (info->live_update
&& (x_mode != GET_MODE (info->x_local[i])
? size >= size_i : size > size_i))
diff --git a/gcc/tree-cfgcleanup.c b/gcc/tree-cfgcleanup.c
index 57315d4..fb1ea76 100644
--- a/gcc/tree-cfgcleanup.c
+++ b/gcc/tree-cfgcleanup.c
@@ -782,7 +782,7 @@ merge_phi_nodes (void)
break;
}
- /* If the loop above iterated thorugh all the PHI nodes
+ /* If the loop above iterated through all the PHI nodes
in BB, then we can merge the PHIs from BB into DEST. */
if (!phi)
*current++ = bb;
diff --git a/gcc/tree-ssa-loop-niter.c b/gcc/tree-ssa-loop-niter.c
index f913df3..cda02dd 100644
--- a/gcc/tree-ssa-loop-niter.c
+++ b/gcc/tree-ssa-loop-niter.c
@@ -623,7 +623,7 @@ number_of_iterations_cond (tree type, affine_iv *iv0, enum tree_code code,
return false;
}
- /* If the loop exits immediatelly, there is nothing to do. */
+ /* If the loop exits immediately, there is nothing to do. */
if (zero_p (fold_build2 (code, boolean_type_node, iv0->base, iv1->base)))
{
niter->niter = build_int_cst_type (unsigned_type_for (type), 0);
diff --git a/gcc/tree-ssa-loop-prefetch.c b/gcc/tree-ssa-loop-prefetch.c
index 911b389..c00aaea 100644
--- a/gcc/tree-ssa-loop-prefetch.c
+++ b/gcc/tree-ssa-loop-prefetch.c
@@ -115,7 +115,7 @@ Software Foundation, 59 Temple Place - Suite 330, Boston, MA
/* Magic constants follow. These should be replaced by machine specific
numbers. */
-/* A number that should rouhgly correspond to the number of instructions
+/* A number that should roughly correspond to the number of instructions
executed before the prefetch is completed. */
#ifndef PREFETCH_LATENCY
@@ -768,7 +768,7 @@ schedule_prefetches (struct mem_ref_group *groups, unsigned unroll_factor,
/* For now we just take memory references one by one and issue
prefetches for as many as possible. The groups are sorted
starting with the largest step, since the references with
- large step are more likely to cause many cache mises. */
+ large step are more likely to cause many cache misses. */
for (; groups; groups = groups->next)
for (ref = groups->refs; ref; ref = ref->next)
diff --git a/gcc/tree-ssa-structalias.c b/gcc/tree-ssa-structalias.c
index a84a32c..a9032a3 100644
--- a/gcc/tree-ssa-structalias.c
+++ b/gcc/tree-ssa-structalias.c
@@ -1040,7 +1040,7 @@ merge_graph_nodes (constraint_graph_t graph, unsigned int to,
graph->zero_weight_succs[from]);
}
- /* Merge all the non-zero weighted predecessor edges. */
+ /* Merge all the nonzero weighted predecessor edges. */
for (i = 0; VEC_iterate (constraint_edge_t, predvec, i, c); i++)
{
unsigned int d = c->dest;
@@ -1064,7 +1064,7 @@ merge_graph_nodes (constraint_graph_t graph, unsigned int to,
}
- /* Merge all the non-zero weighted successor edges. */
+ /* Merge all the nonzero weighted successor edges. */
for (i = 0; VEC_iterate (constraint_edge_t, succvec, i, c); i++)
{
unsigned int d = c->dest;
@@ -4049,7 +4049,7 @@ intra_create_variable_infos (void)
lhs.type = SCALAR;
lhs.var = create_variable_info_for (t, alias_get_name (t));
- /* With flag_argument_noalias greater than one means that the incomming
+ /* With flag_argument_noalias greater than one means that the incoming
argument cannot alias anything except for itself so create a HEAP
variable. */
if (POINTER_TYPE_P (TREE_TYPE (t))
diff --git a/gcc/tree-ssa-threadedge.c b/gcc/tree-ssa-threadedge.c
index cc98867..511eb6a 100644
--- a/gcc/tree-ssa-threadedge.c
+++ b/gcc/tree-ssa-threadedge.c
@@ -120,7 +120,7 @@ remove_temporary_equivalences (VEC(tree, heap) **stack)
dest = VEC_pop (tree, *stack);
- /* A NULL value indicates we should stop unwinding, oherwise
+ /* A NULL value indicates we should stop unwinding, otherwise
pop off the next entry as they're recorded in pairs. */
if (dest == NULL)
break;
diff --git a/gcc/tree-ssa-threadupdate.c b/gcc/tree-ssa-threadupdate.c
index d5777ff..4a92eff 100644
--- a/gcc/tree-ssa-threadupdate.c
+++ b/gcc/tree-ssa-threadupdate.c
@@ -811,7 +811,7 @@ thread_block (basic_block bb)
}
/* Walk through the registered jump threads and convert them into a
- form convienent for this pass.
+ form convenient for this pass.
Any block which has incoming edges threaded to outgoing edges
will have its entry in THREADED_BLOCK set.
diff --git a/gcc/tree-vect-patterns.c b/gcc/tree-vect-patterns.c
index 61d66e1..558908e 100644
--- a/gcc/tree-vect-patterns.c
+++ b/gcc/tree-vect-patterns.c
@@ -41,7 +41,7 @@ Software Foundation, 51 Franklin Street, Fifth Floor, Boston, MA
#include "recog.h"
#include "toplev.h"
-/* Funcion prototypes */
+/* Function prototypes */
static void vect_pattern_recog_1
(tree (* ) (tree, tree *, tree *), block_stmt_iterator);
static bool widened_name_p (tree, tree, tree *, tree *);
@@ -133,7 +133,7 @@ widened_name_p (tree name, tree use_stmt, tree *half_type, tree *def_stmt)
S7 sum_1 = prod + sum_0;
where 'TYPE1' is exactly double the size of type 'type', and 'TYPE2' is the
- same size of 'TYPE1' or bigger. This is a sepcial case of a reduction
+ same size of 'TYPE1' or bigger. This is a special case of a reduction
computation.
Input:
@@ -455,7 +455,7 @@ vect_recog_widen_sum_pattern (tree last_stmt, tree *type_in, tree *type_out)
If 'TYPE_OUT' is also returned by PATTERN_RECOG_FUNC, we check that it fits
to the available target pattern.
- This function also does some bookeeping, as explained in the documentation
+ This function also does some bookkeeping, as explained in the documentation
for vect_recog_pattern. */
static void
@@ -578,7 +578,7 @@ vect_pattern_recog_1 (
remain irrelevant unless used by stmts other than S4.
If vectorization succeeds, vect_transform_stmt will skip over {S1,S2,S3}
- (because they are marked as irrelevent). It will vectorize S6, and record
+ (because they are marked as irrelevant). It will vectorize S6, and record
a pointer to the new vector stmt VS6 both from S6 (as usual), and also
from S4. We do that so that when we get to vectorizing stmts that use the
def of S4 (like S5 that uses a_0), we'll know where to take the relevant
diff --git a/gcc/tree-vect-transform.c b/gcc/tree-vect-transform.c
index 5f2bce3..2d533e6 100644
--- a/gcc/tree-vect-transform.c
+++ b/gcc/tree-vect-transform.c
@@ -857,7 +857,7 @@ vect_create_epilog_for_reduction (tree vect_def, tree stmt,
exit_bsi = bsi_start (exit_bb);
/* 2.2 Get the relevant tree-code to use in the epilog for schemes 2,3
- (i.e. when reduc_code is not available) and in the final adjusment code
+ (i.e. when reduc_code is not available) and in the final adjustment code
(if needed). Also get the original scalar reduction variable as
defined in the loop. In case STMT is a "pattern-stmt" (i.e. - it
represents a reduction pattern), the tree-code and scalar-def are
@@ -2945,7 +2945,7 @@ vect_create_cond_for_align_checks (loop_vec_info loop_vinfo,
append_to_statement_list_force (and_stmt, cond_expr_stmt_list);
/* Make and_tmp the left operand of the conditional test against zero.
- if and_tmp has a non-zero bit then some address is unaligned. */
+ if and_tmp has a nonzero bit then some address is unaligned. */
ptrsize_zero = build_int_cst (int_ptrsize_type, 0);
return build2 (EQ_EXPR, boolean_type_node,
and_tmp_name, ptrsize_zero);
diff --git a/gcc/tree-vectorizer.h b/gcc/tree-vectorizer.h
index ef455d9..405289c 100644
--- a/gcc/tree-vectorizer.h
+++ b/gcc/tree-vectorizer.h
@@ -208,7 +208,7 @@ typedef struct _stmt_vec_info {
/* Stmt is part of some pattern (computation idiom) */
bool in_pattern_p;
- /* Used for various bookeeping purposes, generally holding a pointer to
+ /* Used for various bookkeeping purposes, generally holding a pointer to
some other stmt S that is in some way "related" to this stmt.
Current use of this field is:
If this stmt is part of a pattern (i.e. the field 'in_pattern_p' is
diff --git a/gcc/tree-vrp.c b/gcc/tree-vrp.c
index f6d0a6d..7342d29 100644
--- a/gcc/tree-vrp.c
+++ b/gcc/tree-vrp.c
@@ -1058,7 +1058,7 @@ extract_range_from_assert (value_range_t *vr_p, tree expr)
there are three cases to consider.
- 1. The VR_ANTI_RANGE range is competely within the
+ 1. The VR_ANTI_RANGE range is completely within the
VR_RANGE and the endpoints of the ranges are
different. In that case the resulting range
should be whichever range is more precise.
diff --git a/gcc/unwind-dw2.c b/gcc/unwind-dw2.c
index efae54a..55a6143 100644
--- a/gcc/unwind-dw2.c
+++ b/gcc/unwind-dw2.c
@@ -228,7 +228,7 @@ _Unwind_SetGRValue (struct _Unwind_Context *context, int index,
context->reg[index] = (void *) (_Unwind_Internal_Ptr) val;
}
-/* Return non-zero if register INDEX is stored by value rather than
+/* Return nonzero if register INDEX is stored by value rather than
by reference. */
static inline int