You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@lucene.apache.org by tf...@apache.org on 2019/02/13 05:35:37 UTC

[lucene-solr] branch master updated: Fix some spell check issues

This is an automated email from the ASF dual-hosted git repository.

tflobbe pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/lucene-solr.git


The following commit(s) were added to refs/heads/master by this push:
     new 5310152  Fix some spell check issues
5310152 is described below

commit 531015245042507d71845b6d584e7e7389303093
Author: Bruno P. Kinoshita <br...@yahoo.com.br>
AuthorDate: Mon Oct 30 23:57:52 2017 +1300

    Fix some spell check issues
---
 .../org/apache/lucene/analysis/charfilter/HTMLStripCharFilter.java  | 2 +-
 .../apache/lucene/analysis/miscellaneous/WordDelimiterIterator.java | 2 +-
 .../org/apache/lucene/analysis/pattern/PatternReplaceFilter.java    | 4 ++--
 .../apache/lucene/analysis/pattern/SimplePatternSplitTokenizer.java | 2 +-
 .../org/apache/lucene/analysis/wikipedia/WikipediaTokenizer.java    | 6 +++---
 .../org/apache/lucene/benchmark/byTask/tasks/NewAnalyzerTask.java   | 2 +-
 .../java/org/apache/lucene/benchmark/quality/trec/TrecJudge.java    | 2 +-
 .../src/java/org/apache/lucene/classification/BM25NBClassifier.java | 2 +-
 .../classification/document/SimpleNaiveBayesDocumentClassifier.java | 2 +-
 .../org/apache/lucene/classification/utils/NearestFuzzyQuery.java   | 2 +-
 .../src/java/org/apache/lucene/codecs/PushPostingsWriterBase.java   | 2 +-
 .../src/java/org/apache/lucene/codecs/compressing/Compressor.java   | 2 +-
 .../org/apache/lucene/codecs/lucene60/Lucene60PointsFormat.java     | 2 +-
 .../apache/lucene/codecs/lucene70/Lucene70SegmentInfoFormat.java    | 2 +-
 .../src/java/org/apache/lucene/geo/SimpleGeoJSONPolygonParser.java  | 2 +-
 .../src/java/org/apache/lucene/index/ConcurrentMergeScheduler.java  | 2 +-
 .../src/java/org/apache/lucene/index/EmptyDocValuesProducer.java    | 2 +-
 lucene/core/src/java/org/apache/lucene/index/IndexWriterConfig.java | 2 +-
 .../core/src/java/org/apache/lucene/index/StoredFieldVisitor.java   | 2 +-
 .../org/apache/lucene/search/ControlledRealTimeReopenThread.java    | 2 +-
 .../src/java/org/apache/lucene/search/DisjunctionMaxScorer.java     | 2 +-
 .../src/java/org/apache/lucene/search/SearcherLifetimeManager.java  | 2 +-
 lucene/core/src/java/org/apache/lucene/search/SearcherManager.java  | 2 +-
 lucene/core/src/java/org/apache/lucene/store/FSDirectory.java       | 4 ++--
 lucene/core/src/java/org/apache/lucene/store/IndexInput.java        | 2 +-
 .../core/src/java/org/apache/lucene/store/NativeFSLockFactory.java  | 2 +-
 lucene/core/src/java/org/apache/lucene/util/ByteBlockPool.java      | 2 +-
 .../java/org/apache/lucene/util/automaton/CompiledAutomaton.java    | 2 +-
 .../org/apache/lucene/util/automaton/MinimizationOperations.java    | 2 +-
 lucene/core/src/java/org/apache/lucene/util/automaton/RegExp.java   | 4 ++--
 lucene/core/src/java/org/apache/lucene/util/fst/Util.java           | 2 +-
 lucene/facet/src/java/org/apache/lucene/facet/DrillDownQuery.java   | 2 +-
 .../src/java/org/apache/lucene/facet/LongValueFacetCounts.java      | 2 +-
 .../apache/lucene/facet/taxonomy/writercache/LabelToOrdinal.java    | 4 ++--
 .../lucene/facet/taxonomy/writercache/NameHashIntCacheLRU.java      | 2 +-
 .../apache/lucene/facet/taxonomy/writercache/NameIntCacheLRU.java   | 2 +-
 lucene/misc/src/java/org/apache/lucene/document/LazyDocument.java   | 2 +-
 .../misc/src/java/org/apache/lucene/misc/SweetSpotSimilarity.java   | 2 +-
 .../src/java/org/apache/lucene/queryparser/classic/CharStream.java  | 2 +-
 .../flexible/standard/builders/SynonymQueryNodeBuilder.java         | 2 +-
 .../lucene/queryparser/flexible/standard/nodes/PointQueryNode.java  | 2 +-
 .../lucene/queryparser/flexible/standard/parser/CharStream.java     | 2 +-
 .../org/apache/lucene/queryparser/surround/parser/CharStream.java   | 2 +-
 .../src/java/org/apache/lucene/replicator/ReplicationClient.java    | 2 +-
 .../replicator/src/java/org/apache/lucene/replicator/Revision.java  | 4 ++--
 .../src/java/org/apache/lucene/replicator/nrt/ReplicaNode.java      | 4 ++--
 .../apache/lucene/codecs/idversion/IDVersionSegmentTermsEnum.java   | 2 +-
 .../lucene/search/suggest/analyzing/AnalyzingInfixSuggester.java    | 2 +-
 .../lucene/search/suggest/jaspell/JaspellTernarySearchTrie.java     | 6 +++---
 .../src/test/org/apache/lucene/search/suggest/Top50KWiki.utf8       | 2 +-
 .../apache/solr/handler/dataimport/TestContentStreamDataSource.java | 2 +-
 solr/core/src/java/org/apache/solr/core/CoreContainer.java          | 2 +-
 .../java/org/apache/solr/handler/AnalysisRequestHandlerBase.java    | 4 ++--
 .../apache/solr/update/processor/DistributedUpdateProcessor.java    | 2 +-
 solr/solr-ref-guide/src/highlighting.adoc                           | 2 +-
 .../src/java/org/apache/solr/client/solrj/impl/HttpSolrClient.java  | 4 ++--
 .../solrj/src/java/org/apache/solr/common/ToleratedUpdateError.java | 2 +-
 57 files changed, 69 insertions(+), 69 deletions(-)

diff --git a/lucene/analysis/common/src/java/org/apache/lucene/analysis/charfilter/HTMLStripCharFilter.java b/lucene/analysis/common/src/java/org/apache/lucene/analysis/charfilter/HTMLStripCharFilter.java
index ae67bde..ac42024 100644
--- a/lucene/analysis/common/src/java/org/apache/lucene/analysis/charfilter/HTMLStripCharFilter.java
+++ b/lucene/analysis/common/src/java/org/apache/lucene/analysis/charfilter/HTMLStripCharFilter.java
@@ -30256,7 +30256,7 @@ public final class HTMLStripCharFilter extends BaseCharFilter {
 
 
   /**
-   * Reports an error that occured while scanning.
+   * Reports an error that occurred while scanning.
    *
    * In a wellformed scanner (no or only correct usage of 
    * yypushback(int) and a match-all fallback rule) this method 
diff --git a/lucene/analysis/common/src/java/org/apache/lucene/analysis/miscellaneous/WordDelimiterIterator.java b/lucene/analysis/common/src/java/org/apache/lucene/analysis/miscellaneous/WordDelimiterIterator.java
index 86b983d..f3541ac 100644
--- a/lucene/analysis/common/src/java/org/apache/lucene/analysis/miscellaneous/WordDelimiterIterator.java
+++ b/lucene/analysis/common/src/java/org/apache/lucene/analysis/miscellaneous/WordDelimiterIterator.java
@@ -256,7 +256,7 @@ public final class WordDelimiterIterator {
    * Determines if the text at the given position indicates an English possessive which should be removed
    *
    * @param pos Position in the text to check if it indicates an English possessive
-   * @return {@code true} if the text at the position indicates an English posessive, {@code false} otherwise
+   * @return {@code true} if the text at the position indicates an English possessive, {@code false} otherwise
    */
   private boolean endsWithPossessive(int pos) {
     return (stemEnglishPossessive &&
diff --git a/lucene/analysis/common/src/java/org/apache/lucene/analysis/pattern/PatternReplaceFilter.java b/lucene/analysis/common/src/java/org/apache/lucene/analysis/pattern/PatternReplaceFilter.java
index 5e3ff11..4c4338f 100644
--- a/lucene/analysis/common/src/java/org/apache/lucene/analysis/pattern/PatternReplaceFilter.java
+++ b/lucene/analysis/common/src/java/org/apache/lucene/analysis/pattern/PatternReplaceFilter.java
@@ -26,7 +26,7 @@ import java.io.IOException;
 
 /**
  * A TokenFilter which applies a Pattern to each token in the stream,
- * replacing match occurances with the specified replacement string.
+ * replacing match occurrences with the specified replacement string.
  *
  * <p>
  * <b>Note:</b> Depending on the input and the pattern used and the input
@@ -43,7 +43,7 @@ public final class PatternReplaceFilter extends TokenFilter {
   private final Matcher m;
 
   /**
-   * Constructs an instance to replace either the first, or all occurances
+   * Constructs an instance to replace either the first, or all occurrences
    *
    * @param in the TokenStream to process
    * @param p the patterm to apply to each Token
diff --git a/lucene/analysis/common/src/java/org/apache/lucene/analysis/pattern/SimplePatternSplitTokenizer.java b/lucene/analysis/common/src/java/org/apache/lucene/analysis/pattern/SimplePatternSplitTokenizer.java
index a8a40b2..f5a8fa2 100644
--- a/lucene/analysis/common/src/java/org/apache/lucene/analysis/pattern/SimplePatternSplitTokenizer.java
+++ b/lucene/analysis/common/src/java/org/apache/lucene/analysis/pattern/SimplePatternSplitTokenizer.java
@@ -31,7 +31,7 @@ import org.apache.lucene.util.automaton.RegExp;
 /**
  * This tokenizer uses a Lucene {@link RegExp} or (expert usage) a pre-built determinized {@link Automaton}, to locate tokens.
  * The regexp syntax is more limited than {@link PatternTokenizer}, but the tokenization is quite a bit faster.  This is just
- * like {@link SimplePatternTokenizer} except that the pattern shold make valid token separator characters, like
+ * like {@link SimplePatternTokenizer} except that the pattern should make valid token separator characters, like
  * {@code String.split}.  Empty string tokens are never produced.
  *
  * @lucene.experimental
diff --git a/lucene/analysis/common/src/java/org/apache/lucene/analysis/wikipedia/WikipediaTokenizer.java b/lucene/analysis/common/src/java/org/apache/lucene/analysis/wikipedia/WikipediaTokenizer.java
index 4b662e9..a4abce5 100644
--- a/lucene/analysis/common/src/java/org/apache/lucene/analysis/wikipedia/WikipediaTokenizer.java
+++ b/lucene/analysis/common/src/java/org/apache/lucene/analysis/wikipedia/WikipediaTokenizer.java
@@ -131,7 +131,7 @@ public final class WikipediaTokenizer extends Tokenizer {
 
   /**
    * Creates a new instance of the {@link org.apache.lucene.analysis.wikipedia.WikipediaTokenizer}.  Attaches the
-   * <code>input</code> to a the newly created JFlex scanner.
+   * <code>input</code> to the newly created JFlex scanner.
    *
    * @param tokenOutput One of {@link #TOKENS_ONLY}, {@link #UNTOKENIZED_ONLY}, {@link #BOTH}
    */
@@ -142,7 +142,7 @@ public final class WikipediaTokenizer extends Tokenizer {
 
   /**
    * Creates a new instance of the {@link org.apache.lucene.analysis.wikipedia.WikipediaTokenizer}.  Attaches the
-   * <code>input</code> to a the newly created JFlex scanner. Uses the given {@link org.apache.lucene.util.AttributeFactory}.
+   * <code>input</code> to the newly created JFlex scanner. Uses the given {@link org.apache.lucene.util.AttributeFactory}.
    *
    * @param tokenOutput One of {@link #TOKENS_ONLY}, {@link #UNTOKENIZED_ONLY}, {@link #BOTH}
    */
@@ -314,4 +314,4 @@ public final class WikipediaTokenizer extends Tokenizer {
     final int finalOffset = correctOffset(scanner.yychar() + scanner.yylength());
     this.offsetAtt.setOffset(finalOffset, finalOffset);
   }
-}
\ No newline at end of file
+}
diff --git a/lucene/benchmark/src/java/org/apache/lucene/benchmark/byTask/tasks/NewAnalyzerTask.java b/lucene/benchmark/src/java/org/apache/lucene/benchmark/byTask/tasks/NewAnalyzerTask.java
index b008e70..1673d34 100644
--- a/lucene/benchmark/src/java/org/apache/lucene/benchmark/byTask/tasks/NewAnalyzerTask.java
+++ b/lucene/benchmark/src/java/org/apache/lucene/benchmark/byTask/tasks/NewAnalyzerTask.java
@@ -28,7 +28,7 @@ import java.util.*;
 import java.lang.reflect.Constructor;
 
 /**
- * Create a new {@link org.apache.lucene.analysis.Analyzer} and set it it in the getRunData() for use by all future tasks.
+ * Create a new {@link org.apache.lucene.analysis.Analyzer} and set it in the getRunData() for use by all future tasks.
  *
  */
 public class NewAnalyzerTask extends PerfTask {
diff --git a/lucene/benchmark/src/java/org/apache/lucene/benchmark/quality/trec/TrecJudge.java b/lucene/benchmark/src/java/org/apache/lucene/benchmark/quality/trec/TrecJudge.java
index 156b0d5..377fa35 100644
--- a/lucene/benchmark/src/java/org/apache/lucene/benchmark/quality/trec/TrecJudge.java
+++ b/lucene/benchmark/src/java/org/apache/lucene/benchmark/quality/trec/TrecJudge.java
@@ -45,7 +45,7 @@ public class TrecJudge implements Judge {
    *     19    0   doc303       1
    *     19    0   doc7295      0
    * </pre> 
-   * @param reader where judgments are read from.
+   * @param reader where judgements are read from.
    * @throws IOException If there is a low-level I/O error.
    */
   public TrecJudge (BufferedReader reader) throws IOException {
diff --git a/lucene/classification/src/java/org/apache/lucene/classification/BM25NBClassifier.java b/lucene/classification/src/java/org/apache/lucene/classification/BM25NBClassifier.java
index 9e52a94..d22f612 100644
--- a/lucene/classification/src/java/org/apache/lucene/classification/BM25NBClassifier.java
+++ b/lucene/classification/src/java/org/apache/lucene/classification/BM25NBClassifier.java
@@ -85,7 +85,7 @@ public class BM25NBClassifier implements Classifier<BytesRef> {
    * @param analyzer       an {@link Analyzer} used to analyze unseen text
    * @param query          a {@link Query} to eventually filter the docs used for training the classifier, or {@code null}
    *                       if all the indexed docs should be used
-   * @param classFieldName the name of the field used as the output for the classifier NOTE: must not be havely analyzed
+   * @param classFieldName the name of the field used as the output for the classifier NOTE: must not be heavely analyzed
    *                       as the returned class will be a token indexed for this field
    * @param textFieldNames the name of the fields used as the inputs for the classifier, NO boosting supported per field
    */
diff --git a/lucene/classification/src/java/org/apache/lucene/classification/document/SimpleNaiveBayesDocumentClassifier.java b/lucene/classification/src/java/org/apache/lucene/classification/document/SimpleNaiveBayesDocumentClassifier.java
index 9a67da9..0a60697 100644
--- a/lucene/classification/src/java/org/apache/lucene/classification/document/SimpleNaiveBayesDocumentClassifier.java
+++ b/lucene/classification/src/java/org/apache/lucene/classification/document/SimpleNaiveBayesDocumentClassifier.java
@@ -62,7 +62,7 @@ public class SimpleNaiveBayesDocumentClassifier extends SimpleNaiveBayesClassifi
    * @param indexReader     the reader on the index to be used for classification
    * @param query          a {@link org.apache.lucene.search.Query} to eventually filter the docs used for training the classifier, or {@code null}
    *                       if all the indexed docs should be used
-   * @param classFieldName the name of the field used as the output for the classifier NOTE: must not be havely analyzed
+   * @param classFieldName the name of the field used as the output for the classifier NOTE: must not be heavely analyzed
    *                       as the returned class will be a token indexed for this field
    * @param textFieldNames the name of the fields used as the inputs for the classifier, they can contain boosting indication e.g. title^10
    */
diff --git a/lucene/classification/src/java/org/apache/lucene/classification/utils/NearestFuzzyQuery.java b/lucene/classification/src/java/org/apache/lucene/classification/utils/NearestFuzzyQuery.java
index d286eb9..ad15372 100644
--- a/lucene/classification/src/java/org/apache/lucene/classification/utils/NearestFuzzyQuery.java
+++ b/lucene/classification/src/java/org/apache/lucene/classification/utils/NearestFuzzyQuery.java
@@ -62,7 +62,7 @@ public class NearestFuzzyQuery extends Query {
   /**
    * Default constructor
    *
-   * @param analyzer the analyzer used to proecss the query text
+   * @param analyzer the analyzer used to process the query text
    */
   public NearestFuzzyQuery(Analyzer analyzer) {
     this.analyzer = analyzer;
diff --git a/lucene/core/src/java/org/apache/lucene/codecs/PushPostingsWriterBase.java b/lucene/core/src/java/org/apache/lucene/codecs/PushPostingsWriterBase.java
index 98bdd91..f977086 100644
--- a/lucene/core/src/java/org/apache/lucene/codecs/PushPostingsWriterBase.java
+++ b/lucene/core/src/java/org/apache/lucene/codecs/PushPostingsWriterBase.java
@@ -30,7 +30,7 @@ import org.apache.lucene.util.FixedBitSet;
 /**
  * Extension of {@link PostingsWriterBase}, adding a push
  * API for writing each element of the postings.  This API
- * is somewhat analagous to an XML SAX API, while {@link
+ * is somewhat analogous to an XML SAX API, while {@link
  * PostingsWriterBase} is more like an XML DOM API.
  * 
  * @see PostingsReaderBase
diff --git a/lucene/core/src/java/org/apache/lucene/codecs/compressing/Compressor.java b/lucene/core/src/java/org/apache/lucene/codecs/compressing/Compressor.java
index f95246c..76b7aed 100644
--- a/lucene/core/src/java/org/apache/lucene/codecs/compressing/Compressor.java
+++ b/lucene/core/src/java/org/apache/lucene/codecs/compressing/Compressor.java
@@ -31,7 +31,7 @@ public abstract class Compressor implements Closeable {
   protected Compressor() {}
 
   /**
-   * Compress bytes into <code>out</code>. It it the responsibility of the
+   * Compress bytes into <code>out</code>. It is the responsibility of the
    * compressor to add all necessary information so that a {@link Decompressor}
    * will know when to stop decompressing bytes from the stream.
    */
diff --git a/lucene/core/src/java/org/apache/lucene/codecs/lucene60/Lucene60PointsFormat.java b/lucene/core/src/java/org/apache/lucene/codecs/lucene60/Lucene60PointsFormat.java
index 1d2285c..496dac0 100644
--- a/lucene/core/src/java/org/apache/lucene/codecs/lucene60/Lucene60PointsFormat.java
+++ b/lucene/core/src/java/org/apache/lucene/codecs/lucene60/Lucene60PointsFormat.java
@@ -28,7 +28,7 @@ import org.apache.lucene.index.SegmentWriteState;
 
 /**
  * Lucene 6.0 point format, which encodes dimensional values in a block KD-tree structure
- * for fast 1D range and N dimesional shape intersection filtering.
+ * for fast 1D range and N dimensional shape intersection filtering.
  * See <a href="https://www.cs.duke.edu/~pankaj/publications/papers/bkd-sstd.pdf">this paper</a> for details.
  *
  * <p>This data structure is written as a series of blocks on disk, with an in-memory perfectly balanced
diff --git a/lucene/core/src/java/org/apache/lucene/codecs/lucene70/Lucene70SegmentInfoFormat.java b/lucene/core/src/java/org/apache/lucene/codecs/lucene70/Lucene70SegmentInfoFormat.java
index bd2bf06..2a3b44c 100644
--- a/lucene/core/src/java/org/apache/lucene/codecs/lucene70/Lucene70SegmentInfoFormat.java
+++ b/lucene/core/src/java/org/apache/lucene/codecs/lucene70/Lucene70SegmentInfoFormat.java
@@ -58,7 +58,7 @@ import org.apache.lucene.util.Version;
  *   <li>IsCompoundFile --&gt; {@link DataOutput#writeByte Int8}</li>
  *   <li>IndexSort --&gt; {@link DataOutput#writeVInt Int32} count, followed by {@code count} SortField</li>
  *   <li>SortField --&gt; {@link DataOutput#writeString String} field name, followed by {@link DataOutput#writeVInt Int32} sort type ID,
- *       followed by {@link DataOutput#writeByte Int8} indicatating reversed sort, followed by a type-specific encoding of the optional missing value
+ *       followed by {@link DataOutput#writeByte Int8} indicating reversed sort, followed by a type-specific encoding of the optional missing value
  *   <li>Footer --&gt; {@link CodecUtil#writeFooter CodecFooter}</li>
  * </ul>
  * Field Descriptions:
diff --git a/lucene/core/src/java/org/apache/lucene/geo/SimpleGeoJSONPolygonParser.java b/lucene/core/src/java/org/apache/lucene/geo/SimpleGeoJSONPolygonParser.java
index 278307f..1e35a3d 100644
--- a/lucene/core/src/java/org/apache/lucene/geo/SimpleGeoJSONPolygonParser.java
+++ b/lucene/core/src/java/org/apache/lucene/geo/SimpleGeoJSONPolygonParser.java
@@ -44,7 +44,7 @@ import java.util.List;
    type: MultiPolygon (union of polygons) is also accepted.
 */
 
-/** Does minimal parsing of a GeoJSON object, to extract either Polygon or MultiPolygon, either directly as a the top-level type, or if
+/** Does minimal parsing of a GeoJSON object, to extract either Polygon or MultiPolygon, either directly as the top-level type, or if
  *  the top-level type is Feature, as the geometry of that feature. */
 
 @SuppressWarnings("unchecked")
diff --git a/lucene/core/src/java/org/apache/lucene/index/ConcurrentMergeScheduler.java b/lucene/core/src/java/org/apache/lucene/index/ConcurrentMergeScheduler.java
index eef146a..f20ae31 100644
--- a/lucene/core/src/java/org/apache/lucene/index/ConcurrentMergeScheduler.java
+++ b/lucene/core/src/java/org/apache/lucene/index/ConcurrentMergeScheduler.java
@@ -47,7 +47,7 @@ import org.apache.lucene.util.ThreadInterruptedException;
  *
  *  <p>If more than {@link #getMaxMergeCount} merges are
  *  requested then this class will forcefully throttle the
- *  incoming threads by pausing until one more more merges
+ *  incoming threads by pausing until one more merges
  *  complete.</p>
  *
  *  <p>This class attempts to detect whether the index is
diff --git a/lucene/core/src/java/org/apache/lucene/index/EmptyDocValuesProducer.java b/lucene/core/src/java/org/apache/lucene/index/EmptyDocValuesProducer.java
index 4e9f0e0..b8ccfcf 100644
--- a/lucene/core/src/java/org/apache/lucene/index/EmptyDocValuesProducer.java
+++ b/lucene/core/src/java/org/apache/lucene/index/EmptyDocValuesProducer.java
@@ -21,7 +21,7 @@ import java.io.IOException;
 
 import org.apache.lucene.codecs.DocValuesProducer;
 
-/** Abstrast base class implementing a {@link DocValuesProducer} that has no doc values. */
+/** Abstract base class implementing a {@link DocValuesProducer} that has no doc values. */
 public abstract class EmptyDocValuesProducer extends DocValuesProducer {
 
   /** Sole constructor */
diff --git a/lucene/core/src/java/org/apache/lucene/index/IndexWriterConfig.java b/lucene/core/src/java/org/apache/lucene/index/IndexWriterConfig.java
index a72aeab..2d9b5b1 100644
--- a/lucene/core/src/java/org/apache/lucene/index/IndexWriterConfig.java
+++ b/lucene/core/src/java/org/apache/lucene/index/IndexWriterConfig.java
@@ -427,7 +427,7 @@ public final class IndexWriterConfig extends LiveIndexWriterConfig {
    * Information about merges, deletes and a
    * message when maxFieldLength is reached will be printed
    * to this. Must not be null, but {@link InfoStream#NO_OUTPUT} 
-   * may be used to supress output.
+   * may be used to suppress output.
    */
   public IndexWriterConfig setInfoStream(InfoStream infoStream) {
     if (infoStream == null) {
diff --git a/lucene/core/src/java/org/apache/lucene/index/StoredFieldVisitor.java b/lucene/core/src/java/org/apache/lucene/index/StoredFieldVisitor.java
index 43782fe..911a157 100644
--- a/lucene/core/src/java/org/apache/lucene/index/StoredFieldVisitor.java
+++ b/lucene/core/src/java/org/apache/lucene/index/StoredFieldVisitor.java
@@ -30,7 +30,7 @@ import org.apache.lucene.document.DocumentStoredFieldVisitor;
  * <p><b>NOTE</b>: a {@code StoredFieldVisitor} implementation
  * should not try to load or visit other stored documents in
  * the same reader because the implementation of stored
- * fields for most codecs is not reeentrant and you will see
+ * fields for most codecs is not reentrant and you will see
  * strange exceptions as a result.
  *
  * <p>See {@link DocumentStoredFieldVisitor}, which is a
diff --git a/lucene/core/src/java/org/apache/lucene/search/ControlledRealTimeReopenThread.java b/lucene/core/src/java/org/apache/lucene/search/ControlledRealTimeReopenThread.java
index a98a30d..8499783 100644
--- a/lucene/core/src/java/org/apache/lucene/search/ControlledRealTimeReopenThread.java
+++ b/lucene/core/src/java/org/apache/lucene/search/ControlledRealTimeReopenThread.java
@@ -50,7 +50,7 @@ public class ControlledRealTimeReopenThread<T> extends Thread implements Closeab
   
   /**
    * Create ControlledRealTimeReopenThread, to periodically
-   * reopen the a {@link ReferenceManager}.
+   * reopen the {@link ReferenceManager}.
    *
    * @param targetMaxStaleSec Maximum time until a new
    *        reader must be opened; this sets the upper bound
diff --git a/lucene/core/src/java/org/apache/lucene/search/DisjunctionMaxScorer.java b/lucene/core/src/java/org/apache/lucene/search/DisjunctionMaxScorer.java
index dd04d4c..82dbda0 100644
--- a/lucene/core/src/java/org/apache/lucene/search/DisjunctionMaxScorer.java
+++ b/lucene/core/src/java/org/apache/lucene/search/DisjunctionMaxScorer.java
@@ -24,7 +24,7 @@ import org.apache.lucene.util.MathUtil;
 import static org.apache.lucene.search.DocIdSetIterator.NO_MORE_DOCS;
 
 /**
- * The Scorer for DisjunctionMaxQuery.  The union of all documents generated by the the subquery scorers
+ * The Scorer for DisjunctionMaxQuery.  The union of all documents generated by the subquery scorers
  * is generated in document number order.  The score for each document is the maximum of the scores computed
  * by the subquery scorers that generate that document, plus tieBreakerMultiplier times the sum of the scores
  * for the other subqueries that generate the document.
diff --git a/lucene/core/src/java/org/apache/lucene/search/SearcherLifetimeManager.java b/lucene/core/src/java/org/apache/lucene/search/SearcherLifetimeManager.java
index 0146674..bf84831 100644
--- a/lucene/core/src/java/org/apache/lucene/search/SearcherLifetimeManager.java
+++ b/lucene/core/src/java/org/apache/lucene/search/SearcherLifetimeManager.java
@@ -238,7 +238,7 @@ public class SearcherLifetimeManager implements Closeable {
    *  entries are passed to the Pruner in sorted (newest to
    *  oldest IndexSearcher) order.
    * 
-   *  <p><b>NOTE</b>: you must peridiocally call this, ideally
+   *  <p><b>NOTE</b>: you must periodically call this, ideally
    *  from the same background thread that opens new
    *  searchers. */
   public synchronized void prune(Pruner pruner) throws IOException {
diff --git a/lucene/core/src/java/org/apache/lucene/search/SearcherManager.java b/lucene/core/src/java/org/apache/lucene/search/SearcherManager.java
index 2ebb0a4..70c297f 100644
--- a/lucene/core/src/java/org/apache/lucene/search/SearcherManager.java
+++ b/lucene/core/src/java/org/apache/lucene/search/SearcherManager.java
@@ -172,7 +172,7 @@ public final class SearcherManager extends ReferenceManager<IndexSearcher> {
   }
 
   /**
-   * Returns <code>true</code> if no changes have occured since this searcher
+   * Returns <code>true</code> if no changes have occurred since this searcher
    * ie. reader was opened, otherwise <code>false</code>.
    * @see DirectoryReader#isCurrent() 
    */
diff --git a/lucene/core/src/java/org/apache/lucene/store/FSDirectory.java b/lucene/core/src/java/org/apache/lucene/store/FSDirectory.java
index fb03f85..615c089 100644
--- a/lucene/core/src/java/org/apache/lucene/store/FSDirectory.java
+++ b/lucene/core/src/java/org/apache/lucene/store/FSDirectory.java
@@ -144,7 +144,7 @@ public abstract class FSDirectory extends BaseDirectory {
    * real path to ensure it can correctly lock the index directory and no other process
    * can interfere with changing possible symlinks to the index directory inbetween.
    * If you want to use symlinks and change them dynamically, close all
-   * {@code IndexWriters} and create a new {@code FSDirecory} instance.
+   * {@code IndexWriters} and create a new {@code FSDirectory} instance.
    * @param path the path of the directory
    * @param lockFactory the lock factory to use, or null for the default
    * ({@link NativeFSLockFactory});
@@ -168,7 +168,7 @@ public abstract class FSDirectory extends BaseDirectory {
    * real path to ensure it can correctly lock the index directory and no other process
    * can interfere with changing possible symlinks to the index directory inbetween.
    * If you want to use symlinks and change them dynamically, close all
-   * {@code IndexWriters} and create a new {@code FSDirecory} instance.
+   * {@code IndexWriters} and create a new {@code FSDirectory} instance.
    *
    *  <p>Currently this returns {@link MMapDirectory} for Linux, MacOSX, Solaris,
    *  and Windows 64-bit JREs, {@link NIOFSDirectory} for other
diff --git a/lucene/core/src/java/org/apache/lucene/store/IndexInput.java b/lucene/core/src/java/org/apache/lucene/store/IndexInput.java
index 4e36fe4..ebc5cfe 100644
--- a/lucene/core/src/java/org/apache/lucene/store/IndexInput.java
+++ b/lucene/core/src/java/org/apache/lucene/store/IndexInput.java
@@ -98,7 +98,7 @@ public abstract class IndexInput extends DataInput implements Cloneable,Closeabl
   
   /**
    * Creates a slice of this index input, with the given description, offset, and length. 
-   * The slice is seeked to the beginning.
+   * The slice is sought to the beginning.
    */
   public abstract IndexInput slice(String sliceDescription, long offset, long length) throws IOException;
 
diff --git a/lucene/core/src/java/org/apache/lucene/store/NativeFSLockFactory.java b/lucene/core/src/java/org/apache/lucene/store/NativeFSLockFactory.java
index 106b02e..c61e4a3 100644
--- a/lucene/core/src/java/org/apache/lucene/store/NativeFSLockFactory.java
+++ b/lucene/core/src/java/org/apache/lucene/store/NativeFSLockFactory.java
@@ -45,7 +45,7 @@ import org.apache.lucene.util.IOUtils;
  * could be left when the JVM exits abnormally.</p>
  *
  * <p>The primary benefit of {@link NativeFSLockFactory} is
- * that locks (not the lock file itsself) will be properly
+ * that locks (not the lock file itself) will be properly
  * removed (by the OS) if the JVM has an abnormal exit.</p>
  * 
  * <p>Note that, unlike {@link SimpleFSLockFactory}, the existence of
diff --git a/lucene/core/src/java/org/apache/lucene/util/ByteBlockPool.java b/lucene/core/src/java/org/apache/lucene/util/ByteBlockPool.java
index af8e195..bd0b8ea 100644
--- a/lucene/core/src/java/org/apache/lucene/util/ByteBlockPool.java
+++ b/lucene/core/src/java/org/apache/lucene/util/ByteBlockPool.java
@@ -346,7 +346,7 @@ public final class ByteBlockPool {
   }
   
   /**
-   * Reads bytes bytes out of the pool starting at the given offset with the given  
+   * Reads bytes out of the pool starting at the given offset with the given  
    * length into the given byte array at offset <tt>off</tt>.
    * <p>Note: this method allows to copy across block boundaries.</p>
    */
diff --git a/lucene/core/src/java/org/apache/lucene/util/automaton/CompiledAutomaton.java b/lucene/core/src/java/org/apache/lucene/util/automaton/CompiledAutomaton.java
index bd00a70..41acad4 100644
--- a/lucene/core/src/java/org/apache/lucene/util/automaton/CompiledAutomaton.java
+++ b/lucene/core/src/java/org/apache/lucene/util/automaton/CompiledAutomaton.java
@@ -138,7 +138,7 @@ public class CompiledAutomaton {
    *  to determine whether it is finite.  If simplify is true, we run
    *  possibly expensive operations to determine if the automaton is one
    *  the cases in {@link CompiledAutomaton.AUTOMATON_TYPE}. If simplify
-   *  requires determinizing the autaomaton then only maxDeterminizedStates
+   *  requires determinizing the automaton then only maxDeterminizedStates
    *  will be created.  Any more than that will cause a
    *  TooComplexToDeterminizeException.
    */
diff --git a/lucene/core/src/java/org/apache/lucene/util/automaton/MinimizationOperations.java b/lucene/core/src/java/org/apache/lucene/util/automaton/MinimizationOperations.java
index e911cdc..979f7c5 100644
--- a/lucene/core/src/java/org/apache/lucene/util/automaton/MinimizationOperations.java
+++ b/lucene/core/src/java/org/apache/lucene/util/automaton/MinimizationOperations.java
@@ -45,7 +45,7 @@ final public class MinimizationOperations {
 
   /**
    * Minimizes (and determinizes if not already deterministic) the given
-   * automaton using Hopcroft's algorighm.
+   * automaton using Hopcroft's algorithm.
    * @param maxDeterminizedStates maximum number of states determinizing the
    *  automaton can result in.  Set higher to allow more complex queries and
    *  lower to prevent memory exhaustion.
diff --git a/lucene/core/src/java/org/apache/lucene/util/automaton/RegExp.java b/lucene/core/src/java/org/apache/lucene/util/automaton/RegExp.java
index a643ddb..31cca6d 100644
--- a/lucene/core/src/java/org/apache/lucene/util/automaton/RegExp.java
+++ b/lucene/core/src/java/org/apache/lucene/util/automaton/RegExp.java
@@ -380,7 +380,7 @@ public class RegExp {
    * <code>RegExp(s, ALL)</code>.
    * 
    * @param s regexp string
-   * @exception IllegalArgumentException if an error occured while parsing the
+   * @exception IllegalArgumentException if an error occurred while parsing the
    *              regular expression
    */
   public RegExp(String s) throws IllegalArgumentException {
@@ -393,7 +393,7 @@ public class RegExp {
    * @param s regexp string
    * @param syntax_flags boolean 'or' of optional syntax constructs to be
    *          enabled
-   * @exception IllegalArgumentException if an error occured while parsing the
+   * @exception IllegalArgumentException if an error occurred while parsing the
    *              regular expression
    */
   public RegExp(String s, int syntax_flags) throws IllegalArgumentException {
diff --git a/lucene/core/src/java/org/apache/lucene/util/fst/Util.java b/lucene/core/src/java/org/apache/lucene/util/fst/Util.java
index ba2ff74..8ab3aed 100644
--- a/lucene/core/src/java/org/apache/lucene/util/fst/Util.java
+++ b/lucene/core/src/java/org/apache/lucene/util/fst/Util.java
@@ -103,7 +103,7 @@ public final class Util {
    *  works when the outputs are ascending in order with
    *  the inputs.
    *  For example, simple ordinals (0, 1,
-   *  2, ...), or file offets (when appending to a file)
+   *  2, ...), or file offsets (when appending to a file)
    *  fit this. */
   public static IntsRef getByOutput(FST<Long> fst, long targetOutput) throws IOException {
 
diff --git a/lucene/facet/src/java/org/apache/lucene/facet/DrillDownQuery.java b/lucene/facet/src/java/org/apache/lucene/facet/DrillDownQuery.java
index 7f0acc9..1ea2613 100644
--- a/lucene/facet/src/java/org/apache/lucene/facet/DrillDownQuery.java
+++ b/lucene/facet/src/java/org/apache/lucene/facet/DrillDownQuery.java
@@ -93,7 +93,7 @@ public final class DrillDownQuery extends Query {
 
   /** Adds one dimension of drill downs; if you pass the same
    *  dimension more than once it is OR'd with the previous
-   *  cofnstraints on that dimension, and all dimensions are
+   *  constraints on that dimension, and all dimensions are
    *  AND'd against each other and the base query. */
   public void add(String dim, String... path) {
     String indexedField = config.getDimConfig(dim).indexFieldName;
diff --git a/lucene/facet/src/java/org/apache/lucene/facet/LongValueFacetCounts.java b/lucene/facet/src/java/org/apache/lucene/facet/LongValueFacetCounts.java
index 85fc6c5..e3e33cb 100644
--- a/lucene/facet/src/java/org/apache/lucene/facet/LongValueFacetCounts.java
+++ b/lucene/facet/src/java/org/apache/lucene/facet/LongValueFacetCounts.java
@@ -40,7 +40,7 @@ import org.apache.lucene.util.PriorityQueue;
 
 
 /** {@link Facets} implementation that computes counts for
- *  all uniqute long values, more efficiently counting small values (0-1023) using an int array,
+ *  all unique long values, more efficiently counting small values (0-1023) using an int array,
  *  and switching to a <code>HashMap</code> for values above 1023.
  *  Retrieve all facet counts, in value order, with {@link #getAllChildrenSortByValue},
  *  or get the topN values sorted by count with {@link #getTopChildrenSortByCount}.
diff --git a/lucene/facet/src/java/org/apache/lucene/facet/taxonomy/writercache/LabelToOrdinal.java b/lucene/facet/src/java/org/apache/lucene/facet/taxonomy/writercache/LabelToOrdinal.java
index 40884f6..89746d5 100644
--- a/lucene/facet/src/java/org/apache/lucene/facet/taxonomy/writercache/LabelToOrdinal.java
+++ b/lucene/facet/src/java/org/apache/lucene/facet/taxonomy/writercache/LabelToOrdinal.java
@@ -52,9 +52,9 @@ public abstract class LabelToOrdinal {
   }
 
   /**
-   * Adds a new label if its not yet in the table.
+   * Adds a new label if it is not yet in the table.
    * Throws an {@link IllegalArgumentException} if the same label with
-   * a different ordinal was previoulsy added to this table.
+   * a different ordinal was previously added to this table.
    */
   public abstract void addLabel(FacetLabel label, int ordinal);
 
diff --git a/lucene/facet/src/java/org/apache/lucene/facet/taxonomy/writercache/NameHashIntCacheLRU.java b/lucene/facet/src/java/org/apache/lucene/facet/taxonomy/writercache/NameHashIntCacheLRU.java
index db79497..cca814c 100644
--- a/lucene/facet/src/java/org/apache/lucene/facet/taxonomy/writercache/NameHashIntCacheLRU.java
+++ b/lucene/facet/src/java/org/apache/lucene/facet/taxonomy/writercache/NameHashIntCacheLRU.java
@@ -19,7 +19,7 @@ package org.apache.lucene.facet.taxonomy.writercache;
 import org.apache.lucene.facet.taxonomy.FacetLabel;
 
 /**
- * An an LRU cache of mapping from name to int.
+ * An LRU cache of mapping from name to int.
  * Used to cache Ordinals of category paths.
  * It uses as key, hash of the path instead of the path.
  * This way the cache takes less RAM, but correctness depends on
diff --git a/lucene/facet/src/java/org/apache/lucene/facet/taxonomy/writercache/NameIntCacheLRU.java b/lucene/facet/src/java/org/apache/lucene/facet/taxonomy/writercache/NameIntCacheLRU.java
index 778f0b7..25f2799 100644
--- a/lucene/facet/src/java/org/apache/lucene/facet/taxonomy/writercache/NameIntCacheLRU.java
+++ b/lucene/facet/src/java/org/apache/lucene/facet/taxonomy/writercache/NameIntCacheLRU.java
@@ -22,7 +22,7 @@ import java.util.LinkedHashMap;
 import org.apache.lucene.facet.taxonomy.FacetLabel;
 
 /**
- * An an LRU cache of mapping from name to int.
+ * An LRU cache of mapping from name to int.
  * Used to cache Ordinals of category paths.
  * 
  * @lucene.experimental
diff --git a/lucene/misc/src/java/org/apache/lucene/document/LazyDocument.java b/lucene/misc/src/java/org/apache/lucene/document/LazyDocument.java
index 7fd8d98..7c5690e 100644
--- a/lucene/misc/src/java/org/apache/lucene/document/LazyDocument.java
+++ b/lucene/misc/src/java/org/apache/lucene/document/LazyDocument.java
@@ -59,7 +59,7 @@ public class LazyDocument {
    * <p>
    * <b>NOTE:</b> This method must be called once for each value of the field 
    * name specified in sequence that the values exist.  This method may not be 
-   * used to generate multiple, lazy, StorableField instances refering to 
+   * used to generate multiple, lazy, StorableField instances referring to 
    * the same underlying StorableField instance.
    * </p>
    * <p>
diff --git a/lucene/misc/src/java/org/apache/lucene/misc/SweetSpotSimilarity.java b/lucene/misc/src/java/org/apache/lucene/misc/SweetSpotSimilarity.java
index 8cc962d..5373227 100644
--- a/lucene/misc/src/java/org/apache/lucene/misc/SweetSpotSimilarity.java
+++ b/lucene/misc/src/java/org/apache/lucene/misc/SweetSpotSimilarity.java
@@ -34,7 +34,7 @@ import org.apache.lucene.search.similarities.ClassicSimilarity;
  * subclasses can choose between.
  * </p>
  *
- * @see <a href="doc-files/ss.gnuplot">A Gnuplot file used to generate some of the visualizations refrenced from each function.</a> 
+ * @see <a href="doc-files/ss.gnuplot">A Gnuplot file used to generate some of the visualizations referenced from each function.</a> 
  */
 public class SweetSpotSimilarity extends ClassicSimilarity {
 
diff --git a/lucene/queryparser/src/java/org/apache/lucene/queryparser/classic/CharStream.java b/lucene/queryparser/src/java/org/apache/lucene/queryparser/classic/CharStream.java
index 2c5fcba..443117f 100644
--- a/lucene/queryparser/src/java/org/apache/lucene/queryparser/classic/CharStream.java
+++ b/lucene/queryparser/src/java/org/apache/lucene/queryparser/classic/CharStream.java
@@ -71,7 +71,7 @@ interface CharStream {
    * Backs up the input stream by amount steps. Lexer calls this method if it
    * had already read some characters, but could not use them to match a
    * (longer) token. So, they will be used again as the prefix of the next
-   * token and it is the implemetation's responsibility to do this right.
+   * token and it is the implementation's responsibility to do this right.
    */
   void backup(int amount);
 
diff --git a/lucene/queryparser/src/java/org/apache/lucene/queryparser/flexible/standard/builders/SynonymQueryNodeBuilder.java b/lucene/queryparser/src/java/org/apache/lucene/queryparser/flexible/standard/builders/SynonymQueryNodeBuilder.java
index 716bcfd..ee005a0 100644
--- a/lucene/queryparser/src/java/org/apache/lucene/queryparser/flexible/standard/builders/SynonymQueryNodeBuilder.java
+++ b/lucene/queryparser/src/java/org/apache/lucene/queryparser/flexible/standard/builders/SynonymQueryNodeBuilder.java
@@ -24,7 +24,7 @@ import org.apache.lucene.search.BooleanClause.Occur;
 import org.apache.lucene.search.BooleanQuery;
 import org.apache.lucene.search.Query;
 
-/** Builer for {@link SynonymQueryNode}. */
+/** Builder for {@link SynonymQueryNode}. */
 public class SynonymQueryNodeBuilder implements StandardQueryBuilder {
 
   /** Sole constructor. */
diff --git a/lucene/queryparser/src/java/org/apache/lucene/queryparser/flexible/standard/nodes/PointQueryNode.java b/lucene/queryparser/src/java/org/apache/lucene/queryparser/flexible/standard/nodes/PointQueryNode.java
index 6d4cba7..69a7d09 100644
--- a/lucene/queryparser/src/java/org/apache/lucene/queryparser/flexible/standard/nodes/PointQueryNode.java
+++ b/lucene/queryparser/src/java/org/apache/lucene/queryparser/flexible/standard/nodes/PointQueryNode.java
@@ -88,7 +88,7 @@ public class PointQueryNode extends QueryNodeImpl implements
    * 
    * @param escaper the {@link EscapeQuerySyntax} used to escape the value {@link String}
    * 
-   * @return the value converte to {@link String} and escaped
+   * @return the value converted to {@link String} and escaped
    */
   protected CharSequence getTermEscaped(EscapeQuerySyntax escaper) {
     return escaper.escape(numberFormat.format(this.value),
diff --git a/lucene/queryparser/src/java/org/apache/lucene/queryparser/flexible/standard/parser/CharStream.java b/lucene/queryparser/src/java/org/apache/lucene/queryparser/flexible/standard/parser/CharStream.java
index ab149a6..15aee78 100644
--- a/lucene/queryparser/src/java/org/apache/lucene/queryparser/flexible/standard/parser/CharStream.java
+++ b/lucene/queryparser/src/java/org/apache/lucene/queryparser/flexible/standard/parser/CharStream.java
@@ -71,7 +71,7 @@ interface CharStream {
    * Backs up the input stream by amount steps. Lexer calls this method if it
    * had already read some characters, but could not use them to match a
    * (longer) token. So, they will be used again as the prefix of the next
-   * token and it is the implemetation's responsibility to do this right.
+   * token and it is the implementation's responsibility to do this right.
    */
   void backup(int amount);
 
diff --git a/lucene/queryparser/src/java/org/apache/lucene/queryparser/surround/parser/CharStream.java b/lucene/queryparser/src/java/org/apache/lucene/queryparser/surround/parser/CharStream.java
index 1756f86..4094d88 100644
--- a/lucene/queryparser/src/java/org/apache/lucene/queryparser/surround/parser/CharStream.java
+++ b/lucene/queryparser/src/java/org/apache/lucene/queryparser/surround/parser/CharStream.java
@@ -71,7 +71,7 @@ interface CharStream {
    * Backs up the input stream by amount steps. Lexer calls this method if it
    * had already read some characters, but could not use them to match a
    * (longer) token. So, they will be used again as the prefix of the next
-   * token and it is the implemetation's responsibility to do this right.
+   * token and it is the implementation's responsibility to do this right.
    */
   void backup(int amount);
 
diff --git a/lucene/replicator/src/java/org/apache/lucene/replicator/ReplicationClient.java b/lucene/replicator/src/java/org/apache/lucene/replicator/ReplicationClient.java
index 02fb657..2773eb1 100644
--- a/lucene/replicator/src/java/org/apache/lucene/replicator/ReplicationClient.java
+++ b/lucene/replicator/src/java/org/apache/lucene/replicator/ReplicationClient.java
@@ -391,7 +391,7 @@ public class ReplicationClient implements Closeable {
   }
   
   /**
-   * Executes the update operation immediately, irregardess if an update thread
+   * Executes the update operation immediately, irregardless if an update thread
    * is running or not.
    */
   public void updateNow() throws IOException {
diff --git a/lucene/replicator/src/java/org/apache/lucene/replicator/Revision.java b/lucene/replicator/src/java/org/apache/lucene/replicator/Revision.java
index 93fccd9..c8099a7 100644
--- a/lucene/replicator/src/java/org/apache/lucene/replicator/Revision.java
+++ b/lucene/replicator/src/java/org/apache/lucene/replicator/Revision.java
@@ -30,7 +30,7 @@ import org.apache.lucene.store.IndexInput;
  * revision will contain files from a single source. However, some applications
  * may require to treat a collection of indexes as a single entity so that the
  * files from all sources are replicated together, to guarantee consistency
- * beween them. For example, an application which indexes facets will need to
+ * between them. For example, an application which indexes facets will need to
  * replicate both the search and taxonomy indexes together, to guarantee that
  * they match at the client side.
  * 
@@ -60,7 +60,7 @@ public interface Revision extends Comparable<Revision> {
   
   /**
    * Returns an {@link IndexInput} for the given fileName and source. It is the
-   * caller's respnsibility to close the {@link IndexInput} when it has been
+   * caller's responsibility to close the {@link IndexInput} when it has been
    * consumed.
    */
   public InputStream open(String source, String fileName) throws IOException;
diff --git a/lucene/replicator/src/java/org/apache/lucene/replicator/nrt/ReplicaNode.java b/lucene/replicator/src/java/org/apache/lucene/replicator/nrt/ReplicaNode.java
index 06a8a0c..174a11f 100644
--- a/lucene/replicator/src/java/org/apache/lucene/replicator/nrt/ReplicaNode.java
+++ b/lucene/replicator/src/java/org/apache/lucene/replicator/nrt/ReplicaNode.java
@@ -730,8 +730,8 @@ public abstract class ReplicaNode extends Node {
   }
 
   /** Carefully determine if the file on the primary, identified by its {@code String fileName} along with the {@link FileMetaData}
-   * "summarizing" its contents, is precisely the same file that we have locally.  If the file does not exist locally, or if its its header
-   * (inclues the segment id), length, footer (including checksum) differ, then this returns false, else true. */
+   * "summarizing" its contents, is precisely the same file that we have locally.  If the file does not exist locally, or if its header
+   * (includes the segment id), length, footer (including checksum) differ, then this returns false, else true. */
   private boolean fileIsIdentical(String fileName, FileMetaData srcMetaData) throws IOException {
 
     FileMetaData destMetaData = readLocalFileMetaData(fileName);
diff --git a/lucene/sandbox/src/java/org/apache/lucene/codecs/idversion/IDVersionSegmentTermsEnum.java b/lucene/sandbox/src/java/org/apache/lucene/codecs/idversion/IDVersionSegmentTermsEnum.java
index b5e96ee..1f1b7ec 100644
--- a/lucene/sandbox/src/java/org/apache/lucene/codecs/idversion/IDVersionSegmentTermsEnum.java
+++ b/lucene/sandbox/src/java/org/apache/lucene/codecs/idversion/IDVersionSegmentTermsEnum.java
@@ -37,7 +37,7 @@ import org.apache.lucene.util.fst.Util;
 
 /** Iterates through terms in this field; this class is public so users
  *  can cast it to call {@link #seekExact(BytesRef, long)} for
- *  optimistic-concurreny, and also {@link #getVersion} to get the
+ *  optimistic-concurrency, and also {@link #getVersion} to get the
  *  version of the currently seek'd term. */
 public final class IDVersionSegmentTermsEnum extends TermsEnum {
 
diff --git a/lucene/suggest/src/java/org/apache/lucene/search/suggest/analyzing/AnalyzingInfixSuggester.java b/lucene/suggest/src/java/org/apache/lucene/search/suggest/analyzing/AnalyzingInfixSuggester.java
index 6947675..27ffaf2 100644
--- a/lucene/suggest/src/java/org/apache/lucene/search/suggest/analyzing/AnalyzingInfixSuggester.java
+++ b/lucene/suggest/src/java/org/apache/lucene/search/suggest/analyzing/AnalyzingInfixSuggester.java
@@ -677,7 +677,7 @@ public class AnalyzingInfixSuggester extends Lookup implements Closeable {
   /**
    * Create the results based on the search hits.
    * Can be overridden by subclass to add particular behavior (e.g. weight transformation).
-   * Note that there is no prefix toke (the {@code prefixToken} argument will
+   * Note that there is no prefix token (the {@code prefixToken} argument will
    * be null) whenever the final token in the incoming request was in fact finished
    * (had trailing characters, such as white-space).
    *
diff --git a/lucene/suggest/src/java/org/apache/lucene/search/suggest/jaspell/JaspellTernarySearchTrie.java b/lucene/suggest/src/java/org/apache/lucene/search/suggest/jaspell/JaspellTernarySearchTrie.java
index d30c0b3..3900a62 100644
--- a/lucene/suggest/src/java/org/apache/lucene/search/suggest/jaspell/JaspellTernarySearchTrie.java
+++ b/lucene/suggest/src/java/org/apache/lucene/search/suggest/jaspell/JaspellTernarySearchTrie.java
@@ -113,7 +113,7 @@ public class JaspellTernarySearchTrie implements Accountable {
   }
 
   /**
-   * Compares characters by alfabetical order.
+   * Compares characters by alphabetical order.
    * 
    *@param cCompare2
    *          The first char in the comparison.
@@ -204,7 +204,7 @@ public class JaspellTernarySearchTrie implements Accountable {
    *@param file
    *          The <code>Path</code> with the data to load into the Trie.
    *@exception IOException
-   *              A problem occured while reading the data.
+   *              A problem occurred while reading the data.
    */
   public JaspellTernarySearchTrie(Path file) throws IOException {
     this(file, false);
@@ -221,7 +221,7 @@ public class JaspellTernarySearchTrie implements Accountable {
    *          If true, the file is compressed with the GZIP algorithm, and if
    *          false, the file is a normal text document.
    *@exception IOException
-   *              A problem occured while reading the data.
+   *              A problem occurred while reading the data.
    */
   public JaspellTernarySearchTrie(Path file, boolean compression)
           throws IOException {
diff --git a/lucene/suggest/src/test/org/apache/lucene/search/suggest/Top50KWiki.utf8 b/lucene/suggest/src/test/org/apache/lucene/search/suggest/Top50KWiki.utf8
index 407eac7..1e1119a 100644
--- a/lucene/suggest/src/test/org/apache/lucene/search/suggest/Top50KWiki.utf8
+++ b/lucene/suggest/src/test/org/apache/lucene/search/suggest/Top50KWiki.utf8
@@ -43218,7 +43218,7 @@ phoneme|3088
 aftenposten|3088
 hayek|3088
 groups.yahoo.com|3088
-occured|3087
+occurred|3087
 intimidate|3087
 workable|3087
 groin|3087
diff --git a/solr/contrib/dataimporthandler/src/test/org/apache/solr/handler/dataimport/TestContentStreamDataSource.java b/solr/contrib/dataimporthandler/src/test/org/apache/solr/handler/dataimport/TestContentStreamDataSource.java
index 9617726..7a7ddb5 100644
--- a/solr/contrib/dataimporthandler/src/test/org/apache/solr/handler/dataimport/TestContentStreamDataSource.java
+++ b/solr/contrib/dataimporthandler/src/test/org/apache/solr/handler/dataimport/TestContentStreamDataSource.java
@@ -106,7 +106,7 @@ public class TestContentStreamDataSource extends AbstractDataImportHandlerTestCa
         Thread.sleep(500);
       }
     }
-    fail("Commit should have occured but it did not");
+    fail("Commit should have occurred but it did not");
   }
   
   private static class SolrInstance {
diff --git a/solr/core/src/java/org/apache/solr/core/CoreContainer.java b/solr/core/src/java/org/apache/solr/core/CoreContainer.java
index 3592c6e..05a8c79 100644
--- a/solr/core/src/java/org/apache/solr/core/CoreContainer.java
+++ b/solr/core/src/java/org/apache/solr/core/CoreContainer.java
@@ -1369,7 +1369,7 @@ public class CoreContainer {
   }
 
   /**
-   * Returns an immutable Map of Exceptions that occured when initializing 
+   * Returns an immutable Map of Exceptions that occurred when initializing 
    * SolrCores (either at startup, or do to runtime requests to create cores) 
    * keyed off of the name (String) of the SolrCore that had the Exception 
    * during initialization.
diff --git a/solr/core/src/java/org/apache/solr/handler/AnalysisRequestHandlerBase.java b/solr/core/src/java/org/apache/solr/handler/AnalysisRequestHandlerBase.java
index 54fdf99..814249b 100644
--- a/solr/core/src/java/org/apache/solr/handler/AnalysisRequestHandlerBase.java
+++ b/solr/core/src/java/org/apache/solr/handler/AnalysisRequestHandlerBase.java
@@ -178,7 +178,7 @@ public abstract class AnalysisRequestHandlerBase extends RequestHandlerBase {
       tokenStream.end();
       return tokens;
     } catch (IOException ioe) {
-      throw new RuntimeException("Error occured while iterating over tokenstream", ioe);
+      throw new RuntimeException("Error occurred while iterating over tokenstream", ioe);
     }
   }
 
@@ -206,7 +206,7 @@ public abstract class AnalysisRequestHandlerBase extends RequestHandlerBase {
       }
       tokenStream.end(); // TODO should we capture?
     } catch (IOException ioe) {
-      throw new RuntimeException("Error occured while iterating over tokenstream", ioe);
+      throw new RuntimeException("Error occurred while iterating over tokenstream", ioe);
     } finally {
       IOUtils.closeWhileHandlingException(tokenStream);
     }
diff --git a/solr/core/src/java/org/apache/solr/update/processor/DistributedUpdateProcessor.java b/solr/core/src/java/org/apache/solr/update/processor/DistributedUpdateProcessor.java
index c8ddeb9..2a76ab9 100644
--- a/solr/core/src/java/org/apache/solr/update/processor/DistributedUpdateProcessor.java
+++ b/solr/core/src/java/org/apache/solr/update/processor/DistributedUpdateProcessor.java
@@ -1890,7 +1890,7 @@ public class DistributedUpdateProcessor extends UpdateRequestProcessor {
             // if we aren't the leader, then we need to check that updates were not re-ordered
             if (bucketVersion != 0 && bucketVersion < versionOnUpdate) {
               // we're OK... this update has a version higher than anything we've seen
-              // in this bucket so far, so we know that no reordering has yet occured.
+              // in this bucket so far, so we know that no reordering has yet occurred.
               bucket.updateHighest(versionOnUpdate);
             } else {
               // there have been updates higher than the current update. we need to check
diff --git a/solr/solr-ref-guide/src/highlighting.adoc b/solr/solr-ref-guide/src/highlighting.adoc
index 6764343..c792a91 100644
--- a/solr/solr-ref-guide/src/highlighting.adoc
+++ b/solr/solr-ref-guide/src/highlighting.adoc
@@ -87,7 +87,7 @@ The default is `<em>`.
 The default is `</em>`.
 
 `hl.encoder`::
-If blank, the default, then the stored text will be returned without any escaping/encoding performed by the highlighter. If set to `html` then special HMTL/XML characters will be encoded (e.g., `&` becomes `\&amp;`). The pre/post snippet characters are never encoded.
+If blank, the default, then the stored text will be returned without any escaping/encoding performed by the highlighter. If set to `html` then special HTML/XML characters will be encoded (e.g., `&` becomes `\&amp;`). The pre/post snippet characters are never encoded.
 
 `hl.maxAnalyzedChars`::
 The character limit to look for highlights, after which no highlighting will be done. This is mostly only a performance concern for an _analysis_ based offset source since it's the slowest. See <<Schema Options and Performance Considerations>>.
diff --git a/solr/solrj/src/java/org/apache/solr/client/solrj/impl/HttpSolrClient.java b/solr/solrj/src/java/org/apache/solr/client/solrj/impl/HttpSolrClient.java
index 8831448..3cdcdd4 100644
--- a/solr/solrj/src/java/org/apache/solr/client/solrj/impl/HttpSolrClient.java
+++ b/solr/solrj/src/java/org/apache/solr/client/solrj/impl/HttpSolrClient.java
@@ -657,11 +657,11 @@ public class HttpSolrClient extends SolrClient {
           + getBaseURL(), e);
     } catch (SocketTimeoutException e) {
       throw new SolrServerException(
-          "Timeout occured while waiting response from server at: "
+          "Timeout occurred while waiting response from server at: "
               + getBaseURL(), e);
     } catch (IOException e) {
       throw new SolrServerException(
-          "IOException occured when talking to server at: " + getBaseURL(), e);
+          "IOException occurred when talking to server at: " + getBaseURL(), e);
     } finally {
       if (shouldClose) {
         Utils.consumeFully(entity);
diff --git a/solr/solrj/src/java/org/apache/solr/common/ToleratedUpdateError.java b/solr/solrj/src/java/org/apache/solr/common/ToleratedUpdateError.java
index 9c4ac9a..20ac15d 100644
--- a/solr/solrj/src/java/org/apache/solr/common/ToleratedUpdateError.java
+++ b/solr/solrj/src/java/org/apache/solr/common/ToleratedUpdateError.java
@@ -25,7 +25,7 @@ import org.apache.solr.common.util.SimpleOrderedMap;
 import static org.apache.solr.common.params.CommonParams.ID;
 
 /**
- * Models the basic information related to a single "tolerated" error that occured during updates.  
+ * Models the basic information related to a single "tolerated" error that occurred during updates.  
  * This class is only useful when the <code>TolerantUpdateProcessorFactory</code> is used in an update 
  * processor chain
  */