You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@solr.apache.org by ho...@apache.org on 2023/07/27 18:54:10 UTC

[solr] branch branch_9x updated: MINOR: Fix typos in `org.apache.solr.core` package (#1629)

This is an automated email from the ASF dual-hosted git repository.

houston pushed a commit to branch branch_9x
in repository https://gitbox.apache.org/repos/asf/solr.git


The following commit(s) were added to refs/heads/branch_9x by this push:
     new 38fb272cabf MINOR: Fix typos in `org.apache.solr.core` package (#1629)
38fb272cabf is described below

commit 38fb272cabf0294e4dad4f19c2ddc3d96abf35b0
Author: Andrey Bozhko <an...@gmail.com>
AuthorDate: Mon May 8 14:26:23 2023 -0500

    MINOR: Fix typos in `org.apache.solr.core` package (#1629)
    
    * solr core - fix typos
    
    * solr core - fix typos
    
    * solr core - fix typos
    
    * run ./gradlew tidy
    
    * Update solr/core/src/java/org/apache/solr/core/backup/repository/BackupRepositoryFactory.java
    
    same capitalization
    
    * a few typos.
    
    * add CHANGES entry
    
    * revert "add CHANGES entry"
    
    This reverts commit 07533d68619dfd928a70c91066bcbe39aaf1c653.
    
    * add CHANGES entry
    
    ---------
    
    Co-authored-by: Marcus <ma...@gmail.com>
    (cherry picked from commit 810783d812e53aa6a382e082a0953eb1ca2bfa28)
---
 solr/CHANGES.txt                                   |  2 +
 .../apache/solr/core/CachingDirectoryFactory.java  |  4 +-
 .../src/java/org/apache/solr/core/CloseHook.java   |  4 +-
 .../java/org/apache/solr/core/ConfigOverlay.java   |  2 +-
 .../src/java/org/apache/solr/core/ConfigSet.java   |  2 +-
 .../java/org/apache/solr/core/CoreContainer.java   | 36 +++++++--------
 .../java/org/apache/solr/core/CoreDescriptor.java  |  2 +-
 .../src/java/org/apache/solr/core/CoreSorter.java  |  4 +-
 .../org/apache/solr/core/DirectoryFactory.java     | 10 ++---
 .../solr/core/IndexDeletionPolicyWrapper.java      | 52 +++++++++++-----------
 .../src/java/org/apache/solr/core/NodeConfig.java  |  4 +-
 .../src/java/org/apache/solr/core/PluginBag.java   |  2 +-
 .../src/java/org/apache/solr/core/PluginInfo.java  |  2 +-
 .../src/java/org/apache/solr/core/SolrCore.java    | 46 +++++++++----------
 .../org/apache/solr/core/SolrDeletionPolicy.java   |  4 +-
 .../org/apache/solr/core/SolrResourceLoader.java   |  6 +--
 .../java/org/apache/solr/core/SolrXmlConfig.java   |  8 ++--
 .../org/apache/solr/core/TracerConfigurator.java   |  2 +-
 .../apache/solr/core/TransientSolrCoreCache.java   |  2 +-
 .../java/org/apache/solr/core/XmlConfigFile.java   |  2 +-
 .../core/backup/repository/BackupRepository.java   |  6 +--
 .../backup/repository/BackupRepositoryFactory.java |  2 +-
 22 files changed, 103 insertions(+), 101 deletions(-)

diff --git a/solr/CHANGES.txt b/solr/CHANGES.txt
index 27a97f54c39..86382a0cf64 100644
--- a/solr/CHANGES.txt
+++ b/solr/CHANGES.txt
@@ -40,6 +40,8 @@ Other Changes
 
 * SOLR-16856: The default auto-soft-commit time has been set to 3 seconds. Previously, auto-soft-commit was disabled by default. (Houston Putman)
 
+* PR#1629: Fix typos in org.apache.solr.core package (Andrey Bozhko, Marcus Eagan)
+
 ==================  9.3.0 ==================
 
 Upgrade Notes
diff --git a/solr/core/src/java/org/apache/solr/core/CachingDirectoryFactory.java b/solr/core/src/java/org/apache/solr/core/CachingDirectoryFactory.java
index 9864ad7ee48..8a5766f2f51 100644
--- a/solr/core/src/java/org/apache/solr/core/CachingDirectoryFactory.java
+++ b/solr/core/src/java/org/apache/solr/core/CachingDirectoryFactory.java
@@ -231,7 +231,7 @@ public abstract class CachingDirectoryFactory extends DirectoryFactory {
     byPathCache.remove(v.path);
   }
 
-  // be sure this is called with the this sync lock
+  // be sure the method is called with the sync lock on this object
   // returns true if we closed the cacheValue, false if it will be closed later
   private boolean closeCacheValue(CacheValue cacheValue) {
     log.debug("looking to close {} {}", cacheValue.path, cacheValue.closeEntries);
@@ -441,7 +441,7 @@ public abstract class CachingDirectoryFactory extends DirectoryFactory {
       throw new NullPointerException();
     }
     synchronized (this) {
-      // don't check if already closed here - we need to able to release
+      // don't check if already closed here - we need to be able to release
       // while #close() waits.
 
       CacheValue cacheValue = byDirectoryCache.get(directory);
diff --git a/solr/core/src/java/org/apache/solr/core/CloseHook.java b/solr/core/src/java/org/apache/solr/core/CloseHook.java
index 582e91e5b60..3b1ffeeb895 100644
--- a/solr/core/src/java/org/apache/solr/core/CloseHook.java
+++ b/solr/core/src/java/org/apache/solr/core/CloseHook.java
@@ -32,7 +32,7 @@ public interface CloseHook {
    * handler and searcher(s) are actually closed <br>
    * <b>Important:</b> Keep the method implementation as short as possible. If it were to use any
    * heavy i/o , network connections - it might be a better idea to launch in a separate Thread so
-   * as to not to block the process of shutting down a given SolrCore instance.
+   * as to not block the process of shutting down a given SolrCore instance.
    *
    * @param core SolrCore object that is shutting down / closing
    */
@@ -47,7 +47,7 @@ public interface CloseHook {
    *
    * <p><b>Important:</b> Keep the method implementation as short as possible. If it were to use any
    * heavy i/o , network connections - it might be a better idea to launch in a separate Thread so
-   * as to not to block the process of shutting down a given SolrCore instance.
+   * as to not block the process of shutting down a given SolrCore instance.
    *
    * @param core SolrCore object that is shutting down / closing
    */
diff --git a/solr/core/src/java/org/apache/solr/core/ConfigOverlay.java b/solr/core/src/java/org/apache/solr/core/ConfigOverlay.java
index 478ab92f188..403501c19f8 100644
--- a/solr/core/src/java/org/apache/solr/core/ConfigOverlay.java
+++ b/solr/core/src/java/org/apache/solr/core/ConfigOverlay.java
@@ -31,7 +31,7 @@ import org.apache.solr.common.util.Utils;
 
 /**
  * This class encapsulates the config overlay json file. It is immutable and any edit operations
- * performed on tbhis gives a new copy of the object with the changed value
+ * performed on this gives a new copy of the object with the changed value
  */
 public class ConfigOverlay implements MapSerializable {
   private final int version;
diff --git a/solr/core/src/java/org/apache/solr/core/ConfigSet.java b/solr/core/src/java/org/apache/solr/core/ConfigSet.java
index 857fbb30552..de363f69d90 100644
--- a/solr/core/src/java/org/apache/solr/core/ConfigSet.java
+++ b/solr/core/src/java/org/apache/solr/core/ConfigSet.java
@@ -82,7 +82,7 @@ public class ConfigSet {
   /**
    * Provide a Schema object on demand We want IndexSchema Objects to be lazily instantiated because
    * when a configset is created the {@link SolrResourceLoader} associated with it is not associated
-   * with a core So, we may not be able to update the core if we the schema classes are updated
+   * with a core So, we may not be able to update the core if the schema classes are updated
    */
   interface SchemaSupplier {
     IndexSchema get(boolean forceFetch);
diff --git a/solr/core/src/java/org/apache/solr/core/CoreContainer.java b/solr/core/src/java/org/apache/solr/core/CoreContainer.java
index 99f503258b8..f223f41e36f 100644
--- a/solr/core/src/java/org/apache/solr/core/CoreContainer.java
+++ b/solr/core/src/java/org/apache/solr/core/CoreContainer.java
@@ -307,7 +307,7 @@ public class CoreContainer {
       ExecutorUtil.newMDCAwareCachedThreadPool("Core Container Async Task");
 
   /**
-   * Non empty if the Collection API is executed in a distributed way and not on Overseer, once the
+   * Non-empty if the Collection API is executed in a distributed way and not on Overseer, once the
    * CoreContainer has been initialized properly, i.e. method {@link #load()} called. Until then it
    * is null, and it is not expected to be read.
    */
@@ -579,15 +579,15 @@ public class CoreContainer {
       SolrHttpClientBuilder builder =
           builderPlugin.getHttpClientBuilder(HttpClientUtil.getHttpClientBuilder());
 
-      // this caused plugins like KerberosPlugin to register it's intercepts, but this intercept
-      // logic is also handled by the pki authentication code when it decideds to let the plugin
-      // handle auth via it's intercept - so you would end up with two intercepts
+      // this caused plugins like KerberosPlugin to register its intercepts, but this intercept
+      // logic is also handled by the pki authentication code when it decides to let the plugin
+      // handle auth via its intercept - so you would end up with two intercepts
       // -->
       //  shardHandlerFactory.setSecurityBuilder(builderPlugin); // calls setup for the authcPlugin
       //  updateShardHandler.setSecurityBuilder(builderPlugin);
       // <--
 
-      // This should not happen here at all - it's only currently required due to its affect on
+      // This should not happen here at all - it's only currently required due to its effect on
       // http1 clients in a test or two incorrectly counting on it for their configuration.
       // -->
 
@@ -1400,8 +1400,8 @@ public class CoreContainer {
    * preserve consistency.
    *
    * <p>Jetty already allows a grace period for in-flight requests to complete and our solr cores,
-   * searchers etc are reference counted to allow for graceful shutdown. So we don't worry about any
-   * other kind of requests.
+   * searchers, etc., are reference counted to allow for graceful shutdown. So we don't worry about
+   * any other kind of requests.
    *
    * <p>We do not need to unpause ever because the node is being shut down.
    */
@@ -1514,14 +1514,14 @@ public class CoreContainer {
       assertPathAllowed(cd.getInstanceDir());
       assertPathAllowed(Paths.get(cd.getDataDir()));
 
-      boolean preExisitingZkEntry = false;
+      boolean preExistingZkEntry = false;
       try {
         if (getZkController() != null) {
           if (cd.getCloudDescriptor().getCoreNodeName() == null) {
             throw new SolrException(
                 ErrorCode.SERVER_ERROR, "coreNodeName missing " + parameters.toString());
           }
-          preExisitingZkEntry = getZkController().checkIfCoreNodeNameAlreadyExists(cd);
+          preExistingZkEntry = getZkController().checkIfCoreNodeNameAlreadyExists(cd);
         }
 
         // Much of the logic in core handling pre-supposes that the core.properties file already
@@ -1544,7 +1544,7 @@ public class CoreContainer {
         // First clean up any core descriptor, there should never be an existing core.properties
         // file for any core that failed to be created on-the-fly.
         coresLocator.delete(this, cd);
-        if (isZooKeeperAware() && !preExisitingZkEntry) {
+        if (isZooKeeperAware() && !preExistingZkEntry) {
           try {
             getZkController().unregister(coreName, cd);
           } catch (InterruptedException e) {
@@ -1797,7 +1797,7 @@ public class CoreContainer {
     }
   }
 
-  /** Write a new index directory for the a SolrCore, but do so without loading it. */
+  /** Write a new index directory for the SolrCore, but do so without loading it. */
   private void resetIndexDirectory(CoreDescriptor dcore, ConfigSet coreConfig) {
     SolrConfig config = coreConfig.getSolrConfig();
 
@@ -2087,7 +2087,7 @@ public class CoreContainer {
       // check for core-init errors first
       CoreLoadFailure loadFailure = coreInitFailures.remove(name);
       if (loadFailure != null) {
-        // getting the index directory requires opening a DirectoryFactory with a SolrConfig, etc,
+        // getting the index directory requires opening a DirectoryFactory with a SolrConfig, etc.,
         // which we may not be able to do because of the init error.  So we just go with what we
         // can glean from the CoreDescriptor - datadir and instancedir
         SolrCore.deleteUnloadedCore(loadFailure.cd, deleteDataDir, deleteInstanceDir);
@@ -2236,7 +2236,7 @@ public class CoreContainer {
     // load failure (use reload to cure that). But for
     // TestConfigSetsAPI.testUploadWithScriptUpdateProcessor, this needs to _not_ try to load the
     // core if the core is null and there was an error. If you change this, be sure to run both
-    // TestConfiSetsAPI and TestLazyCores
+    // TestConfigSetsAPI and TestLazyCores
     if (desc == null || zkSys.getZkController() != null) return null;
 
     // This will put an entry in pending core ops if the core isn't loaded. Here's where moving the
@@ -2482,13 +2482,13 @@ public class CoreContainer {
   }
 
   /**
-   * Run an arbitrary task in it's own thread. This is an expert option and is a method you should
+   * Run an arbitrary task in its own thread. This is an expert option and is a method you should
    * use with great care. It would be bad to run something that never stopped or run something that
    * took a very long time. Typically this is intended for actions that take a few seconds, and
    * therefore would be bad to wait for within a request, or actions that need to happen when a core
-   * has zero references, but but would not pose a significant hindrance to server shut down times.
-   * It is not intended for long running tasks and if you are using a Runnable with a loop in it,
-   * you are almost certainly doing it wrong.
+   * has zero references, but would not pose a significant hindrance to server shut down times. It
+   * is not intended for long-running tasks and if you are using a Runnable with a loop in it, you
+   * are almost certainly doing it wrong.
    *
    * <p><br>
    * WARNING: Solr wil not be able to shut down gracefully until this task completes!
@@ -2553,7 +2553,7 @@ class CloserThread extends Thread {
   }
 
   // It's important that this be the _only_ thread removing things from pendingDynamicCloses!
-  // This is single-threaded, but I tried a multi-threaded approach and didn't see any performance
+  // This is single-threaded, but I tried a multithreaded approach and didn't see any performance
   // gains, so there's no good justification for the complexity. I suspect that the locking on
   // things like DefaultSolrCoreState essentially create a single-threaded process anyway.
   @Override
diff --git a/solr/core/src/java/org/apache/solr/core/CoreDescriptor.java b/solr/core/src/java/org/apache/solr/core/CoreDescriptor.java
index ecfb8b55f20..5c4744baeba 100644
--- a/solr/core/src/java/org/apache/solr/core/CoreDescriptor.java
+++ b/solr/core/src/java/org/apache/solr/core/CoreDescriptor.java
@@ -256,7 +256,7 @@ public class CoreDescriptor {
   }
 
   /**
-   * Create the properties object used by resource loaders, etc, for property substitution. The
+   * Create the properties object used by resource loaders, etc., for property substitution. The
    * default solr properties are prefixed with 'solr.core.', so, e.g., 'name' becomes
    * 'solr.core.name'
    */
diff --git a/solr/core/src/java/org/apache/solr/core/CoreSorter.java b/solr/core/src/java/org/apache/solr/core/CoreSorter.java
index c90a3d6540a..399526dfa65 100644
--- a/solr/core/src/java/org/apache/solr/core/CoreSorter.java
+++ b/solr/core/src/java/org/apache/solr/core/CoreSorter.java
@@ -52,14 +52,14 @@ public final class CoreSorter implements Comparator<CoreDescriptor> {
           // with 0 down nodes than 1 down node because it will make the shard
           // complete earlier and avoid waiting by the other live nodes
           if (c1.totalReplicasInLiveNodes > 0) {
-            // means nobody else is waiting for this , so no need to prioritize
+            // means nobody else is waiting for this, so no need to prioritize
             return -1;
           }
         }
         if (c2.totalReplicasInDownNodes < c1.totalReplicasInDownNodes) {
           // same is the above, just to take care of the case where c2 has to be prioritized
           if (c2.totalReplicasInLiveNodes > 0) {
-            // means nobody else is waiting for this , so no need to priotitize
+            // means nobody else is waiting for this, so no need to prioritize
             return 1;
           }
         }
diff --git a/solr/core/src/java/org/apache/solr/core/DirectoryFactory.java b/solr/core/src/java/org/apache/solr/core/DirectoryFactory.java
index 34bed4dc50e..e1d22572bb5 100644
--- a/solr/core/src/java/org/apache/solr/core/DirectoryFactory.java
+++ b/solr/core/src/java/org/apache/solr/core/DirectoryFactory.java
@@ -53,7 +53,7 @@ public abstract class DirectoryFactory implements NamedListInitializedPlugin, Cl
   protected static final String INDEX_W_TIMESTAMP_REGEX =
       "index\\.[0-9]{17}"; // see SnapShooter.DATE_FMT
 
-  // May be set by sub classes as data root, in which case getDataHome will use it as base.
+  // May be set by subclasses as data root, in which case getDataHome will use it as base.
   // Absolute.
   protected Path dataHomePath;
 
@@ -86,7 +86,7 @@ public abstract class DirectoryFactory implements NamedListInitializedPlugin, Cl
   public abstract void addCloseListener(Directory dir, CloseListener closeListener);
 
   /**
-   * Close the this and all of the Directories it contains.
+   * Close this factory and all of the Directories it contains.
    *
    * @throws IOException If there is a low-level I/O error.
    */
@@ -196,7 +196,7 @@ public abstract class DirectoryFactory implements NamedListInitializedPlugin, Cl
     fromDir.deleteFile(fileName);
   }
 
-  // sub classes perform an atomic rename if possible, otherwise fall back to delete + rename
+  // subclasses perform an atomic rename if possible, otherwise fall back to delete + rename
   // this is important to support for index roll over durability after crashes
   public void renameWithOverwrite(Directory dir, String fileName, String toName)
       throws IOException {
@@ -321,7 +321,7 @@ public abstract class DirectoryFactory implements NamedListInitializedPlugin, Cl
    * Get the data home folder. If solr.data.home is set, that is used, else base on instanceDir
    *
    * @param cd core descriptor instance
-   * @return a String with absolute path to data direcotry
+   * @return a String with absolute path to data directory
    */
   public String getDataHome(CoreDescriptor cd) throws IOException {
     Path dataDir;
@@ -391,7 +391,7 @@ public abstract class DirectoryFactory implements NamedListInitializedPlugin, Cl
     }
   }
 
-  // Extension point to allow sub-classes to infuse additional code when deleting old index
+  // Extension point to allow subclasses to infuse additional code when deleting old index
   // directories
   protected boolean deleteOldIndexDirectory(String oldDirPath) throws IOException {
     Path dirToRm = Path.of(oldDirPath);
diff --git a/solr/core/src/java/org/apache/solr/core/IndexDeletionPolicyWrapper.java b/solr/core/src/java/org/apache/solr/core/IndexDeletionPolicyWrapper.java
index 8089372713e..130e41a8317 100644
--- a/solr/core/src/java/org/apache/solr/core/IndexDeletionPolicyWrapper.java
+++ b/solr/core/src/java/org/apache/solr/core/IndexDeletionPolicyWrapper.java
@@ -42,7 +42,7 @@ import org.slf4j.LoggerFactory;
  * A wrapper for an IndexDeletionPolicy instance.
  *
  * <p>Provides features for looking up IndexCommit given a version. Allows reserving index commit
- * points for certain amounts of time to support features such as index replication or snapshooting
+ * points for certain amounts of time to support features such as index replication or snapshotting
  * directly out of a live index directory.
  *
  * <p><b>NOTE</b>: The {@link #clone()} method returns <code>this</code> in order to make this
@@ -72,7 +72,7 @@ public final class IndexDeletionPolicyWrapper extends IndexDeletionPolicy {
   private volatile Map<Long, IndexCommit> knownCommits = new ConcurrentHashMap<>();
 
   /**
-   * The most recent commit included in call to {@link #onInit} or {@link #onCommit} <em>beore</em>
+   * The most recent commit included in call to {@link #onInit} or {@link #onCommit} <em>before</em>
    * delegating to our {@link #getWrappedDeletionPolicy()}.
    *
    * <p><b>NOTE:</b> This may be null if there is not yet a single commit to our index.
@@ -85,7 +85,7 @@ public final class IndexDeletionPolicyWrapper extends IndexDeletionPolicy {
   private volatile IndexCommit latestCommit;
 
   /**
-   * The set of all commit generations htat have been reserved for for some amount of time.
+   * The set of all commit generations that have been reserved for some amount of time.
    *
    * <p>The keys of the {@link IndexCommit#getGeneration} of a commit, the values are the {@link
    * System#nanoTime} that the commit should be reserved until.
@@ -117,7 +117,7 @@ public final class IndexDeletionPolicyWrapper extends IndexDeletionPolicy {
   /**
    * Returns the most recent commit point.
    *
-   * <p><b>NOTE:</b> This method makes no garuntee that the commit returned still exists as the
+   * <p><b>NOTE:</b> This method makes no guarantee that the commit returned still exists as the
    * moment this method completes. Callers are encouraged to use {@link #getAndSaveLatestCommit}
    * instead.
    *
@@ -186,13 +186,13 @@ public final class IndexDeletionPolicyWrapper extends IndexDeletionPolicy {
   /**
    * Set the duration for which commit point is to be reserved by the deletion policy.
    *
-   * <p><b>NOTE:</b> This method does not make any garuntees that the specified index generation
-   * exists, or that the specified generation has not already ben deleted. The only garuntee is that
-   * <em>if</em> the specified generation exists now, or is created at some point in the future,
-   * then it will be resered for <em>at least</em> the specified <code>reserveTime</code>.
+   * <p><b>NOTE:</b> This method does not make any guarantees that the specified index generation
+   * exists, or that the specified generation has not already ben deleted. The only guarantee is
+   * that <em>if</em> the specified generation exists now, or is created at some point in the
+   * future, then it will be reserved for <em>at least</em> the specified <code>reserveTime</code>.
    *
    * @param indexGen gen of the commit point to be reserved
-   * @param reserveTime durration in milliseconds (relative to 'now') for which the commit point is
+   * @param reserveTime duration in milliseconds (relative to 'now') for which the commit point is
    *     to be reserved
    */
   public void setReserveDuration(Long indexGen, long reserveTime) {
@@ -200,7 +200,7 @@ public final class IndexDeletionPolicyWrapper extends IndexDeletionPolicy {
     // all operations on 'reserves' are done atomically.
     //
     // Here we'll use Map.merge to ensure that we atomically replace any existing timestamp if
-    // and only if our new reservation timetsamp is larger.
+    // and only if our new reservation timestamp is larger.
     final long reserveAsNanoTime =
         System.nanoTime() + TimeUnit.NANOSECONDS.convert(reserveTime, TimeUnit.MILLISECONDS);
     reserves.merge(indexGen, reserveAsNanoTime, BinaryOperator.maxBy(Comparator.naturalOrder()));
@@ -220,13 +220,13 @@ public final class IndexDeletionPolicyWrapper extends IndexDeletionPolicy {
 
   /**
    * Permanently prevent this commit point from being deleted (if it has not already) using a
-   * refrence count.
+   * reference count.
    *
    * <p><b>NOTE:</b> Callers <em>MUST</em> call {@link #releaseCommitPoint} when finished using it
    * in order to decrement the reference count, or the commit will be preserved in the Directory
    * forever.
    *
-   * @param generation the generation of the IndexComit to save until released
+   * @param generation the generation of the IndexCommit to save until released
    * @see #getAndSaveLatestCommit
    * @see #getAndSaveCommitPoint
    * @see #releaseCommitPoint
@@ -239,7 +239,7 @@ public final class IndexDeletionPolicyWrapper extends IndexDeletionPolicy {
   /**
    * Release a previously saved commit point.
    *
-   * <p>This is a convinience wrapper around {@link #releaseCommitPoint(Long)} that will ignore null
+   * <p>This is a convenience wrapper around {@link #releaseCommitPoint(Long)} that will ignore null
    * input.
    */
   public synchronized void releaseCommitPoint(IndexCommit commit) {
@@ -251,9 +251,9 @@ public final class IndexDeletionPolicyWrapper extends IndexDeletionPolicy {
   /**
    * Release a previously saved commit point.
    *
-   * <p>This method does not enforce that that the specified generation has previously been saved,
-   * or even that it's 'non-null'. But if both are true then it will decrement the reference count
-   * for the specified generation.
+   * <p>This method does not enforce that the specified generation has previously been saved, or
+   * even that it's 'non-null'. But if both are true then it will decrement the reference count for
+   * the specified generation.
    */
   public synchronized void releaseCommitPoint(Long generation) {
     if (null == generation) {
@@ -382,7 +382,7 @@ public final class IndexDeletionPolicyWrapper extends IndexDeletionPolicy {
   /**
    * Returns the commit with the specified generation <em>if it is known</em>.
    *
-   * <p><b>NOTE:</b> This method makes no garuntee that the commit returned still exists as the
+   * <p><b>NOTE:</b> This method makes no guarantee that the commit returned still exists as the
    * moment this method completes. Callers are encouraged to use {@link #getAndSaveLatestCommit}
    * instead.
    *
@@ -402,7 +402,7 @@ public final class IndexDeletionPolicyWrapper extends IndexDeletionPolicy {
    * <p><b>NOTE:</b> This map instance may change between commits and commit points may be deleted.
    * This API is intended for "informational purposes" only, to provide an "at the moment" view of
    * the current list of known commits. Callers that need to ensure commits exist for an extended
-   * period must wrap this call and all subsequent usage of the results in a synchornization block.
+   * period must wrap this call and all subsequent usage of the results in a synchronization block.
    *
    * @return a Map of generation to commit points
    */
@@ -419,32 +419,32 @@ public final class IndexDeletionPolicyWrapper extends IndexDeletionPolicy {
    * #latestCommit} from being deleted.
    *
    * <p>If we did not do this, and waited to update <code>latestCommit</code> in <code>
-   * updateKnownCommitPoints()</code> then we would need to wrap synchronization completley around
+   * updateKnownCommitPoints()</code> then we would need to wrap synchronization completely around
    * the (delegated) <code>onInit()</code> and <code>onCommit()</code> calls, to ensure there was no
    * window of time when {@link #getAndSaveLatestCommit} might return the "old" latest commit, after
    * our delegate Policy had already deleted it.
    *
    * <p>(Since Saving/Reserving (other) commits is handled indirectly ("by reference") via the
    * generation callers can still safely (try) to reserve "old" commits using an explicit generation
-   * since {@link IndexCommitWrapper#delete} is synchornized on <code>this</code>)
+   * since {@link IndexCommitWrapper#delete} is synchronized on <code>this</code>)
    *
    * @see #latestCommit
    * @see #updateKnownCommitPoints
    */
   private synchronized void updateLatestCommit(final List<IndexCommitWrapper> list) {
-    // NOTE: There's a hypothetical, not neccessarily possible/plausible, situation that
+    // NOTE: There's a hypothetical, not necessarily possible/plausible, situation that
     // could lead to this combination of updateLatestCommit + updateKnownCommitPoints not
-    // being as thread safe as completley synchornizing in onInit/onCommit...
+    // being as thread safe as completely synchronizing in onInit/onCommit...
     //  - knownCommits==(1, 2, 3, 4), latestCommit==4
     //  - onCommit(1, 2, 3, 4, 5, 6, 7) - we immediately update latestCommit=7
     //    - before knownCommits is updated, some client calls getAndSaveCommitPoint(6)
     //      - call fails "too old to be saved" even though it's in flight
     // (this assumes some future caller/use-case that doesn't currently exist)
     //
-    // The upside of this current approach, and not completley synchornizing onInit/onCommit
+    // The upside of this current approach, and not completely synchronizing onInit/onCommit
     // is that we have no control over what delegate is used, or how long those calls might take.
     //
-    // If the hypotehtical situation above ever becomes problematic, then an alternative approach
+    // If the hypothetical situation above ever becomes problematic, then an alternative approach
     // might be to *add* to the Set/Map of all known commits *before* delegating, then *remove*
     // everything except the new (non-deleted) commits *after* delegating.
 
@@ -463,7 +463,7 @@ public final class IndexDeletionPolicyWrapper extends IndexDeletionPolicy {
    * Updates the state of all "current" commits.
    *
    * <p>This method is safe to call <em>after</em> delegating to ou inner <code>IndexDeletionPolicy
-   * </code> (w/o synchornizing the delegate calls) because even if the delegate decides to {@link
+   * </code> (w/o synchronizing the delegate calls) because even if the delegate decides to {@link
    * IndexCommit#delete} a commit that a concurrent thread may wish to reserve/save, that {@link
    * IndexCommitWrapper} will ensure that call is synchronized.
    *
@@ -487,7 +487,7 @@ public final class IndexDeletionPolicyWrapper extends IndexDeletionPolicy {
   }
 
   /**
-   * Helper method for unpacking the timestamp infor from the user data
+   * Helper method for unpacking the timestamp info from the user data
    *
    * @see SolrIndexWriter#COMMIT_TIME_MSEC_KEY
    * @see IndexCommit#getUserData
diff --git a/solr/core/src/java/org/apache/solr/core/NodeConfig.java b/solr/core/src/java/org/apache/solr/core/NodeConfig.java
index 9006951e3dc..8117aad4523 100644
--- a/solr/core/src/java/org/apache/solr/core/NodeConfig.java
+++ b/solr/core/src/java/org/apache/solr/core/NodeConfig.java
@@ -413,8 +413,8 @@ public class NodeConfig {
    * properties, or from the "extra" properties configured explicitly on the SolrDispatchFilter; or
    * null if not specified.
    *
-   * <p>This is the value that would have been used when attempting locate the solr.xml in ZooKeeper
-   * (regardless of wether the file was actaully loaded from ZK or from local disk)
+   * <p>This is the value that would have been used when attempting to locate the solr.xml in
+   * ZooKeeper (regardless of whether the file was actually loaded from ZK or from local disk)
    *
    * <p>(This value should only be used for "accounting" purposes to track where the node config
    * came from if it <em>was</em> loaded from zk -- ie: to check if the chroot has already been
diff --git a/solr/core/src/java/org/apache/solr/core/PluginBag.java b/solr/core/src/java/org/apache/solr/core/PluginBag.java
index f583fc03bd2..c70bdd4af9d 100644
--- a/solr/core/src/java/org/apache/solr/core/PluginBag.java
+++ b/solr/core/src/java/org/apache/solr/core/PluginBag.java
@@ -318,7 +318,7 @@ public class PluginBag<T> implements AutoCloseable {
   }
 
   /**
-   * Initializes the plugins after reading the meta data from {@link
+   * Initializes the plugins after reading the metadata from {@link
    * org.apache.solr.core.SolrConfig}.
    *
    * @param defaults These will be registered if not explicitly specified
diff --git a/solr/core/src/java/org/apache/solr/core/PluginInfo.java b/solr/core/src/java/org/apache/solr/core/PluginInfo.java
index 0331b084045..5c24ee9a8f4 100644
--- a/solr/core/src/java/org/apache/solr/core/PluginInfo.java
+++ b/solr/core/src/java/org/apache/solr/core/PluginInfo.java
@@ -252,7 +252,7 @@ public class PluginInfo implements MapSerializable {
    * Filter children by type
    *
    * @param type The type name. must not be null
-   * @return The mathcing children
+   * @return The matching children
    */
   public List<PluginInfo> getChildren(String type) {
     if (children.isEmpty()) return children;
diff --git a/solr/core/src/java/org/apache/solr/core/SolrCore.java b/solr/core/src/java/org/apache/solr/core/SolrCore.java
index dd8032a4e88..661e9f1e29f 100644
--- a/solr/core/src/java/org/apache/solr/core/SolrCore.java
+++ b/solr/core/src/java/org/apache/solr/core/SolrCore.java
@@ -369,7 +369,7 @@ public class SolrCore implements SolrInfoBean, Closeable {
    */
   public void setLatestSchema(IndexSchema replacementSchema) {
     // 1) For a newly instantiated core, the Similarity needs SolrCore before inform() is called on
-    // any registered SolrCoreAware listeners (which will likeley need to use the SolrIndexSearcher.
+    // any registered SolrCoreAware listeners (which will likely need to use the SolrIndexSearcher.
     //
     // 2) If a new IndexSchema is assigned to an existing live SolrCore (ie: managed schema
     // replacement via SolrCloud) then we need to explicitly inform() the similarity because
@@ -404,12 +404,12 @@ public class SolrCore implements SolrInfoBean, Closeable {
 
   /**
    * Returns the indexdir as given in index.properties. If index.properties exists in dataDir and
-   * there is a property <i>index</i> available and it points to a valid directory in dataDir that
+   * there is a property <i>index</i> available, and it points to a valid directory in dataDir that
    * is returned. Else dataDir/index is returned. Only called for creating new indexSearchers and
    * indexwriters. Use the getIndexDir() method to know the active index directory
    *
    * @return the indexdir as given in index.properties
-   * @throws SolrException if for any reason the a reasonable index directory cannot be determined.
+   * @throws SolrException if for any reason an index directory cannot be determined.
    */
   public String getNewIndexDir() {
     Directory dir = null;
@@ -427,7 +427,7 @@ public class SolrCore implements SolrInfoBean, Closeable {
       String errorMessage = "Error in getNewIndexDir, exception: ";
       log.error(errorMessage, e);
       // See SOLR-11687. It is inadvisable to assume we can do the right thing for any but a small
-      // number of exceptions that ware caught and swallowed in getIndexProperty.
+      // number of exceptions that were caught and swallowed in getIndexProperty.
       throw new SolrException(ErrorCode.SERVER_ERROR, errorMessage, e);
     } finally {
       if (dir != null) {
@@ -1412,9 +1412,9 @@ public class SolrCore implements SolrInfoBean, Closeable {
       // we are evidently running in cloud mode.
       //
       // In cloud mode, version field is required for correct consistency
-      // ideally this check would be more fine grained, and individual features
+      // ideally this check would be more fine-grained, and individual features
       // would assert it when they initialize, but DistributedUpdateProcessor
-      // is currently a big ball of wax that does more then just distributing
+      // is currently a big ball of wax that does more than just distributing
       // updates (ie: partial document updates), so it needs to work in no cloud
       // mode as well, and can't assert version field support on init.
 
@@ -1441,7 +1441,7 @@ public class SolrCore implements SolrInfoBean, Closeable {
    * @param dataDir An optional hint to the data directory location. Will be normalized and used if
    *     not null.
    * @param config A solr config to retrieve the default data directory location, if used.
-   * @param coreDescriptor descriptor to load the actual data dir from, if not using the defualt.
+   * @param coreDescriptor descriptor to load the actual data dir from, if not using the default.
    * @return a normalized data directory name
    * @throws SolrException if the data directory cannot be loaded from the core descriptor
    */
@@ -2514,7 +2514,7 @@ public class SolrCore implements SolrInfoBean, Closeable {
     // if it isn't necessary.
 
     synchronized (searcherLock) {
-      for (; ; ) { // this loop is so w can retry in the event that we exceed maxWarmingSearchers
+      for (; ; ) { // this loop is so we can retry in the event that we exceed maxWarmingSearchers
         // see if we can return the current searcher
         if (_searcher != null && !forceNew) {
           if (returnSearcher) {
@@ -2531,7 +2531,7 @@ public class SolrCore implements SolrInfoBean, Closeable {
             searcherLock.wait();
           } catch (InterruptedException e) {
             if (log.isInfoEnabled()) {
-              log.info("Interupted waiting for searcherLock", e);
+              log.info("Interrupted waiting for searcherLock", e);
             }
           }
         }
@@ -2562,7 +2562,7 @@ public class SolrCore implements SolrInfoBean, Closeable {
             searcherLock.wait();
           } catch (InterruptedException e) {
             if (log.isInfoEnabled()) {
-              log.info("Interupted waiting for searcherLock", e);
+              log.info("Interrupted waiting for searcherLock", e);
             }
           }
           continue; // go back to the top of the loop and retry
@@ -2786,7 +2786,7 @@ public class SolrCore implements SolrInfoBean, Closeable {
   }
 
   // Take control of newSearcherHolder (which should have a reference count of at
-  // least 1 already.  If the caller wishes to use the newSearcherHolder directly
+  // least 1 already). If the caller wishes to use the newSearcherHolder directly
   // after registering it, then they should increment the reference count *before*
   // calling this method.
   //
@@ -3396,20 +3396,20 @@ public class SolrCore implements SolrInfoBean, Closeable {
       ManagedIndexSchema mis = (ManagedIndexSchema) core.getLatestSchema();
       schemaRes = mis.getResourceName();
     }
-    final String managedSchmaResourcePath =
+    final String managedSchemaResourcePath =
         schemaRes == null ? null : zkSolrResourceLoader.getConfigSetZkPath() + "/" + schemaRes;
     return () -> {
       log.info("config update listener called for core {}", coreName);
       SolrZkClient zkClient = cc.getZkController().getZkClient();
-      int solrConfigversion, overlayVersion, managedSchemaVersion = 0;
+      int solrConfigVersion, overlayVersion, managedSchemaVersion = 0;
       SolrConfig cfg = null;
       try (SolrCore solrCore = cc.solrCores.getCoreFromAnyList(coreName, true, coreId)) {
         if (solrCore == null || solrCore.isClosed() || solrCore.getCoreContainer().isShutDown())
           return;
         cfg = solrCore.getSolrConfig();
-        solrConfigversion = solrCore.getSolrConfig().getZnodeVersion();
+        solrConfigVersion = solrCore.getSolrConfig().getZnodeVersion();
         overlayVersion = solrCore.getSolrConfig().getOverlay().getVersion();
-        if (managedSchmaResourcePath != null) {
+        if (managedSchemaResourcePath != null) {
           managedSchemaVersion =
               ((ManagedIndexSchema) solrCore.getLatestSchema()).getSchemaZkVersion();
         }
@@ -3418,8 +3418,8 @@ public class SolrCore implements SolrInfoBean, Closeable {
         cfg.refreshRequestParams();
       }
       if (checkStale(zkClient, overlayPath, overlayVersion)
-          || checkStale(zkClient, solrConfigPath, solrConfigversion)
-          || checkStale(zkClient, managedSchmaResourcePath, managedSchemaVersion)) {
+          || checkStale(zkClient, solrConfigPath, solrConfigVersion)
+          || checkStale(zkClient, managedSchemaResourcePath, managedSchemaVersion)) {
         log.info("core reload {}", coreName);
         SolrConfigHandler configHandler = ((SolrConfigHandler) core.getRequestHandler("/config"));
         if (configHandler.getReloadLock().tryLock()) {
@@ -3580,19 +3580,19 @@ public class SolrCore implements SolrInfoBean, Closeable {
   }
 
   /**
-   * Run an arbitrary task in it's own thread. This is an expert option and is a method you should
+   * Run an arbitrary task in its own thread. This is an expert option and is a method you should
    * use with great care. It would be bad to run something that never stopped or run something that
-   * took a very long time. Typically this is intended for actions that take a few seconds, and
-   * therefore would be bad to wait for within a request, but but would not pose a significant
-   * hindrance to server shut down times. It is not intended for long running tasks and if you are
-   * using a Runnable with a loop in it, you are almost certainly doing it wrong.
+   * took a very long time. Typically, this is intended for actions that take a few seconds, and
+   * therefore would be bad to wait for within a request, but would not pose a significant hindrance
+   * to server shut down times. It is not intended for long-running tasks and if you are using a
+   * Runnable with a loop in it, you are almost certainly doing it wrong.
    *
    * <p>WARNING: Solr wil not be able to shut down gracefully until this task completes!
    *
    * <p>A significant upside of using this method vs creating your own ExecutorService is that your
    * code does not have to properly shutdown executors which typically is risky from a unit testing
    * perspective since the test framework will complain if you don't carefully ensure the executor
-   * shuts down before the end of the test. Also the threads running this task are sure to have a
+   * shuts down before the end of the test. Also, the threads running this task are sure to have a
    * proper MDC for logging.
    *
    * @param r the task to run
diff --git a/solr/core/src/java/org/apache/solr/core/SolrDeletionPolicy.java b/solr/core/src/java/org/apache/solr/core/SolrDeletionPolicy.java
index 50b46211aac..2b8232deef0 100644
--- a/solr/core/src/java/org/apache/solr/core/SolrDeletionPolicy.java
+++ b/solr/core/src/java/org/apache/solr/core/SolrDeletionPolicy.java
@@ -32,7 +32,7 @@ import org.slf4j.LoggerFactory;
 
 /**
  * Standard Solr deletion policy that allows reserving index commit points for certain amounts of
- * time to support features such as index replication or snapshooting directly out of a live index
+ * time to support features such as index replication or snapshotting directly out of a live index
  * directory.
  *
  * @see org.apache.lucene.index.IndexDeletionPolicy
@@ -143,7 +143,7 @@ public class SolrDeletionPolicy extends IndexDeletionPolicy implements NamedList
 
   private void updateCommits(List<? extends IndexCommit> commits) {
     // to be safe, we should only call delete on a commit point passed to us
-    // in this specific call (may be across diff IndexWriter instances).
+    // in this specific call (which may be across diff IndexWriter instances).
     // this will happen rarely, so just synchronize everything
     // for safety and to avoid race conditions
 
diff --git a/solr/core/src/java/org/apache/solr/core/SolrResourceLoader.java b/solr/core/src/java/org/apache/solr/core/SolrResourceLoader.java
index 7017ce47f93..e82a05d66b2 100644
--- a/solr/core/src/java/org/apache/solr/core/SolrResourceLoader.java
+++ b/solr/core/src/java/org/apache/solr/core/SolrResourceLoader.java
@@ -812,7 +812,7 @@ public class SolrResourceLoader
         SolrCoreAware.class,
         new Class<?>[] {
           // DO NOT ADD THINGS TO THIS LIST -- ESPECIALLY THINGS THAT CAN BE CREATED DYNAMICALLY
-          // VIA RUNTIME APIS -- UNTILL CAREFULLY CONSIDERING THE ISSUES MENTIONED IN SOLR-8311
+          // VIA RUNTIME APIS -- UNTIL CAREFULLY CONSIDERING THE ISSUES MENTIONED IN SOLR-8311
           CodecFactory.class,
           DirectoryFactory.class,
           ManagedIndexSchemaFactory.class,
@@ -828,7 +828,7 @@ public class SolrResourceLoader
         ResourceLoaderAware.class,
         new Class<?>[] {
           // DO NOT ADD THINGS TO THIS LIST -- ESPECIALLY THINGS THAT CAN BE CREATED DYNAMICALLY
-          // VIA RUNTIME APIS -- UNTILL CAREFULLY CONSIDERING THE ISSUES MENTIONED IN SOLR-8311
+          // VIA RUNTIME APIS -- UNTIL CAREFULLY CONSIDERING THE ISSUES MENTIONED IN SOLR-8311
           // evaluate if this must go into schemaResourceLoaderComponents
           CharFilterFactory.class,
           TokenFilterFactory.class,
@@ -918,7 +918,7 @@ public class SolrResourceLoader
   }
 
   /**
-   * Create a n instance of a class using an appropriate {@link SolrResourceLoader} depending on the
+   * Create an instance of a class using an appropriate {@link SolrResourceLoader} depending on the
    * package of that class
    *
    * @param registerCoreReloadListener register a listener for the package and reload the core if
diff --git a/solr/core/src/java/org/apache/solr/core/SolrXmlConfig.java b/solr/core/src/java/org/apache/solr/core/SolrXmlConfig.java
index f6d1eb78d93..676026c82ed 100644
--- a/solr/core/src/java/org/apache/solr/core/SolrXmlConfig.java
+++ b/solr/core/src/java/org/apache/solr/core/SolrXmlConfig.java
@@ -70,13 +70,13 @@ public class SolrXmlConfig {
   private static final Pattern COMMA_SEPARATED_PATTERN = Pattern.compile("\\s*,\\s*");
 
   /**
-   * Given some node Properties, checks if non-null and a 'zkHost' is alread included. If so, the
+   * Given some node Properties, checks if non-null and a 'zkHost' is already included. If so, the
    * Properties are returned as is. If not, then the returned value will be a new Properties,
    * wrapping the original Properties, with the 'zkHost' value set based on the value of the
-   * corispond System property (if set)
+   * corresponding System property (if set)
    *
    * <p>In theory we only need this logic once, ideally in SolrDispatchFilter, but we put it here to
-   * re-use redundently because of how much surface area our API has for various tests to poke at
+   * re-use redundantly because of how much surface area our API has for various tests to poke at
    * us.
    */
   public static Properties wrapAndSetZkHostFromSysPropIfNeeded(final Properties props) {
@@ -133,7 +133,7 @@ public class SolrXmlConfig {
     String nodeName = (String) entries.remove("nodeName");
     if (StrUtils.isNullOrEmpty(nodeName) && cloudConfig != null) nodeName = cloudConfig.getHost();
 
-    // It should goes inside the fillSolrSection method but
+    // It should go inside the fillSolrSection method but
     // since it is arranged as a separate section it is placed here
     Map<String, String> coreAdminHandlerActions =
         readNodeListAsNamedList(root.get("coreAdminHandlerActions"), "<coreAdminHandlerActions>")
diff --git a/solr/core/src/java/org/apache/solr/core/TracerConfigurator.java b/solr/core/src/java/org/apache/solr/core/TracerConfigurator.java
index b730c5741e6..ba36ed127ec 100644
--- a/solr/core/src/java/org/apache/solr/core/TracerConfigurator.java
+++ b/solr/core/src/java/org/apache/solr/core/TracerConfigurator.java
@@ -52,7 +52,7 @@ public abstract class TracerConfigurator implements NamedListInitializedPlugin {
     if (GlobalTracer.isRegistered()) {
       // ideally we would furthermore check that it's not a no-op impl either but
       //  GlobalTracer.get() always returns a GlobalTracer implementing Tracer that delegates
-      //  to the real Tracer (that may or may not be a No-Op impl.
+      //  to the real Tracer (that may or may not be a No-Op impl).
       ExecutorUtil.addThreadLocalProvider(new TracerConfigurator.SpanThreadLocalProvider());
     }
 
diff --git a/solr/core/src/java/org/apache/solr/core/TransientSolrCoreCache.java b/solr/core/src/java/org/apache/solr/core/TransientSolrCoreCache.java
index e0119635eda..51a625f2b7f 100644
--- a/solr/core/src/java/org/apache/solr/core/TransientSolrCoreCache.java
+++ b/solr/core/src/java/org/apache/solr/core/TransientSolrCoreCache.java
@@ -21,7 +21,7 @@ import java.util.Collection;
 import java.util.Set;
 
 /**
- * The base class for custom transient core maintenance. Any custom plugin that want's to take
+ * The base class for custom transient core maintenance. Any custom plugin that wants to take
  * control of transient caches (i.e. any core defined with transient=true) should override this
  * class.
  *
diff --git a/solr/core/src/java/org/apache/solr/core/XmlConfigFile.java b/solr/core/src/java/org/apache/solr/core/XmlConfigFile.java
index 554d0c56a6f..bc6ef3650c7 100644
--- a/solr/core/src/java/org/apache/solr/core/XmlConfigFile.java
+++ b/solr/core/src/java/org/apache/solr/core/XmlConfigFile.java
@@ -399,7 +399,7 @@ public class XmlConfigFile { // formerly simply "Config"
     return val != null ? Double.parseDouble(val) : def;
   }
 
-  /** If this config is loaded from zk the version is relevant other wise -1 is returned */
+  /** If this config is loaded from zk the version is relevant otherwise -1 is returned */
   public int getZnodeVersion() {
     return zkVersion;
   }
diff --git a/solr/core/src/java/org/apache/solr/core/backup/repository/BackupRepository.java b/solr/core/src/java/org/apache/solr/core/backup/repository/BackupRepository.java
index b3d2034d4cf..0d968fde0b2 100644
--- a/solr/core/src/java/org/apache/solr/core/backup/repository/BackupRepository.java
+++ b/solr/core/src/java/org/apache/solr/core/backup/repository/BackupRepository.java
@@ -171,7 +171,7 @@ public interface BackupRepository extends NamedListInitializedPlugin, Closeable
    * Copy a file from specified <code>sourceDir</code> to the destination repository (i.e. backup).
    *
    * @param sourceDir The source directory hosting the file to be copied.
-   * @param fileName The name of the file to by copied
+   * @param fileName The name of the file to be copied
    * @param dest The destination backup location.
    * @throws IOException in case of errors
    */
@@ -183,7 +183,7 @@ public interface BackupRepository extends NamedListInitializedPlugin, Closeable
    * Copy a file from specified <code>sourceRepo</code> to the destination directory (i.e. restore).
    *
    * @param sourceRepo The source URI hosting the file to be copied.
-   * @param fileName The name of the file to by copied
+   * @param fileName The name of the file to be copied
    * @param dest The destination where the file should be copied.
    * @throws IOException in case of errors.
    */
@@ -262,7 +262,7 @@ public interface BackupRepository extends NamedListInitializedPlugin, Closeable
    * backup).
    *
    * @param sourceDir The source directory hosting the file to be copied.
-   * @param sourceFileName The name of the file to by copied
+   * @param sourceFileName The name of the file to be copied
    * @param destDir The destination backup location.
    * @throws IOException in case of errors
    * @throws CorruptIndexException in case checksum of the file does not match with precomputed
diff --git a/solr/core/src/java/org/apache/solr/core/backup/repository/BackupRepositoryFactory.java b/solr/core/src/java/org/apache/solr/core/backup/repository/BackupRepositoryFactory.java
index 17fc9b4bd76..5d7be935d06 100644
--- a/solr/core/src/java/org/apache/solr/core/backup/repository/BackupRepositoryFactory.java
+++ b/solr/core/src/java/org/apache/solr/core/backup/repository/BackupRepositoryFactory.java
@@ -78,7 +78,7 @@ public class BackupRepositoryFactory {
     BackupRepository result = loader.newInstance(repo.className, BackupRepository.class);
     if ("trackingBackupRepository".equals(name)) {
       // newInstance can be called by multiple threads, synchronization prevents simultaneous
-      // multi-threaded 'adds' from corrupting the namedlist
+      // multithreaded 'adds' from corrupting the NamedList
       synchronized (repo.initArgs) {
         if (repo.initArgs.get("factory") == null) {
           repo.initArgs.add("factory", this);