You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@solr.apache.org by ho...@apache.org on 2023/07/27 18:54:10 UTC
[solr] branch branch_9x updated: MINOR: Fix typos in `org.apache.solr.core` package (#1629)
This is an automated email from the ASF dual-hosted git repository.
houston pushed a commit to branch branch_9x
in repository https://gitbox.apache.org/repos/asf/solr.git
The following commit(s) were added to refs/heads/branch_9x by this push:
new 38fb272cabf MINOR: Fix typos in `org.apache.solr.core` package (#1629)
38fb272cabf is described below
commit 38fb272cabf0294e4dad4f19c2ddc3d96abf35b0
Author: Andrey Bozhko <an...@gmail.com>
AuthorDate: Mon May 8 14:26:23 2023 -0500
MINOR: Fix typos in `org.apache.solr.core` package (#1629)
* solr core - fix typos
* solr core - fix typos
* solr core - fix typos
* run ./gradlew tidy
* Update solr/core/src/java/org/apache/solr/core/backup/repository/BackupRepositoryFactory.java
same capitalization
* a few typos.
* add CHANGES entry
* revert "add CHANGES entry"
This reverts commit 07533d68619dfd928a70c91066bcbe39aaf1c653.
* add CHANGES entry
---------
Co-authored-by: Marcus <ma...@gmail.com>
(cherry picked from commit 810783d812e53aa6a382e082a0953eb1ca2bfa28)
---
solr/CHANGES.txt | 2 +
.../apache/solr/core/CachingDirectoryFactory.java | 4 +-
.../src/java/org/apache/solr/core/CloseHook.java | 4 +-
.../java/org/apache/solr/core/ConfigOverlay.java | 2 +-
.../src/java/org/apache/solr/core/ConfigSet.java | 2 +-
.../java/org/apache/solr/core/CoreContainer.java | 36 +++++++--------
.../java/org/apache/solr/core/CoreDescriptor.java | 2 +-
.../src/java/org/apache/solr/core/CoreSorter.java | 4 +-
.../org/apache/solr/core/DirectoryFactory.java | 10 ++---
.../solr/core/IndexDeletionPolicyWrapper.java | 52 +++++++++++-----------
.../src/java/org/apache/solr/core/NodeConfig.java | 4 +-
.../src/java/org/apache/solr/core/PluginBag.java | 2 +-
.../src/java/org/apache/solr/core/PluginInfo.java | 2 +-
.../src/java/org/apache/solr/core/SolrCore.java | 46 +++++++++----------
.../org/apache/solr/core/SolrDeletionPolicy.java | 4 +-
.../org/apache/solr/core/SolrResourceLoader.java | 6 +--
.../java/org/apache/solr/core/SolrXmlConfig.java | 8 ++--
.../org/apache/solr/core/TracerConfigurator.java | 2 +-
.../apache/solr/core/TransientSolrCoreCache.java | 2 +-
.../java/org/apache/solr/core/XmlConfigFile.java | 2 +-
.../core/backup/repository/BackupRepository.java | 6 +--
.../backup/repository/BackupRepositoryFactory.java | 2 +-
22 files changed, 103 insertions(+), 101 deletions(-)
diff --git a/solr/CHANGES.txt b/solr/CHANGES.txt
index 27a97f54c39..86382a0cf64 100644
--- a/solr/CHANGES.txt
+++ b/solr/CHANGES.txt
@@ -40,6 +40,8 @@ Other Changes
* SOLR-16856: The default auto-soft-commit time has been set to 3 seconds. Previously, auto-soft-commit was disabled by default. (Houston Putman)
+* PR#1629: Fix typos in org.apache.solr.core package (Andrey Bozhko, Marcus Eagan)
+
================== 9.3.0 ==================
Upgrade Notes
diff --git a/solr/core/src/java/org/apache/solr/core/CachingDirectoryFactory.java b/solr/core/src/java/org/apache/solr/core/CachingDirectoryFactory.java
index 9864ad7ee48..8a5766f2f51 100644
--- a/solr/core/src/java/org/apache/solr/core/CachingDirectoryFactory.java
+++ b/solr/core/src/java/org/apache/solr/core/CachingDirectoryFactory.java
@@ -231,7 +231,7 @@ public abstract class CachingDirectoryFactory extends DirectoryFactory {
byPathCache.remove(v.path);
}
- // be sure this is called with the this sync lock
+ // be sure the method is called with the sync lock on this object
// returns true if we closed the cacheValue, false if it will be closed later
private boolean closeCacheValue(CacheValue cacheValue) {
log.debug("looking to close {} {}", cacheValue.path, cacheValue.closeEntries);
@@ -441,7 +441,7 @@ public abstract class CachingDirectoryFactory extends DirectoryFactory {
throw new NullPointerException();
}
synchronized (this) {
- // don't check if already closed here - we need to able to release
+ // don't check if already closed here - we need to be able to release
// while #close() waits.
CacheValue cacheValue = byDirectoryCache.get(directory);
diff --git a/solr/core/src/java/org/apache/solr/core/CloseHook.java b/solr/core/src/java/org/apache/solr/core/CloseHook.java
index 582e91e5b60..3b1ffeeb895 100644
--- a/solr/core/src/java/org/apache/solr/core/CloseHook.java
+++ b/solr/core/src/java/org/apache/solr/core/CloseHook.java
@@ -32,7 +32,7 @@ public interface CloseHook {
* handler and searcher(s) are actually closed <br>
* <b>Important:</b> Keep the method implementation as short as possible. If it were to use any
* heavy i/o , network connections - it might be a better idea to launch in a separate Thread so
- * as to not to block the process of shutting down a given SolrCore instance.
+ * as to not block the process of shutting down a given SolrCore instance.
*
* @param core SolrCore object that is shutting down / closing
*/
@@ -47,7 +47,7 @@ public interface CloseHook {
*
* <p><b>Important:</b> Keep the method implementation as short as possible. If it were to use any
* heavy i/o , network connections - it might be a better idea to launch in a separate Thread so
- * as to not to block the process of shutting down a given SolrCore instance.
+ * as to not block the process of shutting down a given SolrCore instance.
*
* @param core SolrCore object that is shutting down / closing
*/
diff --git a/solr/core/src/java/org/apache/solr/core/ConfigOverlay.java b/solr/core/src/java/org/apache/solr/core/ConfigOverlay.java
index 478ab92f188..403501c19f8 100644
--- a/solr/core/src/java/org/apache/solr/core/ConfigOverlay.java
+++ b/solr/core/src/java/org/apache/solr/core/ConfigOverlay.java
@@ -31,7 +31,7 @@ import org.apache.solr.common.util.Utils;
/**
* This class encapsulates the config overlay json file. It is immutable and any edit operations
- * performed on tbhis gives a new copy of the object with the changed value
+ * performed on this gives a new copy of the object with the changed value
*/
public class ConfigOverlay implements MapSerializable {
private final int version;
diff --git a/solr/core/src/java/org/apache/solr/core/ConfigSet.java b/solr/core/src/java/org/apache/solr/core/ConfigSet.java
index 857fbb30552..de363f69d90 100644
--- a/solr/core/src/java/org/apache/solr/core/ConfigSet.java
+++ b/solr/core/src/java/org/apache/solr/core/ConfigSet.java
@@ -82,7 +82,7 @@ public class ConfigSet {
/**
* Provide a Schema object on demand We want IndexSchema Objects to be lazily instantiated because
* when a configset is created the {@link SolrResourceLoader} associated with it is not associated
- * with a core So, we may not be able to update the core if we the schema classes are updated
+ * with a core So, we may not be able to update the core if the schema classes are updated
*/
interface SchemaSupplier {
IndexSchema get(boolean forceFetch);
diff --git a/solr/core/src/java/org/apache/solr/core/CoreContainer.java b/solr/core/src/java/org/apache/solr/core/CoreContainer.java
index 99f503258b8..f223f41e36f 100644
--- a/solr/core/src/java/org/apache/solr/core/CoreContainer.java
+++ b/solr/core/src/java/org/apache/solr/core/CoreContainer.java
@@ -307,7 +307,7 @@ public class CoreContainer {
ExecutorUtil.newMDCAwareCachedThreadPool("Core Container Async Task");
/**
- * Non empty if the Collection API is executed in a distributed way and not on Overseer, once the
+ * Non-empty if the Collection API is executed in a distributed way and not on Overseer, once the
* CoreContainer has been initialized properly, i.e. method {@link #load()} called. Until then it
* is null, and it is not expected to be read.
*/
@@ -579,15 +579,15 @@ public class CoreContainer {
SolrHttpClientBuilder builder =
builderPlugin.getHttpClientBuilder(HttpClientUtil.getHttpClientBuilder());
- // this caused plugins like KerberosPlugin to register it's intercepts, but this intercept
- // logic is also handled by the pki authentication code when it decideds to let the plugin
- // handle auth via it's intercept - so you would end up with two intercepts
+ // this caused plugins like KerberosPlugin to register its intercepts, but this intercept
+ // logic is also handled by the pki authentication code when it decides to let the plugin
+ // handle auth via its intercept - so you would end up with two intercepts
// -->
// shardHandlerFactory.setSecurityBuilder(builderPlugin); // calls setup for the authcPlugin
// updateShardHandler.setSecurityBuilder(builderPlugin);
// <--
- // This should not happen here at all - it's only currently required due to its affect on
+ // This should not happen here at all - it's only currently required due to its effect on
// http1 clients in a test or two incorrectly counting on it for their configuration.
// -->
@@ -1400,8 +1400,8 @@ public class CoreContainer {
* preserve consistency.
*
* <p>Jetty already allows a grace period for in-flight requests to complete and our solr cores,
- * searchers etc are reference counted to allow for graceful shutdown. So we don't worry about any
- * other kind of requests.
+ * searchers, etc., are reference counted to allow for graceful shutdown. So we don't worry about
+ * any other kind of requests.
*
* <p>We do not need to unpause ever because the node is being shut down.
*/
@@ -1514,14 +1514,14 @@ public class CoreContainer {
assertPathAllowed(cd.getInstanceDir());
assertPathAllowed(Paths.get(cd.getDataDir()));
- boolean preExisitingZkEntry = false;
+ boolean preExistingZkEntry = false;
try {
if (getZkController() != null) {
if (cd.getCloudDescriptor().getCoreNodeName() == null) {
throw new SolrException(
ErrorCode.SERVER_ERROR, "coreNodeName missing " + parameters.toString());
}
- preExisitingZkEntry = getZkController().checkIfCoreNodeNameAlreadyExists(cd);
+ preExistingZkEntry = getZkController().checkIfCoreNodeNameAlreadyExists(cd);
}
// Much of the logic in core handling pre-supposes that the core.properties file already
@@ -1544,7 +1544,7 @@ public class CoreContainer {
// First clean up any core descriptor, there should never be an existing core.properties
// file for any core that failed to be created on-the-fly.
coresLocator.delete(this, cd);
- if (isZooKeeperAware() && !preExisitingZkEntry) {
+ if (isZooKeeperAware() && !preExistingZkEntry) {
try {
getZkController().unregister(coreName, cd);
} catch (InterruptedException e) {
@@ -1797,7 +1797,7 @@ public class CoreContainer {
}
}
- /** Write a new index directory for the a SolrCore, but do so without loading it. */
+ /** Write a new index directory for the SolrCore, but do so without loading it. */
private void resetIndexDirectory(CoreDescriptor dcore, ConfigSet coreConfig) {
SolrConfig config = coreConfig.getSolrConfig();
@@ -2087,7 +2087,7 @@ public class CoreContainer {
// check for core-init errors first
CoreLoadFailure loadFailure = coreInitFailures.remove(name);
if (loadFailure != null) {
- // getting the index directory requires opening a DirectoryFactory with a SolrConfig, etc,
+ // getting the index directory requires opening a DirectoryFactory with a SolrConfig, etc.,
// which we may not be able to do because of the init error. So we just go with what we
// can glean from the CoreDescriptor - datadir and instancedir
SolrCore.deleteUnloadedCore(loadFailure.cd, deleteDataDir, deleteInstanceDir);
@@ -2236,7 +2236,7 @@ public class CoreContainer {
// load failure (use reload to cure that). But for
// TestConfigSetsAPI.testUploadWithScriptUpdateProcessor, this needs to _not_ try to load the
// core if the core is null and there was an error. If you change this, be sure to run both
- // TestConfiSetsAPI and TestLazyCores
+ // TestConfigSetsAPI and TestLazyCores
if (desc == null || zkSys.getZkController() != null) return null;
// This will put an entry in pending core ops if the core isn't loaded. Here's where moving the
@@ -2482,13 +2482,13 @@ public class CoreContainer {
}
/**
- * Run an arbitrary task in it's own thread. This is an expert option and is a method you should
+ * Run an arbitrary task in its own thread. This is an expert option and is a method you should
* use with great care. It would be bad to run something that never stopped or run something that
* took a very long time. Typically this is intended for actions that take a few seconds, and
* therefore would be bad to wait for within a request, or actions that need to happen when a core
- * has zero references, but but would not pose a significant hindrance to server shut down times.
- * It is not intended for long running tasks and if you are using a Runnable with a loop in it,
- * you are almost certainly doing it wrong.
+ * has zero references, but would not pose a significant hindrance to server shut down times. It
+ * is not intended for long-running tasks and if you are using a Runnable with a loop in it, you
+ * are almost certainly doing it wrong.
*
* <p><br>
* WARNING: Solr wil not be able to shut down gracefully until this task completes!
@@ -2553,7 +2553,7 @@ class CloserThread extends Thread {
}
// It's important that this be the _only_ thread removing things from pendingDynamicCloses!
- // This is single-threaded, but I tried a multi-threaded approach and didn't see any performance
+ // This is single-threaded, but I tried a multithreaded approach and didn't see any performance
// gains, so there's no good justification for the complexity. I suspect that the locking on
// things like DefaultSolrCoreState essentially create a single-threaded process anyway.
@Override
diff --git a/solr/core/src/java/org/apache/solr/core/CoreDescriptor.java b/solr/core/src/java/org/apache/solr/core/CoreDescriptor.java
index ecfb8b55f20..5c4744baeba 100644
--- a/solr/core/src/java/org/apache/solr/core/CoreDescriptor.java
+++ b/solr/core/src/java/org/apache/solr/core/CoreDescriptor.java
@@ -256,7 +256,7 @@ public class CoreDescriptor {
}
/**
- * Create the properties object used by resource loaders, etc, for property substitution. The
+ * Create the properties object used by resource loaders, etc., for property substitution. The
* default solr properties are prefixed with 'solr.core.', so, e.g., 'name' becomes
* 'solr.core.name'
*/
diff --git a/solr/core/src/java/org/apache/solr/core/CoreSorter.java b/solr/core/src/java/org/apache/solr/core/CoreSorter.java
index c90a3d6540a..399526dfa65 100644
--- a/solr/core/src/java/org/apache/solr/core/CoreSorter.java
+++ b/solr/core/src/java/org/apache/solr/core/CoreSorter.java
@@ -52,14 +52,14 @@ public final class CoreSorter implements Comparator<CoreDescriptor> {
// with 0 down nodes than 1 down node because it will make the shard
// complete earlier and avoid waiting by the other live nodes
if (c1.totalReplicasInLiveNodes > 0) {
- // means nobody else is waiting for this , so no need to prioritize
+ // means nobody else is waiting for this, so no need to prioritize
return -1;
}
}
if (c2.totalReplicasInDownNodes < c1.totalReplicasInDownNodes) {
// same is the above, just to take care of the case where c2 has to be prioritized
if (c2.totalReplicasInLiveNodes > 0) {
- // means nobody else is waiting for this , so no need to priotitize
+ // means nobody else is waiting for this, so no need to prioritize
return 1;
}
}
diff --git a/solr/core/src/java/org/apache/solr/core/DirectoryFactory.java b/solr/core/src/java/org/apache/solr/core/DirectoryFactory.java
index 34bed4dc50e..e1d22572bb5 100644
--- a/solr/core/src/java/org/apache/solr/core/DirectoryFactory.java
+++ b/solr/core/src/java/org/apache/solr/core/DirectoryFactory.java
@@ -53,7 +53,7 @@ public abstract class DirectoryFactory implements NamedListInitializedPlugin, Cl
protected static final String INDEX_W_TIMESTAMP_REGEX =
"index\\.[0-9]{17}"; // see SnapShooter.DATE_FMT
- // May be set by sub classes as data root, in which case getDataHome will use it as base.
+ // May be set by subclasses as data root, in which case getDataHome will use it as base.
// Absolute.
protected Path dataHomePath;
@@ -86,7 +86,7 @@ public abstract class DirectoryFactory implements NamedListInitializedPlugin, Cl
public abstract void addCloseListener(Directory dir, CloseListener closeListener);
/**
- * Close the this and all of the Directories it contains.
+ * Close this factory and all of the Directories it contains.
*
* @throws IOException If there is a low-level I/O error.
*/
@@ -196,7 +196,7 @@ public abstract class DirectoryFactory implements NamedListInitializedPlugin, Cl
fromDir.deleteFile(fileName);
}
- // sub classes perform an atomic rename if possible, otherwise fall back to delete + rename
+ // subclasses perform an atomic rename if possible, otherwise fall back to delete + rename
// this is important to support for index roll over durability after crashes
public void renameWithOverwrite(Directory dir, String fileName, String toName)
throws IOException {
@@ -321,7 +321,7 @@ public abstract class DirectoryFactory implements NamedListInitializedPlugin, Cl
* Get the data home folder. If solr.data.home is set, that is used, else base on instanceDir
*
* @param cd core descriptor instance
- * @return a String with absolute path to data direcotry
+ * @return a String with absolute path to data directory
*/
public String getDataHome(CoreDescriptor cd) throws IOException {
Path dataDir;
@@ -391,7 +391,7 @@ public abstract class DirectoryFactory implements NamedListInitializedPlugin, Cl
}
}
- // Extension point to allow sub-classes to infuse additional code when deleting old index
+ // Extension point to allow subclasses to infuse additional code when deleting old index
// directories
protected boolean deleteOldIndexDirectory(String oldDirPath) throws IOException {
Path dirToRm = Path.of(oldDirPath);
diff --git a/solr/core/src/java/org/apache/solr/core/IndexDeletionPolicyWrapper.java b/solr/core/src/java/org/apache/solr/core/IndexDeletionPolicyWrapper.java
index 8089372713e..130e41a8317 100644
--- a/solr/core/src/java/org/apache/solr/core/IndexDeletionPolicyWrapper.java
+++ b/solr/core/src/java/org/apache/solr/core/IndexDeletionPolicyWrapper.java
@@ -42,7 +42,7 @@ import org.slf4j.LoggerFactory;
* A wrapper for an IndexDeletionPolicy instance.
*
* <p>Provides features for looking up IndexCommit given a version. Allows reserving index commit
- * points for certain amounts of time to support features such as index replication or snapshooting
+ * points for certain amounts of time to support features such as index replication or snapshotting
* directly out of a live index directory.
*
* <p><b>NOTE</b>: The {@link #clone()} method returns <code>this</code> in order to make this
@@ -72,7 +72,7 @@ public final class IndexDeletionPolicyWrapper extends IndexDeletionPolicy {
private volatile Map<Long, IndexCommit> knownCommits = new ConcurrentHashMap<>();
/**
- * The most recent commit included in call to {@link #onInit} or {@link #onCommit} <em>beore</em>
+ * The most recent commit included in call to {@link #onInit} or {@link #onCommit} <em>before</em>
* delegating to our {@link #getWrappedDeletionPolicy()}.
*
* <p><b>NOTE:</b> This may be null if there is not yet a single commit to our index.
@@ -85,7 +85,7 @@ public final class IndexDeletionPolicyWrapper extends IndexDeletionPolicy {
private volatile IndexCommit latestCommit;
/**
- * The set of all commit generations htat have been reserved for for some amount of time.
+ * The set of all commit generations that have been reserved for some amount of time.
*
* <p>The keys of the {@link IndexCommit#getGeneration} of a commit, the values are the {@link
* System#nanoTime} that the commit should be reserved until.
@@ -117,7 +117,7 @@ public final class IndexDeletionPolicyWrapper extends IndexDeletionPolicy {
/**
* Returns the most recent commit point.
*
- * <p><b>NOTE:</b> This method makes no garuntee that the commit returned still exists as the
+ * <p><b>NOTE:</b> This method makes no guarantee that the commit returned still exists as the
* moment this method completes. Callers are encouraged to use {@link #getAndSaveLatestCommit}
* instead.
*
@@ -186,13 +186,13 @@ public final class IndexDeletionPolicyWrapper extends IndexDeletionPolicy {
/**
* Set the duration for which commit point is to be reserved by the deletion policy.
*
- * <p><b>NOTE:</b> This method does not make any garuntees that the specified index generation
- * exists, or that the specified generation has not already ben deleted. The only garuntee is that
- * <em>if</em> the specified generation exists now, or is created at some point in the future,
- * then it will be resered for <em>at least</em> the specified <code>reserveTime</code>.
+ * <p><b>NOTE:</b> This method does not make any guarantees that the specified index generation
+ * exists, or that the specified generation has not already ben deleted. The only guarantee is
+ * that <em>if</em> the specified generation exists now, or is created at some point in the
+ * future, then it will be reserved for <em>at least</em> the specified <code>reserveTime</code>.
*
* @param indexGen gen of the commit point to be reserved
- * @param reserveTime durration in milliseconds (relative to 'now') for which the commit point is
+ * @param reserveTime duration in milliseconds (relative to 'now') for which the commit point is
* to be reserved
*/
public void setReserveDuration(Long indexGen, long reserveTime) {
@@ -200,7 +200,7 @@ public final class IndexDeletionPolicyWrapper extends IndexDeletionPolicy {
// all operations on 'reserves' are done atomically.
//
// Here we'll use Map.merge to ensure that we atomically replace any existing timestamp if
- // and only if our new reservation timetsamp is larger.
+ // and only if our new reservation timestamp is larger.
final long reserveAsNanoTime =
System.nanoTime() + TimeUnit.NANOSECONDS.convert(reserveTime, TimeUnit.MILLISECONDS);
reserves.merge(indexGen, reserveAsNanoTime, BinaryOperator.maxBy(Comparator.naturalOrder()));
@@ -220,13 +220,13 @@ public final class IndexDeletionPolicyWrapper extends IndexDeletionPolicy {
/**
* Permanently prevent this commit point from being deleted (if it has not already) using a
- * refrence count.
+ * reference count.
*
* <p><b>NOTE:</b> Callers <em>MUST</em> call {@link #releaseCommitPoint} when finished using it
* in order to decrement the reference count, or the commit will be preserved in the Directory
* forever.
*
- * @param generation the generation of the IndexComit to save until released
+ * @param generation the generation of the IndexCommit to save until released
* @see #getAndSaveLatestCommit
* @see #getAndSaveCommitPoint
* @see #releaseCommitPoint
@@ -239,7 +239,7 @@ public final class IndexDeletionPolicyWrapper extends IndexDeletionPolicy {
/**
* Release a previously saved commit point.
*
- * <p>This is a convinience wrapper around {@link #releaseCommitPoint(Long)} that will ignore null
+ * <p>This is a convenience wrapper around {@link #releaseCommitPoint(Long)} that will ignore null
* input.
*/
public synchronized void releaseCommitPoint(IndexCommit commit) {
@@ -251,9 +251,9 @@ public final class IndexDeletionPolicyWrapper extends IndexDeletionPolicy {
/**
* Release a previously saved commit point.
*
- * <p>This method does not enforce that that the specified generation has previously been saved,
- * or even that it's 'non-null'. But if both are true then it will decrement the reference count
- * for the specified generation.
+ * <p>This method does not enforce that the specified generation has previously been saved, or
+ * even that it's 'non-null'. But if both are true then it will decrement the reference count for
+ * the specified generation.
*/
public synchronized void releaseCommitPoint(Long generation) {
if (null == generation) {
@@ -382,7 +382,7 @@ public final class IndexDeletionPolicyWrapper extends IndexDeletionPolicy {
/**
* Returns the commit with the specified generation <em>if it is known</em>.
*
- * <p><b>NOTE:</b> This method makes no garuntee that the commit returned still exists as the
+ * <p><b>NOTE:</b> This method makes no guarantee that the commit returned still exists as the
* moment this method completes. Callers are encouraged to use {@link #getAndSaveLatestCommit}
* instead.
*
@@ -402,7 +402,7 @@ public final class IndexDeletionPolicyWrapper extends IndexDeletionPolicy {
* <p><b>NOTE:</b> This map instance may change between commits and commit points may be deleted.
* This API is intended for "informational purposes" only, to provide an "at the moment" view of
* the current list of known commits. Callers that need to ensure commits exist for an extended
- * period must wrap this call and all subsequent usage of the results in a synchornization block.
+ * period must wrap this call and all subsequent usage of the results in a synchronization block.
*
* @return a Map of generation to commit points
*/
@@ -419,32 +419,32 @@ public final class IndexDeletionPolicyWrapper extends IndexDeletionPolicy {
* #latestCommit} from being deleted.
*
* <p>If we did not do this, and waited to update <code>latestCommit</code> in <code>
- * updateKnownCommitPoints()</code> then we would need to wrap synchronization completley around
+ * updateKnownCommitPoints()</code> then we would need to wrap synchronization completely around
* the (delegated) <code>onInit()</code> and <code>onCommit()</code> calls, to ensure there was no
* window of time when {@link #getAndSaveLatestCommit} might return the "old" latest commit, after
* our delegate Policy had already deleted it.
*
* <p>(Since Saving/Reserving (other) commits is handled indirectly ("by reference") via the
* generation callers can still safely (try) to reserve "old" commits using an explicit generation
- * since {@link IndexCommitWrapper#delete} is synchornized on <code>this</code>)
+ * since {@link IndexCommitWrapper#delete} is synchronized on <code>this</code>)
*
* @see #latestCommit
* @see #updateKnownCommitPoints
*/
private synchronized void updateLatestCommit(final List<IndexCommitWrapper> list) {
- // NOTE: There's a hypothetical, not neccessarily possible/plausible, situation that
+ // NOTE: There's a hypothetical, not necessarily possible/plausible, situation that
// could lead to this combination of updateLatestCommit + updateKnownCommitPoints not
- // being as thread safe as completley synchornizing in onInit/onCommit...
+ // being as thread safe as completely synchronizing in onInit/onCommit...
// - knownCommits==(1, 2, 3, 4), latestCommit==4
// - onCommit(1, 2, 3, 4, 5, 6, 7) - we immediately update latestCommit=7
// - before knownCommits is updated, some client calls getAndSaveCommitPoint(6)
// - call fails "too old to be saved" even though it's in flight
// (this assumes some future caller/use-case that doesn't currently exist)
//
- // The upside of this current approach, and not completley synchornizing onInit/onCommit
+ // The upside of this current approach, and not completely synchronizing onInit/onCommit
// is that we have no control over what delegate is used, or how long those calls might take.
//
- // If the hypotehtical situation above ever becomes problematic, then an alternative approach
+ // If the hypothetical situation above ever becomes problematic, then an alternative approach
// might be to *add* to the Set/Map of all known commits *before* delegating, then *remove*
// everything except the new (non-deleted) commits *after* delegating.
@@ -463,7 +463,7 @@ public final class IndexDeletionPolicyWrapper extends IndexDeletionPolicy {
* Updates the state of all "current" commits.
*
* <p>This method is safe to call <em>after</em> delegating to ou inner <code>IndexDeletionPolicy
- * </code> (w/o synchornizing the delegate calls) because even if the delegate decides to {@link
+ * </code> (w/o synchronizing the delegate calls) because even if the delegate decides to {@link
* IndexCommit#delete} a commit that a concurrent thread may wish to reserve/save, that {@link
* IndexCommitWrapper} will ensure that call is synchronized.
*
@@ -487,7 +487,7 @@ public final class IndexDeletionPolicyWrapper extends IndexDeletionPolicy {
}
/**
- * Helper method for unpacking the timestamp infor from the user data
+ * Helper method for unpacking the timestamp info from the user data
*
* @see SolrIndexWriter#COMMIT_TIME_MSEC_KEY
* @see IndexCommit#getUserData
diff --git a/solr/core/src/java/org/apache/solr/core/NodeConfig.java b/solr/core/src/java/org/apache/solr/core/NodeConfig.java
index 9006951e3dc..8117aad4523 100644
--- a/solr/core/src/java/org/apache/solr/core/NodeConfig.java
+++ b/solr/core/src/java/org/apache/solr/core/NodeConfig.java
@@ -413,8 +413,8 @@ public class NodeConfig {
* properties, or from the "extra" properties configured explicitly on the SolrDispatchFilter; or
* null if not specified.
*
- * <p>This is the value that would have been used when attempting locate the solr.xml in ZooKeeper
- * (regardless of wether the file was actaully loaded from ZK or from local disk)
+ * <p>This is the value that would have been used when attempting to locate the solr.xml in
+ * ZooKeeper (regardless of whether the file was actually loaded from ZK or from local disk)
*
* <p>(This value should only be used for "accounting" purposes to track where the node config
* came from if it <em>was</em> loaded from zk -- ie: to check if the chroot has already been
diff --git a/solr/core/src/java/org/apache/solr/core/PluginBag.java b/solr/core/src/java/org/apache/solr/core/PluginBag.java
index f583fc03bd2..c70bdd4af9d 100644
--- a/solr/core/src/java/org/apache/solr/core/PluginBag.java
+++ b/solr/core/src/java/org/apache/solr/core/PluginBag.java
@@ -318,7 +318,7 @@ public class PluginBag<T> implements AutoCloseable {
}
/**
- * Initializes the plugins after reading the meta data from {@link
+ * Initializes the plugins after reading the metadata from {@link
* org.apache.solr.core.SolrConfig}.
*
* @param defaults These will be registered if not explicitly specified
diff --git a/solr/core/src/java/org/apache/solr/core/PluginInfo.java b/solr/core/src/java/org/apache/solr/core/PluginInfo.java
index 0331b084045..5c24ee9a8f4 100644
--- a/solr/core/src/java/org/apache/solr/core/PluginInfo.java
+++ b/solr/core/src/java/org/apache/solr/core/PluginInfo.java
@@ -252,7 +252,7 @@ public class PluginInfo implements MapSerializable {
* Filter children by type
*
* @param type The type name. must not be null
- * @return The mathcing children
+ * @return The matching children
*/
public List<PluginInfo> getChildren(String type) {
if (children.isEmpty()) return children;
diff --git a/solr/core/src/java/org/apache/solr/core/SolrCore.java b/solr/core/src/java/org/apache/solr/core/SolrCore.java
index dd8032a4e88..661e9f1e29f 100644
--- a/solr/core/src/java/org/apache/solr/core/SolrCore.java
+++ b/solr/core/src/java/org/apache/solr/core/SolrCore.java
@@ -369,7 +369,7 @@ public class SolrCore implements SolrInfoBean, Closeable {
*/
public void setLatestSchema(IndexSchema replacementSchema) {
// 1) For a newly instantiated core, the Similarity needs SolrCore before inform() is called on
- // any registered SolrCoreAware listeners (which will likeley need to use the SolrIndexSearcher.
+ // any registered SolrCoreAware listeners (which will likely need to use the SolrIndexSearcher.
//
// 2) If a new IndexSchema is assigned to an existing live SolrCore (ie: managed schema
// replacement via SolrCloud) then we need to explicitly inform() the similarity because
@@ -404,12 +404,12 @@ public class SolrCore implements SolrInfoBean, Closeable {
/**
* Returns the indexdir as given in index.properties. If index.properties exists in dataDir and
- * there is a property <i>index</i> available and it points to a valid directory in dataDir that
+ * there is a property <i>index</i> available, and it points to a valid directory in dataDir that
* is returned. Else dataDir/index is returned. Only called for creating new indexSearchers and
* indexwriters. Use the getIndexDir() method to know the active index directory
*
* @return the indexdir as given in index.properties
- * @throws SolrException if for any reason the a reasonable index directory cannot be determined.
+ * @throws SolrException if for any reason an index directory cannot be determined.
*/
public String getNewIndexDir() {
Directory dir = null;
@@ -427,7 +427,7 @@ public class SolrCore implements SolrInfoBean, Closeable {
String errorMessage = "Error in getNewIndexDir, exception: ";
log.error(errorMessage, e);
// See SOLR-11687. It is inadvisable to assume we can do the right thing for any but a small
- // number of exceptions that ware caught and swallowed in getIndexProperty.
+ // number of exceptions that were caught and swallowed in getIndexProperty.
throw new SolrException(ErrorCode.SERVER_ERROR, errorMessage, e);
} finally {
if (dir != null) {
@@ -1412,9 +1412,9 @@ public class SolrCore implements SolrInfoBean, Closeable {
// we are evidently running in cloud mode.
//
// In cloud mode, version field is required for correct consistency
- // ideally this check would be more fine grained, and individual features
+ // ideally this check would be more fine-grained, and individual features
// would assert it when they initialize, but DistributedUpdateProcessor
- // is currently a big ball of wax that does more then just distributing
+ // is currently a big ball of wax that does more than just distributing
// updates (ie: partial document updates), so it needs to work in no cloud
// mode as well, and can't assert version field support on init.
@@ -1441,7 +1441,7 @@ public class SolrCore implements SolrInfoBean, Closeable {
* @param dataDir An optional hint to the data directory location. Will be normalized and used if
* not null.
* @param config A solr config to retrieve the default data directory location, if used.
- * @param coreDescriptor descriptor to load the actual data dir from, if not using the defualt.
+ * @param coreDescriptor descriptor to load the actual data dir from, if not using the default.
* @return a normalized data directory name
* @throws SolrException if the data directory cannot be loaded from the core descriptor
*/
@@ -2514,7 +2514,7 @@ public class SolrCore implements SolrInfoBean, Closeable {
// if it isn't necessary.
synchronized (searcherLock) {
- for (; ; ) { // this loop is so w can retry in the event that we exceed maxWarmingSearchers
+ for (; ; ) { // this loop is so we can retry in the event that we exceed maxWarmingSearchers
// see if we can return the current searcher
if (_searcher != null && !forceNew) {
if (returnSearcher) {
@@ -2531,7 +2531,7 @@ public class SolrCore implements SolrInfoBean, Closeable {
searcherLock.wait();
} catch (InterruptedException e) {
if (log.isInfoEnabled()) {
- log.info("Interupted waiting for searcherLock", e);
+ log.info("Interrupted waiting for searcherLock", e);
}
}
}
@@ -2562,7 +2562,7 @@ public class SolrCore implements SolrInfoBean, Closeable {
searcherLock.wait();
} catch (InterruptedException e) {
if (log.isInfoEnabled()) {
- log.info("Interupted waiting for searcherLock", e);
+ log.info("Interrupted waiting for searcherLock", e);
}
}
continue; // go back to the top of the loop and retry
@@ -2786,7 +2786,7 @@ public class SolrCore implements SolrInfoBean, Closeable {
}
// Take control of newSearcherHolder (which should have a reference count of at
- // least 1 already. If the caller wishes to use the newSearcherHolder directly
+ // least 1 already). If the caller wishes to use the newSearcherHolder directly
// after registering it, then they should increment the reference count *before*
// calling this method.
//
@@ -3396,20 +3396,20 @@ public class SolrCore implements SolrInfoBean, Closeable {
ManagedIndexSchema mis = (ManagedIndexSchema) core.getLatestSchema();
schemaRes = mis.getResourceName();
}
- final String managedSchmaResourcePath =
+ final String managedSchemaResourcePath =
schemaRes == null ? null : zkSolrResourceLoader.getConfigSetZkPath() + "/" + schemaRes;
return () -> {
log.info("config update listener called for core {}", coreName);
SolrZkClient zkClient = cc.getZkController().getZkClient();
- int solrConfigversion, overlayVersion, managedSchemaVersion = 0;
+ int solrConfigVersion, overlayVersion, managedSchemaVersion = 0;
SolrConfig cfg = null;
try (SolrCore solrCore = cc.solrCores.getCoreFromAnyList(coreName, true, coreId)) {
if (solrCore == null || solrCore.isClosed() || solrCore.getCoreContainer().isShutDown())
return;
cfg = solrCore.getSolrConfig();
- solrConfigversion = solrCore.getSolrConfig().getZnodeVersion();
+ solrConfigVersion = solrCore.getSolrConfig().getZnodeVersion();
overlayVersion = solrCore.getSolrConfig().getOverlay().getVersion();
- if (managedSchmaResourcePath != null) {
+ if (managedSchemaResourcePath != null) {
managedSchemaVersion =
((ManagedIndexSchema) solrCore.getLatestSchema()).getSchemaZkVersion();
}
@@ -3418,8 +3418,8 @@ public class SolrCore implements SolrInfoBean, Closeable {
cfg.refreshRequestParams();
}
if (checkStale(zkClient, overlayPath, overlayVersion)
- || checkStale(zkClient, solrConfigPath, solrConfigversion)
- || checkStale(zkClient, managedSchmaResourcePath, managedSchemaVersion)) {
+ || checkStale(zkClient, solrConfigPath, solrConfigVersion)
+ || checkStale(zkClient, managedSchemaResourcePath, managedSchemaVersion)) {
log.info("core reload {}", coreName);
SolrConfigHandler configHandler = ((SolrConfigHandler) core.getRequestHandler("/config"));
if (configHandler.getReloadLock().tryLock()) {
@@ -3580,19 +3580,19 @@ public class SolrCore implements SolrInfoBean, Closeable {
}
/**
- * Run an arbitrary task in it's own thread. This is an expert option and is a method you should
+ * Run an arbitrary task in its own thread. This is an expert option and is a method you should
* use with great care. It would be bad to run something that never stopped or run something that
- * took a very long time. Typically this is intended for actions that take a few seconds, and
- * therefore would be bad to wait for within a request, but but would not pose a significant
- * hindrance to server shut down times. It is not intended for long running tasks and if you are
- * using a Runnable with a loop in it, you are almost certainly doing it wrong.
+ * took a very long time. Typically, this is intended for actions that take a few seconds, and
+ * therefore would be bad to wait for within a request, but would not pose a significant hindrance
+ * to server shut down times. It is not intended for long-running tasks and if you are using a
+ * Runnable with a loop in it, you are almost certainly doing it wrong.
*
* <p>WARNING: Solr wil not be able to shut down gracefully until this task completes!
*
* <p>A significant upside of using this method vs creating your own ExecutorService is that your
* code does not have to properly shutdown executors which typically is risky from a unit testing
* perspective since the test framework will complain if you don't carefully ensure the executor
- * shuts down before the end of the test. Also the threads running this task are sure to have a
+ * shuts down before the end of the test. Also, the threads running this task are sure to have a
* proper MDC for logging.
*
* @param r the task to run
diff --git a/solr/core/src/java/org/apache/solr/core/SolrDeletionPolicy.java b/solr/core/src/java/org/apache/solr/core/SolrDeletionPolicy.java
index 50b46211aac..2b8232deef0 100644
--- a/solr/core/src/java/org/apache/solr/core/SolrDeletionPolicy.java
+++ b/solr/core/src/java/org/apache/solr/core/SolrDeletionPolicy.java
@@ -32,7 +32,7 @@ import org.slf4j.LoggerFactory;
/**
* Standard Solr deletion policy that allows reserving index commit points for certain amounts of
- * time to support features such as index replication or snapshooting directly out of a live index
+ * time to support features such as index replication or snapshotting directly out of a live index
* directory.
*
* @see org.apache.lucene.index.IndexDeletionPolicy
@@ -143,7 +143,7 @@ public class SolrDeletionPolicy extends IndexDeletionPolicy implements NamedList
private void updateCommits(List<? extends IndexCommit> commits) {
// to be safe, we should only call delete on a commit point passed to us
- // in this specific call (may be across diff IndexWriter instances).
+ // in this specific call (which may be across diff IndexWriter instances).
// this will happen rarely, so just synchronize everything
// for safety and to avoid race conditions
diff --git a/solr/core/src/java/org/apache/solr/core/SolrResourceLoader.java b/solr/core/src/java/org/apache/solr/core/SolrResourceLoader.java
index 7017ce47f93..e82a05d66b2 100644
--- a/solr/core/src/java/org/apache/solr/core/SolrResourceLoader.java
+++ b/solr/core/src/java/org/apache/solr/core/SolrResourceLoader.java
@@ -812,7 +812,7 @@ public class SolrResourceLoader
SolrCoreAware.class,
new Class<?>[] {
// DO NOT ADD THINGS TO THIS LIST -- ESPECIALLY THINGS THAT CAN BE CREATED DYNAMICALLY
- // VIA RUNTIME APIS -- UNTILL CAREFULLY CONSIDERING THE ISSUES MENTIONED IN SOLR-8311
+ // VIA RUNTIME APIS -- UNTIL CAREFULLY CONSIDERING THE ISSUES MENTIONED IN SOLR-8311
CodecFactory.class,
DirectoryFactory.class,
ManagedIndexSchemaFactory.class,
@@ -828,7 +828,7 @@ public class SolrResourceLoader
ResourceLoaderAware.class,
new Class<?>[] {
// DO NOT ADD THINGS TO THIS LIST -- ESPECIALLY THINGS THAT CAN BE CREATED DYNAMICALLY
- // VIA RUNTIME APIS -- UNTILL CAREFULLY CONSIDERING THE ISSUES MENTIONED IN SOLR-8311
+ // VIA RUNTIME APIS -- UNTIL CAREFULLY CONSIDERING THE ISSUES MENTIONED IN SOLR-8311
// evaluate if this must go into schemaResourceLoaderComponents
CharFilterFactory.class,
TokenFilterFactory.class,
@@ -918,7 +918,7 @@ public class SolrResourceLoader
}
/**
- * Create a n instance of a class using an appropriate {@link SolrResourceLoader} depending on the
+ * Create an instance of a class using an appropriate {@link SolrResourceLoader} depending on the
* package of that class
*
* @param registerCoreReloadListener register a listener for the package and reload the core if
diff --git a/solr/core/src/java/org/apache/solr/core/SolrXmlConfig.java b/solr/core/src/java/org/apache/solr/core/SolrXmlConfig.java
index f6d1eb78d93..676026c82ed 100644
--- a/solr/core/src/java/org/apache/solr/core/SolrXmlConfig.java
+++ b/solr/core/src/java/org/apache/solr/core/SolrXmlConfig.java
@@ -70,13 +70,13 @@ public class SolrXmlConfig {
private static final Pattern COMMA_SEPARATED_PATTERN = Pattern.compile("\\s*,\\s*");
/**
- * Given some node Properties, checks if non-null and a 'zkHost' is alread included. If so, the
+ * Given some node Properties, checks if non-null and a 'zkHost' is already included. If so, the
* Properties are returned as is. If not, then the returned value will be a new Properties,
* wrapping the original Properties, with the 'zkHost' value set based on the value of the
- * corispond System property (if set)
+ * corresponding System property (if set)
*
* <p>In theory we only need this logic once, ideally in SolrDispatchFilter, but we put it here to
- * re-use redundently because of how much surface area our API has for various tests to poke at
+ * re-use redundantly because of how much surface area our API has for various tests to poke at
* us.
*/
public static Properties wrapAndSetZkHostFromSysPropIfNeeded(final Properties props) {
@@ -133,7 +133,7 @@ public class SolrXmlConfig {
String nodeName = (String) entries.remove("nodeName");
if (StrUtils.isNullOrEmpty(nodeName) && cloudConfig != null) nodeName = cloudConfig.getHost();
- // It should goes inside the fillSolrSection method but
+ // It should go inside the fillSolrSection method but
// since it is arranged as a separate section it is placed here
Map<String, String> coreAdminHandlerActions =
readNodeListAsNamedList(root.get("coreAdminHandlerActions"), "<coreAdminHandlerActions>")
diff --git a/solr/core/src/java/org/apache/solr/core/TracerConfigurator.java b/solr/core/src/java/org/apache/solr/core/TracerConfigurator.java
index b730c5741e6..ba36ed127ec 100644
--- a/solr/core/src/java/org/apache/solr/core/TracerConfigurator.java
+++ b/solr/core/src/java/org/apache/solr/core/TracerConfigurator.java
@@ -52,7 +52,7 @@ public abstract class TracerConfigurator implements NamedListInitializedPlugin {
if (GlobalTracer.isRegistered()) {
// ideally we would furthermore check that it's not a no-op impl either but
// GlobalTracer.get() always returns a GlobalTracer implementing Tracer that delegates
- // to the real Tracer (that may or may not be a No-Op impl.
+ // to the real Tracer (that may or may not be a No-Op impl).
ExecutorUtil.addThreadLocalProvider(new TracerConfigurator.SpanThreadLocalProvider());
}
diff --git a/solr/core/src/java/org/apache/solr/core/TransientSolrCoreCache.java b/solr/core/src/java/org/apache/solr/core/TransientSolrCoreCache.java
index e0119635eda..51a625f2b7f 100644
--- a/solr/core/src/java/org/apache/solr/core/TransientSolrCoreCache.java
+++ b/solr/core/src/java/org/apache/solr/core/TransientSolrCoreCache.java
@@ -21,7 +21,7 @@ import java.util.Collection;
import java.util.Set;
/**
- * The base class for custom transient core maintenance. Any custom plugin that want's to take
+ * The base class for custom transient core maintenance. Any custom plugin that wants to take
* control of transient caches (i.e. any core defined with transient=true) should override this
* class.
*
diff --git a/solr/core/src/java/org/apache/solr/core/XmlConfigFile.java b/solr/core/src/java/org/apache/solr/core/XmlConfigFile.java
index 554d0c56a6f..bc6ef3650c7 100644
--- a/solr/core/src/java/org/apache/solr/core/XmlConfigFile.java
+++ b/solr/core/src/java/org/apache/solr/core/XmlConfigFile.java
@@ -399,7 +399,7 @@ public class XmlConfigFile { // formerly simply "Config"
return val != null ? Double.parseDouble(val) : def;
}
- /** If this config is loaded from zk the version is relevant other wise -1 is returned */
+ /** If this config is loaded from zk the version is relevant otherwise -1 is returned */
public int getZnodeVersion() {
return zkVersion;
}
diff --git a/solr/core/src/java/org/apache/solr/core/backup/repository/BackupRepository.java b/solr/core/src/java/org/apache/solr/core/backup/repository/BackupRepository.java
index b3d2034d4cf..0d968fde0b2 100644
--- a/solr/core/src/java/org/apache/solr/core/backup/repository/BackupRepository.java
+++ b/solr/core/src/java/org/apache/solr/core/backup/repository/BackupRepository.java
@@ -171,7 +171,7 @@ public interface BackupRepository extends NamedListInitializedPlugin, Closeable
* Copy a file from specified <code>sourceDir</code> to the destination repository (i.e. backup).
*
* @param sourceDir The source directory hosting the file to be copied.
- * @param fileName The name of the file to by copied
+ * @param fileName The name of the file to be copied
* @param dest The destination backup location.
* @throws IOException in case of errors
*/
@@ -183,7 +183,7 @@ public interface BackupRepository extends NamedListInitializedPlugin, Closeable
* Copy a file from specified <code>sourceRepo</code> to the destination directory (i.e. restore).
*
* @param sourceRepo The source URI hosting the file to be copied.
- * @param fileName The name of the file to by copied
+ * @param fileName The name of the file to be copied
* @param dest The destination where the file should be copied.
* @throws IOException in case of errors.
*/
@@ -262,7 +262,7 @@ public interface BackupRepository extends NamedListInitializedPlugin, Closeable
* backup).
*
* @param sourceDir The source directory hosting the file to be copied.
- * @param sourceFileName The name of the file to by copied
+ * @param sourceFileName The name of the file to be copied
* @param destDir The destination backup location.
* @throws IOException in case of errors
* @throws CorruptIndexException in case checksum of the file does not match with precomputed
diff --git a/solr/core/src/java/org/apache/solr/core/backup/repository/BackupRepositoryFactory.java b/solr/core/src/java/org/apache/solr/core/backup/repository/BackupRepositoryFactory.java
index 17fc9b4bd76..5d7be935d06 100644
--- a/solr/core/src/java/org/apache/solr/core/backup/repository/BackupRepositoryFactory.java
+++ b/solr/core/src/java/org/apache/solr/core/backup/repository/BackupRepositoryFactory.java
@@ -78,7 +78,7 @@ public class BackupRepositoryFactory {
BackupRepository result = loader.newInstance(repo.className, BackupRepository.class);
if ("trackingBackupRepository".equals(name)) {
// newInstance can be called by multiple threads, synchronization prevents simultaneous
- // multi-threaded 'adds' from corrupting the namedlist
+ // multithreaded 'adds' from corrupting the NamedList
synchronized (repo.initArgs) {
if (repo.initArgs.get("factory") == null) {
repo.initArgs.add("factory", this);