You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@phoenix.apache.org by st...@apache.org on 2021/01/29 06:22:53 UTC
[phoenix] branch 4.x updated: PHOENIX-6346 Javadoc generation fails
with Java 8
This is an automated email from the ASF dual-hosted git repository.
stoty pushed a commit to branch 4.x
in repository https://gitbox.apache.org/repos/asf/phoenix.git
The following commit(s) were added to refs/heads/4.x by this push:
new 409c559 PHOENIX-6346 Javadoc generation fails with Java 8
409c559 is described below
commit 409c559dcae660676abaf211de665a669f58aedd
Author: Istvan Toth <st...@apache.org>
AuthorDate: Thu Jan 28 17:43:07 2021 +0100
PHOENIX-6346 Javadoc generation fails with Java 8
---
.../java/org/apache/hadoop/hbase/PhoenixTagType.java | 2 +-
.../InterRegionServerIndexRpcControllerFactory.java | 3 ++-
...nterRegionServerMetadataRpcControllerFactory.java | 2 +-
.../hbase/regionserver/IndexHalfStoreFileReader.java | 4 ++--
.../hbase/regionserver/IndexKeyValueSkipListSet.java | 6 +++---
.../regionserver/LocalIndexStoreFileScanner.java | 2 +-
.../hbase/regionserver/ScannerContextUtil.java | 2 +-
.../hbase/regionserver/wal/IndexedHLogReader.java | 8 ++++----
.../cache/aggcache/SpillableGroupByCache.java | 8 ++++----
.../java/org/apache/phoenix/compile/BindManager.java | 2 +-
.../org/apache/phoenix/compile/ColumnResolver.java | 3 +--
.../org/apache/phoenix/compile/JoinCompiler.java | 2 +-
.../java/org/apache/phoenix/compile/KeyPart.java | 2 +-
.../org/apache/phoenix/compile/SubqueryRewriter.java | 2 ++
.../apache/phoenix/compile/SubselectRewriter.java | 4 +---
.../org/apache/phoenix/compile/WhereOptimizer.java | 2 +-
.../coprocessor/ChildLinkMetaDataEndpoint.java | 2 +-
.../phoenix/coprocessor/MetaDataEndpointImpl.java | 2 +-
.../apache/phoenix/coprocessor/MetaDataProtocol.java | 2 +-
.../apache/phoenix/expression/CaseExpression.java | 6 ++----
.../phoenix/expression/ComparisonExpression.java | 2 +-
.../apache/phoenix/expression/ModulusExpression.java | 4 ++--
.../apache/phoenix/expression/OrderByExpression.java | 8 +++-----
.../aggregator/PercentileDiscClientAggregator.java | 2 +-
.../expression/function/CeilDateExpression.java | 2 +-
.../expression/function/CountAggregateFunction.java | 2 +-
.../function/DistinctCountAggregateFunction.java | 2 +-
.../function/ExternalSqlTypeIdFunction.java | 2 +-
.../expression/function/FirstValueFunction.java | 2 +-
.../expression/function/FirstValuesFunction.java | 2 +-
.../phoenix/expression/function/LTrimFunction.java | 4 ++--
.../expression/function/LastValueFunction.java | 2 +-
.../expression/function/LastValuesFunction.java | 2 +-
.../phoenix/expression/function/LengthFunction.java | 4 ++--
.../expression/function/NthValueFunction.java | 2 +-
.../function/PercentRankAggregateFunction.java | 2 +-
.../function/PercentileContAggregateFunction.java | 2 +-
.../function/PercentileDiscAggregateFunction.java | 2 +-
.../phoenix/expression/function/RTrimFunction.java | 4 ++--
.../phoenix/expression/function/RandomFunction.java | 2 ++
.../expression/function/RegexpReplaceFunction.java | 2 +-
.../expression/function/RegexpSubstrFunction.java | 10 +++++-----
.../expression/function/RoundDateExpression.java | 2 +-
.../expression/function/StddevPopFunction.java | 2 +-
.../expression/function/StddevSampFunction.java | 2 +-
.../phoenix/expression/function/SubstrFunction.java | 10 +++++-----
.../phoenix/expression/function/TrimFunction.java | 2 +-
.../phoenix/expression/function/TruncFunction.java | 2 +-
.../java/org/apache/phoenix/hbase/index/Indexer.java | 12 ++++++------
.../hbase/index/builder/BaseIndexBuilder.java | 2 --
.../phoenix/hbase/index/builder/IndexBuilder.java | 10 ++++------
.../phoenix/hbase/index/covered/IndexCodec.java | 1 -
.../phoenix/hbase/index/covered/LocalTableState.java | 2 +-
.../hbase/index/covered/data/IndexMemStore.java | 4 ++--
.../index/covered/filter/NewerTimestampFilter.java | 2 +-
.../exception/SingleIndexWriteFailureException.java | 2 +-
.../hbase/index/scanner/FilteredKeyValueScanner.java | 2 +-
.../hbase/index/util/IndexManagementUtil.java | 2 +-
.../phoenix/hbase/index/wal/KeyValueCodec.java | 2 +-
.../hbase/index/write/IndexFailurePolicy.java | 4 ++--
.../phoenix/hbase/index/write/IndexWriter.java | 4 ++--
.../phoenix/hbase/index/write/IndexWriterUtils.java | 2 +-
.../write/TrackingParallelWriterIndexCommitter.java | 2 +-
.../write/recovery/PerRegionIndexWriteCache.java | 2 +-
.../org/apache/phoenix/index/IndexMaintainer.java | 2 +-
.../phoenix/index/IndexMetaDataCacheClient.java | 2 +-
.../apache/phoenix/index/PhoenixIndexBuilder.java | 2 +-
.../phoenix/index/PhoenixIndexFailurePolicy.java | 9 +++------
.../java/org/apache/phoenix/jdbc/PhoenixDriver.java | 2 +-
.../org/apache/phoenix/jdbc/PhoenixStatement.java | 2 +-
.../org/apache/phoenix/join/HashCacheClient.java | 11 ++++++++++-
.../phoenix/mapreduce/CsvToKeyValueMapper.java | 2 +-
.../mapreduce/ImportPreUpsertKeyValueProcessor.java | 4 ++--
.../phoenix/mapreduce/JsonToKeyValueMapper.java | 2 +-
.../apache/phoenix/mapreduce/PhoenixInputFormat.java | 1 -
.../apache/phoenix/mapreduce/PhoenixInputSplit.java | 2 +-
.../phoenix/mapreduce/RegexToKeyValueMapper.java | 2 +-
.../phoenix/monitoring/MutationMetricQueue.java | 2 +-
.../apache/phoenix/monitoring/ReadMetricQueue.java | 2 +-
.../parse/AndRewriterBooleanParseNodeVisitor.java | 2 +-
.../apache/phoenix/parse/ComparisonParseNode.java | 2 +-
.../phoenix/parse/GreaterThanOrEqualParseNode.java | 2 +-
.../apache/phoenix/parse/GreaterThanParseNode.java | 2 +-
.../main/java/org/apache/phoenix/parse/HintNode.java | 2 +-
.../phoenix/parse/LessThanOrEqualParseNode.java | 2 +-
.../org/apache/phoenix/parse/LessThanParseNode.java | 2 +-
.../org/apache/phoenix/parse/NotEqualParseNode.java | 2 +-
.../org/apache/phoenix/parse/ParseNodeFactory.java | 2 +-
.../org/apache/phoenix/parse/ParseNodeRewriter.java | 4 ++--
.../phoenix/query/ConnectionQueryServices.java | 2 +-
.../phoenix/query/ConnectionQueryServicesImpl.java | 1 -
.../query/ConnectionlessQueryServicesImpl.java | 4 ++--
.../apache/phoenix/query/GuidePostsCacheImpl.java | 2 +-
.../apache/phoenix/query/HBaseFactoryProvider.java | 2 +-
.../org/apache/phoenix/query/HConnectionFactory.java | 4 ++--
.../org/apache/phoenix/query/PropertyPolicy.java | 8 +++++---
.../apache/phoenix/query/PropertyPolicyProvider.java | 2 +-
.../java/org/apache/phoenix/query/QueryServices.java | 2 +-
.../org/apache/phoenix/schema/ColumnModifier.java | 2 +-
.../java/org/apache/phoenix/schema/SortOrder.java | 12 ++++++------
.../schema/stats/DefaultStatisticsCollector.java | 1 -
.../apache/phoenix/schema/stats/GuidePostsInfo.java | 2 +-
.../phoenix/schema/stats/GuidePostsInfoBuilder.java | 1 -
.../phoenix/schema/stats/StatisticsCollector.java | 1 -
.../phoenix/schema/stats/StatisticsWriter.java | 2 +-
.../apache/phoenix/schema/types/PArrayDataType.java | 4 ++--
.../org/apache/phoenix/schema/types/PDataType.java | 2 +-
.../transaction/PhoenixTransactionContext.java | 2 --
.../transaction/PhoenixTransactionProvider.java | 1 -
.../apache/phoenix/util/DefaultEnvironmentEdge.java | 1 -
.../java/org/apache/phoenix/util/ExpressionUtil.java | 2 +-
.../java/org/apache/phoenix/util/KeyValueUtil.java | 2 +-
.../java/org/apache/phoenix/util/MetaDataUtil.java | 7 +++----
.../apache/phoenix/util/PhoenixContextExecutor.java | 5 ++---
.../java/org/apache/phoenix/util/PhoenixRuntime.java | 20 ++++++++++++--------
.../java/org/apache/phoenix/util/ReadOnlyProps.java | 4 ++--
.../java/org/apache/phoenix/util/StringUtil.java | 2 +-
.../main/java/org/apache/phoenix/util/ViewUtil.java | 13 +++++++------
.../org/apache/phoenix/pherf/jmx/MonitorManager.java | 4 ++--
.../java/org/apache/phoenix/pherf/result/Result.java | 2 +-
.../apache/phoenix/pherf/result/ResultManager.java | 3 +--
.../phoenix/pherf/rules/RuleBasedDataGenerator.java | 2 +-
.../org/apache/phoenix/pherf/rules/RulesApplier.java | 2 +-
123 files changed, 198 insertions(+), 204 deletions(-)
diff --git a/phoenix-core/src/main/java/org/apache/hadoop/hbase/PhoenixTagType.java b/phoenix-core/src/main/java/org/apache/hadoop/hbase/PhoenixTagType.java
index da2064e..cb49b4c 100644
--- a/phoenix-core/src/main/java/org/apache/hadoop/hbase/PhoenixTagType.java
+++ b/phoenix-core/src/main/java/org/apache/hadoop/hbase/PhoenixTagType.java
@@ -19,7 +19,7 @@ package org.apache.hadoop.hbase;
/**
Used to persist the TagType in HBase Cell Tags.
- All the type present here should be more than @{@link Tag#CUSTOM_TAG_TYPE_RANGE} which is 64.
+ All the type present here should be more than {Tag#CUSTOM_TAG_TYPE_RANGE which is 64.
**/
public final class PhoenixTagType {
/**
diff --git a/phoenix-core/src/main/java/org/apache/hadoop/hbase/ipc/controller/InterRegionServerIndexRpcControllerFactory.java b/phoenix-core/src/main/java/org/apache/hadoop/hbase/ipc/controller/InterRegionServerIndexRpcControllerFactory.java
index 5c07213..eb4f382 100644
--- a/phoenix-core/src/main/java/org/apache/hadoop/hbase/ipc/controller/InterRegionServerIndexRpcControllerFactory.java
+++ b/phoenix-core/src/main/java/org/apache/hadoop/hbase/ipc/controller/InterRegionServerIndexRpcControllerFactory.java
@@ -28,7 +28,8 @@ import org.apache.phoenix.compat.hbase.CompatHBaseRpcController;
import org.apache.phoenix.compat.hbase.CompatRpcControllerFactory;
/**
- * {@link RpcControllerFactory} that should only be used when creating {@link HTable} for
+
+ * RpcControllerFactory that should only be used when creating HTable for
* making remote RPCs to the region servers hosting global mutable index table regions.
* This controller factory shouldn't be globally configured anywhere and is meant to be used
* only internally by Phoenix indexing code.
diff --git a/phoenix-core/src/main/java/org/apache/hadoop/hbase/ipc/controller/InterRegionServerMetadataRpcControllerFactory.java b/phoenix-core/src/main/java/org/apache/hadoop/hbase/ipc/controller/InterRegionServerMetadataRpcControllerFactory.java
index bf961b4..ae2171e 100644
--- a/phoenix-core/src/main/java/org/apache/hadoop/hbase/ipc/controller/InterRegionServerMetadataRpcControllerFactory.java
+++ b/phoenix-core/src/main/java/org/apache/hadoop/hbase/ipc/controller/InterRegionServerMetadataRpcControllerFactory.java
@@ -28,7 +28,7 @@ import org.apache.phoenix.compat.hbase.CompatHBaseRpcController;
import org.apache.phoenix.compat.hbase.CompatRpcControllerFactory;
/**
- * {@link RpcControllerFactory} that should only be used when creating {@link HTable} for
+ * {@link RpcControllerFactory} that should only be used when creating HTable for
* making remote RPCs to the region servers hosting Phoenix SYSTEM tables.
*/
public class InterRegionServerMetadataRpcControllerFactory extends CompatRpcControllerFactory {
diff --git a/phoenix-core/src/main/java/org/apache/hadoop/hbase/regionserver/IndexHalfStoreFileReader.java b/phoenix-core/src/main/java/org/apache/hadoop/hbase/regionserver/IndexHalfStoreFileReader.java
index 87b8277..4bb5f72 100644
--- a/phoenix-core/src/main/java/org/apache/hadoop/hbase/regionserver/IndexHalfStoreFileReader.java
+++ b/phoenix-core/src/main/java/org/apache/hadoop/hbase/regionserver/IndexHalfStoreFileReader.java
@@ -47,8 +47,8 @@ import org.apache.phoenix.index.IndexMaintainer;
* This type works in tandem with the {@link Reference} type. This class is used reading while
* Reference is used writing.
*
- * <p>
- * This file is not splitable. Calls to {@link #midkey()} return null.
+ *
+ * This file is not splitable. Calls to #midkey() return null.
*/
public class IndexHalfStoreFileReader extends StoreFile.Reader {
diff --git a/phoenix-core/src/main/java/org/apache/hadoop/hbase/regionserver/IndexKeyValueSkipListSet.java b/phoenix-core/src/main/java/org/apache/hadoop/hbase/regionserver/IndexKeyValueSkipListSet.java
index c322cb4..63d4113 100644
--- a/phoenix-core/src/main/java/org/apache/hadoop/hbase/regionserver/IndexKeyValueSkipListSet.java
+++ b/phoenix-core/src/main/java/org/apache/hadoop/hbase/regionserver/IndexKeyValueSkipListSet.java
@@ -24,7 +24,7 @@ import org.apache.hadoop.hbase.KeyValue;
/**
* Like a {@link KeyValueSkipListSet}, but also exposes useful, atomic methods (e.g.
- * {@link #putIfAbsent(KeyValue)}).
+ * #putIfAbsent(KeyValue)).
*/
public class IndexKeyValueSkipListSet extends KeyValueSkipListSet {
@@ -54,7 +54,7 @@ public class IndexKeyValueSkipListSet extends KeyValueSkipListSet {
}
/**
- * Add the passed {@link KeyValue} to the set, only if one is not already set. This is equivalent
+ * Add the passed KeyValue to the set, only if one is not already set. This is equivalent
* to
* <pre>
* if (!set.containsKey(key))
@@ -63,7 +63,7 @@ public class IndexKeyValueSkipListSet extends KeyValueSkipListSet {
* return map.set(key);
* </pre>
* except that the action is performed atomically.
- * @param kv {@link KeyValue} to add
+ * @param kv KeyValue to add
* @return the previous value associated with the specified key, or <tt>null</tt> if there was no
* previously stored key
* @throws ClassCastException if the specified key cannot be compared with the keys currently in
diff --git a/phoenix-core/src/main/java/org/apache/hadoop/hbase/regionserver/LocalIndexStoreFileScanner.java b/phoenix-core/src/main/java/org/apache/hadoop/hbase/regionserver/LocalIndexStoreFileScanner.java
index fece89e..faacc3a 100644
--- a/phoenix-core/src/main/java/org/apache/hadoop/hbase/regionserver/LocalIndexStoreFileScanner.java
+++ b/phoenix-core/src/main/java/org/apache/hadoop/hbase/regionserver/LocalIndexStoreFileScanner.java
@@ -212,7 +212,7 @@ public class LocalIndexStoreFileScanner extends CompatStoreFileScanner {
/**
*
- * @param kv
+ * @param cell
* @param isSeek pass true for seek, false for reseek.
* @return
* @throws IOException
diff --git a/phoenix-core/src/main/java/org/apache/hadoop/hbase/regionserver/ScannerContextUtil.java b/phoenix-core/src/main/java/org/apache/hadoop/hbase/regionserver/ScannerContextUtil.java
index 126e0b1..f5c37b1 100644
--- a/phoenix-core/src/main/java/org/apache/hadoop/hbase/regionserver/ScannerContextUtil.java
+++ b/phoenix-core/src/main/java/org/apache/hadoop/hbase/regionserver/ScannerContextUtil.java
@@ -25,7 +25,7 @@ import org.apache.hadoop.hbase.CellUtil;
import java.util.List;
/**
- * @ScannerContext has all methods package visible. To properly update the context progress for our scanners we
+ * ScannerContext has all methods package visible. To properly update the context progress for our scanners we
* need this helper
*/
public class ScannerContextUtil {
diff --git a/phoenix-core/src/main/java/org/apache/hadoop/hbase/regionserver/wal/IndexedHLogReader.java b/phoenix-core/src/main/java/org/apache/hadoop/hbase/regionserver/wal/IndexedHLogReader.java
index cc2fec6..f6abb1e 100644
--- a/phoenix-core/src/main/java/org/apache/hadoop/hbase/regionserver/wal/IndexedHLogReader.java
+++ b/phoenix-core/src/main/java/org/apache/hadoop/hbase/regionserver/wal/IndexedHLogReader.java
@@ -23,10 +23,10 @@ import java.io.IOException;
/**
- * A WALReader that can also deserialize custom {@link WALEdit}s that contain index information.
- * <p>
- * This is basically a wrapper around a {@link SequenceFileLogReader} that has a custom
- * {@link SequenceFileLogReader.WALReader#next(Object)} method that only replaces the creation of the WALEdit with our own custom
+ * A WALReader that can also deserialize custom WALEdit s that contain index information.
+ *
+ * This is basically a wrapper around a SequenceFileLogReader that has a custom
+ * SequenceFileLogReader.WALReader#next(Object) method that only replaces the creation of the WALEdit with our own custom
* type
* <p>
* This is a little bit of a painful way of going about this, but saves the effort of hacking the
diff --git a/phoenix-core/src/main/java/org/apache/phoenix/cache/aggcache/SpillableGroupByCache.java b/phoenix-core/src/main/java/org/apache/phoenix/cache/aggcache/SpillableGroupByCache.java
index 821dc6b..f659194 100644
--- a/phoenix-core/src/main/java/org/apache/phoenix/cache/aggcache/SpillableGroupByCache.java
+++ b/phoenix-core/src/main/java/org/apache/phoenix/cache/aggcache/SpillableGroupByCache.java
@@ -58,7 +58,7 @@ import org.slf4j.LoggerFactory;
/**
* The main entry point is in GroupedAggregateRegionObserver. It instantiates a SpillableGroupByCache and invokes a
- * get() method on it. There is no: "if key not exists -> put into map" case, since the cache is a Loading cache and
+ * get() method on it. There is no: {@code "if key not exists -> put into map" } case, since the cache is a Loading cache and
* therefore handles the put under the covers. I tried to implement the final cache element accesses (RegionScanner
* below) streaming, i.e. there is just an iterator on it and removed the existing result materialization.
* SpillableGroupByCache implements a LRU cache using a LinkedHashMap with access order. There is a configurable an
@@ -120,10 +120,10 @@ public class SpillableGroupByCache implements GroupByCache {
/**
* Instantiates a Loading LRU Cache that stores key / aggregator[] tuples used for group by queries
*
- * @param estSize
- * @param estValueSize
+ * @param env
+ * @param tenantId
* @param aggs
- * @param ctxt
+ * @param estSizeNum
*/
public SpillableGroupByCache(final RegionCoprocessorEnvironment env, ImmutableBytesPtr tenantId,
ServerAggregators aggs, final int estSizeNum) {
diff --git a/phoenix-core/src/main/java/org/apache/phoenix/compile/BindManager.java b/phoenix-core/src/main/java/org/apache/phoenix/compile/BindManager.java
index fc5fbaf..f7a2a6d 100644
--- a/phoenix-core/src/main/java/org/apache/phoenix/compile/BindManager.java
+++ b/phoenix-core/src/main/java/org/apache/phoenix/compile/BindManager.java
@@ -37,7 +37,7 @@ import org.apache.phoenix.schema.PDatum;
* 2) the retrieve param metadata case where we don't have the bind values.
*
* In both cases, during query compilation we figure out what type the bind variable
- * "should" be, based on how it's used in the query. For example foo < ? would expect
+ * "should" be, based on how it's used in the query. For example {@code foo < ? } would expect
* that the bind variable type matches or can be coerced to the type of foo. For (1),
* we check that the bind value has the correct type and for (2) we set the param
* metadata type.
diff --git a/phoenix-core/src/main/java/org/apache/phoenix/compile/ColumnResolver.java b/phoenix-core/src/main/java/org/apache/phoenix/compile/ColumnResolver.java
index 6c91123..a3d8567 100644
--- a/phoenix-core/src/main/java/org/apache/phoenix/compile/ColumnResolver.java
+++ b/phoenix-core/src/main/java/org/apache/phoenix/compile/ColumnResolver.java
@@ -56,8 +56,7 @@ public interface ColumnResolver {
* @param schemaName the schema name
* @param tableName the table name or table alias
* @return the resolved TableRef
- * @throws TableNotFoundException if the table could not be resolved
- * @throws AmbiguousTableException if the table name is ambiguous
+ * @throws SQLException
*/
public TableRef resolveTable(String schemaName, String tableName) throws SQLException;
diff --git a/phoenix-core/src/main/java/org/apache/phoenix/compile/JoinCompiler.java b/phoenix-core/src/main/java/org/apache/phoenix/compile/JoinCompiler.java
index e22fdc4..adfb5dd 100644
--- a/phoenix-core/src/main/java/org/apache/phoenix/compile/JoinCompiler.java
+++ b/phoenix-core/src/main/java/org/apache/phoenix/compile/JoinCompiler.java
@@ -397,7 +397,7 @@ public class JoinCompiler {
/**
* Pruning columns for each {@link JoinCompiler.Table} if
- * {@link @link JoinCompiler.Table#isSubselect()}.
+ * {@link JoinCompiler.Table#isSubselect()}.
* @throws SQLException
*/
public void pruneSubselectAliasedNodes() throws SQLException {
diff --git a/phoenix-core/src/main/java/org/apache/phoenix/compile/KeyPart.java b/phoenix-core/src/main/java/org/apache/phoenix/compile/KeyPart.java
index 6cf9938..c2d8e7e 100644
--- a/phoenix-core/src/main/java/org/apache/phoenix/compile/KeyPart.java
+++ b/phoenix-core/src/main/java/org/apache/phoenix/compile/KeyPart.java
@@ -41,7 +41,7 @@ public interface KeyPart {
* SUBSTR(foo,1,3) = 'bar', the key range would be ['bar','bas'),
* and if foo was fixed length, the upper and lower key range
* bytes would be filled out to the fixed length.
- * @param op comparison operator (=, <=, <, >=, >, !=)
+ * @param op comparison operator {@code (=, <=, <, >=, >, !=) }
* @param rhs the constant on the RHS of an expression.
* @return the key range that encompasses the range for the
* expression for which this keyPart is associated
diff --git a/phoenix-core/src/main/java/org/apache/phoenix/compile/SubqueryRewriter.java b/phoenix-core/src/main/java/org/apache/phoenix/compile/SubqueryRewriter.java
index e2f6f58..9710108 100644
--- a/phoenix-core/src/main/java/org/apache/phoenix/compile/SubqueryRewriter.java
+++ b/phoenix-core/src/main/java/org/apache/phoenix/compile/SubqueryRewriter.java
@@ -135,6 +135,7 @@ public class SubqueryRewriter extends ParseNodeRewriter {
/**
* <pre>
+ * {@code
* Rewrite the In Subquery to semi/anti/left join for both NonCorrelated and Correlated subquery.
*
* 1.If the {@link InParseNode} is the only node in where clause or is the ANDed part of the where clause,
@@ -177,6 +178,7 @@ public class SubqueryRewriter extends ParseNodeRewriter {
* SELECT ITEM_ID,NAME FROM item I Left JOIN
* (SELECT DISTINCT 1 $28, MAX(ITEM_ID) $29,O.PRICE $27 FROM order O GROUP BY O.PRICE,O.CUSTOMER_ID) $26
* ON ((I.ITEM_ID = $26.$29 AND $26.$27 = I.PRICE)) WHERE ($26.$28 IS NOT NULL OR I.DISCOUNT1 > 10)
+ * }
* </pre>
*/
@Override
diff --git a/phoenix-core/src/main/java/org/apache/phoenix/compile/SubselectRewriter.java b/phoenix-core/src/main/java/org/apache/phoenix/compile/SubselectRewriter.java
index a5d501e..28f6a81 100644
--- a/phoenix-core/src/main/java/org/apache/phoenix/compile/SubselectRewriter.java
+++ b/phoenix-core/src/main/java/org/apache/phoenix/compile/SubselectRewriter.java
@@ -68,7 +68,6 @@ public class SubselectRewriter extends ParseNodeRewriter {
* @param preFilterParseNodes
* @param subselectAlias
* @return
- * @throws SQLException
*/
public static SelectStatement applyPreFiltersForSubselect(
SelectStatement subselectStatement,
@@ -111,7 +110,6 @@ public class SubselectRewriter extends ParseNodeRewriter {
* added to the statement by {@link #applyPreFiltersForSubselect}.
* @param statement
* @return
- * @throws SQLException
*/
public static boolean isFilterCanPushDownToSelect(SelectStatement statement) {
return statement.getLimit() == null &&
@@ -282,7 +280,7 @@ public class SubselectRewriter extends ParseNodeRewriter {
/**
* Pruning selectAliasedNodes according to referencedColumnNames,
* Note: the selectStatement is supposed to be a {@link DerivedTableNode} of an Outer SelectStatement,
- * so according to {@link FromCompiler.MultiTableColumnResolver#visit(DerivedTableNode)},
+ * so according to FromCompiler.MultiTableColumnResolver#visit(DerivedTableNode) ,
* wildcard in selectAliasedNode is not supported.
* @param selectStatement
* @param referencedColumnNames
diff --git a/phoenix-core/src/main/java/org/apache/phoenix/compile/WhereOptimizer.java b/phoenix-core/src/main/java/org/apache/phoenix/compile/WhereOptimizer.java
index b1fd238..ffbab01 100644
--- a/phoenix-core/src/main/java/org/apache/phoenix/compile/WhereOptimizer.java
+++ b/phoenix-core/src/main/java/org/apache/phoenix/compile/WhereOptimizer.java
@@ -1809,7 +1809,7 @@ public class WhereOptimizer {
/**
*
* Implementation of KeySlots for AND and OR expressions. The
- * List<KeySlot> will be in PK order.
+ * {@code List<KeySlot> } will be in PK order.
*
*/
public static class MultiKeySlot implements KeySlots {
diff --git a/phoenix-core/src/main/java/org/apache/phoenix/coprocessor/ChildLinkMetaDataEndpoint.java b/phoenix-core/src/main/java/org/apache/phoenix/coprocessor/ChildLinkMetaDataEndpoint.java
index 873a1e3..1df399e 100644
--- a/phoenix-core/src/main/java/org/apache/phoenix/coprocessor/ChildLinkMetaDataEndpoint.java
+++ b/phoenix-core/src/main/java/org/apache/phoenix/coprocessor/ChildLinkMetaDataEndpoint.java
@@ -49,7 +49,7 @@ import static org.apache.phoenix.coprocessor.MetaDataEndpointImpl.mutateRowsWith
/**
* Endpoint co-processor through which Phoenix metadata mutations for SYSTEM.CHILD_LINK flow.
- * The parent->child links ({@link org.apache.phoenix.schema.PTable.LinkType#CHILD_TABLE})
+ * The {@code parent->child } links ({@link org.apache.phoenix.schema.PTable.LinkType#CHILD_TABLE})
* are stored in the SYSTEM.CHILD_LINK table.
*/
public class ChildLinkMetaDataEndpoint extends ChildLinkMetaDataService implements
diff --git a/phoenix-core/src/main/java/org/apache/phoenix/coprocessor/MetaDataEndpointImpl.java b/phoenix-core/src/main/java/org/apache/phoenix/coprocessor/MetaDataEndpointImpl.java
index cde1eaf..900eec7 100644
--- a/phoenix-core/src/main/java/org/apache/phoenix/coprocessor/MetaDataEndpointImpl.java
+++ b/phoenix-core/src/main/java/org/apache/phoenix/coprocessor/MetaDataEndpointImpl.java
@@ -253,7 +253,7 @@ import com.google.protobuf.Service;
* is stored in a single header row. Column information is stored in a separate
* row per column. Linking information (indexes, views etc) are stored using a
* separate row for each link that uses the {@link LinkType} column value. The
- * parent->child links are stored in a separate SYSTEM.CHILD_LINK table.
+ * {@code parent->child } links are stored in a separate SYSTEM.CHILD_LINK table.
* Metadata for all tables/views/indexes in the same schema are stored in a
* single region which is enforced using the {@link MetaDataSplitPolicy}.
* <p>
diff --git a/phoenix-core/src/main/java/org/apache/phoenix/coprocessor/MetaDataProtocol.java b/phoenix-core/src/main/java/org/apache/phoenix/coprocessor/MetaDataProtocol.java
index 7a0c5fb..29a6e90 100644
--- a/phoenix-core/src/main/java/org/apache/phoenix/coprocessor/MetaDataProtocol.java
+++ b/phoenix-core/src/main/java/org/apache/phoenix/coprocessor/MetaDataProtocol.java
@@ -52,7 +52,7 @@ import com.google.protobuf.ByteString;
* an HBase table named SYSTEM.TABLE. Each table is represented by:
* - one row for the table
* - one row per column in the tabe
- * Upto {@link #DEFAULT_MAX_META_DATA_VERSIONS} versions are kept. The time
+ * Upto #DEFAULT_MAX_META_DATA_VERSIONS versions are kept. The time
* stamp of the metadata must always be increasing. The timestamp of the key
* values in the data row corresponds to the schema that it's using.
*
diff --git a/phoenix-core/src/main/java/org/apache/phoenix/expression/CaseExpression.java b/phoenix-core/src/main/java/org/apache/phoenix/expression/CaseExpression.java
index c426d83..966dc63 100644
--- a/phoenix-core/src/main/java/org/apache/phoenix/expression/CaseExpression.java
+++ b/phoenix-core/src/main/java/org/apache/phoenix/expression/CaseExpression.java
@@ -107,10 +107,8 @@ public class CaseExpression extends BaseCompoundExpression {
}
/**
* Construct CASE/WHEN expression
- * @param expressions list of expressions in the form of:
- * ((<result expression>, <boolean expression>)+, [<optional else result expression>])
- * @throws SQLException if return type of case expressions do not match and cannot
- * be coerced to a common type
+ * @param children list of expressions in the form of:
+ * {@code ((<result expression>, <boolean expression>)+, [<optional else result expression>]) }
*/
public CaseExpression(List<Expression> children) {
super(children);
diff --git a/phoenix-core/src/main/java/org/apache/phoenix/expression/ComparisonExpression.java b/phoenix-core/src/main/java/org/apache/phoenix/expression/ComparisonExpression.java
index 1810c3b..18dd4b1 100644
--- a/phoenix-core/src/main/java/org/apache/phoenix/expression/ComparisonExpression.java
+++ b/phoenix-core/src/main/java/org/apache/phoenix/expression/ComparisonExpression.java
@@ -56,7 +56,7 @@ import com.google.common.collect.Lists;
/**
*
- * Implementation for <,<=,>,>=,=,!= comparison expressions
+ * Implementation for {@code <,<=,>,>=,=,!= } comparison expressions
*
* @since 0.1
*/
diff --git a/phoenix-core/src/main/java/org/apache/phoenix/expression/ModulusExpression.java b/phoenix-core/src/main/java/org/apache/phoenix/expression/ModulusExpression.java
index b6b669f..99006c0 100644
--- a/phoenix-core/src/main/java/org/apache/phoenix/expression/ModulusExpression.java
+++ b/phoenix-core/src/main/java/org/apache/phoenix/expression/ModulusExpression.java
@@ -27,11 +27,11 @@ import org.apache.phoenix.schema.types.PLong;
/**
- *
+ * {@code
* Implementation of the LENGTH(<string>) build-in function. <string> is the string
* of characters we want to find the length of. If <string> is NULL or empty, null
* is returned.
- *
+ * }
*
* @since 0.1
*/
diff --git a/phoenix-core/src/main/java/org/apache/phoenix/expression/OrderByExpression.java b/phoenix-core/src/main/java/org/apache/phoenix/expression/OrderByExpression.java
index 2978bee..59d0f14 100644
--- a/phoenix-core/src/main/java/org/apache/phoenix/expression/OrderByExpression.java
+++ b/phoenix-core/src/main/java/org/apache/phoenix/expression/OrderByExpression.java
@@ -51,9 +51,7 @@ public class OrderByExpression implements Writable {
/**
* If {@link Expression#getSortOrder()} is {@link SortOrder#DESC},the isAscending of returned new is reversed,but isNullsLast is untouched.
- * @param expression
- * @param isNullsLast
- * @param isAscending
+ * @param orderByExpression
* @return
*/
public static OrderByExpression convertIfExpressionSortOrderDesc(OrderByExpression orderByExpression) {
@@ -65,7 +63,7 @@ public class OrderByExpression implements Writable {
/**
* If {@link Expression#getSortOrder()} is {@link SortOrder#DESC},reverse the isAscending,but isNullsLast is untouched.
- * A typical case is in {@link OrderByCompiler#compile} to get the compiled {@link OrderByExpression} to used for {@link OrderedResultIterator}.
+ * A typical case is in OrderByCompiler#compile to get the compiled OrderByExpression to used for OrderedResultIterator.
* @param expression
* @param isNullsLast
* @param isAscending
@@ -80,7 +78,7 @@ public class OrderByExpression implements Writable {
/**
* If orderByReverse is true, reverse the isNullsLast and isAscending.
- * A typical case is in {@link AggregatePlan.OrderingResultIteratorFactory#newIterator}
+ * A typical case is in AggregatePlan.OrderingResultIteratorFactory#newIterator
* @param expression
* @param isNullsLast
* @param isAscending
diff --git a/phoenix-core/src/main/java/org/apache/phoenix/expression/aggregator/PercentileDiscClientAggregator.java b/phoenix-core/src/main/java/org/apache/phoenix/expression/aggregator/PercentileDiscClientAggregator.java
index 54541e7..2589795 100644
--- a/phoenix-core/src/main/java/org/apache/phoenix/expression/aggregator/PercentileDiscClientAggregator.java
+++ b/phoenix-core/src/main/java/org/apache/phoenix/expression/aggregator/PercentileDiscClientAggregator.java
@@ -29,7 +29,7 @@ import org.apache.phoenix.schema.tuple.Tuple;
/**
*
- * Built-in function for PERCENTILE_DISC(<expression>) WITHIN GROUP (ORDER BY <expression> ASC/DESC) aggregate function
+ * Built-in function for {@code PERCENTILE_DISC(<expression>) WITHIN GROUP (ORDER BY <expression> ASC/DESC) } aggregate function
*
*
* @since 1.2.1
diff --git a/phoenix-core/src/main/java/org/apache/phoenix/expression/function/CeilDateExpression.java b/phoenix-core/src/main/java/org/apache/phoenix/expression/function/CeilDateExpression.java
index 1b41cb3..73931d8 100644
--- a/phoenix-core/src/main/java/org/apache/phoenix/expression/function/CeilDateExpression.java
+++ b/phoenix-core/src/main/java/org/apache/phoenix/expression/function/CeilDateExpression.java
@@ -37,7 +37,7 @@ import com.google.common.collect.Lists;
/**
*
- * Class encapsulating ceil operation on {@link org.apache.phoenix.schema.types.PDataType#DATE}.
+ * Class encapsulating ceil operation on org.apache.phoenix.schema.types.PDataType#DATE.
*
*
* @since 3.0.0
diff --git a/phoenix-core/src/main/java/org/apache/phoenix/expression/function/CountAggregateFunction.java b/phoenix-core/src/main/java/org/apache/phoenix/expression/function/CountAggregateFunction.java
index 6eef42d..9f6fe27 100644
--- a/phoenix-core/src/main/java/org/apache/phoenix/expression/function/CountAggregateFunction.java
+++ b/phoenix-core/src/main/java/org/apache/phoenix/expression/function/CountAggregateFunction.java
@@ -37,7 +37,7 @@ import org.apache.phoenix.util.SchemaUtil;
/**
*
- * Built-in function for COUNT(<expression>) aggregate function,
+ * Built-in function for {@code COUNT(<expression>) } aggregate function,
* for example COUNT(foo), COUNT(1), COUNT(*)
*
*
diff --git a/phoenix-core/src/main/java/org/apache/phoenix/expression/function/DistinctCountAggregateFunction.java b/phoenix-core/src/main/java/org/apache/phoenix/expression/function/DistinctCountAggregateFunction.java
index 958f8fd..e27973a 100644
--- a/phoenix-core/src/main/java/org/apache/phoenix/expression/function/DistinctCountAggregateFunction.java
+++ b/phoenix-core/src/main/java/org/apache/phoenix/expression/function/DistinctCountAggregateFunction.java
@@ -36,7 +36,7 @@ import org.apache.phoenix.util.SchemaUtil;
/**
*
- * Built-in function for COUNT(distinct <expression>) aggregate function,
+ * Built-in function for {@code COUNT(distinct <expression>) } aggregate function,
*
*
* @since 1.2.1
diff --git a/phoenix-core/src/main/java/org/apache/phoenix/expression/function/ExternalSqlTypeIdFunction.java b/phoenix-core/src/main/java/org/apache/phoenix/expression/function/ExternalSqlTypeIdFunction.java
index f510181..d8b300b 100644
--- a/phoenix-core/src/main/java/org/apache/phoenix/expression/function/ExternalSqlTypeIdFunction.java
+++ b/phoenix-core/src/main/java/org/apache/phoenix/expression/function/ExternalSqlTypeIdFunction.java
@@ -36,7 +36,7 @@ import java.util.List;
* Function used to get the external SQL type id from the internal SQL type integer.
* Typically the external and internal ids are the same, but for some types (e.g. arrays)
* there is are multiple specific internal types to represent multiple external types.
- * <p/>
+ *
* Usage:
* ExternalSqlTypeId(12)
* will return 12 based on {@link java.sql.Types#VARCHAR} being 12
diff --git a/phoenix-core/src/main/java/org/apache/phoenix/expression/function/FirstValueFunction.java b/phoenix-core/src/main/java/org/apache/phoenix/expression/function/FirstValueFunction.java
index 420deba..09871d9 100644
--- a/phoenix-core/src/main/java/org/apache/phoenix/expression/function/FirstValueFunction.java
+++ b/phoenix-core/src/main/java/org/apache/phoenix/expression/function/FirstValueFunction.java
@@ -30,7 +30,7 @@ import org.apache.phoenix.parse.FunctionParseNode;
import org.apache.phoenix.schema.types.PBoolean;
/**
- * Built-in function for FIRST_VALUE(<expression>) WITHIN GROUP (ORDER BY <expression> ASC/DESC) aggregate
+ * Built-in function for {@code FIRST_VALUE(<expression>) WITHIN GROUP (ORDER BY <expression> ASC/DESC) aggregate }
* function
*
*/
diff --git a/phoenix-core/src/main/java/org/apache/phoenix/expression/function/FirstValuesFunction.java b/phoenix-core/src/main/java/org/apache/phoenix/expression/function/FirstValuesFunction.java
index fec2eaf..b9114a2 100644
--- a/phoenix-core/src/main/java/org/apache/phoenix/expression/function/FirstValuesFunction.java
+++ b/phoenix-core/src/main/java/org/apache/phoenix/expression/function/FirstValuesFunction.java
@@ -33,7 +33,7 @@ import org.apache.phoenix.schema.types.PDataType;
import org.apache.phoenix.schema.types.PInteger;
/**
- * Built-in function for FIRST_VALUES(<expression>, <expression>) WITHIN GROUP (ORDER BY <expression> ASC/DESC) aggregate
+ * Built-in function for {@code FIRST_VALUES(<expression>, <expression>) WITHIN GROUP (ORDER BY <expression> ASC/DESC) aggregate }
* function
*
*/
diff --git a/phoenix-core/src/main/java/org/apache/phoenix/expression/function/LTrimFunction.java b/phoenix-core/src/main/java/org/apache/phoenix/expression/function/LTrimFunction.java
index bfbe2f1..fb2706c 100644
--- a/phoenix-core/src/main/java/org/apache/phoenix/expression/function/LTrimFunction.java
+++ b/phoenix-core/src/main/java/org/apache/phoenix/expression/function/LTrimFunction.java
@@ -34,8 +34,8 @@ import org.apache.phoenix.util.StringUtil;
/**
*
- * Implementation of the LTrim(<string>) build-in function. It removes from the left end of
- * <string> space character and other function bytes in single byte utf8 characters
+ * Implementation of the {@code LTrim(<string>) } build-in function. It removes from the left end of
+ * {@code <string> } space character and other function bytes in single byte utf8 characters
* set.
*
*
diff --git a/phoenix-core/src/main/java/org/apache/phoenix/expression/function/LastValueFunction.java b/phoenix-core/src/main/java/org/apache/phoenix/expression/function/LastValueFunction.java
index f5b3193..dd2a4b6 100644
--- a/phoenix-core/src/main/java/org/apache/phoenix/expression/function/LastValueFunction.java
+++ b/phoenix-core/src/main/java/org/apache/phoenix/expression/function/LastValueFunction.java
@@ -30,7 +30,7 @@ import org.apache.phoenix.parse.LastValueAggregateParseNode;
import org.apache.phoenix.schema.types.PBoolean;
/**
- * Built-in function for LAST_VALUE(<expression>) WITHIN GROUP (ORDER BY <expression> ASC/DESC) aggregate
+ * Built-in function for {@code LAST_VALUE(<expression>) WITHIN GROUP (ORDER BY <expression> ASC/DESC) aggregate }
* function
*
*/
diff --git a/phoenix-core/src/main/java/org/apache/phoenix/expression/function/LastValuesFunction.java b/phoenix-core/src/main/java/org/apache/phoenix/expression/function/LastValuesFunction.java
index d4a8061..21cb082 100644
--- a/phoenix-core/src/main/java/org/apache/phoenix/expression/function/LastValuesFunction.java
+++ b/phoenix-core/src/main/java/org/apache/phoenix/expression/function/LastValuesFunction.java
@@ -33,7 +33,7 @@ import org.apache.phoenix.schema.types.PDataType;
import org.apache.phoenix.schema.types.PInteger;
/**
- * Built-in function for FIRST_VALUES(<expression>, <expression>) WITHIN GROUP (ORDER BY <expression> ASC/DESC) aggregate
+ * Built-in function for {@code FIRST_VALUES(<expression>, <expression>) WITHIN GROUP (ORDER BY <expression> ASC/DESC) aggregate }
* function
*
*/
diff --git a/phoenix-core/src/main/java/org/apache/phoenix/expression/function/LengthFunction.java b/phoenix-core/src/main/java/org/apache/phoenix/expression/function/LengthFunction.java
index 76d9e59..e69b45a 100644
--- a/phoenix-core/src/main/java/org/apache/phoenix/expression/function/LengthFunction.java
+++ b/phoenix-core/src/main/java/org/apache/phoenix/expression/function/LengthFunction.java
@@ -36,8 +36,8 @@ import org.apache.phoenix.util.StringUtil;
/**
*
- * Implementation of the LENGTH(<string>) build-in function. <string> is the string
- * of characters we want to find the length of. If <string> is NULL or empty, null
+ * Implementation of the {@code LENGTH(<string>) } build-in function. {@code <string> } is the string
+ * of characters we want to find the length of. If {@code <string> } is NULL or empty, null
* is returned.
*
*
diff --git a/phoenix-core/src/main/java/org/apache/phoenix/expression/function/NthValueFunction.java b/phoenix-core/src/main/java/org/apache/phoenix/expression/function/NthValueFunction.java
index 40502e5..08f6b6e 100644
--- a/phoenix-core/src/main/java/org/apache/phoenix/expression/function/NthValueFunction.java
+++ b/phoenix-core/src/main/java/org/apache/phoenix/expression/function/NthValueFunction.java
@@ -31,7 +31,7 @@ import org.apache.phoenix.schema.types.PBoolean;
import org.apache.phoenix.schema.types.PInteger;
/**
- * Built-in function for NTH_VALUE(<expression>, <expression>) WITHIN GROUP (ORDER BY <expression> ASC/DESC)
+ * Built-in function for {@code NTH_VALUE(<expression>, <expression>) WITHIN GROUP (ORDER BY <expression> ASC/DESC) }
* aggregate function
*
*/
diff --git a/phoenix-core/src/main/java/org/apache/phoenix/expression/function/PercentRankAggregateFunction.java b/phoenix-core/src/main/java/org/apache/phoenix/expression/function/PercentRankAggregateFunction.java
index 51dc816..d56129f 100644
--- a/phoenix-core/src/main/java/org/apache/phoenix/expression/function/PercentRankAggregateFunction.java
+++ b/phoenix-core/src/main/java/org/apache/phoenix/expression/function/PercentRankAggregateFunction.java
@@ -34,7 +34,7 @@ import org.apache.phoenix.schema.types.PDataType;
/**
*
- * PERCENT_RANK(<expression>[,<expression>]) WITHIN GROUP (ORDER BY <expression>[,<expression>] ASC/DESC) aggregate function
+ * {@code PERCENT_RANK(<expression>[,<expression>]) WITHIN GROUP (ORDER BY <expression>[,<expression>] ASC/DESC) } aggregate function
*
*
* @since 1.2.1
diff --git a/phoenix-core/src/main/java/org/apache/phoenix/expression/function/PercentileContAggregateFunction.java b/phoenix-core/src/main/java/org/apache/phoenix/expression/function/PercentileContAggregateFunction.java
index 17bb33c..5a466e1 100644
--- a/phoenix-core/src/main/java/org/apache/phoenix/expression/function/PercentileContAggregateFunction.java
+++ b/phoenix-core/src/main/java/org/apache/phoenix/expression/function/PercentileContAggregateFunction.java
@@ -34,7 +34,7 @@ import org.apache.phoenix.schema.types.PDataType;
/**
*
- * Built-in function for PERCENTILE_CONT(<expression>) WITHIN GROUP (ORDER BY <expression> ASC/DESC) aggregate function
+ * Built-in function for {@code PERCENTILE_CONT(<expression>) WITHIN GROUP (ORDER BY <expression> ASC/DESC) aggregate function }
*
*
* @since 1.2.1
diff --git a/phoenix-core/src/main/java/org/apache/phoenix/expression/function/PercentileDiscAggregateFunction.java b/phoenix-core/src/main/java/org/apache/phoenix/expression/function/PercentileDiscAggregateFunction.java
index 13482f0..b79cf8b 100644
--- a/phoenix-core/src/main/java/org/apache/phoenix/expression/function/PercentileDiscAggregateFunction.java
+++ b/phoenix-core/src/main/java/org/apache/phoenix/expression/function/PercentileDiscAggregateFunction.java
@@ -33,7 +33,7 @@ import org.apache.phoenix.schema.types.PBoolean;
/**
*
- * Built-in function for PERCENTILE_DISC(<expression>) WITHIN GROUP (ORDER BY <expression> ASC/DESC) aggregate function
+ * Built-in function for {@code PERCENTILE_DISC(<expression>) WITHIN GROUP (ORDER BY <expression> ASC/DESC) aggregate function }
*
*
* @since 1.2.1
diff --git a/phoenix-core/src/main/java/org/apache/phoenix/expression/function/RTrimFunction.java b/phoenix-core/src/main/java/org/apache/phoenix/expression/function/RTrimFunction.java
index 81e4f9e..da09eeb 100644
--- a/phoenix-core/src/main/java/org/apache/phoenix/expression/function/RTrimFunction.java
+++ b/phoenix-core/src/main/java/org/apache/phoenix/expression/function/RTrimFunction.java
@@ -42,8 +42,8 @@ import org.apache.phoenix.util.StringUtil;
/**
*
- * Implementation of the RTrim(<string>) build-in function. It removes from the right end of
- * <string> space character and other function bytes in single byte utf8 characters set
+ * Implementation of the {@code RTrim(<string>) } build-in function. It removes from the right end of
+ * {@code <string> } space character and other function bytes in single byte utf8 characters set
*
*
* @since 0.1
diff --git a/phoenix-core/src/main/java/org/apache/phoenix/expression/function/RandomFunction.java b/phoenix-core/src/main/java/org/apache/phoenix/expression/function/RandomFunction.java
index d9048f4..56e0642 100644
--- a/phoenix-core/src/main/java/org/apache/phoenix/expression/function/RandomFunction.java
+++ b/phoenix-core/src/main/java/org/apache/phoenix/expression/function/RandomFunction.java
@@ -40,6 +40,7 @@ import org.apache.phoenix.schema.types.PLong;
* <p>
* Example:
* <pre>
+ * {@code
* 0: jdbc:phoenix:localhost> select rand(), rand(), rand(1), rand(2), rand(1) from t;
* +----------------------------+----------------------------+----------------------------+----------------------------+-----------------------+
* | RAND() | RAND() | RAND(1) | RAND(2) | RAND(1) |
@@ -56,6 +57,7 @@ import org.apache.phoenix.schema.types.PLong;
* | 0.8084646053276106 | 0.6969504742211767 | 0.41008081149220166 | 0.9014476240300544 | 0.41008081149220166 |
* +----------------------------+----------------------------+----------------------------+----------------------------+-----------------------+
* 2 rows selected (0.098 seconds)
+ * }
* </pre>
*/
@BuiltInFunction(name = RandomFunction.NAME, args = {@Argument(allowedTypes={PLong.class},defaultValue="null",isConstant=true)})
diff --git a/phoenix-core/src/main/java/org/apache/phoenix/expression/function/RegexpReplaceFunction.java b/phoenix-core/src/main/java/org/apache/phoenix/expression/function/RegexpReplaceFunction.java
index 265b860..1a29bca 100644
--- a/phoenix-core/src/main/java/org/apache/phoenix/expression/function/RegexpReplaceFunction.java
+++ b/phoenix-core/src/main/java/org/apache/phoenix/expression/function/RegexpReplaceFunction.java
@@ -39,7 +39,7 @@ import org.apache.phoenix.schema.types.PVarchar;
*
* Function similar to the regexp_replace function in Postgres, which is used to pattern
* match a segment of the string. Usage:
- * REGEXP_REPLACE(<source_char>,<pattern>,<replace_string>)
+ * {@code REGEXP_REPLACE(<source_char>,<pattern>,<replace_string>) }
* source_char is the string in which we want to perform string replacement. pattern is a
* Java compatible regular expression string, and we replace all the matching part with
* replace_string. The first 2 arguments are required and are {@link org.apache.phoenix.schema.types.PVarchar},
diff --git a/phoenix-core/src/main/java/org/apache/phoenix/expression/function/RegexpSubstrFunction.java b/phoenix-core/src/main/java/org/apache/phoenix/expression/function/RegexpSubstrFunction.java
index 466fd1a..a699a6b 100644
--- a/phoenix-core/src/main/java/org/apache/phoenix/expression/function/RegexpSubstrFunction.java
+++ b/phoenix-core/src/main/java/org/apache/phoenix/expression/function/RegexpSubstrFunction.java
@@ -39,12 +39,12 @@ import org.apache.phoenix.schema.types.PVarchar;
/**
*
- * Implementation of REGEXP_SUBSTR(<source>, <pattern>, <offset>) built-in function,
- * where <offset> is the offset from the start of <string>. Positive offset is treated as 1-based,
+ * Implementation of {@code REGEXP_SUBSTR(<source>, <pattern>, <offset>) } built-in function,
+ * where {@code <offset> } is the offset from the start of {@code <string> }. Positive offset is treated as 1-based,
* a zero offset is treated as 0-based, and a negative offset starts from the end of the string
- * working backwards. The <pattern> is the pattern we would like to search for in the <source> string.
- * The function returns the first occurrence of any substring in the <source> string that matches
- * the <pattern> input as a VARCHAR.
+ * working backwards. The {@code <pattern> } is the pattern we would like to search for in the {@code <source> } string.
+ * The function returns the first occurrence of any substring in the {@code <source> } string that matches
+ * the {@code <pattern> } input as a VARCHAR.
*
*
* @since 0.1
diff --git a/phoenix-core/src/main/java/org/apache/phoenix/expression/function/RoundDateExpression.java b/phoenix-core/src/main/java/org/apache/phoenix/expression/function/RoundDateExpression.java
index 988b027..372b215 100644
--- a/phoenix-core/src/main/java/org/apache/phoenix/expression/function/RoundDateExpression.java
+++ b/phoenix-core/src/main/java/org/apache/phoenix/expression/function/RoundDateExpression.java
@@ -52,7 +52,7 @@ import com.google.common.collect.Lists;
/**
* Function used to bucketize date/time values by rounding them to
* an even increment. Usage:
- * ROUND(<date/time col ref>,<'day'|'hour'|'minute'|'second'|'millisecond'|'week'|'month'|'year'>,<optional integer multiplier>)
+ * {@code ROUND(<date/time col ref>,<'day'|'hour'|'minute'|'second'|'millisecond'|'week'|'month'|'year'>,<optional integer multiplier>) }
* The integer multiplier is optional and is used to do rollups to a partial time unit (i.e. 10 minute rollup)
* The function returns a {@link org.apache.phoenix.schema.types.PDate}
diff --git a/phoenix-core/src/main/java/org/apache/phoenix/expression/function/StddevPopFunction.java b/phoenix-core/src/main/java/org/apache/phoenix/expression/function/StddevPopFunction.java
index 338031b..e6e79fc 100644
--- a/phoenix-core/src/main/java/org/apache/phoenix/expression/function/StddevPopFunction.java
+++ b/phoenix-core/src/main/java/org/apache/phoenix/expression/function/StddevPopFunction.java
@@ -34,7 +34,7 @@ import org.apache.phoenix.schema.types.PDataType;
/**
*
- * Built-in function for STDDEV_POP(<expression>) aggregate function
+ * Built-in function for {@code STDDEV_POP(<expression>) } aggregate function
*
*
* @since 1.2.1
diff --git a/phoenix-core/src/main/java/org/apache/phoenix/expression/function/StddevSampFunction.java b/phoenix-core/src/main/java/org/apache/phoenix/expression/function/StddevSampFunction.java
index 0f22c75..16d27ad 100644
--- a/phoenix-core/src/main/java/org/apache/phoenix/expression/function/StddevSampFunction.java
+++ b/phoenix-core/src/main/java/org/apache/phoenix/expression/function/StddevSampFunction.java
@@ -34,7 +34,7 @@ import org.apache.phoenix.schema.types.PDataType;
/**
*
- * Built-in function for STDDEV_SAMP(<expression>) aggregate function
+ * Built-in function for {@code STDDEV_SAMP(<expression>) } aggregate function
*
*
* @since 1.2.1
diff --git a/phoenix-core/src/main/java/org/apache/phoenix/expression/function/SubstrFunction.java b/phoenix-core/src/main/java/org/apache/phoenix/expression/function/SubstrFunction.java
index 0d6d1c95..7cf32e0 100644
--- a/phoenix-core/src/main/java/org/apache/phoenix/expression/function/SubstrFunction.java
+++ b/phoenix-core/src/main/java/org/apache/phoenix/expression/function/SubstrFunction.java
@@ -37,13 +37,13 @@ import org.apache.phoenix.util.StringUtil;
/**
*
- * Implementation of the SUBSTR(<string>,<offset>[,<length>]) built-in function
- * where <offset> is the offset from the start of <string>. A positive offset
+ * Implementation of the {@code SUBSTR(<string>,<offset>[,<length>]) } built-in function
+ * where {@code <offset> } is the offset from the start of {@code <string> }. A positive offset
* is treated as 1-based, a zero offset is treated as 0-based, and a negative
* offset starts from the end of the string working backwards. The optional
- * <length> argument is the number of characters to return. In the absence of the
- * <length> argument, the rest of the string starting from <offset> is returned.
- * If <length> is less than 1, null is returned.
+ * {@code <length> } argument is the number of characters to return. In the absence of the
+ * {@code <length> } argument, the rest of the string starting from {@code <offset> } is returned.
+ * If {@code <length> } is less than 1, null is returned.
*
*
* @since 0.1
diff --git a/phoenix-core/src/main/java/org/apache/phoenix/expression/function/TrimFunction.java b/phoenix-core/src/main/java/org/apache/phoenix/expression/function/TrimFunction.java
index 12b53c7..2565445 100644
--- a/phoenix-core/src/main/java/org/apache/phoenix/expression/function/TrimFunction.java
+++ b/phoenix-core/src/main/java/org/apache/phoenix/expression/function/TrimFunction.java
@@ -33,7 +33,7 @@ import org.apache.phoenix.util.StringUtil;
/**
- * Implementation of the Trim(<string>) build-in function. It removes from both end of <string>
+ * Implementation of the {@code Trim(<string>) } build-in function. It removes from both end of {@code <string> }
* space character and other function bytes in single byte utf8 characters set.
*
*
diff --git a/phoenix-core/src/main/java/org/apache/phoenix/expression/function/TruncFunction.java b/phoenix-core/src/main/java/org/apache/phoenix/expression/function/TruncFunction.java
index 83297a1..6d223a7 100644
--- a/phoenix-core/src/main/java/org/apache/phoenix/expression/function/TruncFunction.java
+++ b/phoenix-core/src/main/java/org/apache/phoenix/expression/function/TruncFunction.java
@@ -34,7 +34,7 @@ import org.apache.phoenix.schema.types.PVarchar;
*
* Function used to bucketize date/time values by truncating them to
* an even increment. Usage:
- * TRUNC(<date/time col ref>,<'day'|'hour'|'minute'|'second'|'millisecond'>,[<optional integer multiplier>])
+ * {@code TRUNC(<date/time col ref>,<'day'|'hour'|'minute'|'second'|'millisecond'>,[<optional integer multiplier>]) }
* The integer multiplier is optional and is used to do rollups to a partial time unit (i.e. 10 minute rollup)
* The function returns a {@link org.apache.phoenix.schema.types.PDate}
*
diff --git a/phoenix-core/src/main/java/org/apache/phoenix/hbase/index/Indexer.java b/phoenix-core/src/main/java/org/apache/phoenix/hbase/index/Indexer.java
index a727346..f383976 100644
--- a/phoenix-core/src/main/java/org/apache/phoenix/hbase/index/Indexer.java
+++ b/phoenix-core/src/main/java/org/apache/phoenix/hbase/index/Indexer.java
@@ -99,7 +99,7 @@ import com.google.common.collect.Multimap;
* If the WAL is enabled, these updates are then added to the WALEdit and attempted to be written to
* the WAL after the WALEdit has been saved. If any of the index updates fail, this server is
* immediately terminated and we rely on WAL replay to attempt the index updates again (see
- * {@link #preWALRestore(ObserverContext, HRegionInfo, HLogKey, WALEdit)}).
+ * #preWALRestore(ObserverContext, HRegionInfo, HLogKey, WALEdit)).
* <p>
* If the WAL is disabled, the updates are attempted immediately. No consistency guarantees are made
* if the WAL is disabled - some or none of the index updates may be successful. All updates in a
@@ -760,8 +760,8 @@ public class Indexer extends BaseRegionObserver {
/**
* Validate that the version and configuration parameters are supported
* @param hbaseVersion current version of HBase on which <tt>this</tt> coprocessor is installed
- * @param conf configuration to check for allowed parameters (e.g. WAL Compression only if >=
- * 0.94.9)
+ * @param conf configuration to check for allowed parameters (e.g. WAL Compression only {@code if >=
+ * 0.94.9) }
* @return <tt>null</tt> if the version is supported, the error message to display otherwise
*/
public static String validateVersion(String hbaseVersion, Configuration conf) {
@@ -787,10 +787,10 @@ public class Indexer extends BaseRegionObserver {
/**
* Enable indexing on the given table
* @param desc {@link HTableDescriptor} for the table on which indexing should be enabled
- * @param builder class to use when building the index for this table
- * @param properties map of custom configuration options to make available to your
+ * @param builder class to use when building the index for this table
+ * @param properties map of custom configuration options to make available to your
* {@link IndexBuilder} on the server-side
- * @param priority TODO
+ * @param priority TODO
* @throws IOException the Indexer coprocessor cannot be added
*/
public static void enableIndexing(HTableDescriptor desc, Class<? extends IndexBuilder> builder,
diff --git a/phoenix-core/src/main/java/org/apache/phoenix/hbase/index/builder/BaseIndexBuilder.java b/phoenix-core/src/main/java/org/apache/phoenix/hbase/index/builder/BaseIndexBuilder.java
index e3fdc1b..51a642b 100644
--- a/phoenix-core/src/main/java/org/apache/phoenix/hbase/index/builder/BaseIndexBuilder.java
+++ b/phoenix-core/src/main/java/org/apache/phoenix/hbase/index/builder/BaseIndexBuilder.java
@@ -83,8 +83,6 @@ public abstract class BaseIndexBuilder implements IndexBuilder {
* By default, we always attempt to index the mutation. Commonly this can be slow (because the framework spends the
* time to do the indexing, only to realize that you don't need it) or not ideal (if you want to turn on/off
* indexing on a table without completely reloading it).
- *
- * @throws IOException
*/
@Override
public boolean isEnabled(Mutation m) {
diff --git a/phoenix-core/src/main/java/org/apache/phoenix/hbase/index/builder/IndexBuilder.java b/phoenix-core/src/main/java/org/apache/phoenix/hbase/index/builder/IndexBuilder.java
index 18e361a..e550b3a 100644
--- a/phoenix-core/src/main/java/org/apache/phoenix/hbase/index/builder/IndexBuilder.java
+++ b/phoenix-core/src/main/java/org/apache/phoenix/hbase/index/builder/IndexBuilder.java
@@ -70,13 +70,13 @@ public interface IndexBuilder extends Stoppable {
* called concurrently for different mutations, which may or may not be part of the same batch.
* @param mutation update to the primary table to be indexed.
* @param context index meta data for the mutation
- * @return a Map of the mutations to make -> target index table name
+ * @return a Map of the mutations to {@code make -> target } index table name
* @throws IOException on failure
*/
public Collection<Pair<Mutation, byte[]>> getIndexUpdate(Mutation mutation, IndexMetaData context, LocalHBaseState localHBaseState) throws IOException;
/**
- * Build an index update to cleanup the index when we remove {@link KeyValue}s via the normal flush or compaction
+ * Build an index update to cleanup the index when we remove KeyValue s via the normal flush or compaction
* mechanisms. Currently not implemented by any implementors nor called, but left here to be implemented if we
* ever need it. In Jesse's words:
*
@@ -89,11 +89,11 @@ public interface IndexBuilder extends Stoppable {
* b/c its covered by the delete marker, but an older timestamp based read would actually show the index row, even
* after the primary table row is gone due to MAX_VERSIONS requirement.
*
- * @param filtered {@link KeyValue}s that previously existed, but won't be included
+ * @param filtered KeyValue s that previously existed, but won't be included
* in further output from HBase.
* @param context TODO
*
- * @return a {@link Map} of the mutations to make -> target index table name
+ * @return a {@link Map} of the mutations to {@code make -> target } index table name
* @throws IOException on failure
*/
public Collection<Pair<Mutation, byte[]>> getIndexUpdateForFilteredRows(
@@ -131,7 +131,6 @@ public interface IndexBuilder extends Stoppable {
* @param m mutation that should be indexed.
* @return <tt>true</tt> if indexing is enabled for the given table. This should be on a per-table
* basis, as each codec is instantiated per-region.
- * @throws IOException
*/
public boolean isEnabled(Mutation m);
@@ -139,7 +138,6 @@ public interface IndexBuilder extends Stoppable {
* True if mutation has an ON DUPLICATE KEY clause
* @param m mutation
* @return true if mutation has ON DUPLICATE KEY expression and false otherwise.
- * @throws IOException
*/
public boolean isAtomicOp(Mutation m);
diff --git a/phoenix-core/src/main/java/org/apache/phoenix/hbase/index/covered/IndexCodec.java b/phoenix-core/src/main/java/org/apache/phoenix/hbase/index/covered/IndexCodec.java
index 6762a30..abcb347 100644
--- a/phoenix-core/src/main/java/org/apache/phoenix/hbase/index/covered/IndexCodec.java
+++ b/phoenix-core/src/main/java/org/apache/phoenix/hbase/index/covered/IndexCodec.java
@@ -80,7 +80,6 @@ public interface IndexCodec {
* mutation that should be indexed.
* @return <tt>true</tt> if indexing is enabled for the given table. This should be on a per-table basis, as each
* codec is instantiated per-region.
- * @throws IOException
*/
public boolean isEnabled(Mutation m);
diff --git a/phoenix-core/src/main/java/org/apache/phoenix/hbase/index/covered/LocalTableState.java b/phoenix-core/src/main/java/org/apache/phoenix/hbase/index/covered/LocalTableState.java
index be18cbe..c52d5e0 100644
--- a/phoenix-core/src/main/java/org/apache/phoenix/hbase/index/covered/LocalTableState.java
+++ b/phoenix-core/src/main/java/org/apache/phoenix/hbase/index/covered/LocalTableState.java
@@ -136,7 +136,7 @@ public class LocalTableState implements TableState {
* state for any of the columns you are indexing.
* <p>
* <i>NOTE:</i> This method should <b>not</b> be used during
- * {@link IndexCodec#getIndexDeletes(TableState, BatchState, byte[], byte[])} as the pending update will not yet have been
+ * IndexCodec#getIndexDeletes(TableState, BatchState, byte[], byte[]) as the pending update will not yet have been
* applied - you are merely attempting to cleanup the current state and therefore do <i>not</i>
* need to track the indexed columns.
* <p>
diff --git a/phoenix-core/src/main/java/org/apache/phoenix/hbase/index/covered/data/IndexMemStore.java b/phoenix-core/src/main/java/org/apache/phoenix/hbase/index/covered/data/IndexMemStore.java
index 4a62e14..506f944 100644
--- a/phoenix-core/src/main/java/org/apache/phoenix/hbase/index/covered/data/IndexMemStore.java
+++ b/phoenix-core/src/main/java/org/apache/phoenix/hbase/index/covered/data/IndexMemStore.java
@@ -52,7 +52,7 @@ import org.slf4j.LoggerFactory;
* <li>ignoring memstore timestamps in favor of deciding when we want to overwrite keys based on how
* we obtain them</li>
* <li>ignoring time range updates (so
- * {@link ReseekableScanner#shouldUseScanner(Scan, SortedSet, long)} isn't supported from
+ * ReseekableScanner#shouldUseScanner(Scan, SortedSet, long) isn't supported from
* {@link #getScanner()}).</li>
* </ol>
* <p>
@@ -67,7 +67,7 @@ import org.slf4j.LoggerFactory;
* the previous implementation. Further, by being smart about how we manage the KVs, we can drop the
* extra object creation we were doing to wrap the pending KVs (which we did previously to ensure
* they sorted before the ones we got from the HRegion). We overwrite {@link KeyValue}s when we add
- * them from external sources {@link #add(KeyValue, boolean)}, but then don't overwrite existing
+ * them from external sources #add(KeyValue, boolean), but then don't overwrite existing
* keyvalues when read them from the underlying table (because pending keyvalues should always
* overwrite current ones) - this logic is all contained in LocalTableState.
* @see LocalTableState
diff --git a/phoenix-core/src/main/java/org/apache/phoenix/hbase/index/covered/filter/NewerTimestampFilter.java b/phoenix-core/src/main/java/org/apache/phoenix/hbase/index/covered/filter/NewerTimestampFilter.java
index b67aad6..d8f7c8e 100644
--- a/phoenix-core/src/main/java/org/apache/phoenix/hbase/index/covered/filter/NewerTimestampFilter.java
+++ b/phoenix-core/src/main/java/org/apache/phoenix/hbase/index/covered/filter/NewerTimestampFilter.java
@@ -23,7 +23,7 @@ import org.apache.hadoop.hbase.filter.FilterBase;
/**
* Server-side only class used in the indexer to filter out keyvalues newer than a given timestamp
- * (so allows anything <code><=</code> timestamp through).
+ * (so allows anything {@code <= } timestamp through).
* <p>
*/
public class NewerTimestampFilter extends FilterBase {
diff --git a/phoenix-core/src/main/java/org/apache/phoenix/hbase/index/exception/SingleIndexWriteFailureException.java b/phoenix-core/src/main/java/org/apache/phoenix/hbase/index/exception/SingleIndexWriteFailureException.java
index 67dabba..81235ee 100644
--- a/phoenix-core/src/main/java/org/apache/phoenix/hbase/index/exception/SingleIndexWriteFailureException.java
+++ b/phoenix-core/src/main/java/org/apache/phoenix/hbase/index/exception/SingleIndexWriteFailureException.java
@@ -60,7 +60,7 @@ public class SingleIndexWriteFailureException extends IndexWriteException {
/**
* This constructor used to rematerialize this exception when receiving
* an rpc exception from the server
- * @param message detail message
+ * @param msg detail message
*/
public SingleIndexWriteFailureException(String msg) {
super(IndexWriteException.parseDisableIndexOnFailure(msg));
diff --git a/phoenix-core/src/main/java/org/apache/phoenix/hbase/index/scanner/FilteredKeyValueScanner.java b/phoenix-core/src/main/java/org/apache/phoenix/hbase/index/scanner/FilteredKeyValueScanner.java
index 06e74a3..b1eb2ed 100644
--- a/phoenix-core/src/main/java/org/apache/phoenix/hbase/index/scanner/FilteredKeyValueScanner.java
+++ b/phoenix-core/src/main/java/org/apache/phoenix/hbase/index/scanner/FilteredKeyValueScanner.java
@@ -52,7 +52,7 @@ public class FilteredKeyValueScanner implements ReseekableScanner {
}
/**
- * Same a {@link KeyValueScanner#next()} except that we filter out the next {@link KeyValue} until we find one that
+ * Same a KeyValueScanner#next() except that we filter out the next {@link KeyValue} until we find one that
* passes the filter.
*
* @return the next {@link KeyValue} or <tt>null</tt> if no next {@link KeyValue} is present and passes all the
diff --git a/phoenix-core/src/main/java/org/apache/phoenix/hbase/index/util/IndexManagementUtil.java b/phoenix-core/src/main/java/org/apache/phoenix/hbase/index/util/IndexManagementUtil.java
index f6cf813..9f36df1 100644
--- a/phoenix-core/src/main/java/org/apache/phoenix/hbase/index/util/IndexManagementUtil.java
+++ b/phoenix-core/src/main/java/org/apache/phoenix/hbase/index/util/IndexManagementUtil.java
@@ -184,7 +184,7 @@ public class IndexManagementUtil {
* Propagate the given failure as a generic {@link IOException}, if it isn't already
*
* @param e
- * reason indexing failed. If ,tt>null</tt>, throws a {@link NullPointerException}, which should unload
+ * reason indexing failed. If ,<tt>null</tt>, throws a {@link NullPointerException}, which should unload
* the coprocessor.
*/
public static void rethrowIndexingException(Throwable e) throws IOException {
diff --git a/phoenix-core/src/main/java/org/apache/phoenix/hbase/index/wal/KeyValueCodec.java b/phoenix-core/src/main/java/org/apache/phoenix/hbase/index/wal/KeyValueCodec.java
index 9678c7b..658ee61 100644
--- a/phoenix-core/src/main/java/org/apache/phoenix/hbase/index/wal/KeyValueCodec.java
+++ b/phoenix-core/src/main/java/org/apache/phoenix/hbase/index/wal/KeyValueCodec.java
@@ -33,7 +33,7 @@ import org.apache.hadoop.hbase.util.Bytes;
import org.apache.phoenix.hbase.index.util.ImmutableBytesPtr;
/**
- * Codec to encode/decode {@link KeyValue}s and {@link IndexedKeyValue}s within a {@link WALEdit}
+ * Codec to encode/decode KeyValue s and IndexedKeyValue s within a WALEdit
*/
public class KeyValueCodec {
diff --git a/phoenix-core/src/main/java/org/apache/phoenix/hbase/index/write/IndexFailurePolicy.java b/phoenix-core/src/main/java/org/apache/phoenix/hbase/index/write/IndexFailurePolicy.java
index 3335aaa..b46bdf6 100644
--- a/phoenix-core/src/main/java/org/apache/phoenix/hbase/index/write/IndexFailurePolicy.java
+++ b/phoenix-core/src/main/java/org/apache/phoenix/hbase/index/write/IndexFailurePolicy.java
@@ -34,9 +34,9 @@ public interface IndexFailurePolicy extends Stoppable {
/**
* Handle the failure of the attempted index updates
- * @param attempted map of index table -> mutations to apply
+ * @param attempted map of index {@code table -> mutations } to apply
* @param cause reason why there was a failure
- * @throws IOException
+ * @throws IOException
*/
public void
handleFailure(Multimap<HTableInterfaceReference, Mutation> attempted, Exception cause) throws IOException;
diff --git a/phoenix-core/src/main/java/org/apache/phoenix/hbase/index/write/IndexWriter.java b/phoenix-core/src/main/java/org/apache/phoenix/hbase/index/write/IndexWriter.java
index ca38e43..dab4174 100644
--- a/phoenix-core/src/main/java/org/apache/phoenix/hbase/index/write/IndexWriter.java
+++ b/phoenix-core/src/main/java/org/apache/phoenix/hbase/index/write/IndexWriter.java
@@ -133,7 +133,7 @@ public class IndexWriter implements Stoppable {
}
/**
- * see {@link #writeAndHandleFailure(Collection)}.
+ * see #writeAndHandleFailure(Collection).
* @param toWrite
* @throws IOException
*/
@@ -194,7 +194,7 @@ public class IndexWriter implements Stoppable {
}
/**
- * see {@link #write(Collection)}
+ * see #write(Collection)
* @param toWrite
* @throws IndexWriteException
*/
diff --git a/phoenix-core/src/main/java/org/apache/phoenix/hbase/index/write/IndexWriterUtils.java b/phoenix-core/src/main/java/org/apache/phoenix/hbase/index/write/IndexWriterUtils.java
index dd43cb2..c2bbaf9 100644
--- a/phoenix-core/src/main/java/org/apache/phoenix/hbase/index/write/IndexWriterUtils.java
+++ b/phoenix-core/src/main/java/org/apache/phoenix/hbase/index/write/IndexWriterUtils.java
@@ -34,7 +34,7 @@ public class IndexWriterUtils {
/**
* Maximum number of threads to allow per-table when writing. Each writer thread (from
- * {@link IndexWriterUtils#NUM_CONCURRENT_INDEX_WRITER_THREADS_CONF_KEY}) has a single HTable.
+ * IndexWriterUtils#NUM_CONCURRENT_INDEX_WRITER_THREADS_CONF_KEY) has a single HTable.
* However, each table is backed by a threadpool to manage the updates to that table. this
* specifies the number of threads to allow in each of those tables. Generally, you shouldn't need
* to change this, unless you have a small number of indexes to which most of the writes go.
diff --git a/phoenix-core/src/main/java/org/apache/phoenix/hbase/index/write/TrackingParallelWriterIndexCommitter.java b/phoenix-core/src/main/java/org/apache/phoenix/hbase/index/write/TrackingParallelWriterIndexCommitter.java
index 7bbe3fa..ab65662 100644
--- a/phoenix-core/src/main/java/org/apache/phoenix/hbase/index/write/TrackingParallelWriterIndexCommitter.java
+++ b/phoenix-core/src/main/java/org/apache/phoenix/hbase/index/write/TrackingParallelWriterIndexCommitter.java
@@ -57,7 +57,7 @@ import static org.apache.phoenix.util.ServerUtil.wrapInDoNotRetryIOException;
* you need to at least attempt all writes and know their result; for instance, this is fine for doing WAL recovery -
* it's not a performance intensive situation and we want to limit the the edits we need to retry.
* <p>
- * On failure to {@link #write(Multimap)}, we return a {@link MultiIndexWriteFailureException} that contains the list of
+ * On failure to #write(Multimap), we return a MultiIndexWriteFailureException that contains the list of
* {@link HTableInterfaceReference} that didn't complete successfully.
* <p>
* Failures to write to the index can happen several different ways:
diff --git a/phoenix-core/src/main/java/org/apache/phoenix/hbase/index/write/recovery/PerRegionIndexWriteCache.java b/phoenix-core/src/main/java/org/apache/phoenix/hbase/index/write/recovery/PerRegionIndexWriteCache.java
index 26da2d5..9a213e2 100644
--- a/phoenix-core/src/main/java/org/apache/phoenix/hbase/index/write/recovery/PerRegionIndexWriteCache.java
+++ b/phoenix-core/src/main/java/org/apache/phoenix/hbase/index/write/recovery/PerRegionIndexWriteCache.java
@@ -38,7 +38,7 @@ public class PerRegionIndexWriteCache {
/**
* Get the edits for the current region. Removes the edits from the cache. To add them back, call
- * {@link #addEdits(HRegion, HTableInterfaceReference, Collection)}.
+ * #addEdits(HRegion, HTableInterfaceReference, Collection).
* @param region
* @return Get the edits for the given region. Returns <tt>null</tt> if there are no pending edits
* for the region
diff --git a/phoenix-core/src/main/java/org/apache/phoenix/index/IndexMaintainer.java b/phoenix-core/src/main/java/org/apache/phoenix/index/IndexMaintainer.java
index c0bd10f..3d78e8c 100644
--- a/phoenix-core/src/main/java/org/apache/phoenix/index/IndexMaintainer.java
+++ b/phoenix-core/src/main/java/org/apache/phoenix/index/IndexMaintainer.java
@@ -124,7 +124,7 @@ import com.google.common.collect.Sets;
*
* Class that builds index row key from data row key and current state of
* row and caches any covered columns. Client-side serializes into byte array using
- * @link #serialize(PTable, ImmutableBytesWritable)}
+ * #serialize(PTable, ImmutableBytesWritable)
* and transmits to server-side through either the
* {@link org.apache.phoenix.index.PhoenixIndexCodec#INDEX_PROTO_MD}
* Mutation attribute or as a separate RPC call using
diff --git a/phoenix-core/src/main/java/org/apache/phoenix/index/IndexMetaDataCacheClient.java b/phoenix-core/src/main/java/org/apache/phoenix/index/IndexMetaDataCacheClient.java
index bd308c0..1f6dd73 100644
--- a/phoenix-core/src/main/java/org/apache/phoenix/index/IndexMetaDataCacheClient.java
+++ b/phoenix-core/src/main/java/org/apache/phoenix/index/IndexMetaDataCacheClient.java
@@ -48,7 +48,7 @@ public class IndexMetaDataCacheClient {
* Construct client used to send index metadata to each region server
* for caching during batched put for secondary index maintenance.
* @param connection the client connection
- * @param cacheUsingTableRef table ref to table that will use the cache during its scan
+ * @param cacheUsingTable table ref to table that will use the cache during its scan
*/
public IndexMetaDataCacheClient(PhoenixConnection connection, PTable cacheUsingTable) {
serverCache = new ServerCacheClient(connection);
diff --git a/phoenix-core/src/main/java/org/apache/phoenix/index/PhoenixIndexBuilder.java b/phoenix-core/src/main/java/org/apache/phoenix/index/PhoenixIndexBuilder.java
index ca7c637..3f3c303 100644
--- a/phoenix-core/src/main/java/org/apache/phoenix/index/PhoenixIndexBuilder.java
+++ b/phoenix-core/src/main/java/org/apache/phoenix/index/PhoenixIndexBuilder.java
@@ -286,7 +286,7 @@ public class PhoenixIndexBuilder extends NonTxIndexBuilder {
* UPSERT VALUES ON DUPLICATE KEY IGNORE followed by UPSERT VALUES ON DUPLICATE KEY UPDATE
* 2) Short value tracking how many times the next first clause should be executed. This
* optimizes the same clause be executed many times by only serializing it once.
- * 3) Repeating {List<Expression>, PTable} pairs that encapsulate the ON DUPLICATE KEY clause.
+ * 3) Repeating {@code List<Expression>, PTable } pairs that encapsulate the ON DUPLICATE KEY clause.
* @param table table representing columns being updated
* @param expressions list of expressions to evaluate for updating columns
* @return serialized byte array representation of ON DUPLICATE KEY UPDATE info
diff --git a/phoenix-core/src/main/java/org/apache/phoenix/index/PhoenixIndexFailurePolicy.java b/phoenix-core/src/main/java/org/apache/phoenix/index/PhoenixIndexFailurePolicy.java
index 45bd7dd..3076f70 100644
--- a/phoenix-core/src/main/java/org/apache/phoenix/index/PhoenixIndexFailurePolicy.java
+++ b/phoenix-core/src/main/java/org/apache/phoenix/index/PhoenixIndexFailurePolicy.java
@@ -377,11 +377,8 @@ public class PhoenixIndexFailurePolicy extends DelegateIndexFailurePolicy {
/**
* Check config for whether to disable index on index write failures
- * @param htd
- * @param config
- * @param connection
- * @return The table config for {@link PhoenixIndexFailurePolicy.DISABLE_INDEX_ON_WRITE_FAILURE}
- * @throws SQLException
+ * @param env
+ * @return
*/
public static boolean getDisableIndexOnFailure(RegionCoprocessorEnvironment env) {
HTableDescriptor htd = env.getRegion().getTableDesc();
@@ -460,7 +457,7 @@ public class PhoenixIndexFailurePolicy extends DelegateIndexFailurePolicy {
* @param iwe original IndexWriteException
* @param connection connection to use
* @param config config used to get retry settings
- * @throws Exception
+ * @throws IOException
*/
public static void doBatchWithRetries(MutateCommand mutateCommand,
IndexWriteException iwe, PhoenixConnection connection, ReadOnlyProps config)
diff --git a/phoenix-core/src/main/java/org/apache/phoenix/jdbc/PhoenixDriver.java b/phoenix-core/src/main/java/org/apache/phoenix/jdbc/PhoenixDriver.java
index 8ac5375..3dc3547 100644
--- a/phoenix-core/src/main/java/org/apache/phoenix/jdbc/PhoenixDriver.java
+++ b/phoenix-core/src/main/java/org/apache/phoenix/jdbc/PhoenixDriver.java
@@ -50,7 +50,7 @@ import com.google.common.util.concurrent.ThreadFactoryBuilder;
*
* JDBC Driver implementation of Phoenix for production.
* To use this driver, specify the following URL:
- * jdbc:phoenix:<zookeeper quorum server name>;
+ * {@code jdbc:phoenix:<zookeeper quorum server name>; }
* Only an embedded driver is currently supported (Phoenix client
* runs in the same JVM as the driver). Connections are lightweight
* and are not pooled. The last part of the URL, the hbase zookeeper
diff --git a/phoenix-core/src/main/java/org/apache/phoenix/jdbc/PhoenixStatement.java b/phoenix-core/src/main/java/org/apache/phoenix/jdbc/PhoenixStatement.java
index 3acff61..5cc1f64 100644
--- a/phoenix-core/src/main/java/org/apache/phoenix/jdbc/PhoenixStatement.java
+++ b/phoenix-core/src/main/java/org/apache/phoenix/jdbc/PhoenixStatement.java
@@ -1717,7 +1717,7 @@ public class PhoenixStatement implements Statement, SQLCloseable {
/**
* Execute the current batch of statements. If any exception occurs
- * during execution, a {@link org.apache.phoenix.exception.BatchUpdateException}
+ * during execution, a org.apache.phoenix.exception.BatchUpdateException
* is thrown which includes the index of the statement within the
* batch when the exception occurred.
*/
diff --git a/phoenix-core/src/main/java/org/apache/phoenix/join/HashCacheClient.java b/phoenix-core/src/main/java/org/apache/phoenix/join/HashCacheClient.java
index 315c515..efb7a05 100644
--- a/phoenix-core/src/main/java/org/apache/phoenix/join/HashCacheClient.java
+++ b/phoenix-core/src/main/java/org/apache/phoenix/join/HashCacheClient.java
@@ -85,7 +85,16 @@ public class HashCacheClient {
* Send the results of scanning through the scanner to all
* region servers for regions of the table that will use the cache
* that intersect with the minMaxKeyRange.
- * @param scanner scanner for the table or intermediate results being cached
+ * @param keyRanges
+ * @param cacheId
+ * @param iterator
+ * @param estimatedSize
+ * @param onExpressions
+ * @param singleValueOnly
+ * @param usePersistentCache
+ * @param cacheUsingTable
+ * @param keyRangeRhsExpression
+ * @param keyRangeRhsValues
* @return client-side {@link ServerCache} representing the added hash cache
* @throws SQLException
* @throws MaxServerCacheSizeExceededException if size of hash cache exceeds max allowed
diff --git a/phoenix-core/src/main/java/org/apache/phoenix/mapreduce/CsvToKeyValueMapper.java b/phoenix-core/src/main/java/org/apache/phoenix/mapreduce/CsvToKeyValueMapper.java
index 3b2293e..cb3afdf 100644
--- a/phoenix-core/src/main/java/org/apache/phoenix/mapreduce/CsvToKeyValueMapper.java
+++ b/phoenix-core/src/main/java/org/apache/phoenix/mapreduce/CsvToKeyValueMapper.java
@@ -36,7 +36,7 @@ import com.google.common.collect.Iterables;
/**
* MapReduce mapper that converts CSV input lines into KeyValues that can be written to HFiles.
- * <p/>
+ *
* KeyValues are produced by executing UPSERT statements on a Phoenix connection and then
* extracting the created KeyValues and rolling back the statement execution before it is
* committed to HBase.
diff --git a/phoenix-core/src/main/java/org/apache/phoenix/mapreduce/ImportPreUpsertKeyValueProcessor.java b/phoenix-core/src/main/java/org/apache/phoenix/mapreduce/ImportPreUpsertKeyValueProcessor.java
index dff9ef2..6835b81 100644
--- a/phoenix-core/src/main/java/org/apache/phoenix/mapreduce/ImportPreUpsertKeyValueProcessor.java
+++ b/phoenix-core/src/main/java/org/apache/phoenix/mapreduce/ImportPreUpsertKeyValueProcessor.java
@@ -26,7 +26,7 @@ import java.util.List;
* Implementing this interface and configuring it via the {@link
* org.apache.phoenix.mapreduce.util.PhoenixConfigurationUtil#UPSERT_HOOK_CLASS_CONFKEY}
* configuration key.
- * <p/>
+ *
* The intention of such a hook is to allow coproccessor-style operations to be peformed on
* data that is being bulk-loaded via MapReduce.
*/
@@ -35,7 +35,7 @@ public interface ImportPreUpsertKeyValueProcessor {
/**
* Process a list of KeyValues before they are written to an HFile. The supplied list of
* KeyValues contain all data that is to be written for a single Phoenix row.
- * <p/>
+ *
* Implementors can filter certain KeyValues from the list, augment the list, or return the
* same list.
*
diff --git a/phoenix-core/src/main/java/org/apache/phoenix/mapreduce/JsonToKeyValueMapper.java b/phoenix-core/src/main/java/org/apache/phoenix/mapreduce/JsonToKeyValueMapper.java
index 9702fd0..66b3142 100644
--- a/phoenix-core/src/main/java/org/apache/phoenix/mapreduce/JsonToKeyValueMapper.java
+++ b/phoenix-core/src/main/java/org/apache/phoenix/mapreduce/JsonToKeyValueMapper.java
@@ -32,7 +32,7 @@ import com.google.common.base.Preconditions;
/**
* MapReduce mapper that converts JSON input lines into KeyValues that can be written to HFiles.
- * <p/>
+ *
* KeyValues are produced by executing UPSERT statements on a Phoenix connection and then
* extracting the created KeyValues and rolling back the statement execution before it is
* committed to HBase.
diff --git a/phoenix-core/src/main/java/org/apache/phoenix/mapreduce/PhoenixInputFormat.java b/phoenix-core/src/main/java/org/apache/phoenix/mapreduce/PhoenixInputFormat.java
index 946ae4a..598a1ce 100644
--- a/phoenix-core/src/main/java/org/apache/phoenix/mapreduce/PhoenixInputFormat.java
+++ b/phoenix-core/src/main/java/org/apache/phoenix/mapreduce/PhoenixInputFormat.java
@@ -162,7 +162,6 @@ public class PhoenixInputFormat<T extends DBWritable> extends InputFormat<NullWr
* @param context
* @return
* @throws IOException
- * @throws SQLException
*/
protected QueryPlan getQueryPlan(final JobContext context, final Configuration configuration)
throws IOException {
diff --git a/phoenix-core/src/main/java/org/apache/phoenix/mapreduce/PhoenixInputSplit.java b/phoenix-core/src/main/java/org/apache/phoenix/mapreduce/PhoenixInputSplit.java
index 6d3c5e1..1c8e005 100644
--- a/phoenix-core/src/main/java/org/apache/phoenix/mapreduce/PhoenixInputSplit.java
+++ b/phoenix-core/src/main/java/org/apache/phoenix/mapreduce/PhoenixInputSplit.java
@@ -51,7 +51,7 @@ public class PhoenixInputSplit extends InputSplit implements Writable {
/**
*
- * @param keyRange
+ * @param scans
*/
public PhoenixInputSplit(final List<Scan> scans) {
this(scans, 0, null);
diff --git a/phoenix-core/src/main/java/org/apache/phoenix/mapreduce/RegexToKeyValueMapper.java b/phoenix-core/src/main/java/org/apache/phoenix/mapreduce/RegexToKeyValueMapper.java
index 04de360..708c8c0 100644
--- a/phoenix-core/src/main/java/org/apache/phoenix/mapreduce/RegexToKeyValueMapper.java
+++ b/phoenix-core/src/main/java/org/apache/phoenix/mapreduce/RegexToKeyValueMapper.java
@@ -39,7 +39,7 @@ import com.google.common.base.Preconditions;
/**
* MapReduce mapper that converts input lines into KeyValues based on the Regex that can be written to HFiles.
- * <p/>
+ *
* KeyValues are produced by executing UPSERT statements on a Phoenix connection and then
* extracting the created KeyValues and rolling back the statement execution before it is
* committed to HBase.
diff --git a/phoenix-core/src/main/java/org/apache/phoenix/monitoring/MutationMetricQueue.java b/phoenix-core/src/main/java/org/apache/phoenix/monitoring/MutationMetricQueue.java
index 8387b2c..9efb706 100644
--- a/phoenix-core/src/main/java/org/apache/phoenix/monitoring/MutationMetricQueue.java
+++ b/phoenix-core/src/main/java/org/apache/phoenix/monitoring/MutationMetricQueue.java
@@ -55,7 +55,7 @@ public class MutationMetricQueue {
/**
* Publish the metrics to wherever you want them published. The internal state is cleared out after every publish.
- * @return map of table name -> list of pair of (metric name, metric value)
+ * @return map of table {@code name -> list } of pair of (metric name, metric value)
*/
public Map<String, Map<MetricType, Long>> aggregate() {
Map<String, Map<MetricType, Long>> publishedMetrics = new HashMap<>();
diff --git a/phoenix-core/src/main/java/org/apache/phoenix/monitoring/ReadMetricQueue.java b/phoenix-core/src/main/java/org/apache/phoenix/monitoring/ReadMetricQueue.java
index 6186f6b..02698d8 100644
--- a/phoenix-core/src/main/java/org/apache/phoenix/monitoring/ReadMetricQueue.java
+++ b/phoenix-core/src/main/java/org/apache/phoenix/monitoring/ReadMetricQueue.java
@@ -70,7 +70,7 @@ public class ReadMetricQueue {
}
/**
- * @return map of table name -> list of pair of (metric name, metric value)
+ * @return map of table {@code name -> list } of pair of (metric name, metric value)
*/
public Map<String, Map<MetricType, Long>> aggregate() {
Map<String, Map<MetricType, Long>> publishedMetrics = new HashMap<>();
diff --git a/phoenix-core/src/main/java/org/apache/phoenix/parse/AndRewriterBooleanParseNodeVisitor.java b/phoenix-core/src/main/java/org/apache/phoenix/parse/AndRewriterBooleanParseNodeVisitor.java
index 905e18f..c0d6a4f 100644
--- a/phoenix-core/src/main/java/org/apache/phoenix/parse/AndRewriterBooleanParseNodeVisitor.java
+++ b/phoenix-core/src/main/java/org/apache/phoenix/parse/AndRewriterBooleanParseNodeVisitor.java
@@ -24,7 +24,7 @@ import com.google.common.collect.Lists;
/**
* Base visitor for rewrite {@link ParseNode},only further visit down for {@link AndParseNode}.
- * A example is {@link org.apache.phoenix.optimize.QueryOptimizer.WhereConditionRewriter}, which
+ * A example is org.apache.phoenix.optimize.QueryOptimizer.WhereConditionRewriter, which
* rewrites columns in dataTable to columns in indexTable, and removes parseNodes which have
* columns not in indexTable.
*/
diff --git a/phoenix-core/src/main/java/org/apache/phoenix/parse/ComparisonParseNode.java b/phoenix-core/src/main/java/org/apache/phoenix/parse/ComparisonParseNode.java
index 51cb833..f3d2e3f 100644
--- a/phoenix-core/src/main/java/org/apache/phoenix/parse/ComparisonParseNode.java
+++ b/phoenix-core/src/main/java/org/apache/phoenix/parse/ComparisonParseNode.java
@@ -27,7 +27,7 @@ import org.apache.phoenix.util.QueryUtil;
/**
*
- * Common base class for =, >, >=, <, <=, !=
+ * Common base class {@code =, >, >=, <, <=, != }
*
*
* @since 0.1
diff --git a/phoenix-core/src/main/java/org/apache/phoenix/parse/GreaterThanOrEqualParseNode.java b/phoenix-core/src/main/java/org/apache/phoenix/parse/GreaterThanOrEqualParseNode.java
index 6e0337a..b1cb994 100644
--- a/phoenix-core/src/main/java/org/apache/phoenix/parse/GreaterThanOrEqualParseNode.java
+++ b/phoenix-core/src/main/java/org/apache/phoenix/parse/GreaterThanOrEqualParseNode.java
@@ -23,7 +23,7 @@ import org.apache.hadoop.hbase.filter.CompareFilter.CompareOp;
/**
*
- * Node representing the greater than or equal to operator (>=) in SQL
+ * Node representing the greater than or equal to operator {@code (>=) } in SQL
*
*
* @since 0.1
diff --git a/phoenix-core/src/main/java/org/apache/phoenix/parse/GreaterThanParseNode.java b/phoenix-core/src/main/java/org/apache/phoenix/parse/GreaterThanParseNode.java
index d6b380c..ca4ae0a 100644
--- a/phoenix-core/src/main/java/org/apache/phoenix/parse/GreaterThanParseNode.java
+++ b/phoenix-core/src/main/java/org/apache/phoenix/parse/GreaterThanParseNode.java
@@ -23,7 +23,7 @@ import org.apache.hadoop.hbase.filter.CompareFilter.CompareOp;
/**
*
- * Node representing the greater than operator (>) in SQL
+ * Node representing the greater than operator {@code (>) } in SQL
*
*
* @since 0.1
diff --git a/phoenix-core/src/main/java/org/apache/phoenix/parse/HintNode.java b/phoenix-core/src/main/java/org/apache/phoenix/parse/HintNode.java
index f1dfdb6..c53cda0 100644
--- a/phoenix-core/src/main/java/org/apache/phoenix/parse/HintNode.java
+++ b/phoenix-core/src/main/java/org/apache/phoenix/parse/HintNode.java
@@ -58,7 +58,7 @@ public class HintNode {
*/
NO_INDEX,
/**
- * Hint of the form INDEX(<table_name> <index_name>...)
+ * Hint of the form {@code INDEX(<table_name> <index_name>...) }
* to suggest usage of the index if possible. The first
* usable index in the list of indexes will be choosen.
* Table and index names may be surrounded by double quotes
diff --git a/phoenix-core/src/main/java/org/apache/phoenix/parse/LessThanOrEqualParseNode.java b/phoenix-core/src/main/java/org/apache/phoenix/parse/LessThanOrEqualParseNode.java
index 345a8bd..18aa349 100644
--- a/phoenix-core/src/main/java/org/apache/phoenix/parse/LessThanOrEqualParseNode.java
+++ b/phoenix-core/src/main/java/org/apache/phoenix/parse/LessThanOrEqualParseNode.java
@@ -23,7 +23,7 @@ import org.apache.hadoop.hbase.filter.CompareFilter.CompareOp;
/**
*
- * Node representing the less than or equal to operator (<=) in SQL
+ * Node representing the less than or equal to operator {@code (<=) } in SQL
*
*
* @since 0.1
diff --git a/phoenix-core/src/main/java/org/apache/phoenix/parse/LessThanParseNode.java b/phoenix-core/src/main/java/org/apache/phoenix/parse/LessThanParseNode.java
index 9472d97..e6e86a2 100644
--- a/phoenix-core/src/main/java/org/apache/phoenix/parse/LessThanParseNode.java
+++ b/phoenix-core/src/main/java/org/apache/phoenix/parse/LessThanParseNode.java
@@ -23,7 +23,7 @@ import org.apache.hadoop.hbase.filter.CompareFilter.CompareOp;
/**
*
- * Node representing the less than operator (<) in SQL
+ * Node representing the less than operator {@code (<) } in SQL
*
*
* @since 0.1
diff --git a/phoenix-core/src/main/java/org/apache/phoenix/parse/NotEqualParseNode.java b/phoenix-core/src/main/java/org/apache/phoenix/parse/NotEqualParseNode.java
index 8d95b51..9114aca 100644
--- a/phoenix-core/src/main/java/org/apache/phoenix/parse/NotEqualParseNode.java
+++ b/phoenix-core/src/main/java/org/apache/phoenix/parse/NotEqualParseNode.java
@@ -23,7 +23,7 @@ import org.apache.hadoop.hbase.filter.CompareFilter.CompareOp;
/**
*
- * Node representing a not equal expression (!=,<>) in SQL
+ * Node representing a not equal expression {@code (!=,<>) } in SQL
*
*
* @since 0.1
diff --git a/phoenix-core/src/main/java/org/apache/phoenix/parse/ParseNodeFactory.java b/phoenix-core/src/main/java/org/apache/phoenix/parse/ParseNodeFactory.java
index 6577aac..f117318 100644
--- a/phoenix-core/src/main/java/org/apache/phoenix/parse/ParseNodeFactory.java
+++ b/phoenix-core/src/main/java/org/apache/phoenix/parse/ParseNodeFactory.java
@@ -84,7 +84,7 @@ public class ParseNodeFactory {
*
* Key used to look up a built-in function using the combination of
* the lowercase name and the number of arguments. This disambiguates
- * the aggregate MAX(<col>) from the non aggregate MAX(<col1>,<col2>).
+ * the aggregate {@code MAX(<col>) } from the non aggregate {@code MAX(<col1>,<col2>) }.
*
*
* @since 0.1
diff --git a/phoenix-core/src/main/java/org/apache/phoenix/parse/ParseNodeRewriter.java b/phoenix-core/src/main/java/org/apache/phoenix/parse/ParseNodeRewriter.java
index 01533ef..266ebad 100644
--- a/phoenix-core/src/main/java/org/apache/phoenix/parse/ParseNodeRewriter.java
+++ b/phoenix-core/src/main/java/org/apache/phoenix/parse/ParseNodeRewriter.java
@@ -58,9 +58,9 @@ public class ParseNodeRewriter extends TraverseAllParseNodeVisitor<ParseNode> {
* <pre>
* Resolve the inner alias for the selectStament.
* For following sql:
- * select aid,sum(age) agesum from merge where age >=11 and age<=33 group by aid order by agesum
+ * {@code select aid,sum(age) agesum from merge where age >=11 and age <= 33 group by aid order by agesum }
* "agesum" is an alias of "sum(age)", so for this method, the above sql is rewritten to:
- * select aid,sum(age) agesum from merge where age >=11 and age<=33 group by aid order by sum(age)
+ * {@code select aid,sum(age) agesum from merge where age >= 11 and <= 33 group by aid order by sum(age) }
* </pre>
* @param selectStament
* @param phoenixConnection
diff --git a/phoenix-core/src/main/java/org/apache/phoenix/query/ConnectionQueryServices.java b/phoenix-core/src/main/java/org/apache/phoenix/query/ConnectionQueryServices.java
index 103ce73..1714cfe 100644
--- a/phoenix-core/src/main/java/org/apache/phoenix/query/ConnectionQueryServices.java
+++ b/phoenix-core/src/main/java/org/apache/phoenix/query/ConnectionQueryServices.java
@@ -165,7 +165,7 @@ public interface ConnectionQueryServices extends QueryServices, MetaDataMutated
/**
* Removes cache {@link GuidePostsInfo} for the table with the given name. If no cached guideposts are present, this does nothing.
*
- * @param tableName The table to remove stats for
+ * @param key
*/
void invalidateStats(GuidePostsKey key);
diff --git a/phoenix-core/src/main/java/org/apache/phoenix/query/ConnectionQueryServicesImpl.java b/phoenix-core/src/main/java/org/apache/phoenix/query/ConnectionQueryServicesImpl.java
index 7fdffdd..ea8aca7 100644
--- a/phoenix-core/src/main/java/org/apache/phoenix/query/ConnectionQueryServicesImpl.java
+++ b/phoenix-core/src/main/java/org/apache/phoenix/query/ConnectionQueryServicesImpl.java
@@ -404,7 +404,6 @@ public class ConnectionQueryServicesImpl extends DelegateQueryServices implement
* @param services base services from where we derive our default configuration
* @param connectionInfo to provide connection information
* @param info hbase configuration properties
- * @throws SQLException
*/
public ConnectionQueryServicesImpl(QueryServices services, ConnectionInfo connectionInfo, Properties info) {
super(services);
diff --git a/phoenix-core/src/main/java/org/apache/phoenix/query/ConnectionlessQueryServicesImpl.java b/phoenix-core/src/main/java/org/apache/phoenix/query/ConnectionlessQueryServicesImpl.java
index 9679734..d94622f 100644
--- a/phoenix-core/src/main/java/org/apache/phoenix/query/ConnectionlessQueryServicesImpl.java
+++ b/phoenix-core/src/main/java/org/apache/phoenix/query/ConnectionlessQueryServicesImpl.java
@@ -737,8 +737,8 @@ public class ConnectionlessQueryServicesImpl extends DelegateQueryServices imple
* Manually adds {@link GuidePostsInfo} for a table to the client-side cache. Not a
* {@link ConnectionQueryServices} method. Exposed for testing purposes.
*
- * @param tableName Table name
- * @param stats Stats instance
+ * @param key
+ * @param info
*/
public void addTableStats(GuidePostsKey key, GuidePostsInfo info) {
this.guidePostsCache.put(Objects.requireNonNull(key), info);
diff --git a/phoenix-core/src/main/java/org/apache/phoenix/query/GuidePostsCacheImpl.java b/phoenix-core/src/main/java/org/apache/phoenix/query/GuidePostsCacheImpl.java
index 24b695c..e026591 100644
--- a/phoenix-core/src/main/java/org/apache/phoenix/query/GuidePostsCacheImpl.java
+++ b/phoenix-core/src/main/java/org/apache/phoenix/query/GuidePostsCacheImpl.java
@@ -82,7 +82,7 @@ public class GuidePostsCacheImpl implements GuidePostsCache {
}
/**
- * Returns the PTableStats for the given <code>tableName</code, using the provided
+ * Returns the PTableStats for the given <code>tableName</code>, using the provided
* <code>valueLoader</code> if no such mapping exists.
*
* @see com.google.common.cache.LoadingCache#get(Object)
diff --git a/phoenix-core/src/main/java/org/apache/phoenix/query/HBaseFactoryProvider.java b/phoenix-core/src/main/java/org/apache/phoenix/query/HBaseFactoryProvider.java
index c04db02..780dbfd 100644
--- a/phoenix-core/src/main/java/org/apache/phoenix/query/HBaseFactoryProvider.java
+++ b/phoenix-core/src/main/java/org/apache/phoenix/query/HBaseFactoryProvider.java
@@ -21,7 +21,7 @@ import org.apache.phoenix.util.InstanceResolver;
/**
* Manages factories that provide extension points for HBase.
- * <p/>
+ *
* Dependent modules may register their own implementations of the following using {@link java.util.ServiceLoader}:
* <ul>
* <li>{@link ConfigurationFactory}</li>
diff --git a/phoenix-core/src/main/java/org/apache/phoenix/query/HConnectionFactory.java b/phoenix-core/src/main/java/org/apache/phoenix/query/HConnectionFactory.java
index 72f4182..faae3ad 100644
--- a/phoenix-core/src/main/java/org/apache/phoenix/query/HConnectionFactory.java
+++ b/phoenix-core/src/main/java/org/apache/phoenix/query/HConnectionFactory.java
@@ -24,7 +24,7 @@ import org.apache.hadoop.hbase.client.HConnection;
import org.apache.hadoop.hbase.client.HConnectionManager;
/**
- * Factory for creating {@link HConnection}
+ * Factory for creating HConnection
*
*
*/
@@ -33,7 +33,7 @@ public interface HConnectionFactory {
/**
* Creates HConnection to access HBase clusters.
*
- * @param configuration object
+ * @param conf object
* @return A HConnection instance
*/
HConnection createConnection(Configuration conf) throws IOException;
diff --git a/phoenix-core/src/main/java/org/apache/phoenix/query/PropertyPolicy.java b/phoenix-core/src/main/java/org/apache/phoenix/query/PropertyPolicy.java
index a59f912..239b89f 100644
--- a/phoenix-core/src/main/java/org/apache/phoenix/query/PropertyPolicy.java
+++ b/phoenix-core/src/main/java/org/apache/phoenix/query/PropertyPolicy.java
@@ -27,12 +27,13 @@ import java.util.Properties;
* with exception
*
* An example will be:
- *<code>
+ *<pre>
+ *{@code
*public class Customized PropertyPolicy implements PropertyPolicy {
* final static Set<String> propertiesKeyAllowed = Collections.unmodifiableSet(
* new HashSet<>(asList("DriverMajorVersion","DriverMinorVersion","DriverName","CurrentSCN")));
*
- * @Override public void evaluate(Properties properties) {
+ * public void evaluate(Properties properties) {
* final Set<String> offendingProperties = new HashSet<>();
*
* for(Object k:properties.keySet()){
@@ -44,7 +45,8 @@ import java.util.Properties;
* "properties not allowed. offending properties" + offendingProperties);
* }
*}
- *</code>
+ *}
+ *</pre>
*/
public interface PropertyPolicy {
/**
diff --git a/phoenix-core/src/main/java/org/apache/phoenix/query/PropertyPolicyProvider.java b/phoenix-core/src/main/java/org/apache/phoenix/query/PropertyPolicyProvider.java
index 27cb2d3..4ba4c56 100644
--- a/phoenix-core/src/main/java/org/apache/phoenix/query/PropertyPolicyProvider.java
+++ b/phoenix-core/src/main/java/org/apache/phoenix/query/PropertyPolicyProvider.java
@@ -21,7 +21,7 @@ import org.apache.phoenix.util.InstanceResolver;
/**
* Manages providers that provide property policy for Phoenix.
- * <p/>
+ *
* Dependent modules may register their own implementations of the following using {@link java.util.ServiceLoader}:
* <ul>
* <li>{@link PropertyPolicy}</li>
diff --git a/phoenix-core/src/main/java/org/apache/phoenix/query/QueryServices.java b/phoenix-core/src/main/java/org/apache/phoenix/query/QueryServices.java
index 2bf2350..cc83d16 100644
--- a/phoenix-core/src/main/java/org/apache/phoenix/query/QueryServices.java
+++ b/phoenix-core/src/main/java/org/apache/phoenix/query/QueryServices.java
@@ -71,7 +71,7 @@ public interface QueryServices extends SQLCloseable {
/**
* max size to spool the the result into
* ${java.io.tmpdir}/ResultSpoolerXXX.bin if
- * {@link QueryServices#SPOOL_THRESHOLD_BYTES_ATTRIB } is reached.
+ * QueryServices#SPOOL_THRESHOLD_BYTES_ATTRIB is reached.
* <p>
* default is unlimited(-1)
* <p>
diff --git a/phoenix-core/src/main/java/org/apache/phoenix/schema/ColumnModifier.java b/phoenix-core/src/main/java/org/apache/phoenix/schema/ColumnModifier.java
index 2650cc6..7226798 100644
--- a/phoenix-core/src/main/java/org/apache/phoenix/schema/ColumnModifier.java
+++ b/phoenix-core/src/main/java/org/apache/phoenix/schema/ColumnModifier.java
@@ -22,7 +22,7 @@ import org.apache.hadoop.hbase.filter.CompareFilter.CompareOp;
import com.google.common.base.Preconditions;
/**
- * A ColumnModifier implementation modifies how bytes are stored in a primary key column.</p>
+ * A ColumnModifier implementation modifies how bytes are stored in a primary key column.
* The {@link ColumnModifier#apply apply} method is called when the bytes for a specific column are first written to HBase and again
* when they are read back. Phoenix attemps to minimize calls to apply when bytes are read out of HBase.
*
diff --git a/phoenix-core/src/main/java/org/apache/phoenix/schema/SortOrder.java b/phoenix-core/src/main/java/org/apache/phoenix/schema/SortOrder.java
index 3bfc7ed..ab188c6 100644
--- a/phoenix-core/src/main/java/org/apache/phoenix/schema/SortOrder.java
+++ b/phoenix-core/src/main/java/org/apache/phoenix/schema/SortOrder.java
@@ -22,27 +22,27 @@ import org.apache.hadoop.hbase.filter.CompareFilter.CompareOp;
import com.google.common.base.Preconditions;
/**
- * Specifies the sort order on disk of row key columns. The default is ASC.</p>
+ * Specifies the sort order on disk of row key columns. The default is ASC.
*
* HBase always stores row keys in ascending order resulting in scans to also be
* sorted by ascending row keys. This enum is used to associate a sort order
* with each row key column to allow storing row key columns in descending
- * order.</p>
+ * order.
*
* The often cited example of when you may want to do this is a row key that has
* a date component. If all other parts of the row key are equal, a scan would
* return the data from least recent to most recent; to get the scan to return
* the most recent data first, the time component of the row key can be marked
- * as "desc".</p>
+ * as "desc".
*
* Internally, the bits of values for columns marked as "desc" are inverted before handing
* them to HBase to persist; they are inverted again when read back out.
- * </p>
*
- * Example DDL:</p>
+ *
+ * Example DDL:
*
* CREATE TABLE Events(event_type INTEGER NOT NULL, event_date DATE NOT NULL, event_name VARCHAR NOT NULL
- * CONSTRAINT PK PRIMARY KEY (event_type, event_date DESC))</p>
+ * CONSTRAINT PK PRIMARY KEY (event_type, event_date DESC))
*
* @since 1.2
*/
diff --git a/phoenix-core/src/main/java/org/apache/phoenix/schema/stats/DefaultStatisticsCollector.java b/phoenix-core/src/main/java/org/apache/phoenix/schema/stats/DefaultStatisticsCollector.java
index bafe89d..b20f964 100644
--- a/phoenix-core/src/main/java/org/apache/phoenix/schema/stats/DefaultStatisticsCollector.java
+++ b/phoenix-core/src/main/java/org/apache/phoenix/schema/stats/DefaultStatisticsCollector.java
@@ -306,7 +306,6 @@ public class DefaultStatisticsCollector implements StatisticsCollector {
*
* @param results
* next batch of {@link KeyValue}s
- * @throws IOException
*/
@Override
public void collectStatistics(final List<Cell> results) {
diff --git a/phoenix-core/src/main/java/org/apache/phoenix/schema/stats/GuidePostsInfo.java b/phoenix-core/src/main/java/org/apache/phoenix/schema/stats/GuidePostsInfo.java
index 9492a35..8f31f39 100644
--- a/phoenix-core/src/main/java/org/apache/phoenix/schema/stats/GuidePostsInfo.java
+++ b/phoenix-core/src/main/java/org/apache/phoenix/schema/stats/GuidePostsInfo.java
@@ -85,7 +85,7 @@ public class GuidePostsInfo {
* Maximum length of a guidePost collected
* @param guidePostsCount
* Number of guidePosts
- * @param gpTimestamps
+ * @param updateTimes
* Times at which guidePosts were updated/created
*/
public GuidePostsInfo(List<Long> byteCounts, ImmutableBytesWritable guidePosts, List<Long> rowCounts, int maxLength,
diff --git a/phoenix-core/src/main/java/org/apache/phoenix/schema/stats/GuidePostsInfoBuilder.java b/phoenix-core/src/main/java/org/apache/phoenix/schema/stats/GuidePostsInfoBuilder.java
index 2b82936..8e0bd29 100644
--- a/phoenix-core/src/main/java/org/apache/phoenix/schema/stats/GuidePostsInfoBuilder.java
+++ b/phoenix-core/src/main/java/org/apache/phoenix/schema/stats/GuidePostsInfoBuilder.java
@@ -93,7 +93,6 @@ public class GuidePostsInfoBuilder {
* @param row number of rows in the guidepost
* @param byteCount number of bytes in the guidepost
* @param updateTimestamp time at which guidepost was created/updated.
- * @throws IOException
*/
public boolean trackGuidePost(ImmutableBytesWritable row, long byteCount, long rowCount,
long updateTimestamp) {
diff --git a/phoenix-core/src/main/java/org/apache/phoenix/schema/stats/StatisticsCollector.java b/phoenix-core/src/main/java/org/apache/phoenix/schema/stats/StatisticsCollector.java
index 0fd6390..bed0ee9 100644
--- a/phoenix-core/src/main/java/org/apache/phoenix/schema/stats/StatisticsCollector.java
+++ b/phoenix-core/src/main/java/org/apache/phoenix/schema/stats/StatisticsCollector.java
@@ -50,7 +50,6 @@ public interface StatisticsCollector extends Closeable {
/**
* Collect statistics for the given list of cells. This method can be called multiple times
* during collection of statistics.
- * @throws IOException
*/
void collectStatistics(List<Cell> results);
diff --git a/phoenix-core/src/main/java/org/apache/phoenix/schema/stats/StatisticsWriter.java b/phoenix-core/src/main/java/org/apache/phoenix/schema/stats/StatisticsWriter.java
index 36b33d6..0562623 100644
--- a/phoenix-core/src/main/java/org/apache/phoenix/schema/stats/StatisticsWriter.java
+++ b/phoenix-core/src/main/java/org/apache/phoenix/schema/stats/StatisticsWriter.java
@@ -149,7 +149,7 @@ public class StatisticsWriter implements Closeable {
}
/**
- * Update a list of statistics for a given region. If the UPDATE STATISTICS <tablename> query is issued then we use
+ * Update a list of statistics for a given region. If the UPDATE STATISTICS {@code <tablename> } query is issued then we use
* Upsert queries to update the table If the region gets splitted or the major compaction happens we update using
* HTable.put()
*
diff --git a/phoenix-core/src/main/java/org/apache/phoenix/schema/types/PArrayDataType.java b/phoenix-core/src/main/java/org/apache/phoenix/schema/types/PArrayDataType.java
index 162b235..faddb68 100644
--- a/phoenix-core/src/main/java/org/apache/phoenix/schema/types/PArrayDataType.java
+++ b/phoenix-core/src/main/java/org/apache/phoenix/schema/types/PArrayDataType.java
@@ -44,8 +44,8 @@ import com.google.common.base.Preconditions;
* The datatype for PColummns that are Arrays. Any variable length array would follow the below order. Every element
* would be seperated by a seperator byte '0'. Null elements are counted and once a first non null element appears we
* write the count of the nulls prefixed with a seperator byte. Trailing nulls are not taken into account. The last non
- * null element is followed by two seperator bytes. For eg a, b, null, null, c, null -> 65 0 66 0 0 2 67 0 0 0 a null
- * null null b c null d -> 65 0 0 3 66 0 67 0 0 1 68 0 0 0. The reason we use this serialization format is to allow the
+ * null element is followed by two seperator bytes. For eg {@code a, b, null, null, c, null -> 65 0 66 0 0 2 67 0 0 0 a null
+ * null null b c null d -> 65 0 0 3 66 0 67 0 0 1 68 0 0 0 }. The reason we use this serialization format is to allow the
* byte array of arrays of the same type to be directly comparable against each other. This prevents a costly
* deserialization on compare and allows an array column to be used as the last column in a primary key constraint.
*/
diff --git a/phoenix-core/src/main/java/org/apache/phoenix/schema/types/PDataType.java b/phoenix-core/src/main/java/org/apache/phoenix/schema/types/PDataType.java
index eba6079..415eeef 100644
--- a/phoenix-core/src/main/java/org/apache/phoenix/schema/types/PDataType.java
+++ b/phoenix-core/src/main/java/org/apache/phoenix/schema/types/PDataType.java
@@ -613,7 +613,7 @@ public abstract class PDataType<T> implements DataType<T>, Comparable<PDataType<
/**
* Deserialize a variable length byte array into a BigDecimal. Note that because of the normalization that gets done
* to the scale, if you roundtrip a BigDecimal, it may not be equal before and after. However, the before and after
- * number will always compare to be equal (i.e. <nBefore>.compareTo(<nAfter>) == 0)
+ * number will always compare to be equal {@code (i.e. <nBefore>.compareTo(<nAfter>) == 0) }
*
* @param bytes
* the bytes containing the number
diff --git a/phoenix-core/src/main/java/org/apache/phoenix/transaction/PhoenixTransactionContext.java b/phoenix-core/src/main/java/org/apache/phoenix/transaction/PhoenixTransactionContext.java
index 1882004..fd76cad 100644
--- a/phoenix-core/src/main/java/org/apache/phoenix/transaction/PhoenixTransactionContext.java
+++ b/phoenix-core/src/main/java/org/apache/phoenix/transaction/PhoenixTransactionContext.java
@@ -165,8 +165,6 @@ public interface PhoenixTransactionContext {
*
* @param dataTable the table that the DDL command works on
* @throws SQLException
- * @throws InterruptedException
- * @throws TimeoutException
*/
public void commitDDLFence(PTable dataTable)
throws SQLException;
diff --git a/phoenix-core/src/main/java/org/apache/phoenix/transaction/PhoenixTransactionProvider.java b/phoenix-core/src/main/java/org/apache/phoenix/transaction/PhoenixTransactionProvider.java
index 3af554b..f50b6b5 100644
--- a/phoenix-core/src/main/java/org/apache/phoenix/transaction/PhoenixTransactionProvider.java
+++ b/phoenix-core/src/main/java/org/apache/phoenix/transaction/PhoenixTransactionProvider.java
@@ -63,7 +63,6 @@ public interface PhoenixTransactionProvider {
* @param timestamp - start timestamp
* @param commitTimestamp - commit timestamp
* @return put operation with metadata
- * @throws IOException
*/
public Put markPutAsCommitted(Put put, long timestamp, long commitTimestamp);
}
diff --git a/phoenix-core/src/main/java/org/apache/phoenix/util/DefaultEnvironmentEdge.java b/phoenix-core/src/main/java/org/apache/phoenix/util/DefaultEnvironmentEdge.java
index 8dc63b1..6a1e075 100644
--- a/phoenix-core/src/main/java/org/apache/phoenix/util/DefaultEnvironmentEdge.java
+++ b/phoenix-core/src/main/java/org/apache/phoenix/util/DefaultEnvironmentEdge.java
@@ -24,7 +24,6 @@ package org.apache.phoenix.util;
public class DefaultEnvironmentEdge extends EnvironmentEdge {
/**
* {@inheritDoc}
- * <p/>
* This implementation returns {@link System#currentTimeMillis()}
*/
@Override
diff --git a/phoenix-core/src/main/java/org/apache/phoenix/util/ExpressionUtil.java b/phoenix-core/src/main/java/org/apache/phoenix/util/ExpressionUtil.java
index c6854ab..68a77c0 100644
--- a/phoenix-core/src/main/java/org/apache/phoenix/util/ExpressionUtil.java
+++ b/phoenix-core/src/main/java/org/apache/phoenix/util/ExpressionUtil.java
@@ -104,7 +104,7 @@ public class ExpressionUtil {
/**
* check the whereExpression to see if the columnExpression is constant.
- * eg. for "where a =3 and b > 9", a is constant,but b is not.
+ * eg. for {@code "where a = 3 and b > 9" }, a is constant,but b is not.
* @param columnExpression
* @param whereExpression
* @return
diff --git a/phoenix-core/src/main/java/org/apache/phoenix/util/KeyValueUtil.java b/phoenix-core/src/main/java/org/apache/phoenix/util/KeyValueUtil.java
index 4d8565f..f133ad1 100644
--- a/phoenix-core/src/main/java/org/apache/phoenix/util/KeyValueUtil.java
+++ b/phoenix-core/src/main/java/org/apache/phoenix/util/KeyValueUtil.java
@@ -197,7 +197,7 @@ public class KeyValueUtil {
/**
* Estimates the size of rows stored in RowMutationState (in memory)
- * @param mutations map from table to row to RowMutationState
+ * @param tableMutationMap map from table to row to RowMutationState
* @return estimated row size
*/
public static long
diff --git a/phoenix-core/src/main/java/org/apache/phoenix/util/MetaDataUtil.java b/phoenix-core/src/main/java/org/apache/phoenix/util/MetaDataUtil.java
index 637d2f5..c8c3f8a 100644
--- a/phoenix-core/src/main/java/org/apache/phoenix/util/MetaDataUtil.java
+++ b/phoenix-core/src/main/java/org/apache/phoenix/util/MetaDataUtil.java
@@ -232,8 +232,8 @@ public class MetaDataUtil {
* Encode HBase and Phoenix version along with some server-side config information such as whether WAL codec is
* installed (necessary for non transactional, mutable secondar indexing), and whether systemNamespace mapping is enabled.
*
- * @param env
- * RegionCoprocessorEnvironment to access HBase version and Configuration.
+ * @param hbaseVersionStr
+ * @param config
* @return long value sent back during initialization of a cluster connection.
*/
public static long encodeVersion(String hbaseVersionStr, Configuration config) {
@@ -799,10 +799,9 @@ public class MetaDataUtil {
/**
* This function checks if all regions of a table is online
+ * @param conf
* @param table
* @return true when all regions of a table are online
- * @throws IOException
- * @throws
*/
public static boolean tableRegionsOnline(Configuration conf, PTable table) {
try (HConnection hcon =
diff --git a/phoenix-core/src/main/java/org/apache/phoenix/util/PhoenixContextExecutor.java b/phoenix-core/src/main/java/org/apache/phoenix/util/PhoenixContextExecutor.java
index cd12261..34c9828 100644
--- a/phoenix-core/src/main/java/org/apache/phoenix/util/PhoenixContextExecutor.java
+++ b/phoenix-core/src/main/java/org/apache/phoenix/util/PhoenixContextExecutor.java
@@ -24,7 +24,7 @@ import org.apache.phoenix.call.CallWrapper;
/**
* Executes {@code Callable}s using a context classloader that is set up to load classes from
* Phoenix.
- * <p/>
+ *
* Loading HBase configuration settings and endpoint coprocessor classes is done via the context
* classloader of the calling thread. When Phoenix is being run via a JDBC-enabled GUI, the
* driver is often loaded dynamically and executed via multiple threads, which makes it difficult
@@ -62,7 +62,7 @@ public class PhoenixContextExecutor {
* instead of the currently-set context classloader of the current thread. This allows loading
* dynamically-loaded classes and configuration files using the same classloader used to
* load the rest of the JDBC driver.
- * <p/>
+ *
* The context classloader of the current thread is reset to its original value after the
* callable has been executed.
*
@@ -84,7 +84,6 @@ public class PhoenixContextExecutor {
*
* @param target the callable to be executed
* @return the return value from the callable
- * @throws Exception any exception thrown by the underlying callable
*/
public static <T> T callWithoutPropagation(Callable<T> target) {
try {
diff --git a/phoenix-core/src/main/java/org/apache/phoenix/util/PhoenixRuntime.java b/phoenix-core/src/main/java/org/apache/phoenix/util/PhoenixRuntime.java
index 73c4a20..ac4447f 100644
--- a/phoenix-core/src/main/java/org/apache/phoenix/util/PhoenixRuntime.java
+++ b/phoenix-core/src/main/java/org/apache/phoenix/util/PhoenixRuntime.java
@@ -1093,7 +1093,7 @@ public class PhoenixRuntime {
* of their values in the object array.
* @return values encoded in a byte array
* @throws SQLException
- * @see {@link #decodeValues(Connection, String, byte[], List)}
+ * @see decodeValues(Connection, String, byte[], List)
*/
@Deprecated
public static byte[] encodeValues(Connection conn, String fullTableName, Object[] values, List<Pair<String, String>> columns) throws SQLException {
@@ -1159,7 +1159,7 @@ public class PhoenixRuntime {
* of their values in the object array.
* @return values encoded in a byte array
* @throws SQLException
- * @see {@link #decodeValues(Connection, String, byte[], List)}
+ * @see decodeValues(Connection, String, byte[], List)
*/
public static byte[] encodeColumnValues(Connection conn, String fullTableName, Object[] values, List<Pair<String, String>> columns) throws SQLException {
PTable table = getTable(conn, fullTableName);
@@ -1410,11 +1410,12 @@ public class PhoenixRuntime {
* requestReadMetrics = PhoenixRuntime.getRequestReadMetrics(rs);
* PhoenixRuntime.resetMetrics(rs);
* }
+ * }
* </pre>
*
* @param rs
* result set to get the metrics for
- * @return a map of (table name) -> (map of (metric name) -> (metric value))
+ * @return a map of {@code (table name) -> (map of (metric name) -> (metric value)) }
* @throws SQLException
*/
public static Map<String, Map<MetricType, Long>> getRequestReadMetricInfo(ResultSet rs) throws SQLException {
@@ -1444,11 +1445,12 @@ public class PhoenixRuntime {
* requestReadMetrics = PhoenixRuntime.getRequestReadMetrics(rs);
* PhoenixRuntime.resetMetrics(rs);
* }
+ * }
* </pre>
*
* @param rs
* result set to get the metrics for
- * @return a map of metric name -> metric value
+ * @return a map of {@code metric name -> metric value }
* @throws SQLException
*/
public static Map<MetricType, Long> getOverAllReadRequestMetricInfo(ResultSet rs) throws SQLException {
@@ -1483,11 +1485,12 @@ public class PhoenixRuntime {
* mutationReadMetrics = PhoenixRuntime.getReadMetricsForMutationsSinceLastReset(conn);
* PhoenixRuntime.resetMetrics(rs);
* }
+ * }
* </pre>
*
* @param conn
* connection to get the metrics for
- * @return a map of (table name) -> (map of (metric name) -> (metric value))
+ * @return a map of {@code (table name) -> (map of (metric name) -> (metric value)) }
* @throws SQLException
*/
public static Map<String, Map<MetricType, Long>> getWriteMetricInfoForMutationsSinceLastReset(Connection conn) throws SQLException {
@@ -1522,10 +1525,11 @@ public class PhoenixRuntime {
* mutationReadMetrics = PhoenixRuntime.getReadMetricsForMutationsSinceLastReset(conn);
* PhoenixRuntime.resetMetrics(rs);
* }
+ * }
* </pre>
* @param conn
* connection to get the metrics for
- * @return a map of (table name) -> (map of (metric name) -> (metric value))
+ * @return a map of {@code (table name) -> (map of (metric name) -> (metric value)) }
* @throws SQLException
*/
public static Map<String, Map<MetricType, Long>> getReadMetricInfoForMutationsSinceLastReset(Connection conn) throws SQLException {
@@ -1542,7 +1546,7 @@ public class PhoenixRuntime {
/**
* Reset the read metrics collected in the result set.
*
- * @see {@link #getRequestReadMetrics(ResultSet)} {@link #getOverAllReadRequestMetrics(ResultSet)}
+ * @see #getRequestReadMetrics(ResultSet) #getOverAllReadRequestMetrics(ResultSet)
* @param rs
* @throws SQLException
*/
@@ -1554,7 +1558,7 @@ public class PhoenixRuntime {
/**
* Reset the mutation and reads-for-mutations metrics collected in the connection.
*
- * @see {@link #getReadMetricsForMutationsSinceLastReset(Connection)} {@link #getWriteMetricsForMutationsSinceLastReset(Connection)}
+ * @see #getReadMetricsForMutationsSinceLastReset(Connection) #getWriteMetricsForMutationsSinceLastReset(Connection)
* @param conn
* @throws SQLException
*/
diff --git a/phoenix-core/src/main/java/org/apache/phoenix/util/ReadOnlyProps.java b/phoenix-core/src/main/java/org/apache/phoenix/util/ReadOnlyProps.java
index fe5d045..9c9ed06b 100644
--- a/phoenix-core/src/main/java/org/apache/phoenix/util/ReadOnlyProps.java
+++ b/phoenix-core/src/main/java/org/apache/phoenix/util/ReadOnlyProps.java
@@ -285,9 +285,9 @@ public class ReadOnlyProps implements Iterable<Entry<String, String>> {
}
/**
- * Get the properties as a <code>Map<String,String></code>
+ * Get the properties as a {@code Map<String,String>}
*
- * @return Map<String,String>
+ * @return {@code Map<String,String>}
*/
public Map<String,String> asMap() {
return props;
diff --git a/phoenix-core/src/main/java/org/apache/phoenix/util/StringUtil.java b/phoenix-core/src/main/java/org/apache/phoenix/util/StringUtil.java
index 4d3a36f..7639f6e 100644
--- a/phoenix-core/src/main/java/org/apache/phoenix/util/StringUtil.java
+++ b/phoenix-core/src/main/java/org/apache/phoenix/util/StringUtil.java
@@ -341,7 +341,7 @@ public class StringUtil {
* the values to be stored in all elements of the array
* @param fillFromIdx
* the index of the first element (inclusive) to be used as fill values
- * @param filToIdx
+ * @param fillToIdx
* the index of the last element (exclusive) to be used as fill value
* @param invertFill
* if true inverts the bits in fill before filling the array
diff --git a/phoenix-core/src/main/java/org/apache/phoenix/util/ViewUtil.java b/phoenix-core/src/main/java/org/apache/phoenix/util/ViewUtil.java
index 6464fb6..c04ddc3 100644
--- a/phoenix-core/src/main/java/org/apache/phoenix/util/ViewUtil.java
+++ b/phoenix-core/src/main/java/org/apache/phoenix/util/ViewUtil.java
@@ -92,7 +92,7 @@ public class ViewUtil {
/**
* Find all the descendant views of a given table or view in a depth-first fashion.
- * Note that apart from scanning the parent->child links, we also validate each view
+ * Note that apart from scanning the {@code parent->child } links, we also validate each view
* by trying to resolve it.
* Use {@link ViewUtil#findAllRelatives(Table, byte[], byte[], byte[], LinkType,
* TableViewFinderResult)} if you want to find other links and don't care about orphan results.
@@ -295,7 +295,7 @@ public class ViewUtil {
/**
* Check metadata to find if a given table/view has any immediate child views. Note that this
- * is not resilient to orphan parent->child links.
+ * is not resilient to orphan {@code parent->child } links.
* @param sysCatOrsysChildLink For older (pre-4.15.0) clients, we look for child links inside
* SYSTEM.CATALOG, otherwise we look for them inside
* SYSTEM.CHILD_LINK
@@ -332,14 +332,15 @@ public class ViewUtil {
}
/**
- * Attempt to drop an orphan child view i.e. a child view for which we see a parent->child entry
+ * Attempt to drop an orphan child view i.e. a child view for which we see a
+ * {@code parent->child } entry
* in SYSTEM.CHILD_LINK/SYSTEM.CATALOG (as a child) but for whom the parent no longer exists.
* @param env Region Coprocessor environment
* @param tenantIdBytes tenantId of the parent
* @param schemaName schema of the parent
* @param tableOrViewName parent table/view name
* @param sysCatOrSysChildLink SYSTEM.CATALOG or SYSTEM.CHILD_LINK which is used to find the
- * parent->child linking rows
+ * {@code parent->child } linking rows
* @throws IOException thrown if there is an error scanning SYSTEM.CHILD_LINK or SYSTEM.CATALOG
* @throws SQLException thrown if there is an error getting a connection to the server or
* an error retrieving the PTable for a child view
@@ -431,8 +432,8 @@ public class ViewUtil {
/**
- * Determines whether we should use SYSTEM.CATALOG or SYSTEM.CHILD_LINK to find parent->child
- * links i.e. {@link LinkType#CHILD_TABLE}.
+ * Determines whether we should use SYSTEM.CATALOG or SYSTEM.CHILD_LINK to find
+ * {@code parent->child } links i.e. {@link LinkType#CHILD_TABLE}.
* If the client is older than 4.15.0 and the SYSTEM.CHILD_LINK table does not exist, we use
* the SYSTEM.CATALOG table. In all other cases, we use the SYSTEM.CHILD_LINK table.
* This is required for backwards compatibility.
diff --git a/phoenix-pherf/src/main/java/org/apache/phoenix/pherf/jmx/MonitorManager.java b/phoenix-pherf/src/main/java/org/apache/phoenix/pherf/jmx/MonitorManager.java
index 5c434d8..5800a20 100644
--- a/phoenix-pherf/src/main/java/org/apache/phoenix/pherf/jmx/MonitorManager.java
+++ b/phoenix-pherf/src/main/java/org/apache/phoenix/pherf/jmx/MonitorManager.java
@@ -172,8 +172,8 @@ public class MonitorManager implements Workload {
/**
* This method should really only be used for testing
*
- * @return List < {@link org.apache.phoenix.pherf.result.Result} >
- * @throws IOException
+ * @return {@code List < org.apache.phoenix.pherf.result.Result > }
+ * @throws Exception
*/
public synchronized List<Result> readResults() throws Exception {
ResultHandler handler = null;
diff --git a/phoenix-pherf/src/main/java/org/apache/phoenix/pherf/result/Result.java b/phoenix-pherf/src/main/java/org/apache/phoenix/pherf/result/Result.java
index 158ed11..93225d7 100644
--- a/phoenix-pherf/src/main/java/org/apache/phoenix/pherf/result/Result.java
+++ b/phoenix-pherf/src/main/java/org/apache/phoenix/pherf/result/Result.java
@@ -36,7 +36,7 @@ public class Result {
* @param type {@link org.apache.phoenix.pherf.result.file.ResultFileDetails} Currently unused, but gives metadata about the
* contents of the result.
* @param header Used for CSV, otherwise pass null. For CSV pass comma separated string of header fields.
- * @param messageValues List<{@link ResultValue} All fields combined represent the data
+ * @param messageValues {@code List<ResultValue> } All fields combined represent the data
* for a row to be written.
*/
public Result(ResultFileDetails type, String header, List<ResultValue> messageValues) {
diff --git a/phoenix-pherf/src/main/java/org/apache/phoenix/pherf/result/ResultManager.java b/phoenix-pherf/src/main/java/org/apache/phoenix/pherf/result/ResultManager.java
index 1cf740e..91db782 100644
--- a/phoenix-pherf/src/main/java/org/apache/phoenix/pherf/result/ResultManager.java
+++ b/phoenix-pherf/src/main/java/org/apache/phoenix/pherf/result/ResultManager.java
@@ -121,7 +121,7 @@ public class ResultManager {
/**
* Write a combined set of results for each result in the list.
*
- * @param dataModelResults List<{@link DataModelResult > </>}
+ * @param dataModelResults {@code List<DataModelResult > }
* @throws Exception
*/
public synchronized void write(List<DataModelResult> dataModelResults, RulesApplier rulesApplier) throws Exception {
@@ -145,7 +145,6 @@ public class ResultManager {
/**
* Allows for flushing all the {@link org.apache.phoenix.pherf.result.ResultHandler}
- * @throws Exception
*/
public synchronized void flush(){
for (ResultHandler handler : resultHandlers) {
diff --git a/phoenix-pherf/src/main/java/org/apache/phoenix/pherf/rules/RuleBasedDataGenerator.java b/phoenix-pherf/src/main/java/org/apache/phoenix/pherf/rules/RuleBasedDataGenerator.java
index 24ecd20..d68e468 100644
--- a/phoenix-pherf/src/main/java/org/apache/phoenix/pherf/rules/RuleBasedDataGenerator.java
+++ b/phoenix-pherf/src/main/java/org/apache/phoenix/pherf/rules/RuleBasedDataGenerator.java
@@ -24,7 +24,7 @@ public interface RuleBasedDataGenerator {
* Get data value based on the rules
* Implementations should be thread safe as multiple theads will call it in parallel
*
- * @return {@link org.apache.phoenix.pherf.rules.DataValue} Container Type --> Value mapping
+ * @return {@link org.apache.phoenix.pherf.rules.DataValue} {@code Container Type --> Value } mapping
*/
DataValue getDataValue();
}
diff --git a/phoenix-pherf/src/main/java/org/apache/phoenix/pherf/rules/RulesApplier.java b/phoenix-pherf/src/main/java/org/apache/phoenix/pherf/rules/RulesApplier.java
index b58e2f7..b066e00 100644
--- a/phoenix-pherf/src/main/java/org/apache/phoenix/pherf/rules/RulesApplier.java
+++ b/phoenix-pherf/src/main/java/org/apache/phoenix/pherf/rules/RulesApplier.java
@@ -161,7 +161,7 @@ public class RulesApplier {
* Get data value based on the supplied rule
*
* @param column {@link org.apache.phoenix.pherf.configuration.Column} Column rule to get data for
- * @return {@link org.apache.phoenix.pherf.rules.DataValue} Container Type --> Value mapping
+ * @return {@link org.apache.phoenix.pherf.rules.DataValue} {@code Container Type --> Value mapping }
*/
public DataValue getDataValue(Column column) throws Exception{
DataValue data = null;