You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@drill.apache.org by vi...@apache.org on 2018/08/11 11:51:56 UTC
[drill] 02/03: DRILL-6656: Disallow extra semicolons and multiple
statements on the same line.
This is an automated email from the ASF dual-hosted git repository.
vitalii pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/drill.git
commit 3e3cd2bb76f6a98ca50798bdaa2d86777c7873ad
Author: Timothy Farkas <ti...@apache.org>
AuthorDate: Wed Aug 1 12:32:36 2018 -0700
DRILL-6656: Disallow extra semicolons and multiple statements on the same line.
closes #1415
---
.../java/org/apache/drill/common/types/Types.java | 2 +-
.../apache/drill/common/util/DrillStringUtils.java | 2 +-
.../java/org/apache/drill/test/DrillAssert.java | 9 +-
.../drill/exec/store/hbase/HBaseFilterBuilder.java | 2 +-
.../exec/store/hbase/HBasePushFilterIntoScan.java | 2 +-
.../drill/hbase/TestHBaseConnectionManager.java | 12 +--
.../drill/hbase/TestHBaseFilterPushDown.java | 118 +++++++++++----------
.../drill/hbase/TestHBaseProjectPushDown.java | 16 +--
.../org/apache/drill/hbase/TestHBaseQueries.java | 8 +-
.../hbase/TestHBaseRegionScanAssignments.java | 6 +-
.../org/apache/drill/hbase/TestTableGenerator.java | 17 +--
.../drill/exec/store/hive/HiveUtilities.java | 2 +-
.../initilializers/DefaultReadersInitializer.java | 2 +-
.../drill/exec/fn/hive/TestSampleHiveUDFs.java | 6 +-
.../drill/exec/store/jdbc/JdbcExpressionCheck.java | 1 -
.../drill/exec/store/jdbc/JdbcStoragePlugin.java | 6 +-
.../store/kafka/KafkaPartitionScanSpecBuilder.java | 2 +-
.../drill/yarn/appMaster/BatchScheduler.java | 5 +-
.../yarn/appMaster/http/AMSecurityManagerImpl.java | 1 -
.../org/apache/drill/yarn/client/CleanCommand.java | 1 -
.../org/apache/drill/yarn/core/TestConfig.java | 2 +-
.../org/apache/drill/yarn/scripts/ScriptUtils.java | 2 +-
.../java/org/apache/drill/exec/ExecConstants.java | 2 +-
.../drill/exec/cache/SerializationDefinition.java | 3 +-
.../org/apache/drill/exec/client/DrillClient.java | 2 +-
.../java/org/apache/drill/exec/client/DumpCat.java | 6 +-
.../exec/compile/sig/CodeGeneratorMethod.java | 2 +-
.../apache/drill/exec/dotdrill/DotDrillType.java | 6 +-
.../apache/drill/exec/expr/EvaluationVisitor.java | 4 +-
.../org/apache/drill/exec/expr/SizedJBlock.java | 2 +-
.../drill/exec/expr/fn/DrillAggFuncHolder.java | 6 +-
.../drill/exec/expr/fn/impl/Hash32Functions.java | 2 +-
.../drill/exec/expr/fn/impl/StringFunctions.java | 18 ++--
.../exec/expr/fn/impl/conv/JsonConvertTo.java | 3 +-
.../expr/fn/registry/FunctionRegistryHolder.java | 2 +-
.../drill/exec/physical/config/HashAggregate.java | 2 +-
.../drill/exec/physical/config/HashJoinPOP.java | 2 +-
.../physical/impl/aggregate/HashAggTemplate.java | 59 +++++++----
.../physical/impl/aggregate/HashAggregator.java | 2 +-
.../impl/aggregate/SpilledRecordbatch.java | 4 +-
.../impl/aggregate/StreamingAggregator.java | 3 +-
.../exec/physical/impl/common/HashTableStats.java | 8 +-
.../physical/impl/common/HashTableTemplate.java | 2 +-
.../physical/impl/flatten/FlattenTemplate.java | 2 +-
.../exec/physical/impl/join/HashJoinBatch.java | 2 +-
.../physical/impl/join/HashJoinProbeTemplate.java | 4 +-
.../drill/exec/physical/impl/join/JoinUtils.java | 2 +-
.../exec/physical/impl/join/LateralJoinBatch.java | 4 +-
.../partitionsender/PartitionSenderRootExec.java | 2 +-
.../exec/physical/impl/unnest/UnnestImpl.java | 2 +-
.../impl/xsort/managed/SortMemoryManager.java | 2 +-
.../exec/physical/rowSet/impl/ColumnState.java | 3 +-
.../physical/rowSet/impl/ResultSetLoaderImpl.java | 3 +-
.../exec/planner/AbstractPartitionDescriptor.java | 2 +-
.../drill/exec/planner/StarColumnHelper.java | 2 +-
.../exec/planner/common/DrillJoinRelBase.java | 4 +-
.../exec/planner/common/DrillLimitRelBase.java | 2 +-
.../exec/planner/common/DrillProjectRelBase.java | 6 +-
.../drill/exec/planner/cost/DrillCostBase.java | 2 +-
.../drill/exec/planner/logical/DrillFilterRel.java | 2 +-
.../drill/exec/planner/logical/DrillOptiq.java | 35 ++++--
.../exec/planner/logical/DrillUnionAllRule.java | 2 +-
.../drill/exec/planner/logical/DrillUnionRel.java | 2 +-
.../drill/exec/planner/physical/AggPrelBase.java | 2 +-
.../planner/physical/BroadcastExchangePrel.java | 2 +-
.../planner/physical/DrillDistributionTrait.java | 4 +-
.../drill/exec/planner/physical/HashAggPrule.java | 2 +-
.../planner/physical/HashToMergeExchangePrel.java | 2 +-
.../planner/physical/SingleMergeExchangePrel.java | 2 +-
.../exec/planner/physical/StreamAggPrule.java | 4 +-
.../physical/explain/NumberingRelWriter.java | 4 +-
.../physical/visitor/StarColumnConverter.java | 2 +-
.../apache/drill/exec/rpc/control/Controller.java | 2 +-
.../exec/server/rest/profile/ProfileWrapper.java | 2 +-
.../server/rest/profile/SimpleDurationFormat.java | 2 +-
.../java/org/apache/drill/exec/ssl/SSLConfig.java | 3 +-
.../drill/exec/store/easy/json/JsonProcessor.java | 2 +-
.../store/easy/text/compliant/HeaderBuilder.java | 2 +-
.../exec/store/easy/text/compliant/TextInput.java | 2 +-
.../apache/drill/exec/store/ischema/Records.java | 100 ++++++++++++-----
.../exec/store/parquet/ParquetFilterBuilder.java | 2 +-
.../exec/store/parquet/ParquetPushDownFilter.java | 2 +-
.../store/parquet/columnreaders/ColumnReader.java | 5 +-
.../store/parquet/metadata/MetadataPathUtils.java | 2 +-
.../drill/exec/store/schedule/BlockMapBuilder.java | 2 +-
.../org/apache/drill/exec/util/VectorUtil.java | 2 +-
.../drill/exec/work/batch/DataCollector.java | 2 +-
.../exec/work/batch/UnlimitedRawBatchBuffer.java | 2 +-
.../work/foreman/rm/ResourceManagerBuilder.java | 2 +-
.../test/java/org/apache/drill/PlanningBase.java | 2 +-
.../java/org/apache/drill/TestStarQueries.java | 2 +-
.../test/java/org/apache/drill/TestUnionAll.java | 4 +-
.../java/org/apache/drill/exec/RunRootExec.java | 2 +-
.../fn/registry/FunctionRegistryHolderTest.java | 4 +-
.../drill/exec/fn/impl/TestAggregateFunction.java | 2 +-
.../drill/exec/fn/impl/TestDateTruncFunctions.java | 4 +-
.../exec/fn/interp/ExpressionInterpreterTest.java | 2 +-
.../impersonation/TestImpersonationMetadata.java | 2 +-
.../apache/drill/exec/memory/TestAllocators.java | 3 +-
.../PartitionLimit/TestPartitionLimitBatch.java | 6 +-
.../drill/exec/physical/impl/TestDecimal.java | 4 +-
.../physical/impl/TestDistributedFragmentRun.java | 31 +++---
.../physical/impl/TopN/TestTopNSchemaChanges.java | 4 +-
.../physical/impl/agg/TestHashAggEmitOutcome.java | 4 +-
.../exec/physical/impl/agg/TestHashAggrSpill.java | 2 +-
.../physical/impl/common/HashPartitionTest.java | 6 +-
.../exec/physical/impl/flatten/TestFlatten.java | 2 +-
.../exec/physical/impl/join/TestHashJoinSpill.java | 2 +-
.../impl/protocol/TestOperatorRecordBatch.java | 5 +-
.../physical/impl/writer/TestParquetWriter.java | 2 +-
.../rowSet/impl/TestResultSetLoaderMapArray.java | 9 +-
.../rowSet/impl/TestResultSetLoaderMaps.java | 9 +-
.../rowSet/impl/TestResultSetLoaderTorture.java | 5 +-
.../exec/physical/unit/MiniPlanUnitTestBase.java | 10 +-
.../exec/physical/unit/TestNullInputMiniPlan.java | 2 +-
.../drill/exec/record/TestRecordBatchSizer.java | 2 +-
.../drill/exec/record/vector/TestDateTypes.java | 4 +-
.../drill/exec/store/dfs/TestFileSelection.java | 4 +-
.../exec/store/parquet/ParquetInternalsTest.java | 4 +-
.../parquet/ParquetSimpleTestFileGenerator.java | 14 +--
.../exec/store/parquet/TestParquetComplex.java | 76 ++++++-------
.../store/parquet/TestParquetPhysicalPlan.java | 3 +-
.../store/parquet2/TestDrillParquetReader.java | 82 +++++++-------
.../drill/exec/vector/TestSplitAndTransfer.java | 4 +-
.../exec/vector/complex/writer/TestJsonReader.java | 2 +-
.../org/apache/drill/test/DrillTestWrapper.java | 2 +-
.../java/org/apache/drill/test/ExampleTest.java | 3 +-
.../org/apache/drill/test/QueryRowSetIterator.java | 2 +-
.../java/org/apache/drill/test/TestBuilder.java | 8 +-
.../apache/drill/test/TestGracefulShutdown.java | 4 +-
.../test/rowSet/test/TestFixedWidthWriter.java | 2 +-
.../test/rowSet/test/TestOffsetVectorWriter.java | 2 +-
.../test/rowSet/test/TestVariableWidthWriter.java | 2 +-
.../org/apache/drill/jdbc/ITTestShadedJar.java | 3 +-
.../drill/jdbc/impl/DrillColumnMetaDataList.java | 12 ++-
.../apache/drill/jdbc/proxy/ProxiesManager.java | 3 +-
.../apache/drill/jdbc/PreparedStatementTest.java | 8 +-
.../java/org/apache/drill/jdbc/StatementTest.java | 8 +-
.../apache/drill/exec/memory/BaseAllocator.java | 3 +-
.../java/org/apache/drill/exec/rpc/RpcBus.java | 1 -
.../drill/common/expression/BooleanOperator.java | 2 +-
.../drill/common/expression/IfExpression.java | 2 +-
.../drill/common/expression/NullExpression.java | 2 +-
pom.xml | 2 +-
src/main/resources/checkstyle-config.xml | 4 +
145 files changed, 558 insertions(+), 451 deletions(-)
diff --git a/common/src/main/java/org/apache/drill/common/types/Types.java b/common/src/main/java/org/apache/drill/common/types/Types.java
index e66a340..799f691 100644
--- a/common/src/main/java/org/apache/drill/common/types/Types.java
+++ b/common/src/main/java/org/apache/drill/common/types/Types.java
@@ -57,7 +57,7 @@ public class Types {
}
public static boolean isRepeated(final MajorType type) {
- return type.getMode() == REPEATED ;
+ return type.getMode() == REPEATED;
}
public static boolean isNumericType(final MajorType type) {
diff --git a/common/src/main/java/org/apache/drill/common/util/DrillStringUtils.java b/common/src/main/java/org/apache/drill/common/util/DrillStringUtils.java
index bc3b01b..cf4b4c8 100644
--- a/common/src/main/java/org/apache/drill/common/util/DrillStringUtils.java
+++ b/common/src/main/java/org/apache/drill/common/util/DrillStringUtils.java
@@ -117,7 +117,7 @@ public class DrillStringUtils {
*/
public static String toBinaryString(ByteBuf buf, int strStart, int strEnd) {
StringBuilder result = new StringBuilder();
- for (int i = strStart; i < strEnd ; ++i) {
+ for (int i = strStart; i < strEnd; ++i) {
appendByte(result, buf.getByte(i));
}
return result.toString();
diff --git a/common/src/test/java/org/apache/drill/test/DrillAssert.java b/common/src/test/java/org/apache/drill/test/DrillAssert.java
index 0feca06..5683d29 100644
--- a/common/src/test/java/org/apache/drill/test/DrillAssert.java
+++ b/common/src/test/java/org/apache/drill/test/DrillAssert.java
@@ -36,13 +36,16 @@ public class DrillAssert {
ch1 = expected.charAt(idx1);
ch2 = actual.charAt(idx2);
if (isNewLineChar(ch1)) {
- idx1++; continue;
+ idx1++;
+ continue;
} else if (isNewLineChar(ch2)) {
- idx2++; continue;
+ idx2++;
+ continue;
} else if (ch1 != ch2) {
break outside;
} else {
- idx1++; idx2++;
+ idx1++;
+ idx2++;
}
}
// skip newlines at the end
diff --git a/contrib/storage-hbase/src/main/java/org/apache/drill/exec/store/hbase/HBaseFilterBuilder.java b/contrib/storage-hbase/src/main/java/org/apache/drill/exec/store/hbase/HBaseFilterBuilder.java
index 6e1efe5..e73b512 100644
--- a/contrib/storage-hbase/src/main/java/org/apache/drill/exec/store/hbase/HBaseFilterBuilder.java
+++ b/contrib/storage-hbase/src/main/java/org/apache/drill/exec/store/hbase/HBaseFilterBuilder.java
@@ -292,7 +292,7 @@ public class HBaseFilterBuilder extends AbstractExprVisitor<HBaseScanSpec, Void,
startRow = prefix.getBytes(Charsets.UTF_8);
stopRow = startRow.clone();
boolean isMaxVal = true;
- for (int i = stopRow.length - 1; i >= 0 ; --i) {
+ for (int i = stopRow.length - 1; i >= 0; --i) {
int nextByteValue = (0xff & stopRow[i]) + 1;
if (nextByteValue < 0xff) {
stopRow[i] = (byte) nextByteValue;
diff --git a/contrib/storage-hbase/src/main/java/org/apache/drill/exec/store/hbase/HBasePushFilterIntoScan.java b/contrib/storage-hbase/src/main/java/org/apache/drill/exec/store/hbase/HBasePushFilterIntoScan.java
index 632f046..1128f39 100644
--- a/contrib/storage-hbase/src/main/java/org/apache/drill/exec/store/hbase/HBasePushFilterIntoScan.java
+++ b/contrib/storage-hbase/src/main/java/org/apache/drill/exec/store/hbase/HBasePushFilterIntoScan.java
@@ -125,7 +125,7 @@ public abstract class HBasePushFilterIntoScan extends StoragePluginOptimizerRule
final ScanPrel newScanPrel = ScanPrel.create(scan, filter.getTraitSet(), newGroupsScan, scan.getRowType());
// Depending on whether is a project in the middle, assign either scan or copy of project to childRel.
- final RelNode childRel = project == null ? newScanPrel : project.copy(project.getTraitSet(), ImmutableList.of((RelNode)newScanPrel));;
+ final RelNode childRel = project == null ? newScanPrel : project.copy(project.getTraitSet(), ImmutableList.of(newScanPrel));
if (hbaseFilterBuilder.isAllExpressionsConverted()) {
/*
diff --git a/contrib/storage-hbase/src/test/java/org/apache/drill/hbase/TestHBaseConnectionManager.java b/contrib/storage-hbase/src/test/java/org/apache/drill/hbase/TestHBaseConnectionManager.java
index 36224b3..3227277 100644
--- a/contrib/storage-hbase/src/test/java/org/apache/drill/hbase/TestHBaseConnectionManager.java
+++ b/contrib/storage-hbase/src/test/java/org/apache/drill/hbase/TestHBaseConnectionManager.java
@@ -31,8 +31,8 @@ public class TestHBaseConnectionManager extends BaseHBaseTest {
runHBaseSQLVerifyCount("SELECT\n"
+ "row_key\n"
+ "FROM\n"
- + " hbase.`[TABLE_NAME]` tableName"
- , 8);
+ + " hbase.`[TABLE_NAME]` tableName",
+ 8);
/*
* Simulate HBase connection close and ensure that the connection
@@ -42,8 +42,8 @@ public class TestHBaseConnectionManager extends BaseHBaseTest {
runHBaseSQLVerifyCount("SELECT\n"
+ "row_key\n"
+ "FROM\n"
- + " hbase.`[TABLE_NAME]` tableName"
- , 8);
+ + " hbase.`[TABLE_NAME]` tableName",
+ 8);
/*
* Simulate HBase cluster restart and ensure that running query against
@@ -54,8 +54,8 @@ public class TestHBaseConnectionManager extends BaseHBaseTest {
runHBaseSQLVerifyCount("SELECT\n"
+ "row_key\n"
+ "FROM\n"
- + " hbase.`[TABLE_NAME]` tableName"
- , 8);
+ + " hbase.`[TABLE_NAME]` tableName",
+ 8);
}
diff --git a/contrib/storage-hbase/src/test/java/org/apache/drill/hbase/TestHBaseFilterPushDown.java b/contrib/storage-hbase/src/test/java/org/apache/drill/hbase/TestHBaseFilterPushDown.java
index aee18eb..cccaeb1 100644
--- a/contrib/storage-hbase/src/test/java/org/apache/drill/hbase/TestHBaseFilterPushDown.java
+++ b/contrib/storage-hbase/src/test/java/org/apache/drill/hbase/TestHBaseFilterPushDown.java
@@ -91,8 +91,8 @@ public class TestHBaseFilterPushDown extends BaseHBaseTest {
+ " FROM hbase.`TestTableCompositeDate` tableName\n"
+ " WHERE\n"
+ " CONVERT_FROM(BYTE_SUBSTR(row_key, 1, 8), 'date_epoch_be') < DATE '2015-06-18' AND\n"
- + " CONVERT_FROM(BYTE_SUBSTR(row_key, 1, 8), 'date_epoch_be') > DATE '2015-06-13'"
- , 12);
+ + " CONVERT_FROM(BYTE_SUBSTR(row_key, 1, 8), 'date_epoch_be') > DATE '2015-06-13'",
+ 12);
}
@Test
@@ -104,8 +104,8 @@ public class TestHBaseFilterPushDown extends BaseHBaseTest {
+ ", CONVERT_FROM(tableName.f.c, 'UTF8') \n"
+ " FROM hbase.`TestTableCompositeDate` tableName\n"
+ " WHERE\n"
- + " CONVERT_FROM(BYTE_SUBSTR(row_key, 1, 8), 'date_epoch_be') = DATE '2015-08-22'"
- , 3);
+ + " CONVERT_FROM(BYTE_SUBSTR(row_key, 1, 8), 'date_epoch_be') = DATE '2015-08-22'",
+ 3);
}
@Test
@@ -118,8 +118,8 @@ public class TestHBaseFilterPushDown extends BaseHBaseTest {
+ " FROM hbase.`TestTableCompositeDate` tableName\n"
+ " WHERE\n"
+ " CONVERT_FROM(BYTE_SUBSTR(row_key, 1, 8), 'date_epoch_be') < DATE '2015-06-18' AND\n"
- + " CONVERT_FROM(BYTE_SUBSTR(row_key, 1, 8), 'date_epoch_be') > DATE '2015-06-13'"
- , 1);
+ + " CONVERT_FROM(BYTE_SUBSTR(row_key, 1, 8), 'date_epoch_be') > DATE '2015-06-13'",
+ 1);
}
@Test
@@ -133,8 +133,8 @@ public class TestHBaseFilterPushDown extends BaseHBaseTest {
+ " FROM hbase.`TestTableCompositeDate` tableName\n"
+ " WHERE\n"
+ " CONVERT_FROM(BYTE_SUBSTR(row_key, 1, 8), 'timestamp_epoch_be') >= TIMESTAMP '2015-06-18 08:00:00.000' AND\n"
- + " CONVERT_FROM(BYTE_SUBSTR(row_key, 1, 8), 'timestamp_epoch_be') < TIMESTAMP '2015-06-20 16:00:00.000'"
- , 7);
+ + " CONVERT_FROM(BYTE_SUBSTR(row_key, 1, 8), 'timestamp_epoch_be') < TIMESTAMP '2015-06-20 16:00:00.000'",
+ 7);
}
@Test
@@ -146,8 +146,8 @@ public class TestHBaseFilterPushDown extends BaseHBaseTest {
+ ", CONVERT_FROM(tableName.f.c, 'UTF8') \n"
+ " FROM hbase.`TestTableCompositeTime` tableName\n"
+ " WHERE\n"
- + " CONVERT_FROM(BYTE_SUBSTR(row_key, 1, 8), 'time_epoch_be') = TIME '23:57:15.275'"//convert_from(binary_string('\\x00\\x00\\x00\\x00\\x55\\x4D\\xBE\\x80'), 'BIGINT_BE') \n"
- , 1);
+ + " CONVERT_FROM(BYTE_SUBSTR(row_key, 1, 8), 'time_epoch_be') = TIME '23:57:15.275'",//convert_from(binary_string('\\x00\\x00\\x00\\x00\\x55\\x4D\\xBE\\x80'), 'BIGINT_BE') \n"
+ 1);
}
@Test
@@ -159,8 +159,9 @@ public class TestHBaseFilterPushDown extends BaseHBaseTest {
+ ", CONVERT_FROM(tableName.f.c, 'UTF8') \n"
+ " FROM hbase.`TestTableCompositeTime` tableName\n"
+ " WHERE\n"
- + " CONVERT_FROM(BYTE_SUBSTR(row_key, 1, 8), 'time_epoch_be') = TIME '23:55:51.250'"//convert_from(binary_string('\\x00\\x00\\x00\\x00\\x55\\x4D\\xBE\\x80'), 'BIGINT_BE') \n"
- , 1);
+ + " CONVERT_FROM(BYTE_SUBSTR(row_key, 1, 8), 'time_epoch_be') = TIME '23:55:51.250'",//convert_from(binary_string('\\x00\\x00\\x00\\x00\\x55\\x4D\\xBE\\x80'),
+ // 'BIGINT_BE') \n"
+ 1);
}
@Test
@@ -173,8 +174,8 @@ public class TestHBaseFilterPushDown extends BaseHBaseTest {
+ " FROM hbase.`TestTableCompositeTime` tableName\n"
+ " WHERE\n"
+ " CONVERT_FROM(BYTE_SUBSTR(row_key, 1, 8), 'time_epoch_be') > TIME '23:57:06' AND"//convert_from(binary_string('\\x00\\x00\\x00\\x00\\x55\\x4D\\xBE\\x80'), 'BIGINT_BE') \n"
- + " CONVERT_FROM(BYTE_SUBSTR(row_key, 1, 8), 'time_epoch_be') < TIME '23:59:59'"
- , 8);
+ + " CONVERT_FROM(BYTE_SUBSTR(row_key, 1, 8), 'time_epoch_be') < TIME '23:59:59'",
+ 8);
}
@Test
@@ -186,8 +187,9 @@ public class TestHBaseFilterPushDown extends BaseHBaseTest {
+ ", CONVERT_FROM(tableName.f.c, 'UTF8') \n"
+ " FROM hbase.`TestTableCompositeDate` tableName\n"
+ " WHERE\n"
- + " CONVERT_FROM(BYTE_SUBSTR(row_key, 1, 8), 'bigint_be') = cast(1409040000000 as bigint)"//convert_from(binary_string('\\x00\\x00\\x00\\x00\\x55\\x4D\\xBE\\x80'), 'BIGINT_BE') \n"
- , 1);
+ + " CONVERT_FROM(BYTE_SUBSTR(row_key, 1, 8), 'bigint_be') = cast(1409040000000 as bigint)",//convert_from(binary_string('\\x00\\x00\\x00\\x00\\x55\\x4D\\xBE\\x80'),
+ // 'BIGINT_BE') \n"
+ 1);
}
@Test
@@ -202,8 +204,8 @@ public class TestHBaseFilterPushDown extends BaseHBaseTest {
+ " FROM hbase.`TestTableCompositeDate` tableName\n"
+ " WHERE\n"
+ " CONVERT_FROM(BYTE_SUBSTR(row_key, 1, 8), 'uint8_be') > cast(1438300800000 as bigint) AND\n"
- + " CONVERT_FROM(BYTE_SUBSTR(row_key, 1, 8), 'uint8_be') < cast(1438617600000 as bigint)"
- , 10);
+ + " CONVERT_FROM(BYTE_SUBSTR(row_key, 1, 8), 'uint8_be') < cast(1438617600000 as bigint)",
+ 10);
}
@Test
@@ -216,8 +218,8 @@ public class TestHBaseFilterPushDown extends BaseHBaseTest {
+ " FROM hbase.`TestTableCompositeInt` tableName\n"
+ " WHERE\n"
+ " CONVERT_FROM(BYTE_SUBSTR(row_key, 1, 4), 'uint4_be') >= cast(423 as int) AND"
- + " CONVERT_FROM(BYTE_SUBSTR(row_key, 1, 4), 'uint4_be') < cast(940 as int)"
- , 11);
+ + " CONVERT_FROM(BYTE_SUBSTR(row_key, 1, 4), 'uint4_be') < cast(940 as int)",
+ 11);
}
@Test
@@ -230,8 +232,8 @@ public class TestHBaseFilterPushDown extends BaseHBaseTest {
+ " FROM hbase.`TestTableCompositeInt` tableName\n"
+ " WHERE\n"
+ " CONVERT_FROM(BYTE_SUBSTR(row_key, 1, 4), 'uint4_be') >= cast(300 as int) AND"
- + " CONVERT_FROM(BYTE_SUBSTR(row_key, 1, 4), 'uint4_be') < cast(900 as int)"
- , 1);
+ + " CONVERT_FROM(BYTE_SUBSTR(row_key, 1, 4), 'uint4_be') < cast(900 as int)",
+ 1);
}
@Test
@@ -243,8 +245,8 @@ public class TestHBaseFilterPushDown extends BaseHBaseTest {
+ ", CONVERT_FROM(tableName.f.c, 'UTF8') \n"
+ " FROM hbase.`TestTableCompositeInt` tableName\n"
+ " WHERE\n"
- + " CONVERT_FROM(BYTE_SUBSTR(row_key, 1, 4), 'uint4_be') = cast(658 as int)"
- , 1);
+ + " CONVERT_FROM(BYTE_SUBSTR(row_key, 1, 4), 'uint4_be') = cast(658 as int)",
+ 1);
}
@Test
@@ -256,8 +258,8 @@ public class TestHBaseFilterPushDown extends BaseHBaseTest {
+ "FROM\n"
+ " hbase.`TestTableDoubleOB` t\n"
+ "WHERE\n"
- + " CONVERT_FROM(row_key, 'DOUBLE_OB') > cast(95.54 as DOUBLE)"
- , 6);
+ + " CONVERT_FROM(row_key, 'DOUBLE_OB') > cast(95.54 as DOUBLE)",
+ 6);
}
@Test
@@ -270,8 +272,8 @@ public class TestHBaseFilterPushDown extends BaseHBaseTest {
+ "FROM\n"
+ " hbase.`TestTableDoubleOB` t\n"
+ "WHERE\n"
- + " CONVERT_FROM(row_key, 'DOUBLE_OB') > cast(95.54 as DOUBLE)"
- , 1);
+ + " CONVERT_FROM(row_key, 'DOUBLE_OB') > cast(95.54 as DOUBLE)",
+ 1);
}
@Test
@@ -283,8 +285,8 @@ public class TestHBaseFilterPushDown extends BaseHBaseTest {
+ "FROM\n"
+ " hbase.`TestTableDoubleOBDesc` t\n"
+ "WHERE\n"
- + " CONVERT_FROM(row_key, 'DOUBLE_OBD') > cast(95.54 as DOUBLE)"
- , 6);
+ + " CONVERT_FROM(row_key, 'DOUBLE_OBD') > cast(95.54 as DOUBLE)",
+ 6);
}
@Test
@@ -297,8 +299,8 @@ public class TestHBaseFilterPushDown extends BaseHBaseTest {
+ "FROM\n"
+ " hbase.`TestTableDoubleOBDesc` t\n"
+ "WHERE\n"
- + " CONVERT_FROM(row_key, 'DOUBLE_OBD') > cast(95.54 as DOUBLE)"
- , 1);
+ + " CONVERT_FROM(row_key, 'DOUBLE_OBD') > cast(95.54 as DOUBLE)",
+ 1);
}
@Test
@@ -311,8 +313,8 @@ public class TestHBaseFilterPushDown extends BaseHBaseTest {
+ " hbase.`TestTableIntOB` t\n"
+ "WHERE\n"
+ " CONVERT_FROM(row_key, 'INT_OB') >= cast(-32 as INT) AND"
- + " CONVERT_FROM(row_key, 'INT_OB') < cast(59 as INT)"
- , 91);
+ + " CONVERT_FROM(row_key, 'INT_OB') < cast(59 as INT)",
+ 91);
}
@Test
@@ -325,8 +327,8 @@ public class TestHBaseFilterPushDown extends BaseHBaseTest {
+ " hbase.`TestTableIntOBDesc` t\n"
+ "WHERE\n"
+ " CONVERT_FROM(row_key, 'INT_OBD') >= cast(-32 as INT) AND"
- + " CONVERT_FROM(row_key, 'INT_OBD') < cast(59 as INT)"
- , 91);
+ + " CONVERT_FROM(row_key, 'INT_OBD') < cast(59 as INT)",
+ 91);
}
@Test
@@ -340,8 +342,8 @@ public class TestHBaseFilterPushDown extends BaseHBaseTest {
+ " hbase.`TestTableIntOB` t\n"
+ "WHERE\n"
+ " CONVERT_FROM(row_key, 'INT_OB') > cast(-23 as INT) AND"
- + " CONVERT_FROM(row_key, 'INT_OB') < cast(14 as INT)"
- , 1);
+ + " CONVERT_FROM(row_key, 'INT_OB') < cast(14 as INT)",
+ 1);
}
@Test
@@ -355,8 +357,8 @@ public class TestHBaseFilterPushDown extends BaseHBaseTest {
+ " hbase.`TestTableIntOBDesc` t\n"
+ "WHERE\n"
+ " CONVERT_FROM(row_key, 'INT_OBD') > cast(-23 as INT) AND"
- + " CONVERT_FROM(row_key, 'INT_OBD') < cast(14 as INT)"
- , 1);
+ + " CONVERT_FROM(row_key, 'INT_OBD') < cast(14 as INT)",
+ 1);
}
@Test
@@ -369,8 +371,8 @@ public class TestHBaseFilterPushDown extends BaseHBaseTest {
+ " hbase.`TestTableBigIntOB` t\n"
+ "WHERE\n"
+ " CONVERT_FROM(row_key, 'BIGINT_OB') > cast(1438034423063 as BIGINT) AND"
- + " CONVERT_FROM(row_key, 'BIGINT_OB') <= cast(1438034423097 as BIGINT)"
- , 34);
+ + " CONVERT_FROM(row_key, 'BIGINT_OB') <= cast(1438034423097 as BIGINT)",
+ 34);
}
@Test
@@ -384,8 +386,8 @@ public class TestHBaseFilterPushDown extends BaseHBaseTest {
+ " hbase.`TestTableBigIntOB` t\n"
+ "WHERE\n"
+ " CONVERT_FROM(row_key, 'BIGINT_OB') > cast(1438034423063 as BIGINT) AND"
- + " CONVERT_FROM(row_key, 'BIGINT_OB') < cast(1438034423097 as BIGINT)"
- , 1);
+ + " CONVERT_FROM(row_key, 'BIGINT_OB') < cast(1438034423097 as BIGINT)",
+ 1);
}
@Test
@@ -398,8 +400,8 @@ public class TestHBaseFilterPushDown extends BaseHBaseTest {
+ " hbase.`TestTableFloatOB` t\n"
+ "WHERE\n"
+ " CONVERT_FROM(row_key, 'FLOAT_OB') > cast(95.74 as FLOAT) AND"
- + " CONVERT_FROM(row_key, 'FLOAT_OB') < cast(99.5 as FLOAT)"
- , 5);
+ + " CONVERT_FROM(row_key, 'FLOAT_OB') < cast(99.5 as FLOAT)",
+ 5);
}
@Test
@@ -413,8 +415,8 @@ public class TestHBaseFilterPushDown extends BaseHBaseTest {
+ " hbase.`TestTableFloatOB` t\n"
+ "WHERE\n"
+ " CONVERT_FROM(row_key, 'FLOAT_OB') > cast(95.54 as FLOAT) AND"
- + " CONVERT_FROM(row_key, 'FLOAT_OB') < cast(99.77 as FLOAT)"
- , 1);
+ + " CONVERT_FROM(row_key, 'FLOAT_OB') < cast(99.77 as FLOAT)",
+ 1);
}
@Test
@@ -427,8 +429,8 @@ public class TestHBaseFilterPushDown extends BaseHBaseTest {
+ " hbase.`TestTableBigIntOBDesc` t\n"
+ "WHERE\n"
+ " CONVERT_FROM(row_key, 'BIGINT_OBD') > cast(1438034423063 as BIGINT) AND"
- + " CONVERT_FROM(row_key, 'BIGINT_OBD') <= cast(1438034423097 as BIGINT)"
- , 34);
+ + " CONVERT_FROM(row_key, 'BIGINT_OBD') <= cast(1438034423097 as BIGINT)",
+ 34);
}
@Test
@@ -442,8 +444,8 @@ public class TestHBaseFilterPushDown extends BaseHBaseTest {
+ " hbase.`TestTableBigIntOBDesc` t\n"
+ "WHERE\n"
+ " CONVERT_FROM(row_key, 'BIGINT_OBD') > cast(1438034423063 as BIGINT) AND"
- + " CONVERT_FROM(row_key, 'BIGINT_OBD') < cast(1438034423097 as BIGINT)"
- , 1);
+ + " CONVERT_FROM(row_key, 'BIGINT_OBD') < cast(1438034423097 as BIGINT)",
+ 1);
}
@Test
@@ -456,8 +458,8 @@ public class TestHBaseFilterPushDown extends BaseHBaseTest {
+ " hbase.`TestTableFloatOBDesc` t\n"
+ "WHERE\n"
+ " CONVERT_FROM(row_key, 'FLOAT_OBD') > cast(95.74 as FLOAT) AND"
- + " CONVERT_FROM(row_key, 'FLOAT_OBD') < cast(99.5 as FLOAT)"
- , 5);
+ + " CONVERT_FROM(row_key, 'FLOAT_OBD') < cast(99.5 as FLOAT)",
+ 5);
}
@Test
@@ -471,8 +473,8 @@ public class TestHBaseFilterPushDown extends BaseHBaseTest {
+ " hbase.`TestTableFloatOBDesc` t\n"
+ "WHERE\n"
+ " CONVERT_FROM(row_key, 'FLOAT_OBD') > cast(95.54 as FLOAT) AND"
- + " CONVERT_FROM(row_key, 'FLOAT_OBD') < cast(99.77 as FLOAT)"
- , 1);
+ + " CONVERT_FROM(row_key, 'FLOAT_OBD') < cast(99.77 as FLOAT)",
+ 1);
}
@Test
@@ -680,8 +682,8 @@ public class TestHBaseFilterPushDown extends BaseHBaseTest {
+ "FROM\n"
+ " hbase.`[TABLE_NAME]` tableName\n"
+ "WHERE\n"
- + " convert_from(row_key, 'INT_BE') = 75"
- , 1);
+ + " convert_from(row_key, 'INT_BE') = 75",
+ 1);
}
@Test
diff --git a/contrib/storage-hbase/src/test/java/org/apache/drill/hbase/TestHBaseProjectPushDown.java b/contrib/storage-hbase/src/test/java/org/apache/drill/hbase/TestHBaseProjectPushDown.java
index 28bf036..9c0c707 100644
--- a/contrib/storage-hbase/src/test/java/org/apache/drill/hbase/TestHBaseProjectPushDown.java
+++ b/contrib/storage-hbase/src/test/java/org/apache/drill/hbase/TestHBaseProjectPushDown.java
@@ -31,8 +31,8 @@ public class TestHBaseProjectPushDown extends BaseHBaseTest {
runHBaseSQLVerifyCount("SELECT\n"
+ "row_key\n"
+ "FROM\n"
- + " hbase.`[TABLE_NAME]` tableName"
- , 8);
+ + " hbase.`[TABLE_NAME]` tableName",
+ 8);
}
@Test
@@ -41,8 +41,8 @@ public class TestHBaseProjectPushDown extends BaseHBaseTest {
runHBaseSQLVerifyCount("SELECT\n"
+ "t.f2.c7 as `t.f2.c7`\n"
+ "FROM\n"
- + " hbase.`[TABLE_NAME]` t"
- , 1);
+ + " hbase.`[TABLE_NAME]` t",
+ 1);
}
@Test
@@ -55,8 +55,8 @@ public class TestHBaseProjectPushDown extends BaseHBaseTest {
+ "row_key, t.f.c1 * 31 as `t dot f dot c1 * 31`, "
+ "t.f.c2 as `t dot f dot c2`, 5 as `5`, 'abc' as `'abc'`\n"
+ "FROM\n"
- + " hbase.`[TABLE_NAME]` t"
- , 8);
+ + " hbase.`[TABLE_NAME]` t",
+ 8);
}
@Test
@@ -65,8 +65,8 @@ public class TestHBaseProjectPushDown extends BaseHBaseTest {
runHBaseSQLVerifyCount("SELECT\n"
+ "row_key, f, f2\n"
+ "FROM\n"
- + " hbase.`[TABLE_NAME]` tableName"
- , 8);
+ + " hbase.`[TABLE_NAME]` tableName",
+ 8);
}
}
diff --git a/contrib/storage-hbase/src/test/java/org/apache/drill/hbase/TestHBaseQueries.java b/contrib/storage-hbase/src/test/java/org/apache/drill/hbase/TestHBaseQueries.java
index 27882b5..e8fa925 100644
--- a/contrib/storage-hbase/src/test/java/org/apache/drill/hbase/TestHBaseQueries.java
+++ b/contrib/storage-hbase/src/test/java/org/apache/drill/hbase/TestHBaseQueries.java
@@ -54,8 +54,8 @@ public class TestHBaseQueries extends BaseHBaseTest {
setColumnWidths(new int[] {8, 15});
runHBaseSQLVerifyCount("SELECT *\n"
+ "FROM\n"
- + " hbase.`" + tableName + "` tableName\n"
- , 1);
+ + " hbase.`" + tableName + "` tableName\n",
+ 1);
} finally {
try {
admin.disableTable(tableName);
@@ -78,8 +78,8 @@ public class TestHBaseQueries extends BaseHBaseTest {
setColumnWidths(new int[] {8, 15});
runHBaseSQLVerifyCount("SELECT row_key, count(*)\n"
+ "FROM\n"
- + " hbase.`" + tableName + "` tableName GROUP BY row_key\n"
- , 0);
+ + " hbase.`" + tableName + "` tableName GROUP BY row_key\n",
+ 0);
} finally {
try {
admin.disableTable(tableName);
diff --git a/contrib/storage-hbase/src/test/java/org/apache/drill/hbase/TestHBaseRegionScanAssignments.java b/contrib/storage-hbase/src/test/java/org/apache/drill/hbase/TestHBaseRegionScanAssignments.java
index 112eb87..a414bb7 100644
--- a/contrib/storage-hbase/src/test/java/org/apache/drill/hbase/TestHBaseRegionScanAssignments.java
+++ b/contrib/storage-hbase/src/test/java/org/apache/drill/hbase/TestHBaseRegionScanAssignments.java
@@ -21,6 +21,7 @@ import static org.junit.Assert.assertEquals;
import static org.junit.Assert.assertTrue;
import static org.junit.Assert.fail;
+import java.util.Collections;
import java.util.LinkedList;
import java.util.List;
import java.util.NavigableMap;
@@ -176,8 +177,9 @@ public class TestHBaseRegionScanAssignments extends BaseHBaseTest {
scan.applyAssignments(endpoints);
LinkedList<Integer> sizes = Lists.newLinkedList();
- sizes.add(1); sizes.add(1); sizes.add(1); sizes.add(1); sizes.add(1); sizes.add(1); sizes.add(1); sizes.add(1);
- sizes.add(2); sizes.add(2); sizes.add(2); sizes.add(2); sizes.add(2);
+ Collections.addAll(sizes, 1, 1, 1, 1, 1, 1, 1, 1);
+ Collections.addAll(sizes, 2, 2, 2, 2, 2);
+
for (int i = 0; i < endpoints.size(); i++) {
assertTrue(sizes.remove((Integer)scan.getSpecificScan(i).getRegionScanSpecList().size()));
}
diff --git a/contrib/storage-hbase/src/test/java/org/apache/drill/hbase/TestTableGenerator.java b/contrib/storage-hbase/src/test/java/org/apache/drill/hbase/TestTableGenerator.java
index 26d1c44..3c650db 100644
--- a/contrib/storage-hbase/src/test/java/org/apache/drill/hbase/TestTableGenerator.java
+++ b/contrib/storage-hbase/src/test/java/org/apache/drill/hbase/TestTableGenerator.java
@@ -235,7 +235,8 @@ public class TestTableGenerator {
for (int i = 0; i < numberRegions; i++) {
Put p = new Put((""+rowKeyChar+iteration).getBytes());
for (int j = 1; j <= numColumns; j++) {
- bytes = new byte[5000]; random.nextBytes(bytes);
+ bytes = new byte[5000];
+ random.nextBytes(bytes);
p.addColumn("f".getBytes(), ("c"+j).getBytes(), bytes);
}
table.mutate(p);
@@ -316,7 +317,7 @@ public class TestTableGenerator {
long endTime = startTime + MILLISECONDS_IN_A_YEAR;
long interval = MILLISECONDS_IN_A_DAY / 3;
- for (long ts = startTime, counter = 0; ts < endTime; ts += interval, counter ++) {
+ for (long ts = startTime, counter = 0; ts < endTime; ts += interval, counter++) {
byte[] rowKey = ByteBuffer.allocate(16) .putLong(ts).array();
for(int i = 0; i < 8; ++i) {
@@ -356,7 +357,7 @@ public class TestTableGenerator {
long largeInterval = MILLISECONDS_IN_A_SEC * 42;
long interval = smallInterval;
- for (long ts = startTime, counter = 0; ts < endTime; ts += interval, counter ++) {
+ for (long ts = startTime, counter = 0; ts < endTime; ts += interval, counter++) {
byte[] rowKey = ByteBuffer.allocate(16) .putLong(ts).array();
for(int i = 0; i < 8; ++i) {
@@ -398,7 +399,7 @@ public class TestTableGenerator {
int stopVal = 1000;
int interval = 47;
long counter = 0;
- for (int i = startVal; i < stopVal; i += interval, counter ++) {
+ for (int i = startVal; i < stopVal; i += interval, counter++) {
byte[] rowKey = ByteBuffer.allocate(12).putInt(i).array();
for(int j = 0; j < 8; ++j) {
@@ -492,7 +493,7 @@ public class TestTableGenerator {
BufferedMutator table = conn.getBufferedMutator(tableName);
long startTime = (long)1438034423 * 1000;
- for (long i = startTime; i <= startTime + 100; i ++) {
+ for (long i = startTime; i <= startTime + 100; i++) {
byte[] bytes = new byte[9];
PositionedByteRange br = new SimplePositionedMutableByteRange(bytes, 0, 9);
OrderedBytes.encodeInt64(br, i, Order.ASCENDING);
@@ -523,7 +524,7 @@ public class TestTableGenerator {
BufferedMutator table = conn.getBufferedMutator(tableName);
- for (int i = -49; i <= 100; i ++) {
+ for (int i = -49; i <= 100; i++) {
byte[] bytes = new byte[5];
PositionedByteRange br = new SimplePositionedMutableByteRange(bytes, 0, 5);
OrderedBytes.encodeInt32(br, i, Order.ASCENDING);
@@ -616,7 +617,7 @@ public class TestTableGenerator {
BufferedMutator table = conn.getBufferedMutator(tableName);
long startTime = (long)1438034423 * 1000;
- for (long i = startTime; i <= startTime + 100; i ++) {
+ for (long i = startTime; i <= startTime + 100; i++) {
byte[] bytes = new byte[9];
PositionedByteRange br = new SimplePositionedMutableByteRange(bytes, 0, 9);
OrderedBytes.encodeInt64(br, i, Order.DESCENDING);
@@ -648,7 +649,7 @@ public class TestTableGenerator {
BufferedMutator table = conn.getBufferedMutator(tableName);
- for (int i = -49; i <= 100; i ++) {
+ for (int i = -49; i <= 100; i++) {
byte[] bytes = new byte[5];
PositionedByteRange br = new SimplePositionedMutableByteRange(bytes, 0, 5);
OrderedBytes.encodeInt32(br, i, Order.DESCENDING);
diff --git a/contrib/storage-hive/core/src/main/java/org/apache/drill/exec/store/hive/HiveUtilities.java b/contrib/storage-hive/core/src/main/java/org/apache/drill/exec/store/hive/HiveUtilities.java
index 6fc567e..2beb539 100644
--- a/contrib/storage-hive/core/src/main/java/org/apache/drill/exec/store/hive/HiveUtilities.java
+++ b/contrib/storage-hive/core/src/main/java/org/apache/drill/exec/store/hive/HiveUtilities.java
@@ -436,7 +436,7 @@ public class HiveUtilities {
storageHandler.configureInputJobProperties(tableDesc, table.getParameters());
return (Class<? extends InputFormat<?, ?>>) storageHandler.getInputFormatClass();
} else {
- return (Class<? extends InputFormat<?, ?>>) Class.forName(inputFormatName) ;
+ return (Class<? extends InputFormat<?, ?>>) Class.forName(inputFormatName);
}
}
diff --git a/contrib/storage-hive/core/src/main/java/org/apache/drill/exec/store/hive/readers/initilializers/DefaultReadersInitializer.java b/contrib/storage-hive/core/src/main/java/org/apache/drill/exec/store/hive/readers/initilializers/DefaultReadersInitializer.java
index 52f394f..b383469 100644
--- a/contrib/storage-hive/core/src/main/java/org/apache/drill/exec/store/hive/readers/initilializers/DefaultReadersInitializer.java
+++ b/contrib/storage-hive/core/src/main/java/org/apache/drill/exec/store/hive/readers/initilializers/DefaultReadersInitializer.java
@@ -46,7 +46,7 @@ public class DefaultReadersInitializer extends AbstractReadersInitializer {
List<RecordReader> readers = new LinkedList<>();
Constructor<? extends HiveAbstractReader> readerConstructor = createReaderConstructor();
- for (int i = 0 ; i < inputSplits.size(); i++) {
+ for (int i = 0; i < inputSplits.size(); i++) {
readers.add(createReader(readerConstructor, hasPartitions ? partitions.get(i) : null, inputSplits.get(i)));
}
return readers;
diff --git a/contrib/storage-hive/core/src/test/java/org/apache/drill/exec/fn/hive/TestSampleHiveUDFs.java b/contrib/storage-hive/core/src/test/java/org/apache/drill/exec/fn/hive/TestSampleHiveUDFs.java
index 7808352..932b144 100644
--- a/contrib/storage-hive/core/src/test/java/org/apache/drill/exec/fn/hive/TestSampleHiveUDFs.java
+++ b/contrib/storage-hive/core/src/test/java/org/apache/drill/exec/fn/hive/TestSampleHiveUDFs.java
@@ -84,7 +84,8 @@ public class TestSampleHiveUDFs extends HiveTestBase {
@Test
public void floatInOut() throws Exception{
String query = "SELECT testHiveUDFFloat(float_field) as col1 FROM hive.readtest";
- String expected = "col1\n" + "4.67\n" + "null\n"; helper(query, expected);
+ String expected = "col1\n" + "4.67\n" + "null\n";
+ helper(query, expected);
}
@Test
@@ -105,7 +106,8 @@ public class TestSampleHiveUDFs extends HiveTestBase {
public void binaryInOut() throws Exception{
String query = "SELECT testHiveUDFBinary(binary_field) as col1 FROM hive.readtest";
String expected = "col1\n" + "binaryfield\n" + "null\n";
- helper(query, expected); helper(query, expected);
+ helper(query, expected);
+ helper(query, expected);
}
@Test
diff --git a/contrib/storage-jdbc/src/main/java/org/apache/drill/exec/store/jdbc/JdbcExpressionCheck.java b/contrib/storage-jdbc/src/main/java/org/apache/drill/exec/store/jdbc/JdbcExpressionCheck.java
index e0b01b3..4637abb 100644
--- a/contrib/storage-jdbc/src/main/java/org/apache/drill/exec/store/jdbc/JdbcExpressionCheck.java
+++ b/contrib/storage-jdbc/src/main/java/org/apache/drill/exec/store/jdbc/JdbcExpressionCheck.java
@@ -80,7 +80,6 @@ class JdbcExpressionCheck implements RexVisitor<Boolean> {
if (!visitCall(over)) {
return false;
}
- ;
final RexWindow window = over.getWindow();
for (RexFieldCollation orderKey : window.orderKeys) {
diff --git a/contrib/storage-jdbc/src/main/java/org/apache/drill/exec/store/jdbc/JdbcStoragePlugin.java b/contrib/storage-jdbc/src/main/java/org/apache/drill/exec/store/jdbc/JdbcStoragePlugin.java
index c38ea3b..efb5a5c 100755
--- a/contrib/storage-jdbc/src/main/java/org/apache/drill/exec/store/jdbc/JdbcStoragePlugin.java
+++ b/contrib/storage-jdbc/src/main/java/org/apache/drill/exec/store/jdbc/JdbcStoragePlugin.java
@@ -305,7 +305,8 @@ public class JdbcStoragePlugin extends AbstractStoragePlugin {
public JdbcCatalogSchema(String name) {
super(ImmutableList.<String> of(), name);
- try (Connection con = source.getConnection(); ResultSet set = con.getMetaData().getCatalogs()) {
+ try (Connection con = source.getConnection();
+ ResultSet set = con.getMetaData().getCatalogs()) {
while (set.next()) {
final String catalogName = set.getString(1);
CapitalizingJdbcSchema schema = new CapitalizingJdbcSchema(
@@ -347,7 +348,8 @@ public class JdbcStoragePlugin extends AbstractStoragePlugin {
private boolean addSchemas() {
boolean added = false;
- try (Connection con = source.getConnection(); ResultSet set = con.getMetaData().getSchemas()) {
+ try (Connection con = source.getConnection();
+ ResultSet set = con.getMetaData().getSchemas()) {
while (set.next()) {
final String schemaName = set.getString(1);
final String catalogName = set.getString(2);
diff --git a/contrib/storage-kafka/src/main/java/org/apache/drill/exec/store/kafka/KafkaPartitionScanSpecBuilder.java b/contrib/storage-kafka/src/main/java/org/apache/drill/exec/store/kafka/KafkaPartitionScanSpecBuilder.java
index b52ed44..202670e 100644
--- a/contrib/storage-kafka/src/main/java/org/apache/drill/exec/store/kafka/KafkaPartitionScanSpecBuilder.java
+++ b/contrib/storage-kafka/src/main/java/org/apache/drill/exec/store/kafka/KafkaPartitionScanSpecBuilder.java
@@ -41,7 +41,7 @@ public class KafkaPartitionScanSpecBuilder extends
static final org.slf4j.Logger logger = org.slf4j.LoggerFactory.getLogger(KafkaPartitionScanSpecBuilder.class);
private final LogicalExpression le;
private final KafkaGroupScan groupScan;
- private final KafkaConsumer<? ,?> kafkaConsumer;
+ private final KafkaConsumer<?, ?> kafkaConsumer;
private ImmutableMap<TopicPartition, KafkaPartitionScanSpec> fullScanSpec;
private static final long CLOSE_TIMEOUT_MS = 200;
diff --git a/drill-yarn/src/main/java/org/apache/drill/yarn/appMaster/BatchScheduler.java b/drill-yarn/src/main/java/org/apache/drill/yarn/appMaster/BatchScheduler.java
index 8f3aaab..acb254a 100644
--- a/drill-yarn/src/main/java/org/apache/drill/yarn/appMaster/BatchScheduler.java
+++ b/drill-yarn/src/main/java/org/apache/drill/yarn/appMaster/BatchScheduler.java
@@ -35,7 +35,10 @@ public class BatchScheduler extends AbstractScheduler {
}
@Override
- public int resize(int level) { quantity = level; return quantity; }
+ public int resize(int level) {
+ quantity = level;
+ return quantity;
+ }
@Override
public int getTarget() { return quantity; }
diff --git a/drill-yarn/src/main/java/org/apache/drill/yarn/appMaster/http/AMSecurityManagerImpl.java b/drill-yarn/src/main/java/org/apache/drill/yarn/appMaster/http/AMSecurityManagerImpl.java
index d31690e..2dbba65 100644
--- a/drill-yarn/src/main/java/org/apache/drill/yarn/appMaster/http/AMSecurityManagerImpl.java
+++ b/drill-yarn/src/main/java/org/apache/drill/yarn/appMaster/http/AMSecurityManagerImpl.java
@@ -182,7 +182,6 @@ public class AMSecurityManagerImpl implements AMSecurityManager {
managerImpl.init();
} else if (DoYUtil.isBlank(authType)
|| DrillOnYarnConfig.AUTH_TYPE_NONE.equals(authType)) {
- ;
} else {
LOG.error("Unrecognized authorization type for "
+ DrillOnYarnConfig.HTTP_AUTH_TYPE + ": " + authType
diff --git a/drill-yarn/src/main/java/org/apache/drill/yarn/client/CleanCommand.java b/drill-yarn/src/main/java/org/apache/drill/yarn/client/CleanCommand.java
index 1fcba2d..5a8a8b1 100644
--- a/drill-yarn/src/main/java/org/apache/drill/yarn/client/CleanCommand.java
+++ b/drill-yarn/src/main/java/org/apache/drill/yarn/client/CleanCommand.java
@@ -70,7 +70,6 @@ public class CleanCommand extends ClientCommand {
try {
dfs.removeDrillFile(archiveName);
System.out.println(" Removed");
- ;
} catch (DfsFacadeException e) {
System.out.println();
System.err.println(e.getMessage());
diff --git a/drill-yarn/src/test/java/org/apache/drill/yarn/core/TestConfig.java b/drill-yarn/src/test/java/org/apache/drill/yarn/core/TestConfig.java
index 99cb2c6..b2ec786 100644
--- a/drill-yarn/src/test/java/org/apache/drill/yarn/core/TestConfig.java
+++ b/drill-yarn/src/test/java/org/apache/drill/yarn/core/TestConfig.java
@@ -81,7 +81,7 @@ public class TestConfig {
try {
return file.toURI().toURL();
} catch (MalformedURLException e) {
- ;
+ // noop
}
}
return null;
diff --git a/drill-yarn/src/test/java/org/apache/drill/yarn/scripts/ScriptUtils.java b/drill-yarn/src/test/java/org/apache/drill/yarn/scripts/ScriptUtils.java
index 8a909a5..f9d2ee3 100644
--- a/drill-yarn/src/test/java/org/apache/drill/yarn/scripts/ScriptUtils.java
+++ b/drill-yarn/src/test/java/org/apache/drill/yarn/scripts/ScriptUtils.java
@@ -809,7 +809,7 @@ public class ScriptUtils {
}
result.analyze();
} catch (FileNotFoundException e) {
- ;
+ // noop
}
}
diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/ExecConstants.java b/exec/java-exec/src/main/java/org/apache/drill/exec/ExecConstants.java
index f5556cf..9bec393 100644
--- a/exec/java-exec/src/main/java/org/apache/drill/exec/ExecConstants.java
+++ b/exec/java-exec/src/main/java/org/apache/drill/exec/ExecConstants.java
@@ -49,7 +49,7 @@ public final class ExecConstants {
public static final String ZK_REFRESH = "drill.exec.zk.refresh";
public static final String BIT_RETRY_TIMES = "drill.exec.rpc.bit.server.retry.count";
public static final String BIT_RETRY_DELAY = "drill.exec.rpc.bit.server.retry.delay";
- public static final String BIT_TIMEOUT = "drill.exec.bit.timeout" ;
+ public static final String BIT_TIMEOUT = "drill.exec.bit.timeout";
public static final String SERVICE_NAME = "drill.exec.cluster-id";
public static final String INITIAL_BIT_PORT = "drill.exec.rpc.bit.server.port";
public static final String INITIAL_DATA_PORT = "drill.exec.rpc.bit.server.dataport";
diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/cache/SerializationDefinition.java b/exec/java-exec/src/main/java/org/apache/drill/exec/cache/SerializationDefinition.java
index 84c37ba..8fcb9e9 100644
--- a/exec/java-exec/src/main/java/org/apache/drill/exec/cache/SerializationDefinition.java
+++ b/exec/java-exec/src/main/java/org/apache/drill/exec/cache/SerializationDefinition.java
@@ -29,8 +29,7 @@ public enum SerializationDefinition {
STORAGE_PLUGINS(3003, StoragePlugins.class),
FRAGMENT_STATUS(3004, FragmentStatus.class),
FRAGMENT_HANDLE(3005, FragmentHandle.class),
- PLAN_FRAGMENT(3006, PlanFragment.class)
- ;
+ PLAN_FRAGMENT(3006, PlanFragment.class);
public final int id;
public final Class<?> clazz;
diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/client/DrillClient.java b/exec/java-exec/src/main/java/org/apache/drill/exec/client/DrillClient.java
index ccd7e36..bca59dc 100644
--- a/exec/java-exec/src/main/java/org/apache/drill/exec/client/DrillClient.java
+++ b/exec/java-exec/src/main/java/org/apache/drill/exec/client/DrillClient.java
@@ -824,7 +824,7 @@ public class DrillClient implements Closeable, ConnectionThrottle {
private class ListHoldingResultsListener implements UserResultsListener {
private final Vector<QueryDataBatch> results = new Vector<>();
private final SettableFuture<List<QueryDataBatch>> future = SettableFuture.create();
- private final UserProtos.RunQuery query ;
+ private final UserProtos.RunQuery query;
public ListHoldingResultsListener(UserProtos.RunQuery query) {
logger.debug( "Listener created for query \"\"\"{}\"\"\"", query );
diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/client/DumpCat.java b/exec/java-exec/src/main/java/org/apache/drill/exec/client/DumpCat.java
index ff0d0b2..7458c22 100644
--- a/exec/java-exec/src/main/java/org/apache/drill/exec/client/DumpCat.java
+++ b/exec/java-exec/src/main/java/org/apache/drill/exec/client/DumpCat.java
@@ -180,7 +180,7 @@ public class DumpCat {
aggBatchMetaInfo.add(getBatchMetaInfo(vcSerializable));
if (vectorContainer.getRecordCount() == 0) {
- emptyBatchNum ++;
+ emptyBatchNum++;
}
if (prevSchema != null && !vectorContainer.getSchema().equals(prevSchema)) {
@@ -188,7 +188,7 @@ public class DumpCat {
}
prevSchema = vectorContainer.getSchema();
- batchNum ++;
+ batchNum++;
vectorContainer.zeroVectors();
}
@@ -219,7 +219,7 @@ public class DumpCat {
VectorAccessibleSerializable vcSerializable = null;
- while (input.available() > 0 && batchNum ++ < targetBatchNum) {
+ while (input.available() > 0 && batchNum++ < targetBatchNum) {
vcSerializable = new VectorAccessibleSerializable(DumpCat.allocator);
vcSerializable.readFromStream(input);
diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/compile/sig/CodeGeneratorMethod.java b/exec/java-exec/src/main/java/org/apache/drill/exec/compile/sig/CodeGeneratorMethod.java
index ebb31f4..f15c01a 100644
--- a/exec/java-exec/src/main/java/org/apache/drill/exec/compile/sig/CodeGeneratorMethod.java
+++ b/exec/java-exec/src/main/java/org/apache/drill/exec/compile/sig/CodeGeneratorMethod.java
@@ -58,7 +58,7 @@ public class CodeGeneratorMethod implements Iterable<CodeGeneratorArgument> {
throw new RuntimeException(String.format("Unexpected number of parameter names %s. Expected %s on method %s.", Arrays.toString(parameterNames), Arrays.toString(types), m.toGenericString()));
}
arguments = new CodeGeneratorArgument[parameterNames.length];
- for (int i = 0 ; i < parameterNames.length; i++) {
+ for (int i = 0; i < parameterNames.length; i++) {
arguments[i] = new CodeGeneratorArgument(parameterNames[i], types[i]);
}
exs = m.getExceptionTypes();
diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/dotdrill/DotDrillType.java b/exec/java-exec/src/main/java/org/apache/drill/exec/dotdrill/DotDrillType.java
index e94d9f8..a8b5f4b 100644
--- a/exec/java-exec/src/main/java/org/apache/drill/exec/dotdrill/DotDrillType.java
+++ b/exec/java-exec/src/main/java/org/apache/drill/exec/dotdrill/DotDrillType.java
@@ -21,10 +21,10 @@ import org.apache.hadoop.fs.FileStatus;
import org.apache.hadoop.fs.Path;
public enum DotDrillType {
- VIEW
+ VIEW;
// ,FORMAT
// ,STATS
- ;
+
private final String ending;
@@ -70,4 +70,4 @@ public enum DotDrillType {
b.append("}.drill");
DOT_DRILL_GLOB = b.toString();
}
-}
\ No newline at end of file
+}
diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/expr/EvaluationVisitor.java b/exec/java-exec/src/main/java/org/apache/drill/exec/expr/EvaluationVisitor.java
index 4486972..373cdf6 100644
--- a/exec/java-exec/src/main/java/org/apache/drill/exec/expr/EvaluationVisitor.java
+++ b/exec/java-exec/src/main/java/org/apache/drill/exec/expr/EvaluationVisitor.java
@@ -778,7 +778,7 @@ public class EvaluationVisitor {
setBlock.assign(out.getValue(), JExpr.lit(1));
} else {
assert (e == null);
- eval.assign(out.getValue(), JExpr.lit(1)) ;
+ eval.assign(out.getValue(), JExpr.lit(1));
}
generator.unNestEvalBlock(); // exit from nested block
@@ -841,7 +841,7 @@ public class EvaluationVisitor {
setBlock.assign(out.getValue(), JExpr.lit(0));
} else {
assert (e == null);
- eval.assign(out.getValue(), JExpr.lit(0)) ;
+ eval.assign(out.getValue(), JExpr.lit(0));
}
generator.unNestEvalBlock(); // exit from nested block.
diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/expr/SizedJBlock.java b/exec/java-exec/src/main/java/org/apache/drill/exec/expr/SizedJBlock.java
index fdc4697..6263dec 100644
--- a/exec/java-exec/src/main/java/org/apache/drill/exec/expr/SizedJBlock.java
+++ b/exec/java-exec/src/main/java/org/apache/drill/exec/expr/SizedJBlock.java
@@ -42,7 +42,7 @@ public class SizedJBlock {
}
public void incCounter() {
- this.count ++;
+ this.count++;
}
public int getCount() {
diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/expr/fn/DrillAggFuncHolder.java b/exec/java-exec/src/main/java/org/apache/drill/exec/expr/fn/DrillAggFuncHolder.java
index 1a5df67..5963892 100644
--- a/exec/java-exec/src/main/java/org/apache/drill/exec/expr/fn/DrillAggFuncHolder.java
+++ b/exec/java-exec/src/main/java/org/apache/drill/exec/expr/fn/DrillAggFuncHolder.java
@@ -93,13 +93,13 @@ class DrillAggFuncHolder extends DrillFuncHolder {
//Loop through all workspace vectors, to get the minimum of size of all workspace vectors.
JVar sizeVar = setupBlock.decl(g.getModel().INT, "vectorSize", JExpr.lit(Integer.MAX_VALUE));
JClass mathClass = g.getModel().ref(Math.class);
- for (int id = 0; id < getWorkspaceVars().length; id ++) {
+ for (int id = 0; id < getWorkspaceVars().length; id++) {
if (!getWorkspaceVars()[id].isInject()) {
setupBlock.assign(sizeVar,mathClass.staticInvoke("min").arg(sizeVar).arg(g.getWorkspaceVectors().get(getWorkspaceVars()[id]).invoke("getValueCapacity")));
}
}
- for(int i =0 ; i < getWorkspaceVars().length; i++) {
+ for (int i = 0; i < getWorkspaceVars().length; i++) {
if (!getWorkspaceVars()[i].isInject()) {
setupBlock.assign(workspaceJVars[i], JExpr._new(g.getHolderType(getWorkspaceVars()[i].getMajorType())));
}
@@ -156,7 +156,7 @@ class DrillAggFuncHolder extends DrillFuncHolder {
private JVar[] declareWorkspaceVectors(ClassGenerator<?> g) {
JVar[] workspaceJVars = new JVar[getWorkspaceVars().length];
- for(int i =0 ; i < getWorkspaceVars().length; i++){
+ for (int i = 0; i < getWorkspaceVars().length; i++) {
if (getWorkspaceVars()[i].isInject()) {
workspaceJVars[i] = g.declareClassField("work", g.getModel()._ref(getWorkspaceVars()[i].getType()));
g.getBlock(BlockType.SETUP).assign(workspaceJVars[i], g.getMappingSet().getIncoming().invoke("getContext").invoke("getManagedBuffer"));
diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/expr/fn/impl/Hash32Functions.java b/exec/java-exec/src/main/java/org/apache/drill/exec/expr/fn/impl/Hash32Functions.java
index 412674f..4789a48 100644
--- a/exec/java-exec/src/main/java/org/apache/drill/exec/expr/fn/impl/Hash32Functions.java
+++ b/exec/java-exec/src/main/java/org/apache/drill/exec/expr/fn/impl/Hash32Functions.java
@@ -346,7 +346,7 @@ public class Hash32Functions {
}
}
- @FunctionTemplate(names = {"hash", "hash32" ,"hash32AsDouble"}, scope = FunctionScope.SIMPLE, nulls = FunctionTemplate.NullHandling.INTERNAL)
+ @FunctionTemplate(names = {"hash", "hash32", "hash32AsDouble"}, scope = FunctionScope.SIMPLE, nulls = FunctionTemplate.NullHandling.INTERNAL)
public static class NullableTimeHash implements DrillSimpleFunc {
@Param NullableTimeHolder in;
@Output IntHolder out;
diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/expr/fn/impl/StringFunctions.java b/exec/java-exec/src/main/java/org/apache/drill/exec/expr/fn/impl/StringFunctions.java
index 4309848..c23a1a2 100644
--- a/exec/java-exec/src/main/java/org/apache/drill/exec/expr/fn/impl/StringFunctions.java
+++ b/exec/java-exec/src/main/java/org/apache/drill/exec/expr/fn/impl/StringFunctions.java
@@ -498,7 +498,7 @@ public class StringFunctions{
for (int id = input.start; id < input.end; id++) {
byte currentByte = input.buffer.getByte(id);
- out.buffer.setByte(id - input.start, Character.toLowerCase(currentByte)) ;
+ out.buffer.setByte(id - input.start, Character.toLowerCase(currentByte));
}
}
}
@@ -529,7 +529,7 @@ public class StringFunctions{
for (int id = input.start; id < input.end; id++) {
byte currentByte = input.buffer.getByte(id);
- out.buffer.setByte(id - input.start, Character.toUpperCase(currentByte)) ;
+ out.buffer.setByte(id - input.start, Character.toUpperCase(currentByte));
}
}
}
@@ -711,9 +711,9 @@ public class StringFunctions{
final int charCount = org.apache.drill.exec.expr.fn.impl.StringFunctionUtil.getUTF8CharLength(string.buffer, string.start, string.end);
final int charLen;
if (length.value > 0) {
- charLen = Math.min((int)length.value, charCount); //left('abc', 5) -> 'abc'
+ charLen = Math.min((int) length.value, charCount); //left('abc', 5) -> 'abc'
} else if (length.value < 0) {
- charLen = Math.max(0, charCount + (int)length.value) ; // left('abc', -5) ==> ''
+ charLen = Math.max(0, charCount + (int) length.value); // left('abc', -5) ==> ''
} else {
charLen = 0;
}
@@ -830,7 +830,7 @@ public class StringFunctions{
if (j == from.end ) {
//find a true match ("from" is not empty), copy entire "to" string to out buffer
- for (int k = to.start ; k < to.end; k++) {
+ for (int k = to.start; k < to.end; k++) {
out.buffer.setByte(out.end++, to.buffer.getByte(k));
}
@@ -921,7 +921,7 @@ public class StringFunctions{
(currentByte & 0xE0) == 0xC0 || // 2-byte char. First byte is 110xxxxx
(currentByte & 0xF0) == 0xE0 || // 3-byte char. First byte is 1110xxxx
(currentByte & 0xF8) == 0xF0) { //4-byte char. First byte is 11110xxx
- count ++; //Advance the counter, since we find one char.
+ count++; //Advance the counter, since we find one char.
}
out.buffer.setByte(out.end++, currentByte);
}
@@ -1070,7 +1070,7 @@ public class StringFunctions{
(currentByte & 0xE0) == 0xC0 || // 2-byte char. First byte is 110xxxxx
(currentByte & 0xF0) == 0xE0 || // 3-byte char. First byte is 1110xxxx
(currentByte & 0xF8) == 0xF0) { //4-byte char. First byte is 11110xxx
- count ++; //Advance the counter, since we find one char.
+ count++; //Advance the counter, since we find one char.
}
out.buffer.setByte(out.end++, currentByte);
}
@@ -1429,7 +1429,7 @@ public class StringFunctions{
@Override
public void eval() {
- out.buffer = buffer = buffer.reallocIfNeeded( (left.end - left.start) + (right.end - right.start));
+ out.buffer = buffer = buffer.reallocIfNeeded((left.end - left.start) + (right.end - right.start));
out.start = out.end = 0;
int id = 0;
@@ -1493,7 +1493,7 @@ public class StringFunctions{
@Override
public void eval() {
- out.buffer = buffer = buffer.reallocIfNeeded( (left.end - left.start) + (right.end - right.start));;
+ out.buffer = buffer = buffer.reallocIfNeeded((left.end - left.start) + (right.end - right.start));
out.start = out.end = 0;
int id = 0;
diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/expr/fn/impl/conv/JsonConvertTo.java b/exec/java-exec/src/main/java/org/apache/drill/exec/expr/fn/impl/conv/JsonConvertTo.java
index e384892..1bb2512 100644
--- a/exec/java-exec/src/main/java/org/apache/drill/exec/expr/fn/impl/conv/JsonConvertTo.java
+++ b/exec/java-exec/src/main/java/org/apache/drill/exec/expr/fn/impl/conv/JsonConvertTo.java
@@ -44,7 +44,8 @@ public class JsonConvertTo {
private JsonConvertTo(){}
- @FunctionTemplate(names = { "convert_toJSON", "convert_toSIMPLEJSON" } , scope = FunctionScope.SIMPLE, nulls = NullHandling.NULL_IF_NULL,
+ @FunctionTemplate(names = { "convert_toJSON", "convert_toSIMPLEJSON" },
+ scope = FunctionScope.SIMPLE, nulls = NullHandling.NULL_IF_NULL,
outputWidthCalculatorType = FunctionTemplate.OutputWidthCalculatorType.CUSTOM_FIXED_WIDTH_DEFUALT)
public static class ConvertToJson implements DrillSimpleFunc{
diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/expr/fn/registry/FunctionRegistryHolder.java b/exec/java-exec/src/main/java/org/apache/drill/exec/expr/fn/registry/FunctionRegistryHolder.java
index e127391..c07e6ed 100644
--- a/exec/java-exec/src/main/java/org/apache/drill/exec/expr/fn/registry/FunctionRegistryHolder.java
+++ b/exec/java-exec/src/main/java/org/apache/drill/exec/expr/fn/registry/FunctionRegistryHolder.java
@@ -316,7 +316,7 @@ public class FunctionRegistryHolder {
final String functionName = function.getName();
Queue<String> jarFunctions = jar.get(functionName);
if (jarFunctions == null) {
- jarFunctions = Queues.newConcurrentLinkedQueue();;
+ jarFunctions = Queues.newConcurrentLinkedQueue();
jar.put(functionName, jarFunctions);
}
final String functionSignature = function.getSignature();
diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/physical/config/HashAggregate.java b/exec/java-exec/src/main/java/org/apache/drill/exec/physical/config/HashAggregate.java
index 51f34a0..da988de 100644
--- a/exec/java-exec/src/main/java/org/apache/drill/exec/physical/config/HashAggregate.java
+++ b/exec/java-exec/src/main/java/org/apache/drill/exec/physical/config/HashAggregate.java
@@ -104,6 +104,6 @@ public class HashAggregate extends AbstractSingle {
public boolean isBufferedOperator(QueryContext queryContext) {
// In case forced to use a single partition - do not consider this a buffered op (when memory is divided)
return queryContext == null ||
- 1 < (int)queryContext.getOptions().getOption(ExecConstants.HASHAGG_NUM_PARTITIONS_VALIDATOR) ;
+ 1 < (int) queryContext.getOptions().getOption(ExecConstants.HASHAGG_NUM_PARTITIONS_VALIDATOR);
}
}
diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/physical/config/HashJoinPOP.java b/exec/java-exec/src/main/java/org/apache/drill/exec/physical/config/HashJoinPOP.java
index 48d977e..146bb7f 100644
--- a/exec/java-exec/src/main/java/org/apache/drill/exec/physical/config/HashJoinPOP.java
+++ b/exec/java-exec/src/main/java/org/apache/drill/exec/physical/config/HashJoinPOP.java
@@ -88,6 +88,6 @@ public class HashJoinPOP extends AbstractJoinPop {
public boolean isBufferedOperator(QueryContext queryContext) {
// In case forced to use a single partition - do not consider this a buffered op (when memory is divided)
return queryContext == null ||
- 1 < (int)queryContext.getOptions().getOption(ExecConstants.HASHJOIN_NUM_PARTITIONS_VALIDATOR) ;
+ 1 < (int) queryContext.getOptions().getOption(ExecConstants.HASHJOIN_NUM_PARTITIONS_VALIDATOR);
}
}
diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/aggregate/HashAggTemplate.java b/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/aggregate/HashAggTemplate.java
index 4bbfa05..1954c79 100644
--- a/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/aggregate/HashAggTemplate.java
+++ b/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/aggregate/HashAggTemplate.java
@@ -154,7 +154,13 @@ public abstract class HashAggTemplate implements HashAggregator {
private int cycleNum = 0; // primary, secondary, tertiary, etc.
private int originalPartition = -1; // the partition a secondary reads from
- private static class SpilledPartition { public int spilledBatches; public String spillFile; int cycleNum; int origPartn; int prevOrigPartn; }
+ private static class SpilledPartition {
+ public int spilledBatches;
+ public String spillFile;
+ int cycleNum;
+ int origPartn;
+ int prevOrigPartn;
+ }
private ArrayList<SpilledPartition> spilledPartitionsList;
private int operatorId; // for the spill file name
@@ -193,7 +199,6 @@ public abstract class HashAggTemplate implements HashAggregator {
AVG_OUTPUT_BATCH_BYTES,
AVG_OUTPUT_ROW_BYTES,
OUTPUT_RECORD_COUNT;
- ;
@Override
public int metricId() {
@@ -333,7 +338,7 @@ public abstract class HashAggTemplate implements HashAggregator {
is2ndPhase = hashAggrConfig.getAggPhase() == AggPrelBase.OperatorPhase.PHASE_2of2;
isTwoPhase = hashAggrConfig.getAggPhase() != AggPrelBase.OperatorPhase.PHASE_1of1;
- is1stPhase = isTwoPhase && ! is2ndPhase ;
+ is1stPhase = isTwoPhase && !is2ndPhase;
canSpill = isTwoPhase; // single phase can not spill
// Typically for testing - force a spill after a partition has more than so many batches
@@ -410,8 +415,8 @@ public abstract class HashAggTemplate implements HashAggregator {
updateEstMaxBatchSize(incoming);
}
// create "reserved memory" and adjust the memory limit down
- reserveValueBatchMemory = reserveOutgoingMemory = estValuesBatchSize ;
- long newMemoryLimit = allocator.getLimit() - reserveValueBatchMemory - reserveOutgoingMemory ;
+ reserveValueBatchMemory = reserveOutgoingMemory = estValuesBatchSize;
+ long newMemoryLimit = allocator.getLimit() - reserveValueBatchMemory - reserveOutgoingMemory;
long memAvail = newMemoryLimit - allocator.getAllocatedMemory();
if ( memAvail <= 0 ) { throw new OutOfMemoryException("Too little memory available"); }
allocator.setLimit(newMemoryLimit);
@@ -458,9 +463,9 @@ public abstract class HashAggTemplate implements HashAggregator {
bitsInMask = Integer.bitCount(partitionMask); // e.g. 0x1F -> 5
// Create arrays (one entry per partition)
- htables = new HashTable[numPartitions] ;
- batchHolders = (ArrayList<BatchHolder>[]) new ArrayList<?>[numPartitions] ;
- outBatchIndex = new int[numPartitions] ;
+ htables = new HashTable[numPartitions];
+ batchHolders = (ArrayList<BatchHolder>[]) new ArrayList<?>[numPartitions];
+ outBatchIndex = new int[numPartitions];
writers = new Writer[numPartitions];
spilledBatchesCount = new int[numPartitions];
spillFiles = new String[numPartitions];
@@ -486,7 +491,11 @@ public abstract class HashAggTemplate implements HashAggregator {
this.batchHolders[i] = new ArrayList<BatchHolder>(); // First BatchHolder is created when the first put request is received.
}
// Initialize the value vectors in the generated code (which point to the incoming or outgoing fields)
- try { htables[0].updateBatches(); } catch (SchemaChangeException sc) { throw new UnsupportedOperationException(sc); };
+ try {
+ htables[0].updateBatches();
+ } catch (SchemaChangeException sc) {
+ throw new UnsupportedOperationException(sc);
+ }
}
/**
* get new incoming: (when reading spilled files like an "incoming")
@@ -689,7 +698,7 @@ public abstract class HashAggTemplate implements HashAggregator {
// Either flag buildComplete or handleEmit (or earlyOutput) would cause returning of
// the outgoing batch downstream (see innerNext() in HashAggBatch).
- buildComplete = true ; // now should go and return outgoing
+ buildComplete = true; // now should go and return outgoing
if ( handleEmit ) {
buildComplete = false; // This was not a real NONE - more incoming is expected
@@ -939,7 +948,7 @@ public abstract class HashAggTemplate implements HashAggregator {
}
// Give the current (if already spilled) some priority
if ( ! tryAvoidCurr && isSpilled(currPart) && ( currPartSize + 1 >= maxSizeSpilled )) {
- maxSizeSpilled = currPartSize ;
+ maxSizeSpilled = currPartSize;
indexMaxSpilled = currPart;
}
// now find the largest non-spilled partition
@@ -948,7 +957,7 @@ public abstract class HashAggTemplate implements HashAggregator {
// Use the largest spilled (if found) as a base line, with a factor of 4
if ( indexMaxSpilled > -1 && maxSizeSpilled > 1 ) {
indexMax = indexMaxSpilled;
- maxSize = 4 * maxSizeSpilled ;
+ maxSize = 4 * maxSizeSpilled;
}
for ( int insp = 0; insp < numPartitions; insp++) {
if ( ! isSpilled(insp) && maxSize < batchHolders[insp].size() ) {
@@ -1159,7 +1168,11 @@ public abstract class HashAggTemplate implements HashAggregator {
originalPartition = sp.origPartn; // used for the filename
logger.trace("Reading back spilled original partition {} as an incoming",originalPartition);
// Initialize .... new incoming, new set of partitions
- try { initializeSetup(newIncoming); } catch (Exception e) { throw new RuntimeException(e); }
+ try {
+ initializeSetup(newIncoming);
+ } catch (Exception e) {
+ throw new RuntimeException(e);
+ }
// update the cycle num if needed
// The current cycle num should always be one larger than in the spilled partition
if ( cycleNum == sp.cycleNum ) {
@@ -1179,7 +1192,7 @@ public abstract class HashAggTemplate implements HashAggregator {
return AggIterOutcome.AGG_RESTART;
}
- partitionToReturn = nextPartitionToReturn ;
+ partitionToReturn = nextPartitionToReturn;
}
@@ -1187,7 +1200,7 @@ public abstract class HashAggTemplate implements HashAggregator {
int numPendingOutput = currPartition.get(currOutBatchIndex).getNumPendingOutput();
// The following accounting is for logging, metrics, etc.
- rowsInPartition += numPendingOutput ;
+ rowsInPartition += numPendingOutput;
if ( ! handlingSpills ) { rowsNotSpilled += numPendingOutput; }
else { rowsSpilledReturned += numPendingOutput; }
if ( earlyOutput ) { rowsReturnedEarly += numPendingOutput; }
@@ -1238,7 +1251,7 @@ public abstract class HashAggTemplate implements HashAggregator {
logger.debug("HASH AGG: Finished (early) re-init partition {}, mem allocated: {}", earlyPartition, allocator.getAllocatedMemory());
}
outBatchIndex[earlyPartition] = 0; // reset, for next time
- earlyOutput = false ; // done with early output
+ earlyOutput = false; // done with early output
}
else if ( handleEmit ) {
// When returning the last outgoing batch (following an incoming EMIT), then replace OK with EMIT
@@ -1290,9 +1303,9 @@ public abstract class HashAggTemplate implements HashAggregator {
*/
private String getOOMErrorMsg(String prefix) {
String errmsg;
- if ( !isTwoPhase ) {
- errmsg = "Single Phase Hash Aggregate operator can not spill." ;
- } else if ( ! canSpill ) { // 2nd phase, with only 1 partition
+ if (!isTwoPhase) {
+ errmsg = "Single Phase Hash Aggregate operator can not spill.";
+ } else if (!canSpill) { // 2nd phase, with only 1 partition
errmsg = "Too little memory available to operator to facilitate spilling.";
} else { // a bug ?
errmsg = prefix + " OOM at " + (is2ndPhase ? "Second Phase" : "First Phase") + ". Partitions: " + numPartitions +
@@ -1353,9 +1366,11 @@ public abstract class HashAggTemplate implements HashAggregator {
}
// right shift hash code for secondary (or tertiary...) spilling
- for (int i = 0; i < cycleNum; i++) { hashCode >>>= bitsInMask; }
+ for (int i = 0; i < cycleNum; i++) {
+ hashCode >>>= bitsInMask;
+ }
- int currentPartition = hashCode & partitionMask ;
+ int currentPartition = hashCode & partitionMask;
hashCode >>>= bitsInMask;
HashTable.PutStatus putStatus = null;
long allocatedBeforeHTput = allocator.getAllocatedMemory();
@@ -1398,7 +1413,7 @@ public abstract class HashAggTemplate implements HashAggregator {
throw new UnsupportedOperationException("Unexpected schema change", e);
}
long allocatedBeforeAggCol = allocator.getAllocatedMemory();
- boolean needToCheckIfSpillIsNeeded = allocatedBeforeAggCol > allocatedBeforeHTput ;
+ boolean needToCheckIfSpillIsNeeded = allocatedBeforeAggCol > allocatedBeforeHTput;
// Add an Aggr batch if needed:
//
diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/aggregate/HashAggregator.java b/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/aggregate/HashAggregator.java
index f58be89..4c54650 100644
--- a/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/aggregate/HashAggregator.java
+++ b/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/aggregate/HashAggregator.java
@@ -44,7 +44,7 @@ public interface HashAggregator {
// For returning results from outputCurrentBatch
// OK - batch returned, NONE - end of data, RESTART - call again, EMIT - like OK but EMIT
- enum AggIterOutcome { AGG_OK, AGG_NONE, AGG_RESTART , AGG_EMIT }
+ enum AggIterOutcome { AGG_OK, AGG_NONE, AGG_RESTART, AGG_EMIT }
void setup(HashAggregate hashAggrConfig, HashTableConfig htConfig, FragmentContext context, OperatorContext oContext, RecordBatch incoming, HashAggBatch outgoing,
LogicalExpression[] valueExprs, List<TypedFieldId> valueFieldIds, TypedFieldId[] keyFieldIds, VectorContainer outContainer, int extraRowBytes) throws SchemaChangeException, IOException, ClassTransformationException;
diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/aggregate/SpilledRecordbatch.java b/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/aggregate/SpilledRecordbatch.java
index c78e2c0..7ebce2b 100644
--- a/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/aggregate/SpilledRecordbatch.java
+++ b/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/aggregate/SpilledRecordbatch.java
@@ -135,7 +135,7 @@ public class SpilledRecordbatch implements CloseableRecordBatch {
if ( spillStream == null ) {
throw new IllegalStateException("Spill stream was null");
- };
+ }
if ( spillSet.getPosition(spillStream) < 0 ) {
HashAggTemplate.logger.warn("Position is {} for stream {}", spillSet.getPosition(spillStream), spillStream.toString());
@@ -155,7 +155,7 @@ public class SpilledRecordbatch implements CloseableRecordBatch {
throw UserException.dataReadError(e).addContext("Failed reading from a spill file").build(HashAggTemplate.logger);
}
- spilledBatches-- ; // one less batch to read
+ spilledBatches--; // one less batch to read
return IterOutcome.OK;
}
diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/aggregate/StreamingAggregator.java b/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/aggregate/StreamingAggregator.java
index 2a64b93..23fdcc1 100644
--- a/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/aggregate/StreamingAggregator.java
+++ b/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/aggregate/StreamingAggregator.java
@@ -49,8 +49,7 @@ public interface StreamingAggregator {
RETURN_OUTCOME,
CLEANUP_AND_RETURN,
UPDATE_AGGREGATOR,
- RETURN_AND_RESET
- ;
+ RETURN_AND_RESET;
}
public abstract void setup(OperatorContext context, RecordBatch incoming, StreamingAggBatch outgoing) throws SchemaChangeException;
diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/common/HashTableStats.java b/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/common/HashTableStats.java
index 8f0b7ab..8c93f1b 100644
--- a/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/common/HashTableStats.java
+++ b/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/common/HashTableStats.java
@@ -28,10 +28,10 @@ public class HashTableStats {
}
public void addStats (HashTableStats newStats) {
- this.numBuckets += newStats.numBuckets ;
- this.numEntries += newStats.numEntries ;
- this.numResizing += newStats.numResizing ;
- this.resizingTime += newStats.resizingTime ;
+ this.numBuckets += newStats.numBuckets;
+ this.numEntries += newStats.numEntries;
+ this.numResizing += newStats.numResizing;
+ this.resizingTime += newStats.resizingTime;
}
}
diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/common/HashTableTemplate.java b/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/common/HashTableTemplate.java
index 83b72d7..3c418b9 100644
--- a/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/common/HashTableTemplate.java
+++ b/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/common/HashTableTemplate.java
@@ -476,7 +476,7 @@ public abstract class HashTableTemplate implements HashTable {
if (tableSize > MAXIMUM_CAPACITY) {
tableSize = MAXIMUM_CAPACITY;
}
- originalTableSize = tableSize ; // retain original size
+ originalTableSize = tableSize; // retain original size
threshold = (int) Math.ceil(tableSize * loadf);
diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/flatten/FlattenTemplate.java b/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/flatten/FlattenTemplate.java
index cd58bfd..365451c 100644
--- a/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/flatten/FlattenTemplate.java
+++ b/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/flatten/FlattenTemplate.java
@@ -113,7 +113,7 @@ public abstract class FlattenTemplate implements Flattener {
} catch (OversizedAllocationException ex) {
// unable to flatten due to a soft buffer overflow. split the batch here and resume execution.
logger.debug("Reached allocation limit. Splitting the batch at input index: {} - inner index: {} - current completed index: {}",
- valueIndexLocal, innerValueIndexLocal, currentInnerValueIndexLocal) ;
+ valueIndexLocal, innerValueIndexLocal, currentInnerValueIndexLocal);
/*
* TODO
diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/join/HashJoinBatch.java b/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/join/HashJoinBatch.java
index d4d4f92..b1ea96f 100644
--- a/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/join/HashJoinBatch.java
+++ b/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/join/HashJoinBatch.java
@@ -776,7 +776,7 @@ public class HashJoinBatch extends AbstractBinaryRecordBatch<HashJoinPOP> {
for (int ind = 0; ind < currentRecordCount; ind++) {
int hashCode = ( cycleNum == 0 ) ? partitions[0].getBuildHashCode(ind)
: read_right_HV_vector.getAccessor().get(ind); // get the hash value from the HV column
- int currPart = hashCode & partitionMask ;
+ int currPart = hashCode & partitionMask;
hashCode >>>= bitsInMask;
// Append the new inner row to the appropriate partition; spill (that partition) if needed
partitions[currPart].appendInnerRow(buildBatch.getContainer(), ind, hashCode, buildCalc); // may spill if needed
diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/join/HashJoinProbeTemplate.java b/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/join/HashJoinProbeTemplate.java
index 46f2fa3..639f757 100644
--- a/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/join/HashJoinProbeTemplate.java
+++ b/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/join/HashJoinProbeTemplate.java
@@ -162,7 +162,7 @@ public abstract class HashJoinProbeTemplate implements HashJoinProbe {
*/
private int appendBuild(VectorContainer buildSrcContainer, int buildSrcIndex) {
// "- 1" to skip the last "hash values" added column
- int lastColIndex = buildSrcContainer.getNumberOfColumns() - 1 ;
+ int lastColIndex = buildSrcContainer.getNumberOfColumns() - 1;
for (int vectorIndex = 0; vectorIndex < lastColIndex; vectorIndex++) {
ValueVector destVector = container.getValueVector(vectorIndex).getValueVector();
ValueVector srcVector = buildSrcContainer.getValueVector(vectorIndex).getValueVector();
@@ -292,7 +292,7 @@ public abstract class HashJoinProbeTemplate implements HashJoinProbe {
int hashCode = ( cycleNum == 0 ) ?
partitions[0].getProbeHashCode(recordsProcessed)
: read_left_HV_vector.getAccessor().get(recordsProcessed);
- int currBuildPart = hashCode & partitionMask ;
+ int currBuildPart = hashCode & partitionMask;
hashCode >>>= bitsInMask;
// Set and keep the current partition (may be used again on subsequent probe calls as
diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/join/JoinUtils.java b/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/join/JoinUtils.java
index b974537..52871e2 100644
--- a/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/join/JoinUtils.java
+++ b/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/join/JoinUtils.java
@@ -235,7 +235,7 @@ public class JoinUtils {
if (currentrel instanceof DrillAggregateRel) {
agg = (DrillAggregateRel)currentrel;
} else if (currentrel instanceof RelSubset) {
- currentrel = ((RelSubset) currentrel).getBest() ;
+ currentrel = ((RelSubset) currentrel).getBest();
} else if (currentrel instanceof DrillLimitRel) {
// TODO: Improve this check when DRILL-5691 is fixed.
// The problem is that RelMdMaxRowCount currently cannot be used
diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/join/LateralJoinBatch.java b/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/join/LateralJoinBatch.java
index ff33e2f..0eb15a7 100644
--- a/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/join/LateralJoinBatch.java
+++ b/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/join/LateralJoinBatch.java
@@ -626,8 +626,8 @@ public class LateralJoinBatch extends AbstractBinaryRecordBatch<LateralJoinPOP>
if (leftUpstream == EMIT || leftUpstream == OK_NEW_SCHEMA) {
break;
} else {
- logger.debug("Output batch still has some space left, getting new batches from left and right. OutIndex: {}"
- , outputIndex);
+ logger.debug("Output batch still has some space left, getting new batches from left and right. OutIndex: {}",
+ outputIndex);
// Get both left batch and the right batch and make sure indexes are properly set
leftUpstream = processLeftBatch();
diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/partitionsender/PartitionSenderRootExec.java b/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/partitionsender/PartitionSenderRootExec.java
index 034d6c2..5049b06 100644
--- a/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/partitionsender/PartitionSenderRootExec.java
+++ b/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/partitionsender/PartitionSenderRootExec.java
@@ -269,7 +269,7 @@ public class PartitionSenderRootExec extends BaseRootExec {
// set up partitioning function
final LogicalExpression expr = operator.getExpr();
final ErrorCollector collector = new ErrorCollectorImpl();
- final ClassGenerator<Partitioner> cg ;
+ final ClassGenerator<Partitioner> cg;
cg = CodeGenerator.getRoot(Partitioner.TEMPLATE_DEFINITION, context.getOptions());
cg.getCodeGenerator().plainJavaCapable(true);
diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/unnest/UnnestImpl.java b/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/unnest/UnnestImpl.java
index 02d2f18..f0816b6 100644
--- a/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/unnest/UnnestImpl.java
+++ b/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/unnest/UnnestImpl.java
@@ -54,7 +54,7 @@ public class UnnestImpl implements Unnest {
private RepeatedValueVector.RepeatedAccessor accessor;
private RecordBatch outgoing;
- private IntVector rowIdVector ; // Allocated and owned by the UnnestRecordBatch
+ private IntVector rowIdVector; // Allocated and owned by the UnnestRecordBatch
private IntVector.Mutator rowIdVectorMutator;
/**
diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/xsort/managed/SortMemoryManager.java b/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/xsort/managed/SortMemoryManager.java
index 68b546b..5d2721a 100644
--- a/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/xsort/managed/SortMemoryManager.java
+++ b/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/xsort/managed/SortMemoryManager.java
@@ -266,7 +266,7 @@ public class SortMemoryManager {
memoryLimit = (configMemoryLimit == 0) ? opMemoryLimit
: Math.min(opMemoryLimit, configMemoryLimit);
- preferredSpillBatchSize = config.spillBatchSize();;
+ preferredSpillBatchSize = config.spillBatchSize();
preferredMergeBatchSize = config.mergeBatchSize();
// Initialize the buffer memory limit for the first batch.
diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/physical/rowSet/impl/ColumnState.java b/exec/java-exec/src/main/java/org/apache/drill/exec/physical/rowSet/impl/ColumnState.java
index acc9556..24e270a 100644
--- a/exec/java-exec/src/main/java/org/apache/drill/exec/physical/rowSet/impl/ColumnState.java
+++ b/exec/java-exec/src/main/java/org/apache/drill/exec/physical/rowSet/impl/ColumnState.java
@@ -347,8 +347,7 @@ public abstract class ColumnState {
.attribute("addVersion", addVersion)
.attribute("state", state)
.attributeIdentity("writer", writer)
- .attribute("vectorState")
- ;
+ .attribute("vectorState");
vectorState.dump(format);
format.endObject();
}
diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/physical/rowSet/impl/ResultSetLoaderImpl.java b/exec/java-exec/src/main/java/org/apache/drill/exec/physical/rowSet/impl/ResultSetLoaderImpl.java
index c7c6fdc..2783c5b 100644
--- a/exec/java-exec/src/main/java/org/apache/drill/exec/physical/rowSet/impl/ResultSetLoaderImpl.java
+++ b/exec/java-exec/src/main/java/org/apache/drill/exec/physical/rowSet/impl/ResultSetLoaderImpl.java
@@ -795,8 +795,7 @@ public class ResultSetLoaderImpl implements ResultSetLoader, LoaderInternals {
.attribute("activeSchemaVersion", activeSchemaVersion)
.attribute("harvestSchemaVersion", harvestSchemaVersion)
.attribute("pendingRowCount", pendingRowCount)
- .attribute("targetRowCount", targetRowCount)
- ;
+ .attribute("targetRowCount", targetRowCount);
format.attribute("root");
rootState.dump(format);
format.attribute("rootWriter");
diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/planner/AbstractPartitionDescriptor.java b/exec/java-exec/src/main/java/org/apache/drill/exec/planner/AbstractPartitionDescriptor.java
index b70c764..cf01256 100644
--- a/exec/java-exec/src/main/java/org/apache/drill/exec/planner/AbstractPartitionDescriptor.java
+++ b/exec/java-exec/src/main/java/org/apache/drill/exec/planner/AbstractPartitionDescriptor.java
@@ -44,7 +44,7 @@ public abstract class AbstractPartitionDescriptor implements PartitionDescriptor
* Create sublists of the partition locations, each sublist of size
* at most {@link PartitionDescriptor#PARTITION_BATCH_SIZE}
*/
- protected abstract void createPartitionSublists() ;
+ protected abstract void createPartitionSublists();
/**
* Iterator that traverses over the super list of partition locations and
diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/planner/StarColumnHelper.java b/exec/java-exec/src/main/java/org/apache/drill/exec/planner/StarColumnHelper.java
index 216c8d2..4cd9ed3 100644
--- a/exec/java-exec/src/main/java/org/apache/drill/exec/planner/StarColumnHelper.java
+++ b/exec/java-exec/src/main/java/org/apache/drill/exec/planner/StarColumnHelper.java
@@ -65,7 +65,7 @@ public class StarColumnHelper {
}
public static boolean isPrefixedStarColumn(String fieldName) {
- return fieldName.indexOf(PREFIXED_STAR_COLUMN) > 0 ; // the delimiter * starts at none-zero position.
+ return fieldName.indexOf(PREFIXED_STAR_COLUMN) > 0; // the delimiter * starts at none-zero position.
}
public static boolean isNonPrefixedStarColumn(String fieldName) {
diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/planner/common/DrillJoinRelBase.java b/exec/java-exec/src/main/java/org/apache/drill/exec/planner/common/DrillJoinRelBase.java
index 10c4738..862fb59 100644
--- a/exec/java-exec/src/main/java/org/apache/drill/exec/planner/common/DrillJoinRelBase.java
+++ b/exec/java-exec/src/main/java/org/apache/drill/exec/planner/common/DrillJoinRelBase.java
@@ -137,7 +137,7 @@ public abstract class DrillJoinRelBase extends Join implements DrillRelNode {
// just to make sure Cartesian Join is more expensive
// than Non-Cartesian Join.
- final int keySize = 1 ; // assume having 1 join key, when estimate join cost.
+ final int keySize = 1; // assume having 1 join key, when estimate join cost.
final DrillCostBase cost = (DrillCostBase) computeHashJoinCostWithKeySize(planner, keySize, mq).multiplyBy(mulFactor);
// Cartesian join row count will be product of two inputs. The other factors come from the above estimated DrillCost.
@@ -197,7 +197,7 @@ public abstract class DrillJoinRelBase extends Join implements DrillRelNode {
) * buildRowCount * factor;
double cpuCost = joinConditionCost * (probeRowCount) // probe size determine the join condition comparison cost
- + cpuCostBuild + cpuCostProbe ;
+ + cpuCostBuild + cpuCostProbe;
DrillCostFactory costFactory = (DrillCostFactory) planner.getCostFactory();
diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/planner/common/DrillLimitRelBase.java b/exec/java-exec/src/main/java/org/apache/drill/exec/planner/common/DrillLimitRelBase.java
index 7d070b6..fde3896 100644
--- a/exec/java-exec/src/main/java/org/apache/drill/exec/planner/common/DrillLimitRelBase.java
+++ b/exec/java-exec/src/main/java/org/apache/drill/exec/planner/common/DrillLimitRelBase.java
@@ -84,7 +84,7 @@ public abstract class DrillLimitRelBase extends SingleRel implements DrillRelNod
@Override
public double estimateRowCount(RelMetadataQuery mq) {
- int off = offset != null ? RexLiteral.intValue(offset) : 0 ;
+ int off = offset != null? RexLiteral.intValue(offset): 0;
if (fetch == null) {
// If estimated rowcount is less than offset return 0
diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/planner/common/DrillProjectRelBase.java b/exec/java-exec/src/main/java/org/apache/drill/exec/planner/common/DrillProjectRelBase.java
index b7881c2..6c1b143 100644
--- a/exec/java-exec/src/main/java/org/apache/drill/exec/planner/common/DrillProjectRelBase.java
+++ b/exec/java-exec/src/main/java/org/apache/drill/exec/planner/common/DrillProjectRelBase.java
@@ -60,7 +60,7 @@ import com.google.common.collect.Lists;
* Base class for logical and physical Project implemented in Drill
*/
public abstract class DrillProjectRelBase extends Project implements DrillRelNode {
- private final int nonSimpleFieldCount ;
+ private final int nonSimpleFieldCount;
protected DrillProjectRelBase(Convention convention, RelOptCluster cluster, RelTraitSet traits, RelNode child, List<? extends RexNode> exps,
RelDataType rowType) {
@@ -128,10 +128,10 @@ public abstract class DrillProjectRelBase extends Project implements DrillRelNod
for (RexNode expr : this.getProjects()) {
if (expr instanceof RexInputRef) {
// Simple Field reference.
- cnt ++;
+ cnt++;
} else if (expr instanceof RexCall && expr.accept(complexFieldIdentifer)) {
// Complex field with named segments only.
- cnt ++;
+ cnt++;
}
}
return cnt;
diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/planner/cost/DrillCostBase.java b/exec/java-exec/src/main/java/org/apache/drill/exec/planner/cost/DrillCostBase.java
index ba55fae..f64cd77 100644
--- a/exec/java-exec/src/main/java/org/apache/drill/exec/planner/cost/DrillCostBase.java
+++ b/exec/java-exec/src/main/java/org/apache/drill/exec/planner/cost/DrillCostBase.java
@@ -169,7 +169,7 @@ public class DrillCostBase implements DrillRelOptCost {
|| (this.io == Double.POSITIVE_INFINITY)
|| (this.network == Double.POSITIVE_INFINITY)
|| (this.rowCount == Double.POSITIVE_INFINITY)
- || (this.memory == Double.POSITIVE_INFINITY) ;
+ || (this.memory == Double.POSITIVE_INFINITY);
}
@Override
diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/planner/logical/DrillFilterRel.java b/exec/java-exec/src/main/java/org/apache/drill/exec/planner/logical/DrillFilterRel.java
index c96f4e6..7497783 100644
--- a/exec/java-exec/src/main/java/org/apache/drill/exec/planner/logical/DrillFilterRel.java
+++ b/exec/java-exec/src/main/java/org/apache/drill/exec/planner/logical/DrillFilterRel.java
@@ -51,7 +51,7 @@ public class DrillFilterRel extends DrillFilterRelBase implements DrillRel {
}
public static DrillFilterRel create(RelNode child, RexNode condition) {
- return new DrillFilterRel(child.getCluster(), child.getTraitSet(), child, condition) ;
+ return new DrillFilterRel(child.getCluster(), child.getTraitSet(), child, condition);
}
}
diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/planner/logical/DrillOptiq.java b/exec/java-exec/src/main/java/org/apache/drill/exec/planner/logical/DrillOptiq.java
index 2cd6068..5f2b02c 100644
--- a/exec/java-exec/src/main/java/org/apache/drill/exec/planner/logical/DrillOptiq.java
+++ b/exec/java-exec/src/main/java/org/apache/drill/exec/planner/logical/DrillOptiq.java
@@ -301,10 +301,15 @@ public class DrillOptiq {
case "CHAR":
castType = Types.required(MinorType.VARCHAR).toBuilder().setPrecision(call.getType().getPrecision()).build();
break;
-
- case "INTEGER": castType = Types.required(MinorType.INT); break;
- case "FLOAT": castType = Types.required(MinorType.FLOAT4); break;
- case "DOUBLE": castType = Types.required(MinorType.FLOAT8); break;
+ case "INTEGER":
+ castType = Types.required(MinorType.INT);
+ break;
+ case "FLOAT":
+ castType = Types.required(MinorType.FLOAT4);
+ break;
+ case "DOUBLE":
+ castType = Types.required(MinorType.FLOAT8);
+ break;
case "DECIMAL":
if (!context.getPlannerSettings().getOptions().getOption(PlannerSettings.ENABLE_DECIMAL_DATA_TYPE_KEY).bool_val) {
throw UserException
@@ -327,7 +332,9 @@ public class DrillOptiq {
case "INTERVAL_YEAR":
case "INTERVAL_YEAR_MONTH":
- case "INTERVAL_MONTH": castType = Types.required(MinorType.INTERVALYEAR); break;
+ case "INTERVAL_MONTH":
+ castType = Types.required(MinorType.INTERVALYEAR);
+ break;
case "INTERVAL_DAY":
case "INTERVAL_DAY_HOUR":
case "INTERVAL_DAY_MINUTE":
@@ -337,11 +344,19 @@ public class DrillOptiq {
case "INTERVAL_HOUR_SECOND":
case "INTERVAL_MINUTE":
case "INTERVAL_MINUTE_SECOND":
- case "INTERVAL_SECOND": castType = Types.required(MinorType.INTERVALDAY); break;
- case "BOOLEAN": castType = Types.required(MinorType.BIT); break;
- case "BINARY": castType = Types.required(MinorType.VARBINARY); break;
- case "ANY": return arg; // Type will be same as argument.
- default: castType = Types.required(MinorType.valueOf(call.getType().getSqlTypeName().getName()));
+ case "INTERVAL_SECOND":
+ castType = Types.required(MinorType.INTERVALDAY);
+ break;
+ case "BOOLEAN":
+ castType = Types.required(MinorType.BIT);
+ break;
+ case "BINARY":
+ castType = Types.required(MinorType.VARBINARY);
+ break;
+ case "ANY":
+ return arg; // Type will be same as argument.
+ default:
+ castType = Types.required(MinorType.valueOf(call.getType().getSqlTypeName().getName()));
}
return FunctionCallFactory.createCast(castType, ExpressionPosition.UNKNOWN, arg);
}
diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/planner/logical/DrillUnionAllRule.java b/exec/java-exec/src/main/java/org/apache/drill/exec/planner/logical/DrillUnionAllRule.java
index 293749b..69e9452 100644
--- a/exec/java-exec/src/main/java/org/apache/drill/exec/planner/logical/DrillUnionAllRule.java
+++ b/exec/java-exec/src/main/java/org/apache/drill/exec/planner/logical/DrillUnionAllRule.java
@@ -61,7 +61,7 @@ public class DrillUnionAllRule extends RelOptRule {
call.transformTo(new DrillUnionRel(union.getCluster(), traits, convertedInputs, union.all,
true /* check compatibility */));
} catch (InvalidRelException e) {
- tracer.warn(e.toString()) ;
+ tracer.warn(e.toString());
}
}
}
diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/planner/logical/DrillUnionRel.java b/exec/java-exec/src/main/java/org/apache/drill/exec/planner/logical/DrillUnionRel.java
index c94d2b5..a5a6e03 100644
--- a/exec/java-exec/src/main/java/org/apache/drill/exec/planner/logical/DrillUnionRel.java
+++ b/exec/java-exec/src/main/java/org/apache/drill/exec/planner/logical/DrillUnionRel.java
@@ -50,7 +50,7 @@ public class DrillUnionRel extends DrillUnionRelBase implements DrillRel {
return new DrillUnionRel(getCluster(), traitSet, inputs, all,
false /* don't check compatibility during copy */);
} catch (InvalidRelException e) {
- throw new AssertionError(e) ;
+ throw new AssertionError(e);
}
}
diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/planner/physical/AggPrelBase.java b/exec/java-exec/src/main/java/org/apache/drill/exec/planner/physical/AggPrelBase.java
index ca68a7d..ff640a7 100644
--- a/exec/java-exec/src/main/java/org/apache/drill/exec/planner/physical/AggPrelBase.java
+++ b/exec/java-exec/src/main/java/org/apache/drill/exec/planner/physical/AggPrelBase.java
@@ -49,7 +49,7 @@ public abstract class AggPrelBase extends DrillAggregateRelBase implements Prel
public enum OperatorPhase {PHASE_1of1, PHASE_1of2, PHASE_2of2}
- protected OperatorPhase operPhase = OperatorPhase.PHASE_1of1 ; // default phase
+ protected OperatorPhase operPhase = OperatorPhase.PHASE_1of1; // default phase
protected List<NamedExpression> keys = Lists.newArrayList();
protected List<NamedExpression> aggExprs = Lists.newArrayList();
protected List<AggregateCall> phase2AggCallList = Lists.newArrayList();
diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/planner/physical/BroadcastExchangePrel.java b/exec/java-exec/src/main/java/org/apache/drill/exec/planner/physical/BroadcastExchangePrel.java
index 2c04368..caec426 100644
--- a/exec/java-exec/src/main/java/org/apache/drill/exec/planner/physical/BroadcastExchangePrel.java
+++ b/exec/java-exec/src/main/java/org/apache/drill/exec/planner/physical/BroadcastExchangePrel.java
@@ -56,7 +56,7 @@ public class BroadcastExchangePrel extends ExchangePrel{
final double inputRows = mq.getRowCount(child);
final int rowWidth = child.getRowType().getFieldCount() * DrillCostBase.AVG_FIELD_WIDTH;
- final double cpuCost = broadcastFactor * DrillCostBase.SVR_CPU_COST * inputRows ;
+ final double cpuCost = broadcastFactor * DrillCostBase.SVR_CPU_COST * inputRows;
final double networkCost = broadcastFactor * DrillCostBase.BYTE_NETWORK_COST * inputRows * rowWidth * numEndPoints;
return new DrillCostBase(inputRows, cpuCost, 0, networkCost);
diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/planner/physical/DrillDistributionTrait.java b/exec/java-exec/src/main/java/org/apache/drill/exec/planner/physical/DrillDistributionTrait.java
index d10021c..b250d5a 100644
--- a/exec/java-exec/src/main/java/org/apache/drill/exec/planner/physical/DrillDistributionTrait.java
+++ b/exec/java-exec/src/main/java/org/apache/drill/exec/planner/physical/DrillDistributionTrait.java
@@ -96,7 +96,7 @@ public class DrillDistributionTrait implements RelTrait {
@Override
public int hashCode() {
- return fields == null ? type.hashCode() : type.hashCode() | fields.hashCode() << 4 ;
+ return fields == null? type.hashCode(): type.hashCode() | fields.hashCode() << 4;
}
@Override
@@ -106,7 +106,7 @@ public class DrillDistributionTrait implements RelTrait {
}
if (obj instanceof DrillDistributionTrait) {
DrillDistributionTrait that = (DrillDistributionTrait) obj;
- return this.type == that.type && this.fields.equals(that.fields) ;
+ return this.type == that.type && this.fields.equals(that.fields);
}
return false;
}
diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/planner/physical/HashAggPrule.java b/exec/java-exec/src/main/java/org/apache/drill/exec/planner/physical/HashAggPrule.java
index 19499d6..ade0fb3 100644
--- a/exec/java-exec/src/main/java/org/apache/drill/exec/planner/physical/HashAggPrule.java
+++ b/exec/java-exec/src/main/java/org/apache/drill/exec/planner/physical/HashAggPrule.java
@@ -91,7 +91,7 @@ public class HashAggPrule extends AggPruleBase {
createTransformRequest(call, aggregate, input, traits);
if (create2PhasePlan(call, aggregate)) {
- traits = call.getPlanner().emptyTraitSet().plus(Prel.DRILL_PHYSICAL) ;
+ traits = call.getPlanner().emptyTraitSet().plus(Prel.DRILL_PHYSICAL);
RelNode convertedInput = convert(input, traits);
new TwoPhaseSubset(call, distOnAllKeys).go(aggregate, convertedInput);
diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/planner/physical/HashToMergeExchangePrel.java b/exec/java-exec/src/main/java/org/apache/drill/exec/planner/physical/HashToMergeExchangePrel.java
index 2272a9b..c9be46f 100644
--- a/exec/java-exec/src/main/java/org/apache/drill/exec/planner/physical/HashToMergeExchangePrel.java
+++ b/exec/java-exec/src/main/java/org/apache/drill/exec/planner/physical/HashToMergeExchangePrel.java
@@ -38,7 +38,7 @@ public class HashToMergeExchangePrel extends ExchangePrel {
private final List<DistributionField> distFields;
private int numEndPoints = 0;
- private final RelCollation collation ;
+ private final RelCollation collation;
public HashToMergeExchangePrel(RelOptCluster cluster, RelTraitSet traitSet, RelNode input,
List<DistributionField> fields,
diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/planner/physical/SingleMergeExchangePrel.java b/exec/java-exec/src/main/java/org/apache/drill/exec/planner/physical/SingleMergeExchangePrel.java
index a03c2f5..223882c 100644
--- a/exec/java-exec/src/main/java/org/apache/drill/exec/planner/physical/SingleMergeExchangePrel.java
+++ b/exec/java-exec/src/main/java/org/apache/drill/exec/planner/physical/SingleMergeExchangePrel.java
@@ -40,7 +40,7 @@ import org.apache.drill.exec.server.options.OptionManager;
public class SingleMergeExchangePrel extends ExchangePrel {
- private final RelCollation collation ;
+ private final RelCollation collation;
public SingleMergeExchangePrel(RelOptCluster cluster, RelTraitSet traitSet, RelNode input, RelCollation collation) {
super(cluster, traitSet, input);
diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/planner/physical/StreamAggPrule.java b/exec/java-exec/src/main/java/org/apache/drill/exec/planner/physical/StreamAggPrule.java
index 85f516a..99b6cde 100644
--- a/exec/java-exec/src/main/java/org/apache/drill/exec/planner/physical/StreamAggPrule.java
+++ b/exec/java-exec/src/main/java/org/apache/drill/exec/planner/physical/StreamAggPrule.java
@@ -70,7 +70,7 @@ public class StreamAggPrule extends AggPruleBase {
final RelTraitSet singleDistTrait = call.getPlanner().emptyTraitSet().plus(Prel.DRILL_PHYSICAL).plus(singleDist);
if (create2PhasePlan(call, aggregate)) {
- traits = call.getPlanner().emptyTraitSet().plus(Prel.DRILL_PHYSICAL) ;
+ traits = call.getPlanner().emptyTraitSet().plus(Prel.DRILL_PHYSICAL);
RelNode convertedInput = convert(input, traits);
new SubsetTransformer<DrillAggregateRel, InvalidRelException>(call){
@@ -138,7 +138,7 @@ public class StreamAggPrule extends AggPruleBase {
// createTransformRequest(call, aggregate, input, traits);
if (create2PhasePlan(call, aggregate)) {
- traits = call.getPlanner().emptyTraitSet().plus(Prel.DRILL_PHYSICAL) ;
+ traits = call.getPlanner().emptyTraitSet().plus(Prel.DRILL_PHYSICAL);
RelNode convertedInput = convert(input, traits);
new SubsetTransformer<DrillAggregateRel, InvalidRelException>(call){
diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/planner/physical/explain/NumberingRelWriter.java b/exec/java-exec/src/main/java/org/apache/drill/exec/planner/physical/explain/NumberingRelWriter.java
index 43e0300..eddbe4f 100644
--- a/exec/java-exec/src/main/java/org/apache/drill/exec/planner/physical/explain/NumberingRelWriter.java
+++ b/exec/java-exec/src/main/java/org/apache/drill/exec/planner/physical/explain/NumberingRelWriter.java
@@ -84,7 +84,9 @@ class NumberingRelWriter implements RelWriter {
s.append(" ");
if (id != null && id.opId == 0) {
- for(int i =0; i < spacer.get(); i++){ s.append('-');}
+ for (int i = 0; i < spacer.get(); i++) {
+ s.append('-');
+ }
}else{
spacer.spaces(s);
}
diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/planner/physical/visitor/StarColumnConverter.java b/exec/java-exec/src/main/java/org/apache/drill/exec/planner/physical/visitor/StarColumnConverter.java
index ac491e9..1248578 100644
--- a/exec/java-exec/src/main/java/org/apache/drill/exec/planner/physical/visitor/StarColumnConverter.java
+++ b/exec/java-exec/src/main/java/org/apache/drill/exec/planner/physical/visitor/StarColumnConverter.java
@@ -248,7 +248,7 @@ public class StarColumnConverter extends BasePrelVisitor<Prel, Void, RuntimeExce
for (String s : names) {
if (uniqueNames.contains(s)) {
- for (int i = 0; ; i++ ) {
+ for (int i = 0;; i++) {
s = s + i;
if (! origNames.contains(s) && ! uniqueNames.contains(s)) {
break;
diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/rpc/control/Controller.java b/exec/java-exec/src/main/java/org/apache/drill/exec/rpc/control/Controller.java
index 9bb574c..90de9af 100644
--- a/exec/java-exec/src/main/java/org/apache/drill/exec/rpc/control/Controller.java
+++ b/exec/java-exec/src/main/java/org/apache/drill/exec/rpc/control/Controller.java
@@ -42,7 +42,7 @@ public interface Controller extends AutoCloseable {
* @param node
* @return
*/
- public ControlTunnel getTunnel(DrillbitEndpoint node) ;
+ public ControlTunnel getTunnel(DrillbitEndpoint node);
public DrillbitEndpoint start(DrillbitEndpoint partialEndpoint, boolean allowPortHunting)
throws DrillbitStartupException;
diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/server/rest/profile/ProfileWrapper.java b/exec/java-exec/src/main/java/org/apache/drill/exec/server/rest/profile/ProfileWrapper.java
index bf005b1..22d0c7b 100644
--- a/exec/java-exec/src/main/java/org/apache/drill/exec/server/rest/profile/ProfileWrapper.java
+++ b/exec/java-exec/src/main/java/org/apache/drill/exec/server/rest/profile/ProfileWrapper.java
@@ -65,7 +65,7 @@ public class ProfileWrapper {
this.profile = profile;
this.id = profile.hasQueryId() ? profile.getQueryId() : QueryIdHelper.getQueryId(profile.getId());
//Generating Operator Name map (DRILL-6140)
- String profileTextPlan = profile.hasPlan() ? profile.getPlan() : "" ;
+ String profileTextPlan = profile.hasPlan()? profile.getPlan(): "";
generateOpMap(profileTextPlan);
final List<FragmentWrapper> fragmentProfiles = new ArrayList<>();
diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/server/rest/profile/SimpleDurationFormat.java b/exec/java-exec/src/main/java/org/apache/drill/exec/server/rest/profile/SimpleDurationFormat.java
index 5a2a37b..cddb75a 100644
--- a/exec/java-exec/src/main/java/org/apache/drill/exec/server/rest/profile/SimpleDurationFormat.java
+++ b/exec/java-exec/src/main/java/org/apache/drill/exec/server/rest/profile/SimpleDurationFormat.java
@@ -72,6 +72,6 @@ public class SimpleDurationFormat {
return (days > 0 ? days + " day " : "") +
((hours + days) > 0 ? hours + " hr " : "") +
((minutes + hours + days) > 0 ? String.format("%02d min ", minutes) : "") +
- seconds + "." + String.format("%03d sec", milliSeconds) ;
+ seconds + "." + String.format("%03d sec", milliSeconds);
}
}
diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/ssl/SSLConfig.java b/exec/java-exec/src/main/java/org/apache/drill/exec/ssl/SSLConfig.java
index 28cbe07..c3ff0c5 100644
--- a/exec/java-exec/src/main/java/org/apache/drill/exec/ssl/SSLConfig.java
+++ b/exec/java-exec/src/main/java/org/apache/drill/exec/ssl/SSLConfig.java
@@ -253,8 +253,7 @@ public abstract class SSLConfig {
.append("\n\ttrustStorePassword: ").append(getPrintablePassword(getTrustStorePassword()))
.append("\n\thandshakeTimeout: ").append(getHandshakeTimeout())
.append("\n\tdisableHostVerification: ").append(disableHostVerification())
- .append("\n\tdisableCertificateVerification: ").append(disableCertificateVerification())
- ;
+ .append("\n\tdisableCertificateVerification: ").append(disableCertificateVerification());
}
return sb.toString();
}
diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/json/JsonProcessor.java b/exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/json/JsonProcessor.java
index d35743f..fba80e5 100644
--- a/exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/json/JsonProcessor.java
+++ b/exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/json/JsonProcessor.java
@@ -51,7 +51,7 @@ public interface JsonProcessor {
String msg,
Object... args);
- public boolean ignoreJSONParseError() ;
+ public boolean ignoreJSONParseError();
public void setIgnoreJSONParseErrors(boolean ignoreJSONParseErrors);
}
diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/text/compliant/HeaderBuilder.java b/exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/text/compliant/HeaderBuilder.java
index ef8f861..6d52b6d 100644
--- a/exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/text/compliant/HeaderBuilder.java
+++ b/exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/text/compliant/HeaderBuilder.java
@@ -224,7 +224,7 @@ public class HeaderBuilder extends TextOutput {
// "col", "col_2", "col_2_2", "col_2_2_2".
// No mapping scheme is perfect...
- for (int l = 2; ; l++) {
+ for (int l = 2;; l++) {
final String rewritten = header + "_" + l;
key = rewritten.toLowerCase();
if (! idents.contains(key)) {
diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/text/compliant/TextInput.java b/exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/text/compliant/TextInput.java
index a0043ef..9ee86ca 100644
--- a/exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/text/compliant/TextInput.java
+++ b/exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/text/compliant/TextInput.java
@@ -96,7 +96,7 @@ final class TextInput {
this.lineSeparator = settings.getNewLineDelimiter();
byte normalizedLineSeparator = settings.getNormalizedNewLine();
Preconditions.checkArgument(input instanceof Seekable, "Text input only supports an InputStream that supports Seekable.");
- boolean isCompressed = input instanceof CompressionInputStream ;
+ boolean isCompressed = input instanceof CompressionInputStream;
Preconditions.checkArgument(!isCompressed || startPos == 0, "Cannot use split on compressed stream.");
// splits aren't allowed with compressed data. The split length will be the compressed size which means we'll normally end prematurely.
diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/store/ischema/Records.java b/exec/java-exec/src/main/java/org/apache/drill/exec/store/ischema/Records.java
index c4f8969..c684e7a 100644
--- a/exec/java-exec/src/main/java/org/apache/drill/exec/store/ischema/Records.java
+++ b/exec/java-exec/src/main/java/org/apache/drill/exec/store/ischema/Records.java
@@ -130,9 +130,15 @@ public class Records {
break;
// 2. SqlTypeName enumerators whose names (currently) do not match SQL's
// values for DATA_TYPE:
- case CHAR: this.DATA_TYPE = "CHARACTER"; break;
- case VARCHAR: this.DATA_TYPE = "CHARACTER VARYING"; break;
- case VARBINARY: this.DATA_TYPE = "BINARY VARYING"; break;
+ case CHAR:
+ this.DATA_TYPE = "CHARACTER";
+ break;
+ case VARCHAR:
+ this.DATA_TYPE = "CHARACTER VARYING";
+ break;
+ case VARBINARY:
+ this.DATA_TYPE = "BINARY VARYING";
+ break;
case INTERVAL_YEAR:
case INTERVAL_YEAR_MONTH:
case INTERVAL_MONTH:
@@ -145,7 +151,9 @@ public class Records {
case INTERVAL_HOUR_SECOND:
case INTERVAL_MINUTE:
case INTERVAL_MINUTE_SECOND:
- case INTERVAL_SECOND: this.DATA_TYPE = "INTERVAL"; break;
+ case INTERVAL_SECOND:
+ this.DATA_TYPE = "INTERVAL";
+ break;
// 3: SqlTypeName enumerators not yet seen and confirmed or handled.
default:
logger.warn( "Type not handled explicitly (code needs review): "
@@ -212,10 +220,18 @@ public class Records {
this.CHARACTER_OCTET_LENGTH = null;
// This NUMERIC_PRECISION is in bits since NUMERIC_PRECISION_RADIX is 2.
switch ( sqlTypeName ) {
- case TINYINT: NUMERIC_PRECISION = 8; break;
- case SMALLINT: NUMERIC_PRECISION = 16; break;
- case INTEGER: NUMERIC_PRECISION = 32; break;
- case BIGINT: NUMERIC_PRECISION = 64; break;
+ case TINYINT:
+ NUMERIC_PRECISION = 8;
+ break;
+ case SMALLINT:
+ NUMERIC_PRECISION = 16;
+ break;
+ case INTEGER:
+ NUMERIC_PRECISION = 32;
+ break;
+ case BIGINT:
+ NUMERIC_PRECISION = 64;
+ break;
default:
throw new AssertionError(
"Unexpected " + sqlTypeName.getClass().getName() + " value "
@@ -253,9 +269,15 @@ public class Records {
this.CHARACTER_OCTET_LENGTH = null;
// This NUMERIC_PRECISION is in bits since NUMERIC_PRECISION_RADIX is 2.
switch ( sqlTypeName ) {
- case REAL: NUMERIC_PRECISION = 24; break;
- case FLOAT: NUMERIC_PRECISION = 24; break;
- case DOUBLE: NUMERIC_PRECISION = 53; break;
+ case REAL:
+ NUMERIC_PRECISION = 24;
+ break;
+ case FLOAT:
+ NUMERIC_PRECISION = 24;
+ break;
+ case DOUBLE:
+ NUMERIC_PRECISION = 53;
+ break;
default:
throw new AssertionError(
"Unexpected type " + sqlTypeName + " in approximate-types branch" );
@@ -285,7 +307,9 @@ public class Records {
this.INTERVAL_TYPE = null;
this.INTERVAL_PRECISION = null;
switch(sqlTypeName) {
- case DATE: this.COLUMN_SIZE = 10; break;// yyyy-MM-dd
+ case DATE:
+ this.COLUMN_SIZE = 10;
+ break;// yyyy-MM-dd
case TIME: this.COLUMN_SIZE = this.DATETIME_PRECISION == 0
? 8 // HH::mm::ss
: 8 + 1 + this.DATETIME_PRECISION;
@@ -373,8 +397,12 @@ public class Records {
switch(start) {
case YEAR:
switch(end) {
- case YEAR: this.COLUMN_SIZE = INTERVAL_PRECISION + 2; break;// P..Y
- case MONTH: this.COLUMN_SIZE = this.INTERVAL_PRECISION + 5; break; // P..Y12M
+ case YEAR:
+ this.COLUMN_SIZE = INTERVAL_PRECISION + 2;
+ break;// P..Y
+ case MONTH:
+ this.COLUMN_SIZE = this.INTERVAL_PRECISION + 5;
+ break; // P..Y12M
default:
throw new AssertionError("Unexpected interval type " + this.INTERVAL_TYPE + " in interval-types branch" );
}
@@ -382,7 +410,9 @@ public class Records {
case MONTH:
switch(end) {
- case MONTH: this.COLUMN_SIZE = this.INTERVAL_PRECISION + 2; break; // P..M
+ case MONTH:
+ this.COLUMN_SIZE = this.INTERVAL_PRECISION + 2;
+ break; // P..M
default:
throw new AssertionError("Unexpected interval type " + this.INTERVAL_TYPE + " in interval-types branch" );
}
@@ -390,10 +420,18 @@ public class Records {
case DAY:
switch(end) {
- case DAY: this.COLUMN_SIZE = this.INTERVAL_PRECISION + 2; break; // P..D
- case HOUR: this.COLUMN_SIZE = this.INTERVAL_PRECISION + 6; break; // P..DT12H
- case MINUTE: this.COLUMN_SIZE = this.INTERVAL_PRECISION + 9; break; // P..DT12H60M
- case SECOND: this.COLUMN_SIZE = this.INTERVAL_PRECISION + 12 + extraSecondIntervalSize; break; // P..DT12H60M60....S
+ case DAY:
+ this.COLUMN_SIZE = this.INTERVAL_PRECISION + 2;
+ break; // P..D
+ case HOUR:
+ this.COLUMN_SIZE = this.INTERVAL_PRECISION + 6;
+ break; // P..DT12H
+ case MINUTE:
+ this.COLUMN_SIZE = this.INTERVAL_PRECISION + 9;
+ break; // P..DT12H60M
+ case SECOND:
+ this.COLUMN_SIZE = this.INTERVAL_PRECISION + 12 + extraSecondIntervalSize;
+ break; // P..DT12H60M60....S
default:
throw new AssertionError("Unexpected interval type " + this.INTERVAL_TYPE + " in interval-types branch" );
}
@@ -401,9 +439,15 @@ public class Records {
case HOUR:
switch(end) {
- case HOUR: this.COLUMN_SIZE = this.INTERVAL_PRECISION + 3; break; // PT..H
- case MINUTE: this.COLUMN_SIZE = this.INTERVAL_PRECISION + 6; break; // PT..H60M
- case SECOND: this.COLUMN_SIZE = this.INTERVAL_PRECISION + 9 + extraSecondIntervalSize; break; // PT..H12M60....S
+ case HOUR:
+ this.COLUMN_SIZE = this.INTERVAL_PRECISION + 3;
+ break; // PT..H
+ case MINUTE:
+ this.COLUMN_SIZE = this.INTERVAL_PRECISION + 6;
+ break; // PT..H60M
+ case SECOND:
+ this.COLUMN_SIZE = this.INTERVAL_PRECISION + 9 + extraSecondIntervalSize;
+ break; // PT..H12M60....S
default:
throw new AssertionError("Unexpected interval type " + this.INTERVAL_TYPE + " in interval-types branch" );
}
@@ -411,8 +455,12 @@ public class Records {
case MINUTE:
switch(end) {
- case MINUTE: this.COLUMN_SIZE = this.INTERVAL_PRECISION + 3; break; // PT...M
- case SECOND: this.COLUMN_SIZE = this.INTERVAL_PRECISION + 6 + extraSecondIntervalSize; break; // PT..M60....S
+ case MINUTE:
+ this.COLUMN_SIZE = this.INTERVAL_PRECISION + 3;
+ break; // PT...M
+ case SECOND:
+ this.COLUMN_SIZE = this.INTERVAL_PRECISION + 6 + extraSecondIntervalSize;
+ break; // PT..M60....S
default:
throw new AssertionError("Unexpected interval type " + this.INTERVAL_TYPE + " in interval-types branch" );
}
@@ -421,7 +469,9 @@ public class Records {
case SECOND:
switch(end) {
- case SECOND: this.COLUMN_SIZE = this.INTERVAL_PRECISION + 3 + extraSecondIntervalSize; break; // PT....S
+ case SECOND:
+ this.COLUMN_SIZE = this.INTERVAL_PRECISION + 3 + extraSecondIntervalSize;
+ break; // PT....S
default:
throw new AssertionError("Unexpected interval type " + this.INTERVAL_TYPE + " in interval-types branch" );
}
diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/store/parquet/ParquetFilterBuilder.java b/exec/java-exec/src/main/java/org/apache/drill/exec/store/parquet/ParquetFilterBuilder.java
index f45edbb..34c2fe7 100644
--- a/exec/java-exec/src/main/java/org/apache/drill/exec/store/parquet/ParquetFilterBuilder.java
+++ b/exec/java-exec/src/main/java/org/apache/drill/exec/store/parquet/ParquetFilterBuilder.java
@@ -214,7 +214,7 @@ public class ParquetFilterBuilder extends AbstractExprVisitor<LogicalExpression,
}
if (value.contains(funcHolderExpr)) {
- ValueHolder result ;
+ ValueHolder result;
try {
result = InterpreterEvaluator.evaluateConstantExpr(udfUtilities, funcHolderExpr);
} catch (Exception e) {
diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/store/parquet/ParquetPushDownFilter.java b/exec/java-exec/src/main/java/org/apache/drill/exec/store/parquet/ParquetPushDownFilter.java
index b5f0ca4..7a0dbbe 100644
--- a/exec/java-exec/src/main/java/org/apache/drill/exec/store/parquet/ParquetPushDownFilter.java
+++ b/exec/java-exec/src/main/java/org/apache/drill/exec/store/parquet/ParquetPushDownFilter.java
@@ -167,7 +167,7 @@ public abstract class ParquetPushDownFilter extends StoragePluginOptimizerRule {
return;
}
- RelNode newScan = ScanPrel.create(scan, scan.getTraitSet(), newGroupScan, scan.getRowType());;
+ RelNode newScan = ScanPrel.create(scan, scan.getTraitSet(), newGroupScan, scan.getRowType());
if (project != null) {
newScan = project.copy(project.getTraitSet(), ImmutableList.of(newScan));
diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/store/parquet/columnreaders/ColumnReader.java b/exec/java-exec/src/main/java/org/apache/drill/exec/store/parquet/columnreaders/ColumnReader.java
index c343d31..0474040 100644
--- a/exec/java-exec/src/main/java/org/apache/drill/exec/store/parquet/columnreaders/ColumnReader.java
+++ b/exec/java-exec/src/main/java/org/apache/drill/exec/store/parquet/columnreaders/ColumnReader.java
@@ -78,7 +78,10 @@ public abstract class ColumnReader<V extends ValueVector> {
int currDefLevel;
// variables for a single read pass
- long readStartInBytes = 0, readLength = 0, readLengthInBits = 0, recordsReadInThisIteration = 0;
+ long readStartInBytes = 0;
+ long readLength = 0;
+ long readLengthInBits = 0;
+ long recordsReadInThisIteration = 0;
private ExecutorService threadPool;
volatile boolean isShuttingDown; //Indicate to not submit any new AsyncPageReader Tasks during clear()
diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/store/parquet/metadata/MetadataPathUtils.java b/exec/java-exec/src/main/java/org/apache/drill/exec/store/parquet/metadata/MetadataPathUtils.java
index b9480e8..f7239d6 100644
--- a/exec/java-exec/src/main/java/org/apache/drill/exec/store/parquet/metadata/MetadataPathUtils.java
+++ b/exec/java-exec/src/main/java/org/apache/drill/exec/store/parquet/metadata/MetadataPathUtils.java
@@ -87,7 +87,7 @@ public class MetadataPathUtils {
ParquetTableMetadata_v3 tableMetadataWithAbsolutePaths, String baseDir) {
List<String> directoriesWithRelativePaths = Lists.newArrayList();
for (String directory : tableMetadataWithAbsolutePaths.getDirectories()) {
- directoriesWithRelativePaths.add(relativize(baseDir, directory)) ;
+ directoriesWithRelativePaths.add(relativize(baseDir, directory));
}
List<ParquetFileMetadata_v3> filesWithRelativePaths = Lists.newArrayList();
for (ParquetFileMetadata_v3 file : tableMetadataWithAbsolutePaths.files) {
diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/store/schedule/BlockMapBuilder.java b/exec/java-exec/src/main/java/org/apache/drill/exec/store/schedule/BlockMapBuilder.java
index b2961a5..5499207 100644
--- a/exec/java-exec/src/main/java/org/apache/drill/exec/store/schedule/BlockMapBuilder.java
+++ b/exec/java-exec/src/main/java/org/apache/drill/exec/store/schedule/BlockMapBuilder.java
@@ -192,7 +192,7 @@ public class BlockMapBuilder {
final Timer.Context context = metrics.timer(BLOCK_MAP_BUILDER_TIMER).time();
BlockLocation[] blocks;
ImmutableRangeMap<Long,BlockLocation> blockMap;
- blocks = fs.getFileBlockLocations(status, 0 , status.getLen());
+ blocks = fs.getFileBlockLocations(status, 0, status.getLen());
ImmutableRangeMap.Builder<Long, BlockLocation> blockMapBuilder = new ImmutableRangeMap.Builder<Long,BlockLocation>();
for (BlockLocation block : blocks) {
long start = block.getOffset();
diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/util/VectorUtil.java b/exec/java-exec/src/main/java/org/apache/drill/exec/util/VectorUtil.java
index 8729a39..270841f 100644
--- a/exec/java-exec/src/main/java/org/apache/drill/exec/util/VectorUtil.java
+++ b/exec/java-exec/src/main/java/org/apache/drill/exec/util/VectorUtil.java
@@ -56,7 +56,7 @@ public class VectorUtil {
int columnCounter = 0;
for (VectorWrapper<?> vw : va) {
boolean lastColumn = columnCounter == width - 1;
- Object o ;
+ Object o;
try{
o = vw.getValueVector().getAccessor().getObject(row);
} catch (Exception e) {
diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/work/batch/DataCollector.java b/exec/java-exec/src/main/java/org/apache/drill/exec/work/batch/DataCollector.java
index fa74677..30862ef 100644
--- a/exec/java-exec/src/main/java/org/apache/drill/exec/work/batch/DataCollector.java
+++ b/exec/java-exec/src/main/java/org/apache/drill/exec/work/batch/DataCollector.java
@@ -24,7 +24,7 @@ import org.apache.drill.exec.record.RawFragmentBatch;
public interface DataCollector extends AutoCloseable {
static final org.slf4j.Logger logger = org.slf4j.LoggerFactory.getLogger(DataCollector.class);
- public boolean batchArrived(int minorFragmentId, RawFragmentBatch batch) throws IOException ;
+ public boolean batchArrived(int minorFragmentId, RawFragmentBatch batch) throws IOException;
public int getOppositeMajorFragmentId();
public RawBatchBuffer[] getBuffers();
public int getTotalIncomingFragments();
diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/work/batch/UnlimitedRawBatchBuffer.java b/exec/java-exec/src/main/java/org/apache/drill/exec/work/batch/UnlimitedRawBatchBuffer.java
index 719f367..44b7f53 100644
--- a/exec/java-exec/src/main/java/org/apache/drill/exec/work/batch/UnlimitedRawBatchBuffer.java
+++ b/exec/java-exec/src/main/java/org/apache/drill/exec/work/batch/UnlimitedRawBatchBuffer.java
@@ -40,7 +40,7 @@ public class UnlimitedRawBatchBuffer extends BaseRawBatchBuffer<RawFragmentBatch
}
private class UnlimitedBufferQueue implements BufferQueue<RawFragmentBatch> {
- private final LinkedBlockingDeque<RawFragmentBatch> buffer = Queues.newLinkedBlockingDeque();;
+ private final LinkedBlockingDeque<RawFragmentBatch> buffer = Queues.newLinkedBlockingDeque();
@Override
public void addOomBatch(RawFragmentBatch batch) {
diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/work/foreman/rm/ResourceManagerBuilder.java b/exec/java-exec/src/main/java/org/apache/drill/exec/work/foreman/rm/ResourceManagerBuilder.java
index 4305891..f4e460d 100644
--- a/exec/java-exec/src/main/java/org/apache/drill/exec/work/foreman/rm/ResourceManagerBuilder.java
+++ b/exec/java-exec/src/main/java/org/apache/drill/exec/work/foreman/rm/ResourceManagerBuilder.java
@@ -53,7 +53,7 @@ public class ResourceManagerBuilder {
private static final org.slf4j.Logger logger = org.slf4j.LoggerFactory.getLogger(ResourceManagerBuilder.class);
- private DrillbitContext context ;
+ private DrillbitContext context;
public ResourceManagerBuilder(final DrillbitContext context) {
this.context = context;
diff --git a/exec/java-exec/src/test/java/org/apache/drill/PlanningBase.java b/exec/java-exec/src/test/java/org/apache/drill/PlanningBase.java
index 18fc4e1..802ecce 100644
--- a/exec/java-exec/src/test/java/org/apache/drill/PlanningBase.java
+++ b/exec/java-exec/src/test/java/org/apache/drill/PlanningBase.java
@@ -85,7 +85,7 @@ public class PlanningBase extends ExecTest {
provider.start();
final ScanResult scanResult = ClassPathScanner.fromPrescan(config);
final LogicalPlanPersistence logicalPlanPersistence = new LogicalPlanPersistence(config, scanResult);
- final SystemOptionManager systemOptions = new SystemOptionManager(logicalPlanPersistence , provider, config);
+ final SystemOptionManager systemOptions = new SystemOptionManager(logicalPlanPersistence, provider, config);
systemOptions.init();
@SuppressWarnings("resource")
final UserSession userSession = UserSession.Builder.newBuilder().withOptionManager(systemOptions).build();
diff --git a/exec/java-exec/src/test/java/org/apache/drill/TestStarQueries.java b/exec/java-exec/src/test/java/org/apache/drill/TestStarQueries.java
index d33fbee..683248e 100644
--- a/exec/java-exec/src/test/java/org/apache/drill/TestStarQueries.java
+++ b/exec/java-exec/src/test/java/org/apache/drill/TestStarQueries.java
@@ -293,7 +293,7 @@ public class TestStarQueries extends BaseTestQuery {
@Test(expected = UserException.class) // Should get "At line 1, column 8: Column 'n_nationkey' is ambiguous"
public void testSelStarAmbiguousJoin() throws Exception {
try {
- test("select x.n_nationkey, x.n_name, x.n_regionkey, x.r_name from (select * from cp.`tpch/nation.parquet` n, cp.`tpch/region.parquet` r where n.n_regionkey = r.r_regionkey) x " ) ;
+ test("select x.n_nationkey, x.n_name, x.n_regionkey, x.r_name from (select * from cp.`tpch/nation.parquet` n, cp.`tpch/region.parquet` r where n.n_regionkey = r.r_regionkey) x " );
} catch (UserException e) {
logger.info("***** Test resulted in expected failure: " + e.getMessage());
throw e;
diff --git a/exec/java-exec/src/test/java/org/apache/drill/TestUnionAll.java b/exec/java-exec/src/test/java/org/apache/drill/TestUnionAll.java
index 38f1d3a..d9d647e 100644
--- a/exec/java-exec/src/test/java/org/apache/drill/TestUnionAll.java
+++ b/exec/java-exec/src/test/java/org/apache/drill/TestUnionAll.java
@@ -475,8 +475,8 @@ public class TestUnionAll extends BaseTestQuery {
"(select columns[0] c2 from cp.`%s` t2 \n" +
"where t2.columns[0] is not null \n" +
"group by columns[0])) \n" +
- "group by col0"
- , root, root)
+ "group by col0",
+ root, root)
.unOrdered()
.baselineColumns("col0")
.baselineValues("290")
diff --git a/exec/java-exec/src/test/java/org/apache/drill/exec/RunRootExec.java b/exec/java-exec/src/test/java/org/apache/drill/exec/RunRootExec.java
index 0588a59..5065a57 100644
--- a/exec/java-exec/src/test/java/org/apache/drill/exec/RunRootExec.java
+++ b/exec/java-exec/src/test/java/org/apache/drill/exec/RunRootExec.java
@@ -55,7 +55,7 @@ public class RunRootExec {
FunctionImplementationRegistry registry = bitContext.getFunctionImplementationRegistry();
FragmentContextImpl context = new FragmentContextImpl(bitContext, PlanFragment.getDefaultInstance(), null, registry);
SimpleRootExec exec;
- for (int i = 0; i < iterations; i ++) {
+ for (int i = 0; i < iterations; i++) {
Stopwatch w = Stopwatch.createStarted();
logger.info("STARTITER: {}", i);
exec = new SimpleRootExec(ImplCreator.getExec(context, (FragmentRoot) plan.getSortedOperators(false).iterator().next()));
diff --git a/exec/java-exec/src/test/java/org/apache/drill/exec/expr/fn/registry/FunctionRegistryHolderTest.java b/exec/java-exec/src/test/java/org/apache/drill/exec/expr/fn/registry/FunctionRegistryHolderTest.java
index 883b94a..bc6b742 100644
--- a/exec/java-exec/src/test/java/org/apache/drill/exec/expr/fn/registry/FunctionRegistryHolderTest.java
+++ b/exec/java-exec/src/test/java/org/apache/drill/exec/expr/fn/registry/FunctionRegistryHolderTest.java
@@ -204,7 +204,7 @@ public class FunctionRegistryHolderTest {
for (List<FunctionHolder> functionHolders : newJars.values()) {
for (FunctionHolder functionHolder : functionHolders) {
if ("lower".equals(functionHolder.getName())) {
- expectedResult.add(functionHolder.getHolder()) ;
+ expectedResult.add(functionHolder.getHolder());
}
}
}
@@ -220,7 +220,7 @@ public class FunctionRegistryHolderTest {
for (List<FunctionHolder> functionHolders : newJars.values()) {
for (FunctionHolder functionHolder : functionHolders) {
if ("lower".equals(functionHolder.getName())) {
- expectedResult.add(functionHolder.getHolder()) ;
+ expectedResult.add(functionHolder.getHolder());
}
}
}
diff --git a/exec/java-exec/src/test/java/org/apache/drill/exec/fn/impl/TestAggregateFunction.java b/exec/java-exec/src/test/java/org/apache/drill/exec/fn/impl/TestAggregateFunction.java
index 045c24d..975675a 100644
--- a/exec/java-exec/src/test/java/org/apache/drill/exec/fn/impl/TestAggregateFunction.java
+++ b/exec/java-exec/src/test/java/org/apache/drill/exec/fn/impl/TestAggregateFunction.java
@@ -85,7 +85,7 @@ public class TestAggregateFunction extends PopUnitTestBase {
public void testCovarianceCorrelation() throws Throwable {
String planPath = "/functions/test_covariance.json";
String dataPath = "/covariance_input.json";
- Double expectedValues[] = {4.571428571428571d, 4.857142857142857d, -6.000000000000002d, 4.0d , 4.25d, -5.250000000000002d, 1.0d, 0.9274260335029677d, -1.0000000000000004d};
+ Double expectedValues[] = {4.571428571428571d, 4.857142857142857d, -6.000000000000002d, 4.0d, 4.25d, -5.250000000000002d, 1.0d, 0.9274260335029677d, -1.0000000000000004d};
runTest(expectedValues, planPath, dataPath);
}
diff --git a/exec/java-exec/src/test/java/org/apache/drill/exec/fn/impl/TestDateTruncFunctions.java b/exec/java-exec/src/test/java/org/apache/drill/exec/fn/impl/TestDateTruncFunctions.java
index ff32733..a9da761 100644
--- a/exec/java-exec/src/test/java/org/apache/drill/exec/fn/impl/TestDateTruncFunctions.java
+++ b/exec/java-exec/src/test/java/org/apache/drill/exec/fn/impl/TestDateTruncFunctions.java
@@ -87,7 +87,7 @@ public class TestDateTruncFunctions extends BaseTestQuery {
testBuilder()
.sqlQuery(query)
.unOrdered()
- .baselineColumns("second", "minute", "hour", "day", "month", "week" , "year", "q1", "q2", "q3", "decade1", "decade2", "decade3")
+ .baselineColumns("second", "minute", "hour", "day", "month", "week", "year", "q1", "q2", "q3", "decade1", "decade2", "decade3")
.baselineValues(
DateUtility.parseLocalDate("2011-02-03"), // seconds
DateUtility.parseLocalDate("2011-02-03"), // minute
@@ -183,7 +183,7 @@ public class TestDateTruncFunctions extends BaseTestQuery {
testBuilder()
.sqlQuery(query)
.unOrdered()
- .baselineColumns("second", "minute", "hour", "day", "month", "week" , "year", "q1", "q2", "q3", "decade1", "decade2", "decade3")
+ .baselineColumns("second", "minute", "hour", "day", "month", "week", "year", "q1", "q2", "q3", "decade1", "decade2", "decade3")
.baselineValues(
DateUtility.parseLocalDateTime("2011-02-03 10:11:12.0"), // seconds
DateUtility.parseLocalDateTime("2011-02-03 10:11:00.0"), // minute
diff --git a/exec/java-exec/src/test/java/org/apache/drill/exec/fn/interp/ExpressionInterpreterTest.java b/exec/java-exec/src/test/java/org/apache/drill/exec/fn/interp/ExpressionInterpreterTest.java
index 3b81fa4..249ad5c 100644
--- a/exec/java-exec/src/test/java/org/apache/drill/exec/fn/interp/ExpressionInterpreterTest.java
+++ b/exec/java-exec/src/test/java/org/apache/drill/exec/fn/interp/ExpressionInterpreterTest.java
@@ -227,7 +227,7 @@ public class ExpressionInterpreterTest extends PopUnitTestBase {
}
private void showValueVectorContent(ValueVector vw) {
- for (int row = 0; row < vw.getAccessor().getValueCount(); row ++ ) {
+ for (int row = 0; row < vw.getAccessor().getValueCount(); row++) {
final Object o = vw.getAccessor().getObject(row);
final String cellString;
if (o instanceof byte[]) {
diff --git a/exec/java-exec/src/test/java/org/apache/drill/exec/impersonation/TestImpersonationMetadata.java b/exec/java-exec/src/test/java/org/apache/drill/exec/impersonation/TestImpersonationMetadata.java
index 1ef9c7b..2eb55db 100644
--- a/exec/java-exec/src/test/java/org/apache/drill/exec/impersonation/TestImpersonationMetadata.java
+++ b/exec/java-exec/src/test/java/org/apache/drill/exec/impersonation/TestImpersonationMetadata.java
@@ -65,7 +65,7 @@ public class TestImpersonationMetadata extends BaseTestImpersonation {
addMiniDfsBasedStorage(createTestWorkspaces());
}
- private static Map<String , WorkspaceConfig> createTestWorkspaces() throws Exception {
+ private static Map<String, WorkspaceConfig> createTestWorkspaces() throws Exception {
// Create "/tmp" folder and set permissions to "777"
final Path tmpPath = new Path("/tmp");
fs.delete(tmpPath, true);
diff --git a/exec/java-exec/src/test/java/org/apache/drill/exec/memory/TestAllocators.java b/exec/java-exec/src/test/java/org/apache/drill/exec/memory/TestAllocators.java
index 3501bfc..315bc29 100644
--- a/exec/java-exec/src/test/java/org/apache/drill/exec/memory/TestAllocators.java
+++ b/exec/java-exec/src/test/java/org/apache/drill/exec/memory/TestAllocators.java
@@ -181,8 +181,7 @@ public class TestAllocators extends DrillTest {
final DrillConfig config = DrillConfig.create(TEST_CONFIGURATIONS);
try (final RemoteServiceSet serviceSet = RemoteServiceSet.getLocalServiceSet();
- final Drillbit bit = new Drillbit(config, serviceSet)) {
- ;
+ final Drillbit bit = new Drillbit(config, serviceSet)) {
bit.run();
final DrillbitContext bitContext = bit.getContext();
FunctionImplementationRegistry functionRegistry = bitContext.getFunctionImplementationRegistry();
diff --git a/exec/java-exec/src/test/java/org/apache/drill/exec/physical/impl/PartitionLimit/TestPartitionLimitBatch.java b/exec/java-exec/src/test/java/org/apache/drill/exec/physical/impl/PartitionLimit/TestPartitionLimitBatch.java
index 574ff76..926b7d9 100644
--- a/exec/java-exec/src/test/java/org/apache/drill/exec/physical/impl/PartitionLimit/TestPartitionLimitBatch.java
+++ b/exec/java-exec/src/test/java/org/apache/drill/exec/physical/impl/PartitionLimit/TestPartitionLimitBatch.java
@@ -344,7 +344,7 @@ public class TestPartitionLimitBatch extends BaseTestOpBatchEmitOutcome {
expectedRowSets.add(expectedRowSet1);
expectedRowSets.add(expectedRowSet2);
- testPartitionLimitCommon(0 ,1);
+ testPartitionLimitCommon(0, 1);
}
@Test
@@ -385,7 +385,7 @@ public class TestPartitionLimitBatch extends BaseTestOpBatchEmitOutcome {
expectedRowSets.add(expectedRowSet1);
- testPartitionLimitCommon(2 ,3);
+ testPartitionLimitCommon(2, 3);
}
/**
@@ -440,7 +440,7 @@ public class TestPartitionLimitBatch extends BaseTestOpBatchEmitOutcome {
expectedRowSets.add(expectedRowSet1);
expectedRowSets.add(expectedRowSet2);
- testPartitionLimitCommon(0 ,5);
+ testPartitionLimitCommon(0, 5);
}
/**
diff --git a/exec/java-exec/src/test/java/org/apache/drill/exec/physical/impl/TestDecimal.java b/exec/java-exec/src/test/java/org/apache/drill/exec/physical/impl/TestDecimal.java
index ff1edb9..0f47b80 100644
--- a/exec/java-exec/src/test/java/org/apache/drill/exec/physical/impl/TestDecimal.java
+++ b/exec/java-exec/src/test/java/org/apache/drill/exec/physical/impl/TestDecimal.java
@@ -159,7 +159,7 @@ public class TestDecimal extends PopUnitTestBase {
String addOutput[] = {"123456888.0", "22.2", "0.2", "-0.2", "-987654444.2","-3.0"};
String subtractOutput[] = {"123456690.0", "0.0", "0.0", "0.0", "-987654198.0", "-1.0"};
- String multiplyOutput[] = {"12222222111.00" , "123.21" , "0.01", "0.01", "121580246927.41", "2.00"};
+ String multiplyOutput[] = {"12222222111.00", "123.21", "0.01", "0.01", "121580246927.41", "2.00"};
Iterator<VectorWrapper<?>> itr = batchLoader.iterator();
@@ -208,7 +208,7 @@ public class TestDecimal extends PopUnitTestBase {
QueryDataBatch batch = results.get(0);
assertTrue(batchLoader.load(batch.getHeader().getDef(), batch.getData()));
- String addOutput[] = {"-99999998877.700000000", "11.423456789", "123456789.100000000", "-0.119998000", "100000000112.423456789" , "-99999999879.907000000", "123456789123456801.300000000"};
+ String addOutput[] = {"-99999998877.700000000", "11.423456789", "123456789.100000000", "-0.119998000", "100000000112.423456789", "-99999999879.907000000", "123456789123456801.300000000"};
String subtractOutput[] = {"-100000001124.300000000", "10.823456789", "-123456788.900000000", "-0.120002000", "99999999889.823456789", "-100000000122.093000000", "123456789123456776.700000000"};
Iterator<VectorWrapper<?>> itr = batchLoader.iterator();
diff --git a/exec/java-exec/src/test/java/org/apache/drill/exec/physical/impl/TestDistributedFragmentRun.java b/exec/java-exec/src/test/java/org/apache/drill/exec/physical/impl/TestDistributedFragmentRun.java
index 6fc3dbe..8030573 100644
--- a/exec/java-exec/src/test/java/org/apache/drill/exec/physical/impl/TestDistributedFragmentRun.java
+++ b/exec/java-exec/src/test/java/org/apache/drill/exec/physical/impl/TestDistributedFragmentRun.java
@@ -43,7 +43,8 @@ public class TestDistributedFragmentRun extends PopUnitTestBase{
public void oneBitOneExchangeOneEntryRun() throws Exception{
RemoteServiceSet serviceSet = RemoteServiceSet.getLocalServiceSet();
- try(Drillbit bit1 = new Drillbit(CONFIG, serviceSet); DrillClient client = new DrillClient(CONFIG, serviceSet.getCoordinator());){
+ try (Drillbit bit1 = new Drillbit(CONFIG, serviceSet);
+ DrillClient client = new DrillClient(CONFIG, serviceSet.getCoordinator())) {
bit1.run();
client.connect();
List<QueryDataBatch> results = client.runQuery(QueryType.PHYSICAL, Files.toString(DrillFileUtils.getResourceAsFile("/physical_single_exchange.json"), Charsets.UTF_8));
@@ -63,7 +64,8 @@ public class TestDistributedFragmentRun extends PopUnitTestBase{
public void oneBitOneExchangeTwoEntryRun() throws Exception{
RemoteServiceSet serviceSet = RemoteServiceSet.getLocalServiceSet();
- try(Drillbit bit1 = new Drillbit(CONFIG, serviceSet); DrillClient client = new DrillClient(CONFIG, serviceSet.getCoordinator());){
+ try (Drillbit bit1 = new Drillbit(CONFIG, serviceSet);
+ DrillClient client = new DrillClient(CONFIG, serviceSet.getCoordinator())) {
bit1.run();
client.connect();
List<QueryDataBatch> results = client.runQuery(QueryType.PHYSICAL, Files.toString(DrillFileUtils.getResourceAsFile("/physical_single_exchange_double_entry.json"), Charsets.UTF_8));
@@ -82,16 +84,17 @@ public class TestDistributedFragmentRun extends PopUnitTestBase{
public void oneBitOneExchangeTwoEntryRunLogical() throws Exception{
RemoteServiceSet serviceSet = RemoteServiceSet.getLocalServiceSet();
- try(Drillbit bit1 = new Drillbit(CONFIG, serviceSet); DrillClient client = new DrillClient(CONFIG, serviceSet.getCoordinator());){
- bit1.run();
- client.connect();
- List<QueryDataBatch> results = client.runQuery(QueryType.LOGICAL, Files.toString(DrillFileUtils.getResourceAsFile("/scan_screen_logical.json"), Charsets.UTF_8));
- int count = 0;
- for(QueryDataBatch b : results){
- count += b.getHeader().getRowCount();
- b.release();
- }
- assertEquals(100, count);
+ try (Drillbit bit1 = new Drillbit(CONFIG, serviceSet);
+ DrillClient client = new DrillClient(CONFIG, serviceSet.getCoordinator())) {
+ bit1.run();
+ client.connect();
+ List<QueryDataBatch> results = client.runQuery(QueryType.LOGICAL, Files.toString(DrillFileUtils.getResourceAsFile("/scan_screen_logical.json"), Charsets.UTF_8));
+ int count = 0;
+ for (QueryDataBatch b : results) {
+ count += b.getHeader().getRowCount();
+ b.release();
+ }
+ assertEquals(100, count);
}
@@ -101,7 +104,9 @@ public class TestDistributedFragmentRun extends PopUnitTestBase{
public void twoBitOneExchangeTwoEntryRun() throws Exception{
RemoteServiceSet serviceSet = RemoteServiceSet.getLocalServiceSet();
- try(Drillbit bit1 = new Drillbit(CONFIG, serviceSet); Drillbit bit2 = new Drillbit(CONFIG, serviceSet); DrillClient client = new DrillClient(CONFIG, serviceSet.getCoordinator());){
+ try (Drillbit bit1 = new Drillbit(CONFIG, serviceSet);
+ Drillbit bit2 = new Drillbit(CONFIG, serviceSet);
+ DrillClient client = new DrillClient(CONFIG, serviceSet.getCoordinator())) {
bit1.run();
bit2.run();
client.connect();
diff --git a/exec/java-exec/src/test/java/org/apache/drill/exec/physical/impl/TopN/TestTopNSchemaChanges.java b/exec/java-exec/src/test/java/org/apache/drill/exec/physical/impl/TopN/TestTopNSchemaChanges.java
index 60f1e67..6b16aab 100644
--- a/exec/java-exec/src/test/java/org/apache/drill/exec/physical/impl/TopN/TestTopNSchemaChanges.java
+++ b/exec/java-exec/src/test/java/org/apache/drill/exec/physical/impl/TopN/TestTopNSchemaChanges.java
@@ -72,7 +72,7 @@ public class TestTopNSchemaChanges extends BaseTestQuery {
.ordered()
.baselineColumns("kl", "vl");
- for (long i = 0; i< 12 ; ++i) {
+ for (long i = 0; i < 12; ++i) {
if (i %2 == 0) {
builder.baselineValues(i, i);
} else {
@@ -102,7 +102,7 @@ public class TestTopNSchemaChanges extends BaseTestQuery {
.ordered()
.baselineColumns("kl", "vl");
- for (long i = 0; i< 24 ; i+=2) {
+ for (long i = 0; i < 24; i+=2) {
builder.baselineValues(i, i);
}
diff --git a/exec/java-exec/src/test/java/org/apache/drill/exec/physical/impl/agg/TestHashAggEmitOutcome.java b/exec/java-exec/src/test/java/org/apache/drill/exec/physical/impl/agg/TestHashAggEmitOutcome.java
index b03a989..ebb1cad 100644
--- a/exec/java-exec/src/test/java/org/apache/drill/exec/physical/impl/agg/TestHashAggEmitOutcome.java
+++ b/exec/java-exec/src/test/java/org/apache/drill/exec/physical/impl/agg/TestHashAggEmitOutcome.java
@@ -84,7 +84,7 @@ public class TestHashAggEmitOutcome extends BaseTestOpBatchEmitOutcome {
// First input batch
RowSetBuilder builder2 = operatorFixture.rowSetBuilder(inputSchema);
if ( inp2_1 != null ) {
- for ( int i = 0; i < inp2_1.length ; i++) {
+ for (int i = 0; i < inp2_1.length; i++) {
builder2 = builder2.addRow(inp2_1[i], inp2_2[i], inp2_3[i]);
}
}
@@ -93,7 +93,7 @@ public class TestHashAggEmitOutcome extends BaseTestOpBatchEmitOutcome {
// Second input batch
RowSetBuilder builder3 = operatorFixture.rowSetBuilder(inputSchema);
if ( inp3_1 != null ) {
- for ( int i = 0; i < inp3_1.length ; i++) {
+ for (int i = 0; i < inp3_1.length; i++) {
builder3 = builder3.addRow(inp3_1[i], inp3_2[i], inp3_3[i]);
}
}
diff --git a/exec/java-exec/src/test/java/org/apache/drill/exec/physical/impl/agg/TestHashAggrSpill.java b/exec/java-exec/src/test/java/org/apache/drill/exec/physical/impl/agg/TestHashAggrSpill.java
index 295010a..aac93ea 100644
--- a/exec/java-exec/src/test/java/org/apache/drill/exec/physical/impl/agg/TestHashAggrSpill.java
+++ b/exec/java-exec/src/test/java/org/apache/drill/exec/physical/impl/agg/TestHashAggrSpill.java
@@ -55,7 +55,7 @@ public class TestHashAggrSpill extends DrillTest {
*
* @throws Exception
*/
- private void testSpill(long maxMem, long numPartitions, long minBatches, int maxParallel, boolean fallback ,boolean predict,
+ private void testSpill(long maxMem, long numPartitions, long minBatches, int maxParallel, boolean fallback, boolean predict,
String sql, long expectedRows, int cycle, int fromPart, int toPart) throws Exception {
ClusterFixtureBuilder builder = ClusterFixture.builder(dirTestWatcher)
.sessionOption(ExecConstants.HASHAGG_MAX_MEMORY_KEY,maxMem)
diff --git a/exec/java-exec/src/test/java/org/apache/drill/exec/physical/impl/common/HashPartitionTest.java b/exec/java-exec/src/test/java/org/apache/drill/exec/physical/impl/common/HashPartitionTest.java
index 6d06434..70acf74 100644
--- a/exec/java-exec/src/test/java/org/apache/drill/exec/physical/impl/common/HashPartitionTest.java
+++ b/exec/java-exec/src/test/java/org/apache/drill/exec/physical/impl/common/HashPartitionTest.java
@@ -216,7 +216,7 @@ public class HashPartitionTest {
hashPartition.completeAnInnerBatch(false, false);
hashPartition.spillThisPartition();
final String spillFile = hashPartition.getSpillFile();
- final int batchesCount = hashPartition.getPartitionBatchesCount();;
+ final int batchesCount = hashPartition.getPartitionBatchesCount();
hashPartition.closeWriter();
SpilledRecordbatch spilledBuildBatch = new SpilledRecordbatch(spillFile, batchesCount, context, buildSchema, operatorContext, spillSet);
@@ -270,8 +270,8 @@ public class HashPartitionTest {
final BatchSchema probeSchema = new BatchSchema(BatchSchema.SelectionVectorMode.NONE, probeCols);
final RecordBatch probeBatch = testCase.createProbeBatch(probeSchema, allocator);
- final LogicalExpression buildColExpression = SchemaPath.getSimplePath(buildColB.getName());;
- final LogicalExpression probeColExpression = SchemaPath.getSimplePath(probeColB.getName());;
+ final LogicalExpression buildColExpression = SchemaPath.getSimplePath(buildColB.getName());
+ final LogicalExpression probeColExpression = SchemaPath.getSimplePath(probeColB.getName());
final JoinCondition condition = new JoinCondition(DrillJoinRel.EQUALITY_CONDITION, probeColExpression, buildColExpression);
final List<Comparator> comparators = Lists.newArrayList(JoinUtils.checkAndReturnSupportedJoinComparator(condition));
diff --git a/exec/java-exec/src/test/java/org/apache/drill/exec/physical/impl/flatten/TestFlatten.java b/exec/java-exec/src/test/java/org/apache/drill/exec/physical/impl/flatten/TestFlatten.java
index 4ad2987..35d1010 100644
--- a/exec/java-exec/src/test/java/org/apache/drill/exec/physical/impl/flatten/TestFlatten.java
+++ b/exec/java-exec/src/test/java/org/apache/drill/exec/physical/impl/flatten/TestFlatten.java
@@ -336,7 +336,7 @@ public class TestFlatten extends BaseTestQuery {
// currently runs
// TODO - re-verify results by hand
if(RUN_ADVANCED_TESTS){
- test("select flatten(kvgen(visited_cellid_counts)) as mytb from dfs.`tmp/mapkv.json`") ;
+ test("select flatten(kvgen(visited_cellid_counts)) as mytb from dfs.`tmp/mapkv.json`");
}
}
diff --git a/exec/java-exec/src/test/java/org/apache/drill/exec/physical/impl/join/TestHashJoinSpill.java b/exec/java-exec/src/test/java/org/apache/drill/exec/physical/impl/join/TestHashJoinSpill.java
index 5935d1b..1483b6f 100644
--- a/exec/java-exec/src/test/java/org/apache/drill/exec/physical/impl/join/TestHashJoinSpill.java
+++ b/exec/java-exec/src/test/java/org/apache/drill/exec/physical/impl/join/TestHashJoinSpill.java
@@ -101,7 +101,7 @@ public class TestHashJoinSpill extends PhysicalOpUnitTestBase {
List<String> rightTable = Lists.newArrayList("[{\"rgt\": 0, \"b\" : \"a string\"}]",
"[{\"rgt\": 0, \"b\" : \"a different string\"},{\"rgt\": 0, \"b\" : \"yet another\"}]");
int numRows = 4_000; // 100_000
- for ( int cnt = 1; cnt <= numRows / 2 ; cnt++ ) { // inner use only half, to check the left-outer join
+ for (int cnt = 1; cnt <= numRows / 2; cnt++) { // inner use only half, to check the left-outer join
// leftTable.add("[{\"lft\": " + cnt + ", \"a\" : \"a string\"}]");
rightTable.add("[{\"rgt\": " + cnt + ", \"b\" : \"a string\"}]");
}
diff --git a/exec/java-exec/src/test/java/org/apache/drill/exec/physical/impl/protocol/TestOperatorRecordBatch.java b/exec/java-exec/src/test/java/org/apache/drill/exec/physical/impl/protocol/TestOperatorRecordBatch.java
index 1203f48..108e9a2 100644
--- a/exec/java-exec/src/test/java/org/apache/drill/exec/physical/impl/protocol/TestOperatorRecordBatch.java
+++ b/exec/java-exec/src/test/java/org/apache/drill/exec/physical/impl/protocol/TestOperatorRecordBatch.java
@@ -99,7 +99,10 @@ public class TestOperatorRecordBatch extends SubOperatorTest {
}
@Override
- public boolean buildSchema() { buildSchemaCalled = true; return ! schemaEOF; }
+ public boolean buildSchema() {
+ buildSchemaCalled = true;
+ return !schemaEOF;
+ }
@Override
public boolean next() {
diff --git a/exec/java-exec/src/test/java/org/apache/drill/exec/physical/impl/writer/TestParquetWriter.java b/exec/java-exec/src/test/java/org/apache/drill/exec/physical/impl/writer/TestParquetWriter.java
index 66ee3e4..807f7fd 100644
--- a/exec/java-exec/src/test/java/org/apache/drill/exec/physical/impl/writer/TestParquetWriter.java
+++ b/exec/java-exec/src/test/java/org/apache/drill/exec/physical/impl/writer/TestParquetWriter.java
@@ -164,7 +164,7 @@ public class TestParquetWriter extends BaseTestQuery {
final int numCols = 1000;
String[] colNames = new String[numCols];
Object[] values = new Object[numCols];
- for (int i = 0 ; i < numCols - 1; i++) {
+ for (int i = 0; i < numCols - 1; i++) {
sb.append(String.format("\"col_%d\" : 100,", i));
colNames[i] = "col_" + i;
values[i] = 100L;
diff --git a/exec/java-exec/src/test/java/org/apache/drill/exec/physical/rowSet/impl/TestResultSetLoaderMapArray.java b/exec/java-exec/src/test/java/org/apache/drill/exec/physical/rowSet/impl/TestResultSetLoaderMapArray.java
index 653137e..3d816c3 100644
--- a/exec/java-exec/src/test/java/org/apache/drill/exec/physical/rowSet/impl/TestResultSetLoaderMapArray.java
+++ b/exec/java-exec/src/test/java/org/apache/drill/exec/physical/rowSet/impl/TestResultSetLoaderMapArray.java
@@ -101,8 +101,7 @@ public class TestResultSetLoaderMapArray extends SubOperatorTest {
.addRow(30, mapArray(
mapValue(310, "d3.1"),
mapValue(320, "d3.2"),
- mapValue(330, "d3.3")))
- ;
+ mapValue(330, "d3.3")));
// Verify the first batch
@@ -144,8 +143,7 @@ public class TestResultSetLoaderMapArray extends SubOperatorTest {
.addRow(60, mapArray(
mapValue(610, "d6.1", "e6.1"),
mapValue(620, "d6.2", null),
- mapValue(630, "d6.3", "e6.3")))
- ;
+ mapValue(630, "d6.3", "e6.3")));
// Verify the second batch
@@ -207,8 +205,7 @@ public class TestResultSetLoaderMapArray extends SubOperatorTest {
.addRow(30, mapArray(
mapValue(310, strArray("d3.1.1", "d3.2.2")),
mapValue(320, strArray()),
- mapValue(330, strArray("d3.3.1", "d1.2.2"))))
- ;
+ mapValue(330, strArray("d3.3.1", "d1.2.2"))));
// Verify the batch
diff --git a/exec/java-exec/src/test/java/org/apache/drill/exec/physical/rowSet/impl/TestResultSetLoaderMaps.java b/exec/java-exec/src/test/java/org/apache/drill/exec/physical/rowSet/impl/TestResultSetLoaderMaps.java
index 98ba3ed..f876fb8 100644
--- a/exec/java-exec/src/test/java/org/apache/drill/exec/physical/rowSet/impl/TestResultSetLoaderMaps.java
+++ b/exec/java-exec/src/test/java/org/apache/drill/exec/physical/rowSet/impl/TestResultSetLoaderMaps.java
@@ -553,8 +553,7 @@ public class TestResultSetLoaderMaps extends SubOperatorTest {
.addRow(10, mapValue(intArray(110, 120, 130),
strArray("d1.1", "d1.2", "d1.3", "d1.4")))
.addRow(20, mapValue(intArray(210), strArray()))
- .addRow(30, mapValue(intArray(), strArray("d3.1")))
- ;
+ .addRow(30, mapValue(intArray(), strArray("d3.1")));
// Validate first batch
@@ -573,15 +572,13 @@ public class TestResultSetLoaderMaps extends SubOperatorTest {
rsLoader.startBatch();
rootWriter
.addRow(40, mapValue(intArray(410, 420), strArray("d4.1", "d4.2")))
- .addRow(50, mapValue(intArray(510), strArray("d5.1")))
- ;
+ .addRow(50, mapValue(intArray(510), strArray("d5.1")));
TupleWriter mapWriter = rootWriter.tuple("m");
mapWriter.addColumn(SchemaBuilder.columnSchema("e", MinorType.VARCHAR, DataMode.REPEATED));
rootWriter
.addRow(60, mapValue(intArray(610, 620), strArray("d6.1", "d6.2"), strArray("e6.1", "e6.2")))
- .addRow(70, mapValue(intArray(710), strArray(), strArray("e7.1", "e7.2")))
- ;
+ .addRow(70, mapValue(intArray(710), strArray(), strArray("e7.1", "e7.2")));
// Validate first batch. The new array should have been back-filled with
// empty offsets for the missing rows.
diff --git a/exec/java-exec/src/test/java/org/apache/drill/exec/physical/rowSet/impl/TestResultSetLoaderTorture.java b/exec/java-exec/src/test/java/org/apache/drill/exec/physical/rowSet/impl/TestResultSetLoaderTorture.java
index 826b71e..bcf01de 100644
--- a/exec/java-exec/src/test/java/org/apache/drill/exec/physical/rowSet/impl/TestResultSetLoaderTorture.java
+++ b/exec/java-exec/src/test/java/org/apache/drill/exec/physical/rowSet/impl/TestResultSetLoaderTorture.java
@@ -250,7 +250,7 @@ public class TestResultSetLoaderTorture extends SubOperatorTest {
public BatchReader(TestSetup setup, RowSetReader reader, ReadState readState) {
this.setup = setup;
this.rootReader = reader;
- this.readState = readState;;
+ this.readState = readState;
TupleReader m1Reader = rootReader.tuple("m1");
n1Reader = m1Reader.scalar("n1");
@@ -351,12 +351,11 @@ public class TestResultSetLoaderTorture extends SubOperatorTest {
@Test
public void tortureTest() {
- LogFixtureBuilder logBuilder = new LogFixtureBuilder()
+ LogFixtureBuilder logBuilder = new LogFixtureBuilder();
// Enable to get detailed tracing when things go wrong.
// .logger("org.apache.drill.exec.physical.rowSet", Level.TRACE)
- ;
try (LogFixture logFixture = logBuilder.build()) {
doTortureTest();
}
diff --git a/exec/java-exec/src/test/java/org/apache/drill/exec/physical/unit/MiniPlanUnitTestBase.java b/exec/java-exec/src/test/java/org/apache/drill/exec/physical/unit/MiniPlanUnitTestBase.java
index 6374f1f..79f260f 100644
--- a/exec/java-exec/src/test/java/org/apache/drill/exec/physical/unit/MiniPlanUnitTestBase.java
+++ b/exec/java-exec/src/test/java/org/apache/drill/exec/physical/unit/MiniPlanUnitTestBase.java
@@ -91,14 +91,14 @@ public class MiniPlanUnitTestBase extends PhysicalOpUnitTestBase {
* @param baselineValues
* @return
*/
- public MiniPlanTestBuilder baselineValues(Object ... baselineValues) {
+ public MiniPlanTestBuilder baselineValues(Object... baselineValues) {
if (baselineRecords == null) {
baselineRecords = new ArrayList<>();
}
Map<String, Object> ret = new HashMap<>();
int i = 0;
- Preconditions.checkArgument(expectSchema != null , "Expected schema should be set before specify baseline values.");
+ Preconditions.checkArgument(expectSchema != null, "Expected schema should be set before specify baseline values.");
Preconditions.checkArgument(baselineValues.length == expectSchema.getFieldCount(),
"Must supply the same number of baseline values as columns in expected schema.");
@@ -230,7 +230,7 @@ public class MiniPlanUnitTestBase extends PhysicalOpUnitTestBase {
protected long maxAllocation = MAX_ALLOCATION;
final private List<RecordBatch> inputs = Lists.newArrayList();
- final PopBuilder parent ;
+ final PopBuilder parent;
public PopBuilder() {
this.parent = null;
@@ -334,13 +334,13 @@ public class MiniPlanUnitTestBase extends PhysicalOpUnitTestBase {
}
@SuppressWarnings("unchecked")
- public T columnsToRead(SchemaPath ... columnsToRead) {
+ public T columnsToRead(SchemaPath... columnsToRead) {
this.columnsToRead = Lists.newArrayList(columnsToRead);
return (T) this;
}
@SuppressWarnings("unchecked")
- public T columnsToRead(String ... columnsToRead) {
+ public T columnsToRead(String... columnsToRead) {
this.columnsToRead = Lists.newArrayList();
for (String column : columnsToRead) {
diff --git a/exec/java-exec/src/test/java/org/apache/drill/exec/physical/unit/TestNullInputMiniPlan.java b/exec/java-exec/src/test/java/org/apache/drill/exec/physical/unit/TestNullInputMiniPlan.java
index 3d99adb..4a11657 100644
--- a/exec/java-exec/src/test/java/org/apache/drill/exec/physical/unit/TestNullInputMiniPlan.java
+++ b/exec/java-exec/src/test/java/org/apache/drill/exec/physical/unit/TestNullInputMiniPlan.java
@@ -525,7 +525,7 @@ public class TestNullInputMiniPlan extends MiniPlanUnitTestBase{
.expectNullBatch(true)
.go();
- final RecordBatch input2 = createScanBatchFromJson(SINGLE_EMPTY_JSON, SINGLE_EMPTY_JSON2);;
+ final RecordBatch input2 = createScanBatchFromJson(SINGLE_EMPTY_JSON, SINGLE_EMPTY_JSON2);
RecordBatch batch2 = new PopBuilder()
.physicalOperator(pop)
.addInput(input2)
diff --git a/exec/java-exec/src/test/java/org/apache/drill/exec/record/TestRecordBatchSizer.java b/exec/java-exec/src/test/java/org/apache/drill/exec/record/TestRecordBatchSizer.java
index eb50519..3e47db7 100644
--- a/exec/java-exec/src/test/java/org/apache/drill/exec/record/TestRecordBatchSizer.java
+++ b/exec/java-exec/src/test/java/org/apache/drill/exec/record/TestRecordBatchSizer.java
@@ -1333,7 +1333,7 @@ public class TestRecordBatchSizer extends SubOperatorTest {
ValueVector valueVector1 = mapVector.getChild("value");
assertEquals(((Integer.highestOneBit(testRowCount * STD_REPETITION_FACTOR) << 1)), keyVector.getValueCapacity());
offsetVector = ((VariableWidthVector)valueVector1).getOffsetVector();
- assertEquals((Integer.highestOneBit(testRowCount * STD_REPETITION_FACTOR) << 1) , offsetVector.getValueCapacity());
+ assertEquals((Integer.highestOneBit(testRowCount * STD_REPETITION_FACTOR) << 1), offsetVector.getValueCapacity());
assertEquals(Integer.highestOneBit(testRowCount * STD_REPETITION_FACTOR << 1) - 1, valueVector1.getValueCapacity());
// Allocates the same as value passed since it is already power of two.
diff --git a/exec/java-exec/src/test/java/org/apache/drill/exec/record/vector/TestDateTypes.java b/exec/java-exec/src/test/java/org/apache/drill/exec/record/vector/TestDateTypes.java
index 1da827a..4313ad0 100644
--- a/exec/java-exec/src/test/java/org/apache/drill/exec/record/vector/TestDateTypes.java
+++ b/exec/java-exec/src/test/java/org/apache/drill/exec/record/vector/TestDateTypes.java
@@ -138,8 +138,8 @@ public class TestDateTypes extends PopUnitTestBase {
ValueVector.Accessor accessor = v.getValueVector().getAccessor();
- assertEquals(accessor.getObject(0).toString() ,"1970-01-02 10:20:33.000");
- assertEquals(accessor.getObject(1).toString() ,"2008-12-28 11:34:00.129");
+ assertEquals(accessor.getObject(0).toString(),"1970-01-02 10:20:33.000");
+ assertEquals(accessor.getObject(1).toString(),"2008-12-28 11:34:00.129");
assertEquals(accessor.getObject(2).toString(), "2000-02-27 14:24:00.000");
}
diff --git a/exec/java-exec/src/test/java/org/apache/drill/exec/store/dfs/TestFileSelection.java b/exec/java-exec/src/test/java/org/apache/drill/exec/store/dfs/TestFileSelection.java
index 787584d..f2f550a 100644
--- a/exec/java-exec/src/test/java/org/apache/drill/exec/store/dfs/TestFileSelection.java
+++ b/exec/java-exec/src/test/java/org/apache/drill/exec/store/dfs/TestFileSelection.java
@@ -54,7 +54,7 @@ public class TestFileSelection extends BaseTestQuery {
{"/tmp", "../etc/bad"}, // goes outside parent; resolves to /etc/bad
{"", "/bad"}, // empty parent
{"/", ""}, // empty path
- } ;
+ };
for (int i = 0; i < badPaths.length; i++) {
@@ -82,7 +82,7 @@ public class TestFileSelection extends BaseTestQuery {
{"/", "etc/tmp/../../good"}, // no leading slash in path
{"/", "../good"}, // resolves to /../good which is OK
{"/", "/good"}
- } ;
+ };
for (int i = 0; i < goodPaths.length; i++) {
try {
diff --git a/exec/java-exec/src/test/java/org/apache/drill/exec/store/parquet/ParquetInternalsTest.java b/exec/java-exec/src/test/java/org/apache/drill/exec/store/parquet/ParquetInternalsTest.java
index 5a5207b..caa4e73 100644
--- a/exec/java-exec/src/test/java/org/apache/drill/exec/store/parquet/ParquetInternalsTest.java
+++ b/exec/java-exec/src/test/java/org/apache/drill/exec/store/parquet/ParquetInternalsTest.java
@@ -39,9 +39,9 @@ public class ParquetInternalsTest extends ClusterTest {
@BeforeClass
public static void setup( ) throws Exception {
- ClusterFixtureBuilder builder = ClusterFixture.builder(dirTestWatcher)
+ ClusterFixtureBuilder builder = ClusterFixture.builder(dirTestWatcher);
// Set options, etc.
- ;
+
startCluster(builder);
}
diff --git a/exec/java-exec/src/test/java/org/apache/drill/exec/store/parquet/ParquetSimpleTestFileGenerator.java b/exec/java-exec/src/test/java/org/apache/drill/exec/store/parquet/ParquetSimpleTestFileGenerator.java
index 15dd633..2e6923a 100644
--- a/exec/java-exec/src/test/java/org/apache/drill/exec/store/parquet/ParquetSimpleTestFileGenerator.java
+++ b/exec/java-exec/src/test/java/org/apache/drill/exec/store/parquet/ParquetSimpleTestFileGenerator.java
@@ -87,7 +87,7 @@ public class ParquetSimpleTestFileGenerator {
// " required int64 _TIMESTAMP_MICROS_int64 ( TIMESTAMP_MICROS ) ; \n" +
" required fixed_len_byte_array(12) _INTERVAL_fixed_len_byte_array_12 ( INTERVAL ) ; \n" +
" required int96 _INT96_RAW ; \n" +
- "} \n" ;
+ "} \n";
public static String simpleNullableSchemaMsg =
"message ParquetLogicalDataTypes { \n" +
" required int32 rowKey; \n" +
@@ -115,7 +115,7 @@ public class ParquetSimpleTestFileGenerator {
// " optional int64 _TIMESTAMP_MICROS_int64 ( TIMESTAMP_MICROS ) ; \n" +
" optional fixed_len_byte_array(12) _INTERVAL_fixed_len_byte_array_12 ( INTERVAL ) ; \n" +
" optional int96 _INT96_RAW ; \n" +
- "} \n" ;
+ "} \n";
public static String complexSchemaMsg =
"message ParquetLogicalDataTypes { \n" +
@@ -160,7 +160,7 @@ public class ParquetSimpleTestFileGenerator {
" required int96 _INT96_RAW ; \n" +
" } \n" +
" } \n" +
- "} \n" ;
+ "} \n";
public static String complexNullableSchemaMsg =
"message ParquetLogicalDataTypes { \n" +
" required int32 rowKey; \n" +
@@ -204,7 +204,7 @@ public class ParquetSimpleTestFileGenerator {
" optional int96 _INT96_RAW ; \n" +
" } \n" +
" } \n" +
- "} \n" ;
+ "} \n";
public static MessageType simpleSchema = MessageTypeParser.parseMessageType(simpleSchemaMsg);
public static MessageType complexSchema = MessageTypeParser.parseMessageType(complexSchemaMsg);
@@ -292,7 +292,8 @@ public class ParquetSimpleTestFileGenerator {
.append("_INT_64", 0x7FFFFFFFFFFFFFFFL)
.append("_UINT_64", 0xFFFFFFFFFFFFFFFFL)
.append("_DECIMAL_decimal18", 0xFFFFFFFFFFFFFFFFL);
- byte[] bytes = new byte[30]; Arrays.fill(bytes, (byte)1);
+ byte[] bytes = new byte[30];
+ Arrays.fill(bytes, (byte) 1);
numeric.addGroup("FixedLen").append("_DECIMAL_fixed_n", Binary.fromConstantByteArray(bytes, 0, 20));
numeric.addGroup("Binary").append("_DECIMAL_unlimited", Binary.fromConstantByteArray(bytes, 0, 30));
numeric.addGroup("DateTimeTypes")
@@ -375,7 +376,8 @@ public class ParquetSimpleTestFileGenerator {
}
{
Group simpleGroup = sgf.newGroup();
- byte[] bytes = new byte[30]; Arrays.fill(bytes, (byte)1);
+ byte[] bytes = new byte[30];
+ Arrays.fill(bytes, (byte) 1);
simpleGroup.append("rowKey", ++rowKey);
simpleGroup.append("_UTF8", "UTF8 string" + rowKey)
.append("_Enum", MAX_VALUE.toString())
diff --git a/exec/java-exec/src/test/java/org/apache/drill/exec/store/parquet/TestParquetComplex.java b/exec/java-exec/src/test/java/org/apache/drill/exec/store/parquet/TestParquetComplex.java
index c147980..c309cf3 100644
--- a/exec/java-exec/src/test/java/org/apache/drill/exec/store/parquet/TestParquetComplex.java
+++ b/exec/java-exec/src/test/java/org/apache/drill/exec/store/parquet/TestParquetComplex.java
@@ -215,10 +215,10 @@ public class TestParquetComplex extends BaseTestQuery {
.sqlQuery(query)
.unOrdered()
.baselineColumns(columns)
- .baselineValues(mapOf("a","a","b","b") , 0L , 0 , 0 , 0 , 0L , 0 , 0 ,0 )
- .baselineValues(mapOf("a","a","b","b") , -1L , -1 , -1 , -1 , -1L , -1 , -1 , -1 )
- .baselineValues(mapOf("a","a","b","b") , 1L , 1 , 1 , 1 , -9223372036854775808L , 1 , 1 , 1 )
- .baselineValues(mapOf("a","a","b","b") , 9223372036854775807L , 2147483647 , 65535 , 255 , 9223372036854775807L , -2147483648 , -32768 , -128 )
+ .baselineValues(mapOf("a","a","b","b"), 0L, 0, 0, 0, 0L, 0, 0, 0)
+ .baselineValues(mapOf("a","a","b","b"), -1L, -1, -1, -1, -1L, -1, -1, -1)
+ .baselineValues(mapOf("a","a","b","b"), 1L, 1, 1, 1, -9223372036854775808L, 1, 1, 1)
+ .baselineValues(mapOf("a","a","b","b"), 9223372036854775807L, 2147483647, 65535, 255, 9223372036854775807L, -2147483648, -32768, -128)
.build()
.run();
}
@@ -253,23 +253,23 @@ public class TestParquetComplex extends BaseTestQuery {
" cp.`store/parquet/complex/parquet_logical_types_complex.parquet` t " +
" order by t.rowKey ";
String[] columns = {
- "rowKey " ,
- "_UTF8" ,
- "_Enum" ,
- "_INT32_RAW" ,
- "_INT_8" ,
- "_INT_16" ,
- "_INT_32" ,
- "_UINT_8" ,
- "_UINT_16" ,
- "_UINT_32" ,
- "_INT64_RAW" ,
- "_INT_64" ,
- "_UINT_64" ,
- "_DATE_int32" ,
- "_TIME_MILLIS_int32" ,
- "_TIMESTAMP_MILLIS_int64" ,
- "_INTERVAL_fixed_len_byte_array_12" ,
+ "rowKey ",
+ "_UTF8",
+ "_Enum",
+ "_INT32_RAW",
+ "_INT_8",
+ "_INT_16",
+ "_INT_32",
+ "_UINT_8",
+ "_UINT_16",
+ "_UINT_32",
+ "_INT64_RAW",
+ "_INT_64",
+ "_UINT_64",
+ "_DATE_int32",
+ "_TIME_MILLIS_int32",
+ "_TIMESTAMP_MILLIS_int64",
+ "_INTERVAL_fixed_len_byte_array_12",
"_INT96_RAW"
};
@@ -327,23 +327,23 @@ public class TestParquetComplex extends BaseTestQuery {
" cp.`store/parquet/complex/parquet_logical_types_complex_nullable.parquet` t " +
" order by t.rowKey ";
String[] columns = {
- "rowKey " ,
- "_UTF8" ,
- "_Enum" ,
- "_INT32_RAW" ,
- "_INT_8" ,
- "_INT_16" ,
- "_INT_32" ,
- "_UINT_8" ,
- "_UINT_16" ,
- "_UINT_32" ,
- "_INT64_RAW" ,
- "_INT_64" ,
- "_UINT_64" ,
- "_DATE_int32" ,
- "_TIME_MILLIS_int32" ,
- "_TIMESTAMP_MILLIS_int64" ,
- "_INTERVAL_fixed_len_byte_array_12" ,
+ "rowKey ",
+ "_UTF8",
+ "_Enum",
+ "_INT32_RAW",
+ "_INT_8",
+ "_INT_16",
+ "_INT_32",
+ "_UINT_8",
+ "_UINT_16",
+ "_UINT_32",
+ "_INT64_RAW",
+ "_INT_64",
+ "_UINT_64",
+ "_DATE_int32",
+ "_TIME_MILLIS_int32",
+ "_TIMESTAMP_MILLIS_int64",
+ "_INTERVAL_fixed_len_byte_array_12",
"_INT96_RAW"
};
diff --git a/exec/java-exec/src/test/java/org/apache/drill/exec/store/parquet/TestParquetPhysicalPlan.java b/exec/java-exec/src/test/java/org/apache/drill/exec/store/parquet/TestParquetPhysicalPlan.java
index 9db8398..68c1fce 100644
--- a/exec/java-exec/src/test/java/org/apache/drill/exec/store/parquet/TestParquetPhysicalPlan.java
+++ b/exec/java-exec/src/test/java/org/apache/drill/exec/store/parquet/TestParquetPhysicalPlan.java
@@ -54,7 +54,8 @@ public class TestParquetPhysicalPlan extends ExecTest {
RemoteServiceSet serviceSet = RemoteServiceSet.getLocalServiceSet();
DrillConfig config = DrillConfig.create();
- try (Drillbit bit1 = new Drillbit(config, serviceSet); DrillClient client = new DrillClient(config, serviceSet.getCoordinator())) {
+ try (Drillbit bit1 = new Drillbit(config, serviceSet);
+ DrillClient client = new DrillClient(config, serviceSet.getCoordinator())) {
bit1.run();
client.connect();
List<QueryDataBatch> results = client.runQuery(org.apache.drill.exec.proto.UserBitShared.QueryType.PHYSICAL, Resources.toString(Resources.getResource(fileName),Charsets.UTF_8));
diff --git a/exec/java-exec/src/test/java/org/apache/drill/exec/store/parquet2/TestDrillParquetReader.java b/exec/java-exec/src/test/java/org/apache/drill/exec/store/parquet2/TestDrillParquetReader.java
index bb242ba..941f50f 100644
--- a/exec/java-exec/src/test/java/org/apache/drill/exec/store/parquet2/TestDrillParquetReader.java
+++ b/exec/java-exec/src/test/java/org/apache/drill/exec/store/parquet2/TestDrillParquetReader.java
@@ -117,10 +117,10 @@ public class TestDrillParquetReader extends BaseTestQuery {
.sqlQuery(query)
.unOrdered()
.baselineColumns(columns)
- .baselineValues( 0L , 0 , 0 , 0 , 0L , 0 , 0 ,0 )
- .baselineValues( -1L , -1 , -1 , -1 , -1L , -1 , -1 , -1 )
- .baselineValues( 1L , 1 , 1 , 1 , -9223372036854775808L , 1 , 1 , 1 )
- .baselineValues( 9223372036854775807L , 2147483647 , 65535 , 255 , 9223372036854775807L , -2147483648 , -32768 , -128 )
+ .baselineValues(0L, 0, 0, 0, 0L, 0, 0, 0)
+ .baselineValues(-1L, -1, -1, -1, -1L, -1, -1, -1)
+ .baselineValues(1L, 1, 1, 1, -9223372036854775808L, 1, 1, 1)
+ .baselineValues(9223372036854775807L, 2147483647, 65535, 255, 9223372036854775807L, -2147483648, -32768, -128)
.build()
.run();
}
@@ -128,7 +128,8 @@ public class TestDrillParquetReader extends BaseTestQuery {
@Test //DRILL-5971
public void testLogicalIntTypes2() throws Exception {
byte[] bytes12 = {'1', '2', '3', '4', '5', '6', '7', '8', '9', '0', 'a', 'b' };
- byte[] bytesOnes = new byte[12]; Arrays.fill(bytesOnes, (byte)1);
+ byte[] bytesOnes = new byte[12];
+ Arrays.fill(bytesOnes, (byte)1);
byte[] bytesZeros = new byte[12];
String query = String.format(
" select " +
@@ -155,23 +156,23 @@ public class TestDrillParquetReader extends BaseTestQuery {
" order by t.rowKey "
);
String[] columns = {
- "rowKey " ,
- "_UTF8" ,
- "_Enum" ,
- "_INT32_RAW" ,
- "_INT_8" ,
- "_INT_16" ,
- "_INT_32" ,
- "_UINT_8" ,
- "_UINT_16" ,
- "_UINT_32" ,
- "_INT64_RAW" ,
- "_INT_64" ,
- "_UINT_64" ,
- "_DATE_int32" ,
- "_TIME_MILLIS_int32" ,
- "_TIMESTAMP_MILLIS_int64" ,
- "_INTERVAL_fixed_len_byte_array_12" ,
+ "rowKey ",
+ "_UTF8",
+ "_Enum",
+ "_INT32_RAW",
+ "_INT_8",
+ "_INT_16",
+ "_INT_32",
+ "_UINT_8",
+ "_UINT_16",
+ "_UINT_32",
+ "_INT64_RAW",
+ "_INT_64",
+ "_UINT_64",
+ "_DATE_int32",
+ "_TIME_MILLIS_int32",
+ "_TIMESTAMP_MILLIS_int64",
+ "_INTERVAL_fixed_len_byte_array_12",
"_INT96_RAW"
};
@@ -202,7 +203,8 @@ public class TestDrillParquetReader extends BaseTestQuery {
@Test //DRILL-5971
public void testLogicalIntTypes3() throws Exception {
byte[] bytes12 = {'1', '2', '3', '4', '5', '6', '7', '8', '9', '0', 'a', 'b' };
- byte[] bytesOnes = new byte[12]; Arrays.fill(bytesOnes, (byte)1);
+ byte[] bytesOnes = new byte[12];
+ Arrays.fill(bytesOnes, (byte)1);
byte[] bytesZeros = new byte[12];
String query = String.format(
" select " +
@@ -229,23 +231,23 @@ public class TestDrillParquetReader extends BaseTestQuery {
" order by t.rowKey "
);
String[] columns = {
- "rowKey " ,
- "_UTF8" ,
- "_Enum" ,
- "_INT32_RAW" ,
- "_INT_8" ,
- "_INT_16" ,
- "_INT_32" ,
- "_UINT_8" ,
- "_UINT_16" ,
- "_UINT_32" ,
- "_INT64_RAW" ,
- "_INT_64" ,
- "_UINT_64" ,
- "_DATE_int32" ,
- "_TIME_MILLIS_int32" ,
- "_TIMESTAMP_MILLIS_int64" ,
- "_INTERVAL_fixed_len_byte_array_12" ,
+ "rowKey ",
+ "_UTF8",
+ "_Enum",
+ "_INT32_RAW",
+ "_INT_8",
+ "_INT_16",
+ "_INT_32",
+ "_UINT_8",
+ "_UINT_16",
+ "_UINT_32",
+ "_INT64_RAW",
+ "_INT_64",
+ "_UINT_64",
+ "_DATE_int32",
+ "_TIME_MILLIS_int32",
+ "_TIMESTAMP_MILLIS_int64",
+ "_INTERVAL_fixed_len_byte_array_12",
"_INT96_RAW"
};
diff --git a/exec/java-exec/src/test/java/org/apache/drill/exec/vector/TestSplitAndTransfer.java b/exec/java-exec/src/test/java/org/apache/drill/exec/vector/TestSplitAndTransfer.java
index 96dbd7c..d8a9e03 100644
--- a/exec/java-exec/src/test/java/org/apache/drill/exec/vector/TestSplitAndTransfer.java
+++ b/exec/java-exec/src/test/java/org/apache/drill/exec/vector/TestSplitAndTransfer.java
@@ -170,9 +170,9 @@ public class TestSplitAndTransfer {
bitVector.allocateNew(valueCount + 8); // extra byte at the end that gets filled with junk
final int[] compareArray = new int[valueCount];
- int testBitValue = 0 ;
+ int testBitValue = 0;
final BitVector.Mutator mutator = bitVector.getMutator();
- for (int i = 0; i < valueCount; i ++) {
+ for (int i = 0; i < valueCount; i++) {
testBitValue = getBit(pattern, i);
mutator.set(i, testBitValue);
compareArray[i] = testBitValue;
diff --git a/exec/java-exec/src/test/java/org/apache/drill/exec/vector/complex/writer/TestJsonReader.java b/exec/java-exec/src/test/java/org/apache/drill/exec/vector/complex/writer/TestJsonReader.java
index ccab796..b20c7e2 100644
--- a/exec/java-exec/src/test/java/org/apache/drill/exec/vector/complex/writer/TestJsonReader.java
+++ b/exec/java-exec/src/test/java/org/apache/drill/exec/vector/complex/writer/TestJsonReader.java
@@ -591,7 +591,7 @@ public class TestJsonReader extends BaseTestQuery {
table_dir.mkdir();
BufferedOutputStream os = new BufferedOutputStream(new FileOutputStream(new File(table_dir, "mostlynulls.json")));
// Create an entire batch of null values for 3 columns
- for (int i = 0 ; i < JSONRecordReader.DEFAULT_ROWS_PER_BATCH; i++) {
+ for (int i = 0; i < JSONRecordReader.DEFAULT_ROWS_PER_BATCH; i++) {
os.write("{\"a\": null, \"b\": null, \"c\": null}".getBytes());
}
// Add a row with {bigint, float, string} values
diff --git a/exec/java-exec/src/test/java/org/apache/drill/test/DrillTestWrapper.java b/exec/java-exec/src/test/java/org/apache/drill/test/DrillTestWrapper.java
index e037d02..d7028a5 100644
--- a/exec/java-exec/src/test/java/org/apache/drill/test/DrillTestWrapper.java
+++ b/exec/java-exec/src/test/java/org/apache/drill/test/DrillTestWrapper.java
@@ -627,7 +627,7 @@ public class DrillTestWrapper {
compareMergedVectors(expectedSuperVectors, actualSuperVectors);
} catch (Exception e) {
- throw new Exception(e.getMessage() + "\nFor query: " + query , e);
+ throw new Exception(e.getMessage() + "\nFor query: " + query, e);
} finally {
cleanupBatches(expected, actual);
}
diff --git a/exec/java-exec/src/test/java/org/apache/drill/test/ExampleTest.java b/exec/java-exec/src/test/java/org/apache/drill/test/ExampleTest.java
index 77ee6e9..6b69d2b 100644
--- a/exec/java-exec/src/test/java/org/apache/drill/test/ExampleTest.java
+++ b/exec/java-exec/src/test/java/org/apache/drill/test/ExampleTest.java
@@ -134,7 +134,8 @@ public class ExampleTest {
ClusterFixtureBuilder builder = ClusterFixture.builder(dirTestWatcher).configProperty(ExecConstants.SLICE_TARGET, 10);
- try (ClusterFixture cluster = builder.build(); ClientFixture client = cluster.clientFixture()) {
+ try (ClusterFixture cluster = builder.build();
+ ClientFixture client = cluster.clientFixture()) {
String sql = "SELECT * FROM `dfs`.`test/employee.json`";
logger.info(client.queryBuilder().sql(sql).explainJson());
QuerySummary results = client.queryBuilder().sql(sql).run();
diff --git a/exec/java-exec/src/test/java/org/apache/drill/test/QueryRowSetIterator.java b/exec/java-exec/src/test/java/org/apache/drill/test/QueryRowSetIterator.java
index c1b9253..addd68b 100644
--- a/exec/java-exec/src/test/java/org/apache/drill/test/QueryRowSetIterator.java
+++ b/exec/java-exec/src/test/java/org/apache/drill/test/QueryRowSetIterator.java
@@ -52,7 +52,7 @@ public class QueryRowSetIterator implements Iterator<DirectRowSet>, Iterable<Dir
@Override
public boolean hasNext() {
- for ( ; ; ) {
+ for (;;) {
QueryEvent event = listener.get();
state = event.state;
batch = null;
diff --git a/exec/java-exec/src/test/java/org/apache/drill/test/TestBuilder.java b/exec/java-exec/src/test/java/org/apache/drill/test/TestBuilder.java
index 5060e0a..7b00cb4 100644
--- a/exec/java-exec/src/test/java/org/apache/drill/test/TestBuilder.java
+++ b/exec/java-exec/src/test/java/org/apache/drill/test/TestBuilder.java
@@ -334,7 +334,7 @@ public class TestBuilder {
* @param baselineValues - the baseline values to validate
* @return the test builder
*/
- public TestBuilder baselineValues(Object ... baselineValues) {
+ public TestBuilder baselineValues(Object... baselineValues) {
assert getExpectedSchema() == null : "The expected schema is not needed when baselineValues are provided ";
if (ordered == null) {
throw new RuntimeException("Ordering not set, before specifying baseline data you must explicitly call the ordered() or unOrdered() method on the " + this.getClass().getSimpleName());
@@ -410,7 +410,7 @@ public class TestBuilder {
baselineTypeMap, baselineOptionSettingQueries, testOptionSettingQueries, highPerformanceComparison, expectedNumBatches);
}
- public BaselineQueryTestBuilder sqlBaselineQuery(String query, String ...replacements) {
+ public BaselineQueryTestBuilder sqlBaselineQuery(String query, String... replacements) {
return sqlBaselineQuery(String.format(query, (Object[]) replacements));
}
@@ -442,7 +442,7 @@ public class TestBuilder {
precision = String.format("(%d,%d)", type.getPrecision(), type.getScale());
break;
default:
- ; // do nothing empty string set above
+ // do nothing empty string set above
}
return precision;
}
@@ -471,7 +471,7 @@ public class TestBuilder {
}
// convenience method to convert minor types to major types if no decimals with precisions are needed
- public CSVTestBuilder baselineTypes(TypeProtos.MinorType ... baselineTypes) {
+ public CSVTestBuilder baselineTypes(TypeProtos.MinorType... baselineTypes) {
TypeProtos.MajorType[] majorTypes = new TypeProtos.MajorType[baselineTypes.length];
int i = 0;
for(TypeProtos.MinorType minorType : baselineTypes) {
diff --git a/exec/java-exec/src/test/java/org/apache/drill/test/TestGracefulShutdown.java b/exec/java-exec/src/test/java/org/apache/drill/test/TestGracefulShutdown.java
index bec1691..9e3721c 100644
--- a/exec/java-exec/src/test/java/org/apache/drill/test/TestGracefulShutdown.java
+++ b/exec/java-exec/src/test/java/org/apache/drill/test/TestGracefulShutdown.java
@@ -71,7 +71,7 @@ public class TestGracefulShutdown extends BaseTestQuery {
@Test
public void testOnlineEndPoints() throws Exception {
- String[] drillbits = {"db1" ,"db2","db3"};
+ String[] drillbits = {"db1", "db2", "db3"};
ClusterFixtureBuilder builder = ClusterFixture.bareBuilder(dirTestWatcher).withLocalZk().withBits(drillbits);
enableDrillPortHunting(builder);
@@ -142,7 +142,7 @@ public class TestGracefulShutdown extends BaseTestQuery {
@Test
public void testRestApiShutdown() throws Exception {
- String[] drillbits = {"db1" ,"db2", "db3"};
+ String[] drillbits = {"db1", "db2", "db3"};
ClusterFixtureBuilder builder = ClusterFixture.bareBuilder(dirTestWatcher).withLocalZk().withBits(drillbits);
enableWebServer(builder);
QueryBuilder.QuerySummaryFuture listener;
diff --git a/exec/java-exec/src/test/java/org/apache/drill/test/rowSet/test/TestFixedWidthWriter.java b/exec/java-exec/src/test/java/org/apache/drill/test/rowSet/test/TestFixedWidthWriter.java
index adaeb33..8177d9f 100644
--- a/exec/java-exec/src/test/java/org/apache/drill/test/rowSet/test/TestFixedWidthWriter.java
+++ b/exec/java-exec/src/test/java/org/apache/drill/test/rowSet/test/TestFixedWidthWriter.java
@@ -401,7 +401,7 @@ public class TestFixedWidthWriter extends SubOperatorTest {
});
writer.startWrite();
try {
- for (int i = 0; ; i++ ) {
+ for (int i = 0;; i++ ) {
index.index = i;
writer.startRow();
writer.setInt(i);
diff --git a/exec/java-exec/src/test/java/org/apache/drill/test/rowSet/test/TestOffsetVectorWriter.java b/exec/java-exec/src/test/java/org/apache/drill/test/rowSet/test/TestOffsetVectorWriter.java
index 78257e9..f9b835f 100644
--- a/exec/java-exec/src/test/java/org/apache/drill/test/rowSet/test/TestOffsetVectorWriter.java
+++ b/exec/java-exec/src/test/java/org/apache/drill/test/rowSet/test/TestOffsetVectorWriter.java
@@ -384,7 +384,7 @@ public class TestOffsetVectorWriter extends SubOperatorTest {
});
writer.startWrite();
try {
- for (int i = 0; ; i++ ) {
+ for (int i = 0;; i++ ) {
index.index = i;
writer.startRow();
writer.setNextOffset(i);
diff --git a/exec/java-exec/src/test/java/org/apache/drill/test/rowSet/test/TestVariableWidthWriter.java b/exec/java-exec/src/test/java/org/apache/drill/test/rowSet/test/TestVariableWidthWriter.java
index 5e5e029..2b93b42 100644
--- a/exec/java-exec/src/test/java/org/apache/drill/test/rowSet/test/TestVariableWidthWriter.java
+++ b/exec/java-exec/src/test/java/org/apache/drill/test/rowSet/test/TestVariableWidthWriter.java
@@ -382,7 +382,7 @@ public class TestVariableWidthWriter extends SubOperatorTest {
byte value[] = new byte[423];
Arrays.fill(value, (byte) 'X');
try {
- for (int i = 0; ; i++ ) {
+ for (int i = 0;; i++ ) {
index.index = i;
writer.startRow();
writer.setBytes(value, value.length);
diff --git a/exec/jdbc-all/src/test/java/org/apache/drill/jdbc/ITTestShadedJar.java b/exec/jdbc-all/src/test/java/org/apache/drill/jdbc/ITTestShadedJar.java
index f73a1b9..6319cb3 100644
--- a/exec/jdbc-all/src/test/java/org/apache/drill/jdbc/ITTestShadedJar.java
+++ b/exec/jdbc-all/src/test/java/org/apache/drill/jdbc/ITTestShadedJar.java
@@ -175,7 +175,8 @@ public class ITTestShadedJar {
private static void printQuery(Connection c, String query) throws SQLException {
final StringBuilder sb = new StringBuilder();
- try (Statement s = c.createStatement(); ResultSet result = s.executeQuery(query)) {
+ try (Statement s = c.createStatement();
+ ResultSet result = s.executeQuery(query)) {
while (result.next()) {
final int columnCount = result.getMetaData().getColumnCount();
for(int i = 1; i < columnCount+1; i++){
diff --git a/exec/jdbc/src/main/java/org/apache/drill/jdbc/impl/DrillColumnMetaDataList.java b/exec/jdbc/src/main/java/org/apache/drill/jdbc/impl/DrillColumnMetaDataList.java
index 90dcafe..27132f6 100644
--- a/exec/jdbc/src/main/java/org/apache/drill/jdbc/impl/DrillColumnMetaDataList.java
+++ b/exec/jdbc/src/main/java/org/apache/drill/jdbc/impl/DrillColumnMetaDataList.java
@@ -128,10 +128,16 @@ public class DrillColumnMetaDataList extends BasicList<ColumnMetaData>{
final int nullability;
switch ( field.getDataMode() ) {
- case OPTIONAL: nullability = ResultSetMetaData.columnNullable; break;
- case REQUIRED: nullability = ResultSetMetaData.columnNoNulls; break;
+ case OPTIONAL:
+ nullability = ResultSetMetaData.columnNullable;
+ break;
+ case REQUIRED:
+ nullability = ResultSetMetaData.columnNoNulls;
+ break;
// Should REPEATED still map to columnNoNulls? or to columnNullable?
- case REPEATED: nullability = ResultSetMetaData.columnNoNulls; break;
+ case REPEATED:
+ nullability = ResultSetMetaData.columnNoNulls;
+ break;
default:
throw new AssertionError( "Unexpected new DataMode value '"
+ field.getDataMode().name() + "'" );
diff --git a/exec/jdbc/src/main/java/org/apache/drill/jdbc/proxy/ProxiesManager.java b/exec/jdbc/src/main/java/org/apache/drill/jdbc/proxy/ProxiesManager.java
index c014ebf..3d0bd94 100644
--- a/exec/jdbc/src/main/java/org/apache/drill/jdbc/proxy/ProxiesManager.java
+++ b/exec/jdbc/src/main/java/org/apache/drill/jdbc/proxy/ProxiesManager.java
@@ -107,8 +107,7 @@ class ProxiesManager {
catch ( InstantiationException | IllegalAccessException
| IllegalArgumentException | InvocationTargetException
| NoSuchMethodException | SecurityException e ) {
- throw new RuntimeException(
- "Error creating proxy for " + declaredType + ": " + e , e );
+ throw new RuntimeException("Error creating proxy for " + declaredType + ": " + e, e);
}
}
return proxyInstance;
diff --git a/exec/jdbc/src/test/java/org/apache/drill/jdbc/PreparedStatementTest.java b/exec/jdbc/src/test/java/org/apache/drill/jdbc/PreparedStatementTest.java
index 94155e4..6d2d195 100644
--- a/exec/jdbc/src/test/java/org/apache/drill/jdbc/PreparedStatementTest.java
+++ b/exec/jdbc/src/test/java/org/apache/drill/jdbc/PreparedStatementTest.java
@@ -258,7 +258,7 @@ public class PreparedStatementTest extends JdbcTestBase {
public void testDefaultGetQueryTimeout() throws SQLException {
try (PreparedStatement stmt = connection.prepareStatement(SYS_VERSION_SQL)) {
int timeoutValue = stmt.getQueryTimeout();
- assertEquals( 0L , timeoutValue );
+ assertEquals(0L, timeoutValue);
}
}
@@ -288,7 +288,7 @@ public class PreparedStatementTest extends JdbcTestBase {
int valueToSet = new Random(20150304).nextInt(59)+1;
logger.info("Setting timeout as {} seconds", valueToSet);
stmt.setQueryTimeout(valueToSet);
- assertEquals( valueToSet , stmt.getQueryTimeout() );
+ assertEquals(valueToSet, stmt.getQueryTimeout());
}
}
@@ -306,7 +306,7 @@ public class PreparedStatementTest extends JdbcTestBase {
rs.getBytes(1);
rowCount++;
}
- assertEquals( 3 , rowCount );
+ assertEquals(3, rowCount);
}
}
@@ -406,7 +406,7 @@ public class PreparedStatementTest extends JdbcTestBase {
rs.getBytes(1);
rowCount++;
}
- assertEquals( 1 , rowCount );
+ assertEquals(1, rowCount);
}
}
diff --git a/exec/jdbc/src/test/java/org/apache/drill/jdbc/StatementTest.java b/exec/jdbc/src/test/java/org/apache/drill/jdbc/StatementTest.java
index 9f4b9ad..73757f4 100644
--- a/exec/jdbc/src/test/java/org/apache/drill/jdbc/StatementTest.java
+++ b/exec/jdbc/src/test/java/org/apache/drill/jdbc/StatementTest.java
@@ -83,7 +83,7 @@ public class StatementTest extends JdbcTestBase {
public void testDefaultGetQueryTimeout() throws SQLException {
try(Statement stmt = connection.createStatement()) {
int timeoutValue = stmt.getQueryTimeout();
- assertEquals( 0 , timeoutValue );
+ assertEquals(0, timeoutValue);
}
}
@@ -116,7 +116,7 @@ public class StatementTest extends JdbcTestBase {
int valueToSet = new Random(20150304).nextInt(59)+1;
logger.info("Setting timeout as {} seconds", valueToSet);
stmt.setQueryTimeout(valueToSet);
- assertEquals( valueToSet , stmt.getQueryTimeout() );
+ assertEquals( valueToSet, stmt.getQueryTimeout() );
}
}
@@ -135,7 +135,7 @@ public class StatementTest extends JdbcTestBase {
rs.getBytes(1);
rowCount++;
}
- assertEquals( 3 , rowCount );
+ assertEquals( 3, rowCount );
}
}
@@ -235,7 +235,7 @@ public class StatementTest extends JdbcTestBase {
rs.getBytes(1);
rowCount++;
}
- assertEquals( 1 , rowCount );
+ assertEquals( 1, rowCount );
}
}
diff --git a/exec/memory/base/src/main/java/org/apache/drill/exec/memory/BaseAllocator.java b/exec/memory/base/src/main/java/org/apache/drill/exec/memory/BaseAllocator.java
index eb51dc3..893024a 100644
--- a/exec/memory/base/src/main/java/org/apache/drill/exec/memory/BaseAllocator.java
+++ b/exec/memory/base/src/main/java/org/apache/drill/exec/memory/BaseAllocator.java
@@ -788,8 +788,7 @@ public abstract class BaseAllocator extends Accountant implements BufferAllocato
public static enum Verbosity {
BASIC(false, false), // only include basic information
LOG(true, false), // include basic
- LOG_WITH_STACKTRACE(true, true) //
- ;
+ LOG_WITH_STACKTRACE(true, true);
public final boolean includeHistoricalLog;
public final boolean includeStackTraces;
diff --git a/exec/rpc/src/main/java/org/apache/drill/exec/rpc/RpcBus.java b/exec/rpc/src/main/java/org/apache/drill/exec/rpc/RpcBus.java
index deed7a7..fd27558 100644
--- a/exec/rpc/src/main/java/org/apache/drill/exec/rpc/RpcBus.java
+++ b/exec/rpc/src/main/java/org/apache/drill/exec/rpc/RpcBus.java
@@ -135,7 +135,6 @@ public abstract class RpcBus<T extends EnumLite, C extends RemoteConnection> imp
}
}
- ;
}
}
diff --git a/logical/src/main/java/org/apache/drill/common/expression/BooleanOperator.java b/logical/src/main/java/org/apache/drill/common/expression/BooleanOperator.java
index b107bac..831821e 100644
--- a/logical/src/main/java/org/apache/drill/common/expression/BooleanOperator.java
+++ b/logical/src/main/java/org/apache/drill/common/expression/BooleanOperator.java
@@ -59,7 +59,7 @@ public class BooleanOperator extends FunctionCall{
i++;
}
- return (int) (cost / i) ;
+ return cost / i;
}
}
diff --git a/logical/src/main/java/org/apache/drill/common/expression/IfExpression.java b/logical/src/main/java/org/apache/drill/common/expression/IfExpression.java
index 147129b..6a3dbe5 100644
--- a/logical/src/main/java/org/apache/drill/common/expression/IfExpression.java
+++ b/logical/src/main/java/org/apache/drill/common/expression/IfExpression.java
@@ -153,7 +153,7 @@ public class IfExpression extends LogicalExpressionBase {
i++;
}
- return (int) (cost / i) ;
+ return cost / i;
}
}
diff --git a/logical/src/main/java/org/apache/drill/common/expression/NullExpression.java b/logical/src/main/java/org/apache/drill/common/expression/NullExpression.java
index a346504..b36f3e6 100644
--- a/logical/src/main/java/org/apache/drill/common/expression/NullExpression.java
+++ b/logical/src/main/java/org/apache/drill/common/expression/NullExpression.java
@@ -53,7 +53,7 @@ public class NullExpression implements LogicalExpression{
return Iterators.emptyIterator();
}
- public int getSelfCost() { return 0 ; }
+ public int getSelfCost() { return 0; }
public int getCumulativeCost() { return 0; }
diff --git a/pom.xml b/pom.xml
index 691d9fe..165e7e2 100644
--- a/pom.xml
+++ b/pom.xml
@@ -397,7 +397,7 @@
<dependency>
<groupId>com.puppycrawl.tools</groupId>
<artifactId>checkstyle</artifactId>
- <version>5.9</version>
+ <version>7.8.2</version>
</dependency>
</dependencies>
<configuration>
diff --git a/src/main/resources/checkstyle-config.xml b/src/main/resources/checkstyle-config.xml
index 35628d3..1ee4878 100644
--- a/src/main/resources/checkstyle-config.xml
+++ b/src/main/resources/checkstyle-config.xml
@@ -36,6 +36,10 @@
<module name="AvoidStarImport"/>
<module name="NeedBraces"/>
<module name="PackageDeclaration"/>
+ <module name="EmptyStatement"/>
+ <module name="NoWhitespaceBefore"/>
+ <module name="OneStatementPerLine"/>
+
</module>
<module name="FileTabCharacter"/>