You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@iceberg.apache.org by bl...@apache.org on 2019/04/04 16:24:06 UTC
[incubator-iceberg] branch master updated: Add missing override
annotations (#147)
This is an automated email from the ASF dual-hosted git repository.
blue pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/incubator-iceberg.git
The following commit(s) were added to refs/heads/master by this push:
new ecd35af Add missing override annotations (#147)
ecd35af is described below
commit ecd35afa436f093656fb66c981a51a56ed687af0
Author: lys0716 <lu...@gmail.com>
AuthorDate: Thu Apr 4 09:24:01 2019 -0700
Add missing override annotations (#147)
---
.../java/org/apache/iceberg/expressions/ExpressionVisitors.java | 2 ++
.../org/apache/iceberg/expressions/SerializationProxies.java | 1 +
api/src/main/java/org/apache/iceberg/transforms/Bucket.java | 9 +++++++++
api/src/main/java/org/apache/iceberg/types/Type.java | 4 ++++
common/src/main/java/org/apache/iceberg/common/DynFields.java | 2 ++
common/src/main/java/org/apache/iceberg/common/DynMethods.java | 1 +
core/src/main/java/org/apache/iceberg/BaseTableScan.java | 1 +
core/src/main/java/org/apache/iceberg/PartitionData.java | 1 +
.../java/org/apache/iceberg/hadoop/HadoopTableOperations.java | 1 +
core/src/main/java/org/apache/iceberg/util/BinPacking.java | 2 ++
.../src/main/java/org/apache/iceberg/parquet/ParquetFilters.java | 1 +
.../java/org/apache/iceberg/spark/data/SparkParquetReaders.java | 4 ++--
spark/src/main/java/org/apache/iceberg/spark/source/Writer.java | 1 +
13 files changed, 28 insertions(+), 2 deletions(-)
diff --git a/api/src/main/java/org/apache/iceberg/expressions/ExpressionVisitors.java b/api/src/main/java/org/apache/iceberg/expressions/ExpressionVisitors.java
index 3e6b791..cc45531 100644
--- a/api/src/main/java/org/apache/iceberg/expressions/ExpressionVisitors.java
+++ b/api/src/main/java/org/apache/iceberg/expressions/ExpressionVisitors.java
@@ -97,6 +97,7 @@ public class ExpressionVisitors {
return null;
}
+ @Override
public <T> R predicate(BoundPredicate<T> pred) {
switch (pred.op()) {
case IS_NULL:
@@ -125,6 +126,7 @@ public class ExpressionVisitors {
}
}
+ @Override
public <T> R predicate(UnboundPredicate<T> pred) {
throw new UnsupportedOperationException("Not a bound predicate: " + pred);
}
diff --git a/api/src/main/java/org/apache/iceberg/expressions/SerializationProxies.java b/api/src/main/java/org/apache/iceberg/expressions/SerializationProxies.java
index 6fd7416..bf93a3f 100644
--- a/api/src/main/java/org/apache/iceberg/expressions/SerializationProxies.java
+++ b/api/src/main/java/org/apache/iceberg/expressions/SerializationProxies.java
@@ -62,6 +62,7 @@ class SerializationProxies {
super(buffer);
}
+ @Override
Object readResolve() throws ObjectStreamException {
return new Literals.BinaryLiteral(ByteBuffer.wrap(bytes()));
}
diff --git a/api/src/main/java/org/apache/iceberg/transforms/Bucket.java b/api/src/main/java/org/apache/iceberg/transforms/Bucket.java
index af62a98..4d3bd96 100644
--- a/api/src/main/java/org/apache/iceberg/transforms/Bucket.java
+++ b/api/src/main/java/org/apache/iceberg/transforms/Bucket.java
@@ -145,6 +145,7 @@ abstract class Bucket<T> implements Transform<T, Integer> {
super(numBuckets);
}
+ @Override
public int hash(Integer value) {
return MURMUR3.hashLong(value.longValue()).asInt();
}
@@ -160,6 +161,7 @@ abstract class Bucket<T> implements Transform<T, Integer> {
super(numBuckets);
}
+ @Override
public int hash(Long value) {
return MURMUR3.hashLong(value).asInt();
}
@@ -180,6 +182,7 @@ abstract class Bucket<T> implements Transform<T, Integer> {
super(numBuckets);
}
+ @Override
public int hash(Float value) {
return MURMUR3.hashLong(Double.doubleToRawLongBits((double) value)).asInt();
}
@@ -197,6 +200,7 @@ abstract class Bucket<T> implements Transform<T, Integer> {
super(numBuckets);
}
+ @Override
public int hash(Double value) {
return MURMUR3.hashLong(Double.doubleToRawLongBits(value)).asInt();
}
@@ -212,6 +216,7 @@ abstract class Bucket<T> implements Transform<T, Integer> {
super(numBuckets);
}
+ @Override
public int hash(CharSequence value) {
return MURMUR3.hashString(value, StandardCharsets.UTF_8).asInt();
}
@@ -230,6 +235,7 @@ abstract class Bucket<T> implements Transform<T, Integer> {
super(numBuckets);
}
+ @Override
public int hash(byte[] value) {
return MURMUR3.hashBytes(value).asInt();
}
@@ -248,6 +254,7 @@ abstract class Bucket<T> implements Transform<T, Integer> {
super(numBuckets);
}
+ @Override
public int hash(ByteBuffer value) {
if (value.hasArray()) {
return MURMUR3.hashBytes(value.array(),
@@ -283,6 +290,7 @@ abstract class Bucket<T> implements Transform<T, Integer> {
super(numBuckets);
}
+ @Override
public int hash(UUID value) {
ByteBuffer buffer = BUFFER.get();
buffer.rewind();
@@ -302,6 +310,7 @@ abstract class Bucket<T> implements Transform<T, Integer> {
super(numBuckets);
}
+ @Override
public int hash(BigDecimal value) {
return MURMUR3.hashBytes(value.unscaledValue().toByteArray()).asInt();
}
diff --git a/api/src/main/java/org/apache/iceberg/types/Type.java b/api/src/main/java/org/apache/iceberg/types/Type.java
index 9eac853..7c7f97e 100644
--- a/api/src/main/java/org/apache/iceberg/types/Type.java
+++ b/api/src/main/java/org/apache/iceberg/types/Type.java
@@ -98,10 +98,12 @@ public interface Type extends Serializable {
}
abstract class PrimitiveType implements Type {
+ @Override
public boolean isPrimitiveType() {
return true;
}
+ @Override
public PrimitiveType asPrimitiveType() {
return this;
}
@@ -112,10 +114,12 @@ public interface Type extends Serializable {
}
abstract class NestedType implements Type {
+ @Override
public boolean isNestedType() {
return true;
}
+ @Override
public NestedType asNestedType() {
return this;
}
diff --git a/common/src/main/java/org/apache/iceberg/common/DynFields.java b/common/src/main/java/org/apache/iceberg/common/DynFields.java
index 19cdc92..e2fedbe 100644
--- a/common/src/main/java/org/apache/iceberg/common/DynFields.java
+++ b/common/src/main/java/org/apache/iceberg/common/DynFields.java
@@ -65,6 +65,7 @@ public class DynFields {
}
}
+ @Override
public String toString() {
return Objects.toStringHelper(this)
.add("class", field.getDeclaringClass().toString())
@@ -134,6 +135,7 @@ public class DynFields {
public void set(Object target, Void value) {
}
+ @Override
public String toString() {
return "Field(AlwaysNull)";
}
diff --git a/common/src/main/java/org/apache/iceberg/common/DynMethods.java b/common/src/main/java/org/apache/iceberg/common/DynMethods.java
index 6c5242a..0a64e20 100644
--- a/common/src/main/java/org/apache/iceberg/common/DynMethods.java
+++ b/common/src/main/java/org/apache/iceberg/common/DynMethods.java
@@ -122,6 +122,7 @@ public class DynMethods {
return new StaticMethod(this);
}
+ @Override
public String toString() {
return "DynMethods.UnboundMethod(name=" + name +" method=" +
method.toGenericString() + ")";
diff --git a/core/src/main/java/org/apache/iceberg/BaseTableScan.java b/core/src/main/java/org/apache/iceberg/BaseTableScan.java
index 6ec9dac..74e7e28 100644
--- a/core/src/main/java/org/apache/iceberg/BaseTableScan.java
+++ b/core/src/main/java/org/apache/iceberg/BaseTableScan.java
@@ -126,6 +126,7 @@ class BaseTableScan implements TableScan {
return useSnapshot(lastSnapshotId);
}
+ @Override
public TableScan project(Schema schema) {
return new BaseTableScan(ops, table, snapshotId, schema, rowFilter, caseSensitive, selectedColumns);
}
diff --git a/core/src/main/java/org/apache/iceberg/PartitionData.java b/core/src/main/java/org/apache/iceberg/PartitionData.java
index e638c7c..614cf19 100644
--- a/core/src/main/java/org/apache/iceberg/PartitionData.java
+++ b/core/src/main/java/org/apache/iceberg/PartitionData.java
@@ -83,6 +83,7 @@ class PartitionData
return partitionType;
}
+ @Override
public Schema getSchema() {
if (schema == null) {
this.schema = new Schema.Parser().parse(stringSchema);
diff --git a/core/src/main/java/org/apache/iceberg/hadoop/HadoopTableOperations.java b/core/src/main/java/org/apache/iceberg/hadoop/HadoopTableOperations.java
index 9cd5dff..9d4bb68 100644
--- a/core/src/main/java/org/apache/iceberg/hadoop/HadoopTableOperations.java
+++ b/core/src/main/java/org/apache/iceberg/hadoop/HadoopTableOperations.java
@@ -64,6 +64,7 @@ public class HadoopTableOperations implements TableOperations {
this.location = location;
}
+ @Override
public TableMetadata current() {
if (shouldRefresh) {
return refresh();
diff --git a/core/src/main/java/org/apache/iceberg/util/BinPacking.java b/core/src/main/java/org/apache/iceberg/util/BinPacking.java
index 7573437..8c6a6a5 100644
--- a/core/src/main/java/org/apache/iceberg/util/BinPacking.java
+++ b/core/src/main/java/org/apache/iceberg/util/BinPacking.java
@@ -95,10 +95,12 @@ public class BinPacking {
this.largestBinFirst = largestBinFirst;
}
+ @Override
public boolean hasNext() {
return items.hasNext() || !bins.isEmpty();
}
+ @Override
public List<T> next() {
while (items.hasNext()) {
T item = items.next();
diff --git a/parquet/src/main/java/org/apache/iceberg/parquet/ParquetFilters.java b/parquet/src/main/java/org/apache/iceberg/parquet/ParquetFilters.java
index c856433..84bfe2a 100644
--- a/parquet/src/main/java/org/apache/iceberg/parquet/ParquetFilters.java
+++ b/parquet/src/main/java/org/apache/iceberg/parquet/ParquetFilters.java
@@ -186,6 +186,7 @@ class ParquetFilters {
this.partitionStruct = schema.findField(column).type().asNestedType().asStructType();
}
+ @Override
protected Expression bind(UnboundPredicate<?> pred) {
// instead of binding the predicate using the top-level schema, bind it to the partition data
return pred.bind(partitionStruct, true);
diff --git a/spark/src/main/java/org/apache/iceberg/spark/data/SparkParquetReaders.java b/spark/src/main/java/org/apache/iceberg/spark/data/SparkParquetReaders.java
index 294a849..e01f2c2 100644
--- a/spark/src/main/java/org/apache/iceberg/spark/data/SparkParquetReaders.java
+++ b/spark/src/main/java/org/apache/iceberg/spark/data/SparkParquetReaders.java
@@ -648,12 +648,12 @@ public class SparkParquetReaders {
return Arrays.copyOfRange(values, 0, numElements);
}
-// @Override
+ @Override
public void setNullAt(int i) {
values[i] = null;
}
-// @Override
+ @Override
public void update(int ordinal, Object value) {
values[ordinal] = value;
}
diff --git a/spark/src/main/java/org/apache/iceberg/spark/source/Writer.java b/spark/src/main/java/org/apache/iceberg/spark/source/Writer.java
index 1d1ec01..4124136 100644
--- a/spark/src/main/java/org/apache/iceberg/spark/source/Writer.java
+++ b/spark/src/main/java/org/apache/iceberg/spark/source/Writer.java
@@ -207,6 +207,7 @@ class Writer implements DataSourceWriter {
}
private class SparkAppenderFactory implements AppenderFactory<InternalRow> {
+ @Override
public FileAppender<InternalRow> newAppender(OutputFile file, FileFormat format) {
Schema schema = spec.schema();
try {