You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hive.apache.org by ha...@apache.org on 2013/09/10 17:08:30 UTC
svn commit: r1521520 [2/2] - in /hive/trunk: build.properties
hcatalog/pom.xml ivy/libraries.properties
ql/src/gen/protobuf/gen-java/org/apache/hadoop/hive/ql/io/orc/OrcProto.java
Modified: hive/trunk/ql/src/gen/protobuf/gen-java/org/apache/hadoop/hive/ql/io/orc/OrcProto.java
URL: http://svn.apache.org/viewvc/hive/trunk/ql/src/gen/protobuf/gen-java/org/apache/hadoop/hive/ql/io/orc/OrcProto.java?rev=1521520&r1=1521519&r2=1521520&view=diff
==============================================================================
--- hive/trunk/ql/src/gen/protobuf/gen-java/org/apache/hadoop/hive/ql/io/orc/OrcProto.java (original)
+++ hive/trunk/ql/src/gen/protobuf/gen-java/org/apache/hadoop/hive/ql/io/orc/OrcProto.java Tue Sep 10 15:08:29 2013
@@ -8,22 +8,49 @@ public final class OrcProto {
public static void registerAllExtensions(
com.google.protobuf.ExtensionRegistry registry) {
}
+ /**
+ * Protobuf enum {@code org.apache.hadoop.hive.ql.io.orc.CompressionKind}
+ */
public enum CompressionKind
implements com.google.protobuf.ProtocolMessageEnum {
+ /**
+ * <code>NONE = 0;</code>
+ */
NONE(0, 0),
+ /**
+ * <code>ZLIB = 1;</code>
+ */
ZLIB(1, 1),
+ /**
+ * <code>SNAPPY = 2;</code>
+ */
SNAPPY(2, 2),
+ /**
+ * <code>LZO = 3;</code>
+ */
LZO(3, 3),
;
-
+
+ /**
+ * <code>NONE = 0;</code>
+ */
public static final int NONE_VALUE = 0;
+ /**
+ * <code>ZLIB = 1;</code>
+ */
public static final int ZLIB_VALUE = 1;
+ /**
+ * <code>SNAPPY = 2;</code>
+ */
public static final int SNAPPY_VALUE = 2;
+ /**
+ * <code>LZO = 3;</code>
+ */
public static final int LZO_VALUE = 3;
-
-
+
+
public final int getNumber() { return value; }
-
+
public static CompressionKind valueOf(int value) {
switch (value) {
case 0: return NONE;
@@ -33,7 +60,7 @@ public final class OrcProto {
default: return null;
}
}
-
+
public static com.google.protobuf.Internal.EnumLiteMap<CompressionKind>
internalGetValueMap() {
return internalValueMap;
@@ -45,7 +72,7 @@ public final class OrcProto {
return CompressionKind.valueOf(number);
}
};
-
+
public final com.google.protobuf.Descriptors.EnumValueDescriptor
getValueDescriptor() {
return getDescriptor().getValues().get(index);
@@ -58,11 +85,9 @@ public final class OrcProto {
getDescriptor() {
return org.apache.hadoop.hive.ql.io.orc.OrcProto.getDescriptor().getEnumTypes().get(0);
}
-
- private static final CompressionKind[] VALUES = {
- NONE, ZLIB, SNAPPY, LZO,
- };
-
+
+ private static final CompressionKind[] VALUES = values();
+
public static CompressionKind valueOf(
com.google.protobuf.Descriptors.EnumValueDescriptor desc) {
if (desc.getType() != getDescriptor()) {
@@ -71,92 +96,205 @@ public final class OrcProto {
}
return VALUES[desc.getIndex()];
}
-
+
private final int index;
private final int value;
-
+
private CompressionKind(int index, int value) {
this.index = index;
this.value = value;
}
-
+
// @@protoc_insertion_point(enum_scope:org.apache.hadoop.hive.ql.io.orc.CompressionKind)
}
-
+
public interface IntegerStatisticsOrBuilder
extends com.google.protobuf.MessageOrBuilder {
-
+
// optional sint64 minimum = 1;
+ /**
+ * <code>optional sint64 minimum = 1;</code>
+ */
boolean hasMinimum();
+ /**
+ * <code>optional sint64 minimum = 1;</code>
+ */
long getMinimum();
-
+
// optional sint64 maximum = 2;
+ /**
+ * <code>optional sint64 maximum = 2;</code>
+ */
boolean hasMaximum();
+ /**
+ * <code>optional sint64 maximum = 2;</code>
+ */
long getMaximum();
-
+
// optional sint64 sum = 3;
+ /**
+ * <code>optional sint64 sum = 3;</code>
+ */
boolean hasSum();
+ /**
+ * <code>optional sint64 sum = 3;</code>
+ */
long getSum();
}
+ /**
+ * Protobuf type {@code org.apache.hadoop.hive.ql.io.orc.IntegerStatistics}
+ */
public static final class IntegerStatistics extends
com.google.protobuf.GeneratedMessage
implements IntegerStatisticsOrBuilder {
// Use IntegerStatistics.newBuilder() to construct.
- private IntegerStatistics(Builder builder) {
+ private IntegerStatistics(com.google.protobuf.GeneratedMessage.Builder<?> builder) {
super(builder);
+ this.unknownFields = builder.getUnknownFields();
}
- private IntegerStatistics(boolean noInit) {}
-
+ private IntegerStatistics(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
+
private static final IntegerStatistics defaultInstance;
public static IntegerStatistics getDefaultInstance() {
return defaultInstance;
}
-
+
public IntegerStatistics getDefaultInstanceForType() {
return defaultInstance;
}
-
+
+ private final com.google.protobuf.UnknownFieldSet unknownFields;
+ @java.lang.Override
+ public final com.google.protobuf.UnknownFieldSet
+ getUnknownFields() {
+ return this.unknownFields;
+ }
+ private IntegerStatistics(
+ com.google.protobuf.CodedInputStream input,
+ com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws com.google.protobuf.InvalidProtocolBufferException {
+ initFields();
+ int mutable_bitField0_ = 0;
+ com.google.protobuf.UnknownFieldSet.Builder unknownFields =
+ com.google.protobuf.UnknownFieldSet.newBuilder();
+ try {
+ boolean done = false;
+ while (!done) {
+ int tag = input.readTag();
+ switch (tag) {
+ case 0:
+ done = true;
+ break;
+ default: {
+ if (!parseUnknownField(input, unknownFields,
+ extensionRegistry, tag)) {
+ done = true;
+ }
+ break;
+ }
+ case 8: {
+ bitField0_ |= 0x00000001;
+ minimum_ = input.readSInt64();
+ break;
+ }
+ case 16: {
+ bitField0_ |= 0x00000002;
+ maximum_ = input.readSInt64();
+ break;
+ }
+ case 24: {
+ bitField0_ |= 0x00000004;
+ sum_ = input.readSInt64();
+ break;
+ }
+ }
+ }
+ } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ throw e.setUnfinishedMessage(this);
+ } catch (java.io.IOException e) {
+ throw new com.google.protobuf.InvalidProtocolBufferException(
+ e.getMessage()).setUnfinishedMessage(this);
+ } finally {
+ this.unknownFields = unknownFields.build();
+ makeExtensionsImmutable();
+ }
+ }
public static final com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hive.ql.io.orc.OrcProto.internal_static_org_apache_hadoop_hive_ql_io_orc_IntegerStatistics_descriptor;
}
-
+
protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
internalGetFieldAccessorTable() {
- return org.apache.hadoop.hive.ql.io.orc.OrcProto.internal_static_org_apache_hadoop_hive_ql_io_orc_IntegerStatistics_fieldAccessorTable;
+ return org.apache.hadoop.hive.ql.io.orc.OrcProto.internal_static_org_apache_hadoop_hive_ql_io_orc_IntegerStatistics_fieldAccessorTable
+ .ensureFieldAccessorsInitialized(
+ org.apache.hadoop.hive.ql.io.orc.OrcProto.IntegerStatistics.class, org.apache.hadoop.hive.ql.io.orc.OrcProto.IntegerStatistics.Builder.class);
+ }
+
+ public static com.google.protobuf.Parser<IntegerStatistics> PARSER =
+ new com.google.protobuf.AbstractParser<IntegerStatistics>() {
+ public IntegerStatistics parsePartialFrom(
+ com.google.protobuf.CodedInputStream input,
+ com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws com.google.protobuf.InvalidProtocolBufferException {
+ return new IntegerStatistics(input, extensionRegistry);
+ }
+ };
+
+ @java.lang.Override
+ public com.google.protobuf.Parser<IntegerStatistics> getParserForType() {
+ return PARSER;
}
-
+
private int bitField0_;
// optional sint64 minimum = 1;
public static final int MINIMUM_FIELD_NUMBER = 1;
private long minimum_;
+ /**
+ * <code>optional sint64 minimum = 1;</code>
+ */
public boolean hasMinimum() {
return ((bitField0_ & 0x00000001) == 0x00000001);
}
+ /**
+ * <code>optional sint64 minimum = 1;</code>
+ */
public long getMinimum() {
return minimum_;
}
-
+
// optional sint64 maximum = 2;
public static final int MAXIMUM_FIELD_NUMBER = 2;
private long maximum_;
+ /**
+ * <code>optional sint64 maximum = 2;</code>
+ */
public boolean hasMaximum() {
return ((bitField0_ & 0x00000002) == 0x00000002);
}
+ /**
+ * <code>optional sint64 maximum = 2;</code>
+ */
public long getMaximum() {
return maximum_;
}
-
+
// optional sint64 sum = 3;
public static final int SUM_FIELD_NUMBER = 3;
private long sum_;
+ /**
+ * <code>optional sint64 sum = 3;</code>
+ */
public boolean hasSum() {
return ((bitField0_ & 0x00000004) == 0x00000004);
}
+ /**
+ * <code>optional sint64 sum = 3;</code>
+ */
public long getSum() {
return sum_;
}
-
+
private void initFields() {
minimum_ = 0L;
maximum_ = 0L;
@@ -166,11 +304,11 @@ public final class OrcProto {
public final boolean isInitialized() {
byte isInitialized = memoizedIsInitialized;
if (isInitialized != -1) return isInitialized == 1;
-
+
memoizedIsInitialized = 1;
return true;
}
-
+
public void writeTo(com.google.protobuf.CodedOutputStream output)
throws java.io.IOException {
getSerializedSize();
@@ -185,12 +323,12 @@ public final class OrcProto {
}
getUnknownFields().writeTo(output);
}
-
+
private int memoizedSerializedSize = -1;
public int getSerializedSize() {
int size = memoizedSerializedSize;
if (size != -1) return size;
-
+
size = 0;
if (((bitField0_ & 0x00000001) == 0x00000001)) {
size += com.google.protobuf.CodedOutputStream
@@ -208,94 +346,83 @@ public final class OrcProto {
memoizedSerializedSize = size;
return size;
}
-
+
private static final long serialVersionUID = 0L;
@java.lang.Override
protected java.lang.Object writeReplace()
throws java.io.ObjectStreamException {
return super.writeReplace();
}
-
+
public static org.apache.hadoop.hive.ql.io.orc.OrcProto.IntegerStatistics parseFrom(
com.google.protobuf.ByteString data)
throws com.google.protobuf.InvalidProtocolBufferException {
- return newBuilder().mergeFrom(data).buildParsed();
+ return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hive.ql.io.orc.OrcProto.IntegerStatistics parseFrom(
com.google.protobuf.ByteString data,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws com.google.protobuf.InvalidProtocolBufferException {
- return newBuilder().mergeFrom(data, extensionRegistry)
- .buildParsed();
+ return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hive.ql.io.orc.OrcProto.IntegerStatistics parseFrom(byte[] data)
throws com.google.protobuf.InvalidProtocolBufferException {
- return newBuilder().mergeFrom(data).buildParsed();
+ return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hive.ql.io.orc.OrcProto.IntegerStatistics parseFrom(
byte[] data,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws com.google.protobuf.InvalidProtocolBufferException {
- return newBuilder().mergeFrom(data, extensionRegistry)
- .buildParsed();
+ return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hive.ql.io.orc.OrcProto.IntegerStatistics parseFrom(java.io.InputStream input)
throws java.io.IOException {
- return newBuilder().mergeFrom(input).buildParsed();
+ return PARSER.parseFrom(input);
}
public static org.apache.hadoop.hive.ql.io.orc.OrcProto.IntegerStatistics parseFrom(
java.io.InputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return newBuilder().mergeFrom(input, extensionRegistry)
- .buildParsed();
+ return PARSER.parseFrom(input, extensionRegistry);
}
public static org.apache.hadoop.hive.ql.io.orc.OrcProto.IntegerStatistics parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
- Builder builder = newBuilder();
- if (builder.mergeDelimitedFrom(input)) {
- return builder.buildParsed();
- } else {
- return null;
- }
+ return PARSER.parseDelimitedFrom(input);
}
public static org.apache.hadoop.hive.ql.io.orc.OrcProto.IntegerStatistics parseDelimitedFrom(
java.io.InputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- Builder builder = newBuilder();
- if (builder.mergeDelimitedFrom(input, extensionRegistry)) {
- return builder.buildParsed();
- } else {
- return null;
- }
+ return PARSER.parseDelimitedFrom(input, extensionRegistry);
}
public static org.apache.hadoop.hive.ql.io.orc.OrcProto.IntegerStatistics parseFrom(
com.google.protobuf.CodedInputStream input)
throws java.io.IOException {
- return newBuilder().mergeFrom(input).buildParsed();
+ return PARSER.parseFrom(input);
}
public static org.apache.hadoop.hive.ql.io.orc.OrcProto.IntegerStatistics parseFrom(
com.google.protobuf.CodedInputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return newBuilder().mergeFrom(input, extensionRegistry)
- .buildParsed();
+ return PARSER.parseFrom(input, extensionRegistry);
}
-
+
public static Builder newBuilder() { return Builder.create(); }
public Builder newBuilderForType() { return newBuilder(); }
public static Builder newBuilder(org.apache.hadoop.hive.ql.io.orc.OrcProto.IntegerStatistics prototype) {
return newBuilder().mergeFrom(prototype);
}
public Builder toBuilder() { return newBuilder(this); }
-
+
@java.lang.Override
protected Builder newBuilderForType(
com.google.protobuf.GeneratedMessage.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
+ /**
+ * Protobuf type {@code org.apache.hadoop.hive.ql.io.orc.IntegerStatistics}
+ */
public static final class Builder extends
com.google.protobuf.GeneratedMessage.Builder<Builder>
implements org.apache.hadoop.hive.ql.io.orc.OrcProto.IntegerStatisticsOrBuilder {
@@ -303,18 +430,21 @@ public final class OrcProto {
getDescriptor() {
return org.apache.hadoop.hive.ql.io.orc.OrcProto.internal_static_org_apache_hadoop_hive_ql_io_orc_IntegerStatistics_descriptor;
}
-
+
protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
internalGetFieldAccessorTable() {
- return org.apache.hadoop.hive.ql.io.orc.OrcProto.internal_static_org_apache_hadoop_hive_ql_io_orc_IntegerStatistics_fieldAccessorTable;
+ return org.apache.hadoop.hive.ql.io.orc.OrcProto.internal_static_org_apache_hadoop_hive_ql_io_orc_IntegerStatistics_fieldAccessorTable
+ .ensureFieldAccessorsInitialized(
+ org.apache.hadoop.hive.ql.io.orc.OrcProto.IntegerStatistics.class, org.apache.hadoop.hive.ql.io.orc.OrcProto.IntegerStatistics.Builder.class);
}
-
+
// Construct using org.apache.hadoop.hive.ql.io.orc.OrcProto.IntegerStatistics.newBuilder()
private Builder() {
maybeForceBuilderInitialization();
}
-
- private Builder(BuilderParent parent) {
+
+ private Builder(
+ com.google.protobuf.GeneratedMessage.BuilderParent parent) {
super(parent);
maybeForceBuilderInitialization();
}
@@ -325,7 +455,7 @@ public final class OrcProto {
private static Builder create() {
return new Builder();
}
-
+
public Builder clear() {
super.clear();
minimum_ = 0L;
@@ -336,20 +466,20 @@ public final class OrcProto {
bitField0_ = (bitField0_ & ~0x00000004);
return this;
}
-
+
public Builder clone() {
return create().mergeFrom(buildPartial());
}
-
+
public com.google.protobuf.Descriptors.Descriptor
getDescriptorForType() {
- return org.apache.hadoop.hive.ql.io.orc.OrcProto.IntegerStatistics.getDescriptor();
+ return org.apache.hadoop.hive.ql.io.orc.OrcProto.internal_static_org_apache_hadoop_hive_ql_io_orc_IntegerStatistics_descriptor;
}
-
+
public org.apache.hadoop.hive.ql.io.orc.OrcProto.IntegerStatistics getDefaultInstanceForType() {
return org.apache.hadoop.hive.ql.io.orc.OrcProto.IntegerStatistics.getDefaultInstance();
}
-
+
public org.apache.hadoop.hive.ql.io.orc.OrcProto.IntegerStatistics build() {
org.apache.hadoop.hive.ql.io.orc.OrcProto.IntegerStatistics result = buildPartial();
if (!result.isInitialized()) {
@@ -357,17 +487,7 @@ public final class OrcProto {
}
return result;
}
-
- private org.apache.hadoop.hive.ql.io.orc.OrcProto.IntegerStatistics buildParsed()
- throws com.google.protobuf.InvalidProtocolBufferException {
- org.apache.hadoop.hive.ql.io.orc.OrcProto.IntegerStatistics result = buildPartial();
- if (!result.isInitialized()) {
- throw newUninitializedMessageException(
- result).asInvalidProtocolBufferException();
- }
- return result;
- }
-
+
public org.apache.hadoop.hive.ql.io.orc.OrcProto.IntegerStatistics buildPartial() {
org.apache.hadoop.hive.ql.io.orc.OrcProto.IntegerStatistics result = new org.apache.hadoop.hive.ql.io.orc.OrcProto.IntegerStatistics(this);
int from_bitField0_ = bitField0_;
@@ -388,7 +508,7 @@ public final class OrcProto {
onBuilt();
return result;
}
-
+
public Builder mergeFrom(com.google.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hive.ql.io.orc.OrcProto.IntegerStatistics) {
return mergeFrom((org.apache.hadoop.hive.ql.io.orc.OrcProto.IntegerStatistics)other);
@@ -397,7 +517,7 @@ public final class OrcProto {
return this;
}
}
-
+
public Builder mergeFrom(org.apache.hadoop.hive.ql.io.orc.OrcProto.IntegerStatistics other) {
if (other == org.apache.hadoop.hive.ql.io.orc.OrcProto.IntegerStatistics.getDefaultInstance()) return this;
if (other.hasMinimum()) {
@@ -412,203 +532,327 @@ public final class OrcProto {
this.mergeUnknownFields(other.getUnknownFields());
return this;
}
-
+
public final boolean isInitialized() {
return true;
}
-
+
public Builder mergeFrom(
com.google.protobuf.CodedInputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- com.google.protobuf.UnknownFieldSet.Builder unknownFields =
- com.google.protobuf.UnknownFieldSet.newBuilder(
- this.getUnknownFields());
- while (true) {
- int tag = input.readTag();
- switch (tag) {
- case 0:
- this.setUnknownFields(unknownFields.build());
- onChanged();
- return this;
- default: {
- if (!parseUnknownField(input, unknownFields,
- extensionRegistry, tag)) {
- this.setUnknownFields(unknownFields.build());
- onChanged();
- return this;
- }
- break;
- }
- case 8: {
- bitField0_ |= 0x00000001;
- minimum_ = input.readSInt64();
- break;
- }
- case 16: {
- bitField0_ |= 0x00000002;
- maximum_ = input.readSInt64();
- break;
- }
- case 24: {
- bitField0_ |= 0x00000004;
- sum_ = input.readSInt64();
- break;
- }
+ org.apache.hadoop.hive.ql.io.orc.OrcProto.IntegerStatistics parsedMessage = null;
+ try {
+ parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
+ } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ parsedMessage = (org.apache.hadoop.hive.ql.io.orc.OrcProto.IntegerStatistics) e.getUnfinishedMessage();
+ throw e;
+ } finally {
+ if (parsedMessage != null) {
+ mergeFrom(parsedMessage);
}
}
+ return this;
}
-
private int bitField0_;
-
+
// optional sint64 minimum = 1;
private long minimum_ ;
+ /**
+ * <code>optional sint64 minimum = 1;</code>
+ */
public boolean hasMinimum() {
return ((bitField0_ & 0x00000001) == 0x00000001);
}
+ /**
+ * <code>optional sint64 minimum = 1;</code>
+ */
public long getMinimum() {
return minimum_;
}
+ /**
+ * <code>optional sint64 minimum = 1;</code>
+ */
public Builder setMinimum(long value) {
bitField0_ |= 0x00000001;
minimum_ = value;
onChanged();
return this;
}
+ /**
+ * <code>optional sint64 minimum = 1;</code>
+ */
public Builder clearMinimum() {
bitField0_ = (bitField0_ & ~0x00000001);
minimum_ = 0L;
onChanged();
return this;
}
-
+
// optional sint64 maximum = 2;
private long maximum_ ;
+ /**
+ * <code>optional sint64 maximum = 2;</code>
+ */
public boolean hasMaximum() {
return ((bitField0_ & 0x00000002) == 0x00000002);
}
+ /**
+ * <code>optional sint64 maximum = 2;</code>
+ */
public long getMaximum() {
return maximum_;
}
+ /**
+ * <code>optional sint64 maximum = 2;</code>
+ */
public Builder setMaximum(long value) {
bitField0_ |= 0x00000002;
maximum_ = value;
onChanged();
return this;
}
+ /**
+ * <code>optional sint64 maximum = 2;</code>
+ */
public Builder clearMaximum() {
bitField0_ = (bitField0_ & ~0x00000002);
maximum_ = 0L;
onChanged();
return this;
}
-
+
// optional sint64 sum = 3;
private long sum_ ;
+ /**
+ * <code>optional sint64 sum = 3;</code>
+ */
public boolean hasSum() {
return ((bitField0_ & 0x00000004) == 0x00000004);
}
+ /**
+ * <code>optional sint64 sum = 3;</code>
+ */
public long getSum() {
return sum_;
}
+ /**
+ * <code>optional sint64 sum = 3;</code>
+ */
public Builder setSum(long value) {
bitField0_ |= 0x00000004;
sum_ = value;
onChanged();
return this;
}
+ /**
+ * <code>optional sint64 sum = 3;</code>
+ */
public Builder clearSum() {
bitField0_ = (bitField0_ & ~0x00000004);
sum_ = 0L;
onChanged();
return this;
}
-
+
// @@protoc_insertion_point(builder_scope:org.apache.hadoop.hive.ql.io.orc.IntegerStatistics)
}
-
+
static {
defaultInstance = new IntegerStatistics(true);
defaultInstance.initFields();
}
-
+
// @@protoc_insertion_point(class_scope:org.apache.hadoop.hive.ql.io.orc.IntegerStatistics)
}
-
+
public interface DoubleStatisticsOrBuilder
extends com.google.protobuf.MessageOrBuilder {
-
+
// optional double minimum = 1;
+ /**
+ * <code>optional double minimum = 1;</code>
+ */
boolean hasMinimum();
+ /**
+ * <code>optional double minimum = 1;</code>
+ */
double getMinimum();
-
+
// optional double maximum = 2;
+ /**
+ * <code>optional double maximum = 2;</code>
+ */
boolean hasMaximum();
+ /**
+ * <code>optional double maximum = 2;</code>
+ */
double getMaximum();
-
+
// optional double sum = 3;
+ /**
+ * <code>optional double sum = 3;</code>
+ */
boolean hasSum();
+ /**
+ * <code>optional double sum = 3;</code>
+ */
double getSum();
}
+ /**
+ * Protobuf type {@code org.apache.hadoop.hive.ql.io.orc.DoubleStatistics}
+ */
public static final class DoubleStatistics extends
com.google.protobuf.GeneratedMessage
implements DoubleStatisticsOrBuilder {
// Use DoubleStatistics.newBuilder() to construct.
- private DoubleStatistics(Builder builder) {
+ private DoubleStatistics(com.google.protobuf.GeneratedMessage.Builder<?> builder) {
super(builder);
+ this.unknownFields = builder.getUnknownFields();
}
- private DoubleStatistics(boolean noInit) {}
-
+ private DoubleStatistics(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
+
private static final DoubleStatistics defaultInstance;
public static DoubleStatistics getDefaultInstance() {
return defaultInstance;
}
-
+
public DoubleStatistics getDefaultInstanceForType() {
return defaultInstance;
}
-
+
+ private final com.google.protobuf.UnknownFieldSet unknownFields;
+ @java.lang.Override
+ public final com.google.protobuf.UnknownFieldSet
+ getUnknownFields() {
+ return this.unknownFields;
+ }
+ private DoubleStatistics(
+ com.google.protobuf.CodedInputStream input,
+ com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws com.google.protobuf.InvalidProtocolBufferException {
+ initFields();
+ int mutable_bitField0_ = 0;
+ com.google.protobuf.UnknownFieldSet.Builder unknownFields =
+ com.google.protobuf.UnknownFieldSet.newBuilder();
+ try {
+ boolean done = false;
+ while (!done) {
+ int tag = input.readTag();
+ switch (tag) {
+ case 0:
+ done = true;
+ break;
+ default: {
+ if (!parseUnknownField(input, unknownFields,
+ extensionRegistry, tag)) {
+ done = true;
+ }
+ break;
+ }
+ case 9: {
+ bitField0_ |= 0x00000001;
+ minimum_ = input.readDouble();
+ break;
+ }
+ case 17: {
+ bitField0_ |= 0x00000002;
+ maximum_ = input.readDouble();
+ break;
+ }
+ case 25: {
+ bitField0_ |= 0x00000004;
+ sum_ = input.readDouble();
+ break;
+ }
+ }
+ }
+ } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ throw e.setUnfinishedMessage(this);
+ } catch (java.io.IOException e) {
+ throw new com.google.protobuf.InvalidProtocolBufferException(
+ e.getMessage()).setUnfinishedMessage(this);
+ } finally {
+ this.unknownFields = unknownFields.build();
+ makeExtensionsImmutable();
+ }
+ }
public static final com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hive.ql.io.orc.OrcProto.internal_static_org_apache_hadoop_hive_ql_io_orc_DoubleStatistics_descriptor;
}
-
+
protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
internalGetFieldAccessorTable() {
- return org.apache.hadoop.hive.ql.io.orc.OrcProto.internal_static_org_apache_hadoop_hive_ql_io_orc_DoubleStatistics_fieldAccessorTable;
+ return org.apache.hadoop.hive.ql.io.orc.OrcProto.internal_static_org_apache_hadoop_hive_ql_io_orc_DoubleStatistics_fieldAccessorTable
+ .ensureFieldAccessorsInitialized(
+ org.apache.hadoop.hive.ql.io.orc.OrcProto.DoubleStatistics.class, org.apache.hadoop.hive.ql.io.orc.OrcProto.DoubleStatistics.Builder.class);
+ }
+
+ public static com.google.protobuf.Parser<DoubleStatistics> PARSER =
+ new com.google.protobuf.AbstractParser<DoubleStatistics>() {
+ public DoubleStatistics parsePartialFrom(
+ com.google.protobuf.CodedInputStream input,
+ com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws com.google.protobuf.InvalidProtocolBufferException {
+ return new DoubleStatistics(input, extensionRegistry);
+ }
+ };
+
+ @java.lang.Override
+ public com.google.protobuf.Parser<DoubleStatistics> getParserForType() {
+ return PARSER;
}
-
+
private int bitField0_;
// optional double minimum = 1;
public static final int MINIMUM_FIELD_NUMBER = 1;
private double minimum_;
+ /**
+ * <code>optional double minimum = 1;</code>
+ */
public boolean hasMinimum() {
return ((bitField0_ & 0x00000001) == 0x00000001);
}
+ /**
+ * <code>optional double minimum = 1;</code>
+ */
public double getMinimum() {
return minimum_;
}
-
+
// optional double maximum = 2;
public static final int MAXIMUM_FIELD_NUMBER = 2;
private double maximum_;
+ /**
+ * <code>optional double maximum = 2;</code>
+ */
public boolean hasMaximum() {
return ((bitField0_ & 0x00000002) == 0x00000002);
}
+ /**
+ * <code>optional double maximum = 2;</code>
+ */
public double getMaximum() {
return maximum_;
}
-
+
// optional double sum = 3;
public static final int SUM_FIELD_NUMBER = 3;
private double sum_;
+ /**
+ * <code>optional double sum = 3;</code>
+ */
public boolean hasSum() {
return ((bitField0_ & 0x00000004) == 0x00000004);
}
+ /**
+ * <code>optional double sum = 3;</code>
+ */
public double getSum() {
return sum_;
}
-
+
private void initFields() {
minimum_ = 0D;
maximum_ = 0D;
@@ -618,11 +862,11 @@ public final class OrcProto {
public final boolean isInitialized() {
byte isInitialized = memoizedIsInitialized;
if (isInitialized != -1) return isInitialized == 1;
-
+
memoizedIsInitialized = 1;
return true;
}
-
+
public void writeTo(com.google.protobuf.CodedOutputStream output)
throws java.io.IOException {
getSerializedSize();
@@ -637,12 +881,12 @@ public final class OrcProto {
}
getUnknownFields().writeTo(output);
}
-
+
private int memoizedSerializedSize = -1;
public int getSerializedSize() {
int size = memoizedSerializedSize;
if (size != -1) return size;
-
+
size = 0;
if (((bitField0_ & 0x00000001) == 0x00000001)) {
size += com.google.protobuf.CodedOutputStream
@@ -660,94 +904,83 @@ public final class OrcProto {
memoizedSerializedSize = size;
return size;
}
-
+
private static final long serialVersionUID = 0L;
@java.lang.Override
protected java.lang.Object writeReplace()
throws java.io.ObjectStreamException {
return super.writeReplace();
}
-
+
public static org.apache.hadoop.hive.ql.io.orc.OrcProto.DoubleStatistics parseFrom(
com.google.protobuf.ByteString data)
throws com.google.protobuf.InvalidProtocolBufferException {
- return newBuilder().mergeFrom(data).buildParsed();
+ return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hive.ql.io.orc.OrcProto.DoubleStatistics parseFrom(
com.google.protobuf.ByteString data,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws com.google.protobuf.InvalidProtocolBufferException {
- return newBuilder().mergeFrom(data, extensionRegistry)
- .buildParsed();
+ return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hive.ql.io.orc.OrcProto.DoubleStatistics parseFrom(byte[] data)
throws com.google.protobuf.InvalidProtocolBufferException {
- return newBuilder().mergeFrom(data).buildParsed();
+ return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hive.ql.io.orc.OrcProto.DoubleStatistics parseFrom(
byte[] data,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws com.google.protobuf.InvalidProtocolBufferException {
- return newBuilder().mergeFrom(data, extensionRegistry)
- .buildParsed();
+ return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hive.ql.io.orc.OrcProto.DoubleStatistics parseFrom(java.io.InputStream input)
throws java.io.IOException {
- return newBuilder().mergeFrom(input).buildParsed();
+ return PARSER.parseFrom(input);
}
public static org.apache.hadoop.hive.ql.io.orc.OrcProto.DoubleStatistics parseFrom(
java.io.InputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return newBuilder().mergeFrom(input, extensionRegistry)
- .buildParsed();
+ return PARSER.parseFrom(input, extensionRegistry);
}
public static org.apache.hadoop.hive.ql.io.orc.OrcProto.DoubleStatistics parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
- Builder builder = newBuilder();
- if (builder.mergeDelimitedFrom(input)) {
- return builder.buildParsed();
- } else {
- return null;
- }
+ return PARSER.parseDelimitedFrom(input);
}
public static org.apache.hadoop.hive.ql.io.orc.OrcProto.DoubleStatistics parseDelimitedFrom(
java.io.InputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- Builder builder = newBuilder();
- if (builder.mergeDelimitedFrom(input, extensionRegistry)) {
- return builder.buildParsed();
- } else {
- return null;
- }
+ return PARSER.parseDelimitedFrom(input, extensionRegistry);
}
public static org.apache.hadoop.hive.ql.io.orc.OrcProto.DoubleStatistics parseFrom(
com.google.protobuf.CodedInputStream input)
throws java.io.IOException {
- return newBuilder().mergeFrom(input).buildParsed();
+ return PARSER.parseFrom(input);
}
public static org.apache.hadoop.hive.ql.io.orc.OrcProto.DoubleStatistics parseFrom(
com.google.protobuf.CodedInputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return newBuilder().mergeFrom(input, extensionRegistry)
- .buildParsed();
+ return PARSER.parseFrom(input, extensionRegistry);
}
-
+
public static Builder newBuilder() { return Builder.create(); }
public Builder newBuilderForType() { return newBuilder(); }
public static Builder newBuilder(org.apache.hadoop.hive.ql.io.orc.OrcProto.DoubleStatistics prototype) {
return newBuilder().mergeFrom(prototype);
}
public Builder toBuilder() { return newBuilder(this); }
-
+
@java.lang.Override
protected Builder newBuilderForType(
com.google.protobuf.GeneratedMessage.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
+ /**
+ * Protobuf type {@code org.apache.hadoop.hive.ql.io.orc.DoubleStatistics}
+ */
public static final class Builder extends
com.google.protobuf.GeneratedMessage.Builder<Builder>
implements org.apache.hadoop.hive.ql.io.orc.OrcProto.DoubleStatisticsOrBuilder {
@@ -755,18 +988,21 @@ public final class OrcProto {
getDescriptor() {
return org.apache.hadoop.hive.ql.io.orc.OrcProto.internal_static_org_apache_hadoop_hive_ql_io_orc_DoubleStatistics_descriptor;
}
-
+
protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
internalGetFieldAccessorTable() {
- return org.apache.hadoop.hive.ql.io.orc.OrcProto.internal_static_org_apache_hadoop_hive_ql_io_orc_DoubleStatistics_fieldAccessorTable;
+ return org.apache.hadoop.hive.ql.io.orc.OrcProto.internal_static_org_apache_hadoop_hive_ql_io_orc_DoubleStatistics_fieldAccessorTable
+ .ensureFieldAccessorsInitialized(
+ org.apache.hadoop.hive.ql.io.orc.OrcProto.DoubleStatistics.class, org.apache.hadoop.hive.ql.io.orc.OrcProto.DoubleStatistics.Builder.class);
}
-
+
// Construct using org.apache.hadoop.hive.ql.io.orc.OrcProto.DoubleStatistics.newBuilder()
private Builder() {
maybeForceBuilderInitialization();
}
-
- private Builder(BuilderParent parent) {
+
+ private Builder(
+ com.google.protobuf.GeneratedMessage.BuilderParent parent) {
super(parent);
maybeForceBuilderInitialization();
}
@@ -777,7 +1013,7 @@ public final class OrcProto {
private static Builder create() {
return new Builder();
}
-
+
public Builder clear() {
super.clear();
minimum_ = 0D;
@@ -788,20 +1024,20 @@ public final class OrcProto {
bitField0_ = (bitField0_ & ~0x00000004);
return this;
}
-
+
public Builder clone() {
return create().mergeFrom(buildPartial());
}
-
+
public com.google.protobuf.Descriptors.Descriptor
getDescriptorForType() {
- return org.apache.hadoop.hive.ql.io.orc.OrcProto.DoubleStatistics.getDescriptor();
+ return org.apache.hadoop.hive.ql.io.orc.OrcProto.internal_static_org_apache_hadoop_hive_ql_io_orc_DoubleStatistics_descriptor;
}
-
+
public org.apache.hadoop.hive.ql.io.orc.OrcProto.DoubleStatistics getDefaultInstanceForType() {
return org.apache.hadoop.hive.ql.io.orc.OrcProto.DoubleStatistics.getDefaultInstance();
}
-
+
public org.apache.hadoop.hive.ql.io.orc.OrcProto.DoubleStatistics build() {
org.apache.hadoop.hive.ql.io.orc.OrcProto.DoubleStatistics result = buildPartial();
if (!result.isInitialized()) {
@@ -809,17 +1045,7 @@ public final class OrcProto {
}
return result;
}
-
- private org.apache.hadoop.hive.ql.io.orc.OrcProto.DoubleStatistics buildParsed()
- throws com.google.protobuf.InvalidProtocolBufferException {
- org.apache.hadoop.hive.ql.io.orc.OrcProto.DoubleStatistics result = buildPartial();
- if (!result.isInitialized()) {
- throw newUninitializedMessageException(
- result).asInvalidProtocolBufferException();
- }
- return result;
- }
-
+
public org.apache.hadoop.hive.ql.io.orc.OrcProto.DoubleStatistics buildPartial() {
org.apache.hadoop.hive.ql.io.orc.OrcProto.DoubleStatistics result = new org.apache.hadoop.hive.ql.io.orc.OrcProto.DoubleStatistics(this);
int from_bitField0_ = bitField0_;
@@ -840,7 +1066,7 @@ public final class OrcProto {
onBuilt();
return result;
}
-
+
public Builder mergeFrom(com.google.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hive.ql.io.orc.OrcProto.DoubleStatistics) {
return mergeFrom((org.apache.hadoop.hive.ql.io.orc.OrcProto.DoubleStatistics)other);
@@ -849,7 +1075,7 @@ public final class OrcProto {
return this;
}
}
-
+
public Builder mergeFrom(org.apache.hadoop.hive.ql.io.orc.OrcProto.DoubleStatistics other) {
if (other == org.apache.hadoop.hive.ql.io.orc.OrcProto.DoubleStatistics.getDefaultInstance()) return this;
if (other.hasMinimum()) {
@@ -864,233 +1090,360 @@ public final class OrcProto {
this.mergeUnknownFields(other.getUnknownFields());
return this;
}
-
+
public final boolean isInitialized() {
return true;
}
-
+
public Builder mergeFrom(
com.google.protobuf.CodedInputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- com.google.protobuf.UnknownFieldSet.Builder unknownFields =
- com.google.protobuf.UnknownFieldSet.newBuilder(
- this.getUnknownFields());
- while (true) {
- int tag = input.readTag();
- switch (tag) {
- case 0:
- this.setUnknownFields(unknownFields.build());
- onChanged();
- return this;
- default: {
- if (!parseUnknownField(input, unknownFields,
- extensionRegistry, tag)) {
- this.setUnknownFields(unknownFields.build());
- onChanged();
- return this;
- }
- break;
- }
- case 9: {
- bitField0_ |= 0x00000001;
- minimum_ = input.readDouble();
- break;
- }
- case 17: {
- bitField0_ |= 0x00000002;
- maximum_ = input.readDouble();
- break;
- }
- case 25: {
- bitField0_ |= 0x00000004;
- sum_ = input.readDouble();
- break;
- }
+ org.apache.hadoop.hive.ql.io.orc.OrcProto.DoubleStatistics parsedMessage = null;
+ try {
+ parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
+ } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ parsedMessage = (org.apache.hadoop.hive.ql.io.orc.OrcProto.DoubleStatistics) e.getUnfinishedMessage();
+ throw e;
+ } finally {
+ if (parsedMessage != null) {
+ mergeFrom(parsedMessage);
}
}
+ return this;
}
-
private int bitField0_;
-
+
// optional double minimum = 1;
private double minimum_ ;
+ /**
+ * <code>optional double minimum = 1;</code>
+ */
public boolean hasMinimum() {
return ((bitField0_ & 0x00000001) == 0x00000001);
}
+ /**
+ * <code>optional double minimum = 1;</code>
+ */
public double getMinimum() {
return minimum_;
}
+ /**
+ * <code>optional double minimum = 1;</code>
+ */
public Builder setMinimum(double value) {
bitField0_ |= 0x00000001;
minimum_ = value;
onChanged();
return this;
}
+ /**
+ * <code>optional double minimum = 1;</code>
+ */
public Builder clearMinimum() {
bitField0_ = (bitField0_ & ~0x00000001);
minimum_ = 0D;
onChanged();
return this;
}
-
+
// optional double maximum = 2;
private double maximum_ ;
+ /**
+ * <code>optional double maximum = 2;</code>
+ */
public boolean hasMaximum() {
return ((bitField0_ & 0x00000002) == 0x00000002);
}
+ /**
+ * <code>optional double maximum = 2;</code>
+ */
public double getMaximum() {
return maximum_;
}
+ /**
+ * <code>optional double maximum = 2;</code>
+ */
public Builder setMaximum(double value) {
bitField0_ |= 0x00000002;
maximum_ = value;
onChanged();
return this;
}
+ /**
+ * <code>optional double maximum = 2;</code>
+ */
public Builder clearMaximum() {
bitField0_ = (bitField0_ & ~0x00000002);
maximum_ = 0D;
onChanged();
return this;
}
-
+
// optional double sum = 3;
private double sum_ ;
+ /**
+ * <code>optional double sum = 3;</code>
+ */
public boolean hasSum() {
return ((bitField0_ & 0x00000004) == 0x00000004);
}
+ /**
+ * <code>optional double sum = 3;</code>
+ */
public double getSum() {
return sum_;
}
+ /**
+ * <code>optional double sum = 3;</code>
+ */
public Builder setSum(double value) {
bitField0_ |= 0x00000004;
sum_ = value;
onChanged();
return this;
}
+ /**
+ * <code>optional double sum = 3;</code>
+ */
public Builder clearSum() {
bitField0_ = (bitField0_ & ~0x00000004);
sum_ = 0D;
onChanged();
return this;
}
-
+
// @@protoc_insertion_point(builder_scope:org.apache.hadoop.hive.ql.io.orc.DoubleStatistics)
}
-
+
static {
defaultInstance = new DoubleStatistics(true);
defaultInstance.initFields();
}
-
+
// @@protoc_insertion_point(class_scope:org.apache.hadoop.hive.ql.io.orc.DoubleStatistics)
}
-
+
public interface StringStatisticsOrBuilder
extends com.google.protobuf.MessageOrBuilder {
-
+
// optional string minimum = 1;
+ /**
+ * <code>optional string minimum = 1;</code>
+ */
boolean hasMinimum();
- String getMinimum();
-
+ /**
+ * <code>optional string minimum = 1;</code>
+ */
+ java.lang.String getMinimum();
+ /**
+ * <code>optional string minimum = 1;</code>
+ */
+ com.google.protobuf.ByteString
+ getMinimumBytes();
+
// optional string maximum = 2;
+ /**
+ * <code>optional string maximum = 2;</code>
+ */
boolean hasMaximum();
- String getMaximum();
- }
+ /**
+ * <code>optional string maximum = 2;</code>
+ */
+ java.lang.String getMaximum();
+ /**
+ * <code>optional string maximum = 2;</code>
+ */
+ com.google.protobuf.ByteString
+ getMaximumBytes();
+ }
+ /**
+ * Protobuf type {@code org.apache.hadoop.hive.ql.io.orc.StringStatistics}
+ */
public static final class StringStatistics extends
com.google.protobuf.GeneratedMessage
implements StringStatisticsOrBuilder {
// Use StringStatistics.newBuilder() to construct.
- private StringStatistics(Builder builder) {
+ private StringStatistics(com.google.protobuf.GeneratedMessage.Builder<?> builder) {
super(builder);
+ this.unknownFields = builder.getUnknownFields();
}
- private StringStatistics(boolean noInit) {}
-
+ private StringStatistics(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
+
private static final StringStatistics defaultInstance;
public static StringStatistics getDefaultInstance() {
return defaultInstance;
}
-
+
public StringStatistics getDefaultInstanceForType() {
return defaultInstance;
}
-
+
+ private final com.google.protobuf.UnknownFieldSet unknownFields;
+ @java.lang.Override
+ public final com.google.protobuf.UnknownFieldSet
+ getUnknownFields() {
+ return this.unknownFields;
+ }
+ private StringStatistics(
+ com.google.protobuf.CodedInputStream input,
+ com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws com.google.protobuf.InvalidProtocolBufferException {
+ initFields();
+ int mutable_bitField0_ = 0;
+ com.google.protobuf.UnknownFieldSet.Builder unknownFields =
+ com.google.protobuf.UnknownFieldSet.newBuilder();
+ try {
+ boolean done = false;
+ while (!done) {
+ int tag = input.readTag();
+ switch (tag) {
+ case 0:
+ done = true;
+ break;
+ default: {
+ if (!parseUnknownField(input, unknownFields,
+ extensionRegistry, tag)) {
+ done = true;
+ }
+ break;
+ }
+ case 10: {
+ bitField0_ |= 0x00000001;
+ minimum_ = input.readBytes();
+ break;
+ }
+ case 18: {
+ bitField0_ |= 0x00000002;
+ maximum_ = input.readBytes();
+ break;
+ }
+ }
+ }
+ } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ throw e.setUnfinishedMessage(this);
+ } catch (java.io.IOException e) {
+ throw new com.google.protobuf.InvalidProtocolBufferException(
+ e.getMessage()).setUnfinishedMessage(this);
+ } finally {
+ this.unknownFields = unknownFields.build();
+ makeExtensionsImmutable();
+ }
+ }
public static final com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hive.ql.io.orc.OrcProto.internal_static_org_apache_hadoop_hive_ql_io_orc_StringStatistics_descriptor;
}
-
+
protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
internalGetFieldAccessorTable() {
- return org.apache.hadoop.hive.ql.io.orc.OrcProto.internal_static_org_apache_hadoop_hive_ql_io_orc_StringStatistics_fieldAccessorTable;
+ return org.apache.hadoop.hive.ql.io.orc.OrcProto.internal_static_org_apache_hadoop_hive_ql_io_orc_StringStatistics_fieldAccessorTable
+ .ensureFieldAccessorsInitialized(
+ org.apache.hadoop.hive.ql.io.orc.OrcProto.StringStatistics.class, org.apache.hadoop.hive.ql.io.orc.OrcProto.StringStatistics.Builder.class);
+ }
+
+ public static com.google.protobuf.Parser<StringStatistics> PARSER =
+ new com.google.protobuf.AbstractParser<StringStatistics>() {
+ public StringStatistics parsePartialFrom(
+ com.google.protobuf.CodedInputStream input,
+ com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws com.google.protobuf.InvalidProtocolBufferException {
+ return new StringStatistics(input, extensionRegistry);
+ }
+ };
+
+ @java.lang.Override
+ public com.google.protobuf.Parser<StringStatistics> getParserForType() {
+ return PARSER;
}
-
+
private int bitField0_;
// optional string minimum = 1;
public static final int MINIMUM_FIELD_NUMBER = 1;
private java.lang.Object minimum_;
+ /**
+ * <code>optional string minimum = 1;</code>
+ */
public boolean hasMinimum() {
return ((bitField0_ & 0x00000001) == 0x00000001);
}
- public String getMinimum() {
+ /**
+ * <code>optional string minimum = 1;</code>
+ */
+ public java.lang.String getMinimum() {
java.lang.Object ref = minimum_;
- if (ref instanceof String) {
- return (String) ref;
+ if (ref instanceof java.lang.String) {
+ return (java.lang.String) ref;
} else {
com.google.protobuf.ByteString bs =
(com.google.protobuf.ByteString) ref;
- String s = bs.toStringUtf8();
- if (com.google.protobuf.Internal.isValidUtf8(bs)) {
+ java.lang.String s = bs.toStringUtf8();
+ if (bs.isValidUtf8()) {
minimum_ = s;
}
return s;
}
}
- private com.google.protobuf.ByteString getMinimumBytes() {
+ /**
+ * <code>optional string minimum = 1;</code>
+ */
+ public com.google.protobuf.ByteString
+ getMinimumBytes() {
java.lang.Object ref = minimum_;
- if (ref instanceof String) {
+ if (ref instanceof java.lang.String) {
com.google.protobuf.ByteString b =
- com.google.protobuf.ByteString.copyFromUtf8((String) ref);
+ com.google.protobuf.ByteString.copyFromUtf8(
+ (java.lang.String) ref);
minimum_ = b;
return b;
} else {
return (com.google.protobuf.ByteString) ref;
}
}
-
+
// optional string maximum = 2;
public static final int MAXIMUM_FIELD_NUMBER = 2;
private java.lang.Object maximum_;
+ /**
+ * <code>optional string maximum = 2;</code>
+ */
public boolean hasMaximum() {
return ((bitField0_ & 0x00000002) == 0x00000002);
}
- public String getMaximum() {
+ /**
+ * <code>optional string maximum = 2;</code>
+ */
+ public java.lang.String getMaximum() {
java.lang.Object ref = maximum_;
- if (ref instanceof String) {
- return (String) ref;
+ if (ref instanceof java.lang.String) {
+ return (java.lang.String) ref;
} else {
com.google.protobuf.ByteString bs =
(com.google.protobuf.ByteString) ref;
- String s = bs.toStringUtf8();
- if (com.google.protobuf.Internal.isValidUtf8(bs)) {
+ java.lang.String s = bs.toStringUtf8();
+ if (bs.isValidUtf8()) {
maximum_ = s;
}
return s;
}
}
- private com.google.protobuf.ByteString getMaximumBytes() {
+ /**
+ * <code>optional string maximum = 2;</code>
+ */
+ public com.google.protobuf.ByteString
+ getMaximumBytes() {
java.lang.Object ref = maximum_;
- if (ref instanceof String) {
+ if (ref instanceof java.lang.String) {
com.google.protobuf.ByteString b =
- com.google.protobuf.ByteString.copyFromUtf8((String) ref);
+ com.google.protobuf.ByteString.copyFromUtf8(
+ (java.lang.String) ref);
maximum_ = b;
return b;
} else {
return (com.google.protobuf.ByteString) ref;
}
}
-
+
private void initFields() {
minimum_ = "";
maximum_ = "";
@@ -1099,11 +1452,11 @@ public final class OrcProto {
public final boolean isInitialized() {
byte isInitialized = memoizedIsInitialized;
if (isInitialized != -1) return isInitialized == 1;
-
+
memoizedIsInitialized = 1;
return true;
}
-
+
public void writeTo(com.google.protobuf.CodedOutputStream output)
throws java.io.IOException {
getSerializedSize();
@@ -1115,12 +1468,12 @@ public final class OrcProto {
}
getUnknownFields().writeTo(output);
}
-
+
private int memoizedSerializedSize = -1;
public int getSerializedSize() {
int size = memoizedSerializedSize;
if (size != -1) return size;
-
+
size = 0;
if (((bitField0_ & 0x00000001) == 0x00000001)) {
size += com.google.protobuf.CodedOutputStream
@@ -1134,94 +1487,83 @@ public final class OrcProto {
memoizedSerializedSize = size;
return size;
}
-
+
private static final long serialVersionUID = 0L;
@java.lang.Override
protected java.lang.Object writeReplace()
throws java.io.ObjectStreamException {
return super.writeReplace();
}
-
+
public static org.apache.hadoop.hive.ql.io.orc.OrcProto.StringStatistics parseFrom(
com.google.protobuf.ByteString data)
throws com.google.protobuf.InvalidProtocolBufferException {
- return newBuilder().mergeFrom(data).buildParsed();
+ return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hive.ql.io.orc.OrcProto.StringStatistics parseFrom(
com.google.protobuf.ByteString data,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws com.google.protobuf.InvalidProtocolBufferException {
- return newBuilder().mergeFrom(data, extensionRegistry)
- .buildParsed();
+ return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hive.ql.io.orc.OrcProto.StringStatistics parseFrom(byte[] data)
throws com.google.protobuf.InvalidProtocolBufferException {
- return newBuilder().mergeFrom(data).buildParsed();
+ return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hive.ql.io.orc.OrcProto.StringStatistics parseFrom(
byte[] data,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws com.google.protobuf.InvalidProtocolBufferException {
- return newBuilder().mergeFrom(data, extensionRegistry)
- .buildParsed();
+ return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hive.ql.io.orc.OrcProto.StringStatistics parseFrom(java.io.InputStream input)
throws java.io.IOException {
- return newBuilder().mergeFrom(input).buildParsed();
+ return PARSER.parseFrom(input);
}
public static org.apache.hadoop.hive.ql.io.orc.OrcProto.StringStatistics parseFrom(
java.io.InputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return newBuilder().mergeFrom(input, extensionRegistry)
- .buildParsed();
+ return PARSER.parseFrom(input, extensionRegistry);
}
public static org.apache.hadoop.hive.ql.io.orc.OrcProto.StringStatistics parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
- Builder builder = newBuilder();
- if (builder.mergeDelimitedFrom(input)) {
- return builder.buildParsed();
- } else {
- return null;
- }
+ return PARSER.parseDelimitedFrom(input);
}
public static org.apache.hadoop.hive.ql.io.orc.OrcProto.StringStatistics parseDelimitedFrom(
java.io.InputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- Builder builder = newBuilder();
- if (builder.mergeDelimitedFrom(input, extensionRegistry)) {
- return builder.buildParsed();
- } else {
- return null;
- }
+ return PARSER.parseDelimitedFrom(input, extensionRegistry);
}
public static org.apache.hadoop.hive.ql.io.orc.OrcProto.StringStatistics parseFrom(
com.google.protobuf.CodedInputStream input)
throws java.io.IOException {
- return newBuilder().mergeFrom(input).buildParsed();
+ return PARSER.parseFrom(input);
}
public static org.apache.hadoop.hive.ql.io.orc.OrcProto.StringStatistics parseFrom(
com.google.protobuf.CodedInputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return newBuilder().mergeFrom(input, extensionRegistry)
- .buildParsed();
+ return PARSER.parseFrom(input, extensionRegistry);
}
-
+
public static Builder newBuilder() { return Builder.create(); }
public Builder newBuilderForType() { return newBuilder(); }
public static Builder newBuilder(org.apache.hadoop.hive.ql.io.orc.OrcProto.StringStatistics prototype) {
return newBuilder().mergeFrom(prototype);
}
public Builder toBuilder() { return newBuilder(this); }
-
+
@java.lang.Override
protected Builder newBuilderForType(
com.google.protobuf.GeneratedMessage.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
+ /**
+ * Protobuf type {@code org.apache.hadoop.hive.ql.io.orc.StringStatistics}
+ */
public static final class Builder extends
com.google.protobuf.GeneratedMessage.Builder<Builder>
implements org.apache.hadoop.hive.ql.io.orc.OrcProto.StringStatisticsOrBuilder {
@@ -1229,18 +1571,21 @@ public final class OrcProto {
getDescriptor() {
return org.apache.hadoop.hive.ql.io.orc.OrcProto.internal_static_org_apache_hadoop_hive_ql_io_orc_StringStatistics_descriptor;
}
-
+
protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
internalGetFieldAccessorTable() {
- return org.apache.hadoop.hive.ql.io.orc.OrcProto.internal_static_org_apache_hadoop_hive_ql_io_orc_StringStatistics_fieldAccessorTable;
+ return org.apache.hadoop.hive.ql.io.orc.OrcProto.internal_static_org_apache_hadoop_hive_ql_io_orc_StringStatistics_fieldAccessorTable
+ .ensureFieldAccessorsInitialized(
+ org.apache.hadoop.hive.ql.io.orc.OrcProto.StringStatistics.class, org.apache.hadoop.hive.ql.io.orc.OrcProto.StringStatistics.Builder.class);
}
-
+
// Construct using org.apache.hadoop.hive.ql.io.orc.OrcProto.StringStatistics.newBuilder()
private Builder() {
maybeForceBuilderInitialization();
}
-
- private Builder(BuilderParent parent) {
+
+ private Builder(
+ com.google.protobuf.GeneratedMessage.BuilderParent parent) {
super(parent);
maybeForceBuilderInitialization();
}
@@ -1251,7 +1596,7 @@ public final class OrcProto {
private static Builder create() {
return new Builder();
}
-
+
public Builder clear() {
super.clear();
minimum_ = "";
@@ -1260,20 +1605,20 @@ public final class OrcProto {
bitField0_ = (bitField0_ & ~0x00000002);
return this;
}
-
+
public Builder clone() {
return create().mergeFrom(buildPartial());
}
-
+
public com.google.protobuf.Descriptors.Descriptor
getDescriptorForType() {
- return org.apache.hadoop.hive.ql.io.orc.OrcProto.StringStatistics.getDescriptor();
+ return org.apache.hadoop.hive.ql.io.orc.OrcProto.internal_static_org_apache_hadoop_hive_ql_io_orc_StringStatistics_descriptor;
}
-
+
public org.apache.hadoop.hive.ql.io.orc.OrcProto.StringStatistics getDefaultInstanceForType() {
return org.apache.hadoop.hive.ql.io.orc.OrcProto.StringStatistics.getDefaultInstance();
}
-
+
public org.apache.hadoop.hive.ql.io.orc.OrcProto.StringStatistics build() {
org.apache.hadoop.hive.ql.io.orc.OrcProto.StringStatistics result = buildPartial();
if (!result.isInitialized()) {
@@ -1281,17 +1626,7 @@ public final class OrcProto {
}
return result;
}
-
- private org.apache.hadoop.hive.ql.io.orc.OrcProto.StringStatistics buildParsed()
- throws com.google.protobuf.InvalidProtocolBufferException {
- org.apache.hadoop.hive.ql.io.orc.OrcProto.StringStatistics result = buildPartial();
- if (!result.isInitialized()) {
- throw newUninitializedMessageException(
- result).asInvalidProtocolBufferException();
- }
- return result;
- }
-
+
public org.apache.hadoop.hive.ql.io.orc.OrcProto.StringStatistics buildPartial() {
org.apache.hadoop.hive.ql.io.orc.OrcProto.StringStatistics result = new org.apache.hadoop.hive.ql.io.orc.OrcProto.StringStatistics(this);
int from_bitField0_ = bitField0_;
@@ -1308,7 +1643,7 @@ public final class OrcProto {
onBuilt();
return result;
}
-
+
public Builder mergeFrom(com.google.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hive.ql.io.orc.OrcProto.StringStatistics) {
return mergeFrom((org.apache.hadoop.hive.ql.io.orc.OrcProto.StringStatistics)other);
@@ -1317,78 +1652,89 @@ public final class OrcProto {
return this;
}
}
-
+
public Builder mergeFrom(org.apache.hadoop.hive.ql.io.orc.OrcProto.StringStatistics other) {
if (other == org.apache.hadoop.hive.ql.io.orc.OrcProto.StringStatistics.getDefaultInstance()) return this;
if (other.hasMinimum()) {
- setMinimum(other.getMinimum());
+ bitField0_ |= 0x00000001;
+ minimum_ = other.minimum_;
+ onChanged();
}
if (other.hasMaximum()) {
- setMaximum(other.getMaximum());
+ bitField0_ |= 0x00000002;
+ maximum_ = other.maximum_;
+ onChanged();
}
this.mergeUnknownFields(other.getUnknownFields());
return this;
}
-
+
public final boolean isInitialized() {
return true;
}
-
+
public Builder mergeFrom(
com.google.protobuf.CodedInputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- com.google.protobuf.UnknownFieldSet.Builder unknownFields =
- com.google.protobuf.UnknownFieldSet.newBuilder(
- this.getUnknownFields());
- while (true) {
- int tag = input.readTag();
- switch (tag) {
- case 0:
- this.setUnknownFields(unknownFields.build());
- onChanged();
- return this;
- default: {
- if (!parseUnknownField(input, unknownFields,
- extensionRegistry, tag)) {
- this.setUnknownFields(unknownFields.build());
- onChanged();
- return this;
- }
- break;
- }
- case 10: {
- bitField0_ |= 0x00000001;
- minimum_ = input.readBytes();
- break;
- }
- case 18: {
- bitField0_ |= 0x00000002;
- maximum_ = input.readBytes();
- break;
- }
+ org.apache.hadoop.hive.ql.io.orc.OrcProto.StringStatistics parsedMessage = null;
+ try {
+ parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
+ } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ parsedMessage = (org.apache.hadoop.hive.ql.io.orc.OrcProto.StringStatistics) e.getUnfinishedMessage();
+ throw e;
+ } finally {
+ if (parsedMessage != null) {
+ mergeFrom(parsedMessage);
}
}
+ return this;
}
-
private int bitField0_;
-
+
// optional string minimum = 1;
private java.lang.Object minimum_ = "";
+ /**
+ * <code>optional string minimum = 1;</code>
+ */
public boolean hasMinimum() {
return ((bitField0_ & 0x00000001) == 0x00000001);
}
- public String getMinimum() {
+ /**
+ * <code>optional string minimum = 1;</code>
+ */
+ public java.lang.String getMinimum() {
java.lang.Object ref = minimum_;
- if (!(ref instanceof String)) {
- String s = ((com.google.protobuf.ByteString) ref).toStringUtf8();
+ if (!(ref instanceof java.lang.String)) {
+ java.lang.String s = ((com.google.protobuf.ByteString) ref)
+ .toStringUtf8();
minimum_ = s;
return s;
} else {
- return (String) ref;
+ return (java.lang.String) ref;
+ }
+ }
+ /**
+ * <code>optional string minimum = 1;</code>
+ */
+ public com.google.protobuf.ByteString
+ getMinimumBytes() {
+ java.lang.Object ref = minimum_;
+ if (ref instanceof String) {
+ com.google.protobuf.ByteString b =
+ com.google.protobuf.ByteString.copyFromUtf8(
+ (java.lang.String) ref);
+ minimum_ = b;
+ return b;
+ } else {
+ return (com.google.protobuf.ByteString) ref;
}
}
- public Builder setMinimum(String value) {
+ /**
+ * <code>optional string minimum = 1;</code>
+ */
+ public Builder setMinimum(
+ java.lang.String value) {
if (value == null) {
throw new NullPointerException();
}
@@ -1397,34 +1743,72 @@ public final class OrcProto {
onChanged();
return this;
}
+ /**
+ * <code>optional string minimum = 1;</code>
+ */
public Builder clearMinimum() {
bitField0_ = (bitField0_ & ~0x00000001);
minimum_ = getDefaultInstance().getMinimum();
onChanged();
return this;
}
- void setMinimum(com.google.protobuf.ByteString value) {
- bitField0_ |= 0x00000001;
+ /**
+ * <code>optional string minimum = 1;</code>
+ */
+ public Builder setMinimumBytes(
+ com.google.protobuf.ByteString value) {
+ if (value == null) {
+ throw new NullPointerException();
+ }
+ bitField0_ |= 0x00000001;
minimum_ = value;
onChanged();
+ return this;
}
-
+
// optional string maximum = 2;
private java.lang.Object maximum_ = "";
+ /**
+ * <code>optional string maximum = 2;</code>
+ */
public boolean hasMaximum() {
return ((bitField0_ & 0x00000002) == 0x00000002);
}
- public String getMaximum() {
+ /**
+ * <code>optional string maximum = 2;</code>
+ */
+ public java.lang.String getMaximum() {
java.lang.Object ref = maximum_;
- if (!(ref instanceof String)) {
- String s = ((com.google.protobuf.ByteString) ref).toStringUtf8();
+ if (!(ref instanceof java.lang.String)) {
+ java.lang.String s = ((com.google.protobuf.ByteString) ref)
+ .toStringUtf8();
maximum_ = s;
return s;
} else {
- return (String) ref;
+ return (java.lang.String) ref;
+ }
+ }
+ /**
+ * <code>optional string maximum = 2;</code>
+ */
+ public com.google.protobuf.ByteString
+ getMaximumBytes() {
+ java.lang.Object ref = maximum_;
+ if (ref instanceof String) {
+ com.google.protobuf.ByteString b =
+ com.google.protobuf.ByteString.copyFromUtf8(
+ (java.lang.String) ref);
+ maximum_ = b;
+ return b;
+ } else {
+ return (com.google.protobuf.ByteString) ref;
}
}
- public Builder setMaximum(String value) {
+ /**
+ * <code>optional string maximum = 2;</code>
+ */
+ public Builder setMaximum(
+ java.lang.String value) {
if (value == null) {
throw new NullPointerException();
}
@@ -1433,92 +1817,207 @@ public final class OrcProto {
onChanged();
return this;
}
+ /**
+ * <code>optional string maximum = 2;</code>
+ */
public Builder clearMaximum() {
bitField0_ = (bitField0_ & ~0x00000002);
maximum_ = getDefaultInstance().getMaximum();
onChanged();
return this;
}
- void setMaximum(com.google.protobuf.ByteString value) {
- bitField0_ |= 0x00000002;
+ /**
+ * <code>optional string maximum = 2;</code>
+ */
+ public Builder setMaximumBytes(
+ com.google.protobuf.ByteString value) {
+ if (value == null) {
+ throw new NullPointerException();
+ }
+ bitField0_ |= 0x00000002;
maximum_ = value;
onChanged();
+ return this;
}
-
+
// @@protoc_insertion_point(builder_scope:org.apache.hadoop.hive.ql.io.orc.StringStatistics)
}
-
+
static {
defaultInstance = new StringStatistics(true);
defaultInstance.initFields();
}
-
+
// @@protoc_insertion_point(class_scope:org.apache.hadoop.hive.ql.io.orc.StringStatistics)
}
-
+
public interface BucketStatisticsOrBuilder
extends com.google.protobuf.MessageOrBuilder {
-
+
// repeated uint64 count = 1 [packed = true];
+ /**
+ * <code>repeated uint64 count = 1 [packed = true];</code>
+ */
java.util.List<java.lang.Long> getCountList();
+ /**
+ * <code>repeated uint64 count = 1 [packed = true];</code>
+ */
int getCountCount();
+ /**
+ * <code>repeated uint64 count = 1 [packed = true];</code>
+ */
long getCount(int index);
}
+ /**
+ * Protobuf type {@code org.apache.hadoop.hive.ql.io.orc.BucketStatistics}
+ */
public static final class BucketStatistics extends
com.google.protobuf.GeneratedMessage
implements BucketStatisticsOrBuilder {
// Use BucketStatistics.newBuilder() to construct.
- private BucketStatistics(Builder builder) {
+ private BucketStatistics(com.google.protobuf.GeneratedMessage.Builder<?> builder) {
super(builder);
+ this.unknownFields = builder.getUnknownFields();
}
- private BucketStatistics(boolean noInit) {}
-
+ private BucketStatistics(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
+
private static final BucketStatistics defaultInstance;
public static BucketStatistics getDefaultInstance() {
return defaultInstance;
}
-
+
public BucketStatistics getDefaultInstanceForType() {
return defaultInstance;
}
-
- public static final com.google.protobuf.Descriptors.Descriptor
- getDescriptor() {
- return org.apache.hadoop.hive.ql.io.orc.OrcProto.internal_static_org_apache_hadoop_hive_ql_io_orc_BucketStatistics_descriptor;
- }
-
- protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
- internalGetFieldAccessorTable() {
- return org.apache.hadoop.hive.ql.io.orc.OrcProto.internal_static_org_apache_hadoop_hive_ql_io_orc_BucketStatistics_fieldAccessorTable;
+
+ private final com.google.protobuf.UnknownFieldSet unknownFields;
+ @java.lang.Override
+ public final com.google.protobuf.UnknownFieldSet
+ getUnknownFields() {
+ return this.unknownFields;
}
-
- // repeated uint64 count = 1 [packed = true];
- public static final int COUNT_FIELD_NUMBER = 1;
- private java.util.List<java.lang.Long> count_;
- public java.util.List<java.lang.Long>
- getCountList() {
+ private BucketStatistics(
+ com.google.protobuf.CodedInputStream input,
+ com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws com.google.protobuf.InvalidProtocolBufferException {
+ initFields();
+ int mutable_bitField0_ = 0;
+ com.google.protobuf.UnknownFieldSet.Builder unknownFields =
+ com.google.protobuf.UnknownFieldSet.newBuilder();
+ try {
+ boolean done = false;
+ while (!done) {
+ int tag = input.readTag();
+ switch (tag) {
+ case 0:
+ done = true;
+ break;
+ default: {
+ if (!parseUnknownField(input, unknownFields,
+ extensionRegistry, tag)) {
+ done = true;
+ }
+ break;
+ }
+ case 8: {
+ if (!((mutable_bitField0_ & 0x00000001) == 0x00000001)) {
+ count_ = new java.util.ArrayList<java.lang.Long>();
+ mutable_bitField0_ |= 0x00000001;
+ }
+ count_.add(input.readUInt64());
+ break;
+ }
+ case 10: {
+ int length = input.readRawVarint32();
+ int limit = input.pushLimit(length);
+ if (!((mutable_bitField0_ & 0x00000001) == 0x00000001) && input.getBytesUntilLimit() > 0) {
+ count_ = new java.util.ArrayList<java.lang.Long>();
+ mutable_bitField0_ |= 0x00000001;
+ }
+ while (input.getBytesUntilLimit() > 0) {
+ count_.add(input.readUInt64());
+ }
+ input.popLimit(limit);
+ break;
+ }
+ }
+ }
+ } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ throw e.setUnfinishedMessage(this);
+ } catch (java.io.IOException e) {
+ throw new com.google.protobuf.InvalidProtocolBufferException(
+ e.getMessage()).setUnfinishedMessage(this);
+ } finally {
+ if (((mutable_bitField0_ & 0x00000001) == 0x00000001)) {
+ count_ = java.util.Collections.unmodifiableList(count_);
+ }
+ this.unknownFields = unknownFields.build();
+ makeExtensionsImmutable();
+ }
+ }
+ public static final com.google.protobuf.Descriptors.Descriptor
+ getDescriptor() {
+ return org.apache.hadoop.hive.ql.io.orc.OrcProto.internal_static_org_apache_hadoop_hive_ql_io_orc_BucketStatistics_descriptor;
+ }
+
+ protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
+ internalGetFieldAccessorTable() {
+ return org.apache.hadoop.hive.ql.io.orc.OrcProto.internal_static_org_apache_hadoop_hive_ql_io_orc_BucketStatistics_fieldAccessorTable
+ .ensureFieldAccessorsInitialized(
+ org.apache.hadoop.hive.ql.io.orc.OrcProto.BucketStatistics.class, org.apache.hadoop.hive.ql.io.orc.OrcProto.BucketStatistics.Builder.class);
+ }
+
+ public static com.google.protobuf.Parser<BucketStatistics> PARSER =
+ new com.google.protobuf.AbstractParser<BucketStatistics>() {
+ public BucketStatistics parsePartialFrom(
+ com.google.protobuf.CodedInputStream input,
+ com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws com.google.protobuf.InvalidProtocolBufferException {
+ return new BucketStatistics(input, extensionRegistry);
+ }
+ };
+
+ @java.lang.Override
+ public com.google.protobuf.Parser<BucketStatistics> getParserForType() {
+ return PARSER;
+ }
+
+ // repeated uint64 count = 1 [packed = true];
+ public static final int COUNT_FIELD_NUMBER = 1;
+ private java.util.List<java.lang.Long> count_;
+ /**
+ * <code>repeated uint64 count = 1 [packed = true];</code>
+ */
+ public java.util.List<java.lang.Long>
+ getCountList() {
return count_;
}
+ /**
+ * <code>repeated uint64 count = 1 [packed = true];</code>
+ */
public int getCountCount() {
return count_.size();
}
+ /**
+ * <code>repeated uint64 count = 1 [packed = true];</code>
+ */
public long getCount(int index) {
return count_.get(index);
}
private int countMemoizedSerializedSize = -1;
-
+
private void initFields() {
- count_ = java.util.Collections.emptyList();;
+ count_ = java.util.Collections.emptyList();
}
private byte memoizedIsInitialized = -1;
public final boolean isInitialized() {
byte isInitialized = memoizedIsInitialized;
if (isInitialized != -1) return isInitialized == 1;
-
+
memoizedIsInitialized = 1;
return true;
}
-
+
public void writeTo(com.google.protobuf.CodedOutputStream output)
throws java.io.IOException {
getSerializedSize();
@@ -1531,12 +2030,12 @@ public final class OrcProto {
}
getUnknownFields().writeTo(output);
}
-
+
private int memoizedSerializedSize = -1;
public int getSerializedSize() {
int size = memoizedSerializedSize;
if (size != -1) return size;
-
+
size = 0;
{
int dataSize = 0;
@@ -1556,94 +2055,83 @@ public final class OrcProto {
memoizedSerializedSize = size;
return size;
}
-
+
private static final long serialVersionUID = 0L;
@java.lang.Override
protected java.lang.Object writeReplace()
throws java.io.ObjectStreamException {
return super.writeReplace();
}
-
+
public static org.apache.hadoop.hive.ql.io.orc.OrcProto.BucketStatistics parseFrom(
com.google.protobuf.ByteString data)
throws com.google.protobuf.InvalidProtocolBufferException {
- return newBuilder().mergeFrom(data).buildParsed();
+ return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hive.ql.io.orc.OrcProto.BucketStatistics parseFrom(
com.google.protobuf.ByteString data,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws com.google.protobuf.InvalidProtocolBufferException {
- return newBuilder().mergeFrom(data, extensionRegistry)
- .buildParsed();
+ return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hive.ql.io.orc.OrcProto.BucketStatistics parseFrom(byte[] data)
throws com.google.protobuf.InvalidProtocolBufferException {
- return newBuilder().mergeFrom(data).buildParsed();
+ return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hive.ql.io.orc.OrcProto.BucketStatistics parseFrom(
byte[] data,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws com.google.protobuf.InvalidProtocolBufferException {
- return newBuilder().mergeFrom(data, extensionRegistry)
- .buildParsed();
+ return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hive.ql.io.orc.OrcProto.BucketStatistics parseFrom(java.io.InputStream input)
throws java.io.IOException {
- return newBuilder().mergeFrom(input).buildParsed();
+ return PARSER.parseFrom(input);
}
public static org.apache.hadoop.hive.ql.io.orc.OrcProto.BucketStatistics parseFrom(
java.io.InputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return newBuilder().mergeFrom(input, extensionRegistry)
- .buildParsed();
+ return PARSER.parseFrom(input, extensionRegistry);
}
public static org.apache.hadoop.hive.ql.io.orc.OrcProto.BucketStatistics parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
- Builder builder = newBuilder();
- if (builder.mergeDelimitedFrom(input)) {
- return builder.buildParsed();
- } else {
- return null;
- }
+ return PARSER.parseDelimitedFrom(input);
}
public static org.apache.hadoop.hive.ql.io.orc.OrcProto.BucketStatistics parseDelimitedFrom(
java.io.InputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- Builder builder = newBuilder();
- if (builder.mergeDelimitedFrom(input, extensionRegistry)) {
- return builder.buildParsed();
- } else {
- return null;
- }
+ return PARSER.parseDelimitedFrom(input, extensionRegistry);
}
public static org.apache.hadoop.hive.ql.io.orc.OrcProto.BucketStatistics parseFrom(
com.google.protobuf.CodedInputStream input)
throws java.io.IOException {
- return newBuilder().mergeFrom(input).buildParsed();
+ return PARSER.parseFrom(input);
}
public static org.apache.hadoop.hive.ql.io.orc.OrcProto.BucketStatistics parseFrom(
com.google.protobuf.CodedInputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return newBuilder().mergeFrom(input, extensionRegistry)
- .buildParsed();
+ return PARSER.parseFrom(input, extensionRegistry);
}
-
+
public static Builder newBuilder() { return Builder.create(); }
public Builder newBuilderForType() { return newBuilder(); }
public static Builder newBuilder(org.apache.hadoop.hive.ql.io.orc.OrcProto.BucketStatistics prototype) {
return newBuilder().mergeFrom(prototype);
}
public Builder toBuilder() { return newBuilder(this); }
-
+
@java.lang.Override
protected Builder newBuilderForType(
com.google.protobuf.GeneratedMessage.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
+ /**
+ * Protobuf type {@code org.apache.hadoop.hive.ql.io.orc.BucketStatistics}
+ */
public static final class Builder extends
com.google.protobuf.GeneratedMessage.Builder<Builder>
implements org.apache.hadoop.hive.ql.io.orc.OrcProto.BucketStatisticsOrBuilder {
@@ -1651,18 +2139,21 @@ public final class OrcProto {
getDescriptor() {
return org.apache.hadoop.hive.ql.io.orc.OrcProto.internal_static_org_apache_hadoop_hive_ql_io_orc_BucketStatistics_descriptor;
}
-
+
protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
internalGetFieldAccessorTable() {
- return org.apache.hadoop.hive.ql.io.orc.OrcProto.internal_static_org_apache_hadoop_hive_ql_io_orc_BucketStatistics_fieldAccessorTable;
+ return org.apache.hadoop.hive.ql.io.orc.OrcProto.internal_static_org_apache_hadoop_hive_ql_io_orc_BucketStatistics_fieldAccessorTable
+ .ensureFieldAccessorsInitialized(
+ org.apache.hadoop.hive.ql.io.orc.OrcProto.BucketStatistics.class, org.apache.hadoop.hive.ql.io.orc.OrcProto.BucketStatistics.Builder.class);
}
-
+
// Construct using org.apache.hadoop.hive.ql.io.orc.OrcProto.BucketStatistics.newBuilder()
private Builder() {
maybeForceBuilderInitialization();
}
-
- private Builder(BuilderParent parent) {
+
+ private Builder(
+ com.google.protobuf.GeneratedMessage.BuilderParent parent) {
super(parent);
maybeForceBuilderInitialization();
}
@@ -1673,27 +2164,27 @@ public final class OrcProto {
private static Builder create() {
return new Builder();
}
-
+
public Builder clear() {
super.clear();
- count_ = java.util.Collections.emptyList();;
+ count_ = java.util.Collections.emptyList();
bitField0_ = (bitField0_ & ~0x00000001);
return this;
}
-
+
public Builder clone() {
return create().mergeFrom(buildPartial());
}
-
+
public com.google.protobuf.Descriptors.Descriptor
getDescriptorForType() {
- return org.apache.hadoop.hive.ql.io.orc.OrcProto.BucketStatistics.getDescriptor();
+ return org.apache.hadoop.hive.ql.io.orc.OrcProto.internal_static_org_apache_hadoop_hive_ql_io_orc_BucketStatistics_descriptor;
}
-
+
public org.apache.hadoop.hive.ql.io.orc.OrcProto.BucketStatistics getDefaultInstanceForType() {
return org.apache.hadoop.hive.ql.io.orc.OrcProto.BucketStatistics.getDefaultInstance();
}
-
+
public org.apache.hadoop.hive.ql.io.orc.OrcProto.BucketStatistics build() {
org.apache.hadoop.hive.ql.io.orc.OrcProto.BucketStatistics result = buildPartial();
if (!result.isInitialized()) {
@@ -1701,17 +2192,7 @@ public final class OrcProto {
}
return result;
}
-
- private org.apache.hadoop.hive.ql.io.orc.OrcProto.BucketStatistics buildParsed()
- throws com.google.protobuf.InvalidProtocolBufferException {
- org.apache.hadoop.hive.ql.io.orc.OrcProto.BucketStatistics result = buildPartial();
- if (!result.isInitialized()) {
- throw newUninitializedMessageException(
- result).asInvalidProtocolBufferException();
- }
- return result;
- }
-
+
public org.apache.hadoop.hive.ql.io.orc.OrcProto.BucketStatistics buildPartial() {
org.apache.hadoop.hive.ql.io.orc.OrcProto.BucketStatistics result = new org.apache.hadoop.hive.ql.io.orc.OrcProto.BucketStatistics(this);
int from_bitField0_ = bitField0_;
@@ -1723,7 +2204,7 @@ public final class OrcProto {
onBuilt();
return result;
}
-
+
public Builder mergeFrom(com.google.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hive.ql.io.orc.OrcProto.BucketStatistics) {
return mergeFrom((org.apache.hadoop.hive.ql.io.orc.OrcProto.BucketStatistics)other);
@@ -1732,7 +2213,7 @@ public final class OrcProto {
return this;
}
}
-
+
public Builder mergeFrom(org.apache.hadoop.hive.ql.io.orc.OrcProto.BucketStatistics other) {
if (other == org.apache.hadoop.hive.ql.io.orc.OrcProto.BucketStatistics.getDefaultInstance()) return this;
if (!other.count_.isEmpty()) {
@@ -1748,72 +2229,60 @@ public final class OrcProto {
this.mergeUnknownFields(other.getUnknownFields());
return this;
}
-
+
public final boolean isInitialized() {
return true;
}
-
+
public Builder mergeFrom(
com.google.protobuf.CodedInputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- com.google.protobuf.UnknownFieldSet.Builder unknownFields =
- com.google.protobuf.UnknownFieldSet.newBuilder(
- this.getUnknownFields());
- while (true) {
- int tag = input.readTag();
- switch (tag) {
- case 0:
- this.setUnknownFields(unknownFields.build());
- onChanged();
- return this;
- default: {
[... 12257 lines stripped ...]