You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hbase.apache.org by st...@apache.org on 2017/08/02 16:33:45 UTC
[13/51] [partial] hbase git commit: HBASE-17056 Remove checked in PB
generated files
http://git-wip-us.apache.org/repos/asf/hbase/blob/7a6de1bd/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/TextFormatParseInfoTree.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/TextFormatParseInfoTree.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/TextFormatParseInfoTree.java
deleted file mode 100644
index cd08eab..0000000
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/TextFormatParseInfoTree.java
+++ /dev/null
@@ -1,226 +0,0 @@
-// Protocol Buffers - Google's data interchange format
-// Copyright 2008 Google Inc. All rights reserved.
-// https://developers.google.com/protocol-buffers/
-//
-// Redistribution and use in source and binary forms, with or without
-// modification, are permitted provided that the following conditions are
-// met:
-//
-// * Redistributions of source code must retain the above copyright
-// notice, this list of conditions and the following disclaimer.
-// * Redistributions in binary form must reproduce the above
-// copyright notice, this list of conditions and the following disclaimer
-// in the documentation and/or other materials provided with the
-// distribution.
-// * Neither the name of Google Inc. nor the names of its
-// contributors may be used to endorse or promote products derived from
-// this software without specific prior written permission.
-//
-// THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
-// "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
-// LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
-// A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
-// OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
-// SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
-// LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
-// DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
-// THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-// (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
-// OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-
-package org.apache.hadoop.hbase.shaded.com.google.protobuf;
-
-import org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor;
-
-import java.util.ArrayList;
-import java.util.Collections;
-import java.util.HashMap;
-import java.util.List;
-import java.util.Map;
-import java.util.Map.Entry;
-
-
-/**
- * Data structure which is populated with the locations of each field value parsed from the text.
- *
- * <p>The locations of primary fields values are retrieved by {@code getLocation} or
- * {@code getLocations}. The locations of sub message values are within nested
- * {@code TextFormatParseInfoTree}s and are retrieve by {@code getNestedTree} or
- * {@code getNestedTrees}.
- *
- * <p>The {@code TextFormatParseInfoTree} is created by a Builder.
- */
-public class TextFormatParseInfoTree {
-
- // Defines a mapping between each field's descriptor to the list of locations where
- // its value(s) were was encountered.
- private Map<FieldDescriptor, List<TextFormatParseLocation>> locationsFromField;
-
- // Defines a mapping between a field's descriptor to a list of TextFormatParseInfoTrees for
- // sub message location information.
- Map<FieldDescriptor, List<TextFormatParseInfoTree>> subtreesFromField;
-
- /**
- * Construct a {@code TextFormatParseInfoTree}.
- *
- * @param locationsFromField a map of fields to location in the source code
- * @param subtreeBuildersFromField a map of fields to parse tree location information builders
- */
- private TextFormatParseInfoTree(
- Map<FieldDescriptor, List<TextFormatParseLocation>> locationsFromField,
- Map<FieldDescriptor, List<TextFormatParseInfoTree.Builder>> subtreeBuildersFromField) {
-
- // The maps are unmodifiable. The values in the maps are unmodifiable.
- Map<FieldDescriptor, List<TextFormatParseLocation>> locs =
- new HashMap<FieldDescriptor, List<TextFormatParseLocation>>();
- for (Entry<FieldDescriptor, List<TextFormatParseLocation>> kv : locationsFromField.entrySet()) {
- locs.put(kv.getKey(), Collections.unmodifiableList(kv.getValue()));
- }
- this.locationsFromField = Collections.unmodifiableMap(locs);
-
- Map<FieldDescriptor, List<TextFormatParseInfoTree>> subs =
- new HashMap<FieldDescriptor, List<TextFormatParseInfoTree>>();
- for (Entry<FieldDescriptor, List<Builder>> kv : subtreeBuildersFromField.entrySet()) {
- List<TextFormatParseInfoTree> submessagesOfField = new ArrayList<TextFormatParseInfoTree>();
- for (Builder subBuilder : kv.getValue()) {
- submessagesOfField.add(subBuilder.build());
- }
- subs.put(kv.getKey(), Collections.unmodifiableList(submessagesOfField));
- }
- this.subtreesFromField = Collections.unmodifiableMap(subs);
- }
-
- /**
- * Retrieve all the locations of a field.
- *
- * @param fieldDescriptor the the @{link FieldDescriptor} of the desired field
- * @return a list of the locations of values of the field. If there are not values
- * or the field doesn't exist, an empty list is returned.
- */
- public List<TextFormatParseLocation> getLocations(final FieldDescriptor fieldDescriptor) {
- List<TextFormatParseLocation> result = locationsFromField.get(fieldDescriptor);
- return (result == null) ? Collections.<TextFormatParseLocation>emptyList() : result;
- }
-
- /**
- * Get the location in the source of a field's value.
- *
- * <p>Returns the {@link TextFormatParseLocation} for index-th value of the field in the parsed
- * text.
- *
- * @param fieldDescriptor the @{link FieldDescriptor} of the desired field
- * @param index the index of the value.
- * @return the {@link TextFormatParseLocation} of the value
- * @throws IllegalArgumentException index is out of range
- */
- public TextFormatParseLocation getLocation(final FieldDescriptor fieldDescriptor, int index) {
- return getFromList(getLocations(fieldDescriptor), index, fieldDescriptor);
- }
-
- /**
- * Retrieve a list of all the location information trees for a sub message field.
- *
- * @param fieldDescriptor the @{link FieldDescriptor} of the desired field
- * @return A list of {@link TextFormatParseInfoTree}
- */
- public List<TextFormatParseInfoTree> getNestedTrees(final FieldDescriptor fieldDescriptor) {
- List<TextFormatParseInfoTree> result = subtreesFromField.get(fieldDescriptor);
- return result == null ? Collections.<TextFormatParseInfoTree>emptyList() : result;
- }
-
- /**
- * Returns the parse info tree for the given field, which must be a message type.
- *
- * @param fieldDescriptor the @{link FieldDescriptor} of the desired sub message
- * @param index the index of message value.
- * @return the {@code ParseInfoTree} of the message value. {@code null} is returned if the field
- * doesn't exist or the index is out of range.
- * @throws IllegalArgumentException if index is out of range
- */
- public TextFormatParseInfoTree getNestedTree(final FieldDescriptor fieldDescriptor, int index) {
- return getFromList(getNestedTrees(fieldDescriptor), index, fieldDescriptor);
- }
-
- /**
- * Create a builder for a {@code ParseInfoTree}.
- *
- * @return the builder
- */
- public static Builder builder() {
- return new Builder();
- }
-
- private static <T> T getFromList(List<T> list, int index, FieldDescriptor fieldDescriptor) {
- if (index >= list.size() || index < 0) {
- throw new IllegalArgumentException(String.format("Illegal index field: %s, index %d",
- fieldDescriptor == null ? "<null>" : fieldDescriptor.getName(), index));
- }
- return list.get(index);
- }
-
- /**
- * Builder for a {@link TextFormatParseInfoTree}.
- */
- public static class Builder {
-
- private Map<FieldDescriptor, List<TextFormatParseLocation>> locationsFromField;
-
- // Defines a mapping between a field's descriptor to a list of ParseInfoTrees builders for
- // sub message location information.
- private Map<FieldDescriptor, List<Builder>> subtreeBuildersFromField;
-
- /**
- * Create a root level {@ParseInfoTree} builder.
- */
- private Builder() {
- locationsFromField = new HashMap<FieldDescriptor, List<TextFormatParseLocation>>();
- subtreeBuildersFromField = new HashMap<FieldDescriptor, List<Builder>>();
- }
-
- /**
- * Record the starting location of a single value for a field.
- *
- * @param fieldDescriptor the field
- * @param location source code location information
- */
- public Builder setLocation(
- final FieldDescriptor fieldDescriptor, TextFormatParseLocation location) {
- List<TextFormatParseLocation> fieldLocations = locationsFromField.get(fieldDescriptor);
- if (fieldLocations == null) {
- fieldLocations = new ArrayList<TextFormatParseLocation>();
- locationsFromField.put(fieldDescriptor, fieldLocations);
- }
- fieldLocations.add(location);
- return this;
- }
-
- /**
- * Set for a sub message.
- *
- * <p>A new builder is created for a sub message. The builder that is returned is a new builder.
- * The return is <em>not</em> the invoked {@code builder.getBuilderForSubMessageField}.
- *
- * @param fieldDescriptor the field whose value is the submessage
- * @return a new Builder for the sub message
- */
- public Builder getBuilderForSubMessageField(final FieldDescriptor fieldDescriptor) {
- List<Builder> submessageBuilders = subtreeBuildersFromField.get(fieldDescriptor);
- if (submessageBuilders == null) {
- submessageBuilders = new ArrayList<Builder>();
- subtreeBuildersFromField.put(fieldDescriptor, submessageBuilders);
- }
- Builder subtreeBuilder = new Builder();
- submessageBuilders.add(subtreeBuilder);
- return subtreeBuilder;
- }
-
- /**
- * Build the {@code TextFormatParseInfoTree}.
- *
- * @return the {@code TextFormatParseInfoTree}
- */
- public TextFormatParseInfoTree build() {
- return new TextFormatParseInfoTree(locationsFromField, subtreeBuildersFromField);
- }
- }
-}
http://git-wip-us.apache.org/repos/asf/hbase/blob/7a6de1bd/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/TextFormatParseLocation.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/TextFormatParseLocation.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/TextFormatParseLocation.java
deleted file mode 100644
index 89ca9d2..0000000
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/TextFormatParseLocation.java
+++ /dev/null
@@ -1,104 +0,0 @@
-// Protocol Buffers - Google's data interchange format
-// Copyright 2008 Google Inc. All rights reserved.
-// https://developers.google.com/protocol-buffers/
-//
-// Redistribution and use in source and binary forms, with or without
-// modification, are permitted provided that the following conditions are
-// met:
-//
-// * Redistributions of source code must retain the above copyright
-// notice, this list of conditions and the following disclaimer.
-// * Redistributions in binary form must reproduce the above
-// copyright notice, this list of conditions and the following disclaimer
-// in the documentation and/or other materials provided with the
-// distribution.
-// * Neither the name of Google Inc. nor the names of its
-// contributors may be used to endorse or promote products derived from
-// this software without specific prior written permission.
-//
-// THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
-// "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
-// LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
-// A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
-// OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
-// SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
-// LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
-// DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
-// THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-// (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
-// OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-
-package org.apache.hadoop.hbase.shaded.com.google.protobuf;
-
-import java.util.Arrays;
-
-/**
- * A location in the source code.
- *
- * <p>A location is the starting line number and starting column number.
- */
-public final class TextFormatParseLocation {
-
- /**
- * The empty location.
- */
- public static final TextFormatParseLocation EMPTY = new TextFormatParseLocation(-1, -1);
-
- /**
- * Create a location.
- *
- * @param line the starting line number
- * @param column the starting column number
- * @return a {@code ParseLocation}
- */
- static TextFormatParseLocation create(int line, int column) {
- if (line == -1 && column == -1) {
- return EMPTY;
- }
- if (line < 0 || column < 0) {
- throw new IllegalArgumentException(
- String.format("line and column values must be >= 0: line %d, column: %d", line, column));
- }
- return new TextFormatParseLocation(line, column);
- }
-
- private final int line;
- private final int column;
-
- private TextFormatParseLocation(int line, int column) {
- this.line = line;
- this.column = column;
- }
-
- public int getLine() {
- return line;
- }
-
- public int getColumn() {
- return column;
- }
-
- @Override
- public String toString() {
- return String.format("ParseLocation{line=%d, column=%d}", line, column);
- }
-
- @Override
- public boolean equals(Object o) {
- if (o == this) {
- return true;
- }
- if (!(o instanceof TextFormatParseLocation)) {
- return false;
- }
- TextFormatParseLocation that = (TextFormatParseLocation) o;
- return (this.line == that.getLine())
- && (this.column == that.getColumn());
- }
-
- @Override
- public int hashCode() {
- int[] values = {line, column};
- return Arrays.hashCode(values);
- }
-}
http://git-wip-us.apache.org/repos/asf/hbase/blob/7a6de1bd/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/Timestamp.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/Timestamp.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/Timestamp.java
deleted file mode 100644
index 0023d50..0000000
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/Timestamp.java
+++ /dev/null
@@ -1,616 +0,0 @@
-// Generated by the protocol buffer compiler. DO NOT EDIT!
-// source: google/protobuf/timestamp.proto
-
-package org.apache.hadoop.hbase.shaded.com.google.protobuf;
-
-/**
- * <pre>
- * A Timestamp represents a point in time independent of any time zone
- * or calendar, represented as seconds and fractions of seconds at
- * nanosecond resolution in UTC Epoch time. It is encoded using the
- * Proleptic Gregorian Calendar which extends the Gregorian calendar
- * backwards to year one. It is encoded assuming all minutes are 60
- * seconds long, i.e. leap seconds are "smeared" so that no leap second
- * table is needed for interpretation. Range is from
- * 0001-01-01T00:00:00Z to 9999-12-31T23:59:59.999999999Z.
- * By restricting to that range, we ensure that we can convert to
- * and from RFC 3339 date strings.
- * See [https://www.ietf.org/rfc/rfc3339.txt](https://www.ietf.org/rfc/rfc3339.txt).
- * Example 1: Compute Timestamp from POSIX `time()`.
- * Timestamp timestamp;
- * timestamp.set_seconds(time(NULL));
- * timestamp.set_nanos(0);
- * Example 2: Compute Timestamp from POSIX `gettimeofday()`.
- * struct timeval tv;
- * gettimeofday(&tv, NULL);
- * Timestamp timestamp;
- * timestamp.set_seconds(tv.tv_sec);
- * timestamp.set_nanos(tv.tv_usec * 1000);
- * Example 3: Compute Timestamp from Win32 `GetSystemTimeAsFileTime()`.
- * FILETIME ft;
- * GetSystemTimeAsFileTime(&ft);
- * UINT64 ticks = (((UINT64)ft.dwHighDateTime) << 32) | ft.dwLowDateTime;
- * // A Windows tick is 100 nanoseconds. Windows epoch 1601-01-01T00:00:00Z
- * // is 11644473600 seconds before Unix epoch 1970-01-01T00:00:00Z.
- * Timestamp timestamp;
- * timestamp.set_seconds((INT64) ((ticks / 10000000) - 11644473600LL));
- * timestamp.set_nanos((INT32) ((ticks % 10000000) * 100));
- * Example 4: Compute Timestamp from Java `System.currentTimeMillis()`.
- * long millis = System.currentTimeMillis();
- * Timestamp timestamp = Timestamp.newBuilder().setSeconds(millis / 1000)
- * .setNanos((int) ((millis % 1000) * 1000000)).build();
- * Example 5: Compute Timestamp from current time in Python.
- * timestamp = Timestamp()
- * timestamp.GetCurrentTime()
- * </pre>
- *
- * Protobuf type {@code google.protobuf.Timestamp}
- */
-public final class Timestamp extends
- org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 implements
- // @@protoc_insertion_point(message_implements:google.protobuf.Timestamp)
- TimestampOrBuilder {
- // Use Timestamp.newBuilder() to construct.
- private Timestamp(org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
- super(builder);
- }
- private Timestamp() {
- seconds_ = 0L;
- nanos_ = 0;
- }
-
- @java.lang.Override
- public final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet
- getUnknownFields() {
- return org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.getDefaultInstance();
- }
- private Timestamp(
- org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
- org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
- this();
- int mutable_bitField0_ = 0;
- try {
- boolean done = false;
- while (!done) {
- int tag = input.readTag();
- switch (tag) {
- case 0:
- done = true;
- break;
- default: {
- if (!input.skipField(tag)) {
- done = true;
- }
- break;
- }
- case 8: {
-
- seconds_ = input.readInt64();
- break;
- }
- case 16: {
-
- nanos_ = input.readInt32();
- break;
- }
- }
- }
- } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
- throw e.setUnfinishedMessage(this);
- } catch (java.io.IOException e) {
- throw new org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException(
- e).setUnfinishedMessage(this);
- } finally {
- makeExtensionsImmutable();
- }
- }
- public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
- getDescriptor() {
- return org.apache.hadoop.hbase.shaded.com.google.protobuf.TimestampProto.internal_static_google_protobuf_Timestamp_descriptor;
- }
-
- protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
- internalGetFieldAccessorTable() {
- return org.apache.hadoop.hbase.shaded.com.google.protobuf.TimestampProto.internal_static_google_protobuf_Timestamp_fieldAccessorTable
- .ensureFieldAccessorsInitialized(
- org.apache.hadoop.hbase.shaded.com.google.protobuf.Timestamp.class, org.apache.hadoop.hbase.shaded.com.google.protobuf.Timestamp.Builder.class);
- }
-
- public static final int SECONDS_FIELD_NUMBER = 1;
- private long seconds_;
- /**
- * <pre>
- * Represents seconds of UTC time since Unix epoch
- * 1970-01-01T00:00:00Z. Must be from 0001-01-01T00:00:00Z to
- * 9999-12-31T23:59:59Z inclusive.
- * </pre>
- *
- * <code>int64 seconds = 1;</code>
- */
- public long getSeconds() {
- return seconds_;
- }
-
- public static final int NANOS_FIELD_NUMBER = 2;
- private int nanos_;
- /**
- * <pre>
- * Non-negative fractions of a second at nanosecond resolution. Negative
- * second values with fractions must still have non-negative nanos values
- * that count forward in time. Must be from 0 to 999,999,999
- * inclusive.
- * </pre>
- *
- * <code>int32 nanos = 2;</code>
- */
- public int getNanos() {
- return nanos_;
- }
-
- private byte memoizedIsInitialized = -1;
- public final boolean isInitialized() {
- byte isInitialized = memoizedIsInitialized;
- if (isInitialized == 1) return true;
- if (isInitialized == 0) return false;
-
- memoizedIsInitialized = 1;
- return true;
- }
-
- public void writeTo(org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream output)
- throws java.io.IOException {
- if (seconds_ != 0L) {
- output.writeInt64(1, seconds_);
- }
- if (nanos_ != 0) {
- output.writeInt32(2, nanos_);
- }
- }
-
- public int getSerializedSize() {
- int size = memoizedSize;
- if (size != -1) return size;
-
- size = 0;
- if (seconds_ != 0L) {
- size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
- .computeInt64Size(1, seconds_);
- }
- if (nanos_ != 0) {
- size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
- .computeInt32Size(2, nanos_);
- }
- memoizedSize = size;
- return size;
- }
-
- private static final long serialVersionUID = 0L;
- @java.lang.Override
- public boolean equals(final java.lang.Object obj) {
- if (obj == this) {
- return true;
- }
- if (!(obj instanceof org.apache.hadoop.hbase.shaded.com.google.protobuf.Timestamp)) {
- return super.equals(obj);
- }
- org.apache.hadoop.hbase.shaded.com.google.protobuf.Timestamp other = (org.apache.hadoop.hbase.shaded.com.google.protobuf.Timestamp) obj;
-
- boolean result = true;
- result = result && (getSeconds()
- == other.getSeconds());
- result = result && (getNanos()
- == other.getNanos());
- return result;
- }
-
- @java.lang.Override
- public int hashCode() {
- if (memoizedHashCode != 0) {
- return memoizedHashCode;
- }
- int hash = 41;
- hash = (19 * hash) + getDescriptor().hashCode();
- hash = (37 * hash) + SECONDS_FIELD_NUMBER;
- hash = (53 * hash) + org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.hashLong(
- getSeconds());
- hash = (37 * hash) + NANOS_FIELD_NUMBER;
- hash = (53 * hash) + getNanos();
- hash = (29 * hash) + unknownFields.hashCode();
- memoizedHashCode = hash;
- return hash;
- }
-
- public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Timestamp parseFrom(
- org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data)
- throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
- return PARSER.parseFrom(data);
- }
- public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Timestamp parseFrom(
- org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data,
- org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
- return PARSER.parseFrom(data, extensionRegistry);
- }
- public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Timestamp parseFrom(byte[] data)
- throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
- return PARSER.parseFrom(data);
- }
- public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Timestamp parseFrom(
- byte[] data,
- org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
- return PARSER.parseFrom(data, extensionRegistry);
- }
- public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Timestamp parseFrom(java.io.InputStream input)
- throws java.io.IOException {
- return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
- .parseWithIOException(PARSER, input);
- }
- public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Timestamp parseFrom(
- java.io.InputStream input,
- org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws java.io.IOException {
- return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
- .parseWithIOException(PARSER, input, extensionRegistry);
- }
- public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Timestamp parseDelimitedFrom(java.io.InputStream input)
- throws java.io.IOException {
- return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
- .parseDelimitedWithIOException(PARSER, input);
- }
- public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Timestamp parseDelimitedFrom(
- java.io.InputStream input,
- org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws java.io.IOException {
- return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
- .parseDelimitedWithIOException(PARSER, input, extensionRegistry);
- }
- public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Timestamp parseFrom(
- org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input)
- throws java.io.IOException {
- return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
- .parseWithIOException(PARSER, input);
- }
- public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Timestamp parseFrom(
- org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
- org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws java.io.IOException {
- return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
- .parseWithIOException(PARSER, input, extensionRegistry);
- }
-
- public Builder newBuilderForType() { return newBuilder(); }
- public static Builder newBuilder() {
- return DEFAULT_INSTANCE.toBuilder();
- }
- public static Builder newBuilder(org.apache.hadoop.hbase.shaded.com.google.protobuf.Timestamp prototype) {
- return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype);
- }
- public Builder toBuilder() {
- return this == DEFAULT_INSTANCE
- ? new Builder() : new Builder().mergeFrom(this);
- }
-
- @java.lang.Override
- protected Builder newBuilderForType(
- org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
- Builder builder = new Builder(parent);
- return builder;
- }
- /**
- * <pre>
- * A Timestamp represents a point in time independent of any time zone
- * or calendar, represented as seconds and fractions of seconds at
- * nanosecond resolution in UTC Epoch time. It is encoded using the
- * Proleptic Gregorian Calendar which extends the Gregorian calendar
- * backwards to year one. It is encoded assuming all minutes are 60
- * seconds long, i.e. leap seconds are "smeared" so that no leap second
- * table is needed for interpretation. Range is from
- * 0001-01-01T00:00:00Z to 9999-12-31T23:59:59.999999999Z.
- * By restricting to that range, we ensure that we can convert to
- * and from RFC 3339 date strings.
- * See [https://www.ietf.org/rfc/rfc3339.txt](https://www.ietf.org/rfc/rfc3339.txt).
- * Example 1: Compute Timestamp from POSIX `time()`.
- * Timestamp timestamp;
- * timestamp.set_seconds(time(NULL));
- * timestamp.set_nanos(0);
- * Example 2: Compute Timestamp from POSIX `gettimeofday()`.
- * struct timeval tv;
- * gettimeofday(&tv, NULL);
- * Timestamp timestamp;
- * timestamp.set_seconds(tv.tv_sec);
- * timestamp.set_nanos(tv.tv_usec * 1000);
- * Example 3: Compute Timestamp from Win32 `GetSystemTimeAsFileTime()`.
- * FILETIME ft;
- * GetSystemTimeAsFileTime(&ft);
- * UINT64 ticks = (((UINT64)ft.dwHighDateTime) << 32) | ft.dwLowDateTime;
- * // A Windows tick is 100 nanoseconds. Windows epoch 1601-01-01T00:00:00Z
- * // is 11644473600 seconds before Unix epoch 1970-01-01T00:00:00Z.
- * Timestamp timestamp;
- * timestamp.set_seconds((INT64) ((ticks / 10000000) - 11644473600LL));
- * timestamp.set_nanos((INT32) ((ticks % 10000000) * 100));
- * Example 4: Compute Timestamp from Java `System.currentTimeMillis()`.
- * long millis = System.currentTimeMillis();
- * Timestamp timestamp = Timestamp.newBuilder().setSeconds(millis / 1000)
- * .setNanos((int) ((millis % 1000) * 1000000)).build();
- * Example 5: Compute Timestamp from current time in Python.
- * timestamp = Timestamp()
- * timestamp.GetCurrentTime()
- * </pre>
- *
- * Protobuf type {@code google.protobuf.Timestamp}
- */
- public static final class Builder extends
- org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
- // @@protoc_insertion_point(builder_implements:google.protobuf.Timestamp)
- org.apache.hadoop.hbase.shaded.com.google.protobuf.TimestampOrBuilder {
- public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
- getDescriptor() {
- return org.apache.hadoop.hbase.shaded.com.google.protobuf.TimestampProto.internal_static_google_protobuf_Timestamp_descriptor;
- }
-
- protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
- internalGetFieldAccessorTable() {
- return org.apache.hadoop.hbase.shaded.com.google.protobuf.TimestampProto.internal_static_google_protobuf_Timestamp_fieldAccessorTable
- .ensureFieldAccessorsInitialized(
- org.apache.hadoop.hbase.shaded.com.google.protobuf.Timestamp.class, org.apache.hadoop.hbase.shaded.com.google.protobuf.Timestamp.Builder.class);
- }
-
- // Construct using org.apache.hadoop.hbase.shaded.com.google.protobuf.Timestamp.newBuilder()
- private Builder() {
- maybeForceBuilderInitialization();
- }
-
- private Builder(
- org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
- super(parent);
- maybeForceBuilderInitialization();
- }
- private void maybeForceBuilderInitialization() {
- if (org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
- .alwaysUseFieldBuilders) {
- }
- }
- public Builder clear() {
- super.clear();
- seconds_ = 0L;
-
- nanos_ = 0;
-
- return this;
- }
-
- public org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
- getDescriptorForType() {
- return org.apache.hadoop.hbase.shaded.com.google.protobuf.TimestampProto.internal_static_google_protobuf_Timestamp_descriptor;
- }
-
- public org.apache.hadoop.hbase.shaded.com.google.protobuf.Timestamp getDefaultInstanceForType() {
- return org.apache.hadoop.hbase.shaded.com.google.protobuf.Timestamp.getDefaultInstance();
- }
-
- public org.apache.hadoop.hbase.shaded.com.google.protobuf.Timestamp build() {
- org.apache.hadoop.hbase.shaded.com.google.protobuf.Timestamp result = buildPartial();
- if (!result.isInitialized()) {
- throw newUninitializedMessageException(result);
- }
- return result;
- }
-
- public org.apache.hadoop.hbase.shaded.com.google.protobuf.Timestamp buildPartial() {
- org.apache.hadoop.hbase.shaded.com.google.protobuf.Timestamp result = new org.apache.hadoop.hbase.shaded.com.google.protobuf.Timestamp(this);
- result.seconds_ = seconds_;
- result.nanos_ = nanos_;
- onBuilt();
- return result;
- }
-
- public Builder clone() {
- return (Builder) super.clone();
- }
- public Builder setField(
- org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
- Object value) {
- return (Builder) super.setField(field, value);
- }
- public Builder clearField(
- org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field) {
- return (Builder) super.clearField(field);
- }
- public Builder clearOneof(
- org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.OneofDescriptor oneof) {
- return (Builder) super.clearOneof(oneof);
- }
- public Builder setRepeatedField(
- org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
- int index, Object value) {
- return (Builder) super.setRepeatedField(field, index, value);
- }
- public Builder addRepeatedField(
- org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
- Object value) {
- return (Builder) super.addRepeatedField(field, value);
- }
- public Builder mergeFrom(org.apache.hadoop.hbase.shaded.com.google.protobuf.Message other) {
- if (other instanceof org.apache.hadoop.hbase.shaded.com.google.protobuf.Timestamp) {
- return mergeFrom((org.apache.hadoop.hbase.shaded.com.google.protobuf.Timestamp)other);
- } else {
- super.mergeFrom(other);
- return this;
- }
- }
-
- public Builder mergeFrom(org.apache.hadoop.hbase.shaded.com.google.protobuf.Timestamp other) {
- if (other == org.apache.hadoop.hbase.shaded.com.google.protobuf.Timestamp.getDefaultInstance()) return this;
- if (other.getSeconds() != 0L) {
- setSeconds(other.getSeconds());
- }
- if (other.getNanos() != 0) {
- setNanos(other.getNanos());
- }
- onChanged();
- return this;
- }
-
- public final boolean isInitialized() {
- return true;
- }
-
- public Builder mergeFrom(
- org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
- org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws java.io.IOException {
- org.apache.hadoop.hbase.shaded.com.google.protobuf.Timestamp parsedMessage = null;
- try {
- parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
- } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
- parsedMessage = (org.apache.hadoop.hbase.shaded.com.google.protobuf.Timestamp) e.getUnfinishedMessage();
- throw e.unwrapIOException();
- } finally {
- if (parsedMessage != null) {
- mergeFrom(parsedMessage);
- }
- }
- return this;
- }
-
- private long seconds_ ;
- /**
- * <pre>
- * Represents seconds of UTC time since Unix epoch
- * 1970-01-01T00:00:00Z. Must be from 0001-01-01T00:00:00Z to
- * 9999-12-31T23:59:59Z inclusive.
- * </pre>
- *
- * <code>int64 seconds = 1;</code>
- */
- public long getSeconds() {
- return seconds_;
- }
- /**
- * <pre>
- * Represents seconds of UTC time since Unix epoch
- * 1970-01-01T00:00:00Z. Must be from 0001-01-01T00:00:00Z to
- * 9999-12-31T23:59:59Z inclusive.
- * </pre>
- *
- * <code>int64 seconds = 1;</code>
- */
- public Builder setSeconds(long value) {
-
- seconds_ = value;
- onChanged();
- return this;
- }
- /**
- * <pre>
- * Represents seconds of UTC time since Unix epoch
- * 1970-01-01T00:00:00Z. Must be from 0001-01-01T00:00:00Z to
- * 9999-12-31T23:59:59Z inclusive.
- * </pre>
- *
- * <code>int64 seconds = 1;</code>
- */
- public Builder clearSeconds() {
-
- seconds_ = 0L;
- onChanged();
- return this;
- }
-
- private int nanos_ ;
- /**
- * <pre>
- * Non-negative fractions of a second at nanosecond resolution. Negative
- * second values with fractions must still have non-negative nanos values
- * that count forward in time. Must be from 0 to 999,999,999
- * inclusive.
- * </pre>
- *
- * <code>int32 nanos = 2;</code>
- */
- public int getNanos() {
- return nanos_;
- }
- /**
- * <pre>
- * Non-negative fractions of a second at nanosecond resolution. Negative
- * second values with fractions must still have non-negative nanos values
- * that count forward in time. Must be from 0 to 999,999,999
- * inclusive.
- * </pre>
- *
- * <code>int32 nanos = 2;</code>
- */
- public Builder setNanos(int value) {
-
- nanos_ = value;
- onChanged();
- return this;
- }
- /**
- * <pre>
- * Non-negative fractions of a second at nanosecond resolution. Negative
- * second values with fractions must still have non-negative nanos values
- * that count forward in time. Must be from 0 to 999,999,999
- * inclusive.
- * </pre>
- *
- * <code>int32 nanos = 2;</code>
- */
- public Builder clearNanos() {
-
- nanos_ = 0;
- onChanged();
- return this;
- }
- public final Builder setUnknownFields(
- final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
- return this;
- }
-
- public final Builder mergeUnknownFields(
- final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
- return this;
- }
-
-
- // @@protoc_insertion_point(builder_scope:google.protobuf.Timestamp)
- }
-
- // @@protoc_insertion_point(class_scope:google.protobuf.Timestamp)
- private static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Timestamp DEFAULT_INSTANCE;
- static {
- DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.com.google.protobuf.Timestamp();
- }
-
- public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Timestamp getDefaultInstance() {
- return DEFAULT_INSTANCE;
- }
-
- private static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<Timestamp>
- PARSER = new org.apache.hadoop.hbase.shaded.com.google.protobuf.AbstractParser<Timestamp>() {
- public Timestamp parsePartialFrom(
- org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
- org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
- return new Timestamp(input, extensionRegistry);
- }
- };
-
- public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<Timestamp> parser() {
- return PARSER;
- }
-
- @java.lang.Override
- public org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<Timestamp> getParserForType() {
- return PARSER;
- }
-
- public org.apache.hadoop.hbase.shaded.com.google.protobuf.Timestamp getDefaultInstanceForType() {
- return DEFAULT_INSTANCE;
- }
-
-}
-
http://git-wip-us.apache.org/repos/asf/hbase/blob/7a6de1bd/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/TimestampOrBuilder.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/TimestampOrBuilder.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/TimestampOrBuilder.java
deleted file mode 100644
index fca3c16..0000000
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/TimestampOrBuilder.java
+++ /dev/null
@@ -1,32 +0,0 @@
-// Generated by the protocol buffer compiler. DO NOT EDIT!
-// source: google/protobuf/timestamp.proto
-
-package org.apache.hadoop.hbase.shaded.com.google.protobuf;
-
-public interface TimestampOrBuilder extends
- // @@protoc_insertion_point(interface_extends:google.protobuf.Timestamp)
- org.apache.hadoop.hbase.shaded.com.google.protobuf.MessageOrBuilder {
-
- /**
- * <pre>
- * Represents seconds of UTC time since Unix epoch
- * 1970-01-01T00:00:00Z. Must be from 0001-01-01T00:00:00Z to
- * 9999-12-31T23:59:59Z inclusive.
- * </pre>
- *
- * <code>int64 seconds = 1;</code>
- */
- long getSeconds();
-
- /**
- * <pre>
- * Non-negative fractions of a second at nanosecond resolution. Negative
- * second values with fractions must still have non-negative nanos values
- * that count forward in time. Must be from 0 to 999,999,999
- * inclusive.
- * </pre>
- *
- * <code>int32 nanos = 2;</code>
- */
- int getNanos();
-}
http://git-wip-us.apache.org/repos/asf/hbase/blob/7a6de1bd/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/TimestampProto.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/TimestampProto.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/TimestampProto.java
deleted file mode 100644
index c62c4ce..0000000
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/TimestampProto.java
+++ /dev/null
@@ -1,59 +0,0 @@
-// Generated by the protocol buffer compiler. DO NOT EDIT!
-// source: google/protobuf/timestamp.proto
-
-package org.apache.hadoop.hbase.shaded.com.google.protobuf;
-
-public final class TimestampProto {
- private TimestampProto() {}
- public static void registerAllExtensions(
- org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite registry) {
- }
-
- public static void registerAllExtensions(
- org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistry registry) {
- registerAllExtensions(
- (org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite) registry);
- }
- static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
- internal_static_google_protobuf_Timestamp_descriptor;
- static final
- org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
- internal_static_google_protobuf_Timestamp_fieldAccessorTable;
-
- public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor
- getDescriptor() {
- return descriptor;
- }
- private static org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor
- descriptor;
- static {
- java.lang.String[] descriptorData = {
- "\n\037google/protobuf/timestamp.proto\022\017googl" +
- "e.protobuf\"+\n\tTimestamp\022\017\n\007seconds\030\001 \001(\003" +
- "\022\r\n\005nanos\030\002 \001(\005B~\n\023com.google.protobufB\016" +
- "TimestampProtoP\001Z+github.com/golang/prot" +
- "obuf/ptypes/timestamp\370\001\001\242\002\003GPB\252\002\036Google." +
- "Protobuf.WellKnownTypesb\006proto3"
- };
- org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
- new org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor. InternalDescriptorAssigner() {
- public org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistry assignDescriptors(
- org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor root) {
- descriptor = root;
- return null;
- }
- };
- org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor
- .internalBuildGeneratedFileFrom(descriptorData,
- new org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor[] {
- }, assigner);
- internal_static_google_protobuf_Timestamp_descriptor =
- getDescriptor().getMessageTypes().get(0);
- internal_static_google_protobuf_Timestamp_fieldAccessorTable = new
- org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable(
- internal_static_google_protobuf_Timestamp_descriptor,
- new java.lang.String[] { "Seconds", "Nanos", });
- }
-
- // @@protoc_insertion_point(outer_class_scope)
-}