You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hbase.apache.org by st...@apache.org on 2017/08/02 16:55:04 UTC
[38/52] [partial] hbase git commit: HBASE-17056 Remove checked in PB
generated files
http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/ApiOrBuilder.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/ApiOrBuilder.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/ApiOrBuilder.java
deleted file mode 100644
index b8b7030..0000000
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/ApiOrBuilder.java
+++ /dev/null
@@ -1,258 +0,0 @@
-// Generated by the protocol buffer compiler. DO NOT EDIT!
-// source: google/protobuf/api.proto
-
-package org.apache.hadoop.hbase.shaded.com.google.protobuf;
-
-public interface ApiOrBuilder extends
- // @@protoc_insertion_point(interface_extends:google.protobuf.Api)
- org.apache.hadoop.hbase.shaded.com.google.protobuf.MessageOrBuilder {
-
- /**
- * <pre>
- * The fully qualified name of this api, including package name
- * followed by the api's simple name.
- * </pre>
- *
- * <code>string name = 1;</code>
- */
- java.lang.String getName();
- /**
- * <pre>
- * The fully qualified name of this api, including package name
- * followed by the api's simple name.
- * </pre>
- *
- * <code>string name = 1;</code>
- */
- org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
- getNameBytes();
-
- /**
- * <pre>
- * The methods of this api, in unspecified order.
- * </pre>
- *
- * <code>repeated .google.protobuf.Method methods = 2;</code>
- */
- java.util.List<org.apache.hadoop.hbase.shaded.com.google.protobuf.Method>
- getMethodsList();
- /**
- * <pre>
- * The methods of this api, in unspecified order.
- * </pre>
- *
- * <code>repeated .google.protobuf.Method methods = 2;</code>
- */
- org.apache.hadoop.hbase.shaded.com.google.protobuf.Method getMethods(int index);
- /**
- * <pre>
- * The methods of this api, in unspecified order.
- * </pre>
- *
- * <code>repeated .google.protobuf.Method methods = 2;</code>
- */
- int getMethodsCount();
- /**
- * <pre>
- * The methods of this api, in unspecified order.
- * </pre>
- *
- * <code>repeated .google.protobuf.Method methods = 2;</code>
- */
- java.util.List<? extends org.apache.hadoop.hbase.shaded.com.google.protobuf.MethodOrBuilder>
- getMethodsOrBuilderList();
- /**
- * <pre>
- * The methods of this api, in unspecified order.
- * </pre>
- *
- * <code>repeated .google.protobuf.Method methods = 2;</code>
- */
- org.apache.hadoop.hbase.shaded.com.google.protobuf.MethodOrBuilder getMethodsOrBuilder(
- int index);
-
- /**
- * <pre>
- * Any metadata attached to the API.
- * </pre>
- *
- * <code>repeated .google.protobuf.Option options = 3;</code>
- */
- java.util.List<org.apache.hadoop.hbase.shaded.com.google.protobuf.Option>
- getOptionsList();
- /**
- * <pre>
- * Any metadata attached to the API.
- * </pre>
- *
- * <code>repeated .google.protobuf.Option options = 3;</code>
- */
- org.apache.hadoop.hbase.shaded.com.google.protobuf.Option getOptions(int index);
- /**
- * <pre>
- * Any metadata attached to the API.
- * </pre>
- *
- * <code>repeated .google.protobuf.Option options = 3;</code>
- */
- int getOptionsCount();
- /**
- * <pre>
- * Any metadata attached to the API.
- * </pre>
- *
- * <code>repeated .google.protobuf.Option options = 3;</code>
- */
- java.util.List<? extends org.apache.hadoop.hbase.shaded.com.google.protobuf.OptionOrBuilder>
- getOptionsOrBuilderList();
- /**
- * <pre>
- * Any metadata attached to the API.
- * </pre>
- *
- * <code>repeated .google.protobuf.Option options = 3;</code>
- */
- org.apache.hadoop.hbase.shaded.com.google.protobuf.OptionOrBuilder getOptionsOrBuilder(
- int index);
-
- /**
- * <pre>
- * A version string for this api. If specified, must have the form
- * `major-version.minor-version`, as in `1.10`. If the minor version
- * is omitted, it defaults to zero. If the entire version field is
- * empty, the major version is derived from the package name, as
- * outlined below. If the field is not empty, the version in the
- * package name will be verified to be consistent with what is
- * provided here.
- * The versioning schema uses [semantic
- * versioning](http://semver.org) where the major version number
- * indicates a breaking change and the minor version an additive,
- * non-breaking change. Both version numbers are signals to users
- * what to expect from different versions, and should be carefully
- * chosen based on the product plan.
- * The major version is also reflected in the package name of the
- * API, which must end in `v<major-version>`, as in
- * `google.feature.v1`. For major versions 0 and 1, the suffix can
- * be omitted. Zero major versions must only be used for
- * experimental, none-GA apis.
- * </pre>
- *
- * <code>string version = 4;</code>
- */
- java.lang.String getVersion();
- /**
- * <pre>
- * A version string for this api. If specified, must have the form
- * `major-version.minor-version`, as in `1.10`. If the minor version
- * is omitted, it defaults to zero. If the entire version field is
- * empty, the major version is derived from the package name, as
- * outlined below. If the field is not empty, the version in the
- * package name will be verified to be consistent with what is
- * provided here.
- * The versioning schema uses [semantic
- * versioning](http://semver.org) where the major version number
- * indicates a breaking change and the minor version an additive,
- * non-breaking change. Both version numbers are signals to users
- * what to expect from different versions, and should be carefully
- * chosen based on the product plan.
- * The major version is also reflected in the package name of the
- * API, which must end in `v<major-version>`, as in
- * `google.feature.v1`. For major versions 0 and 1, the suffix can
- * be omitted. Zero major versions must only be used for
- * experimental, none-GA apis.
- * </pre>
- *
- * <code>string version = 4;</code>
- */
- org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
- getVersionBytes();
-
- /**
- * <pre>
- * Source context for the protocol buffer service represented by this
- * message.
- * </pre>
- *
- * <code>.google.protobuf.SourceContext source_context = 5;</code>
- */
- boolean hasSourceContext();
- /**
- * <pre>
- * Source context for the protocol buffer service represented by this
- * message.
- * </pre>
- *
- * <code>.google.protobuf.SourceContext source_context = 5;</code>
- */
- org.apache.hadoop.hbase.shaded.com.google.protobuf.SourceContext getSourceContext();
- /**
- * <pre>
- * Source context for the protocol buffer service represented by this
- * message.
- * </pre>
- *
- * <code>.google.protobuf.SourceContext source_context = 5;</code>
- */
- org.apache.hadoop.hbase.shaded.com.google.protobuf.SourceContextOrBuilder getSourceContextOrBuilder();
-
- /**
- * <pre>
- * Included APIs. See [Mixin][].
- * </pre>
- *
- * <code>repeated .google.protobuf.Mixin mixins = 6;</code>
- */
- java.util.List<org.apache.hadoop.hbase.shaded.com.google.protobuf.Mixin>
- getMixinsList();
- /**
- * <pre>
- * Included APIs. See [Mixin][].
- * </pre>
- *
- * <code>repeated .google.protobuf.Mixin mixins = 6;</code>
- */
- org.apache.hadoop.hbase.shaded.com.google.protobuf.Mixin getMixins(int index);
- /**
- * <pre>
- * Included APIs. See [Mixin][].
- * </pre>
- *
- * <code>repeated .google.protobuf.Mixin mixins = 6;</code>
- */
- int getMixinsCount();
- /**
- * <pre>
- * Included APIs. See [Mixin][].
- * </pre>
- *
- * <code>repeated .google.protobuf.Mixin mixins = 6;</code>
- */
- java.util.List<? extends org.apache.hadoop.hbase.shaded.com.google.protobuf.MixinOrBuilder>
- getMixinsOrBuilderList();
- /**
- * <pre>
- * Included APIs. See [Mixin][].
- * </pre>
- *
- * <code>repeated .google.protobuf.Mixin mixins = 6;</code>
- */
- org.apache.hadoop.hbase.shaded.com.google.protobuf.MixinOrBuilder getMixinsOrBuilder(
- int index);
-
- /**
- * <pre>
- * The source syntax of the service.
- * </pre>
- *
- * <code>.google.protobuf.Syntax syntax = 7;</code>
- */
- int getSyntaxValue();
- /**
- * <pre>
- * The source syntax of the service.
- * </pre>
- *
- * <code>.google.protobuf.Syntax syntax = 7;</code>
- */
- org.apache.hadoop.hbase.shaded.com.google.protobuf.Syntax getSyntax();
-}
http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/ApiProto.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/ApiProto.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/ApiProto.java
deleted file mode 100644
index 817e07d..0000000
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/ApiProto.java
+++ /dev/null
@@ -1,98 +0,0 @@
-// Generated by the protocol buffer compiler. DO NOT EDIT!
-// source: google/protobuf/api.proto
-
-package org.apache.hadoop.hbase.shaded.com.google.protobuf;
-
-public final class ApiProto {
- private ApiProto() {}
- public static void registerAllExtensions(
- org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite registry) {
- }
-
- public static void registerAllExtensions(
- org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistry registry) {
- registerAllExtensions(
- (org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite) registry);
- }
- static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
- internal_static_google_protobuf_Api_descriptor;
- static final
- org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
- internal_static_google_protobuf_Api_fieldAccessorTable;
- static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
- internal_static_google_protobuf_Method_descriptor;
- static final
- org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
- internal_static_google_protobuf_Method_fieldAccessorTable;
- static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
- internal_static_google_protobuf_Mixin_descriptor;
- static final
- org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
- internal_static_google_protobuf_Mixin_fieldAccessorTable;
-
- public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor
- getDescriptor() {
- return descriptor;
- }
- private static org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor
- descriptor;
- static {
- java.lang.String[] descriptorData = {
- "\n\031google/protobuf/api.proto\022\017google.prot" +
- "obuf\032$google/protobuf/source_context.pro" +
- "to\032\032google/protobuf/type.proto\"\201\002\n\003Api\022\014" +
- "\n\004name\030\001 \001(\t\022(\n\007methods\030\002 \003(\0132\027.google.p" +
- "rotobuf.Method\022(\n\007options\030\003 \003(\0132\027.google" +
- ".protobuf.Option\022\017\n\007version\030\004 \001(\t\0226\n\016sou" +
- "rce_context\030\005 \001(\0132\036.google.protobuf.Sour" +
- "ceContext\022&\n\006mixins\030\006 \003(\0132\026.google.proto" +
- "buf.Mixin\022\'\n\006syntax\030\007 \001(\0162\027.google.proto" +
- "buf.Syntax\"\325\001\n\006Method\022\014\n\004name\030\001 \001(\t\022\030\n\020r",
- "equest_type_url\030\002 \001(\t\022\031\n\021request_streami" +
- "ng\030\003 \001(\010\022\031\n\021response_type_url\030\004 \001(\t\022\032\n\022r" +
- "esponse_streaming\030\005 \001(\010\022(\n\007options\030\006 \003(\013" +
- "2\027.google.protobuf.Option\022\'\n\006syntax\030\007 \001(" +
- "\0162\027.google.protobuf.Syntax\"#\n\005Mixin\022\014\n\004n" +
- "ame\030\001 \001(\t\022\014\n\004root\030\002 \001(\tBu\n\023com.google.pr" +
- "otobufB\010ApiProtoP\001Z+google.golang.org/ge" +
- "nproto/protobuf/api;api\242\002\003GPB\252\002\036Google.P" +
- "rotobuf.WellKnownTypesb\006proto3"
- };
- org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
- new org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor. InternalDescriptorAssigner() {
- public org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistry assignDescriptors(
- org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor root) {
- descriptor = root;
- return null;
- }
- };
- org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor
- .internalBuildGeneratedFileFrom(descriptorData,
- new org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor[] {
- org.apache.hadoop.hbase.shaded.com.google.protobuf.SourceContextProto.getDescriptor(),
- org.apache.hadoop.hbase.shaded.com.google.protobuf.TypeProto.getDescriptor(),
- }, assigner);
- internal_static_google_protobuf_Api_descriptor =
- getDescriptor().getMessageTypes().get(0);
- internal_static_google_protobuf_Api_fieldAccessorTable = new
- org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable(
- internal_static_google_protobuf_Api_descriptor,
- new java.lang.String[] { "Name", "Methods", "Options", "Version", "SourceContext", "Mixins", "Syntax", });
- internal_static_google_protobuf_Method_descriptor =
- getDescriptor().getMessageTypes().get(1);
- internal_static_google_protobuf_Method_fieldAccessorTable = new
- org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable(
- internal_static_google_protobuf_Method_descriptor,
- new java.lang.String[] { "Name", "RequestTypeUrl", "RequestStreaming", "ResponseTypeUrl", "ResponseStreaming", "Options", "Syntax", });
- internal_static_google_protobuf_Mixin_descriptor =
- getDescriptor().getMessageTypes().get(2);
- internal_static_google_protobuf_Mixin_fieldAccessorTable = new
- org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable(
- internal_static_google_protobuf_Mixin_descriptor,
- new java.lang.String[] { "Name", "Root", });
- org.apache.hadoop.hbase.shaded.com.google.protobuf.SourceContextProto.getDescriptor();
- org.apache.hadoop.hbase.shaded.com.google.protobuf.TypeProto.getDescriptor();
- }
-
- // @@protoc_insertion_point(outer_class_scope)
-}
http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/BlockingRpcChannel.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/BlockingRpcChannel.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/BlockingRpcChannel.java
deleted file mode 100644
index bd2f2cd..0000000
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/BlockingRpcChannel.java
+++ /dev/null
@@ -1,51 +0,0 @@
-// Protocol Buffers - Google's data interchange format
-// Copyright 2008 Google Inc. All rights reserved.
-// https://developers.google.com/protocol-buffers/
-//
-// Redistribution and use in source and binary forms, with or without
-// modification, are permitted provided that the following conditions are
-// met:
-//
-// * Redistributions of source code must retain the above copyright
-// notice, this list of conditions and the following disclaimer.
-// * Redistributions in binary form must reproduce the above
-// copyright notice, this list of conditions and the following disclaimer
-// in the documentation and/or other materials provided with the
-// distribution.
-// * Neither the name of Google Inc. nor the names of its
-// contributors may be used to endorse or promote products derived from
-// this software without specific prior written permission.
-//
-// THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
-// "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
-// LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
-// A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
-// OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
-// SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
-// LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
-// DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
-// THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-// (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
-// OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-
-package org.apache.hadoop.hbase.shaded.com.google.protobuf;
-
-/**
- * <p>Abstract interface for a blocking RPC channel. {@code BlockingRpcChannel}
- * is the blocking equivalent to {@link RpcChannel}.
- *
- * @author kenton@google.com Kenton Varda
- * @author cpovirk@google.com Chris Povirk
- */
-public interface BlockingRpcChannel {
- /**
- * Call the given method of the remote service and blocks until it returns.
- * {@code callBlockingMethod()} is the blocking equivalent to
- * {@link RpcChannel#callMethod}.
- */
- Message callBlockingMethod(
- Descriptors.MethodDescriptor method,
- RpcController controller,
- Message request,
- Message responsePrototype) throws ServiceException;
-}
http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/BlockingService.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/BlockingService.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/BlockingService.java
deleted file mode 100644
index f496f2d..0000000
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/BlockingService.java
+++ /dev/null
@@ -1,64 +0,0 @@
-// Protocol Buffers - Google's data interchange format
-// Copyright 2008 Google Inc. All rights reserved.
-// https://developers.google.com/protocol-buffers/
-//
-// Redistribution and use in source and binary forms, with or without
-// modification, are permitted provided that the following conditions are
-// met:
-//
-// * Redistributions of source code must retain the above copyright
-// notice, this list of conditions and the following disclaimer.
-// * Redistributions in binary form must reproduce the above
-// copyright notice, this list of conditions and the following disclaimer
-// in the documentation and/or other materials provided with the
-// distribution.
-// * Neither the name of Google Inc. nor the names of its
-// contributors may be used to endorse or promote products derived from
-// this software without specific prior written permission.
-//
-// THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
-// "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
-// LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
-// A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
-// OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
-// SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
-// LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
-// DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
-// THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-// (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
-// OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-
-package org.apache.hadoop.hbase.shaded.com.google.protobuf;
-
-/**
- * Blocking equivalent to {@link Service}.
- *
- * @author kenton@google.com Kenton Varda
- * @author cpovirk@google.com Chris Povirk
- */
-public interface BlockingService {
- /**
- * Equivalent to {@link Service#getDescriptorForType}.
- */
- Descriptors.ServiceDescriptor getDescriptorForType();
-
- /**
- * Equivalent to {@link Service#callMethod}, except that
- * {@code callBlockingMethod()} returns the result of the RPC or throws a
- * {@link ServiceException} if there is a failure, rather than passing the
- * information to a callback.
- */
- Message callBlockingMethod(Descriptors.MethodDescriptor method,
- RpcController controller,
- Message request) throws ServiceException;
-
- /**
- * Equivalent to {@link Service#getRequestPrototype}.
- */
- Message getRequestPrototype(Descriptors.MethodDescriptor method);
-
- /**
- * Equivalent to {@link Service#getResponsePrototype}.
- */
- Message getResponsePrototype(Descriptors.MethodDescriptor method);
-}
http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/BoolValue.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/BoolValue.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/BoolValue.java
deleted file mode 100644
index a2dbe4a..0000000
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/BoolValue.java
+++ /dev/null
@@ -1,452 +0,0 @@
-// Generated by the protocol buffer compiler. DO NOT EDIT!
-// source: google/protobuf/wrappers.proto
-
-package org.apache.hadoop.hbase.shaded.com.google.protobuf;
-
-/**
- * <pre>
- * Wrapper message for `bool`.
- * The JSON representation for `BoolValue` is JSON `true` and `false`.
- * </pre>
- *
- * Protobuf type {@code google.protobuf.BoolValue}
- */
-public final class BoolValue extends
- org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 implements
- // @@protoc_insertion_point(message_implements:google.protobuf.BoolValue)
- BoolValueOrBuilder {
- // Use BoolValue.newBuilder() to construct.
- private BoolValue(org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
- super(builder);
- }
- private BoolValue() {
- value_ = false;
- }
-
- @java.lang.Override
- public final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet
- getUnknownFields() {
- return org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.getDefaultInstance();
- }
- private BoolValue(
- org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
- org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
- this();
- int mutable_bitField0_ = 0;
- try {
- boolean done = false;
- while (!done) {
- int tag = input.readTag();
- switch (tag) {
- case 0:
- done = true;
- break;
- default: {
- if (!input.skipField(tag)) {
- done = true;
- }
- break;
- }
- case 8: {
-
- value_ = input.readBool();
- break;
- }
- }
- }
- } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
- throw e.setUnfinishedMessage(this);
- } catch (java.io.IOException e) {
- throw new org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException(
- e).setUnfinishedMessage(this);
- } finally {
- makeExtensionsImmutable();
- }
- }
- public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
- getDescriptor() {
- return org.apache.hadoop.hbase.shaded.com.google.protobuf.WrappersProto.internal_static_google_protobuf_BoolValue_descriptor;
- }
-
- protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
- internalGetFieldAccessorTable() {
- return org.apache.hadoop.hbase.shaded.com.google.protobuf.WrappersProto.internal_static_google_protobuf_BoolValue_fieldAccessorTable
- .ensureFieldAccessorsInitialized(
- org.apache.hadoop.hbase.shaded.com.google.protobuf.BoolValue.class, org.apache.hadoop.hbase.shaded.com.google.protobuf.BoolValue.Builder.class);
- }
-
- public static final int VALUE_FIELD_NUMBER = 1;
- private boolean value_;
- /**
- * <pre>
- * The bool value.
- * </pre>
- *
- * <code>bool value = 1;</code>
- */
- public boolean getValue() {
- return value_;
- }
-
- private byte memoizedIsInitialized = -1;
- public final boolean isInitialized() {
- byte isInitialized = memoizedIsInitialized;
- if (isInitialized == 1) return true;
- if (isInitialized == 0) return false;
-
- memoizedIsInitialized = 1;
- return true;
- }
-
- public void writeTo(org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream output)
- throws java.io.IOException {
- if (value_ != false) {
- output.writeBool(1, value_);
- }
- }
-
- public int getSerializedSize() {
- int size = memoizedSize;
- if (size != -1) return size;
-
- size = 0;
- if (value_ != false) {
- size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
- .computeBoolSize(1, value_);
- }
- memoizedSize = size;
- return size;
- }
-
- private static final long serialVersionUID = 0L;
- @java.lang.Override
- public boolean equals(final java.lang.Object obj) {
- if (obj == this) {
- return true;
- }
- if (!(obj instanceof org.apache.hadoop.hbase.shaded.com.google.protobuf.BoolValue)) {
- return super.equals(obj);
- }
- org.apache.hadoop.hbase.shaded.com.google.protobuf.BoolValue other = (org.apache.hadoop.hbase.shaded.com.google.protobuf.BoolValue) obj;
-
- boolean result = true;
- result = result && (getValue()
- == other.getValue());
- return result;
- }
-
- @java.lang.Override
- public int hashCode() {
- if (memoizedHashCode != 0) {
- return memoizedHashCode;
- }
- int hash = 41;
- hash = (19 * hash) + getDescriptor().hashCode();
- hash = (37 * hash) + VALUE_FIELD_NUMBER;
- hash = (53 * hash) + org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.hashBoolean(
- getValue());
- hash = (29 * hash) + unknownFields.hashCode();
- memoizedHashCode = hash;
- return hash;
- }
-
- public static org.apache.hadoop.hbase.shaded.com.google.protobuf.BoolValue parseFrom(
- org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data)
- throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
- return PARSER.parseFrom(data);
- }
- public static org.apache.hadoop.hbase.shaded.com.google.protobuf.BoolValue parseFrom(
- org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data,
- org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
- return PARSER.parseFrom(data, extensionRegistry);
- }
- public static org.apache.hadoop.hbase.shaded.com.google.protobuf.BoolValue parseFrom(byte[] data)
- throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
- return PARSER.parseFrom(data);
- }
- public static org.apache.hadoop.hbase.shaded.com.google.protobuf.BoolValue parseFrom(
- byte[] data,
- org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
- return PARSER.parseFrom(data, extensionRegistry);
- }
- public static org.apache.hadoop.hbase.shaded.com.google.protobuf.BoolValue parseFrom(java.io.InputStream input)
- throws java.io.IOException {
- return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
- .parseWithIOException(PARSER, input);
- }
- public static org.apache.hadoop.hbase.shaded.com.google.protobuf.BoolValue parseFrom(
- java.io.InputStream input,
- org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws java.io.IOException {
- return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
- .parseWithIOException(PARSER, input, extensionRegistry);
- }
- public static org.apache.hadoop.hbase.shaded.com.google.protobuf.BoolValue parseDelimitedFrom(java.io.InputStream input)
- throws java.io.IOException {
- return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
- .parseDelimitedWithIOException(PARSER, input);
- }
- public static org.apache.hadoop.hbase.shaded.com.google.protobuf.BoolValue parseDelimitedFrom(
- java.io.InputStream input,
- org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws java.io.IOException {
- return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
- .parseDelimitedWithIOException(PARSER, input, extensionRegistry);
- }
- public static org.apache.hadoop.hbase.shaded.com.google.protobuf.BoolValue parseFrom(
- org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input)
- throws java.io.IOException {
- return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
- .parseWithIOException(PARSER, input);
- }
- public static org.apache.hadoop.hbase.shaded.com.google.protobuf.BoolValue parseFrom(
- org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
- org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws java.io.IOException {
- return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
- .parseWithIOException(PARSER, input, extensionRegistry);
- }
-
- public Builder newBuilderForType() { return newBuilder(); }
- public static Builder newBuilder() {
- return DEFAULT_INSTANCE.toBuilder();
- }
- public static Builder newBuilder(org.apache.hadoop.hbase.shaded.com.google.protobuf.BoolValue prototype) {
- return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype);
- }
- public Builder toBuilder() {
- return this == DEFAULT_INSTANCE
- ? new Builder() : new Builder().mergeFrom(this);
- }
-
- @java.lang.Override
- protected Builder newBuilderForType(
- org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
- Builder builder = new Builder(parent);
- return builder;
- }
- /**
- * <pre>
- * Wrapper message for `bool`.
- * The JSON representation for `BoolValue` is JSON `true` and `false`.
- * </pre>
- *
- * Protobuf type {@code google.protobuf.BoolValue}
- */
- public static final class Builder extends
- org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
- // @@protoc_insertion_point(builder_implements:google.protobuf.BoolValue)
- org.apache.hadoop.hbase.shaded.com.google.protobuf.BoolValueOrBuilder {
- public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
- getDescriptor() {
- return org.apache.hadoop.hbase.shaded.com.google.protobuf.WrappersProto.internal_static_google_protobuf_BoolValue_descriptor;
- }
-
- protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
- internalGetFieldAccessorTable() {
- return org.apache.hadoop.hbase.shaded.com.google.protobuf.WrappersProto.internal_static_google_protobuf_BoolValue_fieldAccessorTable
- .ensureFieldAccessorsInitialized(
- org.apache.hadoop.hbase.shaded.com.google.protobuf.BoolValue.class, org.apache.hadoop.hbase.shaded.com.google.protobuf.BoolValue.Builder.class);
- }
-
- // Construct using org.apache.hadoop.hbase.shaded.com.google.protobuf.BoolValue.newBuilder()
- private Builder() {
- maybeForceBuilderInitialization();
- }
-
- private Builder(
- org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
- super(parent);
- maybeForceBuilderInitialization();
- }
- private void maybeForceBuilderInitialization() {
- if (org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
- .alwaysUseFieldBuilders) {
- }
- }
- public Builder clear() {
- super.clear();
- value_ = false;
-
- return this;
- }
-
- public org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
- getDescriptorForType() {
- return org.apache.hadoop.hbase.shaded.com.google.protobuf.WrappersProto.internal_static_google_protobuf_BoolValue_descriptor;
- }
-
- public org.apache.hadoop.hbase.shaded.com.google.protobuf.BoolValue getDefaultInstanceForType() {
- return org.apache.hadoop.hbase.shaded.com.google.protobuf.BoolValue.getDefaultInstance();
- }
-
- public org.apache.hadoop.hbase.shaded.com.google.protobuf.BoolValue build() {
- org.apache.hadoop.hbase.shaded.com.google.protobuf.BoolValue result = buildPartial();
- if (!result.isInitialized()) {
- throw newUninitializedMessageException(result);
- }
- return result;
- }
-
- public org.apache.hadoop.hbase.shaded.com.google.protobuf.BoolValue buildPartial() {
- org.apache.hadoop.hbase.shaded.com.google.protobuf.BoolValue result = new org.apache.hadoop.hbase.shaded.com.google.protobuf.BoolValue(this);
- result.value_ = value_;
- onBuilt();
- return result;
- }
-
- public Builder clone() {
- return (Builder) super.clone();
- }
- public Builder setField(
- org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
- Object value) {
- return (Builder) super.setField(field, value);
- }
- public Builder clearField(
- org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field) {
- return (Builder) super.clearField(field);
- }
- public Builder clearOneof(
- org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.OneofDescriptor oneof) {
- return (Builder) super.clearOneof(oneof);
- }
- public Builder setRepeatedField(
- org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
- int index, Object value) {
- return (Builder) super.setRepeatedField(field, index, value);
- }
- public Builder addRepeatedField(
- org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
- Object value) {
- return (Builder) super.addRepeatedField(field, value);
- }
- public Builder mergeFrom(org.apache.hadoop.hbase.shaded.com.google.protobuf.Message other) {
- if (other instanceof org.apache.hadoop.hbase.shaded.com.google.protobuf.BoolValue) {
- return mergeFrom((org.apache.hadoop.hbase.shaded.com.google.protobuf.BoolValue)other);
- } else {
- super.mergeFrom(other);
- return this;
- }
- }
-
- public Builder mergeFrom(org.apache.hadoop.hbase.shaded.com.google.protobuf.BoolValue other) {
- if (other == org.apache.hadoop.hbase.shaded.com.google.protobuf.BoolValue.getDefaultInstance()) return this;
- if (other.getValue() != false) {
- setValue(other.getValue());
- }
- onChanged();
- return this;
- }
-
- public final boolean isInitialized() {
- return true;
- }
-
- public Builder mergeFrom(
- org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
- org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws java.io.IOException {
- org.apache.hadoop.hbase.shaded.com.google.protobuf.BoolValue parsedMessage = null;
- try {
- parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
- } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
- parsedMessage = (org.apache.hadoop.hbase.shaded.com.google.protobuf.BoolValue) e.getUnfinishedMessage();
- throw e.unwrapIOException();
- } finally {
- if (parsedMessage != null) {
- mergeFrom(parsedMessage);
- }
- }
- return this;
- }
-
- private boolean value_ ;
- /**
- * <pre>
- * The bool value.
- * </pre>
- *
- * <code>bool value = 1;</code>
- */
- public boolean getValue() {
- return value_;
- }
- /**
- * <pre>
- * The bool value.
- * </pre>
- *
- * <code>bool value = 1;</code>
- */
- public Builder setValue(boolean value) {
-
- value_ = value;
- onChanged();
- return this;
- }
- /**
- * <pre>
- * The bool value.
- * </pre>
- *
- * <code>bool value = 1;</code>
- */
- public Builder clearValue() {
-
- value_ = false;
- onChanged();
- return this;
- }
- public final Builder setUnknownFields(
- final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
- return this;
- }
-
- public final Builder mergeUnknownFields(
- final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
- return this;
- }
-
-
- // @@protoc_insertion_point(builder_scope:google.protobuf.BoolValue)
- }
-
- // @@protoc_insertion_point(class_scope:google.protobuf.BoolValue)
- private static final org.apache.hadoop.hbase.shaded.com.google.protobuf.BoolValue DEFAULT_INSTANCE;
- static {
- DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.com.google.protobuf.BoolValue();
- }
-
- public static org.apache.hadoop.hbase.shaded.com.google.protobuf.BoolValue getDefaultInstance() {
- return DEFAULT_INSTANCE;
- }
-
- private static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<BoolValue>
- PARSER = new org.apache.hadoop.hbase.shaded.com.google.protobuf.AbstractParser<BoolValue>() {
- public BoolValue parsePartialFrom(
- org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
- org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
- return new BoolValue(input, extensionRegistry);
- }
- };
-
- public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<BoolValue> parser() {
- return PARSER;
- }
-
- @java.lang.Override
- public org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<BoolValue> getParserForType() {
- return PARSER;
- }
-
- public org.apache.hadoop.hbase.shaded.com.google.protobuf.BoolValue getDefaultInstanceForType() {
- return DEFAULT_INSTANCE;
- }
-
-}
-
http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/BoolValueOrBuilder.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/BoolValueOrBuilder.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/BoolValueOrBuilder.java
deleted file mode 100644
index c1d3b5f..0000000
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/BoolValueOrBuilder.java
+++ /dev/null
@@ -1,18 +0,0 @@
-// Generated by the protocol buffer compiler. DO NOT EDIT!
-// source: google/protobuf/wrappers.proto
-
-package org.apache.hadoop.hbase.shaded.com.google.protobuf;
-
-public interface BoolValueOrBuilder extends
- // @@protoc_insertion_point(interface_extends:google.protobuf.BoolValue)
- org.apache.hadoop.hbase.shaded.com.google.protobuf.MessageOrBuilder {
-
- /**
- * <pre>
- * The bool value.
- * </pre>
- *
- * <code>bool value = 1;</code>
- */
- boolean getValue();
-}
http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/BooleanArrayList.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/BooleanArrayList.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/BooleanArrayList.java
deleted file mode 100644
index 6bc9641..0000000
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/BooleanArrayList.java
+++ /dev/null
@@ -1,272 +0,0 @@
-// Protocol Buffers - Google's data interchange format
-// Copyright 2008 Google Inc. All rights reserved.
-// https://developers.google.com/protocol-buffers/
-//
-// Redistribution and use in source and binary forms, with or without
-// modification, are permitted provided that the following conditions are
-// met:
-//
-// * Redistributions of source code must retain the above copyright
-// notice, this list of conditions and the following disclaimer.
-// * Redistributions in binary form must reproduce the above
-// copyright notice, this list of conditions and the following disclaimer
-// in the documentation and/or other materials provided with the
-// distribution.
-// * Neither the name of Google Inc. nor the names of its
-// contributors may be used to endorse or promote products derived from
-// this software without specific prior written permission.
-//
-// THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
-// "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
-// LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
-// A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
-// OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
-// SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
-// LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
-// DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
-// THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-// (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
-// OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-
-package org.apache.hadoop.hbase.shaded.com.google.protobuf;
-
-import org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.BooleanList;
-
-import java.util.Arrays;
-import java.util.Collection;
-import java.util.RandomAccess;
-
-/**
- * An implementation of {@link BooleanList} on top of a primitive array.
- *
- * @author dweis@google.com (Daniel Weis)
- */
-final class BooleanArrayList
- extends AbstractProtobufList<Boolean>
- implements BooleanList, RandomAccess {
-
- private static final BooleanArrayList EMPTY_LIST = new BooleanArrayList();
- static {
- EMPTY_LIST.makeImmutable();
- }
-
- public static BooleanArrayList emptyList() {
- return EMPTY_LIST;
- }
-
- /**
- * The backing store for the list.
- */
- private boolean[] array;
-
- /**
- * The size of the list distinct from the length of the array. That is, it is the number of
- * elements set in the list.
- */
- private int size;
-
- /**
- * Constructs a new mutable {@code BooleanArrayList} with default capacity.
- */
- BooleanArrayList() {
- this(new boolean[DEFAULT_CAPACITY], 0);
- }
-
- /**
- * Constructs a new mutable {@code BooleanArrayList}
- * containing the same elements as {@code other}.
- */
- private BooleanArrayList(boolean[] other, int size) {
- array = other;
- this.size = size;
- }
-
- @Override
- public boolean equals(Object o) {
- if (this == o) {
- return true;
- }
- if (!(o instanceof BooleanArrayList)) {
- return super.equals(o);
- }
- BooleanArrayList other = (BooleanArrayList) o;
- if (size != other.size) {
- return false;
- }
-
- final boolean[] arr = other.array;
- for (int i = 0; i < size; i++) {
- if (array[i] != arr[i]) {
- return false;
- }
- }
-
- return true;
- }
-
- @Override
- public int hashCode() {
- int result = 1;
- for (int i = 0; i < size; i++) {
- result = (31 * result) + Internal.hashBoolean(array[i]);
- }
- return result;
- }
-
- @Override
- public BooleanList mutableCopyWithCapacity(int capacity) {
- if (capacity < size) {
- throw new IllegalArgumentException();
- }
- return new BooleanArrayList(Arrays.copyOf(array, capacity), size);
- }
-
- @Override
- public Boolean get(int index) {
- return getBoolean(index);
- }
-
- @Override
- public boolean getBoolean(int index) {
- ensureIndexInRange(index);
- return array[index];
- }
-
- @Override
- public int size() {
- return size;
- }
-
- @Override
- public Boolean set(int index, Boolean element) {
- return setBoolean(index, element);
- }
-
- @Override
- public boolean setBoolean(int index, boolean element) {
- ensureIsMutable();
- ensureIndexInRange(index);
- boolean previousValue = array[index];
- array[index] = element;
- return previousValue;
- }
-
- @Override
- public void add(int index, Boolean element) {
- addBoolean(index, element);
- }
-
- /**
- * Like {@link #add(Boolean)} but more efficient in that it doesn't box the element.
- */
- @Override
- public void addBoolean(boolean element) {
- addBoolean(size, element);
- }
-
- /**
- * Like {@link #add(int, Boolean)} but more efficient in that it doesn't box the element.
- */
- private void addBoolean(int index, boolean element) {
- ensureIsMutable();
- if (index < 0 || index > size) {
- throw new IndexOutOfBoundsException(makeOutOfBoundsExceptionMessage(index));
- }
-
- if (size < array.length) {
- // Shift everything over to make room
- System.arraycopy(array, index, array, index + 1, size - index);
- } else {
- // Resize to 1.5x the size
- int length = ((size * 3) / 2) + 1;
- boolean[] newArray = new boolean[length];
-
- // Copy the first part directly
- System.arraycopy(array, 0, newArray, 0, index);
-
- // Copy the rest shifted over by one to make room
- System.arraycopy(array, index, newArray, index + 1, size - index);
- array = newArray;
- }
-
- array[index] = element;
- size++;
- modCount++;
- }
-
- @Override
- public boolean addAll(Collection<? extends Boolean> collection) {
- ensureIsMutable();
-
- if (collection == null) {
- throw new NullPointerException();
- }
-
- // We specialize when adding another BooleanArrayList to avoid boxing elements.
- if (!(collection instanceof BooleanArrayList)) {
- return super.addAll(collection);
- }
-
- BooleanArrayList list = (BooleanArrayList) collection;
- if (list.size == 0) {
- return false;
- }
-
- int overflow = Integer.MAX_VALUE - size;
- if (overflow < list.size) {
- // We can't actually represent a list this large.
- throw new OutOfMemoryError();
- }
-
- int newSize = size + list.size;
- if (newSize > array.length) {
- array = Arrays.copyOf(array, newSize);
- }
-
- System.arraycopy(list.array, 0, array, size, list.size);
- size = newSize;
- modCount++;
- return true;
- }
-
- @Override
- public boolean remove(Object o) {
- ensureIsMutable();
- for (int i = 0; i < size; i++) {
- if (o.equals(array[i])) {
- System.arraycopy(array, i + 1, array, i, size - i);
- size--;
- modCount++;
- return true;
- }
- }
- return false;
- }
-
- @Override
- public Boolean remove(int index) {
- ensureIsMutable();
- ensureIndexInRange(index);
- boolean value = array[index];
- System.arraycopy(array, index + 1, array, index, size - index);
- size--;
- modCount++;
- return value;
- }
-
- /**
- * Ensures that the provided {@code index} is within the range of {@code [0, size]}. Throws an
- * {@link IndexOutOfBoundsException} if it is not.
- *
- * @param index the index to verify is in range
- */
- private void ensureIndexInRange(int index) {
- if (index < 0 || index >= size) {
- throw new IndexOutOfBoundsException(makeOutOfBoundsExceptionMessage(index));
- }
- }
-
- private String makeOutOfBoundsExceptionMessage(int index) {
- return "Index:" + index + ", Size:" + size;
- }
-}
http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/ByteBufferWriter.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/ByteBufferWriter.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/ByteBufferWriter.java
deleted file mode 100644
index 553243a..0000000
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/ByteBufferWriter.java
+++ /dev/null
@@ -1,185 +0,0 @@
-// Protocol Buffers - Google's data interchange format
-// Copyright 2008 Google Inc. All rights reserved.
-// https://developers.google.com/protocol-buffers/
-//
-// Redistribution and use in source and binary forms, with or without
-// modification, are permitted provided that the following conditions are
-// met:
-//
-// * Redistributions of source code must retain the above copyright
-// notice, this list of conditions and the following disclaimer.
-// * Redistributions in binary form must reproduce the above
-// copyright notice, this list of conditions and the following disclaimer
-// in the documentation and/or other materials provided with the
-// distribution.
-// * Neither the name of Google Inc. nor the names of its
-// contributors may be used to endorse or promote products derived from
-// this software without specific prior written permission.
-//
-// THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
-// "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
-// LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
-// A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
-// OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
-// SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
-// LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
-// DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
-// THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-// (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
-// OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-
-package org.apache.hadoop.hbase.shaded.com.google.protobuf;
-
-import static java.lang.Math.max;
-import static java.lang.Math.min;
-
-import java.io.IOException;
-import java.io.OutputStream;
-import java.lang.ref.SoftReference;
-import java.lang.reflect.Field;
-import java.nio.ByteBuffer;
-import java.nio.channels.WritableByteChannel;
-
-/**
- * Utility class to provide efficient writing of {@link ByteBuffer}s to {@link OutputStream}s.
- */
-final class ByteBufferWriter {
- private ByteBufferWriter() {}
-
- /**
- * Minimum size for a cached buffer. This prevents us from allocating buffers that are too
- * small to be easily reused.
- */
- // TODO(nathanmittler): tune this property or allow configuration?
- private static final int MIN_CACHED_BUFFER_SIZE = 1024;
-
- /**
- * Maximum size for a cached buffer. If a larger buffer is required, it will be allocated
- * but not cached.
- */
- // TODO(nathanmittler): tune this property or allow configuration?
- private static final int MAX_CACHED_BUFFER_SIZE = 16 * 1024;
-
- /**
- * The fraction of the requested buffer size under which the buffer will be reallocated.
- */
- // TODO(nathanmittler): tune this property or allow configuration?
- private static final float BUFFER_REALLOCATION_THRESHOLD = 0.5f;
-
- /**
- * Keeping a soft reference to a thread-local buffer. This buffer is used for writing a
- * {@link ByteBuffer} to an {@link OutputStream} when no zero-copy alternative was available.
- * Using a "soft" reference since VMs may keep this reference around longer than "weak"
- * (e.g. HotSpot will maintain soft references until memory pressure warrants collection).
- */
- private static final ThreadLocal<SoftReference<byte[]>> BUFFER =
- new ThreadLocal<SoftReference<byte[]>>();
-
- /**
- * This is a hack for GAE, where {@code FileOutputStream} is unavailable.
- */
- private static final Class<?> FILE_OUTPUT_STREAM_CLASS = safeGetClass("java.io.FileOutputStream");
- private static final long CHANNEL_FIELD_OFFSET = getChannelFieldOffset(FILE_OUTPUT_STREAM_CLASS);
-
- /**
- * For testing purposes only. Clears the cached buffer to force a new allocation on the next
- * invocation.
- */
- static void clearCachedBuffer() {
- BUFFER.set(null);
- }
-
- /**
- * Writes the remaining content of the buffer to the given stream. The buffer {@code position}
- * will remain unchanged by this method.
- */
- static void write(ByteBuffer buffer, OutputStream output) throws IOException {
- final int initialPos = buffer.position();
- try {
- if (buffer.hasArray()) {
- // Optimized write for array-backed buffers.
- // Note that we're taking the risk that a malicious OutputStream could modify the array.
- output.write(buffer.array(), buffer.arrayOffset() + buffer.position(), buffer.remaining());
- } else if (!writeToChannel(buffer, output)){
- // Read all of the data from the buffer to an array.
- // TODO(nathanmittler): Consider performance improvements for other "known" stream types.
- final byte[] array = getOrCreateBuffer(buffer.remaining());
- while (buffer.hasRemaining()) {
- int length = min(buffer.remaining(), array.length);
- buffer.get(array, 0, length);
- output.write(array, 0, length);
- }
- }
- } finally {
- // Restore the initial position.
- buffer.position(initialPos);
- }
- }
-
- static byte[] getOrCreateBuffer(int requestedSize) {
- requestedSize = max(requestedSize, MIN_CACHED_BUFFER_SIZE);
-
- byte[] buffer = getBuffer();
- // Only allocate if we need to.
- if (buffer == null || needToReallocate(requestedSize, buffer.length)) {
- buffer = new byte[requestedSize];
-
- // Only cache the buffer if it's not too big.
- if (requestedSize <= MAX_CACHED_BUFFER_SIZE) {
- setBuffer(buffer);
- }
- }
- return buffer;
- }
-
- private static boolean needToReallocate(int requestedSize, int bufferLength) {
- // First check against just the requested length to avoid the multiply.
- return bufferLength < requestedSize
- && bufferLength < requestedSize * BUFFER_REALLOCATION_THRESHOLD;
- }
-
- private static byte[] getBuffer() {
- SoftReference<byte[]> sr = BUFFER.get();
- return sr == null ? null : sr.get();
- }
-
- private static void setBuffer(byte[] value) {
- BUFFER.set(new SoftReference<byte[]>(value));
- }
-
- private static boolean writeToChannel(ByteBuffer buffer, OutputStream output) throws IOException {
- if (CHANNEL_FIELD_OFFSET >= 0 && FILE_OUTPUT_STREAM_CLASS.isInstance(output)) {
- // Use a channel to write out the ByteBuffer. This will automatically empty the buffer.
- WritableByteChannel channel = null;
- try {
- channel = (WritableByteChannel) UnsafeUtil.getObject(output, CHANNEL_FIELD_OFFSET);
- } catch (ClassCastException e) {
- // Absorb.
- }
- if (channel != null) {
- channel.write(buffer);
- return true;
- }
- }
- return false;
- }
-
- private static Class<?> safeGetClass(String className) {
- try {
- return Class.forName(className);
- } catch (ClassNotFoundException e) {
- return null;
- }
- }
- private static long getChannelFieldOffset(Class<?> clazz) {
- try {
- if (clazz != null && UnsafeUtil.hasUnsafeArrayOperations()) {
- Field field = clazz.getDeclaredField("channel");
- return UnsafeUtil.objectFieldOffset(field);
- }
- } catch (Throwable e) {
- // Absorb
- }
- return -1;
- }
-}
http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/ByteInput.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/ByteInput.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/ByteInput.java
deleted file mode 100644
index a745d37..0000000
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/ByteInput.java
+++ /dev/null
@@ -1,81 +0,0 @@
-// Protocol Buffers - Google's data interchange format
-// Copyright 2008 Google Inc. All rights reserved.
-// https://developers.google.com/protocol-buffers/
-//
-// Redistribution and use in source and binary forms, with or without
-// modification, are permitted provided that the following conditions are
-// met:
-//
-// * Redistributions of source code must retain the above copyright
-// notice, this list of conditions and the following disclaimer.
-// * Redistributions in binary form must reproduce the above
-// copyright notice, this list of conditions and the following disclaimer
-// in the documentation and/or other materials provided with the
-// distribution.
-// * Neither the name of Google Inc. nor the names of its
-// contributors may be used to endorse or promote products derived from
-// this software without specific prior written permission.
-//
-// THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
-// "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
-// LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
-// A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
-// OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
-// SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
-// LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
-// DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
-// THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-// (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
-// OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-package org.apache.hadoop.hbase.shaded.com.google.protobuf;
-
-import java.io.IOException;
-import java.nio.ByteBuffer;
-
-/**
- * An input for raw bytes. This is similar to an InputStream but it is offset addressable. All the
- * read APIs are relative.
- */
-@ExperimentalApi
-public abstract class ByteInput {
-
- /**
- * Reads a single byte from the given offset.
- * @param offset The offset from where byte to be read
- * @return The byte of data at given offset
- */
- public abstract byte read(int offset);
-
- /**
- * Reads bytes of data from the given offset into an array of bytes.
- * @param offset The src offset within this ByteInput from where data to be read.
- * @param out Destination byte array to read data into.
- * @return The number of bytes read from ByteInput
- */
- public int read(int offset, byte b[]) throws IOException {
- return read(offset, b, 0, b.length);
- }
-
- /**
- * Reads up to <code>len</code> bytes of data from the given offset into an array of bytes.
- * @param offset The src offset within this ByteInput from where data to be read.
- * @param out Destination byte array to read data into.
- * @param outOffset Offset within the the out byte[] where data to be read into.
- * @param len The number of bytes to read.
- * @return The number of bytes read from ByteInput
- */
- public abstract int read(int offset, byte[] out, int outOffset, int len);
-
- /**
- * Reads bytes of data from the given offset into given {@link ByteBuffer}.
- * @param offset he src offset within this ByteInput from where data to be read.
- * @param out Destination {@link ByteBuffer} to read data into.
- * @return The number of bytes read from ByteInput
- */
- public abstract int read(int offset, ByteBuffer out);
-
- /**
- * @return Total number of bytes in this ByteInput.
- */
- public abstract int size();
-}
http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/ByteInputByteString.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/ByteInputByteString.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/ByteInputByteString.java
deleted file mode 100644
index 30de4ec..0000000
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/ByteInputByteString.java
+++ /dev/null
@@ -1,251 +0,0 @@
-// Protocol Buffers - Google's data interchange format
-// Copyright 2008 Google Inc. All rights reserved.
-// https://developers.google.com/protocol-buffers/
-//
-// Redistribution and use in source and binary forms, with or without
-// modification, are permitted provided that the following conditions are
-// met:
-//
-// * Redistributions of source code must retain the above copyright
-// notice, this list of conditions and the following disclaimer.
-// * Redistributions in binary form must reproduce the above
-// copyright notice, this list of conditions and the following disclaimer
-// in the documentation and/or other materials provided with the
-// distribution.
-// * Neither the name of Google Inc. nor the names of its
-// contributors may be used to endorse or promote products derived from
-// this software without specific prior written permission.
-//
-// THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
-// "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
-// LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
-// A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
-// OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
-// SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
-// LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
-// DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
-// THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-// (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
-// OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-
-package org.apache.hadoop.hbase.shaded.com.google.protobuf;
-
-import java.io.IOException;
-import java.io.InputStream;
-import java.io.InvalidObjectException;
-import java.io.ObjectInputStream;
-import java.io.OutputStream;
-import java.nio.ByteBuffer;
-import java.nio.charset.Charset;
-import java.util.Arrays;
-import java.util.Collections;
-import java.util.List;
-
-/**
- * A {@link ByteString} that wraps around a {@link ByteInput}.
- */
-final class ByteInputByteString extends ByteString.LeafByteString {
- private final ByteInput buffer;
- private final int offset, length;
-
- ByteInputByteString(ByteInput buffer, int offset, int length) {
- if (buffer == null) {
- throw new NullPointerException("buffer");
- }
- this.buffer = buffer;
- this.offset = offset;
- this.length = length;
- }
-
- // =================================================================
- // Serializable
-
- /**
- * Magic method that lets us override serialization behavior.
- */
- private Object writeReplace() {
- return ByteString.wrap(toByteArray());
- }
-
- /**
- * Magic method that lets us override deserialization behavior.
- */
- private void readObject(@SuppressWarnings("unused") ObjectInputStream in) throws IOException {
- throw new InvalidObjectException("ByteInputByteString instances are not to be serialized directly");// TODO check here
- }
-
- // =================================================================
-
- @Override
- public byte byteAt(int index) {
- return buffer.read(getAbsoluteOffset(index));
- }
-
- private int getAbsoluteOffset(int relativeOffset) {
- return this.offset + relativeOffset;
- }
-
- @Override
- public int size() {
- return length;
- }
-
- @Override
- public ByteString substring(int beginIndex, int endIndex) {
- if (beginIndex < 0 || beginIndex >= size() || endIndex < beginIndex || endIndex >= size()) {
- throw new IllegalArgumentException(
- String.format("Invalid indices [%d, %d]", beginIndex, endIndex));
- }
- return new ByteInputByteString(this.buffer, getAbsoluteOffset(beginIndex), endIndex - beginIndex);
- }
-
- @Override
- protected void copyToInternal(
- byte[] target, int sourceOffset, int targetOffset, int numberToCopy) {
- this.buffer.read(getAbsoluteOffset(sourceOffset), target, targetOffset, numberToCopy);
- }
-
- @Override
- public void copyTo(ByteBuffer target) {
- this.buffer.read(this.offset, target);
- }
-
- @Override
- public void writeTo(OutputStream out) throws IOException {
- out.write(toByteArray());// TODO
- }
-
- @Override
- boolean equalsRange(ByteString other, int offset, int length) {
- return substring(0, length).equals(other.substring(offset, offset + length));
- }
-
- @Override
- void writeToInternal(OutputStream out, int sourceOffset, int numberToWrite) throws IOException {
- byte[] buf = ByteBufferWriter.getOrCreateBuffer(numberToWrite);
- this.buffer.read(getAbsoluteOffset(sourceOffset), buf, 0, numberToWrite);
- out.write(buf, 0, numberToWrite);
- }
-
- @Override
- void writeTo(ByteOutput output) throws IOException {
- output.writeLazy(toByteArray(), 0, length);
- }
-
- @Override
- public ByteBuffer asReadOnlyByteBuffer() {
- return ByteBuffer.wrap(toByteArray()).asReadOnlyBuffer();
- }
-
- @Override
- public List<ByteBuffer> asReadOnlyByteBufferList() {
- return Collections.singletonList(asReadOnlyByteBuffer());
- }
-
- @Override
- protected String toStringInternal(Charset charset) {
- byte[] bytes = toByteArray();
- return new String(bytes, 0, bytes.length, charset);
- }
-
- @Override
- public boolean isValidUtf8() {
- return Utf8.isValidUtf8(buffer, offset, offset + length);
- }
-
- @Override
- protected int partialIsValidUtf8(int state, int offset, int length) {
- int off = getAbsoluteOffset(offset);
- return Utf8.partialIsValidUtf8(state, buffer, off, off + length);
- }
-
- @Override
- public boolean equals(Object other) {
- if (other == this) {
- return true;
- }
- if (!(other instanceof ByteString)) {
- return false;
- }
- ByteString otherString = ((ByteString) other);
- if (size() != otherString.size()) {
- return false;
- }
- if (size() == 0) {
- return true;
- }
- if (other instanceof RopeByteString) {
- return other.equals(this);
- }
- return Arrays.equals(this.toByteArray(), otherString.toByteArray());
- }
-
- @Override
- protected int partialHash(int h, int offset, int length) {
- offset = getAbsoluteOffset(offset);
- int end = offset + length;
- for (int i = offset; i < end; i++) {
- h = h * 31 + buffer.read(i);
- }
- return h;
- }
-
- @Override
- public InputStream newInput() {
- return new InputStream() {
- private final ByteInput buf = buffer;
- private int pos = offset;
- private int limit = pos + length;
- private int mark = pos;
-
- @Override
- public void mark(int readlimit) {
- this.mark = readlimit;
- }
-
- @Override
- public boolean markSupported() {
- return true;
- }
-
- @Override
- public void reset() throws IOException {
- this.pos = this.mark;
- }
-
- @Override
- public int available() throws IOException {
- return this.limit - this.pos;
- }
-
- @Override
- public int read() throws IOException {
- if (available() <= 0) {
- return -1;
- }
- return this.buf.read(pos++) & 0xFF;
- }
-
- @Override
- public int read(byte[] bytes, int off, int len) throws IOException {
- int remain = available();
- if (remain <= 0) {
- return -1;
- }
- len = Math.min(len, remain);
- buf.read(pos, bytes, off, len);
- pos += len;
- return len;
- }
- };
- }
-
- @Override
- public CodedInputStream newCodedInput() {
- // We trust CodedInputStream not to modify the bytes, or to give anyone
- // else access to them.
- CodedInputStream cis = CodedInputStream.newInstance(buffer, offset, length, true);
- cis.enableAliasing(true);
- return cis;
- }
-}
http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/ByteOutput.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/ByteOutput.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/ByteOutput.java
deleted file mode 100644
index 8a83fb2..0000000
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/ByteOutput.java
+++ /dev/null
@@ -1,116 +0,0 @@
-// Protocol Buffers - Google's data interchange format
-// Copyright 2008 Google Inc. All rights reserved.
-// https://developers.google.com/protocol-buffers/
-//
-// Redistribution and use in source and binary forms, with or without
-// modification, are permitted provided that the following conditions are
-// met:
-//
-// * Redistributions of source code must retain the above copyright
-// notice, this list of conditions and the following disclaimer.
-// * Redistributions in binary form must reproduce the above
-// copyright notice, this list of conditions and the following disclaimer
-// in the documentation and/or other materials provided with the
-// distribution.
-// * Neither the name of Google Inc. nor the names of its
-// contributors may be used to endorse or promote products derived from
-// this software without specific prior written permission.
-//
-// THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
-// "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
-// LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
-// A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
-// OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
-// SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
-// LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
-// DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
-// THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-// (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
-// OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-
-package org.apache.hadoop.hbase.shaded.com.google.protobuf;
-
-import java.io.IOException;
-import java.nio.ByteBuffer;
-
-/**
- * An output target for raw bytes. This interface provides semantics that support two types of
- * writing:
- *
- * <p><b>Traditional write operations:</b>
- * (as defined by {@link java.io.OutputStream}) where the target method is responsible for either
- * copying the data or completing the write before returning from the method call.
- *
- * <p><b>Lazy write operations:</b> where the caller guarantees that it will never modify the
- * provided buffer and it can therefore be considered immutable. The target method is free to
- * maintain a reference to the buffer beyond the scope of the method call (e.g. until the write
- * operation completes).
- */
-@ExperimentalApi
-public abstract class ByteOutput {
- /**
- * Writes a single byte.
- *
- * @param value the byte to be written
- * @throws IOException thrown if an error occurred while writing
- */
- public abstract void write(byte value) throws IOException;
-
- /**
- * Writes a sequence of bytes. The {@link ByteOutput} must copy {@code value} if it will
- * not be processed prior to the return of this method call, since {@code value} may be
- * reused/altered by the caller.
- *
- * <p>NOTE: This method <strong>MUST NOT</strong> modify the {@code value}. Doing so is a
- * programming error and will lead to data corruption which will be difficult to debug.
- *
- * @param value the bytes to be written
- * @param offset the offset of the start of the writable range
- * @param length the number of bytes to write starting from {@code offset}
- * @throws IOException thrown if an error occurred while writing
- */
- public abstract void write(byte[] value, int offset, int length) throws IOException;
-
- /**
- * Writes a sequence of bytes. The {@link ByteOutput} is free to retain a reference to the value
- * beyond the scope of this method call (e.g. write later) since it is considered immutable and is
- * guaranteed not to change by the caller.
- *
- * <p>NOTE: This method <strong>MUST NOT</strong> modify the {@code value}. Doing so is a
- * programming error and will lead to data corruption which will be difficult to debug.
- *
- * @param value the bytes to be written
- * @param offset the offset of the start of the writable range
- * @param length the number of bytes to write starting from {@code offset}
- * @throws IOException thrown if an error occurred while writing
- */
- public abstract void writeLazy(byte[] value, int offset, int length) throws IOException;
-
- /**
- * Writes a sequence of bytes. The {@link ByteOutput} must copy {@code value} if it will
- * not be processed prior to the return of this method call, since {@code value} may be
- * reused/altered by the caller.
- *
- * <p>NOTE: This method <strong>MUST NOT</strong> modify the {@code value}. Doing so is a
- * programming error and will lead to data corruption which will be difficult to debug.
- *
- * @param value the bytes to be written. Upon returning from this call, the {@code position} of
- * this buffer will be set to the {@code limit}
- * @throws IOException thrown if an error occurred while writing
- */
- public abstract void write(ByteBuffer value) throws IOException;
-
- /**
- * Writes a sequence of bytes. The {@link ByteOutput} is free to retain a reference to the value
- * beyond the scope of this method call (e.g. write later) since it is considered immutable and is
- * guaranteed not to change by the caller.
- *
- * <p>NOTE: This method <strong>MUST NOT</strong> modify the {@code value}. Doing so is a
- * programming error and will lead to data corruption which will be difficult to debug.
- *
- * @param value the bytes to be written. Upon returning from this call, the {@code position} of
- * this buffer will be set to the {@code limit}
- * @throws IOException thrown if an error occurred while writing
- */
- public abstract void writeLazy(ByteBuffer value) throws IOException;
-}