You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hbase.apache.org by st...@apache.org on 2016/07/21 17:20:25 UTC
[02/39] hbase git commit: HBASE-16263 Move all to do w/ protobuf --
*.proto files and generated classes -- under hbase-protocol
http://git-wip-us.apache.org/repos/asf/hbase/blob/9d740f7b/hbase-server/src/test/java/org/apache/hadoop/hbase/ipc/protobuf/generated/TestRpcServiceProtos.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/ipc/protobuf/generated/TestRpcServiceProtos.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/ipc/protobuf/generated/TestRpcServiceProtos.java
deleted file mode 100644
index 3fd34e9..0000000
--- a/hbase-server/src/test/java/org/apache/hadoop/hbase/ipc/protobuf/generated/TestRpcServiceProtos.java
+++ /dev/null
@@ -1,424 +0,0 @@
-// Generated by the protocol buffer compiler. DO NOT EDIT!
-// source: test_rpc_service.proto
-
-package org.apache.hadoop.hbase.ipc.protobuf.generated;
-
-public final class TestRpcServiceProtos {
- private TestRpcServiceProtos() {}
- public static void registerAllExtensions(
- com.google.protobuf.ExtensionRegistry registry) {
- }
- /**
- * Protobuf service {@code TestProtobufRpcProto}
- *
- * <pre>
- **
- * A protobuf service for use in tests
- * </pre>
- */
- public static abstract class TestProtobufRpcProto
- implements com.google.protobuf.Service {
- protected TestProtobufRpcProto() {}
-
- public interface Interface {
- /**
- * <code>rpc ping(.EmptyRequestProto) returns (.EmptyResponseProto);</code>
- */
- public abstract void ping(
- com.google.protobuf.RpcController controller,
- org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EmptyRequestProto request,
- com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EmptyResponseProto> done);
-
- /**
- * <code>rpc echo(.EchoRequestProto) returns (.EchoResponseProto);</code>
- */
- public abstract void echo(
- com.google.protobuf.RpcController controller,
- org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EchoRequestProto request,
- com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EchoResponseProto> done);
-
- /**
- * <code>rpc error(.EmptyRequestProto) returns (.EmptyResponseProto);</code>
- */
- public abstract void error(
- com.google.protobuf.RpcController controller,
- org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EmptyRequestProto request,
- com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EmptyResponseProto> done);
-
- }
-
- public static com.google.protobuf.Service newReflectiveService(
- final Interface impl) {
- return new TestProtobufRpcProto() {
- @java.lang.Override
- public void ping(
- com.google.protobuf.RpcController controller,
- org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EmptyRequestProto request,
- com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EmptyResponseProto> done) {
- impl.ping(controller, request, done);
- }
-
- @java.lang.Override
- public void echo(
- com.google.protobuf.RpcController controller,
- org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EchoRequestProto request,
- com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EchoResponseProto> done) {
- impl.echo(controller, request, done);
- }
-
- @java.lang.Override
- public void error(
- com.google.protobuf.RpcController controller,
- org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EmptyRequestProto request,
- com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EmptyResponseProto> done) {
- impl.error(controller, request, done);
- }
-
- };
- }
-
- public static com.google.protobuf.BlockingService
- newReflectiveBlockingService(final BlockingInterface impl) {
- return new com.google.protobuf.BlockingService() {
- public final com.google.protobuf.Descriptors.ServiceDescriptor
- getDescriptorForType() {
- return getDescriptor();
- }
-
- public final com.google.protobuf.Message callBlockingMethod(
- com.google.protobuf.Descriptors.MethodDescriptor method,
- com.google.protobuf.RpcController controller,
- com.google.protobuf.Message request)
- throws com.google.protobuf.ServiceException {
- if (method.getService() != getDescriptor()) {
- throw new java.lang.IllegalArgumentException(
- "Service.callBlockingMethod() given method descriptor for " +
- "wrong service type.");
- }
- switch(method.getIndex()) {
- case 0:
- return impl.ping(controller, (org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EmptyRequestProto)request);
- case 1:
- return impl.echo(controller, (org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EchoRequestProto)request);
- case 2:
- return impl.error(controller, (org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EmptyRequestProto)request);
- default:
- throw new java.lang.AssertionError("Can't get here.");
- }
- }
-
- public final com.google.protobuf.Message
- getRequestPrototype(
- com.google.protobuf.Descriptors.MethodDescriptor method) {
- if (method.getService() != getDescriptor()) {
- throw new java.lang.IllegalArgumentException(
- "Service.getRequestPrototype() given method " +
- "descriptor for wrong service type.");
- }
- switch(method.getIndex()) {
- case 0:
- return org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EmptyRequestProto.getDefaultInstance();
- case 1:
- return org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EchoRequestProto.getDefaultInstance();
- case 2:
- return org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EmptyRequestProto.getDefaultInstance();
- default:
- throw new java.lang.AssertionError("Can't get here.");
- }
- }
-
- public final com.google.protobuf.Message
- getResponsePrototype(
- com.google.protobuf.Descriptors.MethodDescriptor method) {
- if (method.getService() != getDescriptor()) {
- throw new java.lang.IllegalArgumentException(
- "Service.getResponsePrototype() given method " +
- "descriptor for wrong service type.");
- }
- switch(method.getIndex()) {
- case 0:
- return org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EmptyResponseProto.getDefaultInstance();
- case 1:
- return org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EchoResponseProto.getDefaultInstance();
- case 2:
- return org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EmptyResponseProto.getDefaultInstance();
- default:
- throw new java.lang.AssertionError("Can't get here.");
- }
- }
-
- };
- }
-
- /**
- * <code>rpc ping(.EmptyRequestProto) returns (.EmptyResponseProto);</code>
- */
- public abstract void ping(
- com.google.protobuf.RpcController controller,
- org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EmptyRequestProto request,
- com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EmptyResponseProto> done);
-
- /**
- * <code>rpc echo(.EchoRequestProto) returns (.EchoResponseProto);</code>
- */
- public abstract void echo(
- com.google.protobuf.RpcController controller,
- org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EchoRequestProto request,
- com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EchoResponseProto> done);
-
- /**
- * <code>rpc error(.EmptyRequestProto) returns (.EmptyResponseProto);</code>
- */
- public abstract void error(
- com.google.protobuf.RpcController controller,
- org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EmptyRequestProto request,
- com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EmptyResponseProto> done);
-
- public static final
- com.google.protobuf.Descriptors.ServiceDescriptor
- getDescriptor() {
- return org.apache.hadoop.hbase.ipc.protobuf.generated.TestRpcServiceProtos.getDescriptor().getServices().get(0);
- }
- public final com.google.protobuf.Descriptors.ServiceDescriptor
- getDescriptorForType() {
- return getDescriptor();
- }
-
- public final void callMethod(
- com.google.protobuf.Descriptors.MethodDescriptor method,
- com.google.protobuf.RpcController controller,
- com.google.protobuf.Message request,
- com.google.protobuf.RpcCallback<
- com.google.protobuf.Message> done) {
- if (method.getService() != getDescriptor()) {
- throw new java.lang.IllegalArgumentException(
- "Service.callMethod() given method descriptor for wrong " +
- "service type.");
- }
- switch(method.getIndex()) {
- case 0:
- this.ping(controller, (org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EmptyRequestProto)request,
- com.google.protobuf.RpcUtil.<org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EmptyResponseProto>specializeCallback(
- done));
- return;
- case 1:
- this.echo(controller, (org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EchoRequestProto)request,
- com.google.protobuf.RpcUtil.<org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EchoResponseProto>specializeCallback(
- done));
- return;
- case 2:
- this.error(controller, (org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EmptyRequestProto)request,
- com.google.protobuf.RpcUtil.<org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EmptyResponseProto>specializeCallback(
- done));
- return;
- default:
- throw new java.lang.AssertionError("Can't get here.");
- }
- }
-
- public final com.google.protobuf.Message
- getRequestPrototype(
- com.google.protobuf.Descriptors.MethodDescriptor method) {
- if (method.getService() != getDescriptor()) {
- throw new java.lang.IllegalArgumentException(
- "Service.getRequestPrototype() given method " +
- "descriptor for wrong service type.");
- }
- switch(method.getIndex()) {
- case 0:
- return org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EmptyRequestProto.getDefaultInstance();
- case 1:
- return org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EchoRequestProto.getDefaultInstance();
- case 2:
- return org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EmptyRequestProto.getDefaultInstance();
- default:
- throw new java.lang.AssertionError("Can't get here.");
- }
- }
-
- public final com.google.protobuf.Message
- getResponsePrototype(
- com.google.protobuf.Descriptors.MethodDescriptor method) {
- if (method.getService() != getDescriptor()) {
- throw new java.lang.IllegalArgumentException(
- "Service.getResponsePrototype() given method " +
- "descriptor for wrong service type.");
- }
- switch(method.getIndex()) {
- case 0:
- return org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EmptyResponseProto.getDefaultInstance();
- case 1:
- return org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EchoResponseProto.getDefaultInstance();
- case 2:
- return org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EmptyResponseProto.getDefaultInstance();
- default:
- throw new java.lang.AssertionError("Can't get here.");
- }
- }
-
- public static Stub newStub(
- com.google.protobuf.RpcChannel channel) {
- return new Stub(channel);
- }
-
- public static final class Stub extends org.apache.hadoop.hbase.ipc.protobuf.generated.TestRpcServiceProtos.TestProtobufRpcProto implements Interface {
- private Stub(com.google.protobuf.RpcChannel channel) {
- this.channel = channel;
- }
-
- private final com.google.protobuf.RpcChannel channel;
-
- public com.google.protobuf.RpcChannel getChannel() {
- return channel;
- }
-
- public void ping(
- com.google.protobuf.RpcController controller,
- org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EmptyRequestProto request,
- com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EmptyResponseProto> done) {
- channel.callMethod(
- getDescriptor().getMethods().get(0),
- controller,
- request,
- org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EmptyResponseProto.getDefaultInstance(),
- com.google.protobuf.RpcUtil.generalizeCallback(
- done,
- org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EmptyResponseProto.class,
- org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EmptyResponseProto.getDefaultInstance()));
- }
-
- public void echo(
- com.google.protobuf.RpcController controller,
- org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EchoRequestProto request,
- com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EchoResponseProto> done) {
- channel.callMethod(
- getDescriptor().getMethods().get(1),
- controller,
- request,
- org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EchoResponseProto.getDefaultInstance(),
- com.google.protobuf.RpcUtil.generalizeCallback(
- done,
- org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EchoResponseProto.class,
- org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EchoResponseProto.getDefaultInstance()));
- }
-
- public void error(
- com.google.protobuf.RpcController controller,
- org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EmptyRequestProto request,
- com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EmptyResponseProto> done) {
- channel.callMethod(
- getDescriptor().getMethods().get(2),
- controller,
- request,
- org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EmptyResponseProto.getDefaultInstance(),
- com.google.protobuf.RpcUtil.generalizeCallback(
- done,
- org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EmptyResponseProto.class,
- org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EmptyResponseProto.getDefaultInstance()));
- }
- }
-
- public static BlockingInterface newBlockingStub(
- com.google.protobuf.BlockingRpcChannel channel) {
- return new BlockingStub(channel);
- }
-
- public interface BlockingInterface {
- public org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EmptyResponseProto ping(
- com.google.protobuf.RpcController controller,
- org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EmptyRequestProto request)
- throws com.google.protobuf.ServiceException;
-
- public org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EchoResponseProto echo(
- com.google.protobuf.RpcController controller,
- org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EchoRequestProto request)
- throws com.google.protobuf.ServiceException;
-
- public org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EmptyResponseProto error(
- com.google.protobuf.RpcController controller,
- org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EmptyRequestProto request)
- throws com.google.protobuf.ServiceException;
- }
-
- private static final class BlockingStub implements BlockingInterface {
- private BlockingStub(com.google.protobuf.BlockingRpcChannel channel) {
- this.channel = channel;
- }
-
- private final com.google.protobuf.BlockingRpcChannel channel;
-
- public org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EmptyResponseProto ping(
- com.google.protobuf.RpcController controller,
- org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EmptyRequestProto request)
- throws com.google.protobuf.ServiceException {
- return (org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EmptyResponseProto) channel.callBlockingMethod(
- getDescriptor().getMethods().get(0),
- controller,
- request,
- org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EmptyResponseProto.getDefaultInstance());
- }
-
-
- public org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EchoResponseProto echo(
- com.google.protobuf.RpcController controller,
- org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EchoRequestProto request)
- throws com.google.protobuf.ServiceException {
- return (org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EchoResponseProto) channel.callBlockingMethod(
- getDescriptor().getMethods().get(1),
- controller,
- request,
- org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EchoResponseProto.getDefaultInstance());
- }
-
-
- public org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EmptyResponseProto error(
- com.google.protobuf.RpcController controller,
- org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EmptyRequestProto request)
- throws com.google.protobuf.ServiceException {
- return (org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EmptyResponseProto) channel.callBlockingMethod(
- getDescriptor().getMethods().get(2),
- controller,
- request,
- org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EmptyResponseProto.getDefaultInstance());
- }
-
- }
-
- // @@protoc_insertion_point(class_scope:TestProtobufRpcProto)
- }
-
-
- public static com.google.protobuf.Descriptors.FileDescriptor
- getDescriptor() {
- return descriptor;
- }
- private static com.google.protobuf.Descriptors.FileDescriptor
- descriptor;
- static {
- java.lang.String[] descriptorData = {
- "\n\026test_rpc_service.proto\032\ntest.proto2\250\001\n" +
- "\024TestProtobufRpcProto\022/\n\004ping\022\022.EmptyReq" +
- "uestProto\032\023.EmptyResponseProto\022-\n\004echo\022\021" +
- ".EchoRequestProto\032\022.EchoResponseProto\0220\n" +
- "\005error\022\022.EmptyRequestProto\032\023.EmptyRespon" +
- "seProtoBL\n.org.apache.hadoop.hbase.ipc.p" +
- "rotobuf.generatedB\024TestRpcServiceProtos\210" +
- "\001\001\240\001\001"
- };
- com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
- new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() {
- public com.google.protobuf.ExtensionRegistry assignDescriptors(
- com.google.protobuf.Descriptors.FileDescriptor root) {
- descriptor = root;
- return null;
- }
- };
- com.google.protobuf.Descriptors.FileDescriptor
- .internalBuildGeneratedFileFrom(descriptorData,
- new com.google.protobuf.Descriptors.FileDescriptor[] {
- org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.getDescriptor(),
- }, assigner);
- }
-
- // @@protoc_insertion_point(outer_class_scope)
-}
http://git-wip-us.apache.org/repos/asf/hbase/blob/9d740f7b/hbase-server/src/test/protobuf/ColumnAggregationNullResponseProtocol.proto
----------------------------------------------------------------------
diff --git a/hbase-server/src/test/protobuf/ColumnAggregationNullResponseProtocol.proto b/hbase-server/src/test/protobuf/ColumnAggregationNullResponseProtocol.proto
deleted file mode 100644
index b4dc01e..0000000
--- a/hbase-server/src/test/protobuf/ColumnAggregationNullResponseProtocol.proto
+++ /dev/null
@@ -1,38 +0,0 @@
-/**
- * Licensed to the Apache Software Foundation (ASF) under one
- * or more contributor license agreements. See the NOTICE file
- * distributed with this work for additional information
- * regarding copyright ownership. The ASF licenses this file
- * to you under the Apache License, Version 2.0 (the
- * "License"); you may not use this file except in compliance
- * with the License. You may obtain a copy of the License at
- *
- * http://www.apache.org/licenses/LICENSE-2.0
- *
- * Unless required by applicable law or agreed to in writing, software
- * distributed under the License is distributed on an "AS IS" BASIS,
- * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- * See the License for the specific language governing permissions and
- * limitations under the License.
- */
-// Coprocessor test
-option java_package = "org.apache.hadoop.hbase.coprocessor.protobuf.generated";
-option java_outer_classname = "ColumnAggregationWithNullResponseProtos";
-option java_generic_services = true;
-option java_generate_equals_and_hash = true;
-
-// use unique names for messages in ColumnAggregationXXX.protos due to a bug in
-// protoc or hadoop's protoc compiler.
-message ColumnAggregationNullResponseSumRequest {
- required bytes family = 1;
- optional bytes qualifier = 2;
-}
-
-message ColumnAggregationNullResponseSumResponse {
- optional int64 sum = 1;
-}
-
-service ColumnAggregationServiceNullResponse {
- rpc sum(ColumnAggregationNullResponseSumRequest)
- returns(ColumnAggregationNullResponseSumResponse);
-}
http://git-wip-us.apache.org/repos/asf/hbase/blob/9d740f7b/hbase-server/src/test/protobuf/ColumnAggregationProtocol.proto
----------------------------------------------------------------------
diff --git a/hbase-server/src/test/protobuf/ColumnAggregationProtocol.proto b/hbase-server/src/test/protobuf/ColumnAggregationProtocol.proto
deleted file mode 100644
index ad1acda..0000000
--- a/hbase-server/src/test/protobuf/ColumnAggregationProtocol.proto
+++ /dev/null
@@ -1,35 +0,0 @@
-/**
- * Licensed to the Apache Software Foundation (ASF) under one
- * or more contributor license agreements. See the NOTICE file
- * distributed with this work for additional information
- * regarding copyright ownership. The ASF licenses this file
- * to you under the Apache License, Version 2.0 (the
- * "License"); you may not use this file except in compliance
- * with the License. You may obtain a copy of the License at
- *
- * http://www.apache.org/licenses/LICENSE-2.0
- *
- * Unless required by applicable law or agreed to in writing, software
- * distributed under the License is distributed on an "AS IS" BASIS,
- * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- * See the License for the specific language governing permissions and
- * limitations under the License.
- */
-// Coprocessor test
-option java_package = "org.apache.hadoop.hbase.coprocessor.protobuf.generated";
-option java_outer_classname = "ColumnAggregationProtos";
-option java_generic_services = true;
-option java_generate_equals_and_hash = true;
-
-message SumRequest {
- required bytes family = 1;
- optional bytes qualifier = 2;
-}
-
-message SumResponse {
- required int64 sum = 1;
-}
-
-service ColumnAggregationService {
- rpc sum(SumRequest) returns(SumResponse);
-}
http://git-wip-us.apache.org/repos/asf/hbase/blob/9d740f7b/hbase-server/src/test/protobuf/ColumnAggregationWithErrorsProtocol.proto
----------------------------------------------------------------------
diff --git a/hbase-server/src/test/protobuf/ColumnAggregationWithErrorsProtocol.proto b/hbase-server/src/test/protobuf/ColumnAggregationWithErrorsProtocol.proto
deleted file mode 100644
index 7808949..0000000
--- a/hbase-server/src/test/protobuf/ColumnAggregationWithErrorsProtocol.proto
+++ /dev/null
@@ -1,38 +0,0 @@
-/**
- * Licensed to the Apache Software Foundation (ASF) under one
- * or more contributor license agreements. See the NOTICE file
- * distributed with this work for additional information
- * regarding copyright ownership. The ASF licenses this file
- * to you under the Apache License, Version 2.0 (the
- * "License"); you may not use this file except in compliance
- * with the License. You may obtain a copy of the License at
- *
- * http://www.apache.org/licenses/LICENSE-2.0
- *
- * Unless required by applicable law or agreed to in writing, software
- * distributed under the License is distributed on an "AS IS" BASIS,
- * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- * See the License for the specific language governing permissions and
- * limitations under the License.
- */
-// Coprocessor test
-option java_package = "org.apache.hadoop.hbase.coprocessor.protobuf.generated";
-option java_outer_classname = "ColumnAggregationWithErrorsProtos";
-option java_generic_services = true;
-option java_generate_equals_and_hash = true;
-
-// use unique names for messages in ColumnAggregationXXX.protos due to a bug in
-// protoc or hadoop's protoc compiler.
-message ColumnAggregationWithErrorsSumRequest {
- required bytes family = 1;
- optional bytes qualifier = 2;
-}
-
-message ColumnAggregationWithErrorsSumResponse {
- required int64 sum = 1;
-}
-
-service ColumnAggregationServiceWithErrors {
- rpc sum(ColumnAggregationWithErrorsSumRequest)
- returns(ColumnAggregationWithErrorsSumResponse);
-}
http://git-wip-us.apache.org/repos/asf/hbase/blob/9d740f7b/hbase-server/src/test/protobuf/DummyRegionServerEndpoint.proto
----------------------------------------------------------------------
diff --git a/hbase-server/src/test/protobuf/DummyRegionServerEndpoint.proto b/hbase-server/src/test/protobuf/DummyRegionServerEndpoint.proto
deleted file mode 100644
index 539f7da..0000000
--- a/hbase-server/src/test/protobuf/DummyRegionServerEndpoint.proto
+++ /dev/null
@@ -1,37 +0,0 @@
-/**
- * Licensed to the Apache Software Foundation (ASF) under one
- * or more contributor license agreements. See the NOTICE file
- * distributed with this work for additional information
- * regarding copyright ownership. The ASF licenses this file
- * to you under the Apache License, Version 2.0 (the
- * "License"); you may not use this file except in compliance
- * with the License. You may obtain a copy of the License at
- *
- * http://www.apache.org/licenses/LICENSE-2.0
- *
- * Unless required by applicable law or agreed to in writing, software
- * distributed under the License is distributed on an "AS IS" BASIS,
- * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- * See the License for the specific language governing permissions and
- * limitations under the License.
- */
-
-package hbase.test.pb;
-
-// Coprocessor test
-option java_package = "org.apache.hadoop.hbase.coprocessor.protobuf.generated";
-option java_outer_classname = "DummyRegionServerEndpointProtos";
-option java_generic_services = true;
-option java_generate_equals_and_hash = true;
-
-message DummyRequest {
-}
-
-message DummyResponse {
- required string value = 1;
-}
-
-service DummyService {
- rpc dummyCall(DummyRequest) returns(DummyResponse);
- rpc dummyThrow(DummyRequest) returns(DummyResponse);
-}
http://git-wip-us.apache.org/repos/asf/hbase/blob/9d740f7b/hbase-server/src/test/protobuf/IncrementCounterProcessor.proto
----------------------------------------------------------------------
diff --git a/hbase-server/src/test/protobuf/IncrementCounterProcessor.proto b/hbase-server/src/test/protobuf/IncrementCounterProcessor.proto
deleted file mode 100644
index b8c77ca..0000000
--- a/hbase-server/src/test/protobuf/IncrementCounterProcessor.proto
+++ /dev/null
@@ -1,55 +0,0 @@
-/**
- * Licensed to the Apache Software Foundation (ASF) under one
- * or more contributor license agreements. See the NOTICE file
- * distributed with this work for additional information
- * regarding copyright ownership. The ASF licenses this file
- * to you under the Apache License, Version 2.0 (the
- * "License"); you may not use this file except in compliance
- * with the License. You may obtain a copy of the License at
- *
- * http://www.apache.org/licenses/LICENSE-2.0
- *
- * Unless required by applicable law or agreed to in writing, software
- * distributed under the License is distributed on an "AS IS" BASIS,
- * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- * See the License for the specific language governing permissions and
- * limitations under the License.
- */
-
-option java_package = "org.apache.hadoop.hbase.coprocessor.protobuf.generated";
-option java_outer_classname = "IncrementCounterProcessorTestProtos";
-option java_generate_equals_and_hash = true;
-
-message IncCounterProcessorRequest {
- required bytes row = 1;
- required int32 counter = 2;
-}
-
-message IncCounterProcessorResponse {
- required int32 response = 1;
-}
-
-message FriendsOfFriendsProcessorRequest {
- required bytes person = 1;
- required bytes row = 2;
- repeated string result = 3;
-}
-
-message FriendsOfFriendsProcessorResponse {
- repeated string result = 1;
-}
-
-message RowSwapProcessorRequest {
- required bytes row1 = 1;
- required bytes row2 = 2;
-}
-
-message RowSwapProcessorResponse {
-}
-
-message TimeoutProcessorRequest {
- required bytes row = 1;
-}
-
-message TimeoutProcessorResponse {
-}
\ No newline at end of file
http://git-wip-us.apache.org/repos/asf/hbase/blob/9d740f7b/hbase-server/src/test/protobuf/PingProtocol.proto
----------------------------------------------------------------------
diff --git a/hbase-server/src/test/protobuf/PingProtocol.proto b/hbase-server/src/test/protobuf/PingProtocol.proto
deleted file mode 100644
index ef63ee0..0000000
--- a/hbase-server/src/test/protobuf/PingProtocol.proto
+++ /dev/null
@@ -1,66 +0,0 @@
-/**
- * Licensed to the Apache Software Foundation (ASF) under one
- * or more contributor license agreements. See the NOTICE file
- * distributed with this work for additional information
- * regarding copyright ownership. The ASF licenses this file
- * to you under the Apache License, Version 2.0 (the
- * "License"); you may not use this file except in compliance
- * with the License. You may obtain a copy of the License at
- *
- * http://www.apache.org/licenses/LICENSE-2.0
- *
- * Unless required by applicable law or agreed to in writing, software
- * distributed under the License is distributed on an "AS IS" BASIS,
- * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- * See the License for the specific language governing permissions and
- * limitations under the License.
- */
-// Coprocessor test
-option java_package = "org.apache.hadoop.hbase.coprocessor.protobuf.generated";
-option java_outer_classname = "PingProtos";
-option java_generic_services = true;
-option java_generate_equals_and_hash = true;
-
-message PingRequest {
-}
-
-message PingResponse {
- required string pong = 1;
-}
-
-message CountRequest {
-}
-
-message CountResponse {
- required int32 count = 1;
-}
-
-message IncrementCountRequest {
- required int32 diff = 1;
-}
-
-message IncrementCountResponse {
- required int32 count = 1;
-}
-
-message HelloRequest {
- optional string name = 1;
-}
-
-message HelloResponse {
- optional string response = 1;
-}
-
-message NoopRequest {
-}
-
-message NoopResponse {
-}
-
-service PingService {
- rpc ping(PingRequest) returns(PingResponse);
- rpc count(CountRequest) returns(CountResponse);
- rpc increment(IncrementCountRequest) returns(IncrementCountResponse);
- rpc hello(HelloRequest) returns(HelloResponse);
- rpc noop(NoopRequest) returns(NoopResponse);
-}
http://git-wip-us.apache.org/repos/asf/hbase/blob/9d740f7b/hbase-server/src/test/protobuf/TestProcedure.proto
----------------------------------------------------------------------
diff --git a/hbase-server/src/test/protobuf/TestProcedure.proto b/hbase-server/src/test/protobuf/TestProcedure.proto
deleted file mode 100644
index de74f36..0000000
--- a/hbase-server/src/test/protobuf/TestProcedure.proto
+++ /dev/null
@@ -1,24 +0,0 @@
-/**
- * Licensed to the Apache Software Foundation (ASF) under one
- * or more contributor license agreements. See the NOTICE file
- * distributed with this work for additional information
- * regarding copyright ownership. The ASF licenses this file
- * to you under the Apache License, Version 2.0 (the
- * "License"); you may not use this file except in compliance
- * with the License. You may obtain a copy of the License at
- *
- * http://www.apache.org/licenses/LICENSE-2.0
- *
- * Unless required by applicable law or agreed to in writing, software
- * distributed under the License is distributed on an "AS IS" BASIS,
- * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- * See the License for the specific language governing permissions and
- * limitations under the License.
- */
-option java_package = "org.apache.hadoop.hbase.ipc.protobuf.generated";
-option java_outer_classname = "TestProcedureProtos";
-option java_generic_services = true;
-
-message TestTableDDLStateData {
- required string table_name = 1;
-}
http://git-wip-us.apache.org/repos/asf/hbase/blob/9d740f7b/hbase-server/src/test/protobuf/test.proto
----------------------------------------------------------------------
diff --git a/hbase-server/src/test/protobuf/test.proto b/hbase-server/src/test/protobuf/test.proto
deleted file mode 100644
index 566b04b..0000000
--- a/hbase-server/src/test/protobuf/test.proto
+++ /dev/null
@@ -1,35 +0,0 @@
-/**
- * Licensed to the Apache Software Foundation (ASF) under one
- * or more contributor license agreements. See the NOTICE file
- * distributed with this work for additional information
- * regarding copyright ownership. The ASF licenses this file
- * to you under the Apache License, Version 2.0 (the
- * "License"); you may not use this file except in compliance
- * with the License. You may obtain a copy of the License at
- *
- * http://www.apache.org/licenses/LICENSE-2.0
- *
- * Unless required by applicable law or agreed to in writing, software
- * distributed under the License is distributed on an "AS IS" BASIS,
- * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- * See the License for the specific language governing permissions and
- * limitations under the License.
- */
-
-option java_package = "org.apache.hadoop.hbase.ipc.protobuf.generated";
-option java_outer_classname = "TestProtos";
-option java_generate_equals_and_hash = true;
-
-message EmptyRequestProto {
-}
-
-message EmptyResponseProto {
-}
-
-message EchoRequestProto {
- required string message = 1;
-}
-
-message EchoResponseProto {
- required string message = 1;
-}
http://git-wip-us.apache.org/repos/asf/hbase/blob/9d740f7b/hbase-server/src/test/protobuf/test_rpc_service.proto
----------------------------------------------------------------------
diff --git a/hbase-server/src/test/protobuf/test_rpc_service.proto b/hbase-server/src/test/protobuf/test_rpc_service.proto
deleted file mode 100644
index 4ed0380..0000000
--- a/hbase-server/src/test/protobuf/test_rpc_service.proto
+++ /dev/null
@@ -1,33 +0,0 @@
-/**
- * Licensed to the Apache Software Foundation (ASF) under one
- * or more contributor license agreements. See the NOTICE file
- * distributed with this work for additional information
- * regarding copyright ownership. The ASF licenses this file
- * to you under the Apache License, Version 2.0 (the
- * "License"); you may not use this file except in compliance
- * with the License. You may obtain a copy of the License at
- *
- * http://www.apache.org/licenses/LICENSE-2.0
- *
- * Unless required by applicable law or agreed to in writing, software
- * distributed under the License is distributed on an "AS IS" BASIS,
- * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- * See the License for the specific language governing permissions and
- * limitations under the License.
- */
-option java_package = "org.apache.hadoop.hbase.ipc.protobuf.generated";
-option java_outer_classname = "TestRpcServiceProtos";
-option java_generic_services = true;
-option java_generate_equals_and_hash = true;
-
-import "test.proto";
-
-
-/**
- * A protobuf service for use in tests
- */
-service TestProtobufRpcProto {
- rpc ping(EmptyRequestProto) returns (EmptyResponseProto);
- rpc echo(EchoRequestProto) returns (EchoResponseProto);
- rpc error(EmptyRequestProto) returns (EmptyResponseProto);
-}
http://git-wip-us.apache.org/repos/asf/hbase/blob/9d740f7b/hbase-spark/pom.xml
----------------------------------------------------------------------
diff --git a/hbase-spark/pom.xml b/hbase-spark/pom.xml
index 776eadb..89ed186 100644
--- a/hbase-spark/pom.xml
+++ b/hbase-spark/pom.xml
@@ -663,43 +663,5 @@
<surefire.skipSecondPart>true</surefire.skipSecondPart>
</properties>
</profile>
- <profile>
- <id>compile-protobuf</id>
- <activation>
- <property>
- <name>compile-protobuf</name>
- </property>
- </activation>
- <build>
- <plugins>
- <plugin>
- <groupId>org.apache.hadoop</groupId>
- <artifactId>hadoop-maven-plugins</artifactId>
- <executions>
- <execution>
- <id>compile-protoc</id>
- <phase>generate-sources</phase>
- <goals>
- <goal>protoc</goal>
- </goals>
- <configuration>
- <imports>
- <param>${basedir}/src/main/protobuf</param>
- </imports>
- <source>
- <directory>${basedir}/src/main/protobuf</directory>
- <includes>
- <include>Filter.proto</include>
- </includes>
- </source>
- <!--<output>${project.build.directory}/generated-sources/java</output>-->
- <output>${basedir}/src/main/java/</output>
- </configuration>
- </execution>
- </executions>
- </plugin>
- </plugins>
- </build>
- </profile>
</profiles>
</project>
http://git-wip-us.apache.org/repos/asf/hbase/blob/9d740f7b/hbase-spark/src/main/java/org/apache/hadoop/hbase/spark/SparkSQLPushDownFilter.java
----------------------------------------------------------------------
diff --git a/hbase-spark/src/main/java/org/apache/hadoop/hbase/spark/SparkSQLPushDownFilter.java b/hbase-spark/src/main/java/org/apache/hadoop/hbase/spark/SparkSQLPushDownFilter.java
index 071c1ca..fa66d69 100644
--- a/hbase-spark/src/main/java/org/apache/hadoop/hbase/spark/SparkSQLPushDownFilter.java
+++ b/hbase-spark/src/main/java/org/apache/hadoop/hbase/spark/SparkSQLPushDownFilter.java
@@ -26,7 +26,7 @@ import org.apache.hadoop.hbase.exceptions.DeserializationException;
import org.apache.hadoop.hbase.filter.FilterBase;
import org.apache.hadoop.hbase.spark.datasources.BytesEncoder;
import org.apache.hadoop.hbase.spark.datasources.JavaBytesEncoder;
-import org.apache.hadoop.hbase.spark.protobuf.generated.FilterProtos;
+import org.apache.hadoop.hbase.spark.protobuf.generated.SparkFilterProtos;
import org.apache.hadoop.hbase.util.ByteStringer;
import org.apache.hadoop.hbase.util.Bytes;
import org.apache.spark.sql.datasources.hbase.Field;
@@ -183,9 +183,9 @@ public class SparkSQLPushDownFilter extends FilterBase{
public static SparkSQLPushDownFilter parseFrom(final byte[] pbBytes)
throws DeserializationException {
- FilterProtos.SQLPredicatePushDownFilter proto;
+ SparkFilterProtos.SQLPredicatePushDownFilter proto;
try {
- proto = FilterProtos.SQLPredicatePushDownFilter.parseFrom(pbBytes);
+ proto = SparkFilterProtos.SQLPredicatePushDownFilter.parseFrom(pbBytes);
} catch (InvalidProtocolBufferException e) {
throw new DeserializationException(e);
}
@@ -208,7 +208,7 @@ public class SparkSQLPushDownFilter extends FilterBase{
HashMap<ByteArrayComparable, HashMap<ByteArrayComparable, String>>
currentCellToColumnIndexMap = new HashMap<>();
- for (FilterProtos.SQLPredicatePushDownCellToColumnMapping
+ for (SparkFilterProtos.SQLPredicatePushDownCellToColumnMapping
sqlPredicatePushDownCellToColumnMapping :
proto.getCellToColumnMappingList()) {
@@ -242,11 +242,11 @@ public class SparkSQLPushDownFilter extends FilterBase{
*/
public byte[] toByteArray() {
- FilterProtos.SQLPredicatePushDownFilter.Builder builder =
- FilterProtos.SQLPredicatePushDownFilter.newBuilder();
+ SparkFilterProtos.SQLPredicatePushDownFilter.Builder builder =
+ SparkFilterProtos.SQLPredicatePushDownFilter.newBuilder();
- FilterProtos.SQLPredicatePushDownCellToColumnMapping.Builder columnMappingBuilder =
- FilterProtos.SQLPredicatePushDownCellToColumnMapping.newBuilder();
+ SparkFilterProtos.SQLPredicatePushDownCellToColumnMapping.Builder columnMappingBuilder =
+ SparkFilterProtos.SQLPredicatePushDownCellToColumnMapping.newBuilder();
builder.setDynamicLogicExpression(dynamicLogicExpression.toExpressionString());
for (byte[] valueFromQuery: valueFromQueryArray) {