You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hbase.apache.org by st...@apache.org on 2017/08/02 00:05:36 UTC
[1/5] hbase git commit: HBASE-18431 Mitigate compatibility concerns
between branch-1.3 and branch-1.4
Repository: hbase
Updated Branches:
refs/heads/branch-2 f67dae90f -> 3f540f380
http://git-wip-us.apache.org/repos/asf/hbase/blob/5e21e306/hbase-rsgroup/src/main/java/org/apache/hadoop/hbase/protobuf/generated/RSGroupAdminProtos.java
----------------------------------------------------------------------
diff --git a/hbase-rsgroup/src/main/java/org/apache/hadoop/hbase/protobuf/generated/RSGroupAdminProtos.java b/hbase-rsgroup/src/main/java/org/apache/hadoop/hbase/protobuf/generated/RSGroupAdminProtos.java
index ca1db1e..3d2285c 100644
--- a/hbase-rsgroup/src/main/java/org/apache/hadoop/hbase/protobuf/generated/RSGroupAdminProtos.java
+++ b/hbase-rsgroup/src/main/java/org/apache/hadoop/hbase/protobuf/generated/RSGroupAdminProtos.java
@@ -532,12 +532,12 @@ public final class RSGroupAdminProtos {
/**
* <code>repeated .hbase.pb.TableName table_name = 1;</code>
*/
- java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName>
+ java.util.List<org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName>
getTableNameList();
/**
* <code>repeated .hbase.pb.TableName table_name = 1;</code>
*/
- org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName getTableName(int index);
+ org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName getTableName(int index);
/**
* <code>repeated .hbase.pb.TableName table_name = 1;</code>
*/
@@ -545,12 +545,12 @@ public final class RSGroupAdminProtos {
/**
* <code>repeated .hbase.pb.TableName table_name = 1;</code>
*/
- java.util.List<? extends org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder>
+ java.util.List<? extends org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder>
getTableNameOrBuilderList();
/**
* <code>repeated .hbase.pb.TableName table_name = 1;</code>
*/
- org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder(
+ org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder getTableNameOrBuilder(
int index);
}
/**
@@ -606,10 +606,10 @@ public final class RSGroupAdminProtos {
}
case 10: {
if (!((mutable_bitField0_ & 0x00000001) == 0x00000001)) {
- tableName_ = new java.util.ArrayList<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName>();
+ tableName_ = new java.util.ArrayList<org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName>();
mutable_bitField0_ |= 0x00000001;
}
- tableName_.add(input.readMessage(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.PARSER, extensionRegistry));
+ tableName_.add(input.readMessage(org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.PARSER, extensionRegistry));
break;
}
}
@@ -656,17 +656,17 @@ public final class RSGroupAdminProtos {
// repeated .hbase.pb.TableName table_name = 1;
public static final int TABLE_NAME_FIELD_NUMBER = 1;
- private java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName> tableName_;
+ private java.util.List<org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName> tableName_;
/**
* <code>repeated .hbase.pb.TableName table_name = 1;</code>
*/
- public java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName> getTableNameList() {
+ public java.util.List<org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName> getTableNameList() {
return tableName_;
}
/**
* <code>repeated .hbase.pb.TableName table_name = 1;</code>
*/
- public java.util.List<? extends org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder>
+ public java.util.List<? extends org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder>
getTableNameOrBuilderList() {
return tableName_;
}
@@ -679,13 +679,13 @@ public final class RSGroupAdminProtos {
/**
* <code>repeated .hbase.pb.TableName table_name = 1;</code>
*/
- public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName getTableName(int index) {
+ public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName getTableName(int index) {
return tableName_.get(index);
}
/**
* <code>repeated .hbase.pb.TableName table_name = 1;</code>
*/
- public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder(
+ public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder getTableNameOrBuilder(
int index) {
return tableName_.get(index);
}
@@ -996,22 +996,22 @@ public final class RSGroupAdminProtos {
private int bitField0_;
// repeated .hbase.pb.TableName table_name = 1;
- private java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName> tableName_ =
+ private java.util.List<org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName> tableName_ =
java.util.Collections.emptyList();
private void ensureTableNameIsMutable() {
if (!((bitField0_ & 0x00000001) == 0x00000001)) {
- tableName_ = new java.util.ArrayList<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName>(tableName_);
+ tableName_ = new java.util.ArrayList<org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName>(tableName_);
bitField0_ |= 0x00000001;
}
}
private com.google.protobuf.RepeatedFieldBuilder<
- org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder> tableNameBuilder_;
+ org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder> tableNameBuilder_;
/**
* <code>repeated .hbase.pb.TableName table_name = 1;</code>
*/
- public java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName> getTableNameList() {
+ public java.util.List<org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName> getTableNameList() {
if (tableNameBuilder_ == null) {
return java.util.Collections.unmodifiableList(tableName_);
} else {
@@ -1031,7 +1031,7 @@ public final class RSGroupAdminProtos {
/**
* <code>repeated .hbase.pb.TableName table_name = 1;</code>
*/
- public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName getTableName(int index) {
+ public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName getTableName(int index) {
if (tableNameBuilder_ == null) {
return tableName_.get(index);
} else {
@@ -1042,7 +1042,7 @@ public final class RSGroupAdminProtos {
* <code>repeated .hbase.pb.TableName table_name = 1;</code>
*/
public Builder setTableName(
- int index, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName value) {
+ int index, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName value) {
if (tableNameBuilder_ == null) {
if (value == null) {
throw new NullPointerException();
@@ -1059,7 +1059,7 @@ public final class RSGroupAdminProtos {
* <code>repeated .hbase.pb.TableName table_name = 1;</code>
*/
public Builder setTableName(
- int index, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder builderForValue) {
+ int index, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder builderForValue) {
if (tableNameBuilder_ == null) {
ensureTableNameIsMutable();
tableName_.set(index, builderForValue.build());
@@ -1072,7 +1072,7 @@ public final class RSGroupAdminProtos {
/**
* <code>repeated .hbase.pb.TableName table_name = 1;</code>
*/
- public Builder addTableName(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName value) {
+ public Builder addTableName(org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName value) {
if (tableNameBuilder_ == null) {
if (value == null) {
throw new NullPointerException();
@@ -1089,7 +1089,7 @@ public final class RSGroupAdminProtos {
* <code>repeated .hbase.pb.TableName table_name = 1;</code>
*/
public Builder addTableName(
- int index, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName value) {
+ int index, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName value) {
if (tableNameBuilder_ == null) {
if (value == null) {
throw new NullPointerException();
@@ -1106,7 +1106,7 @@ public final class RSGroupAdminProtos {
* <code>repeated .hbase.pb.TableName table_name = 1;</code>
*/
public Builder addTableName(
- org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder builderForValue) {
+ org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder builderForValue) {
if (tableNameBuilder_ == null) {
ensureTableNameIsMutable();
tableName_.add(builderForValue.build());
@@ -1120,7 +1120,7 @@ public final class RSGroupAdminProtos {
* <code>repeated .hbase.pb.TableName table_name = 1;</code>
*/
public Builder addTableName(
- int index, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder builderForValue) {
+ int index, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder builderForValue) {
if (tableNameBuilder_ == null) {
ensureTableNameIsMutable();
tableName_.add(index, builderForValue.build());
@@ -1134,7 +1134,7 @@ public final class RSGroupAdminProtos {
* <code>repeated .hbase.pb.TableName table_name = 1;</code>
*/
public Builder addAllTableName(
- java.lang.Iterable<? extends org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName> values) {
+ java.lang.Iterable<? extends org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName> values) {
if (tableNameBuilder_ == null) {
ensureTableNameIsMutable();
super.addAll(values, tableName_);
@@ -1173,14 +1173,14 @@ public final class RSGroupAdminProtos {
/**
* <code>repeated .hbase.pb.TableName table_name = 1;</code>
*/
- public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder getTableNameBuilder(
+ public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder getTableNameBuilder(
int index) {
return getTableNameFieldBuilder().getBuilder(index);
}
/**
* <code>repeated .hbase.pb.TableName table_name = 1;</code>
*/
- public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder(
+ public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder getTableNameOrBuilder(
int index) {
if (tableNameBuilder_ == null) {
return tableName_.get(index); } else {
@@ -1190,7 +1190,7 @@ public final class RSGroupAdminProtos {
/**
* <code>repeated .hbase.pb.TableName table_name = 1;</code>
*/
- public java.util.List<? extends org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder>
+ public java.util.List<? extends org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder>
getTableNameOrBuilderList() {
if (tableNameBuilder_ != null) {
return tableNameBuilder_.getMessageOrBuilderList();
@@ -1201,31 +1201,31 @@ public final class RSGroupAdminProtos {
/**
* <code>repeated .hbase.pb.TableName table_name = 1;</code>
*/
- public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder addTableNameBuilder() {
+ public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder addTableNameBuilder() {
return getTableNameFieldBuilder().addBuilder(
- org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance());
+ org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance());
}
/**
* <code>repeated .hbase.pb.TableName table_name = 1;</code>
*/
- public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder addTableNameBuilder(
+ public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder addTableNameBuilder(
int index) {
return getTableNameFieldBuilder().addBuilder(
- index, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance());
+ index, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance());
}
/**
* <code>repeated .hbase.pb.TableName table_name = 1;</code>
*/
- public java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder>
+ public java.util.List<org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder>
getTableNameBuilderList() {
return getTableNameFieldBuilder().getBuilderList();
}
private com.google.protobuf.RepeatedFieldBuilder<
- org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder>
+ org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder>
getTableNameFieldBuilder() {
if (tableNameBuilder_ == null) {
tableNameBuilder_ = new com.google.protobuf.RepeatedFieldBuilder<
- org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder>(
+ org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder>(
tableName_,
((bitField0_ & 0x00000001) == 0x00000001),
getParentForChildren(),
@@ -2331,11 +2331,11 @@ public final class RSGroupAdminProtos {
/**
* <code>required .hbase.pb.TableName table_name = 1;</code>
*/
- org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName getTableName();
+ org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName getTableName();
/**
* <code>required .hbase.pb.TableName table_name = 1;</code>
*/
- org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder();
+ org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder getTableNameOrBuilder();
}
/**
* Protobuf type {@code hbase.pb.GetRSGroupInfoOfTableRequest}
@@ -2389,11 +2389,11 @@ public final class RSGroupAdminProtos {
break;
}
case 10: {
- org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder subBuilder = null;
+ org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder subBuilder = null;
if (((bitField0_ & 0x00000001) == 0x00000001)) {
subBuilder = tableName_.toBuilder();
}
- tableName_ = input.readMessage(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.PARSER, extensionRegistry);
+ tableName_ = input.readMessage(org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.PARSER, extensionRegistry);
if (subBuilder != null) {
subBuilder.mergeFrom(tableName_);
tableName_ = subBuilder.buildPartial();
@@ -2443,7 +2443,7 @@ public final class RSGroupAdminProtos {
private int bitField0_;
// required .hbase.pb.TableName table_name = 1;
public static final int TABLE_NAME_FIELD_NUMBER = 1;
- private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName tableName_;
+ private org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName tableName_;
/**
* <code>required .hbase.pb.TableName table_name = 1;</code>
*/
@@ -2453,18 +2453,18 @@ public final class RSGroupAdminProtos {
/**
* <code>required .hbase.pb.TableName table_name = 1;</code>
*/
- public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName getTableName() {
+ public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName getTableName() {
return tableName_;
}
/**
* <code>required .hbase.pb.TableName table_name = 1;</code>
*/
- public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder() {
+ public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder getTableNameOrBuilder() {
return tableName_;
}
private void initFields() {
- tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance();
+ tableName_ = org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance();
}
private byte memoizedIsInitialized = -1;
public final boolean isInitialized() {
@@ -2658,7 +2658,7 @@ public final class RSGroupAdminProtos {
public Builder clear() {
super.clear();
if (tableNameBuilder_ == null) {
- tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance();
+ tableName_ = org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance();
} else {
tableNameBuilder_.clear();
}
@@ -2754,9 +2754,9 @@ public final class RSGroupAdminProtos {
private int bitField0_;
// required .hbase.pb.TableName table_name = 1;
- private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance();
+ private org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName tableName_ = org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance();
private com.google.protobuf.SingleFieldBuilder<
- org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder> tableNameBuilder_;
+ org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder> tableNameBuilder_;
/**
* <code>required .hbase.pb.TableName table_name = 1;</code>
*/
@@ -2766,7 +2766,7 @@ public final class RSGroupAdminProtos {
/**
* <code>required .hbase.pb.TableName table_name = 1;</code>
*/
- public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName getTableName() {
+ public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName getTableName() {
if (tableNameBuilder_ == null) {
return tableName_;
} else {
@@ -2776,7 +2776,7 @@ public final class RSGroupAdminProtos {
/**
* <code>required .hbase.pb.TableName table_name = 1;</code>
*/
- public Builder setTableName(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName value) {
+ public Builder setTableName(org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName value) {
if (tableNameBuilder_ == null) {
if (value == null) {
throw new NullPointerException();
@@ -2793,7 +2793,7 @@ public final class RSGroupAdminProtos {
* <code>required .hbase.pb.TableName table_name = 1;</code>
*/
public Builder setTableName(
- org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder builderForValue) {
+ org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder builderForValue) {
if (tableNameBuilder_ == null) {
tableName_ = builderForValue.build();
onChanged();
@@ -2806,12 +2806,12 @@ public final class RSGroupAdminProtos {
/**
* <code>required .hbase.pb.TableName table_name = 1;</code>
*/
- public Builder mergeTableName(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName value) {
+ public Builder mergeTableName(org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName value) {
if (tableNameBuilder_ == null) {
if (((bitField0_ & 0x00000001) == 0x00000001) &&
- tableName_ != org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance()) {
+ tableName_ != org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance()) {
tableName_ =
- org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.newBuilder(tableName_).mergeFrom(value).buildPartial();
+ org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.newBuilder(tableName_).mergeFrom(value).buildPartial();
} else {
tableName_ = value;
}
@@ -2827,7 +2827,7 @@ public final class RSGroupAdminProtos {
*/
public Builder clearTableName() {
if (tableNameBuilder_ == null) {
- tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance();
+ tableName_ = org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance();
onChanged();
} else {
tableNameBuilder_.clear();
@@ -2838,7 +2838,7 @@ public final class RSGroupAdminProtos {
/**
* <code>required .hbase.pb.TableName table_name = 1;</code>
*/
- public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder getTableNameBuilder() {
+ public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder getTableNameBuilder() {
bitField0_ |= 0x00000001;
onChanged();
return getTableNameFieldBuilder().getBuilder();
@@ -2846,7 +2846,7 @@ public final class RSGroupAdminProtos {
/**
* <code>required .hbase.pb.TableName table_name = 1;</code>
*/
- public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder() {
+ public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder getTableNameOrBuilder() {
if (tableNameBuilder_ != null) {
return tableNameBuilder_.getMessageOrBuilder();
} else {
@@ -2857,11 +2857,11 @@ public final class RSGroupAdminProtos {
* <code>required .hbase.pb.TableName table_name = 1;</code>
*/
private com.google.protobuf.SingleFieldBuilder<
- org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder>
+ org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder>
getTableNameFieldBuilder() {
if (tableNameBuilder_ == null) {
tableNameBuilder_ = new com.google.protobuf.SingleFieldBuilder<
- org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder>(
+ org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder>(
tableName_,
getParentForChildren(),
isClean());
@@ -4695,12 +4695,12 @@ public final class RSGroupAdminProtos {
/**
* <code>repeated .hbase.pb.TableName table_name = 2;</code>
*/
- java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName>
+ java.util.List<org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName>
getTableNameList();
/**
* <code>repeated .hbase.pb.TableName table_name = 2;</code>
*/
- org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName getTableName(int index);
+ org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName getTableName(int index);
/**
* <code>repeated .hbase.pb.TableName table_name = 2;</code>
*/
@@ -4708,12 +4708,12 @@ public final class RSGroupAdminProtos {
/**
* <code>repeated .hbase.pb.TableName table_name = 2;</code>
*/
- java.util.List<? extends org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder>
+ java.util.List<? extends org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder>
getTableNameOrBuilderList();
/**
* <code>repeated .hbase.pb.TableName table_name = 2;</code>
*/
- org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder(
+ org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder getTableNameOrBuilder(
int index);
}
/**
@@ -4774,10 +4774,10 @@ public final class RSGroupAdminProtos {
}
case 18: {
if (!((mutable_bitField0_ & 0x00000002) == 0x00000002)) {
- tableName_ = new java.util.ArrayList<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName>();
+ tableName_ = new java.util.ArrayList<org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName>();
mutable_bitField0_ |= 0x00000002;
}
- tableName_.add(input.readMessage(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.PARSER, extensionRegistry));
+ tableName_.add(input.readMessage(org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.PARSER, extensionRegistry));
break;
}
}
@@ -4868,17 +4868,17 @@ public final class RSGroupAdminProtos {
// repeated .hbase.pb.TableName table_name = 2;
public static final int TABLE_NAME_FIELD_NUMBER = 2;
- private java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName> tableName_;
+ private java.util.List<org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName> tableName_;
/**
* <code>repeated .hbase.pb.TableName table_name = 2;</code>
*/
- public java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName> getTableNameList() {
+ public java.util.List<org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName> getTableNameList() {
return tableName_;
}
/**
* <code>repeated .hbase.pb.TableName table_name = 2;</code>
*/
- public java.util.List<? extends org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder>
+ public java.util.List<? extends org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder>
getTableNameOrBuilderList() {
return tableName_;
}
@@ -4891,13 +4891,13 @@ public final class RSGroupAdminProtos {
/**
* <code>repeated .hbase.pb.TableName table_name = 2;</code>
*/
- public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName getTableName(int index) {
+ public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName getTableName(int index) {
return tableName_.get(index);
}
/**
* <code>repeated .hbase.pb.TableName table_name = 2;</code>
*/
- public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder(
+ public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder getTableNameOrBuilder(
int index) {
return tableName_.get(index);
}
@@ -5320,22 +5320,22 @@ public final class RSGroupAdminProtos {
}
// repeated .hbase.pb.TableName table_name = 2;
- private java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName> tableName_ =
+ private java.util.List<org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName> tableName_ =
java.util.Collections.emptyList();
private void ensureTableNameIsMutable() {
if (!((bitField0_ & 0x00000002) == 0x00000002)) {
- tableName_ = new java.util.ArrayList<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName>(tableName_);
+ tableName_ = new java.util.ArrayList<org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName>(tableName_);
bitField0_ |= 0x00000002;
}
}
private com.google.protobuf.RepeatedFieldBuilder<
- org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder> tableNameBuilder_;
+ org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder> tableNameBuilder_;
/**
* <code>repeated .hbase.pb.TableName table_name = 2;</code>
*/
- public java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName> getTableNameList() {
+ public java.util.List<org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName> getTableNameList() {
if (tableNameBuilder_ == null) {
return java.util.Collections.unmodifiableList(tableName_);
} else {
@@ -5355,7 +5355,7 @@ public final class RSGroupAdminProtos {
/**
* <code>repeated .hbase.pb.TableName table_name = 2;</code>
*/
- public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName getTableName(int index) {
+ public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName getTableName(int index) {
if (tableNameBuilder_ == null) {
return tableName_.get(index);
} else {
@@ -5366,7 +5366,7 @@ public final class RSGroupAdminProtos {
* <code>repeated .hbase.pb.TableName table_name = 2;</code>
*/
public Builder setTableName(
- int index, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName value) {
+ int index, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName value) {
if (tableNameBuilder_ == null) {
if (value == null) {
throw new NullPointerException();
@@ -5383,7 +5383,7 @@ public final class RSGroupAdminProtos {
* <code>repeated .hbase.pb.TableName table_name = 2;</code>
*/
public Builder setTableName(
- int index, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder builderForValue) {
+ int index, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder builderForValue) {
if (tableNameBuilder_ == null) {
ensureTableNameIsMutable();
tableName_.set(index, builderForValue.build());
@@ -5396,7 +5396,7 @@ public final class RSGroupAdminProtos {
/**
* <code>repeated .hbase.pb.TableName table_name = 2;</code>
*/
- public Builder addTableName(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName value) {
+ public Builder addTableName(org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName value) {
if (tableNameBuilder_ == null) {
if (value == null) {
throw new NullPointerException();
@@ -5413,7 +5413,7 @@ public final class RSGroupAdminProtos {
* <code>repeated .hbase.pb.TableName table_name = 2;</code>
*/
public Builder addTableName(
- int index, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName value) {
+ int index, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName value) {
if (tableNameBuilder_ == null) {
if (value == null) {
throw new NullPointerException();
@@ -5430,7 +5430,7 @@ public final class RSGroupAdminProtos {
* <code>repeated .hbase.pb.TableName table_name = 2;</code>
*/
public Builder addTableName(
- org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder builderForValue) {
+ org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder builderForValue) {
if (tableNameBuilder_ == null) {
ensureTableNameIsMutable();
tableName_.add(builderForValue.build());
@@ -5444,7 +5444,7 @@ public final class RSGroupAdminProtos {
* <code>repeated .hbase.pb.TableName table_name = 2;</code>
*/
public Builder addTableName(
- int index, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder builderForValue) {
+ int index, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder builderForValue) {
if (tableNameBuilder_ == null) {
ensureTableNameIsMutable();
tableName_.add(index, builderForValue.build());
@@ -5458,7 +5458,7 @@ public final class RSGroupAdminProtos {
* <code>repeated .hbase.pb.TableName table_name = 2;</code>
*/
public Builder addAllTableName(
- java.lang.Iterable<? extends org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName> values) {
+ java.lang.Iterable<? extends org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName> values) {
if (tableNameBuilder_ == null) {
ensureTableNameIsMutable();
super.addAll(values, tableName_);
@@ -5497,14 +5497,14 @@ public final class RSGroupAdminProtos {
/**
* <code>repeated .hbase.pb.TableName table_name = 2;</code>
*/
- public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder getTableNameBuilder(
+ public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder getTableNameBuilder(
int index) {
return getTableNameFieldBuilder().getBuilder(index);
}
/**
* <code>repeated .hbase.pb.TableName table_name = 2;</code>
*/
- public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder(
+ public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder getTableNameOrBuilder(
int index) {
if (tableNameBuilder_ == null) {
return tableName_.get(index); } else {
@@ -5514,7 +5514,7 @@ public final class RSGroupAdminProtos {
/**
* <code>repeated .hbase.pb.TableName table_name = 2;</code>
*/
- public java.util.List<? extends org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder>
+ public java.util.List<? extends org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder>
getTableNameOrBuilderList() {
if (tableNameBuilder_ != null) {
return tableNameBuilder_.getMessageOrBuilderList();
@@ -5525,31 +5525,31 @@ public final class RSGroupAdminProtos {
/**
* <code>repeated .hbase.pb.TableName table_name = 2;</code>
*/
- public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder addTableNameBuilder() {
+ public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder addTableNameBuilder() {
return getTableNameFieldBuilder().addBuilder(
- org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance());
+ org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance());
}
/**
* <code>repeated .hbase.pb.TableName table_name = 2;</code>
*/
- public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder addTableNameBuilder(
+ public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder addTableNameBuilder(
int index) {
return getTableNameFieldBuilder().addBuilder(
- index, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance());
+ index, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance());
}
/**
* <code>repeated .hbase.pb.TableName table_name = 2;</code>
*/
- public java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder>
+ public java.util.List<org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder>
getTableNameBuilderList() {
return getTableNameFieldBuilder().getBuilderList();
}
private com.google.protobuf.RepeatedFieldBuilder<
- org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder>
+ org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder>
getTableNameFieldBuilder() {
if (tableNameBuilder_ == null) {
tableNameBuilder_ = new com.google.protobuf.RepeatedFieldBuilder<
- org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder>(
+ org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder>(
tableName_,
((bitField0_ & 0x00000002) == 0x00000002),
getParentForChildren(),
@@ -10801,12 +10801,12 @@ public final class RSGroupAdminProtos {
/**
* <code>repeated .hbase.pb.TableName table_name = 3;</code>
*/
- java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName>
+ java.util.List<org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName>
getTableNameList();
/**
* <code>repeated .hbase.pb.TableName table_name = 3;</code>
*/
- org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName getTableName(int index);
+ org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName getTableName(int index);
/**
* <code>repeated .hbase.pb.TableName table_name = 3;</code>
*/
@@ -10814,12 +10814,12 @@ public final class RSGroupAdminProtos {
/**
* <code>repeated .hbase.pb.TableName table_name = 3;</code>
*/
- java.util.List<? extends org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder>
+ java.util.List<? extends org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder>
getTableNameOrBuilderList();
/**
* <code>repeated .hbase.pb.TableName table_name = 3;</code>
*/
- org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder(
+ org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder getTableNameOrBuilder(
int index);
}
/**
@@ -10888,10 +10888,10 @@ public final class RSGroupAdminProtos {
}
case 26: {
if (!((mutable_bitField0_ & 0x00000004) == 0x00000004)) {
- tableName_ = new java.util.ArrayList<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName>();
+ tableName_ = new java.util.ArrayList<org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName>();
mutable_bitField0_ |= 0x00000004;
}
- tableName_.add(input.readMessage(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.PARSER, extensionRegistry));
+ tableName_.add(input.readMessage(org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.PARSER, extensionRegistry));
break;
}
}
@@ -11021,17 +11021,17 @@ public final class RSGroupAdminProtos {
// repeated .hbase.pb.TableName table_name = 3;
public static final int TABLE_NAME_FIELD_NUMBER = 3;
- private java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName> tableName_;
+ private java.util.List<org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName> tableName_;
/**
* <code>repeated .hbase.pb.TableName table_name = 3;</code>
*/
- public java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName> getTableNameList() {
+ public java.util.List<org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName> getTableNameList() {
return tableName_;
}
/**
* <code>repeated .hbase.pb.TableName table_name = 3;</code>
*/
- public java.util.List<? extends org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder>
+ public java.util.List<? extends org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder>
getTableNameOrBuilderList() {
return tableName_;
}
@@ -11044,13 +11044,13 @@ public final class RSGroupAdminProtos {
/**
* <code>repeated .hbase.pb.TableName table_name = 3;</code>
*/
- public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName getTableName(int index) {
+ public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName getTableName(int index) {
return tableName_.get(index);
}
/**
* <code>repeated .hbase.pb.TableName table_name = 3;</code>
*/
- public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder(
+ public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder getTableNameOrBuilder(
int index) {
return tableName_.get(index);
}
@@ -11781,22 +11781,22 @@ public final class RSGroupAdminProtos {
}
// repeated .hbase.pb.TableName table_name = 3;
- private java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName> tableName_ =
+ private java.util.List<org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName> tableName_ =
java.util.Collections.emptyList();
private void ensureTableNameIsMutable() {
if (!((bitField0_ & 0x00000004) == 0x00000004)) {
- tableName_ = new java.util.ArrayList<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName>(tableName_);
+ tableName_ = new java.util.ArrayList<org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName>(tableName_);
bitField0_ |= 0x00000004;
}
}
private com.google.protobuf.RepeatedFieldBuilder<
- org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder> tableNameBuilder_;
+ org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder> tableNameBuilder_;
/**
* <code>repeated .hbase.pb.TableName table_name = 3;</code>
*/
- public java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName> getTableNameList() {
+ public java.util.List<org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName> getTableNameList() {
if (tableNameBuilder_ == null) {
return java.util.Collections.unmodifiableList(tableName_);
} else {
@@ -11816,7 +11816,7 @@ public final class RSGroupAdminProtos {
/**
* <code>repeated .hbase.pb.TableName table_name = 3;</code>
*/
- public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName getTableName(int index) {
+ public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName getTableName(int index) {
if (tableNameBuilder_ == null) {
return tableName_.get(index);
} else {
@@ -11827,7 +11827,7 @@ public final class RSGroupAdminProtos {
* <code>repeated .hbase.pb.TableName table_name = 3;</code>
*/
public Builder setTableName(
- int index, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName value) {
+ int index, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName value) {
if (tableNameBuilder_ == null) {
if (value == null) {
throw new NullPointerException();
@@ -11844,7 +11844,7 @@ public final class RSGroupAdminProtos {
* <code>repeated .hbase.pb.TableName table_name = 3;</code>
*/
public Builder setTableName(
- int index, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder builderForValue) {
+ int index, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder builderForValue) {
if (tableNameBuilder_ == null) {
ensureTableNameIsMutable();
tableName_.set(index, builderForValue.build());
@@ -11857,7 +11857,7 @@ public final class RSGroupAdminProtos {
/**
* <code>repeated .hbase.pb.TableName table_name = 3;</code>
*/
- public Builder addTableName(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName value) {
+ public Builder addTableName(org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName value) {
if (tableNameBuilder_ == null) {
if (value == null) {
throw new NullPointerException();
@@ -11874,7 +11874,7 @@ public final class RSGroupAdminProtos {
* <code>repeated .hbase.pb.TableName table_name = 3;</code>
*/
public Builder addTableName(
- int index, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName value) {
+ int index, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName value) {
if (tableNameBuilder_ == null) {
if (value == null) {
throw new NullPointerException();
@@ -11891,7 +11891,7 @@ public final class RSGroupAdminProtos {
* <code>repeated .hbase.pb.TableName table_name = 3;</code>
*/
public Builder addTableName(
- org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder builderForValue) {
+ org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder builderForValue) {
if (tableNameBuilder_ == null) {
ensureTableNameIsMutable();
tableName_.add(builderForValue.build());
@@ -11905,7 +11905,7 @@ public final class RSGroupAdminProtos {
* <code>repeated .hbase.pb.TableName table_name = 3;</code>
*/
public Builder addTableName(
- int index, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder builderForValue) {
+ int index, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder builderForValue) {
if (tableNameBuilder_ == null) {
ensureTableNameIsMutable();
tableName_.add(index, builderForValue.build());
@@ -11919,7 +11919,7 @@ public final class RSGroupAdminProtos {
* <code>repeated .hbase.pb.TableName table_name = 3;</code>
*/
public Builder addAllTableName(
- java.lang.Iterable<? extends org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName> values) {
+ java.lang.Iterable<? extends org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName> values) {
if (tableNameBuilder_ == null) {
ensureTableNameIsMutable();
super.addAll(values, tableName_);
@@ -11958,14 +11958,14 @@ public final class RSGroupAdminProtos {
/**
* <code>repeated .hbase.pb.TableName table_name = 3;</code>
*/
- public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder getTableNameBuilder(
+ public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder getTableNameBuilder(
int index) {
return getTableNameFieldBuilder().getBuilder(index);
}
/**
* <code>repeated .hbase.pb.TableName table_name = 3;</code>
*/
- public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder(
+ public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder getTableNameOrBuilder(
int index) {
if (tableNameBuilder_ == null) {
return tableName_.get(index); } else {
@@ -11975,7 +11975,7 @@ public final class RSGroupAdminProtos {
/**
* <code>repeated .hbase.pb.TableName table_name = 3;</code>
*/
- public java.util.List<? extends org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder>
+ public java.util.List<? extends org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder>
getTableNameOrBuilderList() {
if (tableNameBuilder_ != null) {
return tableNameBuilder_.getMessageOrBuilderList();
@@ -11986,31 +11986,31 @@ public final class RSGroupAdminProtos {
/**
* <code>repeated .hbase.pb.TableName table_name = 3;</code>
*/
- public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder addTableNameBuilder() {
+ public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder addTableNameBuilder() {
return getTableNameFieldBuilder().addBuilder(
- org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance());
+ org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance());
}
/**
* <code>repeated .hbase.pb.TableName table_name = 3;</code>
*/
- public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder addTableNameBuilder(
+ public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder addTableNameBuilder(
int index) {
return getTableNameFieldBuilder().addBuilder(
- index, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance());
+ index, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance());
}
/**
* <code>repeated .hbase.pb.TableName table_name = 3;</code>
*/
- public java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder>
+ public java.util.List<org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder>
getTableNameBuilderList() {
return getTableNameFieldBuilder().getBuilderList();
}
private com.google.protobuf.RepeatedFieldBuilder<
- org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder>
+ org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder>
getTableNameFieldBuilder() {
if (tableNameBuilder_ == null) {
tableNameBuilder_ = new com.google.protobuf.RepeatedFieldBuilder<
- org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder>(
+ org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder>(
tableName_,
((bitField0_ & 0x00000004) == 0x00000004),
getParentForChildren(),
@@ -13359,64 +13359,64 @@ public final class RSGroupAdminProtos {
descriptor;
static {
java.lang.String[] descriptorData = {
- "\n\022RSGroupAdmin.proto\022\010hbase.pb\032\013HBase.pr" +
- "oto\032\rRSGroup.proto\"4\n\032ListTablesOfRSGrou" +
- "pRequest\022\026\n\016r_s_group_name\030\001 \002(\t\"F\n\033List" +
- "TablesOfRSGroupResponse\022\'\n\ntable_name\030\001 " +
- "\003(\0132\023.hbase.pb.TableName\"/\n\025GetRSGroupIn" +
- "foRequest\022\026\n\016r_s_group_name\030\001 \002(\t\"G\n\026Get" +
- "RSGroupInfoResponse\022-\n\016r_s_group_info\030\001 " +
- "\001(\0132\025.hbase.pb.RSGroupInfo\"G\n\034GetRSGroup" +
- "InfoOfTableRequest\022\'\n\ntable_name\030\001 \002(\0132\023" +
- ".hbase.pb.TableName\"N\n\035GetRSGroupInfoOfT",
- "ableResponse\022-\n\016r_s_group_info\030\001 \001(\0132\025.h" +
- "base.pb.RSGroupInfo\"Q\n\022MoveServersReques" +
- "t\022\024\n\014target_group\030\001 \002(\t\022%\n\007servers\030\003 \003(\013" +
- "2\024.hbase.pb.ServerName\"\025\n\023MoveServersRes" +
- "ponse\"R\n\021MoveTablesRequest\022\024\n\014target_gro" +
- "up\030\001 \002(\t\022\'\n\ntable_name\030\002 \003(\0132\023.hbase.pb." +
- "TableName\"\024\n\022MoveTablesResponse\"+\n\021AddRS" +
- "GroupRequest\022\026\n\016r_s_group_name\030\001 \002(\t\"\024\n\022" +
- "AddRSGroupResponse\".\n\024RemoveRSGroupReque" +
- "st\022\026\n\016r_s_group_name\030\001 \002(\t\"\027\n\025RemoveRSGr",
- "oupResponse\"/\n\025BalanceRSGroupRequest\022\026\n\016" +
- "r_s_group_name\030\001 \002(\t\",\n\026BalanceRSGroupRe" +
- "sponse\022\022\n\nbalanceRan\030\001 \002(\010\"\031\n\027ListRSGrou" +
- "pInfosRequest\"I\n\030ListRSGroupInfosRespons" +
- "e\022-\n\016r_s_group_info\030\001 \003(\0132\025.hbase.pb.RSG" +
- "roupInfo\"E\n\035GetRSGroupInfoOfServerReques" +
- "t\022$\n\006server\030\002 \002(\0132\024.hbase.pb.ServerName\"" +
- "O\n\036GetRSGroupInfoOfServerResponse\022-\n\016r_s" +
- "_group_info\030\001 \001(\0132\025.hbase.pb.RSGroupInfo" +
- "\"\203\001\n\033MoveServersAndTablesRequest\022\024\n\014targ",
- "et_group\030\001 \002(\t\022%\n\007servers\030\002 \003(\0132\024.hbase." +
- "pb.ServerName\022\'\n\ntable_name\030\003 \003(\0132\023.hbas" +
- "e.pb.TableName\"\036\n\034MoveServersAndTablesRe" +
- "sponse2\210\007\n\023RSGroupAdminService\022S\n\016GetRSG" +
- "roupInfo\022\037.hbase.pb.GetRSGroupInfoReques" +
- "t\032 .hbase.pb.GetRSGroupInfoResponse\022h\n\025G" +
- "etRSGroupInfoOfTable\022&.hbase.pb.GetRSGro" +
- "upInfoOfTableRequest\032\'.hbase.pb.GetRSGro" +
- "upInfoOfTableResponse\022k\n\026GetRSGroupInfoO" +
- "fServer\022\'.hbase.pb.GetRSGroupInfoOfServe",
- "rRequest\032(.hbase.pb.GetRSGroupInfoOfServ" +
- "erResponse\022J\n\013MoveServers\022\034.hbase.pb.Mov" +
- "eServersRequest\032\035.hbase.pb.MoveServersRe" +
- "sponse\022G\n\nMoveTables\022\033.hbase.pb.MoveTabl" +
- "esRequest\032\034.hbase.pb.MoveTablesResponse\022" +
- "G\n\nAddRSGroup\022\033.hbase.pb.AddRSGroupReque" +
- "st\032\034.hbase.pb.AddRSGroupResponse\022P\n\rRemo" +
- "veRSGroup\022\036.hbase.pb.RemoveRSGroupReques" +
- "t\032\037.hbase.pb.RemoveRSGroupResponse\022S\n\016Ba" +
- "lanceRSGroup\022\037.hbase.pb.BalanceRSGroupRe",
- "quest\032 .hbase.pb.BalanceRSGroupResponse\022" +
- "Y\n\020ListRSGroupInfos\022!.hbase.pb.ListRSGro" +
- "upInfosRequest\032\".hbase.pb.ListRSGroupInf" +
- "osResponse\022e\n\024MoveServersAndTables\022%.hba" +
- "se.pb.MoveServersAndTablesRequest\032&.hbas" +
- "e.pb.MoveServersAndTablesResponseBH\n*org" +
- ".apache.hadoop.hbase.protobuf.generatedB" +
- "\022RSGroupAdminProtosH\001\210\001\001\240\001\001"
+ "\n\022RSGroupAdmin.proto\022\010hbase.pb\032\013Table.pr" +
+ "oto\032\013HBase.proto\032\rRSGroup.proto\"4\n\032ListT" +
+ "ablesOfRSGroupRequest\022\026\n\016r_s_group_name\030" +
+ "\001 \002(\t\"F\n\033ListTablesOfRSGroupResponse\022\'\n\n" +
+ "table_name\030\001 \003(\0132\023.hbase.pb.TableName\"/\n" +
+ "\025GetRSGroupInfoRequest\022\026\n\016r_s_group_name" +
+ "\030\001 \002(\t\"G\n\026GetRSGroupInfoResponse\022-\n\016r_s_" +
+ "group_info\030\001 \001(\0132\025.hbase.pb.RSGroupInfo\"" +
+ "G\n\034GetRSGroupInfoOfTableRequest\022\'\n\ntable" +
+ "_name\030\001 \002(\0132\023.hbase.pb.TableName\"N\n\035GetR",
+ "SGroupInfoOfTableResponse\022-\n\016r_s_group_i" +
+ "nfo\030\001 \001(\0132\025.hbase.pb.RSGroupInfo\"Q\n\022Move" +
+ "ServersRequest\022\024\n\014target_group\030\001 \002(\t\022%\n\007" +
+ "servers\030\003 \003(\0132\024.hbase.pb.ServerName\"\025\n\023M" +
+ "oveServersResponse\"R\n\021MoveTablesRequest\022" +
+ "\024\n\014target_group\030\001 \002(\t\022\'\n\ntable_name\030\002 \003(" +
+ "\0132\023.hbase.pb.TableName\"\024\n\022MoveTablesResp" +
+ "onse\"+\n\021AddRSGroupRequest\022\026\n\016r_s_group_n" +
+ "ame\030\001 \002(\t\"\024\n\022AddRSGroupResponse\".\n\024Remov" +
+ "eRSGroupRequest\022\026\n\016r_s_group_name\030\001 \002(\t\"",
+ "\027\n\025RemoveRSGroupResponse\"/\n\025BalanceRSGro" +
+ "upRequest\022\026\n\016r_s_group_name\030\001 \002(\t\",\n\026Bal" +
+ "anceRSGroupResponse\022\022\n\nbalanceRan\030\001 \002(\010\"" +
+ "\031\n\027ListRSGroupInfosRequest\"I\n\030ListRSGrou" +
+ "pInfosResponse\022-\n\016r_s_group_info\030\001 \003(\0132\025" +
+ ".hbase.pb.RSGroupInfo\"E\n\035GetRSGroupInfoO" +
+ "fServerRequest\022$\n\006server\030\002 \002(\0132\024.hbase.p" +
+ "b.ServerName\"O\n\036GetRSGroupInfoOfServerRe" +
+ "sponse\022-\n\016r_s_group_info\030\001 \001(\0132\025.hbase.p" +
+ "b.RSGroupInfo\"\203\001\n\033MoveServersAndTablesRe",
+ "quest\022\024\n\014target_group\030\001 \002(\t\022%\n\007servers\030\002" +
+ " \003(\0132\024.hbase.pb.ServerName\022\'\n\ntable_name" +
+ "\030\003 \003(\0132\023.hbase.pb.TableName\"\036\n\034MoveServe" +
+ "rsAndTablesResponse2\210\007\n\023RSGroupAdminServ" +
+ "ice\022S\n\016GetRSGroupInfo\022\037.hbase.pb.GetRSGr" +
+ "oupInfoRequest\032 .hbase.pb.GetRSGroupInfo" +
+ "Response\022h\n\025GetRSGroupInfoOfTable\022&.hbas" +
+ "e.pb.GetRSGroupInfoOfTableRequest\032\'.hbas" +
+ "e.pb.GetRSGroupInfoOfTableResponse\022k\n\026Ge" +
+ "tRSGroupInfoOfServer\022\'.hbase.pb.GetRSGro",
+ "upInfoOfServerRequest\032(.hbase.pb.GetRSGr" +
+ "oupInfoOfServerResponse\022J\n\013MoveServers\022\034" +
+ ".hbase.pb.MoveServersRequest\032\035.hbase.pb." +
+ "MoveServersResponse\022G\n\nMoveTables\022\033.hbas" +
+ "e.pb.MoveTablesRequest\032\034.hbase.pb.MoveTa" +
+ "blesResponse\022G\n\nAddRSGroup\022\033.hbase.pb.Ad" +
+ "dRSGroupRequest\032\034.hbase.pb.AddRSGroupRes" +
+ "ponse\022P\n\rRemoveRSGroup\022\036.hbase.pb.Remove" +
+ "RSGroupRequest\032\037.hbase.pb.RemoveRSGroupR" +
+ "esponse\022S\n\016BalanceRSGroup\022\037.hbase.pb.Bal",
+ "anceRSGroupRequest\032 .hbase.pb.BalanceRSG" +
+ "roupResponse\022Y\n\020ListRSGroupInfos\022!.hbase" +
+ ".pb.ListRSGroupInfosRequest\032\".hbase.pb.L" +
+ "istRSGroupInfosResponse\022e\n\024MoveServersAn" +
+ "dTables\022%.hbase.pb.MoveServersAndTablesR" +
+ "equest\032&.hbase.pb.MoveServersAndTablesRe" +
+ "sponseBH\n*org.apache.hadoop.hbase.protob" +
+ "uf.generatedB\022RSGroupAdminProtosH\001\210\001\001\240\001\001"
};
com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() {
@@ -13561,6 +13561,7 @@ public final class RSGroupAdminProtos {
com.google.protobuf.Descriptors.FileDescriptor
.internalBuildGeneratedFileFrom(descriptorData,
new com.google.protobuf.Descriptors.FileDescriptor[] {
+ org.apache.hadoop.hbase.protobuf.generated.TableProtos.getDescriptor(),
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.getDescriptor(),
org.apache.hadoop.hbase.protobuf.generated.RSGroupProtos.getDescriptor(),
}, assigner);
http://git-wip-us.apache.org/repos/asf/hbase/blob/5e21e306/hbase-rsgroup/src/main/java/org/apache/hadoop/hbase/protobuf/generated/RSGroupProtos.java
----------------------------------------------------------------------
diff --git a/hbase-rsgroup/src/main/java/org/apache/hadoop/hbase/protobuf/generated/RSGroupProtos.java b/hbase-rsgroup/src/main/java/org/apache/hadoop/hbase/protobuf/generated/RSGroupProtos.java
index 979f762..5f5eb3b 100644
--- a/hbase-rsgroup/src/main/java/org/apache/hadoop/hbase/protobuf/generated/RSGroupProtos.java
+++ b/hbase-rsgroup/src/main/java/org/apache/hadoop/hbase/protobuf/generated/RSGroupProtos.java
@@ -55,12 +55,12 @@ public final class RSGroupProtos {
/**
* <code>repeated .hbase.pb.TableName tables = 3;</code>
*/
- java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName>
+ java.util.List<org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName>
getTablesList();
/**
* <code>repeated .hbase.pb.TableName tables = 3;</code>
*/
- org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName getTables(int index);
+ org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName getTables(int index);
/**
* <code>repeated .hbase.pb.TableName tables = 3;</code>
*/
@@ -68,12 +68,12 @@ public final class RSGroupProtos {
/**
* <code>repeated .hbase.pb.TableName tables = 3;</code>
*/
- java.util.List<? extends org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder>
+ java.util.List<? extends org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder>
getTablesOrBuilderList();
/**
* <code>repeated .hbase.pb.TableName tables = 3;</code>
*/
- org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder getTablesOrBuilder(
+ org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder getTablesOrBuilder(
int index);
}
/**
@@ -134,10 +134,10 @@ public final class RSGroupProtos {
}
case 26: {
if (!((mutable_bitField0_ & 0x00000004) == 0x00000004)) {
- tables_ = new java.util.ArrayList<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName>();
+ tables_ = new java.util.ArrayList<org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName>();
mutable_bitField0_ |= 0x00000004;
}
- tables_.add(input.readMessage(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.PARSER, extensionRegistry));
+ tables_.add(input.readMessage(org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.PARSER, extensionRegistry));
break;
}
case 34: {
@@ -275,17 +275,17 @@ public final class RSGroupProtos {
// repeated .hbase.pb.TableName tables = 3;
public static final int TABLES_FIELD_NUMBER = 3;
- private java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName> tables_;
+ private java.util.List<org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName> tables_;
/**
* <code>repeated .hbase.pb.TableName tables = 3;</code>
*/
- public java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName> getTablesList() {
+ public java.util.List<org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName> getTablesList() {
return tables_;
}
/**
* <code>repeated .hbase.pb.TableName tables = 3;</code>
*/
- public java.util.List<? extends org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder>
+ public java.util.List<? extends org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder>
getTablesOrBuilderList() {
return tables_;
}
@@ -298,13 +298,13 @@ public final class RSGroupProtos {
/**
* <code>repeated .hbase.pb.TableName tables = 3;</code>
*/
- public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName getTables(int index) {
+ public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName getTables(int index) {
return tables_.get(index);
}
/**
* <code>repeated .hbase.pb.TableName tables = 3;</code>
*/
- public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder getTablesOrBuilder(
+ public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder getTablesOrBuilder(
int index) {
return tables_.get(index);
}
@@ -1035,22 +1035,22 @@ public final class RSGroupProtos {
}
// repeated .hbase.pb.TableName tables = 3;
- private java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName> tables_ =
+ private java.util.List<org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName> tables_ =
java.util.Collections.emptyList();
private void ensureTablesIsMutable() {
if (!((bitField0_ & 0x00000004) == 0x00000004)) {
- tables_ = new java.util.ArrayList<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName>(tables_);
+ tables_ = new java.util.ArrayList<org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName>(tables_);
bitField0_ |= 0x00000004;
}
}
private com.google.protobuf.RepeatedFieldBuilder<
- org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder> tablesBuilder_;
+ org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder> tablesBuilder_;
/**
* <code>repeated .hbase.pb.TableName tables = 3;</code>
*/
- public java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName> getTablesList() {
+ public java.util.List<org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName> getTablesList() {
if (tablesBuilder_ == null) {
return java.util.Collections.unmodifiableList(tables_);
} else {
@@ -1070,7 +1070,7 @@ public final class RSGroupProtos {
/**
* <code>repeated .hbase.pb.TableName tables = 3;</code>
*/
- public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName getTables(int index) {
+ public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName getTables(int index) {
if (tablesBuilder_ == null) {
return tables_.get(index);
} else {
@@ -1081,7 +1081,7 @@ public final class RSGroupProtos {
* <code>repeated .hbase.pb.TableName tables = 3;</code>
*/
public Builder setTables(
- int index, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName value) {
+ int index, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName value) {
if (tablesBuilder_ == null) {
if (value == null) {
throw new NullPointerException();
@@ -1098,7 +1098,7 @@ public final class RSGroupProtos {
* <code>repeated .hbase.pb.TableName tables = 3;</code>
*/
public Builder setTables(
- int index, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder builderForValue) {
+ int index, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder builderForValue) {
if (tablesBuilder_ == null) {
ensureTablesIsMutable();
tables_.set(index, builderForValue.build());
@@ -1111,7 +1111,7 @@ public final class RSGroupProtos {
/**
* <code>repeated .hbase.pb.TableName tables = 3;</code>
*/
- public Builder addTables(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName value) {
+ public Builder addTables(org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName value) {
if (tablesBuilder_ == null) {
if (value == null) {
throw new NullPointerException();
@@ -1128,7 +1128,7 @@ public final class RSGroupProtos {
* <code>repeated .hbase.pb.TableName tables = 3;</code>
*/
public Builder addTables(
- int index, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName value) {
+ int index, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName value) {
if (tablesBuilder_ == null) {
if (value == null) {
throw new NullPointerException();
@@ -1145,7 +1145,7 @@ public final class RSGroupProtos {
* <code>repeated .hbase.pb.TableName tables = 3;</code>
*/
public Builder addTables(
- org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder builderForValue) {
+ org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder builderForValue) {
if (tablesBuilder_ == null) {
ensureTablesIsMutable();
tables_.add(builderForValue.build());
@@ -1159,7 +1159,7 @@ public final class RSGroupProtos {
* <code>repeated .hbase.pb.TableName tables = 3;</code>
*/
public Builder addTables(
- int index, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder builderForValue) {
+ int index, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder builderForValue) {
if (tablesBuilder_ == null) {
ensureTablesIsMutable();
tables_.add(index, builderForValue.build());
@@ -1173,7 +1173,7 @@ public final class RSGroupProtos {
* <code>repeated .hbase.pb.TableName tables = 3;</code>
*/
public Builder addAllTables(
- java.lang.Iterable<? extends org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName> values) {
+ java.lang.Iterable<? extends org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName> values) {
if (tablesBuilder_ == null) {
ensureTablesIsMutable();
super.addAll(values, tables_);
@@ -1212,14 +1212,14 @@ public final class RSGroupProtos {
/**
* <code>repeated .hbase.pb.TableName tables = 3;</code>
*/
- public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder getTablesBuilder(
+ public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder getTablesBuilder(
int index) {
return getTablesFieldBuilder().getBuilder(index);
}
/**
* <code>repeated .hbase.pb.TableName tables = 3;</code>
*/
- public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder getTablesOrBuilder(
+ public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder getTablesOrBuilder(
int index) {
if (tablesBuilder_ == null) {
return tables_.get(index); } else {
@@ -1229,7 +1229,7 @@ public final class RSGroupProtos {
/**
* <code>repeated .hbase.pb.TableName tables = 3;</code>
*/
- public java.util.List<? extends org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder>
+ public java.util.List<? extends org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder>
getTablesOrBuilderList() {
if (tablesBuilder_ != null) {
return tablesBuilder_.getMessageOrBuilderList();
@@ -1240,31 +1240,31 @@ public final class RSGroupProtos {
/**
* <code>repeated .hbase.pb.TableName tables = 3;</code>
*/
- public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder addTablesBuilder() {
+ public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder addTablesBuilder() {
return getTablesFieldBuilder().addBuilder(
- org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance());
+ org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance());
}
/**
* <code>repeated .hbase.pb.TableName tables = 3;</code>
*/
- public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder addTablesBuilder(
+ public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder addTablesBuilder(
int index) {
return getTablesFieldBuilder().addBuilder(
- index, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance());
+ index, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance());
}
/**
* <code>repeated .hbase.pb.TableName tables = 3;</code>
*/
- public java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder>
+ public java.util.List<org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder>
getTablesBuilderList() {
return getTablesFieldBuilder().getBuilderList();
}
private com.google.protobuf.RepeatedFieldBuilder<
- org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder>
+ org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder>
getTablesFieldBuilder() {
if (tablesBuilder_ == null) {
tablesBuilder_ = new com.google.protobuf.RepeatedFieldBuilder<
- org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder>(
+ org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder>(
tables_,
((bitField0_ & 0x00000004) == 0x00000004),
getParentForChildren(),
@@ -1299,12 +1299,12 @@ public final class RSGroupProtos {
descriptor;
static {
java.lang.String[] descriptorData = {
- "\n\rRSGroup.proto\022\010hbase.pb\032\013HBase.proto\"g" +
- "\n\013RSGroupInfo\022\014\n\004name\030\001 \002(\t\022%\n\007servers\030\004" +
- " \003(\0132\024.hbase.pb.ServerName\022#\n\006tables\030\003 \003" +
- "(\0132\023.hbase.pb.TableNameBC\n*org.apache.ha" +
- "doop.hbase.protobuf.generatedB\rRSGroupPr" +
- "otosH\001\210\001\001\240\001\001"
+ "\n\rRSGroup.proto\022\010hbase.pb\032\013Table.proto\032\013" +
+ "HBase.proto\"g\n\013RSGroupInfo\022\014\n\004name\030\001 \002(\t" +
+ "\022%\n\007servers\030\004 \003(\0132\024.hbase.pb.ServerName\022" +
+ "#\n\006tables\030\003 \003(\0132\023.hbase.pb.TableNameBC\n*" +
+ "org.apache.hadoop.hbase.protobuf.generat" +
+ "edB\rRSGroupProtosH\001\210\001\001\240\001\001"
};
com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() {
@@ -1323,6 +1323,7 @@ public final class RSGroupProtos {
com.google.protobuf.Descriptors.FileDescriptor
.internalBuildGeneratedFileFrom(descriptorData,
new com.google.protobuf.Descriptors.FileDescriptor[] {
+ org.apache.hadoop.hbase.protobuf.generated.TableProtos.getDescriptor(),
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.getDescriptor(),
}, assigner);
}
http://git-wip-us.apache.org/repos/asf/hbase/blob/5e21e306/hbase-rsgroup/src/main/java/org/apache/hadoop/hbase/rsgroup/RSGroupAdminEndpoint.java
----------------------------------------------------------------------
diff --git a/hbase-rsgroup/src/main/java/org/apache/hadoop/hbase/rsgroup/RSGroupAdminEndpoint.java b/hbase-rsgroup/src/main/java/org/apache/hadoop/hbase/rsgroup/RSGroupAdminEndpoint.java
index 91d31d0..7620317 100644
--- a/hbase-rsgroup/src/main/java/org/apache/hadoop/hbase/rsgroup/RSGroupAdminEndpoint.java
+++ b/hbase-rsgroup/src/main/java/org/apache/hadoop/hbase/rsgroup/RSGroupAdminEndpoint.java
@@ -67,6 +67,7 @@ import org.apache.hadoop.hbase.protobuf.generated.RSGroupAdminProtos.MoveTablesR
import org.apache.hadoop.hbase.protobuf.generated.RSGroupAdminProtos.RSGroupAdminService;
import org.apache.hadoop.hbase.protobuf.generated.RSGroupAdminProtos.RemoveRSGroupRequest;
import org.apache.hadoop.hbase.protobuf.generated.RSGroupAdminProtos.RemoveRSGroupResponse;
+import org.apache.hadoop.hbase.protobuf.generated.TableProtos;
import org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription;
@InterfaceAudience.Private
@@ -168,7 +169,7 @@ public class RSGroupAdminEndpoint implements MasterObserver, CoprocessorService
MoveTablesResponse.Builder builder = MoveTablesResponse.newBuilder();
try {
Set<TableName> tables = new HashSet<>(request.getTableNameList().size());
- for (HBaseProtos.TableName tableName : request.getTableNameList()) {
+ for (TableProtos.TableName tableName : request.getTableNameList()) {
tables.add(ProtobufUtil.toTableName(tableName));
}
groupAdminServer.moveTables(tables, request.getTargetGroup());
@@ -257,7 +258,7 @@ public class RSGroupAdminEndpoint implements MasterObserver, CoprocessorService
hostPorts.add(Address.fromParts(el.getHostName(), el.getPort()));
}
Set<TableName> tables = new HashSet<>(request.getTableNameList().size());
- for (HBaseProtos.TableName tableName : request.getTableNameList()) {
+ for (TableProtos.TableName tableName : request.getTableNameList()) {
tables.add(ProtobufUtil.toTableName(tableName));
}
groupAdminServer.moveServersAndTables(hostPorts, tables, request.getTargetGroup());
http://git-wip-us.apache.org/repos/asf/hbase/blob/5e21e306/hbase-rsgroup/src/main/java/org/apache/hadoop/hbase/rsgroup/RSGroupProtobufUtil.java
----------------------------------------------------------------------
diff --git a/hbase-rsgroup/src/main/java/org/apache/hadoop/hbase/rsgroup/RSGroupProtobufUtil.java b/hbase-rsgroup/src/main/java/org/apache/hadoop/hbase/rsgroup/RSGroupProtobufUtil.java
index 24f3976..629a7c4 100644
--- a/hbase-rsgroup/src/main/java/org/apache/hadoop/hbase/rsgroup/RSGroupProtobufUtil.java
+++ b/hbase-rsgroup/src/main/java/org/apache/hadoop/hbase/rsgroup/RSGroupProtobufUtil.java
@@ -27,6 +27,7 @@ import org.apache.hadoop.hbase.net.Address;
import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos;
import org.apache.hadoop.hbase.protobuf.ProtobufUtil;
import org.apache.hadoop.hbase.protobuf.generated.RSGroupProtos;
+import org.apache.hadoop.hbase.protobuf.generated.TableProtos;
@InterfaceAudience.Private
class RSGroupProtobufUtil {
@@ -35,14 +36,14 @@ class RSGroupProtobufUtil {
for(HBaseProtos.ServerName el: proto.getServersList()) {
RSGroupInfo.addServer(Address.fromParts(el.getHostName(), el.getPort()));
}
- for(HBaseProtos.TableName pTableName: proto.getTablesList()) {
+ for(TableProtos.TableName pTableName: proto.getTablesList()) {
RSGroupInfo.addTable(ProtobufUtil.toTableName(pTableName));
}
return RSGroupInfo;
}
static RSGroupProtos.RSGroupInfo toProtoGroupInfo(RSGroupInfo pojo) {
- List<HBaseProtos.TableName> tables = new ArrayList<>(pojo.getTables().size());
+ List<TableProtos.TableName> tables = new ArrayList<>(pojo.getTables().size());
for(TableName arg: pojo.getTables()) {
tables.add(ProtobufUtil.toProtoTableName(arg));
}
http://git-wip-us.apache.org/repos/asf/hbase/blob/5e21e306/hbase-rsgroup/src/main/protobuf/RSGroup.proto
----------------------------------------------------------------------
diff --git a/hbase-rsgroup/src/main/protobuf/RSGroup.proto b/hbase-rsgroup/src/main/protobuf/RSGroup.proto
index 7358941..b88e82c 100644
--- a/hbase-rsgroup/src/main/protobuf/RSGroup.proto
+++ b/hbase-rsgroup/src/main/protobuf/RSGroup.proto
@@ -24,6 +24,7 @@ option java_generic_services = true;
option java_generate_equals_and_hash = true;
option optimize_for = SPEED;
+import "Table.proto";
import "HBase.proto";
message RSGroupInfo {
http://git-wip-us.apache.org/repos/asf/hbase/blob/5e21e306/hbase-rsgroup/src/main/protobuf/RSGroupAdmin.proto
----------------------------------------------------------------------
diff --git a/hbase-rsgroup/src/main/protobuf/RSGroupAdmin.proto b/hbase-rsgroup/src/main/protobuf/RSGroupAdmin.proto
index 0213402..65da657 100644
--- a/hbase-rsgroup/src/main/protobuf/RSGroupAdmin.proto
+++ b/hbase-rsgroup/src/main/protobuf/RSGroupAdmin.proto
@@ -24,6 +24,7 @@ option java_generic_services = true;
option java_generate_equals_and_hash = true;
option optimize_for = SPEED;
+import "Table.proto";
import "HBase.proto";
import "RSGroup.proto";
[5/5] hbase git commit: HBASE-18491 [AMv2] Fail UnassignProcedure if
source Region Server is not online.
Posted by st...@apache.org.
HBASE-18491 [AMv2] Fail UnassignProcedure if source Region Server is not online.
The patch also enables TestServerCrashProcedure.testRecoveryAndDoubleExecutionOnRsWithMeta()
Signed-off-by: Michael Stack <st...@apache.org>
Project: http://git-wip-us.apache.org/repos/asf/hbase/repo
Commit: http://git-wip-us.apache.org/repos/asf/hbase/commit/3f540f38
Tree: http://git-wip-us.apache.org/repos/asf/hbase/tree/3f540f38
Diff: http://git-wip-us.apache.org/repos/asf/hbase/diff/3f540f38
Branch: refs/heads/branch-2
Commit: 3f540f3801c93813a3a018dddb7fdf24d4aff27d
Parents: 5e21e30
Author: Umesh Agashe <ua...@cloudera.com>
Authored: Tue Aug 1 12:36:47 2017 -0700
Committer: Michael Stack <st...@apache.org>
Committed: Tue Aug 1 17:05:22 2017 -0700
----------------------------------------------------------------------
.../hadoop/hbase/master/assignment/UnassignProcedure.java | 7 +++++--
.../hbase/master/procedure/TestServerCrashProcedure.java | 1 -
2 files changed, 5 insertions(+), 3 deletions(-)
----------------------------------------------------------------------
http://git-wip-us.apache.org/repos/asf/hbase/blob/3f540f38/hbase-server/src/main/java/org/apache/hadoop/hbase/master/assignment/UnassignProcedure.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/assignment/UnassignProcedure.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/assignment/UnassignProcedure.java
index 740c7ff..c6b7e4b 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/assignment/UnassignProcedure.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/assignment/UnassignProcedure.java
@@ -161,9 +161,12 @@ public class UnassignProcedure extends RegionTransitionProcedure {
return false;
}
- // if the server is down, mark the operation as complete
+ // if the server is down, mark the operation as failed. region cannot be unassigned
+ // if server is down
if (serverCrashed.get() || !isServerOnline(env, regionNode)) {
- LOG.info("Server already down: " + this + "; " + regionNode.toShortString());
+ LOG.warn("Server already down: " + this + "; " + regionNode.toShortString());
+ setFailure("source region server not online",
+ new ServerCrashException(getProcId(), regionNode.getRegionLocation()));
return false;
}
http://git-wip-us.apache.org/repos/asf/hbase/blob/3f540f38/hbase-server/src/test/java/org/apache/hadoop/hbase/master/procedure/TestServerCrashProcedure.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/master/procedure/TestServerCrashProcedure.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/master/procedure/TestServerCrashProcedure.java
index 0a31a84..9d6e765 100644
--- a/hbase-server/src/test/java/org/apache/hadoop/hbase/master/procedure/TestServerCrashProcedure.java
+++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/master/procedure/TestServerCrashProcedure.java
@@ -87,7 +87,6 @@ public class TestServerCrashProcedure {
testRecoveryAndDoubleExecution(false, false);
}
- @Ignore // HBASE-18366... To be enabled again.
@Test(timeout=60000)
public void testRecoveryAndDoubleExecutionOnRsWithMeta() throws Exception {
testRecoveryAndDoubleExecution(true, true);
[3/5] hbase git commit: HBASE-18431 Mitigate compatibility concerns
between branch-1.3 and branch-1.4
Posted by st...@apache.org.
http://git-wip-us.apache.org/repos/asf/hbase/blob/5e21e306/hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/HBaseProtos.java
----------------------------------------------------------------------
diff --git a/hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/HBaseProtos.java b/hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/HBaseProtos.java
index 72fc7b7..c4aa2b1 100644
--- a/hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/HBaseProtos.java
+++ b/hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/HBaseProtos.java
@@ -266,562 +266,6 @@ public final class HBaseProtos {
// @@protoc_insertion_point(enum_scope:hbase.pb.TimeUnit)
}
- public interface TableNameOrBuilder
- extends com.google.protobuf.MessageOrBuilder {
-
- // required bytes namespace = 1;
- /**
- * <code>required bytes namespace = 1;</code>
- */
- boolean hasNamespace();
- /**
- * <code>required bytes namespace = 1;</code>
- */
- com.google.protobuf.ByteString getNamespace();
-
- // required bytes qualifier = 2;
- /**
- * <code>required bytes qualifier = 2;</code>
- */
- boolean hasQualifier();
- /**
- * <code>required bytes qualifier = 2;</code>
- */
- com.google.protobuf.ByteString getQualifier();
- }
- /**
- * Protobuf type {@code hbase.pb.TableName}
- *
- * <pre>
- **
- * Table Name
- * </pre>
- */
- public static final class TableName extends
- com.google.protobuf.GeneratedMessage
- implements TableNameOrBuilder {
- // Use TableName.newBuilder() to construct.
- private TableName(com.google.protobuf.GeneratedMessage.Builder<?> builder) {
- super(builder);
- this.unknownFields = builder.getUnknownFields();
- }
- private TableName(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
-
- private static final TableName defaultInstance;
- public static TableName getDefaultInstance() {
- return defaultInstance;
- }
-
- public TableName getDefaultInstanceForType() {
- return defaultInstance;
- }
-
- private final com.google.protobuf.UnknownFieldSet unknownFields;
- @java.lang.Override
- public final com.google.protobuf.UnknownFieldSet
- getUnknownFields() {
- return this.unknownFields;
- }
- private TableName(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
- initFields();
- int mutable_bitField0_ = 0;
- com.google.protobuf.UnknownFieldSet.Builder unknownFields =
- com.google.protobuf.UnknownFieldSet.newBuilder();
- try {
- boolean done = false;
- while (!done) {
- int tag = input.readTag();
- switch (tag) {
- case 0:
- done = true;
- break;
- default: {
- if (!parseUnknownField(input, unknownFields,
- extensionRegistry, tag)) {
- done = true;
- }
- break;
- }
- case 10: {
- bitField0_ |= 0x00000001;
- namespace_ = input.readBytes();
- break;
- }
- case 18: {
- bitField0_ |= 0x00000002;
- qualifier_ = input.readBytes();
- break;
- }
- }
- }
- } catch (com.google.protobuf.InvalidProtocolBufferException e) {
- throw e.setUnfinishedMessage(this);
- } catch (java.io.IOException e) {
- throw new com.google.protobuf.InvalidProtocolBufferException(
- e.getMessage()).setUnfinishedMessage(this);
- } finally {
- this.unknownFields = unknownFields.build();
- makeExtensionsImmutable();
- }
- }
- public static final com.google.protobuf.Descriptors.Descriptor
- getDescriptor() {
- return org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.internal_static_hbase_pb_TableName_descriptor;
- }
-
- protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
- internalGetFieldAccessorTable() {
- return org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.internal_static_hbase_pb_TableName_fieldAccessorTable
- .ensureFieldAccessorsInitialized(
- org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.class, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder.class);
- }
-
- public static com.google.protobuf.Parser<TableName> PARSER =
- new com.google.protobuf.AbstractParser<TableName>() {
- public TableName parsePartialFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
- return new TableName(input, extensionRegistry);
- }
- };
-
- @java.lang.Override
- public com.google.protobuf.Parser<TableName> getParserForType() {
- return PARSER;
- }
-
- private int bitField0_;
- // required bytes namespace = 1;
- public static final int NAMESPACE_FIELD_NUMBER = 1;
- private com.google.protobuf.ByteString namespace_;
- /**
- * <code>required bytes namespace = 1;</code>
- */
- public boolean hasNamespace() {
- return ((bitField0_ & 0x00000001) == 0x00000001);
- }
- /**
- * <code>required bytes namespace = 1;</code>
- */
- public com.google.protobuf.ByteString getNamespace() {
- return namespace_;
- }
-
- // required bytes qualifier = 2;
- public static final int QUALIFIER_FIELD_NUMBER = 2;
- private com.google.protobuf.ByteString qualifier_;
- /**
- * <code>required bytes qualifier = 2;</code>
- */
- public boolean hasQualifier() {
- return ((bitField0_ & 0x00000002) == 0x00000002);
- }
- /**
- * <code>required bytes qualifier = 2;</code>
- */
- public com.google.protobuf.ByteString getQualifier() {
- return qualifier_;
- }
-
- private void initFields() {
- namespace_ = com.google.protobuf.ByteString.EMPTY;
- qualifier_ = com.google.protobuf.ByteString.EMPTY;
- }
- private byte memoizedIsInitialized = -1;
- public final boolean isInitialized() {
- byte isInitialized = memoizedIsInitialized;
- if (isInitialized != -1) return isInitialized == 1;
-
- if (!hasNamespace()) {
- memoizedIsInitialized = 0;
- return false;
- }
- if (!hasQualifier()) {
- memoizedIsInitialized = 0;
- return false;
- }
- memoizedIsInitialized = 1;
- return true;
- }
-
- public void writeTo(com.google.protobuf.CodedOutputStream output)
- throws java.io.IOException {
- getSerializedSize();
- if (((bitField0_ & 0x00000001) == 0x00000001)) {
- output.writeBytes(1, namespace_);
- }
- if (((bitField0_ & 0x00000002) == 0x00000002)) {
- output.writeBytes(2, qualifier_);
- }
- getUnknownFields().writeTo(output);
- }
-
- private int memoizedSerializedSize = -1;
- public int getSerializedSize() {
- int size = memoizedSerializedSize;
- if (size != -1) return size;
-
- size = 0;
- if (((bitField0_ & 0x00000001) == 0x00000001)) {
- size += com.google.protobuf.CodedOutputStream
- .computeBytesSize(1, namespace_);
- }
- if (((bitField0_ & 0x00000002) == 0x00000002)) {
- size += com.google.protobuf.CodedOutputStream
- .computeBytesSize(2, qualifier_);
- }
- size += getUnknownFields().getSerializedSize();
- memoizedSerializedSize = size;
- return size;
- }
-
- private static final long serialVersionUID = 0L;
- @java.lang.Override
- protected java.lang.Object writeReplace()
- throws java.io.ObjectStreamException {
- return super.writeReplace();
- }
-
- @java.lang.Override
- public boolean equals(final java.lang.Object obj) {
- if (obj == this) {
- return true;
- }
- if (!(obj instanceof org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName)) {
- return super.equals(obj);
- }
- org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName other = (org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName) obj;
-
- boolean result = true;
- result = result && (hasNamespace() == other.hasNamespace());
- if (hasNamespace()) {
- result = result && getNamespace()
- .equals(other.getNamespace());
- }
- result = result && (hasQualifier() == other.hasQualifier());
- if (hasQualifier()) {
- result = result && getQualifier()
- .equals(other.getQualifier());
- }
- result = result &&
- getUnknownFields().equals(other.getUnknownFields());
- return result;
- }
-
- private int memoizedHashCode = 0;
- @java.lang.Override
- public int hashCode() {
- if (memoizedHashCode != 0) {
- return memoizedHashCode;
- }
- int hash = 41;
- hash = (19 * hash) + getDescriptorForType().hashCode();
- if (hasNamespace()) {
- hash = (37 * hash) + NAMESPACE_FIELD_NUMBER;
- hash = (53 * hash) + getNamespace().hashCode();
- }
- if (hasQualifier()) {
- hash = (37 * hash) + QUALIFIER_FIELD_NUMBER;
- hash = (53 * hash) + getQualifier().hashCode();
- }
- hash = (29 * hash) + getUnknownFields().hashCode();
- memoizedHashCode = hash;
- return hash;
- }
-
- public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName parseFrom(
- com.google.protobuf.ByteString data)
- throws com.google.protobuf.InvalidProtocolBufferException {
- return PARSER.parseFrom(data);
- }
- public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName parseFrom(
- com.google.protobuf.ByteString data,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
- return PARSER.parseFrom(data, extensionRegistry);
- }
- public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName parseFrom(byte[] data)
- throws com.google.protobuf.InvalidProtocolBufferException {
- return PARSER.parseFrom(data);
- }
- public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName parseFrom(
- byte[] data,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
- return PARSER.parseFrom(data, extensionRegistry);
- }
- public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName parseFrom(java.io.InputStream input)
- throws java.io.IOException {
- return PARSER.parseFrom(input);
- }
- public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName parseFrom(
- java.io.InputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws java.io.IOException {
- return PARSER.parseFrom(input, extensionRegistry);
- }
- public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName parseDelimitedFrom(java.io.InputStream input)
- throws java.io.IOException {
- return PARSER.parseDelimitedFrom(input);
- }
- public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName parseDelimitedFrom(
- java.io.InputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws java.io.IOException {
- return PARSER.parseDelimitedFrom(input, extensionRegistry);
- }
- public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName parseFrom(
- com.google.protobuf.CodedInputStream input)
- throws java.io.IOException {
- return PARSER.parseFrom(input);
- }
- public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName parseFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws java.io.IOException {
- return PARSER.parseFrom(input, extensionRegistry);
- }
-
- public static Builder newBuilder() { return Builder.create(); }
- public Builder newBuilderForType() { return newBuilder(); }
- public static Builder newBuilder(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName prototype) {
- return newBuilder().mergeFrom(prototype);
- }
- public Builder toBuilder() { return newBuilder(this); }
-
- @java.lang.Override
- protected Builder newBuilderForType(
- com.google.protobuf.GeneratedMessage.BuilderParent parent) {
- Builder builder = new Builder(parent);
- return builder;
- }
- /**
- * Protobuf type {@code hbase.pb.TableName}
- *
- * <pre>
- **
- * Table Name
- * </pre>
- */
- public static final class Builder extends
- com.google.protobuf.GeneratedMessage.Builder<Builder>
- implements org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder {
- public static final com.google.protobuf.Descriptors.Descriptor
- getDescriptor() {
- return org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.internal_static_hbase_pb_TableName_descriptor;
- }
-
- protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
- internalGetFieldAccessorTable() {
- return org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.internal_static_hbase_pb_TableName_fieldAccessorTable
- .ensureFieldAccessorsInitialized(
- org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.class, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder.class);
- }
-
- // Construct using org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.newBuilder()
- private Builder() {
- maybeForceBuilderInitialization();
- }
-
- private Builder(
- com.google.protobuf.GeneratedMessage.BuilderParent parent) {
- super(parent);
- maybeForceBuilderInitialization();
- }
- private void maybeForceBuilderInitialization() {
- if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
- }
- }
- private static Builder create() {
- return new Builder();
- }
-
- public Builder clear() {
- super.clear();
- namespace_ = com.google.protobuf.ByteString.EMPTY;
- bitField0_ = (bitField0_ & ~0x00000001);
- qualifier_ = com.google.protobuf.ByteString.EMPTY;
- bitField0_ = (bitField0_ & ~0x00000002);
- return this;
- }
-
- public Builder clone() {
- return create().mergeFrom(buildPartial());
- }
-
- public com.google.protobuf.Descriptors.Descriptor
- getDescriptorForType() {
- return org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.internal_static_hbase_pb_TableName_descriptor;
- }
-
- public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName getDefaultInstanceForType() {
- return org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance();
- }
-
- public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName build() {
- org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName result = buildPartial();
- if (!result.isInitialized()) {
- throw newUninitializedMessageException(result);
- }
- return result;
- }
-
- public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName buildPartial() {
- org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName result = new org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName(this);
- int from_bitField0_ = bitField0_;
- int to_bitField0_ = 0;
- if (((from_bitField0_ & 0x00000001) == 0x00000001)) {
- to_bitField0_ |= 0x00000001;
- }
- result.namespace_ = namespace_;
- if (((from_bitField0_ & 0x00000002) == 0x00000002)) {
- to_bitField0_ |= 0x00000002;
- }
- result.qualifier_ = qualifier_;
- result.bitField0_ = to_bitField0_;
- onBuilt();
- return result;
- }
-
- public Builder mergeFrom(com.google.protobuf.Message other) {
- if (other instanceof org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName) {
- return mergeFrom((org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName)other);
- } else {
- super.mergeFrom(other);
- return this;
- }
- }
-
- public Builder mergeFrom(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName other) {
- if (other == org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance()) return this;
- if (other.hasNamespace()) {
- setNamespace(other.getNamespace());
- }
- if (other.hasQualifier()) {
- setQualifier(other.getQualifier());
- }
- this.mergeUnknownFields(other.getUnknownFields());
- return this;
- }
-
- public final boolean isInitialized() {
- if (!hasNamespace()) {
-
- return false;
- }
- if (!hasQualifier()) {
-
- return false;
- }
- return true;
- }
-
- public Builder mergeFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws java.io.IOException {
- org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName parsedMessage = null;
- try {
- parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
- } catch (com.google.protobuf.InvalidProtocolBufferException e) {
- parsedMessage = (org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName) e.getUnfinishedMessage();
- throw e;
- } finally {
- if (parsedMessage != null) {
- mergeFrom(parsedMessage);
- }
- }
- return this;
- }
- private int bitField0_;
-
- // required bytes namespace = 1;
- private com.google.protobuf.ByteString namespace_ = com.google.protobuf.ByteString.EMPTY;
- /**
- * <code>required bytes namespace = 1;</code>
- */
- public boolean hasNamespace() {
- return ((bitField0_ & 0x00000001) == 0x00000001);
- }
- /**
- * <code>required bytes namespace = 1;</code>
- */
- public com.google.protobuf.ByteString getNamespace() {
- return namespace_;
- }
- /**
- * <code>required bytes namespace = 1;</code>
- */
- public Builder setNamespace(com.google.protobuf.ByteString value) {
- if (value == null) {
- throw new NullPointerException();
- }
- bitField0_ |= 0x00000001;
- namespace_ = value;
- onChanged();
- return this;
- }
- /**
- * <code>required bytes namespace = 1;</code>
- */
- public Builder clearNamespace() {
- bitField0_ = (bitField0_ & ~0x00000001);
- namespace_ = getDefaultInstance().getNamespace();
- onChanged();
- return this;
- }
-
- // required bytes qualifier = 2;
- private com.google.protobuf.ByteString qualifier_ = com.google.protobuf.ByteString.EMPTY;
- /**
- * <code>required bytes qualifier = 2;</code>
- */
- public boolean hasQualifier() {
- return ((bitField0_ & 0x00000002) == 0x00000002);
- }
- /**
- * <code>required bytes qualifier = 2;</code>
- */
- public com.google.protobuf.ByteString getQualifier() {
- return qualifier_;
- }
- /**
- * <code>required bytes qualifier = 2;</code>
- */
- public Builder setQualifier(com.google.protobuf.ByteString value) {
- if (value == null) {
- throw new NullPointerException();
- }
- bitField0_ |= 0x00000002;
- qualifier_ = value;
- onChanged();
- return this;
- }
- /**
- * <code>required bytes qualifier = 2;</code>
- */
- public Builder clearQualifier() {
- bitField0_ = (bitField0_ & ~0x00000002);
- qualifier_ = getDefaultInstance().getQualifier();
- onChanged();
- return this;
- }
-
- // @@protoc_insertion_point(builder_scope:hbase.pb.TableName)
- }
-
- static {
- defaultInstance = new TableName(true);
- defaultInstance.initFields();
- }
-
- // @@protoc_insertion_point(class_scope:hbase.pb.TableName)
- }
-
public interface TableSchemaOrBuilder
extends com.google.protobuf.MessageOrBuilder {
@@ -833,11 +277,11 @@ public final class HBaseProtos {
/**
* <code>optional .hbase.pb.TableName table_name = 1;</code>
*/
- org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName getTableName();
+ org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName getTableName();
/**
* <code>optional .hbase.pb.TableName table_name = 1;</code>
*/
- org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder();
+ org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder getTableNameOrBuilder();
// repeated .hbase.pb.BytesBytesPair attributes = 2;
/**
@@ -972,11 +416,11 @@ public final class HBaseProtos {
break;
}
case 10: {
- org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder subBuilder = null;
+ org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder subBuilder = null;
if (((bitField0_ & 0x00000001) == 0x00000001)) {
subBuilder = tableName_.toBuilder();
}
- tableName_ = input.readMessage(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.PARSER, extensionRegistry);
+ tableName_ = input.readMessage(org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.PARSER, extensionRegistry);
if (subBuilder != null) {
subBuilder.mergeFrom(tableName_);
tableName_ = subBuilder.buildPartial();
@@ -1059,7 +503,7 @@ public final class HBaseProtos {
private int bitField0_;
// optional .hbase.pb.TableName table_name = 1;
public static final int TABLE_NAME_FIELD_NUMBER = 1;
- private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName tableName_;
+ private org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName tableName_;
/**
* <code>optional .hbase.pb.TableName table_name = 1;</code>
*/
@@ -1069,13 +513,13 @@ public final class HBaseProtos {
/**
* <code>optional .hbase.pb.TableName table_name = 1;</code>
*/
- public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName getTableName() {
+ public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName getTableName() {
return tableName_;
}
/**
* <code>optional .hbase.pb.TableName table_name = 1;</code>
*/
- public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder() {
+ public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder getTableNameOrBuilder() {
return tableName_;
}
@@ -1188,7 +632,7 @@ public final class HBaseProtos {
}
private void initFields() {
- tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance();
+ tableName_ = org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance();
attributes_ = java.util.Collections.emptyList();
columnFamilies_ = java.util.Collections.emptyList();
configuration_ = java.util.Collections.emptyList();
@@ -1449,7 +893,7 @@ public final class HBaseProtos {
public Builder clear() {
super.clear();
if (tableNameBuilder_ == null) {
- tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance();
+ tableName_ = org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance();
} else {
tableNameBuilder_.clear();
}
@@ -1684,9 +1128,9 @@ public final class HBaseProtos {
private int bitField0_;
// optional .hbase.pb.TableName table_name = 1;
- private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance();
+ private org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName tableName_ = org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance();
private com.google.protobuf.SingleFieldBuilder<
- org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder> tableNameBuilder_;
+ org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder> tableNameBuilder_;
/**
* <code>optional .hbase.pb.TableName table_name = 1;</code>
*/
@@ -1696,7 +1140,7 @@ public final class HBaseProtos {
/**
* <code>optional .hbase.pb.TableName table_name = 1;</code>
*/
- public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName getTableName() {
+ public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName getTableName() {
if (tableNameBuilder_ == null) {
return tableName_;
} else {
@@ -1706,7 +1150,7 @@ public final class HBaseProtos {
/**
* <code>optional .hbase.pb.TableName table_name = 1;</code>
*/
- public Builder setTableName(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName value) {
+ public Builder setTableName(org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName value) {
if (tableNameBuilder_ == null) {
if (value == null) {
throw new NullPointerException();
@@ -1723,7 +1167,7 @@ public final class HBaseProtos {
* <code>optional .hbase.pb.TableName table_name = 1;</code>
*/
public Builder setTableName(
- org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder builderForValue) {
+ org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder builderForValue) {
if (tableNameBuilder_ == null) {
tableName_ = builderForValue.build();
onChanged();
@@ -1736,12 +1180,12 @@ public final class HBaseProtos {
/**
* <code>optional .hbase.pb.TableName table_name = 1;</code>
*/
- public Builder mergeTableName(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName value) {
+ public Builder mergeTableName(org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName value) {
if (tableNameBuilder_ == null) {
if (((bitField0_ & 0x00000001) == 0x00000001) &&
- tableName_ != org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance()) {
+ tableName_ != org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance()) {
tableName_ =
- org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.newBuilder(tableName_).mergeFrom(value).buildPartial();
+ org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.newBuilder(tableName_).mergeFrom(value).buildPartial();
} else {
tableName_ = value;
}
@@ -1757,7 +1201,7 @@ public final class HBaseProtos {
*/
public Builder clearTableName() {
if (tableNameBuilder_ == null) {
- tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance();
+ tableName_ = org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance();
onChanged();
} else {
tableNameBuilder_.clear();
@@ -1768,7 +1212,7 @@ public final class HBaseProtos {
/**
* <code>optional .hbase.pb.TableName table_name = 1;</code>
*/
- public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder getTableNameBuilder() {
+ public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder getTableNameBuilder() {
bitField0_ |= 0x00000001;
onChanged();
return getTableNameFieldBuilder().getBuilder();
@@ -1776,7 +1220,7 @@ public final class HBaseProtos {
/**
* <code>optional .hbase.pb.TableName table_name = 1;</code>
*/
- public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder() {
+ public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder getTableNameOrBuilder() {
if (tableNameBuilder_ != null) {
return tableNameBuilder_.getMessageOrBuilder();
} else {
@@ -1787,11 +1231,11 @@ public final class HBaseProtos {
* <code>optional .hbase.pb.TableName table_name = 1;</code>
*/
private com.google.protobuf.SingleFieldBuilder<
- org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder>
+ org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder>
getTableNameFieldBuilder() {
if (tableNameBuilder_ == null) {
tableNameBuilder_ = new com.google.protobuf.SingleFieldBuilder<
- org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder>(
+ org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder>(
tableName_,
getParentForChildren(),
isClean());
@@ -4364,11 +3808,11 @@ public final class HBaseProtos {
/**
* <code>required .hbase.pb.TableName table_name = 2;</code>
*/
- org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName getTableName();
+ org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName getTableName();
/**
* <code>required .hbase.pb.TableName table_name = 2;</code>
*/
- org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder();
+ org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder getTableNameOrBuilder();
// optional bytes start_key = 3;
/**
@@ -4482,11 +3926,11 @@ public final class HBaseProtos {
break;
}
case 18: {
- org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder subBuilder = null;
+ org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder subBuilder = null;
if (((bitField0_ & 0x00000002) == 0x00000002)) {
subBuilder = tableName_.toBuilder();
}
- tableName_ = input.readMessage(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.PARSER, extensionRegistry);
+ tableName_ = input.readMessage(org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.PARSER, extensionRegistry);
if (subBuilder != null) {
subBuilder.mergeFrom(tableName_);
tableName_ = subBuilder.buildPartial();
@@ -4577,7 +4021,7 @@ public final class HBaseProtos {
// required .hbase.pb.TableName table_name = 2;
public static final int TABLE_NAME_FIELD_NUMBER = 2;
- private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName tableName_;
+ private org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName tableName_;
/**
* <code>required .hbase.pb.TableName table_name = 2;</code>
*/
@@ -4587,13 +4031,13 @@ public final class HBaseProtos {
/**
* <code>required .hbase.pb.TableName table_name = 2;</code>
*/
- public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName getTableName() {
+ public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName getTableName() {
return tableName_;
}
/**
* <code>required .hbase.pb.TableName table_name = 2;</code>
*/
- public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder() {
+ public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder getTableNameOrBuilder() {
return tableName_;
}
@@ -4679,7 +4123,7 @@ public final class HBaseProtos {
private void initFields() {
regionId_ = 0L;
- tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance();
+ tableName_ = org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance();
startKey_ = com.google.protobuf.ByteString.EMPTY;
endKey_ = com.google.protobuf.ByteString.EMPTY;
offline_ = false;
@@ -4985,7 +4429,7 @@ public final class HBaseProtos {
regionId_ = 0L;
bitField0_ = (bitField0_ & ~0x00000001);
if (tableNameBuilder_ == null) {
- tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance();
+ tableName_ = org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance();
} else {
tableNameBuilder_.clear();
}
@@ -5170,9 +4614,9 @@ public final class HBaseProtos {
}
// required .hbase.pb.TableName table_name = 2;
- private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance();
+ private org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName tableName_ = org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance();
private com.google.protobuf.SingleFieldBuilder<
- org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder> tableNameBuilder_;
+ org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder> tableNameBuilder_;
/**
* <code>required .hbase.pb.TableName table_name = 2;</code>
*/
@@ -5182,7 +4626,7 @@ public final class HBaseProtos {
/**
* <code>required .hbase.pb.TableName table_name = 2;</code>
*/
- public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName getTableName() {
+ public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName getTableName() {
if (tableNameBuilder_ == null) {
return tableName_;
} else {
@@ -5192,7 +4636,7 @@ public final class HBaseProtos {
/**
* <code>required .hbase.pb.TableName table_name = 2;</code>
*/
- public Builder setTableName(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName value) {
+ public Builder setTableName(org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName value) {
if (tableNameBuilder_ == null) {
if (value == null) {
throw new NullPointerException();
@@ -5209,7 +4653,7 @@ public final class HBaseProtos {
* <code>required .hbase.pb.TableName table_name = 2;</code>
*/
public Builder setTableName(
- org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder builderForValue) {
+ org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder builderForValue) {
if (tableNameBuilder_ == null) {
tableName_ = builderForValue.build();
onChanged();
@@ -5222,12 +4666,12 @@ public final class HBaseProtos {
/**
* <code>required .hbase.pb.TableName table_name = 2;</code>
*/
- public Builder mergeTableName(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName value) {
+ public Builder mergeTableName(org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName value) {
if (tableNameBuilder_ == null) {
if (((bitField0_ & 0x00000002) == 0x00000002) &&
- tableName_ != org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance()) {
+ tableName_ != org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance()) {
tableName_ =
- org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.newBuilder(tableName_).mergeFrom(value).buildPartial();
+ org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.newBuilder(tableName_).mergeFrom(value).buildPartial();
} else {
tableName_ = value;
}
@@ -5243,7 +4687,7 @@ public final class HBaseProtos {
*/
public Builder clearTableName() {
if (tableNameBuilder_ == null) {
- tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance();
+ tableName_ = org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance();
onChanged();
} else {
tableNameBuilder_.clear();
@@ -5254,7 +4698,7 @@ public final class HBaseProtos {
/**
* <code>required .hbase.pb.TableName table_name = 2;</code>
*/
- public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder getTableNameBuilder() {
+ public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder getTableNameBuilder() {
bitField0_ |= 0x00000002;
onChanged();
return getTableNameFieldBuilder().getBuilder();
@@ -5262,7 +4706,7 @@ public final class HBaseProtos {
/**
* <code>required .hbase.pb.TableName table_name = 2;</code>
*/
- public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder() {
+ public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder getTableNameOrBuilder() {
if (tableNameBuilder_ != null) {
return tableNameBuilder_.getMessageOrBuilder();
} else {
@@ -5273,11 +4717,11 @@ public final class HBaseProtos {
* <code>required .hbase.pb.TableName table_name = 2;</code>
*/
private com.google.protobuf.SingleFieldBuilder<
- org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder>
+ org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder>
getTableNameFieldBuilder() {
if (tableNameBuilder_ == null) {
tableNameBuilder_ = new com.google.protobuf.SingleFieldBuilder<
- org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder>(
+ org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder>(
tableName_,
getParentForChildren(),
isClean());
@@ -19529,11 +18973,6 @@ public final class HBaseProtos {
}
private static com.google.protobuf.Descriptors.Descriptor
- internal_static_hbase_pb_TableName_descriptor;
- private static
- com.google.protobuf.GeneratedMessage.FieldAccessorTable
- internal_static_hbase_pb_TableName_fieldAccessorTable;
- private static com.google.protobuf.Descriptors.Descriptor
internal_static_hbase_pb_TableSchema_descriptor;
private static
com.google.protobuf.GeneratedMessage.FieldAccessorTable
@@ -19662,220 +19101,213 @@ public final class HBaseProtos {
descriptor;
static {
java.lang.String[] descriptorData = {
- "\n\013HBase.proto\022\010hbase.pb\"1\n\tTableName\022\021\n\t" +
- "namespace\030\001 \002(\014\022\021\n\tqualifier\030\002 \002(\014\"\314\001\n\013T" +
- "ableSchema\022\'\n\ntable_name\030\001 \001(\0132\023.hbase.p" +
- "b.TableName\022,\n\nattributes\030\002 \003(\0132\030.hbase." +
- "pb.BytesBytesPair\0225\n\017column_families\030\003 \003" +
- "(\0132\034.hbase.pb.ColumnFamilySchema\022/\n\rconf" +
- "iguration\030\004 \003(\0132\030.hbase.pb.NameStringPai" +
- "r\"x\n\nTableState\022)\n\005state\030\001 \002(\0162\032.hbase.p" +
- "b.TableState.State\"?\n\005State\022\013\n\007ENABLED\020\000" +
- "\022\014\n\010DISABLED\020\001\022\r\n\tDISABLING\020\002\022\014\n\010ENABLIN",
- "G\020\003\"\201\001\n\022ColumnFamilySchema\022\014\n\004name\030\001 \002(\014" +
- "\022,\n\nattributes\030\002 \003(\0132\030.hbase.pb.BytesByt" +
- "esPair\022/\n\rconfiguration\030\003 \003(\0132\030.hbase.pb" +
- ".NameStringPair\"\243\001\n\nRegionInfo\022\021\n\tregion" +
- "_id\030\001 \002(\004\022\'\n\ntable_name\030\002 \002(\0132\023.hbase.pb" +
- ".TableName\022\021\n\tstart_key\030\003 \001(\014\022\017\n\007end_key" +
- "\030\004 \001(\014\022\017\n\007offline\030\005 \001(\010\022\r\n\005split\030\006 \001(\010\022\025" +
- "\n\nreplica_id\030\007 \001(\005:\0010\":\n\014FavoredNodes\022*\n" +
- "\014favored_node\030\001 \003(\0132\024.hbase.pb.ServerNam" +
- "e\"\236\001\n\017RegionSpecifier\022;\n\004type\030\001 \002(\0162-.hb",
- "ase.pb.RegionSpecifier.RegionSpecifierTy" +
- "pe\022\r\n\005value\030\002 \002(\014\"?\n\023RegionSpecifierType" +
- "\022\017\n\013REGION_NAME\020\001\022\027\n\023ENCODED_REGION_NAME" +
- "\020\002\"%\n\tTimeRange\022\014\n\004from\030\001 \001(\004\022\n\n\002to\030\002 \001(" +
- "\004\"W\n\025ColumnFamilyTimeRange\022\025\n\rcolumn_fam" +
- "ily\030\001 \002(\014\022\'\n\ntime_range\030\002 \002(\0132\023.hbase.pb" +
- ".TimeRange\"A\n\nServerName\022\021\n\thost_name\030\001 " +
- "\002(\t\022\014\n\004port\030\002 \001(\r\022\022\n\nstart_code\030\003 \001(\004\"\033\n" +
- "\013Coprocessor\022\014\n\004name\030\001 \002(\t\"-\n\016NameString" +
- "Pair\022\014\n\004name\030\001 \002(\t\022\r\n\005value\030\002 \002(\t\",\n\rNam",
- "eBytesPair\022\014\n\004name\030\001 \002(\t\022\r\n\005value\030\002 \001(\014\"" +
- "/\n\016BytesBytesPair\022\r\n\005first\030\001 \002(\014\022\016\n\006seco" +
- "nd\030\002 \002(\014\",\n\rNameInt64Pair\022\014\n\004name\030\001 \001(\t\022" +
- "\r\n\005value\030\002 \001(\003\"\325\001\n\023SnapshotDescription\022\014" +
- "\n\004name\030\001 \002(\t\022\r\n\005table\030\002 \001(\t\022\030\n\rcreation_" +
- "time\030\003 \001(\003:\0010\0227\n\004type\030\004 \001(\0162\".hbase.pb.S" +
- "napshotDescription.Type:\005FLUSH\022\017\n\007versio" +
- "n\030\005 \001(\005\022\r\n\005owner\030\006 \001(\t\".\n\004Type\022\014\n\010DISABL" +
- "ED\020\000\022\t\n\005FLUSH\020\001\022\r\n\tSKIPFLUSH\020\002\"\206\001\n\024Proce" +
- "dureDescription\022\021\n\tsignature\030\001 \002(\t\022\020\n\010in",
- "stance\030\002 \001(\t\022\030\n\rcreation_time\030\003 \001(\003:\0010\022/" +
- "\n\rconfiguration\030\004 \003(\0132\030.hbase.pb.NameStr" +
- "ingPair\"\n\n\010EmptyMsg\"\033\n\007LongMsg\022\020\n\010long_m" +
- "sg\030\001 \002(\003\"\037\n\tDoubleMsg\022\022\n\ndouble_msg\030\001 \002(" +
- "\001\"\'\n\rBigDecimalMsg\022\026\n\016bigdecimal_msg\030\001 \002" +
- "(\014\"5\n\004UUID\022\026\n\016least_sig_bits\030\001 \002(\004\022\025\n\rmo" +
- "st_sig_bits\030\002 \002(\004\"T\n\023NamespaceDescriptor" +
- "\022\014\n\004name\030\001 \002(\014\022/\n\rconfiguration\030\002 \003(\0132\030." +
- "hbase.pb.NameStringPair\"\235\001\n\013VersionInfo\022" +
- "\017\n\007version\030\001 \002(\t\022\013\n\003url\030\002 \002(\t\022\020\n\010revisio",
- "n\030\003 \002(\t\022\014\n\004user\030\004 \002(\t\022\014\n\004date\030\005 \002(\t\022\024\n\014s" +
- "rc_checksum\030\006 \002(\t\022\025\n\rversion_major\030\007 \001(\r" +
- "\022\025\n\rversion_minor\030\010 \001(\r\"Q\n\020RegionServerI" +
- "nfo\022\020\n\010infoPort\030\001 \001(\005\022+\n\014version_info\030\002 " +
- "\001(\0132\025.hbase.pb.VersionInfo*r\n\013CompareTyp" +
- "e\022\010\n\004LESS\020\000\022\021\n\rLESS_OR_EQUAL\020\001\022\t\n\005EQUAL\020" +
- "\002\022\r\n\tNOT_EQUAL\020\003\022\024\n\020GREATER_OR_EQUAL\020\004\022\013" +
- "\n\007GREATER\020\005\022\t\n\005NO_OP\020\006*n\n\010TimeUnit\022\017\n\013NA" +
- "NOSECONDS\020\001\022\020\n\014MICROSECONDS\020\002\022\020\n\014MILLISE" +
- "CONDS\020\003\022\013\n\007SECONDS\020\004\022\013\n\007MINUTES\020\005\022\t\n\005HOU",
- "RS\020\006\022\010\n\004DAYS\020\007B>\n*org.apache.hadoop.hbas" +
- "e.protobuf.generatedB\013HBaseProtosH\001\240\001\001"
+ "\n\013HBase.proto\022\010hbase.pb\032\013Table.proto\"\314\001\n" +
+ "\013TableSchema\022\'\n\ntable_name\030\001 \001(\0132\023.hbase" +
+ ".pb.TableName\022,\n\nattributes\030\002 \003(\0132\030.hbas" +
+ "e.pb.BytesBytesPair\0225\n\017column_families\030\003" +
+ " \003(\0132\034.hbase.pb.ColumnFamilySchema\022/\n\rco" +
+ "nfiguration\030\004 \003(\0132\030.hbase.pb.NameStringP" +
+ "air\"x\n\nTableState\022)\n\005state\030\001 \002(\0162\032.hbase" +
+ ".pb.TableState.State\"?\n\005State\022\013\n\007ENABLED" +
+ "\020\000\022\014\n\010DISABLED\020\001\022\r\n\tDISABLING\020\002\022\014\n\010ENABL" +
+ "ING\020\003\"\201\001\n\022ColumnFamilySchema\022\014\n\004name\030\001 \002",
+ "(\014\022,\n\nattributes\030\002 \003(\0132\030.hbase.pb.BytesB" +
+ "ytesPair\022/\n\rconfiguration\030\003 \003(\0132\030.hbase." +
+ "pb.NameStringPair\"\243\001\n\nRegionInfo\022\021\n\tregi" +
+ "on_id\030\001 \002(\004\022\'\n\ntable_name\030\002 \002(\0132\023.hbase." +
+ "pb.TableName\022\021\n\tstart_key\030\003 \001(\014\022\017\n\007end_k" +
+ "ey\030\004 \001(\014\022\017\n\007offline\030\005 \001(\010\022\r\n\005split\030\006 \001(\010" +
+ "\022\025\n\nreplica_id\030\007 \001(\005:\0010\":\n\014FavoredNodes\022" +
+ "*\n\014favored_node\030\001 \003(\0132\024.hbase.pb.ServerN" +
+ "ame\"\236\001\n\017RegionSpecifier\022;\n\004type\030\001 \002(\0162-." +
+ "hbase.pb.RegionSpecifier.RegionSpecifier",
+ "Type\022\r\n\005value\030\002 \002(\014\"?\n\023RegionSpecifierTy" +
+ "pe\022\017\n\013REGION_NAME\020\001\022\027\n\023ENCODED_REGION_NA" +
+ "ME\020\002\"%\n\tTimeRange\022\014\n\004from\030\001 \001(\004\022\n\n\002to\030\002 " +
+ "\001(\004\"W\n\025ColumnFamilyTimeRange\022\025\n\rcolumn_f" +
+ "amily\030\001 \002(\014\022\'\n\ntime_range\030\002 \002(\0132\023.hbase." +
+ "pb.TimeRange\"A\n\nServerName\022\021\n\thost_name\030" +
+ "\001 \002(\t\022\014\n\004port\030\002 \001(\r\022\022\n\nstart_code\030\003 \001(\004\"" +
+ "\033\n\013Coprocessor\022\014\n\004name\030\001 \002(\t\"-\n\016NameStri" +
+ "ngPair\022\014\n\004name\030\001 \002(\t\022\r\n\005value\030\002 \002(\t\",\n\rN" +
+ "ameBytesPair\022\014\n\004name\030\001 \002(\t\022\r\n\005value\030\002 \001(",
+ "\014\"/\n\016BytesBytesPair\022\r\n\005first\030\001 \002(\014\022\016\n\006se" +
+ "cond\030\002 \002(\014\",\n\rNameInt64Pair\022\014\n\004name\030\001 \001(" +
+ "\t\022\r\n\005value\030\002 \001(\003\"\325\001\n\023SnapshotDescription" +
+ "\022\014\n\004name\030\001 \002(\t\022\r\n\005table\030\002 \001(\t\022\030\n\rcreatio" +
+ "n_time\030\003 \001(\003:\0010\0227\n\004type\030\004 \001(\0162\".hbase.pb" +
+ ".SnapshotDescription.Type:\005FLUSH\022\017\n\007vers" +
+ "ion\030\005 \001(\005\022\r\n\005owner\030\006 \001(\t\".\n\004Type\022\014\n\010DISA" +
+ "BLED\020\000\022\t\n\005FLUSH\020\001\022\r\n\tSKIPFLUSH\020\002\"\206\001\n\024Pro" +
+ "cedureDescription\022\021\n\tsignature\030\001 \002(\t\022\020\n\010" +
+ "instance\030\002 \001(\t\022\030\n\rcreation_time\030\003 \001(\003:\0010",
+ "\022/\n\rconfiguration\030\004 \003(\0132\030.hbase.pb.NameS" +
+ "tringPair\"\n\n\010EmptyMsg\"\033\n\007LongMsg\022\020\n\010long" +
+ "_msg\030\001 \002(\003\"\037\n\tDoubleMsg\022\022\n\ndouble_msg\030\001 " +
+ "\002(\001\"\'\n\rBigDecimalMsg\022\026\n\016bigdecimal_msg\030\001" +
+ " \002(\014\"5\n\004UUID\022\026\n\016least_sig_bits\030\001 \002(\004\022\025\n\r" +
+ "most_sig_bits\030\002 \002(\004\"T\n\023NamespaceDescript" +
+ "or\022\014\n\004name\030\001 \002(\014\022/\n\rconfiguration\030\002 \003(\0132" +
+ "\030.hbase.pb.NameStringPair\"\235\001\n\013VersionInf" +
+ "o\022\017\n\007version\030\001 \002(\t\022\013\n\003url\030\002 \002(\t\022\020\n\010revis" +
+ "ion\030\003 \002(\t\022\014\n\004user\030\004 \002(\t\022\014\n\004date\030\005 \002(\t\022\024\n",
+ "\014src_checksum\030\006 \002(\t\022\025\n\rversion_major\030\007 \001" +
+ "(\r\022\025\n\rversion_minor\030\010 \001(\r\"Q\n\020RegionServe" +
+ "rInfo\022\020\n\010infoPort\030\001 \001(\005\022+\n\014version_info\030" +
+ "\002 \001(\0132\025.hbase.pb.VersionInfo*r\n\013CompareT" +
+ "ype\022\010\n\004LESS\020\000\022\021\n\rLESS_OR_EQUAL\020\001\022\t\n\005EQUA" +
+ "L\020\002\022\r\n\tNOT_EQUAL\020\003\022\024\n\020GREATER_OR_EQUAL\020\004" +
+ "\022\013\n\007GREATER\020\005\022\t\n\005NO_OP\020\006*n\n\010TimeUnit\022\017\n\013" +
+ "NANOSECONDS\020\001\022\020\n\014MICROSECONDS\020\002\022\020\n\014MILLI" +
+ "SECONDS\020\003\022\013\n\007SECONDS\020\004\022\013\n\007MINUTES\020\005\022\t\n\005H" +
+ "OURS\020\006\022\010\n\004DAYS\020\007B>\n*org.apache.hadoop.hb",
+ "ase.protobuf.generatedB\013HBaseProtosH\001\240\001\001"
};
com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() {
public com.google.protobuf.ExtensionRegistry assignDescriptors(
com.google.protobuf.Descriptors.FileDescriptor root) {
descriptor = root;
- internal_static_hbase_pb_TableName_descriptor =
- getDescriptor().getMessageTypes().get(0);
- internal_static_hbase_pb_TableName_fieldAccessorTable = new
- com.google.protobuf.GeneratedMessage.FieldAccessorTable(
- internal_static_hbase_pb_TableName_descriptor,
- new java.lang.String[] { "Namespace", "Qualifier", });
internal_static_hbase_pb_TableSchema_descriptor =
- getDescriptor().getMessageTypes().get(1);
+ getDescriptor().getMessageTypes().get(0);
internal_static_hbase_pb_TableSchema_fieldAccessorTable = new
com.google.protobuf.GeneratedMessage.FieldAccessorTable(
internal_static_hbase_pb_TableSchema_descriptor,
new java.lang.String[] { "TableName", "Attributes", "ColumnFamilies", "Configuration", });
internal_static_hbase_pb_TableState_descriptor =
- getDescriptor().getMessageTypes().get(2);
+ getDescriptor().getMessageTypes().get(1);
internal_static_hbase_pb_TableState_fieldAccessorTable = new
com.google.protobuf.GeneratedMessage.FieldAccessorTable(
internal_static_hbase_pb_TableState_descriptor,
new java.lang.String[] { "State", });
internal_static_hbase_pb_ColumnFamilySchema_descriptor =
- getDescriptor().getMessageTypes().get(3);
+ getDescriptor().getMessageTypes().get(2);
internal_static_hbase_pb_ColumnFamilySchema_fieldAccessorTable = new
com.google.protobuf.GeneratedMessage.FieldAccessorTable(
internal_static_hbase_pb_ColumnFamilySchema_descriptor,
new java.lang.String[] { "Name", "Attributes", "Configuration", });
internal_static_hbase_pb_RegionInfo_descriptor =
- getDescriptor().getMessageTypes().get(4);
+ getDescriptor().getMessageTypes().get(3);
internal_static_hbase_pb_RegionInfo_fieldAccessorTable = new
com.google.protobuf.GeneratedMessage.FieldAccessorTable(
internal_static_hbase_pb_RegionInfo_descriptor,
new java.lang.String[] { "RegionId", "TableName", "StartKey", "EndKey", "Offline", "Split", "ReplicaId", });
internal_static_hbase_pb_FavoredNodes_descriptor =
- getDescriptor().getMessageTypes().get(5);
+ getDescriptor().getMessageTypes().get(4);
internal_static_hbase_pb_FavoredNodes_fieldAccessorTable = new
com.google.protobuf.GeneratedMessage.FieldAccessorTable(
internal_static_hbase_pb_FavoredNodes_descriptor,
new java.lang.String[] { "FavoredNode", });
internal_static_hbase_pb_RegionSpecifier_descriptor =
- getDescriptor().getMessageTypes().get(6);
+ getDescriptor().getMessageTypes().get(5);
internal_static_hbase_pb_RegionSpecifier_fieldAccessorTable = new
com.google.protobuf.GeneratedMessage.FieldAccessorTable(
internal_static_hbase_pb_RegionSpecifier_descriptor,
new java.lang.String[] { "Type", "Value", });
internal_static_hbase_pb_TimeRange_descriptor =
- getDescriptor().getMessageTypes().get(7);
+ getDescriptor().getMessageTypes().get(6);
internal_static_hbase_pb_TimeRange_fieldAccessorTable = new
com.google.protobuf.GeneratedMessage.FieldAccessorTable(
internal_static_hbase_pb_TimeRange_descriptor,
new java.lang.String[] { "From", "To", });
internal_static_hbase_pb_ColumnFamilyTimeRange_descriptor =
- getDescriptor().getMessageTypes().get(8);
+ getDescriptor().getMessageTypes().get(7);
internal_static_hbase_pb_ColumnFamilyTimeRange_fieldAccessorTable = new
com.google.protobuf.GeneratedMessage.FieldAccessorTable(
internal_static_hbase_pb_ColumnFamilyTimeRange_descriptor,
new java.lang.String[] { "ColumnFamily", "TimeRange", });
internal_static_hbase_pb_ServerName_descriptor =
- getDescriptor().getMessageTypes().get(9);
+ getDescriptor().getMessageTypes().get(8);
internal_static_hbase_pb_ServerName_fieldAccessorTable = new
com.google.protobuf.GeneratedMessage.FieldAccessorTable(
internal_static_hbase_pb_ServerName_descriptor,
new java.lang.String[] { "HostName", "Port", "StartCode", });
internal_static_hbase_pb_Coprocessor_descriptor =
- getDescriptor().getMessageTypes().get(10);
+ getDescriptor().getMessageTypes().get(9);
internal_static_hbase_pb_Coprocessor_fieldAccessorTable = new
com.google.protobuf.GeneratedMessage.FieldAccessorTable(
internal_static_hbase_pb_Coprocessor_descriptor,
new java.lang.String[] { "Name", });
internal_static_hbase_pb_NameStringPair_descriptor =
- getDescriptor().getMessageTypes().get(11);
+ getDescriptor().getMessageTypes().get(10);
internal_static_hbase_pb_NameStringPair_fieldAccessorTable = new
com.google.protobuf.GeneratedMessage.FieldAccessorTable(
internal_static_hbase_pb_NameStringPair_descriptor,
new java.lang.String[] { "Name", "Value", });
internal_static_hbase_pb_NameBytesPair_descriptor =
- getDescriptor().getMessageTypes().get(12);
+ getDescriptor().getMessageTypes().get(11);
internal_static_hbase_pb_NameBytesPair_fieldAccessorTable = new
com.google.protobuf.GeneratedMessage.FieldAccessorTable(
internal_static_hbase_pb_NameBytesPair_descriptor,
new java.lang.String[] { "Name", "Value", });
internal_static_hbase_pb_BytesBytesPair_descriptor =
- getDescriptor().getMessageTypes().get(13);
+ getDescriptor().getMessageTypes().get(12);
internal_static_hbase_pb_BytesBytesPair_fieldAccessorTable = new
com.google.protobuf.GeneratedMessage.FieldAccessorTable(
internal_static_hbase_pb_BytesBytesPair_descriptor,
new java.lang.String[] { "First", "Second", });
internal_static_hbase_pb_NameInt64Pair_descriptor =
- getDescriptor().getMessageTypes().get(14);
+ getDescriptor().getMessageTypes().get(13);
internal_static_hbase_pb_NameInt64Pair_fieldAccessorTable = new
com.google.protobuf.GeneratedMessage.FieldAccessorTable(
internal_static_hbase_pb_NameInt64Pair_descriptor,
new java.lang.String[] { "Name", "Value", });
internal_static_hbase_pb_SnapshotDescription_descriptor =
- getDescriptor().getMessageTypes().get(15);
+ getDescriptor().getMessageTypes().get(14);
internal_static_hbase_pb_SnapshotDescription_fieldAccessorTable = new
com.google.protobuf.GeneratedMessage.FieldAccessorTable(
internal_static_hbase_pb_SnapshotDescription_descriptor,
new java.lang.String[] { "Name", "Table", "CreationTime", "Type", "Version", "Owner", });
internal_static_hbase_pb_ProcedureDescription_descriptor =
- getDescriptor().getMessageTypes().get(16);
+ getDescriptor().getMessageTypes().get(15);
internal_static_hbase_pb_ProcedureDescription_fieldAccessorTable = new
com.google.protobuf.GeneratedMessage.FieldAccessorTable(
internal_static_hbase_pb_ProcedureDescription_descriptor,
new java.lang.String[] { "Signature", "Instance", "CreationTime", "Configuration", });
internal_static_hbase_pb_EmptyMsg_descriptor =
- getDescriptor().getMessageTypes().get(17);
+ getDescriptor().getMessageTypes().get(16);
internal_static_hbase_pb_EmptyMsg_fieldAccessorTable = new
com.google.protobuf.GeneratedMessage.FieldAccessorTable(
internal_static_hbase_pb_EmptyMsg_descriptor,
new java.lang.String[] { });
internal_static_hbase_pb_LongMsg_descriptor =
- getDescriptor().getMessageTypes().get(18);
+ getDescriptor().getMessageTypes().get(17);
internal_static_hbase_pb_LongMsg_fieldAccessorTable = new
com.google.protobuf.GeneratedMessage.FieldAccessorTable(
internal_static_hbase_pb_LongMsg_descriptor,
new java.lang.String[] { "LongMsg", });
internal_static_hbase_pb_DoubleMsg_descriptor =
- getDescriptor().getMessageTypes().get(19);
+ getDescriptor().getMessageTypes().get(18);
internal_static_hbase_pb_DoubleMsg_fieldAccessorTable = new
com.google.protobuf.GeneratedMessage.FieldAccessorTable(
internal_static_hbase_pb_DoubleMsg_descriptor,
new java.lang.String[] { "DoubleMsg", });
internal_static_hbase_pb_BigDecimalMsg_descriptor =
- getDescriptor().getMessageTypes().get(20);
+ getDescriptor().getMessageTypes().get(19);
internal_static_hbase_pb_BigDecimalMsg_fieldAccessorTable = new
com.google.protobuf.GeneratedMessage.FieldAccessorTable(
internal_static_hbase_pb_BigDecimalMsg_descriptor,
new java.lang.String[] { "BigdecimalMsg", });
internal_static_hbase_pb_UUID_descriptor =
- getDescriptor().getMessageTypes().get(21);
+ getDescriptor().getMessageTypes().get(20);
internal_static_hbase_pb_UUID_fieldAccessorTable = new
com.google.protobuf.GeneratedMessage.FieldAccessorTable(
internal_static_hbase_pb_UUID_descriptor,
new java.lang.String[] { "LeastSigBits", "MostSigBits", });
internal_static_hbase_pb_NamespaceDescriptor_descriptor =
- getDescriptor().getMessageTypes().get(22);
+ getDescriptor().getMessageTypes().get(21);
internal_static_hbase_pb_NamespaceDescriptor_fieldAccessorTable = new
com.google.protobuf.GeneratedMessage.FieldAccessorTable(
internal_static_hbase_pb_NamespaceDescriptor_descriptor,
new java.lang.String[] { "Name", "Configuration", });
internal_static_hbase_pb_VersionInfo_descriptor =
- getDescriptor().getMessageTypes().get(23);
+ getDescriptor().getMessageTypes().get(22);
internal_static_hbase_pb_VersionInfo_fieldAccessorTable = new
com.google.protobuf.GeneratedMessage.FieldAccessorTable(
internal_static_hbase_pb_VersionInfo_descriptor,
new java.lang.String[] { "Version", "Url", "Revision", "User", "Date", "SrcChecksum", "VersionMajor", "VersionMinor", });
internal_static_hbase_pb_RegionServerInfo_descriptor =
- getDescriptor().getMessageTypes().get(24);
+ getDescriptor().getMessageTypes().get(23);
internal_static_hbase_pb_RegionServerInfo_fieldAccessorTable = new
com.google.protobuf.GeneratedMessage.FieldAccessorTable(
internal_static_hbase_pb_RegionServerInfo_descriptor,
@@ -19886,6 +19318,7 @@ public final class HBaseProtos {
com.google.protobuf.Descriptors.FileDescriptor
.internalBuildGeneratedFileFrom(descriptorData,
new com.google.protobuf.Descriptors.FileDescriptor[] {
+ org.apache.hadoop.hbase.protobuf.generated.TableProtos.getDescriptor(),
}, assigner);
}
http://git-wip-us.apache.org/repos/asf/hbase/blob/5e21e306/hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/QuotaProtos.java
----------------------------------------------------------------------
diff --git a/hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/QuotaProtos.java b/hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/QuotaProtos.java
index 717ec73..e3bb364 100644
--- a/hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/QuotaProtos.java
+++ b/hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/QuotaProtos.java
@@ -982,7 +982,7 @@ public final class QuotaProtos {
public final boolean isInitialized() {
if (!hasTimeUnit()) {
-
+
return false;
}
return true;
@@ -2009,37 +2009,37 @@ public final class QuotaProtos {
public final boolean isInitialized() {
if (hasReqNum()) {
if (!getReqNum().isInitialized()) {
-
+
return false;
}
}
if (hasReqSize()) {
if (!getReqSize().isInitialized()) {
-
+
return false;
}
}
if (hasWriteNum()) {
if (!getWriteNum().isInitialized()) {
-
+
return false;
}
}
if (hasWriteSize()) {
if (!getWriteSize().isInitialized()) {
-
+
return false;
}
}
if (hasReadNum()) {
if (!getReadNum().isInitialized()) {
-
+
return false;
}
}
if (hasReadSize()) {
if (!getReadSize().isInitialized()) {
-
+
return false;
}
}
@@ -2169,7 +2169,7 @@ public final class QuotaProtos {
* <code>optional .hbase.pb.TimedQuota req_num = 1;</code>
*/
private com.google.protobuf.SingleFieldBuilder<
- org.apache.hadoop.hbase.protobuf.generated.QuotaProtos.TimedQuota, org.apache.hadoop.hbase.protobuf.generated.QuotaProtos.TimedQuota.Builder, org.apache.hadoop.hbase.protobuf.generated.QuotaProtos.TimedQuotaOrBuilder>
+ org.apache.hadoop.hbase.protobuf.generated.QuotaProtos.TimedQuota, org.apache.hadoop.hbase.protobuf.generated.QuotaProtos.TimedQuota.Builder, org.apache.hadoop.hbase.protobuf.generated.QuotaProtos.TimedQuotaOrBuilder>
getReqNumFieldBuilder() {
if (reqNumBuilder_ == null) {
reqNumBuilder_ = new com.google.protobuf.SingleFieldBuilder<
@@ -2286,7 +2286,7 @@ public final class QuotaProtos {
* <code>optional .hbase.pb.TimedQuota req_size = 2;</code>
*/
private com.google.protobuf.SingleFieldBuilder<
- org.apache.hadoop.hbase.protobuf.generated.QuotaProtos.TimedQuota, org.apache.hadoop.hbase.protobuf.generated.QuotaProtos.TimedQuota.Builder, org.apache.hadoop.hbase.protobuf.generated.QuotaProtos.TimedQuotaOrBuilder>
+ org.apache.hadoop.hbase.protobuf.generated.QuotaProtos.TimedQuota, org.apache.hadoop.hbase.protobuf.generated.QuotaProtos.TimedQuota.Builder, org.apache.hadoop.hbase.protobuf.generated.QuotaProtos.TimedQuotaOrBuilder>
getReqSizeFieldBuilder() {
if (reqSizeBuilder_ == null) {
reqSizeBuilder_ = new com.google.protobuf.SingleFieldBuilder<
@@ -2403,7 +2403,7 @@ public final class QuotaProtos {
* <code>optional .hbase.pb.TimedQuota write_num = 3;</code>
*/
private com.google.protobuf.SingleFieldBuilder<
- org.apache.hadoop.hbase.protobuf.generated.QuotaProtos.TimedQuota, org.apache.hadoop.hbase.protobuf.generated.QuotaProtos.TimedQuota.Builder, org.apache.hadoop.hbase.protobuf.generated.QuotaProtos.TimedQuotaOrBuilder>
+ org.apache.hadoop.hbase.protobuf.generated.QuotaProtos.TimedQuota, org.apache.hadoop.hbase.protobuf.generated.QuotaProtos.TimedQuota.Builder, org.apache.hadoop.hbase.protobuf.generated.QuotaProtos.TimedQuotaOrBuilder>
getWriteNumFieldBuilder() {
if (writeNumBuilder_ == null) {
writeNumBuilder_ = new com.google.protobuf.SingleFieldBuilder<
@@ -2520,7 +2520,7 @@ public final class QuotaProtos {
* <code>optional .hbase.pb.TimedQuota write_size = 4;</code>
*/
private com.google.protobuf.SingleFieldBuilder<
- org.apache.hadoop.hbase.protobuf.generated.QuotaProtos.TimedQuota, org.apache.hadoop.hbase.protobuf.generated.QuotaProtos.TimedQuota.Builder, org.apache.hadoop.hbase.protobuf.generated.QuotaProtos.TimedQuotaOrBuilder>
+ org.apache.hadoop.hbase.protobuf.generated.QuotaProtos.TimedQuota, org.apache.hadoop.hbase.protobuf.generated.QuotaProtos.TimedQuota.Builder, org.apache.hadoop.hbase.protobuf.generated.QuotaProtos.TimedQuotaOrBuilder>
getWriteSizeFieldBuilder() {
if (writeSizeBuilder_ == null) {
writeSizeBuilder_ = new com.google.protobuf.SingleFieldBuilder<
@@ -2637,7 +2637,7 @@ public final class QuotaProtos {
* <code>optional .hbase.pb.TimedQuota read_num = 5;</code>
*/
private com.google.protobuf.SingleFieldBuilder<
- org.apache.hadoop.hbase.protobuf.generated.QuotaProtos.TimedQuota, org.apache.hadoop.hbase.protobuf.generated.QuotaProtos.TimedQuota.Builder, org.apache.hadoop.hbase.protobuf.generated.QuotaProtos.TimedQuotaOrBuilder>
+ org.apache.hadoop.hbase.protobuf.generated.QuotaProtos.TimedQuota, org.apache.hadoop.hbase.protobuf.generated.QuotaProtos.TimedQuota.Builder, org.apache.hadoop.hbase.protobuf.generated.QuotaProtos.TimedQuotaOrBuilder>
getReadNumFieldBuilder() {
if (readNumBuilder_ == null) {
readNumBuilder_ = new com.google.protobuf.SingleFieldBuilder<
@@ -2754,7 +2754,7 @@ public final class QuotaProtos {
* <code>optional .hbase.pb.TimedQuota read_size = 6;</code>
*/
private com.google.protobuf.SingleFieldBuilder<
- org.apache.hadoop.hbase.protobuf.generated.QuotaProtos.TimedQuota, org.apache.hadoop.hbase.protobuf.generated.QuotaProtos.TimedQuota.Builder, org.apache.hadoop.hbase.protobuf.generated.QuotaProtos.TimedQuotaOrBuilder>
+ org.apache.hadoop.hbase.protobuf.generated.QuotaProtos.TimedQuota, org.apache.hadoop.hbase.protobuf.generated.QuotaProtos.TimedQuota.Builder, org.apache.hadoop.hbase.protobuf.generated.QuotaProtos.TimedQuotaOrBuilder>
getReadSizeFieldBuilder() {
if (readSizeBuilder_ == null) {
readSizeBuilder_ = new com.google.protobuf.SingleFieldBuilder<
@@ -3244,7 +3244,7 @@ public final class QuotaProtos {
public final boolean isInitialized() {
if (hasTimedQuota()) {
if (!getTimedQuota().isInitialized()) {
-
+
return false;
}
}
@@ -3410,7 +3410,7 @@ public final class QuotaProtos {
* <code>optional .hbase.pb.TimedQuota timed_quota = 2;</code>
*/
private com.google.protobuf.SingleFieldBuilder<
- org.apache.hadoop.hbase.protobuf.generated.QuotaProtos.TimedQuota, org.apache.hadoop.hbase.protobuf.generated.QuotaProtos.TimedQuota.Builder, org.apache.hadoop.hbase.protobuf.generated.QuotaProtos.TimedQuotaOrBuilder>
+ org.apache.hadoop.hbase.protobuf.generated.QuotaProtos.TimedQuota, org.apache.hadoop.hbase.protobuf.generated.QuotaProtos.TimedQuota.Builder, org.apache.hadoop.hbase.protobuf.generated.QuotaProtos.TimedQuotaOrBuilder>
getTimedQuotaFieldBuilder() {
if (timedQuotaBuilder_ == null) {
timedQuotaBuilder_ = new com.google.protobuf.SingleFieldBuilder<
@@ -3978,7 +3978,7 @@ public final class QuotaProtos {
public final boolean isInitialized() {
if (hasThrottle()) {
if (!getThrottle().isInitialized()) {
-
+
return false;
}
}
@@ -4141,7 +4141,7 @@ public final class QuotaProtos {
* <code>optional .hbase.pb.Throttle throttle = 2;</code>
*/
private com.google.protobuf.SingleFieldBuilder<
- org.apache.hadoop.hbase.protobuf.generated.QuotaProtos.Throttle, org.apache.hadoop.hbase.protobuf.generated.QuotaProtos.Throttle.Builder, org.apache.hadoop.hbase.protobuf.generated.QuotaProtos.ThrottleOrBuilder>
+ org.apache.hadoop.hbase.protobuf.generated.QuotaProtos.Throttle, org.apache.hadoop.hbase.protobuf.generated.QuotaProtos.Throttle.Builder, org.apache.hadoop.hbase.protobuf.generated.QuotaProtos.ThrottleOrBuilder>
getThrottleFieldBuilder() {
if (throttleBuilder_ == null) {
throttleBuilder_ = new com.google.protobuf.SingleFieldBuilder<
@@ -4258,7 +4258,7 @@ public final class QuotaProtos {
* <code>optional .hbase.pb.SpaceQuota space = 3;</code>
*/
private com.google.protobuf.SingleFieldBuilder<
- org.apache.hadoop.hbase.protobuf.generated.QuotaProtos.SpaceQuota, org.apache.hadoop.hbase.protobuf.generated.QuotaProtos.SpaceQuota.Builder, org.apache.hadoop.hbase.protobuf.generated.QuotaProtos.SpaceQuotaOrBuilder>
+ org.apache.hadoop.hbase.protobuf.generated.QuotaProtos.SpaceQuota, org.apache.hadoop.hbase.protobuf.generated.QuotaProtos.SpaceQuota.Builder, org.apache.hadoop.hbase.protobuf.generated.QuotaProtos.SpaceQuotaOrBuilder>
getSpaceFieldBuilder() {
if (spaceBuilder_ == null) {
spaceBuilder_ = new com.google.protobuf.SingleFieldBuilder<
@@ -5876,7 +5876,7 @@ public final class QuotaProtos {
* <code>optional .hbase.pb.SpaceQuota quota = 1;</code>
*/
private com.google.protobuf.SingleFieldBuilder<
- org.apache.hadoop.hbase.protobuf.generated.QuotaProtos.SpaceQuota, org.apache.hadoop.hbase.protobuf.generated.QuotaProtos.SpaceQuota.Builder, org.apache.hadoop.hbase.protobuf.generated.QuotaProtos.SpaceQuotaOrBuilder>
+ org.apache.hadoop.hbase.protobuf.generated.QuotaProtos.SpaceQuota, org.apache.hadoop.hbase.protobuf.generated.QuotaProtos.SpaceQuota.Builder, org.apache.hadoop.hbase.protobuf.generated.QuotaProtos.SpaceQuotaOrBuilder>
getQuotaFieldBuilder() {
if (quotaBuilder_ == null) {
quotaBuilder_ = new com.google.protobuf.SingleFieldBuilder<
@@ -7088,7 +7088,7 @@ public final class QuotaProtos {
* <code>optional .hbase.pb.SpaceQuotaStatus quota_status = 1;</code>
*/
private com.google.protobuf.SingleFieldBuilder<
- org.apache.hadoop.hbase.protobuf.generated.QuotaProtos.SpaceQuotaStatus, org.apache.hadoop.hbase.protobuf.generated.QuotaProtos.SpaceQuotaStatus.Builder, org.apache.hadoop.hbase.protobuf.generated.QuotaProtos.SpaceQuotaStatusOrBuilder>
+ org.apache.hadoop.hbase.protobuf.generated.QuotaProtos.SpaceQuotaStatus, org.apache.hadoop.hbase.protobuf.generated.QuotaProtos.SpaceQuotaStatus.Builder, org.apache.hadoop.hbase.protobuf.generated.QuotaProtos.SpaceQuotaStatusOrBuilder>
getQuotaStatusFieldBuilder() {
if (quotaStatusBuilder_ == null) {
quotaStatusBuilder_ = new com.google.protobuf.SingleFieldBuilder<
[2/5] hbase git commit: HBASE-18431 Mitigate compatibility concerns
between branch-1.3 and branch-1.4
Posted by st...@apache.org.
http://git-wip-us.apache.org/repos/asf/hbase/blob/5e21e306/hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/TableProtos.java
----------------------------------------------------------------------
diff --git a/hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/TableProtos.java b/hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/TableProtos.java
new file mode 100644
index 0000000..9507d01
--- /dev/null
+++ b/hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/TableProtos.java
@@ -0,0 +1,607 @@
+// Generated by the protocol buffer compiler. DO NOT EDIT!
+// source: Table.proto
+
+package org.apache.hadoop.hbase.protobuf.generated;
+
+public final class TableProtos {
+ private TableProtos() {}
+ public static void registerAllExtensions(
+ com.google.protobuf.ExtensionRegistry registry) {
+ }
+ public interface TableNameOrBuilder
+ extends com.google.protobuf.MessageOrBuilder {
+
+ // required bytes namespace = 1;
+ /**
+ * <code>required bytes namespace = 1;</code>
+ */
+ boolean hasNamespace();
+ /**
+ * <code>required bytes namespace = 1;</code>
+ */
+ com.google.protobuf.ByteString getNamespace();
+
+ // required bytes qualifier = 2;
+ /**
+ * <code>required bytes qualifier = 2;</code>
+ */
+ boolean hasQualifier();
+ /**
+ * <code>required bytes qualifier = 2;</code>
+ */
+ com.google.protobuf.ByteString getQualifier();
+ }
+ /**
+ * Protobuf type {@code hbase.pb.TableName}
+ *
+ * <pre>
+ **
+ * Table Name
+ * </pre>
+ */
+ public static final class TableName extends
+ com.google.protobuf.GeneratedMessage
+ implements TableNameOrBuilder {
+ // Use TableName.newBuilder() to construct.
+ private TableName(com.google.protobuf.GeneratedMessage.Builder<?> builder) {
+ super(builder);
+ this.unknownFields = builder.getUnknownFields();
+ }
+ private TableName(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
+
+ private static final TableName defaultInstance;
+ public static TableName getDefaultInstance() {
+ return defaultInstance;
+ }
+
+ public TableName getDefaultInstanceForType() {
+ return defaultInstance;
+ }
+
+ private final com.google.protobuf.UnknownFieldSet unknownFields;
+ @java.lang.Override
+ public final com.google.protobuf.UnknownFieldSet
+ getUnknownFields() {
+ return this.unknownFields;
+ }
+ private TableName(
+ com.google.protobuf.CodedInputStream input,
+ com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws com.google.protobuf.InvalidProtocolBufferException {
+ initFields();
+ int mutable_bitField0_ = 0;
+ com.google.protobuf.UnknownFieldSet.Builder unknownFields =
+ com.google.protobuf.UnknownFieldSet.newBuilder();
+ try {
+ boolean done = false;
+ while (!done) {
+ int tag = input.readTag();
+ switch (tag) {
+ case 0:
+ done = true;
+ break;
+ default: {
+ if (!parseUnknownField(input, unknownFields,
+ extensionRegistry, tag)) {
+ done = true;
+ }
+ break;
+ }
+ case 10: {
+ bitField0_ |= 0x00000001;
+ namespace_ = input.readBytes();
+ break;
+ }
+ case 18: {
+ bitField0_ |= 0x00000002;
+ qualifier_ = input.readBytes();
+ break;
+ }
+ }
+ }
+ } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ throw e.setUnfinishedMessage(this);
+ } catch (java.io.IOException e) {
+ throw new com.google.protobuf.InvalidProtocolBufferException(
+ e.getMessage()).setUnfinishedMessage(this);
+ } finally {
+ this.unknownFields = unknownFields.build();
+ makeExtensionsImmutable();
+ }
+ }
+ public static final com.google.protobuf.Descriptors.Descriptor
+ getDescriptor() {
+ return org.apache.hadoop.hbase.protobuf.generated.TableProtos.internal_static_hbase_pb_TableName_descriptor;
+ }
+
+ protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
+ internalGetFieldAccessorTable() {
+ return org.apache.hadoop.hbase.protobuf.generated.TableProtos.internal_static_hbase_pb_TableName_fieldAccessorTable
+ .ensureFieldAccessorsInitialized(
+ org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.class, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder.class);
+ }
+
+ public static com.google.protobuf.Parser<TableName> PARSER =
+ new com.google.protobuf.AbstractParser<TableName>() {
+ public TableName parsePartialFrom(
+ com.google.protobuf.CodedInputStream input,
+ com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws com.google.protobuf.InvalidProtocolBufferException {
+ return new TableName(input, extensionRegistry);
+ }
+ };
+
+ @java.lang.Override
+ public com.google.protobuf.Parser<TableName> getParserForType() {
+ return PARSER;
+ }
+
+ private int bitField0_;
+ // required bytes namespace = 1;
+ public static final int NAMESPACE_FIELD_NUMBER = 1;
+ private com.google.protobuf.ByteString namespace_;
+ /**
+ * <code>required bytes namespace = 1;</code>
+ */
+ public boolean hasNamespace() {
+ return ((bitField0_ & 0x00000001) == 0x00000001);
+ }
+ /**
+ * <code>required bytes namespace = 1;</code>
+ */
+ public com.google.protobuf.ByteString getNamespace() {
+ return namespace_;
+ }
+
+ // required bytes qualifier = 2;
+ public static final int QUALIFIER_FIELD_NUMBER = 2;
+ private com.google.protobuf.ByteString qualifier_;
+ /**
+ * <code>required bytes qualifier = 2;</code>
+ */
+ public boolean hasQualifier() {
+ return ((bitField0_ & 0x00000002) == 0x00000002);
+ }
+ /**
+ * <code>required bytes qualifier = 2;</code>
+ */
+ public com.google.protobuf.ByteString getQualifier() {
+ return qualifier_;
+ }
+
+ private void initFields() {
+ namespace_ = com.google.protobuf.ByteString.EMPTY;
+ qualifier_ = com.google.protobuf.ByteString.EMPTY;
+ }
+ private byte memoizedIsInitialized = -1;
+ public final boolean isInitialized() {
+ byte isInitialized = memoizedIsInitialized;
+ if (isInitialized != -1) return isInitialized == 1;
+
+ if (!hasNamespace()) {
+ memoizedIsInitialized = 0;
+ return false;
+ }
+ if (!hasQualifier()) {
+ memoizedIsInitialized = 0;
+ return false;
+ }
+ memoizedIsInitialized = 1;
+ return true;
+ }
+
+ public void writeTo(com.google.protobuf.CodedOutputStream output)
+ throws java.io.IOException {
+ getSerializedSize();
+ if (((bitField0_ & 0x00000001) == 0x00000001)) {
+ output.writeBytes(1, namespace_);
+ }
+ if (((bitField0_ & 0x00000002) == 0x00000002)) {
+ output.writeBytes(2, qualifier_);
+ }
+ getUnknownFields().writeTo(output);
+ }
+
+ private int memoizedSerializedSize = -1;
+ public int getSerializedSize() {
+ int size = memoizedSerializedSize;
+ if (size != -1) return size;
+
+ size = 0;
+ if (((bitField0_ & 0x00000001) == 0x00000001)) {
+ size += com.google.protobuf.CodedOutputStream
+ .computeBytesSize(1, namespace_);
+ }
+ if (((bitField0_ & 0x00000002) == 0x00000002)) {
+ size += com.google.protobuf.CodedOutputStream
+ .computeBytesSize(2, qualifier_);
+ }
+ size += getUnknownFields().getSerializedSize();
+ memoizedSerializedSize = size;
+ return size;
+ }
+
+ private static final long serialVersionUID = 0L;
+ @java.lang.Override
+ protected java.lang.Object writeReplace()
+ throws java.io.ObjectStreamException {
+ return super.writeReplace();
+ }
+
+ @java.lang.Override
+ public boolean equals(final java.lang.Object obj) {
+ if (obj == this) {
+ return true;
+ }
+ if (!(obj instanceof org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName)) {
+ return super.equals(obj);
+ }
+ org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName other = (org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName) obj;
+
+ boolean result = true;
+ result = result && (hasNamespace() == other.hasNamespace());
+ if (hasNamespace()) {
+ result = result && getNamespace()
+ .equals(other.getNamespace());
+ }
+ result = result && (hasQualifier() == other.hasQualifier());
+ if (hasQualifier()) {
+ result = result && getQualifier()
+ .equals(other.getQualifier());
+ }
+ result = result &&
+ getUnknownFields().equals(other.getUnknownFields());
+ return result;
+ }
+
+ private int memoizedHashCode = 0;
+ @java.lang.Override
+ public int hashCode() {
+ if (memoizedHashCode != 0) {
+ return memoizedHashCode;
+ }
+ int hash = 41;
+ hash = (19 * hash) + getDescriptorForType().hashCode();
+ if (hasNamespace()) {
+ hash = (37 * hash) + NAMESPACE_FIELD_NUMBER;
+ hash = (53 * hash) + getNamespace().hashCode();
+ }
+ if (hasQualifier()) {
+ hash = (37 * hash) + QUALIFIER_FIELD_NUMBER;
+ hash = (53 * hash) + getQualifier().hashCode();
+ }
+ hash = (29 * hash) + getUnknownFields().hashCode();
+ memoizedHashCode = hash;
+ return hash;
+ }
+
+ public static org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName parseFrom(
+ com.google.protobuf.ByteString data)
+ throws com.google.protobuf.InvalidProtocolBufferException {
+ return PARSER.parseFrom(data);
+ }
+ public static org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName parseFrom(
+ com.google.protobuf.ByteString data,
+ com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws com.google.protobuf.InvalidProtocolBufferException {
+ return PARSER.parseFrom(data, extensionRegistry);
+ }
+ public static org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName parseFrom(byte[] data)
+ throws com.google.protobuf.InvalidProtocolBufferException {
+ return PARSER.parseFrom(data);
+ }
+ public static org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName parseFrom(
+ byte[] data,
+ com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws com.google.protobuf.InvalidProtocolBufferException {
+ return PARSER.parseFrom(data, extensionRegistry);
+ }
+ public static org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName parseFrom(java.io.InputStream input)
+ throws java.io.IOException {
+ return PARSER.parseFrom(input);
+ }
+ public static org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName parseFrom(
+ java.io.InputStream input,
+ com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws java.io.IOException {
+ return PARSER.parseFrom(input, extensionRegistry);
+ }
+ public static org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName parseDelimitedFrom(java.io.InputStream input)
+ throws java.io.IOException {
+ return PARSER.parseDelimitedFrom(input);
+ }
+ public static org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName parseDelimitedFrom(
+ java.io.InputStream input,
+ com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws java.io.IOException {
+ return PARSER.parseDelimitedFrom(input, extensionRegistry);
+ }
+ public static org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName parseFrom(
+ com.google.protobuf.CodedInputStream input)
+ throws java.io.IOException {
+ return PARSER.parseFrom(input);
+ }
+ public static org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName parseFrom(
+ com.google.protobuf.CodedInputStream input,
+ com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws java.io.IOException {
+ return PARSER.parseFrom(input, extensionRegistry);
+ }
+
+ public static Builder newBuilder() { return Builder.create(); }
+ public Builder newBuilderForType() { return newBuilder(); }
+ public static Builder newBuilder(org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName prototype) {
+ return newBuilder().mergeFrom(prototype);
+ }
+ public Builder toBuilder() { return newBuilder(this); }
+
+ @java.lang.Override
+ protected Builder newBuilderForType(
+ com.google.protobuf.GeneratedMessage.BuilderParent parent) {
+ Builder builder = new Builder(parent);
+ return builder;
+ }
+ /**
+ * Protobuf type {@code hbase.pb.TableName}
+ *
+ * <pre>
+ **
+ * Table Name
+ * </pre>
+ */
+ public static final class Builder extends
+ com.google.protobuf.GeneratedMessage.Builder<Builder>
+ implements org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder {
+ public static final com.google.protobuf.Descriptors.Descriptor
+ getDescriptor() {
+ return org.apache.hadoop.hbase.protobuf.generated.TableProtos.internal_static_hbase_pb_TableName_descriptor;
+ }
+
+ protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
+ internalGetFieldAccessorTable() {
+ return org.apache.hadoop.hbase.protobuf.generated.TableProtos.internal_static_hbase_pb_TableName_fieldAccessorTable
+ .ensureFieldAccessorsInitialized(
+ org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.class, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder.class);
+ }
+
+ // Construct using org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.newBuilder()
+ private Builder() {
+ maybeForceBuilderInitialization();
+ }
+
+ private Builder(
+ com.google.protobuf.GeneratedMessage.BuilderParent parent) {
+ super(parent);
+ maybeForceBuilderInitialization();
+ }
+ private void maybeForceBuilderInitialization() {
+ if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
+ }
+ }
+ private static Builder create() {
+ return new Builder();
+ }
+
+ public Builder clear() {
+ super.clear();
+ namespace_ = com.google.protobuf.ByteString.EMPTY;
+ bitField0_ = (bitField0_ & ~0x00000001);
+ qualifier_ = com.google.protobuf.ByteString.EMPTY;
+ bitField0_ = (bitField0_ & ~0x00000002);
+ return this;
+ }
+
+ public Builder clone() {
+ return create().mergeFrom(buildPartial());
+ }
+
+ public com.google.protobuf.Descriptors.Descriptor
+ getDescriptorForType() {
+ return org.apache.hadoop.hbase.protobuf.generated.TableProtos.internal_static_hbase_pb_TableName_descriptor;
+ }
+
+ public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName getDefaultInstanceForType() {
+ return org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance();
+ }
+
+ public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName build() {
+ org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName result = buildPartial();
+ if (!result.isInitialized()) {
+ throw newUninitializedMessageException(result);
+ }
+ return result;
+ }
+
+ public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName buildPartial() {
+ org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName result = new org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName(this);
+ int from_bitField0_ = bitField0_;
+ int to_bitField0_ = 0;
+ if (((from_bitField0_ & 0x00000001) == 0x00000001)) {
+ to_bitField0_ |= 0x00000001;
+ }
+ result.namespace_ = namespace_;
+ if (((from_bitField0_ & 0x00000002) == 0x00000002)) {
+ to_bitField0_ |= 0x00000002;
+ }
+ result.qualifier_ = qualifier_;
+ result.bitField0_ = to_bitField0_;
+ onBuilt();
+ return result;
+ }
+
+ public Builder mergeFrom(com.google.protobuf.Message other) {
+ if (other instanceof org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName) {
+ return mergeFrom((org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName)other);
+ } else {
+ super.mergeFrom(other);
+ return this;
+ }
+ }
+
+ public Builder mergeFrom(org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName other) {
+ if (other == org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance()) return this;
+ if (other.hasNamespace()) {
+ setNamespace(other.getNamespace());
+ }
+ if (other.hasQualifier()) {
+ setQualifier(other.getQualifier());
+ }
+ this.mergeUnknownFields(other.getUnknownFields());
+ return this;
+ }
+
+ public final boolean isInitialized() {
+ if (!hasNamespace()) {
+
+ return false;
+ }
+ if (!hasQualifier()) {
+
+ return false;
+ }
+ return true;
+ }
+
+ public Builder mergeFrom(
+ com.google.protobuf.CodedInputStream input,
+ com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws java.io.IOException {
+ org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName parsedMessage = null;
+ try {
+ parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
+ } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ parsedMessage = (org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName) e.getUnfinishedMessage();
+ throw e;
+ } finally {
+ if (parsedMessage != null) {
+ mergeFrom(parsedMessage);
+ }
+ }
+ return this;
+ }
+ private int bitField0_;
+
+ // required bytes namespace = 1;
+ private com.google.protobuf.ByteString namespace_ = com.google.protobuf.ByteString.EMPTY;
+ /**
+ * <code>required bytes namespace = 1;</code>
+ */
+ public boolean hasNamespace() {
+ return ((bitField0_ & 0x00000001) == 0x00000001);
+ }
+ /**
+ * <code>required bytes namespace = 1;</code>
+ */
+ public com.google.protobuf.ByteString getNamespace() {
+ return namespace_;
+ }
+ /**
+ * <code>required bytes namespace = 1;</code>
+ */
+ public Builder setNamespace(com.google.protobuf.ByteString value) {
+ if (value == null) {
+ throw new NullPointerException();
+ }
+ bitField0_ |= 0x00000001;
+ namespace_ = value;
+ onChanged();
+ return this;
+ }
+ /**
+ * <code>required bytes namespace = 1;</code>
+ */
+ public Builder clearNamespace() {
+ bitField0_ = (bitField0_ & ~0x00000001);
+ namespace_ = getDefaultInstance().getNamespace();
+ onChanged();
+ return this;
+ }
+
+ // required bytes qualifier = 2;
+ private com.google.protobuf.ByteString qualifier_ = com.google.protobuf.ByteString.EMPTY;
+ /**
+ * <code>required bytes qualifier = 2;</code>
+ */
+ public boolean hasQualifier() {
+ return ((bitField0_ & 0x00000002) == 0x00000002);
+ }
+ /**
+ * <code>required bytes qualifier = 2;</code>
+ */
+ public com.google.protobuf.ByteString getQualifier() {
+ return qualifier_;
+ }
+ /**
+ * <code>required bytes qualifier = 2;</code>
+ */
+ public Builder setQualifier(com.google.protobuf.ByteString value) {
+ if (value == null) {
+ throw new NullPointerException();
+ }
+ bitField0_ |= 0x00000002;
+ qualifier_ = value;
+ onChanged();
+ return this;
+ }
+ /**
+ * <code>required bytes qualifier = 2;</code>
+ */
+ public Builder clearQualifier() {
+ bitField0_ = (bitField0_ & ~0x00000002);
+ qualifier_ = getDefaultInstance().getQualifier();
+ onChanged();
+ return this;
+ }
+
+ // @@protoc_insertion_point(builder_scope:hbase.pb.TableName)
+ }
+
+ static {
+ defaultInstance = new TableName(true);
+ defaultInstance.initFields();
+ }
+
+ // @@protoc_insertion_point(class_scope:hbase.pb.TableName)
+ }
+
+ private static com.google.protobuf.Descriptors.Descriptor
+ internal_static_hbase_pb_TableName_descriptor;
+ private static
+ com.google.protobuf.GeneratedMessage.FieldAccessorTable
+ internal_static_hbase_pb_TableName_fieldAccessorTable;
+
+ public static com.google.protobuf.Descriptors.FileDescriptor
+ getDescriptor() {
+ return descriptor;
+ }
+ private static com.google.protobuf.Descriptors.FileDescriptor
+ descriptor;
+ static {
+ java.lang.String[] descriptorData = {
+ "\n\013Table.proto\022\010hbase.pb\"1\n\tTableName\022\021\n\t" +
+ "namespace\030\001 \002(\014\022\021\n\tqualifier\030\002 \002(\014B>\n*or" +
+ "g.apache.hadoop.hbase.protobuf.generated" +
+ "B\013TableProtosH\001\240\001\001"
+ };
+ com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
+ new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() {
+ public com.google.protobuf.ExtensionRegistry assignDescriptors(
+ com.google.protobuf.Descriptors.FileDescriptor root) {
+ descriptor = root;
+ internal_static_hbase_pb_TableName_descriptor =
+ getDescriptor().getMessageTypes().get(0);
+ internal_static_hbase_pb_TableName_fieldAccessorTable = new
+ com.google.protobuf.GeneratedMessage.FieldAccessorTable(
+ internal_static_hbase_pb_TableName_descriptor,
+ new java.lang.String[] { "Namespace", "Qualifier", });
+ return null;
+ }
+ };
+ com.google.protobuf.Descriptors.FileDescriptor
+ .internalBuildGeneratedFileFrom(descriptorData,
+ new com.google.protobuf.Descriptors.FileDescriptor[] {
+ }, assigner);
+ }
+
+ // @@protoc_insertion_point(outer_class_scope)
+}
http://git-wip-us.apache.org/repos/asf/hbase/blob/5e21e306/hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/WALProtos.java
----------------------------------------------------------------------
diff --git a/hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/WALProtos.java b/hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/WALProtos.java
index a675b12..a8e2581 100644
--- a/hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/WALProtos.java
+++ b/hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/WALProtos.java
@@ -8813,11 +8813,11 @@ public final class WALProtos {
/**
* <code>required .hbase.pb.TableName table_name = 1;</code>
*/
- org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName getTableName();
+ org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName getTableName();
/**
* <code>required .hbase.pb.TableName table_name = 1;</code>
*/
- org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder();
+ org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder getTableNameOrBuilder();
// required bytes encoded_region_name = 2;
/**
@@ -8921,11 +8921,11 @@ public final class WALProtos {
break;
}
case 10: {
- org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder subBuilder = null;
+ org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder subBuilder = null;
if (((bitField0_ & 0x00000001) == 0x00000001)) {
subBuilder = tableName_.toBuilder();
}
- tableName_ = input.readMessage(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.PARSER, extensionRegistry);
+ tableName_ = input.readMessage(org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.PARSER, extensionRegistry);
if (subBuilder != null) {
subBuilder.mergeFrom(tableName_);
tableName_ = subBuilder.buildPartial();
@@ -8996,7 +8996,7 @@ public final class WALProtos {
private int bitField0_;
// required .hbase.pb.TableName table_name = 1;
public static final int TABLE_NAME_FIELD_NUMBER = 1;
- private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName tableName_;
+ private org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName tableName_;
/**
* <code>required .hbase.pb.TableName table_name = 1;</code>
*/
@@ -9006,13 +9006,13 @@ public final class WALProtos {
/**
* <code>required .hbase.pb.TableName table_name = 1;</code>
*/
- public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName getTableName() {
+ public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName getTableName() {
return tableName_;
}
/**
* <code>required .hbase.pb.TableName table_name = 1;</code>
*/
- public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder() {
+ public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder getTableNameOrBuilder() {
return tableName_;
}
@@ -9085,7 +9085,7 @@ public final class WALProtos {
}
private void initFields() {
- tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance();
+ tableName_ = org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance();
encodedRegionName_ = com.google.protobuf.ByteString.EMPTY;
stores_ = java.util.Collections.emptyList();
bulkloadSeqNum_ = 0L;
@@ -9347,7 +9347,7 @@ public final class WALProtos {
public Builder clear() {
super.clear();
if (tableNameBuilder_ == null) {
- tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance();
+ tableName_ = org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance();
} else {
tableNameBuilder_.clear();
}
@@ -9516,9 +9516,9 @@ public final class WALProtos {
private int bitField0_;
// required .hbase.pb.TableName table_name = 1;
- private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance();
+ private org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName tableName_ = org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance();
private com.google.protobuf.SingleFieldBuilder<
- org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder> tableNameBuilder_;
+ org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder> tableNameBuilder_;
/**
* <code>required .hbase.pb.TableName table_name = 1;</code>
*/
@@ -9528,7 +9528,7 @@ public final class WALProtos {
/**
* <code>required .hbase.pb.TableName table_name = 1;</code>
*/
- public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName getTableName() {
+ public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName getTableName() {
if (tableNameBuilder_ == null) {
return tableName_;
} else {
@@ -9538,7 +9538,7 @@ public final class WALProtos {
/**
* <code>required .hbase.pb.TableName table_name = 1;</code>
*/
- public Builder setTableName(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName value) {
+ public Builder setTableName(org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName value) {
if (tableNameBuilder_ == null) {
if (value == null) {
throw new NullPointerException();
@@ -9555,7 +9555,7 @@ public final class WALProtos {
* <code>required .hbase.pb.TableName table_name = 1;</code>
*/
public Builder setTableName(
- org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder builderForValue) {
+ org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder builderForValue) {
if (tableNameBuilder_ == null) {
tableName_ = builderForValue.build();
onChanged();
@@ -9568,12 +9568,12 @@ public final class WALProtos {
/**
* <code>required .hbase.pb.TableName table_name = 1;</code>
*/
- public Builder mergeTableName(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName value) {
+ public Builder mergeTableName(org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName value) {
if (tableNameBuilder_ == null) {
if (((bitField0_ & 0x00000001) == 0x00000001) &&
- tableName_ != org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance()) {
+ tableName_ != org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance()) {
tableName_ =
- org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.newBuilder(tableName_).mergeFrom(value).buildPartial();
+ org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.newBuilder(tableName_).mergeFrom(value).buildPartial();
} else {
tableName_ = value;
}
@@ -9589,7 +9589,7 @@ public final class WALProtos {
*/
public Builder clearTableName() {
if (tableNameBuilder_ == null) {
- tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance();
+ tableName_ = org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance();
onChanged();
} else {
tableNameBuilder_.clear();
@@ -9600,7 +9600,7 @@ public final class WALProtos {
/**
* <code>required .hbase.pb.TableName table_name = 1;</code>
*/
- public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder getTableNameBuilder() {
+ public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder getTableNameBuilder() {
bitField0_ |= 0x00000001;
onChanged();
return getTableNameFieldBuilder().getBuilder();
@@ -9608,7 +9608,7 @@ public final class WALProtos {
/**
* <code>required .hbase.pb.TableName table_name = 1;</code>
*/
- public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder() {
+ public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder getTableNameOrBuilder() {
if (tableNameBuilder_ != null) {
return tableNameBuilder_.getMessageOrBuilder();
} else {
@@ -9619,11 +9619,11 @@ public final class WALProtos {
* <code>required .hbase.pb.TableName table_name = 1;</code>
*/
private com.google.protobuf.SingleFieldBuilder<
- org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder>
+ org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder>
getTableNameFieldBuilder() {
if (tableNameBuilder_ == null) {
tableNameBuilder_ = new com.google.protobuf.SingleFieldBuilder<
- org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder>(
+ org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder>(
tableName_,
getParentForChildren(),
isClean());
@@ -11978,56 +11978,56 @@ public final class WALProtos {
descriptor;
static {
java.lang.String[] descriptorData = {
- "\n\tWAL.proto\022\010hbase.pb\032\013HBase.proto\"\217\001\n\tW" +
- "ALHeader\022\027\n\017has_compression\030\001 \001(\010\022\026\n\016enc" +
- "ryption_key\030\002 \001(\014\022\033\n\023has_tag_compression" +
- "\030\003 \001(\010\022\027\n\017writer_cls_name\030\004 \001(\t\022\033\n\023cell_" +
- "codec_cls_name\030\005 \001(\t\"\273\002\n\006WALKey\022\033\n\023encod" +
- "ed_region_name\030\001 \002(\014\022\022\n\ntable_name\030\002 \002(\014" +
- "\022\033\n\023log_sequence_number\030\003 \002(\004\022\022\n\nwrite_t" +
- "ime\030\004 \002(\004\022&\n\ncluster_id\030\005 \001(\0132\016.hbase.pb" +
- ".UUIDB\002\030\001\022%\n\006scopes\030\006 \003(\0132\025.hbase.pb.Fam" +
- "ilyScope\022\032\n\022following_kv_count\030\007 \001(\r\022#\n\013",
- "cluster_ids\030\010 \003(\0132\016.hbase.pb.UUID\022\022\n\nnon" +
- "ceGroup\030\t \001(\004\022\r\n\005nonce\030\n \001(\004\022\034\n\024orig_seq" +
- "uence_number\030\013 \001(\004\"F\n\013FamilyScope\022\016\n\006fam" +
- "ily\030\001 \002(\014\022\'\n\nscope_type\030\002 \002(\0162\023.hbase.pb" +
- ".ScopeType\"\276\001\n\024CompactionDescriptor\022\022\n\nt" +
- "able_name\030\001 \002(\014\022\033\n\023encoded_region_name\030\002" +
- " \002(\014\022\023\n\013family_name\030\003 \002(\014\022\030\n\020compaction_" +
- "input\030\004 \003(\t\022\031\n\021compaction_output\030\005 \003(\t\022\026" +
- "\n\016store_home_dir\030\006 \002(\t\022\023\n\013region_name\030\007 " +
- "\001(\014\"\244\003\n\017FlushDescriptor\0225\n\006action\030\001 \002(\0162",
- "%.hbase.pb.FlushDescriptor.FlushAction\022\022" +
- "\n\ntable_name\030\002 \002(\014\022\033\n\023encoded_region_nam" +
- "e\030\003 \002(\014\022\035\n\025flush_sequence_number\030\004 \001(\004\022E" +
- "\n\rstore_flushes\030\005 \003(\0132..hbase.pb.FlushDe" +
- "scriptor.StoreFlushDescriptor\022\023\n\013region_" +
- "name\030\006 \001(\014\032Y\n\024StoreFlushDescriptor\022\023\n\013fa" +
- "mily_name\030\001 \002(\014\022\026\n\016store_home_dir\030\002 \002(\t\022" +
- "\024\n\014flush_output\030\003 \003(\t\"S\n\013FlushAction\022\017\n\013" +
- "START_FLUSH\020\000\022\020\n\014COMMIT_FLUSH\020\001\022\017\n\013ABORT" +
- "_FLUSH\020\002\022\020\n\014CANNOT_FLUSH\020\003\"q\n\017StoreDescr",
- "iptor\022\023\n\013family_name\030\001 \002(\014\022\026\n\016store_home" +
- "_dir\030\002 \002(\t\022\022\n\nstore_file\030\003 \003(\t\022\035\n\025store_" +
- "file_size_bytes\030\004 \001(\004\"\237\001\n\022BulkLoadDescri" +
- "ptor\022\'\n\ntable_name\030\001 \002(\0132\023.hbase.pb.Tabl" +
- "eName\022\033\n\023encoded_region_name\030\002 \002(\014\022)\n\006st" +
- "ores\030\003 \003(\0132\031.hbase.pb.StoreDescriptor\022\030\n" +
- "\020bulkload_seq_num\030\004 \002(\003\"\272\002\n\025RegionEventD" +
- "escriptor\022=\n\nevent_type\030\001 \002(\0162).hbase.pb" +
- ".RegionEventDescriptor.EventType\022\022\n\ntabl" +
- "e_name\030\002 \002(\014\022\033\n\023encoded_region_name\030\003 \002(",
- "\014\022\033\n\023log_sequence_number\030\004 \001(\004\022)\n\006stores" +
- "\030\005 \003(\0132\031.hbase.pb.StoreDescriptor\022$\n\006ser" +
- "ver\030\006 \001(\0132\024.hbase.pb.ServerName\022\023\n\013regio" +
- "n_name\030\007 \001(\014\".\n\tEventType\022\017\n\013REGION_OPEN" +
- "\020\000\022\020\n\014REGION_CLOSE\020\001\"\014\n\nWALTrailer*d\n\tSc" +
- "opeType\022\033\n\027REPLICATION_SCOPE_LOCAL\020\000\022\034\n\030" +
- "REPLICATION_SCOPE_GLOBAL\020\001\022\034\n\030REPLICATIO" +
- "N_SCOPE_SERIAL\020\002B?\n*org.apache.hadoop.hb" +
- "ase.protobuf.generatedB\tWALProtosH\001\210\001\000\240\001" +
- "\001"
+ "\n\tWAL.proto\022\010hbase.pb\032\013Table.proto\032\013HBas" +
+ "e.proto\"\217\001\n\tWALHeader\022\027\n\017has_compression" +
+ "\030\001 \001(\010\022\026\n\016encryption_key\030\002 \001(\014\022\033\n\023has_ta" +
+ "g_compression\030\003 \001(\010\022\027\n\017writer_cls_name\030\004" +
+ " \001(\t\022\033\n\023cell_codec_cls_name\030\005 \001(\t\"\273\002\n\006WA" +
+ "LKey\022\033\n\023encoded_region_name\030\001 \002(\014\022\022\n\ntab" +
+ "le_name\030\002 \002(\014\022\033\n\023log_sequence_number\030\003 \002" +
+ "(\004\022\022\n\nwrite_time\030\004 \002(\004\022&\n\ncluster_id\030\005 \001" +
+ "(\0132\016.hbase.pb.UUIDB\002\030\001\022%\n\006scopes\030\006 \003(\0132\025" +
+ ".hbase.pb.FamilyScope\022\032\n\022following_kv_co",
+ "unt\030\007 \001(\r\022#\n\013cluster_ids\030\010 \003(\0132\016.hbase.p" +
+ "b.UUID\022\022\n\nnonceGroup\030\t \001(\004\022\r\n\005nonce\030\n \001(" +
+ "\004\022\034\n\024orig_sequence_number\030\013 \001(\004\"F\n\013Famil" +
+ "yScope\022\016\n\006family\030\001 \002(\014\022\'\n\nscope_type\030\002 \002" +
+ "(\0162\023.hbase.pb.ScopeType\"\276\001\n\024CompactionDe" +
+ "scriptor\022\022\n\ntable_name\030\001 \002(\014\022\033\n\023encoded_" +
+ "region_name\030\002 \002(\014\022\023\n\013family_name\030\003 \002(\014\022\030" +
+ "\n\020compaction_input\030\004 \003(\t\022\031\n\021compaction_o" +
+ "utput\030\005 \003(\t\022\026\n\016store_home_dir\030\006 \002(\t\022\023\n\013r" +
+ "egion_name\030\007 \001(\014\"\244\003\n\017FlushDescriptor\0225\n\006",
+ "action\030\001 \002(\0162%.hbase.pb.FlushDescriptor." +
+ "FlushAction\022\022\n\ntable_name\030\002 \002(\014\022\033\n\023encod" +
+ "ed_region_name\030\003 \002(\014\022\035\n\025flush_sequence_n" +
+ "umber\030\004 \001(\004\022E\n\rstore_flushes\030\005 \003(\0132..hba" +
+ "se.pb.FlushDescriptor.StoreFlushDescript" +
+ "or\022\023\n\013region_name\030\006 \001(\014\032Y\n\024StoreFlushDes" +
+ "criptor\022\023\n\013family_name\030\001 \002(\014\022\026\n\016store_ho" +
+ "me_dir\030\002 \002(\t\022\024\n\014flush_output\030\003 \003(\t\"S\n\013Fl" +
+ "ushAction\022\017\n\013START_FLUSH\020\000\022\020\n\014COMMIT_FLU" +
+ "SH\020\001\022\017\n\013ABORT_FLUSH\020\002\022\020\n\014CANNOT_FLUSH\020\003\"",
+ "q\n\017StoreDescriptor\022\023\n\013family_name\030\001 \002(\014\022" +
+ "\026\n\016store_home_dir\030\002 \002(\t\022\022\n\nstore_file\030\003 " +
+ "\003(\t\022\035\n\025store_file_size_bytes\030\004 \001(\004\"\237\001\n\022B" +
+ "ulkLoadDescriptor\022\'\n\ntable_name\030\001 \002(\0132\023." +
+ "hbase.pb.TableName\022\033\n\023encoded_region_nam" +
+ "e\030\002 \002(\014\022)\n\006stores\030\003 \003(\0132\031.hbase.pb.Store" +
+ "Descriptor\022\030\n\020bulkload_seq_num\030\004 \002(\003\"\272\002\n" +
+ "\025RegionEventDescriptor\022=\n\nevent_type\030\001 \002" +
+ "(\0162).hbase.pb.RegionEventDescriptor.Even" +
+ "tType\022\022\n\ntable_name\030\002 \002(\014\022\033\n\023encoded_reg",
+ "ion_name\030\003 \002(\014\022\033\n\023log_sequence_number\030\004 " +
+ "\001(\004\022)\n\006stores\030\005 \003(\0132\031.hbase.pb.StoreDesc" +
+ "riptor\022$\n\006server\030\006 \001(\0132\024.hbase.pb.Server" +
+ "Name\022\023\n\013region_name\030\007 \001(\014\".\n\tEventType\022\017" +
+ "\n\013REGION_OPEN\020\000\022\020\n\014REGION_CLOSE\020\001\"\014\n\nWAL" +
+ "Trailer*d\n\tScopeType\022\033\n\027REPLICATION_SCOP" +
+ "E_LOCAL\020\000\022\034\n\030REPLICATION_SCOPE_GLOBAL\020\001\022" +
+ "\034\n\030REPLICATION_SCOPE_SERIAL\020\002B?\n*org.apa" +
+ "che.hadoop.hbase.protobuf.generatedB\tWAL" +
+ "ProtosH\001\210\001\000\240\001\001"
};
com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() {
@@ -12100,6 +12100,7 @@ public final class WALProtos {
com.google.protobuf.Descriptors.FileDescriptor
.internalBuildGeneratedFileFrom(descriptorData,
new com.google.protobuf.Descriptors.FileDescriptor[] {
+ org.apache.hadoop.hbase.protobuf.generated.TableProtos.getDescriptor(),
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.getDescriptor(),
}, assigner);
}
http://git-wip-us.apache.org/repos/asf/hbase/blob/5e21e306/hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/ZooKeeperProtos.java
----------------------------------------------------------------------
diff --git a/hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/ZooKeeperProtos.java b/hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/ZooKeeperProtos.java
index 81cc69b..3d4ad9a7 100644
--- a/hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/ZooKeeperProtos.java
+++ b/hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/ZooKeeperProtos.java
@@ -3958,11 +3958,11 @@ public final class ZooKeeperProtos {
/**
* <code>optional .hbase.pb.TableName table_name = 1;</code>
*/
- org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName getTableName();
+ org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName getTableName();
/**
* <code>optional .hbase.pb.TableName table_name = 1;</code>
*/
- org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder();
+ org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder getTableNameOrBuilder();
// repeated bytes families = 2;
/**
@@ -4030,11 +4030,11 @@ public final class ZooKeeperProtos {
break;
}
case 10: {
- org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder subBuilder = null;
+ org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder subBuilder = null;
if (((bitField0_ & 0x00000001) == 0x00000001)) {
subBuilder = tableName_.toBuilder();
}
- tableName_ = input.readMessage(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.PARSER, extensionRegistry);
+ tableName_ = input.readMessage(org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.PARSER, extensionRegistry);
if (subBuilder != null) {
subBuilder.mergeFrom(tableName_);
tableName_ = subBuilder.buildPartial();
@@ -4095,7 +4095,7 @@ public final class ZooKeeperProtos {
private int bitField0_;
// optional .hbase.pb.TableName table_name = 1;
public static final int TABLE_NAME_FIELD_NUMBER = 1;
- private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName tableName_;
+ private org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName tableName_;
/**
* <code>optional .hbase.pb.TableName table_name = 1;</code>
*/
@@ -4105,13 +4105,13 @@ public final class ZooKeeperProtos {
/**
* <code>optional .hbase.pb.TableName table_name = 1;</code>
*/
- public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName getTableName() {
+ public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName getTableName() {
return tableName_;
}
/**
* <code>optional .hbase.pb.TableName table_name = 1;</code>
*/
- public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder() {
+ public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder getTableNameOrBuilder() {
return tableName_;
}
@@ -4139,7 +4139,7 @@ public final class ZooKeeperProtos {
}
private void initFields() {
- tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance();
+ tableName_ = org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance();
families_ = java.util.Collections.emptyList();
}
private byte memoizedIsInitialized = -1;
@@ -4350,7 +4350,7 @@ public final class ZooKeeperProtos {
public Builder clear() {
super.clear();
if (tableNameBuilder_ == null) {
- tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance();
+ tableName_ = org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance();
} else {
tableNameBuilder_.clear();
}
@@ -4461,9 +4461,9 @@ public final class ZooKeeperProtos {
private int bitField0_;
// optional .hbase.pb.TableName table_name = 1;
- private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance();
+ private org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName tableName_ = org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance();
private com.google.protobuf.SingleFieldBuilder<
- org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder> tableNameBuilder_;
+ org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder> tableNameBuilder_;
/**
* <code>optional .hbase.pb.TableName table_name = 1;</code>
*/
@@ -4473,7 +4473,7 @@ public final class ZooKeeperProtos {
/**
* <code>optional .hbase.pb.TableName table_name = 1;</code>
*/
- public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName getTableName() {
+ public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName getTableName() {
if (tableNameBuilder_ == null) {
return tableName_;
} else {
@@ -4483,7 +4483,7 @@ public final class ZooKeeperProtos {
/**
* <code>optional .hbase.pb.TableName table_name = 1;</code>
*/
- public Builder setTableName(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName value) {
+ public Builder setTableName(org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName value) {
if (tableNameBuilder_ == null) {
if (value == null) {
throw new NullPointerException();
@@ -4500,7 +4500,7 @@ public final class ZooKeeperProtos {
* <code>optional .hbase.pb.TableName table_name = 1;</code>
*/
public Builder setTableName(
- org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder builderForValue) {
+ org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder builderForValue) {
if (tableNameBuilder_ == null) {
tableName_ = builderForValue.build();
onChanged();
@@ -4513,12 +4513,12 @@ public final class ZooKeeperProtos {
/**
* <code>optional .hbase.pb.TableName table_name = 1;</code>
*/
- public Builder mergeTableName(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName value) {
+ public Builder mergeTableName(org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName value) {
if (tableNameBuilder_ == null) {
if (((bitField0_ & 0x00000001) == 0x00000001) &&
- tableName_ != org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance()) {
+ tableName_ != org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance()) {
tableName_ =
- org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.newBuilder(tableName_).mergeFrom(value).buildPartial();
+ org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.newBuilder(tableName_).mergeFrom(value).buildPartial();
} else {
tableName_ = value;
}
@@ -4534,7 +4534,7 @@ public final class ZooKeeperProtos {
*/
public Builder clearTableName() {
if (tableNameBuilder_ == null) {
- tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance();
+ tableName_ = org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance();
onChanged();
} else {
tableNameBuilder_.clear();
@@ -4545,7 +4545,7 @@ public final class ZooKeeperProtos {
/**
* <code>optional .hbase.pb.TableName table_name = 1;</code>
*/
- public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder getTableNameBuilder() {
+ public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder getTableNameBuilder() {
bitField0_ |= 0x00000001;
onChanged();
return getTableNameFieldBuilder().getBuilder();
@@ -4553,7 +4553,7 @@ public final class ZooKeeperProtos {
/**
* <code>optional .hbase.pb.TableName table_name = 1;</code>
*/
- public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder() {
+ public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder getTableNameOrBuilder() {
if (tableNameBuilder_ != null) {
return tableNameBuilder_.getMessageOrBuilder();
} else {
@@ -4564,11 +4564,11 @@ public final class ZooKeeperProtos {
* <code>optional .hbase.pb.TableName table_name = 1;</code>
*/
private com.google.protobuf.SingleFieldBuilder<
- org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder>
+ org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder>
getTableNameFieldBuilder() {
if (tableNameBuilder_ == null) {
tableNameBuilder_ = new com.google.protobuf.SingleFieldBuilder<
- org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder>(
+ org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder>(
tableName_,
getParentForChildren(),
isClean());
@@ -8296,39 +8296,40 @@ public final class ZooKeeperProtos {
descriptor;
static {
java.lang.String[] descriptorData = {
- "\n\017ZooKeeper.proto\022\010hbase.pb\032\013HBase.proto" +
- "\032\023ClusterStatus.proto\"y\n\020MetaRegionServe" +
- "r\022$\n\006server\030\001 \002(\0132\024.hbase.pb.ServerName\022" +
- "\023\n\013rpc_version\030\002 \001(\r\022*\n\005state\030\003 \001(\0162\033.hb" +
- "ase.pb.RegionState.State\"V\n\006Master\022$\n\006ma" +
- "ster\030\001 \002(\0132\024.hbase.pb.ServerName\022\023\n\013rpc_" +
- "version\030\002 \001(\r\022\021\n\tinfo_port\030\003 \001(\r\"\037\n\tClus" +
- "terUp\022\022\n\nstart_date\030\001 \002(\t\"\247\002\n\014SplitLogTa" +
- "sk\022+\n\005state\030\001 \002(\0162\034.hbase.pb.SplitLogTas" +
- "k.State\022)\n\013server_name\030\002 \002(\0132\024.hbase.pb.",
- "ServerName\022:\n\004mode\030\003 \001(\0162#.hbase.pb.Spli" +
- "tLogTask.RecoveryMode:\007UNKNOWN\"C\n\005State\022" +
- "\016\n\nUNASSIGNED\020\000\022\t\n\005OWNED\020\001\022\014\n\010RESIGNED\020\002" +
- "\022\010\n\004DONE\020\003\022\007\n\003ERR\020\004\">\n\014RecoveryMode\022\013\n\007U" +
- "NKNOWN\020\000\022\021\n\rLOG_SPLITTING\020\001\022\016\n\nLOG_REPLA" +
- "Y\020\002\"\225\001\n\024DeprecatedTableState\022<\n\005state\030\001 " +
- "\002(\0162$.hbase.pb.DeprecatedTableState.Stat" +
- "e:\007ENABLED\"?\n\005State\022\013\n\007ENABLED\020\000\022\014\n\010DISA" +
- "BLED\020\001\022\r\n\tDISABLING\020\002\022\014\n\010ENABLING\020\003\"D\n\007T" +
- "ableCF\022\'\n\ntable_name\030\001 \001(\0132\023.hbase.pb.Ta",
- "bleName\022\020\n\010families\030\002 \003(\014\"\354\001\n\017Replicatio" +
- "nPeer\022\022\n\nclusterkey\030\001 \002(\t\022\037\n\027replication" +
- "EndpointImpl\030\002 \001(\t\022&\n\004data\030\003 \003(\0132\030.hbase" +
- ".pb.BytesBytesPair\022/\n\rconfiguration\030\004 \003(" +
- "\0132\030.hbase.pb.NameStringPair\022$\n\ttable_cfs" +
- "\030\005 \003(\0132\021.hbase.pb.TableCF\022\022\n\nnamespaces\030" +
- "\006 \003(\014\022\021\n\tbandwidth\030\007 \001(\003\"g\n\020ReplicationS" +
- "tate\022/\n\005state\030\001 \002(\0162 .hbase.pb.Replicati" +
- "onState.State\"\"\n\005State\022\013\n\007ENABLED\020\000\022\014\n\010D" +
- "ISABLED\020\001\"+\n\027ReplicationHLogPosition\022\020\n\010",
- "position\030\001 \002(\003\"\036\n\013SwitchState\022\017\n\007enabled" +
- "\030\001 \001(\010BE\n*org.apache.hadoop.hbase.protob" +
- "uf.generatedB\017ZooKeeperProtosH\001\210\001\001\240\001\001"
+ "\n\017ZooKeeper.proto\022\010hbase.pb\032\013Table.proto" +
+ "\032\013HBase.proto\032\023ClusterStatus.proto\"y\n\020Me" +
+ "taRegionServer\022$\n\006server\030\001 \002(\0132\024.hbase.p" +
+ "b.ServerName\022\023\n\013rpc_version\030\002 \001(\r\022*\n\005sta" +
+ "te\030\003 \001(\0162\033.hbase.pb.RegionState.State\"V\n" +
+ "\006Master\022$\n\006master\030\001 \002(\0132\024.hbase.pb.Serve" +
+ "rName\022\023\n\013rpc_version\030\002 \001(\r\022\021\n\tinfo_port\030" +
+ "\003 \001(\r\"\037\n\tClusterUp\022\022\n\nstart_date\030\001 \002(\t\"\247" +
+ "\002\n\014SplitLogTask\022+\n\005state\030\001 \002(\0162\034.hbase.p" +
+ "b.SplitLogTask.State\022)\n\013server_name\030\002 \002(",
+ "\0132\024.hbase.pb.ServerName\022:\n\004mode\030\003 \001(\0162#." +
+ "hbase.pb.SplitLogTask.RecoveryMode:\007UNKN" +
+ "OWN\"C\n\005State\022\016\n\nUNASSIGNED\020\000\022\t\n\005OWNED\020\001\022" +
+ "\014\n\010RESIGNED\020\002\022\010\n\004DONE\020\003\022\007\n\003ERR\020\004\">\n\014Reco" +
+ "veryMode\022\013\n\007UNKNOWN\020\000\022\021\n\rLOG_SPLITTING\020\001" +
+ "\022\016\n\nLOG_REPLAY\020\002\"\225\001\n\024DeprecatedTableStat" +
+ "e\022<\n\005state\030\001 \002(\0162$.hbase.pb.DeprecatedTa" +
+ "bleState.State:\007ENABLED\"?\n\005State\022\013\n\007ENAB" +
+ "LED\020\000\022\014\n\010DISABLED\020\001\022\r\n\tDISABLING\020\002\022\014\n\010EN" +
+ "ABLING\020\003\"D\n\007TableCF\022\'\n\ntable_name\030\001 \001(\0132",
+ "\023.hbase.pb.TableName\022\020\n\010families\030\002 \003(\014\"\354" +
+ "\001\n\017ReplicationPeer\022\022\n\nclusterkey\030\001 \002(\t\022\037" +
+ "\n\027replicationEndpointImpl\030\002 \001(\t\022&\n\004data\030" +
+ "\003 \003(\0132\030.hbase.pb.BytesBytesPair\022/\n\rconfi" +
+ "guration\030\004 \003(\0132\030.hbase.pb.NameStringPair" +
+ "\022$\n\ttable_cfs\030\005 \003(\0132\021.hbase.pb.TableCF\022\022" +
+ "\n\nnamespaces\030\006 \003(\014\022\021\n\tbandwidth\030\007 \001(\003\"g\n" +
+ "\020ReplicationState\022/\n\005state\030\001 \002(\0162 .hbase" +
+ ".pb.ReplicationState.State\"\"\n\005State\022\013\n\007E" +
+ "NABLED\020\000\022\014\n\010DISABLED\020\001\"+\n\027ReplicationHLo",
+ "gPosition\022\020\n\010position\030\001 \002(\003\"\036\n\013SwitchSta" +
+ "te\022\017\n\007enabled\030\001 \001(\010BE\n*org.apache.hadoop" +
+ ".hbase.protobuf.generatedB\017ZooKeeperProt" +
+ "osH\001\210\001\001\240\001\001"
};
com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() {
@@ -8401,6 +8402,7 @@ public final class ZooKeeperProtos {
com.google.protobuf.Descriptors.FileDescriptor
.internalBuildGeneratedFileFrom(descriptorData,
new com.google.protobuf.Descriptors.FileDescriptor[] {
+ org.apache.hadoop.hbase.protobuf.generated.TableProtos.getDescriptor(),
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.getDescriptor(),
org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.getDescriptor(),
}, assigner);
http://git-wip-us.apache.org/repos/asf/hbase/blob/5e21e306/hbase-protocol/src/main/protobuf/AccessControl.proto
----------------------------------------------------------------------
diff --git a/hbase-protocol/src/main/protobuf/AccessControl.proto b/hbase-protocol/src/main/protobuf/AccessControl.proto
index cc0d4a5..02e3a5b 100644
--- a/hbase-protocol/src/main/protobuf/AccessControl.proto
+++ b/hbase-protocol/src/main/protobuf/AccessControl.proto
@@ -23,7 +23,7 @@ option java_generic_services = true;
option java_generate_equals_and_hash = true;
option optimize_for = SPEED;
-import "HBase.proto";
+import "Table.proto";
message Permission {
enum Action {
http://git-wip-us.apache.org/repos/asf/hbase/blob/5e21e306/hbase-protocol/src/main/protobuf/Client.proto
----------------------------------------------------------------------
diff --git a/hbase-protocol/src/main/protobuf/Client.proto b/hbase-protocol/src/main/protobuf/Client.proto
index 725a2b5..20a5b63 100644
--- a/hbase-protocol/src/main/protobuf/Client.proto
+++ b/hbase-protocol/src/main/protobuf/Client.proto
@@ -25,6 +25,7 @@ option java_generic_services = true;
option java_generate_equals_and_hash = true;
option optimize_for = SPEED;
+import "Table.proto";
import "HBase.proto";
import "Filter.proto";
import "Cell.proto";
http://git-wip-us.apache.org/repos/asf/hbase/blob/5e21e306/hbase-protocol/src/main/protobuf/HBase.proto
----------------------------------------------------------------------
diff --git a/hbase-protocol/src/main/protobuf/HBase.proto b/hbase-protocol/src/main/protobuf/HBase.proto
index 7d064f3..0dcf363 100644
--- a/hbase-protocol/src/main/protobuf/HBase.proto
+++ b/hbase-protocol/src/main/protobuf/HBase.proto
@@ -24,14 +24,7 @@ option java_outer_classname = "HBaseProtos";
option java_generate_equals_and_hash = true;
option optimize_for = SPEED;
-
-/**
- * Table Name
- */
-message TableName {
- required bytes namespace = 1;
- required bytes qualifier = 2;
-}
+import "Table.proto";
/**
* Table Schema
http://git-wip-us.apache.org/repos/asf/hbase/blob/5e21e306/hbase-protocol/src/main/protobuf/Table.proto
----------------------------------------------------------------------
diff --git a/hbase-protocol/src/main/protobuf/Table.proto b/hbase-protocol/src/main/protobuf/Table.proto
new file mode 100644
index 0000000..4452eb2
--- /dev/null
+++ b/hbase-protocol/src/main/protobuf/Table.proto
@@ -0,0 +1,33 @@
+/**
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements. See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership. The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License. You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+// This file contains protocol buffers that are shared throughout HBase
+package hbase.pb;
+
+option java_package = "org.apache.hadoop.hbase.protobuf.generated";
+option java_outer_classname = "TableProtos";
+option java_generate_equals_and_hash = true;
+option optimize_for = SPEED;
+
+/**
+ * Table Name
+ */
+message TableName {
+ required bytes namespace = 1;
+ required bytes qualifier = 2;
+}
http://git-wip-us.apache.org/repos/asf/hbase/blob/5e21e306/hbase-protocol/src/main/protobuf/WAL.proto
----------------------------------------------------------------------
diff --git a/hbase-protocol/src/main/protobuf/WAL.proto b/hbase-protocol/src/main/protobuf/WAL.proto
index 2494977..9bf4a79 100644
--- a/hbase-protocol/src/main/protobuf/WAL.proto
+++ b/hbase-protocol/src/main/protobuf/WAL.proto
@@ -23,6 +23,7 @@ option java_generic_services = false;
option java_generate_equals_and_hash = true;
option optimize_for = SPEED;
+import "Table.proto";
import "HBase.proto";
message WALHeader {
http://git-wip-us.apache.org/repos/asf/hbase/blob/5e21e306/hbase-protocol/src/main/protobuf/ZooKeeper.proto
----------------------------------------------------------------------
diff --git a/hbase-protocol/src/main/protobuf/ZooKeeper.proto b/hbase-protocol/src/main/protobuf/ZooKeeper.proto
index ee73c84..d2cf24b 100644
--- a/hbase-protocol/src/main/protobuf/ZooKeeper.proto
+++ b/hbase-protocol/src/main/protobuf/ZooKeeper.proto
@@ -26,6 +26,7 @@ option java_generic_services = true;
option java_generate_equals_and_hash = true;
option optimize_for = SPEED;
+import "Table.proto";
import "HBase.proto";
import "ClusterStatus.proto";
[4/5] hbase git commit: HBASE-18431 Mitigate compatibility concerns
between branch-1.3 and branch-1.4
Posted by st...@apache.org.
HBASE-18431 Mitigate compatibility concerns between branch-1.3 and branch-1.4
We needed to move TableName out of HBase.proto to resolve a cycle after
repairing incompatible changes introduced into branch-1. Carry this forward.
Project: http://git-wip-us.apache.org/repos/asf/hbase/repo
Commit: http://git-wip-us.apache.org/repos/asf/hbase/commit/5e21e306
Tree: http://git-wip-us.apache.org/repos/asf/hbase/tree/5e21e306
Diff: http://git-wip-us.apache.org/repos/asf/hbase/diff/5e21e306
Branch: refs/heads/branch-2
Commit: 5e21e30682147a45c50ff7067ed40d21bb52239d
Parents: f67dae9
Author: Andrew Purtell <ap...@apache.org>
Authored: Mon Jul 31 16:12:02 2017 -0700
Committer: Michael Stack <st...@apache.org>
Committed: Tue Aug 1 17:05:14 2017 -0700
----------------------------------------------------------------------
.../hadoop/hbase/protobuf/ProtobufUtil.java | 7 +-
.../protobuf/generated/AccessControlProtos.java | 92 +--
.../hbase/protobuf/generated/ClientProtos.java | 383 ++++-----
.../hbase/protobuf/generated/HBaseProtos.java | 827 +++----------------
.../hbase/protobuf/generated/QuotaProtos.java | 40 +-
.../hbase/protobuf/generated/TableProtos.java | 607 ++++++++++++++
.../hbase/protobuf/generated/WALProtos.java | 145 ++--
.../protobuf/generated/ZooKeeperProtos.java | 112 +--
.../src/main/protobuf/AccessControl.proto | 2 +-
hbase-protocol/src/main/protobuf/Client.proto | 1 +
hbase-protocol/src/main/protobuf/HBase.proto | 9 +-
hbase-protocol/src/main/protobuf/Table.proto | 33 +
hbase-protocol/src/main/protobuf/WAL.proto | 1 +
.../src/main/protobuf/ZooKeeper.proto | 1 +
.../protobuf/generated/RSGroupAdminProtos.java | 359 ++++----
.../hbase/protobuf/generated/RSGroupProtos.java | 79 +-
.../hbase/rsgroup/RSGroupAdminEndpoint.java | 5 +-
.../hbase/rsgroup/RSGroupProtobufUtil.java | 5 +-
hbase-rsgroup/src/main/protobuf/RSGroup.proto | 1 +
.../src/main/protobuf/RSGroupAdmin.proto | 1 +
20 files changed, 1395 insertions(+), 1315 deletions(-)
----------------------------------------------------------------------
http://git-wip-us.apache.org/repos/asf/hbase/blob/5e21e306/hbase-client/src/main/java/org/apache/hadoop/hbase/protobuf/ProtobufUtil.java
----------------------------------------------------------------------
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/protobuf/ProtobufUtil.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/protobuf/ProtobufUtil.java
index fcf2c34..bc03699 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/protobuf/ProtobufUtil.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/protobuf/ProtobufUtil.java
@@ -78,6 +78,7 @@ import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameBytesPair;
import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier;
import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.RegionSpecifierType;
import org.apache.hadoop.hbase.protobuf.generated.MapReduceProtos;
+import org.apache.hadoop.hbase.protobuf.generated.TableProtos;
import org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos;
import org.apache.hadoop.hbase.util.Addressing;
import org.apache.hadoop.hbase.util.ByteStringer;
@@ -1707,13 +1708,13 @@ public final class ProtobufUtil {
", type=" + proto.getMutateType().toString();
}
- public static TableName toTableName(HBaseProtos.TableName tableNamePB) {
+ public static TableName toTableName(TableProtos.TableName tableNamePB) {
return TableName.valueOf(tableNamePB.getNamespace().asReadOnlyByteBuffer(),
tableNamePB.getQualifier().asReadOnlyByteBuffer());
}
- public static HBaseProtos.TableName toProtoTableName(TableName tableName) {
- return HBaseProtos.TableName.newBuilder()
+ public static TableProtos.TableName toProtoTableName(TableName tableName) {
+ return TableProtos.TableName.newBuilder()
.setNamespace(ByteStringer.wrap(tableName.getNamespace()))
.setQualifier(ByteStringer.wrap(tableName.getQualifier())).build();
}
http://git-wip-us.apache.org/repos/asf/hbase/blob/5e21e306/hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/AccessControlProtos.java
----------------------------------------------------------------------
diff --git a/hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/AccessControlProtos.java b/hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/AccessControlProtos.java
index b72e6e5..161282c 100644
--- a/hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/AccessControlProtos.java
+++ b/hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/AccessControlProtos.java
@@ -1285,11 +1285,11 @@ public final class AccessControlProtos {
/**
* <code>optional .hbase.pb.TableName table_name = 1;</code>
*/
- org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName getTableName();
+ org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName getTableName();
/**
* <code>optional .hbase.pb.TableName table_name = 1;</code>
*/
- org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder();
+ org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder getTableNameOrBuilder();
// optional bytes family = 2;
/**
@@ -1377,11 +1377,11 @@ public final class AccessControlProtos {
break;
}
case 10: {
- org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder subBuilder = null;
+ org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder subBuilder = null;
if (((bitField0_ & 0x00000001) == 0x00000001)) {
subBuilder = tableName_.toBuilder();
}
- tableName_ = input.readMessage(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.PARSER, extensionRegistry);
+ tableName_ = input.readMessage(org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.PARSER, extensionRegistry);
if (subBuilder != null) {
subBuilder.mergeFrom(tableName_);
tableName_ = subBuilder.buildPartial();
@@ -1477,7 +1477,7 @@ public final class AccessControlProtos {
private int bitField0_;
// optional .hbase.pb.TableName table_name = 1;
public static final int TABLE_NAME_FIELD_NUMBER = 1;
- private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName tableName_;
+ private org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName tableName_;
/**
* <code>optional .hbase.pb.TableName table_name = 1;</code>
*/
@@ -1487,13 +1487,13 @@ public final class AccessControlProtos {
/**
* <code>optional .hbase.pb.TableName table_name = 1;</code>
*/
- public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName getTableName() {
+ public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName getTableName() {
return tableName_;
}
/**
* <code>optional .hbase.pb.TableName table_name = 1;</code>
*/
- public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder() {
+ public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder getTableNameOrBuilder() {
return tableName_;
}
@@ -1552,7 +1552,7 @@ public final class AccessControlProtos {
}
private void initFields() {
- tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance();
+ tableName_ = org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance();
family_ = com.google.protobuf.ByteString.EMPTY;
qualifier_ = com.google.protobuf.ByteString.EMPTY;
action_ = java.util.Collections.emptyList();
@@ -1797,7 +1797,7 @@ public final class AccessControlProtos {
public Builder clear() {
super.clear();
if (tableNameBuilder_ == null) {
- tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance();
+ tableName_ = org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance();
} else {
tableNameBuilder_.clear();
}
@@ -1926,9 +1926,9 @@ public final class AccessControlProtos {
private int bitField0_;
// optional .hbase.pb.TableName table_name = 1;
- private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance();
+ private org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName tableName_ = org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance();
private com.google.protobuf.SingleFieldBuilder<
- org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder> tableNameBuilder_;
+ org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder> tableNameBuilder_;
/**
* <code>optional .hbase.pb.TableName table_name = 1;</code>
*/
@@ -1938,7 +1938,7 @@ public final class AccessControlProtos {
/**
* <code>optional .hbase.pb.TableName table_name = 1;</code>
*/
- public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName getTableName() {
+ public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName getTableName() {
if (tableNameBuilder_ == null) {
return tableName_;
} else {
@@ -1948,7 +1948,7 @@ public final class AccessControlProtos {
/**
* <code>optional .hbase.pb.TableName table_name = 1;</code>
*/
- public Builder setTableName(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName value) {
+ public Builder setTableName(org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName value) {
if (tableNameBuilder_ == null) {
if (value == null) {
throw new NullPointerException();
@@ -1965,7 +1965,7 @@ public final class AccessControlProtos {
* <code>optional .hbase.pb.TableName table_name = 1;</code>
*/
public Builder setTableName(
- org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder builderForValue) {
+ org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder builderForValue) {
if (tableNameBuilder_ == null) {
tableName_ = builderForValue.build();
onChanged();
@@ -1978,12 +1978,12 @@ public final class AccessControlProtos {
/**
* <code>optional .hbase.pb.TableName table_name = 1;</code>
*/
- public Builder mergeTableName(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName value) {
+ public Builder mergeTableName(org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName value) {
if (tableNameBuilder_ == null) {
if (((bitField0_ & 0x00000001) == 0x00000001) &&
- tableName_ != org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance()) {
+ tableName_ != org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance()) {
tableName_ =
- org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.newBuilder(tableName_).mergeFrom(value).buildPartial();
+ org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.newBuilder(tableName_).mergeFrom(value).buildPartial();
} else {
tableName_ = value;
}
@@ -1999,7 +1999,7 @@ public final class AccessControlProtos {
*/
public Builder clearTableName() {
if (tableNameBuilder_ == null) {
- tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance();
+ tableName_ = org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance();
onChanged();
} else {
tableNameBuilder_.clear();
@@ -2010,7 +2010,7 @@ public final class AccessControlProtos {
/**
* <code>optional .hbase.pb.TableName table_name = 1;</code>
*/
- public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder getTableNameBuilder() {
+ public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder getTableNameBuilder() {
bitField0_ |= 0x00000001;
onChanged();
return getTableNameFieldBuilder().getBuilder();
@@ -2018,7 +2018,7 @@ public final class AccessControlProtos {
/**
* <code>optional .hbase.pb.TableName table_name = 1;</code>
*/
- public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder() {
+ public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder getTableNameOrBuilder() {
if (tableNameBuilder_ != null) {
return tableNameBuilder_.getMessageOrBuilder();
} else {
@@ -2029,11 +2029,11 @@ public final class AccessControlProtos {
* <code>optional .hbase.pb.TableName table_name = 1;</code>
*/
private com.google.protobuf.SingleFieldBuilder<
- org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder>
+ org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder>
getTableNameFieldBuilder() {
if (tableNameBuilder_ == null) {
tableNameBuilder_ = new com.google.protobuf.SingleFieldBuilder<
- org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder>(
+ org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder>(
tableName_,
getParentForChildren(),
isClean());
@@ -7462,11 +7462,11 @@ public final class AccessControlProtos {
/**
* <code>optional .hbase.pb.TableName table_name = 2;</code>
*/
- org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName getTableName();
+ org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName getTableName();
/**
* <code>optional .hbase.pb.TableName table_name = 2;</code>
*/
- org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder();
+ org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder getTableNameOrBuilder();
// optional bytes namespace_name = 3;
/**
@@ -7541,11 +7541,11 @@ public final class AccessControlProtos {
break;
}
case 18: {
- org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder subBuilder = null;
+ org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder subBuilder = null;
if (((bitField0_ & 0x00000002) == 0x00000002)) {
subBuilder = tableName_.toBuilder();
}
- tableName_ = input.readMessage(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.PARSER, extensionRegistry);
+ tableName_ = input.readMessage(org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.PARSER, extensionRegistry);
if (subBuilder != null) {
subBuilder.mergeFrom(tableName_);
tableName_ = subBuilder.buildPartial();
@@ -7616,7 +7616,7 @@ public final class AccessControlProtos {
// optional .hbase.pb.TableName table_name = 2;
public static final int TABLE_NAME_FIELD_NUMBER = 2;
- private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName tableName_;
+ private org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName tableName_;
/**
* <code>optional .hbase.pb.TableName table_name = 2;</code>
*/
@@ -7626,13 +7626,13 @@ public final class AccessControlProtos {
/**
* <code>optional .hbase.pb.TableName table_name = 2;</code>
*/
- public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName getTableName() {
+ public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName getTableName() {
return tableName_;
}
/**
* <code>optional .hbase.pb.TableName table_name = 2;</code>
*/
- public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder() {
+ public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder getTableNameOrBuilder() {
return tableName_;
}
@@ -7654,7 +7654,7 @@ public final class AccessControlProtos {
private void initFields() {
type_ = org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.Permission.Type.Global;
- tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance();
+ tableName_ = org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance();
namespaceName_ = com.google.protobuf.ByteString.EMPTY;
}
private byte memoizedIsInitialized = -1;
@@ -7881,7 +7881,7 @@ public final class AccessControlProtos {
type_ = org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.Permission.Type.Global;
bitField0_ = (bitField0_ & ~0x00000001);
if (tableNameBuilder_ == null) {
- tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance();
+ tableName_ = org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance();
} else {
tableNameBuilder_.clear();
}
@@ -8027,9 +8027,9 @@ public final class AccessControlProtos {
}
// optional .hbase.pb.TableName table_name = 2;
- private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance();
+ private org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName tableName_ = org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance();
private com.google.protobuf.SingleFieldBuilder<
- org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder> tableNameBuilder_;
+ org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder> tableNameBuilder_;
/**
* <code>optional .hbase.pb.TableName table_name = 2;</code>
*/
@@ -8039,7 +8039,7 @@ public final class AccessControlProtos {
/**
* <code>optional .hbase.pb.TableName table_name = 2;</code>
*/
- public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName getTableName() {
+ public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName getTableName() {
if (tableNameBuilder_ == null) {
return tableName_;
} else {
@@ -8049,7 +8049,7 @@ public final class AccessControlProtos {
/**
* <code>optional .hbase.pb.TableName table_name = 2;</code>
*/
- public Builder setTableName(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName value) {
+ public Builder setTableName(org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName value) {
if (tableNameBuilder_ == null) {
if (value == null) {
throw new NullPointerException();
@@ -8066,7 +8066,7 @@ public final class AccessControlProtos {
* <code>optional .hbase.pb.TableName table_name = 2;</code>
*/
public Builder setTableName(
- org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder builderForValue) {
+ org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder builderForValue) {
if (tableNameBuilder_ == null) {
tableName_ = builderForValue.build();
onChanged();
@@ -8079,12 +8079,12 @@ public final class AccessControlProtos {
/**
* <code>optional .hbase.pb.TableName table_name = 2;</code>
*/
- public Builder mergeTableName(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName value) {
+ public Builder mergeTableName(org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName value) {
if (tableNameBuilder_ == null) {
if (((bitField0_ & 0x00000002) == 0x00000002) &&
- tableName_ != org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance()) {
+ tableName_ != org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance()) {
tableName_ =
- org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.newBuilder(tableName_).mergeFrom(value).buildPartial();
+ org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.newBuilder(tableName_).mergeFrom(value).buildPartial();
} else {
tableName_ = value;
}
@@ -8100,7 +8100,7 @@ public final class AccessControlProtos {
*/
public Builder clearTableName() {
if (tableNameBuilder_ == null) {
- tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance();
+ tableName_ = org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance();
onChanged();
} else {
tableNameBuilder_.clear();
@@ -8111,7 +8111,7 @@ public final class AccessControlProtos {
/**
* <code>optional .hbase.pb.TableName table_name = 2;</code>
*/
- public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder getTableNameBuilder() {
+ public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder getTableNameBuilder() {
bitField0_ |= 0x00000002;
onChanged();
return getTableNameFieldBuilder().getBuilder();
@@ -8119,7 +8119,7 @@ public final class AccessControlProtos {
/**
* <code>optional .hbase.pb.TableName table_name = 2;</code>
*/
- public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder() {
+ public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder getTableNameOrBuilder() {
if (tableNameBuilder_ != null) {
return tableNameBuilder_.getMessageOrBuilder();
} else {
@@ -8130,11 +8130,11 @@ public final class AccessControlProtos {
* <code>optional .hbase.pb.TableName table_name = 2;</code>
*/
private com.google.protobuf.SingleFieldBuilder<
- org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder>
+ org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder>
getTableNameFieldBuilder() {
if (tableNameBuilder_ == null) {
tableNameBuilder_ = new com.google.protobuf.SingleFieldBuilder<
- org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder>(
+ org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder>(
tableName_,
getParentForChildren(),
isClean());
@@ -10499,7 +10499,7 @@ public final class AccessControlProtos {
descriptor;
static {
java.lang.String[] descriptorData = {
- "\n\023AccessControl.proto\022\010hbase.pb\032\013HBase.p" +
+ "\n\023AccessControl.proto\022\010hbase.pb\032\013Table.p" +
"roto\"\314\002\n\nPermission\022\'\n\004type\030\001 \002(\0162\031.hbas" +
"e.pb.Permission.Type\0225\n\021global_permissio" +
"n\030\002 \001(\0132\032.hbase.pb.GlobalPermission\022;\n\024n" +
@@ -10648,7 +10648,7 @@ public final class AccessControlProtos {
com.google.protobuf.Descriptors.FileDescriptor
.internalBuildGeneratedFileFrom(descriptorData,
new com.google.protobuf.Descriptors.FileDescriptor[] {
- org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.getDescriptor(),
+ org.apache.hadoop.hbase.protobuf.generated.TableProtos.getDescriptor(),
}, assigner);
}
http://git-wip-us.apache.org/repos/asf/hbase/blob/5e21e306/hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/ClientProtos.java
----------------------------------------------------------------------
diff --git a/hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/ClientProtos.java b/hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/ClientProtos.java
index e25064f..ad4f690 100644
--- a/hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/ClientProtos.java
+++ b/hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/ClientProtos.java
@@ -22636,7 +22636,7 @@ public final class ClientProtos {
* </pre>
*/
private com.google.protobuf.SingleFieldBuilder<
- org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Cursor, org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Cursor.Builder, org.apache.hadoop.hbase.protobuf.generated.ClientProtos.CursorOrBuilder>
+ org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Cursor, org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Cursor.Builder, org.apache.hadoop.hbase.protobuf.generated.ClientProtos.CursorOrBuilder>
getCursorFieldBuilder() {
if (cursorBuilder_ == null) {
cursorBuilder_ = new com.google.protobuf.SingleFieldBuilder<
@@ -26090,11 +26090,11 @@ public final class ClientProtos {
/**
* <code>required .hbase.pb.TableName table_name = 1;</code>
*/
- org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName getTableName();
+ org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName getTableName();
/**
* <code>required .hbase.pb.TableName table_name = 1;</code>
*/
- org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder();
+ org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder getTableNameOrBuilder();
// optional .hbase.pb.RegionSpecifier region = 2;
/**
@@ -26162,11 +26162,11 @@ public final class ClientProtos {
break;
}
case 10: {
- org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder subBuilder = null;
+ org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder subBuilder = null;
if (((bitField0_ & 0x00000001) == 0x00000001)) {
subBuilder = tableName_.toBuilder();
}
- tableName_ = input.readMessage(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.PARSER, extensionRegistry);
+ tableName_ = input.readMessage(org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.PARSER, extensionRegistry);
if (subBuilder != null) {
subBuilder.mergeFrom(tableName_);
tableName_ = subBuilder.buildPartial();
@@ -26229,7 +26229,7 @@ public final class ClientProtos {
private int bitField0_;
// required .hbase.pb.TableName table_name = 1;
public static final int TABLE_NAME_FIELD_NUMBER = 1;
- private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName tableName_;
+ private org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName tableName_;
/**
* <code>required .hbase.pb.TableName table_name = 1;</code>
*/
@@ -26239,13 +26239,13 @@ public final class ClientProtos {
/**
* <code>required .hbase.pb.TableName table_name = 1;</code>
*/
- public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName getTableName() {
+ public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName getTableName() {
return tableName_;
}
/**
* <code>required .hbase.pb.TableName table_name = 1;</code>
*/
- public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder() {
+ public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder getTableNameOrBuilder() {
return tableName_;
}
@@ -26272,7 +26272,7 @@ public final class ClientProtos {
}
private void initFields() {
- tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance();
+ tableName_ = org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance();
region_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance();
}
private byte memoizedIsInitialized = -1;
@@ -26490,7 +26490,7 @@ public final class ClientProtos {
public Builder clear() {
super.clear();
if (tableNameBuilder_ == null) {
- tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance();
+ tableName_ = org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance();
} else {
tableNameBuilder_.clear();
}
@@ -26609,9 +26609,9 @@ public final class ClientProtos {
private int bitField0_;
// required .hbase.pb.TableName table_name = 1;
- private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance();
+ private org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName tableName_ = org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance();
private com.google.protobuf.SingleFieldBuilder<
- org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder> tableNameBuilder_;
+ org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder> tableNameBuilder_;
/**
* <code>required .hbase.pb.TableName table_name = 1;</code>
*/
@@ -26621,7 +26621,7 @@ public final class ClientProtos {
/**
* <code>required .hbase.pb.TableName table_name = 1;</code>
*/
- public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName getTableName() {
+ public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName getTableName() {
if (tableNameBuilder_ == null) {
return tableName_;
} else {
@@ -26631,7 +26631,7 @@ public final class ClientProtos {
/**
* <code>required .hbase.pb.TableName table_name = 1;</code>
*/
- public Builder setTableName(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName value) {
+ public Builder setTableName(org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName value) {
if (tableNameBuilder_ == null) {
if (value == null) {
throw new NullPointerException();
@@ -26648,7 +26648,7 @@ public final class ClientProtos {
* <code>required .hbase.pb.TableName table_name = 1;</code>
*/
public Builder setTableName(
- org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder builderForValue) {
+ org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder builderForValue) {
if (tableNameBuilder_ == null) {
tableName_ = builderForValue.build();
onChanged();
@@ -26661,12 +26661,12 @@ public final class ClientProtos {
/**
* <code>required .hbase.pb.TableName table_name = 1;</code>
*/
- public Builder mergeTableName(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName value) {
+ public Builder mergeTableName(org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName value) {
if (tableNameBuilder_ == null) {
if (((bitField0_ & 0x00000001) == 0x00000001) &&
- tableName_ != org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance()) {
+ tableName_ != org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance()) {
tableName_ =
- org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.newBuilder(tableName_).mergeFrom(value).buildPartial();
+ org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.newBuilder(tableName_).mergeFrom(value).buildPartial();
} else {
tableName_ = value;
}
@@ -26682,7 +26682,7 @@ public final class ClientProtos {
*/
public Builder clearTableName() {
if (tableNameBuilder_ == null) {
- tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance();
+ tableName_ = org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance();
onChanged();
} else {
tableNameBuilder_.clear();
@@ -26693,7 +26693,7 @@ public final class ClientProtos {
/**
* <code>required .hbase.pb.TableName table_name = 1;</code>
*/
- public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder getTableNameBuilder() {
+ public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder getTableNameBuilder() {
bitField0_ |= 0x00000001;
onChanged();
return getTableNameFieldBuilder().getBuilder();
@@ -26701,7 +26701,7 @@ public final class ClientProtos {
/**
* <code>required .hbase.pb.TableName table_name = 1;</code>
*/
- public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder() {
+ public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder getTableNameOrBuilder() {
if (tableNameBuilder_ != null) {
return tableNameBuilder_.getMessageOrBuilder();
} else {
@@ -26712,11 +26712,11 @@ public final class ClientProtos {
* <code>required .hbase.pb.TableName table_name = 1;</code>
*/
private com.google.protobuf.SingleFieldBuilder<
- org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder>
+ org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder>
getTableNameFieldBuilder() {
if (tableNameBuilder_ == null) {
tableNameBuilder_ = new com.google.protobuf.SingleFieldBuilder<
- org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder>(
+ org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder>(
tableName_,
getParentForChildren(),
isClean());
@@ -40995,174 +40995,174 @@ public final class ClientProtos {
descriptor;
static {
java.lang.String[] descriptorData = {
- "\n\014Client.proto\022\010hbase.pb\032\013HBase.proto\032\014F" +
- "ilter.proto\032\nCell.proto\032\020Comparator.prot" +
- "o\032\017MapReduce.proto\"\037\n\016Authorizations\022\r\n\005" +
- "label\030\001 \003(\t\"$\n\016CellVisibility\022\022\n\nexpress" +
- "ion\030\001 \002(\t\"+\n\006Column\022\016\n\006family\030\001 \002(\014\022\021\n\tq" +
- "ualifier\030\002 \003(\014\"\276\003\n\003Get\022\013\n\003row\030\001 \002(\014\022 \n\006c" +
- "olumn\030\002 \003(\0132\020.hbase.pb.Column\022*\n\tattribu" +
- "te\030\003 \003(\0132\027.hbase.pb.NameBytesPair\022 \n\006fil" +
- "ter\030\004 \001(\0132\020.hbase.pb.Filter\022\'\n\ntime_rang" +
- "e\030\005 \001(\0132\023.hbase.pb.TimeRange\022\027\n\014max_vers",
- "ions\030\006 \001(\r:\0011\022\032\n\014cache_blocks\030\007 \001(\010:\004tru" +
- "e\022\023\n\013store_limit\030\010 \001(\r\022\024\n\014store_offset\030\t" +
- " \001(\r\022\035\n\016existence_only\030\n \001(\010:\005false\0222\n\013c" +
- "onsistency\030\014 \001(\0162\025.hbase.pb.Consistency:" +
- "\006STRONG\0226\n\rcf_time_range\030\r \003(\0132\037.hbase.p" +
- "b.ColumnFamilyTimeRange\022&\n\036load_column_f" +
- "amilies_on_demand\030\016 \001(\010\"\203\001\n\006Result\022\034\n\004ce" +
- "ll\030\001 \003(\0132\016.hbase.pb.Cell\022\035\n\025associated_c" +
- "ell_count\030\002 \001(\005\022\016\n\006exists\030\003 \001(\010\022\024\n\005stale" +
- "\030\004 \001(\010:\005false\022\026\n\007partial\030\005 \001(\010:\005false\"S\n",
- "\nGetRequest\022)\n\006region\030\001 \002(\0132\031.hbase.pb.R" +
- "egionSpecifier\022\032\n\003get\030\002 \002(\0132\r.hbase.pb.G" +
- "et\"/\n\013GetResponse\022 \n\006result\030\001 \001(\0132\020.hbas" +
- "e.pb.Result\"\222\001\n\tCondition\022\013\n\003row\030\001 \002(\014\022\016" +
- "\n\006family\030\002 \002(\014\022\021\n\tqualifier\030\003 \002(\014\022+\n\014com" +
- "pare_type\030\004 \002(\0162\025.hbase.pb.CompareType\022(" +
- "\n\ncomparator\030\005 \002(\0132\024.hbase.pb.Comparator" +
- "\"\364\006\n\rMutationProto\022\013\n\003row\030\001 \001(\014\0229\n\013mutat" +
- "e_type\030\002 \001(\0162$.hbase.pb.MutationProto.Mu" +
- "tationType\0229\n\014column_value\030\003 \003(\0132#.hbase",
- ".pb.MutationProto.ColumnValue\022\021\n\ttimesta" +
- "mp\030\004 \001(\004\022*\n\tattribute\030\005 \003(\0132\027.hbase.pb.N" +
- "ameBytesPair\022C\n\ndurability\030\006 \001(\0162\".hbase" +
- ".pb.MutationProto.Durability:\013USE_DEFAUL" +
- "T\022\'\n\ntime_range\030\007 \001(\0132\023.hbase.pb.TimeRan" +
- "ge\022\035\n\025associated_cell_count\030\010 \001(\005\022\r\n\005non" +
- "ce\030\t \001(\004\032\371\001\n\013ColumnValue\022\016\n\006family\030\001 \002(\014" +
- "\022K\n\017qualifier_value\030\002 \003(\01322.hbase.pb.Mut" +
- "ationProto.ColumnValue.QualifierValue\032\214\001" +
- "\n\016QualifierValue\022\021\n\tqualifier\030\001 \001(\014\022\r\n\005v",
- "alue\030\002 \001(\014\022\021\n\ttimestamp\030\003 \001(\004\0227\n\013delete_" +
- "type\030\004 \001(\0162\".hbase.pb.MutationProto.Dele" +
- "teType\022\014\n\004tags\030\005 \001(\014\"W\n\nDurability\022\017\n\013US" +
- "E_DEFAULT\020\000\022\014\n\010SKIP_WAL\020\001\022\r\n\tASYNC_WAL\020\002" +
- "\022\014\n\010SYNC_WAL\020\003\022\r\n\tFSYNC_WAL\020\004\">\n\014Mutatio" +
- "nType\022\n\n\006APPEND\020\000\022\r\n\tINCREMENT\020\001\022\007\n\003PUT\020" +
- "\002\022\n\n\006DELETE\020\003\"p\n\nDeleteType\022\026\n\022DELETE_ON" +
- "E_VERSION\020\000\022\034\n\030DELETE_MULTIPLE_VERSIONS\020" +
- "\001\022\021\n\rDELETE_FAMILY\020\002\022\031\n\025DELETE_FAMILY_VE" +
- "RSION\020\003\"\242\001\n\rMutateRequest\022)\n\006region\030\001 \002(",
- "\0132\031.hbase.pb.RegionSpecifier\022)\n\010mutation" +
- "\030\002 \002(\0132\027.hbase.pb.MutationProto\022&\n\tcondi" +
- "tion\030\003 \001(\0132\023.hbase.pb.Condition\022\023\n\013nonce" +
- "_group\030\004 \001(\004\"E\n\016MutateResponse\022 \n\006result" +
- "\030\001 \001(\0132\020.hbase.pb.Result\022\021\n\tprocessed\030\002 " +
- "\001(\010\"\246\006\n\004Scan\022 \n\006column\030\001 \003(\0132\020.hbase.pb." +
- "Column\022*\n\tattribute\030\002 \003(\0132\027.hbase.pb.Nam" +
- "eBytesPair\022\021\n\tstart_row\030\003 \001(\014\022\020\n\010stop_ro" +
- "w\030\004 \001(\014\022 \n\006filter\030\005 \001(\0132\020.hbase.pb.Filte" +
- "r\022\'\n\ntime_range\030\006 \001(\0132\023.hbase.pb.TimeRan",
- "ge\022\027\n\014max_versions\030\007 \001(\r:\0011\022\032\n\014cache_blo" +
- "cks\030\010 \001(\010:\004true\022\022\n\nbatch_size\030\t \001(\r\022\027\n\017m" +
- "ax_result_size\030\n \001(\004\022\023\n\013store_limit\030\013 \001(" +
- "\r\022\024\n\014store_offset\030\014 \001(\r\022&\n\036load_column_f" +
- "amilies_on_demand\030\r \001(\010\022\021\n\005small\030\016 \001(\010B\002" +
- "\030\001\022\027\n\010reversed\030\017 \001(\010:\005false\0222\n\013consisten" +
- "cy\030\020 \001(\0162\025.hbase.pb.Consistency:\006STRONG\022" +
- "\017\n\007caching\030\021 \001(\r\022\035\n\025allow_partial_result" +
- "s\030\022 \001(\010\0226\n\rcf_time_range\030\023 \003(\0132\037.hbase.p" +
- "b.ColumnFamilyTimeRange\022\032\n\017mvcc_read_poi",
- "nt\030\024 \001(\004:\0010\022\037\n\021include_start_row\030\025 \001(\010:\004" +
- "true\022\037\n\020include_stop_row\030\026 \001(\010:\005false\0222\n" +
- "\010readType\030\027 \001(\0162\027.hbase.pb.Scan.ReadType" +
- ":\007DEFAULT\022!\n\022need_cursor_result\030\030 \001(\010:\005f" +
- "alse\".\n\010ReadType\022\013\n\007DEFAULT\020\000\022\n\n\006STREAM\020" +
- "\001\022\t\n\005PREAD\020\002\"\300\002\n\013ScanRequest\022)\n\006region\030\001" +
- " \001(\0132\031.hbase.pb.RegionSpecifier\022\034\n\004scan\030" +
- "\002 \001(\0132\016.hbase.pb.Scan\022\022\n\nscanner_id\030\003 \001(" +
- "\004\022\026\n\016number_of_rows\030\004 \001(\r\022\025\n\rclose_scann" +
- "er\030\005 \001(\010\022\025\n\rnext_call_seq\030\006 \001(\004\022\037\n\027clien",
- "t_handles_partials\030\007 \001(\010\022!\n\031client_handl" +
- "es_heartbeats\030\010 \001(\010\022\032\n\022track_scan_metric" +
- "s\030\t \001(\010\022\024\n\005renew\030\n \001(\010:\005false\022\030\n\rlimit_o" +
- "f_rows\030\013 \001(\r:\0010\"\025\n\006Cursor\022\013\n\003row\030\001 \001(\014\"\330" +
- "\002\n\014ScanResponse\022\030\n\020cells_per_result\030\001 \003(" +
- "\r\022\022\n\nscanner_id\030\002 \001(\004\022\024\n\014more_results\030\003 " +
- "\001(\010\022\013\n\003ttl\030\004 \001(\r\022!\n\007results\030\005 \003(\0132\020.hbas" +
- "e.pb.Result\022\r\n\005stale\030\006 \001(\010\022\037\n\027partial_fl" +
- "ag_per_result\030\007 \003(\010\022\036\n\026more_results_in_r" +
- "egion\030\010 \001(\010\022\031\n\021heartbeat_message\030\t \001(\010\022+",
- "\n\014scan_metrics\030\n \001(\0132\025.hbase.pb.ScanMetr" +
- "ics\022\032\n\017mvcc_read_point\030\013 \001(\004:\0010\022 \n\006curso" +
- "r\030\014 \001(\0132\020.hbase.pb.Cursor\"\240\002\n\024BulkLoadHF" +
- "ileRequest\022)\n\006region\030\001 \002(\0132\031.hbase.pb.Re" +
- "gionSpecifier\022>\n\013family_path\030\002 \003(\0132).hba" +
- "se.pb.BulkLoadHFileRequest.FamilyPath\022\026\n" +
- "\016assign_seq_num\030\003 \001(\010\022+\n\010fs_token\030\004 \001(\0132" +
- "\031.hbase.pb.DelegationToken\022\022\n\nbulk_token" +
- "\030\005 \001(\t\022\030\n\tcopy_file\030\006 \001(\010:\005false\032*\n\nFami" +
- "lyPath\022\016\n\006family\030\001 \002(\014\022\014\n\004path\030\002 \002(\t\"\'\n\025",
- "BulkLoadHFileResponse\022\016\n\006loaded\030\001 \002(\010\"V\n" +
- "\017DelegationToken\022\022\n\nidentifier\030\001 \001(\014\022\020\n\010" +
- "password\030\002 \001(\014\022\014\n\004kind\030\003 \001(\t\022\017\n\007service\030" +
- "\004 \001(\t\"l\n\026PrepareBulkLoadRequest\022\'\n\ntable" +
- "_name\030\001 \002(\0132\023.hbase.pb.TableName\022)\n\006regi" +
- "on\030\002 \001(\0132\031.hbase.pb.RegionSpecifier\"-\n\027P" +
- "repareBulkLoadResponse\022\022\n\nbulk_token\030\001 \002" +
- "(\t\"W\n\026CleanupBulkLoadRequest\022\022\n\nbulk_tok" +
- "en\030\001 \002(\t\022)\n\006region\030\002 \001(\0132\031.hbase.pb.Regi" +
- "onSpecifier\"\031\n\027CleanupBulkLoadResponse\"a",
- "\n\026CoprocessorServiceCall\022\013\n\003row\030\001 \002(\014\022\024\n" +
- "\014service_name\030\002 \002(\t\022\023\n\013method_name\030\003 \002(\t" +
- "\022\017\n\007request\030\004 \002(\014\"B\n\030CoprocessorServiceR" +
- "esult\022&\n\005value\030\001 \001(\0132\027.hbase.pb.NameByte" +
- "sPair\"v\n\031CoprocessorServiceRequest\022)\n\006re" +
- "gion\030\001 \002(\0132\031.hbase.pb.RegionSpecifier\022.\n" +
- "\004call\030\002 \002(\0132 .hbase.pb.CoprocessorServic" +
- "eCall\"o\n\032CoprocessorServiceResponse\022)\n\006r" +
- "egion\030\001 \002(\0132\031.hbase.pb.RegionSpecifier\022&" +
- "\n\005value\030\002 \002(\0132\027.hbase.pb.NameBytesPair\"\226",
- "\001\n\006Action\022\r\n\005index\030\001 \001(\r\022)\n\010mutation\030\002 \001" +
- "(\0132\027.hbase.pb.MutationProto\022\032\n\003get\030\003 \001(\013" +
- "2\r.hbase.pb.Get\0226\n\014service_call\030\004 \001(\0132 ." +
- "hbase.pb.CoprocessorServiceCall\"k\n\014Regio" +
- "nAction\022)\n\006region\030\001 \002(\0132\031.hbase.pb.Regio" +
- "nSpecifier\022\016\n\006atomic\030\002 \001(\010\022 \n\006action\030\003 \003" +
- "(\0132\020.hbase.pb.Action\"c\n\017RegionLoadStats\022" +
- "\027\n\014memstoreLoad\030\001 \001(\005:\0010\022\030\n\rheapOccupanc" +
- "y\030\002 \001(\005:\0010\022\035\n\022compactionPressure\030\003 \001(\005:\001" +
- "0\"j\n\024MultiRegionLoadStats\022)\n\006region\030\001 \003(",
- "\0132\031.hbase.pb.RegionSpecifier\022\'\n\004stat\030\002 \003" +
- "(\0132\031.hbase.pb.RegionLoadStats\"\336\001\n\021Result" +
- "OrException\022\r\n\005index\030\001 \001(\r\022 \n\006result\030\002 \001" +
- "(\0132\020.hbase.pb.Result\022*\n\texception\030\003 \001(\0132" +
- "\027.hbase.pb.NameBytesPair\022:\n\016service_resu" +
- "lt\030\004 \001(\0132\".hbase.pb.CoprocessorServiceRe" +
- "sult\0220\n\tloadStats\030\005 \001(\0132\031.hbase.pb.Regio" +
- "nLoadStatsB\002\030\001\"x\n\022RegionActionResult\0226\n\021" +
- "resultOrException\030\001 \003(\0132\033.hbase.pb.Resul" +
- "tOrException\022*\n\texception\030\002 \001(\0132\027.hbase.",
- "pb.NameBytesPair\"x\n\014MultiRequest\022,\n\014regi" +
- "onAction\030\001 \003(\0132\026.hbase.pb.RegionAction\022\022" +
- "\n\nnonceGroup\030\002 \001(\004\022&\n\tcondition\030\003 \001(\0132\023." +
- "hbase.pb.Condition\"\226\001\n\rMultiResponse\0228\n\022" +
- "regionActionResult\030\001 \003(\0132\034.hbase.pb.Regi" +
- "onActionResult\022\021\n\tprocessed\030\002 \001(\010\0228\n\020reg" +
- "ionStatistics\030\003 \001(\0132\036.hbase.pb.MultiRegi" +
- "onLoadStats*\'\n\013Consistency\022\n\n\006STRONG\020\000\022\014" +
- "\n\010TIMELINE\020\0012\263\005\n\rClientService\0222\n\003Get\022\024." +
- "hbase.pb.GetRequest\032\025.hbase.pb.GetRespon",
- "se\022;\n\006Mutate\022\027.hbase.pb.MutateRequest\032\030." +
- "hbase.pb.MutateResponse\0225\n\004Scan\022\025.hbase." +
- "pb.ScanRequest\032\026.hbase.pb.ScanResponse\022P" +
- "\n\rBulkLoadHFile\022\036.hbase.pb.BulkLoadHFile" +
- "Request\032\037.hbase.pb.BulkLoadHFileResponse" +
- "\022V\n\017PrepareBulkLoad\022 .hbase.pb.PrepareBu" +
- "lkLoadRequest\032!.hbase.pb.PrepareBulkLoad" +
- "Response\022V\n\017CleanupBulkLoad\022 .hbase.pb.C" +
- "leanupBulkLoadRequest\032!.hbase.pb.Cleanup" +
- "BulkLoadResponse\022X\n\013ExecService\022#.hbase.",
- "pb.CoprocessorServiceRequest\032$.hbase.pb." +
- "CoprocessorServiceResponse\022d\n\027ExecRegion" +
- "ServerService\022#.hbase.pb.CoprocessorServ" +
- "iceRequest\032$.hbase.pb.CoprocessorService" +
- "Response\0228\n\005Multi\022\026.hbase.pb.MultiReques" +
- "t\032\027.hbase.pb.MultiResponseBB\n*org.apache" +
- ".hadoop.hbase.protobuf.generatedB\014Client" +
- "ProtosH\001\210\001\001\240\001\001"
+ "\n\014Client.proto\022\010hbase.pb\032\013Table.proto\032\013H" +
+ "Base.proto\032\014Filter.proto\032\nCell.proto\032\020Co" +
+ "mparator.proto\032\017MapReduce.proto\"\037\n\016Autho" +
+ "rizations\022\r\n\005label\030\001 \003(\t\"$\n\016CellVisibili" +
+ "ty\022\022\n\nexpression\030\001 \002(\t\"+\n\006Column\022\016\n\006fami" +
+ "ly\030\001 \002(\014\022\021\n\tqualifier\030\002 \003(\014\"\276\003\n\003Get\022\013\n\003r" +
+ "ow\030\001 \002(\014\022 \n\006column\030\002 \003(\0132\020.hbase.pb.Colu" +
+ "mn\022*\n\tattribute\030\003 \003(\0132\027.hbase.pb.NameByt" +
+ "esPair\022 \n\006filter\030\004 \001(\0132\020.hbase.pb.Filter" +
+ "\022\'\n\ntime_range\030\005 \001(\0132\023.hbase.pb.TimeRang",
+ "e\022\027\n\014max_versions\030\006 \001(\r:\0011\022\032\n\014cache_bloc" +
+ "ks\030\007 \001(\010:\004true\022\023\n\013store_limit\030\010 \001(\r\022\024\n\014s" +
+ "tore_offset\030\t \001(\r\022\035\n\016existence_only\030\n \001(" +
+ "\010:\005false\0222\n\013consistency\030\014 \001(\0162\025.hbase.pb" +
+ ".Consistency:\006STRONG\0226\n\rcf_time_range\030\r " +
+ "\003(\0132\037.hbase.pb.ColumnFamilyTimeRange\022&\n\036" +
+ "load_column_families_on_demand\030\016 \001(\010\"\203\001\n" +
+ "\006Result\022\034\n\004cell\030\001 \003(\0132\016.hbase.pb.Cell\022\035\n" +
+ "\025associated_cell_count\030\002 \001(\005\022\016\n\006exists\030\003" +
+ " \001(\010\022\024\n\005stale\030\004 \001(\010:\005false\022\026\n\007partial\030\005 ",
+ "\001(\010:\005false\"S\n\nGetRequest\022)\n\006region\030\001 \002(\013" +
+ "2\031.hbase.pb.RegionSpecifier\022\032\n\003get\030\002 \002(\013" +
+ "2\r.hbase.pb.Get\"/\n\013GetResponse\022 \n\006result" +
+ "\030\001 \001(\0132\020.hbase.pb.Result\"\222\001\n\tCondition\022\013" +
+ "\n\003row\030\001 \002(\014\022\016\n\006family\030\002 \002(\014\022\021\n\tqualifier" +
+ "\030\003 \002(\014\022+\n\014compare_type\030\004 \002(\0162\025.hbase.pb." +
+ "CompareType\022(\n\ncomparator\030\005 \002(\0132\024.hbase." +
+ "pb.Comparator\"\364\006\n\rMutationProto\022\013\n\003row\030\001" +
+ " \001(\014\0229\n\013mutate_type\030\002 \001(\0162$.hbase.pb.Mut" +
+ "ationProto.MutationType\0229\n\014column_value\030",
+ "\003 \003(\0132#.hbase.pb.MutationProto.ColumnVal" +
+ "ue\022\021\n\ttimestamp\030\004 \001(\004\022*\n\tattribute\030\005 \003(\013" +
+ "2\027.hbase.pb.NameBytesPair\022C\n\ndurability\030" +
+ "\006 \001(\0162\".hbase.pb.MutationProto.Durabilit" +
+ "y:\013USE_DEFAULT\022\'\n\ntime_range\030\007 \001(\0132\023.hba" +
+ "se.pb.TimeRange\022\035\n\025associated_cell_count" +
+ "\030\010 \001(\005\022\r\n\005nonce\030\t \001(\004\032\371\001\n\013ColumnValue\022\016\n" +
+ "\006family\030\001 \002(\014\022K\n\017qualifier_value\030\002 \003(\01322" +
+ ".hbase.pb.MutationProto.ColumnValue.Qual" +
+ "ifierValue\032\214\001\n\016QualifierValue\022\021\n\tqualifi",
+ "er\030\001 \001(\014\022\r\n\005value\030\002 \001(\014\022\021\n\ttimestamp\030\003 \001" +
+ "(\004\0227\n\013delete_type\030\004 \001(\0162\".hbase.pb.Mutat" +
+ "ionProto.DeleteType\022\014\n\004tags\030\005 \001(\014\"W\n\nDur" +
+ "ability\022\017\n\013USE_DEFAULT\020\000\022\014\n\010SKIP_WAL\020\001\022\r" +
+ "\n\tASYNC_WAL\020\002\022\014\n\010SYNC_WAL\020\003\022\r\n\tFSYNC_WAL" +
+ "\020\004\">\n\014MutationType\022\n\n\006APPEND\020\000\022\r\n\tINCREM" +
+ "ENT\020\001\022\007\n\003PUT\020\002\022\n\n\006DELETE\020\003\"p\n\nDeleteType" +
+ "\022\026\n\022DELETE_ONE_VERSION\020\000\022\034\n\030DELETE_MULTI" +
+ "PLE_VERSIONS\020\001\022\021\n\rDELETE_FAMILY\020\002\022\031\n\025DEL" +
+ "ETE_FAMILY_VERSION\020\003\"\242\001\n\rMutateRequest\022)",
+ "\n\006region\030\001 \002(\0132\031.hbase.pb.RegionSpecifie" +
+ "r\022)\n\010mutation\030\002 \002(\0132\027.hbase.pb.MutationP" +
+ "roto\022&\n\tcondition\030\003 \001(\0132\023.hbase.pb.Condi" +
+ "tion\022\023\n\013nonce_group\030\004 \001(\004\"E\n\016MutateRespo" +
+ "nse\022 \n\006result\030\001 \001(\0132\020.hbase.pb.Result\022\021\n" +
+ "\tprocessed\030\002 \001(\010\"\246\006\n\004Scan\022 \n\006column\030\001 \003(" +
+ "\0132\020.hbase.pb.Column\022*\n\tattribute\030\002 \003(\0132\027" +
+ ".hbase.pb.NameBytesPair\022\021\n\tstart_row\030\003 \001" +
+ "(\014\022\020\n\010stop_row\030\004 \001(\014\022 \n\006filter\030\005 \001(\0132\020.h" +
+ "base.pb.Filter\022\'\n\ntime_range\030\006 \001(\0132\023.hba",
+ "se.pb.TimeRange\022\027\n\014max_versions\030\007 \001(\r:\0011" +
+ "\022\032\n\014cache_blocks\030\010 \001(\010:\004true\022\022\n\nbatch_si" +
+ "ze\030\t \001(\r\022\027\n\017max_result_size\030\n \001(\004\022\023\n\013sto" +
+ "re_limit\030\013 \001(\r\022\024\n\014store_offset\030\014 \001(\r\022&\n\036" +
+ "load_column_families_on_demand\030\r \001(\010\022\021\n\005" +
+ "small\030\016 \001(\010B\002\030\001\022\027\n\010reversed\030\017 \001(\010:\005false" +
+ "\0222\n\013consistency\030\020 \001(\0162\025.hbase.pb.Consist" +
+ "ency:\006STRONG\022\017\n\007caching\030\021 \001(\r\022\035\n\025allow_p" +
+ "artial_results\030\022 \001(\010\0226\n\rcf_time_range\030\023 " +
+ "\003(\0132\037.hbase.pb.ColumnFamilyTimeRange\022\032\n\017",
+ "mvcc_read_point\030\024 \001(\004:\0010\022\037\n\021include_star" +
+ "t_row\030\025 \001(\010:\004true\022\037\n\020include_stop_row\030\026 " +
+ "\001(\010:\005false\0222\n\010readType\030\027 \001(\0162\027.hbase.pb." +
+ "Scan.ReadType:\007DEFAULT\022!\n\022need_cursor_re" +
+ "sult\030\030 \001(\010:\005false\".\n\010ReadType\022\013\n\007DEFAULT" +
+ "\020\000\022\n\n\006STREAM\020\001\022\t\n\005PREAD\020\002\"\300\002\n\013ScanReques" +
+ "t\022)\n\006region\030\001 \001(\0132\031.hbase.pb.RegionSpeci" +
+ "fier\022\034\n\004scan\030\002 \001(\0132\016.hbase.pb.Scan\022\022\n\nsc" +
+ "anner_id\030\003 \001(\004\022\026\n\016number_of_rows\030\004 \001(\r\022\025" +
+ "\n\rclose_scanner\030\005 \001(\010\022\025\n\rnext_call_seq\030\006",
+ " \001(\004\022\037\n\027client_handles_partials\030\007 \001(\010\022!\n" +
+ "\031client_handles_heartbeats\030\010 \001(\010\022\032\n\022trac" +
+ "k_scan_metrics\030\t \001(\010\022\024\n\005renew\030\n \001(\010:\005fal" +
+ "se\022\030\n\rlimit_of_rows\030\013 \001(\r:\0010\"\025\n\006Cursor\022\013" +
+ "\n\003row\030\001 \001(\014\"\330\002\n\014ScanResponse\022\030\n\020cells_pe" +
+ "r_result\030\001 \003(\r\022\022\n\nscanner_id\030\002 \001(\004\022\024\n\014mo" +
+ "re_results\030\003 \001(\010\022\013\n\003ttl\030\004 \001(\r\022!\n\007results" +
+ "\030\005 \003(\0132\020.hbase.pb.Result\022\r\n\005stale\030\006 \001(\010\022" +
+ "\037\n\027partial_flag_per_result\030\007 \003(\010\022\036\n\026more" +
+ "_results_in_region\030\010 \001(\010\022\031\n\021heartbeat_me",
+ "ssage\030\t \001(\010\022+\n\014scan_metrics\030\n \001(\0132\025.hbas" +
+ "e.pb.ScanMetrics\022\032\n\017mvcc_read_point\030\013 \001(" +
+ "\004:\0010\022 \n\006cursor\030\014 \001(\0132\020.hbase.pb.Cursor\"\240" +
+ "\002\n\024BulkLoadHFileRequest\022)\n\006region\030\001 \002(\0132" +
+ "\031.hbase.pb.RegionSpecifier\022>\n\013family_pat" +
+ "h\030\002 \003(\0132).hbase.pb.BulkLoadHFileRequest." +
+ "FamilyPath\022\026\n\016assign_seq_num\030\003 \001(\010\022+\n\010fs" +
+ "_token\030\004 \001(\0132\031.hbase.pb.DelegationToken\022" +
+ "\022\n\nbulk_token\030\005 \001(\t\022\030\n\tcopy_file\030\006 \001(\010:\005" +
+ "false\032*\n\nFamilyPath\022\016\n\006family\030\001 \002(\014\022\014\n\004p",
+ "ath\030\002 \002(\t\"\'\n\025BulkLoadHFileResponse\022\016\n\006lo" +
+ "aded\030\001 \002(\010\"V\n\017DelegationToken\022\022\n\nidentif" +
+ "ier\030\001 \001(\014\022\020\n\010password\030\002 \001(\014\022\014\n\004kind\030\003 \001(" +
+ "\t\022\017\n\007service\030\004 \001(\t\"l\n\026PrepareBulkLoadReq" +
+ "uest\022\'\n\ntable_name\030\001 \002(\0132\023.hbase.pb.Tabl" +
+ "eName\022)\n\006region\030\002 \001(\0132\031.hbase.pb.RegionS" +
+ "pecifier\"-\n\027PrepareBulkLoadResponse\022\022\n\nb" +
+ "ulk_token\030\001 \002(\t\"W\n\026CleanupBulkLoadReques" +
+ "t\022\022\n\nbulk_token\030\001 \002(\t\022)\n\006region\030\002 \001(\0132\031." +
+ "hbase.pb.RegionSpecifier\"\031\n\027CleanupBulkL",
+ "oadResponse\"a\n\026CoprocessorServiceCall\022\013\n" +
+ "\003row\030\001 \002(\014\022\024\n\014service_name\030\002 \002(\t\022\023\n\013meth" +
+ "od_name\030\003 \002(\t\022\017\n\007request\030\004 \002(\014\"B\n\030Coproc" +
+ "essorServiceResult\022&\n\005value\030\001 \001(\0132\027.hbas" +
+ "e.pb.NameBytesPair\"v\n\031CoprocessorService" +
+ "Request\022)\n\006region\030\001 \002(\0132\031.hbase.pb.Regio" +
+ "nSpecifier\022.\n\004call\030\002 \002(\0132 .hbase.pb.Copr" +
+ "ocessorServiceCall\"o\n\032CoprocessorService" +
+ "Response\022)\n\006region\030\001 \002(\0132\031.hbase.pb.Regi" +
+ "onSpecifier\022&\n\005value\030\002 \002(\0132\027.hbase.pb.Na",
+ "meBytesPair\"\226\001\n\006Action\022\r\n\005index\030\001 \001(\r\022)\n" +
+ "\010mutation\030\002 \001(\0132\027.hbase.pb.MutationProto" +
+ "\022\032\n\003get\030\003 \001(\0132\r.hbase.pb.Get\0226\n\014service_" +
+ "call\030\004 \001(\0132 .hbase.pb.CoprocessorService" +
+ "Call\"k\n\014RegionAction\022)\n\006region\030\001 \002(\0132\031.h" +
+ "base.pb.RegionSpecifier\022\016\n\006atomic\030\002 \001(\010\022" +
+ " \n\006action\030\003 \003(\0132\020.hbase.pb.Action\"c\n\017Reg" +
+ "ionLoadStats\022\027\n\014memstoreLoad\030\001 \001(\005:\0010\022\030\n" +
+ "\rheapOccupancy\030\002 \001(\005:\0010\022\035\n\022compactionPre" +
+ "ssure\030\003 \001(\005:\0010\"j\n\024MultiRegionLoadStats\022)",
+ "\n\006region\030\001 \003(\0132\031.hbase.pb.RegionSpecifie" +
+ "r\022\'\n\004stat\030\002 \003(\0132\031.hbase.pb.RegionLoadSta" +
+ "ts\"\336\001\n\021ResultOrException\022\r\n\005index\030\001 \001(\r\022" +
+ " \n\006result\030\002 \001(\0132\020.hbase.pb.Result\022*\n\texc" +
+ "eption\030\003 \001(\0132\027.hbase.pb.NameBytesPair\022:\n" +
+ "\016service_result\030\004 \001(\0132\".hbase.pb.Coproce" +
+ "ssorServiceResult\0220\n\tloadStats\030\005 \001(\0132\031.h" +
+ "base.pb.RegionLoadStatsB\002\030\001\"x\n\022RegionAct" +
+ "ionResult\0226\n\021resultOrException\030\001 \003(\0132\033.h" +
+ "base.pb.ResultOrException\022*\n\texception\030\002",
+ " \001(\0132\027.hbase.pb.NameBytesPair\"x\n\014MultiRe" +
+ "quest\022,\n\014regionAction\030\001 \003(\0132\026.hbase.pb.R" +
+ "egionAction\022\022\n\nnonceGroup\030\002 \001(\004\022&\n\tcondi" +
+ "tion\030\003 \001(\0132\023.hbase.pb.Condition\"\226\001\n\rMult" +
+ "iResponse\0228\n\022regionActionResult\030\001 \003(\0132\034." +
+ "hbase.pb.RegionActionResult\022\021\n\tprocessed" +
+ "\030\002 \001(\010\0228\n\020regionStatistics\030\003 \001(\0132\036.hbase" +
+ ".pb.MultiRegionLoadStats*\'\n\013Consistency\022" +
+ "\n\n\006STRONG\020\000\022\014\n\010TIMELINE\020\0012\263\005\n\rClientServ" +
+ "ice\0222\n\003Get\022\024.hbase.pb.GetRequest\032\025.hbase",
+ ".pb.GetResponse\022;\n\006Mutate\022\027.hbase.pb.Mut" +
+ "ateRequest\032\030.hbase.pb.MutateResponse\0225\n\004" +
+ "Scan\022\025.hbase.pb.ScanRequest\032\026.hbase.pb.S" +
+ "canResponse\022P\n\rBulkLoadHFile\022\036.hbase.pb." +
+ "BulkLoadHFileRequest\032\037.hbase.pb.BulkLoad" +
+ "HFileResponse\022V\n\017PrepareBulkLoad\022 .hbase" +
+ ".pb.PrepareBulkLoadRequest\032!.hbase.pb.Pr" +
+ "epareBulkLoadResponse\022V\n\017CleanupBulkLoad" +
+ "\022 .hbase.pb.CleanupBulkLoadRequest\032!.hba" +
+ "se.pb.CleanupBulkLoadResponse\022X\n\013ExecSer",
+ "vice\022#.hbase.pb.CoprocessorServiceReques" +
+ "t\032$.hbase.pb.CoprocessorServiceResponse\022" +
+ "d\n\027ExecRegionServerService\022#.hbase.pb.Co" +
+ "processorServiceRequest\032$.hbase.pb.Copro" +
+ "cessorServiceResponse\0228\n\005Multi\022\026.hbase.p" +
+ "b.MultiRequest\032\027.hbase.pb.MultiResponseB" +
+ "B\n*org.apache.hadoop.hbase.protobuf.gene" +
+ "ratedB\014ClientProtosH\001\210\001\001\240\001\001"
};
com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() {
@@ -41397,6 +41397,7 @@ public final class ClientProtos {
com.google.protobuf.Descriptors.FileDescriptor
.internalBuildGeneratedFileFrom(descriptorData,
new com.google.protobuf.Descriptors.FileDescriptor[] {
+ org.apache.hadoop.hbase.protobuf.generated.TableProtos.getDescriptor(),
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.getDescriptor(),
org.apache.hadoop.hbase.protobuf.generated.FilterProtos.getDescriptor(),
org.apache.hadoop.hbase.protobuf.generated.CellProtos.getDescriptor(),