HBASE-8915 protobuf message style

git-svn-id: https://svn.apache.org/repos/asf/hbase/trunk@1501834 13f79535-47bb-0310-9956-ffa450edef68
This commit is contained in:
Michael Stack 2013-07-10 16:30:32 +00:00
parent 8bf99de619
commit 457c928d35
52 changed files with 2371 additions and 2360 deletions

View File

@ -6422,11 +6422,11 @@ public final class AccessControlProtos {
"Permission\":\n\027CheckPermissionsRequest\022\037\n" + "Permission\":\n\027CheckPermissionsRequest\022\037\n" +
"\npermission\030\001 \003(\0132\013.Permission\"\032\n\030CheckP" + "\npermission\030\001 \003(\0132\013.Permission\"\032\n\030CheckP" +
"ermissionsResponse2\373\001\n\024AccessControlServ" + "ermissionsResponse2\373\001\n\024AccessControlServ" +
"ice\022&\n\005grant\022\r.GrantRequest\032\016.GrantRespo", "ice\022&\n\005Grant\022\r.GrantRequest\032\016.GrantRespo",
"nse\022)\n\006revoke\022\016.RevokeRequest\032\017.RevokeRe" + "nse\022)\n\006Revoke\022\016.RevokeRequest\032\017.RevokeRe" +
"sponse\022G\n\022getUserPermissions\022\027.UserPermi" + "sponse\022G\n\022GetUserPermissions\022\027.UserPermi" +
"ssionsRequest\032\030.UserPermissionsResponse\022" + "ssionsRequest\032\030.UserPermissionsResponse\022" +
"G\n\020checkPermissions\022\030.CheckPermissionsRe" + "G\n\020CheckPermissions\022\030.CheckPermissionsRe" +
"quest\032\031.CheckPermissionsResponseBI\n*org." + "quest\032\031.CheckPermissionsResponseBI\n*org." +
"apache.hadoop.hbase.protobuf.generatedB\023" + "apache.hadoop.hbase.protobuf.generatedB\023" +
"AccessControlProtosH\001\210\001\001\240\001\001" "AccessControlProtosH\001\210\001\001\240\001\001"

View File

@ -16,7 +16,7 @@ public final class AdminProtos {
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier getRegion(); org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier getRegion();
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder getRegionOrBuilder(); org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder getRegionOrBuilder();
// optional bool compactionState = 2; // optional bool compaction_state = 2;
boolean hasCompactionState(); boolean hasCompactionState();
boolean getCompactionState(); boolean getCompactionState();
} }
@ -62,8 +62,8 @@ public final class AdminProtos {
return region_; return region_;
} }
// optional bool compactionState = 2; // optional bool compaction_state = 2;
public static final int COMPACTIONSTATE_FIELD_NUMBER = 2; public static final int COMPACTION_STATE_FIELD_NUMBER = 2;
private boolean compactionState_; private boolean compactionState_;
public boolean hasCompactionState() { public boolean hasCompactionState() {
return ((bitField0_ & 0x00000002) == 0x00000002); return ((bitField0_ & 0x00000002) == 0x00000002);
@ -166,7 +166,7 @@ public final class AdminProtos {
hash = (53 * hash) + getRegion().hashCode(); hash = (53 * hash) + getRegion().hashCode();
} }
if (hasCompactionState()) { if (hasCompactionState()) {
hash = (37 * hash) + COMPACTIONSTATE_FIELD_NUMBER; hash = (37 * hash) + COMPACTION_STATE_FIELD_NUMBER;
hash = (53 * hash) + hashBoolean(getCompactionState()); hash = (53 * hash) + hashBoolean(getCompactionState());
} }
hash = (29 * hash) + getUnknownFields().hashCode(); hash = (29 * hash) + getUnknownFields().hashCode();
@ -515,7 +515,7 @@ public final class AdminProtos {
return regionBuilder_; return regionBuilder_;
} }
// optional bool compactionState = 2; // optional bool compaction_state = 2;
private boolean compactionState_ ; private boolean compactionState_ ;
public boolean hasCompactionState() { public boolean hasCompactionState() {
return ((bitField0_ & 0x00000002) == 0x00000002); return ((bitField0_ & 0x00000002) == 0x00000002);
@ -550,12 +550,12 @@ public final class AdminProtos {
public interface GetRegionInfoResponseOrBuilder public interface GetRegionInfoResponseOrBuilder
extends com.google.protobuf.MessageOrBuilder { extends com.google.protobuf.MessageOrBuilder {
// required .RegionInfo regionInfo = 1; // required .RegionInfo region_info = 1;
boolean hasRegionInfo(); boolean hasRegionInfo();
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfo getRegionInfo(); org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfo getRegionInfo();
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfoOrBuilder getRegionInfoOrBuilder(); org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfoOrBuilder getRegionInfoOrBuilder();
// optional .GetRegionInfoResponse.CompactionState compactionState = 2; // optional .GetRegionInfoResponse.CompactionState compaction_state = 2;
boolean hasCompactionState(); boolean hasCompactionState();
org.apache.hadoop.hbase.protobuf.generated.AdminProtos.GetRegionInfoResponse.CompactionState getCompactionState(); org.apache.hadoop.hbase.protobuf.generated.AdminProtos.GetRegionInfoResponse.CompactionState getCompactionState();
} }
@ -663,8 +663,8 @@ public final class AdminProtos {
} }
private int bitField0_; private int bitField0_;
// required .RegionInfo regionInfo = 1; // required .RegionInfo region_info = 1;
public static final int REGIONINFO_FIELD_NUMBER = 1; public static final int REGION_INFO_FIELD_NUMBER = 1;
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfo regionInfo_; private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfo regionInfo_;
public boolean hasRegionInfo() { public boolean hasRegionInfo() {
return ((bitField0_ & 0x00000001) == 0x00000001); return ((bitField0_ & 0x00000001) == 0x00000001);
@ -676,8 +676,8 @@ public final class AdminProtos {
return regionInfo_; return regionInfo_;
} }
// optional .GetRegionInfoResponse.CompactionState compactionState = 2; // optional .GetRegionInfoResponse.CompactionState compaction_state = 2;
public static final int COMPACTIONSTATE_FIELD_NUMBER = 2; public static final int COMPACTION_STATE_FIELD_NUMBER = 2;
private org.apache.hadoop.hbase.protobuf.generated.AdminProtos.GetRegionInfoResponse.CompactionState compactionState_; private org.apache.hadoop.hbase.protobuf.generated.AdminProtos.GetRegionInfoResponse.CompactionState compactionState_;
public boolean hasCompactionState() { public boolean hasCompactionState() {
return ((bitField0_ & 0x00000002) == 0x00000002); return ((bitField0_ & 0x00000002) == 0x00000002);
@ -776,11 +776,11 @@ public final class AdminProtos {
int hash = 41; int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode(); hash = (19 * hash) + getDescriptorForType().hashCode();
if (hasRegionInfo()) { if (hasRegionInfo()) {
hash = (37 * hash) + REGIONINFO_FIELD_NUMBER; hash = (37 * hash) + REGION_INFO_FIELD_NUMBER;
hash = (53 * hash) + getRegionInfo().hashCode(); hash = (53 * hash) + getRegionInfo().hashCode();
} }
if (hasCompactionState()) { if (hasCompactionState()) {
hash = (37 * hash) + COMPACTIONSTATE_FIELD_NUMBER; hash = (37 * hash) + COMPACTION_STATE_FIELD_NUMBER;
hash = (53 * hash) + hashEnum(getCompactionState()); hash = (53 * hash) + hashEnum(getCompactionState());
} }
hash = (29 * hash) + getUnknownFields().hashCode(); hash = (29 * hash) + getUnknownFields().hashCode();
@ -1045,7 +1045,7 @@ public final class AdminProtos {
private int bitField0_; private int bitField0_;
// required .RegionInfo regionInfo = 1; // required .RegionInfo region_info = 1;
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfo regionInfo_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfo.getDefaultInstance(); private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfo regionInfo_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfo.getDefaultInstance();
private com.google.protobuf.SingleFieldBuilder< private com.google.protobuf.SingleFieldBuilder<
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfo, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfo.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfoOrBuilder> regionInfoBuilder_; org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfo, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfo.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfoOrBuilder> regionInfoBuilder_;
@ -1135,7 +1135,7 @@ public final class AdminProtos {
return regionInfoBuilder_; return regionInfoBuilder_;
} }
// optional .GetRegionInfoResponse.CompactionState compactionState = 2; // optional .GetRegionInfoResponse.CompactionState compaction_state = 2;
private org.apache.hadoop.hbase.protobuf.generated.AdminProtos.GetRegionInfoResponse.CompactionState compactionState_ = org.apache.hadoop.hbase.protobuf.generated.AdminProtos.GetRegionInfoResponse.CompactionState.NONE; private org.apache.hadoop.hbase.protobuf.generated.AdminProtos.GetRegionInfoResponse.CompactionState compactionState_ = org.apache.hadoop.hbase.protobuf.generated.AdminProtos.GetRegionInfoResponse.CompactionState.NONE;
public boolean hasCompactionState() { public boolean hasCompactionState() {
return ((bitField0_ & 0x00000002) == 0x00000002); return ((bitField0_ & 0x00000002) == 0x00000002);
@ -1757,7 +1757,7 @@ public final class AdminProtos {
public interface GetStoreFileResponseOrBuilder public interface GetStoreFileResponseOrBuilder
extends com.google.protobuf.MessageOrBuilder { extends com.google.protobuf.MessageOrBuilder {
// repeated string storeFile = 1; // repeated string store_file = 1;
java.util.List<String> getStoreFileList(); java.util.List<String> getStoreFileList();
int getStoreFileCount(); int getStoreFileCount();
String getStoreFile(int index); String getStoreFile(int index);
@ -1790,8 +1790,8 @@ public final class AdminProtos {
return org.apache.hadoop.hbase.protobuf.generated.AdminProtos.internal_static_GetStoreFileResponse_fieldAccessorTable; return org.apache.hadoop.hbase.protobuf.generated.AdminProtos.internal_static_GetStoreFileResponse_fieldAccessorTable;
} }
// repeated string storeFile = 1; // repeated string store_file = 1;
public static final int STOREFILE_FIELD_NUMBER = 1; public static final int STORE_FILE_FIELD_NUMBER = 1;
private com.google.protobuf.LazyStringList storeFile_; private com.google.protobuf.LazyStringList storeFile_;
public java.util.List<String> public java.util.List<String>
getStoreFileList() { getStoreFileList() {
@ -1875,7 +1875,7 @@ public final class AdminProtos {
int hash = 41; int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode(); hash = (19 * hash) + getDescriptorForType().hashCode();
if (getStoreFileCount() > 0) { if (getStoreFileCount() > 0) {
hash = (37 * hash) + STOREFILE_FIELD_NUMBER; hash = (37 * hash) + STORE_FILE_FIELD_NUMBER;
hash = (53 * hash) + getStoreFileList().hashCode(); hash = (53 * hash) + getStoreFileList().hashCode();
} }
hash = (29 * hash) + getUnknownFields().hashCode(); hash = (29 * hash) + getUnknownFields().hashCode();
@ -2106,7 +2106,7 @@ public final class AdminProtos {
private int bitField0_; private int bitField0_;
// repeated string storeFile = 1; // repeated string store_file = 1;
private com.google.protobuf.LazyStringList storeFile_ = com.google.protobuf.LazyStringArrayList.EMPTY; private com.google.protobuf.LazyStringList storeFile_ = com.google.protobuf.LazyStringArrayList.EMPTY;
private void ensureStoreFileIsMutable() { private void ensureStoreFileIsMutable() {
if (!((bitField0_ & 0x00000001) == 0x00000001)) { if (!((bitField0_ & 0x00000001) == 0x00000001)) {
@ -2476,7 +2476,7 @@ public final class AdminProtos {
public interface GetOnlineRegionResponseOrBuilder public interface GetOnlineRegionResponseOrBuilder
extends com.google.protobuf.MessageOrBuilder { extends com.google.protobuf.MessageOrBuilder {
// repeated .RegionInfo regionInfo = 1; // repeated .RegionInfo region_info = 1;
java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfo> java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfo>
getRegionInfoList(); getRegionInfoList();
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfo getRegionInfo(int index); org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfo getRegionInfo(int index);
@ -2514,8 +2514,8 @@ public final class AdminProtos {
return org.apache.hadoop.hbase.protobuf.generated.AdminProtos.internal_static_GetOnlineRegionResponse_fieldAccessorTable; return org.apache.hadoop.hbase.protobuf.generated.AdminProtos.internal_static_GetOnlineRegionResponse_fieldAccessorTable;
} }
// repeated .RegionInfo regionInfo = 1; // repeated .RegionInfo region_info = 1;
public static final int REGIONINFO_FIELD_NUMBER = 1; public static final int REGION_INFO_FIELD_NUMBER = 1;
private java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfo> regionInfo_; private java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfo> regionInfo_;
public java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfo> getRegionInfoList() { public java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfo> getRegionInfoList() {
return regionInfo_; return regionInfo_;
@ -2607,7 +2607,7 @@ public final class AdminProtos {
int hash = 41; int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode(); hash = (19 * hash) + getDescriptorForType().hashCode();
if (getRegionInfoCount() > 0) { if (getRegionInfoCount() > 0) {
hash = (37 * hash) + REGIONINFO_FIELD_NUMBER; hash = (37 * hash) + REGION_INFO_FIELD_NUMBER;
hash = (53 * hash) + getRegionInfoList().hashCode(); hash = (53 * hash) + getRegionInfoList().hashCode();
} }
hash = (29 * hash) + getUnknownFields().hashCode(); hash = (29 * hash) + getUnknownFields().hashCode();
@ -2869,7 +2869,7 @@ public final class AdminProtos {
private int bitField0_; private int bitField0_;
// repeated .RegionInfo regionInfo = 1; // repeated .RegionInfo region_info = 1;
private java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfo> regionInfo_ = private java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfo> regionInfo_ =
java.util.Collections.emptyList(); java.util.Collections.emptyList();
private void ensureRegionInfoIsMutable() { private void ensureRegionInfoIsMutable() {
@ -3069,7 +3069,7 @@ public final class AdminProtos {
public interface OpenRegionRequestOrBuilder public interface OpenRegionRequestOrBuilder
extends com.google.protobuf.MessageOrBuilder { extends com.google.protobuf.MessageOrBuilder {
// repeated .OpenRegionRequest.RegionOpenInfo openInfo = 1; // repeated .OpenRegionRequest.RegionOpenInfo open_info = 1;
java.util.List<org.apache.hadoop.hbase.protobuf.generated.AdminProtos.OpenRegionRequest.RegionOpenInfo> java.util.List<org.apache.hadoop.hbase.protobuf.generated.AdminProtos.OpenRegionRequest.RegionOpenInfo>
getOpenInfoList(); getOpenInfoList();
org.apache.hadoop.hbase.protobuf.generated.AdminProtos.OpenRegionRequest.RegionOpenInfo getOpenInfo(int index); org.apache.hadoop.hbase.protobuf.generated.AdminProtos.OpenRegionRequest.RegionOpenInfo getOpenInfo(int index);
@ -3115,11 +3115,11 @@ public final class AdminProtos {
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfo getRegion(); org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfo getRegion();
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfoOrBuilder getRegionOrBuilder(); org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfoOrBuilder getRegionOrBuilder();
// optional uint32 versionOfOfflineNode = 2; // optional uint32 version_of_offline_node = 2;
boolean hasVersionOfOfflineNode(); boolean hasVersionOfOfflineNode();
int getVersionOfOfflineNode(); int getVersionOfOfflineNode();
// repeated .ServerName favoredNodes = 3; // repeated .ServerName favored_nodes = 3;
java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName> java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName>
getFavoredNodesList(); getFavoredNodesList();
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName getFavoredNodes(int index); org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName getFavoredNodes(int index);
@ -3171,8 +3171,8 @@ public final class AdminProtos {
return region_; return region_;
} }
// optional uint32 versionOfOfflineNode = 2; // optional uint32 version_of_offline_node = 2;
public static final int VERSIONOFOFFLINENODE_FIELD_NUMBER = 2; public static final int VERSION_OF_OFFLINE_NODE_FIELD_NUMBER = 2;
private int versionOfOfflineNode_; private int versionOfOfflineNode_;
public boolean hasVersionOfOfflineNode() { public boolean hasVersionOfOfflineNode() {
return ((bitField0_ & 0x00000002) == 0x00000002); return ((bitField0_ & 0x00000002) == 0x00000002);
@ -3181,8 +3181,8 @@ public final class AdminProtos {
return versionOfOfflineNode_; return versionOfOfflineNode_;
} }
// repeated .ServerName favoredNodes = 3; // repeated .ServerName favored_nodes = 3;
public static final int FAVOREDNODES_FIELD_NUMBER = 3; public static final int FAVORED_NODES_FIELD_NUMBER = 3;
private java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName> favoredNodes_; private java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName> favoredNodes_;
public java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName> getFavoredNodesList() { public java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName> getFavoredNodesList() {
return favoredNodes_; return favoredNodes_;
@ -3312,11 +3312,11 @@ public final class AdminProtos {
hash = (53 * hash) + getRegion().hashCode(); hash = (53 * hash) + getRegion().hashCode();
} }
if (hasVersionOfOfflineNode()) { if (hasVersionOfOfflineNode()) {
hash = (37 * hash) + VERSIONOFOFFLINENODE_FIELD_NUMBER; hash = (37 * hash) + VERSION_OF_OFFLINE_NODE_FIELD_NUMBER;
hash = (53 * hash) + getVersionOfOfflineNode(); hash = (53 * hash) + getVersionOfOfflineNode();
} }
if (getFavoredNodesCount() > 0) { if (getFavoredNodesCount() > 0) {
hash = (37 * hash) + FAVOREDNODES_FIELD_NUMBER; hash = (37 * hash) + FAVORED_NODES_FIELD_NUMBER;
hash = (53 * hash) + getFavoredNodesList().hashCode(); hash = (53 * hash) + getFavoredNodesList().hashCode();
} }
hash = (29 * hash) + getUnknownFields().hashCode(); hash = (29 * hash) + getUnknownFields().hashCode();
@ -3719,7 +3719,7 @@ public final class AdminProtos {
return regionBuilder_; return regionBuilder_;
} }
// optional uint32 versionOfOfflineNode = 2; // optional uint32 version_of_offline_node = 2;
private int versionOfOfflineNode_ ; private int versionOfOfflineNode_ ;
public boolean hasVersionOfOfflineNode() { public boolean hasVersionOfOfflineNode() {
return ((bitField0_ & 0x00000002) == 0x00000002); return ((bitField0_ & 0x00000002) == 0x00000002);
@ -3740,7 +3740,7 @@ public final class AdminProtos {
return this; return this;
} }
// repeated .ServerName favoredNodes = 3; // repeated .ServerName favored_nodes = 3;
private java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName> favoredNodes_ = private java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName> favoredNodes_ =
java.util.Collections.emptyList(); java.util.Collections.emptyList();
private void ensureFavoredNodesIsMutable() { private void ensureFavoredNodesIsMutable() {
@ -3937,8 +3937,8 @@ public final class AdminProtos {
// @@protoc_insertion_point(class_scope:OpenRegionRequest.RegionOpenInfo) // @@protoc_insertion_point(class_scope:OpenRegionRequest.RegionOpenInfo)
} }
// repeated .OpenRegionRequest.RegionOpenInfo openInfo = 1; // repeated .OpenRegionRequest.RegionOpenInfo open_info = 1;
public static final int OPENINFO_FIELD_NUMBER = 1; public static final int OPEN_INFO_FIELD_NUMBER = 1;
private java.util.List<org.apache.hadoop.hbase.protobuf.generated.AdminProtos.OpenRegionRequest.RegionOpenInfo> openInfo_; private java.util.List<org.apache.hadoop.hbase.protobuf.generated.AdminProtos.OpenRegionRequest.RegionOpenInfo> openInfo_;
public java.util.List<org.apache.hadoop.hbase.protobuf.generated.AdminProtos.OpenRegionRequest.RegionOpenInfo> getOpenInfoList() { public java.util.List<org.apache.hadoop.hbase.protobuf.generated.AdminProtos.OpenRegionRequest.RegionOpenInfo> getOpenInfoList() {
return openInfo_; return openInfo_;
@ -4030,7 +4030,7 @@ public final class AdminProtos {
int hash = 41; int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode(); hash = (19 * hash) + getDescriptorForType().hashCode();
if (getOpenInfoCount() > 0) { if (getOpenInfoCount() > 0) {
hash = (37 * hash) + OPENINFO_FIELD_NUMBER; hash = (37 * hash) + OPEN_INFO_FIELD_NUMBER;
hash = (53 * hash) + getOpenInfoList().hashCode(); hash = (53 * hash) + getOpenInfoList().hashCode();
} }
hash = (29 * hash) + getUnknownFields().hashCode(); hash = (29 * hash) + getUnknownFields().hashCode();
@ -4292,7 +4292,7 @@ public final class AdminProtos {
private int bitField0_; private int bitField0_;
// repeated .OpenRegionRequest.RegionOpenInfo openInfo = 1; // repeated .OpenRegionRequest.RegionOpenInfo open_info = 1;
private java.util.List<org.apache.hadoop.hbase.protobuf.generated.AdminProtos.OpenRegionRequest.RegionOpenInfo> openInfo_ = private java.util.List<org.apache.hadoop.hbase.protobuf.generated.AdminProtos.OpenRegionRequest.RegionOpenInfo> openInfo_ =
java.util.Collections.emptyList(); java.util.Collections.emptyList();
private void ensureOpenInfoIsMutable() { private void ensureOpenInfoIsMutable() {
@ -4492,7 +4492,7 @@ public final class AdminProtos {
public interface OpenRegionResponseOrBuilder public interface OpenRegionResponseOrBuilder
extends com.google.protobuf.MessageOrBuilder { extends com.google.protobuf.MessageOrBuilder {
// repeated .OpenRegionResponse.RegionOpeningState openingState = 1; // repeated .OpenRegionResponse.RegionOpeningState opening_state = 1;
java.util.List<org.apache.hadoop.hbase.protobuf.generated.AdminProtos.OpenRegionResponse.RegionOpeningState> getOpeningStateList(); java.util.List<org.apache.hadoop.hbase.protobuf.generated.AdminProtos.OpenRegionResponse.RegionOpeningState> getOpeningStateList();
int getOpeningStateCount(); int getOpeningStateCount();
org.apache.hadoop.hbase.protobuf.generated.AdminProtos.OpenRegionResponse.RegionOpeningState getOpeningState(int index); org.apache.hadoop.hbase.protobuf.generated.AdminProtos.OpenRegionResponse.RegionOpeningState getOpeningState(int index);
@ -4597,8 +4597,8 @@ public final class AdminProtos {
// @@protoc_insertion_point(enum_scope:OpenRegionResponse.RegionOpeningState) // @@protoc_insertion_point(enum_scope:OpenRegionResponse.RegionOpeningState)
} }
// repeated .OpenRegionResponse.RegionOpeningState openingState = 1; // repeated .OpenRegionResponse.RegionOpeningState opening_state = 1;
public static final int OPENINGSTATE_FIELD_NUMBER = 1; public static final int OPENING_STATE_FIELD_NUMBER = 1;
private java.util.List<org.apache.hadoop.hbase.protobuf.generated.AdminProtos.OpenRegionResponse.RegionOpeningState> openingState_; private java.util.List<org.apache.hadoop.hbase.protobuf.generated.AdminProtos.OpenRegionResponse.RegionOpeningState> openingState_;
public java.util.List<org.apache.hadoop.hbase.protobuf.generated.AdminProtos.OpenRegionResponse.RegionOpeningState> getOpeningStateList() { public java.util.List<org.apache.hadoop.hbase.protobuf.generated.AdminProtos.OpenRegionResponse.RegionOpeningState> getOpeningStateList() {
return openingState_; return openingState_;
@ -4681,7 +4681,7 @@ public final class AdminProtos {
int hash = 41; int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode(); hash = (19 * hash) + getDescriptorForType().hashCode();
if (getOpeningStateCount() > 0) { if (getOpeningStateCount() > 0) {
hash = (37 * hash) + OPENINGSTATE_FIELD_NUMBER; hash = (37 * hash) + OPENING_STATE_FIELD_NUMBER;
hash = (53 * hash) + hashEnumList(getOpeningStateList()); hash = (53 * hash) + hashEnumList(getOpeningStateList());
} }
hash = (29 * hash) + getUnknownFields().hashCode(); hash = (29 * hash) + getUnknownFields().hashCode();
@ -4931,7 +4931,7 @@ public final class AdminProtos {
private int bitField0_; private int bitField0_;
// repeated .OpenRegionResponse.RegionOpeningState openingState = 1; // repeated .OpenRegionResponse.RegionOpeningState opening_state = 1;
private java.util.List<org.apache.hadoop.hbase.protobuf.generated.AdminProtos.OpenRegionResponse.RegionOpeningState> openingState_ = private java.util.List<org.apache.hadoop.hbase.protobuf.generated.AdminProtos.OpenRegionResponse.RegionOpeningState> openingState_ =
java.util.Collections.emptyList(); java.util.Collections.emptyList();
private void ensureOpeningStateIsMutable() { private void ensureOpeningStateIsMutable() {
@ -5001,15 +5001,15 @@ public final class AdminProtos {
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier getRegion(); org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier getRegion();
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder getRegionOrBuilder(); org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder getRegionOrBuilder();
// optional uint32 versionOfClosingNode = 2; // optional uint32 version_of_closing_node = 2;
boolean hasVersionOfClosingNode(); boolean hasVersionOfClosingNode();
int getVersionOfClosingNode(); int getVersionOfClosingNode();
// optional bool transitionInZK = 3 [default = true]; // optional bool transition_in_ZK = 3 [default = true];
boolean hasTransitionInZK(); boolean hasTransitionInZK();
boolean getTransitionInZK(); boolean getTransitionInZK();
// optional .ServerName destinationServer = 4; // optional .ServerName destination_server = 4;
boolean hasDestinationServer(); boolean hasDestinationServer();
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName getDestinationServer(); org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName getDestinationServer();
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerNameOrBuilder getDestinationServerOrBuilder(); org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerNameOrBuilder getDestinationServerOrBuilder();
@ -5056,8 +5056,8 @@ public final class AdminProtos {
return region_; return region_;
} }
// optional uint32 versionOfClosingNode = 2; // optional uint32 version_of_closing_node = 2;
public static final int VERSIONOFCLOSINGNODE_FIELD_NUMBER = 2; public static final int VERSION_OF_CLOSING_NODE_FIELD_NUMBER = 2;
private int versionOfClosingNode_; private int versionOfClosingNode_;
public boolean hasVersionOfClosingNode() { public boolean hasVersionOfClosingNode() {
return ((bitField0_ & 0x00000002) == 0x00000002); return ((bitField0_ & 0x00000002) == 0x00000002);
@ -5066,8 +5066,8 @@ public final class AdminProtos {
return versionOfClosingNode_; return versionOfClosingNode_;
} }
// optional bool transitionInZK = 3 [default = true]; // optional bool transition_in_ZK = 3 [default = true];
public static final int TRANSITIONINZK_FIELD_NUMBER = 3; public static final int TRANSITION_IN_ZK_FIELD_NUMBER = 3;
private boolean transitionInZK_; private boolean transitionInZK_;
public boolean hasTransitionInZK() { public boolean hasTransitionInZK() {
return ((bitField0_ & 0x00000004) == 0x00000004); return ((bitField0_ & 0x00000004) == 0x00000004);
@ -5076,8 +5076,8 @@ public final class AdminProtos {
return transitionInZK_; return transitionInZK_;
} }
// optional .ServerName destinationServer = 4; // optional .ServerName destination_server = 4;
public static final int DESTINATIONSERVER_FIELD_NUMBER = 4; public static final int DESTINATION_SERVER_FIELD_NUMBER = 4;
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName destinationServer_; private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName destinationServer_;
public boolean hasDestinationServer() { public boolean hasDestinationServer() {
return ((bitField0_ & 0x00000008) == 0x00000008); return ((bitField0_ & 0x00000008) == 0x00000008);
@ -5215,15 +5215,15 @@ public final class AdminProtos {
hash = (53 * hash) + getRegion().hashCode(); hash = (53 * hash) + getRegion().hashCode();
} }
if (hasVersionOfClosingNode()) { if (hasVersionOfClosingNode()) {
hash = (37 * hash) + VERSIONOFCLOSINGNODE_FIELD_NUMBER; hash = (37 * hash) + VERSION_OF_CLOSING_NODE_FIELD_NUMBER;
hash = (53 * hash) + getVersionOfClosingNode(); hash = (53 * hash) + getVersionOfClosingNode();
} }
if (hasTransitionInZK()) { if (hasTransitionInZK()) {
hash = (37 * hash) + TRANSITIONINZK_FIELD_NUMBER; hash = (37 * hash) + TRANSITION_IN_ZK_FIELD_NUMBER;
hash = (53 * hash) + hashBoolean(getTransitionInZK()); hash = (53 * hash) + hashBoolean(getTransitionInZK());
} }
if (hasDestinationServer()) { if (hasDestinationServer()) {
hash = (37 * hash) + DESTINATIONSERVER_FIELD_NUMBER; hash = (37 * hash) + DESTINATION_SERVER_FIELD_NUMBER;
hash = (53 * hash) + getDestinationServer().hashCode(); hash = (53 * hash) + getDestinationServer().hashCode();
} }
hash = (29 * hash) + getUnknownFields().hashCode(); hash = (29 * hash) + getUnknownFields().hashCode();
@ -5619,7 +5619,7 @@ public final class AdminProtos {
return regionBuilder_; return regionBuilder_;
} }
// optional uint32 versionOfClosingNode = 2; // optional uint32 version_of_closing_node = 2;
private int versionOfClosingNode_ ; private int versionOfClosingNode_ ;
public boolean hasVersionOfClosingNode() { public boolean hasVersionOfClosingNode() {
return ((bitField0_ & 0x00000002) == 0x00000002); return ((bitField0_ & 0x00000002) == 0x00000002);
@ -5640,7 +5640,7 @@ public final class AdminProtos {
return this; return this;
} }
// optional bool transitionInZK = 3 [default = true]; // optional bool transition_in_ZK = 3 [default = true];
private boolean transitionInZK_ = true; private boolean transitionInZK_ = true;
public boolean hasTransitionInZK() { public boolean hasTransitionInZK() {
return ((bitField0_ & 0x00000004) == 0x00000004); return ((bitField0_ & 0x00000004) == 0x00000004);
@ -5661,7 +5661,7 @@ public final class AdminProtos {
return this; return this;
} }
// optional .ServerName destinationServer = 4; // optional .ServerName destination_server = 4;
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName destinationServer_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.getDefaultInstance(); private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName destinationServer_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.getDefaultInstance();
private com.google.protobuf.SingleFieldBuilder< private com.google.protobuf.SingleFieldBuilder<
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerNameOrBuilder> destinationServerBuilder_; org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerNameOrBuilder> destinationServerBuilder_;
@ -6149,7 +6149,7 @@ public final class AdminProtos {
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier getRegion(); org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier getRegion();
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder getRegionOrBuilder(); org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder getRegionOrBuilder();
// optional uint64 ifOlderThanTs = 2; // optional uint64 if_older_than_ts = 2;
boolean hasIfOlderThanTs(); boolean hasIfOlderThanTs();
long getIfOlderThanTs(); long getIfOlderThanTs();
} }
@ -6195,8 +6195,8 @@ public final class AdminProtos {
return region_; return region_;
} }
// optional uint64 ifOlderThanTs = 2; // optional uint64 if_older_than_ts = 2;
public static final int IFOLDERTHANTS_FIELD_NUMBER = 2; public static final int IF_OLDER_THAN_TS_FIELD_NUMBER = 2;
private long ifOlderThanTs_; private long ifOlderThanTs_;
public boolean hasIfOlderThanTs() { public boolean hasIfOlderThanTs() {
return ((bitField0_ & 0x00000002) == 0x00000002); return ((bitField0_ & 0x00000002) == 0x00000002);
@ -6299,7 +6299,7 @@ public final class AdminProtos {
hash = (53 * hash) + getRegion().hashCode(); hash = (53 * hash) + getRegion().hashCode();
} }
if (hasIfOlderThanTs()) { if (hasIfOlderThanTs()) {
hash = (37 * hash) + IFOLDERTHANTS_FIELD_NUMBER; hash = (37 * hash) + IF_OLDER_THAN_TS_FIELD_NUMBER;
hash = (53 * hash) + hashLong(getIfOlderThanTs()); hash = (53 * hash) + hashLong(getIfOlderThanTs());
} }
hash = (29 * hash) + getUnknownFields().hashCode(); hash = (29 * hash) + getUnknownFields().hashCode();
@ -6648,7 +6648,7 @@ public final class AdminProtos {
return regionBuilder_; return regionBuilder_;
} }
// optional uint64 ifOlderThanTs = 2; // optional uint64 if_older_than_ts = 2;
private long ifOlderThanTs_ ; private long ifOlderThanTs_ ;
public boolean hasIfOlderThanTs() { public boolean hasIfOlderThanTs() {
return ((bitField0_ & 0x00000002) == 0x00000002); return ((bitField0_ & 0x00000002) == 0x00000002);
@ -6683,7 +6683,7 @@ public final class AdminProtos {
public interface FlushRegionResponseOrBuilder public interface FlushRegionResponseOrBuilder
extends com.google.protobuf.MessageOrBuilder { extends com.google.protobuf.MessageOrBuilder {
// required uint64 lastFlushTime = 1; // required uint64 last_flush_time = 1;
boolean hasLastFlushTime(); boolean hasLastFlushTime();
long getLastFlushTime(); long getLastFlushTime();
@ -6720,8 +6720,8 @@ public final class AdminProtos {
} }
private int bitField0_; private int bitField0_;
// required uint64 lastFlushTime = 1; // required uint64 last_flush_time = 1;
public static final int LASTFLUSHTIME_FIELD_NUMBER = 1; public static final int LAST_FLUSH_TIME_FIELD_NUMBER = 1;
private long lastFlushTime_; private long lastFlushTime_;
public boolean hasLastFlushTime() { public boolean hasLastFlushTime() {
return ((bitField0_ & 0x00000001) == 0x00000001); return ((bitField0_ & 0x00000001) == 0x00000001);
@ -6826,7 +6826,7 @@ public final class AdminProtos {
int hash = 41; int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode(); hash = (19 * hash) + getDescriptorForType().hashCode();
if (hasLastFlushTime()) { if (hasLastFlushTime()) {
hash = (37 * hash) + LASTFLUSHTIME_FIELD_NUMBER; hash = (37 * hash) + LAST_FLUSH_TIME_FIELD_NUMBER;
hash = (53 * hash) + hashLong(getLastFlushTime()); hash = (53 * hash) + hashLong(getLastFlushTime());
} }
if (hasFlushed()) { if (hasFlushed()) {
@ -7072,7 +7072,7 @@ public final class AdminProtos {
private int bitField0_; private int bitField0_;
// required uint64 lastFlushTime = 1; // required uint64 last_flush_time = 1;
private long lastFlushTime_ ; private long lastFlushTime_ ;
public boolean hasLastFlushTime() { public boolean hasLastFlushTime() {
return ((bitField0_ & 0x00000001) == 0x00000001); return ((bitField0_ & 0x00000001) == 0x00000001);
@ -7133,7 +7133,7 @@ public final class AdminProtos {
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier getRegion(); org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier getRegion();
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder getRegionOrBuilder(); org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder getRegionOrBuilder();
// optional bytes splitPoint = 2; // optional bytes split_point = 2;
boolean hasSplitPoint(); boolean hasSplitPoint();
com.google.protobuf.ByteString getSplitPoint(); com.google.protobuf.ByteString getSplitPoint();
} }
@ -7179,8 +7179,8 @@ public final class AdminProtos {
return region_; return region_;
} }
// optional bytes splitPoint = 2; // optional bytes split_point = 2;
public static final int SPLITPOINT_FIELD_NUMBER = 2; public static final int SPLIT_POINT_FIELD_NUMBER = 2;
private com.google.protobuf.ByteString splitPoint_; private com.google.protobuf.ByteString splitPoint_;
public boolean hasSplitPoint() { public boolean hasSplitPoint() {
return ((bitField0_ & 0x00000002) == 0x00000002); return ((bitField0_ & 0x00000002) == 0x00000002);
@ -7283,7 +7283,7 @@ public final class AdminProtos {
hash = (53 * hash) + getRegion().hashCode(); hash = (53 * hash) + getRegion().hashCode();
} }
if (hasSplitPoint()) { if (hasSplitPoint()) {
hash = (37 * hash) + SPLITPOINT_FIELD_NUMBER; hash = (37 * hash) + SPLIT_POINT_FIELD_NUMBER;
hash = (53 * hash) + getSplitPoint().hashCode(); hash = (53 * hash) + getSplitPoint().hashCode();
} }
hash = (29 * hash) + getUnknownFields().hashCode(); hash = (29 * hash) + getUnknownFields().hashCode();
@ -7632,7 +7632,7 @@ public final class AdminProtos {
return regionBuilder_; return regionBuilder_;
} }
// optional bytes splitPoint = 2; // optional bytes split_point = 2;
private com.google.protobuf.ByteString splitPoint_ = com.google.protobuf.ByteString.EMPTY; private com.google.protobuf.ByteString splitPoint_ = com.google.protobuf.ByteString.EMPTY;
public boolean hasSplitPoint() { public boolean hasSplitPoint() {
return ((bitField0_ & 0x00000002) == 0x00000002); return ((bitField0_ & 0x00000002) == 0x00000002);
@ -8878,12 +8878,12 @@ public final class AdminProtos {
public interface MergeRegionsRequestOrBuilder public interface MergeRegionsRequestOrBuilder
extends com.google.protobuf.MessageOrBuilder { extends com.google.protobuf.MessageOrBuilder {
// required .RegionSpecifier regionA = 1; // required .RegionSpecifier region_a = 1;
boolean hasRegionA(); boolean hasRegionA();
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier getRegionA(); org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier getRegionA();
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder getRegionAOrBuilder(); org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder getRegionAOrBuilder();
// required .RegionSpecifier regionB = 2; // required .RegionSpecifier region_b = 2;
boolean hasRegionB(); boolean hasRegionB();
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier getRegionB(); org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier getRegionB();
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder getRegionBOrBuilder(); org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder getRegionBOrBuilder();
@ -8921,8 +8921,8 @@ public final class AdminProtos {
} }
private int bitField0_; private int bitField0_;
// required .RegionSpecifier regionA = 1; // required .RegionSpecifier region_a = 1;
public static final int REGIONA_FIELD_NUMBER = 1; public static final int REGION_A_FIELD_NUMBER = 1;
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier regionA_; private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier regionA_;
public boolean hasRegionA() { public boolean hasRegionA() {
return ((bitField0_ & 0x00000001) == 0x00000001); return ((bitField0_ & 0x00000001) == 0x00000001);
@ -8934,8 +8934,8 @@ public final class AdminProtos {
return regionA_; return regionA_;
} }
// required .RegionSpecifier regionB = 2; // required .RegionSpecifier region_b = 2;
public static final int REGIONB_FIELD_NUMBER = 2; public static final int REGION_B_FIELD_NUMBER = 2;
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier regionB_; private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier regionB_;
public boolean hasRegionB() { public boolean hasRegionB() {
return ((bitField0_ & 0x00000002) == 0x00000002); return ((bitField0_ & 0x00000002) == 0x00000002);
@ -9068,11 +9068,11 @@ public final class AdminProtos {
int hash = 41; int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode(); hash = (19 * hash) + getDescriptorForType().hashCode();
if (hasRegionA()) { if (hasRegionA()) {
hash = (37 * hash) + REGIONA_FIELD_NUMBER; hash = (37 * hash) + REGION_A_FIELD_NUMBER;
hash = (53 * hash) + getRegionA().hashCode(); hash = (53 * hash) + getRegionA().hashCode();
} }
if (hasRegionB()) { if (hasRegionB()) {
hash = (37 * hash) + REGIONB_FIELD_NUMBER; hash = (37 * hash) + REGION_B_FIELD_NUMBER;
hash = (53 * hash) + getRegionB().hashCode(); hash = (53 * hash) + getRegionB().hashCode();
} }
if (hasForcible()) { if (hasForcible()) {
@ -9370,7 +9370,7 @@ public final class AdminProtos {
private int bitField0_; private int bitField0_;
// required .RegionSpecifier regionA = 1; // required .RegionSpecifier region_a = 1;
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier regionA_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance(); private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier regionA_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance();
private com.google.protobuf.SingleFieldBuilder< private com.google.protobuf.SingleFieldBuilder<
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder> regionABuilder_; org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder> regionABuilder_;
@ -9460,7 +9460,7 @@ public final class AdminProtos {
return regionABuilder_; return regionABuilder_;
} }
// required .RegionSpecifier regionB = 2; // required .RegionSpecifier region_b = 2;
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier regionB_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance(); private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier regionB_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance();
private com.google.protobuf.SingleFieldBuilder< private com.google.protobuf.SingleFieldBuilder<
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder> regionBBuilder_; org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder> regionBBuilder_;
@ -9890,12 +9890,12 @@ public final class AdminProtos {
org.apache.hadoop.hbase.protobuf.generated.WALProtos.WALKey getKey(); org.apache.hadoop.hbase.protobuf.generated.WALProtos.WALKey getKey();
org.apache.hadoop.hbase.protobuf.generated.WALProtos.WALKeyOrBuilder getKeyOrBuilder(); org.apache.hadoop.hbase.protobuf.generated.WALProtos.WALKeyOrBuilder getKeyOrBuilder();
// repeated bytes keyValueBytes = 2; // repeated bytes key_value_bytes = 2;
java.util.List<com.google.protobuf.ByteString> getKeyValueBytesList(); java.util.List<com.google.protobuf.ByteString> getKeyValueBytesList();
int getKeyValueBytesCount(); int getKeyValueBytesCount();
com.google.protobuf.ByteString getKeyValueBytes(int index); com.google.protobuf.ByteString getKeyValueBytes(int index);
// optional int32 associatedCellCount = 3; // optional int32 associated_cell_count = 3;
boolean hasAssociatedCellCount(); boolean hasAssociatedCellCount();
int getAssociatedCellCount(); int getAssociatedCellCount();
} }
@ -9941,8 +9941,8 @@ public final class AdminProtos {
return key_; return key_;
} }
// repeated bytes keyValueBytes = 2; // repeated bytes key_value_bytes = 2;
public static final int KEYVALUEBYTES_FIELD_NUMBER = 2; public static final int KEY_VALUE_BYTES_FIELD_NUMBER = 2;
private java.util.List<com.google.protobuf.ByteString> keyValueBytes_; private java.util.List<com.google.protobuf.ByteString> keyValueBytes_;
public java.util.List<com.google.protobuf.ByteString> public java.util.List<com.google.protobuf.ByteString>
getKeyValueBytesList() { getKeyValueBytesList() {
@ -9955,8 +9955,8 @@ public final class AdminProtos {
return keyValueBytes_.get(index); return keyValueBytes_.get(index);
} }
// optional int32 associatedCellCount = 3; // optional int32 associated_cell_count = 3;
public static final int ASSOCIATEDCELLCOUNT_FIELD_NUMBER = 3; public static final int ASSOCIATED_CELL_COUNT_FIELD_NUMBER = 3;
private int associatedCellCount_; private int associatedCellCount_;
public boolean hasAssociatedCellCount() { public boolean hasAssociatedCellCount() {
return ((bitField0_ & 0x00000002) == 0x00000002); return ((bitField0_ & 0x00000002) == 0x00000002);
@ -10074,11 +10074,11 @@ public final class AdminProtos {
hash = (53 * hash) + getKey().hashCode(); hash = (53 * hash) + getKey().hashCode();
} }
if (getKeyValueBytesCount() > 0) { if (getKeyValueBytesCount() > 0) {
hash = (37 * hash) + KEYVALUEBYTES_FIELD_NUMBER; hash = (37 * hash) + KEY_VALUE_BYTES_FIELD_NUMBER;
hash = (53 * hash) + getKeyValueBytesList().hashCode(); hash = (53 * hash) + getKeyValueBytesList().hashCode();
} }
if (hasAssociatedCellCount()) { if (hasAssociatedCellCount()) {
hash = (37 * hash) + ASSOCIATEDCELLCOUNT_FIELD_NUMBER; hash = (37 * hash) + ASSOCIATED_CELL_COUNT_FIELD_NUMBER;
hash = (53 * hash) + getAssociatedCellCount(); hash = (53 * hash) + getAssociatedCellCount();
} }
hash = (29 * hash) + getUnknownFields().hashCode(); hash = (29 * hash) + getUnknownFields().hashCode();
@ -10449,7 +10449,7 @@ public final class AdminProtos {
return keyBuilder_; return keyBuilder_;
} }
// repeated bytes keyValueBytes = 2; // repeated bytes key_value_bytes = 2;
private java.util.List<com.google.protobuf.ByteString> keyValueBytes_ = java.util.Collections.emptyList();; private java.util.List<com.google.protobuf.ByteString> keyValueBytes_ = java.util.Collections.emptyList();;
private void ensureKeyValueBytesIsMutable() { private void ensureKeyValueBytesIsMutable() {
if (!((bitField0_ & 0x00000002) == 0x00000002)) { if (!((bitField0_ & 0x00000002) == 0x00000002)) {
@ -10500,7 +10500,7 @@ public final class AdminProtos {
return this; return this;
} }
// optional int32 associatedCellCount = 3; // optional int32 associated_cell_count = 3;
private int associatedCellCount_ ; private int associatedCellCount_ ;
public boolean hasAssociatedCellCount() { public boolean hasAssociatedCellCount() {
return ((bitField0_ & 0x00000004) == 0x00000004); return ((bitField0_ & 0x00000004) == 0x00000004);
@ -11728,7 +11728,7 @@ public final class AdminProtos {
public interface RollWALWriterResponseOrBuilder public interface RollWALWriterResponseOrBuilder
extends com.google.protobuf.MessageOrBuilder { extends com.google.protobuf.MessageOrBuilder {
// repeated bytes regionToFlush = 1; // repeated bytes region_to_flush = 1;
java.util.List<com.google.protobuf.ByteString> getRegionToFlushList(); java.util.List<com.google.protobuf.ByteString> getRegionToFlushList();
int getRegionToFlushCount(); int getRegionToFlushCount();
com.google.protobuf.ByteString getRegionToFlush(int index); com.google.protobuf.ByteString getRegionToFlush(int index);
@ -11761,8 +11761,8 @@ public final class AdminProtos {
return org.apache.hadoop.hbase.protobuf.generated.AdminProtos.internal_static_RollWALWriterResponse_fieldAccessorTable; return org.apache.hadoop.hbase.protobuf.generated.AdminProtos.internal_static_RollWALWriterResponse_fieldAccessorTable;
} }
// repeated bytes regionToFlush = 1; // repeated bytes region_to_flush = 1;
public static final int REGIONTOFLUSH_FIELD_NUMBER = 1; public static final int REGION_TO_FLUSH_FIELD_NUMBER = 1;
private java.util.List<com.google.protobuf.ByteString> regionToFlush_; private java.util.List<com.google.protobuf.ByteString> regionToFlush_;
public java.util.List<com.google.protobuf.ByteString> public java.util.List<com.google.protobuf.ByteString>
getRegionToFlushList() { getRegionToFlushList() {
@ -11846,7 +11846,7 @@ public final class AdminProtos {
int hash = 41; int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode(); hash = (19 * hash) + getDescriptorForType().hashCode();
if (getRegionToFlushCount() > 0) { if (getRegionToFlushCount() > 0) {
hash = (37 * hash) + REGIONTOFLUSH_FIELD_NUMBER; hash = (37 * hash) + REGION_TO_FLUSH_FIELD_NUMBER;
hash = (53 * hash) + getRegionToFlushList().hashCode(); hash = (53 * hash) + getRegionToFlushList().hashCode();
} }
hash = (29 * hash) + getUnknownFields().hashCode(); hash = (29 * hash) + getUnknownFields().hashCode();
@ -12076,7 +12076,7 @@ public final class AdminProtos {
private int bitField0_; private int bitField0_;
// repeated bytes regionToFlush = 1; // repeated bytes region_to_flush = 1;
private java.util.List<com.google.protobuf.ByteString> regionToFlush_ = java.util.Collections.emptyList();; private java.util.List<com.google.protobuf.ByteString> regionToFlush_ = java.util.Collections.emptyList();;
private void ensureRegionToFlushIsMutable() { private void ensureRegionToFlushIsMutable() {
if (!((bitField0_ & 0x00000001) == 0x00000001)) { if (!((bitField0_ & 0x00000001) == 0x00000001)) {
@ -13157,12 +13157,12 @@ public final class AdminProtos {
public interface ServerInfoOrBuilder public interface ServerInfoOrBuilder
extends com.google.protobuf.MessageOrBuilder { extends com.google.protobuf.MessageOrBuilder {
// required .ServerName serverName = 1; // required .ServerName server_name = 1;
boolean hasServerName(); boolean hasServerName();
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName getServerName(); org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName getServerName();
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerNameOrBuilder getServerNameOrBuilder(); org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerNameOrBuilder getServerNameOrBuilder();
// optional uint32 webuiPort = 2; // optional uint32 webui_port = 2;
boolean hasWebuiPort(); boolean hasWebuiPort();
int getWebuiPort(); int getWebuiPort();
} }
@ -13195,8 +13195,8 @@ public final class AdminProtos {
} }
private int bitField0_; private int bitField0_;
// required .ServerName serverName = 1; // required .ServerName server_name = 1;
public static final int SERVERNAME_FIELD_NUMBER = 1; public static final int SERVER_NAME_FIELD_NUMBER = 1;
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName serverName_; private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName serverName_;
public boolean hasServerName() { public boolean hasServerName() {
return ((bitField0_ & 0x00000001) == 0x00000001); return ((bitField0_ & 0x00000001) == 0x00000001);
@ -13208,8 +13208,8 @@ public final class AdminProtos {
return serverName_; return serverName_;
} }
// optional uint32 webuiPort = 2; // optional uint32 webui_port = 2;
public static final int WEBUIPORT_FIELD_NUMBER = 2; public static final int WEBUI_PORT_FIELD_NUMBER = 2;
private int webuiPort_; private int webuiPort_;
public boolean hasWebuiPort() { public boolean hasWebuiPort() {
return ((bitField0_ & 0x00000002) == 0x00000002); return ((bitField0_ & 0x00000002) == 0x00000002);
@ -13308,11 +13308,11 @@ public final class AdminProtos {
int hash = 41; int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode(); hash = (19 * hash) + getDescriptorForType().hashCode();
if (hasServerName()) { if (hasServerName()) {
hash = (37 * hash) + SERVERNAME_FIELD_NUMBER; hash = (37 * hash) + SERVER_NAME_FIELD_NUMBER;
hash = (53 * hash) + getServerName().hashCode(); hash = (53 * hash) + getServerName().hashCode();
} }
if (hasWebuiPort()) { if (hasWebuiPort()) {
hash = (37 * hash) + WEBUIPORT_FIELD_NUMBER; hash = (37 * hash) + WEBUI_PORT_FIELD_NUMBER;
hash = (53 * hash) + getWebuiPort(); hash = (53 * hash) + getWebuiPort();
} }
hash = (29 * hash) + getUnknownFields().hashCode(); hash = (29 * hash) + getUnknownFields().hashCode();
@ -13571,7 +13571,7 @@ public final class AdminProtos {
private int bitField0_; private int bitField0_;
// required .ServerName serverName = 1; // required .ServerName server_name = 1;
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName serverName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.getDefaultInstance(); private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName serverName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.getDefaultInstance();
private com.google.protobuf.SingleFieldBuilder< private com.google.protobuf.SingleFieldBuilder<
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerNameOrBuilder> serverNameBuilder_; org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerNameOrBuilder> serverNameBuilder_;
@ -13661,7 +13661,7 @@ public final class AdminProtos {
return serverNameBuilder_; return serverNameBuilder_;
} }
// optional uint32 webuiPort = 2; // optional uint32 webui_port = 2;
private int webuiPort_ ; private int webuiPort_ ;
public boolean hasWebuiPort() { public boolean hasWebuiPort() {
return ((bitField0_ & 0x00000002) == 0x00000002); return ((bitField0_ & 0x00000002) == 0x00000002);
@ -13696,7 +13696,7 @@ public final class AdminProtos {
public interface GetServerInfoResponseOrBuilder public interface GetServerInfoResponseOrBuilder
extends com.google.protobuf.MessageOrBuilder { extends com.google.protobuf.MessageOrBuilder {
// required .ServerInfo serverInfo = 1; // required .ServerInfo server_info = 1;
boolean hasServerInfo(); boolean hasServerInfo();
org.apache.hadoop.hbase.protobuf.generated.AdminProtos.ServerInfo getServerInfo(); org.apache.hadoop.hbase.protobuf.generated.AdminProtos.ServerInfo getServerInfo();
org.apache.hadoop.hbase.protobuf.generated.AdminProtos.ServerInfoOrBuilder getServerInfoOrBuilder(); org.apache.hadoop.hbase.protobuf.generated.AdminProtos.ServerInfoOrBuilder getServerInfoOrBuilder();
@ -13730,8 +13730,8 @@ public final class AdminProtos {
} }
private int bitField0_; private int bitField0_;
// required .ServerInfo serverInfo = 1; // required .ServerInfo server_info = 1;
public static final int SERVERINFO_FIELD_NUMBER = 1; public static final int SERVER_INFO_FIELD_NUMBER = 1;
private org.apache.hadoop.hbase.protobuf.generated.AdminProtos.ServerInfo serverInfo_; private org.apache.hadoop.hbase.protobuf.generated.AdminProtos.ServerInfo serverInfo_;
public boolean hasServerInfo() { public boolean hasServerInfo() {
return ((bitField0_ & 0x00000001) == 0x00000001); return ((bitField0_ & 0x00000001) == 0x00000001);
@ -13820,7 +13820,7 @@ public final class AdminProtos {
int hash = 41; int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode(); hash = (19 * hash) + getDescriptorForType().hashCode();
if (hasServerInfo()) { if (hasServerInfo()) {
hash = (37 * hash) + SERVERINFO_FIELD_NUMBER; hash = (37 * hash) + SERVER_INFO_FIELD_NUMBER;
hash = (53 * hash) + getServerInfo().hashCode(); hash = (53 * hash) + getServerInfo().hashCode();
} }
hash = (29 * hash) + getUnknownFields().hashCode(); hash = (29 * hash) + getUnknownFields().hashCode();
@ -14065,7 +14065,7 @@ public final class AdminProtos {
private int bitField0_; private int bitField0_;
// required .ServerInfo serverInfo = 1; // required .ServerInfo server_info = 1;
private org.apache.hadoop.hbase.protobuf.generated.AdminProtos.ServerInfo serverInfo_ = org.apache.hadoop.hbase.protobuf.generated.AdminProtos.ServerInfo.getDefaultInstance(); private org.apache.hadoop.hbase.protobuf.generated.AdminProtos.ServerInfo serverInfo_ = org.apache.hadoop.hbase.protobuf.generated.AdminProtos.ServerInfo.getDefaultInstance();
private com.google.protobuf.SingleFieldBuilder< private com.google.protobuf.SingleFieldBuilder<
org.apache.hadoop.hbase.protobuf.generated.AdminProtos.ServerInfo, org.apache.hadoop.hbase.protobuf.generated.AdminProtos.ServerInfo.Builder, org.apache.hadoop.hbase.protobuf.generated.AdminProtos.ServerInfoOrBuilder> serverInfoBuilder_; org.apache.hadoop.hbase.protobuf.generated.AdminProtos.ServerInfo, org.apache.hadoop.hbase.protobuf.generated.AdminProtos.ServerInfo.Builder, org.apache.hadoop.hbase.protobuf.generated.AdminProtos.ServerInfoOrBuilder> serverInfoBuilder_;
@ -15387,79 +15387,80 @@ public final class AdminProtos {
static { static {
java.lang.String[] descriptorData = { java.lang.String[] descriptorData = {
"\n\013Admin.proto\032\014Client.proto\032\013hbase.proto" + "\n\013Admin.proto\032\014Client.proto\032\013hbase.proto" +
"\032\tWAL.proto\"Q\n\024GetRegionInfoRequest\022 \n\006r" + "\032\tWAL.proto\"R\n\024GetRegionInfoRequest\022 \n\006r" +
"egion\030\001 \002(\0132\020.RegionSpecifier\022\027\n\017compact" + "egion\030\001 \002(\0132\020.RegionSpecifier\022\030\n\020compact" +
"ionState\030\002 \001(\010\"\301\001\n\025GetRegionInfoResponse" + "ion_state\030\002 \001(\010\"\303\001\n\025GetRegionInfoRespons" +
"\022\037\n\nregionInfo\030\001 \002(\0132\013.RegionInfo\022?\n\017com" + "e\022 \n\013region_info\030\001 \002(\0132\013.RegionInfo\022@\n\020c" +
"pactionState\030\002 \001(\0162&.GetRegionInfoRespon" + "ompaction_state\030\002 \001(\0162&.GetRegionInfoRes" +
"se.CompactionState\"F\n\017CompactionState\022\010\n" + "ponse.CompactionState\"F\n\017CompactionState" +
"\004NONE\020\000\022\t\n\005MINOR\020\001\022\t\n\005MAJOR\020\002\022\023\n\017MAJOR_A" + "\022\010\n\004NONE\020\000\022\t\n\005MINOR\020\001\022\t\n\005MAJOR\020\002\022\023\n\017MAJO" +
"ND_MINOR\020\003\"G\n\023GetStoreFileRequest\022 \n\006reg" + "R_AND_MINOR\020\003\"G\n\023GetStoreFileRequest\022 \n\006" +
"ion\030\001 \002(\0132\020.RegionSpecifier\022\016\n\006family\030\002 ", "region\030\001 \002(\0132\020.RegionSpecifier\022\016\n\006family",
"\003(\014\")\n\024GetStoreFileResponse\022\021\n\tstoreFile" + "\030\002 \003(\014\"*\n\024GetStoreFileResponse\022\022\n\nstore_" +
"\030\001 \003(\t\"\030\n\026GetOnlineRegionRequest\":\n\027GetO" + "file\030\001 \003(\t\"\030\n\026GetOnlineRegionRequest\";\n\027" +
"nlineRegionResponse\022\037\n\nregionInfo\030\001 \003(\0132" + "GetOnlineRegionResponse\022 \n\013region_info\030\001" +
"\013.RegionInfo\"\270\001\n\021OpenRegionRequest\0223\n\010op" + " \003(\0132\013.RegionInfo\"\275\001\n\021OpenRegionRequest\022" +
"enInfo\030\001 \003(\0132!.OpenRegionRequest.RegionO" + "4\n\topen_info\030\001 \003(\0132!.OpenRegionRequest.R" +
"penInfo\032n\n\016RegionOpenInfo\022\033\n\006region\030\001 \002(" + "egionOpenInfo\032r\n\016RegionOpenInfo\022\033\n\006regio" +
"\0132\013.RegionInfo\022\034\n\024versionOfOfflineNode\030\002" + "n\030\001 \002(\0132\013.RegionInfo\022\037\n\027version_of_offli" +
" \001(\r\022!\n\014favoredNodes\030\003 \003(\0132\013.ServerName\"" + "ne_node\030\002 \001(\r\022\"\n\rfavored_nodes\030\003 \003(\0132\013.S" +
"\234\001\n\022OpenRegionResponse\022<\n\014openingState\030\001" + "erverName\"\235\001\n\022OpenRegionResponse\022=\n\ropen" +
" \003(\0162&.OpenRegionResponse.RegionOpeningS", "ing_state\030\001 \003(\0162&.OpenRegionResponse.Reg",
"tate\"H\n\022RegionOpeningState\022\n\n\006OPENED\020\000\022\022" + "ionOpeningState\"H\n\022RegionOpeningState\022\n\n" +
"\n\016ALREADY_OPENED\020\001\022\022\n\016FAILED_OPENING\020\002\"\232" + "\006OPENED\020\000\022\022\n\016ALREADY_OPENED\020\001\022\022\n\016FAILED_" +
"\001\n\022CloseRegionRequest\022 \n\006region\030\001 \002(\0132\020." + "OPENING\020\002\"\240\001\n\022CloseRegionRequest\022 \n\006regi" +
"RegionSpecifier\022\034\n\024versionOfClosingNode\030" + "on\030\001 \002(\0132\020.RegionSpecifier\022\037\n\027version_of" +
"\002 \001(\r\022\034\n\016transitionInZK\030\003 \001(\010:\004true\022&\n\021d" + "_closing_node\030\002 \001(\r\022\036\n\020transition_in_ZK\030" +
"estinationServer\030\004 \001(\0132\013.ServerName\"%\n\023C" + "\003 \001(\010:\004true\022\'\n\022destination_server\030\004 \001(\0132" +
"loseRegionResponse\022\016\n\006closed\030\001 \002(\010\"M\n\022Fl" + "\013.ServerName\"%\n\023CloseRegionResponse\022\016\n\006c" +
"ushRegionRequest\022 \n\006region\030\001 \002(\0132\020.Regio" + "losed\030\001 \002(\010\"P\n\022FlushRegionRequest\022 \n\006reg" +
"nSpecifier\022\025\n\rifOlderThanTs\030\002 \001(\004\"=\n\023Flu" + "ion\030\001 \002(\0132\020.RegionSpecifier\022\030\n\020if_older_" +
"shRegionResponse\022\025\n\rlastFlushTime\030\001 \002(\004\022", "than_ts\030\002 \001(\004\"?\n\023FlushRegionResponse\022\027\n\017",
"\017\n\007flushed\030\002 \001(\010\"J\n\022SplitRegionRequest\022 " + "last_flush_time\030\001 \002(\004\022\017\n\007flushed\030\002 \001(\010\"K" +
"\n\006region\030\001 \002(\0132\020.RegionSpecifier\022\022\n\nspli" + "\n\022SplitRegionRequest\022 \n\006region\030\001 \002(\0132\020.R" +
"tPoint\030\002 \001(\014\"\025\n\023SplitRegionResponse\"W\n\024C" + "egionSpecifier\022\023\n\013split_point\030\002 \001(\014\"\025\n\023S" +
"ompactRegionRequest\022 \n\006region\030\001 \002(\0132\020.Re" + "plitRegionResponse\"W\n\024CompactRegionReque" +
"gionSpecifier\022\r\n\005major\030\002 \001(\010\022\016\n\006family\030\003" + "st\022 \n\006region\030\001 \002(\0132\020.RegionSpecifier\022\r\n\005" +
" \001(\014\"\027\n\025CompactRegionResponse\"t\n\023MergeRe" + "major\030\002 \001(\010\022\016\n\006family\030\003 \001(\014\"\027\n\025CompactRe" +
"gionsRequest\022!\n\007regionA\030\001 \002(\0132\020.RegionSp" + "gionResponse\"v\n\023MergeRegionsRequest\022\"\n\010r" +
"ecifier\022!\n\007regionB\030\002 \002(\0132\020.RegionSpecifi" + "egion_a\030\001 \002(\0132\020.RegionSpecifier\022\"\n\010regio" +
"er\022\027\n\010forcible\030\003 \001(\010:\005false\"\026\n\024MergeRegi" + "n_b\030\002 \002(\0132\020.RegionSpecifier\022\027\n\010forcible\030" +
"onsResponse\"T\n\010WALEntry\022\024\n\003key\030\001 \002(\0132\007.W", "\003 \001(\010:\005false\"\026\n\024MergeRegionsResponse\"X\n\010",
"ALKey\022\025\n\rkeyValueBytes\030\002 \003(\014\022\033\n\023associat" + "WALEntry\022\024\n\003key\030\001 \002(\0132\007.WALKey\022\027\n\017key_va" +
"edCellCount\030\003 \001(\005\"4\n\030ReplicateWALEntryRe" + "lue_bytes\030\002 \003(\014\022\035\n\025associated_cell_count" +
"quest\022\030\n\005entry\030\001 \003(\0132\t.WALEntry\"\033\n\031Repli" + "\030\003 \001(\005\"4\n\030ReplicateWALEntryRequest\022\030\n\005en" +
"cateWALEntryResponse\"\026\n\024RollWALWriterReq" + "try\030\001 \003(\0132\t.WALEntry\"\033\n\031ReplicateWALEntr" +
"uest\".\n\025RollWALWriterResponse\022\025\n\rregionT" + "yResponse\"\026\n\024RollWALWriterRequest\"0\n\025Rol" +
"oFlush\030\001 \003(\014\"#\n\021StopServerRequest\022\016\n\006rea" + "lWALWriterResponse\022\027\n\017region_to_flush\030\001 " +
"son\030\001 \002(\t\"\024\n\022StopServerResponse\"\026\n\024GetSe" + "\003(\014\"#\n\021StopServerRequest\022\016\n\006reason\030\001 \002(\t" +
"rverInfoRequest\"@\n\nServerInfo\022\037\n\nserverN" + "\"\024\n\022StopServerResponse\"\026\n\024GetServerInfoR" +
"ame\030\001 \002(\0132\013.ServerName\022\021\n\twebuiPort\030\002 \001(" + "equest\"B\n\nServerInfo\022 \n\013server_name\030\001 \002(" +
"\r\"8\n\025GetServerInfoResponse\022\037\n\nserverInfo", "\0132\013.ServerName\022\022\n\nwebui_port\030\002 \001(\r\"9\n\025Ge",
"\030\001 \002(\0132\013.ServerInfo2\337\006\n\014AdminService\022>\n\r" + "tServerInfoResponse\022 \n\013server_info\030\001 \002(\013" +
"getRegionInfo\022\025.GetRegionInfoRequest\032\026.G" + "2\013.ServerInfo2\337\006\n\014AdminService\022>\n\rGetReg" +
"etRegionInfoResponse\022;\n\014getStoreFile\022\024.G" + "ionInfo\022\025.GetRegionInfoRequest\032\026.GetRegi" +
"etStoreFileRequest\032\025.GetStoreFileRespons" + "onInfoResponse\022;\n\014GetStoreFile\022\024.GetStor" +
"e\022D\n\017getOnlineRegion\022\027.GetOnlineRegionRe" + "eFileRequest\032\025.GetStoreFileResponse\022D\n\017G" +
"quest\032\030.GetOnlineRegionResponse\0225\n\nopenR" + "etOnlineRegion\022\027.GetOnlineRegionRequest\032" +
"egion\022\022.OpenRegionRequest\032\023.OpenRegionRe" + "\030.GetOnlineRegionResponse\0225\n\nOpenRegion\022" +
"sponse\0228\n\013closeRegion\022\023.CloseRegionReque" + "\022.OpenRegionRequest\032\023.OpenRegionResponse" +
"st\032\024.CloseRegionResponse\0228\n\013flushRegion\022" + "\0228\n\013CloseRegion\022\023.CloseRegionRequest\032\024.C" +
"\023.FlushRegionRequest\032\024.FlushRegionRespon", "loseRegionResponse\0228\n\013FlushRegion\022\023.Flus",
"se\0228\n\013splitRegion\022\023.SplitRegionRequest\032\024" + "hRegionRequest\032\024.FlushRegionResponse\0228\n\013" +
".SplitRegionResponse\022>\n\rcompactRegion\022\025." + "SplitRegion\022\023.SplitRegionRequest\032\024.Split" +
"CompactRegionRequest\032\026.CompactRegionResp" + "RegionResponse\022>\n\rCompactRegion\022\025.Compac" +
"onse\022;\n\014mergeRegions\022\024.MergeRegionsReque" + "tRegionRequest\032\026.CompactRegionResponse\022;" +
"st\032\025.MergeRegionsResponse\022J\n\021replicateWA" + "\n\014MergeRegions\022\024.MergeRegionsRequest\032\025.M" +
"LEntry\022\031.ReplicateWALEntryRequest\032\032.Repl" + "ergeRegionsResponse\022J\n\021ReplicateWALEntry" +
"icateWALEntryResponse\022\'\n\006replay\022\r.MultiR" + "\022\031.ReplicateWALEntryRequest\032\032.ReplicateW" +
"equest\032\016.MultiResponse\022>\n\rrollWALWriter\022" + "ALEntryResponse\022\'\n\006Replay\022\r.MultiRequest" +
"\025.RollWALWriterRequest\032\026.RollWALWriterRe" + "\032\016.MultiResponse\022>\n\rRollWALWriter\022\025.Roll" +
"sponse\022>\n\rgetServerInfo\022\025.GetServerInfoR", "WALWriterRequest\032\026.RollWALWriterResponse",
"equest\032\026.GetServerInfoResponse\0225\n\nstopSe" + "\022>\n\rGetServerInfo\022\025.GetServerInfoRequest" +
"rver\022\022.StopServerRequest\032\023.StopServerRes" + "\032\026.GetServerInfoResponse\0225\n\nStopServer\022\022" +
"ponseBA\n*org.apache.hadoop.hbase.protobu" + ".StopServerRequest\032\023.StopServerResponseB" +
"f.generatedB\013AdminProtosH\001\210\001\001\240\001\001" "A\n*org.apache.hadoop.hbase.protobuf.gene" +
"ratedB\013AdminProtosH\001\210\001\001\240\001\001"
}; };
com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner = com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() { new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() {

View File

@ -11,7 +11,7 @@ public final class AggregateProtos {
public interface AggregateArgumentOrBuilder public interface AggregateArgumentOrBuilder
extends com.google.protobuf.MessageOrBuilder { extends com.google.protobuf.MessageOrBuilder {
// required string interpreterClassName = 1; // required string interpreter_class_name = 1;
boolean hasInterpreterClassName(); boolean hasInterpreterClassName();
String getInterpreterClassName(); String getInterpreterClassName();
@ -20,7 +20,7 @@ public final class AggregateProtos {
org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Scan getScan(); org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Scan getScan();
org.apache.hadoop.hbase.protobuf.generated.ClientProtos.ScanOrBuilder getScanOrBuilder(); org.apache.hadoop.hbase.protobuf.generated.ClientProtos.ScanOrBuilder getScanOrBuilder();
// optional bytes interpreterSpecificBytes = 3; // optional bytes interpreter_specific_bytes = 3;
boolean hasInterpreterSpecificBytes(); boolean hasInterpreterSpecificBytes();
com.google.protobuf.ByteString getInterpreterSpecificBytes(); com.google.protobuf.ByteString getInterpreterSpecificBytes();
} }
@ -53,8 +53,8 @@ public final class AggregateProtos {
} }
private int bitField0_; private int bitField0_;
// required string interpreterClassName = 1; // required string interpreter_class_name = 1;
public static final int INTERPRETERCLASSNAME_FIELD_NUMBER = 1; public static final int INTERPRETER_CLASS_NAME_FIELD_NUMBER = 1;
private java.lang.Object interpreterClassName_; private java.lang.Object interpreterClassName_;
public boolean hasInterpreterClassName() { public boolean hasInterpreterClassName() {
return ((bitField0_ & 0x00000001) == 0x00000001); return ((bitField0_ & 0x00000001) == 0x00000001);
@ -98,8 +98,8 @@ public final class AggregateProtos {
return scan_; return scan_;
} }
// optional bytes interpreterSpecificBytes = 3; // optional bytes interpreter_specific_bytes = 3;
public static final int INTERPRETERSPECIFICBYTES_FIELD_NUMBER = 3; public static final int INTERPRETER_SPECIFIC_BYTES_FIELD_NUMBER = 3;
private com.google.protobuf.ByteString interpreterSpecificBytes_; private com.google.protobuf.ByteString interpreterSpecificBytes_;
public boolean hasInterpreterSpecificBytes() { public boolean hasInterpreterSpecificBytes() {
return ((bitField0_ & 0x00000004) == 0x00000004); return ((bitField0_ & 0x00000004) == 0x00000004);
@ -215,7 +215,7 @@ public final class AggregateProtos {
int hash = 41; int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode(); hash = (19 * hash) + getDescriptorForType().hashCode();
if (hasInterpreterClassName()) { if (hasInterpreterClassName()) {
hash = (37 * hash) + INTERPRETERCLASSNAME_FIELD_NUMBER; hash = (37 * hash) + INTERPRETER_CLASS_NAME_FIELD_NUMBER;
hash = (53 * hash) + getInterpreterClassName().hashCode(); hash = (53 * hash) + getInterpreterClassName().hashCode();
} }
if (hasScan()) { if (hasScan()) {
@ -223,7 +223,7 @@ public final class AggregateProtos {
hash = (53 * hash) + getScan().hashCode(); hash = (53 * hash) + getScan().hashCode();
} }
if (hasInterpreterSpecificBytes()) { if (hasInterpreterSpecificBytes()) {
hash = (37 * hash) + INTERPRETERSPECIFICBYTES_FIELD_NUMBER; hash = (37 * hash) + INTERPRETER_SPECIFIC_BYTES_FIELD_NUMBER;
hash = (53 * hash) + getInterpreterSpecificBytes().hashCode(); hash = (53 * hash) + getInterpreterSpecificBytes().hashCode();
} }
hash = (29 * hash) + getUnknownFields().hashCode(); hash = (29 * hash) + getUnknownFields().hashCode();
@ -500,7 +500,7 @@ public final class AggregateProtos {
private int bitField0_; private int bitField0_;
// required string interpreterClassName = 1; // required string interpreter_class_name = 1;
private java.lang.Object interpreterClassName_ = ""; private java.lang.Object interpreterClassName_ = "";
public boolean hasInterpreterClassName() { public boolean hasInterpreterClassName() {
return ((bitField0_ & 0x00000001) == 0x00000001); return ((bitField0_ & 0x00000001) == 0x00000001);
@ -626,7 +626,7 @@ public final class AggregateProtos {
return scanBuilder_; return scanBuilder_;
} }
// optional bytes interpreterSpecificBytes = 3; // optional bytes interpreter_specific_bytes = 3;
private com.google.protobuf.ByteString interpreterSpecificBytes_ = com.google.protobuf.ByteString.EMPTY; private com.google.protobuf.ByteString interpreterSpecificBytes_ = com.google.protobuf.ByteString.EMPTY;
public boolean hasInterpreterSpecificBytes() { public boolean hasInterpreterSpecificBytes() {
return ((bitField0_ & 0x00000004) == 0x00000004); return ((bitField0_ & 0x00000004) == 0x00000004);
@ -664,12 +664,12 @@ public final class AggregateProtos {
public interface AggregateResponseOrBuilder public interface AggregateResponseOrBuilder
extends com.google.protobuf.MessageOrBuilder { extends com.google.protobuf.MessageOrBuilder {
// repeated bytes firstPart = 1; // repeated bytes first_part = 1;
java.util.List<com.google.protobuf.ByteString> getFirstPartList(); java.util.List<com.google.protobuf.ByteString> getFirstPartList();
int getFirstPartCount(); int getFirstPartCount();
com.google.protobuf.ByteString getFirstPart(int index); com.google.protobuf.ByteString getFirstPart(int index);
// optional bytes secondPart = 2; // optional bytes second_part = 2;
boolean hasSecondPart(); boolean hasSecondPart();
com.google.protobuf.ByteString getSecondPart(); com.google.protobuf.ByteString getSecondPart();
} }
@ -702,8 +702,8 @@ public final class AggregateProtos {
} }
private int bitField0_; private int bitField0_;
// repeated bytes firstPart = 1; // repeated bytes first_part = 1;
public static final int FIRSTPART_FIELD_NUMBER = 1; public static final int FIRST_PART_FIELD_NUMBER = 1;
private java.util.List<com.google.protobuf.ByteString> firstPart_; private java.util.List<com.google.protobuf.ByteString> firstPart_;
public java.util.List<com.google.protobuf.ByteString> public java.util.List<com.google.protobuf.ByteString>
getFirstPartList() { getFirstPartList() {
@ -716,8 +716,8 @@ public final class AggregateProtos {
return firstPart_.get(index); return firstPart_.get(index);
} }
// optional bytes secondPart = 2; // optional bytes second_part = 2;
public static final int SECONDPART_FIELD_NUMBER = 2; public static final int SECOND_PART_FIELD_NUMBER = 2;
private com.google.protobuf.ByteString secondPart_; private com.google.protobuf.ByteString secondPart_;
public boolean hasSecondPart() { public boolean hasSecondPart() {
return ((bitField0_ & 0x00000001) == 0x00000001); return ((bitField0_ & 0x00000001) == 0x00000001);
@ -810,11 +810,11 @@ public final class AggregateProtos {
int hash = 41; int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode(); hash = (19 * hash) + getDescriptorForType().hashCode();
if (getFirstPartCount() > 0) { if (getFirstPartCount() > 0) {
hash = (37 * hash) + FIRSTPART_FIELD_NUMBER; hash = (37 * hash) + FIRST_PART_FIELD_NUMBER;
hash = (53 * hash) + getFirstPartList().hashCode(); hash = (53 * hash) + getFirstPartList().hashCode();
} }
if (hasSecondPart()) { if (hasSecondPart()) {
hash = (37 * hash) + SECONDPART_FIELD_NUMBER; hash = (37 * hash) + SECOND_PART_FIELD_NUMBER;
hash = (53 * hash) + getSecondPart().hashCode(); hash = (53 * hash) + getSecondPart().hashCode();
} }
hash = (29 * hash) + getUnknownFields().hashCode(); hash = (29 * hash) + getUnknownFields().hashCode();
@ -1060,7 +1060,7 @@ public final class AggregateProtos {
private int bitField0_; private int bitField0_;
// repeated bytes firstPart = 1; // repeated bytes first_part = 1;
private java.util.List<com.google.protobuf.ByteString> firstPart_ = java.util.Collections.emptyList();; private java.util.List<com.google.protobuf.ByteString> firstPart_ = java.util.Collections.emptyList();;
private void ensureFirstPartIsMutable() { private void ensureFirstPartIsMutable() {
if (!((bitField0_ & 0x00000001) == 0x00000001)) { if (!((bitField0_ & 0x00000001) == 0x00000001)) {
@ -1111,7 +1111,7 @@ public final class AggregateProtos {
return this; return this;
} }
// optional bytes secondPart = 2; // optional bytes second_part = 2;
private com.google.protobuf.ByteString secondPart_ = com.google.protobuf.ByteString.EMPTY; private com.google.protobuf.ByteString secondPart_ = com.google.protobuf.ByteString.EMPTY;
public boolean hasSecondPart() { public boolean hasSecondPart() {
return ((bitField0_ & 0x00000002) == 0x00000002); return ((bitField0_ & 0x00000002) == 0x00000002);
@ -1776,23 +1776,23 @@ public final class AggregateProtos {
descriptor; descriptor;
static { static {
java.lang.String[] descriptorData = { java.lang.String[] descriptorData = {
"\n\017Aggregate.proto\032\014Client.proto\"h\n\021Aggre" + "\n\017Aggregate.proto\032\014Client.proto\"l\n\021Aggre" +
"gateArgument\022\034\n\024interpreterClassName\030\001 \002" + "gateArgument\022\036\n\026interpreter_class_name\030\001" +
"(\t\022\023\n\004scan\030\002 \002(\0132\005.Scan\022 \n\030interpreterSp" + " \002(\t\022\023\n\004scan\030\002 \002(\0132\005.Scan\022\"\n\032interpreter" +
"ecificBytes\030\003 \001(\014\":\n\021AggregateResponse\022\021" + "_specific_bytes\030\003 \001(\014\"<\n\021AggregateRespon" +
"\n\tfirstPart\030\001 \003(\014\022\022\n\nsecondPart\030\002 \001(\0142\366\002" + "se\022\022\n\nfirst_part\030\001 \003(\014\022\023\n\013second_part\030\002 " +
"\n\020AggregateService\0220\n\006getMax\022\022.Aggregate" + "\001(\0142\366\002\n\020AggregateService\0220\n\006GetMax\022\022.Agg" +
"Argument\032\022.AggregateResponse\0220\n\006getMin\022\022" + "regateArgument\032\022.AggregateResponse\0220\n\006Ge" +
".AggregateArgument\032\022.AggregateResponse\0220" + "tMin\022\022.AggregateArgument\032\022.AggregateResp" +
"\n\006getSum\022\022.AggregateArgument\032\022.Aggregate" + "onse\0220\n\006GetSum\022\022.AggregateArgument\032\022.Agg" +
"Response\0223\n\tgetRowNum\022\022.AggregateArgumen", "regateResponse\0223\n\tGetRowNum\022\022.AggregateA",
"t\032\022.AggregateResponse\0220\n\006getAvg\022\022.Aggreg" + "rgument\032\022.AggregateResponse\0220\n\006GetAvg\022\022." +
"ateArgument\032\022.AggregateResponse\0220\n\006getSt" + "AggregateArgument\032\022.AggregateResponse\0220\n" +
"d\022\022.AggregateArgument\032\022.AggregateRespons" + "\006GetStd\022\022.AggregateArgument\032\022.AggregateR" +
"e\0223\n\tgetMedian\022\022.AggregateArgument\032\022.Agg" + "esponse\0223\n\tGetMedian\022\022.AggregateArgument" +
"regateResponseBE\n*org.apache.hadoop.hbas" + "\032\022.AggregateResponseBE\n*org.apache.hadoo" +
"e.protobuf.generatedB\017AggregateProtosH\001\210" + "p.hbase.protobuf.generatedB\017AggregatePro" +
"\001\001\240\001\001" "tosH\001\210\001\001\240\001\001"
}; };
com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner = com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() { new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() {

View File

@ -15,7 +15,7 @@ public final class AuthenticationProtos {
boolean hasId(); boolean hasId();
int getId(); int getId();
// required int64 expirationDate = 2; // required int64 expiration_date = 2;
boolean hasExpirationDate(); boolean hasExpirationDate();
long getExpirationDate(); long getExpirationDate();
@ -62,8 +62,8 @@ public final class AuthenticationProtos {
return id_; return id_;
} }
// required int64 expirationDate = 2; // required int64 expiration_date = 2;
public static final int EXPIRATIONDATE_FIELD_NUMBER = 2; public static final int EXPIRATION_DATE_FIELD_NUMBER = 2;
private long expirationDate_; private long expirationDate_;
public boolean hasExpirationDate() { public boolean hasExpirationDate() {
return ((bitField0_ & 0x00000002) == 0x00000002); return ((bitField0_ & 0x00000002) == 0x00000002);
@ -193,7 +193,7 @@ public final class AuthenticationProtos {
hash = (53 * hash) + getId(); hash = (53 * hash) + getId();
} }
if (hasExpirationDate()) { if (hasExpirationDate()) {
hash = (37 * hash) + EXPIRATIONDATE_FIELD_NUMBER; hash = (37 * hash) + EXPIRATION_DATE_FIELD_NUMBER;
hash = (53 * hash) + hashLong(getExpirationDate()); hash = (53 * hash) + hashLong(getExpirationDate());
} }
if (hasKey()) { if (hasKey()) {
@ -482,7 +482,7 @@ public final class AuthenticationProtos {
return this; return this;
} }
// required int64 expirationDate = 2; // required int64 expiration_date = 2;
private long expirationDate_ ; private long expirationDate_ ;
public boolean hasExpirationDate() { public boolean hasExpirationDate() {
return ((bitField0_ & 0x00000002) == 0x00000002); return ((bitField0_ & 0x00000002) == 0x00000002);
@ -549,19 +549,19 @@ public final class AuthenticationProtos {
boolean hasUsername(); boolean hasUsername();
com.google.protobuf.ByteString getUsername(); com.google.protobuf.ByteString getUsername();
// required int32 keyId = 3; // required int32 key_id = 3;
boolean hasKeyId(); boolean hasKeyId();
int getKeyId(); int getKeyId();
// optional int64 issueDate = 4; // optional int64 issue_date = 4;
boolean hasIssueDate(); boolean hasIssueDate();
long getIssueDate(); long getIssueDate();
// optional int64 expirationDate = 5; // optional int64 expiration_date = 5;
boolean hasExpirationDate(); boolean hasExpirationDate();
long getExpirationDate(); long getExpirationDate();
// optional int64 sequenceNumber = 6; // optional int64 sequence_number = 6;
boolean hasSequenceNumber(); boolean hasSequenceNumber();
long getSequenceNumber(); long getSequenceNumber();
} }
@ -680,8 +680,8 @@ public final class AuthenticationProtos {
return username_; return username_;
} }
// required int32 keyId = 3; // required int32 key_id = 3;
public static final int KEYID_FIELD_NUMBER = 3; public static final int KEY_ID_FIELD_NUMBER = 3;
private int keyId_; private int keyId_;
public boolean hasKeyId() { public boolean hasKeyId() {
return ((bitField0_ & 0x00000004) == 0x00000004); return ((bitField0_ & 0x00000004) == 0x00000004);
@ -690,8 +690,8 @@ public final class AuthenticationProtos {
return keyId_; return keyId_;
} }
// optional int64 issueDate = 4; // optional int64 issue_date = 4;
public static final int ISSUEDATE_FIELD_NUMBER = 4; public static final int ISSUE_DATE_FIELD_NUMBER = 4;
private long issueDate_; private long issueDate_;
public boolean hasIssueDate() { public boolean hasIssueDate() {
return ((bitField0_ & 0x00000008) == 0x00000008); return ((bitField0_ & 0x00000008) == 0x00000008);
@ -700,8 +700,8 @@ public final class AuthenticationProtos {
return issueDate_; return issueDate_;
} }
// optional int64 expirationDate = 5; // optional int64 expiration_date = 5;
public static final int EXPIRATIONDATE_FIELD_NUMBER = 5; public static final int EXPIRATION_DATE_FIELD_NUMBER = 5;
private long expirationDate_; private long expirationDate_;
public boolean hasExpirationDate() { public boolean hasExpirationDate() {
return ((bitField0_ & 0x00000010) == 0x00000010); return ((bitField0_ & 0x00000010) == 0x00000010);
@ -710,8 +710,8 @@ public final class AuthenticationProtos {
return expirationDate_; return expirationDate_;
} }
// optional int64 sequenceNumber = 6; // optional int64 sequence_number = 6;
public static final int SEQUENCENUMBER_FIELD_NUMBER = 6; public static final int SEQUENCE_NUMBER_FIELD_NUMBER = 6;
private long sequenceNumber_; private long sequenceNumber_;
public boolean hasSequenceNumber() { public boolean hasSequenceNumber() {
return ((bitField0_ & 0x00000020) == 0x00000020); return ((bitField0_ & 0x00000020) == 0x00000020);
@ -874,19 +874,19 @@ public final class AuthenticationProtos {
hash = (53 * hash) + getUsername().hashCode(); hash = (53 * hash) + getUsername().hashCode();
} }
if (hasKeyId()) { if (hasKeyId()) {
hash = (37 * hash) + KEYID_FIELD_NUMBER; hash = (37 * hash) + KEY_ID_FIELD_NUMBER;
hash = (53 * hash) + getKeyId(); hash = (53 * hash) + getKeyId();
} }
if (hasIssueDate()) { if (hasIssueDate()) {
hash = (37 * hash) + ISSUEDATE_FIELD_NUMBER; hash = (37 * hash) + ISSUE_DATE_FIELD_NUMBER;
hash = (53 * hash) + hashLong(getIssueDate()); hash = (53 * hash) + hashLong(getIssueDate());
} }
if (hasExpirationDate()) { if (hasExpirationDate()) {
hash = (37 * hash) + EXPIRATIONDATE_FIELD_NUMBER; hash = (37 * hash) + EXPIRATION_DATE_FIELD_NUMBER;
hash = (53 * hash) + hashLong(getExpirationDate()); hash = (53 * hash) + hashLong(getExpirationDate());
} }
if (hasSequenceNumber()) { if (hasSequenceNumber()) {
hash = (37 * hash) + SEQUENCENUMBER_FIELD_NUMBER; hash = (37 * hash) + SEQUENCE_NUMBER_FIELD_NUMBER;
hash = (53 * hash) + hashLong(getSequenceNumber()); hash = (53 * hash) + hashLong(getSequenceNumber());
} }
hash = (29 * hash) + getUnknownFields().hashCode(); hash = (29 * hash) + getUnknownFields().hashCode();
@ -1246,7 +1246,7 @@ public final class AuthenticationProtos {
return this; return this;
} }
// required int32 keyId = 3; // required int32 key_id = 3;
private int keyId_ ; private int keyId_ ;
public boolean hasKeyId() { public boolean hasKeyId() {
return ((bitField0_ & 0x00000004) == 0x00000004); return ((bitField0_ & 0x00000004) == 0x00000004);
@ -1267,7 +1267,7 @@ public final class AuthenticationProtos {
return this; return this;
} }
// optional int64 issueDate = 4; // optional int64 issue_date = 4;
private long issueDate_ ; private long issueDate_ ;
public boolean hasIssueDate() { public boolean hasIssueDate() {
return ((bitField0_ & 0x00000008) == 0x00000008); return ((bitField0_ & 0x00000008) == 0x00000008);
@ -1288,7 +1288,7 @@ public final class AuthenticationProtos {
return this; return this;
} }
// optional int64 expirationDate = 5; // optional int64 expiration_date = 5;
private long expirationDate_ ; private long expirationDate_ ;
public boolean hasExpirationDate() { public boolean hasExpirationDate() {
return ((bitField0_ & 0x00000010) == 0x00000010); return ((bitField0_ & 0x00000010) == 0x00000010);
@ -1309,7 +1309,7 @@ public final class AuthenticationProtos {
return this; return this;
} }
// optional int64 sequenceNumber = 6; // optional int64 sequence_number = 6;
private long sequenceNumber_ ; private long sequenceNumber_ ;
public boolean hasSequenceNumber() { public boolean hasSequenceNumber() {
return ((bitField0_ & 0x00000020) == 0x00000020); return ((bitField0_ & 0x00000020) == 0x00000020);
@ -2917,7 +2917,7 @@ public final class AuthenticationProtos {
boolean hasUsername(); boolean hasUsername();
String getUsername(); String getUsername();
// optional string authMethod = 2; // optional string auth_method = 2;
boolean hasAuthMethod(); boolean hasAuthMethod();
String getAuthMethod(); String getAuthMethod();
} }
@ -2982,8 +2982,8 @@ public final class AuthenticationProtos {
} }
} }
// optional string authMethod = 2; // optional string auth_method = 2;
public static final int AUTHMETHOD_FIELD_NUMBER = 2; public static final int AUTH_METHOD_FIELD_NUMBER = 2;
private java.lang.Object authMethod_; private java.lang.Object authMethod_;
public boolean hasAuthMethod() { public boolean hasAuthMethod() {
return ((bitField0_ & 0x00000002) == 0x00000002); return ((bitField0_ & 0x00000002) == 0x00000002);
@ -3100,7 +3100,7 @@ public final class AuthenticationProtos {
hash = (53 * hash) + getUsername().hashCode(); hash = (53 * hash) + getUsername().hashCode();
} }
if (hasAuthMethod()) { if (hasAuthMethod()) {
hash = (37 * hash) + AUTHMETHOD_FIELD_NUMBER; hash = (37 * hash) + AUTH_METHOD_FIELD_NUMBER;
hash = (53 * hash) + getAuthMethod().hashCode(); hash = (53 * hash) + getAuthMethod().hashCode();
} }
hash = (29 * hash) + getUnknownFields().hashCode(); hash = (29 * hash) + getUnknownFields().hashCode();
@ -3374,7 +3374,7 @@ public final class AuthenticationProtos {
onChanged(); onChanged();
} }
// optional string authMethod = 2; // optional string auth_method = 2;
private java.lang.Object authMethod_ = ""; private java.lang.Object authMethod_ = "";
public boolean hasAuthMethod() { public boolean hasAuthMethod() {
return ((bitField0_ & 0x00000002) == 0x00000002); return ((bitField0_ & 0x00000002) == 0x00000002);
@ -3431,7 +3431,7 @@ public final class AuthenticationProtos {
org.apache.hadoop.hbase.protobuf.generated.AuthenticationProtos.TokenRequest request, org.apache.hadoop.hbase.protobuf.generated.AuthenticationProtos.TokenRequest request,
com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.protobuf.generated.AuthenticationProtos.TokenResponse> done); com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.protobuf.generated.AuthenticationProtos.TokenResponse> done);
public abstract void whoami( public abstract void whoAmI(
com.google.protobuf.RpcController controller, com.google.protobuf.RpcController controller,
org.apache.hadoop.hbase.protobuf.generated.AuthenticationProtos.WhoAmIRequest request, org.apache.hadoop.hbase.protobuf.generated.AuthenticationProtos.WhoAmIRequest request,
com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.protobuf.generated.AuthenticationProtos.WhoAmIResponse> done); com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.protobuf.generated.AuthenticationProtos.WhoAmIResponse> done);
@ -3450,11 +3450,11 @@ public final class AuthenticationProtos {
} }
@java.lang.Override @java.lang.Override
public void whoami( public void whoAmI(
com.google.protobuf.RpcController controller, com.google.protobuf.RpcController controller,
org.apache.hadoop.hbase.protobuf.generated.AuthenticationProtos.WhoAmIRequest request, org.apache.hadoop.hbase.protobuf.generated.AuthenticationProtos.WhoAmIRequest request,
com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.protobuf.generated.AuthenticationProtos.WhoAmIResponse> done) { com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.protobuf.generated.AuthenticationProtos.WhoAmIResponse> done) {
impl.whoami(controller, request, done); impl.whoAmI(controller, request, done);
} }
}; };
@ -3482,7 +3482,7 @@ public final class AuthenticationProtos {
case 0: case 0:
return impl.getAuthenticationToken(controller, (org.apache.hadoop.hbase.protobuf.generated.AuthenticationProtos.TokenRequest)request); return impl.getAuthenticationToken(controller, (org.apache.hadoop.hbase.protobuf.generated.AuthenticationProtos.TokenRequest)request);
case 1: case 1:
return impl.whoami(controller, (org.apache.hadoop.hbase.protobuf.generated.AuthenticationProtos.WhoAmIRequest)request); return impl.whoAmI(controller, (org.apache.hadoop.hbase.protobuf.generated.AuthenticationProtos.WhoAmIRequest)request);
default: default:
throw new java.lang.AssertionError("Can't get here."); throw new java.lang.AssertionError("Can't get here.");
} }
@ -3532,7 +3532,7 @@ public final class AuthenticationProtos {
org.apache.hadoop.hbase.protobuf.generated.AuthenticationProtos.TokenRequest request, org.apache.hadoop.hbase.protobuf.generated.AuthenticationProtos.TokenRequest request,
com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.protobuf.generated.AuthenticationProtos.TokenResponse> done); com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.protobuf.generated.AuthenticationProtos.TokenResponse> done);
public abstract void whoami( public abstract void whoAmI(
com.google.protobuf.RpcController controller, com.google.protobuf.RpcController controller,
org.apache.hadoop.hbase.protobuf.generated.AuthenticationProtos.WhoAmIRequest request, org.apache.hadoop.hbase.protobuf.generated.AuthenticationProtos.WhoAmIRequest request,
com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.protobuf.generated.AuthenticationProtos.WhoAmIResponse> done); com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.protobuf.generated.AuthenticationProtos.WhoAmIResponse> done);
@ -3565,7 +3565,7 @@ public final class AuthenticationProtos {
done)); done));
return; return;
case 1: case 1:
this.whoami(controller, (org.apache.hadoop.hbase.protobuf.generated.AuthenticationProtos.WhoAmIRequest)request, this.whoAmI(controller, (org.apache.hadoop.hbase.protobuf.generated.AuthenticationProtos.WhoAmIRequest)request,
com.google.protobuf.RpcUtil.<org.apache.hadoop.hbase.protobuf.generated.AuthenticationProtos.WhoAmIResponse>specializeCallback( com.google.protobuf.RpcUtil.<org.apache.hadoop.hbase.protobuf.generated.AuthenticationProtos.WhoAmIResponse>specializeCallback(
done)); done));
return; return;
@ -3641,7 +3641,7 @@ public final class AuthenticationProtos {
org.apache.hadoop.hbase.protobuf.generated.AuthenticationProtos.TokenResponse.getDefaultInstance())); org.apache.hadoop.hbase.protobuf.generated.AuthenticationProtos.TokenResponse.getDefaultInstance()));
} }
public void whoami( public void whoAmI(
com.google.protobuf.RpcController controller, com.google.protobuf.RpcController controller,
org.apache.hadoop.hbase.protobuf.generated.AuthenticationProtos.WhoAmIRequest request, org.apache.hadoop.hbase.protobuf.generated.AuthenticationProtos.WhoAmIRequest request,
com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.protobuf.generated.AuthenticationProtos.WhoAmIResponse> done) { com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.protobuf.generated.AuthenticationProtos.WhoAmIResponse> done) {
@ -3668,7 +3668,7 @@ public final class AuthenticationProtos {
org.apache.hadoop.hbase.protobuf.generated.AuthenticationProtos.TokenRequest request) org.apache.hadoop.hbase.protobuf.generated.AuthenticationProtos.TokenRequest request)
throws com.google.protobuf.ServiceException; throws com.google.protobuf.ServiceException;
public org.apache.hadoop.hbase.protobuf.generated.AuthenticationProtos.WhoAmIResponse whoami( public org.apache.hadoop.hbase.protobuf.generated.AuthenticationProtos.WhoAmIResponse whoAmI(
com.google.protobuf.RpcController controller, com.google.protobuf.RpcController controller,
org.apache.hadoop.hbase.protobuf.generated.AuthenticationProtos.WhoAmIRequest request) org.apache.hadoop.hbase.protobuf.generated.AuthenticationProtos.WhoAmIRequest request)
throws com.google.protobuf.ServiceException; throws com.google.protobuf.ServiceException;
@ -3693,7 +3693,7 @@ public final class AuthenticationProtos {
} }
public org.apache.hadoop.hbase.protobuf.generated.AuthenticationProtos.WhoAmIResponse whoami( public org.apache.hadoop.hbase.protobuf.generated.AuthenticationProtos.WhoAmIResponse whoAmI(
com.google.protobuf.RpcController controller, com.google.protobuf.RpcController controller,
org.apache.hadoop.hbase.protobuf.generated.AuthenticationProtos.WhoAmIRequest request) org.apache.hadoop.hbase.protobuf.generated.AuthenticationProtos.WhoAmIRequest request)
throws com.google.protobuf.ServiceException { throws com.google.protobuf.ServiceException {
@ -3751,23 +3751,23 @@ public final class AuthenticationProtos {
descriptor; descriptor;
static { static {
java.lang.String[] descriptorData = { java.lang.String[] descriptorData = {
"\n\024Authentication.proto\"D\n\021Authentication" + "\n\024Authentication.proto\"E\n\021Authentication" +
"Key\022\n\n\002id\030\001 \002(\005\022\026\n\016expirationDate\030\002 \002(\003\022" + "Key\022\n\n\002id\030\001 \002(\005\022\027\n\017expiration_date\030\002 \002(\003" +
"\013\n\003key\030\003 \002(\014\"\270\001\n\017TokenIdentifier\022#\n\004kind" + "\022\013\n\003key\030\003 \002(\014\"\274\001\n\017TokenIdentifier\022#\n\004kin" +
"\030\001 \002(\0162\025.TokenIdentifier.Kind\022\020\n\010usernam" + "d\030\001 \002(\0162\025.TokenIdentifier.Kind\022\020\n\010userna" +
"e\030\002 \002(\014\022\r\n\005keyId\030\003 \002(\005\022\021\n\tissueDate\030\004 \001(" + "me\030\002 \002(\014\022\016\n\006key_id\030\003 \002(\005\022\022\n\nissue_date\030\004" +
"\003\022\026\n\016expirationDate\030\005 \001(\003\022\026\n\016sequenceNum" + " \001(\003\022\027\n\017expiration_date\030\005 \001(\003\022\027\n\017sequenc" +
"ber\030\006 \001(\003\"\034\n\004Kind\022\024\n\020HBASE_AUTH_TOKEN\020\000\"" + "e_number\030\006 \001(\003\"\034\n\004Kind\022\024\n\020HBASE_AUTH_TOK" +
">\n\005Token\022\022\n\nidentifier\030\001 \001(\014\022\020\n\010password" + "EN\020\000\">\n\005Token\022\022\n\nidentifier\030\001 \001(\014\022\020\n\010pas" +
"\030\002 \001(\014\022\017\n\007service\030\003 \001(\014\"\016\n\014TokenRequest\"" + "sword\030\002 \001(\014\022\017\n\007service\030\003 \001(\014\"\016\n\014TokenReq" +
"&\n\rTokenResponse\022\025\n\005token\030\001 \001(\0132\006.Token\"", "uest\"&\n\rTokenResponse\022\025\n\005token\030\001 \001(\0132\006.T",
"\017\n\rWhoAmIRequest\"6\n\016WhoAmIResponse\022\020\n\010us" + "oken\"\017\n\rWhoAmIRequest\"7\n\016WhoAmIResponse\022" +
"ername\030\001 \001(\t\022\022\n\nauthMethod\030\002 \001(\t2{\n\025Auth" + "\020\n\010username\030\001 \001(\t\022\023\n\013auth_method\030\002 \001(\t2{" +
"enticationService\0227\n\026getAuthenticationTo" + "\n\025AuthenticationService\0227\n\026GetAuthentica" +
"ken\022\r.TokenRequest\032\016.TokenResponse\022)\n\006wh" + "tionToken\022\r.TokenRequest\032\016.TokenResponse" +
"oami\022\016.WhoAmIRequest\032\017.WhoAmIResponseBJ\n" + "\022)\n\006WhoAmI\022\016.WhoAmIRequest\032\017.WhoAmIRespo" +
"*org.apache.hadoop.hbase.protobuf.genera" + "nseBJ\n*org.apache.hadoop.hbase.protobuf." +
"tedB\024AuthenticationProtosH\001\210\001\001\240\001\001" "generatedB\024AuthenticationProtosH\001\210\001\001\240\001\001"
}; };
com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner = com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() { new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() {

View File

@ -108,7 +108,7 @@ public final class CellProtos {
boolean hasTimestamp(); boolean hasTimestamp();
long getTimestamp(); long getTimestamp();
// optional .CellType cellType = 5; // optional .CellType cell_type = 5;
boolean hasCellType(); boolean hasCellType();
org.apache.hadoop.hbase.protobuf.generated.CellProtos.CellType getCellType(); org.apache.hadoop.hbase.protobuf.generated.CellProtos.CellType getCellType();
@ -185,8 +185,8 @@ public final class CellProtos {
return timestamp_; return timestamp_;
} }
// optional .CellType cellType = 5; // optional .CellType cell_type = 5;
public static final int CELLTYPE_FIELD_NUMBER = 5; public static final int CELL_TYPE_FIELD_NUMBER = 5;
private org.apache.hadoop.hbase.protobuf.generated.CellProtos.CellType cellType_; private org.apache.hadoop.hbase.protobuf.generated.CellProtos.CellType cellType_;
public boolean hasCellType() { public boolean hasCellType() {
return ((bitField0_ & 0x00000010) == 0x00000010); return ((bitField0_ & 0x00000010) == 0x00000010);
@ -355,7 +355,7 @@ public final class CellProtos {
hash = (53 * hash) + hashLong(getTimestamp()); hash = (53 * hash) + hashLong(getTimestamp());
} }
if (hasCellType()) { if (hasCellType()) {
hash = (37 * hash) + CELLTYPE_FIELD_NUMBER; hash = (37 * hash) + CELL_TYPE_FIELD_NUMBER;
hash = (53 * hash) + hashEnum(getCellType()); hash = (53 * hash) + hashEnum(getCellType());
} }
if (hasValue()) { if (hasValue()) {
@ -752,7 +752,7 @@ public final class CellProtos {
return this; return this;
} }
// optional .CellType cellType = 5; // optional .CellType cell_type = 5;
private org.apache.hadoop.hbase.protobuf.generated.CellProtos.CellType cellType_ = org.apache.hadoop.hbase.protobuf.generated.CellProtos.CellType.MINIMUM; private org.apache.hadoop.hbase.protobuf.generated.CellProtos.CellType cellType_ = org.apache.hadoop.hbase.protobuf.generated.CellProtos.CellType.MINIMUM;
public boolean hasCellType() { public boolean hasCellType() {
return ((bitField0_ & 0x00000010) == 0x00000010); return ((bitField0_ & 0x00000010) == 0x00000010);
@ -830,7 +830,7 @@ public final class CellProtos {
boolean hasTimestamp(); boolean hasTimestamp();
long getTimestamp(); long getTimestamp();
// optional .CellType keyType = 5; // optional .CellType key_type = 5;
boolean hasKeyType(); boolean hasKeyType();
org.apache.hadoop.hbase.protobuf.generated.CellProtos.CellType getKeyType(); org.apache.hadoop.hbase.protobuf.generated.CellProtos.CellType getKeyType();
@ -907,8 +907,8 @@ public final class CellProtos {
return timestamp_; return timestamp_;
} }
// optional .CellType keyType = 5; // optional .CellType key_type = 5;
public static final int KEYTYPE_FIELD_NUMBER = 5; public static final int KEY_TYPE_FIELD_NUMBER = 5;
private org.apache.hadoop.hbase.protobuf.generated.CellProtos.CellType keyType_; private org.apache.hadoop.hbase.protobuf.generated.CellProtos.CellType keyType_;
public boolean hasKeyType() { public boolean hasKeyType() {
return ((bitField0_ & 0x00000010) == 0x00000010); return ((bitField0_ & 0x00000010) == 0x00000010);
@ -1089,7 +1089,7 @@ public final class CellProtos {
hash = (53 * hash) + hashLong(getTimestamp()); hash = (53 * hash) + hashLong(getTimestamp());
} }
if (hasKeyType()) { if (hasKeyType()) {
hash = (37 * hash) + KEYTYPE_FIELD_NUMBER; hash = (37 * hash) + KEY_TYPE_FIELD_NUMBER;
hash = (53 * hash) + hashEnum(getKeyType()); hash = (53 * hash) + hashEnum(getKeyType());
} }
if (hasValue()) { if (hasValue()) {
@ -1498,7 +1498,7 @@ public final class CellProtos {
return this; return this;
} }
// optional .CellType keyType = 5; // optional .CellType key_type = 5;
private org.apache.hadoop.hbase.protobuf.generated.CellProtos.CellType keyType_ = org.apache.hadoop.hbase.protobuf.generated.CellProtos.CellType.MINIMUM; private org.apache.hadoop.hbase.protobuf.generated.CellProtos.CellType keyType_ = org.apache.hadoop.hbase.protobuf.generated.CellProtos.CellType.MINIMUM;
public boolean hasKeyType() { public boolean hasKeyType() {
return ((bitField0_ & 0x00000010) == 0x00000010); return ((bitField0_ & 0x00000010) == 0x00000010);
@ -1576,17 +1576,17 @@ public final class CellProtos {
descriptor; descriptor;
static { static {
java.lang.String[] descriptorData = { java.lang.String[] descriptorData = {
"\n\nCell.proto\"u\n\004Cell\022\013\n\003row\030\001 \001(\014\022\016\n\006fam" + "\n\nCell.proto\"v\n\004Cell\022\013\n\003row\030\001 \001(\014\022\016\n\006fam" +
"ily\030\002 \001(\014\022\021\n\tqualifier\030\003 \001(\014\022\021\n\ttimestam" + "ily\030\002 \001(\014\022\021\n\tqualifier\030\003 \001(\014\022\021\n\ttimestam" +
"p\030\004 \001(\004\022\033\n\010cellType\030\005 \001(\0162\t.CellType\022\r\n\005" + "p\030\004 \001(\004\022\034\n\tcell_type\030\005 \001(\0162\t.CellType\022\r\n" +
"value\030\006 \001(\014\"x\n\010KeyValue\022\013\n\003row\030\001 \002(\014\022\016\n\006" + "\005value\030\006 \001(\014\"y\n\010KeyValue\022\013\n\003row\030\001 \002(\014\022\016\n" +
"family\030\002 \002(\014\022\021\n\tqualifier\030\003 \002(\014\022\021\n\ttimes" + "\006family\030\002 \002(\014\022\021\n\tqualifier\030\003 \002(\014\022\021\n\ttime" +
"tamp\030\004 \001(\004\022\032\n\007keyType\030\005 \001(\0162\t.CellType\022\r" + "stamp\030\004 \001(\004\022\033\n\010key_type\030\005 \001(\0162\t.CellType" +
"\n\005value\030\006 \001(\014*`\n\010CellType\022\013\n\007MINIMUM\020\000\022\007" + "\022\r\n\005value\030\006 \001(\014*`\n\010CellType\022\013\n\007MINIMUM\020\000" +
"\n\003PUT\020\004\022\n\n\006DELETE\020\010\022\021\n\rDELETE_COLUMN\020\014\022\021" + "\022\007\n\003PUT\020\004\022\n\n\006DELETE\020\010\022\021\n\rDELETE_COLUMN\020\014" +
"\n\rDELETE_FAMILY\020\016\022\014\n\007MAXIMUM\020\377\001B=\n*org.a" + "\022\021\n\rDELETE_FAMILY\020\016\022\014\n\007MAXIMUM\020\377\001B=\n*org" +
"pache.hadoop.hbase.protobuf.generatedB\nC", ".apache.hadoop.hbase.protobuf.generatedB",
"ellProtosH\001\240\001\001" "\nCellProtosH\001\240\001\001"
}; };
com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner = com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() { new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() {

View File

@ -11,7 +11,7 @@ public final class ClusterIdProtos {
public interface ClusterIdOrBuilder public interface ClusterIdOrBuilder
extends com.google.protobuf.MessageOrBuilder { extends com.google.protobuf.MessageOrBuilder {
// required string clusterId = 1; // required string cluster_id = 1;
boolean hasClusterId(); boolean hasClusterId();
String getClusterId(); String getClusterId();
} }
@ -44,8 +44,8 @@ public final class ClusterIdProtos {
} }
private int bitField0_; private int bitField0_;
// required string clusterId = 1; // required string cluster_id = 1;
public static final int CLUSTERID_FIELD_NUMBER = 1; public static final int CLUSTER_ID_FIELD_NUMBER = 1;
private java.lang.Object clusterId_; private java.lang.Object clusterId_;
public boolean hasClusterId() { public boolean hasClusterId() {
return ((bitField0_ & 0x00000001) == 0x00000001); return ((bitField0_ & 0x00000001) == 0x00000001);
@ -149,7 +149,7 @@ public final class ClusterIdProtos {
int hash = 41; int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode(); hash = (19 * hash) + getDescriptorForType().hashCode();
if (hasClusterId()) { if (hasClusterId()) {
hash = (37 * hash) + CLUSTERID_FIELD_NUMBER; hash = (37 * hash) + CLUSTER_ID_FIELD_NUMBER;
hash = (53 * hash) + getClusterId().hashCode(); hash = (53 * hash) + getClusterId().hashCode();
} }
hash = (29 * hash) + getUnknownFields().hashCode(); hash = (29 * hash) + getUnknownFields().hashCode();
@ -377,7 +377,7 @@ public final class ClusterIdProtos {
private int bitField0_; private int bitField0_;
// required string clusterId = 1; // required string cluster_id = 1;
private java.lang.Object clusterId_ = ""; private java.lang.Object clusterId_ = "";
public boolean hasClusterId() { public boolean hasClusterId() {
return ((bitField0_ & 0x00000001) == 0x00000001); return ((bitField0_ & 0x00000001) == 0x00000001);
@ -438,9 +438,9 @@ public final class ClusterIdProtos {
descriptor; descriptor;
static { static {
java.lang.String[] descriptorData = { java.lang.String[] descriptorData = {
"\n\017ClusterId.proto\"\036\n\tClusterId\022\021\n\tcluste" + "\n\017ClusterId.proto\"\037\n\tClusterId\022\022\n\ncluste" +
"rId\030\001 \002(\tBB\n*org.apache.hadoop.hbase.pro" + "r_id\030\001 \002(\tBB\n*org.apache.hadoop.hbase.pr" +
"tobuf.generatedB\017ClusterIdProtosH\001\240\001\001" "otobuf.generatedB\017ClusterIdProtosH\001\240\001\001"
}; };
com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner = com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() { new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() {

View File

@ -15,7 +15,7 @@ public final class ComparatorProtos {
boolean hasName(); boolean hasName();
String getName(); String getName();
// optional bytes serializedComparator = 2; // optional bytes serialized_comparator = 2;
boolean hasSerializedComparator(); boolean hasSerializedComparator();
com.google.protobuf.ByteString getSerializedComparator(); com.google.protobuf.ByteString getSerializedComparator();
} }
@ -80,8 +80,8 @@ public final class ComparatorProtos {
} }
} }
// optional bytes serializedComparator = 2; // optional bytes serialized_comparator = 2;
public static final int SERIALIZEDCOMPARATOR_FIELD_NUMBER = 2; public static final int SERIALIZED_COMPARATOR_FIELD_NUMBER = 2;
private com.google.protobuf.ByteString serializedComparator_; private com.google.protobuf.ByteString serializedComparator_;
public boolean hasSerializedComparator() { public boolean hasSerializedComparator() {
return ((bitField0_ & 0x00000002) == 0x00000002); return ((bitField0_ & 0x00000002) == 0x00000002);
@ -180,7 +180,7 @@ public final class ComparatorProtos {
hash = (53 * hash) + getName().hashCode(); hash = (53 * hash) + getName().hashCode();
} }
if (hasSerializedComparator()) { if (hasSerializedComparator()) {
hash = (37 * hash) + SERIALIZEDCOMPARATOR_FIELD_NUMBER; hash = (37 * hash) + SERIALIZED_COMPARATOR_FIELD_NUMBER;
hash = (53 * hash) + getSerializedComparator().hashCode(); hash = (53 * hash) + getSerializedComparator().hashCode();
} }
hash = (29 * hash) + getUnknownFields().hashCode(); hash = (29 * hash) + getUnknownFields().hashCode();
@ -458,7 +458,7 @@ public final class ComparatorProtos {
onChanged(); onChanged();
} }
// optional bytes serializedComparator = 2; // optional bytes serialized_comparator = 2;
private com.google.protobuf.ByteString serializedComparator_ = com.google.protobuf.ByteString.EMPTY; private com.google.protobuf.ByteString serializedComparator_ = com.google.protobuf.ByteString.EMPTY;
public boolean hasSerializedComparator() { public boolean hasSerializedComparator() {
return ((bitField0_ & 0x00000002) == 0x00000002); return ((bitField0_ & 0x00000002) == 0x00000002);
@ -1805,7 +1805,7 @@ public final class ComparatorProtos {
org.apache.hadoop.hbase.protobuf.generated.ComparatorProtos.ByteArrayComparable getComparable(); org.apache.hadoop.hbase.protobuf.generated.ComparatorProtos.ByteArrayComparable getComparable();
org.apache.hadoop.hbase.protobuf.generated.ComparatorProtos.ByteArrayComparableOrBuilder getComparableOrBuilder(); org.apache.hadoop.hbase.protobuf.generated.ComparatorProtos.ByteArrayComparableOrBuilder getComparableOrBuilder();
// required .BitComparator.BitwiseOp bitwiseOp = 2; // required .BitComparator.BitwiseOp bitwise_op = 2;
boolean hasBitwiseOp(); boolean hasBitwiseOp();
org.apache.hadoop.hbase.protobuf.generated.ComparatorProtos.BitComparator.BitwiseOp getBitwiseOp(); org.apache.hadoop.hbase.protobuf.generated.ComparatorProtos.BitComparator.BitwiseOp getBitwiseOp();
} }
@ -1923,8 +1923,8 @@ public final class ComparatorProtos {
return comparable_; return comparable_;
} }
// required .BitComparator.BitwiseOp bitwiseOp = 2; // required .BitComparator.BitwiseOp bitwise_op = 2;
public static final int BITWISEOP_FIELD_NUMBER = 2; public static final int BITWISE_OP_FIELD_NUMBER = 2;
private org.apache.hadoop.hbase.protobuf.generated.ComparatorProtos.BitComparator.BitwiseOp bitwiseOp_; private org.apache.hadoop.hbase.protobuf.generated.ComparatorProtos.BitComparator.BitwiseOp bitwiseOp_;
public boolean hasBitwiseOp() { public boolean hasBitwiseOp() {
return ((bitField0_ & 0x00000002) == 0x00000002); return ((bitField0_ & 0x00000002) == 0x00000002);
@ -2027,7 +2027,7 @@ public final class ComparatorProtos {
hash = (53 * hash) + getComparable().hashCode(); hash = (53 * hash) + getComparable().hashCode();
} }
if (hasBitwiseOp()) { if (hasBitwiseOp()) {
hash = (37 * hash) + BITWISEOP_FIELD_NUMBER; hash = (37 * hash) + BITWISE_OP_FIELD_NUMBER;
hash = (53 * hash) + hashEnum(getBitwiseOp()); hash = (53 * hash) + hashEnum(getBitwiseOp());
} }
hash = (29 * hash) + getUnknownFields().hashCode(); hash = (29 * hash) + getUnknownFields().hashCode();
@ -2382,7 +2382,7 @@ public final class ComparatorProtos {
return comparableBuilder_; return comparableBuilder_;
} }
// required .BitComparator.BitwiseOp bitwiseOp = 2; // required .BitComparator.BitwiseOp bitwise_op = 2;
private org.apache.hadoop.hbase.protobuf.generated.ComparatorProtos.BitComparator.BitwiseOp bitwiseOp_ = org.apache.hadoop.hbase.protobuf.generated.ComparatorProtos.BitComparator.BitwiseOp.AND; private org.apache.hadoop.hbase.protobuf.generated.ComparatorProtos.BitComparator.BitwiseOp bitwiseOp_ = org.apache.hadoop.hbase.protobuf.generated.ComparatorProtos.BitComparator.BitwiseOp.AND;
public boolean hasBitwiseOp() { public boolean hasBitwiseOp() {
return ((bitField0_ & 0x00000002) == 0x00000002); return ((bitField0_ & 0x00000002) == 0x00000002);
@ -2724,7 +2724,7 @@ public final class ComparatorProtos {
boolean hasPattern(); boolean hasPattern();
String getPattern(); String getPattern();
// required int32 patternFlags = 2; // required int32 pattern_flags = 2;
boolean hasPatternFlags(); boolean hasPatternFlags();
int getPatternFlags(); int getPatternFlags();
@ -2793,8 +2793,8 @@ public final class ComparatorProtos {
} }
} }
// required int32 patternFlags = 2; // required int32 pattern_flags = 2;
public static final int PATTERNFLAGS_FIELD_NUMBER = 2; public static final int PATTERN_FLAGS_FIELD_NUMBER = 2;
private int patternFlags_; private int patternFlags_;
public boolean hasPatternFlags() { public boolean hasPatternFlags() {
return ((bitField0_ & 0x00000002) == 0x00000002); return ((bitField0_ & 0x00000002) == 0x00000002);
@ -2946,7 +2946,7 @@ public final class ComparatorProtos {
hash = (53 * hash) + getPattern().hashCode(); hash = (53 * hash) + getPattern().hashCode();
} }
if (hasPatternFlags()) { if (hasPatternFlags()) {
hash = (37 * hash) + PATTERNFLAGS_FIELD_NUMBER; hash = (37 * hash) + PATTERN_FLAGS_FIELD_NUMBER;
hash = (53 * hash) + getPatternFlags(); hash = (53 * hash) + getPatternFlags();
} }
if (hasCharset()) { if (hasCharset()) {
@ -3250,7 +3250,7 @@ public final class ComparatorProtos {
onChanged(); onChanged();
} }
// required int32 patternFlags = 2; // required int32 pattern_flags = 2;
private int patternFlags_ ; private int patternFlags_ ;
public boolean hasPatternFlags() { public boolean hasPatternFlags() {
return ((bitField0_ & 0x00000002) == 0x00000002); return ((bitField0_ & 0x00000002) == 0x00000002);
@ -3783,21 +3783,22 @@ public final class ComparatorProtos {
descriptor; descriptor;
static { static {
java.lang.String[] descriptorData = { java.lang.String[] descriptorData = {
"\n\020Comparator.proto\"8\n\nComparator\022\014\n\004name" + "\n\020Comparator.proto\"9\n\nComparator\022\014\n\004name" +
"\030\001 \002(\t\022\034\n\024serializedComparator\030\002 \001(\014\"$\n\023" + "\030\001 \002(\t\022\035\n\025serialized_comparator\030\002 \001(\014\"$\n" +
"ByteArrayComparable\022\r\n\005value\030\001 \001(\014\"<\n\020Bi" + "\023ByteArrayComparable\022\r\n\005value\030\001 \001(\014\"<\n\020B" +
"naryComparator\022(\n\ncomparable\030\001 \002(\0132\024.Byt" + "inaryComparator\022(\n\ncomparable\030\001 \002(\0132\024.By" +
"eArrayComparable\"B\n\026BinaryPrefixComparat" + "teArrayComparable\"B\n\026BinaryPrefixCompara" +
"or\022(\n\ncomparable\030\001 \002(\0132\024.ByteArrayCompar" + "tor\022(\n\ncomparable\030\001 \002(\0132\024.ByteArrayCompa" +
"able\"\215\001\n\rBitComparator\022(\n\ncomparable\030\001 \002" + "rable\"\216\001\n\rBitComparator\022(\n\ncomparable\030\001 " +
"(\0132\024.ByteArrayComparable\022+\n\tbitwiseOp\030\002 " + "\002(\0132\024.ByteArrayComparable\022,\n\nbitwise_op\030" +
"\002(\0162\030.BitComparator.BitwiseOp\"%\n\tBitwise" + "\002 \002(\0162\030.BitComparator.BitwiseOp\"%\n\tBitwi" +
"Op\022\007\n\003AND\020\001\022\006\n\002OR\020\002\022\007\n\003XOR\020\003\"\020\n\016NullComp", "seOp\022\007\n\003AND\020\001\022\006\n\002OR\020\002\022\007\n\003XOR\020\003\"\020\n\016NullCo",
"arator\"O\n\025RegexStringComparator\022\017\n\007patte" + "mparator\"P\n\025RegexStringComparator\022\017\n\007pat" +
"rn\030\001 \002(\t\022\024\n\014patternFlags\030\002 \002(\005\022\017\n\007charse" + "tern\030\001 \002(\t\022\025\n\rpattern_flags\030\002 \002(\005\022\017\n\007cha" +
"t\030\003 \002(\t\"%\n\023SubstringComparator\022\016\n\006substr" + "rset\030\003 \002(\t\"%\n\023SubstringComparator\022\016\n\006sub" +
"\030\001 \002(\tBF\n*org.apache.hadoop.hbase.protob" + "str\030\001 \002(\tBF\n*org.apache.hadoop.hbase.pro" +
"uf.generatedB\020ComparatorProtosH\001\210\001\001\240\001\001" "tobuf.generatedB\020ComparatorProtosH\001\210\001\001\240\001" +
"\001"
}; };
com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner = com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() { new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() {

View File

@ -11,19 +11,19 @@ public final class ErrorHandlingProtos {
public interface StackTraceElementMessageOrBuilder public interface StackTraceElementMessageOrBuilder
extends com.google.protobuf.MessageOrBuilder { extends com.google.protobuf.MessageOrBuilder {
// optional string declaringClass = 1; // optional string declaring_class = 1;
boolean hasDeclaringClass(); boolean hasDeclaringClass();
String getDeclaringClass(); String getDeclaringClass();
// optional string methodName = 2; // optional string method_name = 2;
boolean hasMethodName(); boolean hasMethodName();
String getMethodName(); String getMethodName();
// optional string fileName = 3; // optional string file_name = 3;
boolean hasFileName(); boolean hasFileName();
String getFileName(); String getFileName();
// optional int32 lineNumber = 4; // optional int32 line_number = 4;
boolean hasLineNumber(); boolean hasLineNumber();
int getLineNumber(); int getLineNumber();
} }
@ -56,8 +56,8 @@ public final class ErrorHandlingProtos {
} }
private int bitField0_; private int bitField0_;
// optional string declaringClass = 1; // optional string declaring_class = 1;
public static final int DECLARINGCLASS_FIELD_NUMBER = 1; public static final int DECLARING_CLASS_FIELD_NUMBER = 1;
private java.lang.Object declaringClass_; private java.lang.Object declaringClass_;
public boolean hasDeclaringClass() { public boolean hasDeclaringClass() {
return ((bitField0_ & 0x00000001) == 0x00000001); return ((bitField0_ & 0x00000001) == 0x00000001);
@ -88,8 +88,8 @@ public final class ErrorHandlingProtos {
} }
} }
// optional string methodName = 2; // optional string method_name = 2;
public static final int METHODNAME_FIELD_NUMBER = 2; public static final int METHOD_NAME_FIELD_NUMBER = 2;
private java.lang.Object methodName_; private java.lang.Object methodName_;
public boolean hasMethodName() { public boolean hasMethodName() {
return ((bitField0_ & 0x00000002) == 0x00000002); return ((bitField0_ & 0x00000002) == 0x00000002);
@ -120,8 +120,8 @@ public final class ErrorHandlingProtos {
} }
} }
// optional string fileName = 3; // optional string file_name = 3;
public static final int FILENAME_FIELD_NUMBER = 3; public static final int FILE_NAME_FIELD_NUMBER = 3;
private java.lang.Object fileName_; private java.lang.Object fileName_;
public boolean hasFileName() { public boolean hasFileName() {
return ((bitField0_ & 0x00000004) == 0x00000004); return ((bitField0_ & 0x00000004) == 0x00000004);
@ -152,8 +152,8 @@ public final class ErrorHandlingProtos {
} }
} }
// optional int32 lineNumber = 4; // optional int32 line_number = 4;
public static final int LINENUMBER_FIELD_NUMBER = 4; public static final int LINE_NUMBER_FIELD_NUMBER = 4;
private int lineNumber_; private int lineNumber_;
public boolean hasLineNumber() { public boolean hasLineNumber() {
return ((bitField0_ & 0x00000008) == 0x00000008); return ((bitField0_ & 0x00000008) == 0x00000008);
@ -270,19 +270,19 @@ public final class ErrorHandlingProtos {
int hash = 41; int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode(); hash = (19 * hash) + getDescriptorForType().hashCode();
if (hasDeclaringClass()) { if (hasDeclaringClass()) {
hash = (37 * hash) + DECLARINGCLASS_FIELD_NUMBER; hash = (37 * hash) + DECLARING_CLASS_FIELD_NUMBER;
hash = (53 * hash) + getDeclaringClass().hashCode(); hash = (53 * hash) + getDeclaringClass().hashCode();
} }
if (hasMethodName()) { if (hasMethodName()) {
hash = (37 * hash) + METHODNAME_FIELD_NUMBER; hash = (37 * hash) + METHOD_NAME_FIELD_NUMBER;
hash = (53 * hash) + getMethodName().hashCode(); hash = (53 * hash) + getMethodName().hashCode();
} }
if (hasFileName()) { if (hasFileName()) {
hash = (37 * hash) + FILENAME_FIELD_NUMBER; hash = (37 * hash) + FILE_NAME_FIELD_NUMBER;
hash = (53 * hash) + getFileName().hashCode(); hash = (53 * hash) + getFileName().hashCode();
} }
if (hasLineNumber()) { if (hasLineNumber()) {
hash = (37 * hash) + LINENUMBER_FIELD_NUMBER; hash = (37 * hash) + LINE_NUMBER_FIELD_NUMBER;
hash = (53 * hash) + getLineNumber(); hash = (53 * hash) + getLineNumber();
} }
hash = (29 * hash) + getUnknownFields().hashCode(); hash = (29 * hash) + getUnknownFields().hashCode();
@ -548,7 +548,7 @@ public final class ErrorHandlingProtos {
private int bitField0_; private int bitField0_;
// optional string declaringClass = 1; // optional string declaring_class = 1;
private java.lang.Object declaringClass_ = ""; private java.lang.Object declaringClass_ = "";
public boolean hasDeclaringClass() { public boolean hasDeclaringClass() {
return ((bitField0_ & 0x00000001) == 0x00000001); return ((bitField0_ & 0x00000001) == 0x00000001);
@ -584,7 +584,7 @@ public final class ErrorHandlingProtos {
onChanged(); onChanged();
} }
// optional string methodName = 2; // optional string method_name = 2;
private java.lang.Object methodName_ = ""; private java.lang.Object methodName_ = "";
public boolean hasMethodName() { public boolean hasMethodName() {
return ((bitField0_ & 0x00000002) == 0x00000002); return ((bitField0_ & 0x00000002) == 0x00000002);
@ -620,7 +620,7 @@ public final class ErrorHandlingProtos {
onChanged(); onChanged();
} }
// optional string fileName = 3; // optional string file_name = 3;
private java.lang.Object fileName_ = ""; private java.lang.Object fileName_ = "";
public boolean hasFileName() { public boolean hasFileName() {
return ((bitField0_ & 0x00000004) == 0x00000004); return ((bitField0_ & 0x00000004) == 0x00000004);
@ -656,7 +656,7 @@ public final class ErrorHandlingProtos {
onChanged(); onChanged();
} }
// optional int32 lineNumber = 4; // optional int32 line_number = 4;
private int lineNumber_ ; private int lineNumber_ ;
public boolean hasLineNumber() { public boolean hasLineNumber() {
return ((bitField0_ & 0x00000008) == 0x00000008); return ((bitField0_ & 0x00000008) == 0x00000008);
@ -691,7 +691,7 @@ public final class ErrorHandlingProtos {
public interface GenericExceptionMessageOrBuilder public interface GenericExceptionMessageOrBuilder
extends com.google.protobuf.MessageOrBuilder { extends com.google.protobuf.MessageOrBuilder {
// optional string className = 1; // optional string class_name = 1;
boolean hasClassName(); boolean hasClassName();
String getClassName(); String getClassName();
@ -699,7 +699,7 @@ public final class ErrorHandlingProtos {
boolean hasMessage(); boolean hasMessage();
String getMessage(); String getMessage();
// optional bytes errorInfo = 3; // optional bytes error_info = 3;
boolean hasErrorInfo(); boolean hasErrorInfo();
com.google.protobuf.ByteString getErrorInfo(); com.google.protobuf.ByteString getErrorInfo();
@ -742,8 +742,8 @@ public final class ErrorHandlingProtos {
} }
private int bitField0_; private int bitField0_;
// optional string className = 1; // optional string class_name = 1;
public static final int CLASSNAME_FIELD_NUMBER = 1; public static final int CLASS_NAME_FIELD_NUMBER = 1;
private java.lang.Object className_; private java.lang.Object className_;
public boolean hasClassName() { public boolean hasClassName() {
return ((bitField0_ & 0x00000001) == 0x00000001); return ((bitField0_ & 0x00000001) == 0x00000001);
@ -806,8 +806,8 @@ public final class ErrorHandlingProtos {
} }
} }
// optional bytes errorInfo = 3; // optional bytes error_info = 3;
public static final int ERRORINFO_FIELD_NUMBER = 3; public static final int ERROR_INFO_FIELD_NUMBER = 3;
private com.google.protobuf.ByteString errorInfo_; private com.google.protobuf.ByteString errorInfo_;
public boolean hasErrorInfo() { public boolean hasErrorInfo() {
return ((bitField0_ & 0x00000004) == 0x00000004); return ((bitField0_ & 0x00000004) == 0x00000004);
@ -942,7 +942,7 @@ public final class ErrorHandlingProtos {
int hash = 41; int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode(); hash = (19 * hash) + getDescriptorForType().hashCode();
if (hasClassName()) { if (hasClassName()) {
hash = (37 * hash) + CLASSNAME_FIELD_NUMBER; hash = (37 * hash) + CLASS_NAME_FIELD_NUMBER;
hash = (53 * hash) + getClassName().hashCode(); hash = (53 * hash) + getClassName().hashCode();
} }
if (hasMessage()) { if (hasMessage()) {
@ -950,7 +950,7 @@ public final class ErrorHandlingProtos {
hash = (53 * hash) + getMessage().hashCode(); hash = (53 * hash) + getMessage().hashCode();
} }
if (hasErrorInfo()) { if (hasErrorInfo()) {
hash = (37 * hash) + ERRORINFO_FIELD_NUMBER; hash = (37 * hash) + ERROR_INFO_FIELD_NUMBER;
hash = (53 * hash) + getErrorInfo().hashCode(); hash = (53 * hash) + getErrorInfo().hashCode();
} }
if (getTraceCount() > 0) { if (getTraceCount() > 0) {
@ -1254,7 +1254,7 @@ public final class ErrorHandlingProtos {
private int bitField0_; private int bitField0_;
// optional string className = 1; // optional string class_name = 1;
private java.lang.Object className_ = ""; private java.lang.Object className_ = "";
public boolean hasClassName() { public boolean hasClassName() {
return ((bitField0_ & 0x00000001) == 0x00000001); return ((bitField0_ & 0x00000001) == 0x00000001);
@ -1326,7 +1326,7 @@ public final class ErrorHandlingProtos {
onChanged(); onChanged();
} }
// optional bytes errorInfo = 3; // optional bytes error_info = 3;
private com.google.protobuf.ByteString errorInfo_ = com.google.protobuf.ByteString.EMPTY; private com.google.protobuf.ByteString errorInfo_ = com.google.protobuf.ByteString.EMPTY;
public boolean hasErrorInfo() { public boolean hasErrorInfo() {
return ((bitField0_ & 0x00000004) == 0x00000004); return ((bitField0_ & 0x00000004) == 0x00000004);
@ -1554,7 +1554,7 @@ public final class ErrorHandlingProtos {
boolean hasSource(); boolean hasSource();
String getSource(); String getSource();
// optional .GenericExceptionMessage genericException = 2; // optional .GenericExceptionMessage generic_exception = 2;
boolean hasGenericException(); boolean hasGenericException();
org.apache.hadoop.hbase.protobuf.generated.ErrorHandlingProtos.GenericExceptionMessage getGenericException(); org.apache.hadoop.hbase.protobuf.generated.ErrorHandlingProtos.GenericExceptionMessage getGenericException();
org.apache.hadoop.hbase.protobuf.generated.ErrorHandlingProtos.GenericExceptionMessageOrBuilder getGenericExceptionOrBuilder(); org.apache.hadoop.hbase.protobuf.generated.ErrorHandlingProtos.GenericExceptionMessageOrBuilder getGenericExceptionOrBuilder();
@ -1620,8 +1620,8 @@ public final class ErrorHandlingProtos {
} }
} }
// optional .GenericExceptionMessage genericException = 2; // optional .GenericExceptionMessage generic_exception = 2;
public static final int GENERICEXCEPTION_FIELD_NUMBER = 2; public static final int GENERIC_EXCEPTION_FIELD_NUMBER = 2;
private org.apache.hadoop.hbase.protobuf.generated.ErrorHandlingProtos.GenericExceptionMessage genericException_; private org.apache.hadoop.hbase.protobuf.generated.ErrorHandlingProtos.GenericExceptionMessage genericException_;
public boolean hasGenericException() { public boolean hasGenericException() {
return ((bitField0_ & 0x00000002) == 0x00000002); return ((bitField0_ & 0x00000002) == 0x00000002);
@ -1719,7 +1719,7 @@ public final class ErrorHandlingProtos {
hash = (53 * hash) + getSource().hashCode(); hash = (53 * hash) + getSource().hashCode();
} }
if (hasGenericException()) { if (hasGenericException()) {
hash = (37 * hash) + GENERICEXCEPTION_FIELD_NUMBER; hash = (37 * hash) + GENERIC_EXCEPTION_FIELD_NUMBER;
hash = (53 * hash) + getGenericException().hashCode(); hash = (53 * hash) + getGenericException().hashCode();
} }
hash = (29 * hash) + getUnknownFields().hashCode(); hash = (29 * hash) + getUnknownFields().hashCode();
@ -2006,7 +2006,7 @@ public final class ErrorHandlingProtos {
onChanged(); onChanged();
} }
// optional .GenericExceptionMessage genericException = 2; // optional .GenericExceptionMessage generic_exception = 2;
private org.apache.hadoop.hbase.protobuf.generated.ErrorHandlingProtos.GenericExceptionMessage genericException_ = org.apache.hadoop.hbase.protobuf.generated.ErrorHandlingProtos.GenericExceptionMessage.getDefaultInstance(); private org.apache.hadoop.hbase.protobuf.generated.ErrorHandlingProtos.GenericExceptionMessage genericException_ = org.apache.hadoop.hbase.protobuf.generated.ErrorHandlingProtos.GenericExceptionMessage.getDefaultInstance();
private com.google.protobuf.SingleFieldBuilder< private com.google.protobuf.SingleFieldBuilder<
org.apache.hadoop.hbase.protobuf.generated.ErrorHandlingProtos.GenericExceptionMessage, org.apache.hadoop.hbase.protobuf.generated.ErrorHandlingProtos.GenericExceptionMessage.Builder, org.apache.hadoop.hbase.protobuf.generated.ErrorHandlingProtos.GenericExceptionMessageOrBuilder> genericExceptionBuilder_; org.apache.hadoop.hbase.protobuf.generated.ErrorHandlingProtos.GenericExceptionMessage, org.apache.hadoop.hbase.protobuf.generated.ErrorHandlingProtos.GenericExceptionMessage.Builder, org.apache.hadoop.hbase.protobuf.generated.ErrorHandlingProtos.GenericExceptionMessageOrBuilder> genericExceptionBuilder_;
@ -2131,17 +2131,17 @@ public final class ErrorHandlingProtos {
descriptor; descriptor;
static { static {
java.lang.String[] descriptorData = { java.lang.String[] descriptorData = {
"\n\023ErrorHandling.proto\"l\n\030StackTraceEleme" + "\n\023ErrorHandling.proto\"p\n\030StackTraceEleme" +
"ntMessage\022\026\n\016declaringClass\030\001 \001(\t\022\022\n\nmet" + "ntMessage\022\027\n\017declaring_class\030\001 \001(\t\022\023\n\013me" +
"hodName\030\002 \001(\t\022\020\n\010fileName\030\003 \001(\t\022\022\n\nlineN" + "thod_name\030\002 \001(\t\022\021\n\tfile_name\030\003 \001(\t\022\023\n\013li" +
"umber\030\004 \001(\005\"z\n\027GenericExceptionMessage\022\021" + "ne_number\030\004 \001(\005\"|\n\027GenericExceptionMessa" +
"\n\tclassName\030\001 \001(\t\022\017\n\007message\030\002 \001(\t\022\021\n\ter" + "ge\022\022\n\nclass_name\030\001 \001(\t\022\017\n\007message\030\002 \001(\t\022" +
"rorInfo\030\003 \001(\014\022(\n\005trace\030\004 \003(\0132\031.StackTrac" + "\022\n\nerror_info\030\003 \001(\014\022(\n\005trace\030\004 \003(\0132\031.Sta" +
"eElementMessage\"]\n\027ForeignExceptionMessa" + "ckTraceElementMessage\"^\n\027ForeignExceptio" +
"ge\022\016\n\006source\030\001 \001(\t\0222\n\020genericException\030\002" + "nMessage\022\016\n\006source\030\001 \001(\t\0223\n\021generic_exce" +
" \001(\0132\030.GenericExceptionMessageBF\n*org.ap" + "ption\030\002 \001(\0132\030.GenericExceptionMessageBF\n" +
"ache.hadoop.hbase.protobuf.generatedB\023Er", "*org.apache.hadoop.hbase.protobuf.genera",
"rorHandlingProtosH\001\240\001\001" "tedB\023ErrorHandlingProtosH\001\240\001\001"
}; };
com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner = com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() { new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() {

View File

@ -15,7 +15,7 @@ public final class FilterProtos {
boolean hasName(); boolean hasName();
String getName(); String getName();
// optional bytes serializedFilter = 2; // optional bytes serialized_filter = 2;
boolean hasSerializedFilter(); boolean hasSerializedFilter();
com.google.protobuf.ByteString getSerializedFilter(); com.google.protobuf.ByteString getSerializedFilter();
} }
@ -80,8 +80,8 @@ public final class FilterProtos {
} }
} }
// optional bytes serializedFilter = 2; // optional bytes serialized_filter = 2;
public static final int SERIALIZEDFILTER_FIELD_NUMBER = 2; public static final int SERIALIZED_FILTER_FIELD_NUMBER = 2;
private com.google.protobuf.ByteString serializedFilter_; private com.google.protobuf.ByteString serializedFilter_;
public boolean hasSerializedFilter() { public boolean hasSerializedFilter() {
return ((bitField0_ & 0x00000002) == 0x00000002); return ((bitField0_ & 0x00000002) == 0x00000002);
@ -180,7 +180,7 @@ public final class FilterProtos {
hash = (53 * hash) + getName().hashCode(); hash = (53 * hash) + getName().hashCode();
} }
if (hasSerializedFilter()) { if (hasSerializedFilter()) {
hash = (37 * hash) + SERIALIZEDFILTER_FIELD_NUMBER; hash = (37 * hash) + SERIALIZED_FILTER_FIELD_NUMBER;
hash = (53 * hash) + getSerializedFilter().hashCode(); hash = (53 * hash) + getSerializedFilter().hashCode();
} }
hash = (29 * hash) + getUnknownFields().hashCode(); hash = (29 * hash) + getUnknownFields().hashCode();
@ -458,7 +458,7 @@ public final class FilterProtos {
onChanged(); onChanged();
} }
// optional bytes serializedFilter = 2; // optional bytes serialized_filter = 2;
private com.google.protobuf.ByteString serializedFilter_ = com.google.protobuf.ByteString.EMPTY; private com.google.protobuf.ByteString serializedFilter_ = com.google.protobuf.ByteString.EMPTY;
public boolean hasSerializedFilter() { public boolean hasSerializedFilter() {
return ((bitField0_ & 0x00000002) == 0x00000002); return ((bitField0_ & 0x00000002) == 0x00000002);
@ -883,7 +883,7 @@ public final class FilterProtos {
boolean hasOffset(); boolean hasOffset();
int getOffset(); int getOffset();
// optional bytes columnOffset = 3; // optional bytes column_offset = 3;
boolean hasColumnOffset(); boolean hasColumnOffset();
com.google.protobuf.ByteString getColumnOffset(); com.google.protobuf.ByteString getColumnOffset();
} }
@ -936,8 +936,8 @@ public final class FilterProtos {
return offset_; return offset_;
} }
// optional bytes columnOffset = 3; // optional bytes column_offset = 3;
public static final int COLUMNOFFSET_FIELD_NUMBER = 3; public static final int COLUMN_OFFSET_FIELD_NUMBER = 3;
private com.google.protobuf.ByteString columnOffset_; private com.google.protobuf.ByteString columnOffset_;
public boolean hasColumnOffset() { public boolean hasColumnOffset() {
return ((bitField0_ & 0x00000004) == 0x00000004); return ((bitField0_ & 0x00000004) == 0x00000004);
@ -1053,7 +1053,7 @@ public final class FilterProtos {
hash = (53 * hash) + getOffset(); hash = (53 * hash) + getOffset();
} }
if (hasColumnOffset()) { if (hasColumnOffset()) {
hash = (37 * hash) + COLUMNOFFSET_FIELD_NUMBER; hash = (37 * hash) + COLUMN_OFFSET_FIELD_NUMBER;
hash = (53 * hash) + getColumnOffset().hashCode(); hash = (53 * hash) + getColumnOffset().hashCode();
} }
hash = (29 * hash) + getUnknownFields().hashCode(); hash = (29 * hash) + getUnknownFields().hashCode();
@ -1351,7 +1351,7 @@ public final class FilterProtos {
return this; return this;
} }
// optional bytes columnOffset = 3; // optional bytes column_offset = 3;
private com.google.protobuf.ByteString columnOffset_ = com.google.protobuf.ByteString.EMPTY; private com.google.protobuf.ByteString columnOffset_ = com.google.protobuf.ByteString.EMPTY;
public boolean hasColumnOffset() { public boolean hasColumnOffset() {
return ((bitField0_ & 0x00000004) == 0x00000004); return ((bitField0_ & 0x00000004) == 0x00000004);
@ -1771,19 +1771,19 @@ public final class FilterProtos {
public interface ColumnRangeFilterOrBuilder public interface ColumnRangeFilterOrBuilder
extends com.google.protobuf.MessageOrBuilder { extends com.google.protobuf.MessageOrBuilder {
// optional bytes minColumn = 1; // optional bytes min_column = 1;
boolean hasMinColumn(); boolean hasMinColumn();
com.google.protobuf.ByteString getMinColumn(); com.google.protobuf.ByteString getMinColumn();
// optional bool minColumnInclusive = 2; // optional bool min_column_inclusive = 2;
boolean hasMinColumnInclusive(); boolean hasMinColumnInclusive();
boolean getMinColumnInclusive(); boolean getMinColumnInclusive();
// optional bytes maxColumn = 3; // optional bytes max_column = 3;
boolean hasMaxColumn(); boolean hasMaxColumn();
com.google.protobuf.ByteString getMaxColumn(); com.google.protobuf.ByteString getMaxColumn();
// optional bool maxColumnInclusive = 4; // optional bool max_column_inclusive = 4;
boolean hasMaxColumnInclusive(); boolean hasMaxColumnInclusive();
boolean getMaxColumnInclusive(); boolean getMaxColumnInclusive();
} }
@ -1816,8 +1816,8 @@ public final class FilterProtos {
} }
private int bitField0_; private int bitField0_;
// optional bytes minColumn = 1; // optional bytes min_column = 1;
public static final int MINCOLUMN_FIELD_NUMBER = 1; public static final int MIN_COLUMN_FIELD_NUMBER = 1;
private com.google.protobuf.ByteString minColumn_; private com.google.protobuf.ByteString minColumn_;
public boolean hasMinColumn() { public boolean hasMinColumn() {
return ((bitField0_ & 0x00000001) == 0x00000001); return ((bitField0_ & 0x00000001) == 0x00000001);
@ -1826,8 +1826,8 @@ public final class FilterProtos {
return minColumn_; return minColumn_;
} }
// optional bool minColumnInclusive = 2; // optional bool min_column_inclusive = 2;
public static final int MINCOLUMNINCLUSIVE_FIELD_NUMBER = 2; public static final int MIN_COLUMN_INCLUSIVE_FIELD_NUMBER = 2;
private boolean minColumnInclusive_; private boolean minColumnInclusive_;
public boolean hasMinColumnInclusive() { public boolean hasMinColumnInclusive() {
return ((bitField0_ & 0x00000002) == 0x00000002); return ((bitField0_ & 0x00000002) == 0x00000002);
@ -1836,8 +1836,8 @@ public final class FilterProtos {
return minColumnInclusive_; return minColumnInclusive_;
} }
// optional bytes maxColumn = 3; // optional bytes max_column = 3;
public static final int MAXCOLUMN_FIELD_NUMBER = 3; public static final int MAX_COLUMN_FIELD_NUMBER = 3;
private com.google.protobuf.ByteString maxColumn_; private com.google.protobuf.ByteString maxColumn_;
public boolean hasMaxColumn() { public boolean hasMaxColumn() {
return ((bitField0_ & 0x00000004) == 0x00000004); return ((bitField0_ & 0x00000004) == 0x00000004);
@ -1846,8 +1846,8 @@ public final class FilterProtos {
return maxColumn_; return maxColumn_;
} }
// optional bool maxColumnInclusive = 4; // optional bool max_column_inclusive = 4;
public static final int MAXCOLUMNINCLUSIVE_FIELD_NUMBER = 4; public static final int MAX_COLUMN_INCLUSIVE_FIELD_NUMBER = 4;
private boolean maxColumnInclusive_; private boolean maxColumnInclusive_;
public boolean hasMaxColumnInclusive() { public boolean hasMaxColumnInclusive() {
return ((bitField0_ & 0x00000008) == 0x00000008); return ((bitField0_ & 0x00000008) == 0x00000008);
@ -1964,19 +1964,19 @@ public final class FilterProtos {
int hash = 41; int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode(); hash = (19 * hash) + getDescriptorForType().hashCode();
if (hasMinColumn()) { if (hasMinColumn()) {
hash = (37 * hash) + MINCOLUMN_FIELD_NUMBER; hash = (37 * hash) + MIN_COLUMN_FIELD_NUMBER;
hash = (53 * hash) + getMinColumn().hashCode(); hash = (53 * hash) + getMinColumn().hashCode();
} }
if (hasMinColumnInclusive()) { if (hasMinColumnInclusive()) {
hash = (37 * hash) + MINCOLUMNINCLUSIVE_FIELD_NUMBER; hash = (37 * hash) + MIN_COLUMN_INCLUSIVE_FIELD_NUMBER;
hash = (53 * hash) + hashBoolean(getMinColumnInclusive()); hash = (53 * hash) + hashBoolean(getMinColumnInclusive());
} }
if (hasMaxColumn()) { if (hasMaxColumn()) {
hash = (37 * hash) + MAXCOLUMN_FIELD_NUMBER; hash = (37 * hash) + MAX_COLUMN_FIELD_NUMBER;
hash = (53 * hash) + getMaxColumn().hashCode(); hash = (53 * hash) + getMaxColumn().hashCode();
} }
if (hasMaxColumnInclusive()) { if (hasMaxColumnInclusive()) {
hash = (37 * hash) + MAXCOLUMNINCLUSIVE_FIELD_NUMBER; hash = (37 * hash) + MAX_COLUMN_INCLUSIVE_FIELD_NUMBER;
hash = (53 * hash) + hashBoolean(getMaxColumnInclusive()); hash = (53 * hash) + hashBoolean(getMaxColumnInclusive());
} }
hash = (29 * hash) + getUnknownFields().hashCode(); hash = (29 * hash) + getUnknownFields().hashCode();
@ -2242,7 +2242,7 @@ public final class FilterProtos {
private int bitField0_; private int bitField0_;
// optional bytes minColumn = 1; // optional bytes min_column = 1;
private com.google.protobuf.ByteString minColumn_ = com.google.protobuf.ByteString.EMPTY; private com.google.protobuf.ByteString minColumn_ = com.google.protobuf.ByteString.EMPTY;
public boolean hasMinColumn() { public boolean hasMinColumn() {
return ((bitField0_ & 0x00000001) == 0x00000001); return ((bitField0_ & 0x00000001) == 0x00000001);
@ -2266,7 +2266,7 @@ public final class FilterProtos {
return this; return this;
} }
// optional bool minColumnInclusive = 2; // optional bool min_column_inclusive = 2;
private boolean minColumnInclusive_ ; private boolean minColumnInclusive_ ;
public boolean hasMinColumnInclusive() { public boolean hasMinColumnInclusive() {
return ((bitField0_ & 0x00000002) == 0x00000002); return ((bitField0_ & 0x00000002) == 0x00000002);
@ -2287,7 +2287,7 @@ public final class FilterProtos {
return this; return this;
} }
// optional bytes maxColumn = 3; // optional bytes max_column = 3;
private com.google.protobuf.ByteString maxColumn_ = com.google.protobuf.ByteString.EMPTY; private com.google.protobuf.ByteString maxColumn_ = com.google.protobuf.ByteString.EMPTY;
public boolean hasMaxColumn() { public boolean hasMaxColumn() {
return ((bitField0_ & 0x00000004) == 0x00000004); return ((bitField0_ & 0x00000004) == 0x00000004);
@ -2311,7 +2311,7 @@ public final class FilterProtos {
return this; return this;
} }
// optional bool maxColumnInclusive = 4; // optional bool max_column_inclusive = 4;
private boolean maxColumnInclusive_ ; private boolean maxColumnInclusive_ ;
public boolean hasMaxColumnInclusive() { public boolean hasMaxColumnInclusive() {
return ((bitField0_ & 0x00000008) == 0x00000008); return ((bitField0_ & 0x00000008) == 0x00000008);
@ -2346,7 +2346,7 @@ public final class FilterProtos {
public interface CompareFilterOrBuilder public interface CompareFilterOrBuilder
extends com.google.protobuf.MessageOrBuilder { extends com.google.protobuf.MessageOrBuilder {
// required .CompareType compareOp = 1; // required .CompareType compare_op = 1;
boolean hasCompareOp(); boolean hasCompareOp();
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.CompareType getCompareOp(); org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.CompareType getCompareOp();
@ -2384,8 +2384,8 @@ public final class FilterProtos {
} }
private int bitField0_; private int bitField0_;
// required .CompareType compareOp = 1; // required .CompareType compare_op = 1;
public static final int COMPAREOP_FIELD_NUMBER = 1; public static final int COMPARE_OP_FIELD_NUMBER = 1;
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.CompareType compareOp_; private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.CompareType compareOp_;
public boolean hasCompareOp() { public boolean hasCompareOp() {
return ((bitField0_ & 0x00000001) == 0x00000001); return ((bitField0_ & 0x00000001) == 0x00000001);
@ -2499,7 +2499,7 @@ public final class FilterProtos {
int hash = 41; int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode(); hash = (19 * hash) + getDescriptorForType().hashCode();
if (hasCompareOp()) { if (hasCompareOp()) {
hash = (37 * hash) + COMPAREOP_FIELD_NUMBER; hash = (37 * hash) + COMPARE_OP_FIELD_NUMBER;
hash = (53 * hash) + hashEnum(getCompareOp()); hash = (53 * hash) + hashEnum(getCompareOp());
} }
if (hasComparator()) { if (hasComparator()) {
@ -2770,7 +2770,7 @@ public final class FilterProtos {
private int bitField0_; private int bitField0_;
// required .CompareType compareOp = 1; // required .CompareType compare_op = 1;
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.CompareType compareOp_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.CompareType.LESS; private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.CompareType compareOp_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.CompareType.LESS;
public boolean hasCompareOp() { public boolean hasCompareOp() {
return ((bitField0_ & 0x00000001) == 0x00000001); return ((bitField0_ & 0x00000001) == 0x00000001);
@ -2898,20 +2898,20 @@ public final class FilterProtos {
public interface DependentColumnFilterOrBuilder public interface DependentColumnFilterOrBuilder
extends com.google.protobuf.MessageOrBuilder { extends com.google.protobuf.MessageOrBuilder {
// required .CompareFilter compareFilter = 1; // required .CompareFilter compare_filter = 1;
boolean hasCompareFilter(); boolean hasCompareFilter();
org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter getCompareFilter(); org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter getCompareFilter();
org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilterOrBuilder getCompareFilterOrBuilder(); org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilterOrBuilder getCompareFilterOrBuilder();
// optional bytes columnFamily = 2; // optional bytes column_family = 2;
boolean hasColumnFamily(); boolean hasColumnFamily();
com.google.protobuf.ByteString getColumnFamily(); com.google.protobuf.ByteString getColumnFamily();
// optional bytes columnQualifier = 3; // optional bytes column_qualifier = 3;
boolean hasColumnQualifier(); boolean hasColumnQualifier();
com.google.protobuf.ByteString getColumnQualifier(); com.google.protobuf.ByteString getColumnQualifier();
// optional bool dropDependentColumn = 4; // optional bool drop_dependent_column = 4;
boolean hasDropDependentColumn(); boolean hasDropDependentColumn();
boolean getDropDependentColumn(); boolean getDropDependentColumn();
} }
@ -2944,8 +2944,8 @@ public final class FilterProtos {
} }
private int bitField0_; private int bitField0_;
// required .CompareFilter compareFilter = 1; // required .CompareFilter compare_filter = 1;
public static final int COMPAREFILTER_FIELD_NUMBER = 1; public static final int COMPARE_FILTER_FIELD_NUMBER = 1;
private org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter compareFilter_; private org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter compareFilter_;
public boolean hasCompareFilter() { public boolean hasCompareFilter() {
return ((bitField0_ & 0x00000001) == 0x00000001); return ((bitField0_ & 0x00000001) == 0x00000001);
@ -2957,8 +2957,8 @@ public final class FilterProtos {
return compareFilter_; return compareFilter_;
} }
// optional bytes columnFamily = 2; // optional bytes column_family = 2;
public static final int COLUMNFAMILY_FIELD_NUMBER = 2; public static final int COLUMN_FAMILY_FIELD_NUMBER = 2;
private com.google.protobuf.ByteString columnFamily_; private com.google.protobuf.ByteString columnFamily_;
public boolean hasColumnFamily() { public boolean hasColumnFamily() {
return ((bitField0_ & 0x00000002) == 0x00000002); return ((bitField0_ & 0x00000002) == 0x00000002);
@ -2967,8 +2967,8 @@ public final class FilterProtos {
return columnFamily_; return columnFamily_;
} }
// optional bytes columnQualifier = 3; // optional bytes column_qualifier = 3;
public static final int COLUMNQUALIFIER_FIELD_NUMBER = 3; public static final int COLUMN_QUALIFIER_FIELD_NUMBER = 3;
private com.google.protobuf.ByteString columnQualifier_; private com.google.protobuf.ByteString columnQualifier_;
public boolean hasColumnQualifier() { public boolean hasColumnQualifier() {
return ((bitField0_ & 0x00000004) == 0x00000004); return ((bitField0_ & 0x00000004) == 0x00000004);
@ -2977,8 +2977,8 @@ public final class FilterProtos {
return columnQualifier_; return columnQualifier_;
} }
// optional bool dropDependentColumn = 4; // optional bool drop_dependent_column = 4;
public static final int DROPDEPENDENTCOLUMN_FIELD_NUMBER = 4; public static final int DROP_DEPENDENT_COLUMN_FIELD_NUMBER = 4;
private boolean dropDependentColumn_; private boolean dropDependentColumn_;
public boolean hasDropDependentColumn() { public boolean hasDropDependentColumn() {
return ((bitField0_ & 0x00000008) == 0x00000008); return ((bitField0_ & 0x00000008) == 0x00000008);
@ -3103,19 +3103,19 @@ public final class FilterProtos {
int hash = 41; int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode(); hash = (19 * hash) + getDescriptorForType().hashCode();
if (hasCompareFilter()) { if (hasCompareFilter()) {
hash = (37 * hash) + COMPAREFILTER_FIELD_NUMBER; hash = (37 * hash) + COMPARE_FILTER_FIELD_NUMBER;
hash = (53 * hash) + getCompareFilter().hashCode(); hash = (53 * hash) + getCompareFilter().hashCode();
} }
if (hasColumnFamily()) { if (hasColumnFamily()) {
hash = (37 * hash) + COLUMNFAMILY_FIELD_NUMBER; hash = (37 * hash) + COLUMN_FAMILY_FIELD_NUMBER;
hash = (53 * hash) + getColumnFamily().hashCode(); hash = (53 * hash) + getColumnFamily().hashCode();
} }
if (hasColumnQualifier()) { if (hasColumnQualifier()) {
hash = (37 * hash) + COLUMNQUALIFIER_FIELD_NUMBER; hash = (37 * hash) + COLUMN_QUALIFIER_FIELD_NUMBER;
hash = (53 * hash) + getColumnQualifier().hashCode(); hash = (53 * hash) + getColumnQualifier().hashCode();
} }
if (hasDropDependentColumn()) { if (hasDropDependentColumn()) {
hash = (37 * hash) + DROPDEPENDENTCOLUMN_FIELD_NUMBER; hash = (37 * hash) + DROP_DEPENDENT_COLUMN_FIELD_NUMBER;
hash = (53 * hash) + hashBoolean(getDropDependentColumn()); hash = (53 * hash) + hashBoolean(getDropDependentColumn());
} }
hash = (29 * hash) + getUnknownFields().hashCode(); hash = (29 * hash) + getUnknownFields().hashCode();
@ -3402,7 +3402,7 @@ public final class FilterProtos {
private int bitField0_; private int bitField0_;
// required .CompareFilter compareFilter = 1; // required .CompareFilter compare_filter = 1;
private org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter compareFilter_ = org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter.getDefaultInstance(); private org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter compareFilter_ = org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter.getDefaultInstance();
private com.google.protobuf.SingleFieldBuilder< private com.google.protobuf.SingleFieldBuilder<
org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter, org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter.Builder, org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilterOrBuilder> compareFilterBuilder_; org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter, org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter.Builder, org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilterOrBuilder> compareFilterBuilder_;
@ -3492,7 +3492,7 @@ public final class FilterProtos {
return compareFilterBuilder_; return compareFilterBuilder_;
} }
// optional bytes columnFamily = 2; // optional bytes column_family = 2;
private com.google.protobuf.ByteString columnFamily_ = com.google.protobuf.ByteString.EMPTY; private com.google.protobuf.ByteString columnFamily_ = com.google.protobuf.ByteString.EMPTY;
public boolean hasColumnFamily() { public boolean hasColumnFamily() {
return ((bitField0_ & 0x00000002) == 0x00000002); return ((bitField0_ & 0x00000002) == 0x00000002);
@ -3516,7 +3516,7 @@ public final class FilterProtos {
return this; return this;
} }
// optional bytes columnQualifier = 3; // optional bytes column_qualifier = 3;
private com.google.protobuf.ByteString columnQualifier_ = com.google.protobuf.ByteString.EMPTY; private com.google.protobuf.ByteString columnQualifier_ = com.google.protobuf.ByteString.EMPTY;
public boolean hasColumnQualifier() { public boolean hasColumnQualifier() {
return ((bitField0_ & 0x00000004) == 0x00000004); return ((bitField0_ & 0x00000004) == 0x00000004);
@ -3540,7 +3540,7 @@ public final class FilterProtos {
return this; return this;
} }
// optional bool dropDependentColumn = 4; // optional bool drop_dependent_column = 4;
private boolean dropDependentColumn_ ; private boolean dropDependentColumn_ ;
public boolean hasDropDependentColumn() { public boolean hasDropDependentColumn() {
return ((bitField0_ & 0x00000008) == 0x00000008); return ((bitField0_ & 0x00000008) == 0x00000008);
@ -3575,7 +3575,7 @@ public final class FilterProtos {
public interface FamilyFilterOrBuilder public interface FamilyFilterOrBuilder
extends com.google.protobuf.MessageOrBuilder { extends com.google.protobuf.MessageOrBuilder {
// required .CompareFilter compareFilter = 1; // required .CompareFilter compare_filter = 1;
boolean hasCompareFilter(); boolean hasCompareFilter();
org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter getCompareFilter(); org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter getCompareFilter();
org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilterOrBuilder getCompareFilterOrBuilder(); org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilterOrBuilder getCompareFilterOrBuilder();
@ -3609,8 +3609,8 @@ public final class FilterProtos {
} }
private int bitField0_; private int bitField0_;
// required .CompareFilter compareFilter = 1; // required .CompareFilter compare_filter = 1;
public static final int COMPAREFILTER_FIELD_NUMBER = 1; public static final int COMPARE_FILTER_FIELD_NUMBER = 1;
private org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter compareFilter_; private org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter compareFilter_;
public boolean hasCompareFilter() { public boolean hasCompareFilter() {
return ((bitField0_ & 0x00000001) == 0x00000001); return ((bitField0_ & 0x00000001) == 0x00000001);
@ -3699,7 +3699,7 @@ public final class FilterProtos {
int hash = 41; int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode(); hash = (19 * hash) + getDescriptorForType().hashCode();
if (hasCompareFilter()) { if (hasCompareFilter()) {
hash = (37 * hash) + COMPAREFILTER_FIELD_NUMBER; hash = (37 * hash) + COMPARE_FILTER_FIELD_NUMBER;
hash = (53 * hash) + getCompareFilter().hashCode(); hash = (53 * hash) + getCompareFilter().hashCode();
} }
hash = (29 * hash) + getUnknownFields().hashCode(); hash = (29 * hash) + getUnknownFields().hashCode();
@ -3944,7 +3944,7 @@ public final class FilterProtos {
private int bitField0_; private int bitField0_;
// required .CompareFilter compareFilter = 1; // required .CompareFilter compare_filter = 1;
private org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter compareFilter_ = org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter.getDefaultInstance(); private org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter compareFilter_ = org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter.getDefaultInstance();
private com.google.protobuf.SingleFieldBuilder< private com.google.protobuf.SingleFieldBuilder<
org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter, org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter.Builder, org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilterOrBuilder> compareFilterBuilder_; org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter, org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter.Builder, org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilterOrBuilder> compareFilterBuilder_;
@ -5982,7 +5982,7 @@ public final class FilterProtos {
public interface FuzzyRowFilterOrBuilder public interface FuzzyRowFilterOrBuilder
extends com.google.protobuf.MessageOrBuilder { extends com.google.protobuf.MessageOrBuilder {
// repeated .BytesBytesPair fuzzyKeysData = 1; // repeated .BytesBytesPair fuzzy_keys_data = 1;
java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair> java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair>
getFuzzyKeysDataList(); getFuzzyKeysDataList();
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair getFuzzyKeysData(int index); org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair getFuzzyKeysData(int index);
@ -6020,8 +6020,8 @@ public final class FilterProtos {
return org.apache.hadoop.hbase.protobuf.generated.FilterProtos.internal_static_FuzzyRowFilter_fieldAccessorTable; return org.apache.hadoop.hbase.protobuf.generated.FilterProtos.internal_static_FuzzyRowFilter_fieldAccessorTable;
} }
// repeated .BytesBytesPair fuzzyKeysData = 1; // repeated .BytesBytesPair fuzzy_keys_data = 1;
public static final int FUZZYKEYSDATA_FIELD_NUMBER = 1; public static final int FUZZY_KEYS_DATA_FIELD_NUMBER = 1;
private java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair> fuzzyKeysData_; private java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair> fuzzyKeysData_;
public java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair> getFuzzyKeysDataList() { public java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair> getFuzzyKeysDataList() {
return fuzzyKeysData_; return fuzzyKeysData_;
@ -6113,7 +6113,7 @@ public final class FilterProtos {
int hash = 41; int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode(); hash = (19 * hash) + getDescriptorForType().hashCode();
if (getFuzzyKeysDataCount() > 0) { if (getFuzzyKeysDataCount() > 0) {
hash = (37 * hash) + FUZZYKEYSDATA_FIELD_NUMBER; hash = (37 * hash) + FUZZY_KEYS_DATA_FIELD_NUMBER;
hash = (53 * hash) + getFuzzyKeysDataList().hashCode(); hash = (53 * hash) + getFuzzyKeysDataList().hashCode();
} }
hash = (29 * hash) + getUnknownFields().hashCode(); hash = (29 * hash) + getUnknownFields().hashCode();
@ -6375,7 +6375,7 @@ public final class FilterProtos {
private int bitField0_; private int bitField0_;
// repeated .BytesBytesPair fuzzyKeysData = 1; // repeated .BytesBytesPair fuzzy_keys_data = 1;
private java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair> fuzzyKeysData_ = private java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair> fuzzyKeysData_ =
java.util.Collections.emptyList(); java.util.Collections.emptyList();
private void ensureFuzzyKeysDataIsMutable() { private void ensureFuzzyKeysDataIsMutable() {
@ -6575,7 +6575,7 @@ public final class FilterProtos {
public interface InclusiveStopFilterOrBuilder public interface InclusiveStopFilterOrBuilder
extends com.google.protobuf.MessageOrBuilder { extends com.google.protobuf.MessageOrBuilder {
// optional bytes stopRowKey = 1; // optional bytes stop_row_key = 1;
boolean hasStopRowKey(); boolean hasStopRowKey();
com.google.protobuf.ByteString getStopRowKey(); com.google.protobuf.ByteString getStopRowKey();
} }
@ -6608,8 +6608,8 @@ public final class FilterProtos {
} }
private int bitField0_; private int bitField0_;
// optional bytes stopRowKey = 1; // optional bytes stop_row_key = 1;
public static final int STOPROWKEY_FIELD_NUMBER = 1; public static final int STOP_ROW_KEY_FIELD_NUMBER = 1;
private com.google.protobuf.ByteString stopRowKey_; private com.google.protobuf.ByteString stopRowKey_;
public boolean hasStopRowKey() { public boolean hasStopRowKey() {
return ((bitField0_ & 0x00000001) == 0x00000001); return ((bitField0_ & 0x00000001) == 0x00000001);
@ -6687,7 +6687,7 @@ public final class FilterProtos {
int hash = 41; int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode(); hash = (19 * hash) + getDescriptorForType().hashCode();
if (hasStopRowKey()) { if (hasStopRowKey()) {
hash = (37 * hash) + STOPROWKEY_FIELD_NUMBER; hash = (37 * hash) + STOP_ROW_KEY_FIELD_NUMBER;
hash = (53 * hash) + getStopRowKey().hashCode(); hash = (53 * hash) + getStopRowKey().hashCode();
} }
hash = (29 * hash) + getUnknownFields().hashCode(); hash = (29 * hash) + getUnknownFields().hashCode();
@ -6911,7 +6911,7 @@ public final class FilterProtos {
private int bitField0_; private int bitField0_;
// optional bytes stopRowKey = 1; // optional bytes stop_row_key = 1;
private com.google.protobuf.ByteString stopRowKey_ = com.google.protobuf.ByteString.EMPTY; private com.google.protobuf.ByteString stopRowKey_ = com.google.protobuf.ByteString.EMPTY;
public boolean hasStopRowKey() { public boolean hasStopRowKey() {
return ((bitField0_ & 0x00000001) == 0x00000001); return ((bitField0_ & 0x00000001) == 0x00000001);
@ -6949,7 +6949,7 @@ public final class FilterProtos {
public interface KeyOnlyFilterOrBuilder public interface KeyOnlyFilterOrBuilder
extends com.google.protobuf.MessageOrBuilder { extends com.google.protobuf.MessageOrBuilder {
// required bool lenAsVal = 1; // required bool len_as_val = 1;
boolean hasLenAsVal(); boolean hasLenAsVal();
boolean getLenAsVal(); boolean getLenAsVal();
} }
@ -6982,8 +6982,8 @@ public final class FilterProtos {
} }
private int bitField0_; private int bitField0_;
// required bool lenAsVal = 1; // required bool len_as_val = 1;
public static final int LENASVAL_FIELD_NUMBER = 1; public static final int LEN_AS_VAL_FIELD_NUMBER = 1;
private boolean lenAsVal_; private boolean lenAsVal_;
public boolean hasLenAsVal() { public boolean hasLenAsVal() {
return ((bitField0_ & 0x00000001) == 0x00000001); return ((bitField0_ & 0x00000001) == 0x00000001);
@ -7065,7 +7065,7 @@ public final class FilterProtos {
int hash = 41; int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode(); hash = (19 * hash) + getDescriptorForType().hashCode();
if (hasLenAsVal()) { if (hasLenAsVal()) {
hash = (37 * hash) + LENASVAL_FIELD_NUMBER; hash = (37 * hash) + LEN_AS_VAL_FIELD_NUMBER;
hash = (53 * hash) + hashBoolean(getLenAsVal()); hash = (53 * hash) + hashBoolean(getLenAsVal());
} }
hash = (29 * hash) + getUnknownFields().hashCode(); hash = (29 * hash) + getUnknownFields().hashCode();
@ -7293,7 +7293,7 @@ public final class FilterProtos {
private int bitField0_; private int bitField0_;
// required bool lenAsVal = 1; // required bool len_as_val = 1;
private boolean lenAsVal_ ; private boolean lenAsVal_ ;
public boolean hasLenAsVal() { public boolean hasLenAsVal() {
return ((bitField0_ & 0x00000001) == 0x00000001); return ((bitField0_ & 0x00000001) == 0x00000001);
@ -7328,7 +7328,7 @@ public final class FilterProtos {
public interface MultipleColumnPrefixFilterOrBuilder public interface MultipleColumnPrefixFilterOrBuilder
extends com.google.protobuf.MessageOrBuilder { extends com.google.protobuf.MessageOrBuilder {
// repeated bytes sortedPrefixes = 1; // repeated bytes sorted_prefixes = 1;
java.util.List<com.google.protobuf.ByteString> getSortedPrefixesList(); java.util.List<com.google.protobuf.ByteString> getSortedPrefixesList();
int getSortedPrefixesCount(); int getSortedPrefixesCount();
com.google.protobuf.ByteString getSortedPrefixes(int index); com.google.protobuf.ByteString getSortedPrefixes(int index);
@ -7361,8 +7361,8 @@ public final class FilterProtos {
return org.apache.hadoop.hbase.protobuf.generated.FilterProtos.internal_static_MultipleColumnPrefixFilter_fieldAccessorTable; return org.apache.hadoop.hbase.protobuf.generated.FilterProtos.internal_static_MultipleColumnPrefixFilter_fieldAccessorTable;
} }
// repeated bytes sortedPrefixes = 1; // repeated bytes sorted_prefixes = 1;
public static final int SORTEDPREFIXES_FIELD_NUMBER = 1; public static final int SORTED_PREFIXES_FIELD_NUMBER = 1;
private java.util.List<com.google.protobuf.ByteString> sortedPrefixes_; private java.util.List<com.google.protobuf.ByteString> sortedPrefixes_;
public java.util.List<com.google.protobuf.ByteString> public java.util.List<com.google.protobuf.ByteString>
getSortedPrefixesList() { getSortedPrefixesList() {
@ -7446,7 +7446,7 @@ public final class FilterProtos {
int hash = 41; int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode(); hash = (19 * hash) + getDescriptorForType().hashCode();
if (getSortedPrefixesCount() > 0) { if (getSortedPrefixesCount() > 0) {
hash = (37 * hash) + SORTEDPREFIXES_FIELD_NUMBER; hash = (37 * hash) + SORTED_PREFIXES_FIELD_NUMBER;
hash = (53 * hash) + getSortedPrefixesList().hashCode(); hash = (53 * hash) + getSortedPrefixesList().hashCode();
} }
hash = (29 * hash) + getUnknownFields().hashCode(); hash = (29 * hash) + getUnknownFields().hashCode();
@ -7676,7 +7676,7 @@ public final class FilterProtos {
private int bitField0_; private int bitField0_;
// repeated bytes sortedPrefixes = 1; // repeated bytes sorted_prefixes = 1;
private java.util.List<com.google.protobuf.ByteString> sortedPrefixes_ = java.util.Collections.emptyList();; private java.util.List<com.google.protobuf.ByteString> sortedPrefixes_ = java.util.Collections.emptyList();;
private void ensureSortedPrefixesIsMutable() { private void ensureSortedPrefixesIsMutable() {
if (!((bitField0_ & 0x00000001) == 0x00000001)) { if (!((bitField0_ & 0x00000001) == 0x00000001)) {
@ -7741,7 +7741,7 @@ public final class FilterProtos {
public interface PageFilterOrBuilder public interface PageFilterOrBuilder
extends com.google.protobuf.MessageOrBuilder { extends com.google.protobuf.MessageOrBuilder {
// required int64 pageSize = 1; // required int64 page_size = 1;
boolean hasPageSize(); boolean hasPageSize();
long getPageSize(); long getPageSize();
} }
@ -7774,8 +7774,8 @@ public final class FilterProtos {
} }
private int bitField0_; private int bitField0_;
// required int64 pageSize = 1; // required int64 page_size = 1;
public static final int PAGESIZE_FIELD_NUMBER = 1; public static final int PAGE_SIZE_FIELD_NUMBER = 1;
private long pageSize_; private long pageSize_;
public boolean hasPageSize() { public boolean hasPageSize() {
return ((bitField0_ & 0x00000001) == 0x00000001); return ((bitField0_ & 0x00000001) == 0x00000001);
@ -7857,7 +7857,7 @@ public final class FilterProtos {
int hash = 41; int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode(); hash = (19 * hash) + getDescriptorForType().hashCode();
if (hasPageSize()) { if (hasPageSize()) {
hash = (37 * hash) + PAGESIZE_FIELD_NUMBER; hash = (37 * hash) + PAGE_SIZE_FIELD_NUMBER;
hash = (53 * hash) + hashLong(getPageSize()); hash = (53 * hash) + hashLong(getPageSize());
} }
hash = (29 * hash) + getUnknownFields().hashCode(); hash = (29 * hash) + getUnknownFields().hashCode();
@ -8085,7 +8085,7 @@ public final class FilterProtos {
private int bitField0_; private int bitField0_;
// required int64 pageSize = 1; // required int64 page_size = 1;
private long pageSize_ ; private long pageSize_ ;
public boolean hasPageSize() { public boolean hasPageSize() {
return ((bitField0_ & 0x00000001) == 0x00000001); return ((bitField0_ & 0x00000001) == 0x00000001);
@ -8494,7 +8494,7 @@ public final class FilterProtos {
public interface QualifierFilterOrBuilder public interface QualifierFilterOrBuilder
extends com.google.protobuf.MessageOrBuilder { extends com.google.protobuf.MessageOrBuilder {
// required .CompareFilter compareFilter = 1; // required .CompareFilter compare_filter = 1;
boolean hasCompareFilter(); boolean hasCompareFilter();
org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter getCompareFilter(); org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter getCompareFilter();
org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilterOrBuilder getCompareFilterOrBuilder(); org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilterOrBuilder getCompareFilterOrBuilder();
@ -8528,8 +8528,8 @@ public final class FilterProtos {
} }
private int bitField0_; private int bitField0_;
// required .CompareFilter compareFilter = 1; // required .CompareFilter compare_filter = 1;
public static final int COMPAREFILTER_FIELD_NUMBER = 1; public static final int COMPARE_FILTER_FIELD_NUMBER = 1;
private org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter compareFilter_; private org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter compareFilter_;
public boolean hasCompareFilter() { public boolean hasCompareFilter() {
return ((bitField0_ & 0x00000001) == 0x00000001); return ((bitField0_ & 0x00000001) == 0x00000001);
@ -8618,7 +8618,7 @@ public final class FilterProtos {
int hash = 41; int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode(); hash = (19 * hash) + getDescriptorForType().hashCode();
if (hasCompareFilter()) { if (hasCompareFilter()) {
hash = (37 * hash) + COMPAREFILTER_FIELD_NUMBER; hash = (37 * hash) + COMPARE_FILTER_FIELD_NUMBER;
hash = (53 * hash) + getCompareFilter().hashCode(); hash = (53 * hash) + getCompareFilter().hashCode();
} }
hash = (29 * hash) + getUnknownFields().hashCode(); hash = (29 * hash) + getUnknownFields().hashCode();
@ -8863,7 +8863,7 @@ public final class FilterProtos {
private int bitField0_; private int bitField0_;
// required .CompareFilter compareFilter = 1; // required .CompareFilter compare_filter = 1;
private org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter compareFilter_ = org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter.getDefaultInstance(); private org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter compareFilter_ = org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter.getDefaultInstance();
private com.google.protobuf.SingleFieldBuilder< private com.google.protobuf.SingleFieldBuilder<
org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter, org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter.Builder, org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilterOrBuilder> compareFilterBuilder_; org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter, org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter.Builder, org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilterOrBuilder> compareFilterBuilder_;
@ -9346,7 +9346,7 @@ public final class FilterProtos {
public interface RowFilterOrBuilder public interface RowFilterOrBuilder
extends com.google.protobuf.MessageOrBuilder { extends com.google.protobuf.MessageOrBuilder {
// required .CompareFilter compareFilter = 1; // required .CompareFilter compare_filter = 1;
boolean hasCompareFilter(); boolean hasCompareFilter();
org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter getCompareFilter(); org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter getCompareFilter();
org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilterOrBuilder getCompareFilterOrBuilder(); org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilterOrBuilder getCompareFilterOrBuilder();
@ -9380,8 +9380,8 @@ public final class FilterProtos {
} }
private int bitField0_; private int bitField0_;
// required .CompareFilter compareFilter = 1; // required .CompareFilter compare_filter = 1;
public static final int COMPAREFILTER_FIELD_NUMBER = 1; public static final int COMPARE_FILTER_FIELD_NUMBER = 1;
private org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter compareFilter_; private org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter compareFilter_;
public boolean hasCompareFilter() { public boolean hasCompareFilter() {
return ((bitField0_ & 0x00000001) == 0x00000001); return ((bitField0_ & 0x00000001) == 0x00000001);
@ -9470,7 +9470,7 @@ public final class FilterProtos {
int hash = 41; int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode(); hash = (19 * hash) + getDescriptorForType().hashCode();
if (hasCompareFilter()) { if (hasCompareFilter()) {
hash = (37 * hash) + COMPAREFILTER_FIELD_NUMBER; hash = (37 * hash) + COMPARE_FILTER_FIELD_NUMBER;
hash = (53 * hash) + getCompareFilter().hashCode(); hash = (53 * hash) + getCompareFilter().hashCode();
} }
hash = (29 * hash) + getUnknownFields().hashCode(); hash = (29 * hash) + getUnknownFields().hashCode();
@ -9715,7 +9715,7 @@ public final class FilterProtos {
private int bitField0_; private int bitField0_;
// required .CompareFilter compareFilter = 1; // required .CompareFilter compare_filter = 1;
private org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter compareFilter_ = org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter.getDefaultInstance(); private org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter compareFilter_ = org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter.getDefaultInstance();
private com.google.protobuf.SingleFieldBuilder< private com.google.protobuf.SingleFieldBuilder<
org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter, org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter.Builder, org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilterOrBuilder> compareFilterBuilder_; org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter, org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter.Builder, org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilterOrBuilder> compareFilterBuilder_;
@ -9819,7 +9819,7 @@ public final class FilterProtos {
public interface SingleColumnValueExcludeFilterOrBuilder public interface SingleColumnValueExcludeFilterOrBuilder
extends com.google.protobuf.MessageOrBuilder { extends com.google.protobuf.MessageOrBuilder {
// required .SingleColumnValueFilter singleColumnValueFilter = 1; // required .SingleColumnValueFilter single_column_value_filter = 1;
boolean hasSingleColumnValueFilter(); boolean hasSingleColumnValueFilter();
org.apache.hadoop.hbase.protobuf.generated.FilterProtos.SingleColumnValueFilter getSingleColumnValueFilter(); org.apache.hadoop.hbase.protobuf.generated.FilterProtos.SingleColumnValueFilter getSingleColumnValueFilter();
org.apache.hadoop.hbase.protobuf.generated.FilterProtos.SingleColumnValueFilterOrBuilder getSingleColumnValueFilterOrBuilder(); org.apache.hadoop.hbase.protobuf.generated.FilterProtos.SingleColumnValueFilterOrBuilder getSingleColumnValueFilterOrBuilder();
@ -9853,8 +9853,8 @@ public final class FilterProtos {
} }
private int bitField0_; private int bitField0_;
// required .SingleColumnValueFilter singleColumnValueFilter = 1; // required .SingleColumnValueFilter single_column_value_filter = 1;
public static final int SINGLECOLUMNVALUEFILTER_FIELD_NUMBER = 1; public static final int SINGLE_COLUMN_VALUE_FILTER_FIELD_NUMBER = 1;
private org.apache.hadoop.hbase.protobuf.generated.FilterProtos.SingleColumnValueFilter singleColumnValueFilter_; private org.apache.hadoop.hbase.protobuf.generated.FilterProtos.SingleColumnValueFilter singleColumnValueFilter_;
public boolean hasSingleColumnValueFilter() { public boolean hasSingleColumnValueFilter() {
return ((bitField0_ & 0x00000001) == 0x00000001); return ((bitField0_ & 0x00000001) == 0x00000001);
@ -9943,7 +9943,7 @@ public final class FilterProtos {
int hash = 41; int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode(); hash = (19 * hash) + getDescriptorForType().hashCode();
if (hasSingleColumnValueFilter()) { if (hasSingleColumnValueFilter()) {
hash = (37 * hash) + SINGLECOLUMNVALUEFILTER_FIELD_NUMBER; hash = (37 * hash) + SINGLE_COLUMN_VALUE_FILTER_FIELD_NUMBER;
hash = (53 * hash) + getSingleColumnValueFilter().hashCode(); hash = (53 * hash) + getSingleColumnValueFilter().hashCode();
} }
hash = (29 * hash) + getUnknownFields().hashCode(); hash = (29 * hash) + getUnknownFields().hashCode();
@ -10188,7 +10188,7 @@ public final class FilterProtos {
private int bitField0_; private int bitField0_;
// required .SingleColumnValueFilter singleColumnValueFilter = 1; // required .SingleColumnValueFilter single_column_value_filter = 1;
private org.apache.hadoop.hbase.protobuf.generated.FilterProtos.SingleColumnValueFilter singleColumnValueFilter_ = org.apache.hadoop.hbase.protobuf.generated.FilterProtos.SingleColumnValueFilter.getDefaultInstance(); private org.apache.hadoop.hbase.protobuf.generated.FilterProtos.SingleColumnValueFilter singleColumnValueFilter_ = org.apache.hadoop.hbase.protobuf.generated.FilterProtos.SingleColumnValueFilter.getDefaultInstance();
private com.google.protobuf.SingleFieldBuilder< private com.google.protobuf.SingleFieldBuilder<
org.apache.hadoop.hbase.protobuf.generated.FilterProtos.SingleColumnValueFilter, org.apache.hadoop.hbase.protobuf.generated.FilterProtos.SingleColumnValueFilter.Builder, org.apache.hadoop.hbase.protobuf.generated.FilterProtos.SingleColumnValueFilterOrBuilder> singleColumnValueFilterBuilder_; org.apache.hadoop.hbase.protobuf.generated.FilterProtos.SingleColumnValueFilter, org.apache.hadoop.hbase.protobuf.generated.FilterProtos.SingleColumnValueFilter.Builder, org.apache.hadoop.hbase.protobuf.generated.FilterProtos.SingleColumnValueFilterOrBuilder> singleColumnValueFilterBuilder_;
@ -10292,15 +10292,15 @@ public final class FilterProtos {
public interface SingleColumnValueFilterOrBuilder public interface SingleColumnValueFilterOrBuilder
extends com.google.protobuf.MessageOrBuilder { extends com.google.protobuf.MessageOrBuilder {
// optional bytes columnFamily = 1; // optional bytes column_family = 1;
boolean hasColumnFamily(); boolean hasColumnFamily();
com.google.protobuf.ByteString getColumnFamily(); com.google.protobuf.ByteString getColumnFamily();
// optional bytes columnQualifier = 2; // optional bytes column_qualifier = 2;
boolean hasColumnQualifier(); boolean hasColumnQualifier();
com.google.protobuf.ByteString getColumnQualifier(); com.google.protobuf.ByteString getColumnQualifier();
// required .CompareType compareOp = 3; // required .CompareType compare_op = 3;
boolean hasCompareOp(); boolean hasCompareOp();
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.CompareType getCompareOp(); org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.CompareType getCompareOp();
@ -10309,11 +10309,11 @@ public final class FilterProtos {
org.apache.hadoop.hbase.protobuf.generated.ComparatorProtos.Comparator getComparator(); org.apache.hadoop.hbase.protobuf.generated.ComparatorProtos.Comparator getComparator();
org.apache.hadoop.hbase.protobuf.generated.ComparatorProtos.ComparatorOrBuilder getComparatorOrBuilder(); org.apache.hadoop.hbase.protobuf.generated.ComparatorProtos.ComparatorOrBuilder getComparatorOrBuilder();
// optional bool filterIfMissing = 5; // optional bool filter_if_missing = 5;
boolean hasFilterIfMissing(); boolean hasFilterIfMissing();
boolean getFilterIfMissing(); boolean getFilterIfMissing();
// optional bool latestVersionOnly = 6; // optional bool latest_version_only = 6;
boolean hasLatestVersionOnly(); boolean hasLatestVersionOnly();
boolean getLatestVersionOnly(); boolean getLatestVersionOnly();
} }
@ -10346,8 +10346,8 @@ public final class FilterProtos {
} }
private int bitField0_; private int bitField0_;
// optional bytes columnFamily = 1; // optional bytes column_family = 1;
public static final int COLUMNFAMILY_FIELD_NUMBER = 1; public static final int COLUMN_FAMILY_FIELD_NUMBER = 1;
private com.google.protobuf.ByteString columnFamily_; private com.google.protobuf.ByteString columnFamily_;
public boolean hasColumnFamily() { public boolean hasColumnFamily() {
return ((bitField0_ & 0x00000001) == 0x00000001); return ((bitField0_ & 0x00000001) == 0x00000001);
@ -10356,8 +10356,8 @@ public final class FilterProtos {
return columnFamily_; return columnFamily_;
} }
// optional bytes columnQualifier = 2; // optional bytes column_qualifier = 2;
public static final int COLUMNQUALIFIER_FIELD_NUMBER = 2; public static final int COLUMN_QUALIFIER_FIELD_NUMBER = 2;
private com.google.protobuf.ByteString columnQualifier_; private com.google.protobuf.ByteString columnQualifier_;
public boolean hasColumnQualifier() { public boolean hasColumnQualifier() {
return ((bitField0_ & 0x00000002) == 0x00000002); return ((bitField0_ & 0x00000002) == 0x00000002);
@ -10366,8 +10366,8 @@ public final class FilterProtos {
return columnQualifier_; return columnQualifier_;
} }
// required .CompareType compareOp = 3; // required .CompareType compare_op = 3;
public static final int COMPAREOP_FIELD_NUMBER = 3; public static final int COMPARE_OP_FIELD_NUMBER = 3;
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.CompareType compareOp_; private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.CompareType compareOp_;
public boolean hasCompareOp() { public boolean hasCompareOp() {
return ((bitField0_ & 0x00000004) == 0x00000004); return ((bitField0_ & 0x00000004) == 0x00000004);
@ -10389,8 +10389,8 @@ public final class FilterProtos {
return comparator_; return comparator_;
} }
// optional bool filterIfMissing = 5; // optional bool filter_if_missing = 5;
public static final int FILTERIFMISSING_FIELD_NUMBER = 5; public static final int FILTER_IF_MISSING_FIELD_NUMBER = 5;
private boolean filterIfMissing_; private boolean filterIfMissing_;
public boolean hasFilterIfMissing() { public boolean hasFilterIfMissing() {
return ((bitField0_ & 0x00000010) == 0x00000010); return ((bitField0_ & 0x00000010) == 0x00000010);
@ -10399,8 +10399,8 @@ public final class FilterProtos {
return filterIfMissing_; return filterIfMissing_;
} }
// optional bool latestVersionOnly = 6; // optional bool latest_version_only = 6;
public static final int LATESTVERSIONONLY_FIELD_NUMBER = 6; public static final int LATEST_VERSION_ONLY_FIELD_NUMBER = 6;
private boolean latestVersionOnly_; private boolean latestVersionOnly_;
public boolean hasLatestVersionOnly() { public boolean hasLatestVersionOnly() {
return ((bitField0_ & 0x00000020) == 0x00000020); return ((bitField0_ & 0x00000020) == 0x00000020);
@ -10555,15 +10555,15 @@ public final class FilterProtos {
int hash = 41; int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode(); hash = (19 * hash) + getDescriptorForType().hashCode();
if (hasColumnFamily()) { if (hasColumnFamily()) {
hash = (37 * hash) + COLUMNFAMILY_FIELD_NUMBER; hash = (37 * hash) + COLUMN_FAMILY_FIELD_NUMBER;
hash = (53 * hash) + getColumnFamily().hashCode(); hash = (53 * hash) + getColumnFamily().hashCode();
} }
if (hasColumnQualifier()) { if (hasColumnQualifier()) {
hash = (37 * hash) + COLUMNQUALIFIER_FIELD_NUMBER; hash = (37 * hash) + COLUMN_QUALIFIER_FIELD_NUMBER;
hash = (53 * hash) + getColumnQualifier().hashCode(); hash = (53 * hash) + getColumnQualifier().hashCode();
} }
if (hasCompareOp()) { if (hasCompareOp()) {
hash = (37 * hash) + COMPAREOP_FIELD_NUMBER; hash = (37 * hash) + COMPARE_OP_FIELD_NUMBER;
hash = (53 * hash) + hashEnum(getCompareOp()); hash = (53 * hash) + hashEnum(getCompareOp());
} }
if (hasComparator()) { if (hasComparator()) {
@ -10571,11 +10571,11 @@ public final class FilterProtos {
hash = (53 * hash) + getComparator().hashCode(); hash = (53 * hash) + getComparator().hashCode();
} }
if (hasFilterIfMissing()) { if (hasFilterIfMissing()) {
hash = (37 * hash) + FILTERIFMISSING_FIELD_NUMBER; hash = (37 * hash) + FILTER_IF_MISSING_FIELD_NUMBER;
hash = (53 * hash) + hashBoolean(getFilterIfMissing()); hash = (53 * hash) + hashBoolean(getFilterIfMissing());
} }
if (hasLatestVersionOnly()) { if (hasLatestVersionOnly()) {
hash = (37 * hash) + LATESTVERSIONONLY_FIELD_NUMBER; hash = (37 * hash) + LATEST_VERSION_ONLY_FIELD_NUMBER;
hash = (53 * hash) + hashBoolean(getLatestVersionOnly()); hash = (53 * hash) + hashBoolean(getLatestVersionOnly());
} }
hash = (29 * hash) + getUnknownFields().hashCode(); hash = (29 * hash) + getUnknownFields().hashCode();
@ -10900,7 +10900,7 @@ public final class FilterProtos {
private int bitField0_; private int bitField0_;
// optional bytes columnFamily = 1; // optional bytes column_family = 1;
private com.google.protobuf.ByteString columnFamily_ = com.google.protobuf.ByteString.EMPTY; private com.google.protobuf.ByteString columnFamily_ = com.google.protobuf.ByteString.EMPTY;
public boolean hasColumnFamily() { public boolean hasColumnFamily() {
return ((bitField0_ & 0x00000001) == 0x00000001); return ((bitField0_ & 0x00000001) == 0x00000001);
@ -10924,7 +10924,7 @@ public final class FilterProtos {
return this; return this;
} }
// optional bytes columnQualifier = 2; // optional bytes column_qualifier = 2;
private com.google.protobuf.ByteString columnQualifier_ = com.google.protobuf.ByteString.EMPTY; private com.google.protobuf.ByteString columnQualifier_ = com.google.protobuf.ByteString.EMPTY;
public boolean hasColumnQualifier() { public boolean hasColumnQualifier() {
return ((bitField0_ & 0x00000002) == 0x00000002); return ((bitField0_ & 0x00000002) == 0x00000002);
@ -10948,7 +10948,7 @@ public final class FilterProtos {
return this; return this;
} }
// required .CompareType compareOp = 3; // required .CompareType compare_op = 3;
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.CompareType compareOp_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.CompareType.LESS; private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.CompareType compareOp_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.CompareType.LESS;
public boolean hasCompareOp() { public boolean hasCompareOp() {
return ((bitField0_ & 0x00000004) == 0x00000004); return ((bitField0_ & 0x00000004) == 0x00000004);
@ -11062,7 +11062,7 @@ public final class FilterProtos {
return comparatorBuilder_; return comparatorBuilder_;
} }
// optional bool filterIfMissing = 5; // optional bool filter_if_missing = 5;
private boolean filterIfMissing_ ; private boolean filterIfMissing_ ;
public boolean hasFilterIfMissing() { public boolean hasFilterIfMissing() {
return ((bitField0_ & 0x00000010) == 0x00000010); return ((bitField0_ & 0x00000010) == 0x00000010);
@ -11083,7 +11083,7 @@ public final class FilterProtos {
return this; return this;
} }
// optional bool latestVersionOnly = 6; // optional bool latest_version_only = 6;
private boolean latestVersionOnly_ ; private boolean latestVersionOnly_ ;
public boolean hasLatestVersionOnly() { public boolean hasLatestVersionOnly() {
return ((bitField0_ & 0x00000020) == 0x00000020); return ((bitField0_ & 0x00000020) == 0x00000020);
@ -12017,7 +12017,7 @@ public final class FilterProtos {
public interface ValueFilterOrBuilder public interface ValueFilterOrBuilder
extends com.google.protobuf.MessageOrBuilder { extends com.google.protobuf.MessageOrBuilder {
// required .CompareFilter compareFilter = 1; // required .CompareFilter compare_filter = 1;
boolean hasCompareFilter(); boolean hasCompareFilter();
org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter getCompareFilter(); org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter getCompareFilter();
org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilterOrBuilder getCompareFilterOrBuilder(); org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilterOrBuilder getCompareFilterOrBuilder();
@ -12051,8 +12051,8 @@ public final class FilterProtos {
} }
private int bitField0_; private int bitField0_;
// required .CompareFilter compareFilter = 1; // required .CompareFilter compare_filter = 1;
public static final int COMPAREFILTER_FIELD_NUMBER = 1; public static final int COMPARE_FILTER_FIELD_NUMBER = 1;
private org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter compareFilter_; private org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter compareFilter_;
public boolean hasCompareFilter() { public boolean hasCompareFilter() {
return ((bitField0_ & 0x00000001) == 0x00000001); return ((bitField0_ & 0x00000001) == 0x00000001);
@ -12141,7 +12141,7 @@ public final class FilterProtos {
int hash = 41; int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode(); hash = (19 * hash) + getDescriptorForType().hashCode();
if (hasCompareFilter()) { if (hasCompareFilter()) {
hash = (37 * hash) + COMPAREFILTER_FIELD_NUMBER; hash = (37 * hash) + COMPARE_FILTER_FIELD_NUMBER;
hash = (53 * hash) + getCompareFilter().hashCode(); hash = (53 * hash) + getCompareFilter().hashCode();
} }
hash = (29 * hash) + getUnknownFields().hashCode(); hash = (29 * hash) + getUnknownFields().hashCode();
@ -12386,7 +12386,7 @@ public final class FilterProtos {
private int bitField0_; private int bitField0_;
// required .CompareFilter compareFilter = 1; // required .CompareFilter compare_filter = 1;
private org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter compareFilter_ = org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter.getDefaultInstance(); private org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter compareFilter_ = org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter.getDefaultInstance();
private com.google.protobuf.SingleFieldBuilder< private com.google.protobuf.SingleFieldBuilder<
org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter, org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter.Builder, org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilterOrBuilder> compareFilterBuilder_; org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter, org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter.Builder, org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilterOrBuilder> compareFilterBuilder_;
@ -13105,51 +13105,52 @@ public final class FilterProtos {
static { static {
java.lang.String[] descriptorData = { java.lang.String[] descriptorData = {
"\n\014Filter.proto\032\013hbase.proto\032\020Comparator." + "\n\014Filter.proto\032\013hbase.proto\032\020Comparator." +
"proto\"0\n\006Filter\022\014\n\004name\030\001 \002(\t\022\030\n\020seriali" + "proto\"1\n\006Filter\022\014\n\004name\030\001 \002(\t\022\031\n\021seriali" +
"zedFilter\030\002 \001(\014\"%\n\024ColumnCountGetFilter\022" + "zed_filter\030\002 \001(\014\"%\n\024ColumnCountGetFilter" +
"\r\n\005limit\030\001 \002(\005\"M\n\026ColumnPaginationFilter" + "\022\r\n\005limit\030\001 \002(\005\"N\n\026ColumnPaginationFilte" +
"\022\r\n\005limit\030\001 \002(\005\022\016\n\006offset\030\002 \001(\005\022\024\n\014colum" + "r\022\r\n\005limit\030\001 \002(\005\022\016\n\006offset\030\002 \001(\005\022\025\n\rcolu" +
"nOffset\030\003 \001(\014\"$\n\022ColumnPrefixFilter\022\016\n\006p" + "mn_offset\030\003 \001(\014\"$\n\022ColumnPrefixFilter\022\016\n" +
"refix\030\001 \002(\014\"q\n\021ColumnRangeFilter\022\021\n\tminC" + "\006prefix\030\001 \002(\014\"w\n\021ColumnRangeFilter\022\022\n\nmi" +
"olumn\030\001 \001(\014\022\032\n\022minColumnInclusive\030\002 \001(\010\022" + "n_column\030\001 \001(\014\022\034\n\024min_column_inclusive\030\002" +
"\021\n\tmaxColumn\030\003 \001(\014\022\032\n\022maxColumnInclusive" + " \001(\010\022\022\n\nmax_column\030\003 \001(\014\022\034\n\024max_column_i" +
"\030\004 \001(\010\"Q\n\rCompareFilter\022\037\n\tcompareOp\030\001 \002", "nclusive\030\004 \001(\010\"R\n\rCompareFilter\022 \n\ncompa",
"(\0162\014.CompareType\022\037\n\ncomparator\030\002 \001(\0132\013.C" + "re_op\030\001 \002(\0162\014.CompareType\022\037\n\ncomparator\030" +
"omparator\"\212\001\n\025DependentColumnFilter\022%\n\rc" + "\002 \001(\0132\013.Comparator\"\217\001\n\025DependentColumnFi" +
"ompareFilter\030\001 \002(\0132\016.CompareFilter\022\024\n\014co" + "lter\022&\n\016compare_filter\030\001 \002(\0132\016.CompareFi" +
"lumnFamily\030\002 \001(\014\022\027\n\017columnQualifier\030\003 \001(" + "lter\022\025\n\rcolumn_family\030\002 \001(\014\022\030\n\020column_qu" +
"\014\022\033\n\023dropDependentColumn\030\004 \001(\010\"5\n\014Family" + "alifier\030\003 \001(\014\022\035\n\025drop_dependent_column\030\004" +
"Filter\022%\n\rcompareFilter\030\001 \002(\0132\016.CompareF" + " \001(\010\"6\n\014FamilyFilter\022&\n\016compare_filter\030\001" +
"ilter\"\200\001\n\nFilterList\022&\n\010operator\030\001 \002(\0162\024" + " \002(\0132\016.CompareFilter\"\200\001\n\nFilterList\022&\n\010o" +
".FilterList.Operator\022\030\n\007filters\030\002 \003(\0132\007." + "perator\030\001 \002(\0162\024.FilterList.Operator\022\030\n\007f" +
"Filter\"0\n\010Operator\022\021\n\rMUST_PASS_ALL\020\001\022\021\n" + "ilters\030\002 \003(\0132\007.Filter\"0\n\010Operator\022\021\n\rMUS" +
"\rMUST_PASS_ONE\020\002\"(\n\rFilterWrapper\022\027\n\006fil", "T_PASS_ALL\020\001\022\021\n\rMUST_PASS_ONE\020\002\"(\n\rFilte",
"ter\030\001 \002(\0132\007.Filter\"\024\n\022FirstKeyOnlyFilter" + "rWrapper\022\027\n\006filter\030\001 \002(\0132\007.Filter\"\024\n\022Fir" +
"\";\n%FirstKeyValueMatchingQualifiersFilte" + "stKeyOnlyFilter\";\n%FirstKeyValueMatching" +
"r\022\022\n\nqualifiers\030\001 \003(\014\"8\n\016FuzzyRowFilter\022" + "QualifiersFilter\022\022\n\nqualifiers\030\001 \003(\014\":\n\016" +
"&\n\rfuzzyKeysData\030\001 \003(\0132\017.BytesBytesPair\"" + "FuzzyRowFilter\022(\n\017fuzzy_keys_data\030\001 \003(\0132" +
")\n\023InclusiveStopFilter\022\022\n\nstopRowKey\030\001 \001" + "\017.BytesBytesPair\"+\n\023InclusiveStopFilter\022" +
"(\014\"!\n\rKeyOnlyFilter\022\020\n\010lenAsVal\030\001 \002(\010\"4\n" + "\024\n\014stop_row_key\030\001 \001(\014\"#\n\rKeyOnlyFilter\022\022" +
"\032MultipleColumnPrefixFilter\022\026\n\016sortedPre" + "\n\nlen_as_val\030\001 \002(\010\"5\n\032MultipleColumnPref" +
"fixes\030\001 \003(\014\"\036\n\nPageFilter\022\020\n\010pageSize\030\001 " + "ixFilter\022\027\n\017sorted_prefixes\030\001 \003(\014\"\037\n\nPag" +
"\002(\003\"\036\n\014PrefixFilter\022\016\n\006prefix\030\001 \001(\014\"8\n\017Q" + "eFilter\022\021\n\tpage_size\030\001 \002(\003\"\036\n\014PrefixFilt" +
"ualifierFilter\022%\n\rcompareFilter\030\001 \002(\0132\016.", "er\022\016\n\006prefix\030\001 \001(\014\"9\n\017QualifierFilter\022&\n",
"CompareFilter\"!\n\017RandomRowFilter\022\016\n\006chan" + "\016compare_filter\030\001 \002(\0132\016.CompareFilter\"!\n" +
"ce\030\001 \002(\002\"2\n\tRowFilter\022%\n\rcompareFilter\030\001" + "\017RandomRowFilter\022\016\n\006chance\030\001 \002(\002\"3\n\tRowF" +
" \002(\0132\016.CompareFilter\"[\n\036SingleColumnValu" + "ilter\022&\n\016compare_filter\030\001 \002(\0132\016.CompareF" +
"eExcludeFilter\0229\n\027singleColumnValueFilte" + "ilter\"^\n\036SingleColumnValueExcludeFilter\022" +
"r\030\001 \002(\0132\030.SingleColumnValueFilter\"\276\001\n\027Si" + "<\n\032single_column_value_filter\030\001 \002(\0132\030.Si" +
"ngleColumnValueFilter\022\024\n\014columnFamily\030\001 " + "ngleColumnValueFilter\"\305\001\n\027SingleColumnVa" +
"\001(\014\022\027\n\017columnQualifier\030\002 \001(\014\022\037\n\tcompareO" + "lueFilter\022\025\n\rcolumn_family\030\001 \001(\014\022\030\n\020colu" +
"p\030\003 \002(\0162\014.CompareType\022\037\n\ncomparator\030\004 \002(" + "mn_qualifier\030\002 \001(\014\022 \n\ncompare_op\030\003 \002(\0162\014" +
"\0132\013.Comparator\022\027\n\017filterIfMissing\030\005 \001(\010\022" + ".CompareType\022\037\n\ncomparator\030\004 \002(\0132\013.Compa" +
"\031\n\021latestVersionOnly\030\006 \001(\010\"%\n\nSkipFilter", "rator\022\031\n\021filter_if_missing\030\005 \001(\010\022\033\n\023late",
"\022\027\n\006filter\030\001 \002(\0132\007.Filter\"*\n\020TimestampsF" + "st_version_only\030\006 \001(\010\"%\n\nSkipFilter\022\027\n\006f" +
"ilter\022\026\n\ntimestamps\030\001 \003(\003B\002\020\001\"4\n\013ValueFi" + "ilter\030\001 \002(\0132\007.Filter\"*\n\020TimestampsFilter" +
"lter\022%\n\rcompareFilter\030\001 \002(\0132\016.CompareFil" + "\022\026\n\ntimestamps\030\001 \003(\003B\002\020\001\"5\n\013ValueFilter\022" +
"ter\"+\n\020WhileMatchFilter\022\027\n\006filter\030\001 \002(\0132" + "&\n\016compare_filter\030\001 \002(\0132\016.CompareFilter\"" +
"\007.FilterBB\n*org.apache.hadoop.hbase.prot" + "+\n\020WhileMatchFilter\022\027\n\006filter\030\001 \002(\0132\007.Fi" +
"obuf.generatedB\014FilterProtosH\001\210\001\001\240\001\001" "lterBB\n*org.apache.hadoop.hbase.protobuf" +
".generatedB\014FilterProtosH\001\210\001\001\240\001\001"
}; };
com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner = com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() { new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() {

View File

@ -109,7 +109,7 @@ public final class HBaseProtos {
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPairOrBuilder getAttributesOrBuilder( org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPairOrBuilder getAttributesOrBuilder(
int index); int index);
// repeated .ColumnFamilySchema columnFamilies = 3; // repeated .ColumnFamilySchema column_families = 3;
java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema> java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema>
getColumnFamiliesList(); getColumnFamiliesList();
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema getColumnFamilies(int index); org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema getColumnFamilies(int index);
@ -189,8 +189,8 @@ public final class HBaseProtos {
return attributes_.get(index); return attributes_.get(index);
} }
// repeated .ColumnFamilySchema columnFamilies = 3; // repeated .ColumnFamilySchema column_families = 3;
public static final int COLUMNFAMILIES_FIELD_NUMBER = 3; public static final int COLUMN_FAMILIES_FIELD_NUMBER = 3;
private java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema> columnFamilies_; private java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema> columnFamilies_;
public java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema> getColumnFamiliesList() { public java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema> getColumnFamiliesList() {
return columnFamilies_; return columnFamilies_;
@ -356,7 +356,7 @@ public final class HBaseProtos {
hash = (53 * hash) + getAttributesList().hashCode(); hash = (53 * hash) + getAttributesList().hashCode();
} }
if (getColumnFamiliesCount() > 0) { if (getColumnFamiliesCount() > 0) {
hash = (37 * hash) + COLUMNFAMILIES_FIELD_NUMBER; hash = (37 * hash) + COLUMN_FAMILIES_FIELD_NUMBER;
hash = (53 * hash) + getColumnFamiliesList().hashCode(); hash = (53 * hash) + getColumnFamiliesList().hashCode();
} }
if (getConfigurationCount() > 0) { if (getConfigurationCount() > 0) {
@ -956,7 +956,7 @@ public final class HBaseProtos {
return attributesBuilder_; return attributesBuilder_;
} }
// repeated .ColumnFamilySchema columnFamilies = 3; // repeated .ColumnFamilySchema column_families = 3;
private java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema> columnFamilies_ = private java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema> columnFamilies_ =
java.util.Collections.emptyList(); java.util.Collections.emptyList();
private void ensureColumnFamiliesIsMutable() { private void ensureColumnFamiliesIsMutable() {
@ -2306,19 +2306,19 @@ public final class HBaseProtos {
public interface RegionInfoOrBuilder public interface RegionInfoOrBuilder
extends com.google.protobuf.MessageOrBuilder { extends com.google.protobuf.MessageOrBuilder {
// required uint64 regionId = 1; // required uint64 region_id = 1;
boolean hasRegionId(); boolean hasRegionId();
long getRegionId(); long getRegionId();
// required bytes tableName = 2; // required bytes table_name = 2;
boolean hasTableName(); boolean hasTableName();
com.google.protobuf.ByteString getTableName(); com.google.protobuf.ByteString getTableName();
// optional bytes startKey = 3; // optional bytes start_key = 3;
boolean hasStartKey(); boolean hasStartKey();
com.google.protobuf.ByteString getStartKey(); com.google.protobuf.ByteString getStartKey();
// optional bytes endKey = 4; // optional bytes end_key = 4;
boolean hasEndKey(); boolean hasEndKey();
com.google.protobuf.ByteString getEndKey(); com.google.protobuf.ByteString getEndKey();
@ -2363,8 +2363,8 @@ public final class HBaseProtos {
} }
private int bitField0_; private int bitField0_;
// required uint64 regionId = 1; // required uint64 region_id = 1;
public static final int REGIONID_FIELD_NUMBER = 1; public static final int REGION_ID_FIELD_NUMBER = 1;
private long regionId_; private long regionId_;
public boolean hasRegionId() { public boolean hasRegionId() {
return ((bitField0_ & 0x00000001) == 0x00000001); return ((bitField0_ & 0x00000001) == 0x00000001);
@ -2373,8 +2373,8 @@ public final class HBaseProtos {
return regionId_; return regionId_;
} }
// required bytes tableName = 2; // required bytes table_name = 2;
public static final int TABLENAME_FIELD_NUMBER = 2; public static final int TABLE_NAME_FIELD_NUMBER = 2;
private com.google.protobuf.ByteString tableName_; private com.google.protobuf.ByteString tableName_;
public boolean hasTableName() { public boolean hasTableName() {
return ((bitField0_ & 0x00000002) == 0x00000002); return ((bitField0_ & 0x00000002) == 0x00000002);
@ -2383,8 +2383,8 @@ public final class HBaseProtos {
return tableName_; return tableName_;
} }
// optional bytes startKey = 3; // optional bytes start_key = 3;
public static final int STARTKEY_FIELD_NUMBER = 3; public static final int START_KEY_FIELD_NUMBER = 3;
private com.google.protobuf.ByteString startKey_; private com.google.protobuf.ByteString startKey_;
public boolean hasStartKey() { public boolean hasStartKey() {
return ((bitField0_ & 0x00000004) == 0x00000004); return ((bitField0_ & 0x00000004) == 0x00000004);
@ -2393,8 +2393,8 @@ public final class HBaseProtos {
return startKey_; return startKey_;
} }
// optional bytes endKey = 4; // optional bytes end_key = 4;
public static final int ENDKEY_FIELD_NUMBER = 4; public static final int END_KEY_FIELD_NUMBER = 4;
private com.google.protobuf.ByteString endKey_; private com.google.protobuf.ByteString endKey_;
public boolean hasEndKey() { public boolean hasEndKey() {
return ((bitField0_ & 0x00000008) == 0x00000008); return ((bitField0_ & 0x00000008) == 0x00000008);
@ -2588,19 +2588,19 @@ public final class HBaseProtos {
int hash = 41; int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode(); hash = (19 * hash) + getDescriptorForType().hashCode();
if (hasRegionId()) { if (hasRegionId()) {
hash = (37 * hash) + REGIONID_FIELD_NUMBER; hash = (37 * hash) + REGION_ID_FIELD_NUMBER;
hash = (53 * hash) + hashLong(getRegionId()); hash = (53 * hash) + hashLong(getRegionId());
} }
if (hasTableName()) { if (hasTableName()) {
hash = (37 * hash) + TABLENAME_FIELD_NUMBER; hash = (37 * hash) + TABLE_NAME_FIELD_NUMBER;
hash = (53 * hash) + getTableName().hashCode(); hash = (53 * hash) + getTableName().hashCode();
} }
if (hasStartKey()) { if (hasStartKey()) {
hash = (37 * hash) + STARTKEY_FIELD_NUMBER; hash = (37 * hash) + START_KEY_FIELD_NUMBER;
hash = (53 * hash) + getStartKey().hashCode(); hash = (53 * hash) + getStartKey().hashCode();
} }
if (hasEndKey()) { if (hasEndKey()) {
hash = (37 * hash) + ENDKEY_FIELD_NUMBER; hash = (37 * hash) + END_KEY_FIELD_NUMBER;
hash = (53 * hash) + getEndKey().hashCode(); hash = (53 * hash) + getEndKey().hashCode();
} }
if (hasOffline()) { if (hasOffline()) {
@ -2928,7 +2928,7 @@ public final class HBaseProtos {
private int bitField0_; private int bitField0_;
// required uint64 regionId = 1; // required uint64 region_id = 1;
private long regionId_ ; private long regionId_ ;
public boolean hasRegionId() { public boolean hasRegionId() {
return ((bitField0_ & 0x00000001) == 0x00000001); return ((bitField0_ & 0x00000001) == 0x00000001);
@ -2949,7 +2949,7 @@ public final class HBaseProtos {
return this; return this;
} }
// required bytes tableName = 2; // required bytes table_name = 2;
private com.google.protobuf.ByteString tableName_ = com.google.protobuf.ByteString.EMPTY; private com.google.protobuf.ByteString tableName_ = com.google.protobuf.ByteString.EMPTY;
public boolean hasTableName() { public boolean hasTableName() {
return ((bitField0_ & 0x00000002) == 0x00000002); return ((bitField0_ & 0x00000002) == 0x00000002);
@ -2973,7 +2973,7 @@ public final class HBaseProtos {
return this; return this;
} }
// optional bytes startKey = 3; // optional bytes start_key = 3;
private com.google.protobuf.ByteString startKey_ = com.google.protobuf.ByteString.EMPTY; private com.google.protobuf.ByteString startKey_ = com.google.protobuf.ByteString.EMPTY;
public boolean hasStartKey() { public boolean hasStartKey() {
return ((bitField0_ & 0x00000004) == 0x00000004); return ((bitField0_ & 0x00000004) == 0x00000004);
@ -2997,7 +2997,7 @@ public final class HBaseProtos {
return this; return this;
} }
// optional bytes endKey = 4; // optional bytes end_key = 4;
private com.google.protobuf.ByteString endKey_ = com.google.protobuf.ByteString.EMPTY; private com.google.protobuf.ByteString endKey_ = com.google.protobuf.ByteString.EMPTY;
public boolean hasEndKey() { public boolean hasEndKey() {
return ((bitField0_ & 0x00000008) == 0x00000008); return ((bitField0_ & 0x00000008) == 0x00000008);
@ -3098,7 +3098,7 @@ public final class HBaseProtos {
public interface FavoredNodesOrBuilder public interface FavoredNodesOrBuilder
extends com.google.protobuf.MessageOrBuilder { extends com.google.protobuf.MessageOrBuilder {
// repeated .ServerName favoredNode = 1; // repeated .ServerName favored_node = 1;
java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName> java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName>
getFavoredNodeList(); getFavoredNodeList();
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName getFavoredNode(int index); org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName getFavoredNode(int index);
@ -3136,8 +3136,8 @@ public final class HBaseProtos {
return org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.internal_static_FavoredNodes_fieldAccessorTable; return org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.internal_static_FavoredNodes_fieldAccessorTable;
} }
// repeated .ServerName favoredNode = 1; // repeated .ServerName favored_node = 1;
public static final int FAVOREDNODE_FIELD_NUMBER = 1; public static final int FAVORED_NODE_FIELD_NUMBER = 1;
private java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName> favoredNode_; private java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName> favoredNode_;
public java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName> getFavoredNodeList() { public java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName> getFavoredNodeList() {
return favoredNode_; return favoredNode_;
@ -3229,7 +3229,7 @@ public final class HBaseProtos {
int hash = 41; int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode(); hash = (19 * hash) + getDescriptorForType().hashCode();
if (getFavoredNodeCount() > 0) { if (getFavoredNodeCount() > 0) {
hash = (37 * hash) + FAVOREDNODE_FIELD_NUMBER; hash = (37 * hash) + FAVORED_NODE_FIELD_NUMBER;
hash = (53 * hash) + getFavoredNodeList().hashCode(); hash = (53 * hash) + getFavoredNodeList().hashCode();
} }
hash = (29 * hash) + getUnknownFields().hashCode(); hash = (29 * hash) + getUnknownFields().hashCode();
@ -3491,7 +3491,7 @@ public final class HBaseProtos {
private int bitField0_; private int bitField0_;
// repeated .ServerName favoredNode = 1; // repeated .ServerName favored_node = 1;
private java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName> favoredNode_ = private java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName> favoredNode_ =
java.util.Collections.emptyList(); java.util.Collections.emptyList();
private void ensureFavoredNodeIsMutable() { private void ensureFavoredNodeIsMutable() {
@ -4662,7 +4662,7 @@ public final class HBaseProtos {
public interface ServerNameOrBuilder public interface ServerNameOrBuilder
extends com.google.protobuf.MessageOrBuilder { extends com.google.protobuf.MessageOrBuilder {
// required string hostName = 1; // required string host_name = 1;
boolean hasHostName(); boolean hasHostName();
String getHostName(); String getHostName();
@ -4670,7 +4670,7 @@ public final class HBaseProtos {
boolean hasPort(); boolean hasPort();
int getPort(); int getPort();
// optional uint64 startCode = 3; // optional uint64 start_code = 3;
boolean hasStartCode(); boolean hasStartCode();
long getStartCode(); long getStartCode();
} }
@ -4703,8 +4703,8 @@ public final class HBaseProtos {
} }
private int bitField0_; private int bitField0_;
// required string hostName = 1; // required string host_name = 1;
public static final int HOSTNAME_FIELD_NUMBER = 1; public static final int HOST_NAME_FIELD_NUMBER = 1;
private java.lang.Object hostName_; private java.lang.Object hostName_;
public boolean hasHostName() { public boolean hasHostName() {
return ((bitField0_ & 0x00000001) == 0x00000001); return ((bitField0_ & 0x00000001) == 0x00000001);
@ -4745,8 +4745,8 @@ public final class HBaseProtos {
return port_; return port_;
} }
// optional uint64 startCode = 3; // optional uint64 start_code = 3;
public static final int STARTCODE_FIELD_NUMBER = 3; public static final int START_CODE_FIELD_NUMBER = 3;
private long startCode_; private long startCode_;
public boolean hasStartCode() { public boolean hasStartCode() {
return ((bitField0_ & 0x00000004) == 0x00000004); return ((bitField0_ & 0x00000004) == 0x00000004);
@ -4854,7 +4854,7 @@ public final class HBaseProtos {
int hash = 41; int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode(); hash = (19 * hash) + getDescriptorForType().hashCode();
if (hasHostName()) { if (hasHostName()) {
hash = (37 * hash) + HOSTNAME_FIELD_NUMBER; hash = (37 * hash) + HOST_NAME_FIELD_NUMBER;
hash = (53 * hash) + getHostName().hashCode(); hash = (53 * hash) + getHostName().hashCode();
} }
if (hasPort()) { if (hasPort()) {
@ -4862,7 +4862,7 @@ public final class HBaseProtos {
hash = (53 * hash) + getPort(); hash = (53 * hash) + getPort();
} }
if (hasStartCode()) { if (hasStartCode()) {
hash = (37 * hash) + STARTCODE_FIELD_NUMBER; hash = (37 * hash) + START_CODE_FIELD_NUMBER;
hash = (53 * hash) + hashLong(getStartCode()); hash = (53 * hash) + hashLong(getStartCode());
} }
hash = (29 * hash) + getUnknownFields().hashCode(); hash = (29 * hash) + getUnknownFields().hashCode();
@ -5118,7 +5118,7 @@ public final class HBaseProtos {
private int bitField0_; private int bitField0_;
// required string hostName = 1; // required string host_name = 1;
private java.lang.Object hostName_ = ""; private java.lang.Object hostName_ = "";
public boolean hasHostName() { public boolean hasHostName() {
return ((bitField0_ & 0x00000001) == 0x00000001); return ((bitField0_ & 0x00000001) == 0x00000001);
@ -5175,7 +5175,7 @@ public final class HBaseProtos {
return this; return this;
} }
// optional uint64 startCode = 3; // optional uint64 start_code = 3;
private long startCode_ ; private long startCode_ ;
public boolean hasStartCode() { public boolean hasStartCode() {
return ((bitField0_ & 0x00000004) == 0x00000004); return ((bitField0_ & 0x00000004) == 0x00000004);
@ -7579,7 +7579,7 @@ public final class HBaseProtos {
boolean hasTable(); boolean hasTable();
String getTable(); String getTable();
// optional int64 creationTime = 3 [default = 0]; // optional int64 creation_time = 3 [default = 0];
boolean hasCreationTime(); boolean hasCreationTime();
long getCreationTime(); long getCreationTime();
@ -7753,8 +7753,8 @@ public final class HBaseProtos {
} }
} }
// optional int64 creationTime = 3 [default = 0]; // optional int64 creation_time = 3 [default = 0];
public static final int CREATIONTIME_FIELD_NUMBER = 3; public static final int CREATION_TIME_FIELD_NUMBER = 3;
private long creationTime_; private long creationTime_;
public boolean hasCreationTime() { public boolean hasCreationTime() {
return ((bitField0_ & 0x00000004) == 0x00000004); return ((bitField0_ & 0x00000004) == 0x00000004);
@ -7916,7 +7916,7 @@ public final class HBaseProtos {
hash = (53 * hash) + getTable().hashCode(); hash = (53 * hash) + getTable().hashCode();
} }
if (hasCreationTime()) { if (hasCreationTime()) {
hash = (37 * hash) + CREATIONTIME_FIELD_NUMBER; hash = (37 * hash) + CREATION_TIME_FIELD_NUMBER;
hash = (53 * hash) + hashLong(getCreationTime()); hash = (53 * hash) + hashLong(getCreationTime());
} }
if (hasType()) { if (hasType()) {
@ -8286,7 +8286,7 @@ public final class HBaseProtos {
onChanged(); onChanged();
} }
// optional int64 creationTime = 3 [default = 0]; // optional int64 creation_time = 3 [default = 0];
private long creationTime_ ; private long creationTime_ ;
public boolean hasCreationTime() { public boolean hasCreationTime() {
return ((bitField0_ & 0x00000004) == 0x00000004); return ((bitField0_ & 0x00000004) == 0x00000004);
@ -8666,7 +8666,7 @@ public final class HBaseProtos {
public interface LongMsgOrBuilder public interface LongMsgOrBuilder
extends com.google.protobuf.MessageOrBuilder { extends com.google.protobuf.MessageOrBuilder {
// required int64 longMsg = 1; // required int64 long_msg = 1;
boolean hasLongMsg(); boolean hasLongMsg();
long getLongMsg(); long getLongMsg();
} }
@ -8699,8 +8699,8 @@ public final class HBaseProtos {
} }
private int bitField0_; private int bitField0_;
// required int64 longMsg = 1; // required int64 long_msg = 1;
public static final int LONGMSG_FIELD_NUMBER = 1; public static final int LONG_MSG_FIELD_NUMBER = 1;
private long longMsg_; private long longMsg_;
public boolean hasLongMsg() { public boolean hasLongMsg() {
return ((bitField0_ & 0x00000001) == 0x00000001); return ((bitField0_ & 0x00000001) == 0x00000001);
@ -8782,7 +8782,7 @@ public final class HBaseProtos {
int hash = 41; int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode(); hash = (19 * hash) + getDescriptorForType().hashCode();
if (hasLongMsg()) { if (hasLongMsg()) {
hash = (37 * hash) + LONGMSG_FIELD_NUMBER; hash = (37 * hash) + LONG_MSG_FIELD_NUMBER;
hash = (53 * hash) + hashLong(getLongMsg()); hash = (53 * hash) + hashLong(getLongMsg());
} }
hash = (29 * hash) + getUnknownFields().hashCode(); hash = (29 * hash) + getUnknownFields().hashCode();
@ -9010,7 +9010,7 @@ public final class HBaseProtos {
private int bitField0_; private int bitField0_;
// required int64 longMsg = 1; // required int64 long_msg = 1;
private long longMsg_ ; private long longMsg_ ;
public boolean hasLongMsg() { public boolean hasLongMsg() {
return ((bitField0_ & 0x00000001) == 0x00000001); return ((bitField0_ & 0x00000001) == 0x00000001);
@ -9045,7 +9045,7 @@ public final class HBaseProtos {
public interface BigDecimalMsgOrBuilder public interface BigDecimalMsgOrBuilder
extends com.google.protobuf.MessageOrBuilder { extends com.google.protobuf.MessageOrBuilder {
// required bytes bigdecimalMsg = 1; // required bytes bigdecimal_msg = 1;
boolean hasBigdecimalMsg(); boolean hasBigdecimalMsg();
com.google.protobuf.ByteString getBigdecimalMsg(); com.google.protobuf.ByteString getBigdecimalMsg();
} }
@ -9078,8 +9078,8 @@ public final class HBaseProtos {
} }
private int bitField0_; private int bitField0_;
// required bytes bigdecimalMsg = 1; // required bytes bigdecimal_msg = 1;
public static final int BIGDECIMALMSG_FIELD_NUMBER = 1; public static final int BIGDECIMAL_MSG_FIELD_NUMBER = 1;
private com.google.protobuf.ByteString bigdecimalMsg_; private com.google.protobuf.ByteString bigdecimalMsg_;
public boolean hasBigdecimalMsg() { public boolean hasBigdecimalMsg() {
return ((bitField0_ & 0x00000001) == 0x00000001); return ((bitField0_ & 0x00000001) == 0x00000001);
@ -9161,7 +9161,7 @@ public final class HBaseProtos {
int hash = 41; int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode(); hash = (19 * hash) + getDescriptorForType().hashCode();
if (hasBigdecimalMsg()) { if (hasBigdecimalMsg()) {
hash = (37 * hash) + BIGDECIMALMSG_FIELD_NUMBER; hash = (37 * hash) + BIGDECIMAL_MSG_FIELD_NUMBER;
hash = (53 * hash) + getBigdecimalMsg().hashCode(); hash = (53 * hash) + getBigdecimalMsg().hashCode();
} }
hash = (29 * hash) + getUnknownFields().hashCode(); hash = (29 * hash) + getUnknownFields().hashCode();
@ -9389,7 +9389,7 @@ public final class HBaseProtos {
private int bitField0_; private int bitField0_;
// required bytes bigdecimalMsg = 1; // required bytes bigdecimal_msg = 1;
private com.google.protobuf.ByteString bigdecimalMsg_ = com.google.protobuf.ByteString.EMPTY; private com.google.protobuf.ByteString bigdecimalMsg_ = com.google.protobuf.ByteString.EMPTY;
public boolean hasBigdecimalMsg() { public boolean hasBigdecimalMsg() {
return ((bitField0_ & 0x00000001) == 0x00000001); return ((bitField0_ & 0x00000001) == 0x00000001);
@ -9427,11 +9427,11 @@ public final class HBaseProtos {
public interface UUIDOrBuilder public interface UUIDOrBuilder
extends com.google.protobuf.MessageOrBuilder { extends com.google.protobuf.MessageOrBuilder {
// required uint64 leastSigBits = 1; // required uint64 least_sig_bits = 1;
boolean hasLeastSigBits(); boolean hasLeastSigBits();
long getLeastSigBits(); long getLeastSigBits();
// required uint64 mostSigBits = 2; // required uint64 most_sig_bits = 2;
boolean hasMostSigBits(); boolean hasMostSigBits();
long getMostSigBits(); long getMostSigBits();
} }
@ -9464,8 +9464,8 @@ public final class HBaseProtos {
} }
private int bitField0_; private int bitField0_;
// required uint64 leastSigBits = 1; // required uint64 least_sig_bits = 1;
public static final int LEASTSIGBITS_FIELD_NUMBER = 1; public static final int LEAST_SIG_BITS_FIELD_NUMBER = 1;
private long leastSigBits_; private long leastSigBits_;
public boolean hasLeastSigBits() { public boolean hasLeastSigBits() {
return ((bitField0_ & 0x00000001) == 0x00000001); return ((bitField0_ & 0x00000001) == 0x00000001);
@ -9474,8 +9474,8 @@ public final class HBaseProtos {
return leastSigBits_; return leastSigBits_;
} }
// required uint64 mostSigBits = 2; // required uint64 most_sig_bits = 2;
public static final int MOSTSIGBITS_FIELD_NUMBER = 2; public static final int MOST_SIG_BITS_FIELD_NUMBER = 2;
private long mostSigBits_; private long mostSigBits_;
public boolean hasMostSigBits() { public boolean hasMostSigBits() {
return ((bitField0_ & 0x00000002) == 0x00000002); return ((bitField0_ & 0x00000002) == 0x00000002);
@ -9574,11 +9574,11 @@ public final class HBaseProtos {
int hash = 41; int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode(); hash = (19 * hash) + getDescriptorForType().hashCode();
if (hasLeastSigBits()) { if (hasLeastSigBits()) {
hash = (37 * hash) + LEASTSIGBITS_FIELD_NUMBER; hash = (37 * hash) + LEAST_SIG_BITS_FIELD_NUMBER;
hash = (53 * hash) + hashLong(getLeastSigBits()); hash = (53 * hash) + hashLong(getLeastSigBits());
} }
if (hasMostSigBits()) { if (hasMostSigBits()) {
hash = (37 * hash) + MOSTSIGBITS_FIELD_NUMBER; hash = (37 * hash) + MOST_SIG_BITS_FIELD_NUMBER;
hash = (53 * hash) + hashLong(getMostSigBits()); hash = (53 * hash) + hashLong(getMostSigBits());
} }
hash = (29 * hash) + getUnknownFields().hashCode(); hash = (29 * hash) + getUnknownFields().hashCode();
@ -9824,7 +9824,7 @@ public final class HBaseProtos {
private int bitField0_; private int bitField0_;
// required uint64 leastSigBits = 1; // required uint64 least_sig_bits = 1;
private long leastSigBits_ ; private long leastSigBits_ ;
public boolean hasLeastSigBits() { public boolean hasLeastSigBits() {
return ((bitField0_ & 0x00000001) == 0x00000001); return ((bitField0_ & 0x00000001) == 0x00000001);
@ -9845,7 +9845,7 @@ public final class HBaseProtos {
return this; return this;
} }
// required uint64 mostSigBits = 2; // required uint64 most_sig_bits = 2;
private long mostSigBits_ ; private long mostSigBits_ ;
public boolean hasMostSigBits() { public boolean hasMostSigBits() {
return ((bitField0_ & 0x00000002) == 0x00000002); return ((bitField0_ & 0x00000002) == 0x00000002);
@ -9971,42 +9971,43 @@ public final class HBaseProtos {
descriptor; descriptor;
static { static {
java.lang.String[] descriptorData = { java.lang.String[] descriptorData = {
"\n\013hbase.proto\032\nCell.proto\"\225\001\n\013TableSchem" + "\n\013hbase.proto\032\nCell.proto\"\226\001\n\013TableSchem" +
"a\022\014\n\004name\030\001 \001(\014\022#\n\nattributes\030\002 \003(\0132\017.By" + "a\022\014\n\004name\030\001 \001(\014\022#\n\nattributes\030\002 \003(\0132\017.By" +
"tesBytesPair\022+\n\016columnFamilies\030\003 \003(\0132\023.C" + "tesBytesPair\022,\n\017column_families\030\003 \003(\0132\023." +
"olumnFamilySchema\022&\n\rconfiguration\030\004 \003(\013" + "ColumnFamilySchema\022&\n\rconfiguration\030\004 \003(" +
"2\017.NameStringPair\"o\n\022ColumnFamilySchema\022" + "\0132\017.NameStringPair\"o\n\022ColumnFamilySchema" +
"\014\n\004name\030\001 \002(\014\022#\n\nattributes\030\002 \003(\0132\017.Byte" + "\022\014\n\004name\030\001 \002(\014\022#\n\nattributes\030\002 \003(\0132\017.Byt" +
"sBytesPair\022&\n\rconfiguration\030\003 \003(\0132\017.Name" + "esBytesPair\022&\n\rconfiguration\030\003 \003(\0132\017.Nam" +
"StringPair\"\207\001\n\nRegionInfo\022\020\n\010regionId\030\001 " + "eStringPair\"\213\001\n\nRegionInfo\022\021\n\tregion_id\030" +
"\002(\004\022\021\n\ttableName\030\002 \002(\014\022\020\n\010startKey\030\003 \001(\014" + "\001 \002(\004\022\022\n\ntable_name\030\002 \002(\014\022\021\n\tstart_key\030\003" +
"\022\016\n\006endKey\030\004 \001(\014\022\017\n\007offline\030\005 \001(\010\022\r\n\005spl", " \001(\014\022\017\n\007end_key\030\004 \001(\014\022\017\n\007offline\030\005 \001(\010\022\r",
"it\030\006 \001(\010\022\022\n\nrecovering\030\007 \001(\010\"0\n\014FavoredN" + "\n\005split\030\006 \001(\010\022\022\n\nrecovering\030\007 \001(\010\"1\n\014Fav" +
"odes\022 \n\013favoredNode\030\001 \003(\0132\013.ServerName\"\225" + "oredNodes\022!\n\014favored_node\030\001 \003(\0132\013.Server" +
"\001\n\017RegionSpecifier\0222\n\004type\030\001 \002(\0162$.Regio" + "Name\"\225\001\n\017RegionSpecifier\0222\n\004type\030\001 \002(\0162$" +
"nSpecifier.RegionSpecifierType\022\r\n\005value\030" + ".RegionSpecifier.RegionSpecifierType\022\r\n\005" +
"\002 \002(\014\"?\n\023RegionSpecifierType\022\017\n\013REGION_N" + "value\030\002 \002(\014\"?\n\023RegionSpecifierType\022\017\n\013RE" +
"AME\020\001\022\027\n\023ENCODED_REGION_NAME\020\002\"%\n\tTimeRa" + "GION_NAME\020\001\022\027\n\023ENCODED_REGION_NAME\020\002\"%\n\t" +
"nge\022\014\n\004from\030\001 \001(\004\022\n\n\002to\030\002 \001(\004\"?\n\nServerN" + "TimeRange\022\014\n\004from\030\001 \001(\004\022\n\n\002to\030\002 \001(\004\"A\n\nS" +
"ame\022\020\n\010hostName\030\001 \002(\t\022\014\n\004port\030\002 \001(\r\022\021\n\ts" + "erverName\022\021\n\thost_name\030\001 \002(\t\022\014\n\004port\030\002 \001" +
"tartCode\030\003 \001(\004\"\033\n\013Coprocessor\022\014\n\004name\030\001 " + "(\r\022\022\n\nstart_code\030\003 \001(\004\"\033\n\013Coprocessor\022\014\n" +
"\002(\t\"-\n\016NameStringPair\022\014\n\004name\030\001 \002(\t\022\r\n\005v", "\004name\030\001 \002(\t\"-\n\016NameStringPair\022\014\n\004name\030\001 ",
"alue\030\002 \002(\t\",\n\rNameBytesPair\022\014\n\004name\030\001 \002(" + "\002(\t\022\r\n\005value\030\002 \002(\t\",\n\rNameBytesPair\022\014\n\004n" +
"\t\022\r\n\005value\030\002 \001(\014\"/\n\016BytesBytesPair\022\r\n\005fi" + "ame\030\001 \002(\t\022\r\n\005value\030\002 \001(\014\"/\n\016BytesBytesPa" +
"rst\030\001 \002(\014\022\016\n\006second\030\002 \002(\014\",\n\rNameInt64Pa" + "ir\022\r\n\005first\030\001 \002(\014\022\016\n\006second\030\002 \002(\014\",\n\rNam" +
"ir\022\014\n\004name\030\001 \001(\t\022\r\n\005value\030\002 \001(\003\"\255\001\n\023Snap" + "eInt64Pair\022\014\n\004name\030\001 \001(\t\022\r\n\005value\030\002 \001(\003\"" +
"shotDescription\022\014\n\004name\030\001 \002(\t\022\r\n\005table\030\002" + "\256\001\n\023SnapshotDescription\022\014\n\004name\030\001 \002(\t\022\r\n" +
" \001(\t\022\027\n\014creationTime\030\003 \001(\003:\0010\022.\n\004type\030\004 " + "\005table\030\002 \001(\t\022\030\n\rcreation_time\030\003 \001(\003:\0010\022." +
"\001(\0162\031.SnapshotDescription.Type:\005FLUSH\022\017\n" + "\n\004type\030\004 \001(\0162\031.SnapshotDescription.Type:" +
"\007version\030\005 \001(\005\"\037\n\004Type\022\014\n\010DISABLED\020\000\022\t\n\005" + "\005FLUSH\022\017\n\007version\030\005 \001(\005\"\037\n\004Type\022\014\n\010DISAB" +
"FLUSH\020\001\"\n\n\010EmptyMsg\"\032\n\007LongMsg\022\017\n\007longMs" + "LED\020\000\022\t\n\005FLUSH\020\001\"\n\n\010EmptyMsg\"\033\n\007LongMsg\022" +
"g\030\001 \002(\003\"&\n\rBigDecimalMsg\022\025\n\rbigdecimalMs", "\020\n\010long_msg\030\001 \002(\003\"\'\n\rBigDecimalMsg\022\026\n\016bi",
"g\030\001 \002(\014\"1\n\004UUID\022\024\n\014leastSigBits\030\001 \002(\004\022\023\n" + "gdecimal_msg\030\001 \002(\014\"5\n\004UUID\022\026\n\016least_sig_" +
"\013mostSigBits\030\002 \002(\004*r\n\013CompareType\022\010\n\004LES" + "bits\030\001 \002(\004\022\025\n\rmost_sig_bits\030\002 \002(\004*r\n\013Com" +
"S\020\000\022\021\n\rLESS_OR_EQUAL\020\001\022\t\n\005EQUAL\020\002\022\r\n\tNOT" + "pareType\022\010\n\004LESS\020\000\022\021\n\rLESS_OR_EQUAL\020\001\022\t\n" +
"_EQUAL\020\003\022\024\n\020GREATER_OR_EQUAL\020\004\022\013\n\007GREATE" + "\005EQUAL\020\002\022\r\n\tNOT_EQUAL\020\003\022\024\n\020GREATER_OR_EQ" +
"R\020\005\022\t\n\005NO_OP\020\006B>\n*org.apache.hadoop.hbas" + "UAL\020\004\022\013\n\007GREATER\020\005\022\t\n\005NO_OP\020\006B>\n*org.apa" +
"e.protobuf.generatedB\013HBaseProtosH\001\240\001\001" "che.hadoop.hbase.protobuf.generatedB\013HBa" +
"seProtosH\001\240\001\001"
}; };
com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner = com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() { new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() {

View File

@ -11,7 +11,7 @@ public final class HFileProtos {
public interface FileInfoProtoOrBuilder public interface FileInfoProtoOrBuilder
extends com.google.protobuf.MessageOrBuilder { extends com.google.protobuf.MessageOrBuilder {
// repeated .BytesBytesPair mapEntry = 1; // repeated .BytesBytesPair map_entry = 1;
java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair> java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair>
getMapEntryList(); getMapEntryList();
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair getMapEntry(int index); org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair getMapEntry(int index);
@ -49,8 +49,8 @@ public final class HFileProtos {
return org.apache.hadoop.hbase.protobuf.generated.HFileProtos.internal_static_FileInfoProto_fieldAccessorTable; return org.apache.hadoop.hbase.protobuf.generated.HFileProtos.internal_static_FileInfoProto_fieldAccessorTable;
} }
// repeated .BytesBytesPair mapEntry = 1; // repeated .BytesBytesPair map_entry = 1;
public static final int MAPENTRY_FIELD_NUMBER = 1; public static final int MAP_ENTRY_FIELD_NUMBER = 1;
private java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair> mapEntry_; private java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair> mapEntry_;
public java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair> getMapEntryList() { public java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair> getMapEntryList() {
return mapEntry_; return mapEntry_;
@ -142,7 +142,7 @@ public final class HFileProtos {
int hash = 41; int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode(); hash = (19 * hash) + getDescriptorForType().hashCode();
if (getMapEntryCount() > 0) { if (getMapEntryCount() > 0) {
hash = (37 * hash) + MAPENTRY_FIELD_NUMBER; hash = (37 * hash) + MAP_ENTRY_FIELD_NUMBER;
hash = (53 * hash) + getMapEntryList().hashCode(); hash = (53 * hash) + getMapEntryList().hashCode();
} }
hash = (29 * hash) + getUnknownFields().hashCode(); hash = (29 * hash) + getUnknownFields().hashCode();
@ -404,7 +404,7 @@ public final class HFileProtos {
private int bitField0_; private int bitField0_;
// repeated .BytesBytesPair mapEntry = 1; // repeated .BytesBytesPair map_entry = 1;
private java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair> mapEntry_ = private java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair> mapEntry_ =
java.util.Collections.emptyList(); java.util.Collections.emptyList();
private void ensureMapEntryIsMutable() { private void ensureMapEntryIsMutable() {
@ -604,51 +604,51 @@ public final class HFileProtos {
public interface FileTrailerProtoOrBuilder public interface FileTrailerProtoOrBuilder
extends com.google.protobuf.MessageOrBuilder { extends com.google.protobuf.MessageOrBuilder {
// optional uint64 fileInfoOffset = 1; // optional uint64 file_info_offset = 1;
boolean hasFileInfoOffset(); boolean hasFileInfoOffset();
long getFileInfoOffset(); long getFileInfoOffset();
// optional uint64 loadOnOpenDataOffset = 2; // optional uint64 load_on_open_data_offset = 2;
boolean hasLoadOnOpenDataOffset(); boolean hasLoadOnOpenDataOffset();
long getLoadOnOpenDataOffset(); long getLoadOnOpenDataOffset();
// optional uint64 uncompressedDataIndexSize = 3; // optional uint64 uncompressed_data_index_size = 3;
boolean hasUncompressedDataIndexSize(); boolean hasUncompressedDataIndexSize();
long getUncompressedDataIndexSize(); long getUncompressedDataIndexSize();
// optional uint64 totalUncompressedBytes = 4; // optional uint64 total_uncompressed_bytes = 4;
boolean hasTotalUncompressedBytes(); boolean hasTotalUncompressedBytes();
long getTotalUncompressedBytes(); long getTotalUncompressedBytes();
// optional uint32 dataIndexCount = 5; // optional uint32 data_index_count = 5;
boolean hasDataIndexCount(); boolean hasDataIndexCount();
int getDataIndexCount(); int getDataIndexCount();
// optional uint32 metaIndexCount = 6; // optional uint32 meta_index_count = 6;
boolean hasMetaIndexCount(); boolean hasMetaIndexCount();
int getMetaIndexCount(); int getMetaIndexCount();
// optional uint64 entryCount = 7; // optional uint64 entry_count = 7;
boolean hasEntryCount(); boolean hasEntryCount();
long getEntryCount(); long getEntryCount();
// optional uint32 numDataIndexLevels = 8; // optional uint32 num_data_index_levels = 8;
boolean hasNumDataIndexLevels(); boolean hasNumDataIndexLevels();
int getNumDataIndexLevels(); int getNumDataIndexLevels();
// optional uint64 firstDataBlockOffset = 9; // optional uint64 first_data_block_offset = 9;
boolean hasFirstDataBlockOffset(); boolean hasFirstDataBlockOffset();
long getFirstDataBlockOffset(); long getFirstDataBlockOffset();
// optional uint64 lastDataBlockOffset = 10; // optional uint64 last_data_block_offset = 10;
boolean hasLastDataBlockOffset(); boolean hasLastDataBlockOffset();
long getLastDataBlockOffset(); long getLastDataBlockOffset();
// optional string comparatorClassName = 11; // optional string comparator_class_name = 11;
boolean hasComparatorClassName(); boolean hasComparatorClassName();
String getComparatorClassName(); String getComparatorClassName();
// optional uint32 compressionCodec = 12; // optional uint32 compression_codec = 12;
boolean hasCompressionCodec(); boolean hasCompressionCodec();
int getCompressionCodec(); int getCompressionCodec();
} }
@ -681,8 +681,8 @@ public final class HFileProtos {
} }
private int bitField0_; private int bitField0_;
// optional uint64 fileInfoOffset = 1; // optional uint64 file_info_offset = 1;
public static final int FILEINFOOFFSET_FIELD_NUMBER = 1; public static final int FILE_INFO_OFFSET_FIELD_NUMBER = 1;
private long fileInfoOffset_; private long fileInfoOffset_;
public boolean hasFileInfoOffset() { public boolean hasFileInfoOffset() {
return ((bitField0_ & 0x00000001) == 0x00000001); return ((bitField0_ & 0x00000001) == 0x00000001);
@ -691,8 +691,8 @@ public final class HFileProtos {
return fileInfoOffset_; return fileInfoOffset_;
} }
// optional uint64 loadOnOpenDataOffset = 2; // optional uint64 load_on_open_data_offset = 2;
public static final int LOADONOPENDATAOFFSET_FIELD_NUMBER = 2; public static final int LOAD_ON_OPEN_DATA_OFFSET_FIELD_NUMBER = 2;
private long loadOnOpenDataOffset_; private long loadOnOpenDataOffset_;
public boolean hasLoadOnOpenDataOffset() { public boolean hasLoadOnOpenDataOffset() {
return ((bitField0_ & 0x00000002) == 0x00000002); return ((bitField0_ & 0x00000002) == 0x00000002);
@ -701,8 +701,8 @@ public final class HFileProtos {
return loadOnOpenDataOffset_; return loadOnOpenDataOffset_;
} }
// optional uint64 uncompressedDataIndexSize = 3; // optional uint64 uncompressed_data_index_size = 3;
public static final int UNCOMPRESSEDDATAINDEXSIZE_FIELD_NUMBER = 3; public static final int UNCOMPRESSED_DATA_INDEX_SIZE_FIELD_NUMBER = 3;
private long uncompressedDataIndexSize_; private long uncompressedDataIndexSize_;
public boolean hasUncompressedDataIndexSize() { public boolean hasUncompressedDataIndexSize() {
return ((bitField0_ & 0x00000004) == 0x00000004); return ((bitField0_ & 0x00000004) == 0x00000004);
@ -711,8 +711,8 @@ public final class HFileProtos {
return uncompressedDataIndexSize_; return uncompressedDataIndexSize_;
} }
// optional uint64 totalUncompressedBytes = 4; // optional uint64 total_uncompressed_bytes = 4;
public static final int TOTALUNCOMPRESSEDBYTES_FIELD_NUMBER = 4; public static final int TOTAL_UNCOMPRESSED_BYTES_FIELD_NUMBER = 4;
private long totalUncompressedBytes_; private long totalUncompressedBytes_;
public boolean hasTotalUncompressedBytes() { public boolean hasTotalUncompressedBytes() {
return ((bitField0_ & 0x00000008) == 0x00000008); return ((bitField0_ & 0x00000008) == 0x00000008);
@ -721,8 +721,8 @@ public final class HFileProtos {
return totalUncompressedBytes_; return totalUncompressedBytes_;
} }
// optional uint32 dataIndexCount = 5; // optional uint32 data_index_count = 5;
public static final int DATAINDEXCOUNT_FIELD_NUMBER = 5; public static final int DATA_INDEX_COUNT_FIELD_NUMBER = 5;
private int dataIndexCount_; private int dataIndexCount_;
public boolean hasDataIndexCount() { public boolean hasDataIndexCount() {
return ((bitField0_ & 0x00000010) == 0x00000010); return ((bitField0_ & 0x00000010) == 0x00000010);
@ -731,8 +731,8 @@ public final class HFileProtos {
return dataIndexCount_; return dataIndexCount_;
} }
// optional uint32 metaIndexCount = 6; // optional uint32 meta_index_count = 6;
public static final int METAINDEXCOUNT_FIELD_NUMBER = 6; public static final int META_INDEX_COUNT_FIELD_NUMBER = 6;
private int metaIndexCount_; private int metaIndexCount_;
public boolean hasMetaIndexCount() { public boolean hasMetaIndexCount() {
return ((bitField0_ & 0x00000020) == 0x00000020); return ((bitField0_ & 0x00000020) == 0x00000020);
@ -741,8 +741,8 @@ public final class HFileProtos {
return metaIndexCount_; return metaIndexCount_;
} }
// optional uint64 entryCount = 7; // optional uint64 entry_count = 7;
public static final int ENTRYCOUNT_FIELD_NUMBER = 7; public static final int ENTRY_COUNT_FIELD_NUMBER = 7;
private long entryCount_; private long entryCount_;
public boolean hasEntryCount() { public boolean hasEntryCount() {
return ((bitField0_ & 0x00000040) == 0x00000040); return ((bitField0_ & 0x00000040) == 0x00000040);
@ -751,8 +751,8 @@ public final class HFileProtos {
return entryCount_; return entryCount_;
} }
// optional uint32 numDataIndexLevels = 8; // optional uint32 num_data_index_levels = 8;
public static final int NUMDATAINDEXLEVELS_FIELD_NUMBER = 8; public static final int NUM_DATA_INDEX_LEVELS_FIELD_NUMBER = 8;
private int numDataIndexLevels_; private int numDataIndexLevels_;
public boolean hasNumDataIndexLevels() { public boolean hasNumDataIndexLevels() {
return ((bitField0_ & 0x00000080) == 0x00000080); return ((bitField0_ & 0x00000080) == 0x00000080);
@ -761,8 +761,8 @@ public final class HFileProtos {
return numDataIndexLevels_; return numDataIndexLevels_;
} }
// optional uint64 firstDataBlockOffset = 9; // optional uint64 first_data_block_offset = 9;
public static final int FIRSTDATABLOCKOFFSET_FIELD_NUMBER = 9; public static final int FIRST_DATA_BLOCK_OFFSET_FIELD_NUMBER = 9;
private long firstDataBlockOffset_; private long firstDataBlockOffset_;
public boolean hasFirstDataBlockOffset() { public boolean hasFirstDataBlockOffset() {
return ((bitField0_ & 0x00000100) == 0x00000100); return ((bitField0_ & 0x00000100) == 0x00000100);
@ -771,8 +771,8 @@ public final class HFileProtos {
return firstDataBlockOffset_; return firstDataBlockOffset_;
} }
// optional uint64 lastDataBlockOffset = 10; // optional uint64 last_data_block_offset = 10;
public static final int LASTDATABLOCKOFFSET_FIELD_NUMBER = 10; public static final int LAST_DATA_BLOCK_OFFSET_FIELD_NUMBER = 10;
private long lastDataBlockOffset_; private long lastDataBlockOffset_;
public boolean hasLastDataBlockOffset() { public boolean hasLastDataBlockOffset() {
return ((bitField0_ & 0x00000200) == 0x00000200); return ((bitField0_ & 0x00000200) == 0x00000200);
@ -781,8 +781,8 @@ public final class HFileProtos {
return lastDataBlockOffset_; return lastDataBlockOffset_;
} }
// optional string comparatorClassName = 11; // optional string comparator_class_name = 11;
public static final int COMPARATORCLASSNAME_FIELD_NUMBER = 11; public static final int COMPARATOR_CLASS_NAME_FIELD_NUMBER = 11;
private java.lang.Object comparatorClassName_; private java.lang.Object comparatorClassName_;
public boolean hasComparatorClassName() { public boolean hasComparatorClassName() {
return ((bitField0_ & 0x00000400) == 0x00000400); return ((bitField0_ & 0x00000400) == 0x00000400);
@ -813,8 +813,8 @@ public final class HFileProtos {
} }
} }
// optional uint32 compressionCodec = 12; // optional uint32 compression_codec = 12;
public static final int COMPRESSIONCODEC_FIELD_NUMBER = 12; public static final int COMPRESSION_CODEC_FIELD_NUMBER = 12;
private int compressionCodec_; private int compressionCodec_;
public boolean hasCompressionCodec() { public boolean hasCompressionCodec() {
return ((bitField0_ & 0x00000800) == 0x00000800); return ((bitField0_ & 0x00000800) == 0x00000800);
@ -1035,51 +1035,51 @@ public final class HFileProtos {
int hash = 41; int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode(); hash = (19 * hash) + getDescriptorForType().hashCode();
if (hasFileInfoOffset()) { if (hasFileInfoOffset()) {
hash = (37 * hash) + FILEINFOOFFSET_FIELD_NUMBER; hash = (37 * hash) + FILE_INFO_OFFSET_FIELD_NUMBER;
hash = (53 * hash) + hashLong(getFileInfoOffset()); hash = (53 * hash) + hashLong(getFileInfoOffset());
} }
if (hasLoadOnOpenDataOffset()) { if (hasLoadOnOpenDataOffset()) {
hash = (37 * hash) + LOADONOPENDATAOFFSET_FIELD_NUMBER; hash = (37 * hash) + LOAD_ON_OPEN_DATA_OFFSET_FIELD_NUMBER;
hash = (53 * hash) + hashLong(getLoadOnOpenDataOffset()); hash = (53 * hash) + hashLong(getLoadOnOpenDataOffset());
} }
if (hasUncompressedDataIndexSize()) { if (hasUncompressedDataIndexSize()) {
hash = (37 * hash) + UNCOMPRESSEDDATAINDEXSIZE_FIELD_NUMBER; hash = (37 * hash) + UNCOMPRESSED_DATA_INDEX_SIZE_FIELD_NUMBER;
hash = (53 * hash) + hashLong(getUncompressedDataIndexSize()); hash = (53 * hash) + hashLong(getUncompressedDataIndexSize());
} }
if (hasTotalUncompressedBytes()) { if (hasTotalUncompressedBytes()) {
hash = (37 * hash) + TOTALUNCOMPRESSEDBYTES_FIELD_NUMBER; hash = (37 * hash) + TOTAL_UNCOMPRESSED_BYTES_FIELD_NUMBER;
hash = (53 * hash) + hashLong(getTotalUncompressedBytes()); hash = (53 * hash) + hashLong(getTotalUncompressedBytes());
} }
if (hasDataIndexCount()) { if (hasDataIndexCount()) {
hash = (37 * hash) + DATAINDEXCOUNT_FIELD_NUMBER; hash = (37 * hash) + DATA_INDEX_COUNT_FIELD_NUMBER;
hash = (53 * hash) + getDataIndexCount(); hash = (53 * hash) + getDataIndexCount();
} }
if (hasMetaIndexCount()) { if (hasMetaIndexCount()) {
hash = (37 * hash) + METAINDEXCOUNT_FIELD_NUMBER; hash = (37 * hash) + META_INDEX_COUNT_FIELD_NUMBER;
hash = (53 * hash) + getMetaIndexCount(); hash = (53 * hash) + getMetaIndexCount();
} }
if (hasEntryCount()) { if (hasEntryCount()) {
hash = (37 * hash) + ENTRYCOUNT_FIELD_NUMBER; hash = (37 * hash) + ENTRY_COUNT_FIELD_NUMBER;
hash = (53 * hash) + hashLong(getEntryCount()); hash = (53 * hash) + hashLong(getEntryCount());
} }
if (hasNumDataIndexLevels()) { if (hasNumDataIndexLevels()) {
hash = (37 * hash) + NUMDATAINDEXLEVELS_FIELD_NUMBER; hash = (37 * hash) + NUM_DATA_INDEX_LEVELS_FIELD_NUMBER;
hash = (53 * hash) + getNumDataIndexLevels(); hash = (53 * hash) + getNumDataIndexLevels();
} }
if (hasFirstDataBlockOffset()) { if (hasFirstDataBlockOffset()) {
hash = (37 * hash) + FIRSTDATABLOCKOFFSET_FIELD_NUMBER; hash = (37 * hash) + FIRST_DATA_BLOCK_OFFSET_FIELD_NUMBER;
hash = (53 * hash) + hashLong(getFirstDataBlockOffset()); hash = (53 * hash) + hashLong(getFirstDataBlockOffset());
} }
if (hasLastDataBlockOffset()) { if (hasLastDataBlockOffset()) {
hash = (37 * hash) + LASTDATABLOCKOFFSET_FIELD_NUMBER; hash = (37 * hash) + LAST_DATA_BLOCK_OFFSET_FIELD_NUMBER;
hash = (53 * hash) + hashLong(getLastDataBlockOffset()); hash = (53 * hash) + hashLong(getLastDataBlockOffset());
} }
if (hasComparatorClassName()) { if (hasComparatorClassName()) {
hash = (37 * hash) + COMPARATORCLASSNAME_FIELD_NUMBER; hash = (37 * hash) + COMPARATOR_CLASS_NAME_FIELD_NUMBER;
hash = (53 * hash) + getComparatorClassName().hashCode(); hash = (53 * hash) + getComparatorClassName().hashCode();
} }
if (hasCompressionCodec()) { if (hasCompressionCodec()) {
hash = (37 * hash) + COMPRESSIONCODEC_FIELD_NUMBER; hash = (37 * hash) + COMPRESSION_CODEC_FIELD_NUMBER;
hash = (53 * hash) + getCompressionCodec(); hash = (53 * hash) + getCompressionCodec();
} }
hash = (29 * hash) + getUnknownFields().hashCode(); hash = (29 * hash) + getUnknownFields().hashCode();
@ -1457,7 +1457,7 @@ public final class HFileProtos {
private int bitField0_; private int bitField0_;
// optional uint64 fileInfoOffset = 1; // optional uint64 file_info_offset = 1;
private long fileInfoOffset_ ; private long fileInfoOffset_ ;
public boolean hasFileInfoOffset() { public boolean hasFileInfoOffset() {
return ((bitField0_ & 0x00000001) == 0x00000001); return ((bitField0_ & 0x00000001) == 0x00000001);
@ -1478,7 +1478,7 @@ public final class HFileProtos {
return this; return this;
} }
// optional uint64 loadOnOpenDataOffset = 2; // optional uint64 load_on_open_data_offset = 2;
private long loadOnOpenDataOffset_ ; private long loadOnOpenDataOffset_ ;
public boolean hasLoadOnOpenDataOffset() { public boolean hasLoadOnOpenDataOffset() {
return ((bitField0_ & 0x00000002) == 0x00000002); return ((bitField0_ & 0x00000002) == 0x00000002);
@ -1499,7 +1499,7 @@ public final class HFileProtos {
return this; return this;
} }
// optional uint64 uncompressedDataIndexSize = 3; // optional uint64 uncompressed_data_index_size = 3;
private long uncompressedDataIndexSize_ ; private long uncompressedDataIndexSize_ ;
public boolean hasUncompressedDataIndexSize() { public boolean hasUncompressedDataIndexSize() {
return ((bitField0_ & 0x00000004) == 0x00000004); return ((bitField0_ & 0x00000004) == 0x00000004);
@ -1520,7 +1520,7 @@ public final class HFileProtos {
return this; return this;
} }
// optional uint64 totalUncompressedBytes = 4; // optional uint64 total_uncompressed_bytes = 4;
private long totalUncompressedBytes_ ; private long totalUncompressedBytes_ ;
public boolean hasTotalUncompressedBytes() { public boolean hasTotalUncompressedBytes() {
return ((bitField0_ & 0x00000008) == 0x00000008); return ((bitField0_ & 0x00000008) == 0x00000008);
@ -1541,7 +1541,7 @@ public final class HFileProtos {
return this; return this;
} }
// optional uint32 dataIndexCount = 5; // optional uint32 data_index_count = 5;
private int dataIndexCount_ ; private int dataIndexCount_ ;
public boolean hasDataIndexCount() { public boolean hasDataIndexCount() {
return ((bitField0_ & 0x00000010) == 0x00000010); return ((bitField0_ & 0x00000010) == 0x00000010);
@ -1562,7 +1562,7 @@ public final class HFileProtos {
return this; return this;
} }
// optional uint32 metaIndexCount = 6; // optional uint32 meta_index_count = 6;
private int metaIndexCount_ ; private int metaIndexCount_ ;
public boolean hasMetaIndexCount() { public boolean hasMetaIndexCount() {
return ((bitField0_ & 0x00000020) == 0x00000020); return ((bitField0_ & 0x00000020) == 0x00000020);
@ -1583,7 +1583,7 @@ public final class HFileProtos {
return this; return this;
} }
// optional uint64 entryCount = 7; // optional uint64 entry_count = 7;
private long entryCount_ ; private long entryCount_ ;
public boolean hasEntryCount() { public boolean hasEntryCount() {
return ((bitField0_ & 0x00000040) == 0x00000040); return ((bitField0_ & 0x00000040) == 0x00000040);
@ -1604,7 +1604,7 @@ public final class HFileProtos {
return this; return this;
} }
// optional uint32 numDataIndexLevels = 8; // optional uint32 num_data_index_levels = 8;
private int numDataIndexLevels_ ; private int numDataIndexLevels_ ;
public boolean hasNumDataIndexLevels() { public boolean hasNumDataIndexLevels() {
return ((bitField0_ & 0x00000080) == 0x00000080); return ((bitField0_ & 0x00000080) == 0x00000080);
@ -1625,7 +1625,7 @@ public final class HFileProtos {
return this; return this;
} }
// optional uint64 firstDataBlockOffset = 9; // optional uint64 first_data_block_offset = 9;
private long firstDataBlockOffset_ ; private long firstDataBlockOffset_ ;
public boolean hasFirstDataBlockOffset() { public boolean hasFirstDataBlockOffset() {
return ((bitField0_ & 0x00000100) == 0x00000100); return ((bitField0_ & 0x00000100) == 0x00000100);
@ -1646,7 +1646,7 @@ public final class HFileProtos {
return this; return this;
} }
// optional uint64 lastDataBlockOffset = 10; // optional uint64 last_data_block_offset = 10;
private long lastDataBlockOffset_ ; private long lastDataBlockOffset_ ;
public boolean hasLastDataBlockOffset() { public boolean hasLastDataBlockOffset() {
return ((bitField0_ & 0x00000200) == 0x00000200); return ((bitField0_ & 0x00000200) == 0x00000200);
@ -1667,7 +1667,7 @@ public final class HFileProtos {
return this; return this;
} }
// optional string comparatorClassName = 11; // optional string comparator_class_name = 11;
private java.lang.Object comparatorClassName_ = ""; private java.lang.Object comparatorClassName_ = "";
public boolean hasComparatorClassName() { public boolean hasComparatorClassName() {
return ((bitField0_ & 0x00000400) == 0x00000400); return ((bitField0_ & 0x00000400) == 0x00000400);
@ -1703,7 +1703,7 @@ public final class HFileProtos {
onChanged(); onChanged();
} }
// optional uint32 compressionCodec = 12; // optional uint32 compression_codec = 12;
private int compressionCodec_ ; private int compressionCodec_ ;
public boolean hasCompressionCodec() { public boolean hasCompressionCodec() {
return ((bitField0_ & 0x00000800) == 0x00000800); return ((bitField0_ & 0x00000800) == 0x00000800);
@ -1754,19 +1754,20 @@ public final class HFileProtos {
descriptor; descriptor;
static { static {
java.lang.String[] descriptorData = { java.lang.String[] descriptorData = {
"\n\013HFile.proto\032\013hbase.proto\"2\n\rFileInfoPr" + "\n\013HFile.proto\032\013hbase.proto\"3\n\rFileInfoPr" +
"oto\022!\n\010mapEntry\030\001 \003(\0132\017.BytesBytesPair\"\335" + "oto\022\"\n\tmap_entry\030\001 \003(\0132\017.BytesBytesPair\"" +
"\002\n\020FileTrailerProto\022\026\n\016fileInfoOffset\030\001 " + "\371\002\n\020FileTrailerProto\022\030\n\020file_info_offset" +
"\001(\004\022\034\n\024loadOnOpenDataOffset\030\002 \001(\004\022!\n\031unc" + "\030\001 \001(\004\022 \n\030load_on_open_data_offset\030\002 \001(\004" +
"ompressedDataIndexSize\030\003 \001(\004\022\036\n\026totalUnc" + "\022$\n\034uncompressed_data_index_size\030\003 \001(\004\022 " +
"ompressedBytes\030\004 \001(\004\022\026\n\016dataIndexCount\030\005" + "\n\030total_uncompressed_bytes\030\004 \001(\004\022\030\n\020data" +
" \001(\r\022\026\n\016metaIndexCount\030\006 \001(\r\022\022\n\nentryCou" + "_index_count\030\005 \001(\r\022\030\n\020meta_index_count\030\006" +
"nt\030\007 \001(\004\022\032\n\022numDataIndexLevels\030\010 \001(\r\022\034\n\024" + " \001(\r\022\023\n\013entry_count\030\007 \001(\004\022\035\n\025num_data_in" +
"firstDataBlockOffset\030\t \001(\004\022\033\n\023lastDataBl" + "dex_levels\030\010 \001(\r\022\037\n\027first_data_block_off" +
"ockOffset\030\n \001(\004\022\033\n\023comparatorClassName\030\013", "set\030\t \001(\004\022\036\n\026last_data_block_offset\030\n \001(",
" \001(\t\022\030\n\020compressionCodec\030\014 \001(\rBA\n*org.ap" + "\004\022\035\n\025comparator_class_name\030\013 \001(\t\022\031\n\021comp" +
"ache.hadoop.hbase.protobuf.generatedB\013HF" + "ression_codec\030\014 \001(\rBA\n*org.apache.hadoop" +
"ileProtosH\001\210\001\001\240\001\001" ".hbase.protobuf.generatedB\013HFileProtosH\001" +
"\210\001\001\240\001\001"
}; };
com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner = com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() { new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() {

View File

@ -11,7 +11,7 @@ public final class LoadBalancerProtos {
public interface LoadBalancerStateOrBuilder public interface LoadBalancerStateOrBuilder
extends com.google.protobuf.MessageOrBuilder { extends com.google.protobuf.MessageOrBuilder {
// optional bool balancerOn = 1; // optional bool balancer_on = 1;
boolean hasBalancerOn(); boolean hasBalancerOn();
boolean getBalancerOn(); boolean getBalancerOn();
} }
@ -44,8 +44,8 @@ public final class LoadBalancerProtos {
} }
private int bitField0_; private int bitField0_;
// optional bool balancerOn = 1; // optional bool balancer_on = 1;
public static final int BALANCERON_FIELD_NUMBER = 1; public static final int BALANCER_ON_FIELD_NUMBER = 1;
private boolean balancerOn_; private boolean balancerOn_;
public boolean hasBalancerOn() { public boolean hasBalancerOn() {
return ((bitField0_ & 0x00000001) == 0x00000001); return ((bitField0_ & 0x00000001) == 0x00000001);
@ -123,7 +123,7 @@ public final class LoadBalancerProtos {
int hash = 41; int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode(); hash = (19 * hash) + getDescriptorForType().hashCode();
if (hasBalancerOn()) { if (hasBalancerOn()) {
hash = (37 * hash) + BALANCERON_FIELD_NUMBER; hash = (37 * hash) + BALANCER_ON_FIELD_NUMBER;
hash = (53 * hash) + hashBoolean(getBalancerOn()); hash = (53 * hash) + hashBoolean(getBalancerOn());
} }
hash = (29 * hash) + getUnknownFields().hashCode(); hash = (29 * hash) + getUnknownFields().hashCode();
@ -347,7 +347,7 @@ public final class LoadBalancerProtos {
private int bitField0_; private int bitField0_;
// optional bool balancerOn = 1; // optional bool balancer_on = 1;
private boolean balancerOn_ ; private boolean balancerOn_ ;
public boolean hasBalancerOn() { public boolean hasBalancerOn() {
return ((bitField0_ & 0x00000001) == 0x00000001); return ((bitField0_ & 0x00000001) == 0x00000001);
@ -393,10 +393,10 @@ public final class LoadBalancerProtos {
descriptor; descriptor;
static { static {
java.lang.String[] descriptorData = { java.lang.String[] descriptorData = {
"\n\022LoadBalancer.proto\"\'\n\021LoadBalancerStat" + "\n\022LoadBalancer.proto\"(\n\021LoadBalancerStat" +
"e\022\022\n\nbalancerOn\030\001 \001(\010BE\n*org.apache.hado" + "e\022\023\n\013balancer_on\030\001 \001(\010BE\n*org.apache.had" +
"op.hbase.protobuf.generatedB\022LoadBalance" + "oop.hbase.protobuf.generatedB\022LoadBalanc" +
"rProtosH\001\240\001\001" "erProtosH\001\240\001\001"
}; };
com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner = com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() { new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() {

View File

@ -11,11 +11,11 @@ public final class MasterAdminProtos {
public interface AddColumnRequestOrBuilder public interface AddColumnRequestOrBuilder
extends com.google.protobuf.MessageOrBuilder { extends com.google.protobuf.MessageOrBuilder {
// required bytes tableName = 1; // required bytes table_name = 1;
boolean hasTableName(); boolean hasTableName();
com.google.protobuf.ByteString getTableName(); com.google.protobuf.ByteString getTableName();
// required .ColumnFamilySchema columnFamilies = 2; // required .ColumnFamilySchema column_families = 2;
boolean hasColumnFamilies(); boolean hasColumnFamilies();
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema getColumnFamilies(); org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema getColumnFamilies();
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchemaOrBuilder getColumnFamiliesOrBuilder(); org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchemaOrBuilder getColumnFamiliesOrBuilder();
@ -49,8 +49,8 @@ public final class MasterAdminProtos {
} }
private int bitField0_; private int bitField0_;
// required bytes tableName = 1; // required bytes table_name = 1;
public static final int TABLENAME_FIELD_NUMBER = 1; public static final int TABLE_NAME_FIELD_NUMBER = 1;
private com.google.protobuf.ByteString tableName_; private com.google.protobuf.ByteString tableName_;
public boolean hasTableName() { public boolean hasTableName() {
return ((bitField0_ & 0x00000001) == 0x00000001); return ((bitField0_ & 0x00000001) == 0x00000001);
@ -59,8 +59,8 @@ public final class MasterAdminProtos {
return tableName_; return tableName_;
} }
// required .ColumnFamilySchema columnFamilies = 2; // required .ColumnFamilySchema column_families = 2;
public static final int COLUMNFAMILIES_FIELD_NUMBER = 2; public static final int COLUMN_FAMILIES_FIELD_NUMBER = 2;
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema columnFamilies_; private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema columnFamilies_;
public boolean hasColumnFamilies() { public boolean hasColumnFamilies() {
return ((bitField0_ & 0x00000002) == 0x00000002); return ((bitField0_ & 0x00000002) == 0x00000002);
@ -166,11 +166,11 @@ public final class MasterAdminProtos {
int hash = 41; int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode(); hash = (19 * hash) + getDescriptorForType().hashCode();
if (hasTableName()) { if (hasTableName()) {
hash = (37 * hash) + TABLENAME_FIELD_NUMBER; hash = (37 * hash) + TABLE_NAME_FIELD_NUMBER;
hash = (53 * hash) + getTableName().hashCode(); hash = (53 * hash) + getTableName().hashCode();
} }
if (hasColumnFamilies()) { if (hasColumnFamilies()) {
hash = (37 * hash) + COLUMNFAMILIES_FIELD_NUMBER; hash = (37 * hash) + COLUMN_FAMILIES_FIELD_NUMBER;
hash = (53 * hash) + getColumnFamilies().hashCode(); hash = (53 * hash) + getColumnFamilies().hashCode();
} }
hash = (29 * hash) + getUnknownFields().hashCode(); hash = (29 * hash) + getUnknownFields().hashCode();
@ -433,7 +433,7 @@ public final class MasterAdminProtos {
private int bitField0_; private int bitField0_;
// required bytes tableName = 1; // required bytes table_name = 1;
private com.google.protobuf.ByteString tableName_ = com.google.protobuf.ByteString.EMPTY; private com.google.protobuf.ByteString tableName_ = com.google.protobuf.ByteString.EMPTY;
public boolean hasTableName() { public boolean hasTableName() {
return ((bitField0_ & 0x00000001) == 0x00000001); return ((bitField0_ & 0x00000001) == 0x00000001);
@ -457,7 +457,7 @@ public final class MasterAdminProtos {
return this; return this;
} }
// required .ColumnFamilySchema columnFamilies = 2; // required .ColumnFamilySchema column_families = 2;
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema columnFamilies_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema.getDefaultInstance(); private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema columnFamilies_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema.getDefaultInstance();
private com.google.protobuf.SingleFieldBuilder< private com.google.protobuf.SingleFieldBuilder<
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchemaOrBuilder> columnFamiliesBuilder_; org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchemaOrBuilder> columnFamiliesBuilder_;
@ -861,11 +861,11 @@ public final class MasterAdminProtos {
public interface DeleteColumnRequestOrBuilder public interface DeleteColumnRequestOrBuilder
extends com.google.protobuf.MessageOrBuilder { extends com.google.protobuf.MessageOrBuilder {
// required bytes tableName = 1; // required bytes table_name = 1;
boolean hasTableName(); boolean hasTableName();
com.google.protobuf.ByteString getTableName(); com.google.protobuf.ByteString getTableName();
// required bytes columnName = 2; // required bytes column_name = 2;
boolean hasColumnName(); boolean hasColumnName();
com.google.protobuf.ByteString getColumnName(); com.google.protobuf.ByteString getColumnName();
} }
@ -898,8 +898,8 @@ public final class MasterAdminProtos {
} }
private int bitField0_; private int bitField0_;
// required bytes tableName = 1; // required bytes table_name = 1;
public static final int TABLENAME_FIELD_NUMBER = 1; public static final int TABLE_NAME_FIELD_NUMBER = 1;
private com.google.protobuf.ByteString tableName_; private com.google.protobuf.ByteString tableName_;
public boolean hasTableName() { public boolean hasTableName() {
return ((bitField0_ & 0x00000001) == 0x00000001); return ((bitField0_ & 0x00000001) == 0x00000001);
@ -908,8 +908,8 @@ public final class MasterAdminProtos {
return tableName_; return tableName_;
} }
// required bytes columnName = 2; // required bytes column_name = 2;
public static final int COLUMNNAME_FIELD_NUMBER = 2; public static final int COLUMN_NAME_FIELD_NUMBER = 2;
private com.google.protobuf.ByteString columnName_; private com.google.protobuf.ByteString columnName_;
public boolean hasColumnName() { public boolean hasColumnName() {
return ((bitField0_ & 0x00000002) == 0x00000002); return ((bitField0_ & 0x00000002) == 0x00000002);
@ -1008,11 +1008,11 @@ public final class MasterAdminProtos {
int hash = 41; int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode(); hash = (19 * hash) + getDescriptorForType().hashCode();
if (hasTableName()) { if (hasTableName()) {
hash = (37 * hash) + TABLENAME_FIELD_NUMBER; hash = (37 * hash) + TABLE_NAME_FIELD_NUMBER;
hash = (53 * hash) + getTableName().hashCode(); hash = (53 * hash) + getTableName().hashCode();
} }
if (hasColumnName()) { if (hasColumnName()) {
hash = (37 * hash) + COLUMNNAME_FIELD_NUMBER; hash = (37 * hash) + COLUMN_NAME_FIELD_NUMBER;
hash = (53 * hash) + getColumnName().hashCode(); hash = (53 * hash) + getColumnName().hashCode();
} }
hash = (29 * hash) + getUnknownFields().hashCode(); hash = (29 * hash) + getUnknownFields().hashCode();
@ -1258,7 +1258,7 @@ public final class MasterAdminProtos {
private int bitField0_; private int bitField0_;
// required bytes tableName = 1; // required bytes table_name = 1;
private com.google.protobuf.ByteString tableName_ = com.google.protobuf.ByteString.EMPTY; private com.google.protobuf.ByteString tableName_ = com.google.protobuf.ByteString.EMPTY;
public boolean hasTableName() { public boolean hasTableName() {
return ((bitField0_ & 0x00000001) == 0x00000001); return ((bitField0_ & 0x00000001) == 0x00000001);
@ -1282,7 +1282,7 @@ public final class MasterAdminProtos {
return this; return this;
} }
// required bytes columnName = 2; // required bytes column_name = 2;
private com.google.protobuf.ByteString columnName_ = com.google.protobuf.ByteString.EMPTY; private com.google.protobuf.ByteString columnName_ = com.google.protobuf.ByteString.EMPTY;
public boolean hasColumnName() { public boolean hasColumnName() {
return ((bitField0_ & 0x00000002) == 0x00000002); return ((bitField0_ & 0x00000002) == 0x00000002);
@ -1620,11 +1620,11 @@ public final class MasterAdminProtos {
public interface ModifyColumnRequestOrBuilder public interface ModifyColumnRequestOrBuilder
extends com.google.protobuf.MessageOrBuilder { extends com.google.protobuf.MessageOrBuilder {
// required bytes tableName = 1; // required bytes table_name = 1;
boolean hasTableName(); boolean hasTableName();
com.google.protobuf.ByteString getTableName(); com.google.protobuf.ByteString getTableName();
// required .ColumnFamilySchema columnFamilies = 2; // required .ColumnFamilySchema column_families = 2;
boolean hasColumnFamilies(); boolean hasColumnFamilies();
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema getColumnFamilies(); org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema getColumnFamilies();
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchemaOrBuilder getColumnFamiliesOrBuilder(); org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchemaOrBuilder getColumnFamiliesOrBuilder();
@ -1658,8 +1658,8 @@ public final class MasterAdminProtos {
} }
private int bitField0_; private int bitField0_;
// required bytes tableName = 1; // required bytes table_name = 1;
public static final int TABLENAME_FIELD_NUMBER = 1; public static final int TABLE_NAME_FIELD_NUMBER = 1;
private com.google.protobuf.ByteString tableName_; private com.google.protobuf.ByteString tableName_;
public boolean hasTableName() { public boolean hasTableName() {
return ((bitField0_ & 0x00000001) == 0x00000001); return ((bitField0_ & 0x00000001) == 0x00000001);
@ -1668,8 +1668,8 @@ public final class MasterAdminProtos {
return tableName_; return tableName_;
} }
// required .ColumnFamilySchema columnFamilies = 2; // required .ColumnFamilySchema column_families = 2;
public static final int COLUMNFAMILIES_FIELD_NUMBER = 2; public static final int COLUMN_FAMILIES_FIELD_NUMBER = 2;
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema columnFamilies_; private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema columnFamilies_;
public boolean hasColumnFamilies() { public boolean hasColumnFamilies() {
return ((bitField0_ & 0x00000002) == 0x00000002); return ((bitField0_ & 0x00000002) == 0x00000002);
@ -1775,11 +1775,11 @@ public final class MasterAdminProtos {
int hash = 41; int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode(); hash = (19 * hash) + getDescriptorForType().hashCode();
if (hasTableName()) { if (hasTableName()) {
hash = (37 * hash) + TABLENAME_FIELD_NUMBER; hash = (37 * hash) + TABLE_NAME_FIELD_NUMBER;
hash = (53 * hash) + getTableName().hashCode(); hash = (53 * hash) + getTableName().hashCode();
} }
if (hasColumnFamilies()) { if (hasColumnFamilies()) {
hash = (37 * hash) + COLUMNFAMILIES_FIELD_NUMBER; hash = (37 * hash) + COLUMN_FAMILIES_FIELD_NUMBER;
hash = (53 * hash) + getColumnFamilies().hashCode(); hash = (53 * hash) + getColumnFamilies().hashCode();
} }
hash = (29 * hash) + getUnknownFields().hashCode(); hash = (29 * hash) + getUnknownFields().hashCode();
@ -2042,7 +2042,7 @@ public final class MasterAdminProtos {
private int bitField0_; private int bitField0_;
// required bytes tableName = 1; // required bytes table_name = 1;
private com.google.protobuf.ByteString tableName_ = com.google.protobuf.ByteString.EMPTY; private com.google.protobuf.ByteString tableName_ = com.google.protobuf.ByteString.EMPTY;
public boolean hasTableName() { public boolean hasTableName() {
return ((bitField0_ & 0x00000001) == 0x00000001); return ((bitField0_ & 0x00000001) == 0x00000001);
@ -2066,7 +2066,7 @@ public final class MasterAdminProtos {
return this; return this;
} }
// required .ColumnFamilySchema columnFamilies = 2; // required .ColumnFamilySchema column_families = 2;
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema columnFamilies_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema.getDefaultInstance(); private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema columnFamilies_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema.getDefaultInstance();
private com.google.protobuf.SingleFieldBuilder< private com.google.protobuf.SingleFieldBuilder<
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchemaOrBuilder> columnFamiliesBuilder_; org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchemaOrBuilder> columnFamiliesBuilder_;
@ -2475,7 +2475,7 @@ public final class MasterAdminProtos {
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier getRegion(); org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier getRegion();
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder getRegionOrBuilder(); org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder getRegionOrBuilder();
// optional .ServerName destServerName = 2; // optional .ServerName dest_server_name = 2;
boolean hasDestServerName(); boolean hasDestServerName();
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName getDestServerName(); org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName getDestServerName();
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerNameOrBuilder getDestServerNameOrBuilder(); org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerNameOrBuilder getDestServerNameOrBuilder();
@ -2522,8 +2522,8 @@ public final class MasterAdminProtos {
return region_; return region_;
} }
// optional .ServerName destServerName = 2; // optional .ServerName dest_server_name = 2;
public static final int DESTSERVERNAME_FIELD_NUMBER = 2; public static final int DEST_SERVER_NAME_FIELD_NUMBER = 2;
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName destServerName_; private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName destServerName_;
public boolean hasDestServerName() { public boolean hasDestServerName() {
return ((bitField0_ & 0x00000002) == 0x00000002); return ((bitField0_ & 0x00000002) == 0x00000002);
@ -2635,7 +2635,7 @@ public final class MasterAdminProtos {
hash = (53 * hash) + getRegion().hashCode(); hash = (53 * hash) + getRegion().hashCode();
} }
if (hasDestServerName()) { if (hasDestServerName()) {
hash = (37 * hash) + DESTSERVERNAME_FIELD_NUMBER; hash = (37 * hash) + DEST_SERVER_NAME_FIELD_NUMBER;
hash = (53 * hash) + getDestServerName().hashCode(); hash = (53 * hash) + getDestServerName().hashCode();
} }
hash = (29 * hash) + getUnknownFields().hashCode(); hash = (29 * hash) + getUnknownFields().hashCode();
@ -3003,7 +3003,7 @@ public final class MasterAdminProtos {
return regionBuilder_; return regionBuilder_;
} }
// optional .ServerName destServerName = 2; // optional .ServerName dest_server_name = 2;
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName destServerName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.getDefaultInstance(); private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName destServerName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.getDefaultInstance();
private com.google.protobuf.SingleFieldBuilder< private com.google.protobuf.SingleFieldBuilder<
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerNameOrBuilder> destServerNameBuilder_; org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerNameOrBuilder> destServerNameBuilder_;
@ -3407,12 +3407,12 @@ public final class MasterAdminProtos {
public interface DispatchMergingRegionsRequestOrBuilder public interface DispatchMergingRegionsRequestOrBuilder
extends com.google.protobuf.MessageOrBuilder { extends com.google.protobuf.MessageOrBuilder {
// required .RegionSpecifier regionA = 1; // required .RegionSpecifier region_a = 1;
boolean hasRegionA(); boolean hasRegionA();
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier getRegionA(); org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier getRegionA();
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder getRegionAOrBuilder(); org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder getRegionAOrBuilder();
// required .RegionSpecifier regionB = 2; // required .RegionSpecifier region_b = 2;
boolean hasRegionB(); boolean hasRegionB();
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier getRegionB(); org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier getRegionB();
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder getRegionBOrBuilder(); org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder getRegionBOrBuilder();
@ -3450,8 +3450,8 @@ public final class MasterAdminProtos {
} }
private int bitField0_; private int bitField0_;
// required .RegionSpecifier regionA = 1; // required .RegionSpecifier region_a = 1;
public static final int REGIONA_FIELD_NUMBER = 1; public static final int REGION_A_FIELD_NUMBER = 1;
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier regionA_; private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier regionA_;
public boolean hasRegionA() { public boolean hasRegionA() {
return ((bitField0_ & 0x00000001) == 0x00000001); return ((bitField0_ & 0x00000001) == 0x00000001);
@ -3463,8 +3463,8 @@ public final class MasterAdminProtos {
return regionA_; return regionA_;
} }
// required .RegionSpecifier regionB = 2; // required .RegionSpecifier region_b = 2;
public static final int REGIONB_FIELD_NUMBER = 2; public static final int REGION_B_FIELD_NUMBER = 2;
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier regionB_; private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier regionB_;
public boolean hasRegionB() { public boolean hasRegionB() {
return ((bitField0_ & 0x00000002) == 0x00000002); return ((bitField0_ & 0x00000002) == 0x00000002);
@ -3597,11 +3597,11 @@ public final class MasterAdminProtos {
int hash = 41; int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode(); hash = (19 * hash) + getDescriptorForType().hashCode();
if (hasRegionA()) { if (hasRegionA()) {
hash = (37 * hash) + REGIONA_FIELD_NUMBER; hash = (37 * hash) + REGION_A_FIELD_NUMBER;
hash = (53 * hash) + getRegionA().hashCode(); hash = (53 * hash) + getRegionA().hashCode();
} }
if (hasRegionB()) { if (hasRegionB()) {
hash = (37 * hash) + REGIONB_FIELD_NUMBER; hash = (37 * hash) + REGION_B_FIELD_NUMBER;
hash = (53 * hash) + getRegionB().hashCode(); hash = (53 * hash) + getRegionB().hashCode();
} }
if (hasForcible()) { if (hasForcible()) {
@ -3899,7 +3899,7 @@ public final class MasterAdminProtos {
private int bitField0_; private int bitField0_;
// required .RegionSpecifier regionA = 1; // required .RegionSpecifier region_a = 1;
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier regionA_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance(); private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier regionA_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance();
private com.google.protobuf.SingleFieldBuilder< private com.google.protobuf.SingleFieldBuilder<
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder> regionABuilder_; org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder> regionABuilder_;
@ -3989,7 +3989,7 @@ public final class MasterAdminProtos {
return regionABuilder_; return regionABuilder_;
} }
// required .RegionSpecifier regionB = 2; // required .RegionSpecifier region_b = 2;
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier regionB_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance(); private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier regionB_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance();
private com.google.protobuf.SingleFieldBuilder< private com.google.protobuf.SingleFieldBuilder<
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder> regionBBuilder_; org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder> regionBBuilder_;
@ -6799,12 +6799,12 @@ public final class MasterAdminProtos {
public interface CreateTableRequestOrBuilder public interface CreateTableRequestOrBuilder
extends com.google.protobuf.MessageOrBuilder { extends com.google.protobuf.MessageOrBuilder {
// required .TableSchema tableSchema = 1; // required .TableSchema table_schema = 1;
boolean hasTableSchema(); boolean hasTableSchema();
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema getTableSchema(); org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema getTableSchema();
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchemaOrBuilder getTableSchemaOrBuilder(); org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchemaOrBuilder getTableSchemaOrBuilder();
// repeated bytes splitKeys = 2; // repeated bytes split_keys = 2;
java.util.List<com.google.protobuf.ByteString> getSplitKeysList(); java.util.List<com.google.protobuf.ByteString> getSplitKeysList();
int getSplitKeysCount(); int getSplitKeysCount();
com.google.protobuf.ByteString getSplitKeys(int index); com.google.protobuf.ByteString getSplitKeys(int index);
@ -6838,8 +6838,8 @@ public final class MasterAdminProtos {
} }
private int bitField0_; private int bitField0_;
// required .TableSchema tableSchema = 1; // required .TableSchema table_schema = 1;
public static final int TABLESCHEMA_FIELD_NUMBER = 1; public static final int TABLE_SCHEMA_FIELD_NUMBER = 1;
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema tableSchema_; private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema tableSchema_;
public boolean hasTableSchema() { public boolean hasTableSchema() {
return ((bitField0_ & 0x00000001) == 0x00000001); return ((bitField0_ & 0x00000001) == 0x00000001);
@ -6851,8 +6851,8 @@ public final class MasterAdminProtos {
return tableSchema_; return tableSchema_;
} }
// repeated bytes splitKeys = 2; // repeated bytes split_keys = 2;
public static final int SPLITKEYS_FIELD_NUMBER = 2; public static final int SPLIT_KEYS_FIELD_NUMBER = 2;
private java.util.List<com.google.protobuf.ByteString> splitKeys_; private java.util.List<com.google.protobuf.ByteString> splitKeys_;
public java.util.List<com.google.protobuf.ByteString> public java.util.List<com.google.protobuf.ByteString>
getSplitKeysList() { getSplitKeysList() {
@ -6957,11 +6957,11 @@ public final class MasterAdminProtos {
int hash = 41; int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode(); hash = (19 * hash) + getDescriptorForType().hashCode();
if (hasTableSchema()) { if (hasTableSchema()) {
hash = (37 * hash) + TABLESCHEMA_FIELD_NUMBER; hash = (37 * hash) + TABLE_SCHEMA_FIELD_NUMBER;
hash = (53 * hash) + getTableSchema().hashCode(); hash = (53 * hash) + getTableSchema().hashCode();
} }
if (getSplitKeysCount() > 0) { if (getSplitKeysCount() > 0) {
hash = (37 * hash) + SPLITKEYS_FIELD_NUMBER; hash = (37 * hash) + SPLIT_KEYS_FIELD_NUMBER;
hash = (53 * hash) + getSplitKeysList().hashCode(); hash = (53 * hash) + getSplitKeysList().hashCode();
} }
hash = (29 * hash) + getUnknownFields().hashCode(); hash = (29 * hash) + getUnknownFields().hashCode();
@ -7228,7 +7228,7 @@ public final class MasterAdminProtos {
private int bitField0_; private int bitField0_;
// required .TableSchema tableSchema = 1; // required .TableSchema table_schema = 1;
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema tableSchema_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema.getDefaultInstance(); private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema tableSchema_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema.getDefaultInstance();
private com.google.protobuf.SingleFieldBuilder< private com.google.protobuf.SingleFieldBuilder<
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchemaOrBuilder> tableSchemaBuilder_; org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchemaOrBuilder> tableSchemaBuilder_;
@ -7318,7 +7318,7 @@ public final class MasterAdminProtos {
return tableSchemaBuilder_; return tableSchemaBuilder_;
} }
// repeated bytes splitKeys = 2; // repeated bytes split_keys = 2;
private java.util.List<com.google.protobuf.ByteString> splitKeys_ = java.util.Collections.emptyList();; private java.util.List<com.google.protobuf.ByteString> splitKeys_ = java.util.Collections.emptyList();;
private void ensureSplitKeysIsMutable() { private void ensureSplitKeysIsMutable() {
if (!((bitField0_ & 0x00000002) == 0x00000002)) { if (!((bitField0_ & 0x00000002) == 0x00000002)) {
@ -7683,7 +7683,7 @@ public final class MasterAdminProtos {
public interface DeleteTableRequestOrBuilder public interface DeleteTableRequestOrBuilder
extends com.google.protobuf.MessageOrBuilder { extends com.google.protobuf.MessageOrBuilder {
// required bytes tableName = 1; // required bytes table_name = 1;
boolean hasTableName(); boolean hasTableName();
com.google.protobuf.ByteString getTableName(); com.google.protobuf.ByteString getTableName();
} }
@ -7716,8 +7716,8 @@ public final class MasterAdminProtos {
} }
private int bitField0_; private int bitField0_;
// required bytes tableName = 1; // required bytes table_name = 1;
public static final int TABLENAME_FIELD_NUMBER = 1; public static final int TABLE_NAME_FIELD_NUMBER = 1;
private com.google.protobuf.ByteString tableName_; private com.google.protobuf.ByteString tableName_;
public boolean hasTableName() { public boolean hasTableName() {
return ((bitField0_ & 0x00000001) == 0x00000001); return ((bitField0_ & 0x00000001) == 0x00000001);
@ -7799,7 +7799,7 @@ public final class MasterAdminProtos {
int hash = 41; int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode(); hash = (19 * hash) + getDescriptorForType().hashCode();
if (hasTableName()) { if (hasTableName()) {
hash = (37 * hash) + TABLENAME_FIELD_NUMBER; hash = (37 * hash) + TABLE_NAME_FIELD_NUMBER;
hash = (53 * hash) + getTableName().hashCode(); hash = (53 * hash) + getTableName().hashCode();
} }
hash = (29 * hash) + getUnknownFields().hashCode(); hash = (29 * hash) + getUnknownFields().hashCode();
@ -8027,7 +8027,7 @@ public final class MasterAdminProtos {
private int bitField0_; private int bitField0_;
// required bytes tableName = 1; // required bytes table_name = 1;
private com.google.protobuf.ByteString tableName_ = com.google.protobuf.ByteString.EMPTY; private com.google.protobuf.ByteString tableName_ = com.google.protobuf.ByteString.EMPTY;
public boolean hasTableName() { public boolean hasTableName() {
return ((bitField0_ & 0x00000001) == 0x00000001); return ((bitField0_ & 0x00000001) == 0x00000001);
@ -8365,7 +8365,7 @@ public final class MasterAdminProtos {
public interface EnableTableRequestOrBuilder public interface EnableTableRequestOrBuilder
extends com.google.protobuf.MessageOrBuilder { extends com.google.protobuf.MessageOrBuilder {
// required bytes tableName = 1; // required bytes table_name = 1;
boolean hasTableName(); boolean hasTableName();
com.google.protobuf.ByteString getTableName(); com.google.protobuf.ByteString getTableName();
} }
@ -8398,8 +8398,8 @@ public final class MasterAdminProtos {
} }
private int bitField0_; private int bitField0_;
// required bytes tableName = 1; // required bytes table_name = 1;
public static final int TABLENAME_FIELD_NUMBER = 1; public static final int TABLE_NAME_FIELD_NUMBER = 1;
private com.google.protobuf.ByteString tableName_; private com.google.protobuf.ByteString tableName_;
public boolean hasTableName() { public boolean hasTableName() {
return ((bitField0_ & 0x00000001) == 0x00000001); return ((bitField0_ & 0x00000001) == 0x00000001);
@ -8481,7 +8481,7 @@ public final class MasterAdminProtos {
int hash = 41; int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode(); hash = (19 * hash) + getDescriptorForType().hashCode();
if (hasTableName()) { if (hasTableName()) {
hash = (37 * hash) + TABLENAME_FIELD_NUMBER; hash = (37 * hash) + TABLE_NAME_FIELD_NUMBER;
hash = (53 * hash) + getTableName().hashCode(); hash = (53 * hash) + getTableName().hashCode();
} }
hash = (29 * hash) + getUnknownFields().hashCode(); hash = (29 * hash) + getUnknownFields().hashCode();
@ -8709,7 +8709,7 @@ public final class MasterAdminProtos {
private int bitField0_; private int bitField0_;
// required bytes tableName = 1; // required bytes table_name = 1;
private com.google.protobuf.ByteString tableName_ = com.google.protobuf.ByteString.EMPTY; private com.google.protobuf.ByteString tableName_ = com.google.protobuf.ByteString.EMPTY;
public boolean hasTableName() { public boolean hasTableName() {
return ((bitField0_ & 0x00000001) == 0x00000001); return ((bitField0_ & 0x00000001) == 0x00000001);
@ -9047,7 +9047,7 @@ public final class MasterAdminProtos {
public interface DisableTableRequestOrBuilder public interface DisableTableRequestOrBuilder
extends com.google.protobuf.MessageOrBuilder { extends com.google.protobuf.MessageOrBuilder {
// required bytes tableName = 1; // required bytes table_name = 1;
boolean hasTableName(); boolean hasTableName();
com.google.protobuf.ByteString getTableName(); com.google.protobuf.ByteString getTableName();
} }
@ -9080,8 +9080,8 @@ public final class MasterAdminProtos {
} }
private int bitField0_; private int bitField0_;
// required bytes tableName = 1; // required bytes table_name = 1;
public static final int TABLENAME_FIELD_NUMBER = 1; public static final int TABLE_NAME_FIELD_NUMBER = 1;
private com.google.protobuf.ByteString tableName_; private com.google.protobuf.ByteString tableName_;
public boolean hasTableName() { public boolean hasTableName() {
return ((bitField0_ & 0x00000001) == 0x00000001); return ((bitField0_ & 0x00000001) == 0x00000001);
@ -9163,7 +9163,7 @@ public final class MasterAdminProtos {
int hash = 41; int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode(); hash = (19 * hash) + getDescriptorForType().hashCode();
if (hasTableName()) { if (hasTableName()) {
hash = (37 * hash) + TABLENAME_FIELD_NUMBER; hash = (37 * hash) + TABLE_NAME_FIELD_NUMBER;
hash = (53 * hash) + getTableName().hashCode(); hash = (53 * hash) + getTableName().hashCode();
} }
hash = (29 * hash) + getUnknownFields().hashCode(); hash = (29 * hash) + getUnknownFields().hashCode();
@ -9391,7 +9391,7 @@ public final class MasterAdminProtos {
private int bitField0_; private int bitField0_;
// required bytes tableName = 1; // required bytes table_name = 1;
private com.google.protobuf.ByteString tableName_ = com.google.protobuf.ByteString.EMPTY; private com.google.protobuf.ByteString tableName_ = com.google.protobuf.ByteString.EMPTY;
public boolean hasTableName() { public boolean hasTableName() {
return ((bitField0_ & 0x00000001) == 0x00000001); return ((bitField0_ & 0x00000001) == 0x00000001);
@ -9729,11 +9729,11 @@ public final class MasterAdminProtos {
public interface ModifyTableRequestOrBuilder public interface ModifyTableRequestOrBuilder
extends com.google.protobuf.MessageOrBuilder { extends com.google.protobuf.MessageOrBuilder {
// required bytes tableName = 1; // required bytes table_name = 1;
boolean hasTableName(); boolean hasTableName();
com.google.protobuf.ByteString getTableName(); com.google.protobuf.ByteString getTableName();
// required .TableSchema tableSchema = 2; // required .TableSchema table_schema = 2;
boolean hasTableSchema(); boolean hasTableSchema();
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema getTableSchema(); org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema getTableSchema();
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchemaOrBuilder getTableSchemaOrBuilder(); org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchemaOrBuilder getTableSchemaOrBuilder();
@ -9767,8 +9767,8 @@ public final class MasterAdminProtos {
} }
private int bitField0_; private int bitField0_;
// required bytes tableName = 1; // required bytes table_name = 1;
public static final int TABLENAME_FIELD_NUMBER = 1; public static final int TABLE_NAME_FIELD_NUMBER = 1;
private com.google.protobuf.ByteString tableName_; private com.google.protobuf.ByteString tableName_;
public boolean hasTableName() { public boolean hasTableName() {
return ((bitField0_ & 0x00000001) == 0x00000001); return ((bitField0_ & 0x00000001) == 0x00000001);
@ -9777,8 +9777,8 @@ public final class MasterAdminProtos {
return tableName_; return tableName_;
} }
// required .TableSchema tableSchema = 2; // required .TableSchema table_schema = 2;
public static final int TABLESCHEMA_FIELD_NUMBER = 2; public static final int TABLE_SCHEMA_FIELD_NUMBER = 2;
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema tableSchema_; private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema tableSchema_;
public boolean hasTableSchema() { public boolean hasTableSchema() {
return ((bitField0_ & 0x00000002) == 0x00000002); return ((bitField0_ & 0x00000002) == 0x00000002);
@ -9884,11 +9884,11 @@ public final class MasterAdminProtos {
int hash = 41; int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode(); hash = (19 * hash) + getDescriptorForType().hashCode();
if (hasTableName()) { if (hasTableName()) {
hash = (37 * hash) + TABLENAME_FIELD_NUMBER; hash = (37 * hash) + TABLE_NAME_FIELD_NUMBER;
hash = (53 * hash) + getTableName().hashCode(); hash = (53 * hash) + getTableName().hashCode();
} }
if (hasTableSchema()) { if (hasTableSchema()) {
hash = (37 * hash) + TABLESCHEMA_FIELD_NUMBER; hash = (37 * hash) + TABLE_SCHEMA_FIELD_NUMBER;
hash = (53 * hash) + getTableSchema().hashCode(); hash = (53 * hash) + getTableSchema().hashCode();
} }
hash = (29 * hash) + getUnknownFields().hashCode(); hash = (29 * hash) + getUnknownFields().hashCode();
@ -10151,7 +10151,7 @@ public final class MasterAdminProtos {
private int bitField0_; private int bitField0_;
// required bytes tableName = 1; // required bytes table_name = 1;
private com.google.protobuf.ByteString tableName_ = com.google.protobuf.ByteString.EMPTY; private com.google.protobuf.ByteString tableName_ = com.google.protobuf.ByteString.EMPTY;
public boolean hasTableName() { public boolean hasTableName() {
return ((bitField0_ & 0x00000001) == 0x00000001); return ((bitField0_ & 0x00000001) == 0x00000001);
@ -10175,7 +10175,7 @@ public final class MasterAdminProtos {
return this; return this;
} }
// required .TableSchema tableSchema = 2; // required .TableSchema table_schema = 2;
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema tableSchema_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema.getDefaultInstance(); private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema tableSchema_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema.getDefaultInstance();
private com.google.protobuf.SingleFieldBuilder< private com.google.protobuf.SingleFieldBuilder<
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchemaOrBuilder> tableSchemaBuilder_; org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchemaOrBuilder> tableSchemaBuilder_;
@ -12079,7 +12079,7 @@ public final class MasterAdminProtos {
public interface BalanceResponseOrBuilder public interface BalanceResponseOrBuilder
extends com.google.protobuf.MessageOrBuilder { extends com.google.protobuf.MessageOrBuilder {
// required bool balancerRan = 1; // required bool balancer_ran = 1;
boolean hasBalancerRan(); boolean hasBalancerRan();
boolean getBalancerRan(); boolean getBalancerRan();
} }
@ -12112,8 +12112,8 @@ public final class MasterAdminProtos {
} }
private int bitField0_; private int bitField0_;
// required bool balancerRan = 1; // required bool balancer_ran = 1;
public static final int BALANCERRAN_FIELD_NUMBER = 1; public static final int BALANCER_RAN_FIELD_NUMBER = 1;
private boolean balancerRan_; private boolean balancerRan_;
public boolean hasBalancerRan() { public boolean hasBalancerRan() {
return ((bitField0_ & 0x00000001) == 0x00000001); return ((bitField0_ & 0x00000001) == 0x00000001);
@ -12195,7 +12195,7 @@ public final class MasterAdminProtos {
int hash = 41; int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode(); hash = (19 * hash) + getDescriptorForType().hashCode();
if (hasBalancerRan()) { if (hasBalancerRan()) {
hash = (37 * hash) + BALANCERRAN_FIELD_NUMBER; hash = (37 * hash) + BALANCER_RAN_FIELD_NUMBER;
hash = (53 * hash) + hashBoolean(getBalancerRan()); hash = (53 * hash) + hashBoolean(getBalancerRan());
} }
hash = (29 * hash) + getUnknownFields().hashCode(); hash = (29 * hash) + getUnknownFields().hashCode();
@ -12423,7 +12423,7 @@ public final class MasterAdminProtos {
private int bitField0_; private int bitField0_;
// required bool balancerRan = 1; // required bool balancer_ran = 1;
private boolean balancerRan_ ; private boolean balancerRan_ ;
public boolean hasBalancerRan() { public boolean hasBalancerRan() {
return ((bitField0_ & 0x00000001) == 0x00000001); return ((bitField0_ & 0x00000001) == 0x00000001);
@ -12903,7 +12903,7 @@ public final class MasterAdminProtos {
public interface SetBalancerRunningResponseOrBuilder public interface SetBalancerRunningResponseOrBuilder
extends com.google.protobuf.MessageOrBuilder { extends com.google.protobuf.MessageOrBuilder {
// optional bool prevBalanceValue = 1; // optional bool prev_balance_value = 1;
boolean hasPrevBalanceValue(); boolean hasPrevBalanceValue();
boolean getPrevBalanceValue(); boolean getPrevBalanceValue();
} }
@ -12936,8 +12936,8 @@ public final class MasterAdminProtos {
} }
private int bitField0_; private int bitField0_;
// optional bool prevBalanceValue = 1; // optional bool prev_balance_value = 1;
public static final int PREVBALANCEVALUE_FIELD_NUMBER = 1; public static final int PREV_BALANCE_VALUE_FIELD_NUMBER = 1;
private boolean prevBalanceValue_; private boolean prevBalanceValue_;
public boolean hasPrevBalanceValue() { public boolean hasPrevBalanceValue() {
return ((bitField0_ & 0x00000001) == 0x00000001); return ((bitField0_ & 0x00000001) == 0x00000001);
@ -13015,7 +13015,7 @@ public final class MasterAdminProtos {
int hash = 41; int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode(); hash = (19 * hash) + getDescriptorForType().hashCode();
if (hasPrevBalanceValue()) { if (hasPrevBalanceValue()) {
hash = (37 * hash) + PREVBALANCEVALUE_FIELD_NUMBER; hash = (37 * hash) + PREV_BALANCE_VALUE_FIELD_NUMBER;
hash = (53 * hash) + hashBoolean(getPrevBalanceValue()); hash = (53 * hash) + hashBoolean(getPrevBalanceValue());
} }
hash = (29 * hash) + getUnknownFields().hashCode(); hash = (29 * hash) + getUnknownFields().hashCode();
@ -13239,7 +13239,7 @@ public final class MasterAdminProtos {
private int bitField0_; private int bitField0_;
// optional bool prevBalanceValue = 1; // optional bool prev_balance_value = 1;
private boolean prevBalanceValue_ ; private boolean prevBalanceValue_ ;
public boolean hasPrevBalanceValue() { public boolean hasPrevBalanceValue() {
return ((bitField0_ & 0x00000001) == 0x00000001); return ((bitField0_ & 0x00000001) == 0x00000001);
@ -13574,7 +13574,7 @@ public final class MasterAdminProtos {
public interface CatalogScanResponseOrBuilder public interface CatalogScanResponseOrBuilder
extends com.google.protobuf.MessageOrBuilder { extends com.google.protobuf.MessageOrBuilder {
// optional int32 scanResult = 1; // optional int32 scan_result = 1;
boolean hasScanResult(); boolean hasScanResult();
int getScanResult(); int getScanResult();
} }
@ -13607,8 +13607,8 @@ public final class MasterAdminProtos {
} }
private int bitField0_; private int bitField0_;
// optional int32 scanResult = 1; // optional int32 scan_result = 1;
public static final int SCANRESULT_FIELD_NUMBER = 1; public static final int SCAN_RESULT_FIELD_NUMBER = 1;
private int scanResult_; private int scanResult_;
public boolean hasScanResult() { public boolean hasScanResult() {
return ((bitField0_ & 0x00000001) == 0x00000001); return ((bitField0_ & 0x00000001) == 0x00000001);
@ -13686,7 +13686,7 @@ public final class MasterAdminProtos {
int hash = 41; int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode(); hash = (19 * hash) + getDescriptorForType().hashCode();
if (hasScanResult()) { if (hasScanResult()) {
hash = (37 * hash) + SCANRESULT_FIELD_NUMBER; hash = (37 * hash) + SCAN_RESULT_FIELD_NUMBER;
hash = (53 * hash) + getScanResult(); hash = (53 * hash) + getScanResult();
} }
hash = (29 * hash) + getUnknownFields().hashCode(); hash = (29 * hash) + getUnknownFields().hashCode();
@ -13910,7 +13910,7 @@ public final class MasterAdminProtos {
private int bitField0_; private int bitField0_;
// optional int32 scanResult = 1; // optional int32 scan_result = 1;
private int scanResult_ ; private int scanResult_ ;
public boolean hasScanResult() { public boolean hasScanResult() {
return ((bitField0_ & 0x00000001) == 0x00000001); return ((bitField0_ & 0x00000001) == 0x00000001);
@ -14324,7 +14324,7 @@ public final class MasterAdminProtos {
public interface EnableCatalogJanitorResponseOrBuilder public interface EnableCatalogJanitorResponseOrBuilder
extends com.google.protobuf.MessageOrBuilder { extends com.google.protobuf.MessageOrBuilder {
// optional bool prevValue = 1; // optional bool prev_value = 1;
boolean hasPrevValue(); boolean hasPrevValue();
boolean getPrevValue(); boolean getPrevValue();
} }
@ -14357,8 +14357,8 @@ public final class MasterAdminProtos {
} }
private int bitField0_; private int bitField0_;
// optional bool prevValue = 1; // optional bool prev_value = 1;
public static final int PREVVALUE_FIELD_NUMBER = 1; public static final int PREV_VALUE_FIELD_NUMBER = 1;
private boolean prevValue_; private boolean prevValue_;
public boolean hasPrevValue() { public boolean hasPrevValue() {
return ((bitField0_ & 0x00000001) == 0x00000001); return ((bitField0_ & 0x00000001) == 0x00000001);
@ -14436,7 +14436,7 @@ public final class MasterAdminProtos {
int hash = 41; int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode(); hash = (19 * hash) + getDescriptorForType().hashCode();
if (hasPrevValue()) { if (hasPrevValue()) {
hash = (37 * hash) + PREVVALUE_FIELD_NUMBER; hash = (37 * hash) + PREV_VALUE_FIELD_NUMBER;
hash = (53 * hash) + hashBoolean(getPrevValue()); hash = (53 * hash) + hashBoolean(getPrevValue());
} }
hash = (29 * hash) + getUnknownFields().hashCode(); hash = (29 * hash) + getUnknownFields().hashCode();
@ -14660,7 +14660,7 @@ public final class MasterAdminProtos {
private int bitField0_; private int bitField0_;
// optional bool prevValue = 1; // optional bool prev_value = 1;
private boolean prevValue_ ; private boolean prevValue_ ;
public boolean hasPrevValue() { public boolean hasPrevValue() {
return ((bitField0_ & 0x00000001) == 0x00000001); return ((bitField0_ & 0x00000001) == 0x00000001);
@ -15847,7 +15847,7 @@ public final class MasterAdminProtos {
public interface TakeSnapshotResponseOrBuilder public interface TakeSnapshotResponseOrBuilder
extends com.google.protobuf.MessageOrBuilder { extends com.google.protobuf.MessageOrBuilder {
// required int64 expectedTimeout = 1; // required int64 expected_timeout = 1;
boolean hasExpectedTimeout(); boolean hasExpectedTimeout();
long getExpectedTimeout(); long getExpectedTimeout();
} }
@ -15880,8 +15880,8 @@ public final class MasterAdminProtos {
} }
private int bitField0_; private int bitField0_;
// required int64 expectedTimeout = 1; // required int64 expected_timeout = 1;
public static final int EXPECTEDTIMEOUT_FIELD_NUMBER = 1; public static final int EXPECTED_TIMEOUT_FIELD_NUMBER = 1;
private long expectedTimeout_; private long expectedTimeout_;
public boolean hasExpectedTimeout() { public boolean hasExpectedTimeout() {
return ((bitField0_ & 0x00000001) == 0x00000001); return ((bitField0_ & 0x00000001) == 0x00000001);
@ -15963,7 +15963,7 @@ public final class MasterAdminProtos {
int hash = 41; int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode(); hash = (19 * hash) + getDescriptorForType().hashCode();
if (hasExpectedTimeout()) { if (hasExpectedTimeout()) {
hash = (37 * hash) + EXPECTEDTIMEOUT_FIELD_NUMBER; hash = (37 * hash) + EXPECTED_TIMEOUT_FIELD_NUMBER;
hash = (53 * hash) + hashLong(getExpectedTimeout()); hash = (53 * hash) + hashLong(getExpectedTimeout());
} }
hash = (29 * hash) + getUnknownFields().hashCode(); hash = (29 * hash) + getUnknownFields().hashCode();
@ -16191,7 +16191,7 @@ public final class MasterAdminProtos {
private int bitField0_; private int bitField0_;
// required int64 expectedTimeout = 1; // required int64 expected_timeout = 1;
private long expectedTimeout_ ; private long expectedTimeout_ ;
public boolean hasExpectedTimeout() { public boolean hasExpectedTimeout() {
return ((bitField0_ & 0x00000001) == 0x00000001); return ((bitField0_ & 0x00000001) == 0x00000001);
@ -22752,116 +22752,116 @@ public final class MasterAdminProtos {
static { static {
java.lang.String[] descriptorData = { java.lang.String[] descriptorData = {
"\n\021MasterAdmin.proto\032\014Master.proto\032\013hbase" + "\n\021MasterAdmin.proto\032\014Master.proto\032\013hbase" +
".proto\032\014Client.proto\"R\n\020AddColumnRequest" + ".proto\032\014Client.proto\"T\n\020AddColumnRequest" +
"\022\021\n\ttableName\030\001 \002(\014\022+\n\016columnFamilies\030\002 " + "\022\022\n\ntable_name\030\001 \002(\014\022,\n\017column_families\030" +
"\002(\0132\023.ColumnFamilySchema\"\023\n\021AddColumnRes" + "\002 \002(\0132\023.ColumnFamilySchema\"\023\n\021AddColumnR" +
"ponse\"<\n\023DeleteColumnRequest\022\021\n\ttableNam" + "esponse\">\n\023DeleteColumnRequest\022\022\n\ntable_" +
"e\030\001 \002(\014\022\022\n\ncolumnName\030\002 \002(\014\"\026\n\024DeleteCol" + "name\030\001 \002(\014\022\023\n\013column_name\030\002 \002(\014\"\026\n\024Delet" +
"umnResponse\"U\n\023ModifyColumnRequest\022\021\n\tta" + "eColumnResponse\"W\n\023ModifyColumnRequest\022\022" +
"bleName\030\001 \002(\014\022+\n\016columnFamilies\030\002 \002(\0132\023." + "\n\ntable_name\030\001 \002(\014\022,\n\017column_families\030\002 " +
"ColumnFamilySchema\"\026\n\024ModifyColumnRespon" + "\002(\0132\023.ColumnFamilySchema\"\026\n\024ModifyColumn" +
"se\"Z\n\021MoveRegionRequest\022 \n\006region\030\001 \002(\0132", "Response\"\\\n\021MoveRegionRequest\022 \n\006region\030",
"\020.RegionSpecifier\022#\n\016destServerName\030\002 \001(" + "\001 \002(\0132\020.RegionSpecifier\022%\n\020dest_server_n" +
"\0132\013.ServerName\"\024\n\022MoveRegionResponse\"~\n\035" + "ame\030\002 \001(\0132\013.ServerName\"\024\n\022MoveRegionResp" +
"DispatchMergingRegionsRequest\022!\n\007regionA" + "onse\"\200\001\n\035DispatchMergingRegionsRequest\022\"" +
"\030\001 \002(\0132\020.RegionSpecifier\022!\n\007regionB\030\002 \002(" + "\n\010region_a\030\001 \002(\0132\020.RegionSpecifier\022\"\n\010re" +
"\0132\020.RegionSpecifier\022\027\n\010forcible\030\003 \001(\010:\005f" + "gion_b\030\002 \002(\0132\020.RegionSpecifier\022\027\n\010forcib" +
"alse\" \n\036DispatchMergingRegionsResponse\"7" + "le\030\003 \001(\010:\005false\" \n\036DispatchMergingRegion" +
"\n\023AssignRegionRequest\022 \n\006region\030\001 \002(\0132\020." + "sResponse\"7\n\023AssignRegionRequest\022 \n\006regi" +
"RegionSpecifier\"\026\n\024AssignRegionResponse\"" + "on\030\001 \002(\0132\020.RegionSpecifier\"\026\n\024AssignRegi" +
"O\n\025UnassignRegionRequest\022 \n\006region\030\001 \002(\013" + "onResponse\"O\n\025UnassignRegionRequest\022 \n\006r" +
"2\020.RegionSpecifier\022\024\n\005force\030\002 \001(\010:\005false", "egion\030\001 \002(\0132\020.RegionSpecifier\022\024\n\005force\030\002",
"\"\030\n\026UnassignRegionResponse\"8\n\024OfflineReg" + " \001(\010:\005false\"\030\n\026UnassignRegionResponse\"8\n" +
"ionRequest\022 \n\006region\030\001 \002(\0132\020.RegionSpeci" + "\024OfflineRegionRequest\022 \n\006region\030\001 \002(\0132\020." +
"fier\"\027\n\025OfflineRegionResponse\"J\n\022CreateT" + "RegionSpecifier\"\027\n\025OfflineRegionResponse" +
"ableRequest\022!\n\013tableSchema\030\001 \002(\0132\014.Table" + "\"L\n\022CreateTableRequest\022\"\n\014table_schema\030\001" +
"Schema\022\021\n\tsplitKeys\030\002 \003(\014\"\025\n\023CreateTable" + " \002(\0132\014.TableSchema\022\022\n\nsplit_keys\030\002 \003(\014\"\025" +
"Response\"\'\n\022DeleteTableRequest\022\021\n\ttableN" + "\n\023CreateTableResponse\"(\n\022DeleteTableRequ" +
"ame\030\001 \002(\014\"\025\n\023DeleteTableResponse\"\'\n\022Enab" + "est\022\022\n\ntable_name\030\001 \002(\014\"\025\n\023DeleteTableRe" +
"leTableRequest\022\021\n\ttableName\030\001 \002(\014\"\025\n\023Ena" + "sponse\"(\n\022EnableTableRequest\022\022\n\ntable_na" +
"bleTableResponse\"(\n\023DisableTableRequest\022" + "me\030\001 \002(\014\"\025\n\023EnableTableResponse\")\n\023Disab" +
"\021\n\ttableName\030\001 \002(\014\"\026\n\024DisableTableRespon", "leTableRequest\022\022\n\ntable_name\030\001 \002(\014\"\026\n\024Di",
"se\"J\n\022ModifyTableRequest\022\021\n\ttableName\030\001 " + "sableTableResponse\"L\n\022ModifyTableRequest" +
"\002(\014\022!\n\013tableSchema\030\002 \002(\0132\014.TableSchema\"\025" + "\022\022\n\ntable_name\030\001 \002(\014\022\"\n\014table_schema\030\002 \002" +
"\n\023ModifyTableResponse\"\021\n\017ShutdownRequest" + "(\0132\014.TableSchema\"\025\n\023ModifyTableResponse\"" +
"\"\022\n\020ShutdownResponse\"\023\n\021StopMasterReques" + "\021\n\017ShutdownRequest\"\022\n\020ShutdownResponse\"\023" +
"t\"\024\n\022StopMasterResponse\"\020\n\016BalanceReques" + "\n\021StopMasterRequest\"\024\n\022StopMasterRespons" +
"t\"&\n\017BalanceResponse\022\023\n\013balancerRan\030\001 \002(" + "e\"\020\n\016BalanceRequest\"\'\n\017BalanceResponse\022\024" +
"\010\"<\n\031SetBalancerRunningRequest\022\n\n\002on\030\001 \002" + "\n\014balancer_ran\030\001 \002(\010\"<\n\031SetBalancerRunni" +
"(\010\022\023\n\013synchronous\030\002 \001(\010\"6\n\032SetBalancerRu" + "ngRequest\022\n\n\002on\030\001 \002(\010\022\023\n\013synchronous\030\002 \001" +
"nningResponse\022\030\n\020prevBalanceValue\030\001 \001(\010\"" + "(\010\"8\n\032SetBalancerRunningResponse\022\032\n\022prev" +
"\024\n\022CatalogScanRequest\")\n\023CatalogScanResp", "_balance_value\030\001 \001(\010\"\024\n\022CatalogScanReque",
"onse\022\022\n\nscanResult\030\001 \001(\005\"-\n\033EnableCatalo" + "st\"*\n\023CatalogScanResponse\022\023\n\013scan_result" +
"gJanitorRequest\022\016\n\006enable\030\001 \002(\010\"1\n\034Enabl" + "\030\001 \001(\005\"-\n\033EnableCatalogJanitorRequest\022\016\n" +
"eCatalogJanitorResponse\022\021\n\tprevValue\030\001 \001" + "\006enable\030\001 \002(\010\"2\n\034EnableCatalogJanitorRes" +
"(\010\" \n\036IsCatalogJanitorEnabledRequest\"0\n\037" + "ponse\022\022\n\nprev_value\030\001 \001(\010\" \n\036IsCatalogJa" +
"IsCatalogJanitorEnabledResponse\022\r\n\005value" + "nitorEnabledRequest\"0\n\037IsCatalogJanitorE" +
"\030\001 \002(\010\"=\n\023TakeSnapshotRequest\022&\n\010snapsho" + "nabledResponse\022\r\n\005value\030\001 \002(\010\"=\n\023TakeSna" +
"t\030\001 \002(\0132\024.SnapshotDescription\"/\n\024TakeSna" + "pshotRequest\022&\n\010snapshot\030\001 \002(\0132\024.Snapsho" +
"pshotResponse\022\027\n\017expectedTimeout\030\001 \002(\003\"\025" + "tDescription\"0\n\024TakeSnapshotResponse\022\030\n\020" +
"\n\023ListSnapshotRequest\"?\n\024ListSnapshotRes" + "expected_timeout\030\001 \002(\003\"\025\n\023ListSnapshotRe" +
"ponse\022\'\n\tsnapshots\030\001 \003(\0132\024.SnapshotDescr", "quest\"?\n\024ListSnapshotResponse\022\'\n\tsnapsho",
"iption\"?\n\025DeleteSnapshotRequest\022&\n\010snaps" + "ts\030\001 \003(\0132\024.SnapshotDescription\"?\n\025Delete" +
"hot\030\001 \002(\0132\024.SnapshotDescription\"\030\n\026Delet" + "SnapshotRequest\022&\n\010snapshot\030\001 \002(\0132\024.Snap" +
"eSnapshotResponse\"@\n\026RestoreSnapshotRequ" + "shotDescription\"\030\n\026DeleteSnapshotRespons" +
"est\022&\n\010snapshot\030\001 \002(\0132\024.SnapshotDescript" + "e\"@\n\026RestoreSnapshotRequest\022&\n\010snapshot\030" +
"ion\"\031\n\027RestoreSnapshotResponse\"?\n\025IsSnap" + "\001 \002(\0132\024.SnapshotDescription\"\031\n\027RestoreSn" +
"shotDoneRequest\022&\n\010snapshot\030\001 \001(\0132\024.Snap" + "apshotResponse\"?\n\025IsSnapshotDoneRequest\022" +
"shotDescription\"U\n\026IsSnapshotDoneRespons" + "&\n\010snapshot\030\001 \001(\0132\024.SnapshotDescription\"" +
"e\022\023\n\004done\030\001 \001(\010:\005false\022&\n\010snapshot\030\002 \001(\013" + "U\n\026IsSnapshotDoneResponse\022\023\n\004done\030\001 \001(\010:" +
"2\024.SnapshotDescription\"F\n\034IsRestoreSnaps" + "\005false\022&\n\010snapshot\030\002 \001(\0132\024.SnapshotDescr" +
"hotDoneRequest\022&\n\010snapshot\030\001 \001(\0132\024.Snaps", "iption\"F\n\034IsRestoreSnapshotDoneRequest\022&",
"hotDescription\"3\n\035IsRestoreSnapshotDoneR" + "\n\010snapshot\030\001 \001(\0132\024.SnapshotDescription\"3" +
"esponse\022\022\n\004done\030\001 \001(\010:\004true2\305\016\n\022MasterAd" + "\n\035IsRestoreSnapshotDoneResponse\022\022\n\004done\030" +
"minService\0222\n\taddColumn\022\021.AddColumnReque" + "\001 \001(\010:\004true2\305\016\n\022MasterAdminService\0222\n\tAd" +
"st\032\022.AddColumnResponse\022;\n\014deleteColumn\022\024" + "dColumn\022\021.AddColumnRequest\032\022.AddColumnRe" +
".DeleteColumnRequest\032\025.DeleteColumnRespo" + "sponse\022;\n\014DeleteColumn\022\024.DeleteColumnReq" +
"nse\022;\n\014modifyColumn\022\024.ModifyColumnReques" + "uest\032\025.DeleteColumnResponse\022;\n\014ModifyCol" +
"t\032\025.ModifyColumnResponse\0225\n\nmoveRegion\022\022" + "umn\022\024.ModifyColumnRequest\032\025.ModifyColumn" +
".MoveRegionRequest\032\023.MoveRegionResponse\022" + "Response\0225\n\nMoveRegion\022\022.MoveRegionReque" +
"Y\n\026dispatchMergingRegions\022\036.DispatchMerg" + "st\032\023.MoveRegionResponse\022Y\n\026DispatchMergi" +
"ingRegionsRequest\032\037.DispatchMergingRegio", "ngRegions\022\036.DispatchMergingRegionsReques",
"nsResponse\022;\n\014assignRegion\022\024.AssignRegio" + "t\032\037.DispatchMergingRegionsResponse\022;\n\014As" +
"nRequest\032\025.AssignRegionResponse\022A\n\016unass" + "signRegion\022\024.AssignRegionRequest\032\025.Assig" +
"ignRegion\022\026.UnassignRegionRequest\032\027.Unas" + "nRegionResponse\022A\n\016UnassignRegion\022\026.Unas" +
"signRegionResponse\022>\n\rofflineRegion\022\025.Of" + "signRegionRequest\032\027.UnassignRegionRespon" +
"flineRegionRequest\032\026.OfflineRegionRespon" + "se\022>\n\rOfflineRegion\022\025.OfflineRegionReque" +
"se\0228\n\013deleteTable\022\023.DeleteTableRequest\032\024" + "st\032\026.OfflineRegionResponse\0228\n\013DeleteTabl" +
".DeleteTableResponse\0228\n\013enableTable\022\023.En" + "e\022\023.DeleteTableRequest\032\024.DeleteTableResp" +
"ableTableRequest\032\024.EnableTableResponse\022;" + "onse\0228\n\013EnableTable\022\023.EnableTableRequest" +
"\n\014disableTable\022\024.DisableTableRequest\032\025.D" + "\032\024.EnableTableResponse\022;\n\014DisableTable\022\024" +
"isableTableResponse\0228\n\013modifyTable\022\023.Mod", ".DisableTableRequest\032\025.DisableTableRespo",
"ifyTableRequest\032\024.ModifyTableResponse\0228\n" + "nse\0228\n\013ModifyTable\022\023.ModifyTableRequest\032" +
"\013createTable\022\023.CreateTableRequest\032\024.Crea" + "\024.ModifyTableResponse\0228\n\013CreateTable\022\023.C" +
"teTableResponse\022/\n\010shutdown\022\020.ShutdownRe" + "reateTableRequest\032\024.CreateTableResponse\022" +
"quest\032\021.ShutdownResponse\0225\n\nstopMaster\022\022" + "/\n\010Shutdown\022\020.ShutdownRequest\032\021.Shutdown" +
".StopMasterRequest\032\023.StopMasterResponse\022" + "Response\0225\n\nStopMaster\022\022.StopMasterReque" +
",\n\007balance\022\017.BalanceRequest\032\020.BalanceRes" + "st\032\023.StopMasterResponse\022,\n\007Balance\022\017.Bal" +
"ponse\022M\n\022setBalancerRunning\022\032.SetBalance" + "anceRequest\032\020.BalanceResponse\022M\n\022SetBala" +
"rRunningRequest\032\033.SetBalancerRunningResp" + "ncerRunning\022\032.SetBalancerRunningRequest\032" +
"onse\022;\n\016runCatalogScan\022\023.CatalogScanRequ" + "\033.SetBalancerRunningResponse\022;\n\016RunCatal" +
"est\032\024.CatalogScanResponse\022S\n\024enableCatal", "ogScan\022\023.CatalogScanRequest\032\024.CatalogSca",
"ogJanitor\022\034.EnableCatalogJanitorRequest\032" + "nResponse\022S\n\024EnableCatalogJanitor\022\034.Enab" +
"\035.EnableCatalogJanitorResponse\022\\\n\027isCata" + "leCatalogJanitorRequest\032\035.EnableCatalogJ" +
"logJanitorEnabled\022\037.IsCatalogJanitorEnab" + "anitorResponse\022\\\n\027IsCatalogJanitorEnable" +
"ledRequest\032 .IsCatalogJanitorEnabledResp" + "d\022\037.IsCatalogJanitorEnabledRequest\032 .IsC" +
"onse\022L\n\021execMasterService\022\032.CoprocessorS" + "atalogJanitorEnabledResponse\022L\n\021ExecMast" +
"erviceRequest\032\033.CoprocessorServiceRespon" + "erService\022\032.CoprocessorServiceRequest\032\033." +
"se\0227\n\010snapshot\022\024.TakeSnapshotRequest\032\025.T" + "CoprocessorServiceResponse\0227\n\010Snapshot\022\024" +
"akeSnapshotResponse\022D\n\025getCompletedSnaps" + ".TakeSnapshotRequest\032\025.TakeSnapshotRespo" +
"hots\022\024.ListSnapshotRequest\032\025.ListSnapsho" + "nse\022D\n\025GetCompletedSnapshots\022\024.ListSnaps" +
"tResponse\022A\n\016deleteSnapshot\022\026.DeleteSnap", "hotRequest\032\025.ListSnapshotResponse\022A\n\016Del",
"shotRequest\032\027.DeleteSnapshotResponse\022A\n\016" + "eteSnapshot\022\026.DeleteSnapshotRequest\032\027.De" +
"isSnapshotDone\022\026.IsSnapshotDoneRequest\032\027" + "leteSnapshotResponse\022A\n\016IsSnapshotDone\022\026" +
".IsSnapshotDoneResponse\022D\n\017restoreSnapsh" + ".IsSnapshotDoneRequest\032\027.IsSnapshotDoneR" +
"ot\022\027.RestoreSnapshotRequest\032\030.RestoreSna" + "esponse\022D\n\017RestoreSnapshot\022\027.RestoreSnap" +
"pshotResponse\022V\n\025isRestoreSnapshotDone\022\035" + "shotRequest\032\030.RestoreSnapshotResponse\022V\n" +
".IsRestoreSnapshotDoneRequest\032\036.IsRestor" + "\025IsRestoreSnapshotDone\022\035.IsRestoreSnapsh" +
"eSnapshotDoneResponse\022D\n\017isMasterRunning" + "otDoneRequest\032\036.IsRestoreSnapshotDoneRes" +
"\022\027.IsMasterRunningRequest\032\030.IsMasterRunn" + "ponse\022D\n\017IsMasterRunning\022\027.IsMasterRunni" +
"ingResponseBG\n*org.apache.hadoop.hbase.p" + "ngRequest\032\030.IsMasterRunningResponseBG\n*o" +
"rotobuf.generatedB\021MasterAdminProtosH\001\210\001", "rg.apache.hadoop.hbase.protobuf.generate",
"\001\240\001\001" "dB\021MasterAdminProtosH\001\210\001\001\240\001\001"
}; };
com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner = com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() { new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() {

View File

@ -11,7 +11,7 @@ public final class MasterMonitorProtos {
public interface GetSchemaAlterStatusRequestOrBuilder public interface GetSchemaAlterStatusRequestOrBuilder
extends com.google.protobuf.MessageOrBuilder { extends com.google.protobuf.MessageOrBuilder {
// required bytes tableName = 1; // required bytes table_name = 1;
boolean hasTableName(); boolean hasTableName();
com.google.protobuf.ByteString getTableName(); com.google.protobuf.ByteString getTableName();
} }
@ -44,8 +44,8 @@ public final class MasterMonitorProtos {
} }
private int bitField0_; private int bitField0_;
// required bytes tableName = 1; // required bytes table_name = 1;
public static final int TABLENAME_FIELD_NUMBER = 1; public static final int TABLE_NAME_FIELD_NUMBER = 1;
private com.google.protobuf.ByteString tableName_; private com.google.protobuf.ByteString tableName_;
public boolean hasTableName() { public boolean hasTableName() {
return ((bitField0_ & 0x00000001) == 0x00000001); return ((bitField0_ & 0x00000001) == 0x00000001);
@ -127,7 +127,7 @@ public final class MasterMonitorProtos {
int hash = 41; int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode(); hash = (19 * hash) + getDescriptorForType().hashCode();
if (hasTableName()) { if (hasTableName()) {
hash = (37 * hash) + TABLENAME_FIELD_NUMBER; hash = (37 * hash) + TABLE_NAME_FIELD_NUMBER;
hash = (53 * hash) + getTableName().hashCode(); hash = (53 * hash) + getTableName().hashCode();
} }
hash = (29 * hash) + getUnknownFields().hashCode(); hash = (29 * hash) + getUnknownFields().hashCode();
@ -355,7 +355,7 @@ public final class MasterMonitorProtos {
private int bitField0_; private int bitField0_;
// required bytes tableName = 1; // required bytes table_name = 1;
private com.google.protobuf.ByteString tableName_ = com.google.protobuf.ByteString.EMPTY; private com.google.protobuf.ByteString tableName_ = com.google.protobuf.ByteString.EMPTY;
public boolean hasTableName() { public boolean hasTableName() {
return ((bitField0_ & 0x00000001) == 0x00000001); return ((bitField0_ & 0x00000001) == 0x00000001);
@ -393,11 +393,11 @@ public final class MasterMonitorProtos {
public interface GetSchemaAlterStatusResponseOrBuilder public interface GetSchemaAlterStatusResponseOrBuilder
extends com.google.protobuf.MessageOrBuilder { extends com.google.protobuf.MessageOrBuilder {
// optional uint32 yetToUpdateRegions = 1; // optional uint32 yet_to_update_regions = 1;
boolean hasYetToUpdateRegions(); boolean hasYetToUpdateRegions();
int getYetToUpdateRegions(); int getYetToUpdateRegions();
// optional uint32 totalRegions = 2; // optional uint32 total_regions = 2;
boolean hasTotalRegions(); boolean hasTotalRegions();
int getTotalRegions(); int getTotalRegions();
} }
@ -430,8 +430,8 @@ public final class MasterMonitorProtos {
} }
private int bitField0_; private int bitField0_;
// optional uint32 yetToUpdateRegions = 1; // optional uint32 yet_to_update_regions = 1;
public static final int YETTOUPDATEREGIONS_FIELD_NUMBER = 1; public static final int YET_TO_UPDATE_REGIONS_FIELD_NUMBER = 1;
private int yetToUpdateRegions_; private int yetToUpdateRegions_;
public boolean hasYetToUpdateRegions() { public boolean hasYetToUpdateRegions() {
return ((bitField0_ & 0x00000001) == 0x00000001); return ((bitField0_ & 0x00000001) == 0x00000001);
@ -440,8 +440,8 @@ public final class MasterMonitorProtos {
return yetToUpdateRegions_; return yetToUpdateRegions_;
} }
// optional uint32 totalRegions = 2; // optional uint32 total_regions = 2;
public static final int TOTALREGIONS_FIELD_NUMBER = 2; public static final int TOTAL_REGIONS_FIELD_NUMBER = 2;
private int totalRegions_; private int totalRegions_;
public boolean hasTotalRegions() { public boolean hasTotalRegions() {
return ((bitField0_ & 0x00000002) == 0x00000002); return ((bitField0_ & 0x00000002) == 0x00000002);
@ -532,11 +532,11 @@ public final class MasterMonitorProtos {
int hash = 41; int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode(); hash = (19 * hash) + getDescriptorForType().hashCode();
if (hasYetToUpdateRegions()) { if (hasYetToUpdateRegions()) {
hash = (37 * hash) + YETTOUPDATEREGIONS_FIELD_NUMBER; hash = (37 * hash) + YET_TO_UPDATE_REGIONS_FIELD_NUMBER;
hash = (53 * hash) + getYetToUpdateRegions(); hash = (53 * hash) + getYetToUpdateRegions();
} }
if (hasTotalRegions()) { if (hasTotalRegions()) {
hash = (37 * hash) + TOTALREGIONS_FIELD_NUMBER; hash = (37 * hash) + TOTAL_REGIONS_FIELD_NUMBER;
hash = (53 * hash) + getTotalRegions(); hash = (53 * hash) + getTotalRegions();
} }
hash = (29 * hash) + getUnknownFields().hashCode(); hash = (29 * hash) + getUnknownFields().hashCode();
@ -774,7 +774,7 @@ public final class MasterMonitorProtos {
private int bitField0_; private int bitField0_;
// optional uint32 yetToUpdateRegions = 1; // optional uint32 yet_to_update_regions = 1;
private int yetToUpdateRegions_ ; private int yetToUpdateRegions_ ;
public boolean hasYetToUpdateRegions() { public boolean hasYetToUpdateRegions() {
return ((bitField0_ & 0x00000001) == 0x00000001); return ((bitField0_ & 0x00000001) == 0x00000001);
@ -795,7 +795,7 @@ public final class MasterMonitorProtos {
return this; return this;
} }
// optional uint32 totalRegions = 2; // optional uint32 total_regions = 2;
private int totalRegions_ ; private int totalRegions_ ;
public boolean hasTotalRegions() { public boolean hasTotalRegions() {
return ((bitField0_ & 0x00000002) == 0x00000002); return ((bitField0_ & 0x00000002) == 0x00000002);
@ -830,7 +830,7 @@ public final class MasterMonitorProtos {
public interface GetTableDescriptorsRequestOrBuilder public interface GetTableDescriptorsRequestOrBuilder
extends com.google.protobuf.MessageOrBuilder { extends com.google.protobuf.MessageOrBuilder {
// repeated string tableNames = 1; // repeated string table_names = 1;
java.util.List<String> getTableNamesList(); java.util.List<String> getTableNamesList();
int getTableNamesCount(); int getTableNamesCount();
String getTableNames(int index); String getTableNames(int index);
@ -863,8 +863,8 @@ public final class MasterMonitorProtos {
return org.apache.hadoop.hbase.protobuf.generated.MasterMonitorProtos.internal_static_GetTableDescriptorsRequest_fieldAccessorTable; return org.apache.hadoop.hbase.protobuf.generated.MasterMonitorProtos.internal_static_GetTableDescriptorsRequest_fieldAccessorTable;
} }
// repeated string tableNames = 1; // repeated string table_names = 1;
public static final int TABLENAMES_FIELD_NUMBER = 1; public static final int TABLE_NAMES_FIELD_NUMBER = 1;
private com.google.protobuf.LazyStringList tableNames_; private com.google.protobuf.LazyStringList tableNames_;
public java.util.List<String> public java.util.List<String>
getTableNamesList() { getTableNamesList() {
@ -948,7 +948,7 @@ public final class MasterMonitorProtos {
int hash = 41; int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode(); hash = (19 * hash) + getDescriptorForType().hashCode();
if (getTableNamesCount() > 0) { if (getTableNamesCount() > 0) {
hash = (37 * hash) + TABLENAMES_FIELD_NUMBER; hash = (37 * hash) + TABLE_NAMES_FIELD_NUMBER;
hash = (53 * hash) + getTableNamesList().hashCode(); hash = (53 * hash) + getTableNamesList().hashCode();
} }
hash = (29 * hash) + getUnknownFields().hashCode(); hash = (29 * hash) + getUnknownFields().hashCode();
@ -1179,7 +1179,7 @@ public final class MasterMonitorProtos {
private int bitField0_; private int bitField0_;
// repeated string tableNames = 1; // repeated string table_names = 1;
private com.google.protobuf.LazyStringList tableNames_ = com.google.protobuf.LazyStringArrayList.EMPTY; private com.google.protobuf.LazyStringList tableNames_ = com.google.protobuf.LazyStringArrayList.EMPTY;
private void ensureTableNamesIsMutable() { private void ensureTableNamesIsMutable() {
if (!((bitField0_ & 0x00000001) == 0x00000001)) { if (!((bitField0_ & 0x00000001) == 0x00000001)) {
@ -1249,7 +1249,7 @@ public final class MasterMonitorProtos {
public interface GetTableDescriptorsResponseOrBuilder public interface GetTableDescriptorsResponseOrBuilder
extends com.google.protobuf.MessageOrBuilder { extends com.google.protobuf.MessageOrBuilder {
// repeated .TableSchema tableSchema = 1; // repeated .TableSchema table_schema = 1;
java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema> java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema>
getTableSchemaList(); getTableSchemaList();
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema getTableSchema(int index); org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema getTableSchema(int index);
@ -1287,8 +1287,8 @@ public final class MasterMonitorProtos {
return org.apache.hadoop.hbase.protobuf.generated.MasterMonitorProtos.internal_static_GetTableDescriptorsResponse_fieldAccessorTable; return org.apache.hadoop.hbase.protobuf.generated.MasterMonitorProtos.internal_static_GetTableDescriptorsResponse_fieldAccessorTable;
} }
// repeated .TableSchema tableSchema = 1; // repeated .TableSchema table_schema = 1;
public static final int TABLESCHEMA_FIELD_NUMBER = 1; public static final int TABLE_SCHEMA_FIELD_NUMBER = 1;
private java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema> tableSchema_; private java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema> tableSchema_;
public java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema> getTableSchemaList() { public java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema> getTableSchemaList() {
return tableSchema_; return tableSchema_;
@ -1380,7 +1380,7 @@ public final class MasterMonitorProtos {
int hash = 41; int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode(); hash = (19 * hash) + getDescriptorForType().hashCode();
if (getTableSchemaCount() > 0) { if (getTableSchemaCount() > 0) {
hash = (37 * hash) + TABLESCHEMA_FIELD_NUMBER; hash = (37 * hash) + TABLE_SCHEMA_FIELD_NUMBER;
hash = (53 * hash) + getTableSchemaList().hashCode(); hash = (53 * hash) + getTableSchemaList().hashCode();
} }
hash = (29 * hash) + getUnknownFields().hashCode(); hash = (29 * hash) + getUnknownFields().hashCode();
@ -1642,7 +1642,7 @@ public final class MasterMonitorProtos {
private int bitField0_; private int bitField0_;
// repeated .TableSchema tableSchema = 1; // repeated .TableSchema table_schema = 1;
private java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema> tableSchema_ = private java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema> tableSchema_ =
java.util.Collections.emptyList(); java.util.Collections.emptyList();
private void ensureTableSchemaIsMutable() { private void ensureTableSchemaIsMutable() {
@ -2142,7 +2142,7 @@ public final class MasterMonitorProtos {
public interface GetClusterStatusResponseOrBuilder public interface GetClusterStatusResponseOrBuilder
extends com.google.protobuf.MessageOrBuilder { extends com.google.protobuf.MessageOrBuilder {
// required .ClusterStatus clusterStatus = 1; // required .ClusterStatus cluster_status = 1;
boolean hasClusterStatus(); boolean hasClusterStatus();
org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.ClusterStatus getClusterStatus(); org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.ClusterStatus getClusterStatus();
org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.ClusterStatusOrBuilder getClusterStatusOrBuilder(); org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.ClusterStatusOrBuilder getClusterStatusOrBuilder();
@ -2176,8 +2176,8 @@ public final class MasterMonitorProtos {
} }
private int bitField0_; private int bitField0_;
// required .ClusterStatus clusterStatus = 1; // required .ClusterStatus cluster_status = 1;
public static final int CLUSTERSTATUS_FIELD_NUMBER = 1; public static final int CLUSTER_STATUS_FIELD_NUMBER = 1;
private org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.ClusterStatus clusterStatus_; private org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.ClusterStatus clusterStatus_;
public boolean hasClusterStatus() { public boolean hasClusterStatus() {
return ((bitField0_ & 0x00000001) == 0x00000001); return ((bitField0_ & 0x00000001) == 0x00000001);
@ -2266,7 +2266,7 @@ public final class MasterMonitorProtos {
int hash = 41; int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode(); hash = (19 * hash) + getDescriptorForType().hashCode();
if (hasClusterStatus()) { if (hasClusterStatus()) {
hash = (37 * hash) + CLUSTERSTATUS_FIELD_NUMBER; hash = (37 * hash) + CLUSTER_STATUS_FIELD_NUMBER;
hash = (53 * hash) + getClusterStatus().hashCode(); hash = (53 * hash) + getClusterStatus().hashCode();
} }
hash = (29 * hash) + getUnknownFields().hashCode(); hash = (29 * hash) + getUnknownFields().hashCode();
@ -2511,7 +2511,7 @@ public final class MasterMonitorProtos {
private int bitField0_; private int bitField0_;
// required .ClusterStatus clusterStatus = 1; // required .ClusterStatus cluster_status = 1;
private org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.ClusterStatus clusterStatus_ = org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.ClusterStatus.getDefaultInstance(); private org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.ClusterStatus clusterStatus_ = org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.ClusterStatus.getDefaultInstance();
private com.google.protobuf.SingleFieldBuilder< private com.google.protobuf.SingleFieldBuilder<
org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.ClusterStatus, org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.ClusterStatus.Builder, org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.ClusterStatusOrBuilder> clusterStatusBuilder_; org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.ClusterStatus, org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.ClusterStatus.Builder, org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.ClusterStatusOrBuilder> clusterStatusBuilder_;
@ -3068,26 +3068,26 @@ public final class MasterMonitorProtos {
static { static {
java.lang.String[] descriptorData = { java.lang.String[] descriptorData = {
"\n\023MasterMonitor.proto\032\014Master.proto\032\013hba" + "\n\023MasterMonitor.proto\032\014Master.proto\032\013hba" +
"se.proto\032\023ClusterStatus.proto\"0\n\033GetSche" + "se.proto\032\023ClusterStatus.proto\"1\n\033GetSche" +
"maAlterStatusRequest\022\021\n\ttableName\030\001 \002(\014\"" + "maAlterStatusRequest\022\022\n\ntable_name\030\001 \002(\014" +
"P\n\034GetSchemaAlterStatusResponse\022\032\n\022yetTo" + "\"T\n\034GetSchemaAlterStatusResponse\022\035\n\025yet_" +
"UpdateRegions\030\001 \001(\r\022\024\n\014totalRegions\030\002 \001(" + "to_update_regions\030\001 \001(\r\022\025\n\rtotal_regions" +
"\r\"0\n\032GetTableDescriptorsRequest\022\022\n\ntable" + "\030\002 \001(\r\"1\n\032GetTableDescriptorsRequest\022\023\n\013" +
"Names\030\001 \003(\t\"@\n\033GetTableDescriptorsRespon" + "table_names\030\001 \003(\t\"A\n\033GetTableDescriptors" +
"se\022!\n\013tableSchema\030\001 \003(\0132\014.TableSchema\"\031\n" + "Response\022\"\n\014table_schema\030\001 \003(\0132\014.TableSc" +
"\027GetClusterStatusRequest\"A\n\030GetClusterSt" + "hema\"\031\n\027GetClusterStatusRequest\"B\n\030GetCl" +
"atusResponse\022%\n\rclusterStatus\030\001 \002(\0132\016.Cl", "usterStatusResponse\022&\n\016cluster_status\030\001 ",
"usterStatus2\314\002\n\024MasterMonitorService\022S\n\024" + "\002(\0132\016.ClusterStatus2\314\002\n\024MasterMonitorSer" +
"getSchemaAlterStatus\022\034.GetSchemaAlterSta" + "vice\022S\n\024GetSchemaAlterStatus\022\034.GetSchema" +
"tusRequest\032\035.GetSchemaAlterStatusRespons" + "AlterStatusRequest\032\035.GetSchemaAlterStatu" +
"e\022P\n\023getTableDescriptors\022\033.GetTableDescr" + "sResponse\022P\n\023GetTableDescriptors\022\033.GetTa" +
"iptorsRequest\032\034.GetTableDescriptorsRespo" + "bleDescriptorsRequest\032\034.GetTableDescript" +
"nse\022G\n\020getClusterStatus\022\030.GetClusterStat" + "orsResponse\022G\n\020GetClusterStatus\022\030.GetClu" +
"usRequest\032\031.GetClusterStatusResponse\022D\n\017" + "sterStatusRequest\032\031.GetClusterStatusResp" +
"isMasterRunning\022\027.IsMasterRunningRequest" + "onse\022D\n\017IsMasterRunning\022\027.IsMasterRunnin" +
"\032\030.IsMasterRunningResponseBI\n*org.apache" + "gRequest\032\030.IsMasterRunningResponseBI\n*or" +
".hadoop.hbase.protobuf.generatedB\023Master", "g.apache.hadoop.hbase.protobuf.generated",
"MonitorProtosH\001\210\001\001\240\001\001" "B\023MasterMonitorProtosH\001\210\001\001\240\001\001"
}; };
com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner = com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() { new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() {

View File

@ -311,7 +311,7 @@ public final class MasterProtos {
public interface IsMasterRunningResponseOrBuilder public interface IsMasterRunningResponseOrBuilder
extends com.google.protobuf.MessageOrBuilder { extends com.google.protobuf.MessageOrBuilder {
// required bool isMasterRunning = 1; // required bool is_master_running = 1;
boolean hasIsMasterRunning(); boolean hasIsMasterRunning();
boolean getIsMasterRunning(); boolean getIsMasterRunning();
} }
@ -344,8 +344,8 @@ public final class MasterProtos {
} }
private int bitField0_; private int bitField0_;
// required bool isMasterRunning = 1; // required bool is_master_running = 1;
public static final int ISMASTERRUNNING_FIELD_NUMBER = 1; public static final int IS_MASTER_RUNNING_FIELD_NUMBER = 1;
private boolean isMasterRunning_; private boolean isMasterRunning_;
public boolean hasIsMasterRunning() { public boolean hasIsMasterRunning() {
return ((bitField0_ & 0x00000001) == 0x00000001); return ((bitField0_ & 0x00000001) == 0x00000001);
@ -427,7 +427,7 @@ public final class MasterProtos {
int hash = 41; int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode(); hash = (19 * hash) + getDescriptorForType().hashCode();
if (hasIsMasterRunning()) { if (hasIsMasterRunning()) {
hash = (37 * hash) + ISMASTERRUNNING_FIELD_NUMBER; hash = (37 * hash) + IS_MASTER_RUNNING_FIELD_NUMBER;
hash = (53 * hash) + hashBoolean(getIsMasterRunning()); hash = (53 * hash) + hashBoolean(getIsMasterRunning());
} }
hash = (29 * hash) + getUnknownFields().hashCode(); hash = (29 * hash) + getUnknownFields().hashCode();
@ -655,7 +655,7 @@ public final class MasterProtos {
private int bitField0_; private int bitField0_;
// required bool isMasterRunning = 1; // required bool is_master_running = 1;
private boolean isMasterRunning_ ; private boolean isMasterRunning_ ;
public boolean hasIsMasterRunning() { public boolean hasIsMasterRunning() {
return ((bitField0_ & 0x00000001) == 0x00000001); return ((bitField0_ & 0x00000001) == 0x00000001);
@ -928,12 +928,12 @@ public final class MasterProtos {
static { static {
java.lang.String[] descriptorData = { java.lang.String[] descriptorData = {
"\n\014Master.proto\"\030\n\026IsMasterRunningRequest" + "\n\014Master.proto\"\030\n\026IsMasterRunningRequest" +
"\"2\n\027IsMasterRunningResponse\022\027\n\017isMasterR" + "\"4\n\027IsMasterRunningResponse\022\031\n\021is_master" +
"unning\030\001 \002(\0102U\n\rMasterService\022D\n\017isMaste" + "_running\030\001 \002(\0102U\n\rMasterService\022D\n\017IsMas" +
"rRunning\022\027.IsMasterRunningRequest\032\030.IsMa" + "terRunning\022\027.IsMasterRunningRequest\032\030.Is" +
"sterRunningResponseBB\n*org.apache.hadoop" + "MasterRunningResponseBB\n*org.apache.hado" +
".hbase.protobuf.generatedB\014MasterProtosH" + "op.hbase.protobuf.generatedB\014MasterProto" +
"\001\210\001\001\240\001\001" "sH\001\210\001\001\240\001\001"
}; };
com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner = com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() { new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() {

View File

@ -11,7 +11,7 @@ public final class MultiRowMutation {
public interface MultiMutateRequestOrBuilder public interface MultiMutateRequestOrBuilder
extends com.google.protobuf.MessageOrBuilder { extends com.google.protobuf.MessageOrBuilder {
// repeated .MutationProto mutationRequest = 1; // repeated .MutationProto mutation_request = 1;
java.util.List<org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MutationProto> java.util.List<org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MutationProto>
getMutationRequestList(); getMutationRequestList();
org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MutationProto getMutationRequest(int index); org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MutationProto getMutationRequest(int index);
@ -49,8 +49,8 @@ public final class MultiRowMutation {
return org.apache.hadoop.hbase.protobuf.generated.MultiRowMutation.internal_static_MultiMutateRequest_fieldAccessorTable; return org.apache.hadoop.hbase.protobuf.generated.MultiRowMutation.internal_static_MultiMutateRequest_fieldAccessorTable;
} }
// repeated .MutationProto mutationRequest = 1; // repeated .MutationProto mutation_request = 1;
public static final int MUTATIONREQUEST_FIELD_NUMBER = 1; public static final int MUTATION_REQUEST_FIELD_NUMBER = 1;
private java.util.List<org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MutationProto> mutationRequest_; private java.util.List<org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MutationProto> mutationRequest_;
public java.util.List<org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MutationProto> getMutationRequestList() { public java.util.List<org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MutationProto> getMutationRequestList() {
return mutationRequest_; return mutationRequest_;
@ -142,7 +142,7 @@ public final class MultiRowMutation {
int hash = 41; int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode(); hash = (19 * hash) + getDescriptorForType().hashCode();
if (getMutationRequestCount() > 0) { if (getMutationRequestCount() > 0) {
hash = (37 * hash) + MUTATIONREQUEST_FIELD_NUMBER; hash = (37 * hash) + MUTATION_REQUEST_FIELD_NUMBER;
hash = (53 * hash) + getMutationRequestList().hashCode(); hash = (53 * hash) + getMutationRequestList().hashCode();
} }
hash = (29 * hash) + getUnknownFields().hashCode(); hash = (29 * hash) + getUnknownFields().hashCode();
@ -404,7 +404,7 @@ public final class MultiRowMutation {
private int bitField0_; private int bitField0_;
// repeated .MutationProto mutationRequest = 1; // repeated .MutationProto mutation_request = 1;
private java.util.List<org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MutationProto> mutationRequest_ = private java.util.List<org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MutationProto> mutationRequest_ =
java.util.Collections.emptyList(); java.util.Collections.emptyList();
private void ensureMutationRequestIsMutable() { private void ensureMutationRequestIsMutable() {
@ -1141,13 +1141,14 @@ public final class MultiRowMutation {
descriptor; descriptor;
static { static {
java.lang.String[] descriptorData = { java.lang.String[] descriptorData = {
"\n\026MultiRowMutation.proto\032\014Client.proto\"=" + "\n\026MultiRowMutation.proto\032\014Client.proto\">" +
"\n\022MultiMutateRequest\022\'\n\017mutationRequest\030" + "\n\022MultiMutateRequest\022(\n\020mutation_request" +
"\001 \003(\0132\016.MutationProto\"\025\n\023MultiMutateResp" + "\030\001 \003(\0132\016.MutationProto\"\025\n\023MultiMutateRes" +
"onse2R\n\027MultiRowMutationService\0227\n\nmutat" + "ponse2R\n\027MultiRowMutationService\0227\n\nMuta" +
"eRows\022\023.MultiMutateRequest\032\024.MultiMutate" + "teRows\022\023.MultiMutateRequest\032\024.MultiMutat" +
"ResponseBF\n*org.apache.hadoop.hbase.prot" + "eResponseBF\n*org.apache.hadoop.hbase.pro" +
"obuf.generatedB\020MultiRowMutationH\001\210\001\001\240\001\001" "tobuf.generatedB\020MultiRowMutationH\001\210\001\001\240\001" +
"\001"
}; };
com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner = com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() { new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() {

View File

@ -11,11 +11,11 @@ public final class RPCProtos {
public interface UserInformationOrBuilder public interface UserInformationOrBuilder
extends com.google.protobuf.MessageOrBuilder { extends com.google.protobuf.MessageOrBuilder {
// required string effectiveUser = 1; // required string effective_user = 1;
boolean hasEffectiveUser(); boolean hasEffectiveUser();
String getEffectiveUser(); String getEffectiveUser();
// optional string realUser = 2; // optional string real_user = 2;
boolean hasRealUser(); boolean hasRealUser();
String getRealUser(); String getRealUser();
} }
@ -48,8 +48,8 @@ public final class RPCProtos {
} }
private int bitField0_; private int bitField0_;
// required string effectiveUser = 1; // required string effective_user = 1;
public static final int EFFECTIVEUSER_FIELD_NUMBER = 1; public static final int EFFECTIVE_USER_FIELD_NUMBER = 1;
private java.lang.Object effectiveUser_; private java.lang.Object effectiveUser_;
public boolean hasEffectiveUser() { public boolean hasEffectiveUser() {
return ((bitField0_ & 0x00000001) == 0x00000001); return ((bitField0_ & 0x00000001) == 0x00000001);
@ -80,8 +80,8 @@ public final class RPCProtos {
} }
} }
// optional string realUser = 2; // optional string real_user = 2;
public static final int REALUSER_FIELD_NUMBER = 2; public static final int REAL_USER_FIELD_NUMBER = 2;
private java.lang.Object realUser_; private java.lang.Object realUser_;
public boolean hasRealUser() { public boolean hasRealUser() {
return ((bitField0_ & 0x00000002) == 0x00000002); return ((bitField0_ & 0x00000002) == 0x00000002);
@ -198,11 +198,11 @@ public final class RPCProtos {
int hash = 41; int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode(); hash = (19 * hash) + getDescriptorForType().hashCode();
if (hasEffectiveUser()) { if (hasEffectiveUser()) {
hash = (37 * hash) + EFFECTIVEUSER_FIELD_NUMBER; hash = (37 * hash) + EFFECTIVE_USER_FIELD_NUMBER;
hash = (53 * hash) + getEffectiveUser().hashCode(); hash = (53 * hash) + getEffectiveUser().hashCode();
} }
if (hasRealUser()) { if (hasRealUser()) {
hash = (37 * hash) + REALUSER_FIELD_NUMBER; hash = (37 * hash) + REAL_USER_FIELD_NUMBER;
hash = (53 * hash) + getRealUser().hashCode(); hash = (53 * hash) + getRealUser().hashCode();
} }
hash = (29 * hash) + getUnknownFields().hashCode(); hash = (29 * hash) + getUnknownFields().hashCode();
@ -444,7 +444,7 @@ public final class RPCProtos {
private int bitField0_; private int bitField0_;
// required string effectiveUser = 1; // required string effective_user = 1;
private java.lang.Object effectiveUser_ = ""; private java.lang.Object effectiveUser_ = "";
public boolean hasEffectiveUser() { public boolean hasEffectiveUser() {
return ((bitField0_ & 0x00000001) == 0x00000001); return ((bitField0_ & 0x00000001) == 0x00000001);
@ -480,7 +480,7 @@ public final class RPCProtos {
onChanged(); onChanged();
} }
// optional string realUser = 2; // optional string real_user = 2;
private java.lang.Object realUser_ = ""; private java.lang.Object realUser_ = "";
public boolean hasRealUser() { public boolean hasRealUser() {
return ((bitField0_ & 0x00000002) == 0x00000002); return ((bitField0_ & 0x00000002) == 0x00000002);
@ -530,20 +530,20 @@ public final class RPCProtos {
public interface ConnectionHeaderOrBuilder public interface ConnectionHeaderOrBuilder
extends com.google.protobuf.MessageOrBuilder { extends com.google.protobuf.MessageOrBuilder {
// optional .UserInformation userInfo = 1; // optional .UserInformation user_info = 1;
boolean hasUserInfo(); boolean hasUserInfo();
org.apache.hadoop.hbase.protobuf.generated.RPCProtos.UserInformation getUserInfo(); org.apache.hadoop.hbase.protobuf.generated.RPCProtos.UserInformation getUserInfo();
org.apache.hadoop.hbase.protobuf.generated.RPCProtos.UserInformationOrBuilder getUserInfoOrBuilder(); org.apache.hadoop.hbase.protobuf.generated.RPCProtos.UserInformationOrBuilder getUserInfoOrBuilder();
// optional string serviceName = 2; // optional string service_name = 2;
boolean hasServiceName(); boolean hasServiceName();
String getServiceName(); String getServiceName();
// optional string cellBlockCodecClass = 3 [default = "org.apache.hadoop.hbase.codec.KeyValueCodec"]; // optional string cell_block_codec_class = 3 [default = "org.apache.hadoop.hbase.codec.KeyValueCodec"];
boolean hasCellBlockCodecClass(); boolean hasCellBlockCodecClass();
String getCellBlockCodecClass(); String getCellBlockCodecClass();
// optional string cellBlockCompressorClass = 4; // optional string cell_block_compressor_class = 4;
boolean hasCellBlockCompressorClass(); boolean hasCellBlockCompressorClass();
String getCellBlockCompressorClass(); String getCellBlockCompressorClass();
} }
@ -576,8 +576,8 @@ public final class RPCProtos {
} }
private int bitField0_; private int bitField0_;
// optional .UserInformation userInfo = 1; // optional .UserInformation user_info = 1;
public static final int USERINFO_FIELD_NUMBER = 1; public static final int USER_INFO_FIELD_NUMBER = 1;
private org.apache.hadoop.hbase.protobuf.generated.RPCProtos.UserInformation userInfo_; private org.apache.hadoop.hbase.protobuf.generated.RPCProtos.UserInformation userInfo_;
public boolean hasUserInfo() { public boolean hasUserInfo() {
return ((bitField0_ & 0x00000001) == 0x00000001); return ((bitField0_ & 0x00000001) == 0x00000001);
@ -589,8 +589,8 @@ public final class RPCProtos {
return userInfo_; return userInfo_;
} }
// optional string serviceName = 2; // optional string service_name = 2;
public static final int SERVICENAME_FIELD_NUMBER = 2; public static final int SERVICE_NAME_FIELD_NUMBER = 2;
private java.lang.Object serviceName_; private java.lang.Object serviceName_;
public boolean hasServiceName() { public boolean hasServiceName() {
return ((bitField0_ & 0x00000002) == 0x00000002); return ((bitField0_ & 0x00000002) == 0x00000002);
@ -621,8 +621,8 @@ public final class RPCProtos {
} }
} }
// optional string cellBlockCodecClass = 3 [default = "org.apache.hadoop.hbase.codec.KeyValueCodec"]; // optional string cell_block_codec_class = 3 [default = "org.apache.hadoop.hbase.codec.KeyValueCodec"];
public static final int CELLBLOCKCODECCLASS_FIELD_NUMBER = 3; public static final int CELL_BLOCK_CODEC_CLASS_FIELD_NUMBER = 3;
private java.lang.Object cellBlockCodecClass_; private java.lang.Object cellBlockCodecClass_;
public boolean hasCellBlockCodecClass() { public boolean hasCellBlockCodecClass() {
return ((bitField0_ & 0x00000004) == 0x00000004); return ((bitField0_ & 0x00000004) == 0x00000004);
@ -653,8 +653,8 @@ public final class RPCProtos {
} }
} }
// optional string cellBlockCompressorClass = 4; // optional string cell_block_compressor_class = 4;
public static final int CELLBLOCKCOMPRESSORCLASS_FIELD_NUMBER = 4; public static final int CELL_BLOCK_COMPRESSOR_CLASS_FIELD_NUMBER = 4;
private java.lang.Object cellBlockCompressorClass_; private java.lang.Object cellBlockCompressorClass_;
public boolean hasCellBlockCompressorClass() { public boolean hasCellBlockCompressorClass() {
return ((bitField0_ & 0x00000008) == 0x00000008); return ((bitField0_ & 0x00000008) == 0x00000008);
@ -799,19 +799,19 @@ public final class RPCProtos {
int hash = 41; int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode(); hash = (19 * hash) + getDescriptorForType().hashCode();
if (hasUserInfo()) { if (hasUserInfo()) {
hash = (37 * hash) + USERINFO_FIELD_NUMBER; hash = (37 * hash) + USER_INFO_FIELD_NUMBER;
hash = (53 * hash) + getUserInfo().hashCode(); hash = (53 * hash) + getUserInfo().hashCode();
} }
if (hasServiceName()) { if (hasServiceName()) {
hash = (37 * hash) + SERVICENAME_FIELD_NUMBER; hash = (37 * hash) + SERVICE_NAME_FIELD_NUMBER;
hash = (53 * hash) + getServiceName().hashCode(); hash = (53 * hash) + getServiceName().hashCode();
} }
if (hasCellBlockCodecClass()) { if (hasCellBlockCodecClass()) {
hash = (37 * hash) + CELLBLOCKCODECCLASS_FIELD_NUMBER; hash = (37 * hash) + CELL_BLOCK_CODEC_CLASS_FIELD_NUMBER;
hash = (53 * hash) + getCellBlockCodecClass().hashCode(); hash = (53 * hash) + getCellBlockCodecClass().hashCode();
} }
if (hasCellBlockCompressorClass()) { if (hasCellBlockCompressorClass()) {
hash = (37 * hash) + CELLBLOCKCOMPRESSORCLASS_FIELD_NUMBER; hash = (37 * hash) + CELL_BLOCK_COMPRESSOR_CLASS_FIELD_NUMBER;
hash = (53 * hash) + getCellBlockCompressorClass().hashCode(); hash = (53 * hash) + getCellBlockCompressorClass().hashCode();
} }
hash = (29 * hash) + getUnknownFields().hashCode(); hash = (29 * hash) + getUnknownFields().hashCode();
@ -1096,7 +1096,7 @@ public final class RPCProtos {
private int bitField0_; private int bitField0_;
// optional .UserInformation userInfo = 1; // optional .UserInformation user_info = 1;
private org.apache.hadoop.hbase.protobuf.generated.RPCProtos.UserInformation userInfo_ = org.apache.hadoop.hbase.protobuf.generated.RPCProtos.UserInformation.getDefaultInstance(); private org.apache.hadoop.hbase.protobuf.generated.RPCProtos.UserInformation userInfo_ = org.apache.hadoop.hbase.protobuf.generated.RPCProtos.UserInformation.getDefaultInstance();
private com.google.protobuf.SingleFieldBuilder< private com.google.protobuf.SingleFieldBuilder<
org.apache.hadoop.hbase.protobuf.generated.RPCProtos.UserInformation, org.apache.hadoop.hbase.protobuf.generated.RPCProtos.UserInformation.Builder, org.apache.hadoop.hbase.protobuf.generated.RPCProtos.UserInformationOrBuilder> userInfoBuilder_; org.apache.hadoop.hbase.protobuf.generated.RPCProtos.UserInformation, org.apache.hadoop.hbase.protobuf.generated.RPCProtos.UserInformation.Builder, org.apache.hadoop.hbase.protobuf.generated.RPCProtos.UserInformationOrBuilder> userInfoBuilder_;
@ -1186,7 +1186,7 @@ public final class RPCProtos {
return userInfoBuilder_; return userInfoBuilder_;
} }
// optional string serviceName = 2; // optional string service_name = 2;
private java.lang.Object serviceName_ = ""; private java.lang.Object serviceName_ = "";
public boolean hasServiceName() { public boolean hasServiceName() {
return ((bitField0_ & 0x00000002) == 0x00000002); return ((bitField0_ & 0x00000002) == 0x00000002);
@ -1222,7 +1222,7 @@ public final class RPCProtos {
onChanged(); onChanged();
} }
// optional string cellBlockCodecClass = 3 [default = "org.apache.hadoop.hbase.codec.KeyValueCodec"]; // optional string cell_block_codec_class = 3 [default = "org.apache.hadoop.hbase.codec.KeyValueCodec"];
private java.lang.Object cellBlockCodecClass_ = "org.apache.hadoop.hbase.codec.KeyValueCodec"; private java.lang.Object cellBlockCodecClass_ = "org.apache.hadoop.hbase.codec.KeyValueCodec";
public boolean hasCellBlockCodecClass() { public boolean hasCellBlockCodecClass() {
return ((bitField0_ & 0x00000004) == 0x00000004); return ((bitField0_ & 0x00000004) == 0x00000004);
@ -1258,7 +1258,7 @@ public final class RPCProtos {
onChanged(); onChanged();
} }
// optional string cellBlockCompressorClass = 4; // optional string cell_block_compressor_class = 4;
private java.lang.Object cellBlockCompressorClass_ = ""; private java.lang.Object cellBlockCompressorClass_ = "";
public boolean hasCellBlockCompressorClass() { public boolean hasCellBlockCompressorClass() {
return ((bitField0_ & 0x00000008) == 0x00000008); return ((bitField0_ & 0x00000008) == 0x00000008);
@ -1679,11 +1679,11 @@ public final class RPCProtos {
public interface ExceptionResponseOrBuilder public interface ExceptionResponseOrBuilder
extends com.google.protobuf.MessageOrBuilder { extends com.google.protobuf.MessageOrBuilder {
// optional string exceptionClassName = 1; // optional string exception_class_name = 1;
boolean hasExceptionClassName(); boolean hasExceptionClassName();
String getExceptionClassName(); String getExceptionClassName();
// optional string stackTrace = 2; // optional string stack_trace = 2;
boolean hasStackTrace(); boolean hasStackTrace();
String getStackTrace(); String getStackTrace();
@ -1695,7 +1695,7 @@ public final class RPCProtos {
boolean hasPort(); boolean hasPort();
int getPort(); int getPort();
// optional bool doNotRetry = 5; // optional bool do_not_retry = 5;
boolean hasDoNotRetry(); boolean hasDoNotRetry();
boolean getDoNotRetry(); boolean getDoNotRetry();
} }
@ -1728,8 +1728,8 @@ public final class RPCProtos {
} }
private int bitField0_; private int bitField0_;
// optional string exceptionClassName = 1; // optional string exception_class_name = 1;
public static final int EXCEPTIONCLASSNAME_FIELD_NUMBER = 1; public static final int EXCEPTION_CLASS_NAME_FIELD_NUMBER = 1;
private java.lang.Object exceptionClassName_; private java.lang.Object exceptionClassName_;
public boolean hasExceptionClassName() { public boolean hasExceptionClassName() {
return ((bitField0_ & 0x00000001) == 0x00000001); return ((bitField0_ & 0x00000001) == 0x00000001);
@ -1760,8 +1760,8 @@ public final class RPCProtos {
} }
} }
// optional string stackTrace = 2; // optional string stack_trace = 2;
public static final int STACKTRACE_FIELD_NUMBER = 2; public static final int STACK_TRACE_FIELD_NUMBER = 2;
private java.lang.Object stackTrace_; private java.lang.Object stackTrace_;
public boolean hasStackTrace() { public boolean hasStackTrace() {
return ((bitField0_ & 0x00000002) == 0x00000002); return ((bitField0_ & 0x00000002) == 0x00000002);
@ -1834,8 +1834,8 @@ public final class RPCProtos {
return port_; return port_;
} }
// optional bool doNotRetry = 5; // optional bool do_not_retry = 5;
public static final int DONOTRETRY_FIELD_NUMBER = 5; public static final int DO_NOT_RETRY_FIELD_NUMBER = 5;
private boolean doNotRetry_; private boolean doNotRetry_;
public boolean hasDoNotRetry() { public boolean hasDoNotRetry() {
return ((bitField0_ & 0x00000010) == 0x00000010); return ((bitField0_ & 0x00000010) == 0x00000010);
@ -1965,11 +1965,11 @@ public final class RPCProtos {
int hash = 41; int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode(); hash = (19 * hash) + getDescriptorForType().hashCode();
if (hasExceptionClassName()) { if (hasExceptionClassName()) {
hash = (37 * hash) + EXCEPTIONCLASSNAME_FIELD_NUMBER; hash = (37 * hash) + EXCEPTION_CLASS_NAME_FIELD_NUMBER;
hash = (53 * hash) + getExceptionClassName().hashCode(); hash = (53 * hash) + getExceptionClassName().hashCode();
} }
if (hasStackTrace()) { if (hasStackTrace()) {
hash = (37 * hash) + STACKTRACE_FIELD_NUMBER; hash = (37 * hash) + STACK_TRACE_FIELD_NUMBER;
hash = (53 * hash) + getStackTrace().hashCode(); hash = (53 * hash) + getStackTrace().hashCode();
} }
if (hasHostname()) { if (hasHostname()) {
@ -1981,7 +1981,7 @@ public final class RPCProtos {
hash = (53 * hash) + getPort(); hash = (53 * hash) + getPort();
} }
if (hasDoNotRetry()) { if (hasDoNotRetry()) {
hash = (37 * hash) + DONOTRETRY_FIELD_NUMBER; hash = (37 * hash) + DO_NOT_RETRY_FIELD_NUMBER;
hash = (53 * hash) + hashBoolean(getDoNotRetry()); hash = (53 * hash) + hashBoolean(getDoNotRetry());
} }
hash = (29 * hash) + getUnknownFields().hashCode(); hash = (29 * hash) + getUnknownFields().hashCode();
@ -2261,7 +2261,7 @@ public final class RPCProtos {
private int bitField0_; private int bitField0_;
// optional string exceptionClassName = 1; // optional string exception_class_name = 1;
private java.lang.Object exceptionClassName_ = ""; private java.lang.Object exceptionClassName_ = "";
public boolean hasExceptionClassName() { public boolean hasExceptionClassName() {
return ((bitField0_ & 0x00000001) == 0x00000001); return ((bitField0_ & 0x00000001) == 0x00000001);
@ -2297,7 +2297,7 @@ public final class RPCProtos {
onChanged(); onChanged();
} }
// optional string stackTrace = 2; // optional string stack_trace = 2;
private java.lang.Object stackTrace_ = ""; private java.lang.Object stackTrace_ = "";
public boolean hasStackTrace() { public boolean hasStackTrace() {
return ((bitField0_ & 0x00000002) == 0x00000002); return ((bitField0_ & 0x00000002) == 0x00000002);
@ -2390,7 +2390,7 @@ public final class RPCProtos {
return this; return this;
} }
// optional bool doNotRetry = 5; // optional bool do_not_retry = 5;
private boolean doNotRetry_ ; private boolean doNotRetry_ ;
public boolean hasDoNotRetry() { public boolean hasDoNotRetry() {
return ((bitField0_ & 0x00000010) == 0x00000010); return ((bitField0_ & 0x00000010) == 0x00000010);
@ -2425,24 +2425,24 @@ public final class RPCProtos {
public interface RequestHeaderOrBuilder public interface RequestHeaderOrBuilder
extends com.google.protobuf.MessageOrBuilder { extends com.google.protobuf.MessageOrBuilder {
// optional uint32 callId = 1; // optional uint32 call_id = 1;
boolean hasCallId(); boolean hasCallId();
int getCallId(); int getCallId();
// optional .RPCTInfo traceInfo = 2; // optional .RPCTInfo trace_info = 2;
boolean hasTraceInfo(); boolean hasTraceInfo();
org.apache.hadoop.hbase.protobuf.generated.Tracing.RPCTInfo getTraceInfo(); org.apache.hadoop.hbase.protobuf.generated.Tracing.RPCTInfo getTraceInfo();
org.apache.hadoop.hbase.protobuf.generated.Tracing.RPCTInfoOrBuilder getTraceInfoOrBuilder(); org.apache.hadoop.hbase.protobuf.generated.Tracing.RPCTInfoOrBuilder getTraceInfoOrBuilder();
// optional string methodName = 3; // optional string method_name = 3;
boolean hasMethodName(); boolean hasMethodName();
String getMethodName(); String getMethodName();
// optional bool requestParam = 4; // optional bool request_param = 4;
boolean hasRequestParam(); boolean hasRequestParam();
boolean getRequestParam(); boolean getRequestParam();
// optional .CellBlockMeta cellBlockMeta = 5; // optional .CellBlockMeta cell_block_meta = 5;
boolean hasCellBlockMeta(); boolean hasCellBlockMeta();
org.apache.hadoop.hbase.protobuf.generated.RPCProtos.CellBlockMeta getCellBlockMeta(); org.apache.hadoop.hbase.protobuf.generated.RPCProtos.CellBlockMeta getCellBlockMeta();
org.apache.hadoop.hbase.protobuf.generated.RPCProtos.CellBlockMetaOrBuilder getCellBlockMetaOrBuilder(); org.apache.hadoop.hbase.protobuf.generated.RPCProtos.CellBlockMetaOrBuilder getCellBlockMetaOrBuilder();
@ -2476,8 +2476,8 @@ public final class RPCProtos {
} }
private int bitField0_; private int bitField0_;
// optional uint32 callId = 1; // optional uint32 call_id = 1;
public static final int CALLID_FIELD_NUMBER = 1; public static final int CALL_ID_FIELD_NUMBER = 1;
private int callId_; private int callId_;
public boolean hasCallId() { public boolean hasCallId() {
return ((bitField0_ & 0x00000001) == 0x00000001); return ((bitField0_ & 0x00000001) == 0x00000001);
@ -2486,8 +2486,8 @@ public final class RPCProtos {
return callId_; return callId_;
} }
// optional .RPCTInfo traceInfo = 2; // optional .RPCTInfo trace_info = 2;
public static final int TRACEINFO_FIELD_NUMBER = 2; public static final int TRACE_INFO_FIELD_NUMBER = 2;
private org.apache.hadoop.hbase.protobuf.generated.Tracing.RPCTInfo traceInfo_; private org.apache.hadoop.hbase.protobuf.generated.Tracing.RPCTInfo traceInfo_;
public boolean hasTraceInfo() { public boolean hasTraceInfo() {
return ((bitField0_ & 0x00000002) == 0x00000002); return ((bitField0_ & 0x00000002) == 0x00000002);
@ -2499,8 +2499,8 @@ public final class RPCProtos {
return traceInfo_; return traceInfo_;
} }
// optional string methodName = 3; // optional string method_name = 3;
public static final int METHODNAME_FIELD_NUMBER = 3; public static final int METHOD_NAME_FIELD_NUMBER = 3;
private java.lang.Object methodName_; private java.lang.Object methodName_;
public boolean hasMethodName() { public boolean hasMethodName() {
return ((bitField0_ & 0x00000004) == 0x00000004); return ((bitField0_ & 0x00000004) == 0x00000004);
@ -2531,8 +2531,8 @@ public final class RPCProtos {
} }
} }
// optional bool requestParam = 4; // optional bool request_param = 4;
public static final int REQUESTPARAM_FIELD_NUMBER = 4; public static final int REQUEST_PARAM_FIELD_NUMBER = 4;
private boolean requestParam_; private boolean requestParam_;
public boolean hasRequestParam() { public boolean hasRequestParam() {
return ((bitField0_ & 0x00000008) == 0x00000008); return ((bitField0_ & 0x00000008) == 0x00000008);
@ -2541,8 +2541,8 @@ public final class RPCProtos {
return requestParam_; return requestParam_;
} }
// optional .CellBlockMeta cellBlockMeta = 5; // optional .CellBlockMeta cell_block_meta = 5;
public static final int CELLBLOCKMETA_FIELD_NUMBER = 5; public static final int CELL_BLOCK_META_FIELD_NUMBER = 5;
private org.apache.hadoop.hbase.protobuf.generated.RPCProtos.CellBlockMeta cellBlockMeta_; private org.apache.hadoop.hbase.protobuf.generated.RPCProtos.CellBlockMeta cellBlockMeta_;
public boolean hasCellBlockMeta() { public boolean hasCellBlockMeta() {
return ((bitField0_ & 0x00000010) == 0x00000010); return ((bitField0_ & 0x00000010) == 0x00000010);
@ -2675,23 +2675,23 @@ public final class RPCProtos {
int hash = 41; int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode(); hash = (19 * hash) + getDescriptorForType().hashCode();
if (hasCallId()) { if (hasCallId()) {
hash = (37 * hash) + CALLID_FIELD_NUMBER; hash = (37 * hash) + CALL_ID_FIELD_NUMBER;
hash = (53 * hash) + getCallId(); hash = (53 * hash) + getCallId();
} }
if (hasTraceInfo()) { if (hasTraceInfo()) {
hash = (37 * hash) + TRACEINFO_FIELD_NUMBER; hash = (37 * hash) + TRACE_INFO_FIELD_NUMBER;
hash = (53 * hash) + getTraceInfo().hashCode(); hash = (53 * hash) + getTraceInfo().hashCode();
} }
if (hasMethodName()) { if (hasMethodName()) {
hash = (37 * hash) + METHODNAME_FIELD_NUMBER; hash = (37 * hash) + METHOD_NAME_FIELD_NUMBER;
hash = (53 * hash) + getMethodName().hashCode(); hash = (53 * hash) + getMethodName().hashCode();
} }
if (hasRequestParam()) { if (hasRequestParam()) {
hash = (37 * hash) + REQUESTPARAM_FIELD_NUMBER; hash = (37 * hash) + REQUEST_PARAM_FIELD_NUMBER;
hash = (53 * hash) + hashBoolean(getRequestParam()); hash = (53 * hash) + hashBoolean(getRequestParam());
} }
if (hasCellBlockMeta()) { if (hasCellBlockMeta()) {
hash = (37 * hash) + CELLBLOCKMETA_FIELD_NUMBER; hash = (37 * hash) + CELL_BLOCK_META_FIELD_NUMBER;
hash = (53 * hash) + getCellBlockMeta().hashCode(); hash = (53 * hash) + getCellBlockMeta().hashCode();
} }
hash = (29 * hash) + getUnknownFields().hashCode(); hash = (29 * hash) + getUnknownFields().hashCode();
@ -2997,7 +2997,7 @@ public final class RPCProtos {
private int bitField0_; private int bitField0_;
// optional uint32 callId = 1; // optional uint32 call_id = 1;
private int callId_ ; private int callId_ ;
public boolean hasCallId() { public boolean hasCallId() {
return ((bitField0_ & 0x00000001) == 0x00000001); return ((bitField0_ & 0x00000001) == 0x00000001);
@ -3018,7 +3018,7 @@ public final class RPCProtos {
return this; return this;
} }
// optional .RPCTInfo traceInfo = 2; // optional .RPCTInfo trace_info = 2;
private org.apache.hadoop.hbase.protobuf.generated.Tracing.RPCTInfo traceInfo_ = org.apache.hadoop.hbase.protobuf.generated.Tracing.RPCTInfo.getDefaultInstance(); private org.apache.hadoop.hbase.protobuf.generated.Tracing.RPCTInfo traceInfo_ = org.apache.hadoop.hbase.protobuf.generated.Tracing.RPCTInfo.getDefaultInstance();
private com.google.protobuf.SingleFieldBuilder< private com.google.protobuf.SingleFieldBuilder<
org.apache.hadoop.hbase.protobuf.generated.Tracing.RPCTInfo, org.apache.hadoop.hbase.protobuf.generated.Tracing.RPCTInfo.Builder, org.apache.hadoop.hbase.protobuf.generated.Tracing.RPCTInfoOrBuilder> traceInfoBuilder_; org.apache.hadoop.hbase.protobuf.generated.Tracing.RPCTInfo, org.apache.hadoop.hbase.protobuf.generated.Tracing.RPCTInfo.Builder, org.apache.hadoop.hbase.protobuf.generated.Tracing.RPCTInfoOrBuilder> traceInfoBuilder_;
@ -3108,7 +3108,7 @@ public final class RPCProtos {
return traceInfoBuilder_; return traceInfoBuilder_;
} }
// optional string methodName = 3; // optional string method_name = 3;
private java.lang.Object methodName_ = ""; private java.lang.Object methodName_ = "";
public boolean hasMethodName() { public boolean hasMethodName() {
return ((bitField0_ & 0x00000004) == 0x00000004); return ((bitField0_ & 0x00000004) == 0x00000004);
@ -3144,7 +3144,7 @@ public final class RPCProtos {
onChanged(); onChanged();
} }
// optional bool requestParam = 4; // optional bool request_param = 4;
private boolean requestParam_ ; private boolean requestParam_ ;
public boolean hasRequestParam() { public boolean hasRequestParam() {
return ((bitField0_ & 0x00000008) == 0x00000008); return ((bitField0_ & 0x00000008) == 0x00000008);
@ -3165,7 +3165,7 @@ public final class RPCProtos {
return this; return this;
} }
// optional .CellBlockMeta cellBlockMeta = 5; // optional .CellBlockMeta cell_block_meta = 5;
private org.apache.hadoop.hbase.protobuf.generated.RPCProtos.CellBlockMeta cellBlockMeta_ = org.apache.hadoop.hbase.protobuf.generated.RPCProtos.CellBlockMeta.getDefaultInstance(); private org.apache.hadoop.hbase.protobuf.generated.RPCProtos.CellBlockMeta cellBlockMeta_ = org.apache.hadoop.hbase.protobuf.generated.RPCProtos.CellBlockMeta.getDefaultInstance();
private com.google.protobuf.SingleFieldBuilder< private com.google.protobuf.SingleFieldBuilder<
org.apache.hadoop.hbase.protobuf.generated.RPCProtos.CellBlockMeta, org.apache.hadoop.hbase.protobuf.generated.RPCProtos.CellBlockMeta.Builder, org.apache.hadoop.hbase.protobuf.generated.RPCProtos.CellBlockMetaOrBuilder> cellBlockMetaBuilder_; org.apache.hadoop.hbase.protobuf.generated.RPCProtos.CellBlockMeta, org.apache.hadoop.hbase.protobuf.generated.RPCProtos.CellBlockMeta.Builder, org.apache.hadoop.hbase.protobuf.generated.RPCProtos.CellBlockMetaOrBuilder> cellBlockMetaBuilder_;
@ -3269,7 +3269,7 @@ public final class RPCProtos {
public interface ResponseHeaderOrBuilder public interface ResponseHeaderOrBuilder
extends com.google.protobuf.MessageOrBuilder { extends com.google.protobuf.MessageOrBuilder {
// optional uint32 callId = 1; // optional uint32 call_id = 1;
boolean hasCallId(); boolean hasCallId();
int getCallId(); int getCallId();
@ -3278,7 +3278,7 @@ public final class RPCProtos {
org.apache.hadoop.hbase.protobuf.generated.RPCProtos.ExceptionResponse getException(); org.apache.hadoop.hbase.protobuf.generated.RPCProtos.ExceptionResponse getException();
org.apache.hadoop.hbase.protobuf.generated.RPCProtos.ExceptionResponseOrBuilder getExceptionOrBuilder(); org.apache.hadoop.hbase.protobuf.generated.RPCProtos.ExceptionResponseOrBuilder getExceptionOrBuilder();
// optional .CellBlockMeta cellBlockMeta = 3; // optional .CellBlockMeta cell_block_meta = 3;
boolean hasCellBlockMeta(); boolean hasCellBlockMeta();
org.apache.hadoop.hbase.protobuf.generated.RPCProtos.CellBlockMeta getCellBlockMeta(); org.apache.hadoop.hbase.protobuf.generated.RPCProtos.CellBlockMeta getCellBlockMeta();
org.apache.hadoop.hbase.protobuf.generated.RPCProtos.CellBlockMetaOrBuilder getCellBlockMetaOrBuilder(); org.apache.hadoop.hbase.protobuf.generated.RPCProtos.CellBlockMetaOrBuilder getCellBlockMetaOrBuilder();
@ -3312,8 +3312,8 @@ public final class RPCProtos {
} }
private int bitField0_; private int bitField0_;
// optional uint32 callId = 1; // optional uint32 call_id = 1;
public static final int CALLID_FIELD_NUMBER = 1; public static final int CALL_ID_FIELD_NUMBER = 1;
private int callId_; private int callId_;
public boolean hasCallId() { public boolean hasCallId() {
return ((bitField0_ & 0x00000001) == 0x00000001); return ((bitField0_ & 0x00000001) == 0x00000001);
@ -3335,8 +3335,8 @@ public final class RPCProtos {
return exception_; return exception_;
} }
// optional .CellBlockMeta cellBlockMeta = 3; // optional .CellBlockMeta cell_block_meta = 3;
public static final int CELLBLOCKMETA_FIELD_NUMBER = 3; public static final int CELL_BLOCK_META_FIELD_NUMBER = 3;
private org.apache.hadoop.hbase.protobuf.generated.RPCProtos.CellBlockMeta cellBlockMeta_; private org.apache.hadoop.hbase.protobuf.generated.RPCProtos.CellBlockMeta cellBlockMeta_;
public boolean hasCellBlockMeta() { public boolean hasCellBlockMeta() {
return ((bitField0_ & 0x00000004) == 0x00000004); return ((bitField0_ & 0x00000004) == 0x00000004);
@ -3443,7 +3443,7 @@ public final class RPCProtos {
int hash = 41; int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode(); hash = (19 * hash) + getDescriptorForType().hashCode();
if (hasCallId()) { if (hasCallId()) {
hash = (37 * hash) + CALLID_FIELD_NUMBER; hash = (37 * hash) + CALL_ID_FIELD_NUMBER;
hash = (53 * hash) + getCallId(); hash = (53 * hash) + getCallId();
} }
if (hasException()) { if (hasException()) {
@ -3451,7 +3451,7 @@ public final class RPCProtos {
hash = (53 * hash) + getException().hashCode(); hash = (53 * hash) + getException().hashCode();
} }
if (hasCellBlockMeta()) { if (hasCellBlockMeta()) {
hash = (37 * hash) + CELLBLOCKMETA_FIELD_NUMBER; hash = (37 * hash) + CELL_BLOCK_META_FIELD_NUMBER;
hash = (53 * hash) + getCellBlockMeta().hashCode(); hash = (53 * hash) + getCellBlockMeta().hashCode();
} }
hash = (29 * hash) + getUnknownFields().hashCode(); hash = (29 * hash) + getUnknownFields().hashCode();
@ -3729,7 +3729,7 @@ public final class RPCProtos {
private int bitField0_; private int bitField0_;
// optional uint32 callId = 1; // optional uint32 call_id = 1;
private int callId_ ; private int callId_ ;
public boolean hasCallId() { public boolean hasCallId() {
return ((bitField0_ & 0x00000001) == 0x00000001); return ((bitField0_ & 0x00000001) == 0x00000001);
@ -3840,7 +3840,7 @@ public final class RPCProtos {
return exceptionBuilder_; return exceptionBuilder_;
} }
// optional .CellBlockMeta cellBlockMeta = 3; // optional .CellBlockMeta cell_block_meta = 3;
private org.apache.hadoop.hbase.protobuf.generated.RPCProtos.CellBlockMeta cellBlockMeta_ = org.apache.hadoop.hbase.protobuf.generated.RPCProtos.CellBlockMeta.getDefaultInstance(); private org.apache.hadoop.hbase.protobuf.generated.RPCProtos.CellBlockMeta cellBlockMeta_ = org.apache.hadoop.hbase.protobuf.generated.RPCProtos.CellBlockMeta.getDefaultInstance();
private com.google.protobuf.SingleFieldBuilder< private com.google.protobuf.SingleFieldBuilder<
org.apache.hadoop.hbase.protobuf.generated.RPCProtos.CellBlockMeta, org.apache.hadoop.hbase.protobuf.generated.RPCProtos.CellBlockMeta.Builder, org.apache.hadoop.hbase.protobuf.generated.RPCProtos.CellBlockMetaOrBuilder> cellBlockMetaBuilder_; org.apache.hadoop.hbase.protobuf.generated.RPCProtos.CellBlockMeta, org.apache.hadoop.hbase.protobuf.generated.RPCProtos.CellBlockMeta.Builder, org.apache.hadoop.hbase.protobuf.generated.RPCProtos.CellBlockMetaOrBuilder> cellBlockMetaBuilder_;
@ -3981,24 +3981,25 @@ public final class RPCProtos {
static { static {
java.lang.String[] descriptorData = { java.lang.String[] descriptorData = {
"\n\tRPC.proto\032\rTracing.proto\032\013hbase.proto\"" + "\n\tRPC.proto\032\rTracing.proto\032\013hbase.proto\"" +
":\n\017UserInformation\022\025\n\reffectiveUser\030\001 \002(" + "<\n\017UserInformation\022\026\n\016effective_user\030\001 \002" +
"\t\022\020\n\010realUser\030\002 \001(\t\"\267\001\n\020ConnectionHeader" + "(\t\022\021\n\treal_user\030\002 \001(\t\"\277\001\n\020ConnectionHead" +
"\022\"\n\010userInfo\030\001 \001(\0132\020.UserInformation\022\023\n\013" + "er\022#\n\tuser_info\030\001 \001(\0132\020.UserInformation\022" +
"serviceName\030\002 \001(\t\022H\n\023cellBlockCodecClass" + "\024\n\014service_name\030\002 \001(\t\022K\n\026cell_block_code" +
"\030\003 \001(\t:+org.apache.hadoop.hbase.codec.Ke" + "c_class\030\003 \001(\t:+org.apache.hadoop.hbase.c" +
"yValueCodec\022 \n\030cellBlockCompressorClass\030" + "odec.KeyValueCodec\022#\n\033cell_block_compres" +
"\004 \001(\t\"\037\n\rCellBlockMeta\022\016\n\006length\030\001 \001(\r\"w" + "sor_class\030\004 \001(\t\"\037\n\rCellBlockMeta\022\016\n\006leng" +
"\n\021ExceptionResponse\022\032\n\022exceptionClassNam" + "th\030\001 \001(\r\"|\n\021ExceptionResponse\022\034\n\024excepti" +
"e\030\001 \001(\t\022\022\n\nstackTrace\030\002 \001(\t\022\020\n\010hostname\030", "on_class_name\030\001 \001(\t\022\023\n\013stack_trace\030\002 \001(\t",
"\003 \001(\t\022\014\n\004port\030\004 \001(\005\022\022\n\ndoNotRetry\030\005 \001(\010\"" + "\022\020\n\010hostname\030\003 \001(\t\022\014\n\004port\030\004 \001(\005\022\024\n\014do_n" +
"\216\001\n\rRequestHeader\022\016\n\006callId\030\001 \001(\r\022\034\n\ttra" + "ot_retry\030\005 \001(\010\"\224\001\n\rRequestHeader\022\017\n\007call" +
"ceInfo\030\002 \001(\0132\t.RPCTInfo\022\022\n\nmethodName\030\003 " + "_id\030\001 \001(\r\022\035\n\ntrace_info\030\002 \001(\0132\t.RPCTInfo" +
"\001(\t\022\024\n\014requestParam\030\004 \001(\010\022%\n\rcellBlockMe" + "\022\023\n\013method_name\030\003 \001(\t\022\025\n\rrequest_param\030\004" +
"ta\030\005 \001(\0132\016.CellBlockMeta\"n\n\016ResponseHead" + " \001(\010\022\'\n\017cell_block_meta\030\005 \001(\0132\016.CellBloc" +
"er\022\016\n\006callId\030\001 \001(\r\022%\n\texception\030\002 \001(\0132\022." + "kMeta\"q\n\016ResponseHeader\022\017\n\007call_id\030\001 \001(\r" +
"ExceptionResponse\022%\n\rcellBlockMeta\030\003 \001(\013" + "\022%\n\texception\030\002 \001(\0132\022.ExceptionResponse\022" +
"2\016.CellBlockMetaB<\n*org.apache.hadoop.hb" + "\'\n\017cell_block_meta\030\003 \001(\0132\016.CellBlockMeta" +
"ase.protobuf.generatedB\tRPCProtosH\001\240\001\001" "B<\n*org.apache.hadoop.hbase.protobuf.gen" +
"eratedB\tRPCProtosH\001\240\001\001"
}; };
com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner = com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() { new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() {

View File

@ -15,11 +15,11 @@ public final class RegionServerStatusProtos {
boolean hasPort(); boolean hasPort();
int getPort(); int getPort();
// required uint64 serverStartCode = 2; // required uint64 server_start_code = 2;
boolean hasServerStartCode(); boolean hasServerStartCode();
long getServerStartCode(); long getServerStartCode();
// required uint64 serverCurrentTime = 3; // required uint64 server_current_time = 3;
boolean hasServerCurrentTime(); boolean hasServerCurrentTime();
long getServerCurrentTime(); long getServerCurrentTime();
} }
@ -62,8 +62,8 @@ public final class RegionServerStatusProtos {
return port_; return port_;
} }
// required uint64 serverStartCode = 2; // required uint64 server_start_code = 2;
public static final int SERVERSTARTCODE_FIELD_NUMBER = 2; public static final int SERVER_START_CODE_FIELD_NUMBER = 2;
private long serverStartCode_; private long serverStartCode_;
public boolean hasServerStartCode() { public boolean hasServerStartCode() {
return ((bitField0_ & 0x00000002) == 0x00000002); return ((bitField0_ & 0x00000002) == 0x00000002);
@ -72,8 +72,8 @@ public final class RegionServerStatusProtos {
return serverStartCode_; return serverStartCode_;
} }
// required uint64 serverCurrentTime = 3; // required uint64 server_current_time = 3;
public static final int SERVERCURRENTTIME_FIELD_NUMBER = 3; public static final int SERVER_CURRENT_TIME_FIELD_NUMBER = 3;
private long serverCurrentTime_; private long serverCurrentTime_;
public boolean hasServerCurrentTime() { public boolean hasServerCurrentTime() {
return ((bitField0_ & 0x00000004) == 0x00000004); return ((bitField0_ & 0x00000004) == 0x00000004);
@ -193,11 +193,11 @@ public final class RegionServerStatusProtos {
hash = (53 * hash) + getPort(); hash = (53 * hash) + getPort();
} }
if (hasServerStartCode()) { if (hasServerStartCode()) {
hash = (37 * hash) + SERVERSTARTCODE_FIELD_NUMBER; hash = (37 * hash) + SERVER_START_CODE_FIELD_NUMBER;
hash = (53 * hash) + hashLong(getServerStartCode()); hash = (53 * hash) + hashLong(getServerStartCode());
} }
if (hasServerCurrentTime()) { if (hasServerCurrentTime()) {
hash = (37 * hash) + SERVERCURRENTTIME_FIELD_NUMBER; hash = (37 * hash) + SERVER_CURRENT_TIME_FIELD_NUMBER;
hash = (53 * hash) + hashLong(getServerCurrentTime()); hash = (53 * hash) + hashLong(getServerCurrentTime());
} }
hash = (29 * hash) + getUnknownFields().hashCode(); hash = (29 * hash) + getUnknownFields().hashCode();
@ -482,7 +482,7 @@ public final class RegionServerStatusProtos {
return this; return this;
} }
// required uint64 serverStartCode = 2; // required uint64 server_start_code = 2;
private long serverStartCode_ ; private long serverStartCode_ ;
public boolean hasServerStartCode() { public boolean hasServerStartCode() {
return ((bitField0_ & 0x00000002) == 0x00000002); return ((bitField0_ & 0x00000002) == 0x00000002);
@ -503,7 +503,7 @@ public final class RegionServerStatusProtos {
return this; return this;
} }
// required uint64 serverCurrentTime = 3; // required uint64 server_current_time = 3;
private long serverCurrentTime_ ; private long serverCurrentTime_ ;
public boolean hasServerCurrentTime() { public boolean hasServerCurrentTime() {
return ((bitField0_ & 0x00000004) == 0x00000004); return ((bitField0_ & 0x00000004) == 0x00000004);
@ -538,7 +538,7 @@ public final class RegionServerStatusProtos {
public interface RegionServerStartupResponseOrBuilder public interface RegionServerStartupResponseOrBuilder
extends com.google.protobuf.MessageOrBuilder { extends com.google.protobuf.MessageOrBuilder {
// repeated .NameStringPair mapEntries = 1; // repeated .NameStringPair map_entries = 1;
java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPair> java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPair>
getMapEntriesList(); getMapEntriesList();
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPair getMapEntries(int index); org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPair getMapEntries(int index);
@ -576,8 +576,8 @@ public final class RegionServerStatusProtos {
return org.apache.hadoop.hbase.protobuf.generated.RegionServerStatusProtos.internal_static_RegionServerStartupResponse_fieldAccessorTable; return org.apache.hadoop.hbase.protobuf.generated.RegionServerStatusProtos.internal_static_RegionServerStartupResponse_fieldAccessorTable;
} }
// repeated .NameStringPair mapEntries = 1; // repeated .NameStringPair map_entries = 1;
public static final int MAPENTRIES_FIELD_NUMBER = 1; public static final int MAP_ENTRIES_FIELD_NUMBER = 1;
private java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPair> mapEntries_; private java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPair> mapEntries_;
public java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPair> getMapEntriesList() { public java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPair> getMapEntriesList() {
return mapEntries_; return mapEntries_;
@ -669,7 +669,7 @@ public final class RegionServerStatusProtos {
int hash = 41; int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode(); hash = (19 * hash) + getDescriptorForType().hashCode();
if (getMapEntriesCount() > 0) { if (getMapEntriesCount() > 0) {
hash = (37 * hash) + MAPENTRIES_FIELD_NUMBER; hash = (37 * hash) + MAP_ENTRIES_FIELD_NUMBER;
hash = (53 * hash) + getMapEntriesList().hashCode(); hash = (53 * hash) + getMapEntriesList().hashCode();
} }
hash = (29 * hash) + getUnknownFields().hashCode(); hash = (29 * hash) + getUnknownFields().hashCode();
@ -931,7 +931,7 @@ public final class RegionServerStatusProtos {
private int bitField0_; private int bitField0_;
// repeated .NameStringPair mapEntries = 1; // repeated .NameStringPair map_entries = 1;
private java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPair> mapEntries_ = private java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPair> mapEntries_ =
java.util.Collections.emptyList(); java.util.Collections.emptyList();
private void ensureMapEntriesIsMutable() { private void ensureMapEntriesIsMutable() {
@ -2073,7 +2073,7 @@ public final class RegionServerStatusProtos {
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName getServer(); org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName getServer();
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerNameOrBuilder getServerOrBuilder(); org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerNameOrBuilder getServerOrBuilder();
// required string errorMessage = 2; // required string error_message = 2;
boolean hasErrorMessage(); boolean hasErrorMessage();
String getErrorMessage(); String getErrorMessage();
} }
@ -2119,8 +2119,8 @@ public final class RegionServerStatusProtos {
return server_; return server_;
} }
// required string errorMessage = 2; // required string error_message = 2;
public static final int ERRORMESSAGE_FIELD_NUMBER = 2; public static final int ERROR_MESSAGE_FIELD_NUMBER = 2;
private java.lang.Object errorMessage_; private java.lang.Object errorMessage_;
public boolean hasErrorMessage() { public boolean hasErrorMessage() {
return ((bitField0_ & 0x00000002) == 0x00000002); return ((bitField0_ & 0x00000002) == 0x00000002);
@ -2249,7 +2249,7 @@ public final class RegionServerStatusProtos {
hash = (53 * hash) + getServer().hashCode(); hash = (53 * hash) + getServer().hashCode();
} }
if (hasErrorMessage()) { if (hasErrorMessage()) {
hash = (37 * hash) + ERRORMESSAGE_FIELD_NUMBER; hash = (37 * hash) + ERROR_MESSAGE_FIELD_NUMBER;
hash = (53 * hash) + getErrorMessage().hashCode(); hash = (53 * hash) + getErrorMessage().hashCode();
} }
hash = (29 * hash) + getUnknownFields().hashCode(); hash = (29 * hash) + getUnknownFields().hashCode();
@ -2602,7 +2602,7 @@ public final class RegionServerStatusProtos {
return serverBuilder_; return serverBuilder_;
} }
// required string errorMessage = 2; // required string error_message = 2;
private java.lang.Object errorMessage_ = ""; private java.lang.Object errorMessage_ = "";
public boolean hasErrorMessage() { public boolean hasErrorMessage() {
return ((bitField0_ & 0x00000002) == 0x00000002); return ((bitField0_ & 0x00000002) == 0x00000002);
@ -2952,7 +2952,7 @@ public final class RegionServerStatusProtos {
public interface GetLastFlushedSequenceIdRequestOrBuilder public interface GetLastFlushedSequenceIdRequestOrBuilder
extends com.google.protobuf.MessageOrBuilder { extends com.google.protobuf.MessageOrBuilder {
// required bytes regionName = 1; // required bytes region_name = 1;
boolean hasRegionName(); boolean hasRegionName();
com.google.protobuf.ByteString getRegionName(); com.google.protobuf.ByteString getRegionName();
} }
@ -2985,8 +2985,8 @@ public final class RegionServerStatusProtos {
} }
private int bitField0_; private int bitField0_;
// required bytes regionName = 1; // required bytes region_name = 1;
public static final int REGIONNAME_FIELD_NUMBER = 1; public static final int REGION_NAME_FIELD_NUMBER = 1;
private com.google.protobuf.ByteString regionName_; private com.google.protobuf.ByteString regionName_;
public boolean hasRegionName() { public boolean hasRegionName() {
return ((bitField0_ & 0x00000001) == 0x00000001); return ((bitField0_ & 0x00000001) == 0x00000001);
@ -3068,7 +3068,7 @@ public final class RegionServerStatusProtos {
int hash = 41; int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode(); hash = (19 * hash) + getDescriptorForType().hashCode();
if (hasRegionName()) { if (hasRegionName()) {
hash = (37 * hash) + REGIONNAME_FIELD_NUMBER; hash = (37 * hash) + REGION_NAME_FIELD_NUMBER;
hash = (53 * hash) + getRegionName().hashCode(); hash = (53 * hash) + getRegionName().hashCode();
} }
hash = (29 * hash) + getUnknownFields().hashCode(); hash = (29 * hash) + getUnknownFields().hashCode();
@ -3296,7 +3296,7 @@ public final class RegionServerStatusProtos {
private int bitField0_; private int bitField0_;
// required bytes regionName = 1; // required bytes region_name = 1;
private com.google.protobuf.ByteString regionName_ = com.google.protobuf.ByteString.EMPTY; private com.google.protobuf.ByteString regionName_ = com.google.protobuf.ByteString.EMPTY;
public boolean hasRegionName() { public boolean hasRegionName() {
return ((bitField0_ & 0x00000001) == 0x00000001); return ((bitField0_ & 0x00000001) == 0x00000001);
@ -3334,7 +3334,7 @@ public final class RegionServerStatusProtos {
public interface GetLastFlushedSequenceIdResponseOrBuilder public interface GetLastFlushedSequenceIdResponseOrBuilder
extends com.google.protobuf.MessageOrBuilder { extends com.google.protobuf.MessageOrBuilder {
// required uint64 lastFlushedSequenceId = 1; // required uint64 last_flushed_sequence_id = 1;
boolean hasLastFlushedSequenceId(); boolean hasLastFlushedSequenceId();
long getLastFlushedSequenceId(); long getLastFlushedSequenceId();
} }
@ -3367,8 +3367,8 @@ public final class RegionServerStatusProtos {
} }
private int bitField0_; private int bitField0_;
// required uint64 lastFlushedSequenceId = 1; // required uint64 last_flushed_sequence_id = 1;
public static final int LASTFLUSHEDSEQUENCEID_FIELD_NUMBER = 1; public static final int LAST_FLUSHED_SEQUENCE_ID_FIELD_NUMBER = 1;
private long lastFlushedSequenceId_; private long lastFlushedSequenceId_;
public boolean hasLastFlushedSequenceId() { public boolean hasLastFlushedSequenceId() {
return ((bitField0_ & 0x00000001) == 0x00000001); return ((bitField0_ & 0x00000001) == 0x00000001);
@ -3450,7 +3450,7 @@ public final class RegionServerStatusProtos {
int hash = 41; int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode(); hash = (19 * hash) + getDescriptorForType().hashCode();
if (hasLastFlushedSequenceId()) { if (hasLastFlushedSequenceId()) {
hash = (37 * hash) + LASTFLUSHEDSEQUENCEID_FIELD_NUMBER; hash = (37 * hash) + LAST_FLUSHED_SEQUENCE_ID_FIELD_NUMBER;
hash = (53 * hash) + hashLong(getLastFlushedSequenceId()); hash = (53 * hash) + hashLong(getLastFlushedSequenceId());
} }
hash = (29 * hash) + getUnknownFields().hashCode(); hash = (29 * hash) + getUnknownFields().hashCode();
@ -3678,7 +3678,7 @@ public final class RegionServerStatusProtos {
private int bitField0_; private int bitField0_;
// required uint64 lastFlushedSequenceId = 1; // required uint64 last_flushed_sequence_id = 1;
private long lastFlushedSequenceId_ ; private long lastFlushedSequenceId_ ;
public boolean hasLastFlushedSequenceId() { public boolean hasLastFlushedSequenceId() {
return ((bitField0_ & 0x00000001) == 0x00000001); return ((bitField0_ & 0x00000001) == 0x00000001);
@ -4176,31 +4176,31 @@ public final class RegionServerStatusProtos {
static { static {
java.lang.String[] descriptorData = { java.lang.String[] descriptorData = {
"\n\030RegionServerStatus.proto\032\013hbase.proto\032" + "\n\030RegionServerStatus.proto\032\013hbase.proto\032" +
"\023ClusterStatus.proto\"^\n\032RegionServerStar" + "\023ClusterStatus.proto\"b\n\032RegionServerStar" +
"tupRequest\022\014\n\004port\030\001 \002(\r\022\027\n\017serverStartC" + "tupRequest\022\014\n\004port\030\001 \002(\r\022\031\n\021server_start" +
"ode\030\002 \002(\004\022\031\n\021serverCurrentTime\030\003 \002(\004\"B\n\033" + "_code\030\002 \002(\004\022\033\n\023server_current_time\030\003 \002(\004" +
"RegionServerStartupResponse\022#\n\nmapEntrie" + "\"C\n\033RegionServerStartupResponse\022$\n\013map_e" +
"s\030\001 \003(\0132\017.NameStringPair\"S\n\031RegionServer" + "ntries\030\001 \003(\0132\017.NameStringPair\"S\n\031RegionS" +
"ReportRequest\022\033\n\006server\030\001 \002(\0132\013.ServerNa" + "erverReportRequest\022\033\n\006server\030\001 \002(\0132\013.Ser" +
"me\022\031\n\004load\030\002 \001(\0132\013.ServerLoad\"\034\n\032RegionS" + "verName\022\031\n\004load\030\002 \001(\0132\013.ServerLoad\"\034\n\032Re" +
"erverReportResponse\"N\n\031ReportRSFatalErro" + "gionServerReportResponse\"O\n\031ReportRSFata" +
"rRequest\022\033\n\006server\030\001 \002(\0132\013.ServerName\022\024\n", "lErrorRequest\022\033\n\006server\030\001 \002(\0132\013.ServerNa",
"\014errorMessage\030\002 \002(\t\"\034\n\032ReportRSFatalErro" + "me\022\025\n\rerror_message\030\002 \002(\t\"\034\n\032ReportRSFat" +
"rResponse\"5\n\037GetLastFlushedSequenceIdReq" + "alErrorResponse\"6\n\037GetLastFlushedSequenc" +
"uest\022\022\n\nregionName\030\001 \002(\014\"A\n GetLastFlush" + "eIdRequest\022\023\n\013region_name\030\001 \002(\014\"D\n GetLa" +
"edSequenceIdResponse\022\035\n\025lastFlushedSeque" + "stFlushedSequenceIdResponse\022 \n\030last_flus" +
"nceId\030\001 \002(\0042\354\002\n\031RegionServerStatusServic" + "hed_sequence_id\030\001 \002(\0042\354\002\n\031RegionServerSt" +
"e\022P\n\023regionServerStartup\022\033.RegionServerS" + "atusService\022P\n\023RegionServerStartup\022\033.Reg" +
"tartupRequest\032\034.RegionServerStartupRespo" + "ionServerStartupRequest\032\034.RegionServerSt" +
"nse\022M\n\022regionServerReport\022\032.RegionServer" + "artupResponse\022M\n\022RegionServerReport\022\032.Re" +
"ReportRequest\032\033.RegionServerReportRespon" + "gionServerReportRequest\032\033.RegionServerRe" +
"se\022M\n\022reportRSFatalError\022\032.ReportRSFatal", "portResponse\022M\n\022ReportRSFatalError\022\032.Rep",
"ErrorRequest\032\033.ReportRSFatalErrorRespons" + "ortRSFatalErrorRequest\032\033.ReportRSFatalEr" +
"e\022_\n\030getLastFlushedSequenceId\022 .GetLastF" + "rorResponse\022_\n\030GetLastFlushedSequenceId\022" +
"lushedSequenceIdRequest\032!.GetLastFlushed" + " .GetLastFlushedSequenceIdRequest\032!.GetL" +
"SequenceIdResponseBN\n*org.apache.hadoop." + "astFlushedSequenceIdResponseBN\n*org.apac" +
"hbase.protobuf.generatedB\030RegionServerSt" + "he.hadoop.hbase.protobuf.generatedB\030Regi" +
"atusProtosH\001\210\001\001\240\001\001" "onServerStatusProtosH\001\210\001\001\240\001\001"
}; };
com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner = com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() { new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() {

View File

@ -11,15 +11,15 @@ public final class RowProcessorProtos {
public interface RowProcessorRequestOrBuilder public interface RowProcessorRequestOrBuilder
extends com.google.protobuf.MessageOrBuilder { extends com.google.protobuf.MessageOrBuilder {
// required string rowProcessorClassName = 1; // required string row_processor_class_name = 1;
boolean hasRowProcessorClassName(); boolean hasRowProcessorClassName();
String getRowProcessorClassName(); String getRowProcessorClassName();
// optional string rowProcessorInitializerMessageName = 2; // optional string row_processor_initializer_message_name = 2;
boolean hasRowProcessorInitializerMessageName(); boolean hasRowProcessorInitializerMessageName();
String getRowProcessorInitializerMessageName(); String getRowProcessorInitializerMessageName();
// optional bytes rowProcessorInitializerMessage = 3; // optional bytes row_processor_initializer_message = 3;
boolean hasRowProcessorInitializerMessage(); boolean hasRowProcessorInitializerMessage();
com.google.protobuf.ByteString getRowProcessorInitializerMessage(); com.google.protobuf.ByteString getRowProcessorInitializerMessage();
} }
@ -52,8 +52,8 @@ public final class RowProcessorProtos {
} }
private int bitField0_; private int bitField0_;
// required string rowProcessorClassName = 1; // required string row_processor_class_name = 1;
public static final int ROWPROCESSORCLASSNAME_FIELD_NUMBER = 1; public static final int ROW_PROCESSOR_CLASS_NAME_FIELD_NUMBER = 1;
private java.lang.Object rowProcessorClassName_; private java.lang.Object rowProcessorClassName_;
public boolean hasRowProcessorClassName() { public boolean hasRowProcessorClassName() {
return ((bitField0_ & 0x00000001) == 0x00000001); return ((bitField0_ & 0x00000001) == 0x00000001);
@ -84,8 +84,8 @@ public final class RowProcessorProtos {
} }
} }
// optional string rowProcessorInitializerMessageName = 2; // optional string row_processor_initializer_message_name = 2;
public static final int ROWPROCESSORINITIALIZERMESSAGENAME_FIELD_NUMBER = 2; public static final int ROW_PROCESSOR_INITIALIZER_MESSAGE_NAME_FIELD_NUMBER = 2;
private java.lang.Object rowProcessorInitializerMessageName_; private java.lang.Object rowProcessorInitializerMessageName_;
public boolean hasRowProcessorInitializerMessageName() { public boolean hasRowProcessorInitializerMessageName() {
return ((bitField0_ & 0x00000002) == 0x00000002); return ((bitField0_ & 0x00000002) == 0x00000002);
@ -116,8 +116,8 @@ public final class RowProcessorProtos {
} }
} }
// optional bytes rowProcessorInitializerMessage = 3; // optional bytes row_processor_initializer_message = 3;
public static final int ROWPROCESSORINITIALIZERMESSAGE_FIELD_NUMBER = 3; public static final int ROW_PROCESSOR_INITIALIZER_MESSAGE_FIELD_NUMBER = 3;
private com.google.protobuf.ByteString rowProcessorInitializerMessage_; private com.google.protobuf.ByteString rowProcessorInitializerMessage_;
public boolean hasRowProcessorInitializerMessage() { public boolean hasRowProcessorInitializerMessage() {
return ((bitField0_ & 0x00000004) == 0x00000004); return ((bitField0_ & 0x00000004) == 0x00000004);
@ -225,15 +225,15 @@ public final class RowProcessorProtos {
int hash = 41; int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode(); hash = (19 * hash) + getDescriptorForType().hashCode();
if (hasRowProcessorClassName()) { if (hasRowProcessorClassName()) {
hash = (37 * hash) + ROWPROCESSORCLASSNAME_FIELD_NUMBER; hash = (37 * hash) + ROW_PROCESSOR_CLASS_NAME_FIELD_NUMBER;
hash = (53 * hash) + getRowProcessorClassName().hashCode(); hash = (53 * hash) + getRowProcessorClassName().hashCode();
} }
if (hasRowProcessorInitializerMessageName()) { if (hasRowProcessorInitializerMessageName()) {
hash = (37 * hash) + ROWPROCESSORINITIALIZERMESSAGENAME_FIELD_NUMBER; hash = (37 * hash) + ROW_PROCESSOR_INITIALIZER_MESSAGE_NAME_FIELD_NUMBER;
hash = (53 * hash) + getRowProcessorInitializerMessageName().hashCode(); hash = (53 * hash) + getRowProcessorInitializerMessageName().hashCode();
} }
if (hasRowProcessorInitializerMessage()) { if (hasRowProcessorInitializerMessage()) {
hash = (37 * hash) + ROWPROCESSORINITIALIZERMESSAGE_FIELD_NUMBER; hash = (37 * hash) + ROW_PROCESSOR_INITIALIZER_MESSAGE_FIELD_NUMBER;
hash = (53 * hash) + getRowProcessorInitializerMessage().hashCode(); hash = (53 * hash) + getRowProcessorInitializerMessage().hashCode();
} }
hash = (29 * hash) + getUnknownFields().hashCode(); hash = (29 * hash) + getUnknownFields().hashCode();
@ -489,7 +489,7 @@ public final class RowProcessorProtos {
private int bitField0_; private int bitField0_;
// required string rowProcessorClassName = 1; // required string row_processor_class_name = 1;
private java.lang.Object rowProcessorClassName_ = ""; private java.lang.Object rowProcessorClassName_ = "";
public boolean hasRowProcessorClassName() { public boolean hasRowProcessorClassName() {
return ((bitField0_ & 0x00000001) == 0x00000001); return ((bitField0_ & 0x00000001) == 0x00000001);
@ -525,7 +525,7 @@ public final class RowProcessorProtos {
onChanged(); onChanged();
} }
// optional string rowProcessorInitializerMessageName = 2; // optional string row_processor_initializer_message_name = 2;
private java.lang.Object rowProcessorInitializerMessageName_ = ""; private java.lang.Object rowProcessorInitializerMessageName_ = "";
public boolean hasRowProcessorInitializerMessageName() { public boolean hasRowProcessorInitializerMessageName() {
return ((bitField0_ & 0x00000002) == 0x00000002); return ((bitField0_ & 0x00000002) == 0x00000002);
@ -561,7 +561,7 @@ public final class RowProcessorProtos {
onChanged(); onChanged();
} }
// optional bytes rowProcessorInitializerMessage = 3; // optional bytes row_processor_initializer_message = 3;
private com.google.protobuf.ByteString rowProcessorInitializerMessage_ = com.google.protobuf.ByteString.EMPTY; private com.google.protobuf.ByteString rowProcessorInitializerMessage_ = com.google.protobuf.ByteString.EMPTY;
public boolean hasRowProcessorInitializerMessage() { public boolean hasRowProcessorInitializerMessage() {
return ((bitField0_ & 0x00000004) == 0x00000004); return ((bitField0_ & 0x00000004) == 0x00000004);
@ -599,7 +599,7 @@ public final class RowProcessorProtos {
public interface RowProcessorResultOrBuilder public interface RowProcessorResultOrBuilder
extends com.google.protobuf.MessageOrBuilder { extends com.google.protobuf.MessageOrBuilder {
// required bytes rowProcessorResult = 1; // required bytes row_processor_result = 1;
boolean hasRowProcessorResult(); boolean hasRowProcessorResult();
com.google.protobuf.ByteString getRowProcessorResult(); com.google.protobuf.ByteString getRowProcessorResult();
} }
@ -632,8 +632,8 @@ public final class RowProcessorProtos {
} }
private int bitField0_; private int bitField0_;
// required bytes rowProcessorResult = 1; // required bytes row_processor_result = 1;
public static final int ROWPROCESSORRESULT_FIELD_NUMBER = 1; public static final int ROW_PROCESSOR_RESULT_FIELD_NUMBER = 1;
private com.google.protobuf.ByteString rowProcessorResult_; private com.google.protobuf.ByteString rowProcessorResult_;
public boolean hasRowProcessorResult() { public boolean hasRowProcessorResult() {
return ((bitField0_ & 0x00000001) == 0x00000001); return ((bitField0_ & 0x00000001) == 0x00000001);
@ -715,7 +715,7 @@ public final class RowProcessorProtos {
int hash = 41; int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode(); hash = (19 * hash) + getDescriptorForType().hashCode();
if (hasRowProcessorResult()) { if (hasRowProcessorResult()) {
hash = (37 * hash) + ROWPROCESSORRESULT_FIELD_NUMBER; hash = (37 * hash) + ROW_PROCESSOR_RESULT_FIELD_NUMBER;
hash = (53 * hash) + getRowProcessorResult().hashCode(); hash = (53 * hash) + getRowProcessorResult().hashCode();
} }
hash = (29 * hash) + getUnknownFields().hashCode(); hash = (29 * hash) + getUnknownFields().hashCode();
@ -943,7 +943,7 @@ public final class RowProcessorProtos {
private int bitField0_; private int bitField0_;
// required bytes rowProcessorResult = 1; // required bytes row_processor_result = 1;
private com.google.protobuf.ByteString rowProcessorResult_ = com.google.protobuf.ByteString.EMPTY; private com.google.protobuf.ByteString rowProcessorResult_ = com.google.protobuf.ByteString.EMPTY;
public boolean hasRowProcessorResult() { public boolean hasRowProcessorResult() {
return ((bitField0_ & 0x00000001) == 0x00000001); return ((bitField0_ & 0x00000001) == 0x00000001);
@ -1218,15 +1218,16 @@ public final class RowProcessorProtos {
descriptor; descriptor;
static { static {
java.lang.String[] descriptorData = { java.lang.String[] descriptorData = {
"\n\022RowProcessor.proto\"\210\001\n\023RowProcessorReq" + "\n\022RowProcessor.proto\"\222\001\n\023RowProcessorReq" +
"uest\022\035\n\025rowProcessorClassName\030\001 \002(\t\022*\n\"r" + "uest\022 \n\030row_processor_class_name\030\001 \002(\t\022." +
"owProcessorInitializerMessageName\030\002 \001(\t\022" + "\n&row_processor_initializer_message_name" +
"&\n\036rowProcessorInitializerMessage\030\003 \001(\014\"" + "\030\002 \001(\t\022)\n!row_processor_initializer_mess" +
"0\n\022RowProcessorResult\022\032\n\022rowProcessorRes" + "age\030\003 \001(\014\"2\n\022RowProcessorResult\022\034\n\024row_p" +
"ult\030\001 \002(\0142K\n\023RowProcessorService\0224\n\007proc" + "rocessor_result\030\001 \002(\0142K\n\023RowProcessorSer" +
"ess\022\024.RowProcessorRequest\032\023.RowProcessor" + "vice\0224\n\007Process\022\024.RowProcessorRequest\032\023." +
"ResultBH\n*org.apache.hadoop.hbase.protob" + "RowProcessorResultBH\n*org.apache.hadoop." +
"uf.generatedB\022RowProcessorProtosH\001\210\001\001\240\001\001" "hbase.protobuf.generatedB\022RowProcessorPr" +
"otosH\001\210\001\001\240\001\001"
}; };
com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner = com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() { new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() {

View File

@ -11,7 +11,7 @@ public final class SecureBulkLoadProtos {
public interface SecureBulkLoadHFilesRequestOrBuilder public interface SecureBulkLoadHFilesRequestOrBuilder
extends com.google.protobuf.MessageOrBuilder { extends com.google.protobuf.MessageOrBuilder {
// repeated .BulkLoadHFileRequest.FamilyPath familyPath = 1; // repeated .BulkLoadHFileRequest.FamilyPath family_path = 1;
java.util.List<org.apache.hadoop.hbase.protobuf.generated.ClientProtos.BulkLoadHFileRequest.FamilyPath> java.util.List<org.apache.hadoop.hbase.protobuf.generated.ClientProtos.BulkLoadHFileRequest.FamilyPath>
getFamilyPathList(); getFamilyPathList();
org.apache.hadoop.hbase.protobuf.generated.ClientProtos.BulkLoadHFileRequest.FamilyPath getFamilyPath(int index); org.apache.hadoop.hbase.protobuf.generated.ClientProtos.BulkLoadHFileRequest.FamilyPath getFamilyPath(int index);
@ -21,16 +21,16 @@ public final class SecureBulkLoadProtos {
org.apache.hadoop.hbase.protobuf.generated.ClientProtos.BulkLoadHFileRequest.FamilyPathOrBuilder getFamilyPathOrBuilder( org.apache.hadoop.hbase.protobuf.generated.ClientProtos.BulkLoadHFileRequest.FamilyPathOrBuilder getFamilyPathOrBuilder(
int index); int index);
// optional bool assignSeqNum = 2; // optional bool assign_seq_num = 2;
boolean hasAssignSeqNum(); boolean hasAssignSeqNum();
boolean getAssignSeqNum(); boolean getAssignSeqNum();
// required .DelegationTokenProto fsToken = 3; // required .DelegationTokenProto fs_token = 3;
boolean hasFsToken(); boolean hasFsToken();
org.apache.hadoop.hbase.protobuf.generated.SecureBulkLoadProtos.DelegationTokenProto getFsToken(); org.apache.hadoop.hbase.protobuf.generated.SecureBulkLoadProtos.DelegationTokenProto getFsToken();
org.apache.hadoop.hbase.protobuf.generated.SecureBulkLoadProtos.DelegationTokenProtoOrBuilder getFsTokenOrBuilder(); org.apache.hadoop.hbase.protobuf.generated.SecureBulkLoadProtos.DelegationTokenProtoOrBuilder getFsTokenOrBuilder();
// required string bulkToken = 4; // required string bulk_token = 4;
boolean hasBulkToken(); boolean hasBulkToken();
String getBulkToken(); String getBulkToken();
} }
@ -63,8 +63,8 @@ public final class SecureBulkLoadProtos {
} }
private int bitField0_; private int bitField0_;
// repeated .BulkLoadHFileRequest.FamilyPath familyPath = 1; // repeated .BulkLoadHFileRequest.FamilyPath family_path = 1;
public static final int FAMILYPATH_FIELD_NUMBER = 1; public static final int FAMILY_PATH_FIELD_NUMBER = 1;
private java.util.List<org.apache.hadoop.hbase.protobuf.generated.ClientProtos.BulkLoadHFileRequest.FamilyPath> familyPath_; private java.util.List<org.apache.hadoop.hbase.protobuf.generated.ClientProtos.BulkLoadHFileRequest.FamilyPath> familyPath_;
public java.util.List<org.apache.hadoop.hbase.protobuf.generated.ClientProtos.BulkLoadHFileRequest.FamilyPath> getFamilyPathList() { public java.util.List<org.apache.hadoop.hbase.protobuf.generated.ClientProtos.BulkLoadHFileRequest.FamilyPath> getFamilyPathList() {
return familyPath_; return familyPath_;
@ -84,8 +84,8 @@ public final class SecureBulkLoadProtos {
return familyPath_.get(index); return familyPath_.get(index);
} }
// optional bool assignSeqNum = 2; // optional bool assign_seq_num = 2;
public static final int ASSIGNSEQNUM_FIELD_NUMBER = 2; public static final int ASSIGN_SEQ_NUM_FIELD_NUMBER = 2;
private boolean assignSeqNum_; private boolean assignSeqNum_;
public boolean hasAssignSeqNum() { public boolean hasAssignSeqNum() {
return ((bitField0_ & 0x00000001) == 0x00000001); return ((bitField0_ & 0x00000001) == 0x00000001);
@ -94,8 +94,8 @@ public final class SecureBulkLoadProtos {
return assignSeqNum_; return assignSeqNum_;
} }
// required .DelegationTokenProto fsToken = 3; // required .DelegationTokenProto fs_token = 3;
public static final int FSTOKEN_FIELD_NUMBER = 3; public static final int FS_TOKEN_FIELD_NUMBER = 3;
private org.apache.hadoop.hbase.protobuf.generated.SecureBulkLoadProtos.DelegationTokenProto fsToken_; private org.apache.hadoop.hbase.protobuf.generated.SecureBulkLoadProtos.DelegationTokenProto fsToken_;
public boolean hasFsToken() { public boolean hasFsToken() {
return ((bitField0_ & 0x00000002) == 0x00000002); return ((bitField0_ & 0x00000002) == 0x00000002);
@ -107,8 +107,8 @@ public final class SecureBulkLoadProtos {
return fsToken_; return fsToken_;
} }
// required string bulkToken = 4; // required string bulk_token = 4;
public static final int BULKTOKEN_FIELD_NUMBER = 4; public static final int BULK_TOKEN_FIELD_NUMBER = 4;
private java.lang.Object bulkToken_; private java.lang.Object bulkToken_;
public boolean hasBulkToken() { public boolean hasBulkToken() {
return ((bitField0_ & 0x00000004) == 0x00000004); return ((bitField0_ & 0x00000004) == 0x00000004);
@ -258,19 +258,19 @@ public final class SecureBulkLoadProtos {
int hash = 41; int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode(); hash = (19 * hash) + getDescriptorForType().hashCode();
if (getFamilyPathCount() > 0) { if (getFamilyPathCount() > 0) {
hash = (37 * hash) + FAMILYPATH_FIELD_NUMBER; hash = (37 * hash) + FAMILY_PATH_FIELD_NUMBER;
hash = (53 * hash) + getFamilyPathList().hashCode(); hash = (53 * hash) + getFamilyPathList().hashCode();
} }
if (hasAssignSeqNum()) { if (hasAssignSeqNum()) {
hash = (37 * hash) + ASSIGNSEQNUM_FIELD_NUMBER; hash = (37 * hash) + ASSIGN_SEQ_NUM_FIELD_NUMBER;
hash = (53 * hash) + hashBoolean(getAssignSeqNum()); hash = (53 * hash) + hashBoolean(getAssignSeqNum());
} }
if (hasFsToken()) { if (hasFsToken()) {
hash = (37 * hash) + FSTOKEN_FIELD_NUMBER; hash = (37 * hash) + FS_TOKEN_FIELD_NUMBER;
hash = (53 * hash) + getFsToken().hashCode(); hash = (53 * hash) + getFsToken().hashCode();
} }
if (hasBulkToken()) { if (hasBulkToken()) {
hash = (37 * hash) + BULKTOKEN_FIELD_NUMBER; hash = (37 * hash) + BULK_TOKEN_FIELD_NUMBER;
hash = (53 * hash) + getBulkToken().hashCode(); hash = (53 * hash) + getBulkToken().hashCode();
} }
hash = (29 * hash) + getUnknownFields().hashCode(); hash = (29 * hash) + getUnknownFields().hashCode();
@ -597,7 +597,7 @@ public final class SecureBulkLoadProtos {
private int bitField0_; private int bitField0_;
// repeated .BulkLoadHFileRequest.FamilyPath familyPath = 1; // repeated .BulkLoadHFileRequest.FamilyPath family_path = 1;
private java.util.List<org.apache.hadoop.hbase.protobuf.generated.ClientProtos.BulkLoadHFileRequest.FamilyPath> familyPath_ = private java.util.List<org.apache.hadoop.hbase.protobuf.generated.ClientProtos.BulkLoadHFileRequest.FamilyPath> familyPath_ =
java.util.Collections.emptyList(); java.util.Collections.emptyList();
private void ensureFamilyPathIsMutable() { private void ensureFamilyPathIsMutable() {
@ -783,7 +783,7 @@ public final class SecureBulkLoadProtos {
return familyPathBuilder_; return familyPathBuilder_;
} }
// optional bool assignSeqNum = 2; // optional bool assign_seq_num = 2;
private boolean assignSeqNum_ ; private boolean assignSeqNum_ ;
public boolean hasAssignSeqNum() { public boolean hasAssignSeqNum() {
return ((bitField0_ & 0x00000002) == 0x00000002); return ((bitField0_ & 0x00000002) == 0x00000002);
@ -804,7 +804,7 @@ public final class SecureBulkLoadProtos {
return this; return this;
} }
// required .DelegationTokenProto fsToken = 3; // required .DelegationTokenProto fs_token = 3;
private org.apache.hadoop.hbase.protobuf.generated.SecureBulkLoadProtos.DelegationTokenProto fsToken_ = org.apache.hadoop.hbase.protobuf.generated.SecureBulkLoadProtos.DelegationTokenProto.getDefaultInstance(); private org.apache.hadoop.hbase.protobuf.generated.SecureBulkLoadProtos.DelegationTokenProto fsToken_ = org.apache.hadoop.hbase.protobuf.generated.SecureBulkLoadProtos.DelegationTokenProto.getDefaultInstance();
private com.google.protobuf.SingleFieldBuilder< private com.google.protobuf.SingleFieldBuilder<
org.apache.hadoop.hbase.protobuf.generated.SecureBulkLoadProtos.DelegationTokenProto, org.apache.hadoop.hbase.protobuf.generated.SecureBulkLoadProtos.DelegationTokenProto.Builder, org.apache.hadoop.hbase.protobuf.generated.SecureBulkLoadProtos.DelegationTokenProtoOrBuilder> fsTokenBuilder_; org.apache.hadoop.hbase.protobuf.generated.SecureBulkLoadProtos.DelegationTokenProto, org.apache.hadoop.hbase.protobuf.generated.SecureBulkLoadProtos.DelegationTokenProto.Builder, org.apache.hadoop.hbase.protobuf.generated.SecureBulkLoadProtos.DelegationTokenProtoOrBuilder> fsTokenBuilder_;
@ -894,7 +894,7 @@ public final class SecureBulkLoadProtos {
return fsTokenBuilder_; return fsTokenBuilder_;
} }
// required string bulkToken = 4; // required string bulk_token = 4;
private java.lang.Object bulkToken_ = ""; private java.lang.Object bulkToken_ = "";
public boolean hasBulkToken() { public boolean hasBulkToken() {
return ((bitField0_ & 0x00000008) == 0x00000008); return ((bitField0_ & 0x00000008) == 0x00000008);
@ -1972,7 +1972,7 @@ public final class SecureBulkLoadProtos {
public interface PrepareBulkLoadRequestOrBuilder public interface PrepareBulkLoadRequestOrBuilder
extends com.google.protobuf.MessageOrBuilder { extends com.google.protobuf.MessageOrBuilder {
// required bytes tableName = 1; // required bytes table_name = 1;
boolean hasTableName(); boolean hasTableName();
com.google.protobuf.ByteString getTableName(); com.google.protobuf.ByteString getTableName();
} }
@ -2005,8 +2005,8 @@ public final class SecureBulkLoadProtos {
} }
private int bitField0_; private int bitField0_;
// required bytes tableName = 1; // required bytes table_name = 1;
public static final int TABLENAME_FIELD_NUMBER = 1; public static final int TABLE_NAME_FIELD_NUMBER = 1;
private com.google.protobuf.ByteString tableName_; private com.google.protobuf.ByteString tableName_;
public boolean hasTableName() { public boolean hasTableName() {
return ((bitField0_ & 0x00000001) == 0x00000001); return ((bitField0_ & 0x00000001) == 0x00000001);
@ -2088,7 +2088,7 @@ public final class SecureBulkLoadProtos {
int hash = 41; int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode(); hash = (19 * hash) + getDescriptorForType().hashCode();
if (hasTableName()) { if (hasTableName()) {
hash = (37 * hash) + TABLENAME_FIELD_NUMBER; hash = (37 * hash) + TABLE_NAME_FIELD_NUMBER;
hash = (53 * hash) + getTableName().hashCode(); hash = (53 * hash) + getTableName().hashCode();
} }
hash = (29 * hash) + getUnknownFields().hashCode(); hash = (29 * hash) + getUnknownFields().hashCode();
@ -2316,7 +2316,7 @@ public final class SecureBulkLoadProtos {
private int bitField0_; private int bitField0_;
// required bytes tableName = 1; // required bytes table_name = 1;
private com.google.protobuf.ByteString tableName_ = com.google.protobuf.ByteString.EMPTY; private com.google.protobuf.ByteString tableName_ = com.google.protobuf.ByteString.EMPTY;
public boolean hasTableName() { public boolean hasTableName() {
return ((bitField0_ & 0x00000001) == 0x00000001); return ((bitField0_ & 0x00000001) == 0x00000001);
@ -2354,7 +2354,7 @@ public final class SecureBulkLoadProtos {
public interface PrepareBulkLoadResponseOrBuilder public interface PrepareBulkLoadResponseOrBuilder
extends com.google.protobuf.MessageOrBuilder { extends com.google.protobuf.MessageOrBuilder {
// required string bulkToken = 1; // required string bulk_token = 1;
boolean hasBulkToken(); boolean hasBulkToken();
String getBulkToken(); String getBulkToken();
} }
@ -2387,8 +2387,8 @@ public final class SecureBulkLoadProtos {
} }
private int bitField0_; private int bitField0_;
// required string bulkToken = 1; // required string bulk_token = 1;
public static final int BULKTOKEN_FIELD_NUMBER = 1; public static final int BULK_TOKEN_FIELD_NUMBER = 1;
private java.lang.Object bulkToken_; private java.lang.Object bulkToken_;
public boolean hasBulkToken() { public boolean hasBulkToken() {
return ((bitField0_ & 0x00000001) == 0x00000001); return ((bitField0_ & 0x00000001) == 0x00000001);
@ -2492,7 +2492,7 @@ public final class SecureBulkLoadProtos {
int hash = 41; int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode(); hash = (19 * hash) + getDescriptorForType().hashCode();
if (hasBulkToken()) { if (hasBulkToken()) {
hash = (37 * hash) + BULKTOKEN_FIELD_NUMBER; hash = (37 * hash) + BULK_TOKEN_FIELD_NUMBER;
hash = (53 * hash) + getBulkToken().hashCode(); hash = (53 * hash) + getBulkToken().hashCode();
} }
hash = (29 * hash) + getUnknownFields().hashCode(); hash = (29 * hash) + getUnknownFields().hashCode();
@ -2720,7 +2720,7 @@ public final class SecureBulkLoadProtos {
private int bitField0_; private int bitField0_;
// required string bulkToken = 1; // required string bulk_token = 1;
private java.lang.Object bulkToken_ = ""; private java.lang.Object bulkToken_ = "";
public boolean hasBulkToken() { public boolean hasBulkToken() {
return ((bitField0_ & 0x00000001) == 0x00000001); return ((bitField0_ & 0x00000001) == 0x00000001);
@ -2770,7 +2770,7 @@ public final class SecureBulkLoadProtos {
public interface CleanupBulkLoadRequestOrBuilder public interface CleanupBulkLoadRequestOrBuilder
extends com.google.protobuf.MessageOrBuilder { extends com.google.protobuf.MessageOrBuilder {
// required string bulkToken = 1; // required string bulk_token = 1;
boolean hasBulkToken(); boolean hasBulkToken();
String getBulkToken(); String getBulkToken();
} }
@ -2803,8 +2803,8 @@ public final class SecureBulkLoadProtos {
} }
private int bitField0_; private int bitField0_;
// required string bulkToken = 1; // required string bulk_token = 1;
public static final int BULKTOKEN_FIELD_NUMBER = 1; public static final int BULK_TOKEN_FIELD_NUMBER = 1;
private java.lang.Object bulkToken_; private java.lang.Object bulkToken_;
public boolean hasBulkToken() { public boolean hasBulkToken() {
return ((bitField0_ & 0x00000001) == 0x00000001); return ((bitField0_ & 0x00000001) == 0x00000001);
@ -2908,7 +2908,7 @@ public final class SecureBulkLoadProtos {
int hash = 41; int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode(); hash = (19 * hash) + getDescriptorForType().hashCode();
if (hasBulkToken()) { if (hasBulkToken()) {
hash = (37 * hash) + BULKTOKEN_FIELD_NUMBER; hash = (37 * hash) + BULK_TOKEN_FIELD_NUMBER;
hash = (53 * hash) + getBulkToken().hashCode(); hash = (53 * hash) + getBulkToken().hashCode();
} }
hash = (29 * hash) + getUnknownFields().hashCode(); hash = (29 * hash) + getUnknownFields().hashCode();
@ -3136,7 +3136,7 @@ public final class SecureBulkLoadProtos {
private int bitField0_; private int bitField0_;
// required string bulkToken = 1; // required string bulk_token = 1;
private java.lang.Object bulkToken_ = ""; private java.lang.Object bulkToken_ = "";
public boolean hasBulkToken() { public boolean hasBulkToken() {
return ((bitField0_ & 0x00000001) == 0x00000001); return ((bitField0_ & 0x00000001) == 0x00000001);
@ -3879,27 +3879,27 @@ public final class SecureBulkLoadProtos {
static { static {
java.lang.String[] descriptorData = { java.lang.String[] descriptorData = {
"\n\024SecureBulkLoad.proto\032\013hbase.proto\032\014Cli" + "\n\024SecureBulkLoad.proto\032\013hbase.proto\032\014Cli" +
"ent.proto\"\244\001\n\033SecureBulkLoadHFilesReques" + "ent.proto\"\251\001\n\033SecureBulkLoadHFilesReques" +
"t\0224\n\nfamilyPath\030\001 \003(\0132 .BulkLoadHFileReq" + "t\0225\n\013family_path\030\001 \003(\0132 .BulkLoadHFileRe" +
"uest.FamilyPath\022\024\n\014assignSeqNum\030\002 \001(\010\022&\n" + "quest.FamilyPath\022\026\n\016assign_seq_num\030\002 \001(\010" +
"\007fsToken\030\003 \002(\0132\025.DelegationTokenProto\022\021\n" + "\022\'\n\010fs_token\030\003 \002(\0132\025.DelegationTokenProt" +
"\tbulkToken\030\004 \002(\t\".\n\034SecureBulkLoadHFiles" + "o\022\022\n\nbulk_token\030\004 \002(\t\".\n\034SecureBulkLoadH" +
"Response\022\016\n\006loaded\030\001 \002(\010\"[\n\024DelegationTo" + "FilesResponse\022\016\n\006loaded\030\001 \002(\010\"[\n\024Delegat" +
"kenProto\022\022\n\nidentifier\030\001 \001(\014\022\020\n\010password" + "ionTokenProto\022\022\n\nidentifier\030\001 \001(\014\022\020\n\010pas" +
"\030\002 \001(\014\022\014\n\004kind\030\003 \001(\t\022\017\n\007service\030\004 \001(\t\"+\n" + "sword\030\002 \001(\014\022\014\n\004kind\030\003 \001(\t\022\017\n\007service\030\004 \001" +
"\026PrepareBulkLoadRequest\022\021\n\ttableName\030\001 \002", "(\t\",\n\026PrepareBulkLoadRequest\022\022\n\ntable_na",
"(\014\",\n\027PrepareBulkLoadResponse\022\021\n\tbulkTok" + "me\030\001 \002(\014\"-\n\027PrepareBulkLoadResponse\022\022\n\nb" +
"en\030\001 \002(\t\"+\n\026CleanupBulkLoadRequest\022\021\n\tbu" + "ulk_token\030\001 \002(\t\",\n\026CleanupBulkLoadReques" +
"lkToken\030\001 \002(\t\"\031\n\027CleanupBulkLoadResponse" + "t\022\022\n\nbulk_token\030\001 \002(\t\"\031\n\027CleanupBulkLoad" +
"2\370\001\n\025SecureBulkLoadService\022D\n\017prepareBul" + "Response2\370\001\n\025SecureBulkLoadService\022D\n\017Pr" +
"kLoad\022\027.PrepareBulkLoadRequest\032\030.Prepare" + "epareBulkLoad\022\027.PrepareBulkLoadRequest\032\030" +
"BulkLoadResponse\022S\n\024secureBulkLoadHFiles" + ".PrepareBulkLoadResponse\022S\n\024SecureBulkLo" +
"\022\034.SecureBulkLoadHFilesRequest\032\035.SecureB" + "adHFiles\022\034.SecureBulkLoadHFilesRequest\032\035" +
"ulkLoadHFilesResponse\022D\n\017cleanupBulkLoad" + ".SecureBulkLoadHFilesResponse\022D\n\017Cleanup" +
"\022\027.CleanupBulkLoadRequest\032\030.CleanupBulkL" + "BulkLoad\022\027.CleanupBulkLoadRequest\032\030.Clea" +
"oadResponseBJ\n*org.apache.hadoop.hbase.p", "nupBulkLoadResponseBJ\n*org.apache.hadoop",
"rotobuf.generatedB\024SecureBulkLoadProtosH" + ".hbase.protobuf.generatedB\024SecureBulkLoa" +
"\001\210\001\001\240\001\001" "dProtosH\001\210\001\001\240\001\001"
}; };
com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner = com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() { new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() {

View File

@ -11,11 +11,11 @@ public final class Tracing {
public interface RPCTInfoOrBuilder public interface RPCTInfoOrBuilder
extends com.google.protobuf.MessageOrBuilder { extends com.google.protobuf.MessageOrBuilder {
// optional int64 traceId = 1; // optional int64 trace_id = 1;
boolean hasTraceId(); boolean hasTraceId();
long getTraceId(); long getTraceId();
// optional int64 parentId = 2; // optional int64 parent_id = 2;
boolean hasParentId(); boolean hasParentId();
long getParentId(); long getParentId();
} }
@ -48,8 +48,8 @@ public final class Tracing {
} }
private int bitField0_; private int bitField0_;
// optional int64 traceId = 1; // optional int64 trace_id = 1;
public static final int TRACEID_FIELD_NUMBER = 1; public static final int TRACE_ID_FIELD_NUMBER = 1;
private long traceId_; private long traceId_;
public boolean hasTraceId() { public boolean hasTraceId() {
return ((bitField0_ & 0x00000001) == 0x00000001); return ((bitField0_ & 0x00000001) == 0x00000001);
@ -58,8 +58,8 @@ public final class Tracing {
return traceId_; return traceId_;
} }
// optional int64 parentId = 2; // optional int64 parent_id = 2;
public static final int PARENTID_FIELD_NUMBER = 2; public static final int PARENT_ID_FIELD_NUMBER = 2;
private long parentId_; private long parentId_;
public boolean hasParentId() { public boolean hasParentId() {
return ((bitField0_ & 0x00000002) == 0x00000002); return ((bitField0_ & 0x00000002) == 0x00000002);
@ -150,11 +150,11 @@ public final class Tracing {
int hash = 41; int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode(); hash = (19 * hash) + getDescriptorForType().hashCode();
if (hasTraceId()) { if (hasTraceId()) {
hash = (37 * hash) + TRACEID_FIELD_NUMBER; hash = (37 * hash) + TRACE_ID_FIELD_NUMBER;
hash = (53 * hash) + hashLong(getTraceId()); hash = (53 * hash) + hashLong(getTraceId());
} }
if (hasParentId()) { if (hasParentId()) {
hash = (37 * hash) + PARENTID_FIELD_NUMBER; hash = (37 * hash) + PARENT_ID_FIELD_NUMBER;
hash = (53 * hash) + hashLong(getParentId()); hash = (53 * hash) + hashLong(getParentId());
} }
hash = (29 * hash) + getUnknownFields().hashCode(); hash = (29 * hash) + getUnknownFields().hashCode();
@ -392,7 +392,7 @@ public final class Tracing {
private int bitField0_; private int bitField0_;
// optional int64 traceId = 1; // optional int64 trace_id = 1;
private long traceId_ ; private long traceId_ ;
public boolean hasTraceId() { public boolean hasTraceId() {
return ((bitField0_ & 0x00000001) == 0x00000001); return ((bitField0_ & 0x00000001) == 0x00000001);
@ -413,7 +413,7 @@ public final class Tracing {
return this; return this;
} }
// optional int64 parentId = 2; // optional int64 parent_id = 2;
private long parentId_ ; private long parentId_ ;
public boolean hasParentId() { public boolean hasParentId() {
return ((bitField0_ & 0x00000002) == 0x00000002); return ((bitField0_ & 0x00000002) == 0x00000002);
@ -459,10 +459,10 @@ public final class Tracing {
descriptor; descriptor;
static { static {
java.lang.String[] descriptorData = { java.lang.String[] descriptorData = {
"\n\rTracing.proto\"-\n\010RPCTInfo\022\017\n\007traceId\030\001" + "\n\rTracing.proto\"/\n\010RPCTInfo\022\020\n\010trace_id\030" +
" \001(\003\022\020\n\010parentId\030\002 \001(\003B:\n*org.apache.had" + "\001 \001(\003\022\021\n\tparent_id\030\002 \001(\003B:\n*org.apache.h" +
"oop.hbase.protobuf.generatedB\007TracingH\001\240" + "adoop.hbase.protobuf.generatedB\007TracingH" +
"\001\001" "\001\240\001\001"
}; };
com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner = com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() { new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() {

View File

@ -80,7 +80,7 @@ public final class WALProtos {
public interface WALHeaderOrBuilder public interface WALHeaderOrBuilder
extends com.google.protobuf.MessageOrBuilder { extends com.google.protobuf.MessageOrBuilder {
// optional bool hasCompression = 1; // optional bool has_compression = 1;
boolean hasHasCompression(); boolean hasHasCompression();
boolean getHasCompression(); boolean getHasCompression();
} }
@ -113,8 +113,8 @@ public final class WALProtos {
} }
private int bitField0_; private int bitField0_;
// optional bool hasCompression = 1; // optional bool has_compression = 1;
public static final int HASCOMPRESSION_FIELD_NUMBER = 1; public static final int HAS_COMPRESSION_FIELD_NUMBER = 1;
private boolean hasCompression_; private boolean hasCompression_;
public boolean hasHasCompression() { public boolean hasHasCompression() {
return ((bitField0_ & 0x00000001) == 0x00000001); return ((bitField0_ & 0x00000001) == 0x00000001);
@ -192,7 +192,7 @@ public final class WALProtos {
int hash = 41; int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode(); hash = (19 * hash) + getDescriptorForType().hashCode();
if (hasHasCompression()) { if (hasHasCompression()) {
hash = (37 * hash) + HASCOMPRESSION_FIELD_NUMBER; hash = (37 * hash) + HAS_COMPRESSION_FIELD_NUMBER;
hash = (53 * hash) + hashBoolean(getHasCompression()); hash = (53 * hash) + hashBoolean(getHasCompression());
} }
hash = (29 * hash) + getUnknownFields().hashCode(); hash = (29 * hash) + getUnknownFields().hashCode();
@ -416,7 +416,7 @@ public final class WALProtos {
private int bitField0_; private int bitField0_;
// optional bool hasCompression = 1; // optional bool has_compression = 1;
private boolean hasCompression_ ; private boolean hasCompression_ ;
public boolean hasHasCompression() { public boolean hasHasCompression() {
return ((bitField0_ & 0x00000001) == 0x00000001); return ((bitField0_ & 0x00000001) == 0x00000001);
@ -451,23 +451,23 @@ public final class WALProtos {
public interface WALKeyOrBuilder public interface WALKeyOrBuilder
extends com.google.protobuf.MessageOrBuilder { extends com.google.protobuf.MessageOrBuilder {
// required bytes encodedRegionName = 1; // required bytes encoded_region_name = 1;
boolean hasEncodedRegionName(); boolean hasEncodedRegionName();
com.google.protobuf.ByteString getEncodedRegionName(); com.google.protobuf.ByteString getEncodedRegionName();
// required bytes tableName = 2; // required bytes table_name = 2;
boolean hasTableName(); boolean hasTableName();
com.google.protobuf.ByteString getTableName(); com.google.protobuf.ByteString getTableName();
// required uint64 logSequenceNumber = 3; // required uint64 log_sequence_number = 3;
boolean hasLogSequenceNumber(); boolean hasLogSequenceNumber();
long getLogSequenceNumber(); long getLogSequenceNumber();
// required uint64 writeTime = 4; // required uint64 write_time = 4;
boolean hasWriteTime(); boolean hasWriteTime();
long getWriteTime(); long getWriteTime();
// optional .UUID clusterId = 5; // optional .UUID cluster_id = 5;
boolean hasClusterId(); boolean hasClusterId();
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.UUID getClusterId(); org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.UUID getClusterId();
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.UUIDOrBuilder getClusterIdOrBuilder(); org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.UUIDOrBuilder getClusterIdOrBuilder();
@ -482,7 +482,7 @@ public final class WALProtos {
org.apache.hadoop.hbase.protobuf.generated.WALProtos.FamilyScopeOrBuilder getScopesOrBuilder( org.apache.hadoop.hbase.protobuf.generated.WALProtos.FamilyScopeOrBuilder getScopesOrBuilder(
int index); int index);
// optional uint32 followingKvCount = 7; // optional uint32 following_kv_count = 7;
boolean hasFollowingKvCount(); boolean hasFollowingKvCount();
int getFollowingKvCount(); int getFollowingKvCount();
} }
@ -515,8 +515,8 @@ public final class WALProtos {
} }
private int bitField0_; private int bitField0_;
// required bytes encodedRegionName = 1; // required bytes encoded_region_name = 1;
public static final int ENCODEDREGIONNAME_FIELD_NUMBER = 1; public static final int ENCODED_REGION_NAME_FIELD_NUMBER = 1;
private com.google.protobuf.ByteString encodedRegionName_; private com.google.protobuf.ByteString encodedRegionName_;
public boolean hasEncodedRegionName() { public boolean hasEncodedRegionName() {
return ((bitField0_ & 0x00000001) == 0x00000001); return ((bitField0_ & 0x00000001) == 0x00000001);
@ -525,8 +525,8 @@ public final class WALProtos {
return encodedRegionName_; return encodedRegionName_;
} }
// required bytes tableName = 2; // required bytes table_name = 2;
public static final int TABLENAME_FIELD_NUMBER = 2; public static final int TABLE_NAME_FIELD_NUMBER = 2;
private com.google.protobuf.ByteString tableName_; private com.google.protobuf.ByteString tableName_;
public boolean hasTableName() { public boolean hasTableName() {
return ((bitField0_ & 0x00000002) == 0x00000002); return ((bitField0_ & 0x00000002) == 0x00000002);
@ -535,8 +535,8 @@ public final class WALProtos {
return tableName_; return tableName_;
} }
// required uint64 logSequenceNumber = 3; // required uint64 log_sequence_number = 3;
public static final int LOGSEQUENCENUMBER_FIELD_NUMBER = 3; public static final int LOG_SEQUENCE_NUMBER_FIELD_NUMBER = 3;
private long logSequenceNumber_; private long logSequenceNumber_;
public boolean hasLogSequenceNumber() { public boolean hasLogSequenceNumber() {
return ((bitField0_ & 0x00000004) == 0x00000004); return ((bitField0_ & 0x00000004) == 0x00000004);
@ -545,8 +545,8 @@ public final class WALProtos {
return logSequenceNumber_; return logSequenceNumber_;
} }
// required uint64 writeTime = 4; // required uint64 write_time = 4;
public static final int WRITETIME_FIELD_NUMBER = 4; public static final int WRITE_TIME_FIELD_NUMBER = 4;
private long writeTime_; private long writeTime_;
public boolean hasWriteTime() { public boolean hasWriteTime() {
return ((bitField0_ & 0x00000008) == 0x00000008); return ((bitField0_ & 0x00000008) == 0x00000008);
@ -555,8 +555,8 @@ public final class WALProtos {
return writeTime_; return writeTime_;
} }
// optional .UUID clusterId = 5; // optional .UUID cluster_id = 5;
public static final int CLUSTERID_FIELD_NUMBER = 5; public static final int CLUSTER_ID_FIELD_NUMBER = 5;
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.UUID clusterId_; private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.UUID clusterId_;
public boolean hasClusterId() { public boolean hasClusterId() {
return ((bitField0_ & 0x00000010) == 0x00000010); return ((bitField0_ & 0x00000010) == 0x00000010);
@ -589,8 +589,8 @@ public final class WALProtos {
return scopes_.get(index); return scopes_.get(index);
} }
// optional uint32 followingKvCount = 7; // optional uint32 following_kv_count = 7;
public static final int FOLLOWINGKVCOUNT_FIELD_NUMBER = 7; public static final int FOLLOWING_KV_COUNT_FIELD_NUMBER = 7;
private int followingKvCount_; private int followingKvCount_;
public boolean hasFollowingKvCount() { public boolean hasFollowingKvCount() {
return ((bitField0_ & 0x00000020) == 0x00000020); return ((bitField0_ & 0x00000020) == 0x00000020);
@ -771,23 +771,23 @@ public final class WALProtos {
int hash = 41; int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode(); hash = (19 * hash) + getDescriptorForType().hashCode();
if (hasEncodedRegionName()) { if (hasEncodedRegionName()) {
hash = (37 * hash) + ENCODEDREGIONNAME_FIELD_NUMBER; hash = (37 * hash) + ENCODED_REGION_NAME_FIELD_NUMBER;
hash = (53 * hash) + getEncodedRegionName().hashCode(); hash = (53 * hash) + getEncodedRegionName().hashCode();
} }
if (hasTableName()) { if (hasTableName()) {
hash = (37 * hash) + TABLENAME_FIELD_NUMBER; hash = (37 * hash) + TABLE_NAME_FIELD_NUMBER;
hash = (53 * hash) + getTableName().hashCode(); hash = (53 * hash) + getTableName().hashCode();
} }
if (hasLogSequenceNumber()) { if (hasLogSequenceNumber()) {
hash = (37 * hash) + LOGSEQUENCENUMBER_FIELD_NUMBER; hash = (37 * hash) + LOG_SEQUENCE_NUMBER_FIELD_NUMBER;
hash = (53 * hash) + hashLong(getLogSequenceNumber()); hash = (53 * hash) + hashLong(getLogSequenceNumber());
} }
if (hasWriteTime()) { if (hasWriteTime()) {
hash = (37 * hash) + WRITETIME_FIELD_NUMBER; hash = (37 * hash) + WRITE_TIME_FIELD_NUMBER;
hash = (53 * hash) + hashLong(getWriteTime()); hash = (53 * hash) + hashLong(getWriteTime());
} }
if (hasClusterId()) { if (hasClusterId()) {
hash = (37 * hash) + CLUSTERID_FIELD_NUMBER; hash = (37 * hash) + CLUSTER_ID_FIELD_NUMBER;
hash = (53 * hash) + getClusterId().hashCode(); hash = (53 * hash) + getClusterId().hashCode();
} }
if (getScopesCount() > 0) { if (getScopesCount() > 0) {
@ -795,7 +795,7 @@ public final class WALProtos {
hash = (53 * hash) + getScopesList().hashCode(); hash = (53 * hash) + getScopesList().hashCode();
} }
if (hasFollowingKvCount()) { if (hasFollowingKvCount()) {
hash = (37 * hash) + FOLLOWINGKVCOUNT_FIELD_NUMBER; hash = (37 * hash) + FOLLOWING_KV_COUNT_FIELD_NUMBER;
hash = (53 * hash) + getFollowingKvCount(); hash = (53 * hash) + getFollowingKvCount();
} }
hash = (29 * hash) + getUnknownFields().hashCode(); hash = (29 * hash) + getUnknownFields().hashCode();
@ -1178,7 +1178,7 @@ public final class WALProtos {
private int bitField0_; private int bitField0_;
// required bytes encodedRegionName = 1; // required bytes encoded_region_name = 1;
private com.google.protobuf.ByteString encodedRegionName_ = com.google.protobuf.ByteString.EMPTY; private com.google.protobuf.ByteString encodedRegionName_ = com.google.protobuf.ByteString.EMPTY;
public boolean hasEncodedRegionName() { public boolean hasEncodedRegionName() {
return ((bitField0_ & 0x00000001) == 0x00000001); return ((bitField0_ & 0x00000001) == 0x00000001);
@ -1202,7 +1202,7 @@ public final class WALProtos {
return this; return this;
} }
// required bytes tableName = 2; // required bytes table_name = 2;
private com.google.protobuf.ByteString tableName_ = com.google.protobuf.ByteString.EMPTY; private com.google.protobuf.ByteString tableName_ = com.google.protobuf.ByteString.EMPTY;
public boolean hasTableName() { public boolean hasTableName() {
return ((bitField0_ & 0x00000002) == 0x00000002); return ((bitField0_ & 0x00000002) == 0x00000002);
@ -1226,7 +1226,7 @@ public final class WALProtos {
return this; return this;
} }
// required uint64 logSequenceNumber = 3; // required uint64 log_sequence_number = 3;
private long logSequenceNumber_ ; private long logSequenceNumber_ ;
public boolean hasLogSequenceNumber() { public boolean hasLogSequenceNumber() {
return ((bitField0_ & 0x00000004) == 0x00000004); return ((bitField0_ & 0x00000004) == 0x00000004);
@ -1247,7 +1247,7 @@ public final class WALProtos {
return this; return this;
} }
// required uint64 writeTime = 4; // required uint64 write_time = 4;
private long writeTime_ ; private long writeTime_ ;
public boolean hasWriteTime() { public boolean hasWriteTime() {
return ((bitField0_ & 0x00000008) == 0x00000008); return ((bitField0_ & 0x00000008) == 0x00000008);
@ -1268,7 +1268,7 @@ public final class WALProtos {
return this; return this;
} }
// optional .UUID clusterId = 5; // optional .UUID cluster_id = 5;
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.UUID clusterId_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.UUID.getDefaultInstance(); private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.UUID clusterId_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.UUID.getDefaultInstance();
private com.google.protobuf.SingleFieldBuilder< private com.google.protobuf.SingleFieldBuilder<
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.UUID, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.UUID.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.UUIDOrBuilder> clusterIdBuilder_; org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.UUID, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.UUID.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.UUIDOrBuilder> clusterIdBuilder_;
@ -1544,7 +1544,7 @@ public final class WALProtos {
return scopesBuilder_; return scopesBuilder_;
} }
// optional uint32 followingKvCount = 7; // optional uint32 following_kv_count = 7;
private int followingKvCount_ ; private int followingKvCount_ ;
public boolean hasFollowingKvCount() { public boolean hasFollowingKvCount() {
return ((bitField0_ & 0x00000040) == 0x00000040); return ((bitField0_ & 0x00000040) == 0x00000040);
@ -1583,7 +1583,7 @@ public final class WALProtos {
boolean hasFamily(); boolean hasFamily();
com.google.protobuf.ByteString getFamily(); com.google.protobuf.ByteString getFamily();
// required .ScopeType scopeType = 2; // required .ScopeType scope_type = 2;
boolean hasScopeType(); boolean hasScopeType();
org.apache.hadoop.hbase.protobuf.generated.WALProtos.ScopeType getScopeType(); org.apache.hadoop.hbase.protobuf.generated.WALProtos.ScopeType getScopeType();
} }
@ -1626,8 +1626,8 @@ public final class WALProtos {
return family_; return family_;
} }
// required .ScopeType scopeType = 2; // required .ScopeType scope_type = 2;
public static final int SCOPETYPE_FIELD_NUMBER = 2; public static final int SCOPE_TYPE_FIELD_NUMBER = 2;
private org.apache.hadoop.hbase.protobuf.generated.WALProtos.ScopeType scopeType_; private org.apache.hadoop.hbase.protobuf.generated.WALProtos.ScopeType scopeType_;
public boolean hasScopeType() { public boolean hasScopeType() {
return ((bitField0_ & 0x00000002) == 0x00000002); return ((bitField0_ & 0x00000002) == 0x00000002);
@ -1730,7 +1730,7 @@ public final class WALProtos {
hash = (53 * hash) + getFamily().hashCode(); hash = (53 * hash) + getFamily().hashCode();
} }
if (hasScopeType()) { if (hasScopeType()) {
hash = (37 * hash) + SCOPETYPE_FIELD_NUMBER; hash = (37 * hash) + SCOPE_TYPE_FIELD_NUMBER;
hash = (53 * hash) + hashEnum(getScopeType()); hash = (53 * hash) + hashEnum(getScopeType());
} }
hash = (29 * hash) + getUnknownFields().hashCode(); hash = (29 * hash) + getUnknownFields().hashCode();
@ -2006,7 +2006,7 @@ public final class WALProtos {
return this; return this;
} }
// required .ScopeType scopeType = 2; // required .ScopeType scope_type = 2;
private org.apache.hadoop.hbase.protobuf.generated.WALProtos.ScopeType scopeType_ = org.apache.hadoop.hbase.protobuf.generated.WALProtos.ScopeType.REPLICATION_SCOPE_LOCAL; private org.apache.hadoop.hbase.protobuf.generated.WALProtos.ScopeType scopeType_ = org.apache.hadoop.hbase.protobuf.generated.WALProtos.ScopeType.REPLICATION_SCOPE_LOCAL;
public boolean hasScopeType() { public boolean hasScopeType() {
return ((bitField0_ & 0x00000002) == 0x00000002); return ((bitField0_ & 0x00000002) == 0x00000002);
@ -2044,29 +2044,29 @@ public final class WALProtos {
public interface CompactionDescriptorOrBuilder public interface CompactionDescriptorOrBuilder
extends com.google.protobuf.MessageOrBuilder { extends com.google.protobuf.MessageOrBuilder {
// required bytes tableName = 1; // required bytes table_name = 1;
boolean hasTableName(); boolean hasTableName();
com.google.protobuf.ByteString getTableName(); com.google.protobuf.ByteString getTableName();
// required bytes encodedRegionName = 2; // required bytes encoded_region_name = 2;
boolean hasEncodedRegionName(); boolean hasEncodedRegionName();
com.google.protobuf.ByteString getEncodedRegionName(); com.google.protobuf.ByteString getEncodedRegionName();
// required bytes familyName = 3; // required bytes family_name = 3;
boolean hasFamilyName(); boolean hasFamilyName();
com.google.protobuf.ByteString getFamilyName(); com.google.protobuf.ByteString getFamilyName();
// repeated string compactionInput = 4; // repeated string compaction_input = 4;
java.util.List<String> getCompactionInputList(); java.util.List<String> getCompactionInputList();
int getCompactionInputCount(); int getCompactionInputCount();
String getCompactionInput(int index); String getCompactionInput(int index);
// repeated string compactionOutput = 5; // repeated string compaction_output = 5;
java.util.List<String> getCompactionOutputList(); java.util.List<String> getCompactionOutputList();
int getCompactionOutputCount(); int getCompactionOutputCount();
String getCompactionOutput(int index); String getCompactionOutput(int index);
// required string storeHomeDir = 6; // required string store_home_dir = 6;
boolean hasStoreHomeDir(); boolean hasStoreHomeDir();
String getStoreHomeDir(); String getStoreHomeDir();
} }
@ -2099,8 +2099,8 @@ public final class WALProtos {
} }
private int bitField0_; private int bitField0_;
// required bytes tableName = 1; // required bytes table_name = 1;
public static final int TABLENAME_FIELD_NUMBER = 1; public static final int TABLE_NAME_FIELD_NUMBER = 1;
private com.google.protobuf.ByteString tableName_; private com.google.protobuf.ByteString tableName_;
public boolean hasTableName() { public boolean hasTableName() {
return ((bitField0_ & 0x00000001) == 0x00000001); return ((bitField0_ & 0x00000001) == 0x00000001);
@ -2109,8 +2109,8 @@ public final class WALProtos {
return tableName_; return tableName_;
} }
// required bytes encodedRegionName = 2; // required bytes encoded_region_name = 2;
public static final int ENCODEDREGIONNAME_FIELD_NUMBER = 2; public static final int ENCODED_REGION_NAME_FIELD_NUMBER = 2;
private com.google.protobuf.ByteString encodedRegionName_; private com.google.protobuf.ByteString encodedRegionName_;
public boolean hasEncodedRegionName() { public boolean hasEncodedRegionName() {
return ((bitField0_ & 0x00000002) == 0x00000002); return ((bitField0_ & 0x00000002) == 0x00000002);
@ -2119,8 +2119,8 @@ public final class WALProtos {
return encodedRegionName_; return encodedRegionName_;
} }
// required bytes familyName = 3; // required bytes family_name = 3;
public static final int FAMILYNAME_FIELD_NUMBER = 3; public static final int FAMILY_NAME_FIELD_NUMBER = 3;
private com.google.protobuf.ByteString familyName_; private com.google.protobuf.ByteString familyName_;
public boolean hasFamilyName() { public boolean hasFamilyName() {
return ((bitField0_ & 0x00000004) == 0x00000004); return ((bitField0_ & 0x00000004) == 0x00000004);
@ -2129,8 +2129,8 @@ public final class WALProtos {
return familyName_; return familyName_;
} }
// repeated string compactionInput = 4; // repeated string compaction_input = 4;
public static final int COMPACTIONINPUT_FIELD_NUMBER = 4; public static final int COMPACTION_INPUT_FIELD_NUMBER = 4;
private com.google.protobuf.LazyStringList compactionInput_; private com.google.protobuf.LazyStringList compactionInput_;
public java.util.List<String> public java.util.List<String>
getCompactionInputList() { getCompactionInputList() {
@ -2143,8 +2143,8 @@ public final class WALProtos {
return compactionInput_.get(index); return compactionInput_.get(index);
} }
// repeated string compactionOutput = 5; // repeated string compaction_output = 5;
public static final int COMPACTIONOUTPUT_FIELD_NUMBER = 5; public static final int COMPACTION_OUTPUT_FIELD_NUMBER = 5;
private com.google.protobuf.LazyStringList compactionOutput_; private com.google.protobuf.LazyStringList compactionOutput_;
public java.util.List<String> public java.util.List<String>
getCompactionOutputList() { getCompactionOutputList() {
@ -2157,8 +2157,8 @@ public final class WALProtos {
return compactionOutput_.get(index); return compactionOutput_.get(index);
} }
// required string storeHomeDir = 6; // required string store_home_dir = 6;
public static final int STOREHOMEDIR_FIELD_NUMBER = 6; public static final int STORE_HOME_DIR_FIELD_NUMBER = 6;
private java.lang.Object storeHomeDir_; private java.lang.Object storeHomeDir_;
public boolean hasStoreHomeDir() { public boolean hasStoreHomeDir() {
return ((bitField0_ & 0x00000008) == 0x00000008); return ((bitField0_ & 0x00000008) == 0x00000008);
@ -2343,27 +2343,27 @@ public final class WALProtos {
int hash = 41; int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode(); hash = (19 * hash) + getDescriptorForType().hashCode();
if (hasTableName()) { if (hasTableName()) {
hash = (37 * hash) + TABLENAME_FIELD_NUMBER; hash = (37 * hash) + TABLE_NAME_FIELD_NUMBER;
hash = (53 * hash) + getTableName().hashCode(); hash = (53 * hash) + getTableName().hashCode();
} }
if (hasEncodedRegionName()) { if (hasEncodedRegionName()) {
hash = (37 * hash) + ENCODEDREGIONNAME_FIELD_NUMBER; hash = (37 * hash) + ENCODED_REGION_NAME_FIELD_NUMBER;
hash = (53 * hash) + getEncodedRegionName().hashCode(); hash = (53 * hash) + getEncodedRegionName().hashCode();
} }
if (hasFamilyName()) { if (hasFamilyName()) {
hash = (37 * hash) + FAMILYNAME_FIELD_NUMBER; hash = (37 * hash) + FAMILY_NAME_FIELD_NUMBER;
hash = (53 * hash) + getFamilyName().hashCode(); hash = (53 * hash) + getFamilyName().hashCode();
} }
if (getCompactionInputCount() > 0) { if (getCompactionInputCount() > 0) {
hash = (37 * hash) + COMPACTIONINPUT_FIELD_NUMBER; hash = (37 * hash) + COMPACTION_INPUT_FIELD_NUMBER;
hash = (53 * hash) + getCompactionInputList().hashCode(); hash = (53 * hash) + getCompactionInputList().hashCode();
} }
if (getCompactionOutputCount() > 0) { if (getCompactionOutputCount() > 0) {
hash = (37 * hash) + COMPACTIONOUTPUT_FIELD_NUMBER; hash = (37 * hash) + COMPACTION_OUTPUT_FIELD_NUMBER;
hash = (53 * hash) + getCompactionOutputList().hashCode(); hash = (53 * hash) + getCompactionOutputList().hashCode();
} }
if (hasStoreHomeDir()) { if (hasStoreHomeDir()) {
hash = (37 * hash) + STOREHOMEDIR_FIELD_NUMBER; hash = (37 * hash) + STORE_HOME_DIR_FIELD_NUMBER;
hash = (53 * hash) + getStoreHomeDir().hashCode(); hash = (53 * hash) + getStoreHomeDir().hashCode();
} }
hash = (29 * hash) + getUnknownFields().hashCode(); hash = (29 * hash) + getUnknownFields().hashCode();
@ -2691,7 +2691,7 @@ public final class WALProtos {
private int bitField0_; private int bitField0_;
// required bytes tableName = 1; // required bytes table_name = 1;
private com.google.protobuf.ByteString tableName_ = com.google.protobuf.ByteString.EMPTY; private com.google.protobuf.ByteString tableName_ = com.google.protobuf.ByteString.EMPTY;
public boolean hasTableName() { public boolean hasTableName() {
return ((bitField0_ & 0x00000001) == 0x00000001); return ((bitField0_ & 0x00000001) == 0x00000001);
@ -2715,7 +2715,7 @@ public final class WALProtos {
return this; return this;
} }
// required bytes encodedRegionName = 2; // required bytes encoded_region_name = 2;
private com.google.protobuf.ByteString encodedRegionName_ = com.google.protobuf.ByteString.EMPTY; private com.google.protobuf.ByteString encodedRegionName_ = com.google.protobuf.ByteString.EMPTY;
public boolean hasEncodedRegionName() { public boolean hasEncodedRegionName() {
return ((bitField0_ & 0x00000002) == 0x00000002); return ((bitField0_ & 0x00000002) == 0x00000002);
@ -2739,7 +2739,7 @@ public final class WALProtos {
return this; return this;
} }
// required bytes familyName = 3; // required bytes family_name = 3;
private com.google.protobuf.ByteString familyName_ = com.google.protobuf.ByteString.EMPTY; private com.google.protobuf.ByteString familyName_ = com.google.protobuf.ByteString.EMPTY;
public boolean hasFamilyName() { public boolean hasFamilyName() {
return ((bitField0_ & 0x00000004) == 0x00000004); return ((bitField0_ & 0x00000004) == 0x00000004);
@ -2763,7 +2763,7 @@ public final class WALProtos {
return this; return this;
} }
// repeated string compactionInput = 4; // repeated string compaction_input = 4;
private com.google.protobuf.LazyStringList compactionInput_ = com.google.protobuf.LazyStringArrayList.EMPTY; private com.google.protobuf.LazyStringList compactionInput_ = com.google.protobuf.LazyStringArrayList.EMPTY;
private void ensureCompactionInputIsMutable() { private void ensureCompactionInputIsMutable() {
if (!((bitField0_ & 0x00000008) == 0x00000008)) { if (!((bitField0_ & 0x00000008) == 0x00000008)) {
@ -2819,7 +2819,7 @@ public final class WALProtos {
onChanged(); onChanged();
} }
// repeated string compactionOutput = 5; // repeated string compaction_output = 5;
private com.google.protobuf.LazyStringList compactionOutput_ = com.google.protobuf.LazyStringArrayList.EMPTY; private com.google.protobuf.LazyStringList compactionOutput_ = com.google.protobuf.LazyStringArrayList.EMPTY;
private void ensureCompactionOutputIsMutable() { private void ensureCompactionOutputIsMutable() {
if (!((bitField0_ & 0x00000010) == 0x00000010)) { if (!((bitField0_ & 0x00000010) == 0x00000010)) {
@ -2875,7 +2875,7 @@ public final class WALProtos {
onChanged(); onChanged();
} }
// required string storeHomeDir = 6; // required string store_home_dir = 6;
private java.lang.Object storeHomeDir_ = ""; private java.lang.Object storeHomeDir_ = "";
public boolean hasStoreHomeDir() { public boolean hasStoreHomeDir() {
return ((bitField0_ & 0x00000020) == 0x00000020); return ((bitField0_ & 0x00000020) == 0x00000020);
@ -3256,22 +3256,23 @@ public final class WALProtos {
descriptor; descriptor;
static { static {
java.lang.String[] descriptorData = { java.lang.String[] descriptorData = {
"\n\tWAL.proto\032\013hbase.proto\"#\n\tWALHeader\022\026\n" + "\n\tWAL.proto\032\013hbase.proto\"$\n\tWALHeader\022\027\n" +
"\016hasCompression\030\001 \001(\010\"\266\001\n\006WALKey\022\031\n\021enco" + "\017has_compression\030\001 \001(\010\"\277\001\n\006WALKey\022\033\n\023enc" +
"dedRegionName\030\001 \002(\014\022\021\n\ttableName\030\002 \002(\014\022\031" + "oded_region_name\030\001 \002(\014\022\022\n\ntable_name\030\002 \002" +
"\n\021logSequenceNumber\030\003 \002(\004\022\021\n\twriteTime\030\004" + "(\014\022\033\n\023log_sequence_number\030\003 \002(\004\022\022\n\nwrite" +
" \002(\004\022\030\n\tclusterId\030\005 \001(\0132\005.UUID\022\034\n\006scopes" + "_time\030\004 \002(\004\022\031\n\ncluster_id\030\005 \001(\0132\005.UUID\022\034" +
"\030\006 \003(\0132\014.FamilyScope\022\030\n\020followingKvCount" + "\n\006scopes\030\006 \003(\0132\014.FamilyScope\022\032\n\022followin" +
"\030\007 \001(\r\"<\n\013FamilyScope\022\016\n\006family\030\001 \002(\014\022\035\n" + "g_kv_count\030\007 \001(\r\"=\n\013FamilyScope\022\016\n\006famil" +
"\tscopeType\030\002 \002(\0162\n.ScopeType\"\241\001\n\024Compact" + "y\030\001 \002(\014\022\036\n\nscope_type\030\002 \002(\0162\n.ScopeType\"" +
"ionDescriptor\022\021\n\ttableName\030\001 \002(\014\022\031\n\021enco" + "\251\001\n\024CompactionDescriptor\022\022\n\ntable_name\030\001" +
"dedRegionName\030\002 \002(\014\022\022\n\nfamilyName\030\003 \002(\014\022", " \002(\014\022\033\n\023encoded_region_name\030\002 \002(\014\022\023\n\013fam",
"\027\n\017compactionInput\030\004 \003(\t\022\030\n\020compactionOu" + "ily_name\030\003 \002(\014\022\030\n\020compaction_input\030\004 \003(\t" +
"tput\030\005 \003(\t\022\024\n\014storeHomeDir\030\006 \002(\t\"\014\n\nWALT" + "\022\031\n\021compaction_output\030\005 \003(\t\022\026\n\016store_hom" +
"railer*F\n\tScopeType\022\033\n\027REPLICATION_SCOPE" + "e_dir\030\006 \002(\t\"\014\n\nWALTrailer*F\n\tScopeType\022\033" +
"_LOCAL\020\000\022\034\n\030REPLICATION_SCOPE_GLOBAL\020\001B?" + "\n\027REPLICATION_SCOPE_LOCAL\020\000\022\034\n\030REPLICATI" +
"\n*org.apache.hadoop.hbase.protobuf.gener" + "ON_SCOPE_GLOBAL\020\001B?\n*org.apache.hadoop.h" +
"atedB\tWALProtosH\001\210\001\000\240\001\001" "base.protobuf.generatedB\tWALProtosH\001\210\001\000\240" +
"\001\001"
}; };
com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner = com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() { new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() {

View File

@ -16,7 +16,7 @@ public final class ZooKeeperProtos {
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName getServer(); org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName getServer();
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerNameOrBuilder getServerOrBuilder(); org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerNameOrBuilder getServerOrBuilder();
// optional uint32 rpcVersion = 2; // optional uint32 rpc_version = 2;
boolean hasRpcVersion(); boolean hasRpcVersion();
int getRpcVersion(); int getRpcVersion();
} }
@ -62,8 +62,8 @@ public final class ZooKeeperProtos {
return server_; return server_;
} }
// optional uint32 rpcVersion = 2; // optional uint32 rpc_version = 2;
public static final int RPCVERSION_FIELD_NUMBER = 2; public static final int RPC_VERSION_FIELD_NUMBER = 2;
private int rpcVersion_; private int rpcVersion_;
public boolean hasRpcVersion() { public boolean hasRpcVersion() {
return ((bitField0_ & 0x00000002) == 0x00000002); return ((bitField0_ & 0x00000002) == 0x00000002);
@ -166,7 +166,7 @@ public final class ZooKeeperProtos {
hash = (53 * hash) + getServer().hashCode(); hash = (53 * hash) + getServer().hashCode();
} }
if (hasRpcVersion()) { if (hasRpcVersion()) {
hash = (37 * hash) + RPCVERSION_FIELD_NUMBER; hash = (37 * hash) + RPC_VERSION_FIELD_NUMBER;
hash = (53 * hash) + getRpcVersion(); hash = (53 * hash) + getRpcVersion();
} }
hash = (29 * hash) + getUnknownFields().hashCode(); hash = (29 * hash) + getUnknownFields().hashCode();
@ -515,7 +515,7 @@ public final class ZooKeeperProtos {
return serverBuilder_; return serverBuilder_;
} }
// optional uint32 rpcVersion = 2; // optional uint32 rpc_version = 2;
private int rpcVersion_ ; private int rpcVersion_ ;
public boolean hasRpcVersion() { public boolean hasRpcVersion() {
return ((bitField0_ & 0x00000002) == 0x00000002); return ((bitField0_ & 0x00000002) == 0x00000002);
@ -555,7 +555,7 @@ public final class ZooKeeperProtos {
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName getMaster(); org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName getMaster();
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerNameOrBuilder getMasterOrBuilder(); org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerNameOrBuilder getMasterOrBuilder();
// optional uint32 rpcVersion = 2; // optional uint32 rpc_version = 2;
boolean hasRpcVersion(); boolean hasRpcVersion();
int getRpcVersion(); int getRpcVersion();
} }
@ -601,8 +601,8 @@ public final class ZooKeeperProtos {
return master_; return master_;
} }
// optional uint32 rpcVersion = 2; // optional uint32 rpc_version = 2;
public static final int RPCVERSION_FIELD_NUMBER = 2; public static final int RPC_VERSION_FIELD_NUMBER = 2;
private int rpcVersion_; private int rpcVersion_;
public boolean hasRpcVersion() { public boolean hasRpcVersion() {
return ((bitField0_ & 0x00000002) == 0x00000002); return ((bitField0_ & 0x00000002) == 0x00000002);
@ -705,7 +705,7 @@ public final class ZooKeeperProtos {
hash = (53 * hash) + getMaster().hashCode(); hash = (53 * hash) + getMaster().hashCode();
} }
if (hasRpcVersion()) { if (hasRpcVersion()) {
hash = (37 * hash) + RPCVERSION_FIELD_NUMBER; hash = (37 * hash) + RPC_VERSION_FIELD_NUMBER;
hash = (53 * hash) + getRpcVersion(); hash = (53 * hash) + getRpcVersion();
} }
hash = (29 * hash) + getUnknownFields().hashCode(); hash = (29 * hash) + getUnknownFields().hashCode();
@ -1054,7 +1054,7 @@ public final class ZooKeeperProtos {
return masterBuilder_; return masterBuilder_;
} }
// optional uint32 rpcVersion = 2; // optional uint32 rpc_version = 2;
private int rpcVersion_ ; private int rpcVersion_ ;
public boolean hasRpcVersion() { public boolean hasRpcVersion() {
return ((bitField0_ & 0x00000002) == 0x00000002); return ((bitField0_ & 0x00000002) == 0x00000002);
@ -1089,7 +1089,7 @@ public final class ZooKeeperProtos {
public interface ClusterUpOrBuilder public interface ClusterUpOrBuilder
extends com.google.protobuf.MessageOrBuilder { extends com.google.protobuf.MessageOrBuilder {
// required string startDate = 1; // required string start_date = 1;
boolean hasStartDate(); boolean hasStartDate();
String getStartDate(); String getStartDate();
} }
@ -1122,8 +1122,8 @@ public final class ZooKeeperProtos {
} }
private int bitField0_; private int bitField0_;
// required string startDate = 1; // required string start_date = 1;
public static final int STARTDATE_FIELD_NUMBER = 1; public static final int START_DATE_FIELD_NUMBER = 1;
private java.lang.Object startDate_; private java.lang.Object startDate_;
public boolean hasStartDate() { public boolean hasStartDate() {
return ((bitField0_ & 0x00000001) == 0x00000001); return ((bitField0_ & 0x00000001) == 0x00000001);
@ -1227,7 +1227,7 @@ public final class ZooKeeperProtos {
int hash = 41; int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode(); hash = (19 * hash) + getDescriptorForType().hashCode();
if (hasStartDate()) { if (hasStartDate()) {
hash = (37 * hash) + STARTDATE_FIELD_NUMBER; hash = (37 * hash) + START_DATE_FIELD_NUMBER;
hash = (53 * hash) + getStartDate().hashCode(); hash = (53 * hash) + getStartDate().hashCode();
} }
hash = (29 * hash) + getUnknownFields().hashCode(); hash = (29 * hash) + getUnknownFields().hashCode();
@ -1455,7 +1455,7 @@ public final class ZooKeeperProtos {
private int bitField0_; private int bitField0_;
// required string startDate = 1; // required string start_date = 1;
private java.lang.Object startDate_ = ""; private java.lang.Object startDate_ = "";
public boolean hasStartDate() { public boolean hasStartDate() {
return ((bitField0_ & 0x00000001) == 0x00000001); return ((bitField0_ & 0x00000001) == 0x00000001);
@ -1505,19 +1505,19 @@ public final class ZooKeeperProtos {
public interface RegionTransitionOrBuilder public interface RegionTransitionOrBuilder
extends com.google.protobuf.MessageOrBuilder { extends com.google.protobuf.MessageOrBuilder {
// required uint32 eventTypeCode = 1; // required uint32 event_type_code = 1;
boolean hasEventTypeCode(); boolean hasEventTypeCode();
int getEventTypeCode(); int getEventTypeCode();
// required bytes regionName = 2; // required bytes region_name = 2;
boolean hasRegionName(); boolean hasRegionName();
com.google.protobuf.ByteString getRegionName(); com.google.protobuf.ByteString getRegionName();
// required uint64 createTime = 3; // required uint64 create_time = 3;
boolean hasCreateTime(); boolean hasCreateTime();
long getCreateTime(); long getCreateTime();
// required .ServerName serverName = 4; // required .ServerName server_name = 4;
boolean hasServerName(); boolean hasServerName();
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName getServerName(); org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName getServerName();
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerNameOrBuilder getServerNameOrBuilder(); org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerNameOrBuilder getServerNameOrBuilder();
@ -1555,8 +1555,8 @@ public final class ZooKeeperProtos {
} }
private int bitField0_; private int bitField0_;
// required uint32 eventTypeCode = 1; // required uint32 event_type_code = 1;
public static final int EVENTTYPECODE_FIELD_NUMBER = 1; public static final int EVENT_TYPE_CODE_FIELD_NUMBER = 1;
private int eventTypeCode_; private int eventTypeCode_;
public boolean hasEventTypeCode() { public boolean hasEventTypeCode() {
return ((bitField0_ & 0x00000001) == 0x00000001); return ((bitField0_ & 0x00000001) == 0x00000001);
@ -1565,8 +1565,8 @@ public final class ZooKeeperProtos {
return eventTypeCode_; return eventTypeCode_;
} }
// required bytes regionName = 2; // required bytes region_name = 2;
public static final int REGIONNAME_FIELD_NUMBER = 2; public static final int REGION_NAME_FIELD_NUMBER = 2;
private com.google.protobuf.ByteString regionName_; private com.google.protobuf.ByteString regionName_;
public boolean hasRegionName() { public boolean hasRegionName() {
return ((bitField0_ & 0x00000002) == 0x00000002); return ((bitField0_ & 0x00000002) == 0x00000002);
@ -1575,8 +1575,8 @@ public final class ZooKeeperProtos {
return regionName_; return regionName_;
} }
// required uint64 createTime = 3; // required uint64 create_time = 3;
public static final int CREATETIME_FIELD_NUMBER = 3; public static final int CREATE_TIME_FIELD_NUMBER = 3;
private long createTime_; private long createTime_;
public boolean hasCreateTime() { public boolean hasCreateTime() {
return ((bitField0_ & 0x00000004) == 0x00000004); return ((bitField0_ & 0x00000004) == 0x00000004);
@ -1585,8 +1585,8 @@ public final class ZooKeeperProtos {
return createTime_; return createTime_;
} }
// required .ServerName serverName = 4; // required .ServerName server_name = 4;
public static final int SERVERNAME_FIELD_NUMBER = 4; public static final int SERVER_NAME_FIELD_NUMBER = 4;
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName serverName_; private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName serverName_;
public boolean hasServerName() { public boolean hasServerName() {
return ((bitField0_ & 0x00000008) == 0x00000008); return ((bitField0_ & 0x00000008) == 0x00000008);
@ -1749,19 +1749,19 @@ public final class ZooKeeperProtos {
int hash = 41; int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode(); hash = (19 * hash) + getDescriptorForType().hashCode();
if (hasEventTypeCode()) { if (hasEventTypeCode()) {
hash = (37 * hash) + EVENTTYPECODE_FIELD_NUMBER; hash = (37 * hash) + EVENT_TYPE_CODE_FIELD_NUMBER;
hash = (53 * hash) + getEventTypeCode(); hash = (53 * hash) + getEventTypeCode();
} }
if (hasRegionName()) { if (hasRegionName()) {
hash = (37 * hash) + REGIONNAME_FIELD_NUMBER; hash = (37 * hash) + REGION_NAME_FIELD_NUMBER;
hash = (53 * hash) + getRegionName().hashCode(); hash = (53 * hash) + getRegionName().hashCode();
} }
if (hasCreateTime()) { if (hasCreateTime()) {
hash = (37 * hash) + CREATETIME_FIELD_NUMBER; hash = (37 * hash) + CREATE_TIME_FIELD_NUMBER;
hash = (53 * hash) + hashLong(getCreateTime()); hash = (53 * hash) + hashLong(getCreateTime());
} }
if (hasServerName()) { if (hasServerName()) {
hash = (37 * hash) + SERVERNAME_FIELD_NUMBER; hash = (37 * hash) + SERVER_NAME_FIELD_NUMBER;
hash = (53 * hash) + getServerName().hashCode(); hash = (53 * hash) + getServerName().hashCode();
} }
if (hasPayload()) { if (hasPayload()) {
@ -2078,7 +2078,7 @@ public final class ZooKeeperProtos {
private int bitField0_; private int bitField0_;
// required uint32 eventTypeCode = 1; // required uint32 event_type_code = 1;
private int eventTypeCode_ ; private int eventTypeCode_ ;
public boolean hasEventTypeCode() { public boolean hasEventTypeCode() {
return ((bitField0_ & 0x00000001) == 0x00000001); return ((bitField0_ & 0x00000001) == 0x00000001);
@ -2099,7 +2099,7 @@ public final class ZooKeeperProtos {
return this; return this;
} }
// required bytes regionName = 2; // required bytes region_name = 2;
private com.google.protobuf.ByteString regionName_ = com.google.protobuf.ByteString.EMPTY; private com.google.protobuf.ByteString regionName_ = com.google.protobuf.ByteString.EMPTY;
public boolean hasRegionName() { public boolean hasRegionName() {
return ((bitField0_ & 0x00000002) == 0x00000002); return ((bitField0_ & 0x00000002) == 0x00000002);
@ -2123,7 +2123,7 @@ public final class ZooKeeperProtos {
return this; return this;
} }
// required uint64 createTime = 3; // required uint64 create_time = 3;
private long createTime_ ; private long createTime_ ;
public boolean hasCreateTime() { public boolean hasCreateTime() {
return ((bitField0_ & 0x00000004) == 0x00000004); return ((bitField0_ & 0x00000004) == 0x00000004);
@ -2144,7 +2144,7 @@ public final class ZooKeeperProtos {
return this; return this;
} }
// required .ServerName serverName = 4; // required .ServerName server_name = 4;
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName serverName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.getDefaultInstance(); private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName serverName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.getDefaultInstance();
private com.google.protobuf.SingleFieldBuilder< private com.google.protobuf.SingleFieldBuilder<
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerNameOrBuilder> serverNameBuilder_; org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerNameOrBuilder> serverNameBuilder_;
@ -2276,7 +2276,7 @@ public final class ZooKeeperProtos {
boolean hasState(); boolean hasState();
org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.SplitLogTask.State getState(); org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.SplitLogTask.State getState();
// required .ServerName serverName = 2; // required .ServerName server_name = 2;
boolean hasServerName(); boolean hasServerName();
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName getServerName(); org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName getServerName();
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerNameOrBuilder getServerNameOrBuilder(); org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerNameOrBuilder getServerNameOrBuilder();
@ -2398,8 +2398,8 @@ public final class ZooKeeperProtos {
return state_; return state_;
} }
// required .ServerName serverName = 2; // required .ServerName server_name = 2;
public static final int SERVERNAME_FIELD_NUMBER = 2; public static final int SERVER_NAME_FIELD_NUMBER = 2;
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName serverName_; private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName serverName_;
public boolean hasServerName() { public boolean hasServerName() {
return ((bitField0_ & 0x00000002) == 0x00000002); return ((bitField0_ & 0x00000002) == 0x00000002);
@ -2509,7 +2509,7 @@ public final class ZooKeeperProtos {
hash = (53 * hash) + hashEnum(getState()); hash = (53 * hash) + hashEnum(getState());
} }
if (hasServerName()) { if (hasServerName()) {
hash = (37 * hash) + SERVERNAME_FIELD_NUMBER; hash = (37 * hash) + SERVER_NAME_FIELD_NUMBER;
hash = (53 * hash) + getServerName().hashCode(); hash = (53 * hash) + getServerName().hashCode();
} }
hash = (29 * hash) + getUnknownFields().hashCode(); hash = (29 * hash) + getUnknownFields().hashCode();
@ -2802,7 +2802,7 @@ public final class ZooKeeperProtos {
return this; return this;
} }
// required .ServerName serverName = 2; // required .ServerName server_name = 2;
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName serverName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.getDefaultInstance(); private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName serverName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.getDefaultInstance();
private com.google.protobuf.SingleFieldBuilder< private com.google.protobuf.SingleFieldBuilder<
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerNameOrBuilder> serverNameBuilder_; org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerNameOrBuilder> serverNameBuilder_;
@ -4621,7 +4621,7 @@ public final class ZooKeeperProtos {
public interface ReplicationLockOrBuilder public interface ReplicationLockOrBuilder
extends com.google.protobuf.MessageOrBuilder { extends com.google.protobuf.MessageOrBuilder {
// required string lockOwner = 1; // required string lock_owner = 1;
boolean hasLockOwner(); boolean hasLockOwner();
String getLockOwner(); String getLockOwner();
} }
@ -4654,8 +4654,8 @@ public final class ZooKeeperProtos {
} }
private int bitField0_; private int bitField0_;
// required string lockOwner = 1; // required string lock_owner = 1;
public static final int LOCKOWNER_FIELD_NUMBER = 1; public static final int LOCK_OWNER_FIELD_NUMBER = 1;
private java.lang.Object lockOwner_; private java.lang.Object lockOwner_;
public boolean hasLockOwner() { public boolean hasLockOwner() {
return ((bitField0_ & 0x00000001) == 0x00000001); return ((bitField0_ & 0x00000001) == 0x00000001);
@ -4759,7 +4759,7 @@ public final class ZooKeeperProtos {
int hash = 41; int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode(); hash = (19 * hash) + getDescriptorForType().hashCode();
if (hasLockOwner()) { if (hasLockOwner()) {
hash = (37 * hash) + LOCKOWNER_FIELD_NUMBER; hash = (37 * hash) + LOCK_OWNER_FIELD_NUMBER;
hash = (53 * hash) + getLockOwner().hashCode(); hash = (53 * hash) + getLockOwner().hashCode();
} }
hash = (29 * hash) + getUnknownFields().hashCode(); hash = (29 * hash) + getUnknownFields().hashCode();
@ -4987,7 +4987,7 @@ public final class ZooKeeperProtos {
private int bitField0_; private int bitField0_;
// required string lockOwner = 1; // required string lock_owner = 1;
private java.lang.Object lockOwner_ = ""; private java.lang.Object lockOwner_ = "";
public boolean hasLockOwner() { public boolean hasLockOwner() {
return ((bitField0_ & 0x00000001) == 0x00000001); return ((bitField0_ & 0x00000001) == 0x00000001);
@ -5037,20 +5037,20 @@ public final class ZooKeeperProtos {
public interface TableLockOrBuilder public interface TableLockOrBuilder
extends com.google.protobuf.MessageOrBuilder { extends com.google.protobuf.MessageOrBuilder {
// optional bytes tableName = 1; // optional bytes table_name = 1;
boolean hasTableName(); boolean hasTableName();
com.google.protobuf.ByteString getTableName(); com.google.protobuf.ByteString getTableName();
// optional .ServerName lockOwner = 2; // optional .ServerName lock_owner = 2;
boolean hasLockOwner(); boolean hasLockOwner();
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName getLockOwner(); org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName getLockOwner();
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerNameOrBuilder getLockOwnerOrBuilder(); org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerNameOrBuilder getLockOwnerOrBuilder();
// optional int64 threadId = 3; // optional int64 thread_id = 3;
boolean hasThreadId(); boolean hasThreadId();
long getThreadId(); long getThreadId();
// optional bool isShared = 4; // optional bool is_shared = 4;
boolean hasIsShared(); boolean hasIsShared();
boolean getIsShared(); boolean getIsShared();
@ -5058,7 +5058,7 @@ public final class ZooKeeperProtos {
boolean hasPurpose(); boolean hasPurpose();
String getPurpose(); String getPurpose();
// optional int64 createTime = 6; // optional int64 create_time = 6;
boolean hasCreateTime(); boolean hasCreateTime();
long getCreateTime(); long getCreateTime();
} }
@ -5091,8 +5091,8 @@ public final class ZooKeeperProtos {
} }
private int bitField0_; private int bitField0_;
// optional bytes tableName = 1; // optional bytes table_name = 1;
public static final int TABLENAME_FIELD_NUMBER = 1; public static final int TABLE_NAME_FIELD_NUMBER = 1;
private com.google.protobuf.ByteString tableName_; private com.google.protobuf.ByteString tableName_;
public boolean hasTableName() { public boolean hasTableName() {
return ((bitField0_ & 0x00000001) == 0x00000001); return ((bitField0_ & 0x00000001) == 0x00000001);
@ -5101,8 +5101,8 @@ public final class ZooKeeperProtos {
return tableName_; return tableName_;
} }
// optional .ServerName lockOwner = 2; // optional .ServerName lock_owner = 2;
public static final int LOCKOWNER_FIELD_NUMBER = 2; public static final int LOCK_OWNER_FIELD_NUMBER = 2;
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName lockOwner_; private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName lockOwner_;
public boolean hasLockOwner() { public boolean hasLockOwner() {
return ((bitField0_ & 0x00000002) == 0x00000002); return ((bitField0_ & 0x00000002) == 0x00000002);
@ -5114,8 +5114,8 @@ public final class ZooKeeperProtos {
return lockOwner_; return lockOwner_;
} }
// optional int64 threadId = 3; // optional int64 thread_id = 3;
public static final int THREADID_FIELD_NUMBER = 3; public static final int THREAD_ID_FIELD_NUMBER = 3;
private long threadId_; private long threadId_;
public boolean hasThreadId() { public boolean hasThreadId() {
return ((bitField0_ & 0x00000004) == 0x00000004); return ((bitField0_ & 0x00000004) == 0x00000004);
@ -5124,8 +5124,8 @@ public final class ZooKeeperProtos {
return threadId_; return threadId_;
} }
// optional bool isShared = 4; // optional bool is_shared = 4;
public static final int ISSHARED_FIELD_NUMBER = 4; public static final int IS_SHARED_FIELD_NUMBER = 4;
private boolean isShared_; private boolean isShared_;
public boolean hasIsShared() { public boolean hasIsShared() {
return ((bitField0_ & 0x00000008) == 0x00000008); return ((bitField0_ & 0x00000008) == 0x00000008);
@ -5166,8 +5166,8 @@ public final class ZooKeeperProtos {
} }
} }
// optional int64 createTime = 6; // optional int64 create_time = 6;
public static final int CREATETIME_FIELD_NUMBER = 6; public static final int CREATE_TIME_FIELD_NUMBER = 6;
private long createTime_; private long createTime_;
public boolean hasCreateTime() { public boolean hasCreateTime() {
return ((bitField0_ & 0x00000020) == 0x00000020); return ((bitField0_ & 0x00000020) == 0x00000020);
@ -5316,19 +5316,19 @@ public final class ZooKeeperProtos {
int hash = 41; int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode(); hash = (19 * hash) + getDescriptorForType().hashCode();
if (hasTableName()) { if (hasTableName()) {
hash = (37 * hash) + TABLENAME_FIELD_NUMBER; hash = (37 * hash) + TABLE_NAME_FIELD_NUMBER;
hash = (53 * hash) + getTableName().hashCode(); hash = (53 * hash) + getTableName().hashCode();
} }
if (hasLockOwner()) { if (hasLockOwner()) {
hash = (37 * hash) + LOCKOWNER_FIELD_NUMBER; hash = (37 * hash) + LOCK_OWNER_FIELD_NUMBER;
hash = (53 * hash) + getLockOwner().hashCode(); hash = (53 * hash) + getLockOwner().hashCode();
} }
if (hasThreadId()) { if (hasThreadId()) {
hash = (37 * hash) + THREADID_FIELD_NUMBER; hash = (37 * hash) + THREAD_ID_FIELD_NUMBER;
hash = (53 * hash) + hashLong(getThreadId()); hash = (53 * hash) + hashLong(getThreadId());
} }
if (hasIsShared()) { if (hasIsShared()) {
hash = (37 * hash) + ISSHARED_FIELD_NUMBER; hash = (37 * hash) + IS_SHARED_FIELD_NUMBER;
hash = (53 * hash) + hashBoolean(getIsShared()); hash = (53 * hash) + hashBoolean(getIsShared());
} }
if (hasPurpose()) { if (hasPurpose()) {
@ -5336,7 +5336,7 @@ public final class ZooKeeperProtos {
hash = (53 * hash) + getPurpose().hashCode(); hash = (53 * hash) + getPurpose().hashCode();
} }
if (hasCreateTime()) { if (hasCreateTime()) {
hash = (37 * hash) + CREATETIME_FIELD_NUMBER; hash = (37 * hash) + CREATE_TIME_FIELD_NUMBER;
hash = (53 * hash) + hashLong(getCreateTime()); hash = (53 * hash) + hashLong(getCreateTime());
} }
hash = (29 * hash) + getUnknownFields().hashCode(); hash = (29 * hash) + getUnknownFields().hashCode();
@ -5649,7 +5649,7 @@ public final class ZooKeeperProtos {
private int bitField0_; private int bitField0_;
// optional bytes tableName = 1; // optional bytes table_name = 1;
private com.google.protobuf.ByteString tableName_ = com.google.protobuf.ByteString.EMPTY; private com.google.protobuf.ByteString tableName_ = com.google.protobuf.ByteString.EMPTY;
public boolean hasTableName() { public boolean hasTableName() {
return ((bitField0_ & 0x00000001) == 0x00000001); return ((bitField0_ & 0x00000001) == 0x00000001);
@ -5673,7 +5673,7 @@ public final class ZooKeeperProtos {
return this; return this;
} }
// optional .ServerName lockOwner = 2; // optional .ServerName lock_owner = 2;
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName lockOwner_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.getDefaultInstance(); private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName lockOwner_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.getDefaultInstance();
private com.google.protobuf.SingleFieldBuilder< private com.google.protobuf.SingleFieldBuilder<
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerNameOrBuilder> lockOwnerBuilder_; org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerNameOrBuilder> lockOwnerBuilder_;
@ -5763,7 +5763,7 @@ public final class ZooKeeperProtos {
return lockOwnerBuilder_; return lockOwnerBuilder_;
} }
// optional int64 threadId = 3; // optional int64 thread_id = 3;
private long threadId_ ; private long threadId_ ;
public boolean hasThreadId() { public boolean hasThreadId() {
return ((bitField0_ & 0x00000004) == 0x00000004); return ((bitField0_ & 0x00000004) == 0x00000004);
@ -5784,7 +5784,7 @@ public final class ZooKeeperProtos {
return this; return this;
} }
// optional bool isShared = 4; // optional bool is_shared = 4;
private boolean isShared_ ; private boolean isShared_ ;
public boolean hasIsShared() { public boolean hasIsShared() {
return ((bitField0_ & 0x00000008) == 0x00000008); return ((bitField0_ & 0x00000008) == 0x00000008);
@ -5841,7 +5841,7 @@ public final class ZooKeeperProtos {
onChanged(); onChanged();
} }
// optional int64 createTime = 6; // optional int64 create_time = 6;
private long createTime_ ; private long createTime_ ;
public boolean hasCreateTime() { public boolean hasCreateTime() {
return ((bitField0_ & 0x00000020) == 0x00000020); return ((bitField0_ & 0x00000020) == 0x00000020);
@ -5876,11 +5876,11 @@ public final class ZooKeeperProtos {
public interface StoreSequenceIdOrBuilder public interface StoreSequenceIdOrBuilder
extends com.google.protobuf.MessageOrBuilder { extends com.google.protobuf.MessageOrBuilder {
// required bytes familyName = 1; // required bytes family_name = 1;
boolean hasFamilyName(); boolean hasFamilyName();
com.google.protobuf.ByteString getFamilyName(); com.google.protobuf.ByteString getFamilyName();
// required uint64 sequenceId = 2; // required uint64 sequence_id = 2;
boolean hasSequenceId(); boolean hasSequenceId();
long getSequenceId(); long getSequenceId();
} }
@ -5913,8 +5913,8 @@ public final class ZooKeeperProtos {
} }
private int bitField0_; private int bitField0_;
// required bytes familyName = 1; // required bytes family_name = 1;
public static final int FAMILYNAME_FIELD_NUMBER = 1; public static final int FAMILY_NAME_FIELD_NUMBER = 1;
private com.google.protobuf.ByteString familyName_; private com.google.protobuf.ByteString familyName_;
public boolean hasFamilyName() { public boolean hasFamilyName() {
return ((bitField0_ & 0x00000001) == 0x00000001); return ((bitField0_ & 0x00000001) == 0x00000001);
@ -5923,8 +5923,8 @@ public final class ZooKeeperProtos {
return familyName_; return familyName_;
} }
// required uint64 sequenceId = 2; // required uint64 sequence_id = 2;
public static final int SEQUENCEID_FIELD_NUMBER = 2; public static final int SEQUENCE_ID_FIELD_NUMBER = 2;
private long sequenceId_; private long sequenceId_;
public boolean hasSequenceId() { public boolean hasSequenceId() {
return ((bitField0_ & 0x00000002) == 0x00000002); return ((bitField0_ & 0x00000002) == 0x00000002);
@ -6023,11 +6023,11 @@ public final class ZooKeeperProtos {
int hash = 41; int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode(); hash = (19 * hash) + getDescriptorForType().hashCode();
if (hasFamilyName()) { if (hasFamilyName()) {
hash = (37 * hash) + FAMILYNAME_FIELD_NUMBER; hash = (37 * hash) + FAMILY_NAME_FIELD_NUMBER;
hash = (53 * hash) + getFamilyName().hashCode(); hash = (53 * hash) + getFamilyName().hashCode();
} }
if (hasSequenceId()) { if (hasSequenceId()) {
hash = (37 * hash) + SEQUENCEID_FIELD_NUMBER; hash = (37 * hash) + SEQUENCE_ID_FIELD_NUMBER;
hash = (53 * hash) + hashLong(getSequenceId()); hash = (53 * hash) + hashLong(getSequenceId());
} }
hash = (29 * hash) + getUnknownFields().hashCode(); hash = (29 * hash) + getUnknownFields().hashCode();
@ -6273,7 +6273,7 @@ public final class ZooKeeperProtos {
private int bitField0_; private int bitField0_;
// required bytes familyName = 1; // required bytes family_name = 1;
private com.google.protobuf.ByteString familyName_ = com.google.protobuf.ByteString.EMPTY; private com.google.protobuf.ByteString familyName_ = com.google.protobuf.ByteString.EMPTY;
public boolean hasFamilyName() { public boolean hasFamilyName() {
return ((bitField0_ & 0x00000001) == 0x00000001); return ((bitField0_ & 0x00000001) == 0x00000001);
@ -6297,7 +6297,7 @@ public final class ZooKeeperProtos {
return this; return this;
} }
// required uint64 sequenceId = 2; // required uint64 sequence_id = 2;
private long sequenceId_ ; private long sequenceId_ ;
public boolean hasSequenceId() { public boolean hasSequenceId() {
return ((bitField0_ & 0x00000002) == 0x00000002); return ((bitField0_ & 0x00000002) == 0x00000002);
@ -6332,11 +6332,11 @@ public final class ZooKeeperProtos {
public interface RegionStoreSequenceIdsOrBuilder public interface RegionStoreSequenceIdsOrBuilder
extends com.google.protobuf.MessageOrBuilder { extends com.google.protobuf.MessageOrBuilder {
// required uint64 lastFlushedSequenceId = 1; // required uint64 last_flushed_sequence_id = 1;
boolean hasLastFlushedSequenceId(); boolean hasLastFlushedSequenceId();
long getLastFlushedSequenceId(); long getLastFlushedSequenceId();
// repeated .StoreSequenceId storeSequenceId = 2; // repeated .StoreSequenceId store_sequence_id = 2;
java.util.List<org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.StoreSequenceId> java.util.List<org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.StoreSequenceId>
getStoreSequenceIdList(); getStoreSequenceIdList();
org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.StoreSequenceId getStoreSequenceId(int index); org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.StoreSequenceId getStoreSequenceId(int index);
@ -6375,8 +6375,8 @@ public final class ZooKeeperProtos {
} }
private int bitField0_; private int bitField0_;
// required uint64 lastFlushedSequenceId = 1; // required uint64 last_flushed_sequence_id = 1;
public static final int LASTFLUSHEDSEQUENCEID_FIELD_NUMBER = 1; public static final int LAST_FLUSHED_SEQUENCE_ID_FIELD_NUMBER = 1;
private long lastFlushedSequenceId_; private long lastFlushedSequenceId_;
public boolean hasLastFlushedSequenceId() { public boolean hasLastFlushedSequenceId() {
return ((bitField0_ & 0x00000001) == 0x00000001); return ((bitField0_ & 0x00000001) == 0x00000001);
@ -6385,8 +6385,8 @@ public final class ZooKeeperProtos {
return lastFlushedSequenceId_; return lastFlushedSequenceId_;
} }
// repeated .StoreSequenceId storeSequenceId = 2; // repeated .StoreSequenceId store_sequence_id = 2;
public static final int STORESEQUENCEID_FIELD_NUMBER = 2; public static final int STORE_SEQUENCE_ID_FIELD_NUMBER = 2;
private java.util.List<org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.StoreSequenceId> storeSequenceId_; private java.util.List<org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.StoreSequenceId> storeSequenceId_;
public java.util.List<org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.StoreSequenceId> getStoreSequenceIdList() { public java.util.List<org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.StoreSequenceId> getStoreSequenceIdList() {
return storeSequenceId_; return storeSequenceId_;
@ -6495,11 +6495,11 @@ public final class ZooKeeperProtos {
int hash = 41; int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode(); hash = (19 * hash) + getDescriptorForType().hashCode();
if (hasLastFlushedSequenceId()) { if (hasLastFlushedSequenceId()) {
hash = (37 * hash) + LASTFLUSHEDSEQUENCEID_FIELD_NUMBER; hash = (37 * hash) + LAST_FLUSHED_SEQUENCE_ID_FIELD_NUMBER;
hash = (53 * hash) + hashLong(getLastFlushedSequenceId()); hash = (53 * hash) + hashLong(getLastFlushedSequenceId());
} }
if (getStoreSequenceIdCount() > 0) { if (getStoreSequenceIdCount() > 0) {
hash = (37 * hash) + STORESEQUENCEID_FIELD_NUMBER; hash = (37 * hash) + STORE_SEQUENCE_ID_FIELD_NUMBER;
hash = (53 * hash) + getStoreSequenceIdList().hashCode(); hash = (53 * hash) + getStoreSequenceIdList().hashCode();
} }
hash = (29 * hash) + getUnknownFields().hashCode(); hash = (29 * hash) + getUnknownFields().hashCode();
@ -6781,7 +6781,7 @@ public final class ZooKeeperProtos {
private int bitField0_; private int bitField0_;
// required uint64 lastFlushedSequenceId = 1; // required uint64 last_flushed_sequence_id = 1;
private long lastFlushedSequenceId_ ; private long lastFlushedSequenceId_ ;
public boolean hasLastFlushedSequenceId() { public boolean hasLastFlushedSequenceId() {
return ((bitField0_ & 0x00000001) == 0x00000001); return ((bitField0_ & 0x00000001) == 0x00000001);
@ -6802,7 +6802,7 @@ public final class ZooKeeperProtos {
return this; return this;
} }
// repeated .StoreSequenceId storeSequenceId = 2; // repeated .StoreSequenceId store_sequence_id = 2;
private java.util.List<org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.StoreSequenceId> storeSequenceId_ = private java.util.List<org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.StoreSequenceId> storeSequenceId_ =
java.util.Collections.emptyList(); java.util.Collections.emptyList();
private void ensureStoreSequenceIdIsMutable() { private void ensureStoreSequenceIdIsMutable() {
@ -7073,36 +7073,36 @@ public final class ZooKeeperProtos {
descriptor; descriptor;
static { static {
java.lang.String[] descriptorData = { java.lang.String[] descriptorData = {
"\n\017ZooKeeper.proto\032\013hbase.proto\"C\n\020MetaRe" + "\n\017ZooKeeper.proto\032\013hbase.proto\"D\n\020MetaRe" +
"gionServer\022\033\n\006server\030\001 \002(\0132\013.ServerName\022" + "gionServer\022\033\n\006server\030\001 \002(\0132\013.ServerName\022" +
"\022\n\nrpcVersion\030\002 \001(\r\"9\n\006Master\022\033\n\006master\030" + "\023\n\013rpc_version\030\002 \001(\r\":\n\006Master\022\033\n\006master" +
"\001 \002(\0132\013.ServerName\022\022\n\nrpcVersion\030\002 \001(\r\"\036" + "\030\001 \002(\0132\013.ServerName\022\023\n\013rpc_version\030\002 \001(\r" +
"\n\tClusterUp\022\021\n\tstartDate\030\001 \002(\t\"\203\001\n\020Regio" + "\"\037\n\tClusterUp\022\022\n\nstart_date\030\001 \002(\t\"\210\001\n\020Re" +
"nTransition\022\025\n\reventTypeCode\030\001 \002(\r\022\022\n\nre" + "gionTransition\022\027\n\017event_type_code\030\001 \002(\r\022" +
"gionName\030\002 \002(\014\022\022\n\ncreateTime\030\003 \002(\004\022\037\n\nse" + "\023\n\013region_name\030\002 \002(\014\022\023\n\013create_time\030\003 \002(" +
"rverName\030\004 \002(\0132\013.ServerName\022\017\n\007payload\030\005" + "\004\022 \n\013server_name\030\004 \002(\0132\013.ServerName\022\017\n\007p" +
" \001(\014\"\230\001\n\014SplitLogTask\022\"\n\005state\030\001 \002(\0162\023.S" + "ayload\030\005 \001(\014\"\231\001\n\014SplitLogTask\022\"\n\005state\030\001" +
"plitLogTask.State\022\037\n\nserverName\030\002 \002(\0132\013.", " \002(\0162\023.SplitLogTask.State\022 \n\013server_name",
"ServerName\"C\n\005State\022\016\n\nUNASSIGNED\020\000\022\t\n\005O" + "\030\002 \002(\0132\013.ServerName\"C\n\005State\022\016\n\nUNASSIGN" +
"WNED\020\001\022\014\n\010RESIGNED\020\002\022\010\n\004DONE\020\003\022\007\n\003ERR\020\004\"" + "ED\020\000\022\t\n\005OWNED\020\001\022\014\n\010RESIGNED\020\002\022\010\n\004DONE\020\003\022" +
"n\n\005Table\022$\n\005state\030\001 \002(\0162\014.Table.State:\007E" + "\007\n\003ERR\020\004\"n\n\005Table\022$\n\005state\030\001 \002(\0162\014.Table" +
"NABLED\"?\n\005State\022\013\n\007ENABLED\020\000\022\014\n\010DISABLED" + ".State:\007ENABLED\"?\n\005State\022\013\n\007ENABLED\020\000\022\014\n" +
"\020\001\022\r\n\tDISABLING\020\002\022\014\n\010ENABLING\020\003\"%\n\017Repli" + "\010DISABLED\020\001\022\r\n\tDISABLING\020\002\022\014\n\010ENABLING\020\003" +
"cationPeer\022\022\n\nclusterkey\030\001 \002(\t\"^\n\020Replic" + "\"%\n\017ReplicationPeer\022\022\n\nclusterkey\030\001 \002(\t\"" +
"ationState\022&\n\005state\030\001 \002(\0162\027.ReplicationS" + "^\n\020ReplicationState\022&\n\005state\030\001 \002(\0162\027.Rep" +
"tate.State\"\"\n\005State\022\013\n\007ENABLED\020\000\022\014\n\010DISA" + "licationState.State\"\"\n\005State\022\013\n\007ENABLED\020" +
"BLED\020\001\"+\n\027ReplicationHLogPosition\022\020\n\010pos" + "\000\022\014\n\010DISABLED\020\001\"+\n\027ReplicationHLogPositi" +
"ition\030\001 \002(\003\"$\n\017ReplicationLock\022\021\n\tlockOw", "on\022\020\n\010position\030\001 \002(\003\"%\n\017ReplicationLock\022",
"ner\030\001 \002(\t\"\207\001\n\tTableLock\022\021\n\ttableName\030\001 \001" + "\022\n\nlock_owner\030\001 \002(\t\"\214\001\n\tTableLock\022\022\n\ntab" +
"(\014\022\036\n\tlockOwner\030\002 \001(\0132\013.ServerName\022\020\n\010th" + "le_name\030\001 \001(\014\022\037\n\nlock_owner\030\002 \001(\0132\013.Serv" +
"readId\030\003 \001(\003\022\020\n\010isShared\030\004 \001(\010\022\017\n\007purpos" + "erName\022\021\n\tthread_id\030\003 \001(\003\022\021\n\tis_shared\030\004" +
"e\030\005 \001(\t\022\022\n\ncreateTime\030\006 \001(\003\"9\n\017StoreSequ" + " \001(\010\022\017\n\007purpose\030\005 \001(\t\022\023\n\013create_time\030\006 \001" +
"enceId\022\022\n\nfamilyName\030\001 \002(\014\022\022\n\nsequenceId" + "(\003\";\n\017StoreSequenceId\022\023\n\013family_name\030\001 \002" +
"\030\002 \002(\004\"b\n\026RegionStoreSequenceIds\022\035\n\025last" + "(\014\022\023\n\013sequence_id\030\002 \002(\004\"g\n\026RegionStoreSe" +
"FlushedSequenceId\030\001 \002(\004\022)\n\017storeSequence" + "quenceIds\022 \n\030last_flushed_sequence_id\030\001 " +
"Id\030\002 \003(\0132\020.StoreSequenceIdBE\n*org.apache" + "\002(\004\022+\n\021store_sequence_id\030\002 \003(\0132\020.StoreSe" +
".hadoop.hbase.protobuf.generatedB\017ZooKee" + "quenceIdBE\n*org.apache.hadoop.hbase.prot" +
"perProtosH\001\210\001\001\240\001\001" "obuf.generatedB\017ZooKeeperProtosH\001\210\001\001\240\001\001"
}; };
com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner = com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() { new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() {

View File

@ -85,15 +85,15 @@ message CheckPermissionsResponse {
} }
service AccessControlService { service AccessControlService {
rpc grant(GrantRequest) rpc Grant(GrantRequest)
returns (GrantResponse); returns (GrantResponse);
rpc revoke(RevokeRequest) rpc Revoke(RevokeRequest)
returns (RevokeResponse); returns (RevokeResponse);
rpc getUserPermissions(UserPermissionsRequest) rpc GetUserPermissions(UserPermissionsRequest)
returns (UserPermissionsResponse); returns (UserPermissionsResponse);
rpc checkPermissions(CheckPermissionsRequest) rpc CheckPermissions(CheckPermissionsRequest)
returns (CheckPermissionsResponse); returns (CheckPermissionsResponse);
} }

View File

@ -30,12 +30,12 @@ import "WAL.proto";
message GetRegionInfoRequest { message GetRegionInfoRequest {
required RegionSpecifier region = 1; required RegionSpecifier region = 1;
optional bool compactionState = 2; optional bool compaction_state = 2;
} }
message GetRegionInfoResponse { message GetRegionInfoResponse {
required RegionInfo regionInfo = 1; required RegionInfo region_info = 1;
optional CompactionState compactionState = 2; optional CompactionState compaction_state = 2;
enum CompactionState { enum CompactionState {
NONE = 0; NONE = 0;
@ -55,28 +55,28 @@ message GetStoreFileRequest {
} }
message GetStoreFileResponse { message GetStoreFileResponse {
repeated string storeFile = 1; repeated string store_file = 1;
} }
message GetOnlineRegionRequest { message GetOnlineRegionRequest {
} }
message GetOnlineRegionResponse { message GetOnlineRegionResponse {
repeated RegionInfo regionInfo = 1; repeated RegionInfo region_info = 1;
} }
message OpenRegionRequest { message OpenRegionRequest {
repeated RegionOpenInfo openInfo = 1; repeated RegionOpenInfo open_info = 1;
message RegionOpenInfo { message RegionOpenInfo {
required RegionInfo region = 1; required RegionInfo region = 1;
optional uint32 versionOfOfflineNode = 2; optional uint32 version_of_offline_node = 2;
repeated ServerName favoredNodes = 3; repeated ServerName favored_nodes = 3;
} }
} }
message OpenRegionResponse { message OpenRegionResponse {
repeated RegionOpeningState openingState = 1; repeated RegionOpeningState opening_state = 1;
enum RegionOpeningState { enum RegionOpeningState {
OPENED = 0; OPENED = 0;
@ -91,9 +91,9 @@ message OpenRegionResponse {
*/ */
message CloseRegionRequest { message CloseRegionRequest {
required RegionSpecifier region = 1; required RegionSpecifier region = 1;
optional uint32 versionOfClosingNode = 2; optional uint32 version_of_closing_node = 2;
optional bool transitionInZK = 3 [default = true]; optional bool transition_in_ZK = 3 [default = true];
optional ServerName destinationServer = 4; optional ServerName destination_server = 4;
} }
message CloseRegionResponse { message CloseRegionResponse {
@ -107,11 +107,11 @@ message CloseRegionResponse {
*/ */
message FlushRegionRequest { message FlushRegionRequest {
required RegionSpecifier region = 1; required RegionSpecifier region = 1;
optional uint64 ifOlderThanTs = 2; optional uint64 if_older_than_ts = 2;
} }
message FlushRegionResponse { message FlushRegionResponse {
required uint64 lastFlushTime = 1; required uint64 last_flush_time = 1;
optional bool flushed = 2; optional bool flushed = 2;
} }
@ -124,7 +124,7 @@ message FlushRegionResponse {
*/ */
message SplitRegionRequest { message SplitRegionRequest {
required RegionSpecifier region = 1; required RegionSpecifier region = 1;
optional bytes splitPoint = 2; optional bytes split_point = 2;
} }
message SplitRegionResponse { message SplitRegionResponse {
@ -150,8 +150,8 @@ message CompactRegionResponse {
* This method currently closes the regions and then merges them * This method currently closes the regions and then merges them
*/ */
message MergeRegionsRequest { message MergeRegionsRequest {
required RegionSpecifier regionA = 1; required RegionSpecifier region_a = 1;
required RegionSpecifier regionB = 2; required RegionSpecifier region_b = 2;
optional bool forcible = 3 [default = false]; optional bool forcible = 3 [default = false];
} }
@ -163,10 +163,10 @@ message WALEntry {
required WALKey key = 1; required WALKey key = 1;
// Following may be null if the KVs/Cells are carried along the side in a cellblock (See // Following may be null if the KVs/Cells are carried along the side in a cellblock (See
// RPC for more on cellblocks). If Cells/KVs are in a cellblock, this next field is null // RPC for more on cellblocks). If Cells/KVs are in a cellblock, this next field is null
// and associatedCellCount has count of Cells associated w/ this WALEntry // and associated_cell_count has count of Cells associated w/ this WALEntry
repeated bytes keyValueBytes = 2; repeated bytes key_value_bytes = 2;
// If Cell data is carried alongside in a cellblock, this is count of Cells in the cellblock. // If Cell data is carried alongside in a cellblock, this is count of Cells in the cellblock.
optional int32 associatedCellCount = 3; optional int32 associated_cell_count = 3;
} }
/** /**
@ -186,7 +186,7 @@ message RollWALWriterRequest {
message RollWALWriterResponse { message RollWALWriterResponse {
// A list of encoded name of regions to flush // A list of encoded name of regions to flush
repeated bytes regionToFlush = 1; repeated bytes region_to_flush = 1;
} }
message StopServerRequest { message StopServerRequest {
@ -200,54 +200,54 @@ message GetServerInfoRequest {
} }
message ServerInfo { message ServerInfo {
required ServerName serverName = 1; required ServerName server_name = 1;
optional uint32 webuiPort = 2; optional uint32 webui_port = 2;
} }
message GetServerInfoResponse { message GetServerInfoResponse {
required ServerInfo serverInfo = 1; required ServerInfo server_info = 1;
} }
service AdminService { service AdminService {
rpc getRegionInfo(GetRegionInfoRequest) rpc GetRegionInfo(GetRegionInfoRequest)
returns(GetRegionInfoResponse); returns(GetRegionInfoResponse);
rpc getStoreFile(GetStoreFileRequest) rpc GetStoreFile(GetStoreFileRequest)
returns(GetStoreFileResponse); returns(GetStoreFileResponse);
rpc getOnlineRegion(GetOnlineRegionRequest) rpc GetOnlineRegion(GetOnlineRegionRequest)
returns(GetOnlineRegionResponse); returns(GetOnlineRegionResponse);
rpc openRegion(OpenRegionRequest) rpc OpenRegion(OpenRegionRequest)
returns(OpenRegionResponse); returns(OpenRegionResponse);
rpc closeRegion(CloseRegionRequest) rpc CloseRegion(CloseRegionRequest)
returns(CloseRegionResponse); returns(CloseRegionResponse);
rpc flushRegion(FlushRegionRequest) rpc FlushRegion(FlushRegionRequest)
returns(FlushRegionResponse); returns(FlushRegionResponse);
rpc splitRegion(SplitRegionRequest) rpc SplitRegion(SplitRegionRequest)
returns(SplitRegionResponse); returns(SplitRegionResponse);
rpc compactRegion(CompactRegionRequest) rpc CompactRegion(CompactRegionRequest)
returns(CompactRegionResponse); returns(CompactRegionResponse);
rpc mergeRegions(MergeRegionsRequest) rpc MergeRegions(MergeRegionsRequest)
returns(MergeRegionsResponse); returns(MergeRegionsResponse);
rpc replicateWALEntry(ReplicateWALEntryRequest) rpc ReplicateWALEntry(ReplicateWALEntryRequest)
returns(ReplicateWALEntryResponse); returns(ReplicateWALEntryResponse);
rpc replay(MultiRequest) rpc Replay(MultiRequest)
returns(MultiResponse); returns(MultiResponse);
rpc rollWALWriter(RollWALWriterRequest) rpc RollWALWriter(RollWALWriterRequest)
returns(RollWALWriterResponse); returns(RollWALWriterResponse);
rpc getServerInfo(GetServerInfoRequest) rpc GetServerInfo(GetServerInfoRequest)
returns(GetServerInfoResponse); returns(GetServerInfoResponse);
rpc stopServer(StopServerRequest) rpc StopServer(StopServerRequest)
returns(StopServerResponse); returns(StopServerResponse);
} }

View File

@ -31,32 +31,32 @@ message AggregateArgument {
* (3) any bytes required to construct the ColumnInterpreter object * (3) any bytes required to construct the ColumnInterpreter object
* properly * properly
*/ */
required string interpreterClassName = 1; required string interpreter_class_name = 1;
required Scan scan = 2; required Scan scan = 2;
optional bytes interpreterSpecificBytes = 3; optional bytes interpreter_specific_bytes = 3;
} }
message AggregateResponse { message AggregateResponse {
/** /**
* The AggregateService methods all have a response that either is a Pair * The AggregateService methods all have a response that either is a Pair
* or a simple object. When it is a Pair both firstPart and secondPart * or a simple object. When it is a Pair both first_part and second_part
* have defined values (and the secondPart is not present in the response * have defined values (and the second_part is not present in the response
* when the response is not a pair). Refer to the AggregateImplementation * when the response is not a pair). Refer to the AggregateImplementation
* class for an overview of the AggregateResponse object constructions. * class for an overview of the AggregateResponse object constructions.
*/ */
repeated bytes firstPart = 1; repeated bytes first_part = 1;
optional bytes secondPart = 2; optional bytes second_part = 2;
} }
/** Refer to the AggregateImplementation class for an overview of the /** Refer to the AggregateImplementation class for an overview of the
* AggregateService method implementations and their functionality. * AggregateService method implementations and their functionality.
*/ */
service AggregateService { service AggregateService {
rpc getMax (AggregateArgument) returns (AggregateResponse); rpc GetMax (AggregateArgument) returns (AggregateResponse);
rpc getMin (AggregateArgument) returns (AggregateResponse); rpc GetMin (AggregateArgument) returns (AggregateResponse);
rpc getSum (AggregateArgument) returns (AggregateResponse); rpc GetSum (AggregateArgument) returns (AggregateResponse);
rpc getRowNum (AggregateArgument) returns (AggregateResponse); rpc GetRowNum (AggregateArgument) returns (AggregateResponse);
rpc getAvg (AggregateArgument) returns (AggregateResponse); rpc GetAvg (AggregateArgument) returns (AggregateResponse);
rpc getStd (AggregateArgument) returns (AggregateResponse); rpc GetStd (AggregateArgument) returns (AggregateResponse);
rpc getMedian (AggregateArgument) returns (AggregateResponse); rpc GetMedian (AggregateArgument) returns (AggregateResponse);
} }

View File

@ -24,7 +24,7 @@ option optimize_for = SPEED;
message AuthenticationKey { message AuthenticationKey {
required int32 id = 1; required int32 id = 1;
required int64 expirationDate = 2; required int64 expiration_date = 2;
required bytes key = 3; required bytes key = 3;
} }
@ -35,10 +35,10 @@ message TokenIdentifier {
} }
required Kind kind = 1; required Kind kind = 1;
required bytes username = 2; required bytes username = 2;
required int32 keyId = 3; required int32 key_id = 3;
optional int64 issueDate = 4; optional int64 issue_date = 4;
optional int64 expirationDate = 5; optional int64 expiration_date = 5;
optional int64 sequenceNumber = 6; optional int64 sequence_number = 6;
} }
@ -67,15 +67,15 @@ message WhoAmIRequest {
message WhoAmIResponse { message WhoAmIResponse {
optional string username = 1; optional string username = 1;
optional string authMethod = 2; optional string auth_method = 2;
} }
// RPC service // RPC service
service AuthenticationService { service AuthenticationService {
rpc getAuthenticationToken(TokenRequest) rpc GetAuthenticationToken(TokenRequest)
returns (TokenResponse); returns (TokenResponse);
rpc whoami(WhoAmIRequest) rpc WhoAmI(WhoAmIRequest)
returns (WhoAmIResponse); returns (WhoAmIResponse);
} }

View File

@ -46,7 +46,7 @@ message Cell {
optional bytes family = 2; optional bytes family = 2;
optional bytes qualifier = 3; optional bytes qualifier = 3;
optional uint64 timestamp = 4; optional uint64 timestamp = 4;
optional CellType cellType = 5; optional CellType cell_type = 5;
optional bytes value = 6; optional bytes value = 6;
} }
@ -59,6 +59,6 @@ message KeyValue {
required bytes family = 2; required bytes family = 2;
required bytes qualifier = 3; required bytes qualifier = 3;
optional uint64 timestamp = 4; optional uint64 timestamp = 4;
optional CellType keyType = 5; optional CellType key_type = 5;
optional bytes value = 6; optional bytes value = 6;
} }

View File

@ -45,11 +45,11 @@ message Get {
repeated Column column = 2; repeated Column column = 2;
repeated NameBytesPair attribute = 3; repeated NameBytesPair attribute = 3;
optional Filter filter = 4; optional Filter filter = 4;
optional TimeRange timeRange = 5; optional TimeRange time_range = 5;
optional uint32 maxVersions = 6 [default = 1]; optional uint32 max_versions = 6 [default = 1];
optional bool cacheBlocks = 7 [default = true]; optional bool cache_blocks = 7 [default = true];
optional uint32 storeLimit = 8; optional uint32 store_limit = 8;
optional uint32 storeOffset = 9; optional uint32 store_offset = 9;
} }
message Result { message Result {
@ -60,18 +60,18 @@ message Result {
// not part of this protobuf message; they are passed alongside // not part of this protobuf message; they are passed alongside
// and then this Message is just a placeholder with metadata. // and then this Message is just a placeholder with metadata.
// The count is needed to know how many to peel off the block of Cells as // The count is needed to know how many to peel off the block of Cells as
// ours. NOTE: This is different from the pb managed cellCount of the // ours. NOTE: This is different from the pb managed cell_count of the
// 'cell' field above which is non-null when the cells are pb'd. // 'cell' field above which is non-null when the cells are pb'd.
optional int32 associatedCellCount = 2; optional int32 associated_cell_count = 2;
} }
/** /**
* The get request. Perform a single Get operation. * The get request. Perform a single Get operation.
* Unless existenceOnly is specified, return all the requested data * Unless existence_only is specified, return all the requested data
* for the row that matches exactly, or the one that immediately * for the row that matches exactly, or the one that immediately
* precedes it if closestRowBefore is specified. * precedes it if closest_row_before is specified.
* *
* If existenceOnly is set, only the existence will be returned. * If existence_only is set, only the existence will be returned.
*/ */
message GetRequest { message GetRequest {
required RegionSpecifier region = 1; required RegionSpecifier region = 1;
@ -79,12 +79,12 @@ message GetRequest {
// If the row to get doesn't exist, return the // If the row to get doesn't exist, return the
// closest row before. // closest row before.
optional bool closestRowBefore = 3; optional bool closest_row_before = 3;
// The result isn't asked for, just check for // The result isn't asked for, just check for
// the existence. If closestRowBefore specified, // the existence. If closest_row_before specified,
// this will be ignored // this will be ignored
optional bool existenceOnly = 4; optional bool existence_only = 4;
} }
message MultiGetRequest { message MultiGetRequest {
@ -93,12 +93,12 @@ message MultiGetRequest {
// If the row to get doesn't exist, return the // If the row to get doesn't exist, return the
// closest row before. // closest row before.
optional bool closestRowBefore = 3; optional bool closest_row_before = 3;
// The result isn't asked for, just check for // The result isn't asked for, just check for
// the existence. If closestRowBefore specified, // the existence. If closest_row_before specified,
// this will be ignored // this will be ignored
optional bool existenceOnly = 4; optional bool existence_only = 4;
} }
message GetResponse { message GetResponse {
@ -125,7 +125,7 @@ message Condition {
required bytes row = 1; required bytes row = 1;
required bytes family = 2; required bytes family = 2;
required bytes qualifier = 3; required bytes qualifier = 3;
required CompareType compareType = 4; required CompareType compare_type = 4;
required Comparator comparator = 5; required Comparator comparator = 5;
} }
@ -139,22 +139,22 @@ message Condition {
*/ */
message MutationProto { message MutationProto {
optional bytes row = 1; optional bytes row = 1;
optional MutationType mutateType = 2; optional MutationType mutate_type = 2;
repeated ColumnValue columnValue = 3; repeated ColumnValue column_value = 3;
optional uint64 timestamp = 4; optional uint64 timestamp = 4;
repeated NameBytesPair attribute = 5; repeated NameBytesPair attribute = 5;
optional Durability durability = 6 [default = USE_DEFAULT]; optional Durability durability = 6 [default = USE_DEFAULT];
// For some mutations, a result may be returned, in which case, // For some mutations, a result may be returned, in which case,
// time range can be specified for potential performance gain // time range can be specified for potential performance gain
optional TimeRange timeRange = 7; optional TimeRange time_range = 7;
// The below count is set when the associated cells are NOT // The below count is set when the associated cells are NOT
// part of this protobuf message; they are passed alongside // part of this protobuf message; they are passed alongside
// and then this Message is a placeholder with metadata. The // and then this Message is a placeholder with metadata. The
// count is needed to know how many to peel off the block of Cells as // count is needed to know how many to peel off the block of Cells as
// ours. NOTE: This is different from the pb managed cellCount of the // ours. NOTE: This is different from the pb managed cell_count of the
// 'cell' field above which is non-null when the cells are pb'd. // 'cell' field above which is non-null when the cells are pb'd.
optional int32 associatedCellCount = 8; optional int32 associated_cell_count = 8;
enum Durability { enum Durability {
USE_DEFAULT = 0; USE_DEFAULT = 0;
@ -180,13 +180,13 @@ message MutationProto {
message ColumnValue { message ColumnValue {
required bytes family = 1; required bytes family = 1;
repeated QualifierValue qualifierValue = 2; repeated QualifierValue qualifier_value = 2;
message QualifierValue { message QualifierValue {
optional bytes qualifier = 1; optional bytes qualifier = 1;
optional bytes value = 2; optional bytes value = 2;
optional uint64 timestamp = 3; optional uint64 timestamp = 3;
optional DeleteType deleteType = 4; optional DeleteType delete_type = 4;
} }
} }
} }
@ -225,18 +225,18 @@ message MutateResponse {
message Scan { message Scan {
repeated Column column = 1; repeated Column column = 1;
repeated NameBytesPair attribute = 2; repeated NameBytesPair attribute = 2;
optional bytes startRow = 3; optional bytes start_row = 3;
optional bytes stopRow = 4; optional bytes stop_row = 4;
optional Filter filter = 5; optional Filter filter = 5;
optional TimeRange timeRange = 6; optional TimeRange time_range = 6;
optional uint32 maxVersions = 7 [default = 1]; optional uint32 max_versions = 7 [default = 1];
optional bool cacheBlocks = 8 [default = true]; optional bool cache_blocks = 8 [default = true];
optional uint32 batchSize = 9; optional uint32 batch_size = 9;
optional uint64 maxResultSize = 10; optional uint64 max_result_size = 10;
optional uint32 storeLimit = 11; optional uint32 store_limit = 11;
optional uint32 storeOffset = 12; optional uint32 store_offset = 12;
optional bool loadColumnFamiliesOnDemand = 13; /* DO NOT add defaults to loadColumnFamiliesOnDemand. */ optional bool load_column_families_on_demand = 13; /* DO NOT add defaults to load_column_families_on_demand. */
optional uint32 cachingCount = 14; optional uint32 caching_count = 14;
optional bool prefetching = 15; optional bool prefetching = 15;
} }
@ -254,25 +254,25 @@ message Scan {
message ScanRequest { message ScanRequest {
optional RegionSpecifier region = 1; optional RegionSpecifier region = 1;
optional Scan scan = 2; optional Scan scan = 2;
optional uint64 scannerId = 3; optional uint64 scanner_id = 3;
optional uint32 numberOfRows = 4; optional uint32 number_of_rows = 4;
optional bool closeScanner = 5; optional bool close_scanner = 5;
optional uint64 nextCallSeq = 6; optional uint64 next_call_seq = 6;
} }
/** /**
* The scan response. If there are no more results, moreResults will * The scan response. If there are no more results, more_results will
* be false. If it is not specified, it means there are more. * be false. If it is not specified, it means there are more.
*/ */
message ScanResponse { message ScanResponse {
optional ResultCellMeta resultCellMeta = 1; optional ResultCellMeta result_cell_meta = 1;
optional uint64 scannerId = 2; optional uint64 scanner_id = 2;
optional bool moreResults = 3; optional bool more_results = 3;
optional uint32 ttl = 4; optional uint32 ttl = 4;
} }
message ResultCellMeta { message ResultCellMeta {
repeated uint32 cellsLength = 1; repeated uint32 cells_length = 1;
} }
/** /**
@ -281,8 +281,8 @@ message ResultCellMeta {
*/ */
message BulkLoadHFileRequest { message BulkLoadHFileRequest {
required RegionSpecifier region = 1; required RegionSpecifier region = 1;
repeated FamilyPath familyPath = 2; repeated FamilyPath family_path = 2;
optional bool assignSeqNum = 3; optional bool assign_seq_num = 3;
message FamilyPath { message FamilyPath {
required bytes family = 1; required bytes family = 1;
@ -296,8 +296,8 @@ message BulkLoadHFileResponse {
message CoprocessorServiceCall { message CoprocessorServiceCall {
required bytes row = 1; required bytes row = 1;
required string serviceName = 2; required string service_name = 2;
required string methodName = 3; required string method_name = 3;
required bytes request = 4; required bytes request = 4;
} }
@ -352,24 +352,24 @@ message MultiResponse {
service ClientService { service ClientService {
rpc get(GetRequest) rpc Get(GetRequest)
returns(GetResponse); returns(GetResponse);
rpc multiGet(MultiGetRequest) rpc MultiGet(MultiGetRequest)
returns(MultiGetResponse); returns(MultiGetResponse);
rpc mutate(MutateRequest) rpc Mutate(MutateRequest)
returns(MutateResponse); returns(MutateResponse);
rpc scan(ScanRequest) rpc Scan(ScanRequest)
returns(ScanResponse); returns(ScanResponse);
rpc bulkLoadHFile(BulkLoadHFileRequest) rpc BulkLoadHFile(BulkLoadHFileRequest)
returns(BulkLoadHFileResponse); returns(BulkLoadHFileResponse);
rpc execService(CoprocessorServiceRequest) rpc ExecService(CoprocessorServiceRequest)
returns(CoprocessorServiceResponse); returns(CoprocessorServiceResponse);
rpc multi(MultiRequest) rpc Multi(MultiRequest)
returns(MultiResponse); returns(MultiResponse);
} }

View File

@ -29,5 +29,5 @@ option optimize_for = SPEED;
*/ */
message ClusterId { message ClusterId {
// This is the cluster id, a uuid as a String // This is the cluster id, a uuid as a String
required string clusterId = 1; required string cluster_id = 1;
} }

View File

@ -28,7 +28,7 @@ import "ClusterId.proto";
import "FS.proto"; import "FS.proto";
message RegionState { message RegionState {
required RegionInfo regionInfo = 1; required RegionInfo region_info = 1;
required State state = 2; required State state = 2;
optional uint64 stamp = 3; optional uint64 stamp = 3;
enum State { enum State {
@ -50,12 +50,12 @@ message RegionState {
message RegionInTransition { message RegionInTransition {
required RegionSpecifier spec = 1; required RegionSpecifier spec = 1;
required RegionState regionState = 2; required RegionState region_state = 2;
} }
message RegionLoad { message RegionLoad {
/** the region specifier */ /** the region specifier */
required RegionSpecifier regionSpecifier = 1; required RegionSpecifier region_specifier = 1;
/** the number of stores for the region */ /** the number of stores for the region */
optional uint32 stores = 2; optional uint32 stores = 2;
@ -64,65 +64,65 @@ message RegionLoad {
optional uint32 storefiles = 3; optional uint32 storefiles = 3;
/** the total size of the store files for the region, uncompressed, in MB */ /** the total size of the store files for the region, uncompressed, in MB */
optional uint32 storeUncompressedSizeMB = 4; optional uint32 store_uncompressed_size_MB = 4;
/** the current total size of the store files for the region, in MB */ /** the current total size of the store files for the region, in MB */
optional uint32 storefileSizeMB = 5; optional uint32 storefile_size_MB = 5;
/** the current size of the memstore for the region, in MB */ /** the current size of the memstore for the region, in MB */
optional uint32 memstoreSizeMB = 6; optional uint32 memstore_size_MB = 6;
/** /**
* The current total size of root-level store file indexes for the region, * The current total size of root-level store file indexes for the region,
* in MB. The same as {@link #rootIndexSizeKB} but in MB. * in MB. The same as {@link #rootIndexSizeKB} but in MB.
*/ */
optional uint32 storefileIndexSizeMB = 7; optional uint32 storefile_index_size_MB = 7;
/** the current total read requests made to region */ /** the current total read requests made to region */
optional uint64 readRequestsCount = 8; optional uint64 read_requests_count = 8;
/** the current total write requests made to region */ /** the current total write requests made to region */
optional uint64 writeRequestsCount = 9; optional uint64 write_requests_count = 9;
/** the total compacting key values in currently running compaction */ /** the total compacting key values in currently running compaction */
optional uint64 totalCompactingKVs = 10; optional uint64 total_compacting_KVs = 10;
/** the completed count of key values in currently running compaction */ /** the completed count of key values in currently running compaction */
optional uint64 currentCompactedKVs = 11; optional uint64 current_compacted_KVs = 11;
/** The current total size of root-level indexes for the region, in KB. */ /** The current total size of root-level indexes for the region, in KB. */
optional uint32 rootIndexSizeKB = 12; optional uint32 root_index_size_KB = 12;
/** The total size of all index blocks, not just the root level, in KB. */ /** The total size of all index blocks, not just the root level, in KB. */
optional uint32 totalStaticIndexSizeKB = 13; optional uint32 total_static_index_size_KB = 13;
/** /**
* The total size of all Bloom filter blocks, not just loaded into the * The total size of all Bloom filter blocks, not just loaded into the
* block cache, in KB. * block cache, in KB.
*/ */
optional uint32 totalStaticBloomSizeKB = 14; optional uint32 total_static_bloom_size_KB = 14;
/** the most recent sequence Id from cache flush */ /** the most recent sequence Id from cache flush */
optional uint64 completeSequenceId = 15; optional uint64 complete_sequence_id = 15;
} }
/* Server-level protobufs */ /* Server-level protobufs */
message ServerLoad { message ServerLoad {
/** Number of requests since last report. */ /** Number of requests since last report. */
optional uint32 numberOfRequests = 1; optional uint32 number_of_requests = 1;
/** Total Number of requests from the start of the region server. */ /** Total Number of requests from the start of the region server. */
optional uint32 totalNumberOfRequests = 2; optional uint32 total_number_of_requests = 2;
/** the amount of used heap, in MB. */ /** the amount of used heap, in MB. */
optional uint32 usedHeapMB = 3; optional uint32 used_heap_MB = 3;
/** the maximum allowable size of the heap, in MB. */ /** the maximum allowable size of the heap, in MB. */
optional uint32 maxHeapMB = 4; optional uint32 max_heap_MB = 4;
/** Information on the load of individual regions. */ /** Information on the load of individual regions. */
repeated RegionLoad regionLoads = 5; repeated RegionLoad region_loads = 5;
/** /**
* Regionserver-level coprocessors, e.g., WALObserver implementations. * Regionserver-level coprocessors, e.g., WALObserver implementations.
@ -132,39 +132,38 @@ message ServerLoad {
repeated Coprocessor coprocessors = 6; repeated Coprocessor coprocessors = 6;
/** /**
* Time when incremental (non-total) counts began being calculated (e.g. numberOfRequests) * Time when incremental (non-total) counts began being calculated (e.g. number_of_requests)
* time is measured as the difference, measured in milliseconds, between the current time * time is measured as the difference, measured in milliseconds, between the current time
* and midnight, January 1, 1970 UTC. * and midnight, January 1, 1970 UTC.
*/ */
optional uint64 reportStartTime = 7; optional uint64 report_start_time = 7;
/** /**
* Time when report was generated. * Time when report was generated.
* time is measured as the difference, measured in milliseconds, between the current time * time is measured as the difference, measured in milliseconds, between the current time
* and midnight, January 1, 1970 UTC. * and midnight, January 1, 1970 UTC.
*/ */
optional uint64 reportEndTime = 8; optional uint64 report_end_time = 8;
/** /**
* The port number that this region server is hosing an info server on. * The port number that this region server is hosing an info server on.
*/ */
optional uint32 infoServerPort = 9; optional uint32 info_server_port = 9;
} }
message LiveServerInfo { message LiveServerInfo {
required ServerName server = 1; required ServerName server = 1;
required ServerLoad serverLoad = 2; required ServerLoad server_load = 2;
} }
message ClusterStatus { message ClusterStatus {
optional HBaseVersionFileContent hbaseVersion = 1; optional HBaseVersionFileContent hbase_version = 1;
repeated LiveServerInfo liveServers = 2; repeated LiveServerInfo live_servers = 2;
repeated ServerName deadServers = 3; repeated ServerName dead_servers = 3;
repeated RegionInTransition regionsInTransition = 4; repeated RegionInTransition regions_in_transition = 4;
optional ClusterId clusterId = 5; optional ClusterId cluster_id = 5;
repeated Coprocessor masterCoprocessors = 6; repeated Coprocessor master_coprocessors = 6;
optional ServerName master = 7; optional ServerName master = 7;
repeated ServerName backupMasters = 8; repeated ServerName backup_masters = 8;
optional bool balancerOn = 9; optional bool balancer_on = 9;
} }

View File

@ -28,7 +28,7 @@ option optimize_for = SPEED;
message Comparator { message Comparator {
required string name = 1; required string name = 1;
optional bytes serializedComparator = 2; optional bytes serialized_comparator = 2;
} }
message ByteArrayComparable { message ByteArrayComparable {
@ -45,7 +45,7 @@ message BinaryPrefixComparator {
message BitComparator { message BitComparator {
required ByteArrayComparable comparable = 1; required ByteArrayComparable comparable = 1;
required BitwiseOp bitwiseOp = 2; required BitwiseOp bitwise_op = 2;
enum BitwiseOp { enum BitwiseOp {
AND = 1; AND = 1;
@ -59,7 +59,7 @@ message NullComparator {
message RegexStringComparator { message RegexStringComparator {
required string pattern = 1; required string pattern = 1;
required int32 patternFlags = 2; required int32 pattern_flags = 2;
required string charset = 3; required string charset = 3;
} }

View File

@ -28,10 +28,10 @@ option optimize_for = SPEED;
* so we can serialize exceptions. * so we can serialize exceptions.
*/ */
message StackTraceElementMessage { message StackTraceElementMessage {
optional string declaringClass = 1; optional string declaring_class = 1;
optional string methodName = 2; optional string method_name = 2;
optional string fileName = 3; optional string file_name = 3;
optional int32 lineNumber = 4; optional int32 line_number = 4;
} }
/** /**
@ -41,9 +41,9 @@ message StackTraceElementMessage {
* (which should be another protobuffed class). * (which should be another protobuffed class).
*/ */
message GenericExceptionMessage { message GenericExceptionMessage {
optional string className = 1; optional string class_name = 1;
optional string message = 2; optional string message = 2;
optional bytes errorInfo = 3; optional bytes error_info = 3;
repeated StackTraceElementMessage trace = 4; repeated StackTraceElementMessage trace = 4;
} }
@ -53,6 +53,5 @@ message GenericExceptionMessage {
*/ */
message ForeignExceptionMessage { message ForeignExceptionMessage {
optional string source = 1; optional string source = 1;
optional GenericExceptionMessage genericException = 2; optional GenericExceptionMessage generic_exception = 2;
} }

View File

@ -29,7 +29,7 @@ import "Comparator.proto";
message Filter { message Filter {
required string name = 1; required string name = 1;
optional bytes serializedFilter = 2; optional bytes serialized_filter = 2;
} }
message ColumnCountGetFilter { message ColumnCountGetFilter {
@ -39,7 +39,7 @@ message ColumnCountGetFilter {
message ColumnPaginationFilter { message ColumnPaginationFilter {
required int32 limit = 1; required int32 limit = 1;
optional int32 offset = 2; optional int32 offset = 2;
optional bytes columnOffset = 3; optional bytes column_offset = 3;
} }
message ColumnPrefixFilter { message ColumnPrefixFilter {
@ -47,26 +47,26 @@ message ColumnPrefixFilter {
} }
message ColumnRangeFilter { message ColumnRangeFilter {
optional bytes minColumn = 1; optional bytes min_column = 1;
optional bool minColumnInclusive = 2; optional bool min_column_inclusive = 2;
optional bytes maxColumn = 3; optional bytes max_column = 3;
optional bool maxColumnInclusive = 4; optional bool max_column_inclusive = 4;
} }
message CompareFilter { message CompareFilter {
required CompareType compareOp = 1; required CompareType compare_op = 1;
optional Comparator comparator = 2; optional Comparator comparator = 2;
} }
message DependentColumnFilter { message DependentColumnFilter {
required CompareFilter compareFilter = 1; required CompareFilter compare_filter = 1;
optional bytes columnFamily = 2; optional bytes column_family = 2;
optional bytes columnQualifier = 3; optional bytes column_qualifier = 3;
optional bool dropDependentColumn = 4; optional bool drop_dependent_column = 4;
} }
message FamilyFilter { message FamilyFilter {
required CompareFilter compareFilter = 1; required CompareFilter compare_filter = 1;
} }
message FilterList { message FilterList {
@ -91,23 +91,23 @@ message FirstKeyValueMatchingQualifiersFilter {
} }
message FuzzyRowFilter { message FuzzyRowFilter {
repeated BytesBytesPair fuzzyKeysData = 1; repeated BytesBytesPair fuzzy_keys_data = 1;
} }
message InclusiveStopFilter { message InclusiveStopFilter {
optional bytes stopRowKey = 1; optional bytes stop_row_key = 1;
} }
message KeyOnlyFilter { message KeyOnlyFilter {
required bool lenAsVal = 1; required bool len_as_val = 1;
} }
message MultipleColumnPrefixFilter { message MultipleColumnPrefixFilter {
repeated bytes sortedPrefixes = 1; repeated bytes sorted_prefixes = 1;
} }
message PageFilter { message PageFilter {
required int64 pageSize = 1; required int64 page_size = 1;
} }
message PrefixFilter { message PrefixFilter {
@ -115,7 +115,7 @@ message PrefixFilter {
} }
message QualifierFilter { message QualifierFilter {
required CompareFilter compareFilter = 1; required CompareFilter compare_filter = 1;
} }
message RandomRowFilter { message RandomRowFilter {
@ -123,20 +123,20 @@ message RandomRowFilter {
} }
message RowFilter { message RowFilter {
required CompareFilter compareFilter = 1; required CompareFilter compare_filter = 1;
} }
message SingleColumnValueExcludeFilter { message SingleColumnValueExcludeFilter {
required SingleColumnValueFilter singleColumnValueFilter = 1; required SingleColumnValueFilter single_column_value_filter = 1;
} }
message SingleColumnValueFilter { message SingleColumnValueFilter {
optional bytes columnFamily = 1; optional bytes column_family = 1;
optional bytes columnQualifier = 2; optional bytes column_qualifier = 2;
required CompareType compareOp = 3; required CompareType compare_op = 3;
required Comparator comparator = 4; required Comparator comparator = 4;
optional bool filterIfMissing = 5; optional bool filter_if_missing = 5;
optional bool latestVersionOnly = 6; optional bool latest_version_only = 6;
} }
message SkipFilter { message SkipFilter {
@ -148,7 +148,7 @@ message TimestampsFilter {
} }
message ValueFilter { message ValueFilter {
required CompareFilter compareFilter = 1; required CompareFilter compare_filter = 1;
} }
message WhileMatchFilter { message WhileMatchFilter {

View File

@ -25,21 +25,21 @@ import "hbase.proto";
// Map of name/values // Map of name/values
message FileInfoProto { message FileInfoProto {
repeated BytesBytesPair mapEntry = 1; repeated BytesBytesPair map_entry = 1;
} }
// HFile file trailer // HFile file trailer
message FileTrailerProto { message FileTrailerProto {
optional uint64 fileInfoOffset = 1; optional uint64 file_info_offset = 1;
optional uint64 loadOnOpenDataOffset = 2; optional uint64 load_on_open_data_offset = 2;
optional uint64 uncompressedDataIndexSize = 3; optional uint64 uncompressed_data_index_size = 3;
optional uint64 totalUncompressedBytes = 4; optional uint64 total_uncompressed_bytes = 4;
optional uint32 dataIndexCount = 5; optional uint32 data_index_count = 5;
optional uint32 metaIndexCount = 6; optional uint32 meta_index_count = 6;
optional uint64 entryCount = 7; optional uint64 entry_count = 7;
optional uint32 numDataIndexLevels = 8; optional uint32 num_data_index_levels = 8;
optional uint64 firstDataBlockOffset = 9; optional uint64 first_data_block_offset = 9;
optional uint64 lastDataBlockOffset = 10; optional uint64 last_data_block_offset = 10;
optional string comparatorClassName = 11; optional string comparator_class_name = 11;
optional uint32 compressionCodec = 12; optional uint32 compression_codec = 12;
} }

View File

@ -24,5 +24,5 @@ option java_generate_equals_and_hash = true;
option optimize_for = SPEED; option optimize_for = SPEED;
message LoadBalancerState { message LoadBalancerState {
optional bool balancerOn = 1; optional bool balancer_on = 1;
} }

View File

@ -28,11 +28,11 @@ message IsMasterRunningRequest {
} }
message IsMasterRunningResponse { message IsMasterRunningResponse {
required bool isMasterRunning = 1; required bool is_master_running = 1;
} }
service MasterService { service MasterService {
/** return true if master is available */ /** return true if master is available */
rpc isMasterRunning(IsMasterRunningRequest) rpc IsMasterRunning(IsMasterRunningRequest)
returns(IsMasterRunningResponse); returns(IsMasterRunningResponse);
} }

View File

@ -32,24 +32,24 @@ import "Client.proto";
/* Column-level protobufs */ /* Column-level protobufs */
message AddColumnRequest { message AddColumnRequest {
required bytes tableName = 1; required bytes table_name = 1;
required ColumnFamilySchema columnFamilies = 2; required ColumnFamilySchema column_families = 2;
} }
message AddColumnResponse { message AddColumnResponse {
} }
message DeleteColumnRequest { message DeleteColumnRequest {
required bytes tableName = 1; required bytes table_name = 1;
required bytes columnName = 2; required bytes column_name = 2;
} }
message DeleteColumnResponse { message DeleteColumnResponse {
} }
message ModifyColumnRequest { message ModifyColumnRequest {
required bytes tableName = 1; required bytes table_name = 1;
required ColumnFamilySchema columnFamilies = 2; required ColumnFamilySchema column_families = 2;
} }
message ModifyColumnResponse { message ModifyColumnResponse {
@ -59,7 +59,7 @@ message ModifyColumnResponse {
message MoveRegionRequest { message MoveRegionRequest {
required RegionSpecifier region = 1; required RegionSpecifier region = 1;
optional ServerName destServerName = 2; optional ServerName dest_server_name = 2;
} }
message MoveRegionResponse { message MoveRegionResponse {
@ -69,8 +69,8 @@ message MoveRegionResponse {
* Dispatch merging the specified regions. * Dispatch merging the specified regions.
*/ */
message DispatchMergingRegionsRequest { message DispatchMergingRegionsRequest {
required RegionSpecifier regionA = 1; required RegionSpecifier region_a = 1;
required RegionSpecifier regionB = 2; required RegionSpecifier region_b = 2;
optional bool forcible = 3 [default = false]; optional bool forcible = 3 [default = false];
} }
@ -102,37 +102,37 @@ message OfflineRegionResponse {
/* Table-level protobufs */ /* Table-level protobufs */
message CreateTableRequest { message CreateTableRequest {
required TableSchema tableSchema = 1; required TableSchema table_schema = 1;
repeated bytes splitKeys = 2; repeated bytes split_keys = 2;
} }
message CreateTableResponse { message CreateTableResponse {
} }
message DeleteTableRequest { message DeleteTableRequest {
required bytes tableName = 1; required bytes table_name = 1;
} }
message DeleteTableResponse { message DeleteTableResponse {
} }
message EnableTableRequest { message EnableTableRequest {
required bytes tableName = 1; required bytes table_name = 1;
} }
message EnableTableResponse { message EnableTableResponse {
} }
message DisableTableRequest { message DisableTableRequest {
required bytes tableName = 1; required bytes table_name = 1;
} }
message DisableTableResponse { message DisableTableResponse {
} }
message ModifyTableRequest { message ModifyTableRequest {
required bytes tableName = 1; required bytes table_name = 1;
required TableSchema tableSchema = 2; required TableSchema table_schema = 2;
} }
message ModifyTableResponse { message ModifyTableResponse {
@ -157,7 +157,7 @@ message BalanceRequest {
} }
message BalanceResponse { message BalanceResponse {
required bool balancerRan = 1; required bool balancer_ran = 1;
} }
message SetBalancerRunningRequest { message SetBalancerRunningRequest {
@ -166,14 +166,14 @@ message SetBalancerRunningRequest {
} }
message SetBalancerRunningResponse { message SetBalancerRunningResponse {
optional bool prevBalanceValue = 1; optional bool prev_balance_value = 1;
} }
message CatalogScanRequest { message CatalogScanRequest {
} }
message CatalogScanResponse { message CatalogScanResponse {
optional int32 scanResult = 1; optional int32 scan_result = 1;
} }
message EnableCatalogJanitorRequest { message EnableCatalogJanitorRequest {
@ -181,7 +181,7 @@ message EnableCatalogJanitorRequest {
} }
message EnableCatalogJanitorResponse { message EnableCatalogJanitorResponse {
optional bool prevValue = 1; optional bool prev_value = 1;
} }
message IsCatalogJanitorEnabledRequest { message IsCatalogJanitorEnabledRequest {
@ -196,7 +196,7 @@ message TakeSnapshotRequest{
} }
message TakeSnapshotResponse{ message TakeSnapshotResponse{
required int64 expectedTimeout = 1; required int64 expected_timeout = 1;
} }
message ListSnapshotRequest{ message ListSnapshotRequest{
@ -242,36 +242,36 @@ message IsRestoreSnapshotDoneResponse {
service MasterAdminService { service MasterAdminService {
/** Adds a column to the specified table. */ /** Adds a column to the specified table. */
rpc addColumn(AddColumnRequest) rpc AddColumn(AddColumnRequest)
returns(AddColumnResponse); returns(AddColumnResponse);
/** Deletes a column from the specified table. Table must be disabled. */ /** Deletes a column from the specified table. Table must be disabled. */
rpc deleteColumn(DeleteColumnRequest) rpc DeleteColumn(DeleteColumnRequest)
returns(DeleteColumnResponse); returns(DeleteColumnResponse);
/** Modifies an existing column on the specified table. */ /** Modifies an existing column on the specified table. */
rpc modifyColumn(ModifyColumnRequest) rpc ModifyColumn(ModifyColumnRequest)
returns(ModifyColumnResponse); returns(ModifyColumnResponse);
/** Move the region region to the destination server. */ /** Move the region region to the destination server. */
rpc moveRegion(MoveRegionRequest) rpc MoveRegion(MoveRegionRequest)
returns(MoveRegionResponse); returns(MoveRegionResponse);
/** Master dispatch merging the regions */ /** Master dispatch merging the regions */
rpc dispatchMergingRegions(DispatchMergingRegionsRequest) rpc DispatchMergingRegions(DispatchMergingRegionsRequest)
returns(DispatchMergingRegionsResponse); returns(DispatchMergingRegionsResponse);
/** Assign a region to a server chosen at random. */ /** Assign a region to a server chosen at random. */
rpc assignRegion(AssignRegionRequest) rpc AssignRegion(AssignRegionRequest)
returns(AssignRegionResponse); returns(AssignRegionResponse);
/** /**
* Unassign a region from current hosting regionserver. Region will then be * Unassign a region from current hosting regionserver. Region will then be
* assigned to a regionserver chosen at random. Region could be reassigned * assigned to a regionserver chosen at random. Region could be reassigned
* back to the same server. Use moveRegion if you want * back to the same server. Use MoveRegion if you want
* to control the region movement. * to control the region movement.
*/ */
rpc unassignRegion(UnassignRegionRequest) rpc UnassignRegion(UnassignRegionRequest)
returns(UnassignRegionResponse); returns(UnassignRegionResponse);
/** /**
@ -280,35 +280,35 @@ service MasterAdminService {
* automatically reassign the region as in unassign. This is a special * automatically reassign the region as in unassign. This is a special
* method, and should only be used by experts or hbck. * method, and should only be used by experts or hbck.
*/ */
rpc offlineRegion(OfflineRegionRequest) rpc OfflineRegion(OfflineRegionRequest)
returns(OfflineRegionResponse); returns(OfflineRegionResponse);
/** Deletes a table */ /** Deletes a table */
rpc deleteTable(DeleteTableRequest) rpc DeleteTable(DeleteTableRequest)
returns(DeleteTableResponse); returns(DeleteTableResponse);
/** Puts the table on-line (only needed if table has been previously taken offline) */ /** Puts the table on-line (only needed if table has been previously taken offline) */
rpc enableTable(EnableTableRequest) rpc EnableTable(EnableTableRequest)
returns(EnableTableResponse); returns(EnableTableResponse);
/** Take table offline */ /** Take table offline */
rpc disableTable(DisableTableRequest) rpc DisableTable(DisableTableRequest)
returns(DisableTableResponse); returns(DisableTableResponse);
/** Modify a table's metadata */ /** Modify a table's metadata */
rpc modifyTable(ModifyTableRequest) rpc ModifyTable(ModifyTableRequest)
returns(ModifyTableResponse); returns(ModifyTableResponse);
/** Creates a new table asynchronously */ /** Creates a new table asynchronously */
rpc createTable(CreateTableRequest) rpc CreateTable(CreateTableRequest)
returns(CreateTableResponse); returns(CreateTableResponse);
/** Shutdown an HBase cluster. */ /** Shutdown an HBase cluster. */
rpc shutdown(ShutdownRequest) rpc Shutdown(ShutdownRequest)
returns(ShutdownResponse); returns(ShutdownResponse);
/** Stop HBase Master only. Does not shutdown the cluster. */ /** Stop HBase Master only. Does not shutdown the cluster. */
rpc stopMaster(StopMasterRequest) rpc StopMaster(StopMasterRequest)
returns(StopMasterResponse); returns(StopMasterResponse);
/** /**
@ -316,72 +316,72 @@ service MasterAdminService {
* go ahead and do the reassignments. Can NOT run for various reasons. * go ahead and do the reassignments. Can NOT run for various reasons.
* Check logs. * Check logs.
*/ */
rpc balance(BalanceRequest) rpc Balance(BalanceRequest)
returns(BalanceResponse); returns(BalanceResponse);
/** /**
* Turn the load balancer on or off. * Turn the load balancer on or off.
* If synchronous is true, it waits until current balance() call, if outstanding, to return. * If synchronous is true, it waits until current balance() call, if outstanding, to return.
*/ */
rpc setBalancerRunning(SetBalancerRunningRequest) rpc SetBalancerRunning(SetBalancerRunningRequest)
returns(SetBalancerRunningResponse); returns(SetBalancerRunningResponse);
/** Get a run of the catalog janitor */ /** Get a run of the catalog janitor */
rpc runCatalogScan(CatalogScanRequest) rpc RunCatalogScan(CatalogScanRequest)
returns(CatalogScanResponse); returns(CatalogScanResponse);
/** /**
* Enable the catalog janitor on or off. * Enable the catalog janitor on or off.
*/ */
rpc enableCatalogJanitor(EnableCatalogJanitorRequest) rpc EnableCatalogJanitor(EnableCatalogJanitorRequest)
returns(EnableCatalogJanitorResponse); returns(EnableCatalogJanitorResponse);
/** /**
* Query whether the catalog janitor is enabled. * Query whether the catalog janitor is enabled.
*/ */
rpc isCatalogJanitorEnabled(IsCatalogJanitorEnabledRequest) rpc IsCatalogJanitorEnabled(IsCatalogJanitorEnabledRequest)
returns(IsCatalogJanitorEnabledResponse); returns(IsCatalogJanitorEnabledResponse);
/** /**
* Call a master coprocessor endpoint * Call a master coprocessor endpoint
*/ */
rpc execMasterService(CoprocessorServiceRequest) rpc ExecMasterService(CoprocessorServiceRequest)
returns(CoprocessorServiceResponse); returns(CoprocessorServiceResponse);
/** /**
* Create a snapshot for the given table. * Create a snapshot for the given table.
* @param snapshot description of the snapshot to take * @param snapshot description of the snapshot to take
*/ */
rpc snapshot(TakeSnapshotRequest) returns(TakeSnapshotResponse); rpc Snapshot(TakeSnapshotRequest) returns(TakeSnapshotResponse);
/** /**
* List completed snapshots. * List completed snapshots.
* Returns a list of snapshot descriptors for completed snapshots * Returns a list of snapshot descriptors for completed snapshots
*/ */
rpc getCompletedSnapshots(ListSnapshotRequest) returns(ListSnapshotResponse); rpc GetCompletedSnapshots(ListSnapshotRequest) returns(ListSnapshotResponse);
/** /**
* Delete an existing snapshot. This method can also be used to clean up an aborted snapshot. * Delete an existing snapshot. This method can also be used to clean up an aborted snapshot.
* @param snapshotName snapshot to delete * @param snapshotName snapshot to delete
*/ */
rpc deleteSnapshot(DeleteSnapshotRequest) returns(DeleteSnapshotResponse); rpc DeleteSnapshot(DeleteSnapshotRequest) returns(DeleteSnapshotResponse);
/** /**
* Determine if the snapshot is done yet. * Determine if the snapshot is done yet.
*/ */
rpc isSnapshotDone(IsSnapshotDoneRequest) returns(IsSnapshotDoneResponse); rpc IsSnapshotDone(IsSnapshotDoneRequest) returns(IsSnapshotDoneResponse);
/** /**
* Restore a snapshot * Restore a snapshot
* @param snapshot description of the snapshot to restore * @param snapshot description of the snapshot to restore
*/ */
rpc restoreSnapshot(RestoreSnapshotRequest) returns(RestoreSnapshotResponse); rpc RestoreSnapshot(RestoreSnapshotRequest) returns(RestoreSnapshotResponse);
/** /**
* Determine if the snapshot restore is done yet. * Determine if the snapshot restore is done yet.
*/ */
rpc isRestoreSnapshotDone(IsRestoreSnapshotDoneRequest) returns(IsRestoreSnapshotDoneResponse); rpc IsRestoreSnapshotDone(IsRestoreSnapshotDoneRequest) returns(IsRestoreSnapshotDoneResponse);
/** return true if master is available */ /** return true if master is available */
rpc isMasterRunning(IsMasterRunningRequest) returns(IsMasterRunningResponse); rpc IsMasterRunning(IsMasterRunningRequest) returns(IsMasterRunningResponse);
} }

View File

@ -29,42 +29,42 @@ import "hbase.proto";
import "ClusterStatus.proto"; import "ClusterStatus.proto";
message GetSchemaAlterStatusRequest { message GetSchemaAlterStatusRequest {
required bytes tableName = 1; required bytes table_name = 1;
} }
message GetSchemaAlterStatusResponse { message GetSchemaAlterStatusResponse {
optional uint32 yetToUpdateRegions = 1; optional uint32 yet_to_update_regions = 1;
optional uint32 totalRegions = 2; optional uint32 total_regions = 2;
} }
message GetTableDescriptorsRequest { message GetTableDescriptorsRequest {
repeated string tableNames = 1; repeated string table_names = 1;
} }
message GetTableDescriptorsResponse { message GetTableDescriptorsResponse {
repeated TableSchema tableSchema = 1; repeated TableSchema table_schema = 1;
} }
message GetClusterStatusRequest { message GetClusterStatusRequest {
} }
message GetClusterStatusResponse { message GetClusterStatusResponse {
required ClusterStatus clusterStatus = 1; required ClusterStatus cluster_status = 1;
} }
service MasterMonitorService { service MasterMonitorService {
/** Used by the client to get the number of regions that have received the updated schema */ /** Used by the client to get the number of regions that have received the updated schema */
rpc getSchemaAlterStatus(GetSchemaAlterStatusRequest) rpc GetSchemaAlterStatus(GetSchemaAlterStatusRequest)
returns(GetSchemaAlterStatusResponse); returns(GetSchemaAlterStatusResponse);
/** Get list of TableDescriptors for requested tables. */ /** Get list of TableDescriptors for requested tables. */
rpc getTableDescriptors(GetTableDescriptorsRequest) rpc GetTableDescriptors(GetTableDescriptorsRequest)
returns(GetTableDescriptorsResponse); returns(GetTableDescriptorsResponse);
/** Return cluster status. */ /** Return cluster status. */
rpc getClusterStatus(GetClusterStatusRequest) rpc GetClusterStatus(GetClusterStatusRequest)
returns(GetClusterStatusResponse); returns(GetClusterStatusResponse);
/** return true if master is available */ /** return true if master is available */
rpc isMasterRunning(IsMasterRunningRequest) returns(IsMasterRunningResponse); rpc IsMasterRunning(IsMasterRunningRequest) returns(IsMasterRunningResponse);
} }

View File

@ -23,13 +23,13 @@ option java_generic_services = true;
option optimize_for = SPEED; option optimize_for = SPEED;
message MultiMutateRequest { message MultiMutateRequest {
repeated MutationProto mutationRequest = 1; repeated MutationProto mutation_request = 1;
} }
message MultiMutateResponse { message MultiMutateResponse {
} }
service MultiRowMutationService { service MultiRowMutationService {
rpc mutateRows(MultiMutateRequest) rpc MutateRows(MultiMutateRequest)
returns(MultiMutateResponse); returns(MultiMutateResponse);
} }

View File

@ -72,20 +72,20 @@ option optimize_for = SPEED;
// User Information proto. Included in ConnectionHeader on connection setup // User Information proto. Included in ConnectionHeader on connection setup
message UserInformation { message UserInformation {
required string effectiveUser = 1; required string effective_user = 1;
optional string realUser = 2; optional string real_user = 2;
} }
// This is sent on connection setup after the connection preamble is sent. // This is sent on connection setup after the connection preamble is sent.
message ConnectionHeader { message ConnectionHeader {
optional UserInformation userInfo = 1; optional UserInformation user_info = 1;
optional string serviceName = 2; optional string service_name = 2;
// Cell block codec we will use sending over optional cell blocks. Server throws exception // Cell block codec we will use sending over optional cell blocks. Server throws exception
// if cannot deal. // if cannot deal.
optional string cellBlockCodecClass = 3 [default = "org.apache.hadoop.hbase.codec.KeyValueCodec"]; optional string cell_block_codec_class = 3 [default = "org.apache.hadoop.hbase.codec.KeyValueCodec"];
// Compressor we will use if cell block is compressed. Server will throw exception if not supported. // Compressor we will use if cell block is compressed. Server will throw exception if not supported.
// Class must implement hadoop's CompressionCodec Interface // Class must implement hadoop's CompressionCodec Interface
optional string cellBlockCompressorClass = 4; optional string cell_block_compressor_class = 4;
} }
// Optional Cell block Message. Included in client RequestHeader // Optional Cell block Message. Included in client RequestHeader
@ -98,34 +98,34 @@ message CellBlockMeta {
// the server side exception to the RPC client. // the server side exception to the RPC client.
message ExceptionResponse { message ExceptionResponse {
// Class name of the exception thrown from the server // Class name of the exception thrown from the server
optional string exceptionClassName = 1; optional string exception_class_name = 1;
// Exception stack trace from the server side // Exception stack trace from the server side
optional string stackTrace = 2; optional string stack_trace = 2;
// Optional hostname. Filled in for some exceptions such as region moved // Optional hostname. Filled in for some exceptions such as region moved
// where exception gives clue on where the region may have moved. // where exception gives clue on where the region may have moved.
optional string hostname = 3; optional string hostname = 3;
optional int32 port = 4; optional int32 port = 4;
// Set if we are NOT to retry on receipt of this exception // Set if we are NOT to retry on receipt of this exception
optional bool doNotRetry = 5; optional bool do_not_retry = 5;
} }
// Header sent making a request. // Header sent making a request.
message RequestHeader { message RequestHeader {
// Monotonically increasing callId to keep track of RPC requests and their response // Monotonically increasing call_id to keep track of RPC requests and their response
optional uint32 callId = 1; optional uint32 call_id = 1;
optional RPCTInfo traceInfo = 2; optional RPCTInfo trace_info = 2;
optional string methodName = 3; optional string method_name = 3;
// If true, then a pb Message param follows. // If true, then a pb Message param follows.
optional bool requestParam = 4; optional bool request_param = 4;
// If present, then an encoded data block follows. // If present, then an encoded data block follows.
optional CellBlockMeta cellBlockMeta = 5; optional CellBlockMeta cell_block_meta = 5;
// TODO: Have client specify priority // TODO: Have client specify priority
} }
message ResponseHeader { message ResponseHeader {
optional uint32 callId = 1; optional uint32 call_id = 1;
// If present, then request threw an exception and no response message (else we presume one) // If present, then request threw an exception and no response message (else we presume one)
optional ExceptionResponse exception = 2; optional ExceptionResponse exception = 2;
// If present, then an encoded data block follows. // If present, then an encoded data block follows.
optional CellBlockMeta cellBlockMeta = 3; optional CellBlockMeta cell_block_meta = 3;
} }

View File

@ -32,10 +32,10 @@ message RegionServerStartupRequest {
required uint32 port = 1; required uint32 port = 1;
/** This servers' startcode */ /** This servers' startcode */
required uint64 serverStartCode = 2; required uint64 server_start_code = 2;
/** Current time of the region server in ms */ /** Current time of the region server in ms */
required uint64 serverCurrentTime = 3; required uint64 server_current_time = 3;
} }
message RegionServerStartupResponse { message RegionServerStartupResponse {
@ -44,7 +44,7 @@ message RegionServerStartupResponse {
* hbase rootdir, the hostname to use creating the RegionServer ServerName, * hbase rootdir, the hostname to use creating the RegionServer ServerName,
* etc * etc
*/ */
repeated NameStringPair mapEntries = 1; repeated NameStringPair map_entries = 1;
} }
message RegionServerReportRequest { message RegionServerReportRequest {
@ -62,7 +62,7 @@ message ReportRSFatalErrorRequest {
required ServerName server = 1; required ServerName server = 1;
/** informative text to expose in the master logs and UI */ /** informative text to expose in the master logs and UI */
required string errorMessage = 2; required string error_message = 2;
} }
message ReportRSFatalErrorResponse { message ReportRSFatalErrorResponse {
@ -70,33 +70,33 @@ message ReportRSFatalErrorResponse {
message GetLastFlushedSequenceIdRequest { message GetLastFlushedSequenceIdRequest {
/** region name */ /** region name */
required bytes regionName = 1; required bytes region_name = 1;
} }
message GetLastFlushedSequenceIdResponse { message GetLastFlushedSequenceIdResponse {
/** the last HLog sequence id flushed from MemStore to HFile for the region */ /** the last HLog sequence id flushed from MemStore to HFile for the region */
required uint64 lastFlushedSequenceId = 1; required uint64 last_flushed_sequence_id = 1;
} }
service RegionServerStatusService { service RegionServerStatusService {
/** Called when a region server first starts. */ /** Called when a region server first starts. */
rpc regionServerStartup(RegionServerStartupRequest) rpc RegionServerStartup(RegionServerStartupRequest)
returns(RegionServerStartupResponse); returns(RegionServerStartupResponse);
/** Called to report the load the RegionServer is under. */ /** Called to report the load the RegionServer is under. */
rpc regionServerReport(RegionServerReportRequest) rpc RegionServerReport(RegionServerReportRequest)
returns(RegionServerReportResponse); returns(RegionServerReportResponse);
/** /**
* Called by a region server to report a fatal error that is causing it to * Called by a region server to report a fatal error that is causing it to
* abort. * abort.
*/ */
rpc reportRSFatalError(ReportRSFatalErrorRequest) rpc ReportRSFatalError(ReportRSFatalErrorRequest)
returns(ReportRSFatalErrorResponse); returns(ReportRSFatalErrorResponse);
/** Called to get the sequence id of the last MemStore entry flushed to an /** Called to get the sequence id of the last MemStore entry flushed to an
* HFile for a specified region. Used by the region server to speed up * HFile for a specified region. Used by the region server to speed up
* log splitting. */ * log splitting. */
rpc getLastFlushedSequenceId(GetLastFlushedSequenceIdRequest) rpc GetLastFlushedSequenceId(GetLastFlushedSequenceIdRequest)
returns(GetLastFlushedSequenceIdResponse); returns(GetLastFlushedSequenceIdResponse);
} }

View File

@ -27,15 +27,15 @@ option java_generate_equals_and_hash = true;
option optimize_for = SPEED; option optimize_for = SPEED;
message RowProcessorRequest { message RowProcessorRequest {
required string rowProcessorClassName = 1; required string row_processor_class_name = 1;
optional string rowProcessorInitializerMessageName = 2; optional string row_processor_initializer_message_name = 2;
optional bytes rowProcessorInitializerMessage = 3; optional bytes row_processor_initializer_message = 3;
} }
message RowProcessorResult { message RowProcessorResult {
required bytes rowProcessorResult = 1; required bytes row_processor_result = 1;
} }
service RowProcessorService { service RowProcessorService {
rpc process (RowProcessorRequest) returns (RowProcessorResult); rpc Process (RowProcessorRequest) returns (RowProcessorResult);
} }

View File

@ -26,10 +26,10 @@ import 'hbase.proto';
import 'Client.proto'; import 'Client.proto';
message SecureBulkLoadHFilesRequest { message SecureBulkLoadHFilesRequest {
repeated BulkLoadHFileRequest.FamilyPath familyPath = 1; repeated BulkLoadHFileRequest.FamilyPath family_path = 1;
optional bool assignSeqNum = 2; optional bool assign_seq_num = 2;
required DelegationTokenProto fsToken = 3; required DelegationTokenProto fs_token = 3;
required string bulkToken = 4; required string bulk_token = 4;
} }
message SecureBulkLoadHFilesResponse { message SecureBulkLoadHFilesResponse {
@ -44,15 +44,15 @@ message DelegationTokenProto {
} }
message PrepareBulkLoadRequest { message PrepareBulkLoadRequest {
required bytes tableName = 1; required bytes table_name = 1;
} }
message PrepareBulkLoadResponse { message PrepareBulkLoadResponse {
required string bulkToken = 1; required string bulk_token = 1;
} }
message CleanupBulkLoadRequest { message CleanupBulkLoadRequest {
required string bulkToken = 1; required string bulk_token = 1;
} }
@ -60,12 +60,12 @@ message CleanupBulkLoadResponse {
} }
service SecureBulkLoadService { service SecureBulkLoadService {
rpc prepareBulkLoad(PrepareBulkLoadRequest) rpc PrepareBulkLoad(PrepareBulkLoadRequest)
returns (PrepareBulkLoadResponse); returns (PrepareBulkLoadResponse);
rpc secureBulkLoadHFiles(SecureBulkLoadHFilesRequest) rpc SecureBulkLoadHFiles(SecureBulkLoadHFilesRequest)
returns (SecureBulkLoadHFilesResponse); returns (SecureBulkLoadHFilesResponse);
rpc cleanupBulkLoad(CleanupBulkLoadRequest) rpc CleanupBulkLoad(CleanupBulkLoadRequest)
returns (CleanupBulkLoadResponse); returns (CleanupBulkLoadResponse);
} }

View File

@ -26,6 +26,6 @@ option optimize_for = SPEED;
//the id of the current span when this message was sent, so we know //the id of the current span when this message was sent, so we know
//what span caused the new span we will create when this message is received. //what span caused the new span we will create when this message is received.
message RPCTInfo { message RPCTInfo {
optional int64 traceId = 1; optional int64 trace_id = 1;
optional int64 parentId = 2; optional int64 parent_id = 2;
} }

View File

@ -24,21 +24,21 @@ option optimize_for = SPEED;
import "hbase.proto"; import "hbase.proto";
message WALHeader { message WALHeader {
optional bool hasCompression = 1; optional bool has_compression = 1;
} }
// Protocol buffer version of HLogKey; see HLogKey comment, not really a key but WALEdit header for some KVs // Protocol buffer version of HLogKey; see HLogKey comment, not really a key but WALEdit header for some KVs
message WALKey { message WALKey {
required bytes encodedRegionName = 1; required bytes encoded_region_name = 1;
required bytes tableName = 2; required bytes table_name = 2;
required uint64 logSequenceNumber = 3; required uint64 log_sequence_number = 3;
required uint64 writeTime = 4; required uint64 write_time = 4;
optional UUID clusterId = 5; optional UUID cluster_id = 5;
repeated FamilyScope scopes = 6; repeated FamilyScope scopes = 6;
optional uint32 followingKvCount = 7; optional uint32 following_kv_count = 7;
/* /*
optional CustomEntryType customEntryType = 8; optional CustomEntryType custom_entry_type = 8;
enum CustomEntryType { enum CustomEntryType {
COMPACTION = 0; COMPACTION = 0;
@ -53,7 +53,7 @@ enum ScopeType {
message FamilyScope { message FamilyScope {
required bytes family = 1; required bytes family = 1;
required ScopeType scopeType = 2; required ScopeType scope_type = 2;
} }
/** /**
@ -67,12 +67,12 @@ message FamilyScope {
* the * compaction should we fail the WAL write. * the * compaction should we fail the WAL write.
*/ */
message CompactionDescriptor { message CompactionDescriptor {
required bytes tableName = 1; // TODO: WALKey already stores these, might remove required bytes table_name = 1; // TODO: WALKey already stores these, might remove
required bytes encodedRegionName = 2; required bytes encoded_region_name = 2;
required bytes familyName = 3; required bytes family_name = 3;
repeated string compactionInput = 4; repeated string compaction_input = 4;
repeated string compactionOutput = 5; repeated string compaction_output = 5;
required string storeHomeDir = 6; required string store_home_dir = 6;
} }
/** /**

View File

@ -36,7 +36,7 @@ message MetaRegionServer {
// The major version of the rpc the server speaks. This is used so that // The major version of the rpc the server speaks. This is used so that
// clients connecting to the cluster can have prior knowledge of what version // clients connecting to the cluster can have prior knowledge of what version
// to send to a RegionServer. AsyncHBase will use this to detect versions. // to send to a RegionServer. AsyncHBase will use this to detect versions.
optional uint32 rpcVersion = 2; optional uint32 rpc_version = 2;
} }
/** /**
@ -46,7 +46,7 @@ message Master {
// The ServerName of the current Master // The ServerName of the current Master
required ServerName master = 1; required ServerName master = 1;
// Major RPC version so that clients can know what version the master can accept. // Major RPC version so that clients can know what version the master can accept.
optional uint32 rpcVersion = 2; optional uint32 rpc_version = 2;
} }
/** /**
@ -54,8 +54,8 @@ message Master {
*/ */
message ClusterUp { message ClusterUp {
// If this znode is present, cluster is up. Currently // If this znode is present, cluster is up. Currently
// the data is cluster startDate. // the data is cluster start_date.
required string startDate = 1; required string start_date = 1;
} }
/** /**
@ -64,12 +64,12 @@ message ClusterUp {
*/ */
message RegionTransition { message RegionTransition {
// Code for EventType gotten by doing o.a.h.h.EventHandler.EventType.getCode() // Code for EventType gotten by doing o.a.h.h.EventHandler.EventType.getCode()
required uint32 eventTypeCode = 1; required uint32 event_type_code = 1;
// Full regionname in bytes // Full regionname in bytes
required bytes regionName = 2; required bytes region_name = 2;
required uint64 createTime = 3; required uint64 create_time = 3;
// The region server where the transition will happen or is happening // The region server where the transition will happen or is happening
required ServerName serverName = 4; required ServerName server_name = 4;
optional bytes payload = 5; optional bytes payload = 5;
} }
@ -86,7 +86,7 @@ message SplitLogTask {
ERR = 4; ERR = 4;
} }
required State state = 1; required State state = 1;
required ServerName serverName = 2; required ServerName server_name = 2;
} }
/** /**
@ -110,7 +110,7 @@ message Table {
* Used by replication. Holds a replication peer key. * Used by replication. Holds a replication peer key.
*/ */
message ReplicationPeer { message ReplicationPeer {
// clusterKey is the concatenation of the slave cluster's // clusterkey is the concatenation of the slave cluster's
// hbase.zookeeper.quorum:hbase.zookeeper.property.clientPort:zookeeper.znode.parent // hbase.zookeeper.quorum:hbase.zookeeper.property.clientPort:zookeeper.znode.parent
required string clusterkey = 1; required string clusterkey = 1;
} }
@ -137,27 +137,27 @@ message ReplicationHLogPosition {
* Used by replication. Used to lock a region server during failover. * Used by replication. Used to lock a region server during failover.
*/ */
message ReplicationLock { message ReplicationLock {
required string lockOwner = 1; required string lock_owner = 1;
} }
/** /**
* Metadata associated with a table lock in zookeeper * Metadata associated with a table lock in zookeeper
*/ */
message TableLock { message TableLock {
optional bytes tableName = 1; optional bytes table_name = 1;
optional ServerName lockOwner = 2; optional ServerName lock_owner = 2;
optional int64 threadId = 3; optional int64 thread_id = 3;
optional bool isShared = 4; optional bool is_shared = 4;
optional string purpose = 5; optional string purpose = 5;
optional int64 createTime = 6; optional int64 create_time = 6;
} }
/** /**
* sequence Id of a store * sequence Id of a store
*/ */
message StoreSequenceId { message StoreSequenceId {
required bytes familyName = 1; required bytes family_name = 1;
required uint64 sequenceId = 2; required uint64 sequence_id = 2;
} }
/** /**
@ -165,6 +165,6 @@ message StoreSequenceId {
* list sequence ids of the region's stores * list sequence ids of the region's stores
*/ */
message RegionStoreSequenceIds { message RegionStoreSequenceIds {
required uint64 lastFlushedSequenceId = 1; required uint64 last_flushed_sequence_id = 1;
repeated StoreSequenceId storeSequenceId = 2; repeated StoreSequenceId store_sequence_id = 2;
} }

View File

@ -32,7 +32,7 @@ import "Cell.proto";
message TableSchema { message TableSchema {
optional bytes name = 1; optional bytes name = 1;
repeated BytesBytesPair attributes = 2; repeated BytesBytesPair attributes = 2;
repeated ColumnFamilySchema columnFamilies = 3; repeated ColumnFamilySchema column_families = 3;
repeated NameStringPair configuration = 4; repeated NameStringPair configuration = 4;
} }
@ -50,10 +50,10 @@ message ColumnFamilySchema {
* Protocol buffer version of HRegionInfo. * Protocol buffer version of HRegionInfo.
*/ */
message RegionInfo { message RegionInfo {
required uint64 regionId = 1; required uint64 region_id = 1;
required bytes tableName = 2; required bytes table_name = 2;
optional bytes startKey = 3; optional bytes start_key = 3;
optional bytes endKey = 4; optional bytes end_key = 4;
optional bool offline = 5; optional bool offline = 5;
optional bool split = 6; optional bool split = 6;
optional bool recovering = 7; optional bool recovering = 7;
@ -63,7 +63,7 @@ message RegionInfo {
* Protocol buffer for favored nodes * Protocol buffer for favored nodes
*/ */
message FavoredNodes { message FavoredNodes {
repeated ServerName favoredNode = 1; repeated ServerName favored_node = 1;
} }
/** /**
@ -111,9 +111,9 @@ enum CompareType {
* Protocol buffer version of ServerName * Protocol buffer version of ServerName
*/ */
message ServerName { message ServerName {
required string hostName = 1; required string host_name = 1;
optional uint32 port = 2; optional uint32 port = 2;
optional uint64 startCode = 3; optional uint64 start_code = 3;
} }
// Comment data structures // Comment data structures
@ -148,7 +148,7 @@ message NameInt64Pair {
message SnapshotDescription { message SnapshotDescription {
required string name = 1; required string name = 1;
optional string table = 2; // not needed for delete, but checked for in taking snapshot optional string table = 2; // not needed for delete, but checked for in taking snapshot
optional int64 creationTime = 3 [default = 0]; optional int64 creation_time = 3 [default = 0];
enum Type { enum Type {
DISABLED = 0; DISABLED = 0;
FLUSH = 1; FLUSH = 1;
@ -161,15 +161,15 @@ message EmptyMsg {
} }
message LongMsg { message LongMsg {
required int64 longMsg = 1; required int64 long_msg = 1;
} }
message BigDecimalMsg { message BigDecimalMsg {
required bytes bigdecimalMsg = 1; required bytes bigdecimal_msg = 1;
} }
message UUID { message UUID {
required uint64 leastSigBits = 1; required uint64 least_sig_bits = 1;
required uint64 mostSigBits = 2; required uint64 most_sig_bits = 2;
} }

View File

@ -133,7 +133,7 @@ public class TokenProvider implements AuthenticationProtos.AuthenticationService
} }
@Override @Override
public void whoami(RpcController controller, AuthenticationProtos.WhoAmIRequest request, public void whoAmI(RpcController controller, AuthenticationProtos.WhoAmIRequest request,
RpcCallback<AuthenticationProtos.WhoAmIResponse> done) { RpcCallback<AuthenticationProtos.WhoAmIResponse> done) {
User requestUser = RequestContext.getRequestUser(); User requestUser = RequestContext.getRequestUser();
AuthenticationProtos.WhoAmIResponse.Builder response = AuthenticationProtos.WhoAmIResponse.Builder response =

View File

@ -274,15 +274,15 @@ public class TestTokenAuthentication {
} }
@Override @Override
public AuthenticationProtos.WhoAmIResponse whoami( public AuthenticationProtos.WhoAmIResponse whoAmI(
RpcController controller, AuthenticationProtos.WhoAmIRequest request) RpcController controller, AuthenticationProtos.WhoAmIRequest request)
throws ServiceException { throws ServiceException {
LOG.debug("whoami() request from "+RequestContext.getRequestUserName()); LOG.debug("whoAmI() request from "+RequestContext.getRequestUserName());
// ignore passed in controller -- it's always null // ignore passed in controller -- it's always null
ServerRpcController serverController = new ServerRpcController(); ServerRpcController serverController = new ServerRpcController();
BlockingRpcCallback<AuthenticationProtos.WhoAmIResponse> callback = BlockingRpcCallback<AuthenticationProtos.WhoAmIResponse> callback =
new BlockingRpcCallback<AuthenticationProtos.WhoAmIResponse>(); new BlockingRpcCallback<AuthenticationProtos.WhoAmIResponse>();
whoami(serverController, request, callback); whoAmI(serverController, request, callback);
try { try {
serverController.checkFailed(); serverController.checkFailed();
return callback.get(); return callback.get();
@ -383,7 +383,7 @@ public class TestTokenAuthentication {
AuthenticationProtos.AuthenticationService.BlockingInterface stub = AuthenticationProtos.AuthenticationService.BlockingInterface stub =
AuthenticationProtos.AuthenticationService.newBlockingStub(channel); AuthenticationProtos.AuthenticationService.newBlockingStub(channel);
AuthenticationProtos.WhoAmIResponse response = AuthenticationProtos.WhoAmIResponse response =
stub.whoami(null, AuthenticationProtos.WhoAmIRequest.getDefaultInstance()); stub.whoAmI(null, AuthenticationProtos.WhoAmIRequest.getDefaultInstance());
String myname = response.getUsername(); String myname = response.getUsername();
assertEquals("testuser", myname); assertEquals("testuser", myname);
String authMethod = response.getAuthMethod(); String authMethod = response.getAuthMethod();