HBASE-8915 protobuf message style

git-svn-id: https://svn.apache.org/repos/asf/hbase/trunk@1501834 13f79535-47bb-0310-9956-ffa450edef68
This commit is contained in:
Michael Stack 2013-07-10 16:30:32 +00:00
parent 8bf99de619
commit 457c928d35
52 changed files with 2371 additions and 2360 deletions

View File

@ -6422,11 +6422,11 @@ public final class AccessControlProtos {
"Permission\":\n\027CheckPermissionsRequest\022\037\n" +
"\npermission\030\001 \003(\0132\013.Permission\"\032\n\030CheckP" +
"ermissionsResponse2\373\001\n\024AccessControlServ" +
"ice\022&\n\005grant\022\r.GrantRequest\032\016.GrantRespo",
"nse\022)\n\006revoke\022\016.RevokeRequest\032\017.RevokeRe" +
"sponse\022G\n\022getUserPermissions\022\027.UserPermi" +
"ice\022&\n\005Grant\022\r.GrantRequest\032\016.GrantRespo",
"nse\022)\n\006Revoke\022\016.RevokeRequest\032\017.RevokeRe" +
"sponse\022G\n\022GetUserPermissions\022\027.UserPermi" +
"ssionsRequest\032\030.UserPermissionsResponse\022" +
"G\n\020checkPermissions\022\030.CheckPermissionsRe" +
"G\n\020CheckPermissions\022\030.CheckPermissionsRe" +
"quest\032\031.CheckPermissionsResponseBI\n*org." +
"apache.hadoop.hbase.protobuf.generatedB\023" +
"AccessControlProtosH\001\210\001\001\240\001\001"

View File

@ -16,7 +16,7 @@ public final class AdminProtos {
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier getRegion();
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder getRegionOrBuilder();
// optional bool compactionState = 2;
// optional bool compaction_state = 2;
boolean hasCompactionState();
boolean getCompactionState();
}
@ -62,8 +62,8 @@ public final class AdminProtos {
return region_;
}
// optional bool compactionState = 2;
public static final int COMPACTIONSTATE_FIELD_NUMBER = 2;
// optional bool compaction_state = 2;
public static final int COMPACTION_STATE_FIELD_NUMBER = 2;
private boolean compactionState_;
public boolean hasCompactionState() {
return ((bitField0_ & 0x00000002) == 0x00000002);
@ -166,7 +166,7 @@ public final class AdminProtos {
hash = (53 * hash) + getRegion().hashCode();
}
if (hasCompactionState()) {
hash = (37 * hash) + COMPACTIONSTATE_FIELD_NUMBER;
hash = (37 * hash) + COMPACTION_STATE_FIELD_NUMBER;
hash = (53 * hash) + hashBoolean(getCompactionState());
}
hash = (29 * hash) + getUnknownFields().hashCode();
@ -515,7 +515,7 @@ public final class AdminProtos {
return regionBuilder_;
}
// optional bool compactionState = 2;
// optional bool compaction_state = 2;
private boolean compactionState_ ;
public boolean hasCompactionState() {
return ((bitField0_ & 0x00000002) == 0x00000002);
@ -550,12 +550,12 @@ public final class AdminProtos {
public interface GetRegionInfoResponseOrBuilder
extends com.google.protobuf.MessageOrBuilder {
// required .RegionInfo regionInfo = 1;
// required .RegionInfo region_info = 1;
boolean hasRegionInfo();
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfo getRegionInfo();
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfoOrBuilder getRegionInfoOrBuilder();
// optional .GetRegionInfoResponse.CompactionState compactionState = 2;
// optional .GetRegionInfoResponse.CompactionState compaction_state = 2;
boolean hasCompactionState();
org.apache.hadoop.hbase.protobuf.generated.AdminProtos.GetRegionInfoResponse.CompactionState getCompactionState();
}
@ -663,8 +663,8 @@ public final class AdminProtos {
}
private int bitField0_;
// required .RegionInfo regionInfo = 1;
public static final int REGIONINFO_FIELD_NUMBER = 1;
// required .RegionInfo region_info = 1;
public static final int REGION_INFO_FIELD_NUMBER = 1;
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfo regionInfo_;
public boolean hasRegionInfo() {
return ((bitField0_ & 0x00000001) == 0x00000001);
@ -676,8 +676,8 @@ public final class AdminProtos {
return regionInfo_;
}
// optional .GetRegionInfoResponse.CompactionState compactionState = 2;
public static final int COMPACTIONSTATE_FIELD_NUMBER = 2;
// optional .GetRegionInfoResponse.CompactionState compaction_state = 2;
public static final int COMPACTION_STATE_FIELD_NUMBER = 2;
private org.apache.hadoop.hbase.protobuf.generated.AdminProtos.GetRegionInfoResponse.CompactionState compactionState_;
public boolean hasCompactionState() {
return ((bitField0_ & 0x00000002) == 0x00000002);
@ -776,11 +776,11 @@ public final class AdminProtos {
int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode();
if (hasRegionInfo()) {
hash = (37 * hash) + REGIONINFO_FIELD_NUMBER;
hash = (37 * hash) + REGION_INFO_FIELD_NUMBER;
hash = (53 * hash) + getRegionInfo().hashCode();
}
if (hasCompactionState()) {
hash = (37 * hash) + COMPACTIONSTATE_FIELD_NUMBER;
hash = (37 * hash) + COMPACTION_STATE_FIELD_NUMBER;
hash = (53 * hash) + hashEnum(getCompactionState());
}
hash = (29 * hash) + getUnknownFields().hashCode();
@ -1045,7 +1045,7 @@ public final class AdminProtos {
private int bitField0_;
// required .RegionInfo regionInfo = 1;
// required .RegionInfo region_info = 1;
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfo regionInfo_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfo.getDefaultInstance();
private com.google.protobuf.SingleFieldBuilder<
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfo, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfo.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfoOrBuilder> regionInfoBuilder_;
@ -1135,7 +1135,7 @@ public final class AdminProtos {
return regionInfoBuilder_;
}
// optional .GetRegionInfoResponse.CompactionState compactionState = 2;
// optional .GetRegionInfoResponse.CompactionState compaction_state = 2;
private org.apache.hadoop.hbase.protobuf.generated.AdminProtos.GetRegionInfoResponse.CompactionState compactionState_ = org.apache.hadoop.hbase.protobuf.generated.AdminProtos.GetRegionInfoResponse.CompactionState.NONE;
public boolean hasCompactionState() {
return ((bitField0_ & 0x00000002) == 0x00000002);
@ -1757,7 +1757,7 @@ public final class AdminProtos {
public interface GetStoreFileResponseOrBuilder
extends com.google.protobuf.MessageOrBuilder {
// repeated string storeFile = 1;
// repeated string store_file = 1;
java.util.List<String> getStoreFileList();
int getStoreFileCount();
String getStoreFile(int index);
@ -1790,8 +1790,8 @@ public final class AdminProtos {
return org.apache.hadoop.hbase.protobuf.generated.AdminProtos.internal_static_GetStoreFileResponse_fieldAccessorTable;
}
// repeated string storeFile = 1;
public static final int STOREFILE_FIELD_NUMBER = 1;
// repeated string store_file = 1;
public static final int STORE_FILE_FIELD_NUMBER = 1;
private com.google.protobuf.LazyStringList storeFile_;
public java.util.List<String>
getStoreFileList() {
@ -1875,7 +1875,7 @@ public final class AdminProtos {
int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode();
if (getStoreFileCount() > 0) {
hash = (37 * hash) + STOREFILE_FIELD_NUMBER;
hash = (37 * hash) + STORE_FILE_FIELD_NUMBER;
hash = (53 * hash) + getStoreFileList().hashCode();
}
hash = (29 * hash) + getUnknownFields().hashCode();
@ -2106,7 +2106,7 @@ public final class AdminProtos {
private int bitField0_;
// repeated string storeFile = 1;
// repeated string store_file = 1;
private com.google.protobuf.LazyStringList storeFile_ = com.google.protobuf.LazyStringArrayList.EMPTY;
private void ensureStoreFileIsMutable() {
if (!((bitField0_ & 0x00000001) == 0x00000001)) {
@ -2476,7 +2476,7 @@ public final class AdminProtos {
public interface GetOnlineRegionResponseOrBuilder
extends com.google.protobuf.MessageOrBuilder {
// repeated .RegionInfo regionInfo = 1;
// repeated .RegionInfo region_info = 1;
java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfo>
getRegionInfoList();
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfo getRegionInfo(int index);
@ -2514,8 +2514,8 @@ public final class AdminProtos {
return org.apache.hadoop.hbase.protobuf.generated.AdminProtos.internal_static_GetOnlineRegionResponse_fieldAccessorTable;
}
// repeated .RegionInfo regionInfo = 1;
public static final int REGIONINFO_FIELD_NUMBER = 1;
// repeated .RegionInfo region_info = 1;
public static final int REGION_INFO_FIELD_NUMBER = 1;
private java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfo> regionInfo_;
public java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfo> getRegionInfoList() {
return regionInfo_;
@ -2607,7 +2607,7 @@ public final class AdminProtos {
int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode();
if (getRegionInfoCount() > 0) {
hash = (37 * hash) + REGIONINFO_FIELD_NUMBER;
hash = (37 * hash) + REGION_INFO_FIELD_NUMBER;
hash = (53 * hash) + getRegionInfoList().hashCode();
}
hash = (29 * hash) + getUnknownFields().hashCode();
@ -2869,7 +2869,7 @@ public final class AdminProtos {
private int bitField0_;
// repeated .RegionInfo regionInfo = 1;
// repeated .RegionInfo region_info = 1;
private java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfo> regionInfo_ =
java.util.Collections.emptyList();
private void ensureRegionInfoIsMutable() {
@ -3069,7 +3069,7 @@ public final class AdminProtos {
public interface OpenRegionRequestOrBuilder
extends com.google.protobuf.MessageOrBuilder {
// repeated .OpenRegionRequest.RegionOpenInfo openInfo = 1;
// repeated .OpenRegionRequest.RegionOpenInfo open_info = 1;
java.util.List<org.apache.hadoop.hbase.protobuf.generated.AdminProtos.OpenRegionRequest.RegionOpenInfo>
getOpenInfoList();
org.apache.hadoop.hbase.protobuf.generated.AdminProtos.OpenRegionRequest.RegionOpenInfo getOpenInfo(int index);
@ -3115,11 +3115,11 @@ public final class AdminProtos {
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfo getRegion();
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfoOrBuilder getRegionOrBuilder();
// optional uint32 versionOfOfflineNode = 2;
// optional uint32 version_of_offline_node = 2;
boolean hasVersionOfOfflineNode();
int getVersionOfOfflineNode();
// repeated .ServerName favoredNodes = 3;
// repeated .ServerName favored_nodes = 3;
java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName>
getFavoredNodesList();
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName getFavoredNodes(int index);
@ -3171,8 +3171,8 @@ public final class AdminProtos {
return region_;
}
// optional uint32 versionOfOfflineNode = 2;
public static final int VERSIONOFOFFLINENODE_FIELD_NUMBER = 2;
// optional uint32 version_of_offline_node = 2;
public static final int VERSION_OF_OFFLINE_NODE_FIELD_NUMBER = 2;
private int versionOfOfflineNode_;
public boolean hasVersionOfOfflineNode() {
return ((bitField0_ & 0x00000002) == 0x00000002);
@ -3181,8 +3181,8 @@ public final class AdminProtos {
return versionOfOfflineNode_;
}
// repeated .ServerName favoredNodes = 3;
public static final int FAVOREDNODES_FIELD_NUMBER = 3;
// repeated .ServerName favored_nodes = 3;
public static final int FAVORED_NODES_FIELD_NUMBER = 3;
private java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName> favoredNodes_;
public java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName> getFavoredNodesList() {
return favoredNodes_;
@ -3312,11 +3312,11 @@ public final class AdminProtos {
hash = (53 * hash) + getRegion().hashCode();
}
if (hasVersionOfOfflineNode()) {
hash = (37 * hash) + VERSIONOFOFFLINENODE_FIELD_NUMBER;
hash = (37 * hash) + VERSION_OF_OFFLINE_NODE_FIELD_NUMBER;
hash = (53 * hash) + getVersionOfOfflineNode();
}
if (getFavoredNodesCount() > 0) {
hash = (37 * hash) + FAVOREDNODES_FIELD_NUMBER;
hash = (37 * hash) + FAVORED_NODES_FIELD_NUMBER;
hash = (53 * hash) + getFavoredNodesList().hashCode();
}
hash = (29 * hash) + getUnknownFields().hashCode();
@ -3719,7 +3719,7 @@ public final class AdminProtos {
return regionBuilder_;
}
// optional uint32 versionOfOfflineNode = 2;
// optional uint32 version_of_offline_node = 2;
private int versionOfOfflineNode_ ;
public boolean hasVersionOfOfflineNode() {
return ((bitField0_ & 0x00000002) == 0x00000002);
@ -3740,7 +3740,7 @@ public final class AdminProtos {
return this;
}
// repeated .ServerName favoredNodes = 3;
// repeated .ServerName favored_nodes = 3;
private java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName> favoredNodes_ =
java.util.Collections.emptyList();
private void ensureFavoredNodesIsMutable() {
@ -3937,8 +3937,8 @@ public final class AdminProtos {
// @@protoc_insertion_point(class_scope:OpenRegionRequest.RegionOpenInfo)
}
// repeated .OpenRegionRequest.RegionOpenInfo openInfo = 1;
public static final int OPENINFO_FIELD_NUMBER = 1;
// repeated .OpenRegionRequest.RegionOpenInfo open_info = 1;
public static final int OPEN_INFO_FIELD_NUMBER = 1;
private java.util.List<org.apache.hadoop.hbase.protobuf.generated.AdminProtos.OpenRegionRequest.RegionOpenInfo> openInfo_;
public java.util.List<org.apache.hadoop.hbase.protobuf.generated.AdminProtos.OpenRegionRequest.RegionOpenInfo> getOpenInfoList() {
return openInfo_;
@ -4030,7 +4030,7 @@ public final class AdminProtos {
int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode();
if (getOpenInfoCount() > 0) {
hash = (37 * hash) + OPENINFO_FIELD_NUMBER;
hash = (37 * hash) + OPEN_INFO_FIELD_NUMBER;
hash = (53 * hash) + getOpenInfoList().hashCode();
}
hash = (29 * hash) + getUnknownFields().hashCode();
@ -4292,7 +4292,7 @@ public final class AdminProtos {
private int bitField0_;
// repeated .OpenRegionRequest.RegionOpenInfo openInfo = 1;
// repeated .OpenRegionRequest.RegionOpenInfo open_info = 1;
private java.util.List<org.apache.hadoop.hbase.protobuf.generated.AdminProtos.OpenRegionRequest.RegionOpenInfo> openInfo_ =
java.util.Collections.emptyList();
private void ensureOpenInfoIsMutable() {
@ -4492,7 +4492,7 @@ public final class AdminProtos {
public interface OpenRegionResponseOrBuilder
extends com.google.protobuf.MessageOrBuilder {
// repeated .OpenRegionResponse.RegionOpeningState openingState = 1;
// repeated .OpenRegionResponse.RegionOpeningState opening_state = 1;
java.util.List<org.apache.hadoop.hbase.protobuf.generated.AdminProtos.OpenRegionResponse.RegionOpeningState> getOpeningStateList();
int getOpeningStateCount();
org.apache.hadoop.hbase.protobuf.generated.AdminProtos.OpenRegionResponse.RegionOpeningState getOpeningState(int index);
@ -4597,8 +4597,8 @@ public final class AdminProtos {
// @@protoc_insertion_point(enum_scope:OpenRegionResponse.RegionOpeningState)
}
// repeated .OpenRegionResponse.RegionOpeningState openingState = 1;
public static final int OPENINGSTATE_FIELD_NUMBER = 1;
// repeated .OpenRegionResponse.RegionOpeningState opening_state = 1;
public static final int OPENING_STATE_FIELD_NUMBER = 1;
private java.util.List<org.apache.hadoop.hbase.protobuf.generated.AdminProtos.OpenRegionResponse.RegionOpeningState> openingState_;
public java.util.List<org.apache.hadoop.hbase.protobuf.generated.AdminProtos.OpenRegionResponse.RegionOpeningState> getOpeningStateList() {
return openingState_;
@ -4681,7 +4681,7 @@ public final class AdminProtos {
int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode();
if (getOpeningStateCount() > 0) {
hash = (37 * hash) + OPENINGSTATE_FIELD_NUMBER;
hash = (37 * hash) + OPENING_STATE_FIELD_NUMBER;
hash = (53 * hash) + hashEnumList(getOpeningStateList());
}
hash = (29 * hash) + getUnknownFields().hashCode();
@ -4931,7 +4931,7 @@ public final class AdminProtos {
private int bitField0_;
// repeated .OpenRegionResponse.RegionOpeningState openingState = 1;
// repeated .OpenRegionResponse.RegionOpeningState opening_state = 1;
private java.util.List<org.apache.hadoop.hbase.protobuf.generated.AdminProtos.OpenRegionResponse.RegionOpeningState> openingState_ =
java.util.Collections.emptyList();
private void ensureOpeningStateIsMutable() {
@ -5001,15 +5001,15 @@ public final class AdminProtos {
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier getRegion();
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder getRegionOrBuilder();
// optional uint32 versionOfClosingNode = 2;
// optional uint32 version_of_closing_node = 2;
boolean hasVersionOfClosingNode();
int getVersionOfClosingNode();
// optional bool transitionInZK = 3 [default = true];
// optional bool transition_in_ZK = 3 [default = true];
boolean hasTransitionInZK();
boolean getTransitionInZK();
// optional .ServerName destinationServer = 4;
// optional .ServerName destination_server = 4;
boolean hasDestinationServer();
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName getDestinationServer();
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerNameOrBuilder getDestinationServerOrBuilder();
@ -5056,8 +5056,8 @@ public final class AdminProtos {
return region_;
}
// optional uint32 versionOfClosingNode = 2;
public static final int VERSIONOFCLOSINGNODE_FIELD_NUMBER = 2;
// optional uint32 version_of_closing_node = 2;
public static final int VERSION_OF_CLOSING_NODE_FIELD_NUMBER = 2;
private int versionOfClosingNode_;
public boolean hasVersionOfClosingNode() {
return ((bitField0_ & 0x00000002) == 0x00000002);
@ -5066,8 +5066,8 @@ public final class AdminProtos {
return versionOfClosingNode_;
}
// optional bool transitionInZK = 3 [default = true];
public static final int TRANSITIONINZK_FIELD_NUMBER = 3;
// optional bool transition_in_ZK = 3 [default = true];
public static final int TRANSITION_IN_ZK_FIELD_NUMBER = 3;
private boolean transitionInZK_;
public boolean hasTransitionInZK() {
return ((bitField0_ & 0x00000004) == 0x00000004);
@ -5076,8 +5076,8 @@ public final class AdminProtos {
return transitionInZK_;
}
// optional .ServerName destinationServer = 4;
public static final int DESTINATIONSERVER_FIELD_NUMBER = 4;
// optional .ServerName destination_server = 4;
public static final int DESTINATION_SERVER_FIELD_NUMBER = 4;
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName destinationServer_;
public boolean hasDestinationServer() {
return ((bitField0_ & 0x00000008) == 0x00000008);
@ -5215,15 +5215,15 @@ public final class AdminProtos {
hash = (53 * hash) + getRegion().hashCode();
}
if (hasVersionOfClosingNode()) {
hash = (37 * hash) + VERSIONOFCLOSINGNODE_FIELD_NUMBER;
hash = (37 * hash) + VERSION_OF_CLOSING_NODE_FIELD_NUMBER;
hash = (53 * hash) + getVersionOfClosingNode();
}
if (hasTransitionInZK()) {
hash = (37 * hash) + TRANSITIONINZK_FIELD_NUMBER;
hash = (37 * hash) + TRANSITION_IN_ZK_FIELD_NUMBER;
hash = (53 * hash) + hashBoolean(getTransitionInZK());
}
if (hasDestinationServer()) {
hash = (37 * hash) + DESTINATIONSERVER_FIELD_NUMBER;
hash = (37 * hash) + DESTINATION_SERVER_FIELD_NUMBER;
hash = (53 * hash) + getDestinationServer().hashCode();
}
hash = (29 * hash) + getUnknownFields().hashCode();
@ -5619,7 +5619,7 @@ public final class AdminProtos {
return regionBuilder_;
}
// optional uint32 versionOfClosingNode = 2;
// optional uint32 version_of_closing_node = 2;
private int versionOfClosingNode_ ;
public boolean hasVersionOfClosingNode() {
return ((bitField0_ & 0x00000002) == 0x00000002);
@ -5640,7 +5640,7 @@ public final class AdminProtos {
return this;
}
// optional bool transitionInZK = 3 [default = true];
// optional bool transition_in_ZK = 3 [default = true];
private boolean transitionInZK_ = true;
public boolean hasTransitionInZK() {
return ((bitField0_ & 0x00000004) == 0x00000004);
@ -5661,7 +5661,7 @@ public final class AdminProtos {
return this;
}
// optional .ServerName destinationServer = 4;
// optional .ServerName destination_server = 4;
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName destinationServer_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.getDefaultInstance();
private com.google.protobuf.SingleFieldBuilder<
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerNameOrBuilder> destinationServerBuilder_;
@ -6149,7 +6149,7 @@ public final class AdminProtos {
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier getRegion();
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder getRegionOrBuilder();
// optional uint64 ifOlderThanTs = 2;
// optional uint64 if_older_than_ts = 2;
boolean hasIfOlderThanTs();
long getIfOlderThanTs();
}
@ -6195,8 +6195,8 @@ public final class AdminProtos {
return region_;
}
// optional uint64 ifOlderThanTs = 2;
public static final int IFOLDERTHANTS_FIELD_NUMBER = 2;
// optional uint64 if_older_than_ts = 2;
public static final int IF_OLDER_THAN_TS_FIELD_NUMBER = 2;
private long ifOlderThanTs_;
public boolean hasIfOlderThanTs() {
return ((bitField0_ & 0x00000002) == 0x00000002);
@ -6299,7 +6299,7 @@ public final class AdminProtos {
hash = (53 * hash) + getRegion().hashCode();
}
if (hasIfOlderThanTs()) {
hash = (37 * hash) + IFOLDERTHANTS_FIELD_NUMBER;
hash = (37 * hash) + IF_OLDER_THAN_TS_FIELD_NUMBER;
hash = (53 * hash) + hashLong(getIfOlderThanTs());
}
hash = (29 * hash) + getUnknownFields().hashCode();
@ -6648,7 +6648,7 @@ public final class AdminProtos {
return regionBuilder_;
}
// optional uint64 ifOlderThanTs = 2;
// optional uint64 if_older_than_ts = 2;
private long ifOlderThanTs_ ;
public boolean hasIfOlderThanTs() {
return ((bitField0_ & 0x00000002) == 0x00000002);
@ -6683,7 +6683,7 @@ public final class AdminProtos {
public interface FlushRegionResponseOrBuilder
extends com.google.protobuf.MessageOrBuilder {
// required uint64 lastFlushTime = 1;
// required uint64 last_flush_time = 1;
boolean hasLastFlushTime();
long getLastFlushTime();
@ -6720,8 +6720,8 @@ public final class AdminProtos {
}
private int bitField0_;
// required uint64 lastFlushTime = 1;
public static final int LASTFLUSHTIME_FIELD_NUMBER = 1;
// required uint64 last_flush_time = 1;
public static final int LAST_FLUSH_TIME_FIELD_NUMBER = 1;
private long lastFlushTime_;
public boolean hasLastFlushTime() {
return ((bitField0_ & 0x00000001) == 0x00000001);
@ -6826,7 +6826,7 @@ public final class AdminProtos {
int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode();
if (hasLastFlushTime()) {
hash = (37 * hash) + LASTFLUSHTIME_FIELD_NUMBER;
hash = (37 * hash) + LAST_FLUSH_TIME_FIELD_NUMBER;
hash = (53 * hash) + hashLong(getLastFlushTime());
}
if (hasFlushed()) {
@ -7072,7 +7072,7 @@ public final class AdminProtos {
private int bitField0_;
// required uint64 lastFlushTime = 1;
// required uint64 last_flush_time = 1;
private long lastFlushTime_ ;
public boolean hasLastFlushTime() {
return ((bitField0_ & 0x00000001) == 0x00000001);
@ -7133,7 +7133,7 @@ public final class AdminProtos {
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier getRegion();
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder getRegionOrBuilder();
// optional bytes splitPoint = 2;
// optional bytes split_point = 2;
boolean hasSplitPoint();
com.google.protobuf.ByteString getSplitPoint();
}
@ -7179,8 +7179,8 @@ public final class AdminProtos {
return region_;
}
// optional bytes splitPoint = 2;
public static final int SPLITPOINT_FIELD_NUMBER = 2;
// optional bytes split_point = 2;
public static final int SPLIT_POINT_FIELD_NUMBER = 2;
private com.google.protobuf.ByteString splitPoint_;
public boolean hasSplitPoint() {
return ((bitField0_ & 0x00000002) == 0x00000002);
@ -7283,7 +7283,7 @@ public final class AdminProtos {
hash = (53 * hash) + getRegion().hashCode();
}
if (hasSplitPoint()) {
hash = (37 * hash) + SPLITPOINT_FIELD_NUMBER;
hash = (37 * hash) + SPLIT_POINT_FIELD_NUMBER;
hash = (53 * hash) + getSplitPoint().hashCode();
}
hash = (29 * hash) + getUnknownFields().hashCode();
@ -7632,7 +7632,7 @@ public final class AdminProtos {
return regionBuilder_;
}
// optional bytes splitPoint = 2;
// optional bytes split_point = 2;
private com.google.protobuf.ByteString splitPoint_ = com.google.protobuf.ByteString.EMPTY;
public boolean hasSplitPoint() {
return ((bitField0_ & 0x00000002) == 0x00000002);
@ -8878,12 +8878,12 @@ public final class AdminProtos {
public interface MergeRegionsRequestOrBuilder
extends com.google.protobuf.MessageOrBuilder {
// required .RegionSpecifier regionA = 1;
// required .RegionSpecifier region_a = 1;
boolean hasRegionA();
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier getRegionA();
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder getRegionAOrBuilder();
// required .RegionSpecifier regionB = 2;
// required .RegionSpecifier region_b = 2;
boolean hasRegionB();
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier getRegionB();
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder getRegionBOrBuilder();
@ -8921,8 +8921,8 @@ public final class AdminProtos {
}
private int bitField0_;
// required .RegionSpecifier regionA = 1;
public static final int REGIONA_FIELD_NUMBER = 1;
// required .RegionSpecifier region_a = 1;
public static final int REGION_A_FIELD_NUMBER = 1;
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier regionA_;
public boolean hasRegionA() {
return ((bitField0_ & 0x00000001) == 0x00000001);
@ -8934,8 +8934,8 @@ public final class AdminProtos {
return regionA_;
}
// required .RegionSpecifier regionB = 2;
public static final int REGIONB_FIELD_NUMBER = 2;
// required .RegionSpecifier region_b = 2;
public static final int REGION_B_FIELD_NUMBER = 2;
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier regionB_;
public boolean hasRegionB() {
return ((bitField0_ & 0x00000002) == 0x00000002);
@ -9068,11 +9068,11 @@ public final class AdminProtos {
int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode();
if (hasRegionA()) {
hash = (37 * hash) + REGIONA_FIELD_NUMBER;
hash = (37 * hash) + REGION_A_FIELD_NUMBER;
hash = (53 * hash) + getRegionA().hashCode();
}
if (hasRegionB()) {
hash = (37 * hash) + REGIONB_FIELD_NUMBER;
hash = (37 * hash) + REGION_B_FIELD_NUMBER;
hash = (53 * hash) + getRegionB().hashCode();
}
if (hasForcible()) {
@ -9370,7 +9370,7 @@ public final class AdminProtos {
private int bitField0_;
// required .RegionSpecifier regionA = 1;
// required .RegionSpecifier region_a = 1;
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier regionA_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance();
private com.google.protobuf.SingleFieldBuilder<
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder> regionABuilder_;
@ -9460,7 +9460,7 @@ public final class AdminProtos {
return regionABuilder_;
}
// required .RegionSpecifier regionB = 2;
// required .RegionSpecifier region_b = 2;
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier regionB_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance();
private com.google.protobuf.SingleFieldBuilder<
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder> regionBBuilder_;
@ -9890,12 +9890,12 @@ public final class AdminProtos {
org.apache.hadoop.hbase.protobuf.generated.WALProtos.WALKey getKey();
org.apache.hadoop.hbase.protobuf.generated.WALProtos.WALKeyOrBuilder getKeyOrBuilder();
// repeated bytes keyValueBytes = 2;
// repeated bytes key_value_bytes = 2;
java.util.List<com.google.protobuf.ByteString> getKeyValueBytesList();
int getKeyValueBytesCount();
com.google.protobuf.ByteString getKeyValueBytes(int index);
// optional int32 associatedCellCount = 3;
// optional int32 associated_cell_count = 3;
boolean hasAssociatedCellCount();
int getAssociatedCellCount();
}
@ -9941,8 +9941,8 @@ public final class AdminProtos {
return key_;
}
// repeated bytes keyValueBytes = 2;
public static final int KEYVALUEBYTES_FIELD_NUMBER = 2;
// repeated bytes key_value_bytes = 2;
public static final int KEY_VALUE_BYTES_FIELD_NUMBER = 2;
private java.util.List<com.google.protobuf.ByteString> keyValueBytes_;
public java.util.List<com.google.protobuf.ByteString>
getKeyValueBytesList() {
@ -9955,8 +9955,8 @@ public final class AdminProtos {
return keyValueBytes_.get(index);
}
// optional int32 associatedCellCount = 3;
public static final int ASSOCIATEDCELLCOUNT_FIELD_NUMBER = 3;
// optional int32 associated_cell_count = 3;
public static final int ASSOCIATED_CELL_COUNT_FIELD_NUMBER = 3;
private int associatedCellCount_;
public boolean hasAssociatedCellCount() {
return ((bitField0_ & 0x00000002) == 0x00000002);
@ -10074,11 +10074,11 @@ public final class AdminProtos {
hash = (53 * hash) + getKey().hashCode();
}
if (getKeyValueBytesCount() > 0) {
hash = (37 * hash) + KEYVALUEBYTES_FIELD_NUMBER;
hash = (37 * hash) + KEY_VALUE_BYTES_FIELD_NUMBER;
hash = (53 * hash) + getKeyValueBytesList().hashCode();
}
if (hasAssociatedCellCount()) {
hash = (37 * hash) + ASSOCIATEDCELLCOUNT_FIELD_NUMBER;
hash = (37 * hash) + ASSOCIATED_CELL_COUNT_FIELD_NUMBER;
hash = (53 * hash) + getAssociatedCellCount();
}
hash = (29 * hash) + getUnknownFields().hashCode();
@ -10449,7 +10449,7 @@ public final class AdminProtos {
return keyBuilder_;
}
// repeated bytes keyValueBytes = 2;
// repeated bytes key_value_bytes = 2;
private java.util.List<com.google.protobuf.ByteString> keyValueBytes_ = java.util.Collections.emptyList();;
private void ensureKeyValueBytesIsMutable() {
if (!((bitField0_ & 0x00000002) == 0x00000002)) {
@ -10500,7 +10500,7 @@ public final class AdminProtos {
return this;
}
// optional int32 associatedCellCount = 3;
// optional int32 associated_cell_count = 3;
private int associatedCellCount_ ;
public boolean hasAssociatedCellCount() {
return ((bitField0_ & 0x00000004) == 0x00000004);
@ -11728,7 +11728,7 @@ public final class AdminProtos {
public interface RollWALWriterResponseOrBuilder
extends com.google.protobuf.MessageOrBuilder {
// repeated bytes regionToFlush = 1;
// repeated bytes region_to_flush = 1;
java.util.List<com.google.protobuf.ByteString> getRegionToFlushList();
int getRegionToFlushCount();
com.google.protobuf.ByteString getRegionToFlush(int index);
@ -11761,8 +11761,8 @@ public final class AdminProtos {
return org.apache.hadoop.hbase.protobuf.generated.AdminProtos.internal_static_RollWALWriterResponse_fieldAccessorTable;
}
// repeated bytes regionToFlush = 1;
public static final int REGIONTOFLUSH_FIELD_NUMBER = 1;
// repeated bytes region_to_flush = 1;
public static final int REGION_TO_FLUSH_FIELD_NUMBER = 1;
private java.util.List<com.google.protobuf.ByteString> regionToFlush_;
public java.util.List<com.google.protobuf.ByteString>
getRegionToFlushList() {
@ -11846,7 +11846,7 @@ public final class AdminProtos {
int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode();
if (getRegionToFlushCount() > 0) {
hash = (37 * hash) + REGIONTOFLUSH_FIELD_NUMBER;
hash = (37 * hash) + REGION_TO_FLUSH_FIELD_NUMBER;
hash = (53 * hash) + getRegionToFlushList().hashCode();
}
hash = (29 * hash) + getUnknownFields().hashCode();
@ -12076,7 +12076,7 @@ public final class AdminProtos {
private int bitField0_;
// repeated bytes regionToFlush = 1;
// repeated bytes region_to_flush = 1;
private java.util.List<com.google.protobuf.ByteString> regionToFlush_ = java.util.Collections.emptyList();;
private void ensureRegionToFlushIsMutable() {
if (!((bitField0_ & 0x00000001) == 0x00000001)) {
@ -13157,12 +13157,12 @@ public final class AdminProtos {
public interface ServerInfoOrBuilder
extends com.google.protobuf.MessageOrBuilder {
// required .ServerName serverName = 1;
// required .ServerName server_name = 1;
boolean hasServerName();
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName getServerName();
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerNameOrBuilder getServerNameOrBuilder();
// optional uint32 webuiPort = 2;
// optional uint32 webui_port = 2;
boolean hasWebuiPort();
int getWebuiPort();
}
@ -13195,8 +13195,8 @@ public final class AdminProtos {
}
private int bitField0_;
// required .ServerName serverName = 1;
public static final int SERVERNAME_FIELD_NUMBER = 1;
// required .ServerName server_name = 1;
public static final int SERVER_NAME_FIELD_NUMBER = 1;
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName serverName_;
public boolean hasServerName() {
return ((bitField0_ & 0x00000001) == 0x00000001);
@ -13208,8 +13208,8 @@ public final class AdminProtos {
return serverName_;
}
// optional uint32 webuiPort = 2;
public static final int WEBUIPORT_FIELD_NUMBER = 2;
// optional uint32 webui_port = 2;
public static final int WEBUI_PORT_FIELD_NUMBER = 2;
private int webuiPort_;
public boolean hasWebuiPort() {
return ((bitField0_ & 0x00000002) == 0x00000002);
@ -13308,11 +13308,11 @@ public final class AdminProtos {
int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode();
if (hasServerName()) {
hash = (37 * hash) + SERVERNAME_FIELD_NUMBER;
hash = (37 * hash) + SERVER_NAME_FIELD_NUMBER;
hash = (53 * hash) + getServerName().hashCode();
}
if (hasWebuiPort()) {
hash = (37 * hash) + WEBUIPORT_FIELD_NUMBER;
hash = (37 * hash) + WEBUI_PORT_FIELD_NUMBER;
hash = (53 * hash) + getWebuiPort();
}
hash = (29 * hash) + getUnknownFields().hashCode();
@ -13571,7 +13571,7 @@ public final class AdminProtos {
private int bitField0_;
// required .ServerName serverName = 1;
// required .ServerName server_name = 1;
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName serverName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.getDefaultInstance();
private com.google.protobuf.SingleFieldBuilder<
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerNameOrBuilder> serverNameBuilder_;
@ -13661,7 +13661,7 @@ public final class AdminProtos {
return serverNameBuilder_;
}
// optional uint32 webuiPort = 2;
// optional uint32 webui_port = 2;
private int webuiPort_ ;
public boolean hasWebuiPort() {
return ((bitField0_ & 0x00000002) == 0x00000002);
@ -13696,7 +13696,7 @@ public final class AdminProtos {
public interface GetServerInfoResponseOrBuilder
extends com.google.protobuf.MessageOrBuilder {
// required .ServerInfo serverInfo = 1;
// required .ServerInfo server_info = 1;
boolean hasServerInfo();
org.apache.hadoop.hbase.protobuf.generated.AdminProtos.ServerInfo getServerInfo();
org.apache.hadoop.hbase.protobuf.generated.AdminProtos.ServerInfoOrBuilder getServerInfoOrBuilder();
@ -13730,8 +13730,8 @@ public final class AdminProtos {
}
private int bitField0_;
// required .ServerInfo serverInfo = 1;
public static final int SERVERINFO_FIELD_NUMBER = 1;
// required .ServerInfo server_info = 1;
public static final int SERVER_INFO_FIELD_NUMBER = 1;
private org.apache.hadoop.hbase.protobuf.generated.AdminProtos.ServerInfo serverInfo_;
public boolean hasServerInfo() {
return ((bitField0_ & 0x00000001) == 0x00000001);
@ -13820,7 +13820,7 @@ public final class AdminProtos {
int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode();
if (hasServerInfo()) {
hash = (37 * hash) + SERVERINFO_FIELD_NUMBER;
hash = (37 * hash) + SERVER_INFO_FIELD_NUMBER;
hash = (53 * hash) + getServerInfo().hashCode();
}
hash = (29 * hash) + getUnknownFields().hashCode();
@ -14065,7 +14065,7 @@ public final class AdminProtos {
private int bitField0_;
// required .ServerInfo serverInfo = 1;
// required .ServerInfo server_info = 1;
private org.apache.hadoop.hbase.protobuf.generated.AdminProtos.ServerInfo serverInfo_ = org.apache.hadoop.hbase.protobuf.generated.AdminProtos.ServerInfo.getDefaultInstance();
private com.google.protobuf.SingleFieldBuilder<
org.apache.hadoop.hbase.protobuf.generated.AdminProtos.ServerInfo, org.apache.hadoop.hbase.protobuf.generated.AdminProtos.ServerInfo.Builder, org.apache.hadoop.hbase.protobuf.generated.AdminProtos.ServerInfoOrBuilder> serverInfoBuilder_;
@ -15387,79 +15387,80 @@ public final class AdminProtos {
static {
java.lang.String[] descriptorData = {
"\n\013Admin.proto\032\014Client.proto\032\013hbase.proto" +
"\032\tWAL.proto\"Q\n\024GetRegionInfoRequest\022 \n\006r" +
"egion\030\001 \002(\0132\020.RegionSpecifier\022\027\n\017compact" +
"ionState\030\002 \001(\010\"\301\001\n\025GetRegionInfoResponse" +
"\022\037\n\nregionInfo\030\001 \002(\0132\013.RegionInfo\022?\n\017com" +
"pactionState\030\002 \001(\0162&.GetRegionInfoRespon" +
"se.CompactionState\"F\n\017CompactionState\022\010\n" +
"\004NONE\020\000\022\t\n\005MINOR\020\001\022\t\n\005MAJOR\020\002\022\023\n\017MAJOR_A" +
"ND_MINOR\020\003\"G\n\023GetStoreFileRequest\022 \n\006reg" +
"ion\030\001 \002(\0132\020.RegionSpecifier\022\016\n\006family\030\002 ",
"\003(\014\")\n\024GetStoreFileResponse\022\021\n\tstoreFile" +
"\030\001 \003(\t\"\030\n\026GetOnlineRegionRequest\":\n\027GetO" +
"nlineRegionResponse\022\037\n\nregionInfo\030\001 \003(\0132" +
"\013.RegionInfo\"\270\001\n\021OpenRegionRequest\0223\n\010op" +
"enInfo\030\001 \003(\0132!.OpenRegionRequest.RegionO" +
"penInfo\032n\n\016RegionOpenInfo\022\033\n\006region\030\001 \002(" +
"\0132\013.RegionInfo\022\034\n\024versionOfOfflineNode\030\002" +
" \001(\r\022!\n\014favoredNodes\030\003 \003(\0132\013.ServerName\"" +
"\234\001\n\022OpenRegionResponse\022<\n\014openingState\030\001" +
" \003(\0162&.OpenRegionResponse.RegionOpeningS",
"tate\"H\n\022RegionOpeningState\022\n\n\006OPENED\020\000\022\022" +
"\n\016ALREADY_OPENED\020\001\022\022\n\016FAILED_OPENING\020\002\"\232" +
"\001\n\022CloseRegionRequest\022 \n\006region\030\001 \002(\0132\020." +
"RegionSpecifier\022\034\n\024versionOfClosingNode\030" +
"\002 \001(\r\022\034\n\016transitionInZK\030\003 \001(\010:\004true\022&\n\021d" +
"estinationServer\030\004 \001(\0132\013.ServerName\"%\n\023C" +
"loseRegionResponse\022\016\n\006closed\030\001 \002(\010\"M\n\022Fl" +
"ushRegionRequest\022 \n\006region\030\001 \002(\0132\020.Regio" +
"nSpecifier\022\025\n\rifOlderThanTs\030\002 \001(\004\"=\n\023Flu" +
"shRegionResponse\022\025\n\rlastFlushTime\030\001 \002(\004\022",
"\017\n\007flushed\030\002 \001(\010\"J\n\022SplitRegionRequest\022 " +
"\n\006region\030\001 \002(\0132\020.RegionSpecifier\022\022\n\nspli" +
"tPoint\030\002 \001(\014\"\025\n\023SplitRegionResponse\"W\n\024C" +
"ompactRegionRequest\022 \n\006region\030\001 \002(\0132\020.Re" +
"gionSpecifier\022\r\n\005major\030\002 \001(\010\022\016\n\006family\030\003" +
" \001(\014\"\027\n\025CompactRegionResponse\"t\n\023MergeRe" +
"gionsRequest\022!\n\007regionA\030\001 \002(\0132\020.RegionSp" +
"ecifier\022!\n\007regionB\030\002 \002(\0132\020.RegionSpecifi" +
"er\022\027\n\010forcible\030\003 \001(\010:\005false\"\026\n\024MergeRegi" +
"onsResponse\"T\n\010WALEntry\022\024\n\003key\030\001 \002(\0132\007.W",
"ALKey\022\025\n\rkeyValueBytes\030\002 \003(\014\022\033\n\023associat" +
"edCellCount\030\003 \001(\005\"4\n\030ReplicateWALEntryRe" +
"quest\022\030\n\005entry\030\001 \003(\0132\t.WALEntry\"\033\n\031Repli" +
"cateWALEntryResponse\"\026\n\024RollWALWriterReq" +
"uest\".\n\025RollWALWriterResponse\022\025\n\rregionT" +
"oFlush\030\001 \003(\014\"#\n\021StopServerRequest\022\016\n\006rea" +
"son\030\001 \002(\t\"\024\n\022StopServerResponse\"\026\n\024GetSe" +
"rverInfoRequest\"@\n\nServerInfo\022\037\n\nserverN" +
"ame\030\001 \002(\0132\013.ServerName\022\021\n\twebuiPort\030\002 \001(" +
"\r\"8\n\025GetServerInfoResponse\022\037\n\nserverInfo",
"\030\001 \002(\0132\013.ServerInfo2\337\006\n\014AdminService\022>\n\r" +
"getRegionInfo\022\025.GetRegionInfoRequest\032\026.G" +
"etRegionInfoResponse\022;\n\014getStoreFile\022\024.G" +
"etStoreFileRequest\032\025.GetStoreFileRespons" +
"e\022D\n\017getOnlineRegion\022\027.GetOnlineRegionRe" +
"quest\032\030.GetOnlineRegionResponse\0225\n\nopenR" +
"egion\022\022.OpenRegionRequest\032\023.OpenRegionRe" +
"sponse\0228\n\013closeRegion\022\023.CloseRegionReque" +
"st\032\024.CloseRegionResponse\0228\n\013flushRegion\022" +
"\023.FlushRegionRequest\032\024.FlushRegionRespon",
"se\0228\n\013splitRegion\022\023.SplitRegionRequest\032\024" +
".SplitRegionResponse\022>\n\rcompactRegion\022\025." +
"CompactRegionRequest\032\026.CompactRegionResp" +
"onse\022;\n\014mergeRegions\022\024.MergeRegionsReque" +
"st\032\025.MergeRegionsResponse\022J\n\021replicateWA" +
"LEntry\022\031.ReplicateWALEntryRequest\032\032.Repl" +
"icateWALEntryResponse\022\'\n\006replay\022\r.MultiR" +
"equest\032\016.MultiResponse\022>\n\rrollWALWriter\022" +
"\025.RollWALWriterRequest\032\026.RollWALWriterRe" +
"sponse\022>\n\rgetServerInfo\022\025.GetServerInfoR",
"equest\032\026.GetServerInfoResponse\0225\n\nstopSe" +
"rver\022\022.StopServerRequest\032\023.StopServerRes" +
"ponseBA\n*org.apache.hadoop.hbase.protobu" +
"f.generatedB\013AdminProtosH\001\210\001\001\240\001\001"
"\032\tWAL.proto\"R\n\024GetRegionInfoRequest\022 \n\006r" +
"egion\030\001 \002(\0132\020.RegionSpecifier\022\030\n\020compact" +
"ion_state\030\002 \001(\010\"\303\001\n\025GetRegionInfoRespons" +
"e\022 \n\013region_info\030\001 \002(\0132\013.RegionInfo\022@\n\020c" +
"ompaction_state\030\002 \001(\0162&.GetRegionInfoRes" +
"ponse.CompactionState\"F\n\017CompactionState" +
"\022\010\n\004NONE\020\000\022\t\n\005MINOR\020\001\022\t\n\005MAJOR\020\002\022\023\n\017MAJO" +
"R_AND_MINOR\020\003\"G\n\023GetStoreFileRequest\022 \n\006" +
"region\030\001 \002(\0132\020.RegionSpecifier\022\016\n\006family",
"\030\002 \003(\014\"*\n\024GetStoreFileResponse\022\022\n\nstore_" +
"file\030\001 \003(\t\"\030\n\026GetOnlineRegionRequest\";\n\027" +
"GetOnlineRegionResponse\022 \n\013region_info\030\001" +
" \003(\0132\013.RegionInfo\"\275\001\n\021OpenRegionRequest\022" +
"4\n\topen_info\030\001 \003(\0132!.OpenRegionRequest.R" +
"egionOpenInfo\032r\n\016RegionOpenInfo\022\033\n\006regio" +
"n\030\001 \002(\0132\013.RegionInfo\022\037\n\027version_of_offli" +
"ne_node\030\002 \001(\r\022\"\n\rfavored_nodes\030\003 \003(\0132\013.S" +
"erverName\"\235\001\n\022OpenRegionResponse\022=\n\ropen" +
"ing_state\030\001 \003(\0162&.OpenRegionResponse.Reg",
"ionOpeningState\"H\n\022RegionOpeningState\022\n\n" +
"\006OPENED\020\000\022\022\n\016ALREADY_OPENED\020\001\022\022\n\016FAILED_" +
"OPENING\020\002\"\240\001\n\022CloseRegionRequest\022 \n\006regi" +
"on\030\001 \002(\0132\020.RegionSpecifier\022\037\n\027version_of" +
"_closing_node\030\002 \001(\r\022\036\n\020transition_in_ZK\030" +
"\003 \001(\010:\004true\022\'\n\022destination_server\030\004 \001(\0132" +
"\013.ServerName\"%\n\023CloseRegionResponse\022\016\n\006c" +
"losed\030\001 \002(\010\"P\n\022FlushRegionRequest\022 \n\006reg" +
"ion\030\001 \002(\0132\020.RegionSpecifier\022\030\n\020if_older_" +
"than_ts\030\002 \001(\004\"?\n\023FlushRegionResponse\022\027\n\017",
"last_flush_time\030\001 \002(\004\022\017\n\007flushed\030\002 \001(\010\"K" +
"\n\022SplitRegionRequest\022 \n\006region\030\001 \002(\0132\020.R" +
"egionSpecifier\022\023\n\013split_point\030\002 \001(\014\"\025\n\023S" +
"plitRegionResponse\"W\n\024CompactRegionReque" +
"st\022 \n\006region\030\001 \002(\0132\020.RegionSpecifier\022\r\n\005" +
"major\030\002 \001(\010\022\016\n\006family\030\003 \001(\014\"\027\n\025CompactRe" +
"gionResponse\"v\n\023MergeRegionsRequest\022\"\n\010r" +
"egion_a\030\001 \002(\0132\020.RegionSpecifier\022\"\n\010regio" +
"n_b\030\002 \002(\0132\020.RegionSpecifier\022\027\n\010forcible\030" +
"\003 \001(\010:\005false\"\026\n\024MergeRegionsResponse\"X\n\010",
"WALEntry\022\024\n\003key\030\001 \002(\0132\007.WALKey\022\027\n\017key_va" +
"lue_bytes\030\002 \003(\014\022\035\n\025associated_cell_count" +
"\030\003 \001(\005\"4\n\030ReplicateWALEntryRequest\022\030\n\005en" +
"try\030\001 \003(\0132\t.WALEntry\"\033\n\031ReplicateWALEntr" +
"yResponse\"\026\n\024RollWALWriterRequest\"0\n\025Rol" +
"lWALWriterResponse\022\027\n\017region_to_flush\030\001 " +
"\003(\014\"#\n\021StopServerRequest\022\016\n\006reason\030\001 \002(\t" +
"\"\024\n\022StopServerResponse\"\026\n\024GetServerInfoR" +
"equest\"B\n\nServerInfo\022 \n\013server_name\030\001 \002(" +
"\0132\013.ServerName\022\022\n\nwebui_port\030\002 \001(\r\"9\n\025Ge",
"tServerInfoResponse\022 \n\013server_info\030\001 \002(\013" +
"2\013.ServerInfo2\337\006\n\014AdminService\022>\n\rGetReg" +
"ionInfo\022\025.GetRegionInfoRequest\032\026.GetRegi" +
"onInfoResponse\022;\n\014GetStoreFile\022\024.GetStor" +
"eFileRequest\032\025.GetStoreFileResponse\022D\n\017G" +
"etOnlineRegion\022\027.GetOnlineRegionRequest\032" +
"\030.GetOnlineRegionResponse\0225\n\nOpenRegion\022" +
"\022.OpenRegionRequest\032\023.OpenRegionResponse" +
"\0228\n\013CloseRegion\022\023.CloseRegionRequest\032\024.C" +
"loseRegionResponse\0228\n\013FlushRegion\022\023.Flus",
"hRegionRequest\032\024.FlushRegionResponse\0228\n\013" +
"SplitRegion\022\023.SplitRegionRequest\032\024.Split" +
"RegionResponse\022>\n\rCompactRegion\022\025.Compac" +
"tRegionRequest\032\026.CompactRegionResponse\022;" +
"\n\014MergeRegions\022\024.MergeRegionsRequest\032\025.M" +
"ergeRegionsResponse\022J\n\021ReplicateWALEntry" +
"\022\031.ReplicateWALEntryRequest\032\032.ReplicateW" +
"ALEntryResponse\022\'\n\006Replay\022\r.MultiRequest" +
"\032\016.MultiResponse\022>\n\rRollWALWriter\022\025.Roll" +
"WALWriterRequest\032\026.RollWALWriterResponse",
"\022>\n\rGetServerInfo\022\025.GetServerInfoRequest" +
"\032\026.GetServerInfoResponse\0225\n\nStopServer\022\022" +
".StopServerRequest\032\023.StopServerResponseB" +
"A\n*org.apache.hadoop.hbase.protobuf.gene" +
"ratedB\013AdminProtosH\001\210\001\001\240\001\001"
};
com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() {

View File

@ -11,7 +11,7 @@ public final class AggregateProtos {
public interface AggregateArgumentOrBuilder
extends com.google.protobuf.MessageOrBuilder {
// required string interpreterClassName = 1;
// required string interpreter_class_name = 1;
boolean hasInterpreterClassName();
String getInterpreterClassName();
@ -20,7 +20,7 @@ public final class AggregateProtos {
org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Scan getScan();
org.apache.hadoop.hbase.protobuf.generated.ClientProtos.ScanOrBuilder getScanOrBuilder();
// optional bytes interpreterSpecificBytes = 3;
// optional bytes interpreter_specific_bytes = 3;
boolean hasInterpreterSpecificBytes();
com.google.protobuf.ByteString getInterpreterSpecificBytes();
}
@ -53,8 +53,8 @@ public final class AggregateProtos {
}
private int bitField0_;
// required string interpreterClassName = 1;
public static final int INTERPRETERCLASSNAME_FIELD_NUMBER = 1;
// required string interpreter_class_name = 1;
public static final int INTERPRETER_CLASS_NAME_FIELD_NUMBER = 1;
private java.lang.Object interpreterClassName_;
public boolean hasInterpreterClassName() {
return ((bitField0_ & 0x00000001) == 0x00000001);
@ -98,8 +98,8 @@ public final class AggregateProtos {
return scan_;
}
// optional bytes interpreterSpecificBytes = 3;
public static final int INTERPRETERSPECIFICBYTES_FIELD_NUMBER = 3;
// optional bytes interpreter_specific_bytes = 3;
public static final int INTERPRETER_SPECIFIC_BYTES_FIELD_NUMBER = 3;
private com.google.protobuf.ByteString interpreterSpecificBytes_;
public boolean hasInterpreterSpecificBytes() {
return ((bitField0_ & 0x00000004) == 0x00000004);
@ -215,7 +215,7 @@ public final class AggregateProtos {
int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode();
if (hasInterpreterClassName()) {
hash = (37 * hash) + INTERPRETERCLASSNAME_FIELD_NUMBER;
hash = (37 * hash) + INTERPRETER_CLASS_NAME_FIELD_NUMBER;
hash = (53 * hash) + getInterpreterClassName().hashCode();
}
if (hasScan()) {
@ -223,7 +223,7 @@ public final class AggregateProtos {
hash = (53 * hash) + getScan().hashCode();
}
if (hasInterpreterSpecificBytes()) {
hash = (37 * hash) + INTERPRETERSPECIFICBYTES_FIELD_NUMBER;
hash = (37 * hash) + INTERPRETER_SPECIFIC_BYTES_FIELD_NUMBER;
hash = (53 * hash) + getInterpreterSpecificBytes().hashCode();
}
hash = (29 * hash) + getUnknownFields().hashCode();
@ -500,7 +500,7 @@ public final class AggregateProtos {
private int bitField0_;
// required string interpreterClassName = 1;
// required string interpreter_class_name = 1;
private java.lang.Object interpreterClassName_ = "";
public boolean hasInterpreterClassName() {
return ((bitField0_ & 0x00000001) == 0x00000001);
@ -626,7 +626,7 @@ public final class AggregateProtos {
return scanBuilder_;
}
// optional bytes interpreterSpecificBytes = 3;
// optional bytes interpreter_specific_bytes = 3;
private com.google.protobuf.ByteString interpreterSpecificBytes_ = com.google.protobuf.ByteString.EMPTY;
public boolean hasInterpreterSpecificBytes() {
return ((bitField0_ & 0x00000004) == 0x00000004);
@ -664,12 +664,12 @@ public final class AggregateProtos {
public interface AggregateResponseOrBuilder
extends com.google.protobuf.MessageOrBuilder {
// repeated bytes firstPart = 1;
// repeated bytes first_part = 1;
java.util.List<com.google.protobuf.ByteString> getFirstPartList();
int getFirstPartCount();
com.google.protobuf.ByteString getFirstPart(int index);
// optional bytes secondPart = 2;
// optional bytes second_part = 2;
boolean hasSecondPart();
com.google.protobuf.ByteString getSecondPart();
}
@ -702,8 +702,8 @@ public final class AggregateProtos {
}
private int bitField0_;
// repeated bytes firstPart = 1;
public static final int FIRSTPART_FIELD_NUMBER = 1;
// repeated bytes first_part = 1;
public static final int FIRST_PART_FIELD_NUMBER = 1;
private java.util.List<com.google.protobuf.ByteString> firstPart_;
public java.util.List<com.google.protobuf.ByteString>
getFirstPartList() {
@ -716,8 +716,8 @@ public final class AggregateProtos {
return firstPart_.get(index);
}
// optional bytes secondPart = 2;
public static final int SECONDPART_FIELD_NUMBER = 2;
// optional bytes second_part = 2;
public static final int SECOND_PART_FIELD_NUMBER = 2;
private com.google.protobuf.ByteString secondPart_;
public boolean hasSecondPart() {
return ((bitField0_ & 0x00000001) == 0x00000001);
@ -810,11 +810,11 @@ public final class AggregateProtos {
int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode();
if (getFirstPartCount() > 0) {
hash = (37 * hash) + FIRSTPART_FIELD_NUMBER;
hash = (37 * hash) + FIRST_PART_FIELD_NUMBER;
hash = (53 * hash) + getFirstPartList().hashCode();
}
if (hasSecondPart()) {
hash = (37 * hash) + SECONDPART_FIELD_NUMBER;
hash = (37 * hash) + SECOND_PART_FIELD_NUMBER;
hash = (53 * hash) + getSecondPart().hashCode();
}
hash = (29 * hash) + getUnknownFields().hashCode();
@ -1060,7 +1060,7 @@ public final class AggregateProtos {
private int bitField0_;
// repeated bytes firstPart = 1;
// repeated bytes first_part = 1;
private java.util.List<com.google.protobuf.ByteString> firstPart_ = java.util.Collections.emptyList();;
private void ensureFirstPartIsMutable() {
if (!((bitField0_ & 0x00000001) == 0x00000001)) {
@ -1111,7 +1111,7 @@ public final class AggregateProtos {
return this;
}
// optional bytes secondPart = 2;
// optional bytes second_part = 2;
private com.google.protobuf.ByteString secondPart_ = com.google.protobuf.ByteString.EMPTY;
public boolean hasSecondPart() {
return ((bitField0_ & 0x00000002) == 0x00000002);
@ -1776,23 +1776,23 @@ public final class AggregateProtos {
descriptor;
static {
java.lang.String[] descriptorData = {
"\n\017Aggregate.proto\032\014Client.proto\"h\n\021Aggre" +
"gateArgument\022\034\n\024interpreterClassName\030\001 \002" +
"(\t\022\023\n\004scan\030\002 \002(\0132\005.Scan\022 \n\030interpreterSp" +
"ecificBytes\030\003 \001(\014\":\n\021AggregateResponse\022\021" +
"\n\tfirstPart\030\001 \003(\014\022\022\n\nsecondPart\030\002 \001(\0142\366\002" +
"\n\020AggregateService\0220\n\006getMax\022\022.Aggregate" +
"Argument\032\022.AggregateResponse\0220\n\006getMin\022\022" +
".AggregateArgument\032\022.AggregateResponse\0220" +
"\n\006getSum\022\022.AggregateArgument\032\022.Aggregate" +
"Response\0223\n\tgetRowNum\022\022.AggregateArgumen",
"t\032\022.AggregateResponse\0220\n\006getAvg\022\022.Aggreg" +
"ateArgument\032\022.AggregateResponse\0220\n\006getSt" +
"d\022\022.AggregateArgument\032\022.AggregateRespons" +
"e\0223\n\tgetMedian\022\022.AggregateArgument\032\022.Agg" +
"regateResponseBE\n*org.apache.hadoop.hbas" +
"e.protobuf.generatedB\017AggregateProtosH\001\210" +
"\001\001\240\001\001"
"\n\017Aggregate.proto\032\014Client.proto\"l\n\021Aggre" +
"gateArgument\022\036\n\026interpreter_class_name\030\001" +
" \002(\t\022\023\n\004scan\030\002 \002(\0132\005.Scan\022\"\n\032interpreter" +
"_specific_bytes\030\003 \001(\014\"<\n\021AggregateRespon" +
"se\022\022\n\nfirst_part\030\001 \003(\014\022\023\n\013second_part\030\002 " +
"\001(\0142\366\002\n\020AggregateService\0220\n\006GetMax\022\022.Agg" +
"regateArgument\032\022.AggregateResponse\0220\n\006Ge" +
"tMin\022\022.AggregateArgument\032\022.AggregateResp" +
"onse\0220\n\006GetSum\022\022.AggregateArgument\032\022.Agg" +
"regateResponse\0223\n\tGetRowNum\022\022.AggregateA",
"rgument\032\022.AggregateResponse\0220\n\006GetAvg\022\022." +
"AggregateArgument\032\022.AggregateResponse\0220\n" +
"\006GetStd\022\022.AggregateArgument\032\022.AggregateR" +
"esponse\0223\n\tGetMedian\022\022.AggregateArgument" +
"\032\022.AggregateResponseBE\n*org.apache.hadoo" +
"p.hbase.protobuf.generatedB\017AggregatePro" +
"tosH\001\210\001\001\240\001\001"
};
com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() {

View File

@ -15,7 +15,7 @@ public final class AuthenticationProtos {
boolean hasId();
int getId();
// required int64 expirationDate = 2;
// required int64 expiration_date = 2;
boolean hasExpirationDate();
long getExpirationDate();
@ -62,8 +62,8 @@ public final class AuthenticationProtos {
return id_;
}
// required int64 expirationDate = 2;
public static final int EXPIRATIONDATE_FIELD_NUMBER = 2;
// required int64 expiration_date = 2;
public static final int EXPIRATION_DATE_FIELD_NUMBER = 2;
private long expirationDate_;
public boolean hasExpirationDate() {
return ((bitField0_ & 0x00000002) == 0x00000002);
@ -193,7 +193,7 @@ public final class AuthenticationProtos {
hash = (53 * hash) + getId();
}
if (hasExpirationDate()) {
hash = (37 * hash) + EXPIRATIONDATE_FIELD_NUMBER;
hash = (37 * hash) + EXPIRATION_DATE_FIELD_NUMBER;
hash = (53 * hash) + hashLong(getExpirationDate());
}
if (hasKey()) {
@ -482,7 +482,7 @@ public final class AuthenticationProtos {
return this;
}
// required int64 expirationDate = 2;
// required int64 expiration_date = 2;
private long expirationDate_ ;
public boolean hasExpirationDate() {
return ((bitField0_ & 0x00000002) == 0x00000002);
@ -549,19 +549,19 @@ public final class AuthenticationProtos {
boolean hasUsername();
com.google.protobuf.ByteString getUsername();
// required int32 keyId = 3;
// required int32 key_id = 3;
boolean hasKeyId();
int getKeyId();
// optional int64 issueDate = 4;
// optional int64 issue_date = 4;
boolean hasIssueDate();
long getIssueDate();
// optional int64 expirationDate = 5;
// optional int64 expiration_date = 5;
boolean hasExpirationDate();
long getExpirationDate();
// optional int64 sequenceNumber = 6;
// optional int64 sequence_number = 6;
boolean hasSequenceNumber();
long getSequenceNumber();
}
@ -680,8 +680,8 @@ public final class AuthenticationProtos {
return username_;
}
// required int32 keyId = 3;
public static final int KEYID_FIELD_NUMBER = 3;
// required int32 key_id = 3;
public static final int KEY_ID_FIELD_NUMBER = 3;
private int keyId_;
public boolean hasKeyId() {
return ((bitField0_ & 0x00000004) == 0x00000004);
@ -690,8 +690,8 @@ public final class AuthenticationProtos {
return keyId_;
}
// optional int64 issueDate = 4;
public static final int ISSUEDATE_FIELD_NUMBER = 4;
// optional int64 issue_date = 4;
public static final int ISSUE_DATE_FIELD_NUMBER = 4;
private long issueDate_;
public boolean hasIssueDate() {
return ((bitField0_ & 0x00000008) == 0x00000008);
@ -700,8 +700,8 @@ public final class AuthenticationProtos {
return issueDate_;
}
// optional int64 expirationDate = 5;
public static final int EXPIRATIONDATE_FIELD_NUMBER = 5;
// optional int64 expiration_date = 5;
public static final int EXPIRATION_DATE_FIELD_NUMBER = 5;
private long expirationDate_;
public boolean hasExpirationDate() {
return ((bitField0_ & 0x00000010) == 0x00000010);
@ -710,8 +710,8 @@ public final class AuthenticationProtos {
return expirationDate_;
}
// optional int64 sequenceNumber = 6;
public static final int SEQUENCENUMBER_FIELD_NUMBER = 6;
// optional int64 sequence_number = 6;
public static final int SEQUENCE_NUMBER_FIELD_NUMBER = 6;
private long sequenceNumber_;
public boolean hasSequenceNumber() {
return ((bitField0_ & 0x00000020) == 0x00000020);
@ -874,19 +874,19 @@ public final class AuthenticationProtos {
hash = (53 * hash) + getUsername().hashCode();
}
if (hasKeyId()) {
hash = (37 * hash) + KEYID_FIELD_NUMBER;
hash = (37 * hash) + KEY_ID_FIELD_NUMBER;
hash = (53 * hash) + getKeyId();
}
if (hasIssueDate()) {
hash = (37 * hash) + ISSUEDATE_FIELD_NUMBER;
hash = (37 * hash) + ISSUE_DATE_FIELD_NUMBER;
hash = (53 * hash) + hashLong(getIssueDate());
}
if (hasExpirationDate()) {
hash = (37 * hash) + EXPIRATIONDATE_FIELD_NUMBER;
hash = (37 * hash) + EXPIRATION_DATE_FIELD_NUMBER;
hash = (53 * hash) + hashLong(getExpirationDate());
}
if (hasSequenceNumber()) {
hash = (37 * hash) + SEQUENCENUMBER_FIELD_NUMBER;
hash = (37 * hash) + SEQUENCE_NUMBER_FIELD_NUMBER;
hash = (53 * hash) + hashLong(getSequenceNumber());
}
hash = (29 * hash) + getUnknownFields().hashCode();
@ -1246,7 +1246,7 @@ public final class AuthenticationProtos {
return this;
}
// required int32 keyId = 3;
// required int32 key_id = 3;
private int keyId_ ;
public boolean hasKeyId() {
return ((bitField0_ & 0x00000004) == 0x00000004);
@ -1267,7 +1267,7 @@ public final class AuthenticationProtos {
return this;
}
// optional int64 issueDate = 4;
// optional int64 issue_date = 4;
private long issueDate_ ;
public boolean hasIssueDate() {
return ((bitField0_ & 0x00000008) == 0x00000008);
@ -1288,7 +1288,7 @@ public final class AuthenticationProtos {
return this;
}
// optional int64 expirationDate = 5;
// optional int64 expiration_date = 5;
private long expirationDate_ ;
public boolean hasExpirationDate() {
return ((bitField0_ & 0x00000010) == 0x00000010);
@ -1309,7 +1309,7 @@ public final class AuthenticationProtos {
return this;
}
// optional int64 sequenceNumber = 6;
// optional int64 sequence_number = 6;
private long sequenceNumber_ ;
public boolean hasSequenceNumber() {
return ((bitField0_ & 0x00000020) == 0x00000020);
@ -2917,7 +2917,7 @@ public final class AuthenticationProtos {
boolean hasUsername();
String getUsername();
// optional string authMethod = 2;
// optional string auth_method = 2;
boolean hasAuthMethod();
String getAuthMethod();
}
@ -2982,8 +2982,8 @@ public final class AuthenticationProtos {
}
}
// optional string authMethod = 2;
public static final int AUTHMETHOD_FIELD_NUMBER = 2;
// optional string auth_method = 2;
public static final int AUTH_METHOD_FIELD_NUMBER = 2;
private java.lang.Object authMethod_;
public boolean hasAuthMethod() {
return ((bitField0_ & 0x00000002) == 0x00000002);
@ -3100,7 +3100,7 @@ public final class AuthenticationProtos {
hash = (53 * hash) + getUsername().hashCode();
}
if (hasAuthMethod()) {
hash = (37 * hash) + AUTHMETHOD_FIELD_NUMBER;
hash = (37 * hash) + AUTH_METHOD_FIELD_NUMBER;
hash = (53 * hash) + getAuthMethod().hashCode();
}
hash = (29 * hash) + getUnknownFields().hashCode();
@ -3374,7 +3374,7 @@ public final class AuthenticationProtos {
onChanged();
}
// optional string authMethod = 2;
// optional string auth_method = 2;
private java.lang.Object authMethod_ = "";
public boolean hasAuthMethod() {
return ((bitField0_ & 0x00000002) == 0x00000002);
@ -3431,7 +3431,7 @@ public final class AuthenticationProtos {
org.apache.hadoop.hbase.protobuf.generated.AuthenticationProtos.TokenRequest request,
com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.protobuf.generated.AuthenticationProtos.TokenResponse> done);
public abstract void whoami(
public abstract void whoAmI(
com.google.protobuf.RpcController controller,
org.apache.hadoop.hbase.protobuf.generated.AuthenticationProtos.WhoAmIRequest request,
com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.protobuf.generated.AuthenticationProtos.WhoAmIResponse> done);
@ -3450,11 +3450,11 @@ public final class AuthenticationProtos {
}
@java.lang.Override
public void whoami(
public void whoAmI(
com.google.protobuf.RpcController controller,
org.apache.hadoop.hbase.protobuf.generated.AuthenticationProtos.WhoAmIRequest request,
com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.protobuf.generated.AuthenticationProtos.WhoAmIResponse> done) {
impl.whoami(controller, request, done);
impl.whoAmI(controller, request, done);
}
};
@ -3482,7 +3482,7 @@ public final class AuthenticationProtos {
case 0:
return impl.getAuthenticationToken(controller, (org.apache.hadoop.hbase.protobuf.generated.AuthenticationProtos.TokenRequest)request);
case 1:
return impl.whoami(controller, (org.apache.hadoop.hbase.protobuf.generated.AuthenticationProtos.WhoAmIRequest)request);
return impl.whoAmI(controller, (org.apache.hadoop.hbase.protobuf.generated.AuthenticationProtos.WhoAmIRequest)request);
default:
throw new java.lang.AssertionError("Can't get here.");
}
@ -3532,7 +3532,7 @@ public final class AuthenticationProtos {
org.apache.hadoop.hbase.protobuf.generated.AuthenticationProtos.TokenRequest request,
com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.protobuf.generated.AuthenticationProtos.TokenResponse> done);
public abstract void whoami(
public abstract void whoAmI(
com.google.protobuf.RpcController controller,
org.apache.hadoop.hbase.protobuf.generated.AuthenticationProtos.WhoAmIRequest request,
com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.protobuf.generated.AuthenticationProtos.WhoAmIResponse> done);
@ -3565,7 +3565,7 @@ public final class AuthenticationProtos {
done));
return;
case 1:
this.whoami(controller, (org.apache.hadoop.hbase.protobuf.generated.AuthenticationProtos.WhoAmIRequest)request,
this.whoAmI(controller, (org.apache.hadoop.hbase.protobuf.generated.AuthenticationProtos.WhoAmIRequest)request,
com.google.protobuf.RpcUtil.<org.apache.hadoop.hbase.protobuf.generated.AuthenticationProtos.WhoAmIResponse>specializeCallback(
done));
return;
@ -3641,7 +3641,7 @@ public final class AuthenticationProtos {
org.apache.hadoop.hbase.protobuf.generated.AuthenticationProtos.TokenResponse.getDefaultInstance()));
}
public void whoami(
public void whoAmI(
com.google.protobuf.RpcController controller,
org.apache.hadoop.hbase.protobuf.generated.AuthenticationProtos.WhoAmIRequest request,
com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.protobuf.generated.AuthenticationProtos.WhoAmIResponse> done) {
@ -3668,7 +3668,7 @@ public final class AuthenticationProtos {
org.apache.hadoop.hbase.protobuf.generated.AuthenticationProtos.TokenRequest request)
throws com.google.protobuf.ServiceException;
public org.apache.hadoop.hbase.protobuf.generated.AuthenticationProtos.WhoAmIResponse whoami(
public org.apache.hadoop.hbase.protobuf.generated.AuthenticationProtos.WhoAmIResponse whoAmI(
com.google.protobuf.RpcController controller,
org.apache.hadoop.hbase.protobuf.generated.AuthenticationProtos.WhoAmIRequest request)
throws com.google.protobuf.ServiceException;
@ -3693,7 +3693,7 @@ public final class AuthenticationProtos {
}
public org.apache.hadoop.hbase.protobuf.generated.AuthenticationProtos.WhoAmIResponse whoami(
public org.apache.hadoop.hbase.protobuf.generated.AuthenticationProtos.WhoAmIResponse whoAmI(
com.google.protobuf.RpcController controller,
org.apache.hadoop.hbase.protobuf.generated.AuthenticationProtos.WhoAmIRequest request)
throws com.google.protobuf.ServiceException {
@ -3751,23 +3751,23 @@ public final class AuthenticationProtos {
descriptor;
static {
java.lang.String[] descriptorData = {
"\n\024Authentication.proto\"D\n\021Authentication" +
"Key\022\n\n\002id\030\001 \002(\005\022\026\n\016expirationDate\030\002 \002(\003\022" +
"\013\n\003key\030\003 \002(\014\"\270\001\n\017TokenIdentifier\022#\n\004kind" +
"\030\001 \002(\0162\025.TokenIdentifier.Kind\022\020\n\010usernam" +
"e\030\002 \002(\014\022\r\n\005keyId\030\003 \002(\005\022\021\n\tissueDate\030\004 \001(" +
"\003\022\026\n\016expirationDate\030\005 \001(\003\022\026\n\016sequenceNum" +
"ber\030\006 \001(\003\"\034\n\004Kind\022\024\n\020HBASE_AUTH_TOKEN\020\000\"" +
">\n\005Token\022\022\n\nidentifier\030\001 \001(\014\022\020\n\010password" +
"\030\002 \001(\014\022\017\n\007service\030\003 \001(\014\"\016\n\014TokenRequest\"" +
"&\n\rTokenResponse\022\025\n\005token\030\001 \001(\0132\006.Token\"",
"\017\n\rWhoAmIRequest\"6\n\016WhoAmIResponse\022\020\n\010us" +
"ername\030\001 \001(\t\022\022\n\nauthMethod\030\002 \001(\t2{\n\025Auth" +
"enticationService\0227\n\026getAuthenticationTo" +
"ken\022\r.TokenRequest\032\016.TokenResponse\022)\n\006wh" +
"oami\022\016.WhoAmIRequest\032\017.WhoAmIResponseBJ\n" +
"*org.apache.hadoop.hbase.protobuf.genera" +
"tedB\024AuthenticationProtosH\001\210\001\001\240\001\001"
"\n\024Authentication.proto\"E\n\021Authentication" +
"Key\022\n\n\002id\030\001 \002(\005\022\027\n\017expiration_date\030\002 \002(\003" +
"\022\013\n\003key\030\003 \002(\014\"\274\001\n\017TokenIdentifier\022#\n\004kin" +
"d\030\001 \002(\0162\025.TokenIdentifier.Kind\022\020\n\010userna" +
"me\030\002 \002(\014\022\016\n\006key_id\030\003 \002(\005\022\022\n\nissue_date\030\004" +
" \001(\003\022\027\n\017expiration_date\030\005 \001(\003\022\027\n\017sequenc" +
"e_number\030\006 \001(\003\"\034\n\004Kind\022\024\n\020HBASE_AUTH_TOK" +
"EN\020\000\">\n\005Token\022\022\n\nidentifier\030\001 \001(\014\022\020\n\010pas" +
"sword\030\002 \001(\014\022\017\n\007service\030\003 \001(\014\"\016\n\014TokenReq" +
"uest\"&\n\rTokenResponse\022\025\n\005token\030\001 \001(\0132\006.T",
"oken\"\017\n\rWhoAmIRequest\"7\n\016WhoAmIResponse\022" +
"\020\n\010username\030\001 \001(\t\022\023\n\013auth_method\030\002 \001(\t2{" +
"\n\025AuthenticationService\0227\n\026GetAuthentica" +
"tionToken\022\r.TokenRequest\032\016.TokenResponse" +
"\022)\n\006WhoAmI\022\016.WhoAmIRequest\032\017.WhoAmIRespo" +
"nseBJ\n*org.apache.hadoop.hbase.protobuf." +
"generatedB\024AuthenticationProtosH\001\210\001\001\240\001\001"
};
com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() {

View File

@ -108,7 +108,7 @@ public final class CellProtos {
boolean hasTimestamp();
long getTimestamp();
// optional .CellType cellType = 5;
// optional .CellType cell_type = 5;
boolean hasCellType();
org.apache.hadoop.hbase.protobuf.generated.CellProtos.CellType getCellType();
@ -185,8 +185,8 @@ public final class CellProtos {
return timestamp_;
}
// optional .CellType cellType = 5;
public static final int CELLTYPE_FIELD_NUMBER = 5;
// optional .CellType cell_type = 5;
public static final int CELL_TYPE_FIELD_NUMBER = 5;
private org.apache.hadoop.hbase.protobuf.generated.CellProtos.CellType cellType_;
public boolean hasCellType() {
return ((bitField0_ & 0x00000010) == 0x00000010);
@ -355,7 +355,7 @@ public final class CellProtos {
hash = (53 * hash) + hashLong(getTimestamp());
}
if (hasCellType()) {
hash = (37 * hash) + CELLTYPE_FIELD_NUMBER;
hash = (37 * hash) + CELL_TYPE_FIELD_NUMBER;
hash = (53 * hash) + hashEnum(getCellType());
}
if (hasValue()) {
@ -752,7 +752,7 @@ public final class CellProtos {
return this;
}
// optional .CellType cellType = 5;
// optional .CellType cell_type = 5;
private org.apache.hadoop.hbase.protobuf.generated.CellProtos.CellType cellType_ = org.apache.hadoop.hbase.protobuf.generated.CellProtos.CellType.MINIMUM;
public boolean hasCellType() {
return ((bitField0_ & 0x00000010) == 0x00000010);
@ -830,7 +830,7 @@ public final class CellProtos {
boolean hasTimestamp();
long getTimestamp();
// optional .CellType keyType = 5;
// optional .CellType key_type = 5;
boolean hasKeyType();
org.apache.hadoop.hbase.protobuf.generated.CellProtos.CellType getKeyType();
@ -907,8 +907,8 @@ public final class CellProtos {
return timestamp_;
}
// optional .CellType keyType = 5;
public static final int KEYTYPE_FIELD_NUMBER = 5;
// optional .CellType key_type = 5;
public static final int KEY_TYPE_FIELD_NUMBER = 5;
private org.apache.hadoop.hbase.protobuf.generated.CellProtos.CellType keyType_;
public boolean hasKeyType() {
return ((bitField0_ & 0x00000010) == 0x00000010);
@ -1089,7 +1089,7 @@ public final class CellProtos {
hash = (53 * hash) + hashLong(getTimestamp());
}
if (hasKeyType()) {
hash = (37 * hash) + KEYTYPE_FIELD_NUMBER;
hash = (37 * hash) + KEY_TYPE_FIELD_NUMBER;
hash = (53 * hash) + hashEnum(getKeyType());
}
if (hasValue()) {
@ -1498,7 +1498,7 @@ public final class CellProtos {
return this;
}
// optional .CellType keyType = 5;
// optional .CellType key_type = 5;
private org.apache.hadoop.hbase.protobuf.generated.CellProtos.CellType keyType_ = org.apache.hadoop.hbase.protobuf.generated.CellProtos.CellType.MINIMUM;
public boolean hasKeyType() {
return ((bitField0_ & 0x00000010) == 0x00000010);
@ -1576,17 +1576,17 @@ public final class CellProtos {
descriptor;
static {
java.lang.String[] descriptorData = {
"\n\nCell.proto\"u\n\004Cell\022\013\n\003row\030\001 \001(\014\022\016\n\006fam" +
"\n\nCell.proto\"v\n\004Cell\022\013\n\003row\030\001 \001(\014\022\016\n\006fam" +
"ily\030\002 \001(\014\022\021\n\tqualifier\030\003 \001(\014\022\021\n\ttimestam" +
"p\030\004 \001(\004\022\033\n\010cellType\030\005 \001(\0162\t.CellType\022\r\n\005" +
"value\030\006 \001(\014\"x\n\010KeyValue\022\013\n\003row\030\001 \002(\014\022\016\n\006" +
"family\030\002 \002(\014\022\021\n\tqualifier\030\003 \002(\014\022\021\n\ttimes" +
"tamp\030\004 \001(\004\022\032\n\007keyType\030\005 \001(\0162\t.CellType\022\r" +
"\n\005value\030\006 \001(\014*`\n\010CellType\022\013\n\007MINIMUM\020\000\022\007" +
"\n\003PUT\020\004\022\n\n\006DELETE\020\010\022\021\n\rDELETE_COLUMN\020\014\022\021" +
"\n\rDELETE_FAMILY\020\016\022\014\n\007MAXIMUM\020\377\001B=\n*org.a" +
"pache.hadoop.hbase.protobuf.generatedB\nC",
"ellProtosH\001\240\001\001"
"p\030\004 \001(\004\022\034\n\tcell_type\030\005 \001(\0162\t.CellType\022\r\n" +
"\005value\030\006 \001(\014\"y\n\010KeyValue\022\013\n\003row\030\001 \002(\014\022\016\n" +
"\006family\030\002 \002(\014\022\021\n\tqualifier\030\003 \002(\014\022\021\n\ttime" +
"stamp\030\004 \001(\004\022\033\n\010key_type\030\005 \001(\0162\t.CellType" +
"\022\r\n\005value\030\006 \001(\014*`\n\010CellType\022\013\n\007MINIMUM\020\000" +
"\022\007\n\003PUT\020\004\022\n\n\006DELETE\020\010\022\021\n\rDELETE_COLUMN\020\014" +
"\022\021\n\rDELETE_FAMILY\020\016\022\014\n\007MAXIMUM\020\377\001B=\n*org" +
".apache.hadoop.hbase.protobuf.generatedB",
"\nCellProtosH\001\240\001\001"
};
com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() {

View File

@ -11,7 +11,7 @@ public final class ClusterIdProtos {
public interface ClusterIdOrBuilder
extends com.google.protobuf.MessageOrBuilder {
// required string clusterId = 1;
// required string cluster_id = 1;
boolean hasClusterId();
String getClusterId();
}
@ -44,8 +44,8 @@ public final class ClusterIdProtos {
}
private int bitField0_;
// required string clusterId = 1;
public static final int CLUSTERID_FIELD_NUMBER = 1;
// required string cluster_id = 1;
public static final int CLUSTER_ID_FIELD_NUMBER = 1;
private java.lang.Object clusterId_;
public boolean hasClusterId() {
return ((bitField0_ & 0x00000001) == 0x00000001);
@ -149,7 +149,7 @@ public final class ClusterIdProtos {
int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode();
if (hasClusterId()) {
hash = (37 * hash) + CLUSTERID_FIELD_NUMBER;
hash = (37 * hash) + CLUSTER_ID_FIELD_NUMBER;
hash = (53 * hash) + getClusterId().hashCode();
}
hash = (29 * hash) + getUnknownFields().hashCode();
@ -377,7 +377,7 @@ public final class ClusterIdProtos {
private int bitField0_;
// required string clusterId = 1;
// required string cluster_id = 1;
private java.lang.Object clusterId_ = "";
public boolean hasClusterId() {
return ((bitField0_ & 0x00000001) == 0x00000001);
@ -438,9 +438,9 @@ public final class ClusterIdProtos {
descriptor;
static {
java.lang.String[] descriptorData = {
"\n\017ClusterId.proto\"\036\n\tClusterId\022\021\n\tcluste" +
"rId\030\001 \002(\tBB\n*org.apache.hadoop.hbase.pro" +
"tobuf.generatedB\017ClusterIdProtosH\001\240\001\001"
"\n\017ClusterId.proto\"\037\n\tClusterId\022\022\n\ncluste" +
"r_id\030\001 \002(\tBB\n*org.apache.hadoop.hbase.pr" +
"otobuf.generatedB\017ClusterIdProtosH\001\240\001\001"
};
com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() {

View File

@ -15,7 +15,7 @@ public final class ComparatorProtos {
boolean hasName();
String getName();
// optional bytes serializedComparator = 2;
// optional bytes serialized_comparator = 2;
boolean hasSerializedComparator();
com.google.protobuf.ByteString getSerializedComparator();
}
@ -80,8 +80,8 @@ public final class ComparatorProtos {
}
}
// optional bytes serializedComparator = 2;
public static final int SERIALIZEDCOMPARATOR_FIELD_NUMBER = 2;
// optional bytes serialized_comparator = 2;
public static final int SERIALIZED_COMPARATOR_FIELD_NUMBER = 2;
private com.google.protobuf.ByteString serializedComparator_;
public boolean hasSerializedComparator() {
return ((bitField0_ & 0x00000002) == 0x00000002);
@ -180,7 +180,7 @@ public final class ComparatorProtos {
hash = (53 * hash) + getName().hashCode();
}
if (hasSerializedComparator()) {
hash = (37 * hash) + SERIALIZEDCOMPARATOR_FIELD_NUMBER;
hash = (37 * hash) + SERIALIZED_COMPARATOR_FIELD_NUMBER;
hash = (53 * hash) + getSerializedComparator().hashCode();
}
hash = (29 * hash) + getUnknownFields().hashCode();
@ -458,7 +458,7 @@ public final class ComparatorProtos {
onChanged();
}
// optional bytes serializedComparator = 2;
// optional bytes serialized_comparator = 2;
private com.google.protobuf.ByteString serializedComparator_ = com.google.protobuf.ByteString.EMPTY;
public boolean hasSerializedComparator() {
return ((bitField0_ & 0x00000002) == 0x00000002);
@ -1805,7 +1805,7 @@ public final class ComparatorProtos {
org.apache.hadoop.hbase.protobuf.generated.ComparatorProtos.ByteArrayComparable getComparable();
org.apache.hadoop.hbase.protobuf.generated.ComparatorProtos.ByteArrayComparableOrBuilder getComparableOrBuilder();
// required .BitComparator.BitwiseOp bitwiseOp = 2;
// required .BitComparator.BitwiseOp bitwise_op = 2;
boolean hasBitwiseOp();
org.apache.hadoop.hbase.protobuf.generated.ComparatorProtos.BitComparator.BitwiseOp getBitwiseOp();
}
@ -1923,8 +1923,8 @@ public final class ComparatorProtos {
return comparable_;
}
// required .BitComparator.BitwiseOp bitwiseOp = 2;
public static final int BITWISEOP_FIELD_NUMBER = 2;
// required .BitComparator.BitwiseOp bitwise_op = 2;
public static final int BITWISE_OP_FIELD_NUMBER = 2;
private org.apache.hadoop.hbase.protobuf.generated.ComparatorProtos.BitComparator.BitwiseOp bitwiseOp_;
public boolean hasBitwiseOp() {
return ((bitField0_ & 0x00000002) == 0x00000002);
@ -2027,7 +2027,7 @@ public final class ComparatorProtos {
hash = (53 * hash) + getComparable().hashCode();
}
if (hasBitwiseOp()) {
hash = (37 * hash) + BITWISEOP_FIELD_NUMBER;
hash = (37 * hash) + BITWISE_OP_FIELD_NUMBER;
hash = (53 * hash) + hashEnum(getBitwiseOp());
}
hash = (29 * hash) + getUnknownFields().hashCode();
@ -2382,7 +2382,7 @@ public final class ComparatorProtos {
return comparableBuilder_;
}
// required .BitComparator.BitwiseOp bitwiseOp = 2;
// required .BitComparator.BitwiseOp bitwise_op = 2;
private org.apache.hadoop.hbase.protobuf.generated.ComparatorProtos.BitComparator.BitwiseOp bitwiseOp_ = org.apache.hadoop.hbase.protobuf.generated.ComparatorProtos.BitComparator.BitwiseOp.AND;
public boolean hasBitwiseOp() {
return ((bitField0_ & 0x00000002) == 0x00000002);
@ -2724,7 +2724,7 @@ public final class ComparatorProtos {
boolean hasPattern();
String getPattern();
// required int32 patternFlags = 2;
// required int32 pattern_flags = 2;
boolean hasPatternFlags();
int getPatternFlags();
@ -2793,8 +2793,8 @@ public final class ComparatorProtos {
}
}
// required int32 patternFlags = 2;
public static final int PATTERNFLAGS_FIELD_NUMBER = 2;
// required int32 pattern_flags = 2;
public static final int PATTERN_FLAGS_FIELD_NUMBER = 2;
private int patternFlags_;
public boolean hasPatternFlags() {
return ((bitField0_ & 0x00000002) == 0x00000002);
@ -2946,7 +2946,7 @@ public final class ComparatorProtos {
hash = (53 * hash) + getPattern().hashCode();
}
if (hasPatternFlags()) {
hash = (37 * hash) + PATTERNFLAGS_FIELD_NUMBER;
hash = (37 * hash) + PATTERN_FLAGS_FIELD_NUMBER;
hash = (53 * hash) + getPatternFlags();
}
if (hasCharset()) {
@ -3250,7 +3250,7 @@ public final class ComparatorProtos {
onChanged();
}
// required int32 patternFlags = 2;
// required int32 pattern_flags = 2;
private int patternFlags_ ;
public boolean hasPatternFlags() {
return ((bitField0_ & 0x00000002) == 0x00000002);
@ -3783,21 +3783,22 @@ public final class ComparatorProtos {
descriptor;
static {
java.lang.String[] descriptorData = {
"\n\020Comparator.proto\"8\n\nComparator\022\014\n\004name" +
"\030\001 \002(\t\022\034\n\024serializedComparator\030\002 \001(\014\"$\n\023" +
"ByteArrayComparable\022\r\n\005value\030\001 \001(\014\"<\n\020Bi" +
"naryComparator\022(\n\ncomparable\030\001 \002(\0132\024.Byt" +
"eArrayComparable\"B\n\026BinaryPrefixComparat" +
"or\022(\n\ncomparable\030\001 \002(\0132\024.ByteArrayCompar" +
"able\"\215\001\n\rBitComparator\022(\n\ncomparable\030\001 \002" +
"(\0132\024.ByteArrayComparable\022+\n\tbitwiseOp\030\002 " +
"\002(\0162\030.BitComparator.BitwiseOp\"%\n\tBitwise" +
"Op\022\007\n\003AND\020\001\022\006\n\002OR\020\002\022\007\n\003XOR\020\003\"\020\n\016NullComp",
"arator\"O\n\025RegexStringComparator\022\017\n\007patte" +
"rn\030\001 \002(\t\022\024\n\014patternFlags\030\002 \002(\005\022\017\n\007charse" +
"t\030\003 \002(\t\"%\n\023SubstringComparator\022\016\n\006substr" +
"\030\001 \002(\tBF\n*org.apache.hadoop.hbase.protob" +
"uf.generatedB\020ComparatorProtosH\001\210\001\001\240\001\001"
"\n\020Comparator.proto\"9\n\nComparator\022\014\n\004name" +
"\030\001 \002(\t\022\035\n\025serialized_comparator\030\002 \001(\014\"$\n" +
"\023ByteArrayComparable\022\r\n\005value\030\001 \001(\014\"<\n\020B" +
"inaryComparator\022(\n\ncomparable\030\001 \002(\0132\024.By" +
"teArrayComparable\"B\n\026BinaryPrefixCompara" +
"tor\022(\n\ncomparable\030\001 \002(\0132\024.ByteArrayCompa" +
"rable\"\216\001\n\rBitComparator\022(\n\ncomparable\030\001 " +
"\002(\0132\024.ByteArrayComparable\022,\n\nbitwise_op\030" +
"\002 \002(\0162\030.BitComparator.BitwiseOp\"%\n\tBitwi" +
"seOp\022\007\n\003AND\020\001\022\006\n\002OR\020\002\022\007\n\003XOR\020\003\"\020\n\016NullCo",
"mparator\"P\n\025RegexStringComparator\022\017\n\007pat" +
"tern\030\001 \002(\t\022\025\n\rpattern_flags\030\002 \002(\005\022\017\n\007cha" +
"rset\030\003 \002(\t\"%\n\023SubstringComparator\022\016\n\006sub" +
"str\030\001 \002(\tBF\n*org.apache.hadoop.hbase.pro" +
"tobuf.generatedB\020ComparatorProtosH\001\210\001\001\240\001" +
"\001"
};
com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() {

View File

@ -11,19 +11,19 @@ public final class ErrorHandlingProtos {
public interface StackTraceElementMessageOrBuilder
extends com.google.protobuf.MessageOrBuilder {
// optional string declaringClass = 1;
// optional string declaring_class = 1;
boolean hasDeclaringClass();
String getDeclaringClass();
// optional string methodName = 2;
// optional string method_name = 2;
boolean hasMethodName();
String getMethodName();
// optional string fileName = 3;
// optional string file_name = 3;
boolean hasFileName();
String getFileName();
// optional int32 lineNumber = 4;
// optional int32 line_number = 4;
boolean hasLineNumber();
int getLineNumber();
}
@ -56,8 +56,8 @@ public final class ErrorHandlingProtos {
}
private int bitField0_;
// optional string declaringClass = 1;
public static final int DECLARINGCLASS_FIELD_NUMBER = 1;
// optional string declaring_class = 1;
public static final int DECLARING_CLASS_FIELD_NUMBER = 1;
private java.lang.Object declaringClass_;
public boolean hasDeclaringClass() {
return ((bitField0_ & 0x00000001) == 0x00000001);
@ -88,8 +88,8 @@ public final class ErrorHandlingProtos {
}
}
// optional string methodName = 2;
public static final int METHODNAME_FIELD_NUMBER = 2;
// optional string method_name = 2;
public static final int METHOD_NAME_FIELD_NUMBER = 2;
private java.lang.Object methodName_;
public boolean hasMethodName() {
return ((bitField0_ & 0x00000002) == 0x00000002);
@ -120,8 +120,8 @@ public final class ErrorHandlingProtos {
}
}
// optional string fileName = 3;
public static final int FILENAME_FIELD_NUMBER = 3;
// optional string file_name = 3;
public static final int FILE_NAME_FIELD_NUMBER = 3;
private java.lang.Object fileName_;
public boolean hasFileName() {
return ((bitField0_ & 0x00000004) == 0x00000004);
@ -152,8 +152,8 @@ public final class ErrorHandlingProtos {
}
}
// optional int32 lineNumber = 4;
public static final int LINENUMBER_FIELD_NUMBER = 4;
// optional int32 line_number = 4;
public static final int LINE_NUMBER_FIELD_NUMBER = 4;
private int lineNumber_;
public boolean hasLineNumber() {
return ((bitField0_ & 0x00000008) == 0x00000008);
@ -270,19 +270,19 @@ public final class ErrorHandlingProtos {
int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode();
if (hasDeclaringClass()) {
hash = (37 * hash) + DECLARINGCLASS_FIELD_NUMBER;
hash = (37 * hash) + DECLARING_CLASS_FIELD_NUMBER;
hash = (53 * hash) + getDeclaringClass().hashCode();
}
if (hasMethodName()) {
hash = (37 * hash) + METHODNAME_FIELD_NUMBER;
hash = (37 * hash) + METHOD_NAME_FIELD_NUMBER;
hash = (53 * hash) + getMethodName().hashCode();
}
if (hasFileName()) {
hash = (37 * hash) + FILENAME_FIELD_NUMBER;
hash = (37 * hash) + FILE_NAME_FIELD_NUMBER;
hash = (53 * hash) + getFileName().hashCode();
}
if (hasLineNumber()) {
hash = (37 * hash) + LINENUMBER_FIELD_NUMBER;
hash = (37 * hash) + LINE_NUMBER_FIELD_NUMBER;
hash = (53 * hash) + getLineNumber();
}
hash = (29 * hash) + getUnknownFields().hashCode();
@ -548,7 +548,7 @@ public final class ErrorHandlingProtos {
private int bitField0_;
// optional string declaringClass = 1;
// optional string declaring_class = 1;
private java.lang.Object declaringClass_ = "";
public boolean hasDeclaringClass() {
return ((bitField0_ & 0x00000001) == 0x00000001);
@ -584,7 +584,7 @@ public final class ErrorHandlingProtos {
onChanged();
}
// optional string methodName = 2;
// optional string method_name = 2;
private java.lang.Object methodName_ = "";
public boolean hasMethodName() {
return ((bitField0_ & 0x00000002) == 0x00000002);
@ -620,7 +620,7 @@ public final class ErrorHandlingProtos {
onChanged();
}
// optional string fileName = 3;
// optional string file_name = 3;
private java.lang.Object fileName_ = "";
public boolean hasFileName() {
return ((bitField0_ & 0x00000004) == 0x00000004);
@ -656,7 +656,7 @@ public final class ErrorHandlingProtos {
onChanged();
}
// optional int32 lineNumber = 4;
// optional int32 line_number = 4;
private int lineNumber_ ;
public boolean hasLineNumber() {
return ((bitField0_ & 0x00000008) == 0x00000008);
@ -691,7 +691,7 @@ public final class ErrorHandlingProtos {
public interface GenericExceptionMessageOrBuilder
extends com.google.protobuf.MessageOrBuilder {
// optional string className = 1;
// optional string class_name = 1;
boolean hasClassName();
String getClassName();
@ -699,7 +699,7 @@ public final class ErrorHandlingProtos {
boolean hasMessage();
String getMessage();
// optional bytes errorInfo = 3;
// optional bytes error_info = 3;
boolean hasErrorInfo();
com.google.protobuf.ByteString getErrorInfo();
@ -742,8 +742,8 @@ public final class ErrorHandlingProtos {
}
private int bitField0_;
// optional string className = 1;
public static final int CLASSNAME_FIELD_NUMBER = 1;
// optional string class_name = 1;
public static final int CLASS_NAME_FIELD_NUMBER = 1;
private java.lang.Object className_;
public boolean hasClassName() {
return ((bitField0_ & 0x00000001) == 0x00000001);
@ -806,8 +806,8 @@ public final class ErrorHandlingProtos {
}
}
// optional bytes errorInfo = 3;
public static final int ERRORINFO_FIELD_NUMBER = 3;
// optional bytes error_info = 3;
public static final int ERROR_INFO_FIELD_NUMBER = 3;
private com.google.protobuf.ByteString errorInfo_;
public boolean hasErrorInfo() {
return ((bitField0_ & 0x00000004) == 0x00000004);
@ -942,7 +942,7 @@ public final class ErrorHandlingProtos {
int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode();
if (hasClassName()) {
hash = (37 * hash) + CLASSNAME_FIELD_NUMBER;
hash = (37 * hash) + CLASS_NAME_FIELD_NUMBER;
hash = (53 * hash) + getClassName().hashCode();
}
if (hasMessage()) {
@ -950,7 +950,7 @@ public final class ErrorHandlingProtos {
hash = (53 * hash) + getMessage().hashCode();
}
if (hasErrorInfo()) {
hash = (37 * hash) + ERRORINFO_FIELD_NUMBER;
hash = (37 * hash) + ERROR_INFO_FIELD_NUMBER;
hash = (53 * hash) + getErrorInfo().hashCode();
}
if (getTraceCount() > 0) {
@ -1254,7 +1254,7 @@ public final class ErrorHandlingProtos {
private int bitField0_;
// optional string className = 1;
// optional string class_name = 1;
private java.lang.Object className_ = "";
public boolean hasClassName() {
return ((bitField0_ & 0x00000001) == 0x00000001);
@ -1326,7 +1326,7 @@ public final class ErrorHandlingProtos {
onChanged();
}
// optional bytes errorInfo = 3;
// optional bytes error_info = 3;
private com.google.protobuf.ByteString errorInfo_ = com.google.protobuf.ByteString.EMPTY;
public boolean hasErrorInfo() {
return ((bitField0_ & 0x00000004) == 0x00000004);
@ -1554,7 +1554,7 @@ public final class ErrorHandlingProtos {
boolean hasSource();
String getSource();
// optional .GenericExceptionMessage genericException = 2;
// optional .GenericExceptionMessage generic_exception = 2;
boolean hasGenericException();
org.apache.hadoop.hbase.protobuf.generated.ErrorHandlingProtos.GenericExceptionMessage getGenericException();
org.apache.hadoop.hbase.protobuf.generated.ErrorHandlingProtos.GenericExceptionMessageOrBuilder getGenericExceptionOrBuilder();
@ -1620,8 +1620,8 @@ public final class ErrorHandlingProtos {
}
}
// optional .GenericExceptionMessage genericException = 2;
public static final int GENERICEXCEPTION_FIELD_NUMBER = 2;
// optional .GenericExceptionMessage generic_exception = 2;
public static final int GENERIC_EXCEPTION_FIELD_NUMBER = 2;
private org.apache.hadoop.hbase.protobuf.generated.ErrorHandlingProtos.GenericExceptionMessage genericException_;
public boolean hasGenericException() {
return ((bitField0_ & 0x00000002) == 0x00000002);
@ -1719,7 +1719,7 @@ public final class ErrorHandlingProtos {
hash = (53 * hash) + getSource().hashCode();
}
if (hasGenericException()) {
hash = (37 * hash) + GENERICEXCEPTION_FIELD_NUMBER;
hash = (37 * hash) + GENERIC_EXCEPTION_FIELD_NUMBER;
hash = (53 * hash) + getGenericException().hashCode();
}
hash = (29 * hash) + getUnknownFields().hashCode();
@ -2006,7 +2006,7 @@ public final class ErrorHandlingProtos {
onChanged();
}
// optional .GenericExceptionMessage genericException = 2;
// optional .GenericExceptionMessage generic_exception = 2;
private org.apache.hadoop.hbase.protobuf.generated.ErrorHandlingProtos.GenericExceptionMessage genericException_ = org.apache.hadoop.hbase.protobuf.generated.ErrorHandlingProtos.GenericExceptionMessage.getDefaultInstance();
private com.google.protobuf.SingleFieldBuilder<
org.apache.hadoop.hbase.protobuf.generated.ErrorHandlingProtos.GenericExceptionMessage, org.apache.hadoop.hbase.protobuf.generated.ErrorHandlingProtos.GenericExceptionMessage.Builder, org.apache.hadoop.hbase.protobuf.generated.ErrorHandlingProtos.GenericExceptionMessageOrBuilder> genericExceptionBuilder_;
@ -2131,17 +2131,17 @@ public final class ErrorHandlingProtos {
descriptor;
static {
java.lang.String[] descriptorData = {
"\n\023ErrorHandling.proto\"l\n\030StackTraceEleme" +
"ntMessage\022\026\n\016declaringClass\030\001 \001(\t\022\022\n\nmet" +
"hodName\030\002 \001(\t\022\020\n\010fileName\030\003 \001(\t\022\022\n\nlineN" +
"umber\030\004 \001(\005\"z\n\027GenericExceptionMessage\022\021" +
"\n\tclassName\030\001 \001(\t\022\017\n\007message\030\002 \001(\t\022\021\n\ter" +
"rorInfo\030\003 \001(\014\022(\n\005trace\030\004 \003(\0132\031.StackTrac" +
"eElementMessage\"]\n\027ForeignExceptionMessa" +
"ge\022\016\n\006source\030\001 \001(\t\0222\n\020genericException\030\002" +
" \001(\0132\030.GenericExceptionMessageBF\n*org.ap" +
"ache.hadoop.hbase.protobuf.generatedB\023Er",
"rorHandlingProtosH\001\240\001\001"
"\n\023ErrorHandling.proto\"p\n\030StackTraceEleme" +
"ntMessage\022\027\n\017declaring_class\030\001 \001(\t\022\023\n\013me" +
"thod_name\030\002 \001(\t\022\021\n\tfile_name\030\003 \001(\t\022\023\n\013li" +
"ne_number\030\004 \001(\005\"|\n\027GenericExceptionMessa" +
"ge\022\022\n\nclass_name\030\001 \001(\t\022\017\n\007message\030\002 \001(\t\022" +
"\022\n\nerror_info\030\003 \001(\014\022(\n\005trace\030\004 \003(\0132\031.Sta" +
"ckTraceElementMessage\"^\n\027ForeignExceptio" +
"nMessage\022\016\n\006source\030\001 \001(\t\0223\n\021generic_exce" +
"ption\030\002 \001(\0132\030.GenericExceptionMessageBF\n" +
"*org.apache.hadoop.hbase.protobuf.genera",
"tedB\023ErrorHandlingProtosH\001\240\001\001"
};
com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() {

View File

@ -15,7 +15,7 @@ public final class FilterProtos {
boolean hasName();
String getName();
// optional bytes serializedFilter = 2;
// optional bytes serialized_filter = 2;
boolean hasSerializedFilter();
com.google.protobuf.ByteString getSerializedFilter();
}
@ -80,8 +80,8 @@ public final class FilterProtos {
}
}
// optional bytes serializedFilter = 2;
public static final int SERIALIZEDFILTER_FIELD_NUMBER = 2;
// optional bytes serialized_filter = 2;
public static final int SERIALIZED_FILTER_FIELD_NUMBER = 2;
private com.google.protobuf.ByteString serializedFilter_;
public boolean hasSerializedFilter() {
return ((bitField0_ & 0x00000002) == 0x00000002);
@ -180,7 +180,7 @@ public final class FilterProtos {
hash = (53 * hash) + getName().hashCode();
}
if (hasSerializedFilter()) {
hash = (37 * hash) + SERIALIZEDFILTER_FIELD_NUMBER;
hash = (37 * hash) + SERIALIZED_FILTER_FIELD_NUMBER;
hash = (53 * hash) + getSerializedFilter().hashCode();
}
hash = (29 * hash) + getUnknownFields().hashCode();
@ -458,7 +458,7 @@ public final class FilterProtos {
onChanged();
}
// optional bytes serializedFilter = 2;
// optional bytes serialized_filter = 2;
private com.google.protobuf.ByteString serializedFilter_ = com.google.protobuf.ByteString.EMPTY;
public boolean hasSerializedFilter() {
return ((bitField0_ & 0x00000002) == 0x00000002);
@ -883,7 +883,7 @@ public final class FilterProtos {
boolean hasOffset();
int getOffset();
// optional bytes columnOffset = 3;
// optional bytes column_offset = 3;
boolean hasColumnOffset();
com.google.protobuf.ByteString getColumnOffset();
}
@ -936,8 +936,8 @@ public final class FilterProtos {
return offset_;
}
// optional bytes columnOffset = 3;
public static final int COLUMNOFFSET_FIELD_NUMBER = 3;
// optional bytes column_offset = 3;
public static final int COLUMN_OFFSET_FIELD_NUMBER = 3;
private com.google.protobuf.ByteString columnOffset_;
public boolean hasColumnOffset() {
return ((bitField0_ & 0x00000004) == 0x00000004);
@ -1053,7 +1053,7 @@ public final class FilterProtos {
hash = (53 * hash) + getOffset();
}
if (hasColumnOffset()) {
hash = (37 * hash) + COLUMNOFFSET_FIELD_NUMBER;
hash = (37 * hash) + COLUMN_OFFSET_FIELD_NUMBER;
hash = (53 * hash) + getColumnOffset().hashCode();
}
hash = (29 * hash) + getUnknownFields().hashCode();
@ -1351,7 +1351,7 @@ public final class FilterProtos {
return this;
}
// optional bytes columnOffset = 3;
// optional bytes column_offset = 3;
private com.google.protobuf.ByteString columnOffset_ = com.google.protobuf.ByteString.EMPTY;
public boolean hasColumnOffset() {
return ((bitField0_ & 0x00000004) == 0x00000004);
@ -1771,19 +1771,19 @@ public final class FilterProtos {
public interface ColumnRangeFilterOrBuilder
extends com.google.protobuf.MessageOrBuilder {
// optional bytes minColumn = 1;
// optional bytes min_column = 1;
boolean hasMinColumn();
com.google.protobuf.ByteString getMinColumn();
// optional bool minColumnInclusive = 2;
// optional bool min_column_inclusive = 2;
boolean hasMinColumnInclusive();
boolean getMinColumnInclusive();
// optional bytes maxColumn = 3;
// optional bytes max_column = 3;
boolean hasMaxColumn();
com.google.protobuf.ByteString getMaxColumn();
// optional bool maxColumnInclusive = 4;
// optional bool max_column_inclusive = 4;
boolean hasMaxColumnInclusive();
boolean getMaxColumnInclusive();
}
@ -1816,8 +1816,8 @@ public final class FilterProtos {
}
private int bitField0_;
// optional bytes minColumn = 1;
public static final int MINCOLUMN_FIELD_NUMBER = 1;
// optional bytes min_column = 1;
public static final int MIN_COLUMN_FIELD_NUMBER = 1;
private com.google.protobuf.ByteString minColumn_;
public boolean hasMinColumn() {
return ((bitField0_ & 0x00000001) == 0x00000001);
@ -1826,8 +1826,8 @@ public final class FilterProtos {
return minColumn_;
}
// optional bool minColumnInclusive = 2;
public static final int MINCOLUMNINCLUSIVE_FIELD_NUMBER = 2;
// optional bool min_column_inclusive = 2;
public static final int MIN_COLUMN_INCLUSIVE_FIELD_NUMBER = 2;
private boolean minColumnInclusive_;
public boolean hasMinColumnInclusive() {
return ((bitField0_ & 0x00000002) == 0x00000002);
@ -1836,8 +1836,8 @@ public final class FilterProtos {
return minColumnInclusive_;
}
// optional bytes maxColumn = 3;
public static final int MAXCOLUMN_FIELD_NUMBER = 3;
// optional bytes max_column = 3;
public static final int MAX_COLUMN_FIELD_NUMBER = 3;
private com.google.protobuf.ByteString maxColumn_;
public boolean hasMaxColumn() {
return ((bitField0_ & 0x00000004) == 0x00000004);
@ -1846,8 +1846,8 @@ public final class FilterProtos {
return maxColumn_;
}
// optional bool maxColumnInclusive = 4;
public static final int MAXCOLUMNINCLUSIVE_FIELD_NUMBER = 4;
// optional bool max_column_inclusive = 4;
public static final int MAX_COLUMN_INCLUSIVE_FIELD_NUMBER = 4;
private boolean maxColumnInclusive_;
public boolean hasMaxColumnInclusive() {
return ((bitField0_ & 0x00000008) == 0x00000008);
@ -1964,19 +1964,19 @@ public final class FilterProtos {
int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode();
if (hasMinColumn()) {
hash = (37 * hash) + MINCOLUMN_FIELD_NUMBER;
hash = (37 * hash) + MIN_COLUMN_FIELD_NUMBER;
hash = (53 * hash) + getMinColumn().hashCode();
}
if (hasMinColumnInclusive()) {
hash = (37 * hash) + MINCOLUMNINCLUSIVE_FIELD_NUMBER;
hash = (37 * hash) + MIN_COLUMN_INCLUSIVE_FIELD_NUMBER;
hash = (53 * hash) + hashBoolean(getMinColumnInclusive());
}
if (hasMaxColumn()) {
hash = (37 * hash) + MAXCOLUMN_FIELD_NUMBER;
hash = (37 * hash) + MAX_COLUMN_FIELD_NUMBER;
hash = (53 * hash) + getMaxColumn().hashCode();
}
if (hasMaxColumnInclusive()) {
hash = (37 * hash) + MAXCOLUMNINCLUSIVE_FIELD_NUMBER;
hash = (37 * hash) + MAX_COLUMN_INCLUSIVE_FIELD_NUMBER;
hash = (53 * hash) + hashBoolean(getMaxColumnInclusive());
}
hash = (29 * hash) + getUnknownFields().hashCode();
@ -2242,7 +2242,7 @@ public final class FilterProtos {
private int bitField0_;
// optional bytes minColumn = 1;
// optional bytes min_column = 1;
private com.google.protobuf.ByteString minColumn_ = com.google.protobuf.ByteString.EMPTY;
public boolean hasMinColumn() {
return ((bitField0_ & 0x00000001) == 0x00000001);
@ -2266,7 +2266,7 @@ public final class FilterProtos {
return this;
}
// optional bool minColumnInclusive = 2;
// optional bool min_column_inclusive = 2;
private boolean minColumnInclusive_ ;
public boolean hasMinColumnInclusive() {
return ((bitField0_ & 0x00000002) == 0x00000002);
@ -2287,7 +2287,7 @@ public final class FilterProtos {
return this;
}
// optional bytes maxColumn = 3;
// optional bytes max_column = 3;
private com.google.protobuf.ByteString maxColumn_ = com.google.protobuf.ByteString.EMPTY;
public boolean hasMaxColumn() {
return ((bitField0_ & 0x00000004) == 0x00000004);
@ -2311,7 +2311,7 @@ public final class FilterProtos {
return this;
}
// optional bool maxColumnInclusive = 4;
// optional bool max_column_inclusive = 4;
private boolean maxColumnInclusive_ ;
public boolean hasMaxColumnInclusive() {
return ((bitField0_ & 0x00000008) == 0x00000008);
@ -2346,7 +2346,7 @@ public final class FilterProtos {
public interface CompareFilterOrBuilder
extends com.google.protobuf.MessageOrBuilder {
// required .CompareType compareOp = 1;
// required .CompareType compare_op = 1;
boolean hasCompareOp();
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.CompareType getCompareOp();
@ -2384,8 +2384,8 @@ public final class FilterProtos {
}
private int bitField0_;
// required .CompareType compareOp = 1;
public static final int COMPAREOP_FIELD_NUMBER = 1;
// required .CompareType compare_op = 1;
public static final int COMPARE_OP_FIELD_NUMBER = 1;
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.CompareType compareOp_;
public boolean hasCompareOp() {
return ((bitField0_ & 0x00000001) == 0x00000001);
@ -2499,7 +2499,7 @@ public final class FilterProtos {
int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode();
if (hasCompareOp()) {
hash = (37 * hash) + COMPAREOP_FIELD_NUMBER;
hash = (37 * hash) + COMPARE_OP_FIELD_NUMBER;
hash = (53 * hash) + hashEnum(getCompareOp());
}
if (hasComparator()) {
@ -2770,7 +2770,7 @@ public final class FilterProtos {
private int bitField0_;
// required .CompareType compareOp = 1;
// required .CompareType compare_op = 1;
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.CompareType compareOp_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.CompareType.LESS;
public boolean hasCompareOp() {
return ((bitField0_ & 0x00000001) == 0x00000001);
@ -2898,20 +2898,20 @@ public final class FilterProtos {
public interface DependentColumnFilterOrBuilder
extends com.google.protobuf.MessageOrBuilder {
// required .CompareFilter compareFilter = 1;
// required .CompareFilter compare_filter = 1;
boolean hasCompareFilter();
org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter getCompareFilter();
org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilterOrBuilder getCompareFilterOrBuilder();
// optional bytes columnFamily = 2;
// optional bytes column_family = 2;
boolean hasColumnFamily();
com.google.protobuf.ByteString getColumnFamily();
// optional bytes columnQualifier = 3;
// optional bytes column_qualifier = 3;
boolean hasColumnQualifier();
com.google.protobuf.ByteString getColumnQualifier();
// optional bool dropDependentColumn = 4;
// optional bool drop_dependent_column = 4;
boolean hasDropDependentColumn();
boolean getDropDependentColumn();
}
@ -2944,8 +2944,8 @@ public final class FilterProtos {
}
private int bitField0_;
// required .CompareFilter compareFilter = 1;
public static final int COMPAREFILTER_FIELD_NUMBER = 1;
// required .CompareFilter compare_filter = 1;
public static final int COMPARE_FILTER_FIELD_NUMBER = 1;
private org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter compareFilter_;
public boolean hasCompareFilter() {
return ((bitField0_ & 0x00000001) == 0x00000001);
@ -2957,8 +2957,8 @@ public final class FilterProtos {
return compareFilter_;
}
// optional bytes columnFamily = 2;
public static final int COLUMNFAMILY_FIELD_NUMBER = 2;
// optional bytes column_family = 2;
public static final int COLUMN_FAMILY_FIELD_NUMBER = 2;
private com.google.protobuf.ByteString columnFamily_;
public boolean hasColumnFamily() {
return ((bitField0_ & 0x00000002) == 0x00000002);
@ -2967,8 +2967,8 @@ public final class FilterProtos {
return columnFamily_;
}
// optional bytes columnQualifier = 3;
public static final int COLUMNQUALIFIER_FIELD_NUMBER = 3;
// optional bytes column_qualifier = 3;
public static final int COLUMN_QUALIFIER_FIELD_NUMBER = 3;
private com.google.protobuf.ByteString columnQualifier_;
public boolean hasColumnQualifier() {
return ((bitField0_ & 0x00000004) == 0x00000004);
@ -2977,8 +2977,8 @@ public final class FilterProtos {
return columnQualifier_;
}
// optional bool dropDependentColumn = 4;
public static final int DROPDEPENDENTCOLUMN_FIELD_NUMBER = 4;
// optional bool drop_dependent_column = 4;
public static final int DROP_DEPENDENT_COLUMN_FIELD_NUMBER = 4;
private boolean dropDependentColumn_;
public boolean hasDropDependentColumn() {
return ((bitField0_ & 0x00000008) == 0x00000008);
@ -3103,19 +3103,19 @@ public final class FilterProtos {
int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode();
if (hasCompareFilter()) {
hash = (37 * hash) + COMPAREFILTER_FIELD_NUMBER;
hash = (37 * hash) + COMPARE_FILTER_FIELD_NUMBER;
hash = (53 * hash) + getCompareFilter().hashCode();
}
if (hasColumnFamily()) {
hash = (37 * hash) + COLUMNFAMILY_FIELD_NUMBER;
hash = (37 * hash) + COLUMN_FAMILY_FIELD_NUMBER;
hash = (53 * hash) + getColumnFamily().hashCode();
}
if (hasColumnQualifier()) {
hash = (37 * hash) + COLUMNQUALIFIER_FIELD_NUMBER;
hash = (37 * hash) + COLUMN_QUALIFIER_FIELD_NUMBER;
hash = (53 * hash) + getColumnQualifier().hashCode();
}
if (hasDropDependentColumn()) {
hash = (37 * hash) + DROPDEPENDENTCOLUMN_FIELD_NUMBER;
hash = (37 * hash) + DROP_DEPENDENT_COLUMN_FIELD_NUMBER;
hash = (53 * hash) + hashBoolean(getDropDependentColumn());
}
hash = (29 * hash) + getUnknownFields().hashCode();
@ -3402,7 +3402,7 @@ public final class FilterProtos {
private int bitField0_;
// required .CompareFilter compareFilter = 1;
// required .CompareFilter compare_filter = 1;
private org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter compareFilter_ = org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter.getDefaultInstance();
private com.google.protobuf.SingleFieldBuilder<
org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter, org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter.Builder, org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilterOrBuilder> compareFilterBuilder_;
@ -3492,7 +3492,7 @@ public final class FilterProtos {
return compareFilterBuilder_;
}
// optional bytes columnFamily = 2;
// optional bytes column_family = 2;
private com.google.protobuf.ByteString columnFamily_ = com.google.protobuf.ByteString.EMPTY;
public boolean hasColumnFamily() {
return ((bitField0_ & 0x00000002) == 0x00000002);
@ -3516,7 +3516,7 @@ public final class FilterProtos {
return this;
}
// optional bytes columnQualifier = 3;
// optional bytes column_qualifier = 3;
private com.google.protobuf.ByteString columnQualifier_ = com.google.protobuf.ByteString.EMPTY;
public boolean hasColumnQualifier() {
return ((bitField0_ & 0x00000004) == 0x00000004);
@ -3540,7 +3540,7 @@ public final class FilterProtos {
return this;
}
// optional bool dropDependentColumn = 4;
// optional bool drop_dependent_column = 4;
private boolean dropDependentColumn_ ;
public boolean hasDropDependentColumn() {
return ((bitField0_ & 0x00000008) == 0x00000008);
@ -3575,7 +3575,7 @@ public final class FilterProtos {
public interface FamilyFilterOrBuilder
extends com.google.protobuf.MessageOrBuilder {
// required .CompareFilter compareFilter = 1;
// required .CompareFilter compare_filter = 1;
boolean hasCompareFilter();
org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter getCompareFilter();
org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilterOrBuilder getCompareFilterOrBuilder();
@ -3609,8 +3609,8 @@ public final class FilterProtos {
}
private int bitField0_;
// required .CompareFilter compareFilter = 1;
public static final int COMPAREFILTER_FIELD_NUMBER = 1;
// required .CompareFilter compare_filter = 1;
public static final int COMPARE_FILTER_FIELD_NUMBER = 1;
private org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter compareFilter_;
public boolean hasCompareFilter() {
return ((bitField0_ & 0x00000001) == 0x00000001);
@ -3699,7 +3699,7 @@ public final class FilterProtos {
int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode();
if (hasCompareFilter()) {
hash = (37 * hash) + COMPAREFILTER_FIELD_NUMBER;
hash = (37 * hash) + COMPARE_FILTER_FIELD_NUMBER;
hash = (53 * hash) + getCompareFilter().hashCode();
}
hash = (29 * hash) + getUnknownFields().hashCode();
@ -3944,7 +3944,7 @@ public final class FilterProtos {
private int bitField0_;
// required .CompareFilter compareFilter = 1;
// required .CompareFilter compare_filter = 1;
private org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter compareFilter_ = org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter.getDefaultInstance();
private com.google.protobuf.SingleFieldBuilder<
org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter, org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter.Builder, org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilterOrBuilder> compareFilterBuilder_;
@ -5982,7 +5982,7 @@ public final class FilterProtos {
public interface FuzzyRowFilterOrBuilder
extends com.google.protobuf.MessageOrBuilder {
// repeated .BytesBytesPair fuzzyKeysData = 1;
// repeated .BytesBytesPair fuzzy_keys_data = 1;
java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair>
getFuzzyKeysDataList();
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair getFuzzyKeysData(int index);
@ -6020,8 +6020,8 @@ public final class FilterProtos {
return org.apache.hadoop.hbase.protobuf.generated.FilterProtos.internal_static_FuzzyRowFilter_fieldAccessorTable;
}
// repeated .BytesBytesPair fuzzyKeysData = 1;
public static final int FUZZYKEYSDATA_FIELD_NUMBER = 1;
// repeated .BytesBytesPair fuzzy_keys_data = 1;
public static final int FUZZY_KEYS_DATA_FIELD_NUMBER = 1;
private java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair> fuzzyKeysData_;
public java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair> getFuzzyKeysDataList() {
return fuzzyKeysData_;
@ -6113,7 +6113,7 @@ public final class FilterProtos {
int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode();
if (getFuzzyKeysDataCount() > 0) {
hash = (37 * hash) + FUZZYKEYSDATA_FIELD_NUMBER;
hash = (37 * hash) + FUZZY_KEYS_DATA_FIELD_NUMBER;
hash = (53 * hash) + getFuzzyKeysDataList().hashCode();
}
hash = (29 * hash) + getUnknownFields().hashCode();
@ -6375,7 +6375,7 @@ public final class FilterProtos {
private int bitField0_;
// repeated .BytesBytesPair fuzzyKeysData = 1;
// repeated .BytesBytesPair fuzzy_keys_data = 1;
private java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair> fuzzyKeysData_ =
java.util.Collections.emptyList();
private void ensureFuzzyKeysDataIsMutable() {
@ -6575,7 +6575,7 @@ public final class FilterProtos {
public interface InclusiveStopFilterOrBuilder
extends com.google.protobuf.MessageOrBuilder {
// optional bytes stopRowKey = 1;
// optional bytes stop_row_key = 1;
boolean hasStopRowKey();
com.google.protobuf.ByteString getStopRowKey();
}
@ -6608,8 +6608,8 @@ public final class FilterProtos {
}
private int bitField0_;
// optional bytes stopRowKey = 1;
public static final int STOPROWKEY_FIELD_NUMBER = 1;
// optional bytes stop_row_key = 1;
public static final int STOP_ROW_KEY_FIELD_NUMBER = 1;
private com.google.protobuf.ByteString stopRowKey_;
public boolean hasStopRowKey() {
return ((bitField0_ & 0x00000001) == 0x00000001);
@ -6687,7 +6687,7 @@ public final class FilterProtos {
int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode();
if (hasStopRowKey()) {
hash = (37 * hash) + STOPROWKEY_FIELD_NUMBER;
hash = (37 * hash) + STOP_ROW_KEY_FIELD_NUMBER;
hash = (53 * hash) + getStopRowKey().hashCode();
}
hash = (29 * hash) + getUnknownFields().hashCode();
@ -6911,7 +6911,7 @@ public final class FilterProtos {
private int bitField0_;
// optional bytes stopRowKey = 1;
// optional bytes stop_row_key = 1;
private com.google.protobuf.ByteString stopRowKey_ = com.google.protobuf.ByteString.EMPTY;
public boolean hasStopRowKey() {
return ((bitField0_ & 0x00000001) == 0x00000001);
@ -6949,7 +6949,7 @@ public final class FilterProtos {
public interface KeyOnlyFilterOrBuilder
extends com.google.protobuf.MessageOrBuilder {
// required bool lenAsVal = 1;
// required bool len_as_val = 1;
boolean hasLenAsVal();
boolean getLenAsVal();
}
@ -6982,8 +6982,8 @@ public final class FilterProtos {
}
private int bitField0_;
// required bool lenAsVal = 1;
public static final int LENASVAL_FIELD_NUMBER = 1;
// required bool len_as_val = 1;
public static final int LEN_AS_VAL_FIELD_NUMBER = 1;
private boolean lenAsVal_;
public boolean hasLenAsVal() {
return ((bitField0_ & 0x00000001) == 0x00000001);
@ -7065,7 +7065,7 @@ public final class FilterProtos {
int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode();
if (hasLenAsVal()) {
hash = (37 * hash) + LENASVAL_FIELD_NUMBER;
hash = (37 * hash) + LEN_AS_VAL_FIELD_NUMBER;
hash = (53 * hash) + hashBoolean(getLenAsVal());
}
hash = (29 * hash) + getUnknownFields().hashCode();
@ -7293,7 +7293,7 @@ public final class FilterProtos {
private int bitField0_;
// required bool lenAsVal = 1;
// required bool len_as_val = 1;
private boolean lenAsVal_ ;
public boolean hasLenAsVal() {
return ((bitField0_ & 0x00000001) == 0x00000001);
@ -7328,7 +7328,7 @@ public final class FilterProtos {
public interface MultipleColumnPrefixFilterOrBuilder
extends com.google.protobuf.MessageOrBuilder {
// repeated bytes sortedPrefixes = 1;
// repeated bytes sorted_prefixes = 1;
java.util.List<com.google.protobuf.ByteString> getSortedPrefixesList();
int getSortedPrefixesCount();
com.google.protobuf.ByteString getSortedPrefixes(int index);
@ -7361,8 +7361,8 @@ public final class FilterProtos {
return org.apache.hadoop.hbase.protobuf.generated.FilterProtos.internal_static_MultipleColumnPrefixFilter_fieldAccessorTable;
}
// repeated bytes sortedPrefixes = 1;
public static final int SORTEDPREFIXES_FIELD_NUMBER = 1;
// repeated bytes sorted_prefixes = 1;
public static final int SORTED_PREFIXES_FIELD_NUMBER = 1;
private java.util.List<com.google.protobuf.ByteString> sortedPrefixes_;
public java.util.List<com.google.protobuf.ByteString>
getSortedPrefixesList() {
@ -7446,7 +7446,7 @@ public final class FilterProtos {
int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode();
if (getSortedPrefixesCount() > 0) {
hash = (37 * hash) + SORTEDPREFIXES_FIELD_NUMBER;
hash = (37 * hash) + SORTED_PREFIXES_FIELD_NUMBER;
hash = (53 * hash) + getSortedPrefixesList().hashCode();
}
hash = (29 * hash) + getUnknownFields().hashCode();
@ -7676,7 +7676,7 @@ public final class FilterProtos {
private int bitField0_;
// repeated bytes sortedPrefixes = 1;
// repeated bytes sorted_prefixes = 1;
private java.util.List<com.google.protobuf.ByteString> sortedPrefixes_ = java.util.Collections.emptyList();;
private void ensureSortedPrefixesIsMutable() {
if (!((bitField0_ & 0x00000001) == 0x00000001)) {
@ -7741,7 +7741,7 @@ public final class FilterProtos {
public interface PageFilterOrBuilder
extends com.google.protobuf.MessageOrBuilder {
// required int64 pageSize = 1;
// required int64 page_size = 1;
boolean hasPageSize();
long getPageSize();
}
@ -7774,8 +7774,8 @@ public final class FilterProtos {
}
private int bitField0_;
// required int64 pageSize = 1;
public static final int PAGESIZE_FIELD_NUMBER = 1;
// required int64 page_size = 1;
public static final int PAGE_SIZE_FIELD_NUMBER = 1;
private long pageSize_;
public boolean hasPageSize() {
return ((bitField0_ & 0x00000001) == 0x00000001);
@ -7857,7 +7857,7 @@ public final class FilterProtos {
int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode();
if (hasPageSize()) {
hash = (37 * hash) + PAGESIZE_FIELD_NUMBER;
hash = (37 * hash) + PAGE_SIZE_FIELD_NUMBER;
hash = (53 * hash) + hashLong(getPageSize());
}
hash = (29 * hash) + getUnknownFields().hashCode();
@ -8085,7 +8085,7 @@ public final class FilterProtos {
private int bitField0_;
// required int64 pageSize = 1;
// required int64 page_size = 1;
private long pageSize_ ;
public boolean hasPageSize() {
return ((bitField0_ & 0x00000001) == 0x00000001);
@ -8494,7 +8494,7 @@ public final class FilterProtos {
public interface QualifierFilterOrBuilder
extends com.google.protobuf.MessageOrBuilder {
// required .CompareFilter compareFilter = 1;
// required .CompareFilter compare_filter = 1;
boolean hasCompareFilter();
org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter getCompareFilter();
org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilterOrBuilder getCompareFilterOrBuilder();
@ -8528,8 +8528,8 @@ public final class FilterProtos {
}
private int bitField0_;
// required .CompareFilter compareFilter = 1;
public static final int COMPAREFILTER_FIELD_NUMBER = 1;
// required .CompareFilter compare_filter = 1;
public static final int COMPARE_FILTER_FIELD_NUMBER = 1;
private org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter compareFilter_;
public boolean hasCompareFilter() {
return ((bitField0_ & 0x00000001) == 0x00000001);
@ -8618,7 +8618,7 @@ public final class FilterProtos {
int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode();
if (hasCompareFilter()) {
hash = (37 * hash) + COMPAREFILTER_FIELD_NUMBER;
hash = (37 * hash) + COMPARE_FILTER_FIELD_NUMBER;
hash = (53 * hash) + getCompareFilter().hashCode();
}
hash = (29 * hash) + getUnknownFields().hashCode();
@ -8863,7 +8863,7 @@ public final class FilterProtos {
private int bitField0_;
// required .CompareFilter compareFilter = 1;
// required .CompareFilter compare_filter = 1;
private org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter compareFilter_ = org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter.getDefaultInstance();
private com.google.protobuf.SingleFieldBuilder<
org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter, org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter.Builder, org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilterOrBuilder> compareFilterBuilder_;
@ -9346,7 +9346,7 @@ public final class FilterProtos {
public interface RowFilterOrBuilder
extends com.google.protobuf.MessageOrBuilder {
// required .CompareFilter compareFilter = 1;
// required .CompareFilter compare_filter = 1;
boolean hasCompareFilter();
org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter getCompareFilter();
org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilterOrBuilder getCompareFilterOrBuilder();
@ -9380,8 +9380,8 @@ public final class FilterProtos {
}
private int bitField0_;
// required .CompareFilter compareFilter = 1;
public static final int COMPAREFILTER_FIELD_NUMBER = 1;
// required .CompareFilter compare_filter = 1;
public static final int COMPARE_FILTER_FIELD_NUMBER = 1;
private org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter compareFilter_;
public boolean hasCompareFilter() {
return ((bitField0_ & 0x00000001) == 0x00000001);
@ -9470,7 +9470,7 @@ public final class FilterProtos {
int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode();
if (hasCompareFilter()) {
hash = (37 * hash) + COMPAREFILTER_FIELD_NUMBER;
hash = (37 * hash) + COMPARE_FILTER_FIELD_NUMBER;
hash = (53 * hash) + getCompareFilter().hashCode();
}
hash = (29 * hash) + getUnknownFields().hashCode();
@ -9715,7 +9715,7 @@ public final class FilterProtos {
private int bitField0_;
// required .CompareFilter compareFilter = 1;
// required .CompareFilter compare_filter = 1;
private org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter compareFilter_ = org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter.getDefaultInstance();
private com.google.protobuf.SingleFieldBuilder<
org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter, org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter.Builder, org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilterOrBuilder> compareFilterBuilder_;
@ -9819,7 +9819,7 @@ public final class FilterProtos {
public interface SingleColumnValueExcludeFilterOrBuilder
extends com.google.protobuf.MessageOrBuilder {
// required .SingleColumnValueFilter singleColumnValueFilter = 1;
// required .SingleColumnValueFilter single_column_value_filter = 1;
boolean hasSingleColumnValueFilter();
org.apache.hadoop.hbase.protobuf.generated.FilterProtos.SingleColumnValueFilter getSingleColumnValueFilter();
org.apache.hadoop.hbase.protobuf.generated.FilterProtos.SingleColumnValueFilterOrBuilder getSingleColumnValueFilterOrBuilder();
@ -9853,8 +9853,8 @@ public final class FilterProtos {
}
private int bitField0_;
// required .SingleColumnValueFilter singleColumnValueFilter = 1;
public static final int SINGLECOLUMNVALUEFILTER_FIELD_NUMBER = 1;
// required .SingleColumnValueFilter single_column_value_filter = 1;
public static final int SINGLE_COLUMN_VALUE_FILTER_FIELD_NUMBER = 1;
private org.apache.hadoop.hbase.protobuf.generated.FilterProtos.SingleColumnValueFilter singleColumnValueFilter_;
public boolean hasSingleColumnValueFilter() {
return ((bitField0_ & 0x00000001) == 0x00000001);
@ -9943,7 +9943,7 @@ public final class FilterProtos {
int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode();
if (hasSingleColumnValueFilter()) {
hash = (37 * hash) + SINGLECOLUMNVALUEFILTER_FIELD_NUMBER;
hash = (37 * hash) + SINGLE_COLUMN_VALUE_FILTER_FIELD_NUMBER;
hash = (53 * hash) + getSingleColumnValueFilter().hashCode();
}
hash = (29 * hash) + getUnknownFields().hashCode();
@ -10188,7 +10188,7 @@ public final class FilterProtos {
private int bitField0_;
// required .SingleColumnValueFilter singleColumnValueFilter = 1;
// required .SingleColumnValueFilter single_column_value_filter = 1;
private org.apache.hadoop.hbase.protobuf.generated.FilterProtos.SingleColumnValueFilter singleColumnValueFilter_ = org.apache.hadoop.hbase.protobuf.generated.FilterProtos.SingleColumnValueFilter.getDefaultInstance();
private com.google.protobuf.SingleFieldBuilder<
org.apache.hadoop.hbase.protobuf.generated.FilterProtos.SingleColumnValueFilter, org.apache.hadoop.hbase.protobuf.generated.FilterProtos.SingleColumnValueFilter.Builder, org.apache.hadoop.hbase.protobuf.generated.FilterProtos.SingleColumnValueFilterOrBuilder> singleColumnValueFilterBuilder_;
@ -10292,15 +10292,15 @@ public final class FilterProtos {
public interface SingleColumnValueFilterOrBuilder
extends com.google.protobuf.MessageOrBuilder {
// optional bytes columnFamily = 1;
// optional bytes column_family = 1;
boolean hasColumnFamily();
com.google.protobuf.ByteString getColumnFamily();
// optional bytes columnQualifier = 2;
// optional bytes column_qualifier = 2;
boolean hasColumnQualifier();
com.google.protobuf.ByteString getColumnQualifier();
// required .CompareType compareOp = 3;
// required .CompareType compare_op = 3;
boolean hasCompareOp();
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.CompareType getCompareOp();
@ -10309,11 +10309,11 @@ public final class FilterProtos {
org.apache.hadoop.hbase.protobuf.generated.ComparatorProtos.Comparator getComparator();
org.apache.hadoop.hbase.protobuf.generated.ComparatorProtos.ComparatorOrBuilder getComparatorOrBuilder();
// optional bool filterIfMissing = 5;
// optional bool filter_if_missing = 5;
boolean hasFilterIfMissing();
boolean getFilterIfMissing();
// optional bool latestVersionOnly = 6;
// optional bool latest_version_only = 6;
boolean hasLatestVersionOnly();
boolean getLatestVersionOnly();
}
@ -10346,8 +10346,8 @@ public final class FilterProtos {
}
private int bitField0_;
// optional bytes columnFamily = 1;
public static final int COLUMNFAMILY_FIELD_NUMBER = 1;
// optional bytes column_family = 1;
public static final int COLUMN_FAMILY_FIELD_NUMBER = 1;
private com.google.protobuf.ByteString columnFamily_;
public boolean hasColumnFamily() {
return ((bitField0_ & 0x00000001) == 0x00000001);
@ -10356,8 +10356,8 @@ public final class FilterProtos {
return columnFamily_;
}
// optional bytes columnQualifier = 2;
public static final int COLUMNQUALIFIER_FIELD_NUMBER = 2;
// optional bytes column_qualifier = 2;
public static final int COLUMN_QUALIFIER_FIELD_NUMBER = 2;
private com.google.protobuf.ByteString columnQualifier_;
public boolean hasColumnQualifier() {
return ((bitField0_ & 0x00000002) == 0x00000002);
@ -10366,8 +10366,8 @@ public final class FilterProtos {
return columnQualifier_;
}
// required .CompareType compareOp = 3;
public static final int COMPAREOP_FIELD_NUMBER = 3;
// required .CompareType compare_op = 3;
public static final int COMPARE_OP_FIELD_NUMBER = 3;
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.CompareType compareOp_;
public boolean hasCompareOp() {
return ((bitField0_ & 0x00000004) == 0x00000004);
@ -10389,8 +10389,8 @@ public final class FilterProtos {
return comparator_;
}
// optional bool filterIfMissing = 5;
public static final int FILTERIFMISSING_FIELD_NUMBER = 5;
// optional bool filter_if_missing = 5;
public static final int FILTER_IF_MISSING_FIELD_NUMBER = 5;
private boolean filterIfMissing_;
public boolean hasFilterIfMissing() {
return ((bitField0_ & 0x00000010) == 0x00000010);
@ -10399,8 +10399,8 @@ public final class FilterProtos {
return filterIfMissing_;
}
// optional bool latestVersionOnly = 6;
public static final int LATESTVERSIONONLY_FIELD_NUMBER = 6;
// optional bool latest_version_only = 6;
public static final int LATEST_VERSION_ONLY_FIELD_NUMBER = 6;
private boolean latestVersionOnly_;
public boolean hasLatestVersionOnly() {
return ((bitField0_ & 0x00000020) == 0x00000020);
@ -10555,15 +10555,15 @@ public final class FilterProtos {
int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode();
if (hasColumnFamily()) {
hash = (37 * hash) + COLUMNFAMILY_FIELD_NUMBER;
hash = (37 * hash) + COLUMN_FAMILY_FIELD_NUMBER;
hash = (53 * hash) + getColumnFamily().hashCode();
}
if (hasColumnQualifier()) {
hash = (37 * hash) + COLUMNQUALIFIER_FIELD_NUMBER;
hash = (37 * hash) + COLUMN_QUALIFIER_FIELD_NUMBER;
hash = (53 * hash) + getColumnQualifier().hashCode();
}
if (hasCompareOp()) {
hash = (37 * hash) + COMPAREOP_FIELD_NUMBER;
hash = (37 * hash) + COMPARE_OP_FIELD_NUMBER;
hash = (53 * hash) + hashEnum(getCompareOp());
}
if (hasComparator()) {
@ -10571,11 +10571,11 @@ public final class FilterProtos {
hash = (53 * hash) + getComparator().hashCode();
}
if (hasFilterIfMissing()) {
hash = (37 * hash) + FILTERIFMISSING_FIELD_NUMBER;
hash = (37 * hash) + FILTER_IF_MISSING_FIELD_NUMBER;
hash = (53 * hash) + hashBoolean(getFilterIfMissing());
}
if (hasLatestVersionOnly()) {
hash = (37 * hash) + LATESTVERSIONONLY_FIELD_NUMBER;
hash = (37 * hash) + LATEST_VERSION_ONLY_FIELD_NUMBER;
hash = (53 * hash) + hashBoolean(getLatestVersionOnly());
}
hash = (29 * hash) + getUnknownFields().hashCode();
@ -10900,7 +10900,7 @@ public final class FilterProtos {
private int bitField0_;
// optional bytes columnFamily = 1;
// optional bytes column_family = 1;
private com.google.protobuf.ByteString columnFamily_ = com.google.protobuf.ByteString.EMPTY;
public boolean hasColumnFamily() {
return ((bitField0_ & 0x00000001) == 0x00000001);
@ -10924,7 +10924,7 @@ public final class FilterProtos {
return this;
}
// optional bytes columnQualifier = 2;
// optional bytes column_qualifier = 2;
private com.google.protobuf.ByteString columnQualifier_ = com.google.protobuf.ByteString.EMPTY;
public boolean hasColumnQualifier() {
return ((bitField0_ & 0x00000002) == 0x00000002);
@ -10948,7 +10948,7 @@ public final class FilterProtos {
return this;
}
// required .CompareType compareOp = 3;
// required .CompareType compare_op = 3;
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.CompareType compareOp_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.CompareType.LESS;
public boolean hasCompareOp() {
return ((bitField0_ & 0x00000004) == 0x00000004);
@ -11062,7 +11062,7 @@ public final class FilterProtos {
return comparatorBuilder_;
}
// optional bool filterIfMissing = 5;
// optional bool filter_if_missing = 5;
private boolean filterIfMissing_ ;
public boolean hasFilterIfMissing() {
return ((bitField0_ & 0x00000010) == 0x00000010);
@ -11083,7 +11083,7 @@ public final class FilterProtos {
return this;
}
// optional bool latestVersionOnly = 6;
// optional bool latest_version_only = 6;
private boolean latestVersionOnly_ ;
public boolean hasLatestVersionOnly() {
return ((bitField0_ & 0x00000020) == 0x00000020);
@ -12017,7 +12017,7 @@ public final class FilterProtos {
public interface ValueFilterOrBuilder
extends com.google.protobuf.MessageOrBuilder {
// required .CompareFilter compareFilter = 1;
// required .CompareFilter compare_filter = 1;
boolean hasCompareFilter();
org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter getCompareFilter();
org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilterOrBuilder getCompareFilterOrBuilder();
@ -12051,8 +12051,8 @@ public final class FilterProtos {
}
private int bitField0_;
// required .CompareFilter compareFilter = 1;
public static final int COMPAREFILTER_FIELD_NUMBER = 1;
// required .CompareFilter compare_filter = 1;
public static final int COMPARE_FILTER_FIELD_NUMBER = 1;
private org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter compareFilter_;
public boolean hasCompareFilter() {
return ((bitField0_ & 0x00000001) == 0x00000001);
@ -12141,7 +12141,7 @@ public final class FilterProtos {
int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode();
if (hasCompareFilter()) {
hash = (37 * hash) + COMPAREFILTER_FIELD_NUMBER;
hash = (37 * hash) + COMPARE_FILTER_FIELD_NUMBER;
hash = (53 * hash) + getCompareFilter().hashCode();
}
hash = (29 * hash) + getUnknownFields().hashCode();
@ -12386,7 +12386,7 @@ public final class FilterProtos {
private int bitField0_;
// required .CompareFilter compareFilter = 1;
// required .CompareFilter compare_filter = 1;
private org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter compareFilter_ = org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter.getDefaultInstance();
private com.google.protobuf.SingleFieldBuilder<
org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter, org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter.Builder, org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilterOrBuilder> compareFilterBuilder_;
@ -13105,51 +13105,52 @@ public final class FilterProtos {
static {
java.lang.String[] descriptorData = {
"\n\014Filter.proto\032\013hbase.proto\032\020Comparator." +
"proto\"0\n\006Filter\022\014\n\004name\030\001 \002(\t\022\030\n\020seriali" +
"zedFilter\030\002 \001(\014\"%\n\024ColumnCountGetFilter\022" +
"\r\n\005limit\030\001 \002(\005\"M\n\026ColumnPaginationFilter" +
"\022\r\n\005limit\030\001 \002(\005\022\016\n\006offset\030\002 \001(\005\022\024\n\014colum" +
"nOffset\030\003 \001(\014\"$\n\022ColumnPrefixFilter\022\016\n\006p" +
"refix\030\001 \002(\014\"q\n\021ColumnRangeFilter\022\021\n\tminC" +
"olumn\030\001 \001(\014\022\032\n\022minColumnInclusive\030\002 \001(\010\022" +
"\021\n\tmaxColumn\030\003 \001(\014\022\032\n\022maxColumnInclusive" +
"\030\004 \001(\010\"Q\n\rCompareFilter\022\037\n\tcompareOp\030\001 \002",
"(\0162\014.CompareType\022\037\n\ncomparator\030\002 \001(\0132\013.C" +
"omparator\"\212\001\n\025DependentColumnFilter\022%\n\rc" +
"ompareFilter\030\001 \002(\0132\016.CompareFilter\022\024\n\014co" +
"lumnFamily\030\002 \001(\014\022\027\n\017columnQualifier\030\003 \001(" +
"\014\022\033\n\023dropDependentColumn\030\004 \001(\010\"5\n\014Family" +
"Filter\022%\n\rcompareFilter\030\001 \002(\0132\016.CompareF" +
"ilter\"\200\001\n\nFilterList\022&\n\010operator\030\001 \002(\0162\024" +
".FilterList.Operator\022\030\n\007filters\030\002 \003(\0132\007." +
"Filter\"0\n\010Operator\022\021\n\rMUST_PASS_ALL\020\001\022\021\n" +
"\rMUST_PASS_ONE\020\002\"(\n\rFilterWrapper\022\027\n\006fil",
"ter\030\001 \002(\0132\007.Filter\"\024\n\022FirstKeyOnlyFilter" +
"\";\n%FirstKeyValueMatchingQualifiersFilte" +
"r\022\022\n\nqualifiers\030\001 \003(\014\"8\n\016FuzzyRowFilter\022" +
"&\n\rfuzzyKeysData\030\001 \003(\0132\017.BytesBytesPair\"" +
")\n\023InclusiveStopFilter\022\022\n\nstopRowKey\030\001 \001" +
"(\014\"!\n\rKeyOnlyFilter\022\020\n\010lenAsVal\030\001 \002(\010\"4\n" +
"\032MultipleColumnPrefixFilter\022\026\n\016sortedPre" +
"fixes\030\001 \003(\014\"\036\n\nPageFilter\022\020\n\010pageSize\030\001 " +
"\002(\003\"\036\n\014PrefixFilter\022\016\n\006prefix\030\001 \001(\014\"8\n\017Q" +
"ualifierFilter\022%\n\rcompareFilter\030\001 \002(\0132\016.",
"CompareFilter\"!\n\017RandomRowFilter\022\016\n\006chan" +
"ce\030\001 \002(\002\"2\n\tRowFilter\022%\n\rcompareFilter\030\001" +
" \002(\0132\016.CompareFilter\"[\n\036SingleColumnValu" +
"eExcludeFilter\0229\n\027singleColumnValueFilte" +
"r\030\001 \002(\0132\030.SingleColumnValueFilter\"\276\001\n\027Si" +
"ngleColumnValueFilter\022\024\n\014columnFamily\030\001 " +
"\001(\014\022\027\n\017columnQualifier\030\002 \001(\014\022\037\n\tcompareO" +
"p\030\003 \002(\0162\014.CompareType\022\037\n\ncomparator\030\004 \002(" +
"\0132\013.Comparator\022\027\n\017filterIfMissing\030\005 \001(\010\022" +
"\031\n\021latestVersionOnly\030\006 \001(\010\"%\n\nSkipFilter",
"\022\027\n\006filter\030\001 \002(\0132\007.Filter\"*\n\020TimestampsF" +
"ilter\022\026\n\ntimestamps\030\001 \003(\003B\002\020\001\"4\n\013ValueFi" +
"lter\022%\n\rcompareFilter\030\001 \002(\0132\016.CompareFil" +
"ter\"+\n\020WhileMatchFilter\022\027\n\006filter\030\001 \002(\0132" +
"\007.FilterBB\n*org.apache.hadoop.hbase.prot" +
"obuf.generatedB\014FilterProtosH\001\210\001\001\240\001\001"
"proto\"1\n\006Filter\022\014\n\004name\030\001 \002(\t\022\031\n\021seriali" +
"zed_filter\030\002 \001(\014\"%\n\024ColumnCountGetFilter" +
"\022\r\n\005limit\030\001 \002(\005\"N\n\026ColumnPaginationFilte" +
"r\022\r\n\005limit\030\001 \002(\005\022\016\n\006offset\030\002 \001(\005\022\025\n\rcolu" +
"mn_offset\030\003 \001(\014\"$\n\022ColumnPrefixFilter\022\016\n" +
"\006prefix\030\001 \002(\014\"w\n\021ColumnRangeFilter\022\022\n\nmi" +
"n_column\030\001 \001(\014\022\034\n\024min_column_inclusive\030\002" +
" \001(\010\022\022\n\nmax_column\030\003 \001(\014\022\034\n\024max_column_i" +
"nclusive\030\004 \001(\010\"R\n\rCompareFilter\022 \n\ncompa",
"re_op\030\001 \002(\0162\014.CompareType\022\037\n\ncomparator\030" +
"\002 \001(\0132\013.Comparator\"\217\001\n\025DependentColumnFi" +
"lter\022&\n\016compare_filter\030\001 \002(\0132\016.CompareFi" +
"lter\022\025\n\rcolumn_family\030\002 \001(\014\022\030\n\020column_qu" +
"alifier\030\003 \001(\014\022\035\n\025drop_dependent_column\030\004" +
" \001(\010\"6\n\014FamilyFilter\022&\n\016compare_filter\030\001" +
" \002(\0132\016.CompareFilter\"\200\001\n\nFilterList\022&\n\010o" +
"perator\030\001 \002(\0162\024.FilterList.Operator\022\030\n\007f" +
"ilters\030\002 \003(\0132\007.Filter\"0\n\010Operator\022\021\n\rMUS" +
"T_PASS_ALL\020\001\022\021\n\rMUST_PASS_ONE\020\002\"(\n\rFilte",
"rWrapper\022\027\n\006filter\030\001 \002(\0132\007.Filter\"\024\n\022Fir" +
"stKeyOnlyFilter\";\n%FirstKeyValueMatching" +
"QualifiersFilter\022\022\n\nqualifiers\030\001 \003(\014\":\n\016" +
"FuzzyRowFilter\022(\n\017fuzzy_keys_data\030\001 \003(\0132" +
"\017.BytesBytesPair\"+\n\023InclusiveStopFilter\022" +
"\024\n\014stop_row_key\030\001 \001(\014\"#\n\rKeyOnlyFilter\022\022" +
"\n\nlen_as_val\030\001 \002(\010\"5\n\032MultipleColumnPref" +
"ixFilter\022\027\n\017sorted_prefixes\030\001 \003(\014\"\037\n\nPag" +
"eFilter\022\021\n\tpage_size\030\001 \002(\003\"\036\n\014PrefixFilt" +
"er\022\016\n\006prefix\030\001 \001(\014\"9\n\017QualifierFilter\022&\n",
"\016compare_filter\030\001 \002(\0132\016.CompareFilter\"!\n" +
"\017RandomRowFilter\022\016\n\006chance\030\001 \002(\002\"3\n\tRowF" +
"ilter\022&\n\016compare_filter\030\001 \002(\0132\016.CompareF" +
"ilter\"^\n\036SingleColumnValueExcludeFilter\022" +
"<\n\032single_column_value_filter\030\001 \002(\0132\030.Si" +
"ngleColumnValueFilter\"\305\001\n\027SingleColumnVa" +
"lueFilter\022\025\n\rcolumn_family\030\001 \001(\014\022\030\n\020colu" +
"mn_qualifier\030\002 \001(\014\022 \n\ncompare_op\030\003 \002(\0162\014" +
".CompareType\022\037\n\ncomparator\030\004 \002(\0132\013.Compa" +
"rator\022\031\n\021filter_if_missing\030\005 \001(\010\022\033\n\023late",
"st_version_only\030\006 \001(\010\"%\n\nSkipFilter\022\027\n\006f" +
"ilter\030\001 \002(\0132\007.Filter\"*\n\020TimestampsFilter" +
"\022\026\n\ntimestamps\030\001 \003(\003B\002\020\001\"5\n\013ValueFilter\022" +
"&\n\016compare_filter\030\001 \002(\0132\016.CompareFilter\"" +
"+\n\020WhileMatchFilter\022\027\n\006filter\030\001 \002(\0132\007.Fi" +
"lterBB\n*org.apache.hadoop.hbase.protobuf" +
".generatedB\014FilterProtosH\001\210\001\001\240\001\001"
};
com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() {

View File

@ -109,7 +109,7 @@ public final class HBaseProtos {
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPairOrBuilder getAttributesOrBuilder(
int index);
// repeated .ColumnFamilySchema columnFamilies = 3;
// repeated .ColumnFamilySchema column_families = 3;
java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema>
getColumnFamiliesList();
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema getColumnFamilies(int index);
@ -189,8 +189,8 @@ public final class HBaseProtos {
return attributes_.get(index);
}
// repeated .ColumnFamilySchema columnFamilies = 3;
public static final int COLUMNFAMILIES_FIELD_NUMBER = 3;
// repeated .ColumnFamilySchema column_families = 3;
public static final int COLUMN_FAMILIES_FIELD_NUMBER = 3;
private java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema> columnFamilies_;
public java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema> getColumnFamiliesList() {
return columnFamilies_;
@ -356,7 +356,7 @@ public final class HBaseProtos {
hash = (53 * hash) + getAttributesList().hashCode();
}
if (getColumnFamiliesCount() > 0) {
hash = (37 * hash) + COLUMNFAMILIES_FIELD_NUMBER;
hash = (37 * hash) + COLUMN_FAMILIES_FIELD_NUMBER;
hash = (53 * hash) + getColumnFamiliesList().hashCode();
}
if (getConfigurationCount() > 0) {
@ -956,7 +956,7 @@ public final class HBaseProtos {
return attributesBuilder_;
}
// repeated .ColumnFamilySchema columnFamilies = 3;
// repeated .ColumnFamilySchema column_families = 3;
private java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema> columnFamilies_ =
java.util.Collections.emptyList();
private void ensureColumnFamiliesIsMutable() {
@ -2306,19 +2306,19 @@ public final class HBaseProtos {
public interface RegionInfoOrBuilder
extends com.google.protobuf.MessageOrBuilder {
// required uint64 regionId = 1;
// required uint64 region_id = 1;
boolean hasRegionId();
long getRegionId();
// required bytes tableName = 2;
// required bytes table_name = 2;
boolean hasTableName();
com.google.protobuf.ByteString getTableName();
// optional bytes startKey = 3;
// optional bytes start_key = 3;
boolean hasStartKey();
com.google.protobuf.ByteString getStartKey();
// optional bytes endKey = 4;
// optional bytes end_key = 4;
boolean hasEndKey();
com.google.protobuf.ByteString getEndKey();
@ -2363,8 +2363,8 @@ public final class HBaseProtos {
}
private int bitField0_;
// required uint64 regionId = 1;
public static final int REGIONID_FIELD_NUMBER = 1;
// required uint64 region_id = 1;
public static final int REGION_ID_FIELD_NUMBER = 1;
private long regionId_;
public boolean hasRegionId() {
return ((bitField0_ & 0x00000001) == 0x00000001);
@ -2373,8 +2373,8 @@ public final class HBaseProtos {
return regionId_;
}
// required bytes tableName = 2;
public static final int TABLENAME_FIELD_NUMBER = 2;
// required bytes table_name = 2;
public static final int TABLE_NAME_FIELD_NUMBER = 2;
private com.google.protobuf.ByteString tableName_;
public boolean hasTableName() {
return ((bitField0_ & 0x00000002) == 0x00000002);
@ -2383,8 +2383,8 @@ public final class HBaseProtos {
return tableName_;
}
// optional bytes startKey = 3;
public static final int STARTKEY_FIELD_NUMBER = 3;
// optional bytes start_key = 3;
public static final int START_KEY_FIELD_NUMBER = 3;
private com.google.protobuf.ByteString startKey_;
public boolean hasStartKey() {
return ((bitField0_ & 0x00000004) == 0x00000004);
@ -2393,8 +2393,8 @@ public final class HBaseProtos {
return startKey_;
}
// optional bytes endKey = 4;
public static final int ENDKEY_FIELD_NUMBER = 4;
// optional bytes end_key = 4;
public static final int END_KEY_FIELD_NUMBER = 4;
private com.google.protobuf.ByteString endKey_;
public boolean hasEndKey() {
return ((bitField0_ & 0x00000008) == 0x00000008);
@ -2588,19 +2588,19 @@ public final class HBaseProtos {
int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode();
if (hasRegionId()) {
hash = (37 * hash) + REGIONID_FIELD_NUMBER;
hash = (37 * hash) + REGION_ID_FIELD_NUMBER;
hash = (53 * hash) + hashLong(getRegionId());
}
if (hasTableName()) {
hash = (37 * hash) + TABLENAME_FIELD_NUMBER;
hash = (37 * hash) + TABLE_NAME_FIELD_NUMBER;
hash = (53 * hash) + getTableName().hashCode();
}
if (hasStartKey()) {
hash = (37 * hash) + STARTKEY_FIELD_NUMBER;
hash = (37 * hash) + START_KEY_FIELD_NUMBER;
hash = (53 * hash) + getStartKey().hashCode();
}
if (hasEndKey()) {
hash = (37 * hash) + ENDKEY_FIELD_NUMBER;
hash = (37 * hash) + END_KEY_FIELD_NUMBER;
hash = (53 * hash) + getEndKey().hashCode();
}
if (hasOffline()) {
@ -2928,7 +2928,7 @@ public final class HBaseProtos {
private int bitField0_;
// required uint64 regionId = 1;
// required uint64 region_id = 1;
private long regionId_ ;
public boolean hasRegionId() {
return ((bitField0_ & 0x00000001) == 0x00000001);
@ -2949,7 +2949,7 @@ public final class HBaseProtos {
return this;
}
// required bytes tableName = 2;
// required bytes table_name = 2;
private com.google.protobuf.ByteString tableName_ = com.google.protobuf.ByteString.EMPTY;
public boolean hasTableName() {
return ((bitField0_ & 0x00000002) == 0x00000002);
@ -2973,7 +2973,7 @@ public final class HBaseProtos {
return this;
}
// optional bytes startKey = 3;
// optional bytes start_key = 3;
private com.google.protobuf.ByteString startKey_ = com.google.protobuf.ByteString.EMPTY;
public boolean hasStartKey() {
return ((bitField0_ & 0x00000004) == 0x00000004);
@ -2997,7 +2997,7 @@ public final class HBaseProtos {
return this;
}
// optional bytes endKey = 4;
// optional bytes end_key = 4;
private com.google.protobuf.ByteString endKey_ = com.google.protobuf.ByteString.EMPTY;
public boolean hasEndKey() {
return ((bitField0_ & 0x00000008) == 0x00000008);
@ -3098,7 +3098,7 @@ public final class HBaseProtos {
public interface FavoredNodesOrBuilder
extends com.google.protobuf.MessageOrBuilder {
// repeated .ServerName favoredNode = 1;
// repeated .ServerName favored_node = 1;
java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName>
getFavoredNodeList();
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName getFavoredNode(int index);
@ -3136,8 +3136,8 @@ public final class HBaseProtos {
return org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.internal_static_FavoredNodes_fieldAccessorTable;
}
// repeated .ServerName favoredNode = 1;
public static final int FAVOREDNODE_FIELD_NUMBER = 1;
// repeated .ServerName favored_node = 1;
public static final int FAVORED_NODE_FIELD_NUMBER = 1;
private java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName> favoredNode_;
public java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName> getFavoredNodeList() {
return favoredNode_;
@ -3229,7 +3229,7 @@ public final class HBaseProtos {
int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode();
if (getFavoredNodeCount() > 0) {
hash = (37 * hash) + FAVOREDNODE_FIELD_NUMBER;
hash = (37 * hash) + FAVORED_NODE_FIELD_NUMBER;
hash = (53 * hash) + getFavoredNodeList().hashCode();
}
hash = (29 * hash) + getUnknownFields().hashCode();
@ -3491,7 +3491,7 @@ public final class HBaseProtos {
private int bitField0_;
// repeated .ServerName favoredNode = 1;
// repeated .ServerName favored_node = 1;
private java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName> favoredNode_ =
java.util.Collections.emptyList();
private void ensureFavoredNodeIsMutable() {
@ -4662,7 +4662,7 @@ public final class HBaseProtos {
public interface ServerNameOrBuilder
extends com.google.protobuf.MessageOrBuilder {
// required string hostName = 1;
// required string host_name = 1;
boolean hasHostName();
String getHostName();
@ -4670,7 +4670,7 @@ public final class HBaseProtos {
boolean hasPort();
int getPort();
// optional uint64 startCode = 3;
// optional uint64 start_code = 3;
boolean hasStartCode();
long getStartCode();
}
@ -4703,8 +4703,8 @@ public final class HBaseProtos {
}
private int bitField0_;
// required string hostName = 1;
public static final int HOSTNAME_FIELD_NUMBER = 1;
// required string host_name = 1;
public static final int HOST_NAME_FIELD_NUMBER = 1;
private java.lang.Object hostName_;
public boolean hasHostName() {
return ((bitField0_ & 0x00000001) == 0x00000001);
@ -4745,8 +4745,8 @@ public final class HBaseProtos {
return port_;
}
// optional uint64 startCode = 3;
public static final int STARTCODE_FIELD_NUMBER = 3;
// optional uint64 start_code = 3;
public static final int START_CODE_FIELD_NUMBER = 3;
private long startCode_;
public boolean hasStartCode() {
return ((bitField0_ & 0x00000004) == 0x00000004);
@ -4854,7 +4854,7 @@ public final class HBaseProtos {
int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode();
if (hasHostName()) {
hash = (37 * hash) + HOSTNAME_FIELD_NUMBER;
hash = (37 * hash) + HOST_NAME_FIELD_NUMBER;
hash = (53 * hash) + getHostName().hashCode();
}
if (hasPort()) {
@ -4862,7 +4862,7 @@ public final class HBaseProtos {
hash = (53 * hash) + getPort();
}
if (hasStartCode()) {
hash = (37 * hash) + STARTCODE_FIELD_NUMBER;
hash = (37 * hash) + START_CODE_FIELD_NUMBER;
hash = (53 * hash) + hashLong(getStartCode());
}
hash = (29 * hash) + getUnknownFields().hashCode();
@ -5118,7 +5118,7 @@ public final class HBaseProtos {
private int bitField0_;
// required string hostName = 1;
// required string host_name = 1;
private java.lang.Object hostName_ = "";
public boolean hasHostName() {
return ((bitField0_ & 0x00000001) == 0x00000001);
@ -5175,7 +5175,7 @@ public final class HBaseProtos {
return this;
}
// optional uint64 startCode = 3;
// optional uint64 start_code = 3;
private long startCode_ ;
public boolean hasStartCode() {
return ((bitField0_ & 0x00000004) == 0x00000004);
@ -7579,7 +7579,7 @@ public final class HBaseProtos {
boolean hasTable();
String getTable();
// optional int64 creationTime = 3 [default = 0];
// optional int64 creation_time = 3 [default = 0];
boolean hasCreationTime();
long getCreationTime();
@ -7753,8 +7753,8 @@ public final class HBaseProtos {
}
}
// optional int64 creationTime = 3 [default = 0];
public static final int CREATIONTIME_FIELD_NUMBER = 3;
// optional int64 creation_time = 3 [default = 0];
public static final int CREATION_TIME_FIELD_NUMBER = 3;
private long creationTime_;
public boolean hasCreationTime() {
return ((bitField0_ & 0x00000004) == 0x00000004);
@ -7916,7 +7916,7 @@ public final class HBaseProtos {
hash = (53 * hash) + getTable().hashCode();
}
if (hasCreationTime()) {
hash = (37 * hash) + CREATIONTIME_FIELD_NUMBER;
hash = (37 * hash) + CREATION_TIME_FIELD_NUMBER;
hash = (53 * hash) + hashLong(getCreationTime());
}
if (hasType()) {
@ -8286,7 +8286,7 @@ public final class HBaseProtos {
onChanged();
}
// optional int64 creationTime = 3 [default = 0];
// optional int64 creation_time = 3 [default = 0];
private long creationTime_ ;
public boolean hasCreationTime() {
return ((bitField0_ & 0x00000004) == 0x00000004);
@ -8666,7 +8666,7 @@ public final class HBaseProtos {
public interface LongMsgOrBuilder
extends com.google.protobuf.MessageOrBuilder {
// required int64 longMsg = 1;
// required int64 long_msg = 1;
boolean hasLongMsg();
long getLongMsg();
}
@ -8699,8 +8699,8 @@ public final class HBaseProtos {
}
private int bitField0_;
// required int64 longMsg = 1;
public static final int LONGMSG_FIELD_NUMBER = 1;
// required int64 long_msg = 1;
public static final int LONG_MSG_FIELD_NUMBER = 1;
private long longMsg_;
public boolean hasLongMsg() {
return ((bitField0_ & 0x00000001) == 0x00000001);
@ -8782,7 +8782,7 @@ public final class HBaseProtos {
int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode();
if (hasLongMsg()) {
hash = (37 * hash) + LONGMSG_FIELD_NUMBER;
hash = (37 * hash) + LONG_MSG_FIELD_NUMBER;
hash = (53 * hash) + hashLong(getLongMsg());
}
hash = (29 * hash) + getUnknownFields().hashCode();
@ -9010,7 +9010,7 @@ public final class HBaseProtos {
private int bitField0_;
// required int64 longMsg = 1;
// required int64 long_msg = 1;
private long longMsg_ ;
public boolean hasLongMsg() {
return ((bitField0_ & 0x00000001) == 0x00000001);
@ -9045,7 +9045,7 @@ public final class HBaseProtos {
public interface BigDecimalMsgOrBuilder
extends com.google.protobuf.MessageOrBuilder {
// required bytes bigdecimalMsg = 1;
// required bytes bigdecimal_msg = 1;
boolean hasBigdecimalMsg();
com.google.protobuf.ByteString getBigdecimalMsg();
}
@ -9078,8 +9078,8 @@ public final class HBaseProtos {
}
private int bitField0_;
// required bytes bigdecimalMsg = 1;
public static final int BIGDECIMALMSG_FIELD_NUMBER = 1;
// required bytes bigdecimal_msg = 1;
public static final int BIGDECIMAL_MSG_FIELD_NUMBER = 1;
private com.google.protobuf.ByteString bigdecimalMsg_;
public boolean hasBigdecimalMsg() {
return ((bitField0_ & 0x00000001) == 0x00000001);
@ -9161,7 +9161,7 @@ public final class HBaseProtos {
int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode();
if (hasBigdecimalMsg()) {
hash = (37 * hash) + BIGDECIMALMSG_FIELD_NUMBER;
hash = (37 * hash) + BIGDECIMAL_MSG_FIELD_NUMBER;
hash = (53 * hash) + getBigdecimalMsg().hashCode();
}
hash = (29 * hash) + getUnknownFields().hashCode();
@ -9389,7 +9389,7 @@ public final class HBaseProtos {
private int bitField0_;
// required bytes bigdecimalMsg = 1;
// required bytes bigdecimal_msg = 1;
private com.google.protobuf.ByteString bigdecimalMsg_ = com.google.protobuf.ByteString.EMPTY;
public boolean hasBigdecimalMsg() {
return ((bitField0_ & 0x00000001) == 0x00000001);
@ -9427,11 +9427,11 @@ public final class HBaseProtos {
public interface UUIDOrBuilder
extends com.google.protobuf.MessageOrBuilder {
// required uint64 leastSigBits = 1;
// required uint64 least_sig_bits = 1;
boolean hasLeastSigBits();
long getLeastSigBits();
// required uint64 mostSigBits = 2;
// required uint64 most_sig_bits = 2;
boolean hasMostSigBits();
long getMostSigBits();
}
@ -9464,8 +9464,8 @@ public final class HBaseProtos {
}
private int bitField0_;
// required uint64 leastSigBits = 1;
public static final int LEASTSIGBITS_FIELD_NUMBER = 1;
// required uint64 least_sig_bits = 1;
public static final int LEAST_SIG_BITS_FIELD_NUMBER = 1;
private long leastSigBits_;
public boolean hasLeastSigBits() {
return ((bitField0_ & 0x00000001) == 0x00000001);
@ -9474,8 +9474,8 @@ public final class HBaseProtos {
return leastSigBits_;
}
// required uint64 mostSigBits = 2;
public static final int MOSTSIGBITS_FIELD_NUMBER = 2;
// required uint64 most_sig_bits = 2;
public static final int MOST_SIG_BITS_FIELD_NUMBER = 2;
private long mostSigBits_;
public boolean hasMostSigBits() {
return ((bitField0_ & 0x00000002) == 0x00000002);
@ -9574,11 +9574,11 @@ public final class HBaseProtos {
int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode();
if (hasLeastSigBits()) {
hash = (37 * hash) + LEASTSIGBITS_FIELD_NUMBER;
hash = (37 * hash) + LEAST_SIG_BITS_FIELD_NUMBER;
hash = (53 * hash) + hashLong(getLeastSigBits());
}
if (hasMostSigBits()) {
hash = (37 * hash) + MOSTSIGBITS_FIELD_NUMBER;
hash = (37 * hash) + MOST_SIG_BITS_FIELD_NUMBER;
hash = (53 * hash) + hashLong(getMostSigBits());
}
hash = (29 * hash) + getUnknownFields().hashCode();
@ -9824,7 +9824,7 @@ public final class HBaseProtos {
private int bitField0_;
// required uint64 leastSigBits = 1;
// required uint64 least_sig_bits = 1;
private long leastSigBits_ ;
public boolean hasLeastSigBits() {
return ((bitField0_ & 0x00000001) == 0x00000001);
@ -9845,7 +9845,7 @@ public final class HBaseProtos {
return this;
}
// required uint64 mostSigBits = 2;
// required uint64 most_sig_bits = 2;
private long mostSigBits_ ;
public boolean hasMostSigBits() {
return ((bitField0_ & 0x00000002) == 0x00000002);
@ -9971,42 +9971,43 @@ public final class HBaseProtos {
descriptor;
static {
java.lang.String[] descriptorData = {
"\n\013hbase.proto\032\nCell.proto\"\225\001\n\013TableSchem" +
"\n\013hbase.proto\032\nCell.proto\"\226\001\n\013TableSchem" +
"a\022\014\n\004name\030\001 \001(\014\022#\n\nattributes\030\002 \003(\0132\017.By" +
"tesBytesPair\022+\n\016columnFamilies\030\003 \003(\0132\023.C" +
"olumnFamilySchema\022&\n\rconfiguration\030\004 \003(\013" +
"2\017.NameStringPair\"o\n\022ColumnFamilySchema\022" +
"\014\n\004name\030\001 \002(\014\022#\n\nattributes\030\002 \003(\0132\017.Byte" +
"sBytesPair\022&\n\rconfiguration\030\003 \003(\0132\017.Name" +
"StringPair\"\207\001\n\nRegionInfo\022\020\n\010regionId\030\001 " +
"\002(\004\022\021\n\ttableName\030\002 \002(\014\022\020\n\010startKey\030\003 \001(\014" +
"\022\016\n\006endKey\030\004 \001(\014\022\017\n\007offline\030\005 \001(\010\022\r\n\005spl",
"it\030\006 \001(\010\022\022\n\nrecovering\030\007 \001(\010\"0\n\014FavoredN" +
"odes\022 \n\013favoredNode\030\001 \003(\0132\013.ServerName\"\225" +
"\001\n\017RegionSpecifier\0222\n\004type\030\001 \002(\0162$.Regio" +
"nSpecifier.RegionSpecifierType\022\r\n\005value\030" +
"\002 \002(\014\"?\n\023RegionSpecifierType\022\017\n\013REGION_N" +
"AME\020\001\022\027\n\023ENCODED_REGION_NAME\020\002\"%\n\tTimeRa" +
"nge\022\014\n\004from\030\001 \001(\004\022\n\n\002to\030\002 \001(\004\"?\n\nServerN" +
"ame\022\020\n\010hostName\030\001 \002(\t\022\014\n\004port\030\002 \001(\r\022\021\n\ts" +
"tartCode\030\003 \001(\004\"\033\n\013Coprocessor\022\014\n\004name\030\001 " +
"\002(\t\"-\n\016NameStringPair\022\014\n\004name\030\001 \002(\t\022\r\n\005v",
"alue\030\002 \002(\t\",\n\rNameBytesPair\022\014\n\004name\030\001 \002(" +
"\t\022\r\n\005value\030\002 \001(\014\"/\n\016BytesBytesPair\022\r\n\005fi" +
"rst\030\001 \002(\014\022\016\n\006second\030\002 \002(\014\",\n\rNameInt64Pa" +
"ir\022\014\n\004name\030\001 \001(\t\022\r\n\005value\030\002 \001(\003\"\255\001\n\023Snap" +
"shotDescription\022\014\n\004name\030\001 \002(\t\022\r\n\005table\030\002" +
" \001(\t\022\027\n\014creationTime\030\003 \001(\003:\0010\022.\n\004type\030\004 " +
"\001(\0162\031.SnapshotDescription.Type:\005FLUSH\022\017\n" +
"\007version\030\005 \001(\005\"\037\n\004Type\022\014\n\010DISABLED\020\000\022\t\n\005" +
"FLUSH\020\001\"\n\n\010EmptyMsg\"\032\n\007LongMsg\022\017\n\007longMs" +
"g\030\001 \002(\003\"&\n\rBigDecimalMsg\022\025\n\rbigdecimalMs",
"g\030\001 \002(\014\"1\n\004UUID\022\024\n\014leastSigBits\030\001 \002(\004\022\023\n" +
"\013mostSigBits\030\002 \002(\004*r\n\013CompareType\022\010\n\004LES" +
"S\020\000\022\021\n\rLESS_OR_EQUAL\020\001\022\t\n\005EQUAL\020\002\022\r\n\tNOT" +
"_EQUAL\020\003\022\024\n\020GREATER_OR_EQUAL\020\004\022\013\n\007GREATE" +
"R\020\005\022\t\n\005NO_OP\020\006B>\n*org.apache.hadoop.hbas" +
"e.protobuf.generatedB\013HBaseProtosH\001\240\001\001"
"tesBytesPair\022,\n\017column_families\030\003 \003(\0132\023." +
"ColumnFamilySchema\022&\n\rconfiguration\030\004 \003(" +
"\0132\017.NameStringPair\"o\n\022ColumnFamilySchema" +
"\022\014\n\004name\030\001 \002(\014\022#\n\nattributes\030\002 \003(\0132\017.Byt" +
"esBytesPair\022&\n\rconfiguration\030\003 \003(\0132\017.Nam" +
"eStringPair\"\213\001\n\nRegionInfo\022\021\n\tregion_id\030" +
"\001 \002(\004\022\022\n\ntable_name\030\002 \002(\014\022\021\n\tstart_key\030\003" +
" \001(\014\022\017\n\007end_key\030\004 \001(\014\022\017\n\007offline\030\005 \001(\010\022\r",
"\n\005split\030\006 \001(\010\022\022\n\nrecovering\030\007 \001(\010\"1\n\014Fav" +
"oredNodes\022!\n\014favored_node\030\001 \003(\0132\013.Server" +
"Name\"\225\001\n\017RegionSpecifier\0222\n\004type\030\001 \002(\0162$" +
".RegionSpecifier.RegionSpecifierType\022\r\n\005" +
"value\030\002 \002(\014\"?\n\023RegionSpecifierType\022\017\n\013RE" +
"GION_NAME\020\001\022\027\n\023ENCODED_REGION_NAME\020\002\"%\n\t" +
"TimeRange\022\014\n\004from\030\001 \001(\004\022\n\n\002to\030\002 \001(\004\"A\n\nS" +
"erverName\022\021\n\thost_name\030\001 \002(\t\022\014\n\004port\030\002 \001" +
"(\r\022\022\n\nstart_code\030\003 \001(\004\"\033\n\013Coprocessor\022\014\n" +
"\004name\030\001 \002(\t\"-\n\016NameStringPair\022\014\n\004name\030\001 ",
"\002(\t\022\r\n\005value\030\002 \002(\t\",\n\rNameBytesPair\022\014\n\004n" +
"ame\030\001 \002(\t\022\r\n\005value\030\002 \001(\014\"/\n\016BytesBytesPa" +
"ir\022\r\n\005first\030\001 \002(\014\022\016\n\006second\030\002 \002(\014\",\n\rNam" +
"eInt64Pair\022\014\n\004name\030\001 \001(\t\022\r\n\005value\030\002 \001(\003\"" +
"\256\001\n\023SnapshotDescription\022\014\n\004name\030\001 \002(\t\022\r\n" +
"\005table\030\002 \001(\t\022\030\n\rcreation_time\030\003 \001(\003:\0010\022." +
"\n\004type\030\004 \001(\0162\031.SnapshotDescription.Type:" +
"\005FLUSH\022\017\n\007version\030\005 \001(\005\"\037\n\004Type\022\014\n\010DISAB" +
"LED\020\000\022\t\n\005FLUSH\020\001\"\n\n\010EmptyMsg\"\033\n\007LongMsg\022" +
"\020\n\010long_msg\030\001 \002(\003\"\'\n\rBigDecimalMsg\022\026\n\016bi",
"gdecimal_msg\030\001 \002(\014\"5\n\004UUID\022\026\n\016least_sig_" +
"bits\030\001 \002(\004\022\025\n\rmost_sig_bits\030\002 \002(\004*r\n\013Com" +
"pareType\022\010\n\004LESS\020\000\022\021\n\rLESS_OR_EQUAL\020\001\022\t\n" +
"\005EQUAL\020\002\022\r\n\tNOT_EQUAL\020\003\022\024\n\020GREATER_OR_EQ" +
"UAL\020\004\022\013\n\007GREATER\020\005\022\t\n\005NO_OP\020\006B>\n*org.apa" +
"che.hadoop.hbase.protobuf.generatedB\013HBa" +
"seProtosH\001\240\001\001"
};
com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() {

View File

@ -11,7 +11,7 @@ public final class HFileProtos {
public interface FileInfoProtoOrBuilder
extends com.google.protobuf.MessageOrBuilder {
// repeated .BytesBytesPair mapEntry = 1;
// repeated .BytesBytesPair map_entry = 1;
java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair>
getMapEntryList();
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair getMapEntry(int index);
@ -49,8 +49,8 @@ public final class HFileProtos {
return org.apache.hadoop.hbase.protobuf.generated.HFileProtos.internal_static_FileInfoProto_fieldAccessorTable;
}
// repeated .BytesBytesPair mapEntry = 1;
public static final int MAPENTRY_FIELD_NUMBER = 1;
// repeated .BytesBytesPair map_entry = 1;
public static final int MAP_ENTRY_FIELD_NUMBER = 1;
private java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair> mapEntry_;
public java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair> getMapEntryList() {
return mapEntry_;
@ -142,7 +142,7 @@ public final class HFileProtos {
int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode();
if (getMapEntryCount() > 0) {
hash = (37 * hash) + MAPENTRY_FIELD_NUMBER;
hash = (37 * hash) + MAP_ENTRY_FIELD_NUMBER;
hash = (53 * hash) + getMapEntryList().hashCode();
}
hash = (29 * hash) + getUnknownFields().hashCode();
@ -404,7 +404,7 @@ public final class HFileProtos {
private int bitField0_;
// repeated .BytesBytesPair mapEntry = 1;
// repeated .BytesBytesPair map_entry = 1;
private java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair> mapEntry_ =
java.util.Collections.emptyList();
private void ensureMapEntryIsMutable() {
@ -604,51 +604,51 @@ public final class HFileProtos {
public interface FileTrailerProtoOrBuilder
extends com.google.protobuf.MessageOrBuilder {
// optional uint64 fileInfoOffset = 1;
// optional uint64 file_info_offset = 1;
boolean hasFileInfoOffset();
long getFileInfoOffset();
// optional uint64 loadOnOpenDataOffset = 2;
// optional uint64 load_on_open_data_offset = 2;
boolean hasLoadOnOpenDataOffset();
long getLoadOnOpenDataOffset();
// optional uint64 uncompressedDataIndexSize = 3;
// optional uint64 uncompressed_data_index_size = 3;
boolean hasUncompressedDataIndexSize();
long getUncompressedDataIndexSize();
// optional uint64 totalUncompressedBytes = 4;
// optional uint64 total_uncompressed_bytes = 4;
boolean hasTotalUncompressedBytes();
long getTotalUncompressedBytes();
// optional uint32 dataIndexCount = 5;
// optional uint32 data_index_count = 5;
boolean hasDataIndexCount();
int getDataIndexCount();
// optional uint32 metaIndexCount = 6;
// optional uint32 meta_index_count = 6;
boolean hasMetaIndexCount();
int getMetaIndexCount();
// optional uint64 entryCount = 7;
// optional uint64 entry_count = 7;
boolean hasEntryCount();
long getEntryCount();
// optional uint32 numDataIndexLevels = 8;
// optional uint32 num_data_index_levels = 8;
boolean hasNumDataIndexLevels();
int getNumDataIndexLevels();
// optional uint64 firstDataBlockOffset = 9;
// optional uint64 first_data_block_offset = 9;
boolean hasFirstDataBlockOffset();
long getFirstDataBlockOffset();
// optional uint64 lastDataBlockOffset = 10;
// optional uint64 last_data_block_offset = 10;
boolean hasLastDataBlockOffset();
long getLastDataBlockOffset();
// optional string comparatorClassName = 11;
// optional string comparator_class_name = 11;
boolean hasComparatorClassName();
String getComparatorClassName();
// optional uint32 compressionCodec = 12;
// optional uint32 compression_codec = 12;
boolean hasCompressionCodec();
int getCompressionCodec();
}
@ -681,8 +681,8 @@ public final class HFileProtos {
}
private int bitField0_;
// optional uint64 fileInfoOffset = 1;
public static final int FILEINFOOFFSET_FIELD_NUMBER = 1;
// optional uint64 file_info_offset = 1;
public static final int FILE_INFO_OFFSET_FIELD_NUMBER = 1;
private long fileInfoOffset_;
public boolean hasFileInfoOffset() {
return ((bitField0_ & 0x00000001) == 0x00000001);
@ -691,8 +691,8 @@ public final class HFileProtos {
return fileInfoOffset_;
}
// optional uint64 loadOnOpenDataOffset = 2;
public static final int LOADONOPENDATAOFFSET_FIELD_NUMBER = 2;
// optional uint64 load_on_open_data_offset = 2;
public static final int LOAD_ON_OPEN_DATA_OFFSET_FIELD_NUMBER = 2;
private long loadOnOpenDataOffset_;
public boolean hasLoadOnOpenDataOffset() {
return ((bitField0_ & 0x00000002) == 0x00000002);
@ -701,8 +701,8 @@ public final class HFileProtos {
return loadOnOpenDataOffset_;
}
// optional uint64 uncompressedDataIndexSize = 3;
public static final int UNCOMPRESSEDDATAINDEXSIZE_FIELD_NUMBER = 3;
// optional uint64 uncompressed_data_index_size = 3;
public static final int UNCOMPRESSED_DATA_INDEX_SIZE_FIELD_NUMBER = 3;
private long uncompressedDataIndexSize_;
public boolean hasUncompressedDataIndexSize() {
return ((bitField0_ & 0x00000004) == 0x00000004);
@ -711,8 +711,8 @@ public final class HFileProtos {
return uncompressedDataIndexSize_;
}
// optional uint64 totalUncompressedBytes = 4;
public static final int TOTALUNCOMPRESSEDBYTES_FIELD_NUMBER = 4;
// optional uint64 total_uncompressed_bytes = 4;
public static final int TOTAL_UNCOMPRESSED_BYTES_FIELD_NUMBER = 4;
private long totalUncompressedBytes_;
public boolean hasTotalUncompressedBytes() {
return ((bitField0_ & 0x00000008) == 0x00000008);
@ -721,8 +721,8 @@ public final class HFileProtos {
return totalUncompressedBytes_;
}
// optional uint32 dataIndexCount = 5;
public static final int DATAINDEXCOUNT_FIELD_NUMBER = 5;
// optional uint32 data_index_count = 5;
public static final int DATA_INDEX_COUNT_FIELD_NUMBER = 5;
private int dataIndexCount_;
public boolean hasDataIndexCount() {
return ((bitField0_ & 0x00000010) == 0x00000010);
@ -731,8 +731,8 @@ public final class HFileProtos {
return dataIndexCount_;
}
// optional uint32 metaIndexCount = 6;
public static final int METAINDEXCOUNT_FIELD_NUMBER = 6;
// optional uint32 meta_index_count = 6;
public static final int META_INDEX_COUNT_FIELD_NUMBER = 6;
private int metaIndexCount_;
public boolean hasMetaIndexCount() {
return ((bitField0_ & 0x00000020) == 0x00000020);
@ -741,8 +741,8 @@ public final class HFileProtos {
return metaIndexCount_;
}
// optional uint64 entryCount = 7;
public static final int ENTRYCOUNT_FIELD_NUMBER = 7;
// optional uint64 entry_count = 7;
public static final int ENTRY_COUNT_FIELD_NUMBER = 7;
private long entryCount_;
public boolean hasEntryCount() {
return ((bitField0_ & 0x00000040) == 0x00000040);
@ -751,8 +751,8 @@ public final class HFileProtos {
return entryCount_;
}
// optional uint32 numDataIndexLevels = 8;
public static final int NUMDATAINDEXLEVELS_FIELD_NUMBER = 8;
// optional uint32 num_data_index_levels = 8;
public static final int NUM_DATA_INDEX_LEVELS_FIELD_NUMBER = 8;
private int numDataIndexLevels_;
public boolean hasNumDataIndexLevels() {
return ((bitField0_ & 0x00000080) == 0x00000080);
@ -761,8 +761,8 @@ public final class HFileProtos {
return numDataIndexLevels_;
}
// optional uint64 firstDataBlockOffset = 9;
public static final int FIRSTDATABLOCKOFFSET_FIELD_NUMBER = 9;
// optional uint64 first_data_block_offset = 9;
public static final int FIRST_DATA_BLOCK_OFFSET_FIELD_NUMBER = 9;
private long firstDataBlockOffset_;
public boolean hasFirstDataBlockOffset() {
return ((bitField0_ & 0x00000100) == 0x00000100);
@ -771,8 +771,8 @@ public final class HFileProtos {
return firstDataBlockOffset_;
}
// optional uint64 lastDataBlockOffset = 10;
public static final int LASTDATABLOCKOFFSET_FIELD_NUMBER = 10;
// optional uint64 last_data_block_offset = 10;
public static final int LAST_DATA_BLOCK_OFFSET_FIELD_NUMBER = 10;
private long lastDataBlockOffset_;
public boolean hasLastDataBlockOffset() {
return ((bitField0_ & 0x00000200) == 0x00000200);
@ -781,8 +781,8 @@ public final class HFileProtos {
return lastDataBlockOffset_;
}
// optional string comparatorClassName = 11;
public static final int COMPARATORCLASSNAME_FIELD_NUMBER = 11;
// optional string comparator_class_name = 11;
public static final int COMPARATOR_CLASS_NAME_FIELD_NUMBER = 11;
private java.lang.Object comparatorClassName_;
public boolean hasComparatorClassName() {
return ((bitField0_ & 0x00000400) == 0x00000400);
@ -813,8 +813,8 @@ public final class HFileProtos {
}
}
// optional uint32 compressionCodec = 12;
public static final int COMPRESSIONCODEC_FIELD_NUMBER = 12;
// optional uint32 compression_codec = 12;
public static final int COMPRESSION_CODEC_FIELD_NUMBER = 12;
private int compressionCodec_;
public boolean hasCompressionCodec() {
return ((bitField0_ & 0x00000800) == 0x00000800);
@ -1035,51 +1035,51 @@ public final class HFileProtos {
int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode();
if (hasFileInfoOffset()) {
hash = (37 * hash) + FILEINFOOFFSET_FIELD_NUMBER;
hash = (37 * hash) + FILE_INFO_OFFSET_FIELD_NUMBER;
hash = (53 * hash) + hashLong(getFileInfoOffset());
}
if (hasLoadOnOpenDataOffset()) {
hash = (37 * hash) + LOADONOPENDATAOFFSET_FIELD_NUMBER;
hash = (37 * hash) + LOAD_ON_OPEN_DATA_OFFSET_FIELD_NUMBER;
hash = (53 * hash) + hashLong(getLoadOnOpenDataOffset());
}
if (hasUncompressedDataIndexSize()) {
hash = (37 * hash) + UNCOMPRESSEDDATAINDEXSIZE_FIELD_NUMBER;
hash = (37 * hash) + UNCOMPRESSED_DATA_INDEX_SIZE_FIELD_NUMBER;
hash = (53 * hash) + hashLong(getUncompressedDataIndexSize());
}
if (hasTotalUncompressedBytes()) {
hash = (37 * hash) + TOTALUNCOMPRESSEDBYTES_FIELD_NUMBER;
hash = (37 * hash) + TOTAL_UNCOMPRESSED_BYTES_FIELD_NUMBER;
hash = (53 * hash) + hashLong(getTotalUncompressedBytes());
}
if (hasDataIndexCount()) {
hash = (37 * hash) + DATAINDEXCOUNT_FIELD_NUMBER;
hash = (37 * hash) + DATA_INDEX_COUNT_FIELD_NUMBER;
hash = (53 * hash) + getDataIndexCount();
}
if (hasMetaIndexCount()) {
hash = (37 * hash) + METAINDEXCOUNT_FIELD_NUMBER;
hash = (37 * hash) + META_INDEX_COUNT_FIELD_NUMBER;
hash = (53 * hash) + getMetaIndexCount();
}
if (hasEntryCount()) {
hash = (37 * hash) + ENTRYCOUNT_FIELD_NUMBER;
hash = (37 * hash) + ENTRY_COUNT_FIELD_NUMBER;
hash = (53 * hash) + hashLong(getEntryCount());
}
if (hasNumDataIndexLevels()) {
hash = (37 * hash) + NUMDATAINDEXLEVELS_FIELD_NUMBER;
hash = (37 * hash) + NUM_DATA_INDEX_LEVELS_FIELD_NUMBER;
hash = (53 * hash) + getNumDataIndexLevels();
}
if (hasFirstDataBlockOffset()) {
hash = (37 * hash) + FIRSTDATABLOCKOFFSET_FIELD_NUMBER;
hash = (37 * hash) + FIRST_DATA_BLOCK_OFFSET_FIELD_NUMBER;
hash = (53 * hash) + hashLong(getFirstDataBlockOffset());
}
if (hasLastDataBlockOffset()) {
hash = (37 * hash) + LASTDATABLOCKOFFSET_FIELD_NUMBER;
hash = (37 * hash) + LAST_DATA_BLOCK_OFFSET_FIELD_NUMBER;
hash = (53 * hash) + hashLong(getLastDataBlockOffset());
}
if (hasComparatorClassName()) {
hash = (37 * hash) + COMPARATORCLASSNAME_FIELD_NUMBER;
hash = (37 * hash) + COMPARATOR_CLASS_NAME_FIELD_NUMBER;
hash = (53 * hash) + getComparatorClassName().hashCode();
}
if (hasCompressionCodec()) {
hash = (37 * hash) + COMPRESSIONCODEC_FIELD_NUMBER;
hash = (37 * hash) + COMPRESSION_CODEC_FIELD_NUMBER;
hash = (53 * hash) + getCompressionCodec();
}
hash = (29 * hash) + getUnknownFields().hashCode();
@ -1457,7 +1457,7 @@ public final class HFileProtos {
private int bitField0_;
// optional uint64 fileInfoOffset = 1;
// optional uint64 file_info_offset = 1;
private long fileInfoOffset_ ;
public boolean hasFileInfoOffset() {
return ((bitField0_ & 0x00000001) == 0x00000001);
@ -1478,7 +1478,7 @@ public final class HFileProtos {
return this;
}
// optional uint64 loadOnOpenDataOffset = 2;
// optional uint64 load_on_open_data_offset = 2;
private long loadOnOpenDataOffset_ ;
public boolean hasLoadOnOpenDataOffset() {
return ((bitField0_ & 0x00000002) == 0x00000002);
@ -1499,7 +1499,7 @@ public final class HFileProtos {
return this;
}
// optional uint64 uncompressedDataIndexSize = 3;
// optional uint64 uncompressed_data_index_size = 3;
private long uncompressedDataIndexSize_ ;
public boolean hasUncompressedDataIndexSize() {
return ((bitField0_ & 0x00000004) == 0x00000004);
@ -1520,7 +1520,7 @@ public final class HFileProtos {
return this;
}
// optional uint64 totalUncompressedBytes = 4;
// optional uint64 total_uncompressed_bytes = 4;
private long totalUncompressedBytes_ ;
public boolean hasTotalUncompressedBytes() {
return ((bitField0_ & 0x00000008) == 0x00000008);
@ -1541,7 +1541,7 @@ public final class HFileProtos {
return this;
}
// optional uint32 dataIndexCount = 5;
// optional uint32 data_index_count = 5;
private int dataIndexCount_ ;
public boolean hasDataIndexCount() {
return ((bitField0_ & 0x00000010) == 0x00000010);
@ -1562,7 +1562,7 @@ public final class HFileProtos {
return this;
}
// optional uint32 metaIndexCount = 6;
// optional uint32 meta_index_count = 6;
private int metaIndexCount_ ;
public boolean hasMetaIndexCount() {
return ((bitField0_ & 0x00000020) == 0x00000020);
@ -1583,7 +1583,7 @@ public final class HFileProtos {
return this;
}
// optional uint64 entryCount = 7;
// optional uint64 entry_count = 7;
private long entryCount_ ;
public boolean hasEntryCount() {
return ((bitField0_ & 0x00000040) == 0x00000040);
@ -1604,7 +1604,7 @@ public final class HFileProtos {
return this;
}
// optional uint32 numDataIndexLevels = 8;
// optional uint32 num_data_index_levels = 8;
private int numDataIndexLevels_ ;
public boolean hasNumDataIndexLevels() {
return ((bitField0_ & 0x00000080) == 0x00000080);
@ -1625,7 +1625,7 @@ public final class HFileProtos {
return this;
}
// optional uint64 firstDataBlockOffset = 9;
// optional uint64 first_data_block_offset = 9;
private long firstDataBlockOffset_ ;
public boolean hasFirstDataBlockOffset() {
return ((bitField0_ & 0x00000100) == 0x00000100);
@ -1646,7 +1646,7 @@ public final class HFileProtos {
return this;
}
// optional uint64 lastDataBlockOffset = 10;
// optional uint64 last_data_block_offset = 10;
private long lastDataBlockOffset_ ;
public boolean hasLastDataBlockOffset() {
return ((bitField0_ & 0x00000200) == 0x00000200);
@ -1667,7 +1667,7 @@ public final class HFileProtos {
return this;
}
// optional string comparatorClassName = 11;
// optional string comparator_class_name = 11;
private java.lang.Object comparatorClassName_ = "";
public boolean hasComparatorClassName() {
return ((bitField0_ & 0x00000400) == 0x00000400);
@ -1703,7 +1703,7 @@ public final class HFileProtos {
onChanged();
}
// optional uint32 compressionCodec = 12;
// optional uint32 compression_codec = 12;
private int compressionCodec_ ;
public boolean hasCompressionCodec() {
return ((bitField0_ & 0x00000800) == 0x00000800);
@ -1754,19 +1754,20 @@ public final class HFileProtos {
descriptor;
static {
java.lang.String[] descriptorData = {
"\n\013HFile.proto\032\013hbase.proto\"2\n\rFileInfoPr" +
"oto\022!\n\010mapEntry\030\001 \003(\0132\017.BytesBytesPair\"\335" +
"\002\n\020FileTrailerProto\022\026\n\016fileInfoOffset\030\001 " +
"\001(\004\022\034\n\024loadOnOpenDataOffset\030\002 \001(\004\022!\n\031unc" +
"ompressedDataIndexSize\030\003 \001(\004\022\036\n\026totalUnc" +
"ompressedBytes\030\004 \001(\004\022\026\n\016dataIndexCount\030\005" +
" \001(\r\022\026\n\016metaIndexCount\030\006 \001(\r\022\022\n\nentryCou" +
"nt\030\007 \001(\004\022\032\n\022numDataIndexLevels\030\010 \001(\r\022\034\n\024" +
"firstDataBlockOffset\030\t \001(\004\022\033\n\023lastDataBl" +
"ockOffset\030\n \001(\004\022\033\n\023comparatorClassName\030\013",
" \001(\t\022\030\n\020compressionCodec\030\014 \001(\rBA\n*org.ap" +
"ache.hadoop.hbase.protobuf.generatedB\013HF" +
"ileProtosH\001\210\001\001\240\001\001"
"\n\013HFile.proto\032\013hbase.proto\"3\n\rFileInfoPr" +
"oto\022\"\n\tmap_entry\030\001 \003(\0132\017.BytesBytesPair\"" +
"\371\002\n\020FileTrailerProto\022\030\n\020file_info_offset" +
"\030\001 \001(\004\022 \n\030load_on_open_data_offset\030\002 \001(\004" +
"\022$\n\034uncompressed_data_index_size\030\003 \001(\004\022 " +
"\n\030total_uncompressed_bytes\030\004 \001(\004\022\030\n\020data" +
"_index_count\030\005 \001(\r\022\030\n\020meta_index_count\030\006" +
" \001(\r\022\023\n\013entry_count\030\007 \001(\004\022\035\n\025num_data_in" +
"dex_levels\030\010 \001(\r\022\037\n\027first_data_block_off" +
"set\030\t \001(\004\022\036\n\026last_data_block_offset\030\n \001(",
"\004\022\035\n\025comparator_class_name\030\013 \001(\t\022\031\n\021comp" +
"ression_codec\030\014 \001(\rBA\n*org.apache.hadoop" +
".hbase.protobuf.generatedB\013HFileProtosH\001" +
"\210\001\001\240\001\001"
};
com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() {

View File

@ -11,7 +11,7 @@ public final class LoadBalancerProtos {
public interface LoadBalancerStateOrBuilder
extends com.google.protobuf.MessageOrBuilder {
// optional bool balancerOn = 1;
// optional bool balancer_on = 1;
boolean hasBalancerOn();
boolean getBalancerOn();
}
@ -44,8 +44,8 @@ public final class LoadBalancerProtos {
}
private int bitField0_;
// optional bool balancerOn = 1;
public static final int BALANCERON_FIELD_NUMBER = 1;
// optional bool balancer_on = 1;
public static final int BALANCER_ON_FIELD_NUMBER = 1;
private boolean balancerOn_;
public boolean hasBalancerOn() {
return ((bitField0_ & 0x00000001) == 0x00000001);
@ -123,7 +123,7 @@ public final class LoadBalancerProtos {
int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode();
if (hasBalancerOn()) {
hash = (37 * hash) + BALANCERON_FIELD_NUMBER;
hash = (37 * hash) + BALANCER_ON_FIELD_NUMBER;
hash = (53 * hash) + hashBoolean(getBalancerOn());
}
hash = (29 * hash) + getUnknownFields().hashCode();
@ -347,7 +347,7 @@ public final class LoadBalancerProtos {
private int bitField0_;
// optional bool balancerOn = 1;
// optional bool balancer_on = 1;
private boolean balancerOn_ ;
public boolean hasBalancerOn() {
return ((bitField0_ & 0x00000001) == 0x00000001);
@ -393,10 +393,10 @@ public final class LoadBalancerProtos {
descriptor;
static {
java.lang.String[] descriptorData = {
"\n\022LoadBalancer.proto\"\'\n\021LoadBalancerStat" +
"e\022\022\n\nbalancerOn\030\001 \001(\010BE\n*org.apache.hado" +
"op.hbase.protobuf.generatedB\022LoadBalance" +
"rProtosH\001\240\001\001"
"\n\022LoadBalancer.proto\"(\n\021LoadBalancerStat" +
"e\022\023\n\013balancer_on\030\001 \001(\010BE\n*org.apache.had" +
"oop.hbase.protobuf.generatedB\022LoadBalanc" +
"erProtosH\001\240\001\001"
};
com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() {

View File

@ -11,11 +11,11 @@ public final class MasterAdminProtos {
public interface AddColumnRequestOrBuilder
extends com.google.protobuf.MessageOrBuilder {
// required bytes tableName = 1;
// required bytes table_name = 1;
boolean hasTableName();
com.google.protobuf.ByteString getTableName();
// required .ColumnFamilySchema columnFamilies = 2;
// required .ColumnFamilySchema column_families = 2;
boolean hasColumnFamilies();
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema getColumnFamilies();
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchemaOrBuilder getColumnFamiliesOrBuilder();
@ -49,8 +49,8 @@ public final class MasterAdminProtos {
}
private int bitField0_;
// required bytes tableName = 1;
public static final int TABLENAME_FIELD_NUMBER = 1;
// required bytes table_name = 1;
public static final int TABLE_NAME_FIELD_NUMBER = 1;
private com.google.protobuf.ByteString tableName_;
public boolean hasTableName() {
return ((bitField0_ & 0x00000001) == 0x00000001);
@ -59,8 +59,8 @@ public final class MasterAdminProtos {
return tableName_;
}
// required .ColumnFamilySchema columnFamilies = 2;
public static final int COLUMNFAMILIES_FIELD_NUMBER = 2;
// required .ColumnFamilySchema column_families = 2;
public static final int COLUMN_FAMILIES_FIELD_NUMBER = 2;
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema columnFamilies_;
public boolean hasColumnFamilies() {
return ((bitField0_ & 0x00000002) == 0x00000002);
@ -166,11 +166,11 @@ public final class MasterAdminProtos {
int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode();
if (hasTableName()) {
hash = (37 * hash) + TABLENAME_FIELD_NUMBER;
hash = (37 * hash) + TABLE_NAME_FIELD_NUMBER;
hash = (53 * hash) + getTableName().hashCode();
}
if (hasColumnFamilies()) {
hash = (37 * hash) + COLUMNFAMILIES_FIELD_NUMBER;
hash = (37 * hash) + COLUMN_FAMILIES_FIELD_NUMBER;
hash = (53 * hash) + getColumnFamilies().hashCode();
}
hash = (29 * hash) + getUnknownFields().hashCode();
@ -433,7 +433,7 @@ public final class MasterAdminProtos {
private int bitField0_;
// required bytes tableName = 1;
// required bytes table_name = 1;
private com.google.protobuf.ByteString tableName_ = com.google.protobuf.ByteString.EMPTY;
public boolean hasTableName() {
return ((bitField0_ & 0x00000001) == 0x00000001);
@ -457,7 +457,7 @@ public final class MasterAdminProtos {
return this;
}
// required .ColumnFamilySchema columnFamilies = 2;
// required .ColumnFamilySchema column_families = 2;
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema columnFamilies_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema.getDefaultInstance();
private com.google.protobuf.SingleFieldBuilder<
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchemaOrBuilder> columnFamiliesBuilder_;
@ -861,11 +861,11 @@ public final class MasterAdminProtos {
public interface DeleteColumnRequestOrBuilder
extends com.google.protobuf.MessageOrBuilder {
// required bytes tableName = 1;
// required bytes table_name = 1;
boolean hasTableName();
com.google.protobuf.ByteString getTableName();
// required bytes columnName = 2;
// required bytes column_name = 2;
boolean hasColumnName();
com.google.protobuf.ByteString getColumnName();
}
@ -898,8 +898,8 @@ public final class MasterAdminProtos {
}
private int bitField0_;
// required bytes tableName = 1;
public static final int TABLENAME_FIELD_NUMBER = 1;
// required bytes table_name = 1;
public static final int TABLE_NAME_FIELD_NUMBER = 1;
private com.google.protobuf.ByteString tableName_;
public boolean hasTableName() {
return ((bitField0_ & 0x00000001) == 0x00000001);
@ -908,8 +908,8 @@ public final class MasterAdminProtos {
return tableName_;
}
// required bytes columnName = 2;
public static final int COLUMNNAME_FIELD_NUMBER = 2;
// required bytes column_name = 2;
public static final int COLUMN_NAME_FIELD_NUMBER = 2;
private com.google.protobuf.ByteString columnName_;
public boolean hasColumnName() {
return ((bitField0_ & 0x00000002) == 0x00000002);
@ -1008,11 +1008,11 @@ public final class MasterAdminProtos {
int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode();
if (hasTableName()) {
hash = (37 * hash) + TABLENAME_FIELD_NUMBER;
hash = (37 * hash) + TABLE_NAME_FIELD_NUMBER;
hash = (53 * hash) + getTableName().hashCode();
}
if (hasColumnName()) {
hash = (37 * hash) + COLUMNNAME_FIELD_NUMBER;
hash = (37 * hash) + COLUMN_NAME_FIELD_NUMBER;
hash = (53 * hash) + getColumnName().hashCode();
}
hash = (29 * hash) + getUnknownFields().hashCode();
@ -1258,7 +1258,7 @@ public final class MasterAdminProtos {
private int bitField0_;
// required bytes tableName = 1;
// required bytes table_name = 1;
private com.google.protobuf.ByteString tableName_ = com.google.protobuf.ByteString.EMPTY;
public boolean hasTableName() {
return ((bitField0_ & 0x00000001) == 0x00000001);
@ -1282,7 +1282,7 @@ public final class MasterAdminProtos {
return this;
}
// required bytes columnName = 2;
// required bytes column_name = 2;
private com.google.protobuf.ByteString columnName_ = com.google.protobuf.ByteString.EMPTY;
public boolean hasColumnName() {
return ((bitField0_ & 0x00000002) == 0x00000002);
@ -1620,11 +1620,11 @@ public final class MasterAdminProtos {
public interface ModifyColumnRequestOrBuilder
extends com.google.protobuf.MessageOrBuilder {
// required bytes tableName = 1;
// required bytes table_name = 1;
boolean hasTableName();
com.google.protobuf.ByteString getTableName();
// required .ColumnFamilySchema columnFamilies = 2;
// required .ColumnFamilySchema column_families = 2;
boolean hasColumnFamilies();
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema getColumnFamilies();
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchemaOrBuilder getColumnFamiliesOrBuilder();
@ -1658,8 +1658,8 @@ public final class MasterAdminProtos {
}
private int bitField0_;
// required bytes tableName = 1;
public static final int TABLENAME_FIELD_NUMBER = 1;
// required bytes table_name = 1;
public static final int TABLE_NAME_FIELD_NUMBER = 1;
private com.google.protobuf.ByteString tableName_;
public boolean hasTableName() {
return ((bitField0_ & 0x00000001) == 0x00000001);
@ -1668,8 +1668,8 @@ public final class MasterAdminProtos {
return tableName_;
}
// required .ColumnFamilySchema columnFamilies = 2;
public static final int COLUMNFAMILIES_FIELD_NUMBER = 2;
// required .ColumnFamilySchema column_families = 2;
public static final int COLUMN_FAMILIES_FIELD_NUMBER = 2;
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema columnFamilies_;
public boolean hasColumnFamilies() {
return ((bitField0_ & 0x00000002) == 0x00000002);
@ -1775,11 +1775,11 @@ public final class MasterAdminProtos {
int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode();
if (hasTableName()) {
hash = (37 * hash) + TABLENAME_FIELD_NUMBER;
hash = (37 * hash) + TABLE_NAME_FIELD_NUMBER;
hash = (53 * hash) + getTableName().hashCode();
}
if (hasColumnFamilies()) {
hash = (37 * hash) + COLUMNFAMILIES_FIELD_NUMBER;
hash = (37 * hash) + COLUMN_FAMILIES_FIELD_NUMBER;
hash = (53 * hash) + getColumnFamilies().hashCode();
}
hash = (29 * hash) + getUnknownFields().hashCode();
@ -2042,7 +2042,7 @@ public final class MasterAdminProtos {
private int bitField0_;
// required bytes tableName = 1;
// required bytes table_name = 1;
private com.google.protobuf.ByteString tableName_ = com.google.protobuf.ByteString.EMPTY;
public boolean hasTableName() {
return ((bitField0_ & 0x00000001) == 0x00000001);
@ -2066,7 +2066,7 @@ public final class MasterAdminProtos {
return this;
}
// required .ColumnFamilySchema columnFamilies = 2;
// required .ColumnFamilySchema column_families = 2;
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema columnFamilies_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema.getDefaultInstance();
private com.google.protobuf.SingleFieldBuilder<
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchemaOrBuilder> columnFamiliesBuilder_;
@ -2475,7 +2475,7 @@ public final class MasterAdminProtos {
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier getRegion();
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder getRegionOrBuilder();
// optional .ServerName destServerName = 2;
// optional .ServerName dest_server_name = 2;
boolean hasDestServerName();
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName getDestServerName();
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerNameOrBuilder getDestServerNameOrBuilder();
@ -2522,8 +2522,8 @@ public final class MasterAdminProtos {
return region_;
}
// optional .ServerName destServerName = 2;
public static final int DESTSERVERNAME_FIELD_NUMBER = 2;
// optional .ServerName dest_server_name = 2;
public static final int DEST_SERVER_NAME_FIELD_NUMBER = 2;
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName destServerName_;
public boolean hasDestServerName() {
return ((bitField0_ & 0x00000002) == 0x00000002);
@ -2635,7 +2635,7 @@ public final class MasterAdminProtos {
hash = (53 * hash) + getRegion().hashCode();
}
if (hasDestServerName()) {
hash = (37 * hash) + DESTSERVERNAME_FIELD_NUMBER;
hash = (37 * hash) + DEST_SERVER_NAME_FIELD_NUMBER;
hash = (53 * hash) + getDestServerName().hashCode();
}
hash = (29 * hash) + getUnknownFields().hashCode();
@ -3003,7 +3003,7 @@ public final class MasterAdminProtos {
return regionBuilder_;
}
// optional .ServerName destServerName = 2;
// optional .ServerName dest_server_name = 2;
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName destServerName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.getDefaultInstance();
private com.google.protobuf.SingleFieldBuilder<
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerNameOrBuilder> destServerNameBuilder_;
@ -3407,12 +3407,12 @@ public final class MasterAdminProtos {
public interface DispatchMergingRegionsRequestOrBuilder
extends com.google.protobuf.MessageOrBuilder {
// required .RegionSpecifier regionA = 1;
// required .RegionSpecifier region_a = 1;
boolean hasRegionA();
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier getRegionA();
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder getRegionAOrBuilder();
// required .RegionSpecifier regionB = 2;
// required .RegionSpecifier region_b = 2;
boolean hasRegionB();
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier getRegionB();
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder getRegionBOrBuilder();
@ -3450,8 +3450,8 @@ public final class MasterAdminProtos {
}
private int bitField0_;
// required .RegionSpecifier regionA = 1;
public static final int REGIONA_FIELD_NUMBER = 1;
// required .RegionSpecifier region_a = 1;
public static final int REGION_A_FIELD_NUMBER = 1;
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier regionA_;
public boolean hasRegionA() {
return ((bitField0_ & 0x00000001) == 0x00000001);
@ -3463,8 +3463,8 @@ public final class MasterAdminProtos {
return regionA_;
}
// required .RegionSpecifier regionB = 2;
public static final int REGIONB_FIELD_NUMBER = 2;
// required .RegionSpecifier region_b = 2;
public static final int REGION_B_FIELD_NUMBER = 2;
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier regionB_;
public boolean hasRegionB() {
return ((bitField0_ & 0x00000002) == 0x00000002);
@ -3597,11 +3597,11 @@ public final class MasterAdminProtos {
int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode();
if (hasRegionA()) {
hash = (37 * hash) + REGIONA_FIELD_NUMBER;
hash = (37 * hash) + REGION_A_FIELD_NUMBER;
hash = (53 * hash) + getRegionA().hashCode();
}
if (hasRegionB()) {
hash = (37 * hash) + REGIONB_FIELD_NUMBER;
hash = (37 * hash) + REGION_B_FIELD_NUMBER;
hash = (53 * hash) + getRegionB().hashCode();
}
if (hasForcible()) {
@ -3899,7 +3899,7 @@ public final class MasterAdminProtos {
private int bitField0_;
// required .RegionSpecifier regionA = 1;
// required .RegionSpecifier region_a = 1;
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier regionA_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance();
private com.google.protobuf.SingleFieldBuilder<
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder> regionABuilder_;
@ -3989,7 +3989,7 @@ public final class MasterAdminProtos {
return regionABuilder_;
}
// required .RegionSpecifier regionB = 2;
// required .RegionSpecifier region_b = 2;
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier regionB_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance();
private com.google.protobuf.SingleFieldBuilder<
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder> regionBBuilder_;
@ -6799,12 +6799,12 @@ public final class MasterAdminProtos {
public interface CreateTableRequestOrBuilder
extends com.google.protobuf.MessageOrBuilder {
// required .TableSchema tableSchema = 1;
// required .TableSchema table_schema = 1;
boolean hasTableSchema();
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema getTableSchema();
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchemaOrBuilder getTableSchemaOrBuilder();
// repeated bytes splitKeys = 2;
// repeated bytes split_keys = 2;
java.util.List<com.google.protobuf.ByteString> getSplitKeysList();
int getSplitKeysCount();
com.google.protobuf.ByteString getSplitKeys(int index);
@ -6838,8 +6838,8 @@ public final class MasterAdminProtos {
}
private int bitField0_;
// required .TableSchema tableSchema = 1;
public static final int TABLESCHEMA_FIELD_NUMBER = 1;
// required .TableSchema table_schema = 1;
public static final int TABLE_SCHEMA_FIELD_NUMBER = 1;
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema tableSchema_;
public boolean hasTableSchema() {
return ((bitField0_ & 0x00000001) == 0x00000001);
@ -6851,8 +6851,8 @@ public final class MasterAdminProtos {
return tableSchema_;
}
// repeated bytes splitKeys = 2;
public static final int SPLITKEYS_FIELD_NUMBER = 2;
// repeated bytes split_keys = 2;
public static final int SPLIT_KEYS_FIELD_NUMBER = 2;
private java.util.List<com.google.protobuf.ByteString> splitKeys_;
public java.util.List<com.google.protobuf.ByteString>
getSplitKeysList() {
@ -6957,11 +6957,11 @@ public final class MasterAdminProtos {
int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode();
if (hasTableSchema()) {
hash = (37 * hash) + TABLESCHEMA_FIELD_NUMBER;
hash = (37 * hash) + TABLE_SCHEMA_FIELD_NUMBER;
hash = (53 * hash) + getTableSchema().hashCode();
}
if (getSplitKeysCount() > 0) {
hash = (37 * hash) + SPLITKEYS_FIELD_NUMBER;
hash = (37 * hash) + SPLIT_KEYS_FIELD_NUMBER;
hash = (53 * hash) + getSplitKeysList().hashCode();
}
hash = (29 * hash) + getUnknownFields().hashCode();
@ -7228,7 +7228,7 @@ public final class MasterAdminProtos {
private int bitField0_;
// required .TableSchema tableSchema = 1;
// required .TableSchema table_schema = 1;
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema tableSchema_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema.getDefaultInstance();
private com.google.protobuf.SingleFieldBuilder<
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchemaOrBuilder> tableSchemaBuilder_;
@ -7318,7 +7318,7 @@ public final class MasterAdminProtos {
return tableSchemaBuilder_;
}
// repeated bytes splitKeys = 2;
// repeated bytes split_keys = 2;
private java.util.List<com.google.protobuf.ByteString> splitKeys_ = java.util.Collections.emptyList();;
private void ensureSplitKeysIsMutable() {
if (!((bitField0_ & 0x00000002) == 0x00000002)) {
@ -7683,7 +7683,7 @@ public final class MasterAdminProtos {
public interface DeleteTableRequestOrBuilder
extends com.google.protobuf.MessageOrBuilder {
// required bytes tableName = 1;
// required bytes table_name = 1;
boolean hasTableName();
com.google.protobuf.ByteString getTableName();
}
@ -7716,8 +7716,8 @@ public final class MasterAdminProtos {
}
private int bitField0_;
// required bytes tableName = 1;
public static final int TABLENAME_FIELD_NUMBER = 1;
// required bytes table_name = 1;
public static final int TABLE_NAME_FIELD_NUMBER = 1;
private com.google.protobuf.ByteString tableName_;
public boolean hasTableName() {
return ((bitField0_ & 0x00000001) == 0x00000001);
@ -7799,7 +7799,7 @@ public final class MasterAdminProtos {
int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode();
if (hasTableName()) {
hash = (37 * hash) + TABLENAME_FIELD_NUMBER;
hash = (37 * hash) + TABLE_NAME_FIELD_NUMBER;
hash = (53 * hash) + getTableName().hashCode();
}
hash = (29 * hash) + getUnknownFields().hashCode();
@ -8027,7 +8027,7 @@ public final class MasterAdminProtos {
private int bitField0_;
// required bytes tableName = 1;
// required bytes table_name = 1;
private com.google.protobuf.ByteString tableName_ = com.google.protobuf.ByteString.EMPTY;
public boolean hasTableName() {
return ((bitField0_ & 0x00000001) == 0x00000001);
@ -8365,7 +8365,7 @@ public final class MasterAdminProtos {
public interface EnableTableRequestOrBuilder
extends com.google.protobuf.MessageOrBuilder {
// required bytes tableName = 1;
// required bytes table_name = 1;
boolean hasTableName();
com.google.protobuf.ByteString getTableName();
}
@ -8398,8 +8398,8 @@ public final class MasterAdminProtos {
}
private int bitField0_;
// required bytes tableName = 1;
public static final int TABLENAME_FIELD_NUMBER = 1;
// required bytes table_name = 1;
public static final int TABLE_NAME_FIELD_NUMBER = 1;
private com.google.protobuf.ByteString tableName_;
public boolean hasTableName() {
return ((bitField0_ & 0x00000001) == 0x00000001);
@ -8481,7 +8481,7 @@ public final class MasterAdminProtos {
int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode();
if (hasTableName()) {
hash = (37 * hash) + TABLENAME_FIELD_NUMBER;
hash = (37 * hash) + TABLE_NAME_FIELD_NUMBER;
hash = (53 * hash) + getTableName().hashCode();
}
hash = (29 * hash) + getUnknownFields().hashCode();
@ -8709,7 +8709,7 @@ public final class MasterAdminProtos {
private int bitField0_;
// required bytes tableName = 1;
// required bytes table_name = 1;
private com.google.protobuf.ByteString tableName_ = com.google.protobuf.ByteString.EMPTY;
public boolean hasTableName() {
return ((bitField0_ & 0x00000001) == 0x00000001);
@ -9047,7 +9047,7 @@ public final class MasterAdminProtos {
public interface DisableTableRequestOrBuilder
extends com.google.protobuf.MessageOrBuilder {
// required bytes tableName = 1;
// required bytes table_name = 1;
boolean hasTableName();
com.google.protobuf.ByteString getTableName();
}
@ -9080,8 +9080,8 @@ public final class MasterAdminProtos {
}
private int bitField0_;
// required bytes tableName = 1;
public static final int TABLENAME_FIELD_NUMBER = 1;
// required bytes table_name = 1;
public static final int TABLE_NAME_FIELD_NUMBER = 1;
private com.google.protobuf.ByteString tableName_;
public boolean hasTableName() {
return ((bitField0_ & 0x00000001) == 0x00000001);
@ -9163,7 +9163,7 @@ public final class MasterAdminProtos {
int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode();
if (hasTableName()) {
hash = (37 * hash) + TABLENAME_FIELD_NUMBER;
hash = (37 * hash) + TABLE_NAME_FIELD_NUMBER;
hash = (53 * hash) + getTableName().hashCode();
}
hash = (29 * hash) + getUnknownFields().hashCode();
@ -9391,7 +9391,7 @@ public final class MasterAdminProtos {
private int bitField0_;
// required bytes tableName = 1;
// required bytes table_name = 1;
private com.google.protobuf.ByteString tableName_ = com.google.protobuf.ByteString.EMPTY;
public boolean hasTableName() {
return ((bitField0_ & 0x00000001) == 0x00000001);
@ -9729,11 +9729,11 @@ public final class MasterAdminProtos {
public interface ModifyTableRequestOrBuilder
extends com.google.protobuf.MessageOrBuilder {
// required bytes tableName = 1;
// required bytes table_name = 1;
boolean hasTableName();
com.google.protobuf.ByteString getTableName();
// required .TableSchema tableSchema = 2;
// required .TableSchema table_schema = 2;
boolean hasTableSchema();
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema getTableSchema();
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchemaOrBuilder getTableSchemaOrBuilder();
@ -9767,8 +9767,8 @@ public final class MasterAdminProtos {
}
private int bitField0_;
// required bytes tableName = 1;
public static final int TABLENAME_FIELD_NUMBER = 1;
// required bytes table_name = 1;
public static final int TABLE_NAME_FIELD_NUMBER = 1;
private com.google.protobuf.ByteString tableName_;
public boolean hasTableName() {
return ((bitField0_ & 0x00000001) == 0x00000001);
@ -9777,8 +9777,8 @@ public final class MasterAdminProtos {
return tableName_;
}
// required .TableSchema tableSchema = 2;
public static final int TABLESCHEMA_FIELD_NUMBER = 2;
// required .TableSchema table_schema = 2;
public static final int TABLE_SCHEMA_FIELD_NUMBER = 2;
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema tableSchema_;
public boolean hasTableSchema() {
return ((bitField0_ & 0x00000002) == 0x00000002);
@ -9884,11 +9884,11 @@ public final class MasterAdminProtos {
int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode();
if (hasTableName()) {
hash = (37 * hash) + TABLENAME_FIELD_NUMBER;
hash = (37 * hash) + TABLE_NAME_FIELD_NUMBER;
hash = (53 * hash) + getTableName().hashCode();
}
if (hasTableSchema()) {
hash = (37 * hash) + TABLESCHEMA_FIELD_NUMBER;
hash = (37 * hash) + TABLE_SCHEMA_FIELD_NUMBER;
hash = (53 * hash) + getTableSchema().hashCode();
}
hash = (29 * hash) + getUnknownFields().hashCode();
@ -10151,7 +10151,7 @@ public final class MasterAdminProtos {
private int bitField0_;
// required bytes tableName = 1;
// required bytes table_name = 1;
private com.google.protobuf.ByteString tableName_ = com.google.protobuf.ByteString.EMPTY;
public boolean hasTableName() {
return ((bitField0_ & 0x00000001) == 0x00000001);
@ -10175,7 +10175,7 @@ public final class MasterAdminProtos {
return this;
}
// required .TableSchema tableSchema = 2;
// required .TableSchema table_schema = 2;
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema tableSchema_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema.getDefaultInstance();
private com.google.protobuf.SingleFieldBuilder<
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchemaOrBuilder> tableSchemaBuilder_;
@ -12079,7 +12079,7 @@ public final class MasterAdminProtos {
public interface BalanceResponseOrBuilder
extends com.google.protobuf.MessageOrBuilder {
// required bool balancerRan = 1;
// required bool balancer_ran = 1;
boolean hasBalancerRan();
boolean getBalancerRan();
}
@ -12112,8 +12112,8 @@ public final class MasterAdminProtos {
}
private int bitField0_;
// required bool balancerRan = 1;
public static final int BALANCERRAN_FIELD_NUMBER = 1;
// required bool balancer_ran = 1;
public static final int BALANCER_RAN_FIELD_NUMBER = 1;
private boolean balancerRan_;
public boolean hasBalancerRan() {
return ((bitField0_ & 0x00000001) == 0x00000001);
@ -12195,7 +12195,7 @@ public final class MasterAdminProtos {
int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode();
if (hasBalancerRan()) {
hash = (37 * hash) + BALANCERRAN_FIELD_NUMBER;
hash = (37 * hash) + BALANCER_RAN_FIELD_NUMBER;
hash = (53 * hash) + hashBoolean(getBalancerRan());
}
hash = (29 * hash) + getUnknownFields().hashCode();
@ -12423,7 +12423,7 @@ public final class MasterAdminProtos {
private int bitField0_;
// required bool balancerRan = 1;
// required bool balancer_ran = 1;
private boolean balancerRan_ ;
public boolean hasBalancerRan() {
return ((bitField0_ & 0x00000001) == 0x00000001);
@ -12903,7 +12903,7 @@ public final class MasterAdminProtos {
public interface SetBalancerRunningResponseOrBuilder
extends com.google.protobuf.MessageOrBuilder {
// optional bool prevBalanceValue = 1;
// optional bool prev_balance_value = 1;
boolean hasPrevBalanceValue();
boolean getPrevBalanceValue();
}
@ -12936,8 +12936,8 @@ public final class MasterAdminProtos {
}
private int bitField0_;
// optional bool prevBalanceValue = 1;
public static final int PREVBALANCEVALUE_FIELD_NUMBER = 1;
// optional bool prev_balance_value = 1;
public static final int PREV_BALANCE_VALUE_FIELD_NUMBER = 1;
private boolean prevBalanceValue_;
public boolean hasPrevBalanceValue() {
return ((bitField0_ & 0x00000001) == 0x00000001);
@ -13015,7 +13015,7 @@ public final class MasterAdminProtos {
int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode();
if (hasPrevBalanceValue()) {
hash = (37 * hash) + PREVBALANCEVALUE_FIELD_NUMBER;
hash = (37 * hash) + PREV_BALANCE_VALUE_FIELD_NUMBER;
hash = (53 * hash) + hashBoolean(getPrevBalanceValue());
}
hash = (29 * hash) + getUnknownFields().hashCode();
@ -13239,7 +13239,7 @@ public final class MasterAdminProtos {
private int bitField0_;
// optional bool prevBalanceValue = 1;
// optional bool prev_balance_value = 1;
private boolean prevBalanceValue_ ;
public boolean hasPrevBalanceValue() {
return ((bitField0_ & 0x00000001) == 0x00000001);
@ -13574,7 +13574,7 @@ public final class MasterAdminProtos {
public interface CatalogScanResponseOrBuilder
extends com.google.protobuf.MessageOrBuilder {
// optional int32 scanResult = 1;
// optional int32 scan_result = 1;
boolean hasScanResult();
int getScanResult();
}
@ -13607,8 +13607,8 @@ public final class MasterAdminProtos {
}
private int bitField0_;
// optional int32 scanResult = 1;
public static final int SCANRESULT_FIELD_NUMBER = 1;
// optional int32 scan_result = 1;
public static final int SCAN_RESULT_FIELD_NUMBER = 1;
private int scanResult_;
public boolean hasScanResult() {
return ((bitField0_ & 0x00000001) == 0x00000001);
@ -13686,7 +13686,7 @@ public final class MasterAdminProtos {
int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode();
if (hasScanResult()) {
hash = (37 * hash) + SCANRESULT_FIELD_NUMBER;
hash = (37 * hash) + SCAN_RESULT_FIELD_NUMBER;
hash = (53 * hash) + getScanResult();
}
hash = (29 * hash) + getUnknownFields().hashCode();
@ -13910,7 +13910,7 @@ public final class MasterAdminProtos {
private int bitField0_;
// optional int32 scanResult = 1;
// optional int32 scan_result = 1;
private int scanResult_ ;
public boolean hasScanResult() {
return ((bitField0_ & 0x00000001) == 0x00000001);
@ -14324,7 +14324,7 @@ public final class MasterAdminProtos {
public interface EnableCatalogJanitorResponseOrBuilder
extends com.google.protobuf.MessageOrBuilder {
// optional bool prevValue = 1;
// optional bool prev_value = 1;
boolean hasPrevValue();
boolean getPrevValue();
}
@ -14357,8 +14357,8 @@ public final class MasterAdminProtos {
}
private int bitField0_;
// optional bool prevValue = 1;
public static final int PREVVALUE_FIELD_NUMBER = 1;
// optional bool prev_value = 1;
public static final int PREV_VALUE_FIELD_NUMBER = 1;
private boolean prevValue_;
public boolean hasPrevValue() {
return ((bitField0_ & 0x00000001) == 0x00000001);
@ -14436,7 +14436,7 @@ public final class MasterAdminProtos {
int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode();
if (hasPrevValue()) {
hash = (37 * hash) + PREVVALUE_FIELD_NUMBER;
hash = (37 * hash) + PREV_VALUE_FIELD_NUMBER;
hash = (53 * hash) + hashBoolean(getPrevValue());
}
hash = (29 * hash) + getUnknownFields().hashCode();
@ -14660,7 +14660,7 @@ public final class MasterAdminProtos {
private int bitField0_;
// optional bool prevValue = 1;
// optional bool prev_value = 1;
private boolean prevValue_ ;
public boolean hasPrevValue() {
return ((bitField0_ & 0x00000001) == 0x00000001);
@ -15847,7 +15847,7 @@ public final class MasterAdminProtos {
public interface TakeSnapshotResponseOrBuilder
extends com.google.protobuf.MessageOrBuilder {
// required int64 expectedTimeout = 1;
// required int64 expected_timeout = 1;
boolean hasExpectedTimeout();
long getExpectedTimeout();
}
@ -15880,8 +15880,8 @@ public final class MasterAdminProtos {
}
private int bitField0_;
// required int64 expectedTimeout = 1;
public static final int EXPECTEDTIMEOUT_FIELD_NUMBER = 1;
// required int64 expected_timeout = 1;
public static final int EXPECTED_TIMEOUT_FIELD_NUMBER = 1;
private long expectedTimeout_;
public boolean hasExpectedTimeout() {
return ((bitField0_ & 0x00000001) == 0x00000001);
@ -15963,7 +15963,7 @@ public final class MasterAdminProtos {
int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode();
if (hasExpectedTimeout()) {
hash = (37 * hash) + EXPECTEDTIMEOUT_FIELD_NUMBER;
hash = (37 * hash) + EXPECTED_TIMEOUT_FIELD_NUMBER;
hash = (53 * hash) + hashLong(getExpectedTimeout());
}
hash = (29 * hash) + getUnknownFields().hashCode();
@ -16191,7 +16191,7 @@ public final class MasterAdminProtos {
private int bitField0_;
// required int64 expectedTimeout = 1;
// required int64 expected_timeout = 1;
private long expectedTimeout_ ;
public boolean hasExpectedTimeout() {
return ((bitField0_ & 0x00000001) == 0x00000001);
@ -22752,116 +22752,116 @@ public final class MasterAdminProtos {
static {
java.lang.String[] descriptorData = {
"\n\021MasterAdmin.proto\032\014Master.proto\032\013hbase" +
".proto\032\014Client.proto\"R\n\020AddColumnRequest" +
"\022\021\n\ttableName\030\001 \002(\014\022+\n\016columnFamilies\030\002 " +
"\002(\0132\023.ColumnFamilySchema\"\023\n\021AddColumnRes" +
"ponse\"<\n\023DeleteColumnRequest\022\021\n\ttableNam" +
"e\030\001 \002(\014\022\022\n\ncolumnName\030\002 \002(\014\"\026\n\024DeleteCol" +
"umnResponse\"U\n\023ModifyColumnRequest\022\021\n\tta" +
"bleName\030\001 \002(\014\022+\n\016columnFamilies\030\002 \002(\0132\023." +
"ColumnFamilySchema\"\026\n\024ModifyColumnRespon" +
"se\"Z\n\021MoveRegionRequest\022 \n\006region\030\001 \002(\0132",
"\020.RegionSpecifier\022#\n\016destServerName\030\002 \001(" +
"\0132\013.ServerName\"\024\n\022MoveRegionResponse\"~\n\035" +
"DispatchMergingRegionsRequest\022!\n\007regionA" +
"\030\001 \002(\0132\020.RegionSpecifier\022!\n\007regionB\030\002 \002(" +
"\0132\020.RegionSpecifier\022\027\n\010forcible\030\003 \001(\010:\005f" +
"alse\" \n\036DispatchMergingRegionsResponse\"7" +
"\n\023AssignRegionRequest\022 \n\006region\030\001 \002(\0132\020." +
"RegionSpecifier\"\026\n\024AssignRegionResponse\"" +
"O\n\025UnassignRegionRequest\022 \n\006region\030\001 \002(\013" +
"2\020.RegionSpecifier\022\024\n\005force\030\002 \001(\010:\005false",
"\"\030\n\026UnassignRegionResponse\"8\n\024OfflineReg" +
"ionRequest\022 \n\006region\030\001 \002(\0132\020.RegionSpeci" +
"fier\"\027\n\025OfflineRegionResponse\"J\n\022CreateT" +
"ableRequest\022!\n\013tableSchema\030\001 \002(\0132\014.Table" +
"Schema\022\021\n\tsplitKeys\030\002 \003(\014\"\025\n\023CreateTable" +
"Response\"\'\n\022DeleteTableRequest\022\021\n\ttableN" +
"ame\030\001 \002(\014\"\025\n\023DeleteTableResponse\"\'\n\022Enab" +
"leTableRequest\022\021\n\ttableName\030\001 \002(\014\"\025\n\023Ena" +
"bleTableResponse\"(\n\023DisableTableRequest\022" +
"\021\n\ttableName\030\001 \002(\014\"\026\n\024DisableTableRespon",
"se\"J\n\022ModifyTableRequest\022\021\n\ttableName\030\001 " +
"\002(\014\022!\n\013tableSchema\030\002 \002(\0132\014.TableSchema\"\025" +
"\n\023ModifyTableResponse\"\021\n\017ShutdownRequest" +
"\"\022\n\020ShutdownResponse\"\023\n\021StopMasterReques" +
"t\"\024\n\022StopMasterResponse\"\020\n\016BalanceReques" +
"t\"&\n\017BalanceResponse\022\023\n\013balancerRan\030\001 \002(" +
"\010\"<\n\031SetBalancerRunningRequest\022\n\n\002on\030\001 \002" +
"(\010\022\023\n\013synchronous\030\002 \001(\010\"6\n\032SetBalancerRu" +
"nningResponse\022\030\n\020prevBalanceValue\030\001 \001(\010\"" +
"\024\n\022CatalogScanRequest\")\n\023CatalogScanResp",
"onse\022\022\n\nscanResult\030\001 \001(\005\"-\n\033EnableCatalo" +
"gJanitorRequest\022\016\n\006enable\030\001 \002(\010\"1\n\034Enabl" +
"eCatalogJanitorResponse\022\021\n\tprevValue\030\001 \001" +
"(\010\" \n\036IsCatalogJanitorEnabledRequest\"0\n\037" +
"IsCatalogJanitorEnabledResponse\022\r\n\005value" +
"\030\001 \002(\010\"=\n\023TakeSnapshotRequest\022&\n\010snapsho" +
"t\030\001 \002(\0132\024.SnapshotDescription\"/\n\024TakeSna" +
"pshotResponse\022\027\n\017expectedTimeout\030\001 \002(\003\"\025" +
"\n\023ListSnapshotRequest\"?\n\024ListSnapshotRes" +
"ponse\022\'\n\tsnapshots\030\001 \003(\0132\024.SnapshotDescr",
"iption\"?\n\025DeleteSnapshotRequest\022&\n\010snaps" +
"hot\030\001 \002(\0132\024.SnapshotDescription\"\030\n\026Delet" +
"eSnapshotResponse\"@\n\026RestoreSnapshotRequ" +
"est\022&\n\010snapshot\030\001 \002(\0132\024.SnapshotDescript" +
"ion\"\031\n\027RestoreSnapshotResponse\"?\n\025IsSnap" +
"shotDoneRequest\022&\n\010snapshot\030\001 \001(\0132\024.Snap" +
"shotDescription\"U\n\026IsSnapshotDoneRespons" +
"e\022\023\n\004done\030\001 \001(\010:\005false\022&\n\010snapshot\030\002 \001(\013" +
"2\024.SnapshotDescription\"F\n\034IsRestoreSnaps" +
"hotDoneRequest\022&\n\010snapshot\030\001 \001(\0132\024.Snaps",
"hotDescription\"3\n\035IsRestoreSnapshotDoneR" +
"esponse\022\022\n\004done\030\001 \001(\010:\004true2\305\016\n\022MasterAd" +
"minService\0222\n\taddColumn\022\021.AddColumnReque" +
"st\032\022.AddColumnResponse\022;\n\014deleteColumn\022\024" +
".DeleteColumnRequest\032\025.DeleteColumnRespo" +
"nse\022;\n\014modifyColumn\022\024.ModifyColumnReques" +
"t\032\025.ModifyColumnResponse\0225\n\nmoveRegion\022\022" +
".MoveRegionRequest\032\023.MoveRegionResponse\022" +
"Y\n\026dispatchMergingRegions\022\036.DispatchMerg" +
"ingRegionsRequest\032\037.DispatchMergingRegio",
"nsResponse\022;\n\014assignRegion\022\024.AssignRegio" +
"nRequest\032\025.AssignRegionResponse\022A\n\016unass" +
"ignRegion\022\026.UnassignRegionRequest\032\027.Unas" +
"signRegionResponse\022>\n\rofflineRegion\022\025.Of" +
"flineRegionRequest\032\026.OfflineRegionRespon" +
"se\0228\n\013deleteTable\022\023.DeleteTableRequest\032\024" +
".DeleteTableResponse\0228\n\013enableTable\022\023.En" +
"ableTableRequest\032\024.EnableTableResponse\022;" +
"\n\014disableTable\022\024.DisableTableRequest\032\025.D" +
"isableTableResponse\0228\n\013modifyTable\022\023.Mod",
"ifyTableRequest\032\024.ModifyTableResponse\0228\n" +
"\013createTable\022\023.CreateTableRequest\032\024.Crea" +
"teTableResponse\022/\n\010shutdown\022\020.ShutdownRe" +
"quest\032\021.ShutdownResponse\0225\n\nstopMaster\022\022" +
".StopMasterRequest\032\023.StopMasterResponse\022" +
",\n\007balance\022\017.BalanceRequest\032\020.BalanceRes" +
"ponse\022M\n\022setBalancerRunning\022\032.SetBalance" +
"rRunningRequest\032\033.SetBalancerRunningResp" +
"onse\022;\n\016runCatalogScan\022\023.CatalogScanRequ" +
"est\032\024.CatalogScanResponse\022S\n\024enableCatal",
"ogJanitor\022\034.EnableCatalogJanitorRequest\032" +
"\035.EnableCatalogJanitorResponse\022\\\n\027isCata" +
"logJanitorEnabled\022\037.IsCatalogJanitorEnab" +
"ledRequest\032 .IsCatalogJanitorEnabledResp" +
"onse\022L\n\021execMasterService\022\032.CoprocessorS" +
"erviceRequest\032\033.CoprocessorServiceRespon" +
"se\0227\n\010snapshot\022\024.TakeSnapshotRequest\032\025.T" +
"akeSnapshotResponse\022D\n\025getCompletedSnaps" +
"hots\022\024.ListSnapshotRequest\032\025.ListSnapsho" +
"tResponse\022A\n\016deleteSnapshot\022\026.DeleteSnap",
"shotRequest\032\027.DeleteSnapshotResponse\022A\n\016" +
"isSnapshotDone\022\026.IsSnapshotDoneRequest\032\027" +
".IsSnapshotDoneResponse\022D\n\017restoreSnapsh" +
"ot\022\027.RestoreSnapshotRequest\032\030.RestoreSna" +
"pshotResponse\022V\n\025isRestoreSnapshotDone\022\035" +
".IsRestoreSnapshotDoneRequest\032\036.IsRestor" +
"eSnapshotDoneResponse\022D\n\017isMasterRunning" +
"\022\027.IsMasterRunningRequest\032\030.IsMasterRunn" +
"ingResponseBG\n*org.apache.hadoop.hbase.p" +
"rotobuf.generatedB\021MasterAdminProtosH\001\210\001",
"\001\240\001\001"
".proto\032\014Client.proto\"T\n\020AddColumnRequest" +
"\022\022\n\ntable_name\030\001 \002(\014\022,\n\017column_families\030" +
"\002 \002(\0132\023.ColumnFamilySchema\"\023\n\021AddColumnR" +
"esponse\">\n\023DeleteColumnRequest\022\022\n\ntable_" +
"name\030\001 \002(\014\022\023\n\013column_name\030\002 \002(\014\"\026\n\024Delet" +
"eColumnResponse\"W\n\023ModifyColumnRequest\022\022" +
"\n\ntable_name\030\001 \002(\014\022,\n\017column_families\030\002 " +
"\002(\0132\023.ColumnFamilySchema\"\026\n\024ModifyColumn" +
"Response\"\\\n\021MoveRegionRequest\022 \n\006region\030",
"\001 \002(\0132\020.RegionSpecifier\022%\n\020dest_server_n" +
"ame\030\002 \001(\0132\013.ServerName\"\024\n\022MoveRegionResp" +
"onse\"\200\001\n\035DispatchMergingRegionsRequest\022\"" +
"\n\010region_a\030\001 \002(\0132\020.RegionSpecifier\022\"\n\010re" +
"gion_b\030\002 \002(\0132\020.RegionSpecifier\022\027\n\010forcib" +
"le\030\003 \001(\010:\005false\" \n\036DispatchMergingRegion" +
"sResponse\"7\n\023AssignRegionRequest\022 \n\006regi" +
"on\030\001 \002(\0132\020.RegionSpecifier\"\026\n\024AssignRegi" +
"onResponse\"O\n\025UnassignRegionRequest\022 \n\006r" +
"egion\030\001 \002(\0132\020.RegionSpecifier\022\024\n\005force\030\002",
" \001(\010:\005false\"\030\n\026UnassignRegionResponse\"8\n" +
"\024OfflineRegionRequest\022 \n\006region\030\001 \002(\0132\020." +
"RegionSpecifier\"\027\n\025OfflineRegionResponse" +
"\"L\n\022CreateTableRequest\022\"\n\014table_schema\030\001" +
" \002(\0132\014.TableSchema\022\022\n\nsplit_keys\030\002 \003(\014\"\025" +
"\n\023CreateTableResponse\"(\n\022DeleteTableRequ" +
"est\022\022\n\ntable_name\030\001 \002(\014\"\025\n\023DeleteTableRe" +
"sponse\"(\n\022EnableTableRequest\022\022\n\ntable_na" +
"me\030\001 \002(\014\"\025\n\023EnableTableResponse\")\n\023Disab" +
"leTableRequest\022\022\n\ntable_name\030\001 \002(\014\"\026\n\024Di",
"sableTableResponse\"L\n\022ModifyTableRequest" +
"\022\022\n\ntable_name\030\001 \002(\014\022\"\n\014table_schema\030\002 \002" +
"(\0132\014.TableSchema\"\025\n\023ModifyTableResponse\"" +
"\021\n\017ShutdownRequest\"\022\n\020ShutdownResponse\"\023" +
"\n\021StopMasterRequest\"\024\n\022StopMasterRespons" +
"e\"\020\n\016BalanceRequest\"\'\n\017BalanceResponse\022\024" +
"\n\014balancer_ran\030\001 \002(\010\"<\n\031SetBalancerRunni" +
"ngRequest\022\n\n\002on\030\001 \002(\010\022\023\n\013synchronous\030\002 \001" +
"(\010\"8\n\032SetBalancerRunningResponse\022\032\n\022prev" +
"_balance_value\030\001 \001(\010\"\024\n\022CatalogScanReque",
"st\"*\n\023CatalogScanResponse\022\023\n\013scan_result" +
"\030\001 \001(\005\"-\n\033EnableCatalogJanitorRequest\022\016\n" +
"\006enable\030\001 \002(\010\"2\n\034EnableCatalogJanitorRes" +
"ponse\022\022\n\nprev_value\030\001 \001(\010\" \n\036IsCatalogJa" +
"nitorEnabledRequest\"0\n\037IsCatalogJanitorE" +
"nabledResponse\022\r\n\005value\030\001 \002(\010\"=\n\023TakeSna" +
"pshotRequest\022&\n\010snapshot\030\001 \002(\0132\024.Snapsho" +
"tDescription\"0\n\024TakeSnapshotResponse\022\030\n\020" +
"expected_timeout\030\001 \002(\003\"\025\n\023ListSnapshotRe" +
"quest\"?\n\024ListSnapshotResponse\022\'\n\tsnapsho",
"ts\030\001 \003(\0132\024.SnapshotDescription\"?\n\025Delete" +
"SnapshotRequest\022&\n\010snapshot\030\001 \002(\0132\024.Snap" +
"shotDescription\"\030\n\026DeleteSnapshotRespons" +
"e\"@\n\026RestoreSnapshotRequest\022&\n\010snapshot\030" +
"\001 \002(\0132\024.SnapshotDescription\"\031\n\027RestoreSn" +
"apshotResponse\"?\n\025IsSnapshotDoneRequest\022" +
"&\n\010snapshot\030\001 \001(\0132\024.SnapshotDescription\"" +
"U\n\026IsSnapshotDoneResponse\022\023\n\004done\030\001 \001(\010:" +
"\005false\022&\n\010snapshot\030\002 \001(\0132\024.SnapshotDescr" +
"iption\"F\n\034IsRestoreSnapshotDoneRequest\022&",
"\n\010snapshot\030\001 \001(\0132\024.SnapshotDescription\"3" +
"\n\035IsRestoreSnapshotDoneResponse\022\022\n\004done\030" +
"\001 \001(\010:\004true2\305\016\n\022MasterAdminService\0222\n\tAd" +
"dColumn\022\021.AddColumnRequest\032\022.AddColumnRe" +
"sponse\022;\n\014DeleteColumn\022\024.DeleteColumnReq" +
"uest\032\025.DeleteColumnResponse\022;\n\014ModifyCol" +
"umn\022\024.ModifyColumnRequest\032\025.ModifyColumn" +
"Response\0225\n\nMoveRegion\022\022.MoveRegionReque" +
"st\032\023.MoveRegionResponse\022Y\n\026DispatchMergi" +
"ngRegions\022\036.DispatchMergingRegionsReques",
"t\032\037.DispatchMergingRegionsResponse\022;\n\014As" +
"signRegion\022\024.AssignRegionRequest\032\025.Assig" +
"nRegionResponse\022A\n\016UnassignRegion\022\026.Unas" +
"signRegionRequest\032\027.UnassignRegionRespon" +
"se\022>\n\rOfflineRegion\022\025.OfflineRegionReque" +
"st\032\026.OfflineRegionResponse\0228\n\013DeleteTabl" +
"e\022\023.DeleteTableRequest\032\024.DeleteTableResp" +
"onse\0228\n\013EnableTable\022\023.EnableTableRequest" +
"\032\024.EnableTableResponse\022;\n\014DisableTable\022\024" +
".DisableTableRequest\032\025.DisableTableRespo",
"nse\0228\n\013ModifyTable\022\023.ModifyTableRequest\032" +
"\024.ModifyTableResponse\0228\n\013CreateTable\022\023.C" +
"reateTableRequest\032\024.CreateTableResponse\022" +
"/\n\010Shutdown\022\020.ShutdownRequest\032\021.Shutdown" +
"Response\0225\n\nStopMaster\022\022.StopMasterReque" +
"st\032\023.StopMasterResponse\022,\n\007Balance\022\017.Bal" +
"anceRequest\032\020.BalanceResponse\022M\n\022SetBala" +
"ncerRunning\022\032.SetBalancerRunningRequest\032" +
"\033.SetBalancerRunningResponse\022;\n\016RunCatal" +
"ogScan\022\023.CatalogScanRequest\032\024.CatalogSca",
"nResponse\022S\n\024EnableCatalogJanitor\022\034.Enab" +
"leCatalogJanitorRequest\032\035.EnableCatalogJ" +
"anitorResponse\022\\\n\027IsCatalogJanitorEnable" +
"d\022\037.IsCatalogJanitorEnabledRequest\032 .IsC" +
"atalogJanitorEnabledResponse\022L\n\021ExecMast" +
"erService\022\032.CoprocessorServiceRequest\032\033." +
"CoprocessorServiceResponse\0227\n\010Snapshot\022\024" +
".TakeSnapshotRequest\032\025.TakeSnapshotRespo" +
"nse\022D\n\025GetCompletedSnapshots\022\024.ListSnaps" +
"hotRequest\032\025.ListSnapshotResponse\022A\n\016Del",
"eteSnapshot\022\026.DeleteSnapshotRequest\032\027.De" +
"leteSnapshotResponse\022A\n\016IsSnapshotDone\022\026" +
".IsSnapshotDoneRequest\032\027.IsSnapshotDoneR" +
"esponse\022D\n\017RestoreSnapshot\022\027.RestoreSnap" +
"shotRequest\032\030.RestoreSnapshotResponse\022V\n" +
"\025IsRestoreSnapshotDone\022\035.IsRestoreSnapsh" +
"otDoneRequest\032\036.IsRestoreSnapshotDoneRes" +
"ponse\022D\n\017IsMasterRunning\022\027.IsMasterRunni" +
"ngRequest\032\030.IsMasterRunningResponseBG\n*o" +
"rg.apache.hadoop.hbase.protobuf.generate",
"dB\021MasterAdminProtosH\001\210\001\001\240\001\001"
};
com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() {

View File

@ -11,7 +11,7 @@ public final class MasterMonitorProtos {
public interface GetSchemaAlterStatusRequestOrBuilder
extends com.google.protobuf.MessageOrBuilder {
// required bytes tableName = 1;
// required bytes table_name = 1;
boolean hasTableName();
com.google.protobuf.ByteString getTableName();
}
@ -44,8 +44,8 @@ public final class MasterMonitorProtos {
}
private int bitField0_;
// required bytes tableName = 1;
public static final int TABLENAME_FIELD_NUMBER = 1;
// required bytes table_name = 1;
public static final int TABLE_NAME_FIELD_NUMBER = 1;
private com.google.protobuf.ByteString tableName_;
public boolean hasTableName() {
return ((bitField0_ & 0x00000001) == 0x00000001);
@ -127,7 +127,7 @@ public final class MasterMonitorProtos {
int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode();
if (hasTableName()) {
hash = (37 * hash) + TABLENAME_FIELD_NUMBER;
hash = (37 * hash) + TABLE_NAME_FIELD_NUMBER;
hash = (53 * hash) + getTableName().hashCode();
}
hash = (29 * hash) + getUnknownFields().hashCode();
@ -355,7 +355,7 @@ public final class MasterMonitorProtos {
private int bitField0_;
// required bytes tableName = 1;
// required bytes table_name = 1;
private com.google.protobuf.ByteString tableName_ = com.google.protobuf.ByteString.EMPTY;
public boolean hasTableName() {
return ((bitField0_ & 0x00000001) == 0x00000001);
@ -393,11 +393,11 @@ public final class MasterMonitorProtos {
public interface GetSchemaAlterStatusResponseOrBuilder
extends com.google.protobuf.MessageOrBuilder {
// optional uint32 yetToUpdateRegions = 1;
// optional uint32 yet_to_update_regions = 1;
boolean hasYetToUpdateRegions();
int getYetToUpdateRegions();
// optional uint32 totalRegions = 2;
// optional uint32 total_regions = 2;
boolean hasTotalRegions();
int getTotalRegions();
}
@ -430,8 +430,8 @@ public final class MasterMonitorProtos {
}
private int bitField0_;
// optional uint32 yetToUpdateRegions = 1;
public static final int YETTOUPDATEREGIONS_FIELD_NUMBER = 1;
// optional uint32 yet_to_update_regions = 1;
public static final int YET_TO_UPDATE_REGIONS_FIELD_NUMBER = 1;
private int yetToUpdateRegions_;
public boolean hasYetToUpdateRegions() {
return ((bitField0_ & 0x00000001) == 0x00000001);
@ -440,8 +440,8 @@ public final class MasterMonitorProtos {
return yetToUpdateRegions_;
}
// optional uint32 totalRegions = 2;
public static final int TOTALREGIONS_FIELD_NUMBER = 2;
// optional uint32 total_regions = 2;
public static final int TOTAL_REGIONS_FIELD_NUMBER = 2;
private int totalRegions_;
public boolean hasTotalRegions() {
return ((bitField0_ & 0x00000002) == 0x00000002);
@ -532,11 +532,11 @@ public final class MasterMonitorProtos {
int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode();
if (hasYetToUpdateRegions()) {
hash = (37 * hash) + YETTOUPDATEREGIONS_FIELD_NUMBER;
hash = (37 * hash) + YET_TO_UPDATE_REGIONS_FIELD_NUMBER;
hash = (53 * hash) + getYetToUpdateRegions();
}
if (hasTotalRegions()) {
hash = (37 * hash) + TOTALREGIONS_FIELD_NUMBER;
hash = (37 * hash) + TOTAL_REGIONS_FIELD_NUMBER;
hash = (53 * hash) + getTotalRegions();
}
hash = (29 * hash) + getUnknownFields().hashCode();
@ -774,7 +774,7 @@ public final class MasterMonitorProtos {
private int bitField0_;
// optional uint32 yetToUpdateRegions = 1;
// optional uint32 yet_to_update_regions = 1;
private int yetToUpdateRegions_ ;
public boolean hasYetToUpdateRegions() {
return ((bitField0_ & 0x00000001) == 0x00000001);
@ -795,7 +795,7 @@ public final class MasterMonitorProtos {
return this;
}
// optional uint32 totalRegions = 2;
// optional uint32 total_regions = 2;
private int totalRegions_ ;
public boolean hasTotalRegions() {
return ((bitField0_ & 0x00000002) == 0x00000002);
@ -830,7 +830,7 @@ public final class MasterMonitorProtos {
public interface GetTableDescriptorsRequestOrBuilder
extends com.google.protobuf.MessageOrBuilder {
// repeated string tableNames = 1;
// repeated string table_names = 1;
java.util.List<String> getTableNamesList();
int getTableNamesCount();
String getTableNames(int index);
@ -863,8 +863,8 @@ public final class MasterMonitorProtos {
return org.apache.hadoop.hbase.protobuf.generated.MasterMonitorProtos.internal_static_GetTableDescriptorsRequest_fieldAccessorTable;
}
// repeated string tableNames = 1;
public static final int TABLENAMES_FIELD_NUMBER = 1;
// repeated string table_names = 1;
public static final int TABLE_NAMES_FIELD_NUMBER = 1;
private com.google.protobuf.LazyStringList tableNames_;
public java.util.List<String>
getTableNamesList() {
@ -948,7 +948,7 @@ public final class MasterMonitorProtos {
int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode();
if (getTableNamesCount() > 0) {
hash = (37 * hash) + TABLENAMES_FIELD_NUMBER;
hash = (37 * hash) + TABLE_NAMES_FIELD_NUMBER;
hash = (53 * hash) + getTableNamesList().hashCode();
}
hash = (29 * hash) + getUnknownFields().hashCode();
@ -1179,7 +1179,7 @@ public final class MasterMonitorProtos {
private int bitField0_;
// repeated string tableNames = 1;
// repeated string table_names = 1;
private com.google.protobuf.LazyStringList tableNames_ = com.google.protobuf.LazyStringArrayList.EMPTY;
private void ensureTableNamesIsMutable() {
if (!((bitField0_ & 0x00000001) == 0x00000001)) {
@ -1249,7 +1249,7 @@ public final class MasterMonitorProtos {
public interface GetTableDescriptorsResponseOrBuilder
extends com.google.protobuf.MessageOrBuilder {
// repeated .TableSchema tableSchema = 1;
// repeated .TableSchema table_schema = 1;
java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema>
getTableSchemaList();
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema getTableSchema(int index);
@ -1287,8 +1287,8 @@ public final class MasterMonitorProtos {
return org.apache.hadoop.hbase.protobuf.generated.MasterMonitorProtos.internal_static_GetTableDescriptorsResponse_fieldAccessorTable;
}
// repeated .TableSchema tableSchema = 1;
public static final int TABLESCHEMA_FIELD_NUMBER = 1;
// repeated .TableSchema table_schema = 1;
public static final int TABLE_SCHEMA_FIELD_NUMBER = 1;
private java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema> tableSchema_;
public java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema> getTableSchemaList() {
return tableSchema_;
@ -1380,7 +1380,7 @@ public final class MasterMonitorProtos {
int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode();
if (getTableSchemaCount() > 0) {
hash = (37 * hash) + TABLESCHEMA_FIELD_NUMBER;
hash = (37 * hash) + TABLE_SCHEMA_FIELD_NUMBER;
hash = (53 * hash) + getTableSchemaList().hashCode();
}
hash = (29 * hash) + getUnknownFields().hashCode();
@ -1642,7 +1642,7 @@ public final class MasterMonitorProtos {
private int bitField0_;
// repeated .TableSchema tableSchema = 1;
// repeated .TableSchema table_schema = 1;
private java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema> tableSchema_ =
java.util.Collections.emptyList();
private void ensureTableSchemaIsMutable() {
@ -2142,7 +2142,7 @@ public final class MasterMonitorProtos {
public interface GetClusterStatusResponseOrBuilder
extends com.google.protobuf.MessageOrBuilder {
// required .ClusterStatus clusterStatus = 1;
// required .ClusterStatus cluster_status = 1;
boolean hasClusterStatus();
org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.ClusterStatus getClusterStatus();
org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.ClusterStatusOrBuilder getClusterStatusOrBuilder();
@ -2176,8 +2176,8 @@ public final class MasterMonitorProtos {
}
private int bitField0_;
// required .ClusterStatus clusterStatus = 1;
public static final int CLUSTERSTATUS_FIELD_NUMBER = 1;
// required .ClusterStatus cluster_status = 1;
public static final int CLUSTER_STATUS_FIELD_NUMBER = 1;
private org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.ClusterStatus clusterStatus_;
public boolean hasClusterStatus() {
return ((bitField0_ & 0x00000001) == 0x00000001);
@ -2266,7 +2266,7 @@ public final class MasterMonitorProtos {
int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode();
if (hasClusterStatus()) {
hash = (37 * hash) + CLUSTERSTATUS_FIELD_NUMBER;
hash = (37 * hash) + CLUSTER_STATUS_FIELD_NUMBER;
hash = (53 * hash) + getClusterStatus().hashCode();
}
hash = (29 * hash) + getUnknownFields().hashCode();
@ -2511,7 +2511,7 @@ public final class MasterMonitorProtos {
private int bitField0_;
// required .ClusterStatus clusterStatus = 1;
// required .ClusterStatus cluster_status = 1;
private org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.ClusterStatus clusterStatus_ = org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.ClusterStatus.getDefaultInstance();
private com.google.protobuf.SingleFieldBuilder<
org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.ClusterStatus, org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.ClusterStatus.Builder, org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.ClusterStatusOrBuilder> clusterStatusBuilder_;
@ -3068,26 +3068,26 @@ public final class MasterMonitorProtos {
static {
java.lang.String[] descriptorData = {
"\n\023MasterMonitor.proto\032\014Master.proto\032\013hba" +
"se.proto\032\023ClusterStatus.proto\"0\n\033GetSche" +
"maAlterStatusRequest\022\021\n\ttableName\030\001 \002(\014\"" +
"P\n\034GetSchemaAlterStatusResponse\022\032\n\022yetTo" +
"UpdateRegions\030\001 \001(\r\022\024\n\014totalRegions\030\002 \001(" +
"\r\"0\n\032GetTableDescriptorsRequest\022\022\n\ntable" +
"Names\030\001 \003(\t\"@\n\033GetTableDescriptorsRespon" +
"se\022!\n\013tableSchema\030\001 \003(\0132\014.TableSchema\"\031\n" +
"\027GetClusterStatusRequest\"A\n\030GetClusterSt" +
"atusResponse\022%\n\rclusterStatus\030\001 \002(\0132\016.Cl",
"usterStatus2\314\002\n\024MasterMonitorService\022S\n\024" +
"getSchemaAlterStatus\022\034.GetSchemaAlterSta" +
"tusRequest\032\035.GetSchemaAlterStatusRespons" +
"e\022P\n\023getTableDescriptors\022\033.GetTableDescr" +
"iptorsRequest\032\034.GetTableDescriptorsRespo" +
"nse\022G\n\020getClusterStatus\022\030.GetClusterStat" +
"usRequest\032\031.GetClusterStatusResponse\022D\n\017" +
"isMasterRunning\022\027.IsMasterRunningRequest" +
"\032\030.IsMasterRunningResponseBI\n*org.apache" +
".hadoop.hbase.protobuf.generatedB\023Master",
"MonitorProtosH\001\210\001\001\240\001\001"
"se.proto\032\023ClusterStatus.proto\"1\n\033GetSche" +
"maAlterStatusRequest\022\022\n\ntable_name\030\001 \002(\014" +
"\"T\n\034GetSchemaAlterStatusResponse\022\035\n\025yet_" +
"to_update_regions\030\001 \001(\r\022\025\n\rtotal_regions" +
"\030\002 \001(\r\"1\n\032GetTableDescriptorsRequest\022\023\n\013" +
"table_names\030\001 \003(\t\"A\n\033GetTableDescriptors" +
"Response\022\"\n\014table_schema\030\001 \003(\0132\014.TableSc" +
"hema\"\031\n\027GetClusterStatusRequest\"B\n\030GetCl" +
"usterStatusResponse\022&\n\016cluster_status\030\001 ",
"\002(\0132\016.ClusterStatus2\314\002\n\024MasterMonitorSer" +
"vice\022S\n\024GetSchemaAlterStatus\022\034.GetSchema" +
"AlterStatusRequest\032\035.GetSchemaAlterStatu" +
"sResponse\022P\n\023GetTableDescriptors\022\033.GetTa" +
"bleDescriptorsRequest\032\034.GetTableDescript" +
"orsResponse\022G\n\020GetClusterStatus\022\030.GetClu" +
"sterStatusRequest\032\031.GetClusterStatusResp" +
"onse\022D\n\017IsMasterRunning\022\027.IsMasterRunnin" +
"gRequest\032\030.IsMasterRunningResponseBI\n*or" +
"g.apache.hadoop.hbase.protobuf.generated",
"B\023MasterMonitorProtosH\001\210\001\001\240\001\001"
};
com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() {

View File

@ -311,7 +311,7 @@ public final class MasterProtos {
public interface IsMasterRunningResponseOrBuilder
extends com.google.protobuf.MessageOrBuilder {
// required bool isMasterRunning = 1;
// required bool is_master_running = 1;
boolean hasIsMasterRunning();
boolean getIsMasterRunning();
}
@ -344,8 +344,8 @@ public final class MasterProtos {
}
private int bitField0_;
// required bool isMasterRunning = 1;
public static final int ISMASTERRUNNING_FIELD_NUMBER = 1;
// required bool is_master_running = 1;
public static final int IS_MASTER_RUNNING_FIELD_NUMBER = 1;
private boolean isMasterRunning_;
public boolean hasIsMasterRunning() {
return ((bitField0_ & 0x00000001) == 0x00000001);
@ -427,7 +427,7 @@ public final class MasterProtos {
int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode();
if (hasIsMasterRunning()) {
hash = (37 * hash) + ISMASTERRUNNING_FIELD_NUMBER;
hash = (37 * hash) + IS_MASTER_RUNNING_FIELD_NUMBER;
hash = (53 * hash) + hashBoolean(getIsMasterRunning());
}
hash = (29 * hash) + getUnknownFields().hashCode();
@ -655,7 +655,7 @@ public final class MasterProtos {
private int bitField0_;
// required bool isMasterRunning = 1;
// required bool is_master_running = 1;
private boolean isMasterRunning_ ;
public boolean hasIsMasterRunning() {
return ((bitField0_ & 0x00000001) == 0x00000001);
@ -928,12 +928,12 @@ public final class MasterProtos {
static {
java.lang.String[] descriptorData = {
"\n\014Master.proto\"\030\n\026IsMasterRunningRequest" +
"\"2\n\027IsMasterRunningResponse\022\027\n\017isMasterR" +
"unning\030\001 \002(\0102U\n\rMasterService\022D\n\017isMaste" +
"rRunning\022\027.IsMasterRunningRequest\032\030.IsMa" +
"sterRunningResponseBB\n*org.apache.hadoop" +
".hbase.protobuf.generatedB\014MasterProtosH" +
"\001\210\001\001\240\001\001"
"\"4\n\027IsMasterRunningResponse\022\031\n\021is_master" +
"_running\030\001 \002(\0102U\n\rMasterService\022D\n\017IsMas" +
"terRunning\022\027.IsMasterRunningRequest\032\030.Is" +
"MasterRunningResponseBB\n*org.apache.hado" +
"op.hbase.protobuf.generatedB\014MasterProto" +
"sH\001\210\001\001\240\001\001"
};
com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() {

View File

@ -11,7 +11,7 @@ public final class MultiRowMutation {
public interface MultiMutateRequestOrBuilder
extends com.google.protobuf.MessageOrBuilder {
// repeated .MutationProto mutationRequest = 1;
// repeated .MutationProto mutation_request = 1;
java.util.List<org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MutationProto>
getMutationRequestList();
org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MutationProto getMutationRequest(int index);
@ -49,8 +49,8 @@ public final class MultiRowMutation {
return org.apache.hadoop.hbase.protobuf.generated.MultiRowMutation.internal_static_MultiMutateRequest_fieldAccessorTable;
}
// repeated .MutationProto mutationRequest = 1;
public static final int MUTATIONREQUEST_FIELD_NUMBER = 1;
// repeated .MutationProto mutation_request = 1;
public static final int MUTATION_REQUEST_FIELD_NUMBER = 1;
private java.util.List<org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MutationProto> mutationRequest_;
public java.util.List<org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MutationProto> getMutationRequestList() {
return mutationRequest_;
@ -142,7 +142,7 @@ public final class MultiRowMutation {
int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode();
if (getMutationRequestCount() > 0) {
hash = (37 * hash) + MUTATIONREQUEST_FIELD_NUMBER;
hash = (37 * hash) + MUTATION_REQUEST_FIELD_NUMBER;
hash = (53 * hash) + getMutationRequestList().hashCode();
}
hash = (29 * hash) + getUnknownFields().hashCode();
@ -404,7 +404,7 @@ public final class MultiRowMutation {
private int bitField0_;
// repeated .MutationProto mutationRequest = 1;
// repeated .MutationProto mutation_request = 1;
private java.util.List<org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MutationProto> mutationRequest_ =
java.util.Collections.emptyList();
private void ensureMutationRequestIsMutable() {
@ -1141,13 +1141,14 @@ public final class MultiRowMutation {
descriptor;
static {
java.lang.String[] descriptorData = {
"\n\026MultiRowMutation.proto\032\014Client.proto\"=" +
"\n\022MultiMutateRequest\022\'\n\017mutationRequest\030" +
"\001 \003(\0132\016.MutationProto\"\025\n\023MultiMutateResp" +
"onse2R\n\027MultiRowMutationService\0227\n\nmutat" +
"eRows\022\023.MultiMutateRequest\032\024.MultiMutate" +
"ResponseBF\n*org.apache.hadoop.hbase.prot" +
"obuf.generatedB\020MultiRowMutationH\001\210\001\001\240\001\001"
"\n\026MultiRowMutation.proto\032\014Client.proto\">" +
"\n\022MultiMutateRequest\022(\n\020mutation_request" +
"\030\001 \003(\0132\016.MutationProto\"\025\n\023MultiMutateRes" +
"ponse2R\n\027MultiRowMutationService\0227\n\nMuta" +
"teRows\022\023.MultiMutateRequest\032\024.MultiMutat" +
"eResponseBF\n*org.apache.hadoop.hbase.pro" +
"tobuf.generatedB\020MultiRowMutationH\001\210\001\001\240\001" +
"\001"
};
com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() {

View File

@ -11,11 +11,11 @@ public final class RPCProtos {
public interface UserInformationOrBuilder
extends com.google.protobuf.MessageOrBuilder {
// required string effectiveUser = 1;
// required string effective_user = 1;
boolean hasEffectiveUser();
String getEffectiveUser();
// optional string realUser = 2;
// optional string real_user = 2;
boolean hasRealUser();
String getRealUser();
}
@ -48,8 +48,8 @@ public final class RPCProtos {
}
private int bitField0_;
// required string effectiveUser = 1;
public static final int EFFECTIVEUSER_FIELD_NUMBER = 1;
// required string effective_user = 1;
public static final int EFFECTIVE_USER_FIELD_NUMBER = 1;
private java.lang.Object effectiveUser_;
public boolean hasEffectiveUser() {
return ((bitField0_ & 0x00000001) == 0x00000001);
@ -80,8 +80,8 @@ public final class RPCProtos {
}
}
// optional string realUser = 2;
public static final int REALUSER_FIELD_NUMBER = 2;
// optional string real_user = 2;
public static final int REAL_USER_FIELD_NUMBER = 2;
private java.lang.Object realUser_;
public boolean hasRealUser() {
return ((bitField0_ & 0x00000002) == 0x00000002);
@ -198,11 +198,11 @@ public final class RPCProtos {
int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode();
if (hasEffectiveUser()) {
hash = (37 * hash) + EFFECTIVEUSER_FIELD_NUMBER;
hash = (37 * hash) + EFFECTIVE_USER_FIELD_NUMBER;
hash = (53 * hash) + getEffectiveUser().hashCode();
}
if (hasRealUser()) {
hash = (37 * hash) + REALUSER_FIELD_NUMBER;
hash = (37 * hash) + REAL_USER_FIELD_NUMBER;
hash = (53 * hash) + getRealUser().hashCode();
}
hash = (29 * hash) + getUnknownFields().hashCode();
@ -444,7 +444,7 @@ public final class RPCProtos {
private int bitField0_;
// required string effectiveUser = 1;
// required string effective_user = 1;
private java.lang.Object effectiveUser_ = "";
public boolean hasEffectiveUser() {
return ((bitField0_ & 0x00000001) == 0x00000001);
@ -480,7 +480,7 @@ public final class RPCProtos {
onChanged();
}
// optional string realUser = 2;
// optional string real_user = 2;
private java.lang.Object realUser_ = "";
public boolean hasRealUser() {
return ((bitField0_ & 0x00000002) == 0x00000002);
@ -530,20 +530,20 @@ public final class RPCProtos {
public interface ConnectionHeaderOrBuilder
extends com.google.protobuf.MessageOrBuilder {
// optional .UserInformation userInfo = 1;
// optional .UserInformation user_info = 1;
boolean hasUserInfo();
org.apache.hadoop.hbase.protobuf.generated.RPCProtos.UserInformation getUserInfo();
org.apache.hadoop.hbase.protobuf.generated.RPCProtos.UserInformationOrBuilder getUserInfoOrBuilder();
// optional string serviceName = 2;
// optional string service_name = 2;
boolean hasServiceName();
String getServiceName();
// optional string cellBlockCodecClass = 3 [default = "org.apache.hadoop.hbase.codec.KeyValueCodec"];
// optional string cell_block_codec_class = 3 [default = "org.apache.hadoop.hbase.codec.KeyValueCodec"];
boolean hasCellBlockCodecClass();
String getCellBlockCodecClass();
// optional string cellBlockCompressorClass = 4;
// optional string cell_block_compressor_class = 4;
boolean hasCellBlockCompressorClass();
String getCellBlockCompressorClass();
}
@ -576,8 +576,8 @@ public final class RPCProtos {
}
private int bitField0_;
// optional .UserInformation userInfo = 1;
public static final int USERINFO_FIELD_NUMBER = 1;
// optional .UserInformation user_info = 1;
public static final int USER_INFO_FIELD_NUMBER = 1;
private org.apache.hadoop.hbase.protobuf.generated.RPCProtos.UserInformation userInfo_;
public boolean hasUserInfo() {
return ((bitField0_ & 0x00000001) == 0x00000001);
@ -589,8 +589,8 @@ public final class RPCProtos {
return userInfo_;
}
// optional string serviceName = 2;
public static final int SERVICENAME_FIELD_NUMBER = 2;
// optional string service_name = 2;
public static final int SERVICE_NAME_FIELD_NUMBER = 2;
private java.lang.Object serviceName_;
public boolean hasServiceName() {
return ((bitField0_ & 0x00000002) == 0x00000002);
@ -621,8 +621,8 @@ public final class RPCProtos {
}
}
// optional string cellBlockCodecClass = 3 [default = "org.apache.hadoop.hbase.codec.KeyValueCodec"];
public static final int CELLBLOCKCODECCLASS_FIELD_NUMBER = 3;
// optional string cell_block_codec_class = 3 [default = "org.apache.hadoop.hbase.codec.KeyValueCodec"];
public static final int CELL_BLOCK_CODEC_CLASS_FIELD_NUMBER = 3;
private java.lang.Object cellBlockCodecClass_;
public boolean hasCellBlockCodecClass() {
return ((bitField0_ & 0x00000004) == 0x00000004);
@ -653,8 +653,8 @@ public final class RPCProtos {
}
}
// optional string cellBlockCompressorClass = 4;
public static final int CELLBLOCKCOMPRESSORCLASS_FIELD_NUMBER = 4;
// optional string cell_block_compressor_class = 4;
public static final int CELL_BLOCK_COMPRESSOR_CLASS_FIELD_NUMBER = 4;
private java.lang.Object cellBlockCompressorClass_;
public boolean hasCellBlockCompressorClass() {
return ((bitField0_ & 0x00000008) == 0x00000008);
@ -799,19 +799,19 @@ public final class RPCProtos {
int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode();
if (hasUserInfo()) {
hash = (37 * hash) + USERINFO_FIELD_NUMBER;
hash = (37 * hash) + USER_INFO_FIELD_NUMBER;
hash = (53 * hash) + getUserInfo().hashCode();
}
if (hasServiceName()) {
hash = (37 * hash) + SERVICENAME_FIELD_NUMBER;
hash = (37 * hash) + SERVICE_NAME_FIELD_NUMBER;
hash = (53 * hash) + getServiceName().hashCode();
}
if (hasCellBlockCodecClass()) {
hash = (37 * hash) + CELLBLOCKCODECCLASS_FIELD_NUMBER;
hash = (37 * hash) + CELL_BLOCK_CODEC_CLASS_FIELD_NUMBER;
hash = (53 * hash) + getCellBlockCodecClass().hashCode();
}
if (hasCellBlockCompressorClass()) {
hash = (37 * hash) + CELLBLOCKCOMPRESSORCLASS_FIELD_NUMBER;
hash = (37 * hash) + CELL_BLOCK_COMPRESSOR_CLASS_FIELD_NUMBER;
hash = (53 * hash) + getCellBlockCompressorClass().hashCode();
}
hash = (29 * hash) + getUnknownFields().hashCode();
@ -1096,7 +1096,7 @@ public final class RPCProtos {
private int bitField0_;
// optional .UserInformation userInfo = 1;
// optional .UserInformation user_info = 1;
private org.apache.hadoop.hbase.protobuf.generated.RPCProtos.UserInformation userInfo_ = org.apache.hadoop.hbase.protobuf.generated.RPCProtos.UserInformation.getDefaultInstance();
private com.google.protobuf.SingleFieldBuilder<
org.apache.hadoop.hbase.protobuf.generated.RPCProtos.UserInformation, org.apache.hadoop.hbase.protobuf.generated.RPCProtos.UserInformation.Builder, org.apache.hadoop.hbase.protobuf.generated.RPCProtos.UserInformationOrBuilder> userInfoBuilder_;
@ -1186,7 +1186,7 @@ public final class RPCProtos {
return userInfoBuilder_;
}
// optional string serviceName = 2;
// optional string service_name = 2;
private java.lang.Object serviceName_ = "";
public boolean hasServiceName() {
return ((bitField0_ & 0x00000002) == 0x00000002);
@ -1222,7 +1222,7 @@ public final class RPCProtos {
onChanged();
}
// optional string cellBlockCodecClass = 3 [default = "org.apache.hadoop.hbase.codec.KeyValueCodec"];
// optional string cell_block_codec_class = 3 [default = "org.apache.hadoop.hbase.codec.KeyValueCodec"];
private java.lang.Object cellBlockCodecClass_ = "org.apache.hadoop.hbase.codec.KeyValueCodec";
public boolean hasCellBlockCodecClass() {
return ((bitField0_ & 0x00000004) == 0x00000004);
@ -1258,7 +1258,7 @@ public final class RPCProtos {
onChanged();
}
// optional string cellBlockCompressorClass = 4;
// optional string cell_block_compressor_class = 4;
private java.lang.Object cellBlockCompressorClass_ = "";
public boolean hasCellBlockCompressorClass() {
return ((bitField0_ & 0x00000008) == 0x00000008);
@ -1679,11 +1679,11 @@ public final class RPCProtos {
public interface ExceptionResponseOrBuilder
extends com.google.protobuf.MessageOrBuilder {
// optional string exceptionClassName = 1;
// optional string exception_class_name = 1;
boolean hasExceptionClassName();
String getExceptionClassName();
// optional string stackTrace = 2;
// optional string stack_trace = 2;
boolean hasStackTrace();
String getStackTrace();
@ -1695,7 +1695,7 @@ public final class RPCProtos {
boolean hasPort();
int getPort();
// optional bool doNotRetry = 5;
// optional bool do_not_retry = 5;
boolean hasDoNotRetry();
boolean getDoNotRetry();
}
@ -1728,8 +1728,8 @@ public final class RPCProtos {
}
private int bitField0_;
// optional string exceptionClassName = 1;
public static final int EXCEPTIONCLASSNAME_FIELD_NUMBER = 1;
// optional string exception_class_name = 1;
public static final int EXCEPTION_CLASS_NAME_FIELD_NUMBER = 1;
private java.lang.Object exceptionClassName_;
public boolean hasExceptionClassName() {
return ((bitField0_ & 0x00000001) == 0x00000001);
@ -1760,8 +1760,8 @@ public final class RPCProtos {
}
}
// optional string stackTrace = 2;
public static final int STACKTRACE_FIELD_NUMBER = 2;
// optional string stack_trace = 2;
public static final int STACK_TRACE_FIELD_NUMBER = 2;
private java.lang.Object stackTrace_;
public boolean hasStackTrace() {
return ((bitField0_ & 0x00000002) == 0x00000002);
@ -1834,8 +1834,8 @@ public final class RPCProtos {
return port_;
}
// optional bool doNotRetry = 5;
public static final int DONOTRETRY_FIELD_NUMBER = 5;
// optional bool do_not_retry = 5;
public static final int DO_NOT_RETRY_FIELD_NUMBER = 5;
private boolean doNotRetry_;
public boolean hasDoNotRetry() {
return ((bitField0_ & 0x00000010) == 0x00000010);
@ -1965,11 +1965,11 @@ public final class RPCProtos {
int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode();
if (hasExceptionClassName()) {
hash = (37 * hash) + EXCEPTIONCLASSNAME_FIELD_NUMBER;
hash = (37 * hash) + EXCEPTION_CLASS_NAME_FIELD_NUMBER;
hash = (53 * hash) + getExceptionClassName().hashCode();
}
if (hasStackTrace()) {
hash = (37 * hash) + STACKTRACE_FIELD_NUMBER;
hash = (37 * hash) + STACK_TRACE_FIELD_NUMBER;
hash = (53 * hash) + getStackTrace().hashCode();
}
if (hasHostname()) {
@ -1981,7 +1981,7 @@ public final class RPCProtos {
hash = (53 * hash) + getPort();
}
if (hasDoNotRetry()) {
hash = (37 * hash) + DONOTRETRY_FIELD_NUMBER;
hash = (37 * hash) + DO_NOT_RETRY_FIELD_NUMBER;
hash = (53 * hash) + hashBoolean(getDoNotRetry());
}
hash = (29 * hash) + getUnknownFields().hashCode();
@ -2261,7 +2261,7 @@ public final class RPCProtos {
private int bitField0_;
// optional string exceptionClassName = 1;
// optional string exception_class_name = 1;
private java.lang.Object exceptionClassName_ = "";
public boolean hasExceptionClassName() {
return ((bitField0_ & 0x00000001) == 0x00000001);
@ -2297,7 +2297,7 @@ public final class RPCProtos {
onChanged();
}
// optional string stackTrace = 2;
// optional string stack_trace = 2;
private java.lang.Object stackTrace_ = "";
public boolean hasStackTrace() {
return ((bitField0_ & 0x00000002) == 0x00000002);
@ -2390,7 +2390,7 @@ public final class RPCProtos {
return this;
}
// optional bool doNotRetry = 5;
// optional bool do_not_retry = 5;
private boolean doNotRetry_ ;
public boolean hasDoNotRetry() {
return ((bitField0_ & 0x00000010) == 0x00000010);
@ -2425,24 +2425,24 @@ public final class RPCProtos {
public interface RequestHeaderOrBuilder
extends com.google.protobuf.MessageOrBuilder {
// optional uint32 callId = 1;
// optional uint32 call_id = 1;
boolean hasCallId();
int getCallId();
// optional .RPCTInfo traceInfo = 2;
// optional .RPCTInfo trace_info = 2;
boolean hasTraceInfo();
org.apache.hadoop.hbase.protobuf.generated.Tracing.RPCTInfo getTraceInfo();
org.apache.hadoop.hbase.protobuf.generated.Tracing.RPCTInfoOrBuilder getTraceInfoOrBuilder();
// optional string methodName = 3;
// optional string method_name = 3;
boolean hasMethodName();
String getMethodName();
// optional bool requestParam = 4;
// optional bool request_param = 4;
boolean hasRequestParam();
boolean getRequestParam();
// optional .CellBlockMeta cellBlockMeta = 5;
// optional .CellBlockMeta cell_block_meta = 5;
boolean hasCellBlockMeta();
org.apache.hadoop.hbase.protobuf.generated.RPCProtos.CellBlockMeta getCellBlockMeta();
org.apache.hadoop.hbase.protobuf.generated.RPCProtos.CellBlockMetaOrBuilder getCellBlockMetaOrBuilder();
@ -2476,8 +2476,8 @@ public final class RPCProtos {
}
private int bitField0_;
// optional uint32 callId = 1;
public static final int CALLID_FIELD_NUMBER = 1;
// optional uint32 call_id = 1;
public static final int CALL_ID_FIELD_NUMBER = 1;
private int callId_;
public boolean hasCallId() {
return ((bitField0_ & 0x00000001) == 0x00000001);
@ -2486,8 +2486,8 @@ public final class RPCProtos {
return callId_;
}
// optional .RPCTInfo traceInfo = 2;
public static final int TRACEINFO_FIELD_NUMBER = 2;
// optional .RPCTInfo trace_info = 2;
public static final int TRACE_INFO_FIELD_NUMBER = 2;
private org.apache.hadoop.hbase.protobuf.generated.Tracing.RPCTInfo traceInfo_;
public boolean hasTraceInfo() {
return ((bitField0_ & 0x00000002) == 0x00000002);
@ -2499,8 +2499,8 @@ public final class RPCProtos {
return traceInfo_;
}
// optional string methodName = 3;
public static final int METHODNAME_FIELD_NUMBER = 3;
// optional string method_name = 3;
public static final int METHOD_NAME_FIELD_NUMBER = 3;
private java.lang.Object methodName_;
public boolean hasMethodName() {
return ((bitField0_ & 0x00000004) == 0x00000004);
@ -2531,8 +2531,8 @@ public final class RPCProtos {
}
}
// optional bool requestParam = 4;
public static final int REQUESTPARAM_FIELD_NUMBER = 4;
// optional bool request_param = 4;
public static final int REQUEST_PARAM_FIELD_NUMBER = 4;
private boolean requestParam_;
public boolean hasRequestParam() {
return ((bitField0_ & 0x00000008) == 0x00000008);
@ -2541,8 +2541,8 @@ public final class RPCProtos {
return requestParam_;
}
// optional .CellBlockMeta cellBlockMeta = 5;
public static final int CELLBLOCKMETA_FIELD_NUMBER = 5;
// optional .CellBlockMeta cell_block_meta = 5;
public static final int CELL_BLOCK_META_FIELD_NUMBER = 5;
private org.apache.hadoop.hbase.protobuf.generated.RPCProtos.CellBlockMeta cellBlockMeta_;
public boolean hasCellBlockMeta() {
return ((bitField0_ & 0x00000010) == 0x00000010);
@ -2675,23 +2675,23 @@ public final class RPCProtos {
int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode();
if (hasCallId()) {
hash = (37 * hash) + CALLID_FIELD_NUMBER;
hash = (37 * hash) + CALL_ID_FIELD_NUMBER;
hash = (53 * hash) + getCallId();
}
if (hasTraceInfo()) {
hash = (37 * hash) + TRACEINFO_FIELD_NUMBER;
hash = (37 * hash) + TRACE_INFO_FIELD_NUMBER;
hash = (53 * hash) + getTraceInfo().hashCode();
}
if (hasMethodName()) {
hash = (37 * hash) + METHODNAME_FIELD_NUMBER;
hash = (37 * hash) + METHOD_NAME_FIELD_NUMBER;
hash = (53 * hash) + getMethodName().hashCode();
}
if (hasRequestParam()) {
hash = (37 * hash) + REQUESTPARAM_FIELD_NUMBER;
hash = (37 * hash) + REQUEST_PARAM_FIELD_NUMBER;
hash = (53 * hash) + hashBoolean(getRequestParam());
}
if (hasCellBlockMeta()) {
hash = (37 * hash) + CELLBLOCKMETA_FIELD_NUMBER;
hash = (37 * hash) + CELL_BLOCK_META_FIELD_NUMBER;
hash = (53 * hash) + getCellBlockMeta().hashCode();
}
hash = (29 * hash) + getUnknownFields().hashCode();
@ -2997,7 +2997,7 @@ public final class RPCProtos {
private int bitField0_;
// optional uint32 callId = 1;
// optional uint32 call_id = 1;
private int callId_ ;
public boolean hasCallId() {
return ((bitField0_ & 0x00000001) == 0x00000001);
@ -3018,7 +3018,7 @@ public final class RPCProtos {
return this;
}
// optional .RPCTInfo traceInfo = 2;
// optional .RPCTInfo trace_info = 2;
private org.apache.hadoop.hbase.protobuf.generated.Tracing.RPCTInfo traceInfo_ = org.apache.hadoop.hbase.protobuf.generated.Tracing.RPCTInfo.getDefaultInstance();
private com.google.protobuf.SingleFieldBuilder<
org.apache.hadoop.hbase.protobuf.generated.Tracing.RPCTInfo, org.apache.hadoop.hbase.protobuf.generated.Tracing.RPCTInfo.Builder, org.apache.hadoop.hbase.protobuf.generated.Tracing.RPCTInfoOrBuilder> traceInfoBuilder_;
@ -3108,7 +3108,7 @@ public final class RPCProtos {
return traceInfoBuilder_;
}
// optional string methodName = 3;
// optional string method_name = 3;
private java.lang.Object methodName_ = "";
public boolean hasMethodName() {
return ((bitField0_ & 0x00000004) == 0x00000004);
@ -3144,7 +3144,7 @@ public final class RPCProtos {
onChanged();
}
// optional bool requestParam = 4;
// optional bool request_param = 4;
private boolean requestParam_ ;
public boolean hasRequestParam() {
return ((bitField0_ & 0x00000008) == 0x00000008);
@ -3165,7 +3165,7 @@ public final class RPCProtos {
return this;
}
// optional .CellBlockMeta cellBlockMeta = 5;
// optional .CellBlockMeta cell_block_meta = 5;
private org.apache.hadoop.hbase.protobuf.generated.RPCProtos.CellBlockMeta cellBlockMeta_ = org.apache.hadoop.hbase.protobuf.generated.RPCProtos.CellBlockMeta.getDefaultInstance();
private com.google.protobuf.SingleFieldBuilder<
org.apache.hadoop.hbase.protobuf.generated.RPCProtos.CellBlockMeta, org.apache.hadoop.hbase.protobuf.generated.RPCProtos.CellBlockMeta.Builder, org.apache.hadoop.hbase.protobuf.generated.RPCProtos.CellBlockMetaOrBuilder> cellBlockMetaBuilder_;
@ -3269,7 +3269,7 @@ public final class RPCProtos {
public interface ResponseHeaderOrBuilder
extends com.google.protobuf.MessageOrBuilder {
// optional uint32 callId = 1;
// optional uint32 call_id = 1;
boolean hasCallId();
int getCallId();
@ -3278,7 +3278,7 @@ public final class RPCProtos {
org.apache.hadoop.hbase.protobuf.generated.RPCProtos.ExceptionResponse getException();
org.apache.hadoop.hbase.protobuf.generated.RPCProtos.ExceptionResponseOrBuilder getExceptionOrBuilder();
// optional .CellBlockMeta cellBlockMeta = 3;
// optional .CellBlockMeta cell_block_meta = 3;
boolean hasCellBlockMeta();
org.apache.hadoop.hbase.protobuf.generated.RPCProtos.CellBlockMeta getCellBlockMeta();
org.apache.hadoop.hbase.protobuf.generated.RPCProtos.CellBlockMetaOrBuilder getCellBlockMetaOrBuilder();
@ -3312,8 +3312,8 @@ public final class RPCProtos {
}
private int bitField0_;
// optional uint32 callId = 1;
public static final int CALLID_FIELD_NUMBER = 1;
// optional uint32 call_id = 1;
public static final int CALL_ID_FIELD_NUMBER = 1;
private int callId_;
public boolean hasCallId() {
return ((bitField0_ & 0x00000001) == 0x00000001);
@ -3335,8 +3335,8 @@ public final class RPCProtos {
return exception_;
}
// optional .CellBlockMeta cellBlockMeta = 3;
public static final int CELLBLOCKMETA_FIELD_NUMBER = 3;
// optional .CellBlockMeta cell_block_meta = 3;
public static final int CELL_BLOCK_META_FIELD_NUMBER = 3;
private org.apache.hadoop.hbase.protobuf.generated.RPCProtos.CellBlockMeta cellBlockMeta_;
public boolean hasCellBlockMeta() {
return ((bitField0_ & 0x00000004) == 0x00000004);
@ -3443,7 +3443,7 @@ public final class RPCProtos {
int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode();
if (hasCallId()) {
hash = (37 * hash) + CALLID_FIELD_NUMBER;
hash = (37 * hash) + CALL_ID_FIELD_NUMBER;
hash = (53 * hash) + getCallId();
}
if (hasException()) {
@ -3451,7 +3451,7 @@ public final class RPCProtos {
hash = (53 * hash) + getException().hashCode();
}
if (hasCellBlockMeta()) {
hash = (37 * hash) + CELLBLOCKMETA_FIELD_NUMBER;
hash = (37 * hash) + CELL_BLOCK_META_FIELD_NUMBER;
hash = (53 * hash) + getCellBlockMeta().hashCode();
}
hash = (29 * hash) + getUnknownFields().hashCode();
@ -3729,7 +3729,7 @@ public final class RPCProtos {
private int bitField0_;
// optional uint32 callId = 1;
// optional uint32 call_id = 1;
private int callId_ ;
public boolean hasCallId() {
return ((bitField0_ & 0x00000001) == 0x00000001);
@ -3840,7 +3840,7 @@ public final class RPCProtos {
return exceptionBuilder_;
}
// optional .CellBlockMeta cellBlockMeta = 3;
// optional .CellBlockMeta cell_block_meta = 3;
private org.apache.hadoop.hbase.protobuf.generated.RPCProtos.CellBlockMeta cellBlockMeta_ = org.apache.hadoop.hbase.protobuf.generated.RPCProtos.CellBlockMeta.getDefaultInstance();
private com.google.protobuf.SingleFieldBuilder<
org.apache.hadoop.hbase.protobuf.generated.RPCProtos.CellBlockMeta, org.apache.hadoop.hbase.protobuf.generated.RPCProtos.CellBlockMeta.Builder, org.apache.hadoop.hbase.protobuf.generated.RPCProtos.CellBlockMetaOrBuilder> cellBlockMetaBuilder_;
@ -3981,24 +3981,25 @@ public final class RPCProtos {
static {
java.lang.String[] descriptorData = {
"\n\tRPC.proto\032\rTracing.proto\032\013hbase.proto\"" +
":\n\017UserInformation\022\025\n\reffectiveUser\030\001 \002(" +
"\t\022\020\n\010realUser\030\002 \001(\t\"\267\001\n\020ConnectionHeader" +
"\022\"\n\010userInfo\030\001 \001(\0132\020.UserInformation\022\023\n\013" +
"serviceName\030\002 \001(\t\022H\n\023cellBlockCodecClass" +
"\030\003 \001(\t:+org.apache.hadoop.hbase.codec.Ke" +
"yValueCodec\022 \n\030cellBlockCompressorClass\030" +
"\004 \001(\t\"\037\n\rCellBlockMeta\022\016\n\006length\030\001 \001(\r\"w" +
"\n\021ExceptionResponse\022\032\n\022exceptionClassNam" +
"e\030\001 \001(\t\022\022\n\nstackTrace\030\002 \001(\t\022\020\n\010hostname\030",
"\003 \001(\t\022\014\n\004port\030\004 \001(\005\022\022\n\ndoNotRetry\030\005 \001(\010\"" +
"\216\001\n\rRequestHeader\022\016\n\006callId\030\001 \001(\r\022\034\n\ttra" +
"ceInfo\030\002 \001(\0132\t.RPCTInfo\022\022\n\nmethodName\030\003 " +
"\001(\t\022\024\n\014requestParam\030\004 \001(\010\022%\n\rcellBlockMe" +
"ta\030\005 \001(\0132\016.CellBlockMeta\"n\n\016ResponseHead" +
"er\022\016\n\006callId\030\001 \001(\r\022%\n\texception\030\002 \001(\0132\022." +
"ExceptionResponse\022%\n\rcellBlockMeta\030\003 \001(\013" +
"2\016.CellBlockMetaB<\n*org.apache.hadoop.hb" +
"ase.protobuf.generatedB\tRPCProtosH\001\240\001\001"
"<\n\017UserInformation\022\026\n\016effective_user\030\001 \002" +
"(\t\022\021\n\treal_user\030\002 \001(\t\"\277\001\n\020ConnectionHead" +
"er\022#\n\tuser_info\030\001 \001(\0132\020.UserInformation\022" +
"\024\n\014service_name\030\002 \001(\t\022K\n\026cell_block_code" +
"c_class\030\003 \001(\t:+org.apache.hadoop.hbase.c" +
"odec.KeyValueCodec\022#\n\033cell_block_compres" +
"sor_class\030\004 \001(\t\"\037\n\rCellBlockMeta\022\016\n\006leng" +
"th\030\001 \001(\r\"|\n\021ExceptionResponse\022\034\n\024excepti" +
"on_class_name\030\001 \001(\t\022\023\n\013stack_trace\030\002 \001(\t",
"\022\020\n\010hostname\030\003 \001(\t\022\014\n\004port\030\004 \001(\005\022\024\n\014do_n" +
"ot_retry\030\005 \001(\010\"\224\001\n\rRequestHeader\022\017\n\007call" +
"_id\030\001 \001(\r\022\035\n\ntrace_info\030\002 \001(\0132\t.RPCTInfo" +
"\022\023\n\013method_name\030\003 \001(\t\022\025\n\rrequest_param\030\004" +
" \001(\010\022\'\n\017cell_block_meta\030\005 \001(\0132\016.CellBloc" +
"kMeta\"q\n\016ResponseHeader\022\017\n\007call_id\030\001 \001(\r" +
"\022%\n\texception\030\002 \001(\0132\022.ExceptionResponse\022" +
"\'\n\017cell_block_meta\030\003 \001(\0132\016.CellBlockMeta" +
"B<\n*org.apache.hadoop.hbase.protobuf.gen" +
"eratedB\tRPCProtosH\001\240\001\001"
};
com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() {

View File

@ -15,11 +15,11 @@ public final class RegionServerStatusProtos {
boolean hasPort();
int getPort();
// required uint64 serverStartCode = 2;
// required uint64 server_start_code = 2;
boolean hasServerStartCode();
long getServerStartCode();
// required uint64 serverCurrentTime = 3;
// required uint64 server_current_time = 3;
boolean hasServerCurrentTime();
long getServerCurrentTime();
}
@ -62,8 +62,8 @@ public final class RegionServerStatusProtos {
return port_;
}
// required uint64 serverStartCode = 2;
public static final int SERVERSTARTCODE_FIELD_NUMBER = 2;
// required uint64 server_start_code = 2;
public static final int SERVER_START_CODE_FIELD_NUMBER = 2;
private long serverStartCode_;
public boolean hasServerStartCode() {
return ((bitField0_ & 0x00000002) == 0x00000002);
@ -72,8 +72,8 @@ public final class RegionServerStatusProtos {
return serverStartCode_;
}
// required uint64 serverCurrentTime = 3;
public static final int SERVERCURRENTTIME_FIELD_NUMBER = 3;
// required uint64 server_current_time = 3;
public static final int SERVER_CURRENT_TIME_FIELD_NUMBER = 3;
private long serverCurrentTime_;
public boolean hasServerCurrentTime() {
return ((bitField0_ & 0x00000004) == 0x00000004);
@ -193,11 +193,11 @@ public final class RegionServerStatusProtos {
hash = (53 * hash) + getPort();
}
if (hasServerStartCode()) {
hash = (37 * hash) + SERVERSTARTCODE_FIELD_NUMBER;
hash = (37 * hash) + SERVER_START_CODE_FIELD_NUMBER;
hash = (53 * hash) + hashLong(getServerStartCode());
}
if (hasServerCurrentTime()) {
hash = (37 * hash) + SERVERCURRENTTIME_FIELD_NUMBER;
hash = (37 * hash) + SERVER_CURRENT_TIME_FIELD_NUMBER;
hash = (53 * hash) + hashLong(getServerCurrentTime());
}
hash = (29 * hash) + getUnknownFields().hashCode();
@ -482,7 +482,7 @@ public final class RegionServerStatusProtos {
return this;
}
// required uint64 serverStartCode = 2;
// required uint64 server_start_code = 2;
private long serverStartCode_ ;
public boolean hasServerStartCode() {
return ((bitField0_ & 0x00000002) == 0x00000002);
@ -503,7 +503,7 @@ public final class RegionServerStatusProtos {
return this;
}
// required uint64 serverCurrentTime = 3;
// required uint64 server_current_time = 3;
private long serverCurrentTime_ ;
public boolean hasServerCurrentTime() {
return ((bitField0_ & 0x00000004) == 0x00000004);
@ -538,7 +538,7 @@ public final class RegionServerStatusProtos {
public interface RegionServerStartupResponseOrBuilder
extends com.google.protobuf.MessageOrBuilder {
// repeated .NameStringPair mapEntries = 1;
// repeated .NameStringPair map_entries = 1;
java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPair>
getMapEntriesList();
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPair getMapEntries(int index);
@ -576,8 +576,8 @@ public final class RegionServerStatusProtos {
return org.apache.hadoop.hbase.protobuf.generated.RegionServerStatusProtos.internal_static_RegionServerStartupResponse_fieldAccessorTable;
}
// repeated .NameStringPair mapEntries = 1;
public static final int MAPENTRIES_FIELD_NUMBER = 1;
// repeated .NameStringPair map_entries = 1;
public static final int MAP_ENTRIES_FIELD_NUMBER = 1;
private java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPair> mapEntries_;
public java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPair> getMapEntriesList() {
return mapEntries_;
@ -669,7 +669,7 @@ public final class RegionServerStatusProtos {
int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode();
if (getMapEntriesCount() > 0) {
hash = (37 * hash) + MAPENTRIES_FIELD_NUMBER;
hash = (37 * hash) + MAP_ENTRIES_FIELD_NUMBER;
hash = (53 * hash) + getMapEntriesList().hashCode();
}
hash = (29 * hash) + getUnknownFields().hashCode();
@ -931,7 +931,7 @@ public final class RegionServerStatusProtos {
private int bitField0_;
// repeated .NameStringPair mapEntries = 1;
// repeated .NameStringPair map_entries = 1;
private java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPair> mapEntries_ =
java.util.Collections.emptyList();
private void ensureMapEntriesIsMutable() {
@ -2073,7 +2073,7 @@ public final class RegionServerStatusProtos {
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName getServer();
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerNameOrBuilder getServerOrBuilder();
// required string errorMessage = 2;
// required string error_message = 2;
boolean hasErrorMessage();
String getErrorMessage();
}
@ -2119,8 +2119,8 @@ public final class RegionServerStatusProtos {
return server_;
}
// required string errorMessage = 2;
public static final int ERRORMESSAGE_FIELD_NUMBER = 2;
// required string error_message = 2;
public static final int ERROR_MESSAGE_FIELD_NUMBER = 2;
private java.lang.Object errorMessage_;
public boolean hasErrorMessage() {
return ((bitField0_ & 0x00000002) == 0x00000002);
@ -2249,7 +2249,7 @@ public final class RegionServerStatusProtos {
hash = (53 * hash) + getServer().hashCode();
}
if (hasErrorMessage()) {
hash = (37 * hash) + ERRORMESSAGE_FIELD_NUMBER;
hash = (37 * hash) + ERROR_MESSAGE_FIELD_NUMBER;
hash = (53 * hash) + getErrorMessage().hashCode();
}
hash = (29 * hash) + getUnknownFields().hashCode();
@ -2602,7 +2602,7 @@ public final class RegionServerStatusProtos {
return serverBuilder_;
}
// required string errorMessage = 2;
// required string error_message = 2;
private java.lang.Object errorMessage_ = "";
public boolean hasErrorMessage() {
return ((bitField0_ & 0x00000002) == 0x00000002);
@ -2952,7 +2952,7 @@ public final class RegionServerStatusProtos {
public interface GetLastFlushedSequenceIdRequestOrBuilder
extends com.google.protobuf.MessageOrBuilder {
// required bytes regionName = 1;
// required bytes region_name = 1;
boolean hasRegionName();
com.google.protobuf.ByteString getRegionName();
}
@ -2985,8 +2985,8 @@ public final class RegionServerStatusProtos {
}
private int bitField0_;
// required bytes regionName = 1;
public static final int REGIONNAME_FIELD_NUMBER = 1;
// required bytes region_name = 1;
public static final int REGION_NAME_FIELD_NUMBER = 1;
private com.google.protobuf.ByteString regionName_;
public boolean hasRegionName() {
return ((bitField0_ & 0x00000001) == 0x00000001);
@ -3068,7 +3068,7 @@ public final class RegionServerStatusProtos {
int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode();
if (hasRegionName()) {
hash = (37 * hash) + REGIONNAME_FIELD_NUMBER;
hash = (37 * hash) + REGION_NAME_FIELD_NUMBER;
hash = (53 * hash) + getRegionName().hashCode();
}
hash = (29 * hash) + getUnknownFields().hashCode();
@ -3296,7 +3296,7 @@ public final class RegionServerStatusProtos {
private int bitField0_;
// required bytes regionName = 1;
// required bytes region_name = 1;
private com.google.protobuf.ByteString regionName_ = com.google.protobuf.ByteString.EMPTY;
public boolean hasRegionName() {
return ((bitField0_ & 0x00000001) == 0x00000001);
@ -3334,7 +3334,7 @@ public final class RegionServerStatusProtos {
public interface GetLastFlushedSequenceIdResponseOrBuilder
extends com.google.protobuf.MessageOrBuilder {
// required uint64 lastFlushedSequenceId = 1;
// required uint64 last_flushed_sequence_id = 1;
boolean hasLastFlushedSequenceId();
long getLastFlushedSequenceId();
}
@ -3367,8 +3367,8 @@ public final class RegionServerStatusProtos {
}
private int bitField0_;
// required uint64 lastFlushedSequenceId = 1;
public static final int LASTFLUSHEDSEQUENCEID_FIELD_NUMBER = 1;
// required uint64 last_flushed_sequence_id = 1;
public static final int LAST_FLUSHED_SEQUENCE_ID_FIELD_NUMBER = 1;
private long lastFlushedSequenceId_;
public boolean hasLastFlushedSequenceId() {
return ((bitField0_ & 0x00000001) == 0x00000001);
@ -3450,7 +3450,7 @@ public final class RegionServerStatusProtos {
int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode();
if (hasLastFlushedSequenceId()) {
hash = (37 * hash) + LASTFLUSHEDSEQUENCEID_FIELD_NUMBER;
hash = (37 * hash) + LAST_FLUSHED_SEQUENCE_ID_FIELD_NUMBER;
hash = (53 * hash) + hashLong(getLastFlushedSequenceId());
}
hash = (29 * hash) + getUnknownFields().hashCode();
@ -3678,7 +3678,7 @@ public final class RegionServerStatusProtos {
private int bitField0_;
// required uint64 lastFlushedSequenceId = 1;
// required uint64 last_flushed_sequence_id = 1;
private long lastFlushedSequenceId_ ;
public boolean hasLastFlushedSequenceId() {
return ((bitField0_ & 0x00000001) == 0x00000001);
@ -4176,31 +4176,31 @@ public final class RegionServerStatusProtos {
static {
java.lang.String[] descriptorData = {
"\n\030RegionServerStatus.proto\032\013hbase.proto\032" +
"\023ClusterStatus.proto\"^\n\032RegionServerStar" +
"tupRequest\022\014\n\004port\030\001 \002(\r\022\027\n\017serverStartC" +
"ode\030\002 \002(\004\022\031\n\021serverCurrentTime\030\003 \002(\004\"B\n\033" +
"RegionServerStartupResponse\022#\n\nmapEntrie" +
"s\030\001 \003(\0132\017.NameStringPair\"S\n\031RegionServer" +
"ReportRequest\022\033\n\006server\030\001 \002(\0132\013.ServerNa" +
"me\022\031\n\004load\030\002 \001(\0132\013.ServerLoad\"\034\n\032RegionS" +
"erverReportResponse\"N\n\031ReportRSFatalErro" +
"rRequest\022\033\n\006server\030\001 \002(\0132\013.ServerName\022\024\n",
"\014errorMessage\030\002 \002(\t\"\034\n\032ReportRSFatalErro" +
"rResponse\"5\n\037GetLastFlushedSequenceIdReq" +
"uest\022\022\n\nregionName\030\001 \002(\014\"A\n GetLastFlush" +
"edSequenceIdResponse\022\035\n\025lastFlushedSeque" +
"nceId\030\001 \002(\0042\354\002\n\031RegionServerStatusServic" +
"e\022P\n\023regionServerStartup\022\033.RegionServerS" +
"tartupRequest\032\034.RegionServerStartupRespo" +
"nse\022M\n\022regionServerReport\022\032.RegionServer" +
"ReportRequest\032\033.RegionServerReportRespon" +
"se\022M\n\022reportRSFatalError\022\032.ReportRSFatal",
"ErrorRequest\032\033.ReportRSFatalErrorRespons" +
"e\022_\n\030getLastFlushedSequenceId\022 .GetLastF" +
"lushedSequenceIdRequest\032!.GetLastFlushed" +
"SequenceIdResponseBN\n*org.apache.hadoop." +
"hbase.protobuf.generatedB\030RegionServerSt" +
"atusProtosH\001\210\001\001\240\001\001"
"\023ClusterStatus.proto\"b\n\032RegionServerStar" +
"tupRequest\022\014\n\004port\030\001 \002(\r\022\031\n\021server_start" +
"_code\030\002 \002(\004\022\033\n\023server_current_time\030\003 \002(\004" +
"\"C\n\033RegionServerStartupResponse\022$\n\013map_e" +
"ntries\030\001 \003(\0132\017.NameStringPair\"S\n\031RegionS" +
"erverReportRequest\022\033\n\006server\030\001 \002(\0132\013.Ser" +
"verName\022\031\n\004load\030\002 \001(\0132\013.ServerLoad\"\034\n\032Re" +
"gionServerReportResponse\"O\n\031ReportRSFata" +
"lErrorRequest\022\033\n\006server\030\001 \002(\0132\013.ServerNa",
"me\022\025\n\rerror_message\030\002 \002(\t\"\034\n\032ReportRSFat" +
"alErrorResponse\"6\n\037GetLastFlushedSequenc" +
"eIdRequest\022\023\n\013region_name\030\001 \002(\014\"D\n GetLa" +
"stFlushedSequenceIdResponse\022 \n\030last_flus" +
"hed_sequence_id\030\001 \002(\0042\354\002\n\031RegionServerSt" +
"atusService\022P\n\023RegionServerStartup\022\033.Reg" +
"ionServerStartupRequest\032\034.RegionServerSt" +
"artupResponse\022M\n\022RegionServerReport\022\032.Re" +
"gionServerReportRequest\032\033.RegionServerRe" +
"portResponse\022M\n\022ReportRSFatalError\022\032.Rep",
"ortRSFatalErrorRequest\032\033.ReportRSFatalEr" +
"rorResponse\022_\n\030GetLastFlushedSequenceId\022" +
" .GetLastFlushedSequenceIdRequest\032!.GetL" +
"astFlushedSequenceIdResponseBN\n*org.apac" +
"he.hadoop.hbase.protobuf.generatedB\030Regi" +
"onServerStatusProtosH\001\210\001\001\240\001\001"
};
com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() {

View File

@ -11,15 +11,15 @@ public final class RowProcessorProtos {
public interface RowProcessorRequestOrBuilder
extends com.google.protobuf.MessageOrBuilder {
// required string rowProcessorClassName = 1;
// required string row_processor_class_name = 1;
boolean hasRowProcessorClassName();
String getRowProcessorClassName();
// optional string rowProcessorInitializerMessageName = 2;
// optional string row_processor_initializer_message_name = 2;
boolean hasRowProcessorInitializerMessageName();
String getRowProcessorInitializerMessageName();
// optional bytes rowProcessorInitializerMessage = 3;
// optional bytes row_processor_initializer_message = 3;
boolean hasRowProcessorInitializerMessage();
com.google.protobuf.ByteString getRowProcessorInitializerMessage();
}
@ -52,8 +52,8 @@ public final class RowProcessorProtos {
}
private int bitField0_;
// required string rowProcessorClassName = 1;
public static final int ROWPROCESSORCLASSNAME_FIELD_NUMBER = 1;
// required string row_processor_class_name = 1;
public static final int ROW_PROCESSOR_CLASS_NAME_FIELD_NUMBER = 1;
private java.lang.Object rowProcessorClassName_;
public boolean hasRowProcessorClassName() {
return ((bitField0_ & 0x00000001) == 0x00000001);
@ -84,8 +84,8 @@ public final class RowProcessorProtos {
}
}
// optional string rowProcessorInitializerMessageName = 2;
public static final int ROWPROCESSORINITIALIZERMESSAGENAME_FIELD_NUMBER = 2;
// optional string row_processor_initializer_message_name = 2;
public static final int ROW_PROCESSOR_INITIALIZER_MESSAGE_NAME_FIELD_NUMBER = 2;
private java.lang.Object rowProcessorInitializerMessageName_;
public boolean hasRowProcessorInitializerMessageName() {
return ((bitField0_ & 0x00000002) == 0x00000002);
@ -116,8 +116,8 @@ public final class RowProcessorProtos {
}
}
// optional bytes rowProcessorInitializerMessage = 3;
public static final int ROWPROCESSORINITIALIZERMESSAGE_FIELD_NUMBER = 3;
// optional bytes row_processor_initializer_message = 3;
public static final int ROW_PROCESSOR_INITIALIZER_MESSAGE_FIELD_NUMBER = 3;
private com.google.protobuf.ByteString rowProcessorInitializerMessage_;
public boolean hasRowProcessorInitializerMessage() {
return ((bitField0_ & 0x00000004) == 0x00000004);
@ -225,15 +225,15 @@ public final class RowProcessorProtos {
int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode();
if (hasRowProcessorClassName()) {
hash = (37 * hash) + ROWPROCESSORCLASSNAME_FIELD_NUMBER;
hash = (37 * hash) + ROW_PROCESSOR_CLASS_NAME_FIELD_NUMBER;
hash = (53 * hash) + getRowProcessorClassName().hashCode();
}
if (hasRowProcessorInitializerMessageName()) {
hash = (37 * hash) + ROWPROCESSORINITIALIZERMESSAGENAME_FIELD_NUMBER;
hash = (37 * hash) + ROW_PROCESSOR_INITIALIZER_MESSAGE_NAME_FIELD_NUMBER;
hash = (53 * hash) + getRowProcessorInitializerMessageName().hashCode();
}
if (hasRowProcessorInitializerMessage()) {
hash = (37 * hash) + ROWPROCESSORINITIALIZERMESSAGE_FIELD_NUMBER;
hash = (37 * hash) + ROW_PROCESSOR_INITIALIZER_MESSAGE_FIELD_NUMBER;
hash = (53 * hash) + getRowProcessorInitializerMessage().hashCode();
}
hash = (29 * hash) + getUnknownFields().hashCode();
@ -489,7 +489,7 @@ public final class RowProcessorProtos {
private int bitField0_;
// required string rowProcessorClassName = 1;
// required string row_processor_class_name = 1;
private java.lang.Object rowProcessorClassName_ = "";
public boolean hasRowProcessorClassName() {
return ((bitField0_ & 0x00000001) == 0x00000001);
@ -525,7 +525,7 @@ public final class RowProcessorProtos {
onChanged();
}
// optional string rowProcessorInitializerMessageName = 2;
// optional string row_processor_initializer_message_name = 2;
private java.lang.Object rowProcessorInitializerMessageName_ = "";
public boolean hasRowProcessorInitializerMessageName() {
return ((bitField0_ & 0x00000002) == 0x00000002);
@ -561,7 +561,7 @@ public final class RowProcessorProtos {
onChanged();
}
// optional bytes rowProcessorInitializerMessage = 3;
// optional bytes row_processor_initializer_message = 3;
private com.google.protobuf.ByteString rowProcessorInitializerMessage_ = com.google.protobuf.ByteString.EMPTY;
public boolean hasRowProcessorInitializerMessage() {
return ((bitField0_ & 0x00000004) == 0x00000004);
@ -599,7 +599,7 @@ public final class RowProcessorProtos {
public interface RowProcessorResultOrBuilder
extends com.google.protobuf.MessageOrBuilder {
// required bytes rowProcessorResult = 1;
// required bytes row_processor_result = 1;
boolean hasRowProcessorResult();
com.google.protobuf.ByteString getRowProcessorResult();
}
@ -632,8 +632,8 @@ public final class RowProcessorProtos {
}
private int bitField0_;
// required bytes rowProcessorResult = 1;
public static final int ROWPROCESSORRESULT_FIELD_NUMBER = 1;
// required bytes row_processor_result = 1;
public static final int ROW_PROCESSOR_RESULT_FIELD_NUMBER = 1;
private com.google.protobuf.ByteString rowProcessorResult_;
public boolean hasRowProcessorResult() {
return ((bitField0_ & 0x00000001) == 0x00000001);
@ -715,7 +715,7 @@ public final class RowProcessorProtos {
int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode();
if (hasRowProcessorResult()) {
hash = (37 * hash) + ROWPROCESSORRESULT_FIELD_NUMBER;
hash = (37 * hash) + ROW_PROCESSOR_RESULT_FIELD_NUMBER;
hash = (53 * hash) + getRowProcessorResult().hashCode();
}
hash = (29 * hash) + getUnknownFields().hashCode();
@ -943,7 +943,7 @@ public final class RowProcessorProtos {
private int bitField0_;
// required bytes rowProcessorResult = 1;
// required bytes row_processor_result = 1;
private com.google.protobuf.ByteString rowProcessorResult_ = com.google.protobuf.ByteString.EMPTY;
public boolean hasRowProcessorResult() {
return ((bitField0_ & 0x00000001) == 0x00000001);
@ -1218,15 +1218,16 @@ public final class RowProcessorProtos {
descriptor;
static {
java.lang.String[] descriptorData = {
"\n\022RowProcessor.proto\"\210\001\n\023RowProcessorReq" +
"uest\022\035\n\025rowProcessorClassName\030\001 \002(\t\022*\n\"r" +
"owProcessorInitializerMessageName\030\002 \001(\t\022" +
"&\n\036rowProcessorInitializerMessage\030\003 \001(\014\"" +
"0\n\022RowProcessorResult\022\032\n\022rowProcessorRes" +
"ult\030\001 \002(\0142K\n\023RowProcessorService\0224\n\007proc" +
"ess\022\024.RowProcessorRequest\032\023.RowProcessor" +
"ResultBH\n*org.apache.hadoop.hbase.protob" +
"uf.generatedB\022RowProcessorProtosH\001\210\001\001\240\001\001"
"\n\022RowProcessor.proto\"\222\001\n\023RowProcessorReq" +
"uest\022 \n\030row_processor_class_name\030\001 \002(\t\022." +
"\n&row_processor_initializer_message_name" +
"\030\002 \001(\t\022)\n!row_processor_initializer_mess" +
"age\030\003 \001(\014\"2\n\022RowProcessorResult\022\034\n\024row_p" +
"rocessor_result\030\001 \002(\0142K\n\023RowProcessorSer" +
"vice\0224\n\007Process\022\024.RowProcessorRequest\032\023." +
"RowProcessorResultBH\n*org.apache.hadoop." +
"hbase.protobuf.generatedB\022RowProcessorPr" +
"otosH\001\210\001\001\240\001\001"
};
com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() {

View File

@ -11,7 +11,7 @@ public final class SecureBulkLoadProtos {
public interface SecureBulkLoadHFilesRequestOrBuilder
extends com.google.protobuf.MessageOrBuilder {
// repeated .BulkLoadHFileRequest.FamilyPath familyPath = 1;
// repeated .BulkLoadHFileRequest.FamilyPath family_path = 1;
java.util.List<org.apache.hadoop.hbase.protobuf.generated.ClientProtos.BulkLoadHFileRequest.FamilyPath>
getFamilyPathList();
org.apache.hadoop.hbase.protobuf.generated.ClientProtos.BulkLoadHFileRequest.FamilyPath getFamilyPath(int index);
@ -21,16 +21,16 @@ public final class SecureBulkLoadProtos {
org.apache.hadoop.hbase.protobuf.generated.ClientProtos.BulkLoadHFileRequest.FamilyPathOrBuilder getFamilyPathOrBuilder(
int index);
// optional bool assignSeqNum = 2;
// optional bool assign_seq_num = 2;
boolean hasAssignSeqNum();
boolean getAssignSeqNum();
// required .DelegationTokenProto fsToken = 3;
// required .DelegationTokenProto fs_token = 3;
boolean hasFsToken();
org.apache.hadoop.hbase.protobuf.generated.SecureBulkLoadProtos.DelegationTokenProto getFsToken();
org.apache.hadoop.hbase.protobuf.generated.SecureBulkLoadProtos.DelegationTokenProtoOrBuilder getFsTokenOrBuilder();
// required string bulkToken = 4;
// required string bulk_token = 4;
boolean hasBulkToken();
String getBulkToken();
}
@ -63,8 +63,8 @@ public final class SecureBulkLoadProtos {
}
private int bitField0_;
// repeated .BulkLoadHFileRequest.FamilyPath familyPath = 1;
public static final int FAMILYPATH_FIELD_NUMBER = 1;
// repeated .BulkLoadHFileRequest.FamilyPath family_path = 1;
public static final int FAMILY_PATH_FIELD_NUMBER = 1;
private java.util.List<org.apache.hadoop.hbase.protobuf.generated.ClientProtos.BulkLoadHFileRequest.FamilyPath> familyPath_;
public java.util.List<org.apache.hadoop.hbase.protobuf.generated.ClientProtos.BulkLoadHFileRequest.FamilyPath> getFamilyPathList() {
return familyPath_;
@ -84,8 +84,8 @@ public final class SecureBulkLoadProtos {
return familyPath_.get(index);
}
// optional bool assignSeqNum = 2;
public static final int ASSIGNSEQNUM_FIELD_NUMBER = 2;
// optional bool assign_seq_num = 2;
public static final int ASSIGN_SEQ_NUM_FIELD_NUMBER = 2;
private boolean assignSeqNum_;
public boolean hasAssignSeqNum() {
return ((bitField0_ & 0x00000001) == 0x00000001);
@ -94,8 +94,8 @@ public final class SecureBulkLoadProtos {
return assignSeqNum_;
}
// required .DelegationTokenProto fsToken = 3;
public static final int FSTOKEN_FIELD_NUMBER = 3;
// required .DelegationTokenProto fs_token = 3;
public static final int FS_TOKEN_FIELD_NUMBER = 3;
private org.apache.hadoop.hbase.protobuf.generated.SecureBulkLoadProtos.DelegationTokenProto fsToken_;
public boolean hasFsToken() {
return ((bitField0_ & 0x00000002) == 0x00000002);
@ -107,8 +107,8 @@ public final class SecureBulkLoadProtos {
return fsToken_;
}
// required string bulkToken = 4;
public static final int BULKTOKEN_FIELD_NUMBER = 4;
// required string bulk_token = 4;
public static final int BULK_TOKEN_FIELD_NUMBER = 4;
private java.lang.Object bulkToken_;
public boolean hasBulkToken() {
return ((bitField0_ & 0x00000004) == 0x00000004);
@ -258,19 +258,19 @@ public final class SecureBulkLoadProtos {
int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode();
if (getFamilyPathCount() > 0) {
hash = (37 * hash) + FAMILYPATH_FIELD_NUMBER;
hash = (37 * hash) + FAMILY_PATH_FIELD_NUMBER;
hash = (53 * hash) + getFamilyPathList().hashCode();
}
if (hasAssignSeqNum()) {
hash = (37 * hash) + ASSIGNSEQNUM_FIELD_NUMBER;
hash = (37 * hash) + ASSIGN_SEQ_NUM_FIELD_NUMBER;
hash = (53 * hash) + hashBoolean(getAssignSeqNum());
}
if (hasFsToken()) {
hash = (37 * hash) + FSTOKEN_FIELD_NUMBER;
hash = (37 * hash) + FS_TOKEN_FIELD_NUMBER;
hash = (53 * hash) + getFsToken().hashCode();
}
if (hasBulkToken()) {
hash = (37 * hash) + BULKTOKEN_FIELD_NUMBER;
hash = (37 * hash) + BULK_TOKEN_FIELD_NUMBER;
hash = (53 * hash) + getBulkToken().hashCode();
}
hash = (29 * hash) + getUnknownFields().hashCode();
@ -597,7 +597,7 @@ public final class SecureBulkLoadProtos {
private int bitField0_;
// repeated .BulkLoadHFileRequest.FamilyPath familyPath = 1;
// repeated .BulkLoadHFileRequest.FamilyPath family_path = 1;
private java.util.List<org.apache.hadoop.hbase.protobuf.generated.ClientProtos.BulkLoadHFileRequest.FamilyPath> familyPath_ =
java.util.Collections.emptyList();
private void ensureFamilyPathIsMutable() {
@ -783,7 +783,7 @@ public final class SecureBulkLoadProtos {
return familyPathBuilder_;
}
// optional bool assignSeqNum = 2;
// optional bool assign_seq_num = 2;
private boolean assignSeqNum_ ;
public boolean hasAssignSeqNum() {
return ((bitField0_ & 0x00000002) == 0x00000002);
@ -804,7 +804,7 @@ public final class SecureBulkLoadProtos {
return this;
}
// required .DelegationTokenProto fsToken = 3;
// required .DelegationTokenProto fs_token = 3;
private org.apache.hadoop.hbase.protobuf.generated.SecureBulkLoadProtos.DelegationTokenProto fsToken_ = org.apache.hadoop.hbase.protobuf.generated.SecureBulkLoadProtos.DelegationTokenProto.getDefaultInstance();
private com.google.protobuf.SingleFieldBuilder<
org.apache.hadoop.hbase.protobuf.generated.SecureBulkLoadProtos.DelegationTokenProto, org.apache.hadoop.hbase.protobuf.generated.SecureBulkLoadProtos.DelegationTokenProto.Builder, org.apache.hadoop.hbase.protobuf.generated.SecureBulkLoadProtos.DelegationTokenProtoOrBuilder> fsTokenBuilder_;
@ -894,7 +894,7 @@ public final class SecureBulkLoadProtos {
return fsTokenBuilder_;
}
// required string bulkToken = 4;
// required string bulk_token = 4;
private java.lang.Object bulkToken_ = "";
public boolean hasBulkToken() {
return ((bitField0_ & 0x00000008) == 0x00000008);
@ -1972,7 +1972,7 @@ public final class SecureBulkLoadProtos {
public interface PrepareBulkLoadRequestOrBuilder
extends com.google.protobuf.MessageOrBuilder {
// required bytes tableName = 1;
// required bytes table_name = 1;
boolean hasTableName();
com.google.protobuf.ByteString getTableName();
}
@ -2005,8 +2005,8 @@ public final class SecureBulkLoadProtos {
}
private int bitField0_;
// required bytes tableName = 1;
public static final int TABLENAME_FIELD_NUMBER = 1;
// required bytes table_name = 1;
public static final int TABLE_NAME_FIELD_NUMBER = 1;
private com.google.protobuf.ByteString tableName_;
public boolean hasTableName() {
return ((bitField0_ & 0x00000001) == 0x00000001);
@ -2088,7 +2088,7 @@ public final class SecureBulkLoadProtos {
int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode();
if (hasTableName()) {
hash = (37 * hash) + TABLENAME_FIELD_NUMBER;
hash = (37 * hash) + TABLE_NAME_FIELD_NUMBER;
hash = (53 * hash) + getTableName().hashCode();
}
hash = (29 * hash) + getUnknownFields().hashCode();
@ -2316,7 +2316,7 @@ public final class SecureBulkLoadProtos {
private int bitField0_;
// required bytes tableName = 1;
// required bytes table_name = 1;
private com.google.protobuf.ByteString tableName_ = com.google.protobuf.ByteString.EMPTY;
public boolean hasTableName() {
return ((bitField0_ & 0x00000001) == 0x00000001);
@ -2354,7 +2354,7 @@ public final class SecureBulkLoadProtos {
public interface PrepareBulkLoadResponseOrBuilder
extends com.google.protobuf.MessageOrBuilder {
// required string bulkToken = 1;
// required string bulk_token = 1;
boolean hasBulkToken();
String getBulkToken();
}
@ -2387,8 +2387,8 @@ public final class SecureBulkLoadProtos {
}
private int bitField0_;
// required string bulkToken = 1;
public static final int BULKTOKEN_FIELD_NUMBER = 1;
// required string bulk_token = 1;
public static final int BULK_TOKEN_FIELD_NUMBER = 1;
private java.lang.Object bulkToken_;
public boolean hasBulkToken() {
return ((bitField0_ & 0x00000001) == 0x00000001);
@ -2492,7 +2492,7 @@ public final class SecureBulkLoadProtos {
int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode();
if (hasBulkToken()) {
hash = (37 * hash) + BULKTOKEN_FIELD_NUMBER;
hash = (37 * hash) + BULK_TOKEN_FIELD_NUMBER;
hash = (53 * hash) + getBulkToken().hashCode();
}
hash = (29 * hash) + getUnknownFields().hashCode();
@ -2720,7 +2720,7 @@ public final class SecureBulkLoadProtos {
private int bitField0_;
// required string bulkToken = 1;
// required string bulk_token = 1;
private java.lang.Object bulkToken_ = "";
public boolean hasBulkToken() {
return ((bitField0_ & 0x00000001) == 0x00000001);
@ -2770,7 +2770,7 @@ public final class SecureBulkLoadProtos {
public interface CleanupBulkLoadRequestOrBuilder
extends com.google.protobuf.MessageOrBuilder {
// required string bulkToken = 1;
// required string bulk_token = 1;
boolean hasBulkToken();
String getBulkToken();
}
@ -2803,8 +2803,8 @@ public final class SecureBulkLoadProtos {
}
private int bitField0_;
// required string bulkToken = 1;
public static final int BULKTOKEN_FIELD_NUMBER = 1;
// required string bulk_token = 1;
public static final int BULK_TOKEN_FIELD_NUMBER = 1;
private java.lang.Object bulkToken_;
public boolean hasBulkToken() {
return ((bitField0_ & 0x00000001) == 0x00000001);
@ -2908,7 +2908,7 @@ public final class SecureBulkLoadProtos {
int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode();
if (hasBulkToken()) {
hash = (37 * hash) + BULKTOKEN_FIELD_NUMBER;
hash = (37 * hash) + BULK_TOKEN_FIELD_NUMBER;
hash = (53 * hash) + getBulkToken().hashCode();
}
hash = (29 * hash) + getUnknownFields().hashCode();
@ -3136,7 +3136,7 @@ public final class SecureBulkLoadProtos {
private int bitField0_;
// required string bulkToken = 1;
// required string bulk_token = 1;
private java.lang.Object bulkToken_ = "";
public boolean hasBulkToken() {
return ((bitField0_ & 0x00000001) == 0x00000001);
@ -3879,27 +3879,27 @@ public final class SecureBulkLoadProtos {
static {
java.lang.String[] descriptorData = {
"\n\024SecureBulkLoad.proto\032\013hbase.proto\032\014Cli" +
"ent.proto\"\244\001\n\033SecureBulkLoadHFilesReques" +
"t\0224\n\nfamilyPath\030\001 \003(\0132 .BulkLoadHFileReq" +
"uest.FamilyPath\022\024\n\014assignSeqNum\030\002 \001(\010\022&\n" +
"\007fsToken\030\003 \002(\0132\025.DelegationTokenProto\022\021\n" +
"\tbulkToken\030\004 \002(\t\".\n\034SecureBulkLoadHFiles" +
"Response\022\016\n\006loaded\030\001 \002(\010\"[\n\024DelegationTo" +
"kenProto\022\022\n\nidentifier\030\001 \001(\014\022\020\n\010password" +
"\030\002 \001(\014\022\014\n\004kind\030\003 \001(\t\022\017\n\007service\030\004 \001(\t\"+\n" +
"\026PrepareBulkLoadRequest\022\021\n\ttableName\030\001 \002",
"(\014\",\n\027PrepareBulkLoadResponse\022\021\n\tbulkTok" +
"en\030\001 \002(\t\"+\n\026CleanupBulkLoadRequest\022\021\n\tbu" +
"lkToken\030\001 \002(\t\"\031\n\027CleanupBulkLoadResponse" +
"2\370\001\n\025SecureBulkLoadService\022D\n\017prepareBul" +
"kLoad\022\027.PrepareBulkLoadRequest\032\030.Prepare" +
"BulkLoadResponse\022S\n\024secureBulkLoadHFiles" +
"\022\034.SecureBulkLoadHFilesRequest\032\035.SecureB" +
"ulkLoadHFilesResponse\022D\n\017cleanupBulkLoad" +
"\022\027.CleanupBulkLoadRequest\032\030.CleanupBulkL" +
"oadResponseBJ\n*org.apache.hadoop.hbase.p",
"rotobuf.generatedB\024SecureBulkLoadProtosH" +
"\001\210\001\001\240\001\001"
"ent.proto\"\251\001\n\033SecureBulkLoadHFilesReques" +
"t\0225\n\013family_path\030\001 \003(\0132 .BulkLoadHFileRe" +
"quest.FamilyPath\022\026\n\016assign_seq_num\030\002 \001(\010" +
"\022\'\n\010fs_token\030\003 \002(\0132\025.DelegationTokenProt" +
"o\022\022\n\nbulk_token\030\004 \002(\t\".\n\034SecureBulkLoadH" +
"FilesResponse\022\016\n\006loaded\030\001 \002(\010\"[\n\024Delegat" +
"ionTokenProto\022\022\n\nidentifier\030\001 \001(\014\022\020\n\010pas" +
"sword\030\002 \001(\014\022\014\n\004kind\030\003 \001(\t\022\017\n\007service\030\004 \001" +
"(\t\",\n\026PrepareBulkLoadRequest\022\022\n\ntable_na",
"me\030\001 \002(\014\"-\n\027PrepareBulkLoadResponse\022\022\n\nb" +
"ulk_token\030\001 \002(\t\",\n\026CleanupBulkLoadReques" +
"t\022\022\n\nbulk_token\030\001 \002(\t\"\031\n\027CleanupBulkLoad" +
"Response2\370\001\n\025SecureBulkLoadService\022D\n\017Pr" +
"epareBulkLoad\022\027.PrepareBulkLoadRequest\032\030" +
".PrepareBulkLoadResponse\022S\n\024SecureBulkLo" +
"adHFiles\022\034.SecureBulkLoadHFilesRequest\032\035" +
".SecureBulkLoadHFilesResponse\022D\n\017Cleanup" +
"BulkLoad\022\027.CleanupBulkLoadRequest\032\030.Clea" +
"nupBulkLoadResponseBJ\n*org.apache.hadoop",
".hbase.protobuf.generatedB\024SecureBulkLoa" +
"dProtosH\001\210\001\001\240\001\001"
};
com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() {

View File

@ -11,11 +11,11 @@ public final class Tracing {
public interface RPCTInfoOrBuilder
extends com.google.protobuf.MessageOrBuilder {
// optional int64 traceId = 1;
// optional int64 trace_id = 1;
boolean hasTraceId();
long getTraceId();
// optional int64 parentId = 2;
// optional int64 parent_id = 2;
boolean hasParentId();
long getParentId();
}
@ -48,8 +48,8 @@ public final class Tracing {
}
private int bitField0_;
// optional int64 traceId = 1;
public static final int TRACEID_FIELD_NUMBER = 1;
// optional int64 trace_id = 1;
public static final int TRACE_ID_FIELD_NUMBER = 1;
private long traceId_;
public boolean hasTraceId() {
return ((bitField0_ & 0x00000001) == 0x00000001);
@ -58,8 +58,8 @@ public final class Tracing {
return traceId_;
}
// optional int64 parentId = 2;
public static final int PARENTID_FIELD_NUMBER = 2;
// optional int64 parent_id = 2;
public static final int PARENT_ID_FIELD_NUMBER = 2;
private long parentId_;
public boolean hasParentId() {
return ((bitField0_ & 0x00000002) == 0x00000002);
@ -150,11 +150,11 @@ public final class Tracing {
int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode();
if (hasTraceId()) {
hash = (37 * hash) + TRACEID_FIELD_NUMBER;
hash = (37 * hash) + TRACE_ID_FIELD_NUMBER;
hash = (53 * hash) + hashLong(getTraceId());
}
if (hasParentId()) {
hash = (37 * hash) + PARENTID_FIELD_NUMBER;
hash = (37 * hash) + PARENT_ID_FIELD_NUMBER;
hash = (53 * hash) + hashLong(getParentId());
}
hash = (29 * hash) + getUnknownFields().hashCode();
@ -392,7 +392,7 @@ public final class Tracing {
private int bitField0_;
// optional int64 traceId = 1;
// optional int64 trace_id = 1;
private long traceId_ ;
public boolean hasTraceId() {
return ((bitField0_ & 0x00000001) == 0x00000001);
@ -413,7 +413,7 @@ public final class Tracing {
return this;
}
// optional int64 parentId = 2;
// optional int64 parent_id = 2;
private long parentId_ ;
public boolean hasParentId() {
return ((bitField0_ & 0x00000002) == 0x00000002);
@ -459,10 +459,10 @@ public final class Tracing {
descriptor;
static {
java.lang.String[] descriptorData = {
"\n\rTracing.proto\"-\n\010RPCTInfo\022\017\n\007traceId\030\001" +
" \001(\003\022\020\n\010parentId\030\002 \001(\003B:\n*org.apache.had" +
"oop.hbase.protobuf.generatedB\007TracingH\001\240" +
"\001\001"
"\n\rTracing.proto\"/\n\010RPCTInfo\022\020\n\010trace_id\030" +
"\001 \001(\003\022\021\n\tparent_id\030\002 \001(\003B:\n*org.apache.h" +
"adoop.hbase.protobuf.generatedB\007TracingH" +
"\001\240\001\001"
};
com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() {

View File

@ -80,7 +80,7 @@ public final class WALProtos {
public interface WALHeaderOrBuilder
extends com.google.protobuf.MessageOrBuilder {
// optional bool hasCompression = 1;
// optional bool has_compression = 1;
boolean hasHasCompression();
boolean getHasCompression();
}
@ -113,8 +113,8 @@ public final class WALProtos {
}
private int bitField0_;
// optional bool hasCompression = 1;
public static final int HASCOMPRESSION_FIELD_NUMBER = 1;
// optional bool has_compression = 1;
public static final int HAS_COMPRESSION_FIELD_NUMBER = 1;
private boolean hasCompression_;
public boolean hasHasCompression() {
return ((bitField0_ & 0x00000001) == 0x00000001);
@ -192,7 +192,7 @@ public final class WALProtos {
int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode();
if (hasHasCompression()) {
hash = (37 * hash) + HASCOMPRESSION_FIELD_NUMBER;
hash = (37 * hash) + HAS_COMPRESSION_FIELD_NUMBER;
hash = (53 * hash) + hashBoolean(getHasCompression());
}
hash = (29 * hash) + getUnknownFields().hashCode();
@ -416,7 +416,7 @@ public final class WALProtos {
private int bitField0_;
// optional bool hasCompression = 1;
// optional bool has_compression = 1;
private boolean hasCompression_ ;
public boolean hasHasCompression() {
return ((bitField0_ & 0x00000001) == 0x00000001);
@ -451,23 +451,23 @@ public final class WALProtos {
public interface WALKeyOrBuilder
extends com.google.protobuf.MessageOrBuilder {
// required bytes encodedRegionName = 1;
// required bytes encoded_region_name = 1;
boolean hasEncodedRegionName();
com.google.protobuf.ByteString getEncodedRegionName();
// required bytes tableName = 2;
// required bytes table_name = 2;
boolean hasTableName();
com.google.protobuf.ByteString getTableName();
// required uint64 logSequenceNumber = 3;
// required uint64 log_sequence_number = 3;
boolean hasLogSequenceNumber();
long getLogSequenceNumber();
// required uint64 writeTime = 4;
// required uint64 write_time = 4;
boolean hasWriteTime();
long getWriteTime();
// optional .UUID clusterId = 5;
// optional .UUID cluster_id = 5;
boolean hasClusterId();
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.UUID getClusterId();
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.UUIDOrBuilder getClusterIdOrBuilder();
@ -482,7 +482,7 @@ public final class WALProtos {
org.apache.hadoop.hbase.protobuf.generated.WALProtos.FamilyScopeOrBuilder getScopesOrBuilder(
int index);
// optional uint32 followingKvCount = 7;
// optional uint32 following_kv_count = 7;
boolean hasFollowingKvCount();
int getFollowingKvCount();
}
@ -515,8 +515,8 @@ public final class WALProtos {
}
private int bitField0_;
// required bytes encodedRegionName = 1;
public static final int ENCODEDREGIONNAME_FIELD_NUMBER = 1;
// required bytes encoded_region_name = 1;
public static final int ENCODED_REGION_NAME_FIELD_NUMBER = 1;
private com.google.protobuf.ByteString encodedRegionName_;
public boolean hasEncodedRegionName() {
return ((bitField0_ & 0x00000001) == 0x00000001);
@ -525,8 +525,8 @@ public final class WALProtos {
return encodedRegionName_;
}
// required bytes tableName = 2;
public static final int TABLENAME_FIELD_NUMBER = 2;
// required bytes table_name = 2;
public static final int TABLE_NAME_FIELD_NUMBER = 2;
private com.google.protobuf.ByteString tableName_;
public boolean hasTableName() {
return ((bitField0_ & 0x00000002) == 0x00000002);
@ -535,8 +535,8 @@ public final class WALProtos {
return tableName_;
}
// required uint64 logSequenceNumber = 3;
public static final int LOGSEQUENCENUMBER_FIELD_NUMBER = 3;
// required uint64 log_sequence_number = 3;
public static final int LOG_SEQUENCE_NUMBER_FIELD_NUMBER = 3;
private long logSequenceNumber_;
public boolean hasLogSequenceNumber() {
return ((bitField0_ & 0x00000004) == 0x00000004);
@ -545,8 +545,8 @@ public final class WALProtos {
return logSequenceNumber_;
}
// required uint64 writeTime = 4;
public static final int WRITETIME_FIELD_NUMBER = 4;
// required uint64 write_time = 4;
public static final int WRITE_TIME_FIELD_NUMBER = 4;
private long writeTime_;
public boolean hasWriteTime() {
return ((bitField0_ & 0x00000008) == 0x00000008);
@ -555,8 +555,8 @@ public final class WALProtos {
return writeTime_;
}
// optional .UUID clusterId = 5;
public static final int CLUSTERID_FIELD_NUMBER = 5;
// optional .UUID cluster_id = 5;
public static final int CLUSTER_ID_FIELD_NUMBER = 5;
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.UUID clusterId_;
public boolean hasClusterId() {
return ((bitField0_ & 0x00000010) == 0x00000010);
@ -589,8 +589,8 @@ public final class WALProtos {
return scopes_.get(index);
}
// optional uint32 followingKvCount = 7;
public static final int FOLLOWINGKVCOUNT_FIELD_NUMBER = 7;
// optional uint32 following_kv_count = 7;
public static final int FOLLOWING_KV_COUNT_FIELD_NUMBER = 7;
private int followingKvCount_;
public boolean hasFollowingKvCount() {
return ((bitField0_ & 0x00000020) == 0x00000020);
@ -771,23 +771,23 @@ public final class WALProtos {
int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode();
if (hasEncodedRegionName()) {
hash = (37 * hash) + ENCODEDREGIONNAME_FIELD_NUMBER;
hash = (37 * hash) + ENCODED_REGION_NAME_FIELD_NUMBER;
hash = (53 * hash) + getEncodedRegionName().hashCode();
}
if (hasTableName()) {
hash = (37 * hash) + TABLENAME_FIELD_NUMBER;
hash = (37 * hash) + TABLE_NAME_FIELD_NUMBER;
hash = (53 * hash) + getTableName().hashCode();
}
if (hasLogSequenceNumber()) {
hash = (37 * hash) + LOGSEQUENCENUMBER_FIELD_NUMBER;
hash = (37 * hash) + LOG_SEQUENCE_NUMBER_FIELD_NUMBER;
hash = (53 * hash) + hashLong(getLogSequenceNumber());
}
if (hasWriteTime()) {
hash = (37 * hash) + WRITETIME_FIELD_NUMBER;
hash = (37 * hash) + WRITE_TIME_FIELD_NUMBER;
hash = (53 * hash) + hashLong(getWriteTime());
}
if (hasClusterId()) {
hash = (37 * hash) + CLUSTERID_FIELD_NUMBER;
hash = (37 * hash) + CLUSTER_ID_FIELD_NUMBER;
hash = (53 * hash) + getClusterId().hashCode();
}
if (getScopesCount() > 0) {
@ -795,7 +795,7 @@ public final class WALProtos {
hash = (53 * hash) + getScopesList().hashCode();
}
if (hasFollowingKvCount()) {
hash = (37 * hash) + FOLLOWINGKVCOUNT_FIELD_NUMBER;
hash = (37 * hash) + FOLLOWING_KV_COUNT_FIELD_NUMBER;
hash = (53 * hash) + getFollowingKvCount();
}
hash = (29 * hash) + getUnknownFields().hashCode();
@ -1178,7 +1178,7 @@ public final class WALProtos {
private int bitField0_;
// required bytes encodedRegionName = 1;
// required bytes encoded_region_name = 1;
private com.google.protobuf.ByteString encodedRegionName_ = com.google.protobuf.ByteString.EMPTY;
public boolean hasEncodedRegionName() {
return ((bitField0_ & 0x00000001) == 0x00000001);
@ -1202,7 +1202,7 @@ public final class WALProtos {
return this;
}
// required bytes tableName = 2;
// required bytes table_name = 2;
private com.google.protobuf.ByteString tableName_ = com.google.protobuf.ByteString.EMPTY;
public boolean hasTableName() {
return ((bitField0_ & 0x00000002) == 0x00000002);
@ -1226,7 +1226,7 @@ public final class WALProtos {
return this;
}
// required uint64 logSequenceNumber = 3;
// required uint64 log_sequence_number = 3;
private long logSequenceNumber_ ;
public boolean hasLogSequenceNumber() {
return ((bitField0_ & 0x00000004) == 0x00000004);
@ -1247,7 +1247,7 @@ public final class WALProtos {
return this;
}
// required uint64 writeTime = 4;
// required uint64 write_time = 4;
private long writeTime_ ;
public boolean hasWriteTime() {
return ((bitField0_ & 0x00000008) == 0x00000008);
@ -1268,7 +1268,7 @@ public final class WALProtos {
return this;
}
// optional .UUID clusterId = 5;
// optional .UUID cluster_id = 5;
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.UUID clusterId_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.UUID.getDefaultInstance();
private com.google.protobuf.SingleFieldBuilder<
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.UUID, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.UUID.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.UUIDOrBuilder> clusterIdBuilder_;
@ -1544,7 +1544,7 @@ public final class WALProtos {
return scopesBuilder_;
}
// optional uint32 followingKvCount = 7;
// optional uint32 following_kv_count = 7;
private int followingKvCount_ ;
public boolean hasFollowingKvCount() {
return ((bitField0_ & 0x00000040) == 0x00000040);
@ -1583,7 +1583,7 @@ public final class WALProtos {
boolean hasFamily();
com.google.protobuf.ByteString getFamily();
// required .ScopeType scopeType = 2;
// required .ScopeType scope_type = 2;
boolean hasScopeType();
org.apache.hadoop.hbase.protobuf.generated.WALProtos.ScopeType getScopeType();
}
@ -1626,8 +1626,8 @@ public final class WALProtos {
return family_;
}
// required .ScopeType scopeType = 2;
public static final int SCOPETYPE_FIELD_NUMBER = 2;
// required .ScopeType scope_type = 2;
public static final int SCOPE_TYPE_FIELD_NUMBER = 2;
private org.apache.hadoop.hbase.protobuf.generated.WALProtos.ScopeType scopeType_;
public boolean hasScopeType() {
return ((bitField0_ & 0x00000002) == 0x00000002);
@ -1730,7 +1730,7 @@ public final class WALProtos {
hash = (53 * hash) + getFamily().hashCode();
}
if (hasScopeType()) {
hash = (37 * hash) + SCOPETYPE_FIELD_NUMBER;
hash = (37 * hash) + SCOPE_TYPE_FIELD_NUMBER;
hash = (53 * hash) + hashEnum(getScopeType());
}
hash = (29 * hash) + getUnknownFields().hashCode();
@ -2006,7 +2006,7 @@ public final class WALProtos {
return this;
}
// required .ScopeType scopeType = 2;
// required .ScopeType scope_type = 2;
private org.apache.hadoop.hbase.protobuf.generated.WALProtos.ScopeType scopeType_ = org.apache.hadoop.hbase.protobuf.generated.WALProtos.ScopeType.REPLICATION_SCOPE_LOCAL;
public boolean hasScopeType() {
return ((bitField0_ & 0x00000002) == 0x00000002);
@ -2044,29 +2044,29 @@ public final class WALProtos {
public interface CompactionDescriptorOrBuilder
extends com.google.protobuf.MessageOrBuilder {
// required bytes tableName = 1;
// required bytes table_name = 1;
boolean hasTableName();
com.google.protobuf.ByteString getTableName();
// required bytes encodedRegionName = 2;
// required bytes encoded_region_name = 2;
boolean hasEncodedRegionName();
com.google.protobuf.ByteString getEncodedRegionName();
// required bytes familyName = 3;
// required bytes family_name = 3;
boolean hasFamilyName();
com.google.protobuf.ByteString getFamilyName();
// repeated string compactionInput = 4;
// repeated string compaction_input = 4;
java.util.List<String> getCompactionInputList();
int getCompactionInputCount();
String getCompactionInput(int index);
// repeated string compactionOutput = 5;
// repeated string compaction_output = 5;
java.util.List<String> getCompactionOutputList();
int getCompactionOutputCount();
String getCompactionOutput(int index);
// required string storeHomeDir = 6;
// required string store_home_dir = 6;
boolean hasStoreHomeDir();
String getStoreHomeDir();
}
@ -2099,8 +2099,8 @@ public final class WALProtos {
}
private int bitField0_;
// required bytes tableName = 1;
public static final int TABLENAME_FIELD_NUMBER = 1;
// required bytes table_name = 1;
public static final int TABLE_NAME_FIELD_NUMBER = 1;
private com.google.protobuf.ByteString tableName_;
public boolean hasTableName() {
return ((bitField0_ & 0x00000001) == 0x00000001);
@ -2109,8 +2109,8 @@ public final class WALProtos {
return tableName_;
}
// required bytes encodedRegionName = 2;
public static final int ENCODEDREGIONNAME_FIELD_NUMBER = 2;
// required bytes encoded_region_name = 2;
public static final int ENCODED_REGION_NAME_FIELD_NUMBER = 2;
private com.google.protobuf.ByteString encodedRegionName_;
public boolean hasEncodedRegionName() {
return ((bitField0_ & 0x00000002) == 0x00000002);
@ -2119,8 +2119,8 @@ public final class WALProtos {
return encodedRegionName_;
}
// required bytes familyName = 3;
public static final int FAMILYNAME_FIELD_NUMBER = 3;
// required bytes family_name = 3;
public static final int FAMILY_NAME_FIELD_NUMBER = 3;
private com.google.protobuf.ByteString familyName_;
public boolean hasFamilyName() {
return ((bitField0_ & 0x00000004) == 0x00000004);
@ -2129,8 +2129,8 @@ public final class WALProtos {
return familyName_;
}
// repeated string compactionInput = 4;
public static final int COMPACTIONINPUT_FIELD_NUMBER = 4;
// repeated string compaction_input = 4;
public static final int COMPACTION_INPUT_FIELD_NUMBER = 4;
private com.google.protobuf.LazyStringList compactionInput_;
public java.util.List<String>
getCompactionInputList() {
@ -2143,8 +2143,8 @@ public final class WALProtos {
return compactionInput_.get(index);
}
// repeated string compactionOutput = 5;
public static final int COMPACTIONOUTPUT_FIELD_NUMBER = 5;
// repeated string compaction_output = 5;
public static final int COMPACTION_OUTPUT_FIELD_NUMBER = 5;
private com.google.protobuf.LazyStringList compactionOutput_;
public java.util.List<String>
getCompactionOutputList() {
@ -2157,8 +2157,8 @@ public final class WALProtos {
return compactionOutput_.get(index);
}
// required string storeHomeDir = 6;
public static final int STOREHOMEDIR_FIELD_NUMBER = 6;
// required string store_home_dir = 6;
public static final int STORE_HOME_DIR_FIELD_NUMBER = 6;
private java.lang.Object storeHomeDir_;
public boolean hasStoreHomeDir() {
return ((bitField0_ & 0x00000008) == 0x00000008);
@ -2343,27 +2343,27 @@ public final class WALProtos {
int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode();
if (hasTableName()) {
hash = (37 * hash) + TABLENAME_FIELD_NUMBER;
hash = (37 * hash) + TABLE_NAME_FIELD_NUMBER;
hash = (53 * hash) + getTableName().hashCode();
}
if (hasEncodedRegionName()) {
hash = (37 * hash) + ENCODEDREGIONNAME_FIELD_NUMBER;
hash = (37 * hash) + ENCODED_REGION_NAME_FIELD_NUMBER;
hash = (53 * hash) + getEncodedRegionName().hashCode();
}
if (hasFamilyName()) {
hash = (37 * hash) + FAMILYNAME_FIELD_NUMBER;
hash = (37 * hash) + FAMILY_NAME_FIELD_NUMBER;
hash = (53 * hash) + getFamilyName().hashCode();
}
if (getCompactionInputCount() > 0) {
hash = (37 * hash) + COMPACTIONINPUT_FIELD_NUMBER;
hash = (37 * hash) + COMPACTION_INPUT_FIELD_NUMBER;
hash = (53 * hash) + getCompactionInputList().hashCode();
}
if (getCompactionOutputCount() > 0) {
hash = (37 * hash) + COMPACTIONOUTPUT_FIELD_NUMBER;
hash = (37 * hash) + COMPACTION_OUTPUT_FIELD_NUMBER;
hash = (53 * hash) + getCompactionOutputList().hashCode();
}
if (hasStoreHomeDir()) {
hash = (37 * hash) + STOREHOMEDIR_FIELD_NUMBER;
hash = (37 * hash) + STORE_HOME_DIR_FIELD_NUMBER;
hash = (53 * hash) + getStoreHomeDir().hashCode();
}
hash = (29 * hash) + getUnknownFields().hashCode();
@ -2691,7 +2691,7 @@ public final class WALProtos {
private int bitField0_;
// required bytes tableName = 1;
// required bytes table_name = 1;
private com.google.protobuf.ByteString tableName_ = com.google.protobuf.ByteString.EMPTY;
public boolean hasTableName() {
return ((bitField0_ & 0x00000001) == 0x00000001);
@ -2715,7 +2715,7 @@ public final class WALProtos {
return this;
}
// required bytes encodedRegionName = 2;
// required bytes encoded_region_name = 2;
private com.google.protobuf.ByteString encodedRegionName_ = com.google.protobuf.ByteString.EMPTY;
public boolean hasEncodedRegionName() {
return ((bitField0_ & 0x00000002) == 0x00000002);
@ -2739,7 +2739,7 @@ public final class WALProtos {
return this;
}
// required bytes familyName = 3;
// required bytes family_name = 3;
private com.google.protobuf.ByteString familyName_ = com.google.protobuf.ByteString.EMPTY;
public boolean hasFamilyName() {
return ((bitField0_ & 0x00000004) == 0x00000004);
@ -2763,7 +2763,7 @@ public final class WALProtos {
return this;
}
// repeated string compactionInput = 4;
// repeated string compaction_input = 4;
private com.google.protobuf.LazyStringList compactionInput_ = com.google.protobuf.LazyStringArrayList.EMPTY;
private void ensureCompactionInputIsMutable() {
if (!((bitField0_ & 0x00000008) == 0x00000008)) {
@ -2819,7 +2819,7 @@ public final class WALProtos {
onChanged();
}
// repeated string compactionOutput = 5;
// repeated string compaction_output = 5;
private com.google.protobuf.LazyStringList compactionOutput_ = com.google.protobuf.LazyStringArrayList.EMPTY;
private void ensureCompactionOutputIsMutable() {
if (!((bitField0_ & 0x00000010) == 0x00000010)) {
@ -2875,7 +2875,7 @@ public final class WALProtos {
onChanged();
}
// required string storeHomeDir = 6;
// required string store_home_dir = 6;
private java.lang.Object storeHomeDir_ = "";
public boolean hasStoreHomeDir() {
return ((bitField0_ & 0x00000020) == 0x00000020);
@ -3256,22 +3256,23 @@ public final class WALProtos {
descriptor;
static {
java.lang.String[] descriptorData = {
"\n\tWAL.proto\032\013hbase.proto\"#\n\tWALHeader\022\026\n" +
"\016hasCompression\030\001 \001(\010\"\266\001\n\006WALKey\022\031\n\021enco" +
"dedRegionName\030\001 \002(\014\022\021\n\ttableName\030\002 \002(\014\022\031" +
"\n\021logSequenceNumber\030\003 \002(\004\022\021\n\twriteTime\030\004" +
" \002(\004\022\030\n\tclusterId\030\005 \001(\0132\005.UUID\022\034\n\006scopes" +
"\030\006 \003(\0132\014.FamilyScope\022\030\n\020followingKvCount" +
"\030\007 \001(\r\"<\n\013FamilyScope\022\016\n\006family\030\001 \002(\014\022\035\n" +
"\tscopeType\030\002 \002(\0162\n.ScopeType\"\241\001\n\024Compact" +
"ionDescriptor\022\021\n\ttableName\030\001 \002(\014\022\031\n\021enco" +
"dedRegionName\030\002 \002(\014\022\022\n\nfamilyName\030\003 \002(\014\022",
"\027\n\017compactionInput\030\004 \003(\t\022\030\n\020compactionOu" +
"tput\030\005 \003(\t\022\024\n\014storeHomeDir\030\006 \002(\t\"\014\n\nWALT" +
"railer*F\n\tScopeType\022\033\n\027REPLICATION_SCOPE" +
"_LOCAL\020\000\022\034\n\030REPLICATION_SCOPE_GLOBAL\020\001B?" +
"\n*org.apache.hadoop.hbase.protobuf.gener" +
"atedB\tWALProtosH\001\210\001\000\240\001\001"
"\n\tWAL.proto\032\013hbase.proto\"$\n\tWALHeader\022\027\n" +
"\017has_compression\030\001 \001(\010\"\277\001\n\006WALKey\022\033\n\023enc" +
"oded_region_name\030\001 \002(\014\022\022\n\ntable_name\030\002 \002" +
"(\014\022\033\n\023log_sequence_number\030\003 \002(\004\022\022\n\nwrite" +
"_time\030\004 \002(\004\022\031\n\ncluster_id\030\005 \001(\0132\005.UUID\022\034" +
"\n\006scopes\030\006 \003(\0132\014.FamilyScope\022\032\n\022followin" +
"g_kv_count\030\007 \001(\r\"=\n\013FamilyScope\022\016\n\006famil" +
"y\030\001 \002(\014\022\036\n\nscope_type\030\002 \002(\0162\n.ScopeType\"" +
"\251\001\n\024CompactionDescriptor\022\022\n\ntable_name\030\001" +
" \002(\014\022\033\n\023encoded_region_name\030\002 \002(\014\022\023\n\013fam",
"ily_name\030\003 \002(\014\022\030\n\020compaction_input\030\004 \003(\t" +
"\022\031\n\021compaction_output\030\005 \003(\t\022\026\n\016store_hom" +
"e_dir\030\006 \002(\t\"\014\n\nWALTrailer*F\n\tScopeType\022\033" +
"\n\027REPLICATION_SCOPE_LOCAL\020\000\022\034\n\030REPLICATI" +
"ON_SCOPE_GLOBAL\020\001B?\n*org.apache.hadoop.h" +
"base.protobuf.generatedB\tWALProtosH\001\210\001\000\240" +
"\001\001"
};
com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() {

View File

@ -16,7 +16,7 @@ public final class ZooKeeperProtos {
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName getServer();
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerNameOrBuilder getServerOrBuilder();
// optional uint32 rpcVersion = 2;
// optional uint32 rpc_version = 2;
boolean hasRpcVersion();
int getRpcVersion();
}
@ -62,8 +62,8 @@ public final class ZooKeeperProtos {
return server_;
}
// optional uint32 rpcVersion = 2;
public static final int RPCVERSION_FIELD_NUMBER = 2;
// optional uint32 rpc_version = 2;
public static final int RPC_VERSION_FIELD_NUMBER = 2;
private int rpcVersion_;
public boolean hasRpcVersion() {
return ((bitField0_ & 0x00000002) == 0x00000002);
@ -166,7 +166,7 @@ public final class ZooKeeperProtos {
hash = (53 * hash) + getServer().hashCode();
}
if (hasRpcVersion()) {
hash = (37 * hash) + RPCVERSION_FIELD_NUMBER;
hash = (37 * hash) + RPC_VERSION_FIELD_NUMBER;
hash = (53 * hash) + getRpcVersion();
}
hash = (29 * hash) + getUnknownFields().hashCode();
@ -515,7 +515,7 @@ public final class ZooKeeperProtos {
return serverBuilder_;
}
// optional uint32 rpcVersion = 2;
// optional uint32 rpc_version = 2;
private int rpcVersion_ ;
public boolean hasRpcVersion() {
return ((bitField0_ & 0x00000002) == 0x00000002);
@ -555,7 +555,7 @@ public final class ZooKeeperProtos {
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName getMaster();
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerNameOrBuilder getMasterOrBuilder();
// optional uint32 rpcVersion = 2;
// optional uint32 rpc_version = 2;
boolean hasRpcVersion();
int getRpcVersion();
}
@ -601,8 +601,8 @@ public final class ZooKeeperProtos {
return master_;
}
// optional uint32 rpcVersion = 2;
public static final int RPCVERSION_FIELD_NUMBER = 2;
// optional uint32 rpc_version = 2;
public static final int RPC_VERSION_FIELD_NUMBER = 2;
private int rpcVersion_;
public boolean hasRpcVersion() {
return ((bitField0_ & 0x00000002) == 0x00000002);
@ -705,7 +705,7 @@ public final class ZooKeeperProtos {
hash = (53 * hash) + getMaster().hashCode();
}
if (hasRpcVersion()) {
hash = (37 * hash) + RPCVERSION_FIELD_NUMBER;
hash = (37 * hash) + RPC_VERSION_FIELD_NUMBER;
hash = (53 * hash) + getRpcVersion();
}
hash = (29 * hash) + getUnknownFields().hashCode();
@ -1054,7 +1054,7 @@ public final class ZooKeeperProtos {
return masterBuilder_;
}
// optional uint32 rpcVersion = 2;
// optional uint32 rpc_version = 2;
private int rpcVersion_ ;
public boolean hasRpcVersion() {
return ((bitField0_ & 0x00000002) == 0x00000002);
@ -1089,7 +1089,7 @@ public final class ZooKeeperProtos {
public interface ClusterUpOrBuilder
extends com.google.protobuf.MessageOrBuilder {
// required string startDate = 1;
// required string start_date = 1;
boolean hasStartDate();
String getStartDate();
}
@ -1122,8 +1122,8 @@ public final class ZooKeeperProtos {
}
private int bitField0_;
// required string startDate = 1;
public static final int STARTDATE_FIELD_NUMBER = 1;
// required string start_date = 1;
public static final int START_DATE_FIELD_NUMBER = 1;
private java.lang.Object startDate_;
public boolean hasStartDate() {
return ((bitField0_ & 0x00000001) == 0x00000001);
@ -1227,7 +1227,7 @@ public final class ZooKeeperProtos {
int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode();
if (hasStartDate()) {
hash = (37 * hash) + STARTDATE_FIELD_NUMBER;
hash = (37 * hash) + START_DATE_FIELD_NUMBER;
hash = (53 * hash) + getStartDate().hashCode();
}
hash = (29 * hash) + getUnknownFields().hashCode();
@ -1455,7 +1455,7 @@ public final class ZooKeeperProtos {
private int bitField0_;
// required string startDate = 1;
// required string start_date = 1;
private java.lang.Object startDate_ = "";
public boolean hasStartDate() {
return ((bitField0_ & 0x00000001) == 0x00000001);
@ -1505,19 +1505,19 @@ public final class ZooKeeperProtos {
public interface RegionTransitionOrBuilder
extends com.google.protobuf.MessageOrBuilder {
// required uint32 eventTypeCode = 1;
// required uint32 event_type_code = 1;
boolean hasEventTypeCode();
int getEventTypeCode();
// required bytes regionName = 2;
// required bytes region_name = 2;
boolean hasRegionName();
com.google.protobuf.ByteString getRegionName();
// required uint64 createTime = 3;
// required uint64 create_time = 3;
boolean hasCreateTime();
long getCreateTime();
// required .ServerName serverName = 4;
// required .ServerName server_name = 4;
boolean hasServerName();
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName getServerName();
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerNameOrBuilder getServerNameOrBuilder();
@ -1555,8 +1555,8 @@ public final class ZooKeeperProtos {
}
private int bitField0_;
// required uint32 eventTypeCode = 1;
public static final int EVENTTYPECODE_FIELD_NUMBER = 1;
// required uint32 event_type_code = 1;
public static final int EVENT_TYPE_CODE_FIELD_NUMBER = 1;
private int eventTypeCode_;
public boolean hasEventTypeCode() {
return ((bitField0_ & 0x00000001) == 0x00000001);
@ -1565,8 +1565,8 @@ public final class ZooKeeperProtos {
return eventTypeCode_;
}
// required bytes regionName = 2;
public static final int REGIONNAME_FIELD_NUMBER = 2;
// required bytes region_name = 2;
public static final int REGION_NAME_FIELD_NUMBER = 2;
private com.google.protobuf.ByteString regionName_;
public boolean hasRegionName() {
return ((bitField0_ & 0x00000002) == 0x00000002);
@ -1575,8 +1575,8 @@ public final class ZooKeeperProtos {
return regionName_;
}
// required uint64 createTime = 3;
public static final int CREATETIME_FIELD_NUMBER = 3;
// required uint64 create_time = 3;
public static final int CREATE_TIME_FIELD_NUMBER = 3;
private long createTime_;
public boolean hasCreateTime() {
return ((bitField0_ & 0x00000004) == 0x00000004);
@ -1585,8 +1585,8 @@ public final class ZooKeeperProtos {
return createTime_;
}
// required .ServerName serverName = 4;
public static final int SERVERNAME_FIELD_NUMBER = 4;
// required .ServerName server_name = 4;
public static final int SERVER_NAME_FIELD_NUMBER = 4;
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName serverName_;
public boolean hasServerName() {
return ((bitField0_ & 0x00000008) == 0x00000008);
@ -1749,19 +1749,19 @@ public final class ZooKeeperProtos {
int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode();
if (hasEventTypeCode()) {
hash = (37 * hash) + EVENTTYPECODE_FIELD_NUMBER;
hash = (37 * hash) + EVENT_TYPE_CODE_FIELD_NUMBER;
hash = (53 * hash) + getEventTypeCode();
}
if (hasRegionName()) {
hash = (37 * hash) + REGIONNAME_FIELD_NUMBER;
hash = (37 * hash) + REGION_NAME_FIELD_NUMBER;
hash = (53 * hash) + getRegionName().hashCode();
}
if (hasCreateTime()) {
hash = (37 * hash) + CREATETIME_FIELD_NUMBER;
hash = (37 * hash) + CREATE_TIME_FIELD_NUMBER;
hash = (53 * hash) + hashLong(getCreateTime());
}
if (hasServerName()) {
hash = (37 * hash) + SERVERNAME_FIELD_NUMBER;
hash = (37 * hash) + SERVER_NAME_FIELD_NUMBER;
hash = (53 * hash) + getServerName().hashCode();
}
if (hasPayload()) {
@ -2078,7 +2078,7 @@ public final class ZooKeeperProtos {
private int bitField0_;
// required uint32 eventTypeCode = 1;
// required uint32 event_type_code = 1;
private int eventTypeCode_ ;
public boolean hasEventTypeCode() {
return ((bitField0_ & 0x00000001) == 0x00000001);
@ -2099,7 +2099,7 @@ public final class ZooKeeperProtos {
return this;
}
// required bytes regionName = 2;
// required bytes region_name = 2;
private com.google.protobuf.ByteString regionName_ = com.google.protobuf.ByteString.EMPTY;
public boolean hasRegionName() {
return ((bitField0_ & 0x00000002) == 0x00000002);
@ -2123,7 +2123,7 @@ public final class ZooKeeperProtos {
return this;
}
// required uint64 createTime = 3;
// required uint64 create_time = 3;
private long createTime_ ;
public boolean hasCreateTime() {
return ((bitField0_ & 0x00000004) == 0x00000004);
@ -2144,7 +2144,7 @@ public final class ZooKeeperProtos {
return this;
}
// required .ServerName serverName = 4;
// required .ServerName server_name = 4;
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName serverName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.getDefaultInstance();
private com.google.protobuf.SingleFieldBuilder<
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerNameOrBuilder> serverNameBuilder_;
@ -2276,7 +2276,7 @@ public final class ZooKeeperProtos {
boolean hasState();
org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.SplitLogTask.State getState();
// required .ServerName serverName = 2;
// required .ServerName server_name = 2;
boolean hasServerName();
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName getServerName();
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerNameOrBuilder getServerNameOrBuilder();
@ -2398,8 +2398,8 @@ public final class ZooKeeperProtos {
return state_;
}
// required .ServerName serverName = 2;
public static final int SERVERNAME_FIELD_NUMBER = 2;
// required .ServerName server_name = 2;
public static final int SERVER_NAME_FIELD_NUMBER = 2;
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName serverName_;
public boolean hasServerName() {
return ((bitField0_ & 0x00000002) == 0x00000002);
@ -2509,7 +2509,7 @@ public final class ZooKeeperProtos {
hash = (53 * hash) + hashEnum(getState());
}
if (hasServerName()) {
hash = (37 * hash) + SERVERNAME_FIELD_NUMBER;
hash = (37 * hash) + SERVER_NAME_FIELD_NUMBER;
hash = (53 * hash) + getServerName().hashCode();
}
hash = (29 * hash) + getUnknownFields().hashCode();
@ -2802,7 +2802,7 @@ public final class ZooKeeperProtos {
return this;
}
// required .ServerName serverName = 2;
// required .ServerName server_name = 2;
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName serverName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.getDefaultInstance();
private com.google.protobuf.SingleFieldBuilder<
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerNameOrBuilder> serverNameBuilder_;
@ -4621,7 +4621,7 @@ public final class ZooKeeperProtos {
public interface ReplicationLockOrBuilder
extends com.google.protobuf.MessageOrBuilder {
// required string lockOwner = 1;
// required string lock_owner = 1;
boolean hasLockOwner();
String getLockOwner();
}
@ -4654,8 +4654,8 @@ public final class ZooKeeperProtos {
}
private int bitField0_;
// required string lockOwner = 1;
public static final int LOCKOWNER_FIELD_NUMBER = 1;
// required string lock_owner = 1;
public static final int LOCK_OWNER_FIELD_NUMBER = 1;
private java.lang.Object lockOwner_;
public boolean hasLockOwner() {
return ((bitField0_ & 0x00000001) == 0x00000001);
@ -4759,7 +4759,7 @@ public final class ZooKeeperProtos {
int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode();
if (hasLockOwner()) {
hash = (37 * hash) + LOCKOWNER_FIELD_NUMBER;
hash = (37 * hash) + LOCK_OWNER_FIELD_NUMBER;
hash = (53 * hash) + getLockOwner().hashCode();
}
hash = (29 * hash) + getUnknownFields().hashCode();
@ -4987,7 +4987,7 @@ public final class ZooKeeperProtos {
private int bitField0_;
// required string lockOwner = 1;
// required string lock_owner = 1;
private java.lang.Object lockOwner_ = "";
public boolean hasLockOwner() {
return ((bitField0_ & 0x00000001) == 0x00000001);
@ -5037,20 +5037,20 @@ public final class ZooKeeperProtos {
public interface TableLockOrBuilder
extends com.google.protobuf.MessageOrBuilder {
// optional bytes tableName = 1;
// optional bytes table_name = 1;
boolean hasTableName();
com.google.protobuf.ByteString getTableName();
// optional .ServerName lockOwner = 2;
// optional .ServerName lock_owner = 2;
boolean hasLockOwner();
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName getLockOwner();
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerNameOrBuilder getLockOwnerOrBuilder();
// optional int64 threadId = 3;
// optional int64 thread_id = 3;
boolean hasThreadId();
long getThreadId();
// optional bool isShared = 4;
// optional bool is_shared = 4;
boolean hasIsShared();
boolean getIsShared();
@ -5058,7 +5058,7 @@ public final class ZooKeeperProtos {
boolean hasPurpose();
String getPurpose();
// optional int64 createTime = 6;
// optional int64 create_time = 6;
boolean hasCreateTime();
long getCreateTime();
}
@ -5091,8 +5091,8 @@ public final class ZooKeeperProtos {
}
private int bitField0_;
// optional bytes tableName = 1;
public static final int TABLENAME_FIELD_NUMBER = 1;
// optional bytes table_name = 1;
public static final int TABLE_NAME_FIELD_NUMBER = 1;
private com.google.protobuf.ByteString tableName_;
public boolean hasTableName() {
return ((bitField0_ & 0x00000001) == 0x00000001);
@ -5101,8 +5101,8 @@ public final class ZooKeeperProtos {
return tableName_;
}
// optional .ServerName lockOwner = 2;
public static final int LOCKOWNER_FIELD_NUMBER = 2;
// optional .ServerName lock_owner = 2;
public static final int LOCK_OWNER_FIELD_NUMBER = 2;
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName lockOwner_;
public boolean hasLockOwner() {
return ((bitField0_ & 0x00000002) == 0x00000002);
@ -5114,8 +5114,8 @@ public final class ZooKeeperProtos {
return lockOwner_;
}
// optional int64 threadId = 3;
public static final int THREADID_FIELD_NUMBER = 3;
// optional int64 thread_id = 3;
public static final int THREAD_ID_FIELD_NUMBER = 3;
private long threadId_;
public boolean hasThreadId() {
return ((bitField0_ & 0x00000004) == 0x00000004);
@ -5124,8 +5124,8 @@ public final class ZooKeeperProtos {
return threadId_;
}
// optional bool isShared = 4;
public static final int ISSHARED_FIELD_NUMBER = 4;
// optional bool is_shared = 4;
public static final int IS_SHARED_FIELD_NUMBER = 4;
private boolean isShared_;
public boolean hasIsShared() {
return ((bitField0_ & 0x00000008) == 0x00000008);
@ -5166,8 +5166,8 @@ public final class ZooKeeperProtos {
}
}
// optional int64 createTime = 6;
public static final int CREATETIME_FIELD_NUMBER = 6;
// optional int64 create_time = 6;
public static final int CREATE_TIME_FIELD_NUMBER = 6;
private long createTime_;
public boolean hasCreateTime() {
return ((bitField0_ & 0x00000020) == 0x00000020);
@ -5316,19 +5316,19 @@ public final class ZooKeeperProtos {
int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode();
if (hasTableName()) {
hash = (37 * hash) + TABLENAME_FIELD_NUMBER;
hash = (37 * hash) + TABLE_NAME_FIELD_NUMBER;
hash = (53 * hash) + getTableName().hashCode();
}
if (hasLockOwner()) {
hash = (37 * hash) + LOCKOWNER_FIELD_NUMBER;
hash = (37 * hash) + LOCK_OWNER_FIELD_NUMBER;
hash = (53 * hash) + getLockOwner().hashCode();
}
if (hasThreadId()) {
hash = (37 * hash) + THREADID_FIELD_NUMBER;
hash = (37 * hash) + THREAD_ID_FIELD_NUMBER;
hash = (53 * hash) + hashLong(getThreadId());
}
if (hasIsShared()) {
hash = (37 * hash) + ISSHARED_FIELD_NUMBER;
hash = (37 * hash) + IS_SHARED_FIELD_NUMBER;
hash = (53 * hash) + hashBoolean(getIsShared());
}
if (hasPurpose()) {
@ -5336,7 +5336,7 @@ public final class ZooKeeperProtos {
hash = (53 * hash) + getPurpose().hashCode();
}
if (hasCreateTime()) {
hash = (37 * hash) + CREATETIME_FIELD_NUMBER;
hash = (37 * hash) + CREATE_TIME_FIELD_NUMBER;
hash = (53 * hash) + hashLong(getCreateTime());
}
hash = (29 * hash) + getUnknownFields().hashCode();
@ -5649,7 +5649,7 @@ public final class ZooKeeperProtos {
private int bitField0_;
// optional bytes tableName = 1;
// optional bytes table_name = 1;
private com.google.protobuf.ByteString tableName_ = com.google.protobuf.ByteString.EMPTY;
public boolean hasTableName() {
return ((bitField0_ & 0x00000001) == 0x00000001);
@ -5673,7 +5673,7 @@ public final class ZooKeeperProtos {
return this;
}
// optional .ServerName lockOwner = 2;
// optional .ServerName lock_owner = 2;
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName lockOwner_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.getDefaultInstance();
private com.google.protobuf.SingleFieldBuilder<
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerNameOrBuilder> lockOwnerBuilder_;
@ -5763,7 +5763,7 @@ public final class ZooKeeperProtos {
return lockOwnerBuilder_;
}
// optional int64 threadId = 3;
// optional int64 thread_id = 3;
private long threadId_ ;
public boolean hasThreadId() {
return ((bitField0_ & 0x00000004) == 0x00000004);
@ -5784,7 +5784,7 @@ public final class ZooKeeperProtos {
return this;
}
// optional bool isShared = 4;
// optional bool is_shared = 4;
private boolean isShared_ ;
public boolean hasIsShared() {
return ((bitField0_ & 0x00000008) == 0x00000008);
@ -5841,7 +5841,7 @@ public final class ZooKeeperProtos {
onChanged();
}
// optional int64 createTime = 6;
// optional int64 create_time = 6;
private long createTime_ ;
public boolean hasCreateTime() {
return ((bitField0_ & 0x00000020) == 0x00000020);
@ -5876,11 +5876,11 @@ public final class ZooKeeperProtos {
public interface StoreSequenceIdOrBuilder
extends com.google.protobuf.MessageOrBuilder {
// required bytes familyName = 1;
// required bytes family_name = 1;
boolean hasFamilyName();
com.google.protobuf.ByteString getFamilyName();
// required uint64 sequenceId = 2;
// required uint64 sequence_id = 2;
boolean hasSequenceId();
long getSequenceId();
}
@ -5913,8 +5913,8 @@ public final class ZooKeeperProtos {
}
private int bitField0_;
// required bytes familyName = 1;
public static final int FAMILYNAME_FIELD_NUMBER = 1;
// required bytes family_name = 1;
public static final int FAMILY_NAME_FIELD_NUMBER = 1;
private com.google.protobuf.ByteString familyName_;
public boolean hasFamilyName() {
return ((bitField0_ & 0x00000001) == 0x00000001);
@ -5923,8 +5923,8 @@ public final class ZooKeeperProtos {
return familyName_;
}
// required uint64 sequenceId = 2;
public static final int SEQUENCEID_FIELD_NUMBER = 2;
// required uint64 sequence_id = 2;
public static final int SEQUENCE_ID_FIELD_NUMBER = 2;
private long sequenceId_;
public boolean hasSequenceId() {
return ((bitField0_ & 0x00000002) == 0x00000002);
@ -6023,11 +6023,11 @@ public final class ZooKeeperProtos {
int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode();
if (hasFamilyName()) {
hash = (37 * hash) + FAMILYNAME_FIELD_NUMBER;
hash = (37 * hash) + FAMILY_NAME_FIELD_NUMBER;
hash = (53 * hash) + getFamilyName().hashCode();
}
if (hasSequenceId()) {
hash = (37 * hash) + SEQUENCEID_FIELD_NUMBER;
hash = (37 * hash) + SEQUENCE_ID_FIELD_NUMBER;
hash = (53 * hash) + hashLong(getSequenceId());
}
hash = (29 * hash) + getUnknownFields().hashCode();
@ -6273,7 +6273,7 @@ public final class ZooKeeperProtos {
private int bitField0_;
// required bytes familyName = 1;
// required bytes family_name = 1;
private com.google.protobuf.ByteString familyName_ = com.google.protobuf.ByteString.EMPTY;
public boolean hasFamilyName() {
return ((bitField0_ & 0x00000001) == 0x00000001);
@ -6297,7 +6297,7 @@ public final class ZooKeeperProtos {
return this;
}
// required uint64 sequenceId = 2;
// required uint64 sequence_id = 2;
private long sequenceId_ ;
public boolean hasSequenceId() {
return ((bitField0_ & 0x00000002) == 0x00000002);
@ -6332,11 +6332,11 @@ public final class ZooKeeperProtos {
public interface RegionStoreSequenceIdsOrBuilder
extends com.google.protobuf.MessageOrBuilder {
// required uint64 lastFlushedSequenceId = 1;
// required uint64 last_flushed_sequence_id = 1;
boolean hasLastFlushedSequenceId();
long getLastFlushedSequenceId();
// repeated .StoreSequenceId storeSequenceId = 2;
// repeated .StoreSequenceId store_sequence_id = 2;
java.util.List<org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.StoreSequenceId>
getStoreSequenceIdList();
org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.StoreSequenceId getStoreSequenceId(int index);
@ -6375,8 +6375,8 @@ public final class ZooKeeperProtos {
}
private int bitField0_;
// required uint64 lastFlushedSequenceId = 1;
public static final int LASTFLUSHEDSEQUENCEID_FIELD_NUMBER = 1;
// required uint64 last_flushed_sequence_id = 1;
public static final int LAST_FLUSHED_SEQUENCE_ID_FIELD_NUMBER = 1;
private long lastFlushedSequenceId_;
public boolean hasLastFlushedSequenceId() {
return ((bitField0_ & 0x00000001) == 0x00000001);
@ -6385,8 +6385,8 @@ public final class ZooKeeperProtos {
return lastFlushedSequenceId_;
}
// repeated .StoreSequenceId storeSequenceId = 2;
public static final int STORESEQUENCEID_FIELD_NUMBER = 2;
// repeated .StoreSequenceId store_sequence_id = 2;
public static final int STORE_SEQUENCE_ID_FIELD_NUMBER = 2;
private java.util.List<org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.StoreSequenceId> storeSequenceId_;
public java.util.List<org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.StoreSequenceId> getStoreSequenceIdList() {
return storeSequenceId_;
@ -6495,11 +6495,11 @@ public final class ZooKeeperProtos {
int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode();
if (hasLastFlushedSequenceId()) {
hash = (37 * hash) + LASTFLUSHEDSEQUENCEID_FIELD_NUMBER;
hash = (37 * hash) + LAST_FLUSHED_SEQUENCE_ID_FIELD_NUMBER;
hash = (53 * hash) + hashLong(getLastFlushedSequenceId());
}
if (getStoreSequenceIdCount() > 0) {
hash = (37 * hash) + STORESEQUENCEID_FIELD_NUMBER;
hash = (37 * hash) + STORE_SEQUENCE_ID_FIELD_NUMBER;
hash = (53 * hash) + getStoreSequenceIdList().hashCode();
}
hash = (29 * hash) + getUnknownFields().hashCode();
@ -6781,7 +6781,7 @@ public final class ZooKeeperProtos {
private int bitField0_;
// required uint64 lastFlushedSequenceId = 1;
// required uint64 last_flushed_sequence_id = 1;
private long lastFlushedSequenceId_ ;
public boolean hasLastFlushedSequenceId() {
return ((bitField0_ & 0x00000001) == 0x00000001);
@ -6802,7 +6802,7 @@ public final class ZooKeeperProtos {
return this;
}
// repeated .StoreSequenceId storeSequenceId = 2;
// repeated .StoreSequenceId store_sequence_id = 2;
private java.util.List<org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.StoreSequenceId> storeSequenceId_ =
java.util.Collections.emptyList();
private void ensureStoreSequenceIdIsMutable() {
@ -7073,36 +7073,36 @@ public final class ZooKeeperProtos {
descriptor;
static {
java.lang.String[] descriptorData = {
"\n\017ZooKeeper.proto\032\013hbase.proto\"C\n\020MetaRe" +
"\n\017ZooKeeper.proto\032\013hbase.proto\"D\n\020MetaRe" +
"gionServer\022\033\n\006server\030\001 \002(\0132\013.ServerName\022" +
"\022\n\nrpcVersion\030\002 \001(\r\"9\n\006Master\022\033\n\006master\030" +
"\001 \002(\0132\013.ServerName\022\022\n\nrpcVersion\030\002 \001(\r\"\036" +
"\n\tClusterUp\022\021\n\tstartDate\030\001 \002(\t\"\203\001\n\020Regio" +
"nTransition\022\025\n\reventTypeCode\030\001 \002(\r\022\022\n\nre" +
"gionName\030\002 \002(\014\022\022\n\ncreateTime\030\003 \002(\004\022\037\n\nse" +
"rverName\030\004 \002(\0132\013.ServerName\022\017\n\007payload\030\005" +
" \001(\014\"\230\001\n\014SplitLogTask\022\"\n\005state\030\001 \002(\0162\023.S" +
"plitLogTask.State\022\037\n\nserverName\030\002 \002(\0132\013.",
"ServerName\"C\n\005State\022\016\n\nUNASSIGNED\020\000\022\t\n\005O" +
"WNED\020\001\022\014\n\010RESIGNED\020\002\022\010\n\004DONE\020\003\022\007\n\003ERR\020\004\"" +
"n\n\005Table\022$\n\005state\030\001 \002(\0162\014.Table.State:\007E" +
"NABLED\"?\n\005State\022\013\n\007ENABLED\020\000\022\014\n\010DISABLED" +
"\020\001\022\r\n\tDISABLING\020\002\022\014\n\010ENABLING\020\003\"%\n\017Repli" +
"cationPeer\022\022\n\nclusterkey\030\001 \002(\t\"^\n\020Replic" +
"ationState\022&\n\005state\030\001 \002(\0162\027.ReplicationS" +
"tate.State\"\"\n\005State\022\013\n\007ENABLED\020\000\022\014\n\010DISA" +
"BLED\020\001\"+\n\027ReplicationHLogPosition\022\020\n\010pos" +
"ition\030\001 \002(\003\"$\n\017ReplicationLock\022\021\n\tlockOw",
"ner\030\001 \002(\t\"\207\001\n\tTableLock\022\021\n\ttableName\030\001 \001" +
"(\014\022\036\n\tlockOwner\030\002 \001(\0132\013.ServerName\022\020\n\010th" +
"readId\030\003 \001(\003\022\020\n\010isShared\030\004 \001(\010\022\017\n\007purpos" +
"e\030\005 \001(\t\022\022\n\ncreateTime\030\006 \001(\003\"9\n\017StoreSequ" +
"enceId\022\022\n\nfamilyName\030\001 \002(\014\022\022\n\nsequenceId" +
"\030\002 \002(\004\"b\n\026RegionStoreSequenceIds\022\035\n\025last" +
"FlushedSequenceId\030\001 \002(\004\022)\n\017storeSequence" +
"Id\030\002 \003(\0132\020.StoreSequenceIdBE\n*org.apache" +
".hadoop.hbase.protobuf.generatedB\017ZooKee" +
"perProtosH\001\210\001\001\240\001\001"
"\023\n\013rpc_version\030\002 \001(\r\":\n\006Master\022\033\n\006master" +
"\030\001 \002(\0132\013.ServerName\022\023\n\013rpc_version\030\002 \001(\r" +
"\"\037\n\tClusterUp\022\022\n\nstart_date\030\001 \002(\t\"\210\001\n\020Re" +
"gionTransition\022\027\n\017event_type_code\030\001 \002(\r\022" +
"\023\n\013region_name\030\002 \002(\014\022\023\n\013create_time\030\003 \002(" +
"\004\022 \n\013server_name\030\004 \002(\0132\013.ServerName\022\017\n\007p" +
"ayload\030\005 \001(\014\"\231\001\n\014SplitLogTask\022\"\n\005state\030\001" +
" \002(\0162\023.SplitLogTask.State\022 \n\013server_name",
"\030\002 \002(\0132\013.ServerName\"C\n\005State\022\016\n\nUNASSIGN" +
"ED\020\000\022\t\n\005OWNED\020\001\022\014\n\010RESIGNED\020\002\022\010\n\004DONE\020\003\022" +
"\007\n\003ERR\020\004\"n\n\005Table\022$\n\005state\030\001 \002(\0162\014.Table" +
".State:\007ENABLED\"?\n\005State\022\013\n\007ENABLED\020\000\022\014\n" +
"\010DISABLED\020\001\022\r\n\tDISABLING\020\002\022\014\n\010ENABLING\020\003" +
"\"%\n\017ReplicationPeer\022\022\n\nclusterkey\030\001 \002(\t\"" +
"^\n\020ReplicationState\022&\n\005state\030\001 \002(\0162\027.Rep" +
"licationState.State\"\"\n\005State\022\013\n\007ENABLED\020" +
"\000\022\014\n\010DISABLED\020\001\"+\n\027ReplicationHLogPositi" +
"on\022\020\n\010position\030\001 \002(\003\"%\n\017ReplicationLock\022",
"\022\n\nlock_owner\030\001 \002(\t\"\214\001\n\tTableLock\022\022\n\ntab" +
"le_name\030\001 \001(\014\022\037\n\nlock_owner\030\002 \001(\0132\013.Serv" +
"erName\022\021\n\tthread_id\030\003 \001(\003\022\021\n\tis_shared\030\004" +
" \001(\010\022\017\n\007purpose\030\005 \001(\t\022\023\n\013create_time\030\006 \001" +
"(\003\";\n\017StoreSequenceId\022\023\n\013family_name\030\001 \002" +
"(\014\022\023\n\013sequence_id\030\002 \002(\004\"g\n\026RegionStoreSe" +
"quenceIds\022 \n\030last_flushed_sequence_id\030\001 " +
"\002(\004\022+\n\021store_sequence_id\030\002 \003(\0132\020.StoreSe" +
"quenceIdBE\n*org.apache.hadoop.hbase.prot" +
"obuf.generatedB\017ZooKeeperProtosH\001\210\001\001\240\001\001"
};
com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() {

View File

@ -85,15 +85,15 @@ message CheckPermissionsResponse {
}
service AccessControlService {
rpc grant(GrantRequest)
rpc Grant(GrantRequest)
returns (GrantResponse);
rpc revoke(RevokeRequest)
rpc Revoke(RevokeRequest)
returns (RevokeResponse);
rpc getUserPermissions(UserPermissionsRequest)
rpc GetUserPermissions(UserPermissionsRequest)
returns (UserPermissionsResponse);
rpc checkPermissions(CheckPermissionsRequest)
rpc CheckPermissions(CheckPermissionsRequest)
returns (CheckPermissionsResponse);
}

View File

@ -30,12 +30,12 @@ import "WAL.proto";
message GetRegionInfoRequest {
required RegionSpecifier region = 1;
optional bool compactionState = 2;
optional bool compaction_state = 2;
}
message GetRegionInfoResponse {
required RegionInfo regionInfo = 1;
optional CompactionState compactionState = 2;
required RegionInfo region_info = 1;
optional CompactionState compaction_state = 2;
enum CompactionState {
NONE = 0;
@ -55,28 +55,28 @@ message GetStoreFileRequest {
}
message GetStoreFileResponse {
repeated string storeFile = 1;
repeated string store_file = 1;
}
message GetOnlineRegionRequest {
}
message GetOnlineRegionResponse {
repeated RegionInfo regionInfo = 1;
repeated RegionInfo region_info = 1;
}
message OpenRegionRequest {
repeated RegionOpenInfo openInfo = 1;
repeated RegionOpenInfo open_info = 1;
message RegionOpenInfo {
required RegionInfo region = 1;
optional uint32 versionOfOfflineNode = 2;
repeated ServerName favoredNodes = 3;
optional uint32 version_of_offline_node = 2;
repeated ServerName favored_nodes = 3;
}
}
message OpenRegionResponse {
repeated RegionOpeningState openingState = 1;
repeated RegionOpeningState opening_state = 1;
enum RegionOpeningState {
OPENED = 0;
@ -91,9 +91,9 @@ message OpenRegionResponse {
*/
message CloseRegionRequest {
required RegionSpecifier region = 1;
optional uint32 versionOfClosingNode = 2;
optional bool transitionInZK = 3 [default = true];
optional ServerName destinationServer = 4;
optional uint32 version_of_closing_node = 2;
optional bool transition_in_ZK = 3 [default = true];
optional ServerName destination_server = 4;
}
message CloseRegionResponse {
@ -107,11 +107,11 @@ message CloseRegionResponse {
*/
message FlushRegionRequest {
required RegionSpecifier region = 1;
optional uint64 ifOlderThanTs = 2;
optional uint64 if_older_than_ts = 2;
}
message FlushRegionResponse {
required uint64 lastFlushTime = 1;
required uint64 last_flush_time = 1;
optional bool flushed = 2;
}
@ -124,7 +124,7 @@ message FlushRegionResponse {
*/
message SplitRegionRequest {
required RegionSpecifier region = 1;
optional bytes splitPoint = 2;
optional bytes split_point = 2;
}
message SplitRegionResponse {
@ -150,8 +150,8 @@ message CompactRegionResponse {
* This method currently closes the regions and then merges them
*/
message MergeRegionsRequest {
required RegionSpecifier regionA = 1;
required RegionSpecifier regionB = 2;
required RegionSpecifier region_a = 1;
required RegionSpecifier region_b = 2;
optional bool forcible = 3 [default = false];
}
@ -163,10 +163,10 @@ message WALEntry {
required WALKey key = 1;
// Following may be null if the KVs/Cells are carried along the side in a cellblock (See
// RPC for more on cellblocks). If Cells/KVs are in a cellblock, this next field is null
// and associatedCellCount has count of Cells associated w/ this WALEntry
repeated bytes keyValueBytes = 2;
// and associated_cell_count has count of Cells associated w/ this WALEntry
repeated bytes key_value_bytes = 2;
// If Cell data is carried alongside in a cellblock, this is count of Cells in the cellblock.
optional int32 associatedCellCount = 3;
optional int32 associated_cell_count = 3;
}
/**
@ -186,7 +186,7 @@ message RollWALWriterRequest {
message RollWALWriterResponse {
// A list of encoded name of regions to flush
repeated bytes regionToFlush = 1;
repeated bytes region_to_flush = 1;
}
message StopServerRequest {
@ -200,54 +200,54 @@ message GetServerInfoRequest {
}
message ServerInfo {
required ServerName serverName = 1;
optional uint32 webuiPort = 2;
required ServerName server_name = 1;
optional uint32 webui_port = 2;
}
message GetServerInfoResponse {
required ServerInfo serverInfo = 1;
required ServerInfo server_info = 1;
}
service AdminService {
rpc getRegionInfo(GetRegionInfoRequest)
rpc GetRegionInfo(GetRegionInfoRequest)
returns(GetRegionInfoResponse);
rpc getStoreFile(GetStoreFileRequest)
rpc GetStoreFile(GetStoreFileRequest)
returns(GetStoreFileResponse);
rpc getOnlineRegion(GetOnlineRegionRequest)
rpc GetOnlineRegion(GetOnlineRegionRequest)
returns(GetOnlineRegionResponse);
rpc openRegion(OpenRegionRequest)
rpc OpenRegion(OpenRegionRequest)
returns(OpenRegionResponse);
rpc closeRegion(CloseRegionRequest)
rpc CloseRegion(CloseRegionRequest)
returns(CloseRegionResponse);
rpc flushRegion(FlushRegionRequest)
rpc FlushRegion(FlushRegionRequest)
returns(FlushRegionResponse);
rpc splitRegion(SplitRegionRequest)
rpc SplitRegion(SplitRegionRequest)
returns(SplitRegionResponse);
rpc compactRegion(CompactRegionRequest)
rpc CompactRegion(CompactRegionRequest)
returns(CompactRegionResponse);
rpc mergeRegions(MergeRegionsRequest)
rpc MergeRegions(MergeRegionsRequest)
returns(MergeRegionsResponse);
rpc replicateWALEntry(ReplicateWALEntryRequest)
rpc ReplicateWALEntry(ReplicateWALEntryRequest)
returns(ReplicateWALEntryResponse);
rpc replay(MultiRequest)
rpc Replay(MultiRequest)
returns(MultiResponse);
rpc rollWALWriter(RollWALWriterRequest)
rpc RollWALWriter(RollWALWriterRequest)
returns(RollWALWriterResponse);
rpc getServerInfo(GetServerInfoRequest)
rpc GetServerInfo(GetServerInfoRequest)
returns(GetServerInfoResponse);
rpc stopServer(StopServerRequest)
rpc StopServer(StopServerRequest)
returns(StopServerResponse);
}

View File

@ -31,32 +31,32 @@ message AggregateArgument {
* (3) any bytes required to construct the ColumnInterpreter object
* properly
*/
required string interpreterClassName = 1;
required string interpreter_class_name = 1;
required Scan scan = 2;
optional bytes interpreterSpecificBytes = 3;
optional bytes interpreter_specific_bytes = 3;
}
message AggregateResponse {
/**
* The AggregateService methods all have a response that either is a Pair
* or a simple object. When it is a Pair both firstPart and secondPart
* have defined values (and the secondPart is not present in the response
* or a simple object. When it is a Pair both first_part and second_part
* have defined values (and the second_part is not present in the response
* when the response is not a pair). Refer to the AggregateImplementation
* class for an overview of the AggregateResponse object constructions.
*/
repeated bytes firstPart = 1;
optional bytes secondPart = 2;
repeated bytes first_part = 1;
optional bytes second_part = 2;
}
/** Refer to the AggregateImplementation class for an overview of the
* AggregateService method implementations and their functionality.
*/
service AggregateService {
rpc getMax (AggregateArgument) returns (AggregateResponse);
rpc getMin (AggregateArgument) returns (AggregateResponse);
rpc getSum (AggregateArgument) returns (AggregateResponse);
rpc getRowNum (AggregateArgument) returns (AggregateResponse);
rpc getAvg (AggregateArgument) returns (AggregateResponse);
rpc getStd (AggregateArgument) returns (AggregateResponse);
rpc getMedian (AggregateArgument) returns (AggregateResponse);
rpc GetMax (AggregateArgument) returns (AggregateResponse);
rpc GetMin (AggregateArgument) returns (AggregateResponse);
rpc GetSum (AggregateArgument) returns (AggregateResponse);
rpc GetRowNum (AggregateArgument) returns (AggregateResponse);
rpc GetAvg (AggregateArgument) returns (AggregateResponse);
rpc GetStd (AggregateArgument) returns (AggregateResponse);
rpc GetMedian (AggregateArgument) returns (AggregateResponse);
}

View File

@ -24,7 +24,7 @@ option optimize_for = SPEED;
message AuthenticationKey {
required int32 id = 1;
required int64 expirationDate = 2;
required int64 expiration_date = 2;
required bytes key = 3;
}
@ -35,10 +35,10 @@ message TokenIdentifier {
}
required Kind kind = 1;
required bytes username = 2;
required int32 keyId = 3;
optional int64 issueDate = 4;
optional int64 expirationDate = 5;
optional int64 sequenceNumber = 6;
required int32 key_id = 3;
optional int64 issue_date = 4;
optional int64 expiration_date = 5;
optional int64 sequence_number = 6;
}
@ -67,15 +67,15 @@ message WhoAmIRequest {
message WhoAmIResponse {
optional string username = 1;
optional string authMethod = 2;
optional string auth_method = 2;
}
// RPC service
service AuthenticationService {
rpc getAuthenticationToken(TokenRequest)
rpc GetAuthenticationToken(TokenRequest)
returns (TokenResponse);
rpc whoami(WhoAmIRequest)
rpc WhoAmI(WhoAmIRequest)
returns (WhoAmIResponse);
}

View File

@ -46,7 +46,7 @@ message Cell {
optional bytes family = 2;
optional bytes qualifier = 3;
optional uint64 timestamp = 4;
optional CellType cellType = 5;
optional CellType cell_type = 5;
optional bytes value = 6;
}
@ -59,6 +59,6 @@ message KeyValue {
required bytes family = 2;
required bytes qualifier = 3;
optional uint64 timestamp = 4;
optional CellType keyType = 5;
optional CellType key_type = 5;
optional bytes value = 6;
}

View File

@ -45,11 +45,11 @@ message Get {
repeated Column column = 2;
repeated NameBytesPair attribute = 3;
optional Filter filter = 4;
optional TimeRange timeRange = 5;
optional uint32 maxVersions = 6 [default = 1];
optional bool cacheBlocks = 7 [default = true];
optional uint32 storeLimit = 8;
optional uint32 storeOffset = 9;
optional TimeRange time_range = 5;
optional uint32 max_versions = 6 [default = 1];
optional bool cache_blocks = 7 [default = true];
optional uint32 store_limit = 8;
optional uint32 store_offset = 9;
}
message Result {
@ -60,18 +60,18 @@ message Result {
// not part of this protobuf message; they are passed alongside
// and then this Message is just a placeholder with metadata.
// The count is needed to know how many to peel off the block of Cells as
// ours. NOTE: This is different from the pb managed cellCount of the
// ours. NOTE: This is different from the pb managed cell_count of the
// 'cell' field above which is non-null when the cells are pb'd.
optional int32 associatedCellCount = 2;
optional int32 associated_cell_count = 2;
}
/**
* The get request. Perform a single Get operation.
* Unless existenceOnly is specified, return all the requested data
* Unless existence_only is specified, return all the requested data
* for the row that matches exactly, or the one that immediately
* precedes it if closestRowBefore is specified.
* precedes it if closest_row_before is specified.
*
* If existenceOnly is set, only the existence will be returned.
* If existence_only is set, only the existence will be returned.
*/
message GetRequest {
required RegionSpecifier region = 1;
@ -79,12 +79,12 @@ message GetRequest {
// If the row to get doesn't exist, return the
// closest row before.
optional bool closestRowBefore = 3;
optional bool closest_row_before = 3;
// The result isn't asked for, just check for
// the existence. If closestRowBefore specified,
// the existence. If closest_row_before specified,
// this will be ignored
optional bool existenceOnly = 4;
optional bool existence_only = 4;
}
message MultiGetRequest {
@ -93,12 +93,12 @@ message MultiGetRequest {
// If the row to get doesn't exist, return the
// closest row before.
optional bool closestRowBefore = 3;
optional bool closest_row_before = 3;
// The result isn't asked for, just check for
// the existence. If closestRowBefore specified,
// the existence. If closest_row_before specified,
// this will be ignored
optional bool existenceOnly = 4;
optional bool existence_only = 4;
}
message GetResponse {
@ -125,7 +125,7 @@ message Condition {
required bytes row = 1;
required bytes family = 2;
required bytes qualifier = 3;
required CompareType compareType = 4;
required CompareType compare_type = 4;
required Comparator comparator = 5;
}
@ -139,22 +139,22 @@ message Condition {
*/
message MutationProto {
optional bytes row = 1;
optional MutationType mutateType = 2;
repeated ColumnValue columnValue = 3;
optional MutationType mutate_type = 2;
repeated ColumnValue column_value = 3;
optional uint64 timestamp = 4;
repeated NameBytesPair attribute = 5;
optional Durability durability = 6 [default = USE_DEFAULT];
// For some mutations, a result may be returned, in which case,
// time range can be specified for potential performance gain
optional TimeRange timeRange = 7;
optional TimeRange time_range = 7;
// The below count is set when the associated cells are NOT
// part of this protobuf message; they are passed alongside
// and then this Message is a placeholder with metadata. The
// count is needed to know how many to peel off the block of Cells as
// ours. NOTE: This is different from the pb managed cellCount of the
// ours. NOTE: This is different from the pb managed cell_count of the
// 'cell' field above which is non-null when the cells are pb'd.
optional int32 associatedCellCount = 8;
optional int32 associated_cell_count = 8;
enum Durability {
USE_DEFAULT = 0;
@ -180,13 +180,13 @@ message MutationProto {
message ColumnValue {
required bytes family = 1;
repeated QualifierValue qualifierValue = 2;
repeated QualifierValue qualifier_value = 2;
message QualifierValue {
optional bytes qualifier = 1;
optional bytes value = 2;
optional uint64 timestamp = 3;
optional DeleteType deleteType = 4;
optional DeleteType delete_type = 4;
}
}
}
@ -225,18 +225,18 @@ message MutateResponse {
message Scan {
repeated Column column = 1;
repeated NameBytesPair attribute = 2;
optional bytes startRow = 3;
optional bytes stopRow = 4;
optional bytes start_row = 3;
optional bytes stop_row = 4;
optional Filter filter = 5;
optional TimeRange timeRange = 6;
optional uint32 maxVersions = 7 [default = 1];
optional bool cacheBlocks = 8 [default = true];
optional uint32 batchSize = 9;
optional uint64 maxResultSize = 10;
optional uint32 storeLimit = 11;
optional uint32 storeOffset = 12;
optional bool loadColumnFamiliesOnDemand = 13; /* DO NOT add defaults to loadColumnFamiliesOnDemand. */
optional uint32 cachingCount = 14;
optional TimeRange time_range = 6;
optional uint32 max_versions = 7 [default = 1];
optional bool cache_blocks = 8 [default = true];
optional uint32 batch_size = 9;
optional uint64 max_result_size = 10;
optional uint32 store_limit = 11;
optional uint32 store_offset = 12;
optional bool load_column_families_on_demand = 13; /* DO NOT add defaults to load_column_families_on_demand. */
optional uint32 caching_count = 14;
optional bool prefetching = 15;
}
@ -254,25 +254,25 @@ message Scan {
message ScanRequest {
optional RegionSpecifier region = 1;
optional Scan scan = 2;
optional uint64 scannerId = 3;
optional uint32 numberOfRows = 4;
optional bool closeScanner = 5;
optional uint64 nextCallSeq = 6;
optional uint64 scanner_id = 3;
optional uint32 number_of_rows = 4;
optional bool close_scanner = 5;
optional uint64 next_call_seq = 6;
}
/**
* The scan response. If there are no more results, moreResults will
* The scan response. If there are no more results, more_results will
* be false. If it is not specified, it means there are more.
*/
message ScanResponse {
optional ResultCellMeta resultCellMeta = 1;
optional uint64 scannerId = 2;
optional bool moreResults = 3;
optional ResultCellMeta result_cell_meta = 1;
optional uint64 scanner_id = 2;
optional bool more_results = 3;
optional uint32 ttl = 4;
}
message ResultCellMeta {
repeated uint32 cellsLength = 1;
repeated uint32 cells_length = 1;
}
/**
@ -281,8 +281,8 @@ message ResultCellMeta {
*/
message BulkLoadHFileRequest {
required RegionSpecifier region = 1;
repeated FamilyPath familyPath = 2;
optional bool assignSeqNum = 3;
repeated FamilyPath family_path = 2;
optional bool assign_seq_num = 3;
message FamilyPath {
required bytes family = 1;
@ -296,8 +296,8 @@ message BulkLoadHFileResponse {
message CoprocessorServiceCall {
required bytes row = 1;
required string serviceName = 2;
required string methodName = 3;
required string service_name = 2;
required string method_name = 3;
required bytes request = 4;
}
@ -352,24 +352,24 @@ message MultiResponse {
service ClientService {
rpc get(GetRequest)
rpc Get(GetRequest)
returns(GetResponse);
rpc multiGet(MultiGetRequest)
rpc MultiGet(MultiGetRequest)
returns(MultiGetResponse);
rpc mutate(MutateRequest)
rpc Mutate(MutateRequest)
returns(MutateResponse);
rpc scan(ScanRequest)
rpc Scan(ScanRequest)
returns(ScanResponse);
rpc bulkLoadHFile(BulkLoadHFileRequest)
rpc BulkLoadHFile(BulkLoadHFileRequest)
returns(BulkLoadHFileResponse);
rpc execService(CoprocessorServiceRequest)
rpc ExecService(CoprocessorServiceRequest)
returns(CoprocessorServiceResponse);
rpc multi(MultiRequest)
rpc Multi(MultiRequest)
returns(MultiResponse);
}

View File

@ -29,5 +29,5 @@ option optimize_for = SPEED;
*/
message ClusterId {
// This is the cluster id, a uuid as a String
required string clusterId = 1;
required string cluster_id = 1;
}

View File

@ -28,7 +28,7 @@ import "ClusterId.proto";
import "FS.proto";
message RegionState {
required RegionInfo regionInfo = 1;
required RegionInfo region_info = 1;
required State state = 2;
optional uint64 stamp = 3;
enum State {
@ -50,12 +50,12 @@ message RegionState {
message RegionInTransition {
required RegionSpecifier spec = 1;
required RegionState regionState = 2;
required RegionState region_state = 2;
}
message RegionLoad {
/** the region specifier */
required RegionSpecifier regionSpecifier = 1;
required RegionSpecifier region_specifier = 1;
/** the number of stores for the region */
optional uint32 stores = 2;
@ -64,65 +64,65 @@ message RegionLoad {
optional uint32 storefiles = 3;
/** the total size of the store files for the region, uncompressed, in MB */
optional uint32 storeUncompressedSizeMB = 4;
optional uint32 store_uncompressed_size_MB = 4;
/** the current total size of the store files for the region, in MB */
optional uint32 storefileSizeMB = 5;
optional uint32 storefile_size_MB = 5;
/** the current size of the memstore for the region, in MB */
optional uint32 memstoreSizeMB = 6;
optional uint32 memstore_size_MB = 6;
/**
* The current total size of root-level store file indexes for the region,
* in MB. The same as {@link #rootIndexSizeKB} but in MB.
*/
optional uint32 storefileIndexSizeMB = 7;
optional uint32 storefile_index_size_MB = 7;
/** the current total read requests made to region */
optional uint64 readRequestsCount = 8;
optional uint64 read_requests_count = 8;
/** the current total write requests made to region */
optional uint64 writeRequestsCount = 9;
optional uint64 write_requests_count = 9;
/** the total compacting key values in currently running compaction */
optional uint64 totalCompactingKVs = 10;
optional uint64 total_compacting_KVs = 10;
/** the completed count of key values in currently running compaction */
optional uint64 currentCompactedKVs = 11;
optional uint64 current_compacted_KVs = 11;
/** The current total size of root-level indexes for the region, in KB. */
optional uint32 rootIndexSizeKB = 12;
optional uint32 root_index_size_KB = 12;
/** The total size of all index blocks, not just the root level, in KB. */
optional uint32 totalStaticIndexSizeKB = 13;
optional uint32 total_static_index_size_KB = 13;
/**
* The total size of all Bloom filter blocks, not just loaded into the
* block cache, in KB.
*/
optional uint32 totalStaticBloomSizeKB = 14;
optional uint32 total_static_bloom_size_KB = 14;
/** the most recent sequence Id from cache flush */
optional uint64 completeSequenceId = 15;
optional uint64 complete_sequence_id = 15;
}
/* Server-level protobufs */
message ServerLoad {
/** Number of requests since last report. */
optional uint32 numberOfRequests = 1;
optional uint32 number_of_requests = 1;
/** Total Number of requests from the start of the region server. */
optional uint32 totalNumberOfRequests = 2;
optional uint32 total_number_of_requests = 2;
/** the amount of used heap, in MB. */
optional uint32 usedHeapMB = 3;
optional uint32 used_heap_MB = 3;
/** the maximum allowable size of the heap, in MB. */
optional uint32 maxHeapMB = 4;
optional uint32 max_heap_MB = 4;
/** Information on the load of individual regions. */
repeated RegionLoad regionLoads = 5;
repeated RegionLoad region_loads = 5;
/**
* Regionserver-level coprocessors, e.g., WALObserver implementations.
@ -132,39 +132,38 @@ message ServerLoad {
repeated Coprocessor coprocessors = 6;
/**
* Time when incremental (non-total) counts began being calculated (e.g. numberOfRequests)
* Time when incremental (non-total) counts began being calculated (e.g. number_of_requests)
* time is measured as the difference, measured in milliseconds, between the current time
* and midnight, January 1, 1970 UTC.
*/
optional uint64 reportStartTime = 7;
optional uint64 report_start_time = 7;
/**
* Time when report was generated.
* time is measured as the difference, measured in milliseconds, between the current time
* and midnight, January 1, 1970 UTC.
*/
optional uint64 reportEndTime = 8;
optional uint64 report_end_time = 8;
/**
* The port number that this region server is hosing an info server on.
*/
optional uint32 infoServerPort = 9;
optional uint32 info_server_port = 9;
}
message LiveServerInfo {
required ServerName server = 1;
required ServerLoad serverLoad = 2;
required ServerLoad server_load = 2;
}
message ClusterStatus {
optional HBaseVersionFileContent hbaseVersion = 1;
repeated LiveServerInfo liveServers = 2;
repeated ServerName deadServers = 3;
repeated RegionInTransition regionsInTransition = 4;
optional ClusterId clusterId = 5;
repeated Coprocessor masterCoprocessors = 6;
optional HBaseVersionFileContent hbase_version = 1;
repeated LiveServerInfo live_servers = 2;
repeated ServerName dead_servers = 3;
repeated RegionInTransition regions_in_transition = 4;
optional ClusterId cluster_id = 5;
repeated Coprocessor master_coprocessors = 6;
optional ServerName master = 7;
repeated ServerName backupMasters = 8;
optional bool balancerOn = 9;
repeated ServerName backup_masters = 8;
optional bool balancer_on = 9;
}

View File

@ -28,7 +28,7 @@ option optimize_for = SPEED;
message Comparator {
required string name = 1;
optional bytes serializedComparator = 2;
optional bytes serialized_comparator = 2;
}
message ByteArrayComparable {
@ -45,7 +45,7 @@ message BinaryPrefixComparator {
message BitComparator {
required ByteArrayComparable comparable = 1;
required BitwiseOp bitwiseOp = 2;
required BitwiseOp bitwise_op = 2;
enum BitwiseOp {
AND = 1;
@ -59,7 +59,7 @@ message NullComparator {
message RegexStringComparator {
required string pattern = 1;
required int32 patternFlags = 2;
required int32 pattern_flags = 2;
required string charset = 3;
}

View File

@ -28,10 +28,10 @@ option optimize_for = SPEED;
* so we can serialize exceptions.
*/
message StackTraceElementMessage {
optional string declaringClass = 1;
optional string methodName = 2;
optional string fileName = 3;
optional int32 lineNumber = 4;
optional string declaring_class = 1;
optional string method_name = 2;
optional string file_name = 3;
optional int32 line_number = 4;
}
/**
@ -41,9 +41,9 @@ message StackTraceElementMessage {
* (which should be another protobuffed class).
*/
message GenericExceptionMessage {
optional string className = 1;
optional string class_name = 1;
optional string message = 2;
optional bytes errorInfo = 3;
optional bytes error_info = 3;
repeated StackTraceElementMessage trace = 4;
}
@ -53,6 +53,5 @@ message GenericExceptionMessage {
*/
message ForeignExceptionMessage {
optional string source = 1;
optional GenericExceptionMessage genericException = 2;
optional GenericExceptionMessage generic_exception = 2;
}

View File

@ -29,7 +29,7 @@ import "Comparator.proto";
message Filter {
required string name = 1;
optional bytes serializedFilter = 2;
optional bytes serialized_filter = 2;
}
message ColumnCountGetFilter {
@ -39,7 +39,7 @@ message ColumnCountGetFilter {
message ColumnPaginationFilter {
required int32 limit = 1;
optional int32 offset = 2;
optional bytes columnOffset = 3;
optional bytes column_offset = 3;
}
message ColumnPrefixFilter {
@ -47,26 +47,26 @@ message ColumnPrefixFilter {
}
message ColumnRangeFilter {
optional bytes minColumn = 1;
optional bool minColumnInclusive = 2;
optional bytes maxColumn = 3;
optional bool maxColumnInclusive = 4;
optional bytes min_column = 1;
optional bool min_column_inclusive = 2;
optional bytes max_column = 3;
optional bool max_column_inclusive = 4;
}
message CompareFilter {
required CompareType compareOp = 1;
required CompareType compare_op = 1;
optional Comparator comparator = 2;
}
message DependentColumnFilter {
required CompareFilter compareFilter = 1;
optional bytes columnFamily = 2;
optional bytes columnQualifier = 3;
optional bool dropDependentColumn = 4;
required CompareFilter compare_filter = 1;
optional bytes column_family = 2;
optional bytes column_qualifier = 3;
optional bool drop_dependent_column = 4;
}
message FamilyFilter {
required CompareFilter compareFilter = 1;
required CompareFilter compare_filter = 1;
}
message FilterList {
@ -91,23 +91,23 @@ message FirstKeyValueMatchingQualifiersFilter {
}
message FuzzyRowFilter {
repeated BytesBytesPair fuzzyKeysData = 1;
repeated BytesBytesPair fuzzy_keys_data = 1;
}
message InclusiveStopFilter {
optional bytes stopRowKey = 1;
optional bytes stop_row_key = 1;
}
message KeyOnlyFilter {
required bool lenAsVal = 1;
required bool len_as_val = 1;
}
message MultipleColumnPrefixFilter {
repeated bytes sortedPrefixes = 1;
repeated bytes sorted_prefixes = 1;
}
message PageFilter {
required int64 pageSize = 1;
required int64 page_size = 1;
}
message PrefixFilter {
@ -115,7 +115,7 @@ message PrefixFilter {
}
message QualifierFilter {
required CompareFilter compareFilter = 1;
required CompareFilter compare_filter = 1;
}
message RandomRowFilter {
@ -123,20 +123,20 @@ message RandomRowFilter {
}
message RowFilter {
required CompareFilter compareFilter = 1;
required CompareFilter compare_filter = 1;
}
message SingleColumnValueExcludeFilter {
required SingleColumnValueFilter singleColumnValueFilter = 1;
required SingleColumnValueFilter single_column_value_filter = 1;
}
message SingleColumnValueFilter {
optional bytes columnFamily = 1;
optional bytes columnQualifier = 2;
required CompareType compareOp = 3;
optional bytes column_family = 1;
optional bytes column_qualifier = 2;
required CompareType compare_op = 3;
required Comparator comparator = 4;
optional bool filterIfMissing = 5;
optional bool latestVersionOnly = 6;
optional bool filter_if_missing = 5;
optional bool latest_version_only = 6;
}
message SkipFilter {
@ -148,7 +148,7 @@ message TimestampsFilter {
}
message ValueFilter {
required CompareFilter compareFilter = 1;
required CompareFilter compare_filter = 1;
}
message WhileMatchFilter {

View File

@ -25,21 +25,21 @@ import "hbase.proto";
// Map of name/values
message FileInfoProto {
repeated BytesBytesPair mapEntry = 1;
repeated BytesBytesPair map_entry = 1;
}
// HFile file trailer
message FileTrailerProto {
optional uint64 fileInfoOffset = 1;
optional uint64 loadOnOpenDataOffset = 2;
optional uint64 uncompressedDataIndexSize = 3;
optional uint64 totalUncompressedBytes = 4;
optional uint32 dataIndexCount = 5;
optional uint32 metaIndexCount = 6;
optional uint64 entryCount = 7;
optional uint32 numDataIndexLevels = 8;
optional uint64 firstDataBlockOffset = 9;
optional uint64 lastDataBlockOffset = 10;
optional string comparatorClassName = 11;
optional uint32 compressionCodec = 12;
optional uint64 file_info_offset = 1;
optional uint64 load_on_open_data_offset = 2;
optional uint64 uncompressed_data_index_size = 3;
optional uint64 total_uncompressed_bytes = 4;
optional uint32 data_index_count = 5;
optional uint32 meta_index_count = 6;
optional uint64 entry_count = 7;
optional uint32 num_data_index_levels = 8;
optional uint64 first_data_block_offset = 9;
optional uint64 last_data_block_offset = 10;
optional string comparator_class_name = 11;
optional uint32 compression_codec = 12;
}

View File

@ -24,5 +24,5 @@ option java_generate_equals_and_hash = true;
option optimize_for = SPEED;
message LoadBalancerState {
optional bool balancerOn = 1;
optional bool balancer_on = 1;
}

View File

@ -28,11 +28,11 @@ message IsMasterRunningRequest {
}
message IsMasterRunningResponse {
required bool isMasterRunning = 1;
required bool is_master_running = 1;
}
service MasterService {
/** return true if master is available */
rpc isMasterRunning(IsMasterRunningRequest)
rpc IsMasterRunning(IsMasterRunningRequest)
returns(IsMasterRunningResponse);
}

View File

@ -32,24 +32,24 @@ import "Client.proto";
/* Column-level protobufs */
message AddColumnRequest {
required bytes tableName = 1;
required ColumnFamilySchema columnFamilies = 2;
required bytes table_name = 1;
required ColumnFamilySchema column_families = 2;
}
message AddColumnResponse {
}
message DeleteColumnRequest {
required bytes tableName = 1;
required bytes columnName = 2;
required bytes table_name = 1;
required bytes column_name = 2;
}
message DeleteColumnResponse {
}
message ModifyColumnRequest {
required bytes tableName = 1;
required ColumnFamilySchema columnFamilies = 2;
required bytes table_name = 1;
required ColumnFamilySchema column_families = 2;
}
message ModifyColumnResponse {
@ -59,7 +59,7 @@ message ModifyColumnResponse {
message MoveRegionRequest {
required RegionSpecifier region = 1;
optional ServerName destServerName = 2;
optional ServerName dest_server_name = 2;
}
message MoveRegionResponse {
@ -69,8 +69,8 @@ message MoveRegionResponse {
* Dispatch merging the specified regions.
*/
message DispatchMergingRegionsRequest {
required RegionSpecifier regionA = 1;
required RegionSpecifier regionB = 2;
required RegionSpecifier region_a = 1;
required RegionSpecifier region_b = 2;
optional bool forcible = 3 [default = false];
}
@ -102,37 +102,37 @@ message OfflineRegionResponse {
/* Table-level protobufs */
message CreateTableRequest {
required TableSchema tableSchema = 1;
repeated bytes splitKeys = 2;
required TableSchema table_schema = 1;
repeated bytes split_keys = 2;
}
message CreateTableResponse {
}
message DeleteTableRequest {
required bytes tableName = 1;
required bytes table_name = 1;
}
message DeleteTableResponse {
}
message EnableTableRequest {
required bytes tableName = 1;
required bytes table_name = 1;
}
message EnableTableResponse {
}
message DisableTableRequest {
required bytes tableName = 1;
required bytes table_name = 1;
}
message DisableTableResponse {
}
message ModifyTableRequest {
required bytes tableName = 1;
required TableSchema tableSchema = 2;
required bytes table_name = 1;
required TableSchema table_schema = 2;
}
message ModifyTableResponse {
@ -157,7 +157,7 @@ message BalanceRequest {
}
message BalanceResponse {
required bool balancerRan = 1;
required bool balancer_ran = 1;
}
message SetBalancerRunningRequest {
@ -166,14 +166,14 @@ message SetBalancerRunningRequest {
}
message SetBalancerRunningResponse {
optional bool prevBalanceValue = 1;
optional bool prev_balance_value = 1;
}
message CatalogScanRequest {
}
message CatalogScanResponse {
optional int32 scanResult = 1;
optional int32 scan_result = 1;
}
message EnableCatalogJanitorRequest {
@ -181,7 +181,7 @@ message EnableCatalogJanitorRequest {
}
message EnableCatalogJanitorResponse {
optional bool prevValue = 1;
optional bool prev_value = 1;
}
message IsCatalogJanitorEnabledRequest {
@ -196,7 +196,7 @@ message TakeSnapshotRequest{
}
message TakeSnapshotResponse{
required int64 expectedTimeout = 1;
required int64 expected_timeout = 1;
}
message ListSnapshotRequest{
@ -242,36 +242,36 @@ message IsRestoreSnapshotDoneResponse {
service MasterAdminService {
/** Adds a column to the specified table. */
rpc addColumn(AddColumnRequest)
rpc AddColumn(AddColumnRequest)
returns(AddColumnResponse);
/** Deletes a column from the specified table. Table must be disabled. */
rpc deleteColumn(DeleteColumnRequest)
rpc DeleteColumn(DeleteColumnRequest)
returns(DeleteColumnResponse);
/** Modifies an existing column on the specified table. */
rpc modifyColumn(ModifyColumnRequest)
rpc ModifyColumn(ModifyColumnRequest)
returns(ModifyColumnResponse);
/** Move the region region to the destination server. */
rpc moveRegion(MoveRegionRequest)
rpc MoveRegion(MoveRegionRequest)
returns(MoveRegionResponse);
/** Master dispatch merging the regions */
rpc dispatchMergingRegions(DispatchMergingRegionsRequest)
rpc DispatchMergingRegions(DispatchMergingRegionsRequest)
returns(DispatchMergingRegionsResponse);
/** Assign a region to a server chosen at random. */
rpc assignRegion(AssignRegionRequest)
rpc AssignRegion(AssignRegionRequest)
returns(AssignRegionResponse);
/**
* Unassign a region from current hosting regionserver. Region will then be
* assigned to a regionserver chosen at random. Region could be reassigned
* back to the same server. Use moveRegion if you want
* back to the same server. Use MoveRegion if you want
* to control the region movement.
*/
rpc unassignRegion(UnassignRegionRequest)
rpc UnassignRegion(UnassignRegionRequest)
returns(UnassignRegionResponse);
/**
@ -280,35 +280,35 @@ service MasterAdminService {
* automatically reassign the region as in unassign. This is a special
* method, and should only be used by experts or hbck.
*/
rpc offlineRegion(OfflineRegionRequest)
rpc OfflineRegion(OfflineRegionRequest)
returns(OfflineRegionResponse);
/** Deletes a table */
rpc deleteTable(DeleteTableRequest)
rpc DeleteTable(DeleteTableRequest)
returns(DeleteTableResponse);
/** Puts the table on-line (only needed if table has been previously taken offline) */
rpc enableTable(EnableTableRequest)
rpc EnableTable(EnableTableRequest)
returns(EnableTableResponse);
/** Take table offline */
rpc disableTable(DisableTableRequest)
rpc DisableTable(DisableTableRequest)
returns(DisableTableResponse);
/** Modify a table's metadata */
rpc modifyTable(ModifyTableRequest)
rpc ModifyTable(ModifyTableRequest)
returns(ModifyTableResponse);
/** Creates a new table asynchronously */
rpc createTable(CreateTableRequest)
rpc CreateTable(CreateTableRequest)
returns(CreateTableResponse);
/** Shutdown an HBase cluster. */
rpc shutdown(ShutdownRequest)
rpc Shutdown(ShutdownRequest)
returns(ShutdownResponse);
/** Stop HBase Master only. Does not shutdown the cluster. */
rpc stopMaster(StopMasterRequest)
rpc StopMaster(StopMasterRequest)
returns(StopMasterResponse);
/**
@ -316,72 +316,72 @@ service MasterAdminService {
* go ahead and do the reassignments. Can NOT run for various reasons.
* Check logs.
*/
rpc balance(BalanceRequest)
rpc Balance(BalanceRequest)
returns(BalanceResponse);
/**
* Turn the load balancer on or off.
* If synchronous is true, it waits until current balance() call, if outstanding, to return.
*/
rpc setBalancerRunning(SetBalancerRunningRequest)
rpc SetBalancerRunning(SetBalancerRunningRequest)
returns(SetBalancerRunningResponse);
/** Get a run of the catalog janitor */
rpc runCatalogScan(CatalogScanRequest)
rpc RunCatalogScan(CatalogScanRequest)
returns(CatalogScanResponse);
/**
* Enable the catalog janitor on or off.
*/
rpc enableCatalogJanitor(EnableCatalogJanitorRequest)
rpc EnableCatalogJanitor(EnableCatalogJanitorRequest)
returns(EnableCatalogJanitorResponse);
/**
* Query whether the catalog janitor is enabled.
*/
rpc isCatalogJanitorEnabled(IsCatalogJanitorEnabledRequest)
rpc IsCatalogJanitorEnabled(IsCatalogJanitorEnabledRequest)
returns(IsCatalogJanitorEnabledResponse);
/**
* Call a master coprocessor endpoint
*/
rpc execMasterService(CoprocessorServiceRequest)
rpc ExecMasterService(CoprocessorServiceRequest)
returns(CoprocessorServiceResponse);
/**
* Create a snapshot for the given table.
* @param snapshot description of the snapshot to take
*/
rpc snapshot(TakeSnapshotRequest) returns(TakeSnapshotResponse);
rpc Snapshot(TakeSnapshotRequest) returns(TakeSnapshotResponse);
/**
* List completed snapshots.
* Returns a list of snapshot descriptors for completed snapshots
*/
rpc getCompletedSnapshots(ListSnapshotRequest) returns(ListSnapshotResponse);
rpc GetCompletedSnapshots(ListSnapshotRequest) returns(ListSnapshotResponse);
/**
* Delete an existing snapshot. This method can also be used to clean up an aborted snapshot.
* @param snapshotName snapshot to delete
*/
rpc deleteSnapshot(DeleteSnapshotRequest) returns(DeleteSnapshotResponse);
rpc DeleteSnapshot(DeleteSnapshotRequest) returns(DeleteSnapshotResponse);
/**
* Determine if the snapshot is done yet.
*/
rpc isSnapshotDone(IsSnapshotDoneRequest) returns(IsSnapshotDoneResponse);
rpc IsSnapshotDone(IsSnapshotDoneRequest) returns(IsSnapshotDoneResponse);
/**
* Restore a snapshot
* @param snapshot description of the snapshot to restore
*/
rpc restoreSnapshot(RestoreSnapshotRequest) returns(RestoreSnapshotResponse);
rpc RestoreSnapshot(RestoreSnapshotRequest) returns(RestoreSnapshotResponse);
/**
* Determine if the snapshot restore is done yet.
*/
rpc isRestoreSnapshotDone(IsRestoreSnapshotDoneRequest) returns(IsRestoreSnapshotDoneResponse);
rpc IsRestoreSnapshotDone(IsRestoreSnapshotDoneRequest) returns(IsRestoreSnapshotDoneResponse);
/** return true if master is available */
rpc isMasterRunning(IsMasterRunningRequest) returns(IsMasterRunningResponse);
rpc IsMasterRunning(IsMasterRunningRequest) returns(IsMasterRunningResponse);
}

View File

@ -29,42 +29,42 @@ import "hbase.proto";
import "ClusterStatus.proto";
message GetSchemaAlterStatusRequest {
required bytes tableName = 1;
required bytes table_name = 1;
}
message GetSchemaAlterStatusResponse {
optional uint32 yetToUpdateRegions = 1;
optional uint32 totalRegions = 2;
optional uint32 yet_to_update_regions = 1;
optional uint32 total_regions = 2;
}
message GetTableDescriptorsRequest {
repeated string tableNames = 1;
repeated string table_names = 1;
}
message GetTableDescriptorsResponse {
repeated TableSchema tableSchema = 1;
repeated TableSchema table_schema = 1;
}
message GetClusterStatusRequest {
}
message GetClusterStatusResponse {
required ClusterStatus clusterStatus = 1;
required ClusterStatus cluster_status = 1;
}
service MasterMonitorService {
/** Used by the client to get the number of regions that have received the updated schema */
rpc getSchemaAlterStatus(GetSchemaAlterStatusRequest)
rpc GetSchemaAlterStatus(GetSchemaAlterStatusRequest)
returns(GetSchemaAlterStatusResponse);
/** Get list of TableDescriptors for requested tables. */
rpc getTableDescriptors(GetTableDescriptorsRequest)
rpc GetTableDescriptors(GetTableDescriptorsRequest)
returns(GetTableDescriptorsResponse);
/** Return cluster status. */
rpc getClusterStatus(GetClusterStatusRequest)
rpc GetClusterStatus(GetClusterStatusRequest)
returns(GetClusterStatusResponse);
/** return true if master is available */
rpc isMasterRunning(IsMasterRunningRequest) returns(IsMasterRunningResponse);
rpc IsMasterRunning(IsMasterRunningRequest) returns(IsMasterRunningResponse);
}

View File

@ -23,13 +23,13 @@ option java_generic_services = true;
option optimize_for = SPEED;
message MultiMutateRequest {
repeated MutationProto mutationRequest = 1;
repeated MutationProto mutation_request = 1;
}
message MultiMutateResponse {
}
service MultiRowMutationService {
rpc mutateRows(MultiMutateRequest)
rpc MutateRows(MultiMutateRequest)
returns(MultiMutateResponse);
}

View File

@ -72,20 +72,20 @@ option optimize_for = SPEED;
// User Information proto. Included in ConnectionHeader on connection setup
message UserInformation {
required string effectiveUser = 1;
optional string realUser = 2;
required string effective_user = 1;
optional string real_user = 2;
}
// This is sent on connection setup after the connection preamble is sent.
message ConnectionHeader {
optional UserInformation userInfo = 1;
optional string serviceName = 2;
optional UserInformation user_info = 1;
optional string service_name = 2;
// Cell block codec we will use sending over optional cell blocks. Server throws exception
// if cannot deal.
optional string cellBlockCodecClass = 3 [default = "org.apache.hadoop.hbase.codec.KeyValueCodec"];
optional string cell_block_codec_class = 3 [default = "org.apache.hadoop.hbase.codec.KeyValueCodec"];
// Compressor we will use if cell block is compressed. Server will throw exception if not supported.
// Class must implement hadoop's CompressionCodec Interface
optional string cellBlockCompressorClass = 4;
optional string cell_block_compressor_class = 4;
}
// Optional Cell block Message. Included in client RequestHeader
@ -98,34 +98,34 @@ message CellBlockMeta {
// the server side exception to the RPC client.
message ExceptionResponse {
// Class name of the exception thrown from the server
optional string exceptionClassName = 1;
optional string exception_class_name = 1;
// Exception stack trace from the server side
optional string stackTrace = 2;
optional string stack_trace = 2;
// Optional hostname. Filled in for some exceptions such as region moved
// where exception gives clue on where the region may have moved.
optional string hostname = 3;
optional int32 port = 4;
// Set if we are NOT to retry on receipt of this exception
optional bool doNotRetry = 5;
optional bool do_not_retry = 5;
}
// Header sent making a request.
message RequestHeader {
// Monotonically increasing callId to keep track of RPC requests and their response
optional uint32 callId = 1;
optional RPCTInfo traceInfo = 2;
optional string methodName = 3;
// Monotonically increasing call_id to keep track of RPC requests and their response
optional uint32 call_id = 1;
optional RPCTInfo trace_info = 2;
optional string method_name = 3;
// If true, then a pb Message param follows.
optional bool requestParam = 4;
optional bool request_param = 4;
// If present, then an encoded data block follows.
optional CellBlockMeta cellBlockMeta = 5;
optional CellBlockMeta cell_block_meta = 5;
// TODO: Have client specify priority
}
message ResponseHeader {
optional uint32 callId = 1;
optional uint32 call_id = 1;
// If present, then request threw an exception and no response message (else we presume one)
optional ExceptionResponse exception = 2;
// If present, then an encoded data block follows.
optional CellBlockMeta cellBlockMeta = 3;
optional CellBlockMeta cell_block_meta = 3;
}

View File

@ -32,10 +32,10 @@ message RegionServerStartupRequest {
required uint32 port = 1;
/** This servers' startcode */
required uint64 serverStartCode = 2;
required uint64 server_start_code = 2;
/** Current time of the region server in ms */
required uint64 serverCurrentTime = 3;
required uint64 server_current_time = 3;
}
message RegionServerStartupResponse {
@ -44,7 +44,7 @@ message RegionServerStartupResponse {
* hbase rootdir, the hostname to use creating the RegionServer ServerName,
* etc
*/
repeated NameStringPair mapEntries = 1;
repeated NameStringPair map_entries = 1;
}
message RegionServerReportRequest {
@ -62,7 +62,7 @@ message ReportRSFatalErrorRequest {
required ServerName server = 1;
/** informative text to expose in the master logs and UI */
required string errorMessage = 2;
required string error_message = 2;
}
message ReportRSFatalErrorResponse {
@ -70,33 +70,33 @@ message ReportRSFatalErrorResponse {
message GetLastFlushedSequenceIdRequest {
/** region name */
required bytes regionName = 1;
required bytes region_name = 1;
}
message GetLastFlushedSequenceIdResponse {
/** the last HLog sequence id flushed from MemStore to HFile for the region */
required uint64 lastFlushedSequenceId = 1;
required uint64 last_flushed_sequence_id = 1;
}
service RegionServerStatusService {
/** Called when a region server first starts. */
rpc regionServerStartup(RegionServerStartupRequest)
rpc RegionServerStartup(RegionServerStartupRequest)
returns(RegionServerStartupResponse);
/** Called to report the load the RegionServer is under. */
rpc regionServerReport(RegionServerReportRequest)
rpc RegionServerReport(RegionServerReportRequest)
returns(RegionServerReportResponse);
/**
* Called by a region server to report a fatal error that is causing it to
* abort.
*/
rpc reportRSFatalError(ReportRSFatalErrorRequest)
rpc ReportRSFatalError(ReportRSFatalErrorRequest)
returns(ReportRSFatalErrorResponse);
/** Called to get the sequence id of the last MemStore entry flushed to an
* HFile for a specified region. Used by the region server to speed up
* log splitting. */
rpc getLastFlushedSequenceId(GetLastFlushedSequenceIdRequest)
rpc GetLastFlushedSequenceId(GetLastFlushedSequenceIdRequest)
returns(GetLastFlushedSequenceIdResponse);
}

View File

@ -27,15 +27,15 @@ option java_generate_equals_and_hash = true;
option optimize_for = SPEED;
message RowProcessorRequest {
required string rowProcessorClassName = 1;
optional string rowProcessorInitializerMessageName = 2;
optional bytes rowProcessorInitializerMessage = 3;
required string row_processor_class_name = 1;
optional string row_processor_initializer_message_name = 2;
optional bytes row_processor_initializer_message = 3;
}
message RowProcessorResult {
required bytes rowProcessorResult = 1;
required bytes row_processor_result = 1;
}
service RowProcessorService {
rpc process (RowProcessorRequest) returns (RowProcessorResult);
rpc Process (RowProcessorRequest) returns (RowProcessorResult);
}

View File

@ -26,10 +26,10 @@ import 'hbase.proto';
import 'Client.proto';
message SecureBulkLoadHFilesRequest {
repeated BulkLoadHFileRequest.FamilyPath familyPath = 1;
optional bool assignSeqNum = 2;
required DelegationTokenProto fsToken = 3;
required string bulkToken = 4;
repeated BulkLoadHFileRequest.FamilyPath family_path = 1;
optional bool assign_seq_num = 2;
required DelegationTokenProto fs_token = 3;
required string bulk_token = 4;
}
message SecureBulkLoadHFilesResponse {
@ -44,15 +44,15 @@ message DelegationTokenProto {
}
message PrepareBulkLoadRequest {
required bytes tableName = 1;
required bytes table_name = 1;
}
message PrepareBulkLoadResponse {
required string bulkToken = 1;
required string bulk_token = 1;
}
message CleanupBulkLoadRequest {
required string bulkToken = 1;
required string bulk_token = 1;
}
@ -60,12 +60,12 @@ message CleanupBulkLoadResponse {
}
service SecureBulkLoadService {
rpc prepareBulkLoad(PrepareBulkLoadRequest)
rpc PrepareBulkLoad(PrepareBulkLoadRequest)
returns (PrepareBulkLoadResponse);
rpc secureBulkLoadHFiles(SecureBulkLoadHFilesRequest)
rpc SecureBulkLoadHFiles(SecureBulkLoadHFilesRequest)
returns (SecureBulkLoadHFilesResponse);
rpc cleanupBulkLoad(CleanupBulkLoadRequest)
rpc CleanupBulkLoad(CleanupBulkLoadRequest)
returns (CleanupBulkLoadResponse);
}

View File

@ -26,6 +26,6 @@ option optimize_for = SPEED;
//the id of the current span when this message was sent, so we know
//what span caused the new span we will create when this message is received.
message RPCTInfo {
optional int64 traceId = 1;
optional int64 parentId = 2;
optional int64 trace_id = 1;
optional int64 parent_id = 2;
}

View File

@ -24,21 +24,21 @@ option optimize_for = SPEED;
import "hbase.proto";
message WALHeader {
optional bool hasCompression = 1;
optional bool has_compression = 1;
}
// Protocol buffer version of HLogKey; see HLogKey comment, not really a key but WALEdit header for some KVs
message WALKey {
required bytes encodedRegionName = 1;
required bytes tableName = 2;
required uint64 logSequenceNumber = 3;
required uint64 writeTime = 4;
optional UUID clusterId = 5;
required bytes encoded_region_name = 1;
required bytes table_name = 2;
required uint64 log_sequence_number = 3;
required uint64 write_time = 4;
optional UUID cluster_id = 5;
repeated FamilyScope scopes = 6;
optional uint32 followingKvCount = 7;
optional uint32 following_kv_count = 7;
/*
optional CustomEntryType customEntryType = 8;
optional CustomEntryType custom_entry_type = 8;
enum CustomEntryType {
COMPACTION = 0;
@ -53,7 +53,7 @@ enum ScopeType {
message FamilyScope {
required bytes family = 1;
required ScopeType scopeType = 2;
required ScopeType scope_type = 2;
}
/**
@ -67,12 +67,12 @@ message FamilyScope {
* the * compaction should we fail the WAL write.
*/
message CompactionDescriptor {
required bytes tableName = 1; // TODO: WALKey already stores these, might remove
required bytes encodedRegionName = 2;
required bytes familyName = 3;
repeated string compactionInput = 4;
repeated string compactionOutput = 5;
required string storeHomeDir = 6;
required bytes table_name = 1; // TODO: WALKey already stores these, might remove
required bytes encoded_region_name = 2;
required bytes family_name = 3;
repeated string compaction_input = 4;
repeated string compaction_output = 5;
required string store_home_dir = 6;
}
/**

View File

@ -36,7 +36,7 @@ message MetaRegionServer {
// The major version of the rpc the server speaks. This is used so that
// clients connecting to the cluster can have prior knowledge of what version
// to send to a RegionServer. AsyncHBase will use this to detect versions.
optional uint32 rpcVersion = 2;
optional uint32 rpc_version = 2;
}
/**
@ -46,7 +46,7 @@ message Master {
// The ServerName of the current Master
required ServerName master = 1;
// Major RPC version so that clients can know what version the master can accept.
optional uint32 rpcVersion = 2;
optional uint32 rpc_version = 2;
}
/**
@ -54,8 +54,8 @@ message Master {
*/
message ClusterUp {
// If this znode is present, cluster is up. Currently
// the data is cluster startDate.
required string startDate = 1;
// the data is cluster start_date.
required string start_date = 1;
}
/**
@ -64,12 +64,12 @@ message ClusterUp {
*/
message RegionTransition {
// Code for EventType gotten by doing o.a.h.h.EventHandler.EventType.getCode()
required uint32 eventTypeCode = 1;
required uint32 event_type_code = 1;
// Full regionname in bytes
required bytes regionName = 2;
required uint64 createTime = 3;
required bytes region_name = 2;
required uint64 create_time = 3;
// The region server where the transition will happen or is happening
required ServerName serverName = 4;
required ServerName server_name = 4;
optional bytes payload = 5;
}
@ -86,7 +86,7 @@ message SplitLogTask {
ERR = 4;
}
required State state = 1;
required ServerName serverName = 2;
required ServerName server_name = 2;
}
/**
@ -110,7 +110,7 @@ message Table {
* Used by replication. Holds a replication peer key.
*/
message ReplicationPeer {
// clusterKey is the concatenation of the slave cluster's
// clusterkey is the concatenation of the slave cluster's
// hbase.zookeeper.quorum:hbase.zookeeper.property.clientPort:zookeeper.znode.parent
required string clusterkey = 1;
}
@ -137,27 +137,27 @@ message ReplicationHLogPosition {
* Used by replication. Used to lock a region server during failover.
*/
message ReplicationLock {
required string lockOwner = 1;
required string lock_owner = 1;
}
/**
* Metadata associated with a table lock in zookeeper
*/
message TableLock {
optional bytes tableName = 1;
optional ServerName lockOwner = 2;
optional int64 threadId = 3;
optional bool isShared = 4;
optional bytes table_name = 1;
optional ServerName lock_owner = 2;
optional int64 thread_id = 3;
optional bool is_shared = 4;
optional string purpose = 5;
optional int64 createTime = 6;
optional int64 create_time = 6;
}
/**
* sequence Id of a store
*/
message StoreSequenceId {
required bytes familyName = 1;
required uint64 sequenceId = 2;
required bytes family_name = 1;
required uint64 sequence_id = 2;
}
/**
@ -165,6 +165,6 @@ message StoreSequenceId {
* list sequence ids of the region's stores
*/
message RegionStoreSequenceIds {
required uint64 lastFlushedSequenceId = 1;
repeated StoreSequenceId storeSequenceId = 2;
required uint64 last_flushed_sequence_id = 1;
repeated StoreSequenceId store_sequence_id = 2;
}

View File

@ -32,7 +32,7 @@ import "Cell.proto";
message TableSchema {
optional bytes name = 1;
repeated BytesBytesPair attributes = 2;
repeated ColumnFamilySchema columnFamilies = 3;
repeated ColumnFamilySchema column_families = 3;
repeated NameStringPair configuration = 4;
}
@ -50,10 +50,10 @@ message ColumnFamilySchema {
* Protocol buffer version of HRegionInfo.
*/
message RegionInfo {
required uint64 regionId = 1;
required bytes tableName = 2;
optional bytes startKey = 3;
optional bytes endKey = 4;
required uint64 region_id = 1;
required bytes table_name = 2;
optional bytes start_key = 3;
optional bytes end_key = 4;
optional bool offline = 5;
optional bool split = 6;
optional bool recovering = 7;
@ -63,7 +63,7 @@ message RegionInfo {
* Protocol buffer for favored nodes
*/
message FavoredNodes {
repeated ServerName favoredNode = 1;
repeated ServerName favored_node = 1;
}
/**
@ -111,9 +111,9 @@ enum CompareType {
* Protocol buffer version of ServerName
*/
message ServerName {
required string hostName = 1;
required string host_name = 1;
optional uint32 port = 2;
optional uint64 startCode = 3;
optional uint64 start_code = 3;
}
// Comment data structures
@ -148,7 +148,7 @@ message NameInt64Pair {
message SnapshotDescription {
required string name = 1;
optional string table = 2; // not needed for delete, but checked for in taking snapshot
optional int64 creationTime = 3 [default = 0];
optional int64 creation_time = 3 [default = 0];
enum Type {
DISABLED = 0;
FLUSH = 1;
@ -161,15 +161,15 @@ message EmptyMsg {
}
message LongMsg {
required int64 longMsg = 1;
required int64 long_msg = 1;
}
message BigDecimalMsg {
required bytes bigdecimalMsg = 1;
required bytes bigdecimal_msg = 1;
}
message UUID {
required uint64 leastSigBits = 1;
required uint64 mostSigBits = 2;
required uint64 least_sig_bits = 1;
required uint64 most_sig_bits = 2;
}

View File

@ -133,7 +133,7 @@ public class TokenProvider implements AuthenticationProtos.AuthenticationService
}
@Override
public void whoami(RpcController controller, AuthenticationProtos.WhoAmIRequest request,
public void whoAmI(RpcController controller, AuthenticationProtos.WhoAmIRequest request,
RpcCallback<AuthenticationProtos.WhoAmIResponse> done) {
User requestUser = RequestContext.getRequestUser();
AuthenticationProtos.WhoAmIResponse.Builder response =

View File

@ -274,15 +274,15 @@ public class TestTokenAuthentication {
}
@Override
public AuthenticationProtos.WhoAmIResponse whoami(
public AuthenticationProtos.WhoAmIResponse whoAmI(
RpcController controller, AuthenticationProtos.WhoAmIRequest request)
throws ServiceException {
LOG.debug("whoami() request from "+RequestContext.getRequestUserName());
LOG.debug("whoAmI() request from "+RequestContext.getRequestUserName());
// ignore passed in controller -- it's always null
ServerRpcController serverController = new ServerRpcController();
BlockingRpcCallback<AuthenticationProtos.WhoAmIResponse> callback =
new BlockingRpcCallback<AuthenticationProtos.WhoAmIResponse>();
whoami(serverController, request, callback);
whoAmI(serverController, request, callback);
try {
serverController.checkFailed();
return callback.get();
@ -383,7 +383,7 @@ public class TestTokenAuthentication {
AuthenticationProtos.AuthenticationService.BlockingInterface stub =
AuthenticationProtos.AuthenticationService.newBlockingStub(channel);
AuthenticationProtos.WhoAmIResponse response =
stub.whoami(null, AuthenticationProtos.WhoAmIRequest.getDefaultInstance());
stub.whoAmI(null, AuthenticationProtos.WhoAmIRequest.getDefaultInstance());
String myname = response.getUsername();
assertEquals("testuser", myname);
String authMethod = response.getAuthMethod();