HBASE-8915 protobuf message style
git-svn-id: https://svn.apache.org/repos/asf/hbase/trunk@1501834 13f79535-47bb-0310-9956-ffa450edef68
This commit is contained in:
parent
8bf99de619
commit
457c928d35
|
@ -6422,11 +6422,11 @@ public final class AccessControlProtos {
|
|||
"Permission\":\n\027CheckPermissionsRequest\022\037\n" +
|
||||
"\npermission\030\001 \003(\0132\013.Permission\"\032\n\030CheckP" +
|
||||
"ermissionsResponse2\373\001\n\024AccessControlServ" +
|
||||
"ice\022&\n\005grant\022\r.GrantRequest\032\016.GrantRespo",
|
||||
"nse\022)\n\006revoke\022\016.RevokeRequest\032\017.RevokeRe" +
|
||||
"sponse\022G\n\022getUserPermissions\022\027.UserPermi" +
|
||||
"ice\022&\n\005Grant\022\r.GrantRequest\032\016.GrantRespo",
|
||||
"nse\022)\n\006Revoke\022\016.RevokeRequest\032\017.RevokeRe" +
|
||||
"sponse\022G\n\022GetUserPermissions\022\027.UserPermi" +
|
||||
"ssionsRequest\032\030.UserPermissionsResponse\022" +
|
||||
"G\n\020checkPermissions\022\030.CheckPermissionsRe" +
|
||||
"G\n\020CheckPermissions\022\030.CheckPermissionsRe" +
|
||||
"quest\032\031.CheckPermissionsResponseBI\n*org." +
|
||||
"apache.hadoop.hbase.protobuf.generatedB\023" +
|
||||
"AccessControlProtosH\001\210\001\001\240\001\001"
|
||||
|
|
|
@ -16,7 +16,7 @@ public final class AdminProtos {
|
|||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier getRegion();
|
||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder getRegionOrBuilder();
|
||||
|
||||
// optional bool compactionState = 2;
|
||||
// optional bool compaction_state = 2;
|
||||
boolean hasCompactionState();
|
||||
boolean getCompactionState();
|
||||
}
|
||||
|
@ -62,8 +62,8 @@ public final class AdminProtos {
|
|||
return region_;
|
||||
}
|
||||
|
||||
// optional bool compactionState = 2;
|
||||
public static final int COMPACTIONSTATE_FIELD_NUMBER = 2;
|
||||
// optional bool compaction_state = 2;
|
||||
public static final int COMPACTION_STATE_FIELD_NUMBER = 2;
|
||||
private boolean compactionState_;
|
||||
public boolean hasCompactionState() {
|
||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||
|
@ -166,7 +166,7 @@ public final class AdminProtos {
|
|||
hash = (53 * hash) + getRegion().hashCode();
|
||||
}
|
||||
if (hasCompactionState()) {
|
||||
hash = (37 * hash) + COMPACTIONSTATE_FIELD_NUMBER;
|
||||
hash = (37 * hash) + COMPACTION_STATE_FIELD_NUMBER;
|
||||
hash = (53 * hash) + hashBoolean(getCompactionState());
|
||||
}
|
||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||
|
@ -515,7 +515,7 @@ public final class AdminProtos {
|
|||
return regionBuilder_;
|
||||
}
|
||||
|
||||
// optional bool compactionState = 2;
|
||||
// optional bool compaction_state = 2;
|
||||
private boolean compactionState_ ;
|
||||
public boolean hasCompactionState() {
|
||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||
|
@ -550,12 +550,12 @@ public final class AdminProtos {
|
|||
public interface GetRegionInfoResponseOrBuilder
|
||||
extends com.google.protobuf.MessageOrBuilder {
|
||||
|
||||
// required .RegionInfo regionInfo = 1;
|
||||
// required .RegionInfo region_info = 1;
|
||||
boolean hasRegionInfo();
|
||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfo getRegionInfo();
|
||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfoOrBuilder getRegionInfoOrBuilder();
|
||||
|
||||
// optional .GetRegionInfoResponse.CompactionState compactionState = 2;
|
||||
// optional .GetRegionInfoResponse.CompactionState compaction_state = 2;
|
||||
boolean hasCompactionState();
|
||||
org.apache.hadoop.hbase.protobuf.generated.AdminProtos.GetRegionInfoResponse.CompactionState getCompactionState();
|
||||
}
|
||||
|
@ -663,8 +663,8 @@ public final class AdminProtos {
|
|||
}
|
||||
|
||||
private int bitField0_;
|
||||
// required .RegionInfo regionInfo = 1;
|
||||
public static final int REGIONINFO_FIELD_NUMBER = 1;
|
||||
// required .RegionInfo region_info = 1;
|
||||
public static final int REGION_INFO_FIELD_NUMBER = 1;
|
||||
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfo regionInfo_;
|
||||
public boolean hasRegionInfo() {
|
||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||
|
@ -676,8 +676,8 @@ public final class AdminProtos {
|
|||
return regionInfo_;
|
||||
}
|
||||
|
||||
// optional .GetRegionInfoResponse.CompactionState compactionState = 2;
|
||||
public static final int COMPACTIONSTATE_FIELD_NUMBER = 2;
|
||||
// optional .GetRegionInfoResponse.CompactionState compaction_state = 2;
|
||||
public static final int COMPACTION_STATE_FIELD_NUMBER = 2;
|
||||
private org.apache.hadoop.hbase.protobuf.generated.AdminProtos.GetRegionInfoResponse.CompactionState compactionState_;
|
||||
public boolean hasCompactionState() {
|
||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||
|
@ -776,11 +776,11 @@ public final class AdminProtos {
|
|||
int hash = 41;
|
||||
hash = (19 * hash) + getDescriptorForType().hashCode();
|
||||
if (hasRegionInfo()) {
|
||||
hash = (37 * hash) + REGIONINFO_FIELD_NUMBER;
|
||||
hash = (37 * hash) + REGION_INFO_FIELD_NUMBER;
|
||||
hash = (53 * hash) + getRegionInfo().hashCode();
|
||||
}
|
||||
if (hasCompactionState()) {
|
||||
hash = (37 * hash) + COMPACTIONSTATE_FIELD_NUMBER;
|
||||
hash = (37 * hash) + COMPACTION_STATE_FIELD_NUMBER;
|
||||
hash = (53 * hash) + hashEnum(getCompactionState());
|
||||
}
|
||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||
|
@ -1045,7 +1045,7 @@ public final class AdminProtos {
|
|||
|
||||
private int bitField0_;
|
||||
|
||||
// required .RegionInfo regionInfo = 1;
|
||||
// required .RegionInfo region_info = 1;
|
||||
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfo regionInfo_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfo.getDefaultInstance();
|
||||
private com.google.protobuf.SingleFieldBuilder<
|
||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfo, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfo.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfoOrBuilder> regionInfoBuilder_;
|
||||
|
@ -1135,7 +1135,7 @@ public final class AdminProtos {
|
|||
return regionInfoBuilder_;
|
||||
}
|
||||
|
||||
// optional .GetRegionInfoResponse.CompactionState compactionState = 2;
|
||||
// optional .GetRegionInfoResponse.CompactionState compaction_state = 2;
|
||||
private org.apache.hadoop.hbase.protobuf.generated.AdminProtos.GetRegionInfoResponse.CompactionState compactionState_ = org.apache.hadoop.hbase.protobuf.generated.AdminProtos.GetRegionInfoResponse.CompactionState.NONE;
|
||||
public boolean hasCompactionState() {
|
||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||
|
@ -1757,7 +1757,7 @@ public final class AdminProtos {
|
|||
public interface GetStoreFileResponseOrBuilder
|
||||
extends com.google.protobuf.MessageOrBuilder {
|
||||
|
||||
// repeated string storeFile = 1;
|
||||
// repeated string store_file = 1;
|
||||
java.util.List<String> getStoreFileList();
|
||||
int getStoreFileCount();
|
||||
String getStoreFile(int index);
|
||||
|
@ -1790,8 +1790,8 @@ public final class AdminProtos {
|
|||
return org.apache.hadoop.hbase.protobuf.generated.AdminProtos.internal_static_GetStoreFileResponse_fieldAccessorTable;
|
||||
}
|
||||
|
||||
// repeated string storeFile = 1;
|
||||
public static final int STOREFILE_FIELD_NUMBER = 1;
|
||||
// repeated string store_file = 1;
|
||||
public static final int STORE_FILE_FIELD_NUMBER = 1;
|
||||
private com.google.protobuf.LazyStringList storeFile_;
|
||||
public java.util.List<String>
|
||||
getStoreFileList() {
|
||||
|
@ -1875,7 +1875,7 @@ public final class AdminProtos {
|
|||
int hash = 41;
|
||||
hash = (19 * hash) + getDescriptorForType().hashCode();
|
||||
if (getStoreFileCount() > 0) {
|
||||
hash = (37 * hash) + STOREFILE_FIELD_NUMBER;
|
||||
hash = (37 * hash) + STORE_FILE_FIELD_NUMBER;
|
||||
hash = (53 * hash) + getStoreFileList().hashCode();
|
||||
}
|
||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||
|
@ -2106,7 +2106,7 @@ public final class AdminProtos {
|
|||
|
||||
private int bitField0_;
|
||||
|
||||
// repeated string storeFile = 1;
|
||||
// repeated string store_file = 1;
|
||||
private com.google.protobuf.LazyStringList storeFile_ = com.google.protobuf.LazyStringArrayList.EMPTY;
|
||||
private void ensureStoreFileIsMutable() {
|
||||
if (!((bitField0_ & 0x00000001) == 0x00000001)) {
|
||||
|
@ -2476,7 +2476,7 @@ public final class AdminProtos {
|
|||
public interface GetOnlineRegionResponseOrBuilder
|
||||
extends com.google.protobuf.MessageOrBuilder {
|
||||
|
||||
// repeated .RegionInfo regionInfo = 1;
|
||||
// repeated .RegionInfo region_info = 1;
|
||||
java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfo>
|
||||
getRegionInfoList();
|
||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfo getRegionInfo(int index);
|
||||
|
@ -2514,8 +2514,8 @@ public final class AdminProtos {
|
|||
return org.apache.hadoop.hbase.protobuf.generated.AdminProtos.internal_static_GetOnlineRegionResponse_fieldAccessorTable;
|
||||
}
|
||||
|
||||
// repeated .RegionInfo regionInfo = 1;
|
||||
public static final int REGIONINFO_FIELD_NUMBER = 1;
|
||||
// repeated .RegionInfo region_info = 1;
|
||||
public static final int REGION_INFO_FIELD_NUMBER = 1;
|
||||
private java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfo> regionInfo_;
|
||||
public java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfo> getRegionInfoList() {
|
||||
return regionInfo_;
|
||||
|
@ -2607,7 +2607,7 @@ public final class AdminProtos {
|
|||
int hash = 41;
|
||||
hash = (19 * hash) + getDescriptorForType().hashCode();
|
||||
if (getRegionInfoCount() > 0) {
|
||||
hash = (37 * hash) + REGIONINFO_FIELD_NUMBER;
|
||||
hash = (37 * hash) + REGION_INFO_FIELD_NUMBER;
|
||||
hash = (53 * hash) + getRegionInfoList().hashCode();
|
||||
}
|
||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||
|
@ -2869,7 +2869,7 @@ public final class AdminProtos {
|
|||
|
||||
private int bitField0_;
|
||||
|
||||
// repeated .RegionInfo regionInfo = 1;
|
||||
// repeated .RegionInfo region_info = 1;
|
||||
private java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfo> regionInfo_ =
|
||||
java.util.Collections.emptyList();
|
||||
private void ensureRegionInfoIsMutable() {
|
||||
|
@ -3069,7 +3069,7 @@ public final class AdminProtos {
|
|||
public interface OpenRegionRequestOrBuilder
|
||||
extends com.google.protobuf.MessageOrBuilder {
|
||||
|
||||
// repeated .OpenRegionRequest.RegionOpenInfo openInfo = 1;
|
||||
// repeated .OpenRegionRequest.RegionOpenInfo open_info = 1;
|
||||
java.util.List<org.apache.hadoop.hbase.protobuf.generated.AdminProtos.OpenRegionRequest.RegionOpenInfo>
|
||||
getOpenInfoList();
|
||||
org.apache.hadoop.hbase.protobuf.generated.AdminProtos.OpenRegionRequest.RegionOpenInfo getOpenInfo(int index);
|
||||
|
@ -3115,11 +3115,11 @@ public final class AdminProtos {
|
|||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfo getRegion();
|
||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfoOrBuilder getRegionOrBuilder();
|
||||
|
||||
// optional uint32 versionOfOfflineNode = 2;
|
||||
// optional uint32 version_of_offline_node = 2;
|
||||
boolean hasVersionOfOfflineNode();
|
||||
int getVersionOfOfflineNode();
|
||||
|
||||
// repeated .ServerName favoredNodes = 3;
|
||||
// repeated .ServerName favored_nodes = 3;
|
||||
java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName>
|
||||
getFavoredNodesList();
|
||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName getFavoredNodes(int index);
|
||||
|
@ -3171,8 +3171,8 @@ public final class AdminProtos {
|
|||
return region_;
|
||||
}
|
||||
|
||||
// optional uint32 versionOfOfflineNode = 2;
|
||||
public static final int VERSIONOFOFFLINENODE_FIELD_NUMBER = 2;
|
||||
// optional uint32 version_of_offline_node = 2;
|
||||
public static final int VERSION_OF_OFFLINE_NODE_FIELD_NUMBER = 2;
|
||||
private int versionOfOfflineNode_;
|
||||
public boolean hasVersionOfOfflineNode() {
|
||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||
|
@ -3181,8 +3181,8 @@ public final class AdminProtos {
|
|||
return versionOfOfflineNode_;
|
||||
}
|
||||
|
||||
// repeated .ServerName favoredNodes = 3;
|
||||
public static final int FAVOREDNODES_FIELD_NUMBER = 3;
|
||||
// repeated .ServerName favored_nodes = 3;
|
||||
public static final int FAVORED_NODES_FIELD_NUMBER = 3;
|
||||
private java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName> favoredNodes_;
|
||||
public java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName> getFavoredNodesList() {
|
||||
return favoredNodes_;
|
||||
|
@ -3312,11 +3312,11 @@ public final class AdminProtos {
|
|||
hash = (53 * hash) + getRegion().hashCode();
|
||||
}
|
||||
if (hasVersionOfOfflineNode()) {
|
||||
hash = (37 * hash) + VERSIONOFOFFLINENODE_FIELD_NUMBER;
|
||||
hash = (37 * hash) + VERSION_OF_OFFLINE_NODE_FIELD_NUMBER;
|
||||
hash = (53 * hash) + getVersionOfOfflineNode();
|
||||
}
|
||||
if (getFavoredNodesCount() > 0) {
|
||||
hash = (37 * hash) + FAVOREDNODES_FIELD_NUMBER;
|
||||
hash = (37 * hash) + FAVORED_NODES_FIELD_NUMBER;
|
||||
hash = (53 * hash) + getFavoredNodesList().hashCode();
|
||||
}
|
||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||
|
@ -3719,7 +3719,7 @@ public final class AdminProtos {
|
|||
return regionBuilder_;
|
||||
}
|
||||
|
||||
// optional uint32 versionOfOfflineNode = 2;
|
||||
// optional uint32 version_of_offline_node = 2;
|
||||
private int versionOfOfflineNode_ ;
|
||||
public boolean hasVersionOfOfflineNode() {
|
||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||
|
@ -3740,7 +3740,7 @@ public final class AdminProtos {
|
|||
return this;
|
||||
}
|
||||
|
||||
// repeated .ServerName favoredNodes = 3;
|
||||
// repeated .ServerName favored_nodes = 3;
|
||||
private java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName> favoredNodes_ =
|
||||
java.util.Collections.emptyList();
|
||||
private void ensureFavoredNodesIsMutable() {
|
||||
|
@ -3937,8 +3937,8 @@ public final class AdminProtos {
|
|||
// @@protoc_insertion_point(class_scope:OpenRegionRequest.RegionOpenInfo)
|
||||
}
|
||||
|
||||
// repeated .OpenRegionRequest.RegionOpenInfo openInfo = 1;
|
||||
public static final int OPENINFO_FIELD_NUMBER = 1;
|
||||
// repeated .OpenRegionRequest.RegionOpenInfo open_info = 1;
|
||||
public static final int OPEN_INFO_FIELD_NUMBER = 1;
|
||||
private java.util.List<org.apache.hadoop.hbase.protobuf.generated.AdminProtos.OpenRegionRequest.RegionOpenInfo> openInfo_;
|
||||
public java.util.List<org.apache.hadoop.hbase.protobuf.generated.AdminProtos.OpenRegionRequest.RegionOpenInfo> getOpenInfoList() {
|
||||
return openInfo_;
|
||||
|
@ -4030,7 +4030,7 @@ public final class AdminProtos {
|
|||
int hash = 41;
|
||||
hash = (19 * hash) + getDescriptorForType().hashCode();
|
||||
if (getOpenInfoCount() > 0) {
|
||||
hash = (37 * hash) + OPENINFO_FIELD_NUMBER;
|
||||
hash = (37 * hash) + OPEN_INFO_FIELD_NUMBER;
|
||||
hash = (53 * hash) + getOpenInfoList().hashCode();
|
||||
}
|
||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||
|
@ -4292,7 +4292,7 @@ public final class AdminProtos {
|
|||
|
||||
private int bitField0_;
|
||||
|
||||
// repeated .OpenRegionRequest.RegionOpenInfo openInfo = 1;
|
||||
// repeated .OpenRegionRequest.RegionOpenInfo open_info = 1;
|
||||
private java.util.List<org.apache.hadoop.hbase.protobuf.generated.AdminProtos.OpenRegionRequest.RegionOpenInfo> openInfo_ =
|
||||
java.util.Collections.emptyList();
|
||||
private void ensureOpenInfoIsMutable() {
|
||||
|
@ -4492,7 +4492,7 @@ public final class AdminProtos {
|
|||
public interface OpenRegionResponseOrBuilder
|
||||
extends com.google.protobuf.MessageOrBuilder {
|
||||
|
||||
// repeated .OpenRegionResponse.RegionOpeningState openingState = 1;
|
||||
// repeated .OpenRegionResponse.RegionOpeningState opening_state = 1;
|
||||
java.util.List<org.apache.hadoop.hbase.protobuf.generated.AdminProtos.OpenRegionResponse.RegionOpeningState> getOpeningStateList();
|
||||
int getOpeningStateCount();
|
||||
org.apache.hadoop.hbase.protobuf.generated.AdminProtos.OpenRegionResponse.RegionOpeningState getOpeningState(int index);
|
||||
|
@ -4597,8 +4597,8 @@ public final class AdminProtos {
|
|||
// @@protoc_insertion_point(enum_scope:OpenRegionResponse.RegionOpeningState)
|
||||
}
|
||||
|
||||
// repeated .OpenRegionResponse.RegionOpeningState openingState = 1;
|
||||
public static final int OPENINGSTATE_FIELD_NUMBER = 1;
|
||||
// repeated .OpenRegionResponse.RegionOpeningState opening_state = 1;
|
||||
public static final int OPENING_STATE_FIELD_NUMBER = 1;
|
||||
private java.util.List<org.apache.hadoop.hbase.protobuf.generated.AdminProtos.OpenRegionResponse.RegionOpeningState> openingState_;
|
||||
public java.util.List<org.apache.hadoop.hbase.protobuf.generated.AdminProtos.OpenRegionResponse.RegionOpeningState> getOpeningStateList() {
|
||||
return openingState_;
|
||||
|
@ -4681,7 +4681,7 @@ public final class AdminProtos {
|
|||
int hash = 41;
|
||||
hash = (19 * hash) + getDescriptorForType().hashCode();
|
||||
if (getOpeningStateCount() > 0) {
|
||||
hash = (37 * hash) + OPENINGSTATE_FIELD_NUMBER;
|
||||
hash = (37 * hash) + OPENING_STATE_FIELD_NUMBER;
|
||||
hash = (53 * hash) + hashEnumList(getOpeningStateList());
|
||||
}
|
||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||
|
@ -4931,7 +4931,7 @@ public final class AdminProtos {
|
|||
|
||||
private int bitField0_;
|
||||
|
||||
// repeated .OpenRegionResponse.RegionOpeningState openingState = 1;
|
||||
// repeated .OpenRegionResponse.RegionOpeningState opening_state = 1;
|
||||
private java.util.List<org.apache.hadoop.hbase.protobuf.generated.AdminProtos.OpenRegionResponse.RegionOpeningState> openingState_ =
|
||||
java.util.Collections.emptyList();
|
||||
private void ensureOpeningStateIsMutable() {
|
||||
|
@ -5001,15 +5001,15 @@ public final class AdminProtos {
|
|||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier getRegion();
|
||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder getRegionOrBuilder();
|
||||
|
||||
// optional uint32 versionOfClosingNode = 2;
|
||||
// optional uint32 version_of_closing_node = 2;
|
||||
boolean hasVersionOfClosingNode();
|
||||
int getVersionOfClosingNode();
|
||||
|
||||
// optional bool transitionInZK = 3 [default = true];
|
||||
// optional bool transition_in_ZK = 3 [default = true];
|
||||
boolean hasTransitionInZK();
|
||||
boolean getTransitionInZK();
|
||||
|
||||
// optional .ServerName destinationServer = 4;
|
||||
// optional .ServerName destination_server = 4;
|
||||
boolean hasDestinationServer();
|
||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName getDestinationServer();
|
||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerNameOrBuilder getDestinationServerOrBuilder();
|
||||
|
@ -5056,8 +5056,8 @@ public final class AdminProtos {
|
|||
return region_;
|
||||
}
|
||||
|
||||
// optional uint32 versionOfClosingNode = 2;
|
||||
public static final int VERSIONOFCLOSINGNODE_FIELD_NUMBER = 2;
|
||||
// optional uint32 version_of_closing_node = 2;
|
||||
public static final int VERSION_OF_CLOSING_NODE_FIELD_NUMBER = 2;
|
||||
private int versionOfClosingNode_;
|
||||
public boolean hasVersionOfClosingNode() {
|
||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||
|
@ -5066,8 +5066,8 @@ public final class AdminProtos {
|
|||
return versionOfClosingNode_;
|
||||
}
|
||||
|
||||
// optional bool transitionInZK = 3 [default = true];
|
||||
public static final int TRANSITIONINZK_FIELD_NUMBER = 3;
|
||||
// optional bool transition_in_ZK = 3 [default = true];
|
||||
public static final int TRANSITION_IN_ZK_FIELD_NUMBER = 3;
|
||||
private boolean transitionInZK_;
|
||||
public boolean hasTransitionInZK() {
|
||||
return ((bitField0_ & 0x00000004) == 0x00000004);
|
||||
|
@ -5076,8 +5076,8 @@ public final class AdminProtos {
|
|||
return transitionInZK_;
|
||||
}
|
||||
|
||||
// optional .ServerName destinationServer = 4;
|
||||
public static final int DESTINATIONSERVER_FIELD_NUMBER = 4;
|
||||
// optional .ServerName destination_server = 4;
|
||||
public static final int DESTINATION_SERVER_FIELD_NUMBER = 4;
|
||||
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName destinationServer_;
|
||||
public boolean hasDestinationServer() {
|
||||
return ((bitField0_ & 0x00000008) == 0x00000008);
|
||||
|
@ -5215,15 +5215,15 @@ public final class AdminProtos {
|
|||
hash = (53 * hash) + getRegion().hashCode();
|
||||
}
|
||||
if (hasVersionOfClosingNode()) {
|
||||
hash = (37 * hash) + VERSIONOFCLOSINGNODE_FIELD_NUMBER;
|
||||
hash = (37 * hash) + VERSION_OF_CLOSING_NODE_FIELD_NUMBER;
|
||||
hash = (53 * hash) + getVersionOfClosingNode();
|
||||
}
|
||||
if (hasTransitionInZK()) {
|
||||
hash = (37 * hash) + TRANSITIONINZK_FIELD_NUMBER;
|
||||
hash = (37 * hash) + TRANSITION_IN_ZK_FIELD_NUMBER;
|
||||
hash = (53 * hash) + hashBoolean(getTransitionInZK());
|
||||
}
|
||||
if (hasDestinationServer()) {
|
||||
hash = (37 * hash) + DESTINATIONSERVER_FIELD_NUMBER;
|
||||
hash = (37 * hash) + DESTINATION_SERVER_FIELD_NUMBER;
|
||||
hash = (53 * hash) + getDestinationServer().hashCode();
|
||||
}
|
||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||
|
@ -5619,7 +5619,7 @@ public final class AdminProtos {
|
|||
return regionBuilder_;
|
||||
}
|
||||
|
||||
// optional uint32 versionOfClosingNode = 2;
|
||||
// optional uint32 version_of_closing_node = 2;
|
||||
private int versionOfClosingNode_ ;
|
||||
public boolean hasVersionOfClosingNode() {
|
||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||
|
@ -5640,7 +5640,7 @@ public final class AdminProtos {
|
|||
return this;
|
||||
}
|
||||
|
||||
// optional bool transitionInZK = 3 [default = true];
|
||||
// optional bool transition_in_ZK = 3 [default = true];
|
||||
private boolean transitionInZK_ = true;
|
||||
public boolean hasTransitionInZK() {
|
||||
return ((bitField0_ & 0x00000004) == 0x00000004);
|
||||
|
@ -5661,7 +5661,7 @@ public final class AdminProtos {
|
|||
return this;
|
||||
}
|
||||
|
||||
// optional .ServerName destinationServer = 4;
|
||||
// optional .ServerName destination_server = 4;
|
||||
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName destinationServer_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.getDefaultInstance();
|
||||
private com.google.protobuf.SingleFieldBuilder<
|
||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerNameOrBuilder> destinationServerBuilder_;
|
||||
|
@ -6149,7 +6149,7 @@ public final class AdminProtos {
|
|||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier getRegion();
|
||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder getRegionOrBuilder();
|
||||
|
||||
// optional uint64 ifOlderThanTs = 2;
|
||||
// optional uint64 if_older_than_ts = 2;
|
||||
boolean hasIfOlderThanTs();
|
||||
long getIfOlderThanTs();
|
||||
}
|
||||
|
@ -6195,8 +6195,8 @@ public final class AdminProtos {
|
|||
return region_;
|
||||
}
|
||||
|
||||
// optional uint64 ifOlderThanTs = 2;
|
||||
public static final int IFOLDERTHANTS_FIELD_NUMBER = 2;
|
||||
// optional uint64 if_older_than_ts = 2;
|
||||
public static final int IF_OLDER_THAN_TS_FIELD_NUMBER = 2;
|
||||
private long ifOlderThanTs_;
|
||||
public boolean hasIfOlderThanTs() {
|
||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||
|
@ -6299,7 +6299,7 @@ public final class AdminProtos {
|
|||
hash = (53 * hash) + getRegion().hashCode();
|
||||
}
|
||||
if (hasIfOlderThanTs()) {
|
||||
hash = (37 * hash) + IFOLDERTHANTS_FIELD_NUMBER;
|
||||
hash = (37 * hash) + IF_OLDER_THAN_TS_FIELD_NUMBER;
|
||||
hash = (53 * hash) + hashLong(getIfOlderThanTs());
|
||||
}
|
||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||
|
@ -6648,7 +6648,7 @@ public final class AdminProtos {
|
|||
return regionBuilder_;
|
||||
}
|
||||
|
||||
// optional uint64 ifOlderThanTs = 2;
|
||||
// optional uint64 if_older_than_ts = 2;
|
||||
private long ifOlderThanTs_ ;
|
||||
public boolean hasIfOlderThanTs() {
|
||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||
|
@ -6683,7 +6683,7 @@ public final class AdminProtos {
|
|||
public interface FlushRegionResponseOrBuilder
|
||||
extends com.google.protobuf.MessageOrBuilder {
|
||||
|
||||
// required uint64 lastFlushTime = 1;
|
||||
// required uint64 last_flush_time = 1;
|
||||
boolean hasLastFlushTime();
|
||||
long getLastFlushTime();
|
||||
|
||||
|
@ -6720,8 +6720,8 @@ public final class AdminProtos {
|
|||
}
|
||||
|
||||
private int bitField0_;
|
||||
// required uint64 lastFlushTime = 1;
|
||||
public static final int LASTFLUSHTIME_FIELD_NUMBER = 1;
|
||||
// required uint64 last_flush_time = 1;
|
||||
public static final int LAST_FLUSH_TIME_FIELD_NUMBER = 1;
|
||||
private long lastFlushTime_;
|
||||
public boolean hasLastFlushTime() {
|
||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||
|
@ -6826,7 +6826,7 @@ public final class AdminProtos {
|
|||
int hash = 41;
|
||||
hash = (19 * hash) + getDescriptorForType().hashCode();
|
||||
if (hasLastFlushTime()) {
|
||||
hash = (37 * hash) + LASTFLUSHTIME_FIELD_NUMBER;
|
||||
hash = (37 * hash) + LAST_FLUSH_TIME_FIELD_NUMBER;
|
||||
hash = (53 * hash) + hashLong(getLastFlushTime());
|
||||
}
|
||||
if (hasFlushed()) {
|
||||
|
@ -7072,7 +7072,7 @@ public final class AdminProtos {
|
|||
|
||||
private int bitField0_;
|
||||
|
||||
// required uint64 lastFlushTime = 1;
|
||||
// required uint64 last_flush_time = 1;
|
||||
private long lastFlushTime_ ;
|
||||
public boolean hasLastFlushTime() {
|
||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||
|
@ -7133,7 +7133,7 @@ public final class AdminProtos {
|
|||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier getRegion();
|
||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder getRegionOrBuilder();
|
||||
|
||||
// optional bytes splitPoint = 2;
|
||||
// optional bytes split_point = 2;
|
||||
boolean hasSplitPoint();
|
||||
com.google.protobuf.ByteString getSplitPoint();
|
||||
}
|
||||
|
@ -7179,8 +7179,8 @@ public final class AdminProtos {
|
|||
return region_;
|
||||
}
|
||||
|
||||
// optional bytes splitPoint = 2;
|
||||
public static final int SPLITPOINT_FIELD_NUMBER = 2;
|
||||
// optional bytes split_point = 2;
|
||||
public static final int SPLIT_POINT_FIELD_NUMBER = 2;
|
||||
private com.google.protobuf.ByteString splitPoint_;
|
||||
public boolean hasSplitPoint() {
|
||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||
|
@ -7283,7 +7283,7 @@ public final class AdminProtos {
|
|||
hash = (53 * hash) + getRegion().hashCode();
|
||||
}
|
||||
if (hasSplitPoint()) {
|
||||
hash = (37 * hash) + SPLITPOINT_FIELD_NUMBER;
|
||||
hash = (37 * hash) + SPLIT_POINT_FIELD_NUMBER;
|
||||
hash = (53 * hash) + getSplitPoint().hashCode();
|
||||
}
|
||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||
|
@ -7632,7 +7632,7 @@ public final class AdminProtos {
|
|||
return regionBuilder_;
|
||||
}
|
||||
|
||||
// optional bytes splitPoint = 2;
|
||||
// optional bytes split_point = 2;
|
||||
private com.google.protobuf.ByteString splitPoint_ = com.google.protobuf.ByteString.EMPTY;
|
||||
public boolean hasSplitPoint() {
|
||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||
|
@ -8878,12 +8878,12 @@ public final class AdminProtos {
|
|||
public interface MergeRegionsRequestOrBuilder
|
||||
extends com.google.protobuf.MessageOrBuilder {
|
||||
|
||||
// required .RegionSpecifier regionA = 1;
|
||||
// required .RegionSpecifier region_a = 1;
|
||||
boolean hasRegionA();
|
||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier getRegionA();
|
||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder getRegionAOrBuilder();
|
||||
|
||||
// required .RegionSpecifier regionB = 2;
|
||||
// required .RegionSpecifier region_b = 2;
|
||||
boolean hasRegionB();
|
||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier getRegionB();
|
||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder getRegionBOrBuilder();
|
||||
|
@ -8921,8 +8921,8 @@ public final class AdminProtos {
|
|||
}
|
||||
|
||||
private int bitField0_;
|
||||
// required .RegionSpecifier regionA = 1;
|
||||
public static final int REGIONA_FIELD_NUMBER = 1;
|
||||
// required .RegionSpecifier region_a = 1;
|
||||
public static final int REGION_A_FIELD_NUMBER = 1;
|
||||
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier regionA_;
|
||||
public boolean hasRegionA() {
|
||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||
|
@ -8934,8 +8934,8 @@ public final class AdminProtos {
|
|||
return regionA_;
|
||||
}
|
||||
|
||||
// required .RegionSpecifier regionB = 2;
|
||||
public static final int REGIONB_FIELD_NUMBER = 2;
|
||||
// required .RegionSpecifier region_b = 2;
|
||||
public static final int REGION_B_FIELD_NUMBER = 2;
|
||||
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier regionB_;
|
||||
public boolean hasRegionB() {
|
||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||
|
@ -9068,11 +9068,11 @@ public final class AdminProtos {
|
|||
int hash = 41;
|
||||
hash = (19 * hash) + getDescriptorForType().hashCode();
|
||||
if (hasRegionA()) {
|
||||
hash = (37 * hash) + REGIONA_FIELD_NUMBER;
|
||||
hash = (37 * hash) + REGION_A_FIELD_NUMBER;
|
||||
hash = (53 * hash) + getRegionA().hashCode();
|
||||
}
|
||||
if (hasRegionB()) {
|
||||
hash = (37 * hash) + REGIONB_FIELD_NUMBER;
|
||||
hash = (37 * hash) + REGION_B_FIELD_NUMBER;
|
||||
hash = (53 * hash) + getRegionB().hashCode();
|
||||
}
|
||||
if (hasForcible()) {
|
||||
|
@ -9370,7 +9370,7 @@ public final class AdminProtos {
|
|||
|
||||
private int bitField0_;
|
||||
|
||||
// required .RegionSpecifier regionA = 1;
|
||||
// required .RegionSpecifier region_a = 1;
|
||||
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier regionA_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance();
|
||||
private com.google.protobuf.SingleFieldBuilder<
|
||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder> regionABuilder_;
|
||||
|
@ -9460,7 +9460,7 @@ public final class AdminProtos {
|
|||
return regionABuilder_;
|
||||
}
|
||||
|
||||
// required .RegionSpecifier regionB = 2;
|
||||
// required .RegionSpecifier region_b = 2;
|
||||
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier regionB_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance();
|
||||
private com.google.protobuf.SingleFieldBuilder<
|
||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder> regionBBuilder_;
|
||||
|
@ -9890,12 +9890,12 @@ public final class AdminProtos {
|
|||
org.apache.hadoop.hbase.protobuf.generated.WALProtos.WALKey getKey();
|
||||
org.apache.hadoop.hbase.protobuf.generated.WALProtos.WALKeyOrBuilder getKeyOrBuilder();
|
||||
|
||||
// repeated bytes keyValueBytes = 2;
|
||||
// repeated bytes key_value_bytes = 2;
|
||||
java.util.List<com.google.protobuf.ByteString> getKeyValueBytesList();
|
||||
int getKeyValueBytesCount();
|
||||
com.google.protobuf.ByteString getKeyValueBytes(int index);
|
||||
|
||||
// optional int32 associatedCellCount = 3;
|
||||
// optional int32 associated_cell_count = 3;
|
||||
boolean hasAssociatedCellCount();
|
||||
int getAssociatedCellCount();
|
||||
}
|
||||
|
@ -9941,8 +9941,8 @@ public final class AdminProtos {
|
|||
return key_;
|
||||
}
|
||||
|
||||
// repeated bytes keyValueBytes = 2;
|
||||
public static final int KEYVALUEBYTES_FIELD_NUMBER = 2;
|
||||
// repeated bytes key_value_bytes = 2;
|
||||
public static final int KEY_VALUE_BYTES_FIELD_NUMBER = 2;
|
||||
private java.util.List<com.google.protobuf.ByteString> keyValueBytes_;
|
||||
public java.util.List<com.google.protobuf.ByteString>
|
||||
getKeyValueBytesList() {
|
||||
|
@ -9955,8 +9955,8 @@ public final class AdminProtos {
|
|||
return keyValueBytes_.get(index);
|
||||
}
|
||||
|
||||
// optional int32 associatedCellCount = 3;
|
||||
public static final int ASSOCIATEDCELLCOUNT_FIELD_NUMBER = 3;
|
||||
// optional int32 associated_cell_count = 3;
|
||||
public static final int ASSOCIATED_CELL_COUNT_FIELD_NUMBER = 3;
|
||||
private int associatedCellCount_;
|
||||
public boolean hasAssociatedCellCount() {
|
||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||
|
@ -10074,11 +10074,11 @@ public final class AdminProtos {
|
|||
hash = (53 * hash) + getKey().hashCode();
|
||||
}
|
||||
if (getKeyValueBytesCount() > 0) {
|
||||
hash = (37 * hash) + KEYVALUEBYTES_FIELD_NUMBER;
|
||||
hash = (37 * hash) + KEY_VALUE_BYTES_FIELD_NUMBER;
|
||||
hash = (53 * hash) + getKeyValueBytesList().hashCode();
|
||||
}
|
||||
if (hasAssociatedCellCount()) {
|
||||
hash = (37 * hash) + ASSOCIATEDCELLCOUNT_FIELD_NUMBER;
|
||||
hash = (37 * hash) + ASSOCIATED_CELL_COUNT_FIELD_NUMBER;
|
||||
hash = (53 * hash) + getAssociatedCellCount();
|
||||
}
|
||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||
|
@ -10449,7 +10449,7 @@ public final class AdminProtos {
|
|||
return keyBuilder_;
|
||||
}
|
||||
|
||||
// repeated bytes keyValueBytes = 2;
|
||||
// repeated bytes key_value_bytes = 2;
|
||||
private java.util.List<com.google.protobuf.ByteString> keyValueBytes_ = java.util.Collections.emptyList();;
|
||||
private void ensureKeyValueBytesIsMutable() {
|
||||
if (!((bitField0_ & 0x00000002) == 0x00000002)) {
|
||||
|
@ -10500,7 +10500,7 @@ public final class AdminProtos {
|
|||
return this;
|
||||
}
|
||||
|
||||
// optional int32 associatedCellCount = 3;
|
||||
// optional int32 associated_cell_count = 3;
|
||||
private int associatedCellCount_ ;
|
||||
public boolean hasAssociatedCellCount() {
|
||||
return ((bitField0_ & 0x00000004) == 0x00000004);
|
||||
|
@ -11728,7 +11728,7 @@ public final class AdminProtos {
|
|||
public interface RollWALWriterResponseOrBuilder
|
||||
extends com.google.protobuf.MessageOrBuilder {
|
||||
|
||||
// repeated bytes regionToFlush = 1;
|
||||
// repeated bytes region_to_flush = 1;
|
||||
java.util.List<com.google.protobuf.ByteString> getRegionToFlushList();
|
||||
int getRegionToFlushCount();
|
||||
com.google.protobuf.ByteString getRegionToFlush(int index);
|
||||
|
@ -11761,8 +11761,8 @@ public final class AdminProtos {
|
|||
return org.apache.hadoop.hbase.protobuf.generated.AdminProtos.internal_static_RollWALWriterResponse_fieldAccessorTable;
|
||||
}
|
||||
|
||||
// repeated bytes regionToFlush = 1;
|
||||
public static final int REGIONTOFLUSH_FIELD_NUMBER = 1;
|
||||
// repeated bytes region_to_flush = 1;
|
||||
public static final int REGION_TO_FLUSH_FIELD_NUMBER = 1;
|
||||
private java.util.List<com.google.protobuf.ByteString> regionToFlush_;
|
||||
public java.util.List<com.google.protobuf.ByteString>
|
||||
getRegionToFlushList() {
|
||||
|
@ -11846,7 +11846,7 @@ public final class AdminProtos {
|
|||
int hash = 41;
|
||||
hash = (19 * hash) + getDescriptorForType().hashCode();
|
||||
if (getRegionToFlushCount() > 0) {
|
||||
hash = (37 * hash) + REGIONTOFLUSH_FIELD_NUMBER;
|
||||
hash = (37 * hash) + REGION_TO_FLUSH_FIELD_NUMBER;
|
||||
hash = (53 * hash) + getRegionToFlushList().hashCode();
|
||||
}
|
||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||
|
@ -12076,7 +12076,7 @@ public final class AdminProtos {
|
|||
|
||||
private int bitField0_;
|
||||
|
||||
// repeated bytes regionToFlush = 1;
|
||||
// repeated bytes region_to_flush = 1;
|
||||
private java.util.List<com.google.protobuf.ByteString> regionToFlush_ = java.util.Collections.emptyList();;
|
||||
private void ensureRegionToFlushIsMutable() {
|
||||
if (!((bitField0_ & 0x00000001) == 0x00000001)) {
|
||||
|
@ -13157,12 +13157,12 @@ public final class AdminProtos {
|
|||
public interface ServerInfoOrBuilder
|
||||
extends com.google.protobuf.MessageOrBuilder {
|
||||
|
||||
// required .ServerName serverName = 1;
|
||||
// required .ServerName server_name = 1;
|
||||
boolean hasServerName();
|
||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName getServerName();
|
||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerNameOrBuilder getServerNameOrBuilder();
|
||||
|
||||
// optional uint32 webuiPort = 2;
|
||||
// optional uint32 webui_port = 2;
|
||||
boolean hasWebuiPort();
|
||||
int getWebuiPort();
|
||||
}
|
||||
|
@ -13195,8 +13195,8 @@ public final class AdminProtos {
|
|||
}
|
||||
|
||||
private int bitField0_;
|
||||
// required .ServerName serverName = 1;
|
||||
public static final int SERVERNAME_FIELD_NUMBER = 1;
|
||||
// required .ServerName server_name = 1;
|
||||
public static final int SERVER_NAME_FIELD_NUMBER = 1;
|
||||
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName serverName_;
|
||||
public boolean hasServerName() {
|
||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||
|
@ -13208,8 +13208,8 @@ public final class AdminProtos {
|
|||
return serverName_;
|
||||
}
|
||||
|
||||
// optional uint32 webuiPort = 2;
|
||||
public static final int WEBUIPORT_FIELD_NUMBER = 2;
|
||||
// optional uint32 webui_port = 2;
|
||||
public static final int WEBUI_PORT_FIELD_NUMBER = 2;
|
||||
private int webuiPort_;
|
||||
public boolean hasWebuiPort() {
|
||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||
|
@ -13308,11 +13308,11 @@ public final class AdminProtos {
|
|||
int hash = 41;
|
||||
hash = (19 * hash) + getDescriptorForType().hashCode();
|
||||
if (hasServerName()) {
|
||||
hash = (37 * hash) + SERVERNAME_FIELD_NUMBER;
|
||||
hash = (37 * hash) + SERVER_NAME_FIELD_NUMBER;
|
||||
hash = (53 * hash) + getServerName().hashCode();
|
||||
}
|
||||
if (hasWebuiPort()) {
|
||||
hash = (37 * hash) + WEBUIPORT_FIELD_NUMBER;
|
||||
hash = (37 * hash) + WEBUI_PORT_FIELD_NUMBER;
|
||||
hash = (53 * hash) + getWebuiPort();
|
||||
}
|
||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||
|
@ -13571,7 +13571,7 @@ public final class AdminProtos {
|
|||
|
||||
private int bitField0_;
|
||||
|
||||
// required .ServerName serverName = 1;
|
||||
// required .ServerName server_name = 1;
|
||||
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName serverName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.getDefaultInstance();
|
||||
private com.google.protobuf.SingleFieldBuilder<
|
||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerNameOrBuilder> serverNameBuilder_;
|
||||
|
@ -13661,7 +13661,7 @@ public final class AdminProtos {
|
|||
return serverNameBuilder_;
|
||||
}
|
||||
|
||||
// optional uint32 webuiPort = 2;
|
||||
// optional uint32 webui_port = 2;
|
||||
private int webuiPort_ ;
|
||||
public boolean hasWebuiPort() {
|
||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||
|
@ -13696,7 +13696,7 @@ public final class AdminProtos {
|
|||
public interface GetServerInfoResponseOrBuilder
|
||||
extends com.google.protobuf.MessageOrBuilder {
|
||||
|
||||
// required .ServerInfo serverInfo = 1;
|
||||
// required .ServerInfo server_info = 1;
|
||||
boolean hasServerInfo();
|
||||
org.apache.hadoop.hbase.protobuf.generated.AdminProtos.ServerInfo getServerInfo();
|
||||
org.apache.hadoop.hbase.protobuf.generated.AdminProtos.ServerInfoOrBuilder getServerInfoOrBuilder();
|
||||
|
@ -13730,8 +13730,8 @@ public final class AdminProtos {
|
|||
}
|
||||
|
||||
private int bitField0_;
|
||||
// required .ServerInfo serverInfo = 1;
|
||||
public static final int SERVERINFO_FIELD_NUMBER = 1;
|
||||
// required .ServerInfo server_info = 1;
|
||||
public static final int SERVER_INFO_FIELD_NUMBER = 1;
|
||||
private org.apache.hadoop.hbase.protobuf.generated.AdminProtos.ServerInfo serverInfo_;
|
||||
public boolean hasServerInfo() {
|
||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||
|
@ -13820,7 +13820,7 @@ public final class AdminProtos {
|
|||
int hash = 41;
|
||||
hash = (19 * hash) + getDescriptorForType().hashCode();
|
||||
if (hasServerInfo()) {
|
||||
hash = (37 * hash) + SERVERINFO_FIELD_NUMBER;
|
||||
hash = (37 * hash) + SERVER_INFO_FIELD_NUMBER;
|
||||
hash = (53 * hash) + getServerInfo().hashCode();
|
||||
}
|
||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||
|
@ -14065,7 +14065,7 @@ public final class AdminProtos {
|
|||
|
||||
private int bitField0_;
|
||||
|
||||
// required .ServerInfo serverInfo = 1;
|
||||
// required .ServerInfo server_info = 1;
|
||||
private org.apache.hadoop.hbase.protobuf.generated.AdminProtos.ServerInfo serverInfo_ = org.apache.hadoop.hbase.protobuf.generated.AdminProtos.ServerInfo.getDefaultInstance();
|
||||
private com.google.protobuf.SingleFieldBuilder<
|
||||
org.apache.hadoop.hbase.protobuf.generated.AdminProtos.ServerInfo, org.apache.hadoop.hbase.protobuf.generated.AdminProtos.ServerInfo.Builder, org.apache.hadoop.hbase.protobuf.generated.AdminProtos.ServerInfoOrBuilder> serverInfoBuilder_;
|
||||
|
@ -15387,79 +15387,80 @@ public final class AdminProtos {
|
|||
static {
|
||||
java.lang.String[] descriptorData = {
|
||||
"\n\013Admin.proto\032\014Client.proto\032\013hbase.proto" +
|
||||
"\032\tWAL.proto\"Q\n\024GetRegionInfoRequest\022 \n\006r" +
|
||||
"egion\030\001 \002(\0132\020.RegionSpecifier\022\027\n\017compact" +
|
||||
"ionState\030\002 \001(\010\"\301\001\n\025GetRegionInfoResponse" +
|
||||
"\022\037\n\nregionInfo\030\001 \002(\0132\013.RegionInfo\022?\n\017com" +
|
||||
"pactionState\030\002 \001(\0162&.GetRegionInfoRespon" +
|
||||
"se.CompactionState\"F\n\017CompactionState\022\010\n" +
|
||||
"\004NONE\020\000\022\t\n\005MINOR\020\001\022\t\n\005MAJOR\020\002\022\023\n\017MAJOR_A" +
|
||||
"ND_MINOR\020\003\"G\n\023GetStoreFileRequest\022 \n\006reg" +
|
||||
"ion\030\001 \002(\0132\020.RegionSpecifier\022\016\n\006family\030\002 ",
|
||||
"\003(\014\")\n\024GetStoreFileResponse\022\021\n\tstoreFile" +
|
||||
"\030\001 \003(\t\"\030\n\026GetOnlineRegionRequest\":\n\027GetO" +
|
||||
"nlineRegionResponse\022\037\n\nregionInfo\030\001 \003(\0132" +
|
||||
"\013.RegionInfo\"\270\001\n\021OpenRegionRequest\0223\n\010op" +
|
||||
"enInfo\030\001 \003(\0132!.OpenRegionRequest.RegionO" +
|
||||
"penInfo\032n\n\016RegionOpenInfo\022\033\n\006region\030\001 \002(" +
|
||||
"\0132\013.RegionInfo\022\034\n\024versionOfOfflineNode\030\002" +
|
||||
" \001(\r\022!\n\014favoredNodes\030\003 \003(\0132\013.ServerName\"" +
|
||||
"\234\001\n\022OpenRegionResponse\022<\n\014openingState\030\001" +
|
||||
" \003(\0162&.OpenRegionResponse.RegionOpeningS",
|
||||
"tate\"H\n\022RegionOpeningState\022\n\n\006OPENED\020\000\022\022" +
|
||||
"\n\016ALREADY_OPENED\020\001\022\022\n\016FAILED_OPENING\020\002\"\232" +
|
||||
"\001\n\022CloseRegionRequest\022 \n\006region\030\001 \002(\0132\020." +
|
||||
"RegionSpecifier\022\034\n\024versionOfClosingNode\030" +
|
||||
"\002 \001(\r\022\034\n\016transitionInZK\030\003 \001(\010:\004true\022&\n\021d" +
|
||||
"estinationServer\030\004 \001(\0132\013.ServerName\"%\n\023C" +
|
||||
"loseRegionResponse\022\016\n\006closed\030\001 \002(\010\"M\n\022Fl" +
|
||||
"ushRegionRequest\022 \n\006region\030\001 \002(\0132\020.Regio" +
|
||||
"nSpecifier\022\025\n\rifOlderThanTs\030\002 \001(\004\"=\n\023Flu" +
|
||||
"shRegionResponse\022\025\n\rlastFlushTime\030\001 \002(\004\022",
|
||||
"\017\n\007flushed\030\002 \001(\010\"J\n\022SplitRegionRequest\022 " +
|
||||
"\n\006region\030\001 \002(\0132\020.RegionSpecifier\022\022\n\nspli" +
|
||||
"tPoint\030\002 \001(\014\"\025\n\023SplitRegionResponse\"W\n\024C" +
|
||||
"ompactRegionRequest\022 \n\006region\030\001 \002(\0132\020.Re" +
|
||||
"gionSpecifier\022\r\n\005major\030\002 \001(\010\022\016\n\006family\030\003" +
|
||||
" \001(\014\"\027\n\025CompactRegionResponse\"t\n\023MergeRe" +
|
||||
"gionsRequest\022!\n\007regionA\030\001 \002(\0132\020.RegionSp" +
|
||||
"ecifier\022!\n\007regionB\030\002 \002(\0132\020.RegionSpecifi" +
|
||||
"er\022\027\n\010forcible\030\003 \001(\010:\005false\"\026\n\024MergeRegi" +
|
||||
"onsResponse\"T\n\010WALEntry\022\024\n\003key\030\001 \002(\0132\007.W",
|
||||
"ALKey\022\025\n\rkeyValueBytes\030\002 \003(\014\022\033\n\023associat" +
|
||||
"edCellCount\030\003 \001(\005\"4\n\030ReplicateWALEntryRe" +
|
||||
"quest\022\030\n\005entry\030\001 \003(\0132\t.WALEntry\"\033\n\031Repli" +
|
||||
"cateWALEntryResponse\"\026\n\024RollWALWriterReq" +
|
||||
"uest\".\n\025RollWALWriterResponse\022\025\n\rregionT" +
|
||||
"oFlush\030\001 \003(\014\"#\n\021StopServerRequest\022\016\n\006rea" +
|
||||
"son\030\001 \002(\t\"\024\n\022StopServerResponse\"\026\n\024GetSe" +
|
||||
"rverInfoRequest\"@\n\nServerInfo\022\037\n\nserverN" +
|
||||
"ame\030\001 \002(\0132\013.ServerName\022\021\n\twebuiPort\030\002 \001(" +
|
||||
"\r\"8\n\025GetServerInfoResponse\022\037\n\nserverInfo",
|
||||
"\030\001 \002(\0132\013.ServerInfo2\337\006\n\014AdminService\022>\n\r" +
|
||||
"getRegionInfo\022\025.GetRegionInfoRequest\032\026.G" +
|
||||
"etRegionInfoResponse\022;\n\014getStoreFile\022\024.G" +
|
||||
"etStoreFileRequest\032\025.GetStoreFileRespons" +
|
||||
"e\022D\n\017getOnlineRegion\022\027.GetOnlineRegionRe" +
|
||||
"quest\032\030.GetOnlineRegionResponse\0225\n\nopenR" +
|
||||
"egion\022\022.OpenRegionRequest\032\023.OpenRegionRe" +
|
||||
"sponse\0228\n\013closeRegion\022\023.CloseRegionReque" +
|
||||
"st\032\024.CloseRegionResponse\0228\n\013flushRegion\022" +
|
||||
"\023.FlushRegionRequest\032\024.FlushRegionRespon",
|
||||
"se\0228\n\013splitRegion\022\023.SplitRegionRequest\032\024" +
|
||||
".SplitRegionResponse\022>\n\rcompactRegion\022\025." +
|
||||
"CompactRegionRequest\032\026.CompactRegionResp" +
|
||||
"onse\022;\n\014mergeRegions\022\024.MergeRegionsReque" +
|
||||
"st\032\025.MergeRegionsResponse\022J\n\021replicateWA" +
|
||||
"LEntry\022\031.ReplicateWALEntryRequest\032\032.Repl" +
|
||||
"icateWALEntryResponse\022\'\n\006replay\022\r.MultiR" +
|
||||
"equest\032\016.MultiResponse\022>\n\rrollWALWriter\022" +
|
||||
"\025.RollWALWriterRequest\032\026.RollWALWriterRe" +
|
||||
"sponse\022>\n\rgetServerInfo\022\025.GetServerInfoR",
|
||||
"equest\032\026.GetServerInfoResponse\0225\n\nstopSe" +
|
||||
"rver\022\022.StopServerRequest\032\023.StopServerRes" +
|
||||
"ponseBA\n*org.apache.hadoop.hbase.protobu" +
|
||||
"f.generatedB\013AdminProtosH\001\210\001\001\240\001\001"
|
||||
"\032\tWAL.proto\"R\n\024GetRegionInfoRequest\022 \n\006r" +
|
||||
"egion\030\001 \002(\0132\020.RegionSpecifier\022\030\n\020compact" +
|
||||
"ion_state\030\002 \001(\010\"\303\001\n\025GetRegionInfoRespons" +
|
||||
"e\022 \n\013region_info\030\001 \002(\0132\013.RegionInfo\022@\n\020c" +
|
||||
"ompaction_state\030\002 \001(\0162&.GetRegionInfoRes" +
|
||||
"ponse.CompactionState\"F\n\017CompactionState" +
|
||||
"\022\010\n\004NONE\020\000\022\t\n\005MINOR\020\001\022\t\n\005MAJOR\020\002\022\023\n\017MAJO" +
|
||||
"R_AND_MINOR\020\003\"G\n\023GetStoreFileRequest\022 \n\006" +
|
||||
"region\030\001 \002(\0132\020.RegionSpecifier\022\016\n\006family",
|
||||
"\030\002 \003(\014\"*\n\024GetStoreFileResponse\022\022\n\nstore_" +
|
||||
"file\030\001 \003(\t\"\030\n\026GetOnlineRegionRequest\";\n\027" +
|
||||
"GetOnlineRegionResponse\022 \n\013region_info\030\001" +
|
||||
" \003(\0132\013.RegionInfo\"\275\001\n\021OpenRegionRequest\022" +
|
||||
"4\n\topen_info\030\001 \003(\0132!.OpenRegionRequest.R" +
|
||||
"egionOpenInfo\032r\n\016RegionOpenInfo\022\033\n\006regio" +
|
||||
"n\030\001 \002(\0132\013.RegionInfo\022\037\n\027version_of_offli" +
|
||||
"ne_node\030\002 \001(\r\022\"\n\rfavored_nodes\030\003 \003(\0132\013.S" +
|
||||
"erverName\"\235\001\n\022OpenRegionResponse\022=\n\ropen" +
|
||||
"ing_state\030\001 \003(\0162&.OpenRegionResponse.Reg",
|
||||
"ionOpeningState\"H\n\022RegionOpeningState\022\n\n" +
|
||||
"\006OPENED\020\000\022\022\n\016ALREADY_OPENED\020\001\022\022\n\016FAILED_" +
|
||||
"OPENING\020\002\"\240\001\n\022CloseRegionRequest\022 \n\006regi" +
|
||||
"on\030\001 \002(\0132\020.RegionSpecifier\022\037\n\027version_of" +
|
||||
"_closing_node\030\002 \001(\r\022\036\n\020transition_in_ZK\030" +
|
||||
"\003 \001(\010:\004true\022\'\n\022destination_server\030\004 \001(\0132" +
|
||||
"\013.ServerName\"%\n\023CloseRegionResponse\022\016\n\006c" +
|
||||
"losed\030\001 \002(\010\"P\n\022FlushRegionRequest\022 \n\006reg" +
|
||||
"ion\030\001 \002(\0132\020.RegionSpecifier\022\030\n\020if_older_" +
|
||||
"than_ts\030\002 \001(\004\"?\n\023FlushRegionResponse\022\027\n\017",
|
||||
"last_flush_time\030\001 \002(\004\022\017\n\007flushed\030\002 \001(\010\"K" +
|
||||
"\n\022SplitRegionRequest\022 \n\006region\030\001 \002(\0132\020.R" +
|
||||
"egionSpecifier\022\023\n\013split_point\030\002 \001(\014\"\025\n\023S" +
|
||||
"plitRegionResponse\"W\n\024CompactRegionReque" +
|
||||
"st\022 \n\006region\030\001 \002(\0132\020.RegionSpecifier\022\r\n\005" +
|
||||
"major\030\002 \001(\010\022\016\n\006family\030\003 \001(\014\"\027\n\025CompactRe" +
|
||||
"gionResponse\"v\n\023MergeRegionsRequest\022\"\n\010r" +
|
||||
"egion_a\030\001 \002(\0132\020.RegionSpecifier\022\"\n\010regio" +
|
||||
"n_b\030\002 \002(\0132\020.RegionSpecifier\022\027\n\010forcible\030" +
|
||||
"\003 \001(\010:\005false\"\026\n\024MergeRegionsResponse\"X\n\010",
|
||||
"WALEntry\022\024\n\003key\030\001 \002(\0132\007.WALKey\022\027\n\017key_va" +
|
||||
"lue_bytes\030\002 \003(\014\022\035\n\025associated_cell_count" +
|
||||
"\030\003 \001(\005\"4\n\030ReplicateWALEntryRequest\022\030\n\005en" +
|
||||
"try\030\001 \003(\0132\t.WALEntry\"\033\n\031ReplicateWALEntr" +
|
||||
"yResponse\"\026\n\024RollWALWriterRequest\"0\n\025Rol" +
|
||||
"lWALWriterResponse\022\027\n\017region_to_flush\030\001 " +
|
||||
"\003(\014\"#\n\021StopServerRequest\022\016\n\006reason\030\001 \002(\t" +
|
||||
"\"\024\n\022StopServerResponse\"\026\n\024GetServerInfoR" +
|
||||
"equest\"B\n\nServerInfo\022 \n\013server_name\030\001 \002(" +
|
||||
"\0132\013.ServerName\022\022\n\nwebui_port\030\002 \001(\r\"9\n\025Ge",
|
||||
"tServerInfoResponse\022 \n\013server_info\030\001 \002(\013" +
|
||||
"2\013.ServerInfo2\337\006\n\014AdminService\022>\n\rGetReg" +
|
||||
"ionInfo\022\025.GetRegionInfoRequest\032\026.GetRegi" +
|
||||
"onInfoResponse\022;\n\014GetStoreFile\022\024.GetStor" +
|
||||
"eFileRequest\032\025.GetStoreFileResponse\022D\n\017G" +
|
||||
"etOnlineRegion\022\027.GetOnlineRegionRequest\032" +
|
||||
"\030.GetOnlineRegionResponse\0225\n\nOpenRegion\022" +
|
||||
"\022.OpenRegionRequest\032\023.OpenRegionResponse" +
|
||||
"\0228\n\013CloseRegion\022\023.CloseRegionRequest\032\024.C" +
|
||||
"loseRegionResponse\0228\n\013FlushRegion\022\023.Flus",
|
||||
"hRegionRequest\032\024.FlushRegionResponse\0228\n\013" +
|
||||
"SplitRegion\022\023.SplitRegionRequest\032\024.Split" +
|
||||
"RegionResponse\022>\n\rCompactRegion\022\025.Compac" +
|
||||
"tRegionRequest\032\026.CompactRegionResponse\022;" +
|
||||
"\n\014MergeRegions\022\024.MergeRegionsRequest\032\025.M" +
|
||||
"ergeRegionsResponse\022J\n\021ReplicateWALEntry" +
|
||||
"\022\031.ReplicateWALEntryRequest\032\032.ReplicateW" +
|
||||
"ALEntryResponse\022\'\n\006Replay\022\r.MultiRequest" +
|
||||
"\032\016.MultiResponse\022>\n\rRollWALWriter\022\025.Roll" +
|
||||
"WALWriterRequest\032\026.RollWALWriterResponse",
|
||||
"\022>\n\rGetServerInfo\022\025.GetServerInfoRequest" +
|
||||
"\032\026.GetServerInfoResponse\0225\n\nStopServer\022\022" +
|
||||
".StopServerRequest\032\023.StopServerResponseB" +
|
||||
"A\n*org.apache.hadoop.hbase.protobuf.gene" +
|
||||
"ratedB\013AdminProtosH\001\210\001\001\240\001\001"
|
||||
};
|
||||
com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
|
||||
new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() {
|
||||
|
|
|
@ -11,7 +11,7 @@ public final class AggregateProtos {
|
|||
public interface AggregateArgumentOrBuilder
|
||||
extends com.google.protobuf.MessageOrBuilder {
|
||||
|
||||
// required string interpreterClassName = 1;
|
||||
// required string interpreter_class_name = 1;
|
||||
boolean hasInterpreterClassName();
|
||||
String getInterpreterClassName();
|
||||
|
||||
|
@ -20,7 +20,7 @@ public final class AggregateProtos {
|
|||
org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Scan getScan();
|
||||
org.apache.hadoop.hbase.protobuf.generated.ClientProtos.ScanOrBuilder getScanOrBuilder();
|
||||
|
||||
// optional bytes interpreterSpecificBytes = 3;
|
||||
// optional bytes interpreter_specific_bytes = 3;
|
||||
boolean hasInterpreterSpecificBytes();
|
||||
com.google.protobuf.ByteString getInterpreterSpecificBytes();
|
||||
}
|
||||
|
@ -53,8 +53,8 @@ public final class AggregateProtos {
|
|||
}
|
||||
|
||||
private int bitField0_;
|
||||
// required string interpreterClassName = 1;
|
||||
public static final int INTERPRETERCLASSNAME_FIELD_NUMBER = 1;
|
||||
// required string interpreter_class_name = 1;
|
||||
public static final int INTERPRETER_CLASS_NAME_FIELD_NUMBER = 1;
|
||||
private java.lang.Object interpreterClassName_;
|
||||
public boolean hasInterpreterClassName() {
|
||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||
|
@ -98,8 +98,8 @@ public final class AggregateProtos {
|
|||
return scan_;
|
||||
}
|
||||
|
||||
// optional bytes interpreterSpecificBytes = 3;
|
||||
public static final int INTERPRETERSPECIFICBYTES_FIELD_NUMBER = 3;
|
||||
// optional bytes interpreter_specific_bytes = 3;
|
||||
public static final int INTERPRETER_SPECIFIC_BYTES_FIELD_NUMBER = 3;
|
||||
private com.google.protobuf.ByteString interpreterSpecificBytes_;
|
||||
public boolean hasInterpreterSpecificBytes() {
|
||||
return ((bitField0_ & 0x00000004) == 0x00000004);
|
||||
|
@ -215,7 +215,7 @@ public final class AggregateProtos {
|
|||
int hash = 41;
|
||||
hash = (19 * hash) + getDescriptorForType().hashCode();
|
||||
if (hasInterpreterClassName()) {
|
||||
hash = (37 * hash) + INTERPRETERCLASSNAME_FIELD_NUMBER;
|
||||
hash = (37 * hash) + INTERPRETER_CLASS_NAME_FIELD_NUMBER;
|
||||
hash = (53 * hash) + getInterpreterClassName().hashCode();
|
||||
}
|
||||
if (hasScan()) {
|
||||
|
@ -223,7 +223,7 @@ public final class AggregateProtos {
|
|||
hash = (53 * hash) + getScan().hashCode();
|
||||
}
|
||||
if (hasInterpreterSpecificBytes()) {
|
||||
hash = (37 * hash) + INTERPRETERSPECIFICBYTES_FIELD_NUMBER;
|
||||
hash = (37 * hash) + INTERPRETER_SPECIFIC_BYTES_FIELD_NUMBER;
|
||||
hash = (53 * hash) + getInterpreterSpecificBytes().hashCode();
|
||||
}
|
||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||
|
@ -500,7 +500,7 @@ public final class AggregateProtos {
|
|||
|
||||
private int bitField0_;
|
||||
|
||||
// required string interpreterClassName = 1;
|
||||
// required string interpreter_class_name = 1;
|
||||
private java.lang.Object interpreterClassName_ = "";
|
||||
public boolean hasInterpreterClassName() {
|
||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||
|
@ -626,7 +626,7 @@ public final class AggregateProtos {
|
|||
return scanBuilder_;
|
||||
}
|
||||
|
||||
// optional bytes interpreterSpecificBytes = 3;
|
||||
// optional bytes interpreter_specific_bytes = 3;
|
||||
private com.google.protobuf.ByteString interpreterSpecificBytes_ = com.google.protobuf.ByteString.EMPTY;
|
||||
public boolean hasInterpreterSpecificBytes() {
|
||||
return ((bitField0_ & 0x00000004) == 0x00000004);
|
||||
|
@ -664,12 +664,12 @@ public final class AggregateProtos {
|
|||
public interface AggregateResponseOrBuilder
|
||||
extends com.google.protobuf.MessageOrBuilder {
|
||||
|
||||
// repeated bytes firstPart = 1;
|
||||
// repeated bytes first_part = 1;
|
||||
java.util.List<com.google.protobuf.ByteString> getFirstPartList();
|
||||
int getFirstPartCount();
|
||||
com.google.protobuf.ByteString getFirstPart(int index);
|
||||
|
||||
// optional bytes secondPart = 2;
|
||||
// optional bytes second_part = 2;
|
||||
boolean hasSecondPart();
|
||||
com.google.protobuf.ByteString getSecondPart();
|
||||
}
|
||||
|
@ -702,8 +702,8 @@ public final class AggregateProtos {
|
|||
}
|
||||
|
||||
private int bitField0_;
|
||||
// repeated bytes firstPart = 1;
|
||||
public static final int FIRSTPART_FIELD_NUMBER = 1;
|
||||
// repeated bytes first_part = 1;
|
||||
public static final int FIRST_PART_FIELD_NUMBER = 1;
|
||||
private java.util.List<com.google.protobuf.ByteString> firstPart_;
|
||||
public java.util.List<com.google.protobuf.ByteString>
|
||||
getFirstPartList() {
|
||||
|
@ -716,8 +716,8 @@ public final class AggregateProtos {
|
|||
return firstPart_.get(index);
|
||||
}
|
||||
|
||||
// optional bytes secondPart = 2;
|
||||
public static final int SECONDPART_FIELD_NUMBER = 2;
|
||||
// optional bytes second_part = 2;
|
||||
public static final int SECOND_PART_FIELD_NUMBER = 2;
|
||||
private com.google.protobuf.ByteString secondPart_;
|
||||
public boolean hasSecondPart() {
|
||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||
|
@ -810,11 +810,11 @@ public final class AggregateProtos {
|
|||
int hash = 41;
|
||||
hash = (19 * hash) + getDescriptorForType().hashCode();
|
||||
if (getFirstPartCount() > 0) {
|
||||
hash = (37 * hash) + FIRSTPART_FIELD_NUMBER;
|
||||
hash = (37 * hash) + FIRST_PART_FIELD_NUMBER;
|
||||
hash = (53 * hash) + getFirstPartList().hashCode();
|
||||
}
|
||||
if (hasSecondPart()) {
|
||||
hash = (37 * hash) + SECONDPART_FIELD_NUMBER;
|
||||
hash = (37 * hash) + SECOND_PART_FIELD_NUMBER;
|
||||
hash = (53 * hash) + getSecondPart().hashCode();
|
||||
}
|
||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||
|
@ -1060,7 +1060,7 @@ public final class AggregateProtos {
|
|||
|
||||
private int bitField0_;
|
||||
|
||||
// repeated bytes firstPart = 1;
|
||||
// repeated bytes first_part = 1;
|
||||
private java.util.List<com.google.protobuf.ByteString> firstPart_ = java.util.Collections.emptyList();;
|
||||
private void ensureFirstPartIsMutable() {
|
||||
if (!((bitField0_ & 0x00000001) == 0x00000001)) {
|
||||
|
@ -1111,7 +1111,7 @@ public final class AggregateProtos {
|
|||
return this;
|
||||
}
|
||||
|
||||
// optional bytes secondPart = 2;
|
||||
// optional bytes second_part = 2;
|
||||
private com.google.protobuf.ByteString secondPart_ = com.google.protobuf.ByteString.EMPTY;
|
||||
public boolean hasSecondPart() {
|
||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||
|
@ -1776,23 +1776,23 @@ public final class AggregateProtos {
|
|||
descriptor;
|
||||
static {
|
||||
java.lang.String[] descriptorData = {
|
||||
"\n\017Aggregate.proto\032\014Client.proto\"h\n\021Aggre" +
|
||||
"gateArgument\022\034\n\024interpreterClassName\030\001 \002" +
|
||||
"(\t\022\023\n\004scan\030\002 \002(\0132\005.Scan\022 \n\030interpreterSp" +
|
||||
"ecificBytes\030\003 \001(\014\":\n\021AggregateResponse\022\021" +
|
||||
"\n\tfirstPart\030\001 \003(\014\022\022\n\nsecondPart\030\002 \001(\0142\366\002" +
|
||||
"\n\020AggregateService\0220\n\006getMax\022\022.Aggregate" +
|
||||
"Argument\032\022.AggregateResponse\0220\n\006getMin\022\022" +
|
||||
".AggregateArgument\032\022.AggregateResponse\0220" +
|
||||
"\n\006getSum\022\022.AggregateArgument\032\022.Aggregate" +
|
||||
"Response\0223\n\tgetRowNum\022\022.AggregateArgumen",
|
||||
"t\032\022.AggregateResponse\0220\n\006getAvg\022\022.Aggreg" +
|
||||
"ateArgument\032\022.AggregateResponse\0220\n\006getSt" +
|
||||
"d\022\022.AggregateArgument\032\022.AggregateRespons" +
|
||||
"e\0223\n\tgetMedian\022\022.AggregateArgument\032\022.Agg" +
|
||||
"regateResponseBE\n*org.apache.hadoop.hbas" +
|
||||
"e.protobuf.generatedB\017AggregateProtosH\001\210" +
|
||||
"\001\001\240\001\001"
|
||||
"\n\017Aggregate.proto\032\014Client.proto\"l\n\021Aggre" +
|
||||
"gateArgument\022\036\n\026interpreter_class_name\030\001" +
|
||||
" \002(\t\022\023\n\004scan\030\002 \002(\0132\005.Scan\022\"\n\032interpreter" +
|
||||
"_specific_bytes\030\003 \001(\014\"<\n\021AggregateRespon" +
|
||||
"se\022\022\n\nfirst_part\030\001 \003(\014\022\023\n\013second_part\030\002 " +
|
||||
"\001(\0142\366\002\n\020AggregateService\0220\n\006GetMax\022\022.Agg" +
|
||||
"regateArgument\032\022.AggregateResponse\0220\n\006Ge" +
|
||||
"tMin\022\022.AggregateArgument\032\022.AggregateResp" +
|
||||
"onse\0220\n\006GetSum\022\022.AggregateArgument\032\022.Agg" +
|
||||
"regateResponse\0223\n\tGetRowNum\022\022.AggregateA",
|
||||
"rgument\032\022.AggregateResponse\0220\n\006GetAvg\022\022." +
|
||||
"AggregateArgument\032\022.AggregateResponse\0220\n" +
|
||||
"\006GetStd\022\022.AggregateArgument\032\022.AggregateR" +
|
||||
"esponse\0223\n\tGetMedian\022\022.AggregateArgument" +
|
||||
"\032\022.AggregateResponseBE\n*org.apache.hadoo" +
|
||||
"p.hbase.protobuf.generatedB\017AggregatePro" +
|
||||
"tosH\001\210\001\001\240\001\001"
|
||||
};
|
||||
com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
|
||||
new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() {
|
||||
|
|
|
@ -15,7 +15,7 @@ public final class AuthenticationProtos {
|
|||
boolean hasId();
|
||||
int getId();
|
||||
|
||||
// required int64 expirationDate = 2;
|
||||
// required int64 expiration_date = 2;
|
||||
boolean hasExpirationDate();
|
||||
long getExpirationDate();
|
||||
|
||||
|
@ -62,8 +62,8 @@ public final class AuthenticationProtos {
|
|||
return id_;
|
||||
}
|
||||
|
||||
// required int64 expirationDate = 2;
|
||||
public static final int EXPIRATIONDATE_FIELD_NUMBER = 2;
|
||||
// required int64 expiration_date = 2;
|
||||
public static final int EXPIRATION_DATE_FIELD_NUMBER = 2;
|
||||
private long expirationDate_;
|
||||
public boolean hasExpirationDate() {
|
||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||
|
@ -193,7 +193,7 @@ public final class AuthenticationProtos {
|
|||
hash = (53 * hash) + getId();
|
||||
}
|
||||
if (hasExpirationDate()) {
|
||||
hash = (37 * hash) + EXPIRATIONDATE_FIELD_NUMBER;
|
||||
hash = (37 * hash) + EXPIRATION_DATE_FIELD_NUMBER;
|
||||
hash = (53 * hash) + hashLong(getExpirationDate());
|
||||
}
|
||||
if (hasKey()) {
|
||||
|
@ -482,7 +482,7 @@ public final class AuthenticationProtos {
|
|||
return this;
|
||||
}
|
||||
|
||||
// required int64 expirationDate = 2;
|
||||
// required int64 expiration_date = 2;
|
||||
private long expirationDate_ ;
|
||||
public boolean hasExpirationDate() {
|
||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||
|
@ -549,19 +549,19 @@ public final class AuthenticationProtos {
|
|||
boolean hasUsername();
|
||||
com.google.protobuf.ByteString getUsername();
|
||||
|
||||
// required int32 keyId = 3;
|
||||
// required int32 key_id = 3;
|
||||
boolean hasKeyId();
|
||||
int getKeyId();
|
||||
|
||||
// optional int64 issueDate = 4;
|
||||
// optional int64 issue_date = 4;
|
||||
boolean hasIssueDate();
|
||||
long getIssueDate();
|
||||
|
||||
// optional int64 expirationDate = 5;
|
||||
// optional int64 expiration_date = 5;
|
||||
boolean hasExpirationDate();
|
||||
long getExpirationDate();
|
||||
|
||||
// optional int64 sequenceNumber = 6;
|
||||
// optional int64 sequence_number = 6;
|
||||
boolean hasSequenceNumber();
|
||||
long getSequenceNumber();
|
||||
}
|
||||
|
@ -680,8 +680,8 @@ public final class AuthenticationProtos {
|
|||
return username_;
|
||||
}
|
||||
|
||||
// required int32 keyId = 3;
|
||||
public static final int KEYID_FIELD_NUMBER = 3;
|
||||
// required int32 key_id = 3;
|
||||
public static final int KEY_ID_FIELD_NUMBER = 3;
|
||||
private int keyId_;
|
||||
public boolean hasKeyId() {
|
||||
return ((bitField0_ & 0x00000004) == 0x00000004);
|
||||
|
@ -690,8 +690,8 @@ public final class AuthenticationProtos {
|
|||
return keyId_;
|
||||
}
|
||||
|
||||
// optional int64 issueDate = 4;
|
||||
public static final int ISSUEDATE_FIELD_NUMBER = 4;
|
||||
// optional int64 issue_date = 4;
|
||||
public static final int ISSUE_DATE_FIELD_NUMBER = 4;
|
||||
private long issueDate_;
|
||||
public boolean hasIssueDate() {
|
||||
return ((bitField0_ & 0x00000008) == 0x00000008);
|
||||
|
@ -700,8 +700,8 @@ public final class AuthenticationProtos {
|
|||
return issueDate_;
|
||||
}
|
||||
|
||||
// optional int64 expirationDate = 5;
|
||||
public static final int EXPIRATIONDATE_FIELD_NUMBER = 5;
|
||||
// optional int64 expiration_date = 5;
|
||||
public static final int EXPIRATION_DATE_FIELD_NUMBER = 5;
|
||||
private long expirationDate_;
|
||||
public boolean hasExpirationDate() {
|
||||
return ((bitField0_ & 0x00000010) == 0x00000010);
|
||||
|
@ -710,8 +710,8 @@ public final class AuthenticationProtos {
|
|||
return expirationDate_;
|
||||
}
|
||||
|
||||
// optional int64 sequenceNumber = 6;
|
||||
public static final int SEQUENCENUMBER_FIELD_NUMBER = 6;
|
||||
// optional int64 sequence_number = 6;
|
||||
public static final int SEQUENCE_NUMBER_FIELD_NUMBER = 6;
|
||||
private long sequenceNumber_;
|
||||
public boolean hasSequenceNumber() {
|
||||
return ((bitField0_ & 0x00000020) == 0x00000020);
|
||||
|
@ -874,19 +874,19 @@ public final class AuthenticationProtos {
|
|||
hash = (53 * hash) + getUsername().hashCode();
|
||||
}
|
||||
if (hasKeyId()) {
|
||||
hash = (37 * hash) + KEYID_FIELD_NUMBER;
|
||||
hash = (37 * hash) + KEY_ID_FIELD_NUMBER;
|
||||
hash = (53 * hash) + getKeyId();
|
||||
}
|
||||
if (hasIssueDate()) {
|
||||
hash = (37 * hash) + ISSUEDATE_FIELD_NUMBER;
|
||||
hash = (37 * hash) + ISSUE_DATE_FIELD_NUMBER;
|
||||
hash = (53 * hash) + hashLong(getIssueDate());
|
||||
}
|
||||
if (hasExpirationDate()) {
|
||||
hash = (37 * hash) + EXPIRATIONDATE_FIELD_NUMBER;
|
||||
hash = (37 * hash) + EXPIRATION_DATE_FIELD_NUMBER;
|
||||
hash = (53 * hash) + hashLong(getExpirationDate());
|
||||
}
|
||||
if (hasSequenceNumber()) {
|
||||
hash = (37 * hash) + SEQUENCENUMBER_FIELD_NUMBER;
|
||||
hash = (37 * hash) + SEQUENCE_NUMBER_FIELD_NUMBER;
|
||||
hash = (53 * hash) + hashLong(getSequenceNumber());
|
||||
}
|
||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||
|
@ -1246,7 +1246,7 @@ public final class AuthenticationProtos {
|
|||
return this;
|
||||
}
|
||||
|
||||
// required int32 keyId = 3;
|
||||
// required int32 key_id = 3;
|
||||
private int keyId_ ;
|
||||
public boolean hasKeyId() {
|
||||
return ((bitField0_ & 0x00000004) == 0x00000004);
|
||||
|
@ -1267,7 +1267,7 @@ public final class AuthenticationProtos {
|
|||
return this;
|
||||
}
|
||||
|
||||
// optional int64 issueDate = 4;
|
||||
// optional int64 issue_date = 4;
|
||||
private long issueDate_ ;
|
||||
public boolean hasIssueDate() {
|
||||
return ((bitField0_ & 0x00000008) == 0x00000008);
|
||||
|
@ -1288,7 +1288,7 @@ public final class AuthenticationProtos {
|
|||
return this;
|
||||
}
|
||||
|
||||
// optional int64 expirationDate = 5;
|
||||
// optional int64 expiration_date = 5;
|
||||
private long expirationDate_ ;
|
||||
public boolean hasExpirationDate() {
|
||||
return ((bitField0_ & 0x00000010) == 0x00000010);
|
||||
|
@ -1309,7 +1309,7 @@ public final class AuthenticationProtos {
|
|||
return this;
|
||||
}
|
||||
|
||||
// optional int64 sequenceNumber = 6;
|
||||
// optional int64 sequence_number = 6;
|
||||
private long sequenceNumber_ ;
|
||||
public boolean hasSequenceNumber() {
|
||||
return ((bitField0_ & 0x00000020) == 0x00000020);
|
||||
|
@ -2917,7 +2917,7 @@ public final class AuthenticationProtos {
|
|||
boolean hasUsername();
|
||||
String getUsername();
|
||||
|
||||
// optional string authMethod = 2;
|
||||
// optional string auth_method = 2;
|
||||
boolean hasAuthMethod();
|
||||
String getAuthMethod();
|
||||
}
|
||||
|
@ -2982,8 +2982,8 @@ public final class AuthenticationProtos {
|
|||
}
|
||||
}
|
||||
|
||||
// optional string authMethod = 2;
|
||||
public static final int AUTHMETHOD_FIELD_NUMBER = 2;
|
||||
// optional string auth_method = 2;
|
||||
public static final int AUTH_METHOD_FIELD_NUMBER = 2;
|
||||
private java.lang.Object authMethod_;
|
||||
public boolean hasAuthMethod() {
|
||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||
|
@ -3100,7 +3100,7 @@ public final class AuthenticationProtos {
|
|||
hash = (53 * hash) + getUsername().hashCode();
|
||||
}
|
||||
if (hasAuthMethod()) {
|
||||
hash = (37 * hash) + AUTHMETHOD_FIELD_NUMBER;
|
||||
hash = (37 * hash) + AUTH_METHOD_FIELD_NUMBER;
|
||||
hash = (53 * hash) + getAuthMethod().hashCode();
|
||||
}
|
||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||
|
@ -3374,7 +3374,7 @@ public final class AuthenticationProtos {
|
|||
onChanged();
|
||||
}
|
||||
|
||||
// optional string authMethod = 2;
|
||||
// optional string auth_method = 2;
|
||||
private java.lang.Object authMethod_ = "";
|
||||
public boolean hasAuthMethod() {
|
||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||
|
@ -3431,7 +3431,7 @@ public final class AuthenticationProtos {
|
|||
org.apache.hadoop.hbase.protobuf.generated.AuthenticationProtos.TokenRequest request,
|
||||
com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.protobuf.generated.AuthenticationProtos.TokenResponse> done);
|
||||
|
||||
public abstract void whoami(
|
||||
public abstract void whoAmI(
|
||||
com.google.protobuf.RpcController controller,
|
||||
org.apache.hadoop.hbase.protobuf.generated.AuthenticationProtos.WhoAmIRequest request,
|
||||
com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.protobuf.generated.AuthenticationProtos.WhoAmIResponse> done);
|
||||
|
@ -3450,11 +3450,11 @@ public final class AuthenticationProtos {
|
|||
}
|
||||
|
||||
@java.lang.Override
|
||||
public void whoami(
|
||||
public void whoAmI(
|
||||
com.google.protobuf.RpcController controller,
|
||||
org.apache.hadoop.hbase.protobuf.generated.AuthenticationProtos.WhoAmIRequest request,
|
||||
com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.protobuf.generated.AuthenticationProtos.WhoAmIResponse> done) {
|
||||
impl.whoami(controller, request, done);
|
||||
impl.whoAmI(controller, request, done);
|
||||
}
|
||||
|
||||
};
|
||||
|
@ -3482,7 +3482,7 @@ public final class AuthenticationProtos {
|
|||
case 0:
|
||||
return impl.getAuthenticationToken(controller, (org.apache.hadoop.hbase.protobuf.generated.AuthenticationProtos.TokenRequest)request);
|
||||
case 1:
|
||||
return impl.whoami(controller, (org.apache.hadoop.hbase.protobuf.generated.AuthenticationProtos.WhoAmIRequest)request);
|
||||
return impl.whoAmI(controller, (org.apache.hadoop.hbase.protobuf.generated.AuthenticationProtos.WhoAmIRequest)request);
|
||||
default:
|
||||
throw new java.lang.AssertionError("Can't get here.");
|
||||
}
|
||||
|
@ -3532,7 +3532,7 @@ public final class AuthenticationProtos {
|
|||
org.apache.hadoop.hbase.protobuf.generated.AuthenticationProtos.TokenRequest request,
|
||||
com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.protobuf.generated.AuthenticationProtos.TokenResponse> done);
|
||||
|
||||
public abstract void whoami(
|
||||
public abstract void whoAmI(
|
||||
com.google.protobuf.RpcController controller,
|
||||
org.apache.hadoop.hbase.protobuf.generated.AuthenticationProtos.WhoAmIRequest request,
|
||||
com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.protobuf.generated.AuthenticationProtos.WhoAmIResponse> done);
|
||||
|
@ -3565,7 +3565,7 @@ public final class AuthenticationProtos {
|
|||
done));
|
||||
return;
|
||||
case 1:
|
||||
this.whoami(controller, (org.apache.hadoop.hbase.protobuf.generated.AuthenticationProtos.WhoAmIRequest)request,
|
||||
this.whoAmI(controller, (org.apache.hadoop.hbase.protobuf.generated.AuthenticationProtos.WhoAmIRequest)request,
|
||||
com.google.protobuf.RpcUtil.<org.apache.hadoop.hbase.protobuf.generated.AuthenticationProtos.WhoAmIResponse>specializeCallback(
|
||||
done));
|
||||
return;
|
||||
|
@ -3641,7 +3641,7 @@ public final class AuthenticationProtos {
|
|||
org.apache.hadoop.hbase.protobuf.generated.AuthenticationProtos.TokenResponse.getDefaultInstance()));
|
||||
}
|
||||
|
||||
public void whoami(
|
||||
public void whoAmI(
|
||||
com.google.protobuf.RpcController controller,
|
||||
org.apache.hadoop.hbase.protobuf.generated.AuthenticationProtos.WhoAmIRequest request,
|
||||
com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.protobuf.generated.AuthenticationProtos.WhoAmIResponse> done) {
|
||||
|
@ -3668,7 +3668,7 @@ public final class AuthenticationProtos {
|
|||
org.apache.hadoop.hbase.protobuf.generated.AuthenticationProtos.TokenRequest request)
|
||||
throws com.google.protobuf.ServiceException;
|
||||
|
||||
public org.apache.hadoop.hbase.protobuf.generated.AuthenticationProtos.WhoAmIResponse whoami(
|
||||
public org.apache.hadoop.hbase.protobuf.generated.AuthenticationProtos.WhoAmIResponse whoAmI(
|
||||
com.google.protobuf.RpcController controller,
|
||||
org.apache.hadoop.hbase.protobuf.generated.AuthenticationProtos.WhoAmIRequest request)
|
||||
throws com.google.protobuf.ServiceException;
|
||||
|
@ -3693,7 +3693,7 @@ public final class AuthenticationProtos {
|
|||
}
|
||||
|
||||
|
||||
public org.apache.hadoop.hbase.protobuf.generated.AuthenticationProtos.WhoAmIResponse whoami(
|
||||
public org.apache.hadoop.hbase.protobuf.generated.AuthenticationProtos.WhoAmIResponse whoAmI(
|
||||
com.google.protobuf.RpcController controller,
|
||||
org.apache.hadoop.hbase.protobuf.generated.AuthenticationProtos.WhoAmIRequest request)
|
||||
throws com.google.protobuf.ServiceException {
|
||||
|
@ -3751,23 +3751,23 @@ public final class AuthenticationProtos {
|
|||
descriptor;
|
||||
static {
|
||||
java.lang.String[] descriptorData = {
|
||||
"\n\024Authentication.proto\"D\n\021Authentication" +
|
||||
"Key\022\n\n\002id\030\001 \002(\005\022\026\n\016expirationDate\030\002 \002(\003\022" +
|
||||
"\013\n\003key\030\003 \002(\014\"\270\001\n\017TokenIdentifier\022#\n\004kind" +
|
||||
"\030\001 \002(\0162\025.TokenIdentifier.Kind\022\020\n\010usernam" +
|
||||
"e\030\002 \002(\014\022\r\n\005keyId\030\003 \002(\005\022\021\n\tissueDate\030\004 \001(" +
|
||||
"\003\022\026\n\016expirationDate\030\005 \001(\003\022\026\n\016sequenceNum" +
|
||||
"ber\030\006 \001(\003\"\034\n\004Kind\022\024\n\020HBASE_AUTH_TOKEN\020\000\"" +
|
||||
">\n\005Token\022\022\n\nidentifier\030\001 \001(\014\022\020\n\010password" +
|
||||
"\030\002 \001(\014\022\017\n\007service\030\003 \001(\014\"\016\n\014TokenRequest\"" +
|
||||
"&\n\rTokenResponse\022\025\n\005token\030\001 \001(\0132\006.Token\"",
|
||||
"\017\n\rWhoAmIRequest\"6\n\016WhoAmIResponse\022\020\n\010us" +
|
||||
"ername\030\001 \001(\t\022\022\n\nauthMethod\030\002 \001(\t2{\n\025Auth" +
|
||||
"enticationService\0227\n\026getAuthenticationTo" +
|
||||
"ken\022\r.TokenRequest\032\016.TokenResponse\022)\n\006wh" +
|
||||
"oami\022\016.WhoAmIRequest\032\017.WhoAmIResponseBJ\n" +
|
||||
"*org.apache.hadoop.hbase.protobuf.genera" +
|
||||
"tedB\024AuthenticationProtosH\001\210\001\001\240\001\001"
|
||||
"\n\024Authentication.proto\"E\n\021Authentication" +
|
||||
"Key\022\n\n\002id\030\001 \002(\005\022\027\n\017expiration_date\030\002 \002(\003" +
|
||||
"\022\013\n\003key\030\003 \002(\014\"\274\001\n\017TokenIdentifier\022#\n\004kin" +
|
||||
"d\030\001 \002(\0162\025.TokenIdentifier.Kind\022\020\n\010userna" +
|
||||
"me\030\002 \002(\014\022\016\n\006key_id\030\003 \002(\005\022\022\n\nissue_date\030\004" +
|
||||
" \001(\003\022\027\n\017expiration_date\030\005 \001(\003\022\027\n\017sequenc" +
|
||||
"e_number\030\006 \001(\003\"\034\n\004Kind\022\024\n\020HBASE_AUTH_TOK" +
|
||||
"EN\020\000\">\n\005Token\022\022\n\nidentifier\030\001 \001(\014\022\020\n\010pas" +
|
||||
"sword\030\002 \001(\014\022\017\n\007service\030\003 \001(\014\"\016\n\014TokenReq" +
|
||||
"uest\"&\n\rTokenResponse\022\025\n\005token\030\001 \001(\0132\006.T",
|
||||
"oken\"\017\n\rWhoAmIRequest\"7\n\016WhoAmIResponse\022" +
|
||||
"\020\n\010username\030\001 \001(\t\022\023\n\013auth_method\030\002 \001(\t2{" +
|
||||
"\n\025AuthenticationService\0227\n\026GetAuthentica" +
|
||||
"tionToken\022\r.TokenRequest\032\016.TokenResponse" +
|
||||
"\022)\n\006WhoAmI\022\016.WhoAmIRequest\032\017.WhoAmIRespo" +
|
||||
"nseBJ\n*org.apache.hadoop.hbase.protobuf." +
|
||||
"generatedB\024AuthenticationProtosH\001\210\001\001\240\001\001"
|
||||
};
|
||||
com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
|
||||
new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() {
|
||||
|
|
|
@ -108,7 +108,7 @@ public final class CellProtos {
|
|||
boolean hasTimestamp();
|
||||
long getTimestamp();
|
||||
|
||||
// optional .CellType cellType = 5;
|
||||
// optional .CellType cell_type = 5;
|
||||
boolean hasCellType();
|
||||
org.apache.hadoop.hbase.protobuf.generated.CellProtos.CellType getCellType();
|
||||
|
||||
|
@ -185,8 +185,8 @@ public final class CellProtos {
|
|||
return timestamp_;
|
||||
}
|
||||
|
||||
// optional .CellType cellType = 5;
|
||||
public static final int CELLTYPE_FIELD_NUMBER = 5;
|
||||
// optional .CellType cell_type = 5;
|
||||
public static final int CELL_TYPE_FIELD_NUMBER = 5;
|
||||
private org.apache.hadoop.hbase.protobuf.generated.CellProtos.CellType cellType_;
|
||||
public boolean hasCellType() {
|
||||
return ((bitField0_ & 0x00000010) == 0x00000010);
|
||||
|
@ -355,7 +355,7 @@ public final class CellProtos {
|
|||
hash = (53 * hash) + hashLong(getTimestamp());
|
||||
}
|
||||
if (hasCellType()) {
|
||||
hash = (37 * hash) + CELLTYPE_FIELD_NUMBER;
|
||||
hash = (37 * hash) + CELL_TYPE_FIELD_NUMBER;
|
||||
hash = (53 * hash) + hashEnum(getCellType());
|
||||
}
|
||||
if (hasValue()) {
|
||||
|
@ -752,7 +752,7 @@ public final class CellProtos {
|
|||
return this;
|
||||
}
|
||||
|
||||
// optional .CellType cellType = 5;
|
||||
// optional .CellType cell_type = 5;
|
||||
private org.apache.hadoop.hbase.protobuf.generated.CellProtos.CellType cellType_ = org.apache.hadoop.hbase.protobuf.generated.CellProtos.CellType.MINIMUM;
|
||||
public boolean hasCellType() {
|
||||
return ((bitField0_ & 0x00000010) == 0x00000010);
|
||||
|
@ -830,7 +830,7 @@ public final class CellProtos {
|
|||
boolean hasTimestamp();
|
||||
long getTimestamp();
|
||||
|
||||
// optional .CellType keyType = 5;
|
||||
// optional .CellType key_type = 5;
|
||||
boolean hasKeyType();
|
||||
org.apache.hadoop.hbase.protobuf.generated.CellProtos.CellType getKeyType();
|
||||
|
||||
|
@ -907,8 +907,8 @@ public final class CellProtos {
|
|||
return timestamp_;
|
||||
}
|
||||
|
||||
// optional .CellType keyType = 5;
|
||||
public static final int KEYTYPE_FIELD_NUMBER = 5;
|
||||
// optional .CellType key_type = 5;
|
||||
public static final int KEY_TYPE_FIELD_NUMBER = 5;
|
||||
private org.apache.hadoop.hbase.protobuf.generated.CellProtos.CellType keyType_;
|
||||
public boolean hasKeyType() {
|
||||
return ((bitField0_ & 0x00000010) == 0x00000010);
|
||||
|
@ -1089,7 +1089,7 @@ public final class CellProtos {
|
|||
hash = (53 * hash) + hashLong(getTimestamp());
|
||||
}
|
||||
if (hasKeyType()) {
|
||||
hash = (37 * hash) + KEYTYPE_FIELD_NUMBER;
|
||||
hash = (37 * hash) + KEY_TYPE_FIELD_NUMBER;
|
||||
hash = (53 * hash) + hashEnum(getKeyType());
|
||||
}
|
||||
if (hasValue()) {
|
||||
|
@ -1498,7 +1498,7 @@ public final class CellProtos {
|
|||
return this;
|
||||
}
|
||||
|
||||
// optional .CellType keyType = 5;
|
||||
// optional .CellType key_type = 5;
|
||||
private org.apache.hadoop.hbase.protobuf.generated.CellProtos.CellType keyType_ = org.apache.hadoop.hbase.protobuf.generated.CellProtos.CellType.MINIMUM;
|
||||
public boolean hasKeyType() {
|
||||
return ((bitField0_ & 0x00000010) == 0x00000010);
|
||||
|
@ -1576,17 +1576,17 @@ public final class CellProtos {
|
|||
descriptor;
|
||||
static {
|
||||
java.lang.String[] descriptorData = {
|
||||
"\n\nCell.proto\"u\n\004Cell\022\013\n\003row\030\001 \001(\014\022\016\n\006fam" +
|
||||
"\n\nCell.proto\"v\n\004Cell\022\013\n\003row\030\001 \001(\014\022\016\n\006fam" +
|
||||
"ily\030\002 \001(\014\022\021\n\tqualifier\030\003 \001(\014\022\021\n\ttimestam" +
|
||||
"p\030\004 \001(\004\022\033\n\010cellType\030\005 \001(\0162\t.CellType\022\r\n\005" +
|
||||
"value\030\006 \001(\014\"x\n\010KeyValue\022\013\n\003row\030\001 \002(\014\022\016\n\006" +
|
||||
"family\030\002 \002(\014\022\021\n\tqualifier\030\003 \002(\014\022\021\n\ttimes" +
|
||||
"tamp\030\004 \001(\004\022\032\n\007keyType\030\005 \001(\0162\t.CellType\022\r" +
|
||||
"\n\005value\030\006 \001(\014*`\n\010CellType\022\013\n\007MINIMUM\020\000\022\007" +
|
||||
"\n\003PUT\020\004\022\n\n\006DELETE\020\010\022\021\n\rDELETE_COLUMN\020\014\022\021" +
|
||||
"\n\rDELETE_FAMILY\020\016\022\014\n\007MAXIMUM\020\377\001B=\n*org.a" +
|
||||
"pache.hadoop.hbase.protobuf.generatedB\nC",
|
||||
"ellProtosH\001\240\001\001"
|
||||
"p\030\004 \001(\004\022\034\n\tcell_type\030\005 \001(\0162\t.CellType\022\r\n" +
|
||||
"\005value\030\006 \001(\014\"y\n\010KeyValue\022\013\n\003row\030\001 \002(\014\022\016\n" +
|
||||
"\006family\030\002 \002(\014\022\021\n\tqualifier\030\003 \002(\014\022\021\n\ttime" +
|
||||
"stamp\030\004 \001(\004\022\033\n\010key_type\030\005 \001(\0162\t.CellType" +
|
||||
"\022\r\n\005value\030\006 \001(\014*`\n\010CellType\022\013\n\007MINIMUM\020\000" +
|
||||
"\022\007\n\003PUT\020\004\022\n\n\006DELETE\020\010\022\021\n\rDELETE_COLUMN\020\014" +
|
||||
"\022\021\n\rDELETE_FAMILY\020\016\022\014\n\007MAXIMUM\020\377\001B=\n*org" +
|
||||
".apache.hadoop.hbase.protobuf.generatedB",
|
||||
"\nCellProtosH\001\240\001\001"
|
||||
};
|
||||
com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
|
||||
new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() {
|
||||
|
|
File diff suppressed because it is too large
Load Diff
|
@ -11,7 +11,7 @@ public final class ClusterIdProtos {
|
|||
public interface ClusterIdOrBuilder
|
||||
extends com.google.protobuf.MessageOrBuilder {
|
||||
|
||||
// required string clusterId = 1;
|
||||
// required string cluster_id = 1;
|
||||
boolean hasClusterId();
|
||||
String getClusterId();
|
||||
}
|
||||
|
@ -44,8 +44,8 @@ public final class ClusterIdProtos {
|
|||
}
|
||||
|
||||
private int bitField0_;
|
||||
// required string clusterId = 1;
|
||||
public static final int CLUSTERID_FIELD_NUMBER = 1;
|
||||
// required string cluster_id = 1;
|
||||
public static final int CLUSTER_ID_FIELD_NUMBER = 1;
|
||||
private java.lang.Object clusterId_;
|
||||
public boolean hasClusterId() {
|
||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||
|
@ -149,7 +149,7 @@ public final class ClusterIdProtos {
|
|||
int hash = 41;
|
||||
hash = (19 * hash) + getDescriptorForType().hashCode();
|
||||
if (hasClusterId()) {
|
||||
hash = (37 * hash) + CLUSTERID_FIELD_NUMBER;
|
||||
hash = (37 * hash) + CLUSTER_ID_FIELD_NUMBER;
|
||||
hash = (53 * hash) + getClusterId().hashCode();
|
||||
}
|
||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||
|
@ -377,7 +377,7 @@ public final class ClusterIdProtos {
|
|||
|
||||
private int bitField0_;
|
||||
|
||||
// required string clusterId = 1;
|
||||
// required string cluster_id = 1;
|
||||
private java.lang.Object clusterId_ = "";
|
||||
public boolean hasClusterId() {
|
||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||
|
@ -438,9 +438,9 @@ public final class ClusterIdProtos {
|
|||
descriptor;
|
||||
static {
|
||||
java.lang.String[] descriptorData = {
|
||||
"\n\017ClusterId.proto\"\036\n\tClusterId\022\021\n\tcluste" +
|
||||
"rId\030\001 \002(\tBB\n*org.apache.hadoop.hbase.pro" +
|
||||
"tobuf.generatedB\017ClusterIdProtosH\001\240\001\001"
|
||||
"\n\017ClusterId.proto\"\037\n\tClusterId\022\022\n\ncluste" +
|
||||
"r_id\030\001 \002(\tBB\n*org.apache.hadoop.hbase.pr" +
|
||||
"otobuf.generatedB\017ClusterIdProtosH\001\240\001\001"
|
||||
};
|
||||
com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
|
||||
new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() {
|
||||
|
|
File diff suppressed because it is too large
Load Diff
|
@ -15,7 +15,7 @@ public final class ComparatorProtos {
|
|||
boolean hasName();
|
||||
String getName();
|
||||
|
||||
// optional bytes serializedComparator = 2;
|
||||
// optional bytes serialized_comparator = 2;
|
||||
boolean hasSerializedComparator();
|
||||
com.google.protobuf.ByteString getSerializedComparator();
|
||||
}
|
||||
|
@ -80,8 +80,8 @@ public final class ComparatorProtos {
|
|||
}
|
||||
}
|
||||
|
||||
// optional bytes serializedComparator = 2;
|
||||
public static final int SERIALIZEDCOMPARATOR_FIELD_NUMBER = 2;
|
||||
// optional bytes serialized_comparator = 2;
|
||||
public static final int SERIALIZED_COMPARATOR_FIELD_NUMBER = 2;
|
||||
private com.google.protobuf.ByteString serializedComparator_;
|
||||
public boolean hasSerializedComparator() {
|
||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||
|
@ -180,7 +180,7 @@ public final class ComparatorProtos {
|
|||
hash = (53 * hash) + getName().hashCode();
|
||||
}
|
||||
if (hasSerializedComparator()) {
|
||||
hash = (37 * hash) + SERIALIZEDCOMPARATOR_FIELD_NUMBER;
|
||||
hash = (37 * hash) + SERIALIZED_COMPARATOR_FIELD_NUMBER;
|
||||
hash = (53 * hash) + getSerializedComparator().hashCode();
|
||||
}
|
||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||
|
@ -458,7 +458,7 @@ public final class ComparatorProtos {
|
|||
onChanged();
|
||||
}
|
||||
|
||||
// optional bytes serializedComparator = 2;
|
||||
// optional bytes serialized_comparator = 2;
|
||||
private com.google.protobuf.ByteString serializedComparator_ = com.google.protobuf.ByteString.EMPTY;
|
||||
public boolean hasSerializedComparator() {
|
||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||
|
@ -1805,7 +1805,7 @@ public final class ComparatorProtos {
|
|||
org.apache.hadoop.hbase.protobuf.generated.ComparatorProtos.ByteArrayComparable getComparable();
|
||||
org.apache.hadoop.hbase.protobuf.generated.ComparatorProtos.ByteArrayComparableOrBuilder getComparableOrBuilder();
|
||||
|
||||
// required .BitComparator.BitwiseOp bitwiseOp = 2;
|
||||
// required .BitComparator.BitwiseOp bitwise_op = 2;
|
||||
boolean hasBitwiseOp();
|
||||
org.apache.hadoop.hbase.protobuf.generated.ComparatorProtos.BitComparator.BitwiseOp getBitwiseOp();
|
||||
}
|
||||
|
@ -1923,8 +1923,8 @@ public final class ComparatorProtos {
|
|||
return comparable_;
|
||||
}
|
||||
|
||||
// required .BitComparator.BitwiseOp bitwiseOp = 2;
|
||||
public static final int BITWISEOP_FIELD_NUMBER = 2;
|
||||
// required .BitComparator.BitwiseOp bitwise_op = 2;
|
||||
public static final int BITWISE_OP_FIELD_NUMBER = 2;
|
||||
private org.apache.hadoop.hbase.protobuf.generated.ComparatorProtos.BitComparator.BitwiseOp bitwiseOp_;
|
||||
public boolean hasBitwiseOp() {
|
||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||
|
@ -2027,7 +2027,7 @@ public final class ComparatorProtos {
|
|||
hash = (53 * hash) + getComparable().hashCode();
|
||||
}
|
||||
if (hasBitwiseOp()) {
|
||||
hash = (37 * hash) + BITWISEOP_FIELD_NUMBER;
|
||||
hash = (37 * hash) + BITWISE_OP_FIELD_NUMBER;
|
||||
hash = (53 * hash) + hashEnum(getBitwiseOp());
|
||||
}
|
||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||
|
@ -2382,7 +2382,7 @@ public final class ComparatorProtos {
|
|||
return comparableBuilder_;
|
||||
}
|
||||
|
||||
// required .BitComparator.BitwiseOp bitwiseOp = 2;
|
||||
// required .BitComparator.BitwiseOp bitwise_op = 2;
|
||||
private org.apache.hadoop.hbase.protobuf.generated.ComparatorProtos.BitComparator.BitwiseOp bitwiseOp_ = org.apache.hadoop.hbase.protobuf.generated.ComparatorProtos.BitComparator.BitwiseOp.AND;
|
||||
public boolean hasBitwiseOp() {
|
||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||
|
@ -2724,7 +2724,7 @@ public final class ComparatorProtos {
|
|||
boolean hasPattern();
|
||||
String getPattern();
|
||||
|
||||
// required int32 patternFlags = 2;
|
||||
// required int32 pattern_flags = 2;
|
||||
boolean hasPatternFlags();
|
||||
int getPatternFlags();
|
||||
|
||||
|
@ -2793,8 +2793,8 @@ public final class ComparatorProtos {
|
|||
}
|
||||
}
|
||||
|
||||
// required int32 patternFlags = 2;
|
||||
public static final int PATTERNFLAGS_FIELD_NUMBER = 2;
|
||||
// required int32 pattern_flags = 2;
|
||||
public static final int PATTERN_FLAGS_FIELD_NUMBER = 2;
|
||||
private int patternFlags_;
|
||||
public boolean hasPatternFlags() {
|
||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||
|
@ -2946,7 +2946,7 @@ public final class ComparatorProtos {
|
|||
hash = (53 * hash) + getPattern().hashCode();
|
||||
}
|
||||
if (hasPatternFlags()) {
|
||||
hash = (37 * hash) + PATTERNFLAGS_FIELD_NUMBER;
|
||||
hash = (37 * hash) + PATTERN_FLAGS_FIELD_NUMBER;
|
||||
hash = (53 * hash) + getPatternFlags();
|
||||
}
|
||||
if (hasCharset()) {
|
||||
|
@ -3250,7 +3250,7 @@ public final class ComparatorProtos {
|
|||
onChanged();
|
||||
}
|
||||
|
||||
// required int32 patternFlags = 2;
|
||||
// required int32 pattern_flags = 2;
|
||||
private int patternFlags_ ;
|
||||
public boolean hasPatternFlags() {
|
||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||
|
@ -3783,21 +3783,22 @@ public final class ComparatorProtos {
|
|||
descriptor;
|
||||
static {
|
||||
java.lang.String[] descriptorData = {
|
||||
"\n\020Comparator.proto\"8\n\nComparator\022\014\n\004name" +
|
||||
"\030\001 \002(\t\022\034\n\024serializedComparator\030\002 \001(\014\"$\n\023" +
|
||||
"ByteArrayComparable\022\r\n\005value\030\001 \001(\014\"<\n\020Bi" +
|
||||
"naryComparator\022(\n\ncomparable\030\001 \002(\0132\024.Byt" +
|
||||
"eArrayComparable\"B\n\026BinaryPrefixComparat" +
|
||||
"or\022(\n\ncomparable\030\001 \002(\0132\024.ByteArrayCompar" +
|
||||
"able\"\215\001\n\rBitComparator\022(\n\ncomparable\030\001 \002" +
|
||||
"(\0132\024.ByteArrayComparable\022+\n\tbitwiseOp\030\002 " +
|
||||
"\002(\0162\030.BitComparator.BitwiseOp\"%\n\tBitwise" +
|
||||
"Op\022\007\n\003AND\020\001\022\006\n\002OR\020\002\022\007\n\003XOR\020\003\"\020\n\016NullComp",
|
||||
"arator\"O\n\025RegexStringComparator\022\017\n\007patte" +
|
||||
"rn\030\001 \002(\t\022\024\n\014patternFlags\030\002 \002(\005\022\017\n\007charse" +
|
||||
"t\030\003 \002(\t\"%\n\023SubstringComparator\022\016\n\006substr" +
|
||||
"\030\001 \002(\tBF\n*org.apache.hadoop.hbase.protob" +
|
||||
"uf.generatedB\020ComparatorProtosH\001\210\001\001\240\001\001"
|
||||
"\n\020Comparator.proto\"9\n\nComparator\022\014\n\004name" +
|
||||
"\030\001 \002(\t\022\035\n\025serialized_comparator\030\002 \001(\014\"$\n" +
|
||||
"\023ByteArrayComparable\022\r\n\005value\030\001 \001(\014\"<\n\020B" +
|
||||
"inaryComparator\022(\n\ncomparable\030\001 \002(\0132\024.By" +
|
||||
"teArrayComparable\"B\n\026BinaryPrefixCompara" +
|
||||
"tor\022(\n\ncomparable\030\001 \002(\0132\024.ByteArrayCompa" +
|
||||
"rable\"\216\001\n\rBitComparator\022(\n\ncomparable\030\001 " +
|
||||
"\002(\0132\024.ByteArrayComparable\022,\n\nbitwise_op\030" +
|
||||
"\002 \002(\0162\030.BitComparator.BitwiseOp\"%\n\tBitwi" +
|
||||
"seOp\022\007\n\003AND\020\001\022\006\n\002OR\020\002\022\007\n\003XOR\020\003\"\020\n\016NullCo",
|
||||
"mparator\"P\n\025RegexStringComparator\022\017\n\007pat" +
|
||||
"tern\030\001 \002(\t\022\025\n\rpattern_flags\030\002 \002(\005\022\017\n\007cha" +
|
||||
"rset\030\003 \002(\t\"%\n\023SubstringComparator\022\016\n\006sub" +
|
||||
"str\030\001 \002(\tBF\n*org.apache.hadoop.hbase.pro" +
|
||||
"tobuf.generatedB\020ComparatorProtosH\001\210\001\001\240\001" +
|
||||
"\001"
|
||||
};
|
||||
com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
|
||||
new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() {
|
||||
|
|
|
@ -11,19 +11,19 @@ public final class ErrorHandlingProtos {
|
|||
public interface StackTraceElementMessageOrBuilder
|
||||
extends com.google.protobuf.MessageOrBuilder {
|
||||
|
||||
// optional string declaringClass = 1;
|
||||
// optional string declaring_class = 1;
|
||||
boolean hasDeclaringClass();
|
||||
String getDeclaringClass();
|
||||
|
||||
// optional string methodName = 2;
|
||||
// optional string method_name = 2;
|
||||
boolean hasMethodName();
|
||||
String getMethodName();
|
||||
|
||||
// optional string fileName = 3;
|
||||
// optional string file_name = 3;
|
||||
boolean hasFileName();
|
||||
String getFileName();
|
||||
|
||||
// optional int32 lineNumber = 4;
|
||||
// optional int32 line_number = 4;
|
||||
boolean hasLineNumber();
|
||||
int getLineNumber();
|
||||
}
|
||||
|
@ -56,8 +56,8 @@ public final class ErrorHandlingProtos {
|
|||
}
|
||||
|
||||
private int bitField0_;
|
||||
// optional string declaringClass = 1;
|
||||
public static final int DECLARINGCLASS_FIELD_NUMBER = 1;
|
||||
// optional string declaring_class = 1;
|
||||
public static final int DECLARING_CLASS_FIELD_NUMBER = 1;
|
||||
private java.lang.Object declaringClass_;
|
||||
public boolean hasDeclaringClass() {
|
||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||
|
@ -88,8 +88,8 @@ public final class ErrorHandlingProtos {
|
|||
}
|
||||
}
|
||||
|
||||
// optional string methodName = 2;
|
||||
public static final int METHODNAME_FIELD_NUMBER = 2;
|
||||
// optional string method_name = 2;
|
||||
public static final int METHOD_NAME_FIELD_NUMBER = 2;
|
||||
private java.lang.Object methodName_;
|
||||
public boolean hasMethodName() {
|
||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||
|
@ -120,8 +120,8 @@ public final class ErrorHandlingProtos {
|
|||
}
|
||||
}
|
||||
|
||||
// optional string fileName = 3;
|
||||
public static final int FILENAME_FIELD_NUMBER = 3;
|
||||
// optional string file_name = 3;
|
||||
public static final int FILE_NAME_FIELD_NUMBER = 3;
|
||||
private java.lang.Object fileName_;
|
||||
public boolean hasFileName() {
|
||||
return ((bitField0_ & 0x00000004) == 0x00000004);
|
||||
|
@ -152,8 +152,8 @@ public final class ErrorHandlingProtos {
|
|||
}
|
||||
}
|
||||
|
||||
// optional int32 lineNumber = 4;
|
||||
public static final int LINENUMBER_FIELD_NUMBER = 4;
|
||||
// optional int32 line_number = 4;
|
||||
public static final int LINE_NUMBER_FIELD_NUMBER = 4;
|
||||
private int lineNumber_;
|
||||
public boolean hasLineNumber() {
|
||||
return ((bitField0_ & 0x00000008) == 0x00000008);
|
||||
|
@ -270,19 +270,19 @@ public final class ErrorHandlingProtos {
|
|||
int hash = 41;
|
||||
hash = (19 * hash) + getDescriptorForType().hashCode();
|
||||
if (hasDeclaringClass()) {
|
||||
hash = (37 * hash) + DECLARINGCLASS_FIELD_NUMBER;
|
||||
hash = (37 * hash) + DECLARING_CLASS_FIELD_NUMBER;
|
||||
hash = (53 * hash) + getDeclaringClass().hashCode();
|
||||
}
|
||||
if (hasMethodName()) {
|
||||
hash = (37 * hash) + METHODNAME_FIELD_NUMBER;
|
||||
hash = (37 * hash) + METHOD_NAME_FIELD_NUMBER;
|
||||
hash = (53 * hash) + getMethodName().hashCode();
|
||||
}
|
||||
if (hasFileName()) {
|
||||
hash = (37 * hash) + FILENAME_FIELD_NUMBER;
|
||||
hash = (37 * hash) + FILE_NAME_FIELD_NUMBER;
|
||||
hash = (53 * hash) + getFileName().hashCode();
|
||||
}
|
||||
if (hasLineNumber()) {
|
||||
hash = (37 * hash) + LINENUMBER_FIELD_NUMBER;
|
||||
hash = (37 * hash) + LINE_NUMBER_FIELD_NUMBER;
|
||||
hash = (53 * hash) + getLineNumber();
|
||||
}
|
||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||
|
@ -548,7 +548,7 @@ public final class ErrorHandlingProtos {
|
|||
|
||||
private int bitField0_;
|
||||
|
||||
// optional string declaringClass = 1;
|
||||
// optional string declaring_class = 1;
|
||||
private java.lang.Object declaringClass_ = "";
|
||||
public boolean hasDeclaringClass() {
|
||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||
|
@ -584,7 +584,7 @@ public final class ErrorHandlingProtos {
|
|||
onChanged();
|
||||
}
|
||||
|
||||
// optional string methodName = 2;
|
||||
// optional string method_name = 2;
|
||||
private java.lang.Object methodName_ = "";
|
||||
public boolean hasMethodName() {
|
||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||
|
@ -620,7 +620,7 @@ public final class ErrorHandlingProtos {
|
|||
onChanged();
|
||||
}
|
||||
|
||||
// optional string fileName = 3;
|
||||
// optional string file_name = 3;
|
||||
private java.lang.Object fileName_ = "";
|
||||
public boolean hasFileName() {
|
||||
return ((bitField0_ & 0x00000004) == 0x00000004);
|
||||
|
@ -656,7 +656,7 @@ public final class ErrorHandlingProtos {
|
|||
onChanged();
|
||||
}
|
||||
|
||||
// optional int32 lineNumber = 4;
|
||||
// optional int32 line_number = 4;
|
||||
private int lineNumber_ ;
|
||||
public boolean hasLineNumber() {
|
||||
return ((bitField0_ & 0x00000008) == 0x00000008);
|
||||
|
@ -691,7 +691,7 @@ public final class ErrorHandlingProtos {
|
|||
public interface GenericExceptionMessageOrBuilder
|
||||
extends com.google.protobuf.MessageOrBuilder {
|
||||
|
||||
// optional string className = 1;
|
||||
// optional string class_name = 1;
|
||||
boolean hasClassName();
|
||||
String getClassName();
|
||||
|
||||
|
@ -699,7 +699,7 @@ public final class ErrorHandlingProtos {
|
|||
boolean hasMessage();
|
||||
String getMessage();
|
||||
|
||||
// optional bytes errorInfo = 3;
|
||||
// optional bytes error_info = 3;
|
||||
boolean hasErrorInfo();
|
||||
com.google.protobuf.ByteString getErrorInfo();
|
||||
|
||||
|
@ -742,8 +742,8 @@ public final class ErrorHandlingProtos {
|
|||
}
|
||||
|
||||
private int bitField0_;
|
||||
// optional string className = 1;
|
||||
public static final int CLASSNAME_FIELD_NUMBER = 1;
|
||||
// optional string class_name = 1;
|
||||
public static final int CLASS_NAME_FIELD_NUMBER = 1;
|
||||
private java.lang.Object className_;
|
||||
public boolean hasClassName() {
|
||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||
|
@ -806,8 +806,8 @@ public final class ErrorHandlingProtos {
|
|||
}
|
||||
}
|
||||
|
||||
// optional bytes errorInfo = 3;
|
||||
public static final int ERRORINFO_FIELD_NUMBER = 3;
|
||||
// optional bytes error_info = 3;
|
||||
public static final int ERROR_INFO_FIELD_NUMBER = 3;
|
||||
private com.google.protobuf.ByteString errorInfo_;
|
||||
public boolean hasErrorInfo() {
|
||||
return ((bitField0_ & 0x00000004) == 0x00000004);
|
||||
|
@ -942,7 +942,7 @@ public final class ErrorHandlingProtos {
|
|||
int hash = 41;
|
||||
hash = (19 * hash) + getDescriptorForType().hashCode();
|
||||
if (hasClassName()) {
|
||||
hash = (37 * hash) + CLASSNAME_FIELD_NUMBER;
|
||||
hash = (37 * hash) + CLASS_NAME_FIELD_NUMBER;
|
||||
hash = (53 * hash) + getClassName().hashCode();
|
||||
}
|
||||
if (hasMessage()) {
|
||||
|
@ -950,7 +950,7 @@ public final class ErrorHandlingProtos {
|
|||
hash = (53 * hash) + getMessage().hashCode();
|
||||
}
|
||||
if (hasErrorInfo()) {
|
||||
hash = (37 * hash) + ERRORINFO_FIELD_NUMBER;
|
||||
hash = (37 * hash) + ERROR_INFO_FIELD_NUMBER;
|
||||
hash = (53 * hash) + getErrorInfo().hashCode();
|
||||
}
|
||||
if (getTraceCount() > 0) {
|
||||
|
@ -1254,7 +1254,7 @@ public final class ErrorHandlingProtos {
|
|||
|
||||
private int bitField0_;
|
||||
|
||||
// optional string className = 1;
|
||||
// optional string class_name = 1;
|
||||
private java.lang.Object className_ = "";
|
||||
public boolean hasClassName() {
|
||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||
|
@ -1326,7 +1326,7 @@ public final class ErrorHandlingProtos {
|
|||
onChanged();
|
||||
}
|
||||
|
||||
// optional bytes errorInfo = 3;
|
||||
// optional bytes error_info = 3;
|
||||
private com.google.protobuf.ByteString errorInfo_ = com.google.protobuf.ByteString.EMPTY;
|
||||
public boolean hasErrorInfo() {
|
||||
return ((bitField0_ & 0x00000004) == 0x00000004);
|
||||
|
@ -1554,7 +1554,7 @@ public final class ErrorHandlingProtos {
|
|||
boolean hasSource();
|
||||
String getSource();
|
||||
|
||||
// optional .GenericExceptionMessage genericException = 2;
|
||||
// optional .GenericExceptionMessage generic_exception = 2;
|
||||
boolean hasGenericException();
|
||||
org.apache.hadoop.hbase.protobuf.generated.ErrorHandlingProtos.GenericExceptionMessage getGenericException();
|
||||
org.apache.hadoop.hbase.protobuf.generated.ErrorHandlingProtos.GenericExceptionMessageOrBuilder getGenericExceptionOrBuilder();
|
||||
|
@ -1620,8 +1620,8 @@ public final class ErrorHandlingProtos {
|
|||
}
|
||||
}
|
||||
|
||||
// optional .GenericExceptionMessage genericException = 2;
|
||||
public static final int GENERICEXCEPTION_FIELD_NUMBER = 2;
|
||||
// optional .GenericExceptionMessage generic_exception = 2;
|
||||
public static final int GENERIC_EXCEPTION_FIELD_NUMBER = 2;
|
||||
private org.apache.hadoop.hbase.protobuf.generated.ErrorHandlingProtos.GenericExceptionMessage genericException_;
|
||||
public boolean hasGenericException() {
|
||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||
|
@ -1719,7 +1719,7 @@ public final class ErrorHandlingProtos {
|
|||
hash = (53 * hash) + getSource().hashCode();
|
||||
}
|
||||
if (hasGenericException()) {
|
||||
hash = (37 * hash) + GENERICEXCEPTION_FIELD_NUMBER;
|
||||
hash = (37 * hash) + GENERIC_EXCEPTION_FIELD_NUMBER;
|
||||
hash = (53 * hash) + getGenericException().hashCode();
|
||||
}
|
||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||
|
@ -2006,7 +2006,7 @@ public final class ErrorHandlingProtos {
|
|||
onChanged();
|
||||
}
|
||||
|
||||
// optional .GenericExceptionMessage genericException = 2;
|
||||
// optional .GenericExceptionMessage generic_exception = 2;
|
||||
private org.apache.hadoop.hbase.protobuf.generated.ErrorHandlingProtos.GenericExceptionMessage genericException_ = org.apache.hadoop.hbase.protobuf.generated.ErrorHandlingProtos.GenericExceptionMessage.getDefaultInstance();
|
||||
private com.google.protobuf.SingleFieldBuilder<
|
||||
org.apache.hadoop.hbase.protobuf.generated.ErrorHandlingProtos.GenericExceptionMessage, org.apache.hadoop.hbase.protobuf.generated.ErrorHandlingProtos.GenericExceptionMessage.Builder, org.apache.hadoop.hbase.protobuf.generated.ErrorHandlingProtos.GenericExceptionMessageOrBuilder> genericExceptionBuilder_;
|
||||
|
@ -2131,17 +2131,17 @@ public final class ErrorHandlingProtos {
|
|||
descriptor;
|
||||
static {
|
||||
java.lang.String[] descriptorData = {
|
||||
"\n\023ErrorHandling.proto\"l\n\030StackTraceEleme" +
|
||||
"ntMessage\022\026\n\016declaringClass\030\001 \001(\t\022\022\n\nmet" +
|
||||
"hodName\030\002 \001(\t\022\020\n\010fileName\030\003 \001(\t\022\022\n\nlineN" +
|
||||
"umber\030\004 \001(\005\"z\n\027GenericExceptionMessage\022\021" +
|
||||
"\n\tclassName\030\001 \001(\t\022\017\n\007message\030\002 \001(\t\022\021\n\ter" +
|
||||
"rorInfo\030\003 \001(\014\022(\n\005trace\030\004 \003(\0132\031.StackTrac" +
|
||||
"eElementMessage\"]\n\027ForeignExceptionMessa" +
|
||||
"ge\022\016\n\006source\030\001 \001(\t\0222\n\020genericException\030\002" +
|
||||
" \001(\0132\030.GenericExceptionMessageBF\n*org.ap" +
|
||||
"ache.hadoop.hbase.protobuf.generatedB\023Er",
|
||||
"rorHandlingProtosH\001\240\001\001"
|
||||
"\n\023ErrorHandling.proto\"p\n\030StackTraceEleme" +
|
||||
"ntMessage\022\027\n\017declaring_class\030\001 \001(\t\022\023\n\013me" +
|
||||
"thod_name\030\002 \001(\t\022\021\n\tfile_name\030\003 \001(\t\022\023\n\013li" +
|
||||
"ne_number\030\004 \001(\005\"|\n\027GenericExceptionMessa" +
|
||||
"ge\022\022\n\nclass_name\030\001 \001(\t\022\017\n\007message\030\002 \001(\t\022" +
|
||||
"\022\n\nerror_info\030\003 \001(\014\022(\n\005trace\030\004 \003(\0132\031.Sta" +
|
||||
"ckTraceElementMessage\"^\n\027ForeignExceptio" +
|
||||
"nMessage\022\016\n\006source\030\001 \001(\t\0223\n\021generic_exce" +
|
||||
"ption\030\002 \001(\0132\030.GenericExceptionMessageBF\n" +
|
||||
"*org.apache.hadoop.hbase.protobuf.genera",
|
||||
"tedB\023ErrorHandlingProtosH\001\240\001\001"
|
||||
};
|
||||
com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
|
||||
new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() {
|
||||
|
|
|
@ -15,7 +15,7 @@ public final class FilterProtos {
|
|||
boolean hasName();
|
||||
String getName();
|
||||
|
||||
// optional bytes serializedFilter = 2;
|
||||
// optional bytes serialized_filter = 2;
|
||||
boolean hasSerializedFilter();
|
||||
com.google.protobuf.ByteString getSerializedFilter();
|
||||
}
|
||||
|
@ -80,8 +80,8 @@ public final class FilterProtos {
|
|||
}
|
||||
}
|
||||
|
||||
// optional bytes serializedFilter = 2;
|
||||
public static final int SERIALIZEDFILTER_FIELD_NUMBER = 2;
|
||||
// optional bytes serialized_filter = 2;
|
||||
public static final int SERIALIZED_FILTER_FIELD_NUMBER = 2;
|
||||
private com.google.protobuf.ByteString serializedFilter_;
|
||||
public boolean hasSerializedFilter() {
|
||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||
|
@ -180,7 +180,7 @@ public final class FilterProtos {
|
|||
hash = (53 * hash) + getName().hashCode();
|
||||
}
|
||||
if (hasSerializedFilter()) {
|
||||
hash = (37 * hash) + SERIALIZEDFILTER_FIELD_NUMBER;
|
||||
hash = (37 * hash) + SERIALIZED_FILTER_FIELD_NUMBER;
|
||||
hash = (53 * hash) + getSerializedFilter().hashCode();
|
||||
}
|
||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||
|
@ -458,7 +458,7 @@ public final class FilterProtos {
|
|||
onChanged();
|
||||
}
|
||||
|
||||
// optional bytes serializedFilter = 2;
|
||||
// optional bytes serialized_filter = 2;
|
||||
private com.google.protobuf.ByteString serializedFilter_ = com.google.protobuf.ByteString.EMPTY;
|
||||
public boolean hasSerializedFilter() {
|
||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||
|
@ -883,7 +883,7 @@ public final class FilterProtos {
|
|||
boolean hasOffset();
|
||||
int getOffset();
|
||||
|
||||
// optional bytes columnOffset = 3;
|
||||
// optional bytes column_offset = 3;
|
||||
boolean hasColumnOffset();
|
||||
com.google.protobuf.ByteString getColumnOffset();
|
||||
}
|
||||
|
@ -936,8 +936,8 @@ public final class FilterProtos {
|
|||
return offset_;
|
||||
}
|
||||
|
||||
// optional bytes columnOffset = 3;
|
||||
public static final int COLUMNOFFSET_FIELD_NUMBER = 3;
|
||||
// optional bytes column_offset = 3;
|
||||
public static final int COLUMN_OFFSET_FIELD_NUMBER = 3;
|
||||
private com.google.protobuf.ByteString columnOffset_;
|
||||
public boolean hasColumnOffset() {
|
||||
return ((bitField0_ & 0x00000004) == 0x00000004);
|
||||
|
@ -1053,7 +1053,7 @@ public final class FilterProtos {
|
|||
hash = (53 * hash) + getOffset();
|
||||
}
|
||||
if (hasColumnOffset()) {
|
||||
hash = (37 * hash) + COLUMNOFFSET_FIELD_NUMBER;
|
||||
hash = (37 * hash) + COLUMN_OFFSET_FIELD_NUMBER;
|
||||
hash = (53 * hash) + getColumnOffset().hashCode();
|
||||
}
|
||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||
|
@ -1351,7 +1351,7 @@ public final class FilterProtos {
|
|||
return this;
|
||||
}
|
||||
|
||||
// optional bytes columnOffset = 3;
|
||||
// optional bytes column_offset = 3;
|
||||
private com.google.protobuf.ByteString columnOffset_ = com.google.protobuf.ByteString.EMPTY;
|
||||
public boolean hasColumnOffset() {
|
||||
return ((bitField0_ & 0x00000004) == 0x00000004);
|
||||
|
@ -1771,19 +1771,19 @@ public final class FilterProtos {
|
|||
public interface ColumnRangeFilterOrBuilder
|
||||
extends com.google.protobuf.MessageOrBuilder {
|
||||
|
||||
// optional bytes minColumn = 1;
|
||||
// optional bytes min_column = 1;
|
||||
boolean hasMinColumn();
|
||||
com.google.protobuf.ByteString getMinColumn();
|
||||
|
||||
// optional bool minColumnInclusive = 2;
|
||||
// optional bool min_column_inclusive = 2;
|
||||
boolean hasMinColumnInclusive();
|
||||
boolean getMinColumnInclusive();
|
||||
|
||||
// optional bytes maxColumn = 3;
|
||||
// optional bytes max_column = 3;
|
||||
boolean hasMaxColumn();
|
||||
com.google.protobuf.ByteString getMaxColumn();
|
||||
|
||||
// optional bool maxColumnInclusive = 4;
|
||||
// optional bool max_column_inclusive = 4;
|
||||
boolean hasMaxColumnInclusive();
|
||||
boolean getMaxColumnInclusive();
|
||||
}
|
||||
|
@ -1816,8 +1816,8 @@ public final class FilterProtos {
|
|||
}
|
||||
|
||||
private int bitField0_;
|
||||
// optional bytes minColumn = 1;
|
||||
public static final int MINCOLUMN_FIELD_NUMBER = 1;
|
||||
// optional bytes min_column = 1;
|
||||
public static final int MIN_COLUMN_FIELD_NUMBER = 1;
|
||||
private com.google.protobuf.ByteString minColumn_;
|
||||
public boolean hasMinColumn() {
|
||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||
|
@ -1826,8 +1826,8 @@ public final class FilterProtos {
|
|||
return minColumn_;
|
||||
}
|
||||
|
||||
// optional bool minColumnInclusive = 2;
|
||||
public static final int MINCOLUMNINCLUSIVE_FIELD_NUMBER = 2;
|
||||
// optional bool min_column_inclusive = 2;
|
||||
public static final int MIN_COLUMN_INCLUSIVE_FIELD_NUMBER = 2;
|
||||
private boolean minColumnInclusive_;
|
||||
public boolean hasMinColumnInclusive() {
|
||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||
|
@ -1836,8 +1836,8 @@ public final class FilterProtos {
|
|||
return minColumnInclusive_;
|
||||
}
|
||||
|
||||
// optional bytes maxColumn = 3;
|
||||
public static final int MAXCOLUMN_FIELD_NUMBER = 3;
|
||||
// optional bytes max_column = 3;
|
||||
public static final int MAX_COLUMN_FIELD_NUMBER = 3;
|
||||
private com.google.protobuf.ByteString maxColumn_;
|
||||
public boolean hasMaxColumn() {
|
||||
return ((bitField0_ & 0x00000004) == 0x00000004);
|
||||
|
@ -1846,8 +1846,8 @@ public final class FilterProtos {
|
|||
return maxColumn_;
|
||||
}
|
||||
|
||||
// optional bool maxColumnInclusive = 4;
|
||||
public static final int MAXCOLUMNINCLUSIVE_FIELD_NUMBER = 4;
|
||||
// optional bool max_column_inclusive = 4;
|
||||
public static final int MAX_COLUMN_INCLUSIVE_FIELD_NUMBER = 4;
|
||||
private boolean maxColumnInclusive_;
|
||||
public boolean hasMaxColumnInclusive() {
|
||||
return ((bitField0_ & 0x00000008) == 0x00000008);
|
||||
|
@ -1964,19 +1964,19 @@ public final class FilterProtos {
|
|||
int hash = 41;
|
||||
hash = (19 * hash) + getDescriptorForType().hashCode();
|
||||
if (hasMinColumn()) {
|
||||
hash = (37 * hash) + MINCOLUMN_FIELD_NUMBER;
|
||||
hash = (37 * hash) + MIN_COLUMN_FIELD_NUMBER;
|
||||
hash = (53 * hash) + getMinColumn().hashCode();
|
||||
}
|
||||
if (hasMinColumnInclusive()) {
|
||||
hash = (37 * hash) + MINCOLUMNINCLUSIVE_FIELD_NUMBER;
|
||||
hash = (37 * hash) + MIN_COLUMN_INCLUSIVE_FIELD_NUMBER;
|
||||
hash = (53 * hash) + hashBoolean(getMinColumnInclusive());
|
||||
}
|
||||
if (hasMaxColumn()) {
|
||||
hash = (37 * hash) + MAXCOLUMN_FIELD_NUMBER;
|
||||
hash = (37 * hash) + MAX_COLUMN_FIELD_NUMBER;
|
||||
hash = (53 * hash) + getMaxColumn().hashCode();
|
||||
}
|
||||
if (hasMaxColumnInclusive()) {
|
||||
hash = (37 * hash) + MAXCOLUMNINCLUSIVE_FIELD_NUMBER;
|
||||
hash = (37 * hash) + MAX_COLUMN_INCLUSIVE_FIELD_NUMBER;
|
||||
hash = (53 * hash) + hashBoolean(getMaxColumnInclusive());
|
||||
}
|
||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||
|
@ -2242,7 +2242,7 @@ public final class FilterProtos {
|
|||
|
||||
private int bitField0_;
|
||||
|
||||
// optional bytes minColumn = 1;
|
||||
// optional bytes min_column = 1;
|
||||
private com.google.protobuf.ByteString minColumn_ = com.google.protobuf.ByteString.EMPTY;
|
||||
public boolean hasMinColumn() {
|
||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||
|
@ -2266,7 +2266,7 @@ public final class FilterProtos {
|
|||
return this;
|
||||
}
|
||||
|
||||
// optional bool minColumnInclusive = 2;
|
||||
// optional bool min_column_inclusive = 2;
|
||||
private boolean minColumnInclusive_ ;
|
||||
public boolean hasMinColumnInclusive() {
|
||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||
|
@ -2287,7 +2287,7 @@ public final class FilterProtos {
|
|||
return this;
|
||||
}
|
||||
|
||||
// optional bytes maxColumn = 3;
|
||||
// optional bytes max_column = 3;
|
||||
private com.google.protobuf.ByteString maxColumn_ = com.google.protobuf.ByteString.EMPTY;
|
||||
public boolean hasMaxColumn() {
|
||||
return ((bitField0_ & 0x00000004) == 0x00000004);
|
||||
|
@ -2311,7 +2311,7 @@ public final class FilterProtos {
|
|||
return this;
|
||||
}
|
||||
|
||||
// optional bool maxColumnInclusive = 4;
|
||||
// optional bool max_column_inclusive = 4;
|
||||
private boolean maxColumnInclusive_ ;
|
||||
public boolean hasMaxColumnInclusive() {
|
||||
return ((bitField0_ & 0x00000008) == 0x00000008);
|
||||
|
@ -2346,7 +2346,7 @@ public final class FilterProtos {
|
|||
public interface CompareFilterOrBuilder
|
||||
extends com.google.protobuf.MessageOrBuilder {
|
||||
|
||||
// required .CompareType compareOp = 1;
|
||||
// required .CompareType compare_op = 1;
|
||||
boolean hasCompareOp();
|
||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.CompareType getCompareOp();
|
||||
|
||||
|
@ -2384,8 +2384,8 @@ public final class FilterProtos {
|
|||
}
|
||||
|
||||
private int bitField0_;
|
||||
// required .CompareType compareOp = 1;
|
||||
public static final int COMPAREOP_FIELD_NUMBER = 1;
|
||||
// required .CompareType compare_op = 1;
|
||||
public static final int COMPARE_OP_FIELD_NUMBER = 1;
|
||||
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.CompareType compareOp_;
|
||||
public boolean hasCompareOp() {
|
||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||
|
@ -2499,7 +2499,7 @@ public final class FilterProtos {
|
|||
int hash = 41;
|
||||
hash = (19 * hash) + getDescriptorForType().hashCode();
|
||||
if (hasCompareOp()) {
|
||||
hash = (37 * hash) + COMPAREOP_FIELD_NUMBER;
|
||||
hash = (37 * hash) + COMPARE_OP_FIELD_NUMBER;
|
||||
hash = (53 * hash) + hashEnum(getCompareOp());
|
||||
}
|
||||
if (hasComparator()) {
|
||||
|
@ -2770,7 +2770,7 @@ public final class FilterProtos {
|
|||
|
||||
private int bitField0_;
|
||||
|
||||
// required .CompareType compareOp = 1;
|
||||
// required .CompareType compare_op = 1;
|
||||
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.CompareType compareOp_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.CompareType.LESS;
|
||||
public boolean hasCompareOp() {
|
||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||
|
@ -2898,20 +2898,20 @@ public final class FilterProtos {
|
|||
public interface DependentColumnFilterOrBuilder
|
||||
extends com.google.protobuf.MessageOrBuilder {
|
||||
|
||||
// required .CompareFilter compareFilter = 1;
|
||||
// required .CompareFilter compare_filter = 1;
|
||||
boolean hasCompareFilter();
|
||||
org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter getCompareFilter();
|
||||
org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilterOrBuilder getCompareFilterOrBuilder();
|
||||
|
||||
// optional bytes columnFamily = 2;
|
||||
// optional bytes column_family = 2;
|
||||
boolean hasColumnFamily();
|
||||
com.google.protobuf.ByteString getColumnFamily();
|
||||
|
||||
// optional bytes columnQualifier = 3;
|
||||
// optional bytes column_qualifier = 3;
|
||||
boolean hasColumnQualifier();
|
||||
com.google.protobuf.ByteString getColumnQualifier();
|
||||
|
||||
// optional bool dropDependentColumn = 4;
|
||||
// optional bool drop_dependent_column = 4;
|
||||
boolean hasDropDependentColumn();
|
||||
boolean getDropDependentColumn();
|
||||
}
|
||||
|
@ -2944,8 +2944,8 @@ public final class FilterProtos {
|
|||
}
|
||||
|
||||
private int bitField0_;
|
||||
// required .CompareFilter compareFilter = 1;
|
||||
public static final int COMPAREFILTER_FIELD_NUMBER = 1;
|
||||
// required .CompareFilter compare_filter = 1;
|
||||
public static final int COMPARE_FILTER_FIELD_NUMBER = 1;
|
||||
private org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter compareFilter_;
|
||||
public boolean hasCompareFilter() {
|
||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||
|
@ -2957,8 +2957,8 @@ public final class FilterProtos {
|
|||
return compareFilter_;
|
||||
}
|
||||
|
||||
// optional bytes columnFamily = 2;
|
||||
public static final int COLUMNFAMILY_FIELD_NUMBER = 2;
|
||||
// optional bytes column_family = 2;
|
||||
public static final int COLUMN_FAMILY_FIELD_NUMBER = 2;
|
||||
private com.google.protobuf.ByteString columnFamily_;
|
||||
public boolean hasColumnFamily() {
|
||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||
|
@ -2967,8 +2967,8 @@ public final class FilterProtos {
|
|||
return columnFamily_;
|
||||
}
|
||||
|
||||
// optional bytes columnQualifier = 3;
|
||||
public static final int COLUMNQUALIFIER_FIELD_NUMBER = 3;
|
||||
// optional bytes column_qualifier = 3;
|
||||
public static final int COLUMN_QUALIFIER_FIELD_NUMBER = 3;
|
||||
private com.google.protobuf.ByteString columnQualifier_;
|
||||
public boolean hasColumnQualifier() {
|
||||
return ((bitField0_ & 0x00000004) == 0x00000004);
|
||||
|
@ -2977,8 +2977,8 @@ public final class FilterProtos {
|
|||
return columnQualifier_;
|
||||
}
|
||||
|
||||
// optional bool dropDependentColumn = 4;
|
||||
public static final int DROPDEPENDENTCOLUMN_FIELD_NUMBER = 4;
|
||||
// optional bool drop_dependent_column = 4;
|
||||
public static final int DROP_DEPENDENT_COLUMN_FIELD_NUMBER = 4;
|
||||
private boolean dropDependentColumn_;
|
||||
public boolean hasDropDependentColumn() {
|
||||
return ((bitField0_ & 0x00000008) == 0x00000008);
|
||||
|
@ -3103,19 +3103,19 @@ public final class FilterProtos {
|
|||
int hash = 41;
|
||||
hash = (19 * hash) + getDescriptorForType().hashCode();
|
||||
if (hasCompareFilter()) {
|
||||
hash = (37 * hash) + COMPAREFILTER_FIELD_NUMBER;
|
||||
hash = (37 * hash) + COMPARE_FILTER_FIELD_NUMBER;
|
||||
hash = (53 * hash) + getCompareFilter().hashCode();
|
||||
}
|
||||
if (hasColumnFamily()) {
|
||||
hash = (37 * hash) + COLUMNFAMILY_FIELD_NUMBER;
|
||||
hash = (37 * hash) + COLUMN_FAMILY_FIELD_NUMBER;
|
||||
hash = (53 * hash) + getColumnFamily().hashCode();
|
||||
}
|
||||
if (hasColumnQualifier()) {
|
||||
hash = (37 * hash) + COLUMNQUALIFIER_FIELD_NUMBER;
|
||||
hash = (37 * hash) + COLUMN_QUALIFIER_FIELD_NUMBER;
|
||||
hash = (53 * hash) + getColumnQualifier().hashCode();
|
||||
}
|
||||
if (hasDropDependentColumn()) {
|
||||
hash = (37 * hash) + DROPDEPENDENTCOLUMN_FIELD_NUMBER;
|
||||
hash = (37 * hash) + DROP_DEPENDENT_COLUMN_FIELD_NUMBER;
|
||||
hash = (53 * hash) + hashBoolean(getDropDependentColumn());
|
||||
}
|
||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||
|
@ -3402,7 +3402,7 @@ public final class FilterProtos {
|
|||
|
||||
private int bitField0_;
|
||||
|
||||
// required .CompareFilter compareFilter = 1;
|
||||
// required .CompareFilter compare_filter = 1;
|
||||
private org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter compareFilter_ = org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter.getDefaultInstance();
|
||||
private com.google.protobuf.SingleFieldBuilder<
|
||||
org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter, org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter.Builder, org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilterOrBuilder> compareFilterBuilder_;
|
||||
|
@ -3492,7 +3492,7 @@ public final class FilterProtos {
|
|||
return compareFilterBuilder_;
|
||||
}
|
||||
|
||||
// optional bytes columnFamily = 2;
|
||||
// optional bytes column_family = 2;
|
||||
private com.google.protobuf.ByteString columnFamily_ = com.google.protobuf.ByteString.EMPTY;
|
||||
public boolean hasColumnFamily() {
|
||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||
|
@ -3516,7 +3516,7 @@ public final class FilterProtos {
|
|||
return this;
|
||||
}
|
||||
|
||||
// optional bytes columnQualifier = 3;
|
||||
// optional bytes column_qualifier = 3;
|
||||
private com.google.protobuf.ByteString columnQualifier_ = com.google.protobuf.ByteString.EMPTY;
|
||||
public boolean hasColumnQualifier() {
|
||||
return ((bitField0_ & 0x00000004) == 0x00000004);
|
||||
|
@ -3540,7 +3540,7 @@ public final class FilterProtos {
|
|||
return this;
|
||||
}
|
||||
|
||||
// optional bool dropDependentColumn = 4;
|
||||
// optional bool drop_dependent_column = 4;
|
||||
private boolean dropDependentColumn_ ;
|
||||
public boolean hasDropDependentColumn() {
|
||||
return ((bitField0_ & 0x00000008) == 0x00000008);
|
||||
|
@ -3575,7 +3575,7 @@ public final class FilterProtos {
|
|||
public interface FamilyFilterOrBuilder
|
||||
extends com.google.protobuf.MessageOrBuilder {
|
||||
|
||||
// required .CompareFilter compareFilter = 1;
|
||||
// required .CompareFilter compare_filter = 1;
|
||||
boolean hasCompareFilter();
|
||||
org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter getCompareFilter();
|
||||
org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilterOrBuilder getCompareFilterOrBuilder();
|
||||
|
@ -3609,8 +3609,8 @@ public final class FilterProtos {
|
|||
}
|
||||
|
||||
private int bitField0_;
|
||||
// required .CompareFilter compareFilter = 1;
|
||||
public static final int COMPAREFILTER_FIELD_NUMBER = 1;
|
||||
// required .CompareFilter compare_filter = 1;
|
||||
public static final int COMPARE_FILTER_FIELD_NUMBER = 1;
|
||||
private org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter compareFilter_;
|
||||
public boolean hasCompareFilter() {
|
||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||
|
@ -3699,7 +3699,7 @@ public final class FilterProtos {
|
|||
int hash = 41;
|
||||
hash = (19 * hash) + getDescriptorForType().hashCode();
|
||||
if (hasCompareFilter()) {
|
||||
hash = (37 * hash) + COMPAREFILTER_FIELD_NUMBER;
|
||||
hash = (37 * hash) + COMPARE_FILTER_FIELD_NUMBER;
|
||||
hash = (53 * hash) + getCompareFilter().hashCode();
|
||||
}
|
||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||
|
@ -3944,7 +3944,7 @@ public final class FilterProtos {
|
|||
|
||||
private int bitField0_;
|
||||
|
||||
// required .CompareFilter compareFilter = 1;
|
||||
// required .CompareFilter compare_filter = 1;
|
||||
private org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter compareFilter_ = org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter.getDefaultInstance();
|
||||
private com.google.protobuf.SingleFieldBuilder<
|
||||
org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter, org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter.Builder, org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilterOrBuilder> compareFilterBuilder_;
|
||||
|
@ -5982,7 +5982,7 @@ public final class FilterProtos {
|
|||
public interface FuzzyRowFilterOrBuilder
|
||||
extends com.google.protobuf.MessageOrBuilder {
|
||||
|
||||
// repeated .BytesBytesPair fuzzyKeysData = 1;
|
||||
// repeated .BytesBytesPair fuzzy_keys_data = 1;
|
||||
java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair>
|
||||
getFuzzyKeysDataList();
|
||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair getFuzzyKeysData(int index);
|
||||
|
@ -6020,8 +6020,8 @@ public final class FilterProtos {
|
|||
return org.apache.hadoop.hbase.protobuf.generated.FilterProtos.internal_static_FuzzyRowFilter_fieldAccessorTable;
|
||||
}
|
||||
|
||||
// repeated .BytesBytesPair fuzzyKeysData = 1;
|
||||
public static final int FUZZYKEYSDATA_FIELD_NUMBER = 1;
|
||||
// repeated .BytesBytesPair fuzzy_keys_data = 1;
|
||||
public static final int FUZZY_KEYS_DATA_FIELD_NUMBER = 1;
|
||||
private java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair> fuzzyKeysData_;
|
||||
public java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair> getFuzzyKeysDataList() {
|
||||
return fuzzyKeysData_;
|
||||
|
@ -6113,7 +6113,7 @@ public final class FilterProtos {
|
|||
int hash = 41;
|
||||
hash = (19 * hash) + getDescriptorForType().hashCode();
|
||||
if (getFuzzyKeysDataCount() > 0) {
|
||||
hash = (37 * hash) + FUZZYKEYSDATA_FIELD_NUMBER;
|
||||
hash = (37 * hash) + FUZZY_KEYS_DATA_FIELD_NUMBER;
|
||||
hash = (53 * hash) + getFuzzyKeysDataList().hashCode();
|
||||
}
|
||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||
|
@ -6375,7 +6375,7 @@ public final class FilterProtos {
|
|||
|
||||
private int bitField0_;
|
||||
|
||||
// repeated .BytesBytesPair fuzzyKeysData = 1;
|
||||
// repeated .BytesBytesPair fuzzy_keys_data = 1;
|
||||
private java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair> fuzzyKeysData_ =
|
||||
java.util.Collections.emptyList();
|
||||
private void ensureFuzzyKeysDataIsMutable() {
|
||||
|
@ -6575,7 +6575,7 @@ public final class FilterProtos {
|
|||
public interface InclusiveStopFilterOrBuilder
|
||||
extends com.google.protobuf.MessageOrBuilder {
|
||||
|
||||
// optional bytes stopRowKey = 1;
|
||||
// optional bytes stop_row_key = 1;
|
||||
boolean hasStopRowKey();
|
||||
com.google.protobuf.ByteString getStopRowKey();
|
||||
}
|
||||
|
@ -6608,8 +6608,8 @@ public final class FilterProtos {
|
|||
}
|
||||
|
||||
private int bitField0_;
|
||||
// optional bytes stopRowKey = 1;
|
||||
public static final int STOPROWKEY_FIELD_NUMBER = 1;
|
||||
// optional bytes stop_row_key = 1;
|
||||
public static final int STOP_ROW_KEY_FIELD_NUMBER = 1;
|
||||
private com.google.protobuf.ByteString stopRowKey_;
|
||||
public boolean hasStopRowKey() {
|
||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||
|
@ -6687,7 +6687,7 @@ public final class FilterProtos {
|
|||
int hash = 41;
|
||||
hash = (19 * hash) + getDescriptorForType().hashCode();
|
||||
if (hasStopRowKey()) {
|
||||
hash = (37 * hash) + STOPROWKEY_FIELD_NUMBER;
|
||||
hash = (37 * hash) + STOP_ROW_KEY_FIELD_NUMBER;
|
||||
hash = (53 * hash) + getStopRowKey().hashCode();
|
||||
}
|
||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||
|
@ -6911,7 +6911,7 @@ public final class FilterProtos {
|
|||
|
||||
private int bitField0_;
|
||||
|
||||
// optional bytes stopRowKey = 1;
|
||||
// optional bytes stop_row_key = 1;
|
||||
private com.google.protobuf.ByteString stopRowKey_ = com.google.protobuf.ByteString.EMPTY;
|
||||
public boolean hasStopRowKey() {
|
||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||
|
@ -6949,7 +6949,7 @@ public final class FilterProtos {
|
|||
public interface KeyOnlyFilterOrBuilder
|
||||
extends com.google.protobuf.MessageOrBuilder {
|
||||
|
||||
// required bool lenAsVal = 1;
|
||||
// required bool len_as_val = 1;
|
||||
boolean hasLenAsVal();
|
||||
boolean getLenAsVal();
|
||||
}
|
||||
|
@ -6982,8 +6982,8 @@ public final class FilterProtos {
|
|||
}
|
||||
|
||||
private int bitField0_;
|
||||
// required bool lenAsVal = 1;
|
||||
public static final int LENASVAL_FIELD_NUMBER = 1;
|
||||
// required bool len_as_val = 1;
|
||||
public static final int LEN_AS_VAL_FIELD_NUMBER = 1;
|
||||
private boolean lenAsVal_;
|
||||
public boolean hasLenAsVal() {
|
||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||
|
@ -7065,7 +7065,7 @@ public final class FilterProtos {
|
|||
int hash = 41;
|
||||
hash = (19 * hash) + getDescriptorForType().hashCode();
|
||||
if (hasLenAsVal()) {
|
||||
hash = (37 * hash) + LENASVAL_FIELD_NUMBER;
|
||||
hash = (37 * hash) + LEN_AS_VAL_FIELD_NUMBER;
|
||||
hash = (53 * hash) + hashBoolean(getLenAsVal());
|
||||
}
|
||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||
|
@ -7293,7 +7293,7 @@ public final class FilterProtos {
|
|||
|
||||
private int bitField0_;
|
||||
|
||||
// required bool lenAsVal = 1;
|
||||
// required bool len_as_val = 1;
|
||||
private boolean lenAsVal_ ;
|
||||
public boolean hasLenAsVal() {
|
||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||
|
@ -7328,7 +7328,7 @@ public final class FilterProtos {
|
|||
public interface MultipleColumnPrefixFilterOrBuilder
|
||||
extends com.google.protobuf.MessageOrBuilder {
|
||||
|
||||
// repeated bytes sortedPrefixes = 1;
|
||||
// repeated bytes sorted_prefixes = 1;
|
||||
java.util.List<com.google.protobuf.ByteString> getSortedPrefixesList();
|
||||
int getSortedPrefixesCount();
|
||||
com.google.protobuf.ByteString getSortedPrefixes(int index);
|
||||
|
@ -7361,8 +7361,8 @@ public final class FilterProtos {
|
|||
return org.apache.hadoop.hbase.protobuf.generated.FilterProtos.internal_static_MultipleColumnPrefixFilter_fieldAccessorTable;
|
||||
}
|
||||
|
||||
// repeated bytes sortedPrefixes = 1;
|
||||
public static final int SORTEDPREFIXES_FIELD_NUMBER = 1;
|
||||
// repeated bytes sorted_prefixes = 1;
|
||||
public static final int SORTED_PREFIXES_FIELD_NUMBER = 1;
|
||||
private java.util.List<com.google.protobuf.ByteString> sortedPrefixes_;
|
||||
public java.util.List<com.google.protobuf.ByteString>
|
||||
getSortedPrefixesList() {
|
||||
|
@ -7446,7 +7446,7 @@ public final class FilterProtos {
|
|||
int hash = 41;
|
||||
hash = (19 * hash) + getDescriptorForType().hashCode();
|
||||
if (getSortedPrefixesCount() > 0) {
|
||||
hash = (37 * hash) + SORTEDPREFIXES_FIELD_NUMBER;
|
||||
hash = (37 * hash) + SORTED_PREFIXES_FIELD_NUMBER;
|
||||
hash = (53 * hash) + getSortedPrefixesList().hashCode();
|
||||
}
|
||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||
|
@ -7676,7 +7676,7 @@ public final class FilterProtos {
|
|||
|
||||
private int bitField0_;
|
||||
|
||||
// repeated bytes sortedPrefixes = 1;
|
||||
// repeated bytes sorted_prefixes = 1;
|
||||
private java.util.List<com.google.protobuf.ByteString> sortedPrefixes_ = java.util.Collections.emptyList();;
|
||||
private void ensureSortedPrefixesIsMutable() {
|
||||
if (!((bitField0_ & 0x00000001) == 0x00000001)) {
|
||||
|
@ -7741,7 +7741,7 @@ public final class FilterProtos {
|
|||
public interface PageFilterOrBuilder
|
||||
extends com.google.protobuf.MessageOrBuilder {
|
||||
|
||||
// required int64 pageSize = 1;
|
||||
// required int64 page_size = 1;
|
||||
boolean hasPageSize();
|
||||
long getPageSize();
|
||||
}
|
||||
|
@ -7774,8 +7774,8 @@ public final class FilterProtos {
|
|||
}
|
||||
|
||||
private int bitField0_;
|
||||
// required int64 pageSize = 1;
|
||||
public static final int PAGESIZE_FIELD_NUMBER = 1;
|
||||
// required int64 page_size = 1;
|
||||
public static final int PAGE_SIZE_FIELD_NUMBER = 1;
|
||||
private long pageSize_;
|
||||
public boolean hasPageSize() {
|
||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||
|
@ -7857,7 +7857,7 @@ public final class FilterProtos {
|
|||
int hash = 41;
|
||||
hash = (19 * hash) + getDescriptorForType().hashCode();
|
||||
if (hasPageSize()) {
|
||||
hash = (37 * hash) + PAGESIZE_FIELD_NUMBER;
|
||||
hash = (37 * hash) + PAGE_SIZE_FIELD_NUMBER;
|
||||
hash = (53 * hash) + hashLong(getPageSize());
|
||||
}
|
||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||
|
@ -8085,7 +8085,7 @@ public final class FilterProtos {
|
|||
|
||||
private int bitField0_;
|
||||
|
||||
// required int64 pageSize = 1;
|
||||
// required int64 page_size = 1;
|
||||
private long pageSize_ ;
|
||||
public boolean hasPageSize() {
|
||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||
|
@ -8494,7 +8494,7 @@ public final class FilterProtos {
|
|||
public interface QualifierFilterOrBuilder
|
||||
extends com.google.protobuf.MessageOrBuilder {
|
||||
|
||||
// required .CompareFilter compareFilter = 1;
|
||||
// required .CompareFilter compare_filter = 1;
|
||||
boolean hasCompareFilter();
|
||||
org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter getCompareFilter();
|
||||
org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilterOrBuilder getCompareFilterOrBuilder();
|
||||
|
@ -8528,8 +8528,8 @@ public final class FilterProtos {
|
|||
}
|
||||
|
||||
private int bitField0_;
|
||||
// required .CompareFilter compareFilter = 1;
|
||||
public static final int COMPAREFILTER_FIELD_NUMBER = 1;
|
||||
// required .CompareFilter compare_filter = 1;
|
||||
public static final int COMPARE_FILTER_FIELD_NUMBER = 1;
|
||||
private org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter compareFilter_;
|
||||
public boolean hasCompareFilter() {
|
||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||
|
@ -8618,7 +8618,7 @@ public final class FilterProtos {
|
|||
int hash = 41;
|
||||
hash = (19 * hash) + getDescriptorForType().hashCode();
|
||||
if (hasCompareFilter()) {
|
||||
hash = (37 * hash) + COMPAREFILTER_FIELD_NUMBER;
|
||||
hash = (37 * hash) + COMPARE_FILTER_FIELD_NUMBER;
|
||||
hash = (53 * hash) + getCompareFilter().hashCode();
|
||||
}
|
||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||
|
@ -8863,7 +8863,7 @@ public final class FilterProtos {
|
|||
|
||||
private int bitField0_;
|
||||
|
||||
// required .CompareFilter compareFilter = 1;
|
||||
// required .CompareFilter compare_filter = 1;
|
||||
private org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter compareFilter_ = org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter.getDefaultInstance();
|
||||
private com.google.protobuf.SingleFieldBuilder<
|
||||
org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter, org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter.Builder, org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilterOrBuilder> compareFilterBuilder_;
|
||||
|
@ -9346,7 +9346,7 @@ public final class FilterProtos {
|
|||
public interface RowFilterOrBuilder
|
||||
extends com.google.protobuf.MessageOrBuilder {
|
||||
|
||||
// required .CompareFilter compareFilter = 1;
|
||||
// required .CompareFilter compare_filter = 1;
|
||||
boolean hasCompareFilter();
|
||||
org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter getCompareFilter();
|
||||
org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilterOrBuilder getCompareFilterOrBuilder();
|
||||
|
@ -9380,8 +9380,8 @@ public final class FilterProtos {
|
|||
}
|
||||
|
||||
private int bitField0_;
|
||||
// required .CompareFilter compareFilter = 1;
|
||||
public static final int COMPAREFILTER_FIELD_NUMBER = 1;
|
||||
// required .CompareFilter compare_filter = 1;
|
||||
public static final int COMPARE_FILTER_FIELD_NUMBER = 1;
|
||||
private org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter compareFilter_;
|
||||
public boolean hasCompareFilter() {
|
||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||
|
@ -9470,7 +9470,7 @@ public final class FilterProtos {
|
|||
int hash = 41;
|
||||
hash = (19 * hash) + getDescriptorForType().hashCode();
|
||||
if (hasCompareFilter()) {
|
||||
hash = (37 * hash) + COMPAREFILTER_FIELD_NUMBER;
|
||||
hash = (37 * hash) + COMPARE_FILTER_FIELD_NUMBER;
|
||||
hash = (53 * hash) + getCompareFilter().hashCode();
|
||||
}
|
||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||
|
@ -9715,7 +9715,7 @@ public final class FilterProtos {
|
|||
|
||||
private int bitField0_;
|
||||
|
||||
// required .CompareFilter compareFilter = 1;
|
||||
// required .CompareFilter compare_filter = 1;
|
||||
private org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter compareFilter_ = org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter.getDefaultInstance();
|
||||
private com.google.protobuf.SingleFieldBuilder<
|
||||
org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter, org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter.Builder, org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilterOrBuilder> compareFilterBuilder_;
|
||||
|
@ -9819,7 +9819,7 @@ public final class FilterProtos {
|
|||
public interface SingleColumnValueExcludeFilterOrBuilder
|
||||
extends com.google.protobuf.MessageOrBuilder {
|
||||
|
||||
// required .SingleColumnValueFilter singleColumnValueFilter = 1;
|
||||
// required .SingleColumnValueFilter single_column_value_filter = 1;
|
||||
boolean hasSingleColumnValueFilter();
|
||||
org.apache.hadoop.hbase.protobuf.generated.FilterProtos.SingleColumnValueFilter getSingleColumnValueFilter();
|
||||
org.apache.hadoop.hbase.protobuf.generated.FilterProtos.SingleColumnValueFilterOrBuilder getSingleColumnValueFilterOrBuilder();
|
||||
|
@ -9853,8 +9853,8 @@ public final class FilterProtos {
|
|||
}
|
||||
|
||||
private int bitField0_;
|
||||
// required .SingleColumnValueFilter singleColumnValueFilter = 1;
|
||||
public static final int SINGLECOLUMNVALUEFILTER_FIELD_NUMBER = 1;
|
||||
// required .SingleColumnValueFilter single_column_value_filter = 1;
|
||||
public static final int SINGLE_COLUMN_VALUE_FILTER_FIELD_NUMBER = 1;
|
||||
private org.apache.hadoop.hbase.protobuf.generated.FilterProtos.SingleColumnValueFilter singleColumnValueFilter_;
|
||||
public boolean hasSingleColumnValueFilter() {
|
||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||
|
@ -9943,7 +9943,7 @@ public final class FilterProtos {
|
|||
int hash = 41;
|
||||
hash = (19 * hash) + getDescriptorForType().hashCode();
|
||||
if (hasSingleColumnValueFilter()) {
|
||||
hash = (37 * hash) + SINGLECOLUMNVALUEFILTER_FIELD_NUMBER;
|
||||
hash = (37 * hash) + SINGLE_COLUMN_VALUE_FILTER_FIELD_NUMBER;
|
||||
hash = (53 * hash) + getSingleColumnValueFilter().hashCode();
|
||||
}
|
||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||
|
@ -10188,7 +10188,7 @@ public final class FilterProtos {
|
|||
|
||||
private int bitField0_;
|
||||
|
||||
// required .SingleColumnValueFilter singleColumnValueFilter = 1;
|
||||
// required .SingleColumnValueFilter single_column_value_filter = 1;
|
||||
private org.apache.hadoop.hbase.protobuf.generated.FilterProtos.SingleColumnValueFilter singleColumnValueFilter_ = org.apache.hadoop.hbase.protobuf.generated.FilterProtos.SingleColumnValueFilter.getDefaultInstance();
|
||||
private com.google.protobuf.SingleFieldBuilder<
|
||||
org.apache.hadoop.hbase.protobuf.generated.FilterProtos.SingleColumnValueFilter, org.apache.hadoop.hbase.protobuf.generated.FilterProtos.SingleColumnValueFilter.Builder, org.apache.hadoop.hbase.protobuf.generated.FilterProtos.SingleColumnValueFilterOrBuilder> singleColumnValueFilterBuilder_;
|
||||
|
@ -10292,15 +10292,15 @@ public final class FilterProtos {
|
|||
public interface SingleColumnValueFilterOrBuilder
|
||||
extends com.google.protobuf.MessageOrBuilder {
|
||||
|
||||
// optional bytes columnFamily = 1;
|
||||
// optional bytes column_family = 1;
|
||||
boolean hasColumnFamily();
|
||||
com.google.protobuf.ByteString getColumnFamily();
|
||||
|
||||
// optional bytes columnQualifier = 2;
|
||||
// optional bytes column_qualifier = 2;
|
||||
boolean hasColumnQualifier();
|
||||
com.google.protobuf.ByteString getColumnQualifier();
|
||||
|
||||
// required .CompareType compareOp = 3;
|
||||
// required .CompareType compare_op = 3;
|
||||
boolean hasCompareOp();
|
||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.CompareType getCompareOp();
|
||||
|
||||
|
@ -10309,11 +10309,11 @@ public final class FilterProtos {
|
|||
org.apache.hadoop.hbase.protobuf.generated.ComparatorProtos.Comparator getComparator();
|
||||
org.apache.hadoop.hbase.protobuf.generated.ComparatorProtos.ComparatorOrBuilder getComparatorOrBuilder();
|
||||
|
||||
// optional bool filterIfMissing = 5;
|
||||
// optional bool filter_if_missing = 5;
|
||||
boolean hasFilterIfMissing();
|
||||
boolean getFilterIfMissing();
|
||||
|
||||
// optional bool latestVersionOnly = 6;
|
||||
// optional bool latest_version_only = 6;
|
||||
boolean hasLatestVersionOnly();
|
||||
boolean getLatestVersionOnly();
|
||||
}
|
||||
|
@ -10346,8 +10346,8 @@ public final class FilterProtos {
|
|||
}
|
||||
|
||||
private int bitField0_;
|
||||
// optional bytes columnFamily = 1;
|
||||
public static final int COLUMNFAMILY_FIELD_NUMBER = 1;
|
||||
// optional bytes column_family = 1;
|
||||
public static final int COLUMN_FAMILY_FIELD_NUMBER = 1;
|
||||
private com.google.protobuf.ByteString columnFamily_;
|
||||
public boolean hasColumnFamily() {
|
||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||
|
@ -10356,8 +10356,8 @@ public final class FilterProtos {
|
|||
return columnFamily_;
|
||||
}
|
||||
|
||||
// optional bytes columnQualifier = 2;
|
||||
public static final int COLUMNQUALIFIER_FIELD_NUMBER = 2;
|
||||
// optional bytes column_qualifier = 2;
|
||||
public static final int COLUMN_QUALIFIER_FIELD_NUMBER = 2;
|
||||
private com.google.protobuf.ByteString columnQualifier_;
|
||||
public boolean hasColumnQualifier() {
|
||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||
|
@ -10366,8 +10366,8 @@ public final class FilterProtos {
|
|||
return columnQualifier_;
|
||||
}
|
||||
|
||||
// required .CompareType compareOp = 3;
|
||||
public static final int COMPAREOP_FIELD_NUMBER = 3;
|
||||
// required .CompareType compare_op = 3;
|
||||
public static final int COMPARE_OP_FIELD_NUMBER = 3;
|
||||
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.CompareType compareOp_;
|
||||
public boolean hasCompareOp() {
|
||||
return ((bitField0_ & 0x00000004) == 0x00000004);
|
||||
|
@ -10389,8 +10389,8 @@ public final class FilterProtos {
|
|||
return comparator_;
|
||||
}
|
||||
|
||||
// optional bool filterIfMissing = 5;
|
||||
public static final int FILTERIFMISSING_FIELD_NUMBER = 5;
|
||||
// optional bool filter_if_missing = 5;
|
||||
public static final int FILTER_IF_MISSING_FIELD_NUMBER = 5;
|
||||
private boolean filterIfMissing_;
|
||||
public boolean hasFilterIfMissing() {
|
||||
return ((bitField0_ & 0x00000010) == 0x00000010);
|
||||
|
@ -10399,8 +10399,8 @@ public final class FilterProtos {
|
|||
return filterIfMissing_;
|
||||
}
|
||||
|
||||
// optional bool latestVersionOnly = 6;
|
||||
public static final int LATESTVERSIONONLY_FIELD_NUMBER = 6;
|
||||
// optional bool latest_version_only = 6;
|
||||
public static final int LATEST_VERSION_ONLY_FIELD_NUMBER = 6;
|
||||
private boolean latestVersionOnly_;
|
||||
public boolean hasLatestVersionOnly() {
|
||||
return ((bitField0_ & 0x00000020) == 0x00000020);
|
||||
|
@ -10555,15 +10555,15 @@ public final class FilterProtos {
|
|||
int hash = 41;
|
||||
hash = (19 * hash) + getDescriptorForType().hashCode();
|
||||
if (hasColumnFamily()) {
|
||||
hash = (37 * hash) + COLUMNFAMILY_FIELD_NUMBER;
|
||||
hash = (37 * hash) + COLUMN_FAMILY_FIELD_NUMBER;
|
||||
hash = (53 * hash) + getColumnFamily().hashCode();
|
||||
}
|
||||
if (hasColumnQualifier()) {
|
||||
hash = (37 * hash) + COLUMNQUALIFIER_FIELD_NUMBER;
|
||||
hash = (37 * hash) + COLUMN_QUALIFIER_FIELD_NUMBER;
|
||||
hash = (53 * hash) + getColumnQualifier().hashCode();
|
||||
}
|
||||
if (hasCompareOp()) {
|
||||
hash = (37 * hash) + COMPAREOP_FIELD_NUMBER;
|
||||
hash = (37 * hash) + COMPARE_OP_FIELD_NUMBER;
|
||||
hash = (53 * hash) + hashEnum(getCompareOp());
|
||||
}
|
||||
if (hasComparator()) {
|
||||
|
@ -10571,11 +10571,11 @@ public final class FilterProtos {
|
|||
hash = (53 * hash) + getComparator().hashCode();
|
||||
}
|
||||
if (hasFilterIfMissing()) {
|
||||
hash = (37 * hash) + FILTERIFMISSING_FIELD_NUMBER;
|
||||
hash = (37 * hash) + FILTER_IF_MISSING_FIELD_NUMBER;
|
||||
hash = (53 * hash) + hashBoolean(getFilterIfMissing());
|
||||
}
|
||||
if (hasLatestVersionOnly()) {
|
||||
hash = (37 * hash) + LATESTVERSIONONLY_FIELD_NUMBER;
|
||||
hash = (37 * hash) + LATEST_VERSION_ONLY_FIELD_NUMBER;
|
||||
hash = (53 * hash) + hashBoolean(getLatestVersionOnly());
|
||||
}
|
||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||
|
@ -10900,7 +10900,7 @@ public final class FilterProtos {
|
|||
|
||||
private int bitField0_;
|
||||
|
||||
// optional bytes columnFamily = 1;
|
||||
// optional bytes column_family = 1;
|
||||
private com.google.protobuf.ByteString columnFamily_ = com.google.protobuf.ByteString.EMPTY;
|
||||
public boolean hasColumnFamily() {
|
||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||
|
@ -10924,7 +10924,7 @@ public final class FilterProtos {
|
|||
return this;
|
||||
}
|
||||
|
||||
// optional bytes columnQualifier = 2;
|
||||
// optional bytes column_qualifier = 2;
|
||||
private com.google.protobuf.ByteString columnQualifier_ = com.google.protobuf.ByteString.EMPTY;
|
||||
public boolean hasColumnQualifier() {
|
||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||
|
@ -10948,7 +10948,7 @@ public final class FilterProtos {
|
|||
return this;
|
||||
}
|
||||
|
||||
// required .CompareType compareOp = 3;
|
||||
// required .CompareType compare_op = 3;
|
||||
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.CompareType compareOp_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.CompareType.LESS;
|
||||
public boolean hasCompareOp() {
|
||||
return ((bitField0_ & 0x00000004) == 0x00000004);
|
||||
|
@ -11062,7 +11062,7 @@ public final class FilterProtos {
|
|||
return comparatorBuilder_;
|
||||
}
|
||||
|
||||
// optional bool filterIfMissing = 5;
|
||||
// optional bool filter_if_missing = 5;
|
||||
private boolean filterIfMissing_ ;
|
||||
public boolean hasFilterIfMissing() {
|
||||
return ((bitField0_ & 0x00000010) == 0x00000010);
|
||||
|
@ -11083,7 +11083,7 @@ public final class FilterProtos {
|
|||
return this;
|
||||
}
|
||||
|
||||
// optional bool latestVersionOnly = 6;
|
||||
// optional bool latest_version_only = 6;
|
||||
private boolean latestVersionOnly_ ;
|
||||
public boolean hasLatestVersionOnly() {
|
||||
return ((bitField0_ & 0x00000020) == 0x00000020);
|
||||
|
@ -12017,7 +12017,7 @@ public final class FilterProtos {
|
|||
public interface ValueFilterOrBuilder
|
||||
extends com.google.protobuf.MessageOrBuilder {
|
||||
|
||||
// required .CompareFilter compareFilter = 1;
|
||||
// required .CompareFilter compare_filter = 1;
|
||||
boolean hasCompareFilter();
|
||||
org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter getCompareFilter();
|
||||
org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilterOrBuilder getCompareFilterOrBuilder();
|
||||
|
@ -12051,8 +12051,8 @@ public final class FilterProtos {
|
|||
}
|
||||
|
||||
private int bitField0_;
|
||||
// required .CompareFilter compareFilter = 1;
|
||||
public static final int COMPAREFILTER_FIELD_NUMBER = 1;
|
||||
// required .CompareFilter compare_filter = 1;
|
||||
public static final int COMPARE_FILTER_FIELD_NUMBER = 1;
|
||||
private org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter compareFilter_;
|
||||
public boolean hasCompareFilter() {
|
||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||
|
@ -12141,7 +12141,7 @@ public final class FilterProtos {
|
|||
int hash = 41;
|
||||
hash = (19 * hash) + getDescriptorForType().hashCode();
|
||||
if (hasCompareFilter()) {
|
||||
hash = (37 * hash) + COMPAREFILTER_FIELD_NUMBER;
|
||||
hash = (37 * hash) + COMPARE_FILTER_FIELD_NUMBER;
|
||||
hash = (53 * hash) + getCompareFilter().hashCode();
|
||||
}
|
||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||
|
@ -12386,7 +12386,7 @@ public final class FilterProtos {
|
|||
|
||||
private int bitField0_;
|
||||
|
||||
// required .CompareFilter compareFilter = 1;
|
||||
// required .CompareFilter compare_filter = 1;
|
||||
private org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter compareFilter_ = org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter.getDefaultInstance();
|
||||
private com.google.protobuf.SingleFieldBuilder<
|
||||
org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter, org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter.Builder, org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilterOrBuilder> compareFilterBuilder_;
|
||||
|
@ -13105,51 +13105,52 @@ public final class FilterProtos {
|
|||
static {
|
||||
java.lang.String[] descriptorData = {
|
||||
"\n\014Filter.proto\032\013hbase.proto\032\020Comparator." +
|
||||
"proto\"0\n\006Filter\022\014\n\004name\030\001 \002(\t\022\030\n\020seriali" +
|
||||
"zedFilter\030\002 \001(\014\"%\n\024ColumnCountGetFilter\022" +
|
||||
"\r\n\005limit\030\001 \002(\005\"M\n\026ColumnPaginationFilter" +
|
||||
"\022\r\n\005limit\030\001 \002(\005\022\016\n\006offset\030\002 \001(\005\022\024\n\014colum" +
|
||||
"nOffset\030\003 \001(\014\"$\n\022ColumnPrefixFilter\022\016\n\006p" +
|
||||
"refix\030\001 \002(\014\"q\n\021ColumnRangeFilter\022\021\n\tminC" +
|
||||
"olumn\030\001 \001(\014\022\032\n\022minColumnInclusive\030\002 \001(\010\022" +
|
||||
"\021\n\tmaxColumn\030\003 \001(\014\022\032\n\022maxColumnInclusive" +
|
||||
"\030\004 \001(\010\"Q\n\rCompareFilter\022\037\n\tcompareOp\030\001 \002",
|
||||
"(\0162\014.CompareType\022\037\n\ncomparator\030\002 \001(\0132\013.C" +
|
||||
"omparator\"\212\001\n\025DependentColumnFilter\022%\n\rc" +
|
||||
"ompareFilter\030\001 \002(\0132\016.CompareFilter\022\024\n\014co" +
|
||||
"lumnFamily\030\002 \001(\014\022\027\n\017columnQualifier\030\003 \001(" +
|
||||
"\014\022\033\n\023dropDependentColumn\030\004 \001(\010\"5\n\014Family" +
|
||||
"Filter\022%\n\rcompareFilter\030\001 \002(\0132\016.CompareF" +
|
||||
"ilter\"\200\001\n\nFilterList\022&\n\010operator\030\001 \002(\0162\024" +
|
||||
".FilterList.Operator\022\030\n\007filters\030\002 \003(\0132\007." +
|
||||
"Filter\"0\n\010Operator\022\021\n\rMUST_PASS_ALL\020\001\022\021\n" +
|
||||
"\rMUST_PASS_ONE\020\002\"(\n\rFilterWrapper\022\027\n\006fil",
|
||||
"ter\030\001 \002(\0132\007.Filter\"\024\n\022FirstKeyOnlyFilter" +
|
||||
"\";\n%FirstKeyValueMatchingQualifiersFilte" +
|
||||
"r\022\022\n\nqualifiers\030\001 \003(\014\"8\n\016FuzzyRowFilter\022" +
|
||||
"&\n\rfuzzyKeysData\030\001 \003(\0132\017.BytesBytesPair\"" +
|
||||
")\n\023InclusiveStopFilter\022\022\n\nstopRowKey\030\001 \001" +
|
||||
"(\014\"!\n\rKeyOnlyFilter\022\020\n\010lenAsVal\030\001 \002(\010\"4\n" +
|
||||
"\032MultipleColumnPrefixFilter\022\026\n\016sortedPre" +
|
||||
"fixes\030\001 \003(\014\"\036\n\nPageFilter\022\020\n\010pageSize\030\001 " +
|
||||
"\002(\003\"\036\n\014PrefixFilter\022\016\n\006prefix\030\001 \001(\014\"8\n\017Q" +
|
||||
"ualifierFilter\022%\n\rcompareFilter\030\001 \002(\0132\016.",
|
||||
"CompareFilter\"!\n\017RandomRowFilter\022\016\n\006chan" +
|
||||
"ce\030\001 \002(\002\"2\n\tRowFilter\022%\n\rcompareFilter\030\001" +
|
||||
" \002(\0132\016.CompareFilter\"[\n\036SingleColumnValu" +
|
||||
"eExcludeFilter\0229\n\027singleColumnValueFilte" +
|
||||
"r\030\001 \002(\0132\030.SingleColumnValueFilter\"\276\001\n\027Si" +
|
||||
"ngleColumnValueFilter\022\024\n\014columnFamily\030\001 " +
|
||||
"\001(\014\022\027\n\017columnQualifier\030\002 \001(\014\022\037\n\tcompareO" +
|
||||
"p\030\003 \002(\0162\014.CompareType\022\037\n\ncomparator\030\004 \002(" +
|
||||
"\0132\013.Comparator\022\027\n\017filterIfMissing\030\005 \001(\010\022" +
|
||||
"\031\n\021latestVersionOnly\030\006 \001(\010\"%\n\nSkipFilter",
|
||||
"\022\027\n\006filter\030\001 \002(\0132\007.Filter\"*\n\020TimestampsF" +
|
||||
"ilter\022\026\n\ntimestamps\030\001 \003(\003B\002\020\001\"4\n\013ValueFi" +
|
||||
"lter\022%\n\rcompareFilter\030\001 \002(\0132\016.CompareFil" +
|
||||
"ter\"+\n\020WhileMatchFilter\022\027\n\006filter\030\001 \002(\0132" +
|
||||
"\007.FilterBB\n*org.apache.hadoop.hbase.prot" +
|
||||
"obuf.generatedB\014FilterProtosH\001\210\001\001\240\001\001"
|
||||
"proto\"1\n\006Filter\022\014\n\004name\030\001 \002(\t\022\031\n\021seriali" +
|
||||
"zed_filter\030\002 \001(\014\"%\n\024ColumnCountGetFilter" +
|
||||
"\022\r\n\005limit\030\001 \002(\005\"N\n\026ColumnPaginationFilte" +
|
||||
"r\022\r\n\005limit\030\001 \002(\005\022\016\n\006offset\030\002 \001(\005\022\025\n\rcolu" +
|
||||
"mn_offset\030\003 \001(\014\"$\n\022ColumnPrefixFilter\022\016\n" +
|
||||
"\006prefix\030\001 \002(\014\"w\n\021ColumnRangeFilter\022\022\n\nmi" +
|
||||
"n_column\030\001 \001(\014\022\034\n\024min_column_inclusive\030\002" +
|
||||
" \001(\010\022\022\n\nmax_column\030\003 \001(\014\022\034\n\024max_column_i" +
|
||||
"nclusive\030\004 \001(\010\"R\n\rCompareFilter\022 \n\ncompa",
|
||||
"re_op\030\001 \002(\0162\014.CompareType\022\037\n\ncomparator\030" +
|
||||
"\002 \001(\0132\013.Comparator\"\217\001\n\025DependentColumnFi" +
|
||||
"lter\022&\n\016compare_filter\030\001 \002(\0132\016.CompareFi" +
|
||||
"lter\022\025\n\rcolumn_family\030\002 \001(\014\022\030\n\020column_qu" +
|
||||
"alifier\030\003 \001(\014\022\035\n\025drop_dependent_column\030\004" +
|
||||
" \001(\010\"6\n\014FamilyFilter\022&\n\016compare_filter\030\001" +
|
||||
" \002(\0132\016.CompareFilter\"\200\001\n\nFilterList\022&\n\010o" +
|
||||
"perator\030\001 \002(\0162\024.FilterList.Operator\022\030\n\007f" +
|
||||
"ilters\030\002 \003(\0132\007.Filter\"0\n\010Operator\022\021\n\rMUS" +
|
||||
"T_PASS_ALL\020\001\022\021\n\rMUST_PASS_ONE\020\002\"(\n\rFilte",
|
||||
"rWrapper\022\027\n\006filter\030\001 \002(\0132\007.Filter\"\024\n\022Fir" +
|
||||
"stKeyOnlyFilter\";\n%FirstKeyValueMatching" +
|
||||
"QualifiersFilter\022\022\n\nqualifiers\030\001 \003(\014\":\n\016" +
|
||||
"FuzzyRowFilter\022(\n\017fuzzy_keys_data\030\001 \003(\0132" +
|
||||
"\017.BytesBytesPair\"+\n\023InclusiveStopFilter\022" +
|
||||
"\024\n\014stop_row_key\030\001 \001(\014\"#\n\rKeyOnlyFilter\022\022" +
|
||||
"\n\nlen_as_val\030\001 \002(\010\"5\n\032MultipleColumnPref" +
|
||||
"ixFilter\022\027\n\017sorted_prefixes\030\001 \003(\014\"\037\n\nPag" +
|
||||
"eFilter\022\021\n\tpage_size\030\001 \002(\003\"\036\n\014PrefixFilt" +
|
||||
"er\022\016\n\006prefix\030\001 \001(\014\"9\n\017QualifierFilter\022&\n",
|
||||
"\016compare_filter\030\001 \002(\0132\016.CompareFilter\"!\n" +
|
||||
"\017RandomRowFilter\022\016\n\006chance\030\001 \002(\002\"3\n\tRowF" +
|
||||
"ilter\022&\n\016compare_filter\030\001 \002(\0132\016.CompareF" +
|
||||
"ilter\"^\n\036SingleColumnValueExcludeFilter\022" +
|
||||
"<\n\032single_column_value_filter\030\001 \002(\0132\030.Si" +
|
||||
"ngleColumnValueFilter\"\305\001\n\027SingleColumnVa" +
|
||||
"lueFilter\022\025\n\rcolumn_family\030\001 \001(\014\022\030\n\020colu" +
|
||||
"mn_qualifier\030\002 \001(\014\022 \n\ncompare_op\030\003 \002(\0162\014" +
|
||||
".CompareType\022\037\n\ncomparator\030\004 \002(\0132\013.Compa" +
|
||||
"rator\022\031\n\021filter_if_missing\030\005 \001(\010\022\033\n\023late",
|
||||
"st_version_only\030\006 \001(\010\"%\n\nSkipFilter\022\027\n\006f" +
|
||||
"ilter\030\001 \002(\0132\007.Filter\"*\n\020TimestampsFilter" +
|
||||
"\022\026\n\ntimestamps\030\001 \003(\003B\002\020\001\"5\n\013ValueFilter\022" +
|
||||
"&\n\016compare_filter\030\001 \002(\0132\016.CompareFilter\"" +
|
||||
"+\n\020WhileMatchFilter\022\027\n\006filter\030\001 \002(\0132\007.Fi" +
|
||||
"lterBB\n*org.apache.hadoop.hbase.protobuf" +
|
||||
".generatedB\014FilterProtosH\001\210\001\001\240\001\001"
|
||||
};
|
||||
com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
|
||||
new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() {
|
||||
|
|
|
@ -109,7 +109,7 @@ public final class HBaseProtos {
|
|||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPairOrBuilder getAttributesOrBuilder(
|
||||
int index);
|
||||
|
||||
// repeated .ColumnFamilySchema columnFamilies = 3;
|
||||
// repeated .ColumnFamilySchema column_families = 3;
|
||||
java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema>
|
||||
getColumnFamiliesList();
|
||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema getColumnFamilies(int index);
|
||||
|
@ -189,8 +189,8 @@ public final class HBaseProtos {
|
|||
return attributes_.get(index);
|
||||
}
|
||||
|
||||
// repeated .ColumnFamilySchema columnFamilies = 3;
|
||||
public static final int COLUMNFAMILIES_FIELD_NUMBER = 3;
|
||||
// repeated .ColumnFamilySchema column_families = 3;
|
||||
public static final int COLUMN_FAMILIES_FIELD_NUMBER = 3;
|
||||
private java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema> columnFamilies_;
|
||||
public java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema> getColumnFamiliesList() {
|
||||
return columnFamilies_;
|
||||
|
@ -356,7 +356,7 @@ public final class HBaseProtos {
|
|||
hash = (53 * hash) + getAttributesList().hashCode();
|
||||
}
|
||||
if (getColumnFamiliesCount() > 0) {
|
||||
hash = (37 * hash) + COLUMNFAMILIES_FIELD_NUMBER;
|
||||
hash = (37 * hash) + COLUMN_FAMILIES_FIELD_NUMBER;
|
||||
hash = (53 * hash) + getColumnFamiliesList().hashCode();
|
||||
}
|
||||
if (getConfigurationCount() > 0) {
|
||||
|
@ -956,7 +956,7 @@ public final class HBaseProtos {
|
|||
return attributesBuilder_;
|
||||
}
|
||||
|
||||
// repeated .ColumnFamilySchema columnFamilies = 3;
|
||||
// repeated .ColumnFamilySchema column_families = 3;
|
||||
private java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema> columnFamilies_ =
|
||||
java.util.Collections.emptyList();
|
||||
private void ensureColumnFamiliesIsMutable() {
|
||||
|
@ -2306,19 +2306,19 @@ public final class HBaseProtos {
|
|||
public interface RegionInfoOrBuilder
|
||||
extends com.google.protobuf.MessageOrBuilder {
|
||||
|
||||
// required uint64 regionId = 1;
|
||||
// required uint64 region_id = 1;
|
||||
boolean hasRegionId();
|
||||
long getRegionId();
|
||||
|
||||
// required bytes tableName = 2;
|
||||
// required bytes table_name = 2;
|
||||
boolean hasTableName();
|
||||
com.google.protobuf.ByteString getTableName();
|
||||
|
||||
// optional bytes startKey = 3;
|
||||
// optional bytes start_key = 3;
|
||||
boolean hasStartKey();
|
||||
com.google.protobuf.ByteString getStartKey();
|
||||
|
||||
// optional bytes endKey = 4;
|
||||
// optional bytes end_key = 4;
|
||||
boolean hasEndKey();
|
||||
com.google.protobuf.ByteString getEndKey();
|
||||
|
||||
|
@ -2363,8 +2363,8 @@ public final class HBaseProtos {
|
|||
}
|
||||
|
||||
private int bitField0_;
|
||||
// required uint64 regionId = 1;
|
||||
public static final int REGIONID_FIELD_NUMBER = 1;
|
||||
// required uint64 region_id = 1;
|
||||
public static final int REGION_ID_FIELD_NUMBER = 1;
|
||||
private long regionId_;
|
||||
public boolean hasRegionId() {
|
||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||
|
@ -2373,8 +2373,8 @@ public final class HBaseProtos {
|
|||
return regionId_;
|
||||
}
|
||||
|
||||
// required bytes tableName = 2;
|
||||
public static final int TABLENAME_FIELD_NUMBER = 2;
|
||||
// required bytes table_name = 2;
|
||||
public static final int TABLE_NAME_FIELD_NUMBER = 2;
|
||||
private com.google.protobuf.ByteString tableName_;
|
||||
public boolean hasTableName() {
|
||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||
|
@ -2383,8 +2383,8 @@ public final class HBaseProtos {
|
|||
return tableName_;
|
||||
}
|
||||
|
||||
// optional bytes startKey = 3;
|
||||
public static final int STARTKEY_FIELD_NUMBER = 3;
|
||||
// optional bytes start_key = 3;
|
||||
public static final int START_KEY_FIELD_NUMBER = 3;
|
||||
private com.google.protobuf.ByteString startKey_;
|
||||
public boolean hasStartKey() {
|
||||
return ((bitField0_ & 0x00000004) == 0x00000004);
|
||||
|
@ -2393,8 +2393,8 @@ public final class HBaseProtos {
|
|||
return startKey_;
|
||||
}
|
||||
|
||||
// optional bytes endKey = 4;
|
||||
public static final int ENDKEY_FIELD_NUMBER = 4;
|
||||
// optional bytes end_key = 4;
|
||||
public static final int END_KEY_FIELD_NUMBER = 4;
|
||||
private com.google.protobuf.ByteString endKey_;
|
||||
public boolean hasEndKey() {
|
||||
return ((bitField0_ & 0x00000008) == 0x00000008);
|
||||
|
@ -2588,19 +2588,19 @@ public final class HBaseProtos {
|
|||
int hash = 41;
|
||||
hash = (19 * hash) + getDescriptorForType().hashCode();
|
||||
if (hasRegionId()) {
|
||||
hash = (37 * hash) + REGIONID_FIELD_NUMBER;
|
||||
hash = (37 * hash) + REGION_ID_FIELD_NUMBER;
|
||||
hash = (53 * hash) + hashLong(getRegionId());
|
||||
}
|
||||
if (hasTableName()) {
|
||||
hash = (37 * hash) + TABLENAME_FIELD_NUMBER;
|
||||
hash = (37 * hash) + TABLE_NAME_FIELD_NUMBER;
|
||||
hash = (53 * hash) + getTableName().hashCode();
|
||||
}
|
||||
if (hasStartKey()) {
|
||||
hash = (37 * hash) + STARTKEY_FIELD_NUMBER;
|
||||
hash = (37 * hash) + START_KEY_FIELD_NUMBER;
|
||||
hash = (53 * hash) + getStartKey().hashCode();
|
||||
}
|
||||
if (hasEndKey()) {
|
||||
hash = (37 * hash) + ENDKEY_FIELD_NUMBER;
|
||||
hash = (37 * hash) + END_KEY_FIELD_NUMBER;
|
||||
hash = (53 * hash) + getEndKey().hashCode();
|
||||
}
|
||||
if (hasOffline()) {
|
||||
|
@ -2928,7 +2928,7 @@ public final class HBaseProtos {
|
|||
|
||||
private int bitField0_;
|
||||
|
||||
// required uint64 regionId = 1;
|
||||
// required uint64 region_id = 1;
|
||||
private long regionId_ ;
|
||||
public boolean hasRegionId() {
|
||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||
|
@ -2949,7 +2949,7 @@ public final class HBaseProtos {
|
|||
return this;
|
||||
}
|
||||
|
||||
// required bytes tableName = 2;
|
||||
// required bytes table_name = 2;
|
||||
private com.google.protobuf.ByteString tableName_ = com.google.protobuf.ByteString.EMPTY;
|
||||
public boolean hasTableName() {
|
||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||
|
@ -2973,7 +2973,7 @@ public final class HBaseProtos {
|
|||
return this;
|
||||
}
|
||||
|
||||
// optional bytes startKey = 3;
|
||||
// optional bytes start_key = 3;
|
||||
private com.google.protobuf.ByteString startKey_ = com.google.protobuf.ByteString.EMPTY;
|
||||
public boolean hasStartKey() {
|
||||
return ((bitField0_ & 0x00000004) == 0x00000004);
|
||||
|
@ -2997,7 +2997,7 @@ public final class HBaseProtos {
|
|||
return this;
|
||||
}
|
||||
|
||||
// optional bytes endKey = 4;
|
||||
// optional bytes end_key = 4;
|
||||
private com.google.protobuf.ByteString endKey_ = com.google.protobuf.ByteString.EMPTY;
|
||||
public boolean hasEndKey() {
|
||||
return ((bitField0_ & 0x00000008) == 0x00000008);
|
||||
|
@ -3098,7 +3098,7 @@ public final class HBaseProtos {
|
|||
public interface FavoredNodesOrBuilder
|
||||
extends com.google.protobuf.MessageOrBuilder {
|
||||
|
||||
// repeated .ServerName favoredNode = 1;
|
||||
// repeated .ServerName favored_node = 1;
|
||||
java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName>
|
||||
getFavoredNodeList();
|
||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName getFavoredNode(int index);
|
||||
|
@ -3136,8 +3136,8 @@ public final class HBaseProtos {
|
|||
return org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.internal_static_FavoredNodes_fieldAccessorTable;
|
||||
}
|
||||
|
||||
// repeated .ServerName favoredNode = 1;
|
||||
public static final int FAVOREDNODE_FIELD_NUMBER = 1;
|
||||
// repeated .ServerName favored_node = 1;
|
||||
public static final int FAVORED_NODE_FIELD_NUMBER = 1;
|
||||
private java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName> favoredNode_;
|
||||
public java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName> getFavoredNodeList() {
|
||||
return favoredNode_;
|
||||
|
@ -3229,7 +3229,7 @@ public final class HBaseProtos {
|
|||
int hash = 41;
|
||||
hash = (19 * hash) + getDescriptorForType().hashCode();
|
||||
if (getFavoredNodeCount() > 0) {
|
||||
hash = (37 * hash) + FAVOREDNODE_FIELD_NUMBER;
|
||||
hash = (37 * hash) + FAVORED_NODE_FIELD_NUMBER;
|
||||
hash = (53 * hash) + getFavoredNodeList().hashCode();
|
||||
}
|
||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||
|
@ -3491,7 +3491,7 @@ public final class HBaseProtos {
|
|||
|
||||
private int bitField0_;
|
||||
|
||||
// repeated .ServerName favoredNode = 1;
|
||||
// repeated .ServerName favored_node = 1;
|
||||
private java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName> favoredNode_ =
|
||||
java.util.Collections.emptyList();
|
||||
private void ensureFavoredNodeIsMutable() {
|
||||
|
@ -4662,7 +4662,7 @@ public final class HBaseProtos {
|
|||
public interface ServerNameOrBuilder
|
||||
extends com.google.protobuf.MessageOrBuilder {
|
||||
|
||||
// required string hostName = 1;
|
||||
// required string host_name = 1;
|
||||
boolean hasHostName();
|
||||
String getHostName();
|
||||
|
||||
|
@ -4670,7 +4670,7 @@ public final class HBaseProtos {
|
|||
boolean hasPort();
|
||||
int getPort();
|
||||
|
||||
// optional uint64 startCode = 3;
|
||||
// optional uint64 start_code = 3;
|
||||
boolean hasStartCode();
|
||||
long getStartCode();
|
||||
}
|
||||
|
@ -4703,8 +4703,8 @@ public final class HBaseProtos {
|
|||
}
|
||||
|
||||
private int bitField0_;
|
||||
// required string hostName = 1;
|
||||
public static final int HOSTNAME_FIELD_NUMBER = 1;
|
||||
// required string host_name = 1;
|
||||
public static final int HOST_NAME_FIELD_NUMBER = 1;
|
||||
private java.lang.Object hostName_;
|
||||
public boolean hasHostName() {
|
||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||
|
@ -4745,8 +4745,8 @@ public final class HBaseProtos {
|
|||
return port_;
|
||||
}
|
||||
|
||||
// optional uint64 startCode = 3;
|
||||
public static final int STARTCODE_FIELD_NUMBER = 3;
|
||||
// optional uint64 start_code = 3;
|
||||
public static final int START_CODE_FIELD_NUMBER = 3;
|
||||
private long startCode_;
|
||||
public boolean hasStartCode() {
|
||||
return ((bitField0_ & 0x00000004) == 0x00000004);
|
||||
|
@ -4854,7 +4854,7 @@ public final class HBaseProtos {
|
|||
int hash = 41;
|
||||
hash = (19 * hash) + getDescriptorForType().hashCode();
|
||||
if (hasHostName()) {
|
||||
hash = (37 * hash) + HOSTNAME_FIELD_NUMBER;
|
||||
hash = (37 * hash) + HOST_NAME_FIELD_NUMBER;
|
||||
hash = (53 * hash) + getHostName().hashCode();
|
||||
}
|
||||
if (hasPort()) {
|
||||
|
@ -4862,7 +4862,7 @@ public final class HBaseProtos {
|
|||
hash = (53 * hash) + getPort();
|
||||
}
|
||||
if (hasStartCode()) {
|
||||
hash = (37 * hash) + STARTCODE_FIELD_NUMBER;
|
||||
hash = (37 * hash) + START_CODE_FIELD_NUMBER;
|
||||
hash = (53 * hash) + hashLong(getStartCode());
|
||||
}
|
||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||
|
@ -5118,7 +5118,7 @@ public final class HBaseProtos {
|
|||
|
||||
private int bitField0_;
|
||||
|
||||
// required string hostName = 1;
|
||||
// required string host_name = 1;
|
||||
private java.lang.Object hostName_ = "";
|
||||
public boolean hasHostName() {
|
||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||
|
@ -5175,7 +5175,7 @@ public final class HBaseProtos {
|
|||
return this;
|
||||
}
|
||||
|
||||
// optional uint64 startCode = 3;
|
||||
// optional uint64 start_code = 3;
|
||||
private long startCode_ ;
|
||||
public boolean hasStartCode() {
|
||||
return ((bitField0_ & 0x00000004) == 0x00000004);
|
||||
|
@ -7579,7 +7579,7 @@ public final class HBaseProtos {
|
|||
boolean hasTable();
|
||||
String getTable();
|
||||
|
||||
// optional int64 creationTime = 3 [default = 0];
|
||||
// optional int64 creation_time = 3 [default = 0];
|
||||
boolean hasCreationTime();
|
||||
long getCreationTime();
|
||||
|
||||
|
@ -7753,8 +7753,8 @@ public final class HBaseProtos {
|
|||
}
|
||||
}
|
||||
|
||||
// optional int64 creationTime = 3 [default = 0];
|
||||
public static final int CREATIONTIME_FIELD_NUMBER = 3;
|
||||
// optional int64 creation_time = 3 [default = 0];
|
||||
public static final int CREATION_TIME_FIELD_NUMBER = 3;
|
||||
private long creationTime_;
|
||||
public boolean hasCreationTime() {
|
||||
return ((bitField0_ & 0x00000004) == 0x00000004);
|
||||
|
@ -7916,7 +7916,7 @@ public final class HBaseProtos {
|
|||
hash = (53 * hash) + getTable().hashCode();
|
||||
}
|
||||
if (hasCreationTime()) {
|
||||
hash = (37 * hash) + CREATIONTIME_FIELD_NUMBER;
|
||||
hash = (37 * hash) + CREATION_TIME_FIELD_NUMBER;
|
||||
hash = (53 * hash) + hashLong(getCreationTime());
|
||||
}
|
||||
if (hasType()) {
|
||||
|
@ -8286,7 +8286,7 @@ public final class HBaseProtos {
|
|||
onChanged();
|
||||
}
|
||||
|
||||
// optional int64 creationTime = 3 [default = 0];
|
||||
// optional int64 creation_time = 3 [default = 0];
|
||||
private long creationTime_ ;
|
||||
public boolean hasCreationTime() {
|
||||
return ((bitField0_ & 0x00000004) == 0x00000004);
|
||||
|
@ -8666,7 +8666,7 @@ public final class HBaseProtos {
|
|||
public interface LongMsgOrBuilder
|
||||
extends com.google.protobuf.MessageOrBuilder {
|
||||
|
||||
// required int64 longMsg = 1;
|
||||
// required int64 long_msg = 1;
|
||||
boolean hasLongMsg();
|
||||
long getLongMsg();
|
||||
}
|
||||
|
@ -8699,8 +8699,8 @@ public final class HBaseProtos {
|
|||
}
|
||||
|
||||
private int bitField0_;
|
||||
// required int64 longMsg = 1;
|
||||
public static final int LONGMSG_FIELD_NUMBER = 1;
|
||||
// required int64 long_msg = 1;
|
||||
public static final int LONG_MSG_FIELD_NUMBER = 1;
|
||||
private long longMsg_;
|
||||
public boolean hasLongMsg() {
|
||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||
|
@ -8782,7 +8782,7 @@ public final class HBaseProtos {
|
|||
int hash = 41;
|
||||
hash = (19 * hash) + getDescriptorForType().hashCode();
|
||||
if (hasLongMsg()) {
|
||||
hash = (37 * hash) + LONGMSG_FIELD_NUMBER;
|
||||
hash = (37 * hash) + LONG_MSG_FIELD_NUMBER;
|
||||
hash = (53 * hash) + hashLong(getLongMsg());
|
||||
}
|
||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||
|
@ -9010,7 +9010,7 @@ public final class HBaseProtos {
|
|||
|
||||
private int bitField0_;
|
||||
|
||||
// required int64 longMsg = 1;
|
||||
// required int64 long_msg = 1;
|
||||
private long longMsg_ ;
|
||||
public boolean hasLongMsg() {
|
||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||
|
@ -9045,7 +9045,7 @@ public final class HBaseProtos {
|
|||
public interface BigDecimalMsgOrBuilder
|
||||
extends com.google.protobuf.MessageOrBuilder {
|
||||
|
||||
// required bytes bigdecimalMsg = 1;
|
||||
// required bytes bigdecimal_msg = 1;
|
||||
boolean hasBigdecimalMsg();
|
||||
com.google.protobuf.ByteString getBigdecimalMsg();
|
||||
}
|
||||
|
@ -9078,8 +9078,8 @@ public final class HBaseProtos {
|
|||
}
|
||||
|
||||
private int bitField0_;
|
||||
// required bytes bigdecimalMsg = 1;
|
||||
public static final int BIGDECIMALMSG_FIELD_NUMBER = 1;
|
||||
// required bytes bigdecimal_msg = 1;
|
||||
public static final int BIGDECIMAL_MSG_FIELD_NUMBER = 1;
|
||||
private com.google.protobuf.ByteString bigdecimalMsg_;
|
||||
public boolean hasBigdecimalMsg() {
|
||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||
|
@ -9161,7 +9161,7 @@ public final class HBaseProtos {
|
|||
int hash = 41;
|
||||
hash = (19 * hash) + getDescriptorForType().hashCode();
|
||||
if (hasBigdecimalMsg()) {
|
||||
hash = (37 * hash) + BIGDECIMALMSG_FIELD_NUMBER;
|
||||
hash = (37 * hash) + BIGDECIMAL_MSG_FIELD_NUMBER;
|
||||
hash = (53 * hash) + getBigdecimalMsg().hashCode();
|
||||
}
|
||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||
|
@ -9389,7 +9389,7 @@ public final class HBaseProtos {
|
|||
|
||||
private int bitField0_;
|
||||
|
||||
// required bytes bigdecimalMsg = 1;
|
||||
// required bytes bigdecimal_msg = 1;
|
||||
private com.google.protobuf.ByteString bigdecimalMsg_ = com.google.protobuf.ByteString.EMPTY;
|
||||
public boolean hasBigdecimalMsg() {
|
||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||
|
@ -9427,11 +9427,11 @@ public final class HBaseProtos {
|
|||
public interface UUIDOrBuilder
|
||||
extends com.google.protobuf.MessageOrBuilder {
|
||||
|
||||
// required uint64 leastSigBits = 1;
|
||||
// required uint64 least_sig_bits = 1;
|
||||
boolean hasLeastSigBits();
|
||||
long getLeastSigBits();
|
||||
|
||||
// required uint64 mostSigBits = 2;
|
||||
// required uint64 most_sig_bits = 2;
|
||||
boolean hasMostSigBits();
|
||||
long getMostSigBits();
|
||||
}
|
||||
|
@ -9464,8 +9464,8 @@ public final class HBaseProtos {
|
|||
}
|
||||
|
||||
private int bitField0_;
|
||||
// required uint64 leastSigBits = 1;
|
||||
public static final int LEASTSIGBITS_FIELD_NUMBER = 1;
|
||||
// required uint64 least_sig_bits = 1;
|
||||
public static final int LEAST_SIG_BITS_FIELD_NUMBER = 1;
|
||||
private long leastSigBits_;
|
||||
public boolean hasLeastSigBits() {
|
||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||
|
@ -9474,8 +9474,8 @@ public final class HBaseProtos {
|
|||
return leastSigBits_;
|
||||
}
|
||||
|
||||
// required uint64 mostSigBits = 2;
|
||||
public static final int MOSTSIGBITS_FIELD_NUMBER = 2;
|
||||
// required uint64 most_sig_bits = 2;
|
||||
public static final int MOST_SIG_BITS_FIELD_NUMBER = 2;
|
||||
private long mostSigBits_;
|
||||
public boolean hasMostSigBits() {
|
||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||
|
@ -9574,11 +9574,11 @@ public final class HBaseProtos {
|
|||
int hash = 41;
|
||||
hash = (19 * hash) + getDescriptorForType().hashCode();
|
||||
if (hasLeastSigBits()) {
|
||||
hash = (37 * hash) + LEASTSIGBITS_FIELD_NUMBER;
|
||||
hash = (37 * hash) + LEAST_SIG_BITS_FIELD_NUMBER;
|
||||
hash = (53 * hash) + hashLong(getLeastSigBits());
|
||||
}
|
||||
if (hasMostSigBits()) {
|
||||
hash = (37 * hash) + MOSTSIGBITS_FIELD_NUMBER;
|
||||
hash = (37 * hash) + MOST_SIG_BITS_FIELD_NUMBER;
|
||||
hash = (53 * hash) + hashLong(getMostSigBits());
|
||||
}
|
||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||
|
@ -9824,7 +9824,7 @@ public final class HBaseProtos {
|
|||
|
||||
private int bitField0_;
|
||||
|
||||
// required uint64 leastSigBits = 1;
|
||||
// required uint64 least_sig_bits = 1;
|
||||
private long leastSigBits_ ;
|
||||
public boolean hasLeastSigBits() {
|
||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||
|
@ -9845,7 +9845,7 @@ public final class HBaseProtos {
|
|||
return this;
|
||||
}
|
||||
|
||||
// required uint64 mostSigBits = 2;
|
||||
// required uint64 most_sig_bits = 2;
|
||||
private long mostSigBits_ ;
|
||||
public boolean hasMostSigBits() {
|
||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||
|
@ -9971,42 +9971,43 @@ public final class HBaseProtos {
|
|||
descriptor;
|
||||
static {
|
||||
java.lang.String[] descriptorData = {
|
||||
"\n\013hbase.proto\032\nCell.proto\"\225\001\n\013TableSchem" +
|
||||
"\n\013hbase.proto\032\nCell.proto\"\226\001\n\013TableSchem" +
|
||||
"a\022\014\n\004name\030\001 \001(\014\022#\n\nattributes\030\002 \003(\0132\017.By" +
|
||||
"tesBytesPair\022+\n\016columnFamilies\030\003 \003(\0132\023.C" +
|
||||
"olumnFamilySchema\022&\n\rconfiguration\030\004 \003(\013" +
|
||||
"2\017.NameStringPair\"o\n\022ColumnFamilySchema\022" +
|
||||
"\014\n\004name\030\001 \002(\014\022#\n\nattributes\030\002 \003(\0132\017.Byte" +
|
||||
"sBytesPair\022&\n\rconfiguration\030\003 \003(\0132\017.Name" +
|
||||
"StringPair\"\207\001\n\nRegionInfo\022\020\n\010regionId\030\001 " +
|
||||
"\002(\004\022\021\n\ttableName\030\002 \002(\014\022\020\n\010startKey\030\003 \001(\014" +
|
||||
"\022\016\n\006endKey\030\004 \001(\014\022\017\n\007offline\030\005 \001(\010\022\r\n\005spl",
|
||||
"it\030\006 \001(\010\022\022\n\nrecovering\030\007 \001(\010\"0\n\014FavoredN" +
|
||||
"odes\022 \n\013favoredNode\030\001 \003(\0132\013.ServerName\"\225" +
|
||||
"\001\n\017RegionSpecifier\0222\n\004type\030\001 \002(\0162$.Regio" +
|
||||
"nSpecifier.RegionSpecifierType\022\r\n\005value\030" +
|
||||
"\002 \002(\014\"?\n\023RegionSpecifierType\022\017\n\013REGION_N" +
|
||||
"AME\020\001\022\027\n\023ENCODED_REGION_NAME\020\002\"%\n\tTimeRa" +
|
||||
"nge\022\014\n\004from\030\001 \001(\004\022\n\n\002to\030\002 \001(\004\"?\n\nServerN" +
|
||||
"ame\022\020\n\010hostName\030\001 \002(\t\022\014\n\004port\030\002 \001(\r\022\021\n\ts" +
|
||||
"tartCode\030\003 \001(\004\"\033\n\013Coprocessor\022\014\n\004name\030\001 " +
|
||||
"\002(\t\"-\n\016NameStringPair\022\014\n\004name\030\001 \002(\t\022\r\n\005v",
|
||||
"alue\030\002 \002(\t\",\n\rNameBytesPair\022\014\n\004name\030\001 \002(" +
|
||||
"\t\022\r\n\005value\030\002 \001(\014\"/\n\016BytesBytesPair\022\r\n\005fi" +
|
||||
"rst\030\001 \002(\014\022\016\n\006second\030\002 \002(\014\",\n\rNameInt64Pa" +
|
||||
"ir\022\014\n\004name\030\001 \001(\t\022\r\n\005value\030\002 \001(\003\"\255\001\n\023Snap" +
|
||||
"shotDescription\022\014\n\004name\030\001 \002(\t\022\r\n\005table\030\002" +
|
||||
" \001(\t\022\027\n\014creationTime\030\003 \001(\003:\0010\022.\n\004type\030\004 " +
|
||||
"\001(\0162\031.SnapshotDescription.Type:\005FLUSH\022\017\n" +
|
||||
"\007version\030\005 \001(\005\"\037\n\004Type\022\014\n\010DISABLED\020\000\022\t\n\005" +
|
||||
"FLUSH\020\001\"\n\n\010EmptyMsg\"\032\n\007LongMsg\022\017\n\007longMs" +
|
||||
"g\030\001 \002(\003\"&\n\rBigDecimalMsg\022\025\n\rbigdecimalMs",
|
||||
"g\030\001 \002(\014\"1\n\004UUID\022\024\n\014leastSigBits\030\001 \002(\004\022\023\n" +
|
||||
"\013mostSigBits\030\002 \002(\004*r\n\013CompareType\022\010\n\004LES" +
|
||||
"S\020\000\022\021\n\rLESS_OR_EQUAL\020\001\022\t\n\005EQUAL\020\002\022\r\n\tNOT" +
|
||||
"_EQUAL\020\003\022\024\n\020GREATER_OR_EQUAL\020\004\022\013\n\007GREATE" +
|
||||
"R\020\005\022\t\n\005NO_OP\020\006B>\n*org.apache.hadoop.hbas" +
|
||||
"e.protobuf.generatedB\013HBaseProtosH\001\240\001\001"
|
||||
"tesBytesPair\022,\n\017column_families\030\003 \003(\0132\023." +
|
||||
"ColumnFamilySchema\022&\n\rconfiguration\030\004 \003(" +
|
||||
"\0132\017.NameStringPair\"o\n\022ColumnFamilySchema" +
|
||||
"\022\014\n\004name\030\001 \002(\014\022#\n\nattributes\030\002 \003(\0132\017.Byt" +
|
||||
"esBytesPair\022&\n\rconfiguration\030\003 \003(\0132\017.Nam" +
|
||||
"eStringPair\"\213\001\n\nRegionInfo\022\021\n\tregion_id\030" +
|
||||
"\001 \002(\004\022\022\n\ntable_name\030\002 \002(\014\022\021\n\tstart_key\030\003" +
|
||||
" \001(\014\022\017\n\007end_key\030\004 \001(\014\022\017\n\007offline\030\005 \001(\010\022\r",
|
||||
"\n\005split\030\006 \001(\010\022\022\n\nrecovering\030\007 \001(\010\"1\n\014Fav" +
|
||||
"oredNodes\022!\n\014favored_node\030\001 \003(\0132\013.Server" +
|
||||
"Name\"\225\001\n\017RegionSpecifier\0222\n\004type\030\001 \002(\0162$" +
|
||||
".RegionSpecifier.RegionSpecifierType\022\r\n\005" +
|
||||
"value\030\002 \002(\014\"?\n\023RegionSpecifierType\022\017\n\013RE" +
|
||||
"GION_NAME\020\001\022\027\n\023ENCODED_REGION_NAME\020\002\"%\n\t" +
|
||||
"TimeRange\022\014\n\004from\030\001 \001(\004\022\n\n\002to\030\002 \001(\004\"A\n\nS" +
|
||||
"erverName\022\021\n\thost_name\030\001 \002(\t\022\014\n\004port\030\002 \001" +
|
||||
"(\r\022\022\n\nstart_code\030\003 \001(\004\"\033\n\013Coprocessor\022\014\n" +
|
||||
"\004name\030\001 \002(\t\"-\n\016NameStringPair\022\014\n\004name\030\001 ",
|
||||
"\002(\t\022\r\n\005value\030\002 \002(\t\",\n\rNameBytesPair\022\014\n\004n" +
|
||||
"ame\030\001 \002(\t\022\r\n\005value\030\002 \001(\014\"/\n\016BytesBytesPa" +
|
||||
"ir\022\r\n\005first\030\001 \002(\014\022\016\n\006second\030\002 \002(\014\",\n\rNam" +
|
||||
"eInt64Pair\022\014\n\004name\030\001 \001(\t\022\r\n\005value\030\002 \001(\003\"" +
|
||||
"\256\001\n\023SnapshotDescription\022\014\n\004name\030\001 \002(\t\022\r\n" +
|
||||
"\005table\030\002 \001(\t\022\030\n\rcreation_time\030\003 \001(\003:\0010\022." +
|
||||
"\n\004type\030\004 \001(\0162\031.SnapshotDescription.Type:" +
|
||||
"\005FLUSH\022\017\n\007version\030\005 \001(\005\"\037\n\004Type\022\014\n\010DISAB" +
|
||||
"LED\020\000\022\t\n\005FLUSH\020\001\"\n\n\010EmptyMsg\"\033\n\007LongMsg\022" +
|
||||
"\020\n\010long_msg\030\001 \002(\003\"\'\n\rBigDecimalMsg\022\026\n\016bi",
|
||||
"gdecimal_msg\030\001 \002(\014\"5\n\004UUID\022\026\n\016least_sig_" +
|
||||
"bits\030\001 \002(\004\022\025\n\rmost_sig_bits\030\002 \002(\004*r\n\013Com" +
|
||||
"pareType\022\010\n\004LESS\020\000\022\021\n\rLESS_OR_EQUAL\020\001\022\t\n" +
|
||||
"\005EQUAL\020\002\022\r\n\tNOT_EQUAL\020\003\022\024\n\020GREATER_OR_EQ" +
|
||||
"UAL\020\004\022\013\n\007GREATER\020\005\022\t\n\005NO_OP\020\006B>\n*org.apa" +
|
||||
"che.hadoop.hbase.protobuf.generatedB\013HBa" +
|
||||
"seProtosH\001\240\001\001"
|
||||
};
|
||||
com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
|
||||
new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() {
|
||||
|
|
|
@ -11,7 +11,7 @@ public final class HFileProtos {
|
|||
public interface FileInfoProtoOrBuilder
|
||||
extends com.google.protobuf.MessageOrBuilder {
|
||||
|
||||
// repeated .BytesBytesPair mapEntry = 1;
|
||||
// repeated .BytesBytesPair map_entry = 1;
|
||||
java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair>
|
||||
getMapEntryList();
|
||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair getMapEntry(int index);
|
||||
|
@ -49,8 +49,8 @@ public final class HFileProtos {
|
|||
return org.apache.hadoop.hbase.protobuf.generated.HFileProtos.internal_static_FileInfoProto_fieldAccessorTable;
|
||||
}
|
||||
|
||||
// repeated .BytesBytesPair mapEntry = 1;
|
||||
public static final int MAPENTRY_FIELD_NUMBER = 1;
|
||||
// repeated .BytesBytesPair map_entry = 1;
|
||||
public static final int MAP_ENTRY_FIELD_NUMBER = 1;
|
||||
private java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair> mapEntry_;
|
||||
public java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair> getMapEntryList() {
|
||||
return mapEntry_;
|
||||
|
@ -142,7 +142,7 @@ public final class HFileProtos {
|
|||
int hash = 41;
|
||||
hash = (19 * hash) + getDescriptorForType().hashCode();
|
||||
if (getMapEntryCount() > 0) {
|
||||
hash = (37 * hash) + MAPENTRY_FIELD_NUMBER;
|
||||
hash = (37 * hash) + MAP_ENTRY_FIELD_NUMBER;
|
||||
hash = (53 * hash) + getMapEntryList().hashCode();
|
||||
}
|
||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||
|
@ -404,7 +404,7 @@ public final class HFileProtos {
|
|||
|
||||
private int bitField0_;
|
||||
|
||||
// repeated .BytesBytesPair mapEntry = 1;
|
||||
// repeated .BytesBytesPair map_entry = 1;
|
||||
private java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair> mapEntry_ =
|
||||
java.util.Collections.emptyList();
|
||||
private void ensureMapEntryIsMutable() {
|
||||
|
@ -604,51 +604,51 @@ public final class HFileProtos {
|
|||
public interface FileTrailerProtoOrBuilder
|
||||
extends com.google.protobuf.MessageOrBuilder {
|
||||
|
||||
// optional uint64 fileInfoOffset = 1;
|
||||
// optional uint64 file_info_offset = 1;
|
||||
boolean hasFileInfoOffset();
|
||||
long getFileInfoOffset();
|
||||
|
||||
// optional uint64 loadOnOpenDataOffset = 2;
|
||||
// optional uint64 load_on_open_data_offset = 2;
|
||||
boolean hasLoadOnOpenDataOffset();
|
||||
long getLoadOnOpenDataOffset();
|
||||
|
||||
// optional uint64 uncompressedDataIndexSize = 3;
|
||||
// optional uint64 uncompressed_data_index_size = 3;
|
||||
boolean hasUncompressedDataIndexSize();
|
||||
long getUncompressedDataIndexSize();
|
||||
|
||||
// optional uint64 totalUncompressedBytes = 4;
|
||||
// optional uint64 total_uncompressed_bytes = 4;
|
||||
boolean hasTotalUncompressedBytes();
|
||||
long getTotalUncompressedBytes();
|
||||
|
||||
// optional uint32 dataIndexCount = 5;
|
||||
// optional uint32 data_index_count = 5;
|
||||
boolean hasDataIndexCount();
|
||||
int getDataIndexCount();
|
||||
|
||||
// optional uint32 metaIndexCount = 6;
|
||||
// optional uint32 meta_index_count = 6;
|
||||
boolean hasMetaIndexCount();
|
||||
int getMetaIndexCount();
|
||||
|
||||
// optional uint64 entryCount = 7;
|
||||
// optional uint64 entry_count = 7;
|
||||
boolean hasEntryCount();
|
||||
long getEntryCount();
|
||||
|
||||
// optional uint32 numDataIndexLevels = 8;
|
||||
// optional uint32 num_data_index_levels = 8;
|
||||
boolean hasNumDataIndexLevels();
|
||||
int getNumDataIndexLevels();
|
||||
|
||||
// optional uint64 firstDataBlockOffset = 9;
|
||||
// optional uint64 first_data_block_offset = 9;
|
||||
boolean hasFirstDataBlockOffset();
|
||||
long getFirstDataBlockOffset();
|
||||
|
||||
// optional uint64 lastDataBlockOffset = 10;
|
||||
// optional uint64 last_data_block_offset = 10;
|
||||
boolean hasLastDataBlockOffset();
|
||||
long getLastDataBlockOffset();
|
||||
|
||||
// optional string comparatorClassName = 11;
|
||||
// optional string comparator_class_name = 11;
|
||||
boolean hasComparatorClassName();
|
||||
String getComparatorClassName();
|
||||
|
||||
// optional uint32 compressionCodec = 12;
|
||||
// optional uint32 compression_codec = 12;
|
||||
boolean hasCompressionCodec();
|
||||
int getCompressionCodec();
|
||||
}
|
||||
|
@ -681,8 +681,8 @@ public final class HFileProtos {
|
|||
}
|
||||
|
||||
private int bitField0_;
|
||||
// optional uint64 fileInfoOffset = 1;
|
||||
public static final int FILEINFOOFFSET_FIELD_NUMBER = 1;
|
||||
// optional uint64 file_info_offset = 1;
|
||||
public static final int FILE_INFO_OFFSET_FIELD_NUMBER = 1;
|
||||
private long fileInfoOffset_;
|
||||
public boolean hasFileInfoOffset() {
|
||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||
|
@ -691,8 +691,8 @@ public final class HFileProtos {
|
|||
return fileInfoOffset_;
|
||||
}
|
||||
|
||||
// optional uint64 loadOnOpenDataOffset = 2;
|
||||
public static final int LOADONOPENDATAOFFSET_FIELD_NUMBER = 2;
|
||||
// optional uint64 load_on_open_data_offset = 2;
|
||||
public static final int LOAD_ON_OPEN_DATA_OFFSET_FIELD_NUMBER = 2;
|
||||
private long loadOnOpenDataOffset_;
|
||||
public boolean hasLoadOnOpenDataOffset() {
|
||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||
|
@ -701,8 +701,8 @@ public final class HFileProtos {
|
|||
return loadOnOpenDataOffset_;
|
||||
}
|
||||
|
||||
// optional uint64 uncompressedDataIndexSize = 3;
|
||||
public static final int UNCOMPRESSEDDATAINDEXSIZE_FIELD_NUMBER = 3;
|
||||
// optional uint64 uncompressed_data_index_size = 3;
|
||||
public static final int UNCOMPRESSED_DATA_INDEX_SIZE_FIELD_NUMBER = 3;
|
||||
private long uncompressedDataIndexSize_;
|
||||
public boolean hasUncompressedDataIndexSize() {
|
||||
return ((bitField0_ & 0x00000004) == 0x00000004);
|
||||
|
@ -711,8 +711,8 @@ public final class HFileProtos {
|
|||
return uncompressedDataIndexSize_;
|
||||
}
|
||||
|
||||
// optional uint64 totalUncompressedBytes = 4;
|
||||
public static final int TOTALUNCOMPRESSEDBYTES_FIELD_NUMBER = 4;
|
||||
// optional uint64 total_uncompressed_bytes = 4;
|
||||
public static final int TOTAL_UNCOMPRESSED_BYTES_FIELD_NUMBER = 4;
|
||||
private long totalUncompressedBytes_;
|
||||
public boolean hasTotalUncompressedBytes() {
|
||||
return ((bitField0_ & 0x00000008) == 0x00000008);
|
||||
|
@ -721,8 +721,8 @@ public final class HFileProtos {
|
|||
return totalUncompressedBytes_;
|
||||
}
|
||||
|
||||
// optional uint32 dataIndexCount = 5;
|
||||
public static final int DATAINDEXCOUNT_FIELD_NUMBER = 5;
|
||||
// optional uint32 data_index_count = 5;
|
||||
public static final int DATA_INDEX_COUNT_FIELD_NUMBER = 5;
|
||||
private int dataIndexCount_;
|
||||
public boolean hasDataIndexCount() {
|
||||
return ((bitField0_ & 0x00000010) == 0x00000010);
|
||||
|
@ -731,8 +731,8 @@ public final class HFileProtos {
|
|||
return dataIndexCount_;
|
||||
}
|
||||
|
||||
// optional uint32 metaIndexCount = 6;
|
||||
public static final int METAINDEXCOUNT_FIELD_NUMBER = 6;
|
||||
// optional uint32 meta_index_count = 6;
|
||||
public static final int META_INDEX_COUNT_FIELD_NUMBER = 6;
|
||||
private int metaIndexCount_;
|
||||
public boolean hasMetaIndexCount() {
|
||||
return ((bitField0_ & 0x00000020) == 0x00000020);
|
||||
|
@ -741,8 +741,8 @@ public final class HFileProtos {
|
|||
return metaIndexCount_;
|
||||
}
|
||||
|
||||
// optional uint64 entryCount = 7;
|
||||
public static final int ENTRYCOUNT_FIELD_NUMBER = 7;
|
||||
// optional uint64 entry_count = 7;
|
||||
public static final int ENTRY_COUNT_FIELD_NUMBER = 7;
|
||||
private long entryCount_;
|
||||
public boolean hasEntryCount() {
|
||||
return ((bitField0_ & 0x00000040) == 0x00000040);
|
||||
|
@ -751,8 +751,8 @@ public final class HFileProtos {
|
|||
return entryCount_;
|
||||
}
|
||||
|
||||
// optional uint32 numDataIndexLevels = 8;
|
||||
public static final int NUMDATAINDEXLEVELS_FIELD_NUMBER = 8;
|
||||
// optional uint32 num_data_index_levels = 8;
|
||||
public static final int NUM_DATA_INDEX_LEVELS_FIELD_NUMBER = 8;
|
||||
private int numDataIndexLevels_;
|
||||
public boolean hasNumDataIndexLevels() {
|
||||
return ((bitField0_ & 0x00000080) == 0x00000080);
|
||||
|
@ -761,8 +761,8 @@ public final class HFileProtos {
|
|||
return numDataIndexLevels_;
|
||||
}
|
||||
|
||||
// optional uint64 firstDataBlockOffset = 9;
|
||||
public static final int FIRSTDATABLOCKOFFSET_FIELD_NUMBER = 9;
|
||||
// optional uint64 first_data_block_offset = 9;
|
||||
public static final int FIRST_DATA_BLOCK_OFFSET_FIELD_NUMBER = 9;
|
||||
private long firstDataBlockOffset_;
|
||||
public boolean hasFirstDataBlockOffset() {
|
||||
return ((bitField0_ & 0x00000100) == 0x00000100);
|
||||
|
@ -771,8 +771,8 @@ public final class HFileProtos {
|
|||
return firstDataBlockOffset_;
|
||||
}
|
||||
|
||||
// optional uint64 lastDataBlockOffset = 10;
|
||||
public static final int LASTDATABLOCKOFFSET_FIELD_NUMBER = 10;
|
||||
// optional uint64 last_data_block_offset = 10;
|
||||
public static final int LAST_DATA_BLOCK_OFFSET_FIELD_NUMBER = 10;
|
||||
private long lastDataBlockOffset_;
|
||||
public boolean hasLastDataBlockOffset() {
|
||||
return ((bitField0_ & 0x00000200) == 0x00000200);
|
||||
|
@ -781,8 +781,8 @@ public final class HFileProtos {
|
|||
return lastDataBlockOffset_;
|
||||
}
|
||||
|
||||
// optional string comparatorClassName = 11;
|
||||
public static final int COMPARATORCLASSNAME_FIELD_NUMBER = 11;
|
||||
// optional string comparator_class_name = 11;
|
||||
public static final int COMPARATOR_CLASS_NAME_FIELD_NUMBER = 11;
|
||||
private java.lang.Object comparatorClassName_;
|
||||
public boolean hasComparatorClassName() {
|
||||
return ((bitField0_ & 0x00000400) == 0x00000400);
|
||||
|
@ -813,8 +813,8 @@ public final class HFileProtos {
|
|||
}
|
||||
}
|
||||
|
||||
// optional uint32 compressionCodec = 12;
|
||||
public static final int COMPRESSIONCODEC_FIELD_NUMBER = 12;
|
||||
// optional uint32 compression_codec = 12;
|
||||
public static final int COMPRESSION_CODEC_FIELD_NUMBER = 12;
|
||||
private int compressionCodec_;
|
||||
public boolean hasCompressionCodec() {
|
||||
return ((bitField0_ & 0x00000800) == 0x00000800);
|
||||
|
@ -1035,51 +1035,51 @@ public final class HFileProtos {
|
|||
int hash = 41;
|
||||
hash = (19 * hash) + getDescriptorForType().hashCode();
|
||||
if (hasFileInfoOffset()) {
|
||||
hash = (37 * hash) + FILEINFOOFFSET_FIELD_NUMBER;
|
||||
hash = (37 * hash) + FILE_INFO_OFFSET_FIELD_NUMBER;
|
||||
hash = (53 * hash) + hashLong(getFileInfoOffset());
|
||||
}
|
||||
if (hasLoadOnOpenDataOffset()) {
|
||||
hash = (37 * hash) + LOADONOPENDATAOFFSET_FIELD_NUMBER;
|
||||
hash = (37 * hash) + LOAD_ON_OPEN_DATA_OFFSET_FIELD_NUMBER;
|
||||
hash = (53 * hash) + hashLong(getLoadOnOpenDataOffset());
|
||||
}
|
||||
if (hasUncompressedDataIndexSize()) {
|
||||
hash = (37 * hash) + UNCOMPRESSEDDATAINDEXSIZE_FIELD_NUMBER;
|
||||
hash = (37 * hash) + UNCOMPRESSED_DATA_INDEX_SIZE_FIELD_NUMBER;
|
||||
hash = (53 * hash) + hashLong(getUncompressedDataIndexSize());
|
||||
}
|
||||
if (hasTotalUncompressedBytes()) {
|
||||
hash = (37 * hash) + TOTALUNCOMPRESSEDBYTES_FIELD_NUMBER;
|
||||
hash = (37 * hash) + TOTAL_UNCOMPRESSED_BYTES_FIELD_NUMBER;
|
||||
hash = (53 * hash) + hashLong(getTotalUncompressedBytes());
|
||||
}
|
||||
if (hasDataIndexCount()) {
|
||||
hash = (37 * hash) + DATAINDEXCOUNT_FIELD_NUMBER;
|
||||
hash = (37 * hash) + DATA_INDEX_COUNT_FIELD_NUMBER;
|
||||
hash = (53 * hash) + getDataIndexCount();
|
||||
}
|
||||
if (hasMetaIndexCount()) {
|
||||
hash = (37 * hash) + METAINDEXCOUNT_FIELD_NUMBER;
|
||||
hash = (37 * hash) + META_INDEX_COUNT_FIELD_NUMBER;
|
||||
hash = (53 * hash) + getMetaIndexCount();
|
||||
}
|
||||
if (hasEntryCount()) {
|
||||
hash = (37 * hash) + ENTRYCOUNT_FIELD_NUMBER;
|
||||
hash = (37 * hash) + ENTRY_COUNT_FIELD_NUMBER;
|
||||
hash = (53 * hash) + hashLong(getEntryCount());
|
||||
}
|
||||
if (hasNumDataIndexLevels()) {
|
||||
hash = (37 * hash) + NUMDATAINDEXLEVELS_FIELD_NUMBER;
|
||||
hash = (37 * hash) + NUM_DATA_INDEX_LEVELS_FIELD_NUMBER;
|
||||
hash = (53 * hash) + getNumDataIndexLevels();
|
||||
}
|
||||
if (hasFirstDataBlockOffset()) {
|
||||
hash = (37 * hash) + FIRSTDATABLOCKOFFSET_FIELD_NUMBER;
|
||||
hash = (37 * hash) + FIRST_DATA_BLOCK_OFFSET_FIELD_NUMBER;
|
||||
hash = (53 * hash) + hashLong(getFirstDataBlockOffset());
|
||||
}
|
||||
if (hasLastDataBlockOffset()) {
|
||||
hash = (37 * hash) + LASTDATABLOCKOFFSET_FIELD_NUMBER;
|
||||
hash = (37 * hash) + LAST_DATA_BLOCK_OFFSET_FIELD_NUMBER;
|
||||
hash = (53 * hash) + hashLong(getLastDataBlockOffset());
|
||||
}
|
||||
if (hasComparatorClassName()) {
|
||||
hash = (37 * hash) + COMPARATORCLASSNAME_FIELD_NUMBER;
|
||||
hash = (37 * hash) + COMPARATOR_CLASS_NAME_FIELD_NUMBER;
|
||||
hash = (53 * hash) + getComparatorClassName().hashCode();
|
||||
}
|
||||
if (hasCompressionCodec()) {
|
||||
hash = (37 * hash) + COMPRESSIONCODEC_FIELD_NUMBER;
|
||||
hash = (37 * hash) + COMPRESSION_CODEC_FIELD_NUMBER;
|
||||
hash = (53 * hash) + getCompressionCodec();
|
||||
}
|
||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||
|
@ -1457,7 +1457,7 @@ public final class HFileProtos {
|
|||
|
||||
private int bitField0_;
|
||||
|
||||
// optional uint64 fileInfoOffset = 1;
|
||||
// optional uint64 file_info_offset = 1;
|
||||
private long fileInfoOffset_ ;
|
||||
public boolean hasFileInfoOffset() {
|
||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||
|
@ -1478,7 +1478,7 @@ public final class HFileProtos {
|
|||
return this;
|
||||
}
|
||||
|
||||
// optional uint64 loadOnOpenDataOffset = 2;
|
||||
// optional uint64 load_on_open_data_offset = 2;
|
||||
private long loadOnOpenDataOffset_ ;
|
||||
public boolean hasLoadOnOpenDataOffset() {
|
||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||
|
@ -1499,7 +1499,7 @@ public final class HFileProtos {
|
|||
return this;
|
||||
}
|
||||
|
||||
// optional uint64 uncompressedDataIndexSize = 3;
|
||||
// optional uint64 uncompressed_data_index_size = 3;
|
||||
private long uncompressedDataIndexSize_ ;
|
||||
public boolean hasUncompressedDataIndexSize() {
|
||||
return ((bitField0_ & 0x00000004) == 0x00000004);
|
||||
|
@ -1520,7 +1520,7 @@ public final class HFileProtos {
|
|||
return this;
|
||||
}
|
||||
|
||||
// optional uint64 totalUncompressedBytes = 4;
|
||||
// optional uint64 total_uncompressed_bytes = 4;
|
||||
private long totalUncompressedBytes_ ;
|
||||
public boolean hasTotalUncompressedBytes() {
|
||||
return ((bitField0_ & 0x00000008) == 0x00000008);
|
||||
|
@ -1541,7 +1541,7 @@ public final class HFileProtos {
|
|||
return this;
|
||||
}
|
||||
|
||||
// optional uint32 dataIndexCount = 5;
|
||||
// optional uint32 data_index_count = 5;
|
||||
private int dataIndexCount_ ;
|
||||
public boolean hasDataIndexCount() {
|
||||
return ((bitField0_ & 0x00000010) == 0x00000010);
|
||||
|
@ -1562,7 +1562,7 @@ public final class HFileProtos {
|
|||
return this;
|
||||
}
|
||||
|
||||
// optional uint32 metaIndexCount = 6;
|
||||
// optional uint32 meta_index_count = 6;
|
||||
private int metaIndexCount_ ;
|
||||
public boolean hasMetaIndexCount() {
|
||||
return ((bitField0_ & 0x00000020) == 0x00000020);
|
||||
|
@ -1583,7 +1583,7 @@ public final class HFileProtos {
|
|||
return this;
|
||||
}
|
||||
|
||||
// optional uint64 entryCount = 7;
|
||||
// optional uint64 entry_count = 7;
|
||||
private long entryCount_ ;
|
||||
public boolean hasEntryCount() {
|
||||
return ((bitField0_ & 0x00000040) == 0x00000040);
|
||||
|
@ -1604,7 +1604,7 @@ public final class HFileProtos {
|
|||
return this;
|
||||
}
|
||||
|
||||
// optional uint32 numDataIndexLevels = 8;
|
||||
// optional uint32 num_data_index_levels = 8;
|
||||
private int numDataIndexLevels_ ;
|
||||
public boolean hasNumDataIndexLevels() {
|
||||
return ((bitField0_ & 0x00000080) == 0x00000080);
|
||||
|
@ -1625,7 +1625,7 @@ public final class HFileProtos {
|
|||
return this;
|
||||
}
|
||||
|
||||
// optional uint64 firstDataBlockOffset = 9;
|
||||
// optional uint64 first_data_block_offset = 9;
|
||||
private long firstDataBlockOffset_ ;
|
||||
public boolean hasFirstDataBlockOffset() {
|
||||
return ((bitField0_ & 0x00000100) == 0x00000100);
|
||||
|
@ -1646,7 +1646,7 @@ public final class HFileProtos {
|
|||
return this;
|
||||
}
|
||||
|
||||
// optional uint64 lastDataBlockOffset = 10;
|
||||
// optional uint64 last_data_block_offset = 10;
|
||||
private long lastDataBlockOffset_ ;
|
||||
public boolean hasLastDataBlockOffset() {
|
||||
return ((bitField0_ & 0x00000200) == 0x00000200);
|
||||
|
@ -1667,7 +1667,7 @@ public final class HFileProtos {
|
|||
return this;
|
||||
}
|
||||
|
||||
// optional string comparatorClassName = 11;
|
||||
// optional string comparator_class_name = 11;
|
||||
private java.lang.Object comparatorClassName_ = "";
|
||||
public boolean hasComparatorClassName() {
|
||||
return ((bitField0_ & 0x00000400) == 0x00000400);
|
||||
|
@ -1703,7 +1703,7 @@ public final class HFileProtos {
|
|||
onChanged();
|
||||
}
|
||||
|
||||
// optional uint32 compressionCodec = 12;
|
||||
// optional uint32 compression_codec = 12;
|
||||
private int compressionCodec_ ;
|
||||
public boolean hasCompressionCodec() {
|
||||
return ((bitField0_ & 0x00000800) == 0x00000800);
|
||||
|
@ -1754,19 +1754,20 @@ public final class HFileProtos {
|
|||
descriptor;
|
||||
static {
|
||||
java.lang.String[] descriptorData = {
|
||||
"\n\013HFile.proto\032\013hbase.proto\"2\n\rFileInfoPr" +
|
||||
"oto\022!\n\010mapEntry\030\001 \003(\0132\017.BytesBytesPair\"\335" +
|
||||
"\002\n\020FileTrailerProto\022\026\n\016fileInfoOffset\030\001 " +
|
||||
"\001(\004\022\034\n\024loadOnOpenDataOffset\030\002 \001(\004\022!\n\031unc" +
|
||||
"ompressedDataIndexSize\030\003 \001(\004\022\036\n\026totalUnc" +
|
||||
"ompressedBytes\030\004 \001(\004\022\026\n\016dataIndexCount\030\005" +
|
||||
" \001(\r\022\026\n\016metaIndexCount\030\006 \001(\r\022\022\n\nentryCou" +
|
||||
"nt\030\007 \001(\004\022\032\n\022numDataIndexLevels\030\010 \001(\r\022\034\n\024" +
|
||||
"firstDataBlockOffset\030\t \001(\004\022\033\n\023lastDataBl" +
|
||||
"ockOffset\030\n \001(\004\022\033\n\023comparatorClassName\030\013",
|
||||
" \001(\t\022\030\n\020compressionCodec\030\014 \001(\rBA\n*org.ap" +
|
||||
"ache.hadoop.hbase.protobuf.generatedB\013HF" +
|
||||
"ileProtosH\001\210\001\001\240\001\001"
|
||||
"\n\013HFile.proto\032\013hbase.proto\"3\n\rFileInfoPr" +
|
||||
"oto\022\"\n\tmap_entry\030\001 \003(\0132\017.BytesBytesPair\"" +
|
||||
"\371\002\n\020FileTrailerProto\022\030\n\020file_info_offset" +
|
||||
"\030\001 \001(\004\022 \n\030load_on_open_data_offset\030\002 \001(\004" +
|
||||
"\022$\n\034uncompressed_data_index_size\030\003 \001(\004\022 " +
|
||||
"\n\030total_uncompressed_bytes\030\004 \001(\004\022\030\n\020data" +
|
||||
"_index_count\030\005 \001(\r\022\030\n\020meta_index_count\030\006" +
|
||||
" \001(\r\022\023\n\013entry_count\030\007 \001(\004\022\035\n\025num_data_in" +
|
||||
"dex_levels\030\010 \001(\r\022\037\n\027first_data_block_off" +
|
||||
"set\030\t \001(\004\022\036\n\026last_data_block_offset\030\n \001(",
|
||||
"\004\022\035\n\025comparator_class_name\030\013 \001(\t\022\031\n\021comp" +
|
||||
"ression_codec\030\014 \001(\rBA\n*org.apache.hadoop" +
|
||||
".hbase.protobuf.generatedB\013HFileProtosH\001" +
|
||||
"\210\001\001\240\001\001"
|
||||
};
|
||||
com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
|
||||
new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() {
|
||||
|
|
|
@ -11,7 +11,7 @@ public final class LoadBalancerProtos {
|
|||
public interface LoadBalancerStateOrBuilder
|
||||
extends com.google.protobuf.MessageOrBuilder {
|
||||
|
||||
// optional bool balancerOn = 1;
|
||||
// optional bool balancer_on = 1;
|
||||
boolean hasBalancerOn();
|
||||
boolean getBalancerOn();
|
||||
}
|
||||
|
@ -44,8 +44,8 @@ public final class LoadBalancerProtos {
|
|||
}
|
||||
|
||||
private int bitField0_;
|
||||
// optional bool balancerOn = 1;
|
||||
public static final int BALANCERON_FIELD_NUMBER = 1;
|
||||
// optional bool balancer_on = 1;
|
||||
public static final int BALANCER_ON_FIELD_NUMBER = 1;
|
||||
private boolean balancerOn_;
|
||||
public boolean hasBalancerOn() {
|
||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||
|
@ -123,7 +123,7 @@ public final class LoadBalancerProtos {
|
|||
int hash = 41;
|
||||
hash = (19 * hash) + getDescriptorForType().hashCode();
|
||||
if (hasBalancerOn()) {
|
||||
hash = (37 * hash) + BALANCERON_FIELD_NUMBER;
|
||||
hash = (37 * hash) + BALANCER_ON_FIELD_NUMBER;
|
||||
hash = (53 * hash) + hashBoolean(getBalancerOn());
|
||||
}
|
||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||
|
@ -347,7 +347,7 @@ public final class LoadBalancerProtos {
|
|||
|
||||
private int bitField0_;
|
||||
|
||||
// optional bool balancerOn = 1;
|
||||
// optional bool balancer_on = 1;
|
||||
private boolean balancerOn_ ;
|
||||
public boolean hasBalancerOn() {
|
||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||
|
@ -393,10 +393,10 @@ public final class LoadBalancerProtos {
|
|||
descriptor;
|
||||
static {
|
||||
java.lang.String[] descriptorData = {
|
||||
"\n\022LoadBalancer.proto\"\'\n\021LoadBalancerStat" +
|
||||
"e\022\022\n\nbalancerOn\030\001 \001(\010BE\n*org.apache.hado" +
|
||||
"op.hbase.protobuf.generatedB\022LoadBalance" +
|
||||
"rProtosH\001\240\001\001"
|
||||
"\n\022LoadBalancer.proto\"(\n\021LoadBalancerStat" +
|
||||
"e\022\023\n\013balancer_on\030\001 \001(\010BE\n*org.apache.had" +
|
||||
"oop.hbase.protobuf.generatedB\022LoadBalanc" +
|
||||
"erProtosH\001\240\001\001"
|
||||
};
|
||||
com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
|
||||
new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() {
|
||||
|
|
|
@ -11,11 +11,11 @@ public final class MasterAdminProtos {
|
|||
public interface AddColumnRequestOrBuilder
|
||||
extends com.google.protobuf.MessageOrBuilder {
|
||||
|
||||
// required bytes tableName = 1;
|
||||
// required bytes table_name = 1;
|
||||
boolean hasTableName();
|
||||
com.google.protobuf.ByteString getTableName();
|
||||
|
||||
// required .ColumnFamilySchema columnFamilies = 2;
|
||||
// required .ColumnFamilySchema column_families = 2;
|
||||
boolean hasColumnFamilies();
|
||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema getColumnFamilies();
|
||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchemaOrBuilder getColumnFamiliesOrBuilder();
|
||||
|
@ -49,8 +49,8 @@ public final class MasterAdminProtos {
|
|||
}
|
||||
|
||||
private int bitField0_;
|
||||
// required bytes tableName = 1;
|
||||
public static final int TABLENAME_FIELD_NUMBER = 1;
|
||||
// required bytes table_name = 1;
|
||||
public static final int TABLE_NAME_FIELD_NUMBER = 1;
|
||||
private com.google.protobuf.ByteString tableName_;
|
||||
public boolean hasTableName() {
|
||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||
|
@ -59,8 +59,8 @@ public final class MasterAdminProtos {
|
|||
return tableName_;
|
||||
}
|
||||
|
||||
// required .ColumnFamilySchema columnFamilies = 2;
|
||||
public static final int COLUMNFAMILIES_FIELD_NUMBER = 2;
|
||||
// required .ColumnFamilySchema column_families = 2;
|
||||
public static final int COLUMN_FAMILIES_FIELD_NUMBER = 2;
|
||||
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema columnFamilies_;
|
||||
public boolean hasColumnFamilies() {
|
||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||
|
@ -166,11 +166,11 @@ public final class MasterAdminProtos {
|
|||
int hash = 41;
|
||||
hash = (19 * hash) + getDescriptorForType().hashCode();
|
||||
if (hasTableName()) {
|
||||
hash = (37 * hash) + TABLENAME_FIELD_NUMBER;
|
||||
hash = (37 * hash) + TABLE_NAME_FIELD_NUMBER;
|
||||
hash = (53 * hash) + getTableName().hashCode();
|
||||
}
|
||||
if (hasColumnFamilies()) {
|
||||
hash = (37 * hash) + COLUMNFAMILIES_FIELD_NUMBER;
|
||||
hash = (37 * hash) + COLUMN_FAMILIES_FIELD_NUMBER;
|
||||
hash = (53 * hash) + getColumnFamilies().hashCode();
|
||||
}
|
||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||
|
@ -433,7 +433,7 @@ public final class MasterAdminProtos {
|
|||
|
||||
private int bitField0_;
|
||||
|
||||
// required bytes tableName = 1;
|
||||
// required bytes table_name = 1;
|
||||
private com.google.protobuf.ByteString tableName_ = com.google.protobuf.ByteString.EMPTY;
|
||||
public boolean hasTableName() {
|
||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||
|
@ -457,7 +457,7 @@ public final class MasterAdminProtos {
|
|||
return this;
|
||||
}
|
||||
|
||||
// required .ColumnFamilySchema columnFamilies = 2;
|
||||
// required .ColumnFamilySchema column_families = 2;
|
||||
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema columnFamilies_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema.getDefaultInstance();
|
||||
private com.google.protobuf.SingleFieldBuilder<
|
||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchemaOrBuilder> columnFamiliesBuilder_;
|
||||
|
@ -861,11 +861,11 @@ public final class MasterAdminProtos {
|
|||
public interface DeleteColumnRequestOrBuilder
|
||||
extends com.google.protobuf.MessageOrBuilder {
|
||||
|
||||
// required bytes tableName = 1;
|
||||
// required bytes table_name = 1;
|
||||
boolean hasTableName();
|
||||
com.google.protobuf.ByteString getTableName();
|
||||
|
||||
// required bytes columnName = 2;
|
||||
// required bytes column_name = 2;
|
||||
boolean hasColumnName();
|
||||
com.google.protobuf.ByteString getColumnName();
|
||||
}
|
||||
|
@ -898,8 +898,8 @@ public final class MasterAdminProtos {
|
|||
}
|
||||
|
||||
private int bitField0_;
|
||||
// required bytes tableName = 1;
|
||||
public static final int TABLENAME_FIELD_NUMBER = 1;
|
||||
// required bytes table_name = 1;
|
||||
public static final int TABLE_NAME_FIELD_NUMBER = 1;
|
||||
private com.google.protobuf.ByteString tableName_;
|
||||
public boolean hasTableName() {
|
||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||
|
@ -908,8 +908,8 @@ public final class MasterAdminProtos {
|
|||
return tableName_;
|
||||
}
|
||||
|
||||
// required bytes columnName = 2;
|
||||
public static final int COLUMNNAME_FIELD_NUMBER = 2;
|
||||
// required bytes column_name = 2;
|
||||
public static final int COLUMN_NAME_FIELD_NUMBER = 2;
|
||||
private com.google.protobuf.ByteString columnName_;
|
||||
public boolean hasColumnName() {
|
||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||
|
@ -1008,11 +1008,11 @@ public final class MasterAdminProtos {
|
|||
int hash = 41;
|
||||
hash = (19 * hash) + getDescriptorForType().hashCode();
|
||||
if (hasTableName()) {
|
||||
hash = (37 * hash) + TABLENAME_FIELD_NUMBER;
|
||||
hash = (37 * hash) + TABLE_NAME_FIELD_NUMBER;
|
||||
hash = (53 * hash) + getTableName().hashCode();
|
||||
}
|
||||
if (hasColumnName()) {
|
||||
hash = (37 * hash) + COLUMNNAME_FIELD_NUMBER;
|
||||
hash = (37 * hash) + COLUMN_NAME_FIELD_NUMBER;
|
||||
hash = (53 * hash) + getColumnName().hashCode();
|
||||
}
|
||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||
|
@ -1258,7 +1258,7 @@ public final class MasterAdminProtos {
|
|||
|
||||
private int bitField0_;
|
||||
|
||||
// required bytes tableName = 1;
|
||||
// required bytes table_name = 1;
|
||||
private com.google.protobuf.ByteString tableName_ = com.google.protobuf.ByteString.EMPTY;
|
||||
public boolean hasTableName() {
|
||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||
|
@ -1282,7 +1282,7 @@ public final class MasterAdminProtos {
|
|||
return this;
|
||||
}
|
||||
|
||||
// required bytes columnName = 2;
|
||||
// required bytes column_name = 2;
|
||||
private com.google.protobuf.ByteString columnName_ = com.google.protobuf.ByteString.EMPTY;
|
||||
public boolean hasColumnName() {
|
||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||
|
@ -1620,11 +1620,11 @@ public final class MasterAdminProtos {
|
|||
public interface ModifyColumnRequestOrBuilder
|
||||
extends com.google.protobuf.MessageOrBuilder {
|
||||
|
||||
// required bytes tableName = 1;
|
||||
// required bytes table_name = 1;
|
||||
boolean hasTableName();
|
||||
com.google.protobuf.ByteString getTableName();
|
||||
|
||||
// required .ColumnFamilySchema columnFamilies = 2;
|
||||
// required .ColumnFamilySchema column_families = 2;
|
||||
boolean hasColumnFamilies();
|
||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema getColumnFamilies();
|
||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchemaOrBuilder getColumnFamiliesOrBuilder();
|
||||
|
@ -1658,8 +1658,8 @@ public final class MasterAdminProtos {
|
|||
}
|
||||
|
||||
private int bitField0_;
|
||||
// required bytes tableName = 1;
|
||||
public static final int TABLENAME_FIELD_NUMBER = 1;
|
||||
// required bytes table_name = 1;
|
||||
public static final int TABLE_NAME_FIELD_NUMBER = 1;
|
||||
private com.google.protobuf.ByteString tableName_;
|
||||
public boolean hasTableName() {
|
||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||
|
@ -1668,8 +1668,8 @@ public final class MasterAdminProtos {
|
|||
return tableName_;
|
||||
}
|
||||
|
||||
// required .ColumnFamilySchema columnFamilies = 2;
|
||||
public static final int COLUMNFAMILIES_FIELD_NUMBER = 2;
|
||||
// required .ColumnFamilySchema column_families = 2;
|
||||
public static final int COLUMN_FAMILIES_FIELD_NUMBER = 2;
|
||||
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema columnFamilies_;
|
||||
public boolean hasColumnFamilies() {
|
||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||
|
@ -1775,11 +1775,11 @@ public final class MasterAdminProtos {
|
|||
int hash = 41;
|
||||
hash = (19 * hash) + getDescriptorForType().hashCode();
|
||||
if (hasTableName()) {
|
||||
hash = (37 * hash) + TABLENAME_FIELD_NUMBER;
|
||||
hash = (37 * hash) + TABLE_NAME_FIELD_NUMBER;
|
||||
hash = (53 * hash) + getTableName().hashCode();
|
||||
}
|
||||
if (hasColumnFamilies()) {
|
||||
hash = (37 * hash) + COLUMNFAMILIES_FIELD_NUMBER;
|
||||
hash = (37 * hash) + COLUMN_FAMILIES_FIELD_NUMBER;
|
||||
hash = (53 * hash) + getColumnFamilies().hashCode();
|
||||
}
|
||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||
|
@ -2042,7 +2042,7 @@ public final class MasterAdminProtos {
|
|||
|
||||
private int bitField0_;
|
||||
|
||||
// required bytes tableName = 1;
|
||||
// required bytes table_name = 1;
|
||||
private com.google.protobuf.ByteString tableName_ = com.google.protobuf.ByteString.EMPTY;
|
||||
public boolean hasTableName() {
|
||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||
|
@ -2066,7 +2066,7 @@ public final class MasterAdminProtos {
|
|||
return this;
|
||||
}
|
||||
|
||||
// required .ColumnFamilySchema columnFamilies = 2;
|
||||
// required .ColumnFamilySchema column_families = 2;
|
||||
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema columnFamilies_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema.getDefaultInstance();
|
||||
private com.google.protobuf.SingleFieldBuilder<
|
||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchemaOrBuilder> columnFamiliesBuilder_;
|
||||
|
@ -2475,7 +2475,7 @@ public final class MasterAdminProtos {
|
|||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier getRegion();
|
||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder getRegionOrBuilder();
|
||||
|
||||
// optional .ServerName destServerName = 2;
|
||||
// optional .ServerName dest_server_name = 2;
|
||||
boolean hasDestServerName();
|
||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName getDestServerName();
|
||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerNameOrBuilder getDestServerNameOrBuilder();
|
||||
|
@ -2522,8 +2522,8 @@ public final class MasterAdminProtos {
|
|||
return region_;
|
||||
}
|
||||
|
||||
// optional .ServerName destServerName = 2;
|
||||
public static final int DESTSERVERNAME_FIELD_NUMBER = 2;
|
||||
// optional .ServerName dest_server_name = 2;
|
||||
public static final int DEST_SERVER_NAME_FIELD_NUMBER = 2;
|
||||
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName destServerName_;
|
||||
public boolean hasDestServerName() {
|
||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||
|
@ -2635,7 +2635,7 @@ public final class MasterAdminProtos {
|
|||
hash = (53 * hash) + getRegion().hashCode();
|
||||
}
|
||||
if (hasDestServerName()) {
|
||||
hash = (37 * hash) + DESTSERVERNAME_FIELD_NUMBER;
|
||||
hash = (37 * hash) + DEST_SERVER_NAME_FIELD_NUMBER;
|
||||
hash = (53 * hash) + getDestServerName().hashCode();
|
||||
}
|
||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||
|
@ -3003,7 +3003,7 @@ public final class MasterAdminProtos {
|
|||
return regionBuilder_;
|
||||
}
|
||||
|
||||
// optional .ServerName destServerName = 2;
|
||||
// optional .ServerName dest_server_name = 2;
|
||||
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName destServerName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.getDefaultInstance();
|
||||
private com.google.protobuf.SingleFieldBuilder<
|
||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerNameOrBuilder> destServerNameBuilder_;
|
||||
|
@ -3407,12 +3407,12 @@ public final class MasterAdminProtos {
|
|||
public interface DispatchMergingRegionsRequestOrBuilder
|
||||
extends com.google.protobuf.MessageOrBuilder {
|
||||
|
||||
// required .RegionSpecifier regionA = 1;
|
||||
// required .RegionSpecifier region_a = 1;
|
||||
boolean hasRegionA();
|
||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier getRegionA();
|
||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder getRegionAOrBuilder();
|
||||
|
||||
// required .RegionSpecifier regionB = 2;
|
||||
// required .RegionSpecifier region_b = 2;
|
||||
boolean hasRegionB();
|
||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier getRegionB();
|
||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder getRegionBOrBuilder();
|
||||
|
@ -3450,8 +3450,8 @@ public final class MasterAdminProtos {
|
|||
}
|
||||
|
||||
private int bitField0_;
|
||||
// required .RegionSpecifier regionA = 1;
|
||||
public static final int REGIONA_FIELD_NUMBER = 1;
|
||||
// required .RegionSpecifier region_a = 1;
|
||||
public static final int REGION_A_FIELD_NUMBER = 1;
|
||||
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier regionA_;
|
||||
public boolean hasRegionA() {
|
||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||
|
@ -3463,8 +3463,8 @@ public final class MasterAdminProtos {
|
|||
return regionA_;
|
||||
}
|
||||
|
||||
// required .RegionSpecifier regionB = 2;
|
||||
public static final int REGIONB_FIELD_NUMBER = 2;
|
||||
// required .RegionSpecifier region_b = 2;
|
||||
public static final int REGION_B_FIELD_NUMBER = 2;
|
||||
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier regionB_;
|
||||
public boolean hasRegionB() {
|
||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||
|
@ -3597,11 +3597,11 @@ public final class MasterAdminProtos {
|
|||
int hash = 41;
|
||||
hash = (19 * hash) + getDescriptorForType().hashCode();
|
||||
if (hasRegionA()) {
|
||||
hash = (37 * hash) + REGIONA_FIELD_NUMBER;
|
||||
hash = (37 * hash) + REGION_A_FIELD_NUMBER;
|
||||
hash = (53 * hash) + getRegionA().hashCode();
|
||||
}
|
||||
if (hasRegionB()) {
|
||||
hash = (37 * hash) + REGIONB_FIELD_NUMBER;
|
||||
hash = (37 * hash) + REGION_B_FIELD_NUMBER;
|
||||
hash = (53 * hash) + getRegionB().hashCode();
|
||||
}
|
||||
if (hasForcible()) {
|
||||
|
@ -3899,7 +3899,7 @@ public final class MasterAdminProtos {
|
|||
|
||||
private int bitField0_;
|
||||
|
||||
// required .RegionSpecifier regionA = 1;
|
||||
// required .RegionSpecifier region_a = 1;
|
||||
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier regionA_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance();
|
||||
private com.google.protobuf.SingleFieldBuilder<
|
||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder> regionABuilder_;
|
||||
|
@ -3989,7 +3989,7 @@ public final class MasterAdminProtos {
|
|||
return regionABuilder_;
|
||||
}
|
||||
|
||||
// required .RegionSpecifier regionB = 2;
|
||||
// required .RegionSpecifier region_b = 2;
|
||||
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier regionB_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance();
|
||||
private com.google.protobuf.SingleFieldBuilder<
|
||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder> regionBBuilder_;
|
||||
|
@ -6799,12 +6799,12 @@ public final class MasterAdminProtos {
|
|||
public interface CreateTableRequestOrBuilder
|
||||
extends com.google.protobuf.MessageOrBuilder {
|
||||
|
||||
// required .TableSchema tableSchema = 1;
|
||||
// required .TableSchema table_schema = 1;
|
||||
boolean hasTableSchema();
|
||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema getTableSchema();
|
||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchemaOrBuilder getTableSchemaOrBuilder();
|
||||
|
||||
// repeated bytes splitKeys = 2;
|
||||
// repeated bytes split_keys = 2;
|
||||
java.util.List<com.google.protobuf.ByteString> getSplitKeysList();
|
||||
int getSplitKeysCount();
|
||||
com.google.protobuf.ByteString getSplitKeys(int index);
|
||||
|
@ -6838,8 +6838,8 @@ public final class MasterAdminProtos {
|
|||
}
|
||||
|
||||
private int bitField0_;
|
||||
// required .TableSchema tableSchema = 1;
|
||||
public static final int TABLESCHEMA_FIELD_NUMBER = 1;
|
||||
// required .TableSchema table_schema = 1;
|
||||
public static final int TABLE_SCHEMA_FIELD_NUMBER = 1;
|
||||
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema tableSchema_;
|
||||
public boolean hasTableSchema() {
|
||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||
|
@ -6851,8 +6851,8 @@ public final class MasterAdminProtos {
|
|||
return tableSchema_;
|
||||
}
|
||||
|
||||
// repeated bytes splitKeys = 2;
|
||||
public static final int SPLITKEYS_FIELD_NUMBER = 2;
|
||||
// repeated bytes split_keys = 2;
|
||||
public static final int SPLIT_KEYS_FIELD_NUMBER = 2;
|
||||
private java.util.List<com.google.protobuf.ByteString> splitKeys_;
|
||||
public java.util.List<com.google.protobuf.ByteString>
|
||||
getSplitKeysList() {
|
||||
|
@ -6957,11 +6957,11 @@ public final class MasterAdminProtos {
|
|||
int hash = 41;
|
||||
hash = (19 * hash) + getDescriptorForType().hashCode();
|
||||
if (hasTableSchema()) {
|
||||
hash = (37 * hash) + TABLESCHEMA_FIELD_NUMBER;
|
||||
hash = (37 * hash) + TABLE_SCHEMA_FIELD_NUMBER;
|
||||
hash = (53 * hash) + getTableSchema().hashCode();
|
||||
}
|
||||
if (getSplitKeysCount() > 0) {
|
||||
hash = (37 * hash) + SPLITKEYS_FIELD_NUMBER;
|
||||
hash = (37 * hash) + SPLIT_KEYS_FIELD_NUMBER;
|
||||
hash = (53 * hash) + getSplitKeysList().hashCode();
|
||||
}
|
||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||
|
@ -7228,7 +7228,7 @@ public final class MasterAdminProtos {
|
|||
|
||||
private int bitField0_;
|
||||
|
||||
// required .TableSchema tableSchema = 1;
|
||||
// required .TableSchema table_schema = 1;
|
||||
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema tableSchema_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema.getDefaultInstance();
|
||||
private com.google.protobuf.SingleFieldBuilder<
|
||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchemaOrBuilder> tableSchemaBuilder_;
|
||||
|
@ -7318,7 +7318,7 @@ public final class MasterAdminProtos {
|
|||
return tableSchemaBuilder_;
|
||||
}
|
||||
|
||||
// repeated bytes splitKeys = 2;
|
||||
// repeated bytes split_keys = 2;
|
||||
private java.util.List<com.google.protobuf.ByteString> splitKeys_ = java.util.Collections.emptyList();;
|
||||
private void ensureSplitKeysIsMutable() {
|
||||
if (!((bitField0_ & 0x00000002) == 0x00000002)) {
|
||||
|
@ -7683,7 +7683,7 @@ public final class MasterAdminProtos {
|
|||
public interface DeleteTableRequestOrBuilder
|
||||
extends com.google.protobuf.MessageOrBuilder {
|
||||
|
||||
// required bytes tableName = 1;
|
||||
// required bytes table_name = 1;
|
||||
boolean hasTableName();
|
||||
com.google.protobuf.ByteString getTableName();
|
||||
}
|
||||
|
@ -7716,8 +7716,8 @@ public final class MasterAdminProtos {
|
|||
}
|
||||
|
||||
private int bitField0_;
|
||||
// required bytes tableName = 1;
|
||||
public static final int TABLENAME_FIELD_NUMBER = 1;
|
||||
// required bytes table_name = 1;
|
||||
public static final int TABLE_NAME_FIELD_NUMBER = 1;
|
||||
private com.google.protobuf.ByteString tableName_;
|
||||
public boolean hasTableName() {
|
||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||
|
@ -7799,7 +7799,7 @@ public final class MasterAdminProtos {
|
|||
int hash = 41;
|
||||
hash = (19 * hash) + getDescriptorForType().hashCode();
|
||||
if (hasTableName()) {
|
||||
hash = (37 * hash) + TABLENAME_FIELD_NUMBER;
|
||||
hash = (37 * hash) + TABLE_NAME_FIELD_NUMBER;
|
||||
hash = (53 * hash) + getTableName().hashCode();
|
||||
}
|
||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||
|
@ -8027,7 +8027,7 @@ public final class MasterAdminProtos {
|
|||
|
||||
private int bitField0_;
|
||||
|
||||
// required bytes tableName = 1;
|
||||
// required bytes table_name = 1;
|
||||
private com.google.protobuf.ByteString tableName_ = com.google.protobuf.ByteString.EMPTY;
|
||||
public boolean hasTableName() {
|
||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||
|
@ -8365,7 +8365,7 @@ public final class MasterAdminProtos {
|
|||
public interface EnableTableRequestOrBuilder
|
||||
extends com.google.protobuf.MessageOrBuilder {
|
||||
|
||||
// required bytes tableName = 1;
|
||||
// required bytes table_name = 1;
|
||||
boolean hasTableName();
|
||||
com.google.protobuf.ByteString getTableName();
|
||||
}
|
||||
|
@ -8398,8 +8398,8 @@ public final class MasterAdminProtos {
|
|||
}
|
||||
|
||||
private int bitField0_;
|
||||
// required bytes tableName = 1;
|
||||
public static final int TABLENAME_FIELD_NUMBER = 1;
|
||||
// required bytes table_name = 1;
|
||||
public static final int TABLE_NAME_FIELD_NUMBER = 1;
|
||||
private com.google.protobuf.ByteString tableName_;
|
||||
public boolean hasTableName() {
|
||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||
|
@ -8481,7 +8481,7 @@ public final class MasterAdminProtos {
|
|||
int hash = 41;
|
||||
hash = (19 * hash) + getDescriptorForType().hashCode();
|
||||
if (hasTableName()) {
|
||||
hash = (37 * hash) + TABLENAME_FIELD_NUMBER;
|
||||
hash = (37 * hash) + TABLE_NAME_FIELD_NUMBER;
|
||||
hash = (53 * hash) + getTableName().hashCode();
|
||||
}
|
||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||
|
@ -8709,7 +8709,7 @@ public final class MasterAdminProtos {
|
|||
|
||||
private int bitField0_;
|
||||
|
||||
// required bytes tableName = 1;
|
||||
// required bytes table_name = 1;
|
||||
private com.google.protobuf.ByteString tableName_ = com.google.protobuf.ByteString.EMPTY;
|
||||
public boolean hasTableName() {
|
||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||
|
@ -9047,7 +9047,7 @@ public final class MasterAdminProtos {
|
|||
public interface DisableTableRequestOrBuilder
|
||||
extends com.google.protobuf.MessageOrBuilder {
|
||||
|
||||
// required bytes tableName = 1;
|
||||
// required bytes table_name = 1;
|
||||
boolean hasTableName();
|
||||
com.google.protobuf.ByteString getTableName();
|
||||
}
|
||||
|
@ -9080,8 +9080,8 @@ public final class MasterAdminProtos {
|
|||
}
|
||||
|
||||
private int bitField0_;
|
||||
// required bytes tableName = 1;
|
||||
public static final int TABLENAME_FIELD_NUMBER = 1;
|
||||
// required bytes table_name = 1;
|
||||
public static final int TABLE_NAME_FIELD_NUMBER = 1;
|
||||
private com.google.protobuf.ByteString tableName_;
|
||||
public boolean hasTableName() {
|
||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||
|
@ -9163,7 +9163,7 @@ public final class MasterAdminProtos {
|
|||
int hash = 41;
|
||||
hash = (19 * hash) + getDescriptorForType().hashCode();
|
||||
if (hasTableName()) {
|
||||
hash = (37 * hash) + TABLENAME_FIELD_NUMBER;
|
||||
hash = (37 * hash) + TABLE_NAME_FIELD_NUMBER;
|
||||
hash = (53 * hash) + getTableName().hashCode();
|
||||
}
|
||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||
|
@ -9391,7 +9391,7 @@ public final class MasterAdminProtos {
|
|||
|
||||
private int bitField0_;
|
||||
|
||||
// required bytes tableName = 1;
|
||||
// required bytes table_name = 1;
|
||||
private com.google.protobuf.ByteString tableName_ = com.google.protobuf.ByteString.EMPTY;
|
||||
public boolean hasTableName() {
|
||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||
|
@ -9729,11 +9729,11 @@ public final class MasterAdminProtos {
|
|||
public interface ModifyTableRequestOrBuilder
|
||||
extends com.google.protobuf.MessageOrBuilder {
|
||||
|
||||
// required bytes tableName = 1;
|
||||
// required bytes table_name = 1;
|
||||
boolean hasTableName();
|
||||
com.google.protobuf.ByteString getTableName();
|
||||
|
||||
// required .TableSchema tableSchema = 2;
|
||||
// required .TableSchema table_schema = 2;
|
||||
boolean hasTableSchema();
|
||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema getTableSchema();
|
||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchemaOrBuilder getTableSchemaOrBuilder();
|
||||
|
@ -9767,8 +9767,8 @@ public final class MasterAdminProtos {
|
|||
}
|
||||
|
||||
private int bitField0_;
|
||||
// required bytes tableName = 1;
|
||||
public static final int TABLENAME_FIELD_NUMBER = 1;
|
||||
// required bytes table_name = 1;
|
||||
public static final int TABLE_NAME_FIELD_NUMBER = 1;
|
||||
private com.google.protobuf.ByteString tableName_;
|
||||
public boolean hasTableName() {
|
||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||
|
@ -9777,8 +9777,8 @@ public final class MasterAdminProtos {
|
|||
return tableName_;
|
||||
}
|
||||
|
||||
// required .TableSchema tableSchema = 2;
|
||||
public static final int TABLESCHEMA_FIELD_NUMBER = 2;
|
||||
// required .TableSchema table_schema = 2;
|
||||
public static final int TABLE_SCHEMA_FIELD_NUMBER = 2;
|
||||
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema tableSchema_;
|
||||
public boolean hasTableSchema() {
|
||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||
|
@ -9884,11 +9884,11 @@ public final class MasterAdminProtos {
|
|||
int hash = 41;
|
||||
hash = (19 * hash) + getDescriptorForType().hashCode();
|
||||
if (hasTableName()) {
|
||||
hash = (37 * hash) + TABLENAME_FIELD_NUMBER;
|
||||
hash = (37 * hash) + TABLE_NAME_FIELD_NUMBER;
|
||||
hash = (53 * hash) + getTableName().hashCode();
|
||||
}
|
||||
if (hasTableSchema()) {
|
||||
hash = (37 * hash) + TABLESCHEMA_FIELD_NUMBER;
|
||||
hash = (37 * hash) + TABLE_SCHEMA_FIELD_NUMBER;
|
||||
hash = (53 * hash) + getTableSchema().hashCode();
|
||||
}
|
||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||
|
@ -10151,7 +10151,7 @@ public final class MasterAdminProtos {
|
|||
|
||||
private int bitField0_;
|
||||
|
||||
// required bytes tableName = 1;
|
||||
// required bytes table_name = 1;
|
||||
private com.google.protobuf.ByteString tableName_ = com.google.protobuf.ByteString.EMPTY;
|
||||
public boolean hasTableName() {
|
||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||
|
@ -10175,7 +10175,7 @@ public final class MasterAdminProtos {
|
|||
return this;
|
||||
}
|
||||
|
||||
// required .TableSchema tableSchema = 2;
|
||||
// required .TableSchema table_schema = 2;
|
||||
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema tableSchema_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema.getDefaultInstance();
|
||||
private com.google.protobuf.SingleFieldBuilder<
|
||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchemaOrBuilder> tableSchemaBuilder_;
|
||||
|
@ -12079,7 +12079,7 @@ public final class MasterAdminProtos {
|
|||
public interface BalanceResponseOrBuilder
|
||||
extends com.google.protobuf.MessageOrBuilder {
|
||||
|
||||
// required bool balancerRan = 1;
|
||||
// required bool balancer_ran = 1;
|
||||
boolean hasBalancerRan();
|
||||
boolean getBalancerRan();
|
||||
}
|
||||
|
@ -12112,8 +12112,8 @@ public final class MasterAdminProtos {
|
|||
}
|
||||
|
||||
private int bitField0_;
|
||||
// required bool balancerRan = 1;
|
||||
public static final int BALANCERRAN_FIELD_NUMBER = 1;
|
||||
// required bool balancer_ran = 1;
|
||||
public static final int BALANCER_RAN_FIELD_NUMBER = 1;
|
||||
private boolean balancerRan_;
|
||||
public boolean hasBalancerRan() {
|
||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||
|
@ -12195,7 +12195,7 @@ public final class MasterAdminProtos {
|
|||
int hash = 41;
|
||||
hash = (19 * hash) + getDescriptorForType().hashCode();
|
||||
if (hasBalancerRan()) {
|
||||
hash = (37 * hash) + BALANCERRAN_FIELD_NUMBER;
|
||||
hash = (37 * hash) + BALANCER_RAN_FIELD_NUMBER;
|
||||
hash = (53 * hash) + hashBoolean(getBalancerRan());
|
||||
}
|
||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||
|
@ -12423,7 +12423,7 @@ public final class MasterAdminProtos {
|
|||
|
||||
private int bitField0_;
|
||||
|
||||
// required bool balancerRan = 1;
|
||||
// required bool balancer_ran = 1;
|
||||
private boolean balancerRan_ ;
|
||||
public boolean hasBalancerRan() {
|
||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||
|
@ -12903,7 +12903,7 @@ public final class MasterAdminProtos {
|
|||
public interface SetBalancerRunningResponseOrBuilder
|
||||
extends com.google.protobuf.MessageOrBuilder {
|
||||
|
||||
// optional bool prevBalanceValue = 1;
|
||||
// optional bool prev_balance_value = 1;
|
||||
boolean hasPrevBalanceValue();
|
||||
boolean getPrevBalanceValue();
|
||||
}
|
||||
|
@ -12936,8 +12936,8 @@ public final class MasterAdminProtos {
|
|||
}
|
||||
|
||||
private int bitField0_;
|
||||
// optional bool prevBalanceValue = 1;
|
||||
public static final int PREVBALANCEVALUE_FIELD_NUMBER = 1;
|
||||
// optional bool prev_balance_value = 1;
|
||||
public static final int PREV_BALANCE_VALUE_FIELD_NUMBER = 1;
|
||||
private boolean prevBalanceValue_;
|
||||
public boolean hasPrevBalanceValue() {
|
||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||
|
@ -13015,7 +13015,7 @@ public final class MasterAdminProtos {
|
|||
int hash = 41;
|
||||
hash = (19 * hash) + getDescriptorForType().hashCode();
|
||||
if (hasPrevBalanceValue()) {
|
||||
hash = (37 * hash) + PREVBALANCEVALUE_FIELD_NUMBER;
|
||||
hash = (37 * hash) + PREV_BALANCE_VALUE_FIELD_NUMBER;
|
||||
hash = (53 * hash) + hashBoolean(getPrevBalanceValue());
|
||||
}
|
||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||
|
@ -13239,7 +13239,7 @@ public final class MasterAdminProtos {
|
|||
|
||||
private int bitField0_;
|
||||
|
||||
// optional bool prevBalanceValue = 1;
|
||||
// optional bool prev_balance_value = 1;
|
||||
private boolean prevBalanceValue_ ;
|
||||
public boolean hasPrevBalanceValue() {
|
||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||
|
@ -13574,7 +13574,7 @@ public final class MasterAdminProtos {
|
|||
public interface CatalogScanResponseOrBuilder
|
||||
extends com.google.protobuf.MessageOrBuilder {
|
||||
|
||||
// optional int32 scanResult = 1;
|
||||
// optional int32 scan_result = 1;
|
||||
boolean hasScanResult();
|
||||
int getScanResult();
|
||||
}
|
||||
|
@ -13607,8 +13607,8 @@ public final class MasterAdminProtos {
|
|||
}
|
||||
|
||||
private int bitField0_;
|
||||
// optional int32 scanResult = 1;
|
||||
public static final int SCANRESULT_FIELD_NUMBER = 1;
|
||||
// optional int32 scan_result = 1;
|
||||
public static final int SCAN_RESULT_FIELD_NUMBER = 1;
|
||||
private int scanResult_;
|
||||
public boolean hasScanResult() {
|
||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||
|
@ -13686,7 +13686,7 @@ public final class MasterAdminProtos {
|
|||
int hash = 41;
|
||||
hash = (19 * hash) + getDescriptorForType().hashCode();
|
||||
if (hasScanResult()) {
|
||||
hash = (37 * hash) + SCANRESULT_FIELD_NUMBER;
|
||||
hash = (37 * hash) + SCAN_RESULT_FIELD_NUMBER;
|
||||
hash = (53 * hash) + getScanResult();
|
||||
}
|
||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||
|
@ -13910,7 +13910,7 @@ public final class MasterAdminProtos {
|
|||
|
||||
private int bitField0_;
|
||||
|
||||
// optional int32 scanResult = 1;
|
||||
// optional int32 scan_result = 1;
|
||||
private int scanResult_ ;
|
||||
public boolean hasScanResult() {
|
||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||
|
@ -14324,7 +14324,7 @@ public final class MasterAdminProtos {
|
|||
public interface EnableCatalogJanitorResponseOrBuilder
|
||||
extends com.google.protobuf.MessageOrBuilder {
|
||||
|
||||
// optional bool prevValue = 1;
|
||||
// optional bool prev_value = 1;
|
||||
boolean hasPrevValue();
|
||||
boolean getPrevValue();
|
||||
}
|
||||
|
@ -14357,8 +14357,8 @@ public final class MasterAdminProtos {
|
|||
}
|
||||
|
||||
private int bitField0_;
|
||||
// optional bool prevValue = 1;
|
||||
public static final int PREVVALUE_FIELD_NUMBER = 1;
|
||||
// optional bool prev_value = 1;
|
||||
public static final int PREV_VALUE_FIELD_NUMBER = 1;
|
||||
private boolean prevValue_;
|
||||
public boolean hasPrevValue() {
|
||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||
|
@ -14436,7 +14436,7 @@ public final class MasterAdminProtos {
|
|||
int hash = 41;
|
||||
hash = (19 * hash) + getDescriptorForType().hashCode();
|
||||
if (hasPrevValue()) {
|
||||
hash = (37 * hash) + PREVVALUE_FIELD_NUMBER;
|
||||
hash = (37 * hash) + PREV_VALUE_FIELD_NUMBER;
|
||||
hash = (53 * hash) + hashBoolean(getPrevValue());
|
||||
}
|
||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||
|
@ -14660,7 +14660,7 @@ public final class MasterAdminProtos {
|
|||
|
||||
private int bitField0_;
|
||||
|
||||
// optional bool prevValue = 1;
|
||||
// optional bool prev_value = 1;
|
||||
private boolean prevValue_ ;
|
||||
public boolean hasPrevValue() {
|
||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||
|
@ -15847,7 +15847,7 @@ public final class MasterAdminProtos {
|
|||
public interface TakeSnapshotResponseOrBuilder
|
||||
extends com.google.protobuf.MessageOrBuilder {
|
||||
|
||||
// required int64 expectedTimeout = 1;
|
||||
// required int64 expected_timeout = 1;
|
||||
boolean hasExpectedTimeout();
|
||||
long getExpectedTimeout();
|
||||
}
|
||||
|
@ -15880,8 +15880,8 @@ public final class MasterAdminProtos {
|
|||
}
|
||||
|
||||
private int bitField0_;
|
||||
// required int64 expectedTimeout = 1;
|
||||
public static final int EXPECTEDTIMEOUT_FIELD_NUMBER = 1;
|
||||
// required int64 expected_timeout = 1;
|
||||
public static final int EXPECTED_TIMEOUT_FIELD_NUMBER = 1;
|
||||
private long expectedTimeout_;
|
||||
public boolean hasExpectedTimeout() {
|
||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||
|
@ -15963,7 +15963,7 @@ public final class MasterAdminProtos {
|
|||
int hash = 41;
|
||||
hash = (19 * hash) + getDescriptorForType().hashCode();
|
||||
if (hasExpectedTimeout()) {
|
||||
hash = (37 * hash) + EXPECTEDTIMEOUT_FIELD_NUMBER;
|
||||
hash = (37 * hash) + EXPECTED_TIMEOUT_FIELD_NUMBER;
|
||||
hash = (53 * hash) + hashLong(getExpectedTimeout());
|
||||
}
|
||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||
|
@ -16191,7 +16191,7 @@ public final class MasterAdminProtos {
|
|||
|
||||
private int bitField0_;
|
||||
|
||||
// required int64 expectedTimeout = 1;
|
||||
// required int64 expected_timeout = 1;
|
||||
private long expectedTimeout_ ;
|
||||
public boolean hasExpectedTimeout() {
|
||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||
|
@ -22752,116 +22752,116 @@ public final class MasterAdminProtos {
|
|||
static {
|
||||
java.lang.String[] descriptorData = {
|
||||
"\n\021MasterAdmin.proto\032\014Master.proto\032\013hbase" +
|
||||
".proto\032\014Client.proto\"R\n\020AddColumnRequest" +
|
||||
"\022\021\n\ttableName\030\001 \002(\014\022+\n\016columnFamilies\030\002 " +
|
||||
"\002(\0132\023.ColumnFamilySchema\"\023\n\021AddColumnRes" +
|
||||
"ponse\"<\n\023DeleteColumnRequest\022\021\n\ttableNam" +
|
||||
"e\030\001 \002(\014\022\022\n\ncolumnName\030\002 \002(\014\"\026\n\024DeleteCol" +
|
||||
"umnResponse\"U\n\023ModifyColumnRequest\022\021\n\tta" +
|
||||
"bleName\030\001 \002(\014\022+\n\016columnFamilies\030\002 \002(\0132\023." +
|
||||
"ColumnFamilySchema\"\026\n\024ModifyColumnRespon" +
|
||||
"se\"Z\n\021MoveRegionRequest\022 \n\006region\030\001 \002(\0132",
|
||||
"\020.RegionSpecifier\022#\n\016destServerName\030\002 \001(" +
|
||||
"\0132\013.ServerName\"\024\n\022MoveRegionResponse\"~\n\035" +
|
||||
"DispatchMergingRegionsRequest\022!\n\007regionA" +
|
||||
"\030\001 \002(\0132\020.RegionSpecifier\022!\n\007regionB\030\002 \002(" +
|
||||
"\0132\020.RegionSpecifier\022\027\n\010forcible\030\003 \001(\010:\005f" +
|
||||
"alse\" \n\036DispatchMergingRegionsResponse\"7" +
|
||||
"\n\023AssignRegionRequest\022 \n\006region\030\001 \002(\0132\020." +
|
||||
"RegionSpecifier\"\026\n\024AssignRegionResponse\"" +
|
||||
"O\n\025UnassignRegionRequest\022 \n\006region\030\001 \002(\013" +
|
||||
"2\020.RegionSpecifier\022\024\n\005force\030\002 \001(\010:\005false",
|
||||
"\"\030\n\026UnassignRegionResponse\"8\n\024OfflineReg" +
|
||||
"ionRequest\022 \n\006region\030\001 \002(\0132\020.RegionSpeci" +
|
||||
"fier\"\027\n\025OfflineRegionResponse\"J\n\022CreateT" +
|
||||
"ableRequest\022!\n\013tableSchema\030\001 \002(\0132\014.Table" +
|
||||
"Schema\022\021\n\tsplitKeys\030\002 \003(\014\"\025\n\023CreateTable" +
|
||||
"Response\"\'\n\022DeleteTableRequest\022\021\n\ttableN" +
|
||||
"ame\030\001 \002(\014\"\025\n\023DeleteTableResponse\"\'\n\022Enab" +
|
||||
"leTableRequest\022\021\n\ttableName\030\001 \002(\014\"\025\n\023Ena" +
|
||||
"bleTableResponse\"(\n\023DisableTableRequest\022" +
|
||||
"\021\n\ttableName\030\001 \002(\014\"\026\n\024DisableTableRespon",
|
||||
"se\"J\n\022ModifyTableRequest\022\021\n\ttableName\030\001 " +
|
||||
"\002(\014\022!\n\013tableSchema\030\002 \002(\0132\014.TableSchema\"\025" +
|
||||
"\n\023ModifyTableResponse\"\021\n\017ShutdownRequest" +
|
||||
"\"\022\n\020ShutdownResponse\"\023\n\021StopMasterReques" +
|
||||
"t\"\024\n\022StopMasterResponse\"\020\n\016BalanceReques" +
|
||||
"t\"&\n\017BalanceResponse\022\023\n\013balancerRan\030\001 \002(" +
|
||||
"\010\"<\n\031SetBalancerRunningRequest\022\n\n\002on\030\001 \002" +
|
||||
"(\010\022\023\n\013synchronous\030\002 \001(\010\"6\n\032SetBalancerRu" +
|
||||
"nningResponse\022\030\n\020prevBalanceValue\030\001 \001(\010\"" +
|
||||
"\024\n\022CatalogScanRequest\")\n\023CatalogScanResp",
|
||||
"onse\022\022\n\nscanResult\030\001 \001(\005\"-\n\033EnableCatalo" +
|
||||
"gJanitorRequest\022\016\n\006enable\030\001 \002(\010\"1\n\034Enabl" +
|
||||
"eCatalogJanitorResponse\022\021\n\tprevValue\030\001 \001" +
|
||||
"(\010\" \n\036IsCatalogJanitorEnabledRequest\"0\n\037" +
|
||||
"IsCatalogJanitorEnabledResponse\022\r\n\005value" +
|
||||
"\030\001 \002(\010\"=\n\023TakeSnapshotRequest\022&\n\010snapsho" +
|
||||
"t\030\001 \002(\0132\024.SnapshotDescription\"/\n\024TakeSna" +
|
||||
"pshotResponse\022\027\n\017expectedTimeout\030\001 \002(\003\"\025" +
|
||||
"\n\023ListSnapshotRequest\"?\n\024ListSnapshotRes" +
|
||||
"ponse\022\'\n\tsnapshots\030\001 \003(\0132\024.SnapshotDescr",
|
||||
"iption\"?\n\025DeleteSnapshotRequest\022&\n\010snaps" +
|
||||
"hot\030\001 \002(\0132\024.SnapshotDescription\"\030\n\026Delet" +
|
||||
"eSnapshotResponse\"@\n\026RestoreSnapshotRequ" +
|
||||
"est\022&\n\010snapshot\030\001 \002(\0132\024.SnapshotDescript" +
|
||||
"ion\"\031\n\027RestoreSnapshotResponse\"?\n\025IsSnap" +
|
||||
"shotDoneRequest\022&\n\010snapshot\030\001 \001(\0132\024.Snap" +
|
||||
"shotDescription\"U\n\026IsSnapshotDoneRespons" +
|
||||
"e\022\023\n\004done\030\001 \001(\010:\005false\022&\n\010snapshot\030\002 \001(\013" +
|
||||
"2\024.SnapshotDescription\"F\n\034IsRestoreSnaps" +
|
||||
"hotDoneRequest\022&\n\010snapshot\030\001 \001(\0132\024.Snaps",
|
||||
"hotDescription\"3\n\035IsRestoreSnapshotDoneR" +
|
||||
"esponse\022\022\n\004done\030\001 \001(\010:\004true2\305\016\n\022MasterAd" +
|
||||
"minService\0222\n\taddColumn\022\021.AddColumnReque" +
|
||||
"st\032\022.AddColumnResponse\022;\n\014deleteColumn\022\024" +
|
||||
".DeleteColumnRequest\032\025.DeleteColumnRespo" +
|
||||
"nse\022;\n\014modifyColumn\022\024.ModifyColumnReques" +
|
||||
"t\032\025.ModifyColumnResponse\0225\n\nmoveRegion\022\022" +
|
||||
".MoveRegionRequest\032\023.MoveRegionResponse\022" +
|
||||
"Y\n\026dispatchMergingRegions\022\036.DispatchMerg" +
|
||||
"ingRegionsRequest\032\037.DispatchMergingRegio",
|
||||
"nsResponse\022;\n\014assignRegion\022\024.AssignRegio" +
|
||||
"nRequest\032\025.AssignRegionResponse\022A\n\016unass" +
|
||||
"ignRegion\022\026.UnassignRegionRequest\032\027.Unas" +
|
||||
"signRegionResponse\022>\n\rofflineRegion\022\025.Of" +
|
||||
"flineRegionRequest\032\026.OfflineRegionRespon" +
|
||||
"se\0228\n\013deleteTable\022\023.DeleteTableRequest\032\024" +
|
||||
".DeleteTableResponse\0228\n\013enableTable\022\023.En" +
|
||||
"ableTableRequest\032\024.EnableTableResponse\022;" +
|
||||
"\n\014disableTable\022\024.DisableTableRequest\032\025.D" +
|
||||
"isableTableResponse\0228\n\013modifyTable\022\023.Mod",
|
||||
"ifyTableRequest\032\024.ModifyTableResponse\0228\n" +
|
||||
"\013createTable\022\023.CreateTableRequest\032\024.Crea" +
|
||||
"teTableResponse\022/\n\010shutdown\022\020.ShutdownRe" +
|
||||
"quest\032\021.ShutdownResponse\0225\n\nstopMaster\022\022" +
|
||||
".StopMasterRequest\032\023.StopMasterResponse\022" +
|
||||
",\n\007balance\022\017.BalanceRequest\032\020.BalanceRes" +
|
||||
"ponse\022M\n\022setBalancerRunning\022\032.SetBalance" +
|
||||
"rRunningRequest\032\033.SetBalancerRunningResp" +
|
||||
"onse\022;\n\016runCatalogScan\022\023.CatalogScanRequ" +
|
||||
"est\032\024.CatalogScanResponse\022S\n\024enableCatal",
|
||||
"ogJanitor\022\034.EnableCatalogJanitorRequest\032" +
|
||||
"\035.EnableCatalogJanitorResponse\022\\\n\027isCata" +
|
||||
"logJanitorEnabled\022\037.IsCatalogJanitorEnab" +
|
||||
"ledRequest\032 .IsCatalogJanitorEnabledResp" +
|
||||
"onse\022L\n\021execMasterService\022\032.CoprocessorS" +
|
||||
"erviceRequest\032\033.CoprocessorServiceRespon" +
|
||||
"se\0227\n\010snapshot\022\024.TakeSnapshotRequest\032\025.T" +
|
||||
"akeSnapshotResponse\022D\n\025getCompletedSnaps" +
|
||||
"hots\022\024.ListSnapshotRequest\032\025.ListSnapsho" +
|
||||
"tResponse\022A\n\016deleteSnapshot\022\026.DeleteSnap",
|
||||
"shotRequest\032\027.DeleteSnapshotResponse\022A\n\016" +
|
||||
"isSnapshotDone\022\026.IsSnapshotDoneRequest\032\027" +
|
||||
".IsSnapshotDoneResponse\022D\n\017restoreSnapsh" +
|
||||
"ot\022\027.RestoreSnapshotRequest\032\030.RestoreSna" +
|
||||
"pshotResponse\022V\n\025isRestoreSnapshotDone\022\035" +
|
||||
".IsRestoreSnapshotDoneRequest\032\036.IsRestor" +
|
||||
"eSnapshotDoneResponse\022D\n\017isMasterRunning" +
|
||||
"\022\027.IsMasterRunningRequest\032\030.IsMasterRunn" +
|
||||
"ingResponseBG\n*org.apache.hadoop.hbase.p" +
|
||||
"rotobuf.generatedB\021MasterAdminProtosH\001\210\001",
|
||||
"\001\240\001\001"
|
||||
".proto\032\014Client.proto\"T\n\020AddColumnRequest" +
|
||||
"\022\022\n\ntable_name\030\001 \002(\014\022,\n\017column_families\030" +
|
||||
"\002 \002(\0132\023.ColumnFamilySchema\"\023\n\021AddColumnR" +
|
||||
"esponse\">\n\023DeleteColumnRequest\022\022\n\ntable_" +
|
||||
"name\030\001 \002(\014\022\023\n\013column_name\030\002 \002(\014\"\026\n\024Delet" +
|
||||
"eColumnResponse\"W\n\023ModifyColumnRequest\022\022" +
|
||||
"\n\ntable_name\030\001 \002(\014\022,\n\017column_families\030\002 " +
|
||||
"\002(\0132\023.ColumnFamilySchema\"\026\n\024ModifyColumn" +
|
||||
"Response\"\\\n\021MoveRegionRequest\022 \n\006region\030",
|
||||
"\001 \002(\0132\020.RegionSpecifier\022%\n\020dest_server_n" +
|
||||
"ame\030\002 \001(\0132\013.ServerName\"\024\n\022MoveRegionResp" +
|
||||
"onse\"\200\001\n\035DispatchMergingRegionsRequest\022\"" +
|
||||
"\n\010region_a\030\001 \002(\0132\020.RegionSpecifier\022\"\n\010re" +
|
||||
"gion_b\030\002 \002(\0132\020.RegionSpecifier\022\027\n\010forcib" +
|
||||
"le\030\003 \001(\010:\005false\" \n\036DispatchMergingRegion" +
|
||||
"sResponse\"7\n\023AssignRegionRequest\022 \n\006regi" +
|
||||
"on\030\001 \002(\0132\020.RegionSpecifier\"\026\n\024AssignRegi" +
|
||||
"onResponse\"O\n\025UnassignRegionRequest\022 \n\006r" +
|
||||
"egion\030\001 \002(\0132\020.RegionSpecifier\022\024\n\005force\030\002",
|
||||
" \001(\010:\005false\"\030\n\026UnassignRegionResponse\"8\n" +
|
||||
"\024OfflineRegionRequest\022 \n\006region\030\001 \002(\0132\020." +
|
||||
"RegionSpecifier\"\027\n\025OfflineRegionResponse" +
|
||||
"\"L\n\022CreateTableRequest\022\"\n\014table_schema\030\001" +
|
||||
" \002(\0132\014.TableSchema\022\022\n\nsplit_keys\030\002 \003(\014\"\025" +
|
||||
"\n\023CreateTableResponse\"(\n\022DeleteTableRequ" +
|
||||
"est\022\022\n\ntable_name\030\001 \002(\014\"\025\n\023DeleteTableRe" +
|
||||
"sponse\"(\n\022EnableTableRequest\022\022\n\ntable_na" +
|
||||
"me\030\001 \002(\014\"\025\n\023EnableTableResponse\")\n\023Disab" +
|
||||
"leTableRequest\022\022\n\ntable_name\030\001 \002(\014\"\026\n\024Di",
|
||||
"sableTableResponse\"L\n\022ModifyTableRequest" +
|
||||
"\022\022\n\ntable_name\030\001 \002(\014\022\"\n\014table_schema\030\002 \002" +
|
||||
"(\0132\014.TableSchema\"\025\n\023ModifyTableResponse\"" +
|
||||
"\021\n\017ShutdownRequest\"\022\n\020ShutdownResponse\"\023" +
|
||||
"\n\021StopMasterRequest\"\024\n\022StopMasterRespons" +
|
||||
"e\"\020\n\016BalanceRequest\"\'\n\017BalanceResponse\022\024" +
|
||||
"\n\014balancer_ran\030\001 \002(\010\"<\n\031SetBalancerRunni" +
|
||||
"ngRequest\022\n\n\002on\030\001 \002(\010\022\023\n\013synchronous\030\002 \001" +
|
||||
"(\010\"8\n\032SetBalancerRunningResponse\022\032\n\022prev" +
|
||||
"_balance_value\030\001 \001(\010\"\024\n\022CatalogScanReque",
|
||||
"st\"*\n\023CatalogScanResponse\022\023\n\013scan_result" +
|
||||
"\030\001 \001(\005\"-\n\033EnableCatalogJanitorRequest\022\016\n" +
|
||||
"\006enable\030\001 \002(\010\"2\n\034EnableCatalogJanitorRes" +
|
||||
"ponse\022\022\n\nprev_value\030\001 \001(\010\" \n\036IsCatalogJa" +
|
||||
"nitorEnabledRequest\"0\n\037IsCatalogJanitorE" +
|
||||
"nabledResponse\022\r\n\005value\030\001 \002(\010\"=\n\023TakeSna" +
|
||||
"pshotRequest\022&\n\010snapshot\030\001 \002(\0132\024.Snapsho" +
|
||||
"tDescription\"0\n\024TakeSnapshotResponse\022\030\n\020" +
|
||||
"expected_timeout\030\001 \002(\003\"\025\n\023ListSnapshotRe" +
|
||||
"quest\"?\n\024ListSnapshotResponse\022\'\n\tsnapsho",
|
||||
"ts\030\001 \003(\0132\024.SnapshotDescription\"?\n\025Delete" +
|
||||
"SnapshotRequest\022&\n\010snapshot\030\001 \002(\0132\024.Snap" +
|
||||
"shotDescription\"\030\n\026DeleteSnapshotRespons" +
|
||||
"e\"@\n\026RestoreSnapshotRequest\022&\n\010snapshot\030" +
|
||||
"\001 \002(\0132\024.SnapshotDescription\"\031\n\027RestoreSn" +
|
||||
"apshotResponse\"?\n\025IsSnapshotDoneRequest\022" +
|
||||
"&\n\010snapshot\030\001 \001(\0132\024.SnapshotDescription\"" +
|
||||
"U\n\026IsSnapshotDoneResponse\022\023\n\004done\030\001 \001(\010:" +
|
||||
"\005false\022&\n\010snapshot\030\002 \001(\0132\024.SnapshotDescr" +
|
||||
"iption\"F\n\034IsRestoreSnapshotDoneRequest\022&",
|
||||
"\n\010snapshot\030\001 \001(\0132\024.SnapshotDescription\"3" +
|
||||
"\n\035IsRestoreSnapshotDoneResponse\022\022\n\004done\030" +
|
||||
"\001 \001(\010:\004true2\305\016\n\022MasterAdminService\0222\n\tAd" +
|
||||
"dColumn\022\021.AddColumnRequest\032\022.AddColumnRe" +
|
||||
"sponse\022;\n\014DeleteColumn\022\024.DeleteColumnReq" +
|
||||
"uest\032\025.DeleteColumnResponse\022;\n\014ModifyCol" +
|
||||
"umn\022\024.ModifyColumnRequest\032\025.ModifyColumn" +
|
||||
"Response\0225\n\nMoveRegion\022\022.MoveRegionReque" +
|
||||
"st\032\023.MoveRegionResponse\022Y\n\026DispatchMergi" +
|
||||
"ngRegions\022\036.DispatchMergingRegionsReques",
|
||||
"t\032\037.DispatchMergingRegionsResponse\022;\n\014As" +
|
||||
"signRegion\022\024.AssignRegionRequest\032\025.Assig" +
|
||||
"nRegionResponse\022A\n\016UnassignRegion\022\026.Unas" +
|
||||
"signRegionRequest\032\027.UnassignRegionRespon" +
|
||||
"se\022>\n\rOfflineRegion\022\025.OfflineRegionReque" +
|
||||
"st\032\026.OfflineRegionResponse\0228\n\013DeleteTabl" +
|
||||
"e\022\023.DeleteTableRequest\032\024.DeleteTableResp" +
|
||||
"onse\0228\n\013EnableTable\022\023.EnableTableRequest" +
|
||||
"\032\024.EnableTableResponse\022;\n\014DisableTable\022\024" +
|
||||
".DisableTableRequest\032\025.DisableTableRespo",
|
||||
"nse\0228\n\013ModifyTable\022\023.ModifyTableRequest\032" +
|
||||
"\024.ModifyTableResponse\0228\n\013CreateTable\022\023.C" +
|
||||
"reateTableRequest\032\024.CreateTableResponse\022" +
|
||||
"/\n\010Shutdown\022\020.ShutdownRequest\032\021.Shutdown" +
|
||||
"Response\0225\n\nStopMaster\022\022.StopMasterReque" +
|
||||
"st\032\023.StopMasterResponse\022,\n\007Balance\022\017.Bal" +
|
||||
"anceRequest\032\020.BalanceResponse\022M\n\022SetBala" +
|
||||
"ncerRunning\022\032.SetBalancerRunningRequest\032" +
|
||||
"\033.SetBalancerRunningResponse\022;\n\016RunCatal" +
|
||||
"ogScan\022\023.CatalogScanRequest\032\024.CatalogSca",
|
||||
"nResponse\022S\n\024EnableCatalogJanitor\022\034.Enab" +
|
||||
"leCatalogJanitorRequest\032\035.EnableCatalogJ" +
|
||||
"anitorResponse\022\\\n\027IsCatalogJanitorEnable" +
|
||||
"d\022\037.IsCatalogJanitorEnabledRequest\032 .IsC" +
|
||||
"atalogJanitorEnabledResponse\022L\n\021ExecMast" +
|
||||
"erService\022\032.CoprocessorServiceRequest\032\033." +
|
||||
"CoprocessorServiceResponse\0227\n\010Snapshot\022\024" +
|
||||
".TakeSnapshotRequest\032\025.TakeSnapshotRespo" +
|
||||
"nse\022D\n\025GetCompletedSnapshots\022\024.ListSnaps" +
|
||||
"hotRequest\032\025.ListSnapshotResponse\022A\n\016Del",
|
||||
"eteSnapshot\022\026.DeleteSnapshotRequest\032\027.De" +
|
||||
"leteSnapshotResponse\022A\n\016IsSnapshotDone\022\026" +
|
||||
".IsSnapshotDoneRequest\032\027.IsSnapshotDoneR" +
|
||||
"esponse\022D\n\017RestoreSnapshot\022\027.RestoreSnap" +
|
||||
"shotRequest\032\030.RestoreSnapshotResponse\022V\n" +
|
||||
"\025IsRestoreSnapshotDone\022\035.IsRestoreSnapsh" +
|
||||
"otDoneRequest\032\036.IsRestoreSnapshotDoneRes" +
|
||||
"ponse\022D\n\017IsMasterRunning\022\027.IsMasterRunni" +
|
||||
"ngRequest\032\030.IsMasterRunningResponseBG\n*o" +
|
||||
"rg.apache.hadoop.hbase.protobuf.generate",
|
||||
"dB\021MasterAdminProtosH\001\210\001\001\240\001\001"
|
||||
};
|
||||
com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
|
||||
new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() {
|
||||
|
|
|
@ -11,7 +11,7 @@ public final class MasterMonitorProtos {
|
|||
public interface GetSchemaAlterStatusRequestOrBuilder
|
||||
extends com.google.protobuf.MessageOrBuilder {
|
||||
|
||||
// required bytes tableName = 1;
|
||||
// required bytes table_name = 1;
|
||||
boolean hasTableName();
|
||||
com.google.protobuf.ByteString getTableName();
|
||||
}
|
||||
|
@ -44,8 +44,8 @@ public final class MasterMonitorProtos {
|
|||
}
|
||||
|
||||
private int bitField0_;
|
||||
// required bytes tableName = 1;
|
||||
public static final int TABLENAME_FIELD_NUMBER = 1;
|
||||
// required bytes table_name = 1;
|
||||
public static final int TABLE_NAME_FIELD_NUMBER = 1;
|
||||
private com.google.protobuf.ByteString tableName_;
|
||||
public boolean hasTableName() {
|
||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||
|
@ -127,7 +127,7 @@ public final class MasterMonitorProtos {
|
|||
int hash = 41;
|
||||
hash = (19 * hash) + getDescriptorForType().hashCode();
|
||||
if (hasTableName()) {
|
||||
hash = (37 * hash) + TABLENAME_FIELD_NUMBER;
|
||||
hash = (37 * hash) + TABLE_NAME_FIELD_NUMBER;
|
||||
hash = (53 * hash) + getTableName().hashCode();
|
||||
}
|
||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||
|
@ -355,7 +355,7 @@ public final class MasterMonitorProtos {
|
|||
|
||||
private int bitField0_;
|
||||
|
||||
// required bytes tableName = 1;
|
||||
// required bytes table_name = 1;
|
||||
private com.google.protobuf.ByteString tableName_ = com.google.protobuf.ByteString.EMPTY;
|
||||
public boolean hasTableName() {
|
||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||
|
@ -393,11 +393,11 @@ public final class MasterMonitorProtos {
|
|||
public interface GetSchemaAlterStatusResponseOrBuilder
|
||||
extends com.google.protobuf.MessageOrBuilder {
|
||||
|
||||
// optional uint32 yetToUpdateRegions = 1;
|
||||
// optional uint32 yet_to_update_regions = 1;
|
||||
boolean hasYetToUpdateRegions();
|
||||
int getYetToUpdateRegions();
|
||||
|
||||
// optional uint32 totalRegions = 2;
|
||||
// optional uint32 total_regions = 2;
|
||||
boolean hasTotalRegions();
|
||||
int getTotalRegions();
|
||||
}
|
||||
|
@ -430,8 +430,8 @@ public final class MasterMonitorProtos {
|
|||
}
|
||||
|
||||
private int bitField0_;
|
||||
// optional uint32 yetToUpdateRegions = 1;
|
||||
public static final int YETTOUPDATEREGIONS_FIELD_NUMBER = 1;
|
||||
// optional uint32 yet_to_update_regions = 1;
|
||||
public static final int YET_TO_UPDATE_REGIONS_FIELD_NUMBER = 1;
|
||||
private int yetToUpdateRegions_;
|
||||
public boolean hasYetToUpdateRegions() {
|
||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||
|
@ -440,8 +440,8 @@ public final class MasterMonitorProtos {
|
|||
return yetToUpdateRegions_;
|
||||
}
|
||||
|
||||
// optional uint32 totalRegions = 2;
|
||||
public static final int TOTALREGIONS_FIELD_NUMBER = 2;
|
||||
// optional uint32 total_regions = 2;
|
||||
public static final int TOTAL_REGIONS_FIELD_NUMBER = 2;
|
||||
private int totalRegions_;
|
||||
public boolean hasTotalRegions() {
|
||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||
|
@ -532,11 +532,11 @@ public final class MasterMonitorProtos {
|
|||
int hash = 41;
|
||||
hash = (19 * hash) + getDescriptorForType().hashCode();
|
||||
if (hasYetToUpdateRegions()) {
|
||||
hash = (37 * hash) + YETTOUPDATEREGIONS_FIELD_NUMBER;
|
||||
hash = (37 * hash) + YET_TO_UPDATE_REGIONS_FIELD_NUMBER;
|
||||
hash = (53 * hash) + getYetToUpdateRegions();
|
||||
}
|
||||
if (hasTotalRegions()) {
|
||||
hash = (37 * hash) + TOTALREGIONS_FIELD_NUMBER;
|
||||
hash = (37 * hash) + TOTAL_REGIONS_FIELD_NUMBER;
|
||||
hash = (53 * hash) + getTotalRegions();
|
||||
}
|
||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||
|
@ -774,7 +774,7 @@ public final class MasterMonitorProtos {
|
|||
|
||||
private int bitField0_;
|
||||
|
||||
// optional uint32 yetToUpdateRegions = 1;
|
||||
// optional uint32 yet_to_update_regions = 1;
|
||||
private int yetToUpdateRegions_ ;
|
||||
public boolean hasYetToUpdateRegions() {
|
||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||
|
@ -795,7 +795,7 @@ public final class MasterMonitorProtos {
|
|||
return this;
|
||||
}
|
||||
|
||||
// optional uint32 totalRegions = 2;
|
||||
// optional uint32 total_regions = 2;
|
||||
private int totalRegions_ ;
|
||||
public boolean hasTotalRegions() {
|
||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||
|
@ -830,7 +830,7 @@ public final class MasterMonitorProtos {
|
|||
public interface GetTableDescriptorsRequestOrBuilder
|
||||
extends com.google.protobuf.MessageOrBuilder {
|
||||
|
||||
// repeated string tableNames = 1;
|
||||
// repeated string table_names = 1;
|
||||
java.util.List<String> getTableNamesList();
|
||||
int getTableNamesCount();
|
||||
String getTableNames(int index);
|
||||
|
@ -863,8 +863,8 @@ public final class MasterMonitorProtos {
|
|||
return org.apache.hadoop.hbase.protobuf.generated.MasterMonitorProtos.internal_static_GetTableDescriptorsRequest_fieldAccessorTable;
|
||||
}
|
||||
|
||||
// repeated string tableNames = 1;
|
||||
public static final int TABLENAMES_FIELD_NUMBER = 1;
|
||||
// repeated string table_names = 1;
|
||||
public static final int TABLE_NAMES_FIELD_NUMBER = 1;
|
||||
private com.google.protobuf.LazyStringList tableNames_;
|
||||
public java.util.List<String>
|
||||
getTableNamesList() {
|
||||
|
@ -948,7 +948,7 @@ public final class MasterMonitorProtos {
|
|||
int hash = 41;
|
||||
hash = (19 * hash) + getDescriptorForType().hashCode();
|
||||
if (getTableNamesCount() > 0) {
|
||||
hash = (37 * hash) + TABLENAMES_FIELD_NUMBER;
|
||||
hash = (37 * hash) + TABLE_NAMES_FIELD_NUMBER;
|
||||
hash = (53 * hash) + getTableNamesList().hashCode();
|
||||
}
|
||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||
|
@ -1179,7 +1179,7 @@ public final class MasterMonitorProtos {
|
|||
|
||||
private int bitField0_;
|
||||
|
||||
// repeated string tableNames = 1;
|
||||
// repeated string table_names = 1;
|
||||
private com.google.protobuf.LazyStringList tableNames_ = com.google.protobuf.LazyStringArrayList.EMPTY;
|
||||
private void ensureTableNamesIsMutable() {
|
||||
if (!((bitField0_ & 0x00000001) == 0x00000001)) {
|
||||
|
@ -1249,7 +1249,7 @@ public final class MasterMonitorProtos {
|
|||
public interface GetTableDescriptorsResponseOrBuilder
|
||||
extends com.google.protobuf.MessageOrBuilder {
|
||||
|
||||
// repeated .TableSchema tableSchema = 1;
|
||||
// repeated .TableSchema table_schema = 1;
|
||||
java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema>
|
||||
getTableSchemaList();
|
||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema getTableSchema(int index);
|
||||
|
@ -1287,8 +1287,8 @@ public final class MasterMonitorProtos {
|
|||
return org.apache.hadoop.hbase.protobuf.generated.MasterMonitorProtos.internal_static_GetTableDescriptorsResponse_fieldAccessorTable;
|
||||
}
|
||||
|
||||
// repeated .TableSchema tableSchema = 1;
|
||||
public static final int TABLESCHEMA_FIELD_NUMBER = 1;
|
||||
// repeated .TableSchema table_schema = 1;
|
||||
public static final int TABLE_SCHEMA_FIELD_NUMBER = 1;
|
||||
private java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema> tableSchema_;
|
||||
public java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema> getTableSchemaList() {
|
||||
return tableSchema_;
|
||||
|
@ -1380,7 +1380,7 @@ public final class MasterMonitorProtos {
|
|||
int hash = 41;
|
||||
hash = (19 * hash) + getDescriptorForType().hashCode();
|
||||
if (getTableSchemaCount() > 0) {
|
||||
hash = (37 * hash) + TABLESCHEMA_FIELD_NUMBER;
|
||||
hash = (37 * hash) + TABLE_SCHEMA_FIELD_NUMBER;
|
||||
hash = (53 * hash) + getTableSchemaList().hashCode();
|
||||
}
|
||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||
|
@ -1642,7 +1642,7 @@ public final class MasterMonitorProtos {
|
|||
|
||||
private int bitField0_;
|
||||
|
||||
// repeated .TableSchema tableSchema = 1;
|
||||
// repeated .TableSchema table_schema = 1;
|
||||
private java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema> tableSchema_ =
|
||||
java.util.Collections.emptyList();
|
||||
private void ensureTableSchemaIsMutable() {
|
||||
|
@ -2142,7 +2142,7 @@ public final class MasterMonitorProtos {
|
|||
public interface GetClusterStatusResponseOrBuilder
|
||||
extends com.google.protobuf.MessageOrBuilder {
|
||||
|
||||
// required .ClusterStatus clusterStatus = 1;
|
||||
// required .ClusterStatus cluster_status = 1;
|
||||
boolean hasClusterStatus();
|
||||
org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.ClusterStatus getClusterStatus();
|
||||
org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.ClusterStatusOrBuilder getClusterStatusOrBuilder();
|
||||
|
@ -2176,8 +2176,8 @@ public final class MasterMonitorProtos {
|
|||
}
|
||||
|
||||
private int bitField0_;
|
||||
// required .ClusterStatus clusterStatus = 1;
|
||||
public static final int CLUSTERSTATUS_FIELD_NUMBER = 1;
|
||||
// required .ClusterStatus cluster_status = 1;
|
||||
public static final int CLUSTER_STATUS_FIELD_NUMBER = 1;
|
||||
private org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.ClusterStatus clusterStatus_;
|
||||
public boolean hasClusterStatus() {
|
||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||
|
@ -2266,7 +2266,7 @@ public final class MasterMonitorProtos {
|
|||
int hash = 41;
|
||||
hash = (19 * hash) + getDescriptorForType().hashCode();
|
||||
if (hasClusterStatus()) {
|
||||
hash = (37 * hash) + CLUSTERSTATUS_FIELD_NUMBER;
|
||||
hash = (37 * hash) + CLUSTER_STATUS_FIELD_NUMBER;
|
||||
hash = (53 * hash) + getClusterStatus().hashCode();
|
||||
}
|
||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||
|
@ -2511,7 +2511,7 @@ public final class MasterMonitorProtos {
|
|||
|
||||
private int bitField0_;
|
||||
|
||||
// required .ClusterStatus clusterStatus = 1;
|
||||
// required .ClusterStatus cluster_status = 1;
|
||||
private org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.ClusterStatus clusterStatus_ = org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.ClusterStatus.getDefaultInstance();
|
||||
private com.google.protobuf.SingleFieldBuilder<
|
||||
org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.ClusterStatus, org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.ClusterStatus.Builder, org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.ClusterStatusOrBuilder> clusterStatusBuilder_;
|
||||
|
@ -3068,26 +3068,26 @@ public final class MasterMonitorProtos {
|
|||
static {
|
||||
java.lang.String[] descriptorData = {
|
||||
"\n\023MasterMonitor.proto\032\014Master.proto\032\013hba" +
|
||||
"se.proto\032\023ClusterStatus.proto\"0\n\033GetSche" +
|
||||
"maAlterStatusRequest\022\021\n\ttableName\030\001 \002(\014\"" +
|
||||
"P\n\034GetSchemaAlterStatusResponse\022\032\n\022yetTo" +
|
||||
"UpdateRegions\030\001 \001(\r\022\024\n\014totalRegions\030\002 \001(" +
|
||||
"\r\"0\n\032GetTableDescriptorsRequest\022\022\n\ntable" +
|
||||
"Names\030\001 \003(\t\"@\n\033GetTableDescriptorsRespon" +
|
||||
"se\022!\n\013tableSchema\030\001 \003(\0132\014.TableSchema\"\031\n" +
|
||||
"\027GetClusterStatusRequest\"A\n\030GetClusterSt" +
|
||||
"atusResponse\022%\n\rclusterStatus\030\001 \002(\0132\016.Cl",
|
||||
"usterStatus2\314\002\n\024MasterMonitorService\022S\n\024" +
|
||||
"getSchemaAlterStatus\022\034.GetSchemaAlterSta" +
|
||||
"tusRequest\032\035.GetSchemaAlterStatusRespons" +
|
||||
"e\022P\n\023getTableDescriptors\022\033.GetTableDescr" +
|
||||
"iptorsRequest\032\034.GetTableDescriptorsRespo" +
|
||||
"nse\022G\n\020getClusterStatus\022\030.GetClusterStat" +
|
||||
"usRequest\032\031.GetClusterStatusResponse\022D\n\017" +
|
||||
"isMasterRunning\022\027.IsMasterRunningRequest" +
|
||||
"\032\030.IsMasterRunningResponseBI\n*org.apache" +
|
||||
".hadoop.hbase.protobuf.generatedB\023Master",
|
||||
"MonitorProtosH\001\210\001\001\240\001\001"
|
||||
"se.proto\032\023ClusterStatus.proto\"1\n\033GetSche" +
|
||||
"maAlterStatusRequest\022\022\n\ntable_name\030\001 \002(\014" +
|
||||
"\"T\n\034GetSchemaAlterStatusResponse\022\035\n\025yet_" +
|
||||
"to_update_regions\030\001 \001(\r\022\025\n\rtotal_regions" +
|
||||
"\030\002 \001(\r\"1\n\032GetTableDescriptorsRequest\022\023\n\013" +
|
||||
"table_names\030\001 \003(\t\"A\n\033GetTableDescriptors" +
|
||||
"Response\022\"\n\014table_schema\030\001 \003(\0132\014.TableSc" +
|
||||
"hema\"\031\n\027GetClusterStatusRequest\"B\n\030GetCl" +
|
||||
"usterStatusResponse\022&\n\016cluster_status\030\001 ",
|
||||
"\002(\0132\016.ClusterStatus2\314\002\n\024MasterMonitorSer" +
|
||||
"vice\022S\n\024GetSchemaAlterStatus\022\034.GetSchema" +
|
||||
"AlterStatusRequest\032\035.GetSchemaAlterStatu" +
|
||||
"sResponse\022P\n\023GetTableDescriptors\022\033.GetTa" +
|
||||
"bleDescriptorsRequest\032\034.GetTableDescript" +
|
||||
"orsResponse\022G\n\020GetClusterStatus\022\030.GetClu" +
|
||||
"sterStatusRequest\032\031.GetClusterStatusResp" +
|
||||
"onse\022D\n\017IsMasterRunning\022\027.IsMasterRunnin" +
|
||||
"gRequest\032\030.IsMasterRunningResponseBI\n*or" +
|
||||
"g.apache.hadoop.hbase.protobuf.generated",
|
||||
"B\023MasterMonitorProtosH\001\210\001\001\240\001\001"
|
||||
};
|
||||
com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
|
||||
new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() {
|
||||
|
|
|
@ -311,7 +311,7 @@ public final class MasterProtos {
|
|||
public interface IsMasterRunningResponseOrBuilder
|
||||
extends com.google.protobuf.MessageOrBuilder {
|
||||
|
||||
// required bool isMasterRunning = 1;
|
||||
// required bool is_master_running = 1;
|
||||
boolean hasIsMasterRunning();
|
||||
boolean getIsMasterRunning();
|
||||
}
|
||||
|
@ -344,8 +344,8 @@ public final class MasterProtos {
|
|||
}
|
||||
|
||||
private int bitField0_;
|
||||
// required bool isMasterRunning = 1;
|
||||
public static final int ISMASTERRUNNING_FIELD_NUMBER = 1;
|
||||
// required bool is_master_running = 1;
|
||||
public static final int IS_MASTER_RUNNING_FIELD_NUMBER = 1;
|
||||
private boolean isMasterRunning_;
|
||||
public boolean hasIsMasterRunning() {
|
||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||
|
@ -427,7 +427,7 @@ public final class MasterProtos {
|
|||
int hash = 41;
|
||||
hash = (19 * hash) + getDescriptorForType().hashCode();
|
||||
if (hasIsMasterRunning()) {
|
||||
hash = (37 * hash) + ISMASTERRUNNING_FIELD_NUMBER;
|
||||
hash = (37 * hash) + IS_MASTER_RUNNING_FIELD_NUMBER;
|
||||
hash = (53 * hash) + hashBoolean(getIsMasterRunning());
|
||||
}
|
||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||
|
@ -655,7 +655,7 @@ public final class MasterProtos {
|
|||
|
||||
private int bitField0_;
|
||||
|
||||
// required bool isMasterRunning = 1;
|
||||
// required bool is_master_running = 1;
|
||||
private boolean isMasterRunning_ ;
|
||||
public boolean hasIsMasterRunning() {
|
||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||
|
@ -928,12 +928,12 @@ public final class MasterProtos {
|
|||
static {
|
||||
java.lang.String[] descriptorData = {
|
||||
"\n\014Master.proto\"\030\n\026IsMasterRunningRequest" +
|
||||
"\"2\n\027IsMasterRunningResponse\022\027\n\017isMasterR" +
|
||||
"unning\030\001 \002(\0102U\n\rMasterService\022D\n\017isMaste" +
|
||||
"rRunning\022\027.IsMasterRunningRequest\032\030.IsMa" +
|
||||
"sterRunningResponseBB\n*org.apache.hadoop" +
|
||||
".hbase.protobuf.generatedB\014MasterProtosH" +
|
||||
"\001\210\001\001\240\001\001"
|
||||
"\"4\n\027IsMasterRunningResponse\022\031\n\021is_master" +
|
||||
"_running\030\001 \002(\0102U\n\rMasterService\022D\n\017IsMas" +
|
||||
"terRunning\022\027.IsMasterRunningRequest\032\030.Is" +
|
||||
"MasterRunningResponseBB\n*org.apache.hado" +
|
||||
"op.hbase.protobuf.generatedB\014MasterProto" +
|
||||
"sH\001\210\001\001\240\001\001"
|
||||
};
|
||||
com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
|
||||
new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() {
|
||||
|
|
|
@ -11,7 +11,7 @@ public final class MultiRowMutation {
|
|||
public interface MultiMutateRequestOrBuilder
|
||||
extends com.google.protobuf.MessageOrBuilder {
|
||||
|
||||
// repeated .MutationProto mutationRequest = 1;
|
||||
// repeated .MutationProto mutation_request = 1;
|
||||
java.util.List<org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MutationProto>
|
||||
getMutationRequestList();
|
||||
org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MutationProto getMutationRequest(int index);
|
||||
|
@ -49,8 +49,8 @@ public final class MultiRowMutation {
|
|||
return org.apache.hadoop.hbase.protobuf.generated.MultiRowMutation.internal_static_MultiMutateRequest_fieldAccessorTable;
|
||||
}
|
||||
|
||||
// repeated .MutationProto mutationRequest = 1;
|
||||
public static final int MUTATIONREQUEST_FIELD_NUMBER = 1;
|
||||
// repeated .MutationProto mutation_request = 1;
|
||||
public static final int MUTATION_REQUEST_FIELD_NUMBER = 1;
|
||||
private java.util.List<org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MutationProto> mutationRequest_;
|
||||
public java.util.List<org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MutationProto> getMutationRequestList() {
|
||||
return mutationRequest_;
|
||||
|
@ -142,7 +142,7 @@ public final class MultiRowMutation {
|
|||
int hash = 41;
|
||||
hash = (19 * hash) + getDescriptorForType().hashCode();
|
||||
if (getMutationRequestCount() > 0) {
|
||||
hash = (37 * hash) + MUTATIONREQUEST_FIELD_NUMBER;
|
||||
hash = (37 * hash) + MUTATION_REQUEST_FIELD_NUMBER;
|
||||
hash = (53 * hash) + getMutationRequestList().hashCode();
|
||||
}
|
||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||
|
@ -404,7 +404,7 @@ public final class MultiRowMutation {
|
|||
|
||||
private int bitField0_;
|
||||
|
||||
// repeated .MutationProto mutationRequest = 1;
|
||||
// repeated .MutationProto mutation_request = 1;
|
||||
private java.util.List<org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MutationProto> mutationRequest_ =
|
||||
java.util.Collections.emptyList();
|
||||
private void ensureMutationRequestIsMutable() {
|
||||
|
@ -1141,13 +1141,14 @@ public final class MultiRowMutation {
|
|||
descriptor;
|
||||
static {
|
||||
java.lang.String[] descriptorData = {
|
||||
"\n\026MultiRowMutation.proto\032\014Client.proto\"=" +
|
||||
"\n\022MultiMutateRequest\022\'\n\017mutationRequest\030" +
|
||||
"\001 \003(\0132\016.MutationProto\"\025\n\023MultiMutateResp" +
|
||||
"onse2R\n\027MultiRowMutationService\0227\n\nmutat" +
|
||||
"eRows\022\023.MultiMutateRequest\032\024.MultiMutate" +
|
||||
"ResponseBF\n*org.apache.hadoop.hbase.prot" +
|
||||
"obuf.generatedB\020MultiRowMutationH\001\210\001\001\240\001\001"
|
||||
"\n\026MultiRowMutation.proto\032\014Client.proto\">" +
|
||||
"\n\022MultiMutateRequest\022(\n\020mutation_request" +
|
||||
"\030\001 \003(\0132\016.MutationProto\"\025\n\023MultiMutateRes" +
|
||||
"ponse2R\n\027MultiRowMutationService\0227\n\nMuta" +
|
||||
"teRows\022\023.MultiMutateRequest\032\024.MultiMutat" +
|
||||
"eResponseBF\n*org.apache.hadoop.hbase.pro" +
|
||||
"tobuf.generatedB\020MultiRowMutationH\001\210\001\001\240\001" +
|
||||
"\001"
|
||||
};
|
||||
com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
|
||||
new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() {
|
||||
|
|
|
@ -11,11 +11,11 @@ public final class RPCProtos {
|
|||
public interface UserInformationOrBuilder
|
||||
extends com.google.protobuf.MessageOrBuilder {
|
||||
|
||||
// required string effectiveUser = 1;
|
||||
// required string effective_user = 1;
|
||||
boolean hasEffectiveUser();
|
||||
String getEffectiveUser();
|
||||
|
||||
// optional string realUser = 2;
|
||||
// optional string real_user = 2;
|
||||
boolean hasRealUser();
|
||||
String getRealUser();
|
||||
}
|
||||
|
@ -48,8 +48,8 @@ public final class RPCProtos {
|
|||
}
|
||||
|
||||
private int bitField0_;
|
||||
// required string effectiveUser = 1;
|
||||
public static final int EFFECTIVEUSER_FIELD_NUMBER = 1;
|
||||
// required string effective_user = 1;
|
||||
public static final int EFFECTIVE_USER_FIELD_NUMBER = 1;
|
||||
private java.lang.Object effectiveUser_;
|
||||
public boolean hasEffectiveUser() {
|
||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||
|
@ -80,8 +80,8 @@ public final class RPCProtos {
|
|||
}
|
||||
}
|
||||
|
||||
// optional string realUser = 2;
|
||||
public static final int REALUSER_FIELD_NUMBER = 2;
|
||||
// optional string real_user = 2;
|
||||
public static final int REAL_USER_FIELD_NUMBER = 2;
|
||||
private java.lang.Object realUser_;
|
||||
public boolean hasRealUser() {
|
||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||
|
@ -198,11 +198,11 @@ public final class RPCProtos {
|
|||
int hash = 41;
|
||||
hash = (19 * hash) + getDescriptorForType().hashCode();
|
||||
if (hasEffectiveUser()) {
|
||||
hash = (37 * hash) + EFFECTIVEUSER_FIELD_NUMBER;
|
||||
hash = (37 * hash) + EFFECTIVE_USER_FIELD_NUMBER;
|
||||
hash = (53 * hash) + getEffectiveUser().hashCode();
|
||||
}
|
||||
if (hasRealUser()) {
|
||||
hash = (37 * hash) + REALUSER_FIELD_NUMBER;
|
||||
hash = (37 * hash) + REAL_USER_FIELD_NUMBER;
|
||||
hash = (53 * hash) + getRealUser().hashCode();
|
||||
}
|
||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||
|
@ -444,7 +444,7 @@ public final class RPCProtos {
|
|||
|
||||
private int bitField0_;
|
||||
|
||||
// required string effectiveUser = 1;
|
||||
// required string effective_user = 1;
|
||||
private java.lang.Object effectiveUser_ = "";
|
||||
public boolean hasEffectiveUser() {
|
||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||
|
@ -480,7 +480,7 @@ public final class RPCProtos {
|
|||
onChanged();
|
||||
}
|
||||
|
||||
// optional string realUser = 2;
|
||||
// optional string real_user = 2;
|
||||
private java.lang.Object realUser_ = "";
|
||||
public boolean hasRealUser() {
|
||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||
|
@ -530,20 +530,20 @@ public final class RPCProtos {
|
|||
public interface ConnectionHeaderOrBuilder
|
||||
extends com.google.protobuf.MessageOrBuilder {
|
||||
|
||||
// optional .UserInformation userInfo = 1;
|
||||
// optional .UserInformation user_info = 1;
|
||||
boolean hasUserInfo();
|
||||
org.apache.hadoop.hbase.protobuf.generated.RPCProtos.UserInformation getUserInfo();
|
||||
org.apache.hadoop.hbase.protobuf.generated.RPCProtos.UserInformationOrBuilder getUserInfoOrBuilder();
|
||||
|
||||
// optional string serviceName = 2;
|
||||
// optional string service_name = 2;
|
||||
boolean hasServiceName();
|
||||
String getServiceName();
|
||||
|
||||
// optional string cellBlockCodecClass = 3 [default = "org.apache.hadoop.hbase.codec.KeyValueCodec"];
|
||||
// optional string cell_block_codec_class = 3 [default = "org.apache.hadoop.hbase.codec.KeyValueCodec"];
|
||||
boolean hasCellBlockCodecClass();
|
||||
String getCellBlockCodecClass();
|
||||
|
||||
// optional string cellBlockCompressorClass = 4;
|
||||
// optional string cell_block_compressor_class = 4;
|
||||
boolean hasCellBlockCompressorClass();
|
||||
String getCellBlockCompressorClass();
|
||||
}
|
||||
|
@ -576,8 +576,8 @@ public final class RPCProtos {
|
|||
}
|
||||
|
||||
private int bitField0_;
|
||||
// optional .UserInformation userInfo = 1;
|
||||
public static final int USERINFO_FIELD_NUMBER = 1;
|
||||
// optional .UserInformation user_info = 1;
|
||||
public static final int USER_INFO_FIELD_NUMBER = 1;
|
||||
private org.apache.hadoop.hbase.protobuf.generated.RPCProtos.UserInformation userInfo_;
|
||||
public boolean hasUserInfo() {
|
||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||
|
@ -589,8 +589,8 @@ public final class RPCProtos {
|
|||
return userInfo_;
|
||||
}
|
||||
|
||||
// optional string serviceName = 2;
|
||||
public static final int SERVICENAME_FIELD_NUMBER = 2;
|
||||
// optional string service_name = 2;
|
||||
public static final int SERVICE_NAME_FIELD_NUMBER = 2;
|
||||
private java.lang.Object serviceName_;
|
||||
public boolean hasServiceName() {
|
||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||
|
@ -621,8 +621,8 @@ public final class RPCProtos {
|
|||
}
|
||||
}
|
||||
|
||||
// optional string cellBlockCodecClass = 3 [default = "org.apache.hadoop.hbase.codec.KeyValueCodec"];
|
||||
public static final int CELLBLOCKCODECCLASS_FIELD_NUMBER = 3;
|
||||
// optional string cell_block_codec_class = 3 [default = "org.apache.hadoop.hbase.codec.KeyValueCodec"];
|
||||
public static final int CELL_BLOCK_CODEC_CLASS_FIELD_NUMBER = 3;
|
||||
private java.lang.Object cellBlockCodecClass_;
|
||||
public boolean hasCellBlockCodecClass() {
|
||||
return ((bitField0_ & 0x00000004) == 0x00000004);
|
||||
|
@ -653,8 +653,8 @@ public final class RPCProtos {
|
|||
}
|
||||
}
|
||||
|
||||
// optional string cellBlockCompressorClass = 4;
|
||||
public static final int CELLBLOCKCOMPRESSORCLASS_FIELD_NUMBER = 4;
|
||||
// optional string cell_block_compressor_class = 4;
|
||||
public static final int CELL_BLOCK_COMPRESSOR_CLASS_FIELD_NUMBER = 4;
|
||||
private java.lang.Object cellBlockCompressorClass_;
|
||||
public boolean hasCellBlockCompressorClass() {
|
||||
return ((bitField0_ & 0x00000008) == 0x00000008);
|
||||
|
@ -799,19 +799,19 @@ public final class RPCProtos {
|
|||
int hash = 41;
|
||||
hash = (19 * hash) + getDescriptorForType().hashCode();
|
||||
if (hasUserInfo()) {
|
||||
hash = (37 * hash) + USERINFO_FIELD_NUMBER;
|
||||
hash = (37 * hash) + USER_INFO_FIELD_NUMBER;
|
||||
hash = (53 * hash) + getUserInfo().hashCode();
|
||||
}
|
||||
if (hasServiceName()) {
|
||||
hash = (37 * hash) + SERVICENAME_FIELD_NUMBER;
|
||||
hash = (37 * hash) + SERVICE_NAME_FIELD_NUMBER;
|
||||
hash = (53 * hash) + getServiceName().hashCode();
|
||||
}
|
||||
if (hasCellBlockCodecClass()) {
|
||||
hash = (37 * hash) + CELLBLOCKCODECCLASS_FIELD_NUMBER;
|
||||
hash = (37 * hash) + CELL_BLOCK_CODEC_CLASS_FIELD_NUMBER;
|
||||
hash = (53 * hash) + getCellBlockCodecClass().hashCode();
|
||||
}
|
||||
if (hasCellBlockCompressorClass()) {
|
||||
hash = (37 * hash) + CELLBLOCKCOMPRESSORCLASS_FIELD_NUMBER;
|
||||
hash = (37 * hash) + CELL_BLOCK_COMPRESSOR_CLASS_FIELD_NUMBER;
|
||||
hash = (53 * hash) + getCellBlockCompressorClass().hashCode();
|
||||
}
|
||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||
|
@ -1096,7 +1096,7 @@ public final class RPCProtos {
|
|||
|
||||
private int bitField0_;
|
||||
|
||||
// optional .UserInformation userInfo = 1;
|
||||
// optional .UserInformation user_info = 1;
|
||||
private org.apache.hadoop.hbase.protobuf.generated.RPCProtos.UserInformation userInfo_ = org.apache.hadoop.hbase.protobuf.generated.RPCProtos.UserInformation.getDefaultInstance();
|
||||
private com.google.protobuf.SingleFieldBuilder<
|
||||
org.apache.hadoop.hbase.protobuf.generated.RPCProtos.UserInformation, org.apache.hadoop.hbase.protobuf.generated.RPCProtos.UserInformation.Builder, org.apache.hadoop.hbase.protobuf.generated.RPCProtos.UserInformationOrBuilder> userInfoBuilder_;
|
||||
|
@ -1186,7 +1186,7 @@ public final class RPCProtos {
|
|||
return userInfoBuilder_;
|
||||
}
|
||||
|
||||
// optional string serviceName = 2;
|
||||
// optional string service_name = 2;
|
||||
private java.lang.Object serviceName_ = "";
|
||||
public boolean hasServiceName() {
|
||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||
|
@ -1222,7 +1222,7 @@ public final class RPCProtos {
|
|||
onChanged();
|
||||
}
|
||||
|
||||
// optional string cellBlockCodecClass = 3 [default = "org.apache.hadoop.hbase.codec.KeyValueCodec"];
|
||||
// optional string cell_block_codec_class = 3 [default = "org.apache.hadoop.hbase.codec.KeyValueCodec"];
|
||||
private java.lang.Object cellBlockCodecClass_ = "org.apache.hadoop.hbase.codec.KeyValueCodec";
|
||||
public boolean hasCellBlockCodecClass() {
|
||||
return ((bitField0_ & 0x00000004) == 0x00000004);
|
||||
|
@ -1258,7 +1258,7 @@ public final class RPCProtos {
|
|||
onChanged();
|
||||
}
|
||||
|
||||
// optional string cellBlockCompressorClass = 4;
|
||||
// optional string cell_block_compressor_class = 4;
|
||||
private java.lang.Object cellBlockCompressorClass_ = "";
|
||||
public boolean hasCellBlockCompressorClass() {
|
||||
return ((bitField0_ & 0x00000008) == 0x00000008);
|
||||
|
@ -1679,11 +1679,11 @@ public final class RPCProtos {
|
|||
public interface ExceptionResponseOrBuilder
|
||||
extends com.google.protobuf.MessageOrBuilder {
|
||||
|
||||
// optional string exceptionClassName = 1;
|
||||
// optional string exception_class_name = 1;
|
||||
boolean hasExceptionClassName();
|
||||
String getExceptionClassName();
|
||||
|
||||
// optional string stackTrace = 2;
|
||||
// optional string stack_trace = 2;
|
||||
boolean hasStackTrace();
|
||||
String getStackTrace();
|
||||
|
||||
|
@ -1695,7 +1695,7 @@ public final class RPCProtos {
|
|||
boolean hasPort();
|
||||
int getPort();
|
||||
|
||||
// optional bool doNotRetry = 5;
|
||||
// optional bool do_not_retry = 5;
|
||||
boolean hasDoNotRetry();
|
||||
boolean getDoNotRetry();
|
||||
}
|
||||
|
@ -1728,8 +1728,8 @@ public final class RPCProtos {
|
|||
}
|
||||
|
||||
private int bitField0_;
|
||||
// optional string exceptionClassName = 1;
|
||||
public static final int EXCEPTIONCLASSNAME_FIELD_NUMBER = 1;
|
||||
// optional string exception_class_name = 1;
|
||||
public static final int EXCEPTION_CLASS_NAME_FIELD_NUMBER = 1;
|
||||
private java.lang.Object exceptionClassName_;
|
||||
public boolean hasExceptionClassName() {
|
||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||
|
@ -1760,8 +1760,8 @@ public final class RPCProtos {
|
|||
}
|
||||
}
|
||||
|
||||
// optional string stackTrace = 2;
|
||||
public static final int STACKTRACE_FIELD_NUMBER = 2;
|
||||
// optional string stack_trace = 2;
|
||||
public static final int STACK_TRACE_FIELD_NUMBER = 2;
|
||||
private java.lang.Object stackTrace_;
|
||||
public boolean hasStackTrace() {
|
||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||
|
@ -1834,8 +1834,8 @@ public final class RPCProtos {
|
|||
return port_;
|
||||
}
|
||||
|
||||
// optional bool doNotRetry = 5;
|
||||
public static final int DONOTRETRY_FIELD_NUMBER = 5;
|
||||
// optional bool do_not_retry = 5;
|
||||
public static final int DO_NOT_RETRY_FIELD_NUMBER = 5;
|
||||
private boolean doNotRetry_;
|
||||
public boolean hasDoNotRetry() {
|
||||
return ((bitField0_ & 0x00000010) == 0x00000010);
|
||||
|
@ -1965,11 +1965,11 @@ public final class RPCProtos {
|
|||
int hash = 41;
|
||||
hash = (19 * hash) + getDescriptorForType().hashCode();
|
||||
if (hasExceptionClassName()) {
|
||||
hash = (37 * hash) + EXCEPTIONCLASSNAME_FIELD_NUMBER;
|
||||
hash = (37 * hash) + EXCEPTION_CLASS_NAME_FIELD_NUMBER;
|
||||
hash = (53 * hash) + getExceptionClassName().hashCode();
|
||||
}
|
||||
if (hasStackTrace()) {
|
||||
hash = (37 * hash) + STACKTRACE_FIELD_NUMBER;
|
||||
hash = (37 * hash) + STACK_TRACE_FIELD_NUMBER;
|
||||
hash = (53 * hash) + getStackTrace().hashCode();
|
||||
}
|
||||
if (hasHostname()) {
|
||||
|
@ -1981,7 +1981,7 @@ public final class RPCProtos {
|
|||
hash = (53 * hash) + getPort();
|
||||
}
|
||||
if (hasDoNotRetry()) {
|
||||
hash = (37 * hash) + DONOTRETRY_FIELD_NUMBER;
|
||||
hash = (37 * hash) + DO_NOT_RETRY_FIELD_NUMBER;
|
||||
hash = (53 * hash) + hashBoolean(getDoNotRetry());
|
||||
}
|
||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||
|
@ -2261,7 +2261,7 @@ public final class RPCProtos {
|
|||
|
||||
private int bitField0_;
|
||||
|
||||
// optional string exceptionClassName = 1;
|
||||
// optional string exception_class_name = 1;
|
||||
private java.lang.Object exceptionClassName_ = "";
|
||||
public boolean hasExceptionClassName() {
|
||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||
|
@ -2297,7 +2297,7 @@ public final class RPCProtos {
|
|||
onChanged();
|
||||
}
|
||||
|
||||
// optional string stackTrace = 2;
|
||||
// optional string stack_trace = 2;
|
||||
private java.lang.Object stackTrace_ = "";
|
||||
public boolean hasStackTrace() {
|
||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||
|
@ -2390,7 +2390,7 @@ public final class RPCProtos {
|
|||
return this;
|
||||
}
|
||||
|
||||
// optional bool doNotRetry = 5;
|
||||
// optional bool do_not_retry = 5;
|
||||
private boolean doNotRetry_ ;
|
||||
public boolean hasDoNotRetry() {
|
||||
return ((bitField0_ & 0x00000010) == 0x00000010);
|
||||
|
@ -2425,24 +2425,24 @@ public final class RPCProtos {
|
|||
public interface RequestHeaderOrBuilder
|
||||
extends com.google.protobuf.MessageOrBuilder {
|
||||
|
||||
// optional uint32 callId = 1;
|
||||
// optional uint32 call_id = 1;
|
||||
boolean hasCallId();
|
||||
int getCallId();
|
||||
|
||||
// optional .RPCTInfo traceInfo = 2;
|
||||
// optional .RPCTInfo trace_info = 2;
|
||||
boolean hasTraceInfo();
|
||||
org.apache.hadoop.hbase.protobuf.generated.Tracing.RPCTInfo getTraceInfo();
|
||||
org.apache.hadoop.hbase.protobuf.generated.Tracing.RPCTInfoOrBuilder getTraceInfoOrBuilder();
|
||||
|
||||
// optional string methodName = 3;
|
||||
// optional string method_name = 3;
|
||||
boolean hasMethodName();
|
||||
String getMethodName();
|
||||
|
||||
// optional bool requestParam = 4;
|
||||
// optional bool request_param = 4;
|
||||
boolean hasRequestParam();
|
||||
boolean getRequestParam();
|
||||
|
||||
// optional .CellBlockMeta cellBlockMeta = 5;
|
||||
// optional .CellBlockMeta cell_block_meta = 5;
|
||||
boolean hasCellBlockMeta();
|
||||
org.apache.hadoop.hbase.protobuf.generated.RPCProtos.CellBlockMeta getCellBlockMeta();
|
||||
org.apache.hadoop.hbase.protobuf.generated.RPCProtos.CellBlockMetaOrBuilder getCellBlockMetaOrBuilder();
|
||||
|
@ -2476,8 +2476,8 @@ public final class RPCProtos {
|
|||
}
|
||||
|
||||
private int bitField0_;
|
||||
// optional uint32 callId = 1;
|
||||
public static final int CALLID_FIELD_NUMBER = 1;
|
||||
// optional uint32 call_id = 1;
|
||||
public static final int CALL_ID_FIELD_NUMBER = 1;
|
||||
private int callId_;
|
||||
public boolean hasCallId() {
|
||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||
|
@ -2486,8 +2486,8 @@ public final class RPCProtos {
|
|||
return callId_;
|
||||
}
|
||||
|
||||
// optional .RPCTInfo traceInfo = 2;
|
||||
public static final int TRACEINFO_FIELD_NUMBER = 2;
|
||||
// optional .RPCTInfo trace_info = 2;
|
||||
public static final int TRACE_INFO_FIELD_NUMBER = 2;
|
||||
private org.apache.hadoop.hbase.protobuf.generated.Tracing.RPCTInfo traceInfo_;
|
||||
public boolean hasTraceInfo() {
|
||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||
|
@ -2499,8 +2499,8 @@ public final class RPCProtos {
|
|||
return traceInfo_;
|
||||
}
|
||||
|
||||
// optional string methodName = 3;
|
||||
public static final int METHODNAME_FIELD_NUMBER = 3;
|
||||
// optional string method_name = 3;
|
||||
public static final int METHOD_NAME_FIELD_NUMBER = 3;
|
||||
private java.lang.Object methodName_;
|
||||
public boolean hasMethodName() {
|
||||
return ((bitField0_ & 0x00000004) == 0x00000004);
|
||||
|
@ -2531,8 +2531,8 @@ public final class RPCProtos {
|
|||
}
|
||||
}
|
||||
|
||||
// optional bool requestParam = 4;
|
||||
public static final int REQUESTPARAM_FIELD_NUMBER = 4;
|
||||
// optional bool request_param = 4;
|
||||
public static final int REQUEST_PARAM_FIELD_NUMBER = 4;
|
||||
private boolean requestParam_;
|
||||
public boolean hasRequestParam() {
|
||||
return ((bitField0_ & 0x00000008) == 0x00000008);
|
||||
|
@ -2541,8 +2541,8 @@ public final class RPCProtos {
|
|||
return requestParam_;
|
||||
}
|
||||
|
||||
// optional .CellBlockMeta cellBlockMeta = 5;
|
||||
public static final int CELLBLOCKMETA_FIELD_NUMBER = 5;
|
||||
// optional .CellBlockMeta cell_block_meta = 5;
|
||||
public static final int CELL_BLOCK_META_FIELD_NUMBER = 5;
|
||||
private org.apache.hadoop.hbase.protobuf.generated.RPCProtos.CellBlockMeta cellBlockMeta_;
|
||||
public boolean hasCellBlockMeta() {
|
||||
return ((bitField0_ & 0x00000010) == 0x00000010);
|
||||
|
@ -2675,23 +2675,23 @@ public final class RPCProtos {
|
|||
int hash = 41;
|
||||
hash = (19 * hash) + getDescriptorForType().hashCode();
|
||||
if (hasCallId()) {
|
||||
hash = (37 * hash) + CALLID_FIELD_NUMBER;
|
||||
hash = (37 * hash) + CALL_ID_FIELD_NUMBER;
|
||||
hash = (53 * hash) + getCallId();
|
||||
}
|
||||
if (hasTraceInfo()) {
|
||||
hash = (37 * hash) + TRACEINFO_FIELD_NUMBER;
|
||||
hash = (37 * hash) + TRACE_INFO_FIELD_NUMBER;
|
||||
hash = (53 * hash) + getTraceInfo().hashCode();
|
||||
}
|
||||
if (hasMethodName()) {
|
||||
hash = (37 * hash) + METHODNAME_FIELD_NUMBER;
|
||||
hash = (37 * hash) + METHOD_NAME_FIELD_NUMBER;
|
||||
hash = (53 * hash) + getMethodName().hashCode();
|
||||
}
|
||||
if (hasRequestParam()) {
|
||||
hash = (37 * hash) + REQUESTPARAM_FIELD_NUMBER;
|
||||
hash = (37 * hash) + REQUEST_PARAM_FIELD_NUMBER;
|
||||
hash = (53 * hash) + hashBoolean(getRequestParam());
|
||||
}
|
||||
if (hasCellBlockMeta()) {
|
||||
hash = (37 * hash) + CELLBLOCKMETA_FIELD_NUMBER;
|
||||
hash = (37 * hash) + CELL_BLOCK_META_FIELD_NUMBER;
|
||||
hash = (53 * hash) + getCellBlockMeta().hashCode();
|
||||
}
|
||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||
|
@ -2997,7 +2997,7 @@ public final class RPCProtos {
|
|||
|
||||
private int bitField0_;
|
||||
|
||||
// optional uint32 callId = 1;
|
||||
// optional uint32 call_id = 1;
|
||||
private int callId_ ;
|
||||
public boolean hasCallId() {
|
||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||
|
@ -3018,7 +3018,7 @@ public final class RPCProtos {
|
|||
return this;
|
||||
}
|
||||
|
||||
// optional .RPCTInfo traceInfo = 2;
|
||||
// optional .RPCTInfo trace_info = 2;
|
||||
private org.apache.hadoop.hbase.protobuf.generated.Tracing.RPCTInfo traceInfo_ = org.apache.hadoop.hbase.protobuf.generated.Tracing.RPCTInfo.getDefaultInstance();
|
||||
private com.google.protobuf.SingleFieldBuilder<
|
||||
org.apache.hadoop.hbase.protobuf.generated.Tracing.RPCTInfo, org.apache.hadoop.hbase.protobuf.generated.Tracing.RPCTInfo.Builder, org.apache.hadoop.hbase.protobuf.generated.Tracing.RPCTInfoOrBuilder> traceInfoBuilder_;
|
||||
|
@ -3108,7 +3108,7 @@ public final class RPCProtos {
|
|||
return traceInfoBuilder_;
|
||||
}
|
||||
|
||||
// optional string methodName = 3;
|
||||
// optional string method_name = 3;
|
||||
private java.lang.Object methodName_ = "";
|
||||
public boolean hasMethodName() {
|
||||
return ((bitField0_ & 0x00000004) == 0x00000004);
|
||||
|
@ -3144,7 +3144,7 @@ public final class RPCProtos {
|
|||
onChanged();
|
||||
}
|
||||
|
||||
// optional bool requestParam = 4;
|
||||
// optional bool request_param = 4;
|
||||
private boolean requestParam_ ;
|
||||
public boolean hasRequestParam() {
|
||||
return ((bitField0_ & 0x00000008) == 0x00000008);
|
||||
|
@ -3165,7 +3165,7 @@ public final class RPCProtos {
|
|||
return this;
|
||||
}
|
||||
|
||||
// optional .CellBlockMeta cellBlockMeta = 5;
|
||||
// optional .CellBlockMeta cell_block_meta = 5;
|
||||
private org.apache.hadoop.hbase.protobuf.generated.RPCProtos.CellBlockMeta cellBlockMeta_ = org.apache.hadoop.hbase.protobuf.generated.RPCProtos.CellBlockMeta.getDefaultInstance();
|
||||
private com.google.protobuf.SingleFieldBuilder<
|
||||
org.apache.hadoop.hbase.protobuf.generated.RPCProtos.CellBlockMeta, org.apache.hadoop.hbase.protobuf.generated.RPCProtos.CellBlockMeta.Builder, org.apache.hadoop.hbase.protobuf.generated.RPCProtos.CellBlockMetaOrBuilder> cellBlockMetaBuilder_;
|
||||
|
@ -3269,7 +3269,7 @@ public final class RPCProtos {
|
|||
public interface ResponseHeaderOrBuilder
|
||||
extends com.google.protobuf.MessageOrBuilder {
|
||||
|
||||
// optional uint32 callId = 1;
|
||||
// optional uint32 call_id = 1;
|
||||
boolean hasCallId();
|
||||
int getCallId();
|
||||
|
||||
|
@ -3278,7 +3278,7 @@ public final class RPCProtos {
|
|||
org.apache.hadoop.hbase.protobuf.generated.RPCProtos.ExceptionResponse getException();
|
||||
org.apache.hadoop.hbase.protobuf.generated.RPCProtos.ExceptionResponseOrBuilder getExceptionOrBuilder();
|
||||
|
||||
// optional .CellBlockMeta cellBlockMeta = 3;
|
||||
// optional .CellBlockMeta cell_block_meta = 3;
|
||||
boolean hasCellBlockMeta();
|
||||
org.apache.hadoop.hbase.protobuf.generated.RPCProtos.CellBlockMeta getCellBlockMeta();
|
||||
org.apache.hadoop.hbase.protobuf.generated.RPCProtos.CellBlockMetaOrBuilder getCellBlockMetaOrBuilder();
|
||||
|
@ -3312,8 +3312,8 @@ public final class RPCProtos {
|
|||
}
|
||||
|
||||
private int bitField0_;
|
||||
// optional uint32 callId = 1;
|
||||
public static final int CALLID_FIELD_NUMBER = 1;
|
||||
// optional uint32 call_id = 1;
|
||||
public static final int CALL_ID_FIELD_NUMBER = 1;
|
||||
private int callId_;
|
||||
public boolean hasCallId() {
|
||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||
|
@ -3335,8 +3335,8 @@ public final class RPCProtos {
|
|||
return exception_;
|
||||
}
|
||||
|
||||
// optional .CellBlockMeta cellBlockMeta = 3;
|
||||
public static final int CELLBLOCKMETA_FIELD_NUMBER = 3;
|
||||
// optional .CellBlockMeta cell_block_meta = 3;
|
||||
public static final int CELL_BLOCK_META_FIELD_NUMBER = 3;
|
||||
private org.apache.hadoop.hbase.protobuf.generated.RPCProtos.CellBlockMeta cellBlockMeta_;
|
||||
public boolean hasCellBlockMeta() {
|
||||
return ((bitField0_ & 0x00000004) == 0x00000004);
|
||||
|
@ -3443,7 +3443,7 @@ public final class RPCProtos {
|
|||
int hash = 41;
|
||||
hash = (19 * hash) + getDescriptorForType().hashCode();
|
||||
if (hasCallId()) {
|
||||
hash = (37 * hash) + CALLID_FIELD_NUMBER;
|
||||
hash = (37 * hash) + CALL_ID_FIELD_NUMBER;
|
||||
hash = (53 * hash) + getCallId();
|
||||
}
|
||||
if (hasException()) {
|
||||
|
@ -3451,7 +3451,7 @@ public final class RPCProtos {
|
|||
hash = (53 * hash) + getException().hashCode();
|
||||
}
|
||||
if (hasCellBlockMeta()) {
|
||||
hash = (37 * hash) + CELLBLOCKMETA_FIELD_NUMBER;
|
||||
hash = (37 * hash) + CELL_BLOCK_META_FIELD_NUMBER;
|
||||
hash = (53 * hash) + getCellBlockMeta().hashCode();
|
||||
}
|
||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||
|
@ -3729,7 +3729,7 @@ public final class RPCProtos {
|
|||
|
||||
private int bitField0_;
|
||||
|
||||
// optional uint32 callId = 1;
|
||||
// optional uint32 call_id = 1;
|
||||
private int callId_ ;
|
||||
public boolean hasCallId() {
|
||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||
|
@ -3840,7 +3840,7 @@ public final class RPCProtos {
|
|||
return exceptionBuilder_;
|
||||
}
|
||||
|
||||
// optional .CellBlockMeta cellBlockMeta = 3;
|
||||
// optional .CellBlockMeta cell_block_meta = 3;
|
||||
private org.apache.hadoop.hbase.protobuf.generated.RPCProtos.CellBlockMeta cellBlockMeta_ = org.apache.hadoop.hbase.protobuf.generated.RPCProtos.CellBlockMeta.getDefaultInstance();
|
||||
private com.google.protobuf.SingleFieldBuilder<
|
||||
org.apache.hadoop.hbase.protobuf.generated.RPCProtos.CellBlockMeta, org.apache.hadoop.hbase.protobuf.generated.RPCProtos.CellBlockMeta.Builder, org.apache.hadoop.hbase.protobuf.generated.RPCProtos.CellBlockMetaOrBuilder> cellBlockMetaBuilder_;
|
||||
|
@ -3981,24 +3981,25 @@ public final class RPCProtos {
|
|||
static {
|
||||
java.lang.String[] descriptorData = {
|
||||
"\n\tRPC.proto\032\rTracing.proto\032\013hbase.proto\"" +
|
||||
":\n\017UserInformation\022\025\n\reffectiveUser\030\001 \002(" +
|
||||
"\t\022\020\n\010realUser\030\002 \001(\t\"\267\001\n\020ConnectionHeader" +
|
||||
"\022\"\n\010userInfo\030\001 \001(\0132\020.UserInformation\022\023\n\013" +
|
||||
"serviceName\030\002 \001(\t\022H\n\023cellBlockCodecClass" +
|
||||
"\030\003 \001(\t:+org.apache.hadoop.hbase.codec.Ke" +
|
||||
"yValueCodec\022 \n\030cellBlockCompressorClass\030" +
|
||||
"\004 \001(\t\"\037\n\rCellBlockMeta\022\016\n\006length\030\001 \001(\r\"w" +
|
||||
"\n\021ExceptionResponse\022\032\n\022exceptionClassNam" +
|
||||
"e\030\001 \001(\t\022\022\n\nstackTrace\030\002 \001(\t\022\020\n\010hostname\030",
|
||||
"\003 \001(\t\022\014\n\004port\030\004 \001(\005\022\022\n\ndoNotRetry\030\005 \001(\010\"" +
|
||||
"\216\001\n\rRequestHeader\022\016\n\006callId\030\001 \001(\r\022\034\n\ttra" +
|
||||
"ceInfo\030\002 \001(\0132\t.RPCTInfo\022\022\n\nmethodName\030\003 " +
|
||||
"\001(\t\022\024\n\014requestParam\030\004 \001(\010\022%\n\rcellBlockMe" +
|
||||
"ta\030\005 \001(\0132\016.CellBlockMeta\"n\n\016ResponseHead" +
|
||||
"er\022\016\n\006callId\030\001 \001(\r\022%\n\texception\030\002 \001(\0132\022." +
|
||||
"ExceptionResponse\022%\n\rcellBlockMeta\030\003 \001(\013" +
|
||||
"2\016.CellBlockMetaB<\n*org.apache.hadoop.hb" +
|
||||
"ase.protobuf.generatedB\tRPCProtosH\001\240\001\001"
|
||||
"<\n\017UserInformation\022\026\n\016effective_user\030\001 \002" +
|
||||
"(\t\022\021\n\treal_user\030\002 \001(\t\"\277\001\n\020ConnectionHead" +
|
||||
"er\022#\n\tuser_info\030\001 \001(\0132\020.UserInformation\022" +
|
||||
"\024\n\014service_name\030\002 \001(\t\022K\n\026cell_block_code" +
|
||||
"c_class\030\003 \001(\t:+org.apache.hadoop.hbase.c" +
|
||||
"odec.KeyValueCodec\022#\n\033cell_block_compres" +
|
||||
"sor_class\030\004 \001(\t\"\037\n\rCellBlockMeta\022\016\n\006leng" +
|
||||
"th\030\001 \001(\r\"|\n\021ExceptionResponse\022\034\n\024excepti" +
|
||||
"on_class_name\030\001 \001(\t\022\023\n\013stack_trace\030\002 \001(\t",
|
||||
"\022\020\n\010hostname\030\003 \001(\t\022\014\n\004port\030\004 \001(\005\022\024\n\014do_n" +
|
||||
"ot_retry\030\005 \001(\010\"\224\001\n\rRequestHeader\022\017\n\007call" +
|
||||
"_id\030\001 \001(\r\022\035\n\ntrace_info\030\002 \001(\0132\t.RPCTInfo" +
|
||||
"\022\023\n\013method_name\030\003 \001(\t\022\025\n\rrequest_param\030\004" +
|
||||
" \001(\010\022\'\n\017cell_block_meta\030\005 \001(\0132\016.CellBloc" +
|
||||
"kMeta\"q\n\016ResponseHeader\022\017\n\007call_id\030\001 \001(\r" +
|
||||
"\022%\n\texception\030\002 \001(\0132\022.ExceptionResponse\022" +
|
||||
"\'\n\017cell_block_meta\030\003 \001(\0132\016.CellBlockMeta" +
|
||||
"B<\n*org.apache.hadoop.hbase.protobuf.gen" +
|
||||
"eratedB\tRPCProtosH\001\240\001\001"
|
||||
};
|
||||
com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
|
||||
new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() {
|
||||
|
|
|
@ -15,11 +15,11 @@ public final class RegionServerStatusProtos {
|
|||
boolean hasPort();
|
||||
int getPort();
|
||||
|
||||
// required uint64 serverStartCode = 2;
|
||||
// required uint64 server_start_code = 2;
|
||||
boolean hasServerStartCode();
|
||||
long getServerStartCode();
|
||||
|
||||
// required uint64 serverCurrentTime = 3;
|
||||
// required uint64 server_current_time = 3;
|
||||
boolean hasServerCurrentTime();
|
||||
long getServerCurrentTime();
|
||||
}
|
||||
|
@ -62,8 +62,8 @@ public final class RegionServerStatusProtos {
|
|||
return port_;
|
||||
}
|
||||
|
||||
// required uint64 serverStartCode = 2;
|
||||
public static final int SERVERSTARTCODE_FIELD_NUMBER = 2;
|
||||
// required uint64 server_start_code = 2;
|
||||
public static final int SERVER_START_CODE_FIELD_NUMBER = 2;
|
||||
private long serverStartCode_;
|
||||
public boolean hasServerStartCode() {
|
||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||
|
@ -72,8 +72,8 @@ public final class RegionServerStatusProtos {
|
|||
return serverStartCode_;
|
||||
}
|
||||
|
||||
// required uint64 serverCurrentTime = 3;
|
||||
public static final int SERVERCURRENTTIME_FIELD_NUMBER = 3;
|
||||
// required uint64 server_current_time = 3;
|
||||
public static final int SERVER_CURRENT_TIME_FIELD_NUMBER = 3;
|
||||
private long serverCurrentTime_;
|
||||
public boolean hasServerCurrentTime() {
|
||||
return ((bitField0_ & 0x00000004) == 0x00000004);
|
||||
|
@ -193,11 +193,11 @@ public final class RegionServerStatusProtos {
|
|||
hash = (53 * hash) + getPort();
|
||||
}
|
||||
if (hasServerStartCode()) {
|
||||
hash = (37 * hash) + SERVERSTARTCODE_FIELD_NUMBER;
|
||||
hash = (37 * hash) + SERVER_START_CODE_FIELD_NUMBER;
|
||||
hash = (53 * hash) + hashLong(getServerStartCode());
|
||||
}
|
||||
if (hasServerCurrentTime()) {
|
||||
hash = (37 * hash) + SERVERCURRENTTIME_FIELD_NUMBER;
|
||||
hash = (37 * hash) + SERVER_CURRENT_TIME_FIELD_NUMBER;
|
||||
hash = (53 * hash) + hashLong(getServerCurrentTime());
|
||||
}
|
||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||
|
@ -482,7 +482,7 @@ public final class RegionServerStatusProtos {
|
|||
return this;
|
||||
}
|
||||
|
||||
// required uint64 serverStartCode = 2;
|
||||
// required uint64 server_start_code = 2;
|
||||
private long serverStartCode_ ;
|
||||
public boolean hasServerStartCode() {
|
||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||
|
@ -503,7 +503,7 @@ public final class RegionServerStatusProtos {
|
|||
return this;
|
||||
}
|
||||
|
||||
// required uint64 serverCurrentTime = 3;
|
||||
// required uint64 server_current_time = 3;
|
||||
private long serverCurrentTime_ ;
|
||||
public boolean hasServerCurrentTime() {
|
||||
return ((bitField0_ & 0x00000004) == 0x00000004);
|
||||
|
@ -538,7 +538,7 @@ public final class RegionServerStatusProtos {
|
|||
public interface RegionServerStartupResponseOrBuilder
|
||||
extends com.google.protobuf.MessageOrBuilder {
|
||||
|
||||
// repeated .NameStringPair mapEntries = 1;
|
||||
// repeated .NameStringPair map_entries = 1;
|
||||
java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPair>
|
||||
getMapEntriesList();
|
||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPair getMapEntries(int index);
|
||||
|
@ -576,8 +576,8 @@ public final class RegionServerStatusProtos {
|
|||
return org.apache.hadoop.hbase.protobuf.generated.RegionServerStatusProtos.internal_static_RegionServerStartupResponse_fieldAccessorTable;
|
||||
}
|
||||
|
||||
// repeated .NameStringPair mapEntries = 1;
|
||||
public static final int MAPENTRIES_FIELD_NUMBER = 1;
|
||||
// repeated .NameStringPair map_entries = 1;
|
||||
public static final int MAP_ENTRIES_FIELD_NUMBER = 1;
|
||||
private java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPair> mapEntries_;
|
||||
public java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPair> getMapEntriesList() {
|
||||
return mapEntries_;
|
||||
|
@ -669,7 +669,7 @@ public final class RegionServerStatusProtos {
|
|||
int hash = 41;
|
||||
hash = (19 * hash) + getDescriptorForType().hashCode();
|
||||
if (getMapEntriesCount() > 0) {
|
||||
hash = (37 * hash) + MAPENTRIES_FIELD_NUMBER;
|
||||
hash = (37 * hash) + MAP_ENTRIES_FIELD_NUMBER;
|
||||
hash = (53 * hash) + getMapEntriesList().hashCode();
|
||||
}
|
||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||
|
@ -931,7 +931,7 @@ public final class RegionServerStatusProtos {
|
|||
|
||||
private int bitField0_;
|
||||
|
||||
// repeated .NameStringPair mapEntries = 1;
|
||||
// repeated .NameStringPair map_entries = 1;
|
||||
private java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPair> mapEntries_ =
|
||||
java.util.Collections.emptyList();
|
||||
private void ensureMapEntriesIsMutable() {
|
||||
|
@ -2073,7 +2073,7 @@ public final class RegionServerStatusProtos {
|
|||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName getServer();
|
||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerNameOrBuilder getServerOrBuilder();
|
||||
|
||||
// required string errorMessage = 2;
|
||||
// required string error_message = 2;
|
||||
boolean hasErrorMessage();
|
||||
String getErrorMessage();
|
||||
}
|
||||
|
@ -2119,8 +2119,8 @@ public final class RegionServerStatusProtos {
|
|||
return server_;
|
||||
}
|
||||
|
||||
// required string errorMessage = 2;
|
||||
public static final int ERRORMESSAGE_FIELD_NUMBER = 2;
|
||||
// required string error_message = 2;
|
||||
public static final int ERROR_MESSAGE_FIELD_NUMBER = 2;
|
||||
private java.lang.Object errorMessage_;
|
||||
public boolean hasErrorMessage() {
|
||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||
|
@ -2249,7 +2249,7 @@ public final class RegionServerStatusProtos {
|
|||
hash = (53 * hash) + getServer().hashCode();
|
||||
}
|
||||
if (hasErrorMessage()) {
|
||||
hash = (37 * hash) + ERRORMESSAGE_FIELD_NUMBER;
|
||||
hash = (37 * hash) + ERROR_MESSAGE_FIELD_NUMBER;
|
||||
hash = (53 * hash) + getErrorMessage().hashCode();
|
||||
}
|
||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||
|
@ -2602,7 +2602,7 @@ public final class RegionServerStatusProtos {
|
|||
return serverBuilder_;
|
||||
}
|
||||
|
||||
// required string errorMessage = 2;
|
||||
// required string error_message = 2;
|
||||
private java.lang.Object errorMessage_ = "";
|
||||
public boolean hasErrorMessage() {
|
||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||
|
@ -2952,7 +2952,7 @@ public final class RegionServerStatusProtos {
|
|||
public interface GetLastFlushedSequenceIdRequestOrBuilder
|
||||
extends com.google.protobuf.MessageOrBuilder {
|
||||
|
||||
// required bytes regionName = 1;
|
||||
// required bytes region_name = 1;
|
||||
boolean hasRegionName();
|
||||
com.google.protobuf.ByteString getRegionName();
|
||||
}
|
||||
|
@ -2985,8 +2985,8 @@ public final class RegionServerStatusProtos {
|
|||
}
|
||||
|
||||
private int bitField0_;
|
||||
// required bytes regionName = 1;
|
||||
public static final int REGIONNAME_FIELD_NUMBER = 1;
|
||||
// required bytes region_name = 1;
|
||||
public static final int REGION_NAME_FIELD_NUMBER = 1;
|
||||
private com.google.protobuf.ByteString regionName_;
|
||||
public boolean hasRegionName() {
|
||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||
|
@ -3068,7 +3068,7 @@ public final class RegionServerStatusProtos {
|
|||
int hash = 41;
|
||||
hash = (19 * hash) + getDescriptorForType().hashCode();
|
||||
if (hasRegionName()) {
|
||||
hash = (37 * hash) + REGIONNAME_FIELD_NUMBER;
|
||||
hash = (37 * hash) + REGION_NAME_FIELD_NUMBER;
|
||||
hash = (53 * hash) + getRegionName().hashCode();
|
||||
}
|
||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||
|
@ -3296,7 +3296,7 @@ public final class RegionServerStatusProtos {
|
|||
|
||||
private int bitField0_;
|
||||
|
||||
// required bytes regionName = 1;
|
||||
// required bytes region_name = 1;
|
||||
private com.google.protobuf.ByteString regionName_ = com.google.protobuf.ByteString.EMPTY;
|
||||
public boolean hasRegionName() {
|
||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||
|
@ -3334,7 +3334,7 @@ public final class RegionServerStatusProtos {
|
|||
public interface GetLastFlushedSequenceIdResponseOrBuilder
|
||||
extends com.google.protobuf.MessageOrBuilder {
|
||||
|
||||
// required uint64 lastFlushedSequenceId = 1;
|
||||
// required uint64 last_flushed_sequence_id = 1;
|
||||
boolean hasLastFlushedSequenceId();
|
||||
long getLastFlushedSequenceId();
|
||||
}
|
||||
|
@ -3367,8 +3367,8 @@ public final class RegionServerStatusProtos {
|
|||
}
|
||||
|
||||
private int bitField0_;
|
||||
// required uint64 lastFlushedSequenceId = 1;
|
||||
public static final int LASTFLUSHEDSEQUENCEID_FIELD_NUMBER = 1;
|
||||
// required uint64 last_flushed_sequence_id = 1;
|
||||
public static final int LAST_FLUSHED_SEQUENCE_ID_FIELD_NUMBER = 1;
|
||||
private long lastFlushedSequenceId_;
|
||||
public boolean hasLastFlushedSequenceId() {
|
||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||
|
@ -3450,7 +3450,7 @@ public final class RegionServerStatusProtos {
|
|||
int hash = 41;
|
||||
hash = (19 * hash) + getDescriptorForType().hashCode();
|
||||
if (hasLastFlushedSequenceId()) {
|
||||
hash = (37 * hash) + LASTFLUSHEDSEQUENCEID_FIELD_NUMBER;
|
||||
hash = (37 * hash) + LAST_FLUSHED_SEQUENCE_ID_FIELD_NUMBER;
|
||||
hash = (53 * hash) + hashLong(getLastFlushedSequenceId());
|
||||
}
|
||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||
|
@ -3678,7 +3678,7 @@ public final class RegionServerStatusProtos {
|
|||
|
||||
private int bitField0_;
|
||||
|
||||
// required uint64 lastFlushedSequenceId = 1;
|
||||
// required uint64 last_flushed_sequence_id = 1;
|
||||
private long lastFlushedSequenceId_ ;
|
||||
public boolean hasLastFlushedSequenceId() {
|
||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||
|
@ -4176,31 +4176,31 @@ public final class RegionServerStatusProtos {
|
|||
static {
|
||||
java.lang.String[] descriptorData = {
|
||||
"\n\030RegionServerStatus.proto\032\013hbase.proto\032" +
|
||||
"\023ClusterStatus.proto\"^\n\032RegionServerStar" +
|
||||
"tupRequest\022\014\n\004port\030\001 \002(\r\022\027\n\017serverStartC" +
|
||||
"ode\030\002 \002(\004\022\031\n\021serverCurrentTime\030\003 \002(\004\"B\n\033" +
|
||||
"RegionServerStartupResponse\022#\n\nmapEntrie" +
|
||||
"s\030\001 \003(\0132\017.NameStringPair\"S\n\031RegionServer" +
|
||||
"ReportRequest\022\033\n\006server\030\001 \002(\0132\013.ServerNa" +
|
||||
"me\022\031\n\004load\030\002 \001(\0132\013.ServerLoad\"\034\n\032RegionS" +
|
||||
"erverReportResponse\"N\n\031ReportRSFatalErro" +
|
||||
"rRequest\022\033\n\006server\030\001 \002(\0132\013.ServerName\022\024\n",
|
||||
"\014errorMessage\030\002 \002(\t\"\034\n\032ReportRSFatalErro" +
|
||||
"rResponse\"5\n\037GetLastFlushedSequenceIdReq" +
|
||||
"uest\022\022\n\nregionName\030\001 \002(\014\"A\n GetLastFlush" +
|
||||
"edSequenceIdResponse\022\035\n\025lastFlushedSeque" +
|
||||
"nceId\030\001 \002(\0042\354\002\n\031RegionServerStatusServic" +
|
||||
"e\022P\n\023regionServerStartup\022\033.RegionServerS" +
|
||||
"tartupRequest\032\034.RegionServerStartupRespo" +
|
||||
"nse\022M\n\022regionServerReport\022\032.RegionServer" +
|
||||
"ReportRequest\032\033.RegionServerReportRespon" +
|
||||
"se\022M\n\022reportRSFatalError\022\032.ReportRSFatal",
|
||||
"ErrorRequest\032\033.ReportRSFatalErrorRespons" +
|
||||
"e\022_\n\030getLastFlushedSequenceId\022 .GetLastF" +
|
||||
"lushedSequenceIdRequest\032!.GetLastFlushed" +
|
||||
"SequenceIdResponseBN\n*org.apache.hadoop." +
|
||||
"hbase.protobuf.generatedB\030RegionServerSt" +
|
||||
"atusProtosH\001\210\001\001\240\001\001"
|
||||
"\023ClusterStatus.proto\"b\n\032RegionServerStar" +
|
||||
"tupRequest\022\014\n\004port\030\001 \002(\r\022\031\n\021server_start" +
|
||||
"_code\030\002 \002(\004\022\033\n\023server_current_time\030\003 \002(\004" +
|
||||
"\"C\n\033RegionServerStartupResponse\022$\n\013map_e" +
|
||||
"ntries\030\001 \003(\0132\017.NameStringPair\"S\n\031RegionS" +
|
||||
"erverReportRequest\022\033\n\006server\030\001 \002(\0132\013.Ser" +
|
||||
"verName\022\031\n\004load\030\002 \001(\0132\013.ServerLoad\"\034\n\032Re" +
|
||||
"gionServerReportResponse\"O\n\031ReportRSFata" +
|
||||
"lErrorRequest\022\033\n\006server\030\001 \002(\0132\013.ServerNa",
|
||||
"me\022\025\n\rerror_message\030\002 \002(\t\"\034\n\032ReportRSFat" +
|
||||
"alErrorResponse\"6\n\037GetLastFlushedSequenc" +
|
||||
"eIdRequest\022\023\n\013region_name\030\001 \002(\014\"D\n GetLa" +
|
||||
"stFlushedSequenceIdResponse\022 \n\030last_flus" +
|
||||
"hed_sequence_id\030\001 \002(\0042\354\002\n\031RegionServerSt" +
|
||||
"atusService\022P\n\023RegionServerStartup\022\033.Reg" +
|
||||
"ionServerStartupRequest\032\034.RegionServerSt" +
|
||||
"artupResponse\022M\n\022RegionServerReport\022\032.Re" +
|
||||
"gionServerReportRequest\032\033.RegionServerRe" +
|
||||
"portResponse\022M\n\022ReportRSFatalError\022\032.Rep",
|
||||
"ortRSFatalErrorRequest\032\033.ReportRSFatalEr" +
|
||||
"rorResponse\022_\n\030GetLastFlushedSequenceId\022" +
|
||||
" .GetLastFlushedSequenceIdRequest\032!.GetL" +
|
||||
"astFlushedSequenceIdResponseBN\n*org.apac" +
|
||||
"he.hadoop.hbase.protobuf.generatedB\030Regi" +
|
||||
"onServerStatusProtosH\001\210\001\001\240\001\001"
|
||||
};
|
||||
com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
|
||||
new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() {
|
||||
|
|
|
@ -11,15 +11,15 @@ public final class RowProcessorProtos {
|
|||
public interface RowProcessorRequestOrBuilder
|
||||
extends com.google.protobuf.MessageOrBuilder {
|
||||
|
||||
// required string rowProcessorClassName = 1;
|
||||
// required string row_processor_class_name = 1;
|
||||
boolean hasRowProcessorClassName();
|
||||
String getRowProcessorClassName();
|
||||
|
||||
// optional string rowProcessorInitializerMessageName = 2;
|
||||
// optional string row_processor_initializer_message_name = 2;
|
||||
boolean hasRowProcessorInitializerMessageName();
|
||||
String getRowProcessorInitializerMessageName();
|
||||
|
||||
// optional bytes rowProcessorInitializerMessage = 3;
|
||||
// optional bytes row_processor_initializer_message = 3;
|
||||
boolean hasRowProcessorInitializerMessage();
|
||||
com.google.protobuf.ByteString getRowProcessorInitializerMessage();
|
||||
}
|
||||
|
@ -52,8 +52,8 @@ public final class RowProcessorProtos {
|
|||
}
|
||||
|
||||
private int bitField0_;
|
||||
// required string rowProcessorClassName = 1;
|
||||
public static final int ROWPROCESSORCLASSNAME_FIELD_NUMBER = 1;
|
||||
// required string row_processor_class_name = 1;
|
||||
public static final int ROW_PROCESSOR_CLASS_NAME_FIELD_NUMBER = 1;
|
||||
private java.lang.Object rowProcessorClassName_;
|
||||
public boolean hasRowProcessorClassName() {
|
||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||
|
@ -84,8 +84,8 @@ public final class RowProcessorProtos {
|
|||
}
|
||||
}
|
||||
|
||||
// optional string rowProcessorInitializerMessageName = 2;
|
||||
public static final int ROWPROCESSORINITIALIZERMESSAGENAME_FIELD_NUMBER = 2;
|
||||
// optional string row_processor_initializer_message_name = 2;
|
||||
public static final int ROW_PROCESSOR_INITIALIZER_MESSAGE_NAME_FIELD_NUMBER = 2;
|
||||
private java.lang.Object rowProcessorInitializerMessageName_;
|
||||
public boolean hasRowProcessorInitializerMessageName() {
|
||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||
|
@ -116,8 +116,8 @@ public final class RowProcessorProtos {
|
|||
}
|
||||
}
|
||||
|
||||
// optional bytes rowProcessorInitializerMessage = 3;
|
||||
public static final int ROWPROCESSORINITIALIZERMESSAGE_FIELD_NUMBER = 3;
|
||||
// optional bytes row_processor_initializer_message = 3;
|
||||
public static final int ROW_PROCESSOR_INITIALIZER_MESSAGE_FIELD_NUMBER = 3;
|
||||
private com.google.protobuf.ByteString rowProcessorInitializerMessage_;
|
||||
public boolean hasRowProcessorInitializerMessage() {
|
||||
return ((bitField0_ & 0x00000004) == 0x00000004);
|
||||
|
@ -225,15 +225,15 @@ public final class RowProcessorProtos {
|
|||
int hash = 41;
|
||||
hash = (19 * hash) + getDescriptorForType().hashCode();
|
||||
if (hasRowProcessorClassName()) {
|
||||
hash = (37 * hash) + ROWPROCESSORCLASSNAME_FIELD_NUMBER;
|
||||
hash = (37 * hash) + ROW_PROCESSOR_CLASS_NAME_FIELD_NUMBER;
|
||||
hash = (53 * hash) + getRowProcessorClassName().hashCode();
|
||||
}
|
||||
if (hasRowProcessorInitializerMessageName()) {
|
||||
hash = (37 * hash) + ROWPROCESSORINITIALIZERMESSAGENAME_FIELD_NUMBER;
|
||||
hash = (37 * hash) + ROW_PROCESSOR_INITIALIZER_MESSAGE_NAME_FIELD_NUMBER;
|
||||
hash = (53 * hash) + getRowProcessorInitializerMessageName().hashCode();
|
||||
}
|
||||
if (hasRowProcessorInitializerMessage()) {
|
||||
hash = (37 * hash) + ROWPROCESSORINITIALIZERMESSAGE_FIELD_NUMBER;
|
||||
hash = (37 * hash) + ROW_PROCESSOR_INITIALIZER_MESSAGE_FIELD_NUMBER;
|
||||
hash = (53 * hash) + getRowProcessorInitializerMessage().hashCode();
|
||||
}
|
||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||
|
@ -489,7 +489,7 @@ public final class RowProcessorProtos {
|
|||
|
||||
private int bitField0_;
|
||||
|
||||
// required string rowProcessorClassName = 1;
|
||||
// required string row_processor_class_name = 1;
|
||||
private java.lang.Object rowProcessorClassName_ = "";
|
||||
public boolean hasRowProcessorClassName() {
|
||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||
|
@ -525,7 +525,7 @@ public final class RowProcessorProtos {
|
|||
onChanged();
|
||||
}
|
||||
|
||||
// optional string rowProcessorInitializerMessageName = 2;
|
||||
// optional string row_processor_initializer_message_name = 2;
|
||||
private java.lang.Object rowProcessorInitializerMessageName_ = "";
|
||||
public boolean hasRowProcessorInitializerMessageName() {
|
||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||
|
@ -561,7 +561,7 @@ public final class RowProcessorProtos {
|
|||
onChanged();
|
||||
}
|
||||
|
||||
// optional bytes rowProcessorInitializerMessage = 3;
|
||||
// optional bytes row_processor_initializer_message = 3;
|
||||
private com.google.protobuf.ByteString rowProcessorInitializerMessage_ = com.google.protobuf.ByteString.EMPTY;
|
||||
public boolean hasRowProcessorInitializerMessage() {
|
||||
return ((bitField0_ & 0x00000004) == 0x00000004);
|
||||
|
@ -599,7 +599,7 @@ public final class RowProcessorProtos {
|
|||
public interface RowProcessorResultOrBuilder
|
||||
extends com.google.protobuf.MessageOrBuilder {
|
||||
|
||||
// required bytes rowProcessorResult = 1;
|
||||
// required bytes row_processor_result = 1;
|
||||
boolean hasRowProcessorResult();
|
||||
com.google.protobuf.ByteString getRowProcessorResult();
|
||||
}
|
||||
|
@ -632,8 +632,8 @@ public final class RowProcessorProtos {
|
|||
}
|
||||
|
||||
private int bitField0_;
|
||||
// required bytes rowProcessorResult = 1;
|
||||
public static final int ROWPROCESSORRESULT_FIELD_NUMBER = 1;
|
||||
// required bytes row_processor_result = 1;
|
||||
public static final int ROW_PROCESSOR_RESULT_FIELD_NUMBER = 1;
|
||||
private com.google.protobuf.ByteString rowProcessorResult_;
|
||||
public boolean hasRowProcessorResult() {
|
||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||
|
@ -715,7 +715,7 @@ public final class RowProcessorProtos {
|
|||
int hash = 41;
|
||||
hash = (19 * hash) + getDescriptorForType().hashCode();
|
||||
if (hasRowProcessorResult()) {
|
||||
hash = (37 * hash) + ROWPROCESSORRESULT_FIELD_NUMBER;
|
||||
hash = (37 * hash) + ROW_PROCESSOR_RESULT_FIELD_NUMBER;
|
||||
hash = (53 * hash) + getRowProcessorResult().hashCode();
|
||||
}
|
||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||
|
@ -943,7 +943,7 @@ public final class RowProcessorProtos {
|
|||
|
||||
private int bitField0_;
|
||||
|
||||
// required bytes rowProcessorResult = 1;
|
||||
// required bytes row_processor_result = 1;
|
||||
private com.google.protobuf.ByteString rowProcessorResult_ = com.google.protobuf.ByteString.EMPTY;
|
||||
public boolean hasRowProcessorResult() {
|
||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||
|
@ -1218,15 +1218,16 @@ public final class RowProcessorProtos {
|
|||
descriptor;
|
||||
static {
|
||||
java.lang.String[] descriptorData = {
|
||||
"\n\022RowProcessor.proto\"\210\001\n\023RowProcessorReq" +
|
||||
"uest\022\035\n\025rowProcessorClassName\030\001 \002(\t\022*\n\"r" +
|
||||
"owProcessorInitializerMessageName\030\002 \001(\t\022" +
|
||||
"&\n\036rowProcessorInitializerMessage\030\003 \001(\014\"" +
|
||||
"0\n\022RowProcessorResult\022\032\n\022rowProcessorRes" +
|
||||
"ult\030\001 \002(\0142K\n\023RowProcessorService\0224\n\007proc" +
|
||||
"ess\022\024.RowProcessorRequest\032\023.RowProcessor" +
|
||||
"ResultBH\n*org.apache.hadoop.hbase.protob" +
|
||||
"uf.generatedB\022RowProcessorProtosH\001\210\001\001\240\001\001"
|
||||
"\n\022RowProcessor.proto\"\222\001\n\023RowProcessorReq" +
|
||||
"uest\022 \n\030row_processor_class_name\030\001 \002(\t\022." +
|
||||
"\n&row_processor_initializer_message_name" +
|
||||
"\030\002 \001(\t\022)\n!row_processor_initializer_mess" +
|
||||
"age\030\003 \001(\014\"2\n\022RowProcessorResult\022\034\n\024row_p" +
|
||||
"rocessor_result\030\001 \002(\0142K\n\023RowProcessorSer" +
|
||||
"vice\0224\n\007Process\022\024.RowProcessorRequest\032\023." +
|
||||
"RowProcessorResultBH\n*org.apache.hadoop." +
|
||||
"hbase.protobuf.generatedB\022RowProcessorPr" +
|
||||
"otosH\001\210\001\001\240\001\001"
|
||||
};
|
||||
com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
|
||||
new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() {
|
||||
|
|
|
@ -11,7 +11,7 @@ public final class SecureBulkLoadProtos {
|
|||
public interface SecureBulkLoadHFilesRequestOrBuilder
|
||||
extends com.google.protobuf.MessageOrBuilder {
|
||||
|
||||
// repeated .BulkLoadHFileRequest.FamilyPath familyPath = 1;
|
||||
// repeated .BulkLoadHFileRequest.FamilyPath family_path = 1;
|
||||
java.util.List<org.apache.hadoop.hbase.protobuf.generated.ClientProtos.BulkLoadHFileRequest.FamilyPath>
|
||||
getFamilyPathList();
|
||||
org.apache.hadoop.hbase.protobuf.generated.ClientProtos.BulkLoadHFileRequest.FamilyPath getFamilyPath(int index);
|
||||
|
@ -21,16 +21,16 @@ public final class SecureBulkLoadProtos {
|
|||
org.apache.hadoop.hbase.protobuf.generated.ClientProtos.BulkLoadHFileRequest.FamilyPathOrBuilder getFamilyPathOrBuilder(
|
||||
int index);
|
||||
|
||||
// optional bool assignSeqNum = 2;
|
||||
// optional bool assign_seq_num = 2;
|
||||
boolean hasAssignSeqNum();
|
||||
boolean getAssignSeqNum();
|
||||
|
||||
// required .DelegationTokenProto fsToken = 3;
|
||||
// required .DelegationTokenProto fs_token = 3;
|
||||
boolean hasFsToken();
|
||||
org.apache.hadoop.hbase.protobuf.generated.SecureBulkLoadProtos.DelegationTokenProto getFsToken();
|
||||
org.apache.hadoop.hbase.protobuf.generated.SecureBulkLoadProtos.DelegationTokenProtoOrBuilder getFsTokenOrBuilder();
|
||||
|
||||
// required string bulkToken = 4;
|
||||
// required string bulk_token = 4;
|
||||
boolean hasBulkToken();
|
||||
String getBulkToken();
|
||||
}
|
||||
|
@ -63,8 +63,8 @@ public final class SecureBulkLoadProtos {
|
|||
}
|
||||
|
||||
private int bitField0_;
|
||||
// repeated .BulkLoadHFileRequest.FamilyPath familyPath = 1;
|
||||
public static final int FAMILYPATH_FIELD_NUMBER = 1;
|
||||
// repeated .BulkLoadHFileRequest.FamilyPath family_path = 1;
|
||||
public static final int FAMILY_PATH_FIELD_NUMBER = 1;
|
||||
private java.util.List<org.apache.hadoop.hbase.protobuf.generated.ClientProtos.BulkLoadHFileRequest.FamilyPath> familyPath_;
|
||||
public java.util.List<org.apache.hadoop.hbase.protobuf.generated.ClientProtos.BulkLoadHFileRequest.FamilyPath> getFamilyPathList() {
|
||||
return familyPath_;
|
||||
|
@ -84,8 +84,8 @@ public final class SecureBulkLoadProtos {
|
|||
return familyPath_.get(index);
|
||||
}
|
||||
|
||||
// optional bool assignSeqNum = 2;
|
||||
public static final int ASSIGNSEQNUM_FIELD_NUMBER = 2;
|
||||
// optional bool assign_seq_num = 2;
|
||||
public static final int ASSIGN_SEQ_NUM_FIELD_NUMBER = 2;
|
||||
private boolean assignSeqNum_;
|
||||
public boolean hasAssignSeqNum() {
|
||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||
|
@ -94,8 +94,8 @@ public final class SecureBulkLoadProtos {
|
|||
return assignSeqNum_;
|
||||
}
|
||||
|
||||
// required .DelegationTokenProto fsToken = 3;
|
||||
public static final int FSTOKEN_FIELD_NUMBER = 3;
|
||||
// required .DelegationTokenProto fs_token = 3;
|
||||
public static final int FS_TOKEN_FIELD_NUMBER = 3;
|
||||
private org.apache.hadoop.hbase.protobuf.generated.SecureBulkLoadProtos.DelegationTokenProto fsToken_;
|
||||
public boolean hasFsToken() {
|
||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||
|
@ -107,8 +107,8 @@ public final class SecureBulkLoadProtos {
|
|||
return fsToken_;
|
||||
}
|
||||
|
||||
// required string bulkToken = 4;
|
||||
public static final int BULKTOKEN_FIELD_NUMBER = 4;
|
||||
// required string bulk_token = 4;
|
||||
public static final int BULK_TOKEN_FIELD_NUMBER = 4;
|
||||
private java.lang.Object bulkToken_;
|
||||
public boolean hasBulkToken() {
|
||||
return ((bitField0_ & 0x00000004) == 0x00000004);
|
||||
|
@ -258,19 +258,19 @@ public final class SecureBulkLoadProtos {
|
|||
int hash = 41;
|
||||
hash = (19 * hash) + getDescriptorForType().hashCode();
|
||||
if (getFamilyPathCount() > 0) {
|
||||
hash = (37 * hash) + FAMILYPATH_FIELD_NUMBER;
|
||||
hash = (37 * hash) + FAMILY_PATH_FIELD_NUMBER;
|
||||
hash = (53 * hash) + getFamilyPathList().hashCode();
|
||||
}
|
||||
if (hasAssignSeqNum()) {
|
||||
hash = (37 * hash) + ASSIGNSEQNUM_FIELD_NUMBER;
|
||||
hash = (37 * hash) + ASSIGN_SEQ_NUM_FIELD_NUMBER;
|
||||
hash = (53 * hash) + hashBoolean(getAssignSeqNum());
|
||||
}
|
||||
if (hasFsToken()) {
|
||||
hash = (37 * hash) + FSTOKEN_FIELD_NUMBER;
|
||||
hash = (37 * hash) + FS_TOKEN_FIELD_NUMBER;
|
||||
hash = (53 * hash) + getFsToken().hashCode();
|
||||
}
|
||||
if (hasBulkToken()) {
|
||||
hash = (37 * hash) + BULKTOKEN_FIELD_NUMBER;
|
||||
hash = (37 * hash) + BULK_TOKEN_FIELD_NUMBER;
|
||||
hash = (53 * hash) + getBulkToken().hashCode();
|
||||
}
|
||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||
|
@ -597,7 +597,7 @@ public final class SecureBulkLoadProtos {
|
|||
|
||||
private int bitField0_;
|
||||
|
||||
// repeated .BulkLoadHFileRequest.FamilyPath familyPath = 1;
|
||||
// repeated .BulkLoadHFileRequest.FamilyPath family_path = 1;
|
||||
private java.util.List<org.apache.hadoop.hbase.protobuf.generated.ClientProtos.BulkLoadHFileRequest.FamilyPath> familyPath_ =
|
||||
java.util.Collections.emptyList();
|
||||
private void ensureFamilyPathIsMutable() {
|
||||
|
@ -783,7 +783,7 @@ public final class SecureBulkLoadProtos {
|
|||
return familyPathBuilder_;
|
||||
}
|
||||
|
||||
// optional bool assignSeqNum = 2;
|
||||
// optional bool assign_seq_num = 2;
|
||||
private boolean assignSeqNum_ ;
|
||||
public boolean hasAssignSeqNum() {
|
||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||
|
@ -804,7 +804,7 @@ public final class SecureBulkLoadProtos {
|
|||
return this;
|
||||
}
|
||||
|
||||
// required .DelegationTokenProto fsToken = 3;
|
||||
// required .DelegationTokenProto fs_token = 3;
|
||||
private org.apache.hadoop.hbase.protobuf.generated.SecureBulkLoadProtos.DelegationTokenProto fsToken_ = org.apache.hadoop.hbase.protobuf.generated.SecureBulkLoadProtos.DelegationTokenProto.getDefaultInstance();
|
||||
private com.google.protobuf.SingleFieldBuilder<
|
||||
org.apache.hadoop.hbase.protobuf.generated.SecureBulkLoadProtos.DelegationTokenProto, org.apache.hadoop.hbase.protobuf.generated.SecureBulkLoadProtos.DelegationTokenProto.Builder, org.apache.hadoop.hbase.protobuf.generated.SecureBulkLoadProtos.DelegationTokenProtoOrBuilder> fsTokenBuilder_;
|
||||
|
@ -894,7 +894,7 @@ public final class SecureBulkLoadProtos {
|
|||
return fsTokenBuilder_;
|
||||
}
|
||||
|
||||
// required string bulkToken = 4;
|
||||
// required string bulk_token = 4;
|
||||
private java.lang.Object bulkToken_ = "";
|
||||
public boolean hasBulkToken() {
|
||||
return ((bitField0_ & 0x00000008) == 0x00000008);
|
||||
|
@ -1972,7 +1972,7 @@ public final class SecureBulkLoadProtos {
|
|||
public interface PrepareBulkLoadRequestOrBuilder
|
||||
extends com.google.protobuf.MessageOrBuilder {
|
||||
|
||||
// required bytes tableName = 1;
|
||||
// required bytes table_name = 1;
|
||||
boolean hasTableName();
|
||||
com.google.protobuf.ByteString getTableName();
|
||||
}
|
||||
|
@ -2005,8 +2005,8 @@ public final class SecureBulkLoadProtos {
|
|||
}
|
||||
|
||||
private int bitField0_;
|
||||
// required bytes tableName = 1;
|
||||
public static final int TABLENAME_FIELD_NUMBER = 1;
|
||||
// required bytes table_name = 1;
|
||||
public static final int TABLE_NAME_FIELD_NUMBER = 1;
|
||||
private com.google.protobuf.ByteString tableName_;
|
||||
public boolean hasTableName() {
|
||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||
|
@ -2088,7 +2088,7 @@ public final class SecureBulkLoadProtos {
|
|||
int hash = 41;
|
||||
hash = (19 * hash) + getDescriptorForType().hashCode();
|
||||
if (hasTableName()) {
|
||||
hash = (37 * hash) + TABLENAME_FIELD_NUMBER;
|
||||
hash = (37 * hash) + TABLE_NAME_FIELD_NUMBER;
|
||||
hash = (53 * hash) + getTableName().hashCode();
|
||||
}
|
||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||
|
@ -2316,7 +2316,7 @@ public final class SecureBulkLoadProtos {
|
|||
|
||||
private int bitField0_;
|
||||
|
||||
// required bytes tableName = 1;
|
||||
// required bytes table_name = 1;
|
||||
private com.google.protobuf.ByteString tableName_ = com.google.protobuf.ByteString.EMPTY;
|
||||
public boolean hasTableName() {
|
||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||
|
@ -2354,7 +2354,7 @@ public final class SecureBulkLoadProtos {
|
|||
public interface PrepareBulkLoadResponseOrBuilder
|
||||
extends com.google.protobuf.MessageOrBuilder {
|
||||
|
||||
// required string bulkToken = 1;
|
||||
// required string bulk_token = 1;
|
||||
boolean hasBulkToken();
|
||||
String getBulkToken();
|
||||
}
|
||||
|
@ -2387,8 +2387,8 @@ public final class SecureBulkLoadProtos {
|
|||
}
|
||||
|
||||
private int bitField0_;
|
||||
// required string bulkToken = 1;
|
||||
public static final int BULKTOKEN_FIELD_NUMBER = 1;
|
||||
// required string bulk_token = 1;
|
||||
public static final int BULK_TOKEN_FIELD_NUMBER = 1;
|
||||
private java.lang.Object bulkToken_;
|
||||
public boolean hasBulkToken() {
|
||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||
|
@ -2492,7 +2492,7 @@ public final class SecureBulkLoadProtos {
|
|||
int hash = 41;
|
||||
hash = (19 * hash) + getDescriptorForType().hashCode();
|
||||
if (hasBulkToken()) {
|
||||
hash = (37 * hash) + BULKTOKEN_FIELD_NUMBER;
|
||||
hash = (37 * hash) + BULK_TOKEN_FIELD_NUMBER;
|
||||
hash = (53 * hash) + getBulkToken().hashCode();
|
||||
}
|
||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||
|
@ -2720,7 +2720,7 @@ public final class SecureBulkLoadProtos {
|
|||
|
||||
private int bitField0_;
|
||||
|
||||
// required string bulkToken = 1;
|
||||
// required string bulk_token = 1;
|
||||
private java.lang.Object bulkToken_ = "";
|
||||
public boolean hasBulkToken() {
|
||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||
|
@ -2770,7 +2770,7 @@ public final class SecureBulkLoadProtos {
|
|||
public interface CleanupBulkLoadRequestOrBuilder
|
||||
extends com.google.protobuf.MessageOrBuilder {
|
||||
|
||||
// required string bulkToken = 1;
|
||||
// required string bulk_token = 1;
|
||||
boolean hasBulkToken();
|
||||
String getBulkToken();
|
||||
}
|
||||
|
@ -2803,8 +2803,8 @@ public final class SecureBulkLoadProtos {
|
|||
}
|
||||
|
||||
private int bitField0_;
|
||||
// required string bulkToken = 1;
|
||||
public static final int BULKTOKEN_FIELD_NUMBER = 1;
|
||||
// required string bulk_token = 1;
|
||||
public static final int BULK_TOKEN_FIELD_NUMBER = 1;
|
||||
private java.lang.Object bulkToken_;
|
||||
public boolean hasBulkToken() {
|
||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||
|
@ -2908,7 +2908,7 @@ public final class SecureBulkLoadProtos {
|
|||
int hash = 41;
|
||||
hash = (19 * hash) + getDescriptorForType().hashCode();
|
||||
if (hasBulkToken()) {
|
||||
hash = (37 * hash) + BULKTOKEN_FIELD_NUMBER;
|
||||
hash = (37 * hash) + BULK_TOKEN_FIELD_NUMBER;
|
||||
hash = (53 * hash) + getBulkToken().hashCode();
|
||||
}
|
||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||
|
@ -3136,7 +3136,7 @@ public final class SecureBulkLoadProtos {
|
|||
|
||||
private int bitField0_;
|
||||
|
||||
// required string bulkToken = 1;
|
||||
// required string bulk_token = 1;
|
||||
private java.lang.Object bulkToken_ = "";
|
||||
public boolean hasBulkToken() {
|
||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||
|
@ -3879,27 +3879,27 @@ public final class SecureBulkLoadProtos {
|
|||
static {
|
||||
java.lang.String[] descriptorData = {
|
||||
"\n\024SecureBulkLoad.proto\032\013hbase.proto\032\014Cli" +
|
||||
"ent.proto\"\244\001\n\033SecureBulkLoadHFilesReques" +
|
||||
"t\0224\n\nfamilyPath\030\001 \003(\0132 .BulkLoadHFileReq" +
|
||||
"uest.FamilyPath\022\024\n\014assignSeqNum\030\002 \001(\010\022&\n" +
|
||||
"\007fsToken\030\003 \002(\0132\025.DelegationTokenProto\022\021\n" +
|
||||
"\tbulkToken\030\004 \002(\t\".\n\034SecureBulkLoadHFiles" +
|
||||
"Response\022\016\n\006loaded\030\001 \002(\010\"[\n\024DelegationTo" +
|
||||
"kenProto\022\022\n\nidentifier\030\001 \001(\014\022\020\n\010password" +
|
||||
"\030\002 \001(\014\022\014\n\004kind\030\003 \001(\t\022\017\n\007service\030\004 \001(\t\"+\n" +
|
||||
"\026PrepareBulkLoadRequest\022\021\n\ttableName\030\001 \002",
|
||||
"(\014\",\n\027PrepareBulkLoadResponse\022\021\n\tbulkTok" +
|
||||
"en\030\001 \002(\t\"+\n\026CleanupBulkLoadRequest\022\021\n\tbu" +
|
||||
"lkToken\030\001 \002(\t\"\031\n\027CleanupBulkLoadResponse" +
|
||||
"2\370\001\n\025SecureBulkLoadService\022D\n\017prepareBul" +
|
||||
"kLoad\022\027.PrepareBulkLoadRequest\032\030.Prepare" +
|
||||
"BulkLoadResponse\022S\n\024secureBulkLoadHFiles" +
|
||||
"\022\034.SecureBulkLoadHFilesRequest\032\035.SecureB" +
|
||||
"ulkLoadHFilesResponse\022D\n\017cleanupBulkLoad" +
|
||||
"\022\027.CleanupBulkLoadRequest\032\030.CleanupBulkL" +
|
||||
"oadResponseBJ\n*org.apache.hadoop.hbase.p",
|
||||
"rotobuf.generatedB\024SecureBulkLoadProtosH" +
|
||||
"\001\210\001\001\240\001\001"
|
||||
"ent.proto\"\251\001\n\033SecureBulkLoadHFilesReques" +
|
||||
"t\0225\n\013family_path\030\001 \003(\0132 .BulkLoadHFileRe" +
|
||||
"quest.FamilyPath\022\026\n\016assign_seq_num\030\002 \001(\010" +
|
||||
"\022\'\n\010fs_token\030\003 \002(\0132\025.DelegationTokenProt" +
|
||||
"o\022\022\n\nbulk_token\030\004 \002(\t\".\n\034SecureBulkLoadH" +
|
||||
"FilesResponse\022\016\n\006loaded\030\001 \002(\010\"[\n\024Delegat" +
|
||||
"ionTokenProto\022\022\n\nidentifier\030\001 \001(\014\022\020\n\010pas" +
|
||||
"sword\030\002 \001(\014\022\014\n\004kind\030\003 \001(\t\022\017\n\007service\030\004 \001" +
|
||||
"(\t\",\n\026PrepareBulkLoadRequest\022\022\n\ntable_na",
|
||||
"me\030\001 \002(\014\"-\n\027PrepareBulkLoadResponse\022\022\n\nb" +
|
||||
"ulk_token\030\001 \002(\t\",\n\026CleanupBulkLoadReques" +
|
||||
"t\022\022\n\nbulk_token\030\001 \002(\t\"\031\n\027CleanupBulkLoad" +
|
||||
"Response2\370\001\n\025SecureBulkLoadService\022D\n\017Pr" +
|
||||
"epareBulkLoad\022\027.PrepareBulkLoadRequest\032\030" +
|
||||
".PrepareBulkLoadResponse\022S\n\024SecureBulkLo" +
|
||||
"adHFiles\022\034.SecureBulkLoadHFilesRequest\032\035" +
|
||||
".SecureBulkLoadHFilesResponse\022D\n\017Cleanup" +
|
||||
"BulkLoad\022\027.CleanupBulkLoadRequest\032\030.Clea" +
|
||||
"nupBulkLoadResponseBJ\n*org.apache.hadoop",
|
||||
".hbase.protobuf.generatedB\024SecureBulkLoa" +
|
||||
"dProtosH\001\210\001\001\240\001\001"
|
||||
};
|
||||
com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
|
||||
new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() {
|
||||
|
|
|
@ -11,11 +11,11 @@ public final class Tracing {
|
|||
public interface RPCTInfoOrBuilder
|
||||
extends com.google.protobuf.MessageOrBuilder {
|
||||
|
||||
// optional int64 traceId = 1;
|
||||
// optional int64 trace_id = 1;
|
||||
boolean hasTraceId();
|
||||
long getTraceId();
|
||||
|
||||
// optional int64 parentId = 2;
|
||||
// optional int64 parent_id = 2;
|
||||
boolean hasParentId();
|
||||
long getParentId();
|
||||
}
|
||||
|
@ -48,8 +48,8 @@ public final class Tracing {
|
|||
}
|
||||
|
||||
private int bitField0_;
|
||||
// optional int64 traceId = 1;
|
||||
public static final int TRACEID_FIELD_NUMBER = 1;
|
||||
// optional int64 trace_id = 1;
|
||||
public static final int TRACE_ID_FIELD_NUMBER = 1;
|
||||
private long traceId_;
|
||||
public boolean hasTraceId() {
|
||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||
|
@ -58,8 +58,8 @@ public final class Tracing {
|
|||
return traceId_;
|
||||
}
|
||||
|
||||
// optional int64 parentId = 2;
|
||||
public static final int PARENTID_FIELD_NUMBER = 2;
|
||||
// optional int64 parent_id = 2;
|
||||
public static final int PARENT_ID_FIELD_NUMBER = 2;
|
||||
private long parentId_;
|
||||
public boolean hasParentId() {
|
||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||
|
@ -150,11 +150,11 @@ public final class Tracing {
|
|||
int hash = 41;
|
||||
hash = (19 * hash) + getDescriptorForType().hashCode();
|
||||
if (hasTraceId()) {
|
||||
hash = (37 * hash) + TRACEID_FIELD_NUMBER;
|
||||
hash = (37 * hash) + TRACE_ID_FIELD_NUMBER;
|
||||
hash = (53 * hash) + hashLong(getTraceId());
|
||||
}
|
||||
if (hasParentId()) {
|
||||
hash = (37 * hash) + PARENTID_FIELD_NUMBER;
|
||||
hash = (37 * hash) + PARENT_ID_FIELD_NUMBER;
|
||||
hash = (53 * hash) + hashLong(getParentId());
|
||||
}
|
||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||
|
@ -392,7 +392,7 @@ public final class Tracing {
|
|||
|
||||
private int bitField0_;
|
||||
|
||||
// optional int64 traceId = 1;
|
||||
// optional int64 trace_id = 1;
|
||||
private long traceId_ ;
|
||||
public boolean hasTraceId() {
|
||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||
|
@ -413,7 +413,7 @@ public final class Tracing {
|
|||
return this;
|
||||
}
|
||||
|
||||
// optional int64 parentId = 2;
|
||||
// optional int64 parent_id = 2;
|
||||
private long parentId_ ;
|
||||
public boolean hasParentId() {
|
||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||
|
@ -459,10 +459,10 @@ public final class Tracing {
|
|||
descriptor;
|
||||
static {
|
||||
java.lang.String[] descriptorData = {
|
||||
"\n\rTracing.proto\"-\n\010RPCTInfo\022\017\n\007traceId\030\001" +
|
||||
" \001(\003\022\020\n\010parentId\030\002 \001(\003B:\n*org.apache.had" +
|
||||
"oop.hbase.protobuf.generatedB\007TracingH\001\240" +
|
||||
"\001\001"
|
||||
"\n\rTracing.proto\"/\n\010RPCTInfo\022\020\n\010trace_id\030" +
|
||||
"\001 \001(\003\022\021\n\tparent_id\030\002 \001(\003B:\n*org.apache.h" +
|
||||
"adoop.hbase.protobuf.generatedB\007TracingH" +
|
||||
"\001\240\001\001"
|
||||
};
|
||||
com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
|
||||
new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() {
|
||||
|
|
|
@ -80,7 +80,7 @@ public final class WALProtos {
|
|||
public interface WALHeaderOrBuilder
|
||||
extends com.google.protobuf.MessageOrBuilder {
|
||||
|
||||
// optional bool hasCompression = 1;
|
||||
// optional bool has_compression = 1;
|
||||
boolean hasHasCompression();
|
||||
boolean getHasCompression();
|
||||
}
|
||||
|
@ -113,8 +113,8 @@ public final class WALProtos {
|
|||
}
|
||||
|
||||
private int bitField0_;
|
||||
// optional bool hasCompression = 1;
|
||||
public static final int HASCOMPRESSION_FIELD_NUMBER = 1;
|
||||
// optional bool has_compression = 1;
|
||||
public static final int HAS_COMPRESSION_FIELD_NUMBER = 1;
|
||||
private boolean hasCompression_;
|
||||
public boolean hasHasCompression() {
|
||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||
|
@ -192,7 +192,7 @@ public final class WALProtos {
|
|||
int hash = 41;
|
||||
hash = (19 * hash) + getDescriptorForType().hashCode();
|
||||
if (hasHasCompression()) {
|
||||
hash = (37 * hash) + HASCOMPRESSION_FIELD_NUMBER;
|
||||
hash = (37 * hash) + HAS_COMPRESSION_FIELD_NUMBER;
|
||||
hash = (53 * hash) + hashBoolean(getHasCompression());
|
||||
}
|
||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||
|
@ -416,7 +416,7 @@ public final class WALProtos {
|
|||
|
||||
private int bitField0_;
|
||||
|
||||
// optional bool hasCompression = 1;
|
||||
// optional bool has_compression = 1;
|
||||
private boolean hasCompression_ ;
|
||||
public boolean hasHasCompression() {
|
||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||
|
@ -451,23 +451,23 @@ public final class WALProtos {
|
|||
public interface WALKeyOrBuilder
|
||||
extends com.google.protobuf.MessageOrBuilder {
|
||||
|
||||
// required bytes encodedRegionName = 1;
|
||||
// required bytes encoded_region_name = 1;
|
||||
boolean hasEncodedRegionName();
|
||||
com.google.protobuf.ByteString getEncodedRegionName();
|
||||
|
||||
// required bytes tableName = 2;
|
||||
// required bytes table_name = 2;
|
||||
boolean hasTableName();
|
||||
com.google.protobuf.ByteString getTableName();
|
||||
|
||||
// required uint64 logSequenceNumber = 3;
|
||||
// required uint64 log_sequence_number = 3;
|
||||
boolean hasLogSequenceNumber();
|
||||
long getLogSequenceNumber();
|
||||
|
||||
// required uint64 writeTime = 4;
|
||||
// required uint64 write_time = 4;
|
||||
boolean hasWriteTime();
|
||||
long getWriteTime();
|
||||
|
||||
// optional .UUID clusterId = 5;
|
||||
// optional .UUID cluster_id = 5;
|
||||
boolean hasClusterId();
|
||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.UUID getClusterId();
|
||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.UUIDOrBuilder getClusterIdOrBuilder();
|
||||
|
@ -482,7 +482,7 @@ public final class WALProtos {
|
|||
org.apache.hadoop.hbase.protobuf.generated.WALProtos.FamilyScopeOrBuilder getScopesOrBuilder(
|
||||
int index);
|
||||
|
||||
// optional uint32 followingKvCount = 7;
|
||||
// optional uint32 following_kv_count = 7;
|
||||
boolean hasFollowingKvCount();
|
||||
int getFollowingKvCount();
|
||||
}
|
||||
|
@ -515,8 +515,8 @@ public final class WALProtos {
|
|||
}
|
||||
|
||||
private int bitField0_;
|
||||
// required bytes encodedRegionName = 1;
|
||||
public static final int ENCODEDREGIONNAME_FIELD_NUMBER = 1;
|
||||
// required bytes encoded_region_name = 1;
|
||||
public static final int ENCODED_REGION_NAME_FIELD_NUMBER = 1;
|
||||
private com.google.protobuf.ByteString encodedRegionName_;
|
||||
public boolean hasEncodedRegionName() {
|
||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||
|
@ -525,8 +525,8 @@ public final class WALProtos {
|
|||
return encodedRegionName_;
|
||||
}
|
||||
|
||||
// required bytes tableName = 2;
|
||||
public static final int TABLENAME_FIELD_NUMBER = 2;
|
||||
// required bytes table_name = 2;
|
||||
public static final int TABLE_NAME_FIELD_NUMBER = 2;
|
||||
private com.google.protobuf.ByteString tableName_;
|
||||
public boolean hasTableName() {
|
||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||
|
@ -535,8 +535,8 @@ public final class WALProtos {
|
|||
return tableName_;
|
||||
}
|
||||
|
||||
// required uint64 logSequenceNumber = 3;
|
||||
public static final int LOGSEQUENCENUMBER_FIELD_NUMBER = 3;
|
||||
// required uint64 log_sequence_number = 3;
|
||||
public static final int LOG_SEQUENCE_NUMBER_FIELD_NUMBER = 3;
|
||||
private long logSequenceNumber_;
|
||||
public boolean hasLogSequenceNumber() {
|
||||
return ((bitField0_ & 0x00000004) == 0x00000004);
|
||||
|
@ -545,8 +545,8 @@ public final class WALProtos {
|
|||
return logSequenceNumber_;
|
||||
}
|
||||
|
||||
// required uint64 writeTime = 4;
|
||||
public static final int WRITETIME_FIELD_NUMBER = 4;
|
||||
// required uint64 write_time = 4;
|
||||
public static final int WRITE_TIME_FIELD_NUMBER = 4;
|
||||
private long writeTime_;
|
||||
public boolean hasWriteTime() {
|
||||
return ((bitField0_ & 0x00000008) == 0x00000008);
|
||||
|
@ -555,8 +555,8 @@ public final class WALProtos {
|
|||
return writeTime_;
|
||||
}
|
||||
|
||||
// optional .UUID clusterId = 5;
|
||||
public static final int CLUSTERID_FIELD_NUMBER = 5;
|
||||
// optional .UUID cluster_id = 5;
|
||||
public static final int CLUSTER_ID_FIELD_NUMBER = 5;
|
||||
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.UUID clusterId_;
|
||||
public boolean hasClusterId() {
|
||||
return ((bitField0_ & 0x00000010) == 0x00000010);
|
||||
|
@ -589,8 +589,8 @@ public final class WALProtos {
|
|||
return scopes_.get(index);
|
||||
}
|
||||
|
||||
// optional uint32 followingKvCount = 7;
|
||||
public static final int FOLLOWINGKVCOUNT_FIELD_NUMBER = 7;
|
||||
// optional uint32 following_kv_count = 7;
|
||||
public static final int FOLLOWING_KV_COUNT_FIELD_NUMBER = 7;
|
||||
private int followingKvCount_;
|
||||
public boolean hasFollowingKvCount() {
|
||||
return ((bitField0_ & 0x00000020) == 0x00000020);
|
||||
|
@ -771,23 +771,23 @@ public final class WALProtos {
|
|||
int hash = 41;
|
||||
hash = (19 * hash) + getDescriptorForType().hashCode();
|
||||
if (hasEncodedRegionName()) {
|
||||
hash = (37 * hash) + ENCODEDREGIONNAME_FIELD_NUMBER;
|
||||
hash = (37 * hash) + ENCODED_REGION_NAME_FIELD_NUMBER;
|
||||
hash = (53 * hash) + getEncodedRegionName().hashCode();
|
||||
}
|
||||
if (hasTableName()) {
|
||||
hash = (37 * hash) + TABLENAME_FIELD_NUMBER;
|
||||
hash = (37 * hash) + TABLE_NAME_FIELD_NUMBER;
|
||||
hash = (53 * hash) + getTableName().hashCode();
|
||||
}
|
||||
if (hasLogSequenceNumber()) {
|
||||
hash = (37 * hash) + LOGSEQUENCENUMBER_FIELD_NUMBER;
|
||||
hash = (37 * hash) + LOG_SEQUENCE_NUMBER_FIELD_NUMBER;
|
||||
hash = (53 * hash) + hashLong(getLogSequenceNumber());
|
||||
}
|
||||
if (hasWriteTime()) {
|
||||
hash = (37 * hash) + WRITETIME_FIELD_NUMBER;
|
||||
hash = (37 * hash) + WRITE_TIME_FIELD_NUMBER;
|
||||
hash = (53 * hash) + hashLong(getWriteTime());
|
||||
}
|
||||
if (hasClusterId()) {
|
||||
hash = (37 * hash) + CLUSTERID_FIELD_NUMBER;
|
||||
hash = (37 * hash) + CLUSTER_ID_FIELD_NUMBER;
|
||||
hash = (53 * hash) + getClusterId().hashCode();
|
||||
}
|
||||
if (getScopesCount() > 0) {
|
||||
|
@ -795,7 +795,7 @@ public final class WALProtos {
|
|||
hash = (53 * hash) + getScopesList().hashCode();
|
||||
}
|
||||
if (hasFollowingKvCount()) {
|
||||
hash = (37 * hash) + FOLLOWINGKVCOUNT_FIELD_NUMBER;
|
||||
hash = (37 * hash) + FOLLOWING_KV_COUNT_FIELD_NUMBER;
|
||||
hash = (53 * hash) + getFollowingKvCount();
|
||||
}
|
||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||
|
@ -1178,7 +1178,7 @@ public final class WALProtos {
|
|||
|
||||
private int bitField0_;
|
||||
|
||||
// required bytes encodedRegionName = 1;
|
||||
// required bytes encoded_region_name = 1;
|
||||
private com.google.protobuf.ByteString encodedRegionName_ = com.google.protobuf.ByteString.EMPTY;
|
||||
public boolean hasEncodedRegionName() {
|
||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||
|
@ -1202,7 +1202,7 @@ public final class WALProtos {
|
|||
return this;
|
||||
}
|
||||
|
||||
// required bytes tableName = 2;
|
||||
// required bytes table_name = 2;
|
||||
private com.google.protobuf.ByteString tableName_ = com.google.protobuf.ByteString.EMPTY;
|
||||
public boolean hasTableName() {
|
||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||
|
@ -1226,7 +1226,7 @@ public final class WALProtos {
|
|||
return this;
|
||||
}
|
||||
|
||||
// required uint64 logSequenceNumber = 3;
|
||||
// required uint64 log_sequence_number = 3;
|
||||
private long logSequenceNumber_ ;
|
||||
public boolean hasLogSequenceNumber() {
|
||||
return ((bitField0_ & 0x00000004) == 0x00000004);
|
||||
|
@ -1247,7 +1247,7 @@ public final class WALProtos {
|
|||
return this;
|
||||
}
|
||||
|
||||
// required uint64 writeTime = 4;
|
||||
// required uint64 write_time = 4;
|
||||
private long writeTime_ ;
|
||||
public boolean hasWriteTime() {
|
||||
return ((bitField0_ & 0x00000008) == 0x00000008);
|
||||
|
@ -1268,7 +1268,7 @@ public final class WALProtos {
|
|||
return this;
|
||||
}
|
||||
|
||||
// optional .UUID clusterId = 5;
|
||||
// optional .UUID cluster_id = 5;
|
||||
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.UUID clusterId_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.UUID.getDefaultInstance();
|
||||
private com.google.protobuf.SingleFieldBuilder<
|
||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.UUID, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.UUID.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.UUIDOrBuilder> clusterIdBuilder_;
|
||||
|
@ -1544,7 +1544,7 @@ public final class WALProtos {
|
|||
return scopesBuilder_;
|
||||
}
|
||||
|
||||
// optional uint32 followingKvCount = 7;
|
||||
// optional uint32 following_kv_count = 7;
|
||||
private int followingKvCount_ ;
|
||||
public boolean hasFollowingKvCount() {
|
||||
return ((bitField0_ & 0x00000040) == 0x00000040);
|
||||
|
@ -1583,7 +1583,7 @@ public final class WALProtos {
|
|||
boolean hasFamily();
|
||||
com.google.protobuf.ByteString getFamily();
|
||||
|
||||
// required .ScopeType scopeType = 2;
|
||||
// required .ScopeType scope_type = 2;
|
||||
boolean hasScopeType();
|
||||
org.apache.hadoop.hbase.protobuf.generated.WALProtos.ScopeType getScopeType();
|
||||
}
|
||||
|
@ -1626,8 +1626,8 @@ public final class WALProtos {
|
|||
return family_;
|
||||
}
|
||||
|
||||
// required .ScopeType scopeType = 2;
|
||||
public static final int SCOPETYPE_FIELD_NUMBER = 2;
|
||||
// required .ScopeType scope_type = 2;
|
||||
public static final int SCOPE_TYPE_FIELD_NUMBER = 2;
|
||||
private org.apache.hadoop.hbase.protobuf.generated.WALProtos.ScopeType scopeType_;
|
||||
public boolean hasScopeType() {
|
||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||
|
@ -1730,7 +1730,7 @@ public final class WALProtos {
|
|||
hash = (53 * hash) + getFamily().hashCode();
|
||||
}
|
||||
if (hasScopeType()) {
|
||||
hash = (37 * hash) + SCOPETYPE_FIELD_NUMBER;
|
||||
hash = (37 * hash) + SCOPE_TYPE_FIELD_NUMBER;
|
||||
hash = (53 * hash) + hashEnum(getScopeType());
|
||||
}
|
||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||
|
@ -2006,7 +2006,7 @@ public final class WALProtos {
|
|||
return this;
|
||||
}
|
||||
|
||||
// required .ScopeType scopeType = 2;
|
||||
// required .ScopeType scope_type = 2;
|
||||
private org.apache.hadoop.hbase.protobuf.generated.WALProtos.ScopeType scopeType_ = org.apache.hadoop.hbase.protobuf.generated.WALProtos.ScopeType.REPLICATION_SCOPE_LOCAL;
|
||||
public boolean hasScopeType() {
|
||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||
|
@ -2044,29 +2044,29 @@ public final class WALProtos {
|
|||
public interface CompactionDescriptorOrBuilder
|
||||
extends com.google.protobuf.MessageOrBuilder {
|
||||
|
||||
// required bytes tableName = 1;
|
||||
// required bytes table_name = 1;
|
||||
boolean hasTableName();
|
||||
com.google.protobuf.ByteString getTableName();
|
||||
|
||||
// required bytes encodedRegionName = 2;
|
||||
// required bytes encoded_region_name = 2;
|
||||
boolean hasEncodedRegionName();
|
||||
com.google.protobuf.ByteString getEncodedRegionName();
|
||||
|
||||
// required bytes familyName = 3;
|
||||
// required bytes family_name = 3;
|
||||
boolean hasFamilyName();
|
||||
com.google.protobuf.ByteString getFamilyName();
|
||||
|
||||
// repeated string compactionInput = 4;
|
||||
// repeated string compaction_input = 4;
|
||||
java.util.List<String> getCompactionInputList();
|
||||
int getCompactionInputCount();
|
||||
String getCompactionInput(int index);
|
||||
|
||||
// repeated string compactionOutput = 5;
|
||||
// repeated string compaction_output = 5;
|
||||
java.util.List<String> getCompactionOutputList();
|
||||
int getCompactionOutputCount();
|
||||
String getCompactionOutput(int index);
|
||||
|
||||
// required string storeHomeDir = 6;
|
||||
// required string store_home_dir = 6;
|
||||
boolean hasStoreHomeDir();
|
||||
String getStoreHomeDir();
|
||||
}
|
||||
|
@ -2099,8 +2099,8 @@ public final class WALProtos {
|
|||
}
|
||||
|
||||
private int bitField0_;
|
||||
// required bytes tableName = 1;
|
||||
public static final int TABLENAME_FIELD_NUMBER = 1;
|
||||
// required bytes table_name = 1;
|
||||
public static final int TABLE_NAME_FIELD_NUMBER = 1;
|
||||
private com.google.protobuf.ByteString tableName_;
|
||||
public boolean hasTableName() {
|
||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||
|
@ -2109,8 +2109,8 @@ public final class WALProtos {
|
|||
return tableName_;
|
||||
}
|
||||
|
||||
// required bytes encodedRegionName = 2;
|
||||
public static final int ENCODEDREGIONNAME_FIELD_NUMBER = 2;
|
||||
// required bytes encoded_region_name = 2;
|
||||
public static final int ENCODED_REGION_NAME_FIELD_NUMBER = 2;
|
||||
private com.google.protobuf.ByteString encodedRegionName_;
|
||||
public boolean hasEncodedRegionName() {
|
||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||
|
@ -2119,8 +2119,8 @@ public final class WALProtos {
|
|||
return encodedRegionName_;
|
||||
}
|
||||
|
||||
// required bytes familyName = 3;
|
||||
public static final int FAMILYNAME_FIELD_NUMBER = 3;
|
||||
// required bytes family_name = 3;
|
||||
public static final int FAMILY_NAME_FIELD_NUMBER = 3;
|
||||
private com.google.protobuf.ByteString familyName_;
|
||||
public boolean hasFamilyName() {
|
||||
return ((bitField0_ & 0x00000004) == 0x00000004);
|
||||
|
@ -2129,8 +2129,8 @@ public final class WALProtos {
|
|||
return familyName_;
|
||||
}
|
||||
|
||||
// repeated string compactionInput = 4;
|
||||
public static final int COMPACTIONINPUT_FIELD_NUMBER = 4;
|
||||
// repeated string compaction_input = 4;
|
||||
public static final int COMPACTION_INPUT_FIELD_NUMBER = 4;
|
||||
private com.google.protobuf.LazyStringList compactionInput_;
|
||||
public java.util.List<String>
|
||||
getCompactionInputList() {
|
||||
|
@ -2143,8 +2143,8 @@ public final class WALProtos {
|
|||
return compactionInput_.get(index);
|
||||
}
|
||||
|
||||
// repeated string compactionOutput = 5;
|
||||
public static final int COMPACTIONOUTPUT_FIELD_NUMBER = 5;
|
||||
// repeated string compaction_output = 5;
|
||||
public static final int COMPACTION_OUTPUT_FIELD_NUMBER = 5;
|
||||
private com.google.protobuf.LazyStringList compactionOutput_;
|
||||
public java.util.List<String>
|
||||
getCompactionOutputList() {
|
||||
|
@ -2157,8 +2157,8 @@ public final class WALProtos {
|
|||
return compactionOutput_.get(index);
|
||||
}
|
||||
|
||||
// required string storeHomeDir = 6;
|
||||
public static final int STOREHOMEDIR_FIELD_NUMBER = 6;
|
||||
// required string store_home_dir = 6;
|
||||
public static final int STORE_HOME_DIR_FIELD_NUMBER = 6;
|
||||
private java.lang.Object storeHomeDir_;
|
||||
public boolean hasStoreHomeDir() {
|
||||
return ((bitField0_ & 0x00000008) == 0x00000008);
|
||||
|
@ -2343,27 +2343,27 @@ public final class WALProtos {
|
|||
int hash = 41;
|
||||
hash = (19 * hash) + getDescriptorForType().hashCode();
|
||||
if (hasTableName()) {
|
||||
hash = (37 * hash) + TABLENAME_FIELD_NUMBER;
|
||||
hash = (37 * hash) + TABLE_NAME_FIELD_NUMBER;
|
||||
hash = (53 * hash) + getTableName().hashCode();
|
||||
}
|
||||
if (hasEncodedRegionName()) {
|
||||
hash = (37 * hash) + ENCODEDREGIONNAME_FIELD_NUMBER;
|
||||
hash = (37 * hash) + ENCODED_REGION_NAME_FIELD_NUMBER;
|
||||
hash = (53 * hash) + getEncodedRegionName().hashCode();
|
||||
}
|
||||
if (hasFamilyName()) {
|
||||
hash = (37 * hash) + FAMILYNAME_FIELD_NUMBER;
|
||||
hash = (37 * hash) + FAMILY_NAME_FIELD_NUMBER;
|
||||
hash = (53 * hash) + getFamilyName().hashCode();
|
||||
}
|
||||
if (getCompactionInputCount() > 0) {
|
||||
hash = (37 * hash) + COMPACTIONINPUT_FIELD_NUMBER;
|
||||
hash = (37 * hash) + COMPACTION_INPUT_FIELD_NUMBER;
|
||||
hash = (53 * hash) + getCompactionInputList().hashCode();
|
||||
}
|
||||
if (getCompactionOutputCount() > 0) {
|
||||
hash = (37 * hash) + COMPACTIONOUTPUT_FIELD_NUMBER;
|
||||
hash = (37 * hash) + COMPACTION_OUTPUT_FIELD_NUMBER;
|
||||
hash = (53 * hash) + getCompactionOutputList().hashCode();
|
||||
}
|
||||
if (hasStoreHomeDir()) {
|
||||
hash = (37 * hash) + STOREHOMEDIR_FIELD_NUMBER;
|
||||
hash = (37 * hash) + STORE_HOME_DIR_FIELD_NUMBER;
|
||||
hash = (53 * hash) + getStoreHomeDir().hashCode();
|
||||
}
|
||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||
|
@ -2691,7 +2691,7 @@ public final class WALProtos {
|
|||
|
||||
private int bitField0_;
|
||||
|
||||
// required bytes tableName = 1;
|
||||
// required bytes table_name = 1;
|
||||
private com.google.protobuf.ByteString tableName_ = com.google.protobuf.ByteString.EMPTY;
|
||||
public boolean hasTableName() {
|
||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||
|
@ -2715,7 +2715,7 @@ public final class WALProtos {
|
|||
return this;
|
||||
}
|
||||
|
||||
// required bytes encodedRegionName = 2;
|
||||
// required bytes encoded_region_name = 2;
|
||||
private com.google.protobuf.ByteString encodedRegionName_ = com.google.protobuf.ByteString.EMPTY;
|
||||
public boolean hasEncodedRegionName() {
|
||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||
|
@ -2739,7 +2739,7 @@ public final class WALProtos {
|
|||
return this;
|
||||
}
|
||||
|
||||
// required bytes familyName = 3;
|
||||
// required bytes family_name = 3;
|
||||
private com.google.protobuf.ByteString familyName_ = com.google.protobuf.ByteString.EMPTY;
|
||||
public boolean hasFamilyName() {
|
||||
return ((bitField0_ & 0x00000004) == 0x00000004);
|
||||
|
@ -2763,7 +2763,7 @@ public final class WALProtos {
|
|||
return this;
|
||||
}
|
||||
|
||||
// repeated string compactionInput = 4;
|
||||
// repeated string compaction_input = 4;
|
||||
private com.google.protobuf.LazyStringList compactionInput_ = com.google.protobuf.LazyStringArrayList.EMPTY;
|
||||
private void ensureCompactionInputIsMutable() {
|
||||
if (!((bitField0_ & 0x00000008) == 0x00000008)) {
|
||||
|
@ -2819,7 +2819,7 @@ public final class WALProtos {
|
|||
onChanged();
|
||||
}
|
||||
|
||||
// repeated string compactionOutput = 5;
|
||||
// repeated string compaction_output = 5;
|
||||
private com.google.protobuf.LazyStringList compactionOutput_ = com.google.protobuf.LazyStringArrayList.EMPTY;
|
||||
private void ensureCompactionOutputIsMutable() {
|
||||
if (!((bitField0_ & 0x00000010) == 0x00000010)) {
|
||||
|
@ -2875,7 +2875,7 @@ public final class WALProtos {
|
|||
onChanged();
|
||||
}
|
||||
|
||||
// required string storeHomeDir = 6;
|
||||
// required string store_home_dir = 6;
|
||||
private java.lang.Object storeHomeDir_ = "";
|
||||
public boolean hasStoreHomeDir() {
|
||||
return ((bitField0_ & 0x00000020) == 0x00000020);
|
||||
|
@ -3256,22 +3256,23 @@ public final class WALProtos {
|
|||
descriptor;
|
||||
static {
|
||||
java.lang.String[] descriptorData = {
|
||||
"\n\tWAL.proto\032\013hbase.proto\"#\n\tWALHeader\022\026\n" +
|
||||
"\016hasCompression\030\001 \001(\010\"\266\001\n\006WALKey\022\031\n\021enco" +
|
||||
"dedRegionName\030\001 \002(\014\022\021\n\ttableName\030\002 \002(\014\022\031" +
|
||||
"\n\021logSequenceNumber\030\003 \002(\004\022\021\n\twriteTime\030\004" +
|
||||
" \002(\004\022\030\n\tclusterId\030\005 \001(\0132\005.UUID\022\034\n\006scopes" +
|
||||
"\030\006 \003(\0132\014.FamilyScope\022\030\n\020followingKvCount" +
|
||||
"\030\007 \001(\r\"<\n\013FamilyScope\022\016\n\006family\030\001 \002(\014\022\035\n" +
|
||||
"\tscopeType\030\002 \002(\0162\n.ScopeType\"\241\001\n\024Compact" +
|
||||
"ionDescriptor\022\021\n\ttableName\030\001 \002(\014\022\031\n\021enco" +
|
||||
"dedRegionName\030\002 \002(\014\022\022\n\nfamilyName\030\003 \002(\014\022",
|
||||
"\027\n\017compactionInput\030\004 \003(\t\022\030\n\020compactionOu" +
|
||||
"tput\030\005 \003(\t\022\024\n\014storeHomeDir\030\006 \002(\t\"\014\n\nWALT" +
|
||||
"railer*F\n\tScopeType\022\033\n\027REPLICATION_SCOPE" +
|
||||
"_LOCAL\020\000\022\034\n\030REPLICATION_SCOPE_GLOBAL\020\001B?" +
|
||||
"\n*org.apache.hadoop.hbase.protobuf.gener" +
|
||||
"atedB\tWALProtosH\001\210\001\000\240\001\001"
|
||||
"\n\tWAL.proto\032\013hbase.proto\"$\n\tWALHeader\022\027\n" +
|
||||
"\017has_compression\030\001 \001(\010\"\277\001\n\006WALKey\022\033\n\023enc" +
|
||||
"oded_region_name\030\001 \002(\014\022\022\n\ntable_name\030\002 \002" +
|
||||
"(\014\022\033\n\023log_sequence_number\030\003 \002(\004\022\022\n\nwrite" +
|
||||
"_time\030\004 \002(\004\022\031\n\ncluster_id\030\005 \001(\0132\005.UUID\022\034" +
|
||||
"\n\006scopes\030\006 \003(\0132\014.FamilyScope\022\032\n\022followin" +
|
||||
"g_kv_count\030\007 \001(\r\"=\n\013FamilyScope\022\016\n\006famil" +
|
||||
"y\030\001 \002(\014\022\036\n\nscope_type\030\002 \002(\0162\n.ScopeType\"" +
|
||||
"\251\001\n\024CompactionDescriptor\022\022\n\ntable_name\030\001" +
|
||||
" \002(\014\022\033\n\023encoded_region_name\030\002 \002(\014\022\023\n\013fam",
|
||||
"ily_name\030\003 \002(\014\022\030\n\020compaction_input\030\004 \003(\t" +
|
||||
"\022\031\n\021compaction_output\030\005 \003(\t\022\026\n\016store_hom" +
|
||||
"e_dir\030\006 \002(\t\"\014\n\nWALTrailer*F\n\tScopeType\022\033" +
|
||||
"\n\027REPLICATION_SCOPE_LOCAL\020\000\022\034\n\030REPLICATI" +
|
||||
"ON_SCOPE_GLOBAL\020\001B?\n*org.apache.hadoop.h" +
|
||||
"base.protobuf.generatedB\tWALProtosH\001\210\001\000\240" +
|
||||
"\001\001"
|
||||
};
|
||||
com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
|
||||
new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() {
|
||||
|
|
|
@ -16,7 +16,7 @@ public final class ZooKeeperProtos {
|
|||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName getServer();
|
||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerNameOrBuilder getServerOrBuilder();
|
||||
|
||||
// optional uint32 rpcVersion = 2;
|
||||
// optional uint32 rpc_version = 2;
|
||||
boolean hasRpcVersion();
|
||||
int getRpcVersion();
|
||||
}
|
||||
|
@ -62,8 +62,8 @@ public final class ZooKeeperProtos {
|
|||
return server_;
|
||||
}
|
||||
|
||||
// optional uint32 rpcVersion = 2;
|
||||
public static final int RPCVERSION_FIELD_NUMBER = 2;
|
||||
// optional uint32 rpc_version = 2;
|
||||
public static final int RPC_VERSION_FIELD_NUMBER = 2;
|
||||
private int rpcVersion_;
|
||||
public boolean hasRpcVersion() {
|
||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||
|
@ -166,7 +166,7 @@ public final class ZooKeeperProtos {
|
|||
hash = (53 * hash) + getServer().hashCode();
|
||||
}
|
||||
if (hasRpcVersion()) {
|
||||
hash = (37 * hash) + RPCVERSION_FIELD_NUMBER;
|
||||
hash = (37 * hash) + RPC_VERSION_FIELD_NUMBER;
|
||||
hash = (53 * hash) + getRpcVersion();
|
||||
}
|
||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||
|
@ -515,7 +515,7 @@ public final class ZooKeeperProtos {
|
|||
return serverBuilder_;
|
||||
}
|
||||
|
||||
// optional uint32 rpcVersion = 2;
|
||||
// optional uint32 rpc_version = 2;
|
||||
private int rpcVersion_ ;
|
||||
public boolean hasRpcVersion() {
|
||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||
|
@ -555,7 +555,7 @@ public final class ZooKeeperProtos {
|
|||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName getMaster();
|
||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerNameOrBuilder getMasterOrBuilder();
|
||||
|
||||
// optional uint32 rpcVersion = 2;
|
||||
// optional uint32 rpc_version = 2;
|
||||
boolean hasRpcVersion();
|
||||
int getRpcVersion();
|
||||
}
|
||||
|
@ -601,8 +601,8 @@ public final class ZooKeeperProtos {
|
|||
return master_;
|
||||
}
|
||||
|
||||
// optional uint32 rpcVersion = 2;
|
||||
public static final int RPCVERSION_FIELD_NUMBER = 2;
|
||||
// optional uint32 rpc_version = 2;
|
||||
public static final int RPC_VERSION_FIELD_NUMBER = 2;
|
||||
private int rpcVersion_;
|
||||
public boolean hasRpcVersion() {
|
||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||
|
@ -705,7 +705,7 @@ public final class ZooKeeperProtos {
|
|||
hash = (53 * hash) + getMaster().hashCode();
|
||||
}
|
||||
if (hasRpcVersion()) {
|
||||
hash = (37 * hash) + RPCVERSION_FIELD_NUMBER;
|
||||
hash = (37 * hash) + RPC_VERSION_FIELD_NUMBER;
|
||||
hash = (53 * hash) + getRpcVersion();
|
||||
}
|
||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||
|
@ -1054,7 +1054,7 @@ public final class ZooKeeperProtos {
|
|||
return masterBuilder_;
|
||||
}
|
||||
|
||||
// optional uint32 rpcVersion = 2;
|
||||
// optional uint32 rpc_version = 2;
|
||||
private int rpcVersion_ ;
|
||||
public boolean hasRpcVersion() {
|
||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||
|
@ -1089,7 +1089,7 @@ public final class ZooKeeperProtos {
|
|||
public interface ClusterUpOrBuilder
|
||||
extends com.google.protobuf.MessageOrBuilder {
|
||||
|
||||
// required string startDate = 1;
|
||||
// required string start_date = 1;
|
||||
boolean hasStartDate();
|
||||
String getStartDate();
|
||||
}
|
||||
|
@ -1122,8 +1122,8 @@ public final class ZooKeeperProtos {
|
|||
}
|
||||
|
||||
private int bitField0_;
|
||||
// required string startDate = 1;
|
||||
public static final int STARTDATE_FIELD_NUMBER = 1;
|
||||
// required string start_date = 1;
|
||||
public static final int START_DATE_FIELD_NUMBER = 1;
|
||||
private java.lang.Object startDate_;
|
||||
public boolean hasStartDate() {
|
||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||
|
@ -1227,7 +1227,7 @@ public final class ZooKeeperProtos {
|
|||
int hash = 41;
|
||||
hash = (19 * hash) + getDescriptorForType().hashCode();
|
||||
if (hasStartDate()) {
|
||||
hash = (37 * hash) + STARTDATE_FIELD_NUMBER;
|
||||
hash = (37 * hash) + START_DATE_FIELD_NUMBER;
|
||||
hash = (53 * hash) + getStartDate().hashCode();
|
||||
}
|
||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||
|
@ -1455,7 +1455,7 @@ public final class ZooKeeperProtos {
|
|||
|
||||
private int bitField0_;
|
||||
|
||||
// required string startDate = 1;
|
||||
// required string start_date = 1;
|
||||
private java.lang.Object startDate_ = "";
|
||||
public boolean hasStartDate() {
|
||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||
|
@ -1505,19 +1505,19 @@ public final class ZooKeeperProtos {
|
|||
public interface RegionTransitionOrBuilder
|
||||
extends com.google.protobuf.MessageOrBuilder {
|
||||
|
||||
// required uint32 eventTypeCode = 1;
|
||||
// required uint32 event_type_code = 1;
|
||||
boolean hasEventTypeCode();
|
||||
int getEventTypeCode();
|
||||
|
||||
// required bytes regionName = 2;
|
||||
// required bytes region_name = 2;
|
||||
boolean hasRegionName();
|
||||
com.google.protobuf.ByteString getRegionName();
|
||||
|
||||
// required uint64 createTime = 3;
|
||||
// required uint64 create_time = 3;
|
||||
boolean hasCreateTime();
|
||||
long getCreateTime();
|
||||
|
||||
// required .ServerName serverName = 4;
|
||||
// required .ServerName server_name = 4;
|
||||
boolean hasServerName();
|
||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName getServerName();
|
||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerNameOrBuilder getServerNameOrBuilder();
|
||||
|
@ -1555,8 +1555,8 @@ public final class ZooKeeperProtos {
|
|||
}
|
||||
|
||||
private int bitField0_;
|
||||
// required uint32 eventTypeCode = 1;
|
||||
public static final int EVENTTYPECODE_FIELD_NUMBER = 1;
|
||||
// required uint32 event_type_code = 1;
|
||||
public static final int EVENT_TYPE_CODE_FIELD_NUMBER = 1;
|
||||
private int eventTypeCode_;
|
||||
public boolean hasEventTypeCode() {
|
||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||
|
@ -1565,8 +1565,8 @@ public final class ZooKeeperProtos {
|
|||
return eventTypeCode_;
|
||||
}
|
||||
|
||||
// required bytes regionName = 2;
|
||||
public static final int REGIONNAME_FIELD_NUMBER = 2;
|
||||
// required bytes region_name = 2;
|
||||
public static final int REGION_NAME_FIELD_NUMBER = 2;
|
||||
private com.google.protobuf.ByteString regionName_;
|
||||
public boolean hasRegionName() {
|
||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||
|
@ -1575,8 +1575,8 @@ public final class ZooKeeperProtos {
|
|||
return regionName_;
|
||||
}
|
||||
|
||||
// required uint64 createTime = 3;
|
||||
public static final int CREATETIME_FIELD_NUMBER = 3;
|
||||
// required uint64 create_time = 3;
|
||||
public static final int CREATE_TIME_FIELD_NUMBER = 3;
|
||||
private long createTime_;
|
||||
public boolean hasCreateTime() {
|
||||
return ((bitField0_ & 0x00000004) == 0x00000004);
|
||||
|
@ -1585,8 +1585,8 @@ public final class ZooKeeperProtos {
|
|||
return createTime_;
|
||||
}
|
||||
|
||||
// required .ServerName serverName = 4;
|
||||
public static final int SERVERNAME_FIELD_NUMBER = 4;
|
||||
// required .ServerName server_name = 4;
|
||||
public static final int SERVER_NAME_FIELD_NUMBER = 4;
|
||||
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName serverName_;
|
||||
public boolean hasServerName() {
|
||||
return ((bitField0_ & 0x00000008) == 0x00000008);
|
||||
|
@ -1749,19 +1749,19 @@ public final class ZooKeeperProtos {
|
|||
int hash = 41;
|
||||
hash = (19 * hash) + getDescriptorForType().hashCode();
|
||||
if (hasEventTypeCode()) {
|
||||
hash = (37 * hash) + EVENTTYPECODE_FIELD_NUMBER;
|
||||
hash = (37 * hash) + EVENT_TYPE_CODE_FIELD_NUMBER;
|
||||
hash = (53 * hash) + getEventTypeCode();
|
||||
}
|
||||
if (hasRegionName()) {
|
||||
hash = (37 * hash) + REGIONNAME_FIELD_NUMBER;
|
||||
hash = (37 * hash) + REGION_NAME_FIELD_NUMBER;
|
||||
hash = (53 * hash) + getRegionName().hashCode();
|
||||
}
|
||||
if (hasCreateTime()) {
|
||||
hash = (37 * hash) + CREATETIME_FIELD_NUMBER;
|
||||
hash = (37 * hash) + CREATE_TIME_FIELD_NUMBER;
|
||||
hash = (53 * hash) + hashLong(getCreateTime());
|
||||
}
|
||||
if (hasServerName()) {
|
||||
hash = (37 * hash) + SERVERNAME_FIELD_NUMBER;
|
||||
hash = (37 * hash) + SERVER_NAME_FIELD_NUMBER;
|
||||
hash = (53 * hash) + getServerName().hashCode();
|
||||
}
|
||||
if (hasPayload()) {
|
||||
|
@ -2078,7 +2078,7 @@ public final class ZooKeeperProtos {
|
|||
|
||||
private int bitField0_;
|
||||
|
||||
// required uint32 eventTypeCode = 1;
|
||||
// required uint32 event_type_code = 1;
|
||||
private int eventTypeCode_ ;
|
||||
public boolean hasEventTypeCode() {
|
||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||
|
@ -2099,7 +2099,7 @@ public final class ZooKeeperProtos {
|
|||
return this;
|
||||
}
|
||||
|
||||
// required bytes regionName = 2;
|
||||
// required bytes region_name = 2;
|
||||
private com.google.protobuf.ByteString regionName_ = com.google.protobuf.ByteString.EMPTY;
|
||||
public boolean hasRegionName() {
|
||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||
|
@ -2123,7 +2123,7 @@ public final class ZooKeeperProtos {
|
|||
return this;
|
||||
}
|
||||
|
||||
// required uint64 createTime = 3;
|
||||
// required uint64 create_time = 3;
|
||||
private long createTime_ ;
|
||||
public boolean hasCreateTime() {
|
||||
return ((bitField0_ & 0x00000004) == 0x00000004);
|
||||
|
@ -2144,7 +2144,7 @@ public final class ZooKeeperProtos {
|
|||
return this;
|
||||
}
|
||||
|
||||
// required .ServerName serverName = 4;
|
||||
// required .ServerName server_name = 4;
|
||||
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName serverName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.getDefaultInstance();
|
||||
private com.google.protobuf.SingleFieldBuilder<
|
||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerNameOrBuilder> serverNameBuilder_;
|
||||
|
@ -2276,7 +2276,7 @@ public final class ZooKeeperProtos {
|
|||
boolean hasState();
|
||||
org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.SplitLogTask.State getState();
|
||||
|
||||
// required .ServerName serverName = 2;
|
||||
// required .ServerName server_name = 2;
|
||||
boolean hasServerName();
|
||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName getServerName();
|
||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerNameOrBuilder getServerNameOrBuilder();
|
||||
|
@ -2398,8 +2398,8 @@ public final class ZooKeeperProtos {
|
|||
return state_;
|
||||
}
|
||||
|
||||
// required .ServerName serverName = 2;
|
||||
public static final int SERVERNAME_FIELD_NUMBER = 2;
|
||||
// required .ServerName server_name = 2;
|
||||
public static final int SERVER_NAME_FIELD_NUMBER = 2;
|
||||
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName serverName_;
|
||||
public boolean hasServerName() {
|
||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||
|
@ -2509,7 +2509,7 @@ public final class ZooKeeperProtos {
|
|||
hash = (53 * hash) + hashEnum(getState());
|
||||
}
|
||||
if (hasServerName()) {
|
||||
hash = (37 * hash) + SERVERNAME_FIELD_NUMBER;
|
||||
hash = (37 * hash) + SERVER_NAME_FIELD_NUMBER;
|
||||
hash = (53 * hash) + getServerName().hashCode();
|
||||
}
|
||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||
|
@ -2802,7 +2802,7 @@ public final class ZooKeeperProtos {
|
|||
return this;
|
||||
}
|
||||
|
||||
// required .ServerName serverName = 2;
|
||||
// required .ServerName server_name = 2;
|
||||
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName serverName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.getDefaultInstance();
|
||||
private com.google.protobuf.SingleFieldBuilder<
|
||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerNameOrBuilder> serverNameBuilder_;
|
||||
|
@ -4621,7 +4621,7 @@ public final class ZooKeeperProtos {
|
|||
public interface ReplicationLockOrBuilder
|
||||
extends com.google.protobuf.MessageOrBuilder {
|
||||
|
||||
// required string lockOwner = 1;
|
||||
// required string lock_owner = 1;
|
||||
boolean hasLockOwner();
|
||||
String getLockOwner();
|
||||
}
|
||||
|
@ -4654,8 +4654,8 @@ public final class ZooKeeperProtos {
|
|||
}
|
||||
|
||||
private int bitField0_;
|
||||
// required string lockOwner = 1;
|
||||
public static final int LOCKOWNER_FIELD_NUMBER = 1;
|
||||
// required string lock_owner = 1;
|
||||
public static final int LOCK_OWNER_FIELD_NUMBER = 1;
|
||||
private java.lang.Object lockOwner_;
|
||||
public boolean hasLockOwner() {
|
||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||
|
@ -4759,7 +4759,7 @@ public final class ZooKeeperProtos {
|
|||
int hash = 41;
|
||||
hash = (19 * hash) + getDescriptorForType().hashCode();
|
||||
if (hasLockOwner()) {
|
||||
hash = (37 * hash) + LOCKOWNER_FIELD_NUMBER;
|
||||
hash = (37 * hash) + LOCK_OWNER_FIELD_NUMBER;
|
||||
hash = (53 * hash) + getLockOwner().hashCode();
|
||||
}
|
||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||
|
@ -4987,7 +4987,7 @@ public final class ZooKeeperProtos {
|
|||
|
||||
private int bitField0_;
|
||||
|
||||
// required string lockOwner = 1;
|
||||
// required string lock_owner = 1;
|
||||
private java.lang.Object lockOwner_ = "";
|
||||
public boolean hasLockOwner() {
|
||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||
|
@ -5037,20 +5037,20 @@ public final class ZooKeeperProtos {
|
|||
public interface TableLockOrBuilder
|
||||
extends com.google.protobuf.MessageOrBuilder {
|
||||
|
||||
// optional bytes tableName = 1;
|
||||
// optional bytes table_name = 1;
|
||||
boolean hasTableName();
|
||||
com.google.protobuf.ByteString getTableName();
|
||||
|
||||
// optional .ServerName lockOwner = 2;
|
||||
// optional .ServerName lock_owner = 2;
|
||||
boolean hasLockOwner();
|
||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName getLockOwner();
|
||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerNameOrBuilder getLockOwnerOrBuilder();
|
||||
|
||||
// optional int64 threadId = 3;
|
||||
// optional int64 thread_id = 3;
|
||||
boolean hasThreadId();
|
||||
long getThreadId();
|
||||
|
||||
// optional bool isShared = 4;
|
||||
// optional bool is_shared = 4;
|
||||
boolean hasIsShared();
|
||||
boolean getIsShared();
|
||||
|
||||
|
@ -5058,7 +5058,7 @@ public final class ZooKeeperProtos {
|
|||
boolean hasPurpose();
|
||||
String getPurpose();
|
||||
|
||||
// optional int64 createTime = 6;
|
||||
// optional int64 create_time = 6;
|
||||
boolean hasCreateTime();
|
||||
long getCreateTime();
|
||||
}
|
||||
|
@ -5091,8 +5091,8 @@ public final class ZooKeeperProtos {
|
|||
}
|
||||
|
||||
private int bitField0_;
|
||||
// optional bytes tableName = 1;
|
||||
public static final int TABLENAME_FIELD_NUMBER = 1;
|
||||
// optional bytes table_name = 1;
|
||||
public static final int TABLE_NAME_FIELD_NUMBER = 1;
|
||||
private com.google.protobuf.ByteString tableName_;
|
||||
public boolean hasTableName() {
|
||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||
|
@ -5101,8 +5101,8 @@ public final class ZooKeeperProtos {
|
|||
return tableName_;
|
||||
}
|
||||
|
||||
// optional .ServerName lockOwner = 2;
|
||||
public static final int LOCKOWNER_FIELD_NUMBER = 2;
|
||||
// optional .ServerName lock_owner = 2;
|
||||
public static final int LOCK_OWNER_FIELD_NUMBER = 2;
|
||||
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName lockOwner_;
|
||||
public boolean hasLockOwner() {
|
||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||
|
@ -5114,8 +5114,8 @@ public final class ZooKeeperProtos {
|
|||
return lockOwner_;
|
||||
}
|
||||
|
||||
// optional int64 threadId = 3;
|
||||
public static final int THREADID_FIELD_NUMBER = 3;
|
||||
// optional int64 thread_id = 3;
|
||||
public static final int THREAD_ID_FIELD_NUMBER = 3;
|
||||
private long threadId_;
|
||||
public boolean hasThreadId() {
|
||||
return ((bitField0_ & 0x00000004) == 0x00000004);
|
||||
|
@ -5124,8 +5124,8 @@ public final class ZooKeeperProtos {
|
|||
return threadId_;
|
||||
}
|
||||
|
||||
// optional bool isShared = 4;
|
||||
public static final int ISSHARED_FIELD_NUMBER = 4;
|
||||
// optional bool is_shared = 4;
|
||||
public static final int IS_SHARED_FIELD_NUMBER = 4;
|
||||
private boolean isShared_;
|
||||
public boolean hasIsShared() {
|
||||
return ((bitField0_ & 0x00000008) == 0x00000008);
|
||||
|
@ -5166,8 +5166,8 @@ public final class ZooKeeperProtos {
|
|||
}
|
||||
}
|
||||
|
||||
// optional int64 createTime = 6;
|
||||
public static final int CREATETIME_FIELD_NUMBER = 6;
|
||||
// optional int64 create_time = 6;
|
||||
public static final int CREATE_TIME_FIELD_NUMBER = 6;
|
||||
private long createTime_;
|
||||
public boolean hasCreateTime() {
|
||||
return ((bitField0_ & 0x00000020) == 0x00000020);
|
||||
|
@ -5316,19 +5316,19 @@ public final class ZooKeeperProtos {
|
|||
int hash = 41;
|
||||
hash = (19 * hash) + getDescriptorForType().hashCode();
|
||||
if (hasTableName()) {
|
||||
hash = (37 * hash) + TABLENAME_FIELD_NUMBER;
|
||||
hash = (37 * hash) + TABLE_NAME_FIELD_NUMBER;
|
||||
hash = (53 * hash) + getTableName().hashCode();
|
||||
}
|
||||
if (hasLockOwner()) {
|
||||
hash = (37 * hash) + LOCKOWNER_FIELD_NUMBER;
|
||||
hash = (37 * hash) + LOCK_OWNER_FIELD_NUMBER;
|
||||
hash = (53 * hash) + getLockOwner().hashCode();
|
||||
}
|
||||
if (hasThreadId()) {
|
||||
hash = (37 * hash) + THREADID_FIELD_NUMBER;
|
||||
hash = (37 * hash) + THREAD_ID_FIELD_NUMBER;
|
||||
hash = (53 * hash) + hashLong(getThreadId());
|
||||
}
|
||||
if (hasIsShared()) {
|
||||
hash = (37 * hash) + ISSHARED_FIELD_NUMBER;
|
||||
hash = (37 * hash) + IS_SHARED_FIELD_NUMBER;
|
||||
hash = (53 * hash) + hashBoolean(getIsShared());
|
||||
}
|
||||
if (hasPurpose()) {
|
||||
|
@ -5336,7 +5336,7 @@ public final class ZooKeeperProtos {
|
|||
hash = (53 * hash) + getPurpose().hashCode();
|
||||
}
|
||||
if (hasCreateTime()) {
|
||||
hash = (37 * hash) + CREATETIME_FIELD_NUMBER;
|
||||
hash = (37 * hash) + CREATE_TIME_FIELD_NUMBER;
|
||||
hash = (53 * hash) + hashLong(getCreateTime());
|
||||
}
|
||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||
|
@ -5649,7 +5649,7 @@ public final class ZooKeeperProtos {
|
|||
|
||||
private int bitField0_;
|
||||
|
||||
// optional bytes tableName = 1;
|
||||
// optional bytes table_name = 1;
|
||||
private com.google.protobuf.ByteString tableName_ = com.google.protobuf.ByteString.EMPTY;
|
||||
public boolean hasTableName() {
|
||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||
|
@ -5673,7 +5673,7 @@ public final class ZooKeeperProtos {
|
|||
return this;
|
||||
}
|
||||
|
||||
// optional .ServerName lockOwner = 2;
|
||||
// optional .ServerName lock_owner = 2;
|
||||
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName lockOwner_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.getDefaultInstance();
|
||||
private com.google.protobuf.SingleFieldBuilder<
|
||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerNameOrBuilder> lockOwnerBuilder_;
|
||||
|
@ -5763,7 +5763,7 @@ public final class ZooKeeperProtos {
|
|||
return lockOwnerBuilder_;
|
||||
}
|
||||
|
||||
// optional int64 threadId = 3;
|
||||
// optional int64 thread_id = 3;
|
||||
private long threadId_ ;
|
||||
public boolean hasThreadId() {
|
||||
return ((bitField0_ & 0x00000004) == 0x00000004);
|
||||
|
@ -5784,7 +5784,7 @@ public final class ZooKeeperProtos {
|
|||
return this;
|
||||
}
|
||||
|
||||
// optional bool isShared = 4;
|
||||
// optional bool is_shared = 4;
|
||||
private boolean isShared_ ;
|
||||
public boolean hasIsShared() {
|
||||
return ((bitField0_ & 0x00000008) == 0x00000008);
|
||||
|
@ -5841,7 +5841,7 @@ public final class ZooKeeperProtos {
|
|||
onChanged();
|
||||
}
|
||||
|
||||
// optional int64 createTime = 6;
|
||||
// optional int64 create_time = 6;
|
||||
private long createTime_ ;
|
||||
public boolean hasCreateTime() {
|
||||
return ((bitField0_ & 0x00000020) == 0x00000020);
|
||||
|
@ -5876,11 +5876,11 @@ public final class ZooKeeperProtos {
|
|||
public interface StoreSequenceIdOrBuilder
|
||||
extends com.google.protobuf.MessageOrBuilder {
|
||||
|
||||
// required bytes familyName = 1;
|
||||
// required bytes family_name = 1;
|
||||
boolean hasFamilyName();
|
||||
com.google.protobuf.ByteString getFamilyName();
|
||||
|
||||
// required uint64 sequenceId = 2;
|
||||
// required uint64 sequence_id = 2;
|
||||
boolean hasSequenceId();
|
||||
long getSequenceId();
|
||||
}
|
||||
|
@ -5913,8 +5913,8 @@ public final class ZooKeeperProtos {
|
|||
}
|
||||
|
||||
private int bitField0_;
|
||||
// required bytes familyName = 1;
|
||||
public static final int FAMILYNAME_FIELD_NUMBER = 1;
|
||||
// required bytes family_name = 1;
|
||||
public static final int FAMILY_NAME_FIELD_NUMBER = 1;
|
||||
private com.google.protobuf.ByteString familyName_;
|
||||
public boolean hasFamilyName() {
|
||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||
|
@ -5923,8 +5923,8 @@ public final class ZooKeeperProtos {
|
|||
return familyName_;
|
||||
}
|
||||
|
||||
// required uint64 sequenceId = 2;
|
||||
public static final int SEQUENCEID_FIELD_NUMBER = 2;
|
||||
// required uint64 sequence_id = 2;
|
||||
public static final int SEQUENCE_ID_FIELD_NUMBER = 2;
|
||||
private long sequenceId_;
|
||||
public boolean hasSequenceId() {
|
||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||
|
@ -6023,11 +6023,11 @@ public final class ZooKeeperProtos {
|
|||
int hash = 41;
|
||||
hash = (19 * hash) + getDescriptorForType().hashCode();
|
||||
if (hasFamilyName()) {
|
||||
hash = (37 * hash) + FAMILYNAME_FIELD_NUMBER;
|
||||
hash = (37 * hash) + FAMILY_NAME_FIELD_NUMBER;
|
||||
hash = (53 * hash) + getFamilyName().hashCode();
|
||||
}
|
||||
if (hasSequenceId()) {
|
||||
hash = (37 * hash) + SEQUENCEID_FIELD_NUMBER;
|
||||
hash = (37 * hash) + SEQUENCE_ID_FIELD_NUMBER;
|
||||
hash = (53 * hash) + hashLong(getSequenceId());
|
||||
}
|
||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||
|
@ -6273,7 +6273,7 @@ public final class ZooKeeperProtos {
|
|||
|
||||
private int bitField0_;
|
||||
|
||||
// required bytes familyName = 1;
|
||||
// required bytes family_name = 1;
|
||||
private com.google.protobuf.ByteString familyName_ = com.google.protobuf.ByteString.EMPTY;
|
||||
public boolean hasFamilyName() {
|
||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||
|
@ -6297,7 +6297,7 @@ public final class ZooKeeperProtos {
|
|||
return this;
|
||||
}
|
||||
|
||||
// required uint64 sequenceId = 2;
|
||||
// required uint64 sequence_id = 2;
|
||||
private long sequenceId_ ;
|
||||
public boolean hasSequenceId() {
|
||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||
|
@ -6332,11 +6332,11 @@ public final class ZooKeeperProtos {
|
|||
public interface RegionStoreSequenceIdsOrBuilder
|
||||
extends com.google.protobuf.MessageOrBuilder {
|
||||
|
||||
// required uint64 lastFlushedSequenceId = 1;
|
||||
// required uint64 last_flushed_sequence_id = 1;
|
||||
boolean hasLastFlushedSequenceId();
|
||||
long getLastFlushedSequenceId();
|
||||
|
||||
// repeated .StoreSequenceId storeSequenceId = 2;
|
||||
// repeated .StoreSequenceId store_sequence_id = 2;
|
||||
java.util.List<org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.StoreSequenceId>
|
||||
getStoreSequenceIdList();
|
||||
org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.StoreSequenceId getStoreSequenceId(int index);
|
||||
|
@ -6375,8 +6375,8 @@ public final class ZooKeeperProtos {
|
|||
}
|
||||
|
||||
private int bitField0_;
|
||||
// required uint64 lastFlushedSequenceId = 1;
|
||||
public static final int LASTFLUSHEDSEQUENCEID_FIELD_NUMBER = 1;
|
||||
// required uint64 last_flushed_sequence_id = 1;
|
||||
public static final int LAST_FLUSHED_SEQUENCE_ID_FIELD_NUMBER = 1;
|
||||
private long lastFlushedSequenceId_;
|
||||
public boolean hasLastFlushedSequenceId() {
|
||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||
|
@ -6385,8 +6385,8 @@ public final class ZooKeeperProtos {
|
|||
return lastFlushedSequenceId_;
|
||||
}
|
||||
|
||||
// repeated .StoreSequenceId storeSequenceId = 2;
|
||||
public static final int STORESEQUENCEID_FIELD_NUMBER = 2;
|
||||
// repeated .StoreSequenceId store_sequence_id = 2;
|
||||
public static final int STORE_SEQUENCE_ID_FIELD_NUMBER = 2;
|
||||
private java.util.List<org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.StoreSequenceId> storeSequenceId_;
|
||||
public java.util.List<org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.StoreSequenceId> getStoreSequenceIdList() {
|
||||
return storeSequenceId_;
|
||||
|
@ -6495,11 +6495,11 @@ public final class ZooKeeperProtos {
|
|||
int hash = 41;
|
||||
hash = (19 * hash) + getDescriptorForType().hashCode();
|
||||
if (hasLastFlushedSequenceId()) {
|
||||
hash = (37 * hash) + LASTFLUSHEDSEQUENCEID_FIELD_NUMBER;
|
||||
hash = (37 * hash) + LAST_FLUSHED_SEQUENCE_ID_FIELD_NUMBER;
|
||||
hash = (53 * hash) + hashLong(getLastFlushedSequenceId());
|
||||
}
|
||||
if (getStoreSequenceIdCount() > 0) {
|
||||
hash = (37 * hash) + STORESEQUENCEID_FIELD_NUMBER;
|
||||
hash = (37 * hash) + STORE_SEQUENCE_ID_FIELD_NUMBER;
|
||||
hash = (53 * hash) + getStoreSequenceIdList().hashCode();
|
||||
}
|
||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||
|
@ -6781,7 +6781,7 @@ public final class ZooKeeperProtos {
|
|||
|
||||
private int bitField0_;
|
||||
|
||||
// required uint64 lastFlushedSequenceId = 1;
|
||||
// required uint64 last_flushed_sequence_id = 1;
|
||||
private long lastFlushedSequenceId_ ;
|
||||
public boolean hasLastFlushedSequenceId() {
|
||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||
|
@ -6802,7 +6802,7 @@ public final class ZooKeeperProtos {
|
|||
return this;
|
||||
}
|
||||
|
||||
// repeated .StoreSequenceId storeSequenceId = 2;
|
||||
// repeated .StoreSequenceId store_sequence_id = 2;
|
||||
private java.util.List<org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.StoreSequenceId> storeSequenceId_ =
|
||||
java.util.Collections.emptyList();
|
||||
private void ensureStoreSequenceIdIsMutable() {
|
||||
|
@ -7073,36 +7073,36 @@ public final class ZooKeeperProtos {
|
|||
descriptor;
|
||||
static {
|
||||
java.lang.String[] descriptorData = {
|
||||
"\n\017ZooKeeper.proto\032\013hbase.proto\"C\n\020MetaRe" +
|
||||
"\n\017ZooKeeper.proto\032\013hbase.proto\"D\n\020MetaRe" +
|
||||
"gionServer\022\033\n\006server\030\001 \002(\0132\013.ServerName\022" +
|
||||
"\022\n\nrpcVersion\030\002 \001(\r\"9\n\006Master\022\033\n\006master\030" +
|
||||
"\001 \002(\0132\013.ServerName\022\022\n\nrpcVersion\030\002 \001(\r\"\036" +
|
||||
"\n\tClusterUp\022\021\n\tstartDate\030\001 \002(\t\"\203\001\n\020Regio" +
|
||||
"nTransition\022\025\n\reventTypeCode\030\001 \002(\r\022\022\n\nre" +
|
||||
"gionName\030\002 \002(\014\022\022\n\ncreateTime\030\003 \002(\004\022\037\n\nse" +
|
||||
"rverName\030\004 \002(\0132\013.ServerName\022\017\n\007payload\030\005" +
|
||||
" \001(\014\"\230\001\n\014SplitLogTask\022\"\n\005state\030\001 \002(\0162\023.S" +
|
||||
"plitLogTask.State\022\037\n\nserverName\030\002 \002(\0132\013.",
|
||||
"ServerName\"C\n\005State\022\016\n\nUNASSIGNED\020\000\022\t\n\005O" +
|
||||
"WNED\020\001\022\014\n\010RESIGNED\020\002\022\010\n\004DONE\020\003\022\007\n\003ERR\020\004\"" +
|
||||
"n\n\005Table\022$\n\005state\030\001 \002(\0162\014.Table.State:\007E" +
|
||||
"NABLED\"?\n\005State\022\013\n\007ENABLED\020\000\022\014\n\010DISABLED" +
|
||||
"\020\001\022\r\n\tDISABLING\020\002\022\014\n\010ENABLING\020\003\"%\n\017Repli" +
|
||||
"cationPeer\022\022\n\nclusterkey\030\001 \002(\t\"^\n\020Replic" +
|
||||
"ationState\022&\n\005state\030\001 \002(\0162\027.ReplicationS" +
|
||||
"tate.State\"\"\n\005State\022\013\n\007ENABLED\020\000\022\014\n\010DISA" +
|
||||
"BLED\020\001\"+\n\027ReplicationHLogPosition\022\020\n\010pos" +
|
||||
"ition\030\001 \002(\003\"$\n\017ReplicationLock\022\021\n\tlockOw",
|
||||
"ner\030\001 \002(\t\"\207\001\n\tTableLock\022\021\n\ttableName\030\001 \001" +
|
||||
"(\014\022\036\n\tlockOwner\030\002 \001(\0132\013.ServerName\022\020\n\010th" +
|
||||
"readId\030\003 \001(\003\022\020\n\010isShared\030\004 \001(\010\022\017\n\007purpos" +
|
||||
"e\030\005 \001(\t\022\022\n\ncreateTime\030\006 \001(\003\"9\n\017StoreSequ" +
|
||||
"enceId\022\022\n\nfamilyName\030\001 \002(\014\022\022\n\nsequenceId" +
|
||||
"\030\002 \002(\004\"b\n\026RegionStoreSequenceIds\022\035\n\025last" +
|
||||
"FlushedSequenceId\030\001 \002(\004\022)\n\017storeSequence" +
|
||||
"Id\030\002 \003(\0132\020.StoreSequenceIdBE\n*org.apache" +
|
||||
".hadoop.hbase.protobuf.generatedB\017ZooKee" +
|
||||
"perProtosH\001\210\001\001\240\001\001"
|
||||
"\023\n\013rpc_version\030\002 \001(\r\":\n\006Master\022\033\n\006master" +
|
||||
"\030\001 \002(\0132\013.ServerName\022\023\n\013rpc_version\030\002 \001(\r" +
|
||||
"\"\037\n\tClusterUp\022\022\n\nstart_date\030\001 \002(\t\"\210\001\n\020Re" +
|
||||
"gionTransition\022\027\n\017event_type_code\030\001 \002(\r\022" +
|
||||
"\023\n\013region_name\030\002 \002(\014\022\023\n\013create_time\030\003 \002(" +
|
||||
"\004\022 \n\013server_name\030\004 \002(\0132\013.ServerName\022\017\n\007p" +
|
||||
"ayload\030\005 \001(\014\"\231\001\n\014SplitLogTask\022\"\n\005state\030\001" +
|
||||
" \002(\0162\023.SplitLogTask.State\022 \n\013server_name",
|
||||
"\030\002 \002(\0132\013.ServerName\"C\n\005State\022\016\n\nUNASSIGN" +
|
||||
"ED\020\000\022\t\n\005OWNED\020\001\022\014\n\010RESIGNED\020\002\022\010\n\004DONE\020\003\022" +
|
||||
"\007\n\003ERR\020\004\"n\n\005Table\022$\n\005state\030\001 \002(\0162\014.Table" +
|
||||
".State:\007ENABLED\"?\n\005State\022\013\n\007ENABLED\020\000\022\014\n" +
|
||||
"\010DISABLED\020\001\022\r\n\tDISABLING\020\002\022\014\n\010ENABLING\020\003" +
|
||||
"\"%\n\017ReplicationPeer\022\022\n\nclusterkey\030\001 \002(\t\"" +
|
||||
"^\n\020ReplicationState\022&\n\005state\030\001 \002(\0162\027.Rep" +
|
||||
"licationState.State\"\"\n\005State\022\013\n\007ENABLED\020" +
|
||||
"\000\022\014\n\010DISABLED\020\001\"+\n\027ReplicationHLogPositi" +
|
||||
"on\022\020\n\010position\030\001 \002(\003\"%\n\017ReplicationLock\022",
|
||||
"\022\n\nlock_owner\030\001 \002(\t\"\214\001\n\tTableLock\022\022\n\ntab" +
|
||||
"le_name\030\001 \001(\014\022\037\n\nlock_owner\030\002 \001(\0132\013.Serv" +
|
||||
"erName\022\021\n\tthread_id\030\003 \001(\003\022\021\n\tis_shared\030\004" +
|
||||
" \001(\010\022\017\n\007purpose\030\005 \001(\t\022\023\n\013create_time\030\006 \001" +
|
||||
"(\003\";\n\017StoreSequenceId\022\023\n\013family_name\030\001 \002" +
|
||||
"(\014\022\023\n\013sequence_id\030\002 \002(\004\"g\n\026RegionStoreSe" +
|
||||
"quenceIds\022 \n\030last_flushed_sequence_id\030\001 " +
|
||||
"\002(\004\022+\n\021store_sequence_id\030\002 \003(\0132\020.StoreSe" +
|
||||
"quenceIdBE\n*org.apache.hadoop.hbase.prot" +
|
||||
"obuf.generatedB\017ZooKeeperProtosH\001\210\001\001\240\001\001"
|
||||
};
|
||||
com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
|
||||
new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() {
|
||||
|
|
|
@ -85,15 +85,15 @@ message CheckPermissionsResponse {
|
|||
}
|
||||
|
||||
service AccessControlService {
|
||||
rpc grant(GrantRequest)
|
||||
rpc Grant(GrantRequest)
|
||||
returns (GrantResponse);
|
||||
|
||||
rpc revoke(RevokeRequest)
|
||||
rpc Revoke(RevokeRequest)
|
||||
returns (RevokeResponse);
|
||||
|
||||
rpc getUserPermissions(UserPermissionsRequest)
|
||||
rpc GetUserPermissions(UserPermissionsRequest)
|
||||
returns (UserPermissionsResponse);
|
||||
|
||||
rpc checkPermissions(CheckPermissionsRequest)
|
||||
rpc CheckPermissions(CheckPermissionsRequest)
|
||||
returns (CheckPermissionsResponse);
|
||||
}
|
||||
|
|
|
@ -30,12 +30,12 @@ import "WAL.proto";
|
|||
|
||||
message GetRegionInfoRequest {
|
||||
required RegionSpecifier region = 1;
|
||||
optional bool compactionState = 2;
|
||||
optional bool compaction_state = 2;
|
||||
}
|
||||
|
||||
message GetRegionInfoResponse {
|
||||
required RegionInfo regionInfo = 1;
|
||||
optional CompactionState compactionState = 2;
|
||||
required RegionInfo region_info = 1;
|
||||
optional CompactionState compaction_state = 2;
|
||||
|
||||
enum CompactionState {
|
||||
NONE = 0;
|
||||
|
@ -55,28 +55,28 @@ message GetStoreFileRequest {
|
|||
}
|
||||
|
||||
message GetStoreFileResponse {
|
||||
repeated string storeFile = 1;
|
||||
repeated string store_file = 1;
|
||||
}
|
||||
|
||||
message GetOnlineRegionRequest {
|
||||
}
|
||||
|
||||
message GetOnlineRegionResponse {
|
||||
repeated RegionInfo regionInfo = 1;
|
||||
repeated RegionInfo region_info = 1;
|
||||
}
|
||||
|
||||
message OpenRegionRequest {
|
||||
repeated RegionOpenInfo openInfo = 1;
|
||||
repeated RegionOpenInfo open_info = 1;
|
||||
|
||||
message RegionOpenInfo {
|
||||
required RegionInfo region = 1;
|
||||
optional uint32 versionOfOfflineNode = 2;
|
||||
repeated ServerName favoredNodes = 3;
|
||||
optional uint32 version_of_offline_node = 2;
|
||||
repeated ServerName favored_nodes = 3;
|
||||
}
|
||||
}
|
||||
|
||||
message OpenRegionResponse {
|
||||
repeated RegionOpeningState openingState = 1;
|
||||
repeated RegionOpeningState opening_state = 1;
|
||||
|
||||
enum RegionOpeningState {
|
||||
OPENED = 0;
|
||||
|
@ -91,9 +91,9 @@ message OpenRegionResponse {
|
|||
*/
|
||||
message CloseRegionRequest {
|
||||
required RegionSpecifier region = 1;
|
||||
optional uint32 versionOfClosingNode = 2;
|
||||
optional bool transitionInZK = 3 [default = true];
|
||||
optional ServerName destinationServer = 4;
|
||||
optional uint32 version_of_closing_node = 2;
|
||||
optional bool transition_in_ZK = 3 [default = true];
|
||||
optional ServerName destination_server = 4;
|
||||
}
|
||||
|
||||
message CloseRegionResponse {
|
||||
|
@ -107,11 +107,11 @@ message CloseRegionResponse {
|
|||
*/
|
||||
message FlushRegionRequest {
|
||||
required RegionSpecifier region = 1;
|
||||
optional uint64 ifOlderThanTs = 2;
|
||||
optional uint64 if_older_than_ts = 2;
|
||||
}
|
||||
|
||||
message FlushRegionResponse {
|
||||
required uint64 lastFlushTime = 1;
|
||||
required uint64 last_flush_time = 1;
|
||||
optional bool flushed = 2;
|
||||
}
|
||||
|
||||
|
@ -124,7 +124,7 @@ message FlushRegionResponse {
|
|||
*/
|
||||
message SplitRegionRequest {
|
||||
required RegionSpecifier region = 1;
|
||||
optional bytes splitPoint = 2;
|
||||
optional bytes split_point = 2;
|
||||
}
|
||||
|
||||
message SplitRegionResponse {
|
||||
|
@ -150,8 +150,8 @@ message CompactRegionResponse {
|
|||
* This method currently closes the regions and then merges them
|
||||
*/
|
||||
message MergeRegionsRequest {
|
||||
required RegionSpecifier regionA = 1;
|
||||
required RegionSpecifier regionB = 2;
|
||||
required RegionSpecifier region_a = 1;
|
||||
required RegionSpecifier region_b = 2;
|
||||
optional bool forcible = 3 [default = false];
|
||||
}
|
||||
|
||||
|
@ -163,10 +163,10 @@ message WALEntry {
|
|||
required WALKey key = 1;
|
||||
// Following may be null if the KVs/Cells are carried along the side in a cellblock (See
|
||||
// RPC for more on cellblocks). If Cells/KVs are in a cellblock, this next field is null
|
||||
// and associatedCellCount has count of Cells associated w/ this WALEntry
|
||||
repeated bytes keyValueBytes = 2;
|
||||
// and associated_cell_count has count of Cells associated w/ this WALEntry
|
||||
repeated bytes key_value_bytes = 2;
|
||||
// If Cell data is carried alongside in a cellblock, this is count of Cells in the cellblock.
|
||||
optional int32 associatedCellCount = 3;
|
||||
optional int32 associated_cell_count = 3;
|
||||
}
|
||||
|
||||
/**
|
||||
|
@ -186,7 +186,7 @@ message RollWALWriterRequest {
|
|||
|
||||
message RollWALWriterResponse {
|
||||
// A list of encoded name of regions to flush
|
||||
repeated bytes regionToFlush = 1;
|
||||
repeated bytes region_to_flush = 1;
|
||||
}
|
||||
|
||||
message StopServerRequest {
|
||||
|
@ -200,54 +200,54 @@ message GetServerInfoRequest {
|
|||
}
|
||||
|
||||
message ServerInfo {
|
||||
required ServerName serverName = 1;
|
||||
optional uint32 webuiPort = 2;
|
||||
required ServerName server_name = 1;
|
||||
optional uint32 webui_port = 2;
|
||||
}
|
||||
|
||||
message GetServerInfoResponse {
|
||||
required ServerInfo serverInfo = 1;
|
||||
required ServerInfo server_info = 1;
|
||||
}
|
||||
|
||||
service AdminService {
|
||||
rpc getRegionInfo(GetRegionInfoRequest)
|
||||
rpc GetRegionInfo(GetRegionInfoRequest)
|
||||
returns(GetRegionInfoResponse);
|
||||
|
||||
rpc getStoreFile(GetStoreFileRequest)
|
||||
rpc GetStoreFile(GetStoreFileRequest)
|
||||
returns(GetStoreFileResponse);
|
||||
|
||||
rpc getOnlineRegion(GetOnlineRegionRequest)
|
||||
rpc GetOnlineRegion(GetOnlineRegionRequest)
|
||||
returns(GetOnlineRegionResponse);
|
||||
|
||||
rpc openRegion(OpenRegionRequest)
|
||||
rpc OpenRegion(OpenRegionRequest)
|
||||
returns(OpenRegionResponse);
|
||||
|
||||
rpc closeRegion(CloseRegionRequest)
|
||||
rpc CloseRegion(CloseRegionRequest)
|
||||
returns(CloseRegionResponse);
|
||||
|
||||
rpc flushRegion(FlushRegionRequest)
|
||||
rpc FlushRegion(FlushRegionRequest)
|
||||
returns(FlushRegionResponse);
|
||||
|
||||
rpc splitRegion(SplitRegionRequest)
|
||||
rpc SplitRegion(SplitRegionRequest)
|
||||
returns(SplitRegionResponse);
|
||||
|
||||
rpc compactRegion(CompactRegionRequest)
|
||||
rpc CompactRegion(CompactRegionRequest)
|
||||
returns(CompactRegionResponse);
|
||||
|
||||
rpc mergeRegions(MergeRegionsRequest)
|
||||
|
||||
rpc MergeRegions(MergeRegionsRequest)
|
||||
returns(MergeRegionsResponse);
|
||||
|
||||
rpc replicateWALEntry(ReplicateWALEntryRequest)
|
||||
rpc ReplicateWALEntry(ReplicateWALEntryRequest)
|
||||
returns(ReplicateWALEntryResponse);
|
||||
|
||||
rpc replay(MultiRequest)
|
||||
returns(MultiResponse);
|
||||
|
||||
rpc rollWALWriter(RollWALWriterRequest)
|
||||
rpc Replay(MultiRequest)
|
||||
returns(MultiResponse);
|
||||
|
||||
rpc RollWALWriter(RollWALWriterRequest)
|
||||
returns(RollWALWriterResponse);
|
||||
|
||||
rpc getServerInfo(GetServerInfoRequest)
|
||||
rpc GetServerInfo(GetServerInfoRequest)
|
||||
returns(GetServerInfoResponse);
|
||||
|
||||
rpc stopServer(StopServerRequest)
|
||||
rpc StopServer(StopServerRequest)
|
||||
returns(StopServerResponse);
|
||||
}
|
||||
|
|
|
@ -31,32 +31,32 @@ message AggregateArgument {
|
|||
* (3) any bytes required to construct the ColumnInterpreter object
|
||||
* properly
|
||||
*/
|
||||
required string interpreterClassName = 1;
|
||||
required string interpreter_class_name = 1;
|
||||
required Scan scan = 2;
|
||||
optional bytes interpreterSpecificBytes = 3;
|
||||
optional bytes interpreter_specific_bytes = 3;
|
||||
}
|
||||
|
||||
message AggregateResponse {
|
||||
/**
|
||||
* The AggregateService methods all have a response that either is a Pair
|
||||
* or a simple object. When it is a Pair both firstPart and secondPart
|
||||
* have defined values (and the secondPart is not present in the response
|
||||
* or a simple object. When it is a Pair both first_part and second_part
|
||||
* have defined values (and the second_part is not present in the response
|
||||
* when the response is not a pair). Refer to the AggregateImplementation
|
||||
* class for an overview of the AggregateResponse object constructions.
|
||||
*/
|
||||
repeated bytes firstPart = 1;
|
||||
optional bytes secondPart = 2;
|
||||
repeated bytes first_part = 1;
|
||||
optional bytes second_part = 2;
|
||||
}
|
||||
|
||||
/** Refer to the AggregateImplementation class for an overview of the
|
||||
* AggregateService method implementations and their functionality.
|
||||
*/
|
||||
service AggregateService {
|
||||
rpc getMax (AggregateArgument) returns (AggregateResponse);
|
||||
rpc getMin (AggregateArgument) returns (AggregateResponse);
|
||||
rpc getSum (AggregateArgument) returns (AggregateResponse);
|
||||
rpc getRowNum (AggregateArgument) returns (AggregateResponse);
|
||||
rpc getAvg (AggregateArgument) returns (AggregateResponse);
|
||||
rpc getStd (AggregateArgument) returns (AggregateResponse);
|
||||
rpc getMedian (AggregateArgument) returns (AggregateResponse);
|
||||
}
|
||||
rpc GetMax (AggregateArgument) returns (AggregateResponse);
|
||||
rpc GetMin (AggregateArgument) returns (AggregateResponse);
|
||||
rpc GetSum (AggregateArgument) returns (AggregateResponse);
|
||||
rpc GetRowNum (AggregateArgument) returns (AggregateResponse);
|
||||
rpc GetAvg (AggregateArgument) returns (AggregateResponse);
|
||||
rpc GetStd (AggregateArgument) returns (AggregateResponse);
|
||||
rpc GetMedian (AggregateArgument) returns (AggregateResponse);
|
||||
}
|
||||
|
|
|
@ -24,7 +24,7 @@ option optimize_for = SPEED;
|
|||
|
||||
message AuthenticationKey {
|
||||
required int32 id = 1;
|
||||
required int64 expirationDate = 2;
|
||||
required int64 expiration_date = 2;
|
||||
required bytes key = 3;
|
||||
}
|
||||
|
||||
|
@ -35,10 +35,10 @@ message TokenIdentifier {
|
|||
}
|
||||
required Kind kind = 1;
|
||||
required bytes username = 2;
|
||||
required int32 keyId = 3;
|
||||
optional int64 issueDate = 4;
|
||||
optional int64 expirationDate = 5;
|
||||
optional int64 sequenceNumber = 6;
|
||||
required int32 key_id = 3;
|
||||
optional int64 issue_date = 4;
|
||||
optional int64 expiration_date = 5;
|
||||
optional int64 sequence_number = 6;
|
||||
}
|
||||
|
||||
|
||||
|
@ -67,15 +67,15 @@ message WhoAmIRequest {
|
|||
|
||||
message WhoAmIResponse {
|
||||
optional string username = 1;
|
||||
optional string authMethod = 2;
|
||||
optional string auth_method = 2;
|
||||
}
|
||||
|
||||
|
||||
// RPC service
|
||||
service AuthenticationService {
|
||||
rpc getAuthenticationToken(TokenRequest)
|
||||
rpc GetAuthenticationToken(TokenRequest)
|
||||
returns (TokenResponse);
|
||||
|
||||
rpc whoami(WhoAmIRequest)
|
||||
rpc WhoAmI(WhoAmIRequest)
|
||||
returns (WhoAmIResponse);
|
||||
}
|
||||
}
|
||||
|
|
|
@ -46,7 +46,7 @@ message Cell {
|
|||
optional bytes family = 2;
|
||||
optional bytes qualifier = 3;
|
||||
optional uint64 timestamp = 4;
|
||||
optional CellType cellType = 5;
|
||||
optional CellType cell_type = 5;
|
||||
optional bytes value = 6;
|
||||
}
|
||||
|
||||
|
@ -59,6 +59,6 @@ message KeyValue {
|
|||
required bytes family = 2;
|
||||
required bytes qualifier = 3;
|
||||
optional uint64 timestamp = 4;
|
||||
optional CellType keyType = 5;
|
||||
optional CellType key_type = 5;
|
||||
optional bytes value = 6;
|
||||
}
|
||||
|
|
|
@ -45,11 +45,11 @@ message Get {
|
|||
repeated Column column = 2;
|
||||
repeated NameBytesPair attribute = 3;
|
||||
optional Filter filter = 4;
|
||||
optional TimeRange timeRange = 5;
|
||||
optional uint32 maxVersions = 6 [default = 1];
|
||||
optional bool cacheBlocks = 7 [default = true];
|
||||
optional uint32 storeLimit = 8;
|
||||
optional uint32 storeOffset = 9;
|
||||
optional TimeRange time_range = 5;
|
||||
optional uint32 max_versions = 6 [default = 1];
|
||||
optional bool cache_blocks = 7 [default = true];
|
||||
optional uint32 store_limit = 8;
|
||||
optional uint32 store_offset = 9;
|
||||
}
|
||||
|
||||
message Result {
|
||||
|
@ -60,18 +60,18 @@ message Result {
|
|||
// not part of this protobuf message; they are passed alongside
|
||||
// and then this Message is just a placeholder with metadata.
|
||||
// The count is needed to know how many to peel off the block of Cells as
|
||||
// ours. NOTE: This is different from the pb managed cellCount of the
|
||||
// ours. NOTE: This is different from the pb managed cell_count of the
|
||||
// 'cell' field above which is non-null when the cells are pb'd.
|
||||
optional int32 associatedCellCount = 2;
|
||||
optional int32 associated_cell_count = 2;
|
||||
}
|
||||
|
||||
/**
|
||||
* The get request. Perform a single Get operation.
|
||||
* Unless existenceOnly is specified, return all the requested data
|
||||
* Unless existence_only is specified, return all the requested data
|
||||
* for the row that matches exactly, or the one that immediately
|
||||
* precedes it if closestRowBefore is specified.
|
||||
* precedes it if closest_row_before is specified.
|
||||
*
|
||||
* If existenceOnly is set, only the existence will be returned.
|
||||
* If existence_only is set, only the existence will be returned.
|
||||
*/
|
||||
message GetRequest {
|
||||
required RegionSpecifier region = 1;
|
||||
|
@ -79,12 +79,12 @@ message GetRequest {
|
|||
|
||||
// If the row to get doesn't exist, return the
|
||||
// closest row before.
|
||||
optional bool closestRowBefore = 3;
|
||||
optional bool closest_row_before = 3;
|
||||
|
||||
// The result isn't asked for, just check for
|
||||
// the existence. If closestRowBefore specified,
|
||||
// the existence. If closest_row_before specified,
|
||||
// this will be ignored
|
||||
optional bool existenceOnly = 4;
|
||||
optional bool existence_only = 4;
|
||||
}
|
||||
|
||||
message MultiGetRequest {
|
||||
|
@ -93,12 +93,12 @@ message MultiGetRequest {
|
|||
|
||||
// If the row to get doesn't exist, return the
|
||||
// closest row before.
|
||||
optional bool closestRowBefore = 3;
|
||||
optional bool closest_row_before = 3;
|
||||
|
||||
// The result isn't asked for, just check for
|
||||
// the existence. If closestRowBefore specified,
|
||||
// the existence. If closest_row_before specified,
|
||||
// this will be ignored
|
||||
optional bool existenceOnly = 4;
|
||||
optional bool existence_only = 4;
|
||||
}
|
||||
|
||||
message GetResponse {
|
||||
|
@ -125,7 +125,7 @@ message Condition {
|
|||
required bytes row = 1;
|
||||
required bytes family = 2;
|
||||
required bytes qualifier = 3;
|
||||
required CompareType compareType = 4;
|
||||
required CompareType compare_type = 4;
|
||||
required Comparator comparator = 5;
|
||||
}
|
||||
|
||||
|
@ -139,22 +139,22 @@ message Condition {
|
|||
*/
|
||||
message MutationProto {
|
||||
optional bytes row = 1;
|
||||
optional MutationType mutateType = 2;
|
||||
repeated ColumnValue columnValue = 3;
|
||||
optional MutationType mutate_type = 2;
|
||||
repeated ColumnValue column_value = 3;
|
||||
optional uint64 timestamp = 4;
|
||||
repeated NameBytesPair attribute = 5;
|
||||
optional Durability durability = 6 [default = USE_DEFAULT];
|
||||
|
||||
// For some mutations, a result may be returned, in which case,
|
||||
// time range can be specified for potential performance gain
|
||||
optional TimeRange timeRange = 7;
|
||||
optional TimeRange time_range = 7;
|
||||
// The below count is set when the associated cells are NOT
|
||||
// part of this protobuf message; they are passed alongside
|
||||
// and then this Message is a placeholder with metadata. The
|
||||
// count is needed to know how many to peel off the block of Cells as
|
||||
// ours. NOTE: This is different from the pb managed cellCount of the
|
||||
// ours. NOTE: This is different from the pb managed cell_count of the
|
||||
// 'cell' field above which is non-null when the cells are pb'd.
|
||||
optional int32 associatedCellCount = 8;
|
||||
optional int32 associated_cell_count = 8;
|
||||
|
||||
enum Durability {
|
||||
USE_DEFAULT = 0;
|
||||
|
@ -180,13 +180,13 @@ message MutationProto {
|
|||
|
||||
message ColumnValue {
|
||||
required bytes family = 1;
|
||||
repeated QualifierValue qualifierValue = 2;
|
||||
repeated QualifierValue qualifier_value = 2;
|
||||
|
||||
message QualifierValue {
|
||||
optional bytes qualifier = 1;
|
||||
optional bytes value = 2;
|
||||
optional uint64 timestamp = 3;
|
||||
optional DeleteType deleteType = 4;
|
||||
optional DeleteType delete_type = 4;
|
||||
}
|
||||
}
|
||||
}
|
||||
|
@ -225,18 +225,18 @@ message MutateResponse {
|
|||
message Scan {
|
||||
repeated Column column = 1;
|
||||
repeated NameBytesPair attribute = 2;
|
||||
optional bytes startRow = 3;
|
||||
optional bytes stopRow = 4;
|
||||
optional bytes start_row = 3;
|
||||
optional bytes stop_row = 4;
|
||||
optional Filter filter = 5;
|
||||
optional TimeRange timeRange = 6;
|
||||
optional uint32 maxVersions = 7 [default = 1];
|
||||
optional bool cacheBlocks = 8 [default = true];
|
||||
optional uint32 batchSize = 9;
|
||||
optional uint64 maxResultSize = 10;
|
||||
optional uint32 storeLimit = 11;
|
||||
optional uint32 storeOffset = 12;
|
||||
optional bool loadColumnFamiliesOnDemand = 13; /* DO NOT add defaults to loadColumnFamiliesOnDemand. */
|
||||
optional uint32 cachingCount = 14;
|
||||
optional TimeRange time_range = 6;
|
||||
optional uint32 max_versions = 7 [default = 1];
|
||||
optional bool cache_blocks = 8 [default = true];
|
||||
optional uint32 batch_size = 9;
|
||||
optional uint64 max_result_size = 10;
|
||||
optional uint32 store_limit = 11;
|
||||
optional uint32 store_offset = 12;
|
||||
optional bool load_column_families_on_demand = 13; /* DO NOT add defaults to load_column_families_on_demand. */
|
||||
optional uint32 caching_count = 14;
|
||||
optional bool prefetching = 15;
|
||||
}
|
||||
|
||||
|
@ -254,25 +254,25 @@ message Scan {
|
|||
message ScanRequest {
|
||||
optional RegionSpecifier region = 1;
|
||||
optional Scan scan = 2;
|
||||
optional uint64 scannerId = 3;
|
||||
optional uint32 numberOfRows = 4;
|
||||
optional bool closeScanner = 5;
|
||||
optional uint64 nextCallSeq = 6;
|
||||
optional uint64 scanner_id = 3;
|
||||
optional uint32 number_of_rows = 4;
|
||||
optional bool close_scanner = 5;
|
||||
optional uint64 next_call_seq = 6;
|
||||
}
|
||||
|
||||
/**
|
||||
* The scan response. If there are no more results, moreResults will
|
||||
* The scan response. If there are no more results, more_results will
|
||||
* be false. If it is not specified, it means there are more.
|
||||
*/
|
||||
message ScanResponse {
|
||||
optional ResultCellMeta resultCellMeta = 1;
|
||||
optional uint64 scannerId = 2;
|
||||
optional bool moreResults = 3;
|
||||
optional ResultCellMeta result_cell_meta = 1;
|
||||
optional uint64 scanner_id = 2;
|
||||
optional bool more_results = 3;
|
||||
optional uint32 ttl = 4;
|
||||
}
|
||||
|
||||
message ResultCellMeta {
|
||||
repeated uint32 cellsLength = 1;
|
||||
repeated uint32 cells_length = 1;
|
||||
}
|
||||
|
||||
/**
|
||||
|
@ -281,8 +281,8 @@ message ResultCellMeta {
|
|||
*/
|
||||
message BulkLoadHFileRequest {
|
||||
required RegionSpecifier region = 1;
|
||||
repeated FamilyPath familyPath = 2;
|
||||
optional bool assignSeqNum = 3;
|
||||
repeated FamilyPath family_path = 2;
|
||||
optional bool assign_seq_num = 3;
|
||||
|
||||
message FamilyPath {
|
||||
required bytes family = 1;
|
||||
|
@ -296,8 +296,8 @@ message BulkLoadHFileResponse {
|
|||
|
||||
message CoprocessorServiceCall {
|
||||
required bytes row = 1;
|
||||
required string serviceName = 2;
|
||||
required string methodName = 3;
|
||||
required string service_name = 2;
|
||||
required string method_name = 3;
|
||||
required bytes request = 4;
|
||||
}
|
||||
|
||||
|
@ -352,24 +352,24 @@ message MultiResponse {
|
|||
|
||||
|
||||
service ClientService {
|
||||
rpc get(GetRequest)
|
||||
rpc Get(GetRequest)
|
||||
returns(GetResponse);
|
||||
|
||||
rpc multiGet(MultiGetRequest)
|
||||
rpc MultiGet(MultiGetRequest)
|
||||
returns(MultiGetResponse);
|
||||
|
||||
rpc mutate(MutateRequest)
|
||||
rpc Mutate(MutateRequest)
|
||||
returns(MutateResponse);
|
||||
|
||||
rpc scan(ScanRequest)
|
||||
rpc Scan(ScanRequest)
|
||||
returns(ScanResponse);
|
||||
|
||||
rpc bulkLoadHFile(BulkLoadHFileRequest)
|
||||
rpc BulkLoadHFile(BulkLoadHFileRequest)
|
||||
returns(BulkLoadHFileResponse);
|
||||
|
||||
rpc execService(CoprocessorServiceRequest)
|
||||
rpc ExecService(CoprocessorServiceRequest)
|
||||
returns(CoprocessorServiceResponse);
|
||||
|
||||
rpc multi(MultiRequest)
|
||||
rpc Multi(MultiRequest)
|
||||
returns(MultiResponse);
|
||||
}
|
||||
|
|
|
@ -29,5 +29,5 @@ option optimize_for = SPEED;
|
|||
*/
|
||||
message ClusterId {
|
||||
// This is the cluster id, a uuid as a String
|
||||
required string clusterId = 1;
|
||||
required string cluster_id = 1;
|
||||
}
|
||||
|
|
|
@ -28,7 +28,7 @@ import "ClusterId.proto";
|
|||
import "FS.proto";
|
||||
|
||||
message RegionState {
|
||||
required RegionInfo regionInfo = 1;
|
||||
required RegionInfo region_info = 1;
|
||||
required State state = 2;
|
||||
optional uint64 stamp = 3;
|
||||
enum State {
|
||||
|
@ -50,12 +50,12 @@ message RegionState {
|
|||
|
||||
message RegionInTransition {
|
||||
required RegionSpecifier spec = 1;
|
||||
required RegionState regionState = 2;
|
||||
required RegionState region_state = 2;
|
||||
}
|
||||
|
||||
message RegionLoad {
|
||||
/** the region specifier */
|
||||
required RegionSpecifier regionSpecifier = 1;
|
||||
required RegionSpecifier region_specifier = 1;
|
||||
|
||||
/** the number of stores for the region */
|
||||
optional uint32 stores = 2;
|
||||
|
@ -64,65 +64,65 @@ message RegionLoad {
|
|||
optional uint32 storefiles = 3;
|
||||
|
||||
/** the total size of the store files for the region, uncompressed, in MB */
|
||||
optional uint32 storeUncompressedSizeMB = 4;
|
||||
optional uint32 store_uncompressed_size_MB = 4;
|
||||
|
||||
/** the current total size of the store files for the region, in MB */
|
||||
optional uint32 storefileSizeMB = 5;
|
||||
optional uint32 storefile_size_MB = 5;
|
||||
|
||||
/** the current size of the memstore for the region, in MB */
|
||||
optional uint32 memstoreSizeMB = 6;
|
||||
optional uint32 memstore_size_MB = 6;
|
||||
|
||||
/**
|
||||
* The current total size of root-level store file indexes for the region,
|
||||
* in MB. The same as {@link #rootIndexSizeKB} but in MB.
|
||||
*/
|
||||
optional uint32 storefileIndexSizeMB = 7;
|
||||
optional uint32 storefile_index_size_MB = 7;
|
||||
|
||||
/** the current total read requests made to region */
|
||||
optional uint64 readRequestsCount = 8;
|
||||
optional uint64 read_requests_count = 8;
|
||||
|
||||
/** the current total write requests made to region */
|
||||
optional uint64 writeRequestsCount = 9;
|
||||
optional uint64 write_requests_count = 9;
|
||||
|
||||
/** the total compacting key values in currently running compaction */
|
||||
optional uint64 totalCompactingKVs = 10;
|
||||
optional uint64 total_compacting_KVs = 10;
|
||||
|
||||
/** the completed count of key values in currently running compaction */
|
||||
optional uint64 currentCompactedKVs = 11;
|
||||
optional uint64 current_compacted_KVs = 11;
|
||||
|
||||
/** The current total size of root-level indexes for the region, in KB. */
|
||||
optional uint32 rootIndexSizeKB = 12;
|
||||
optional uint32 root_index_size_KB = 12;
|
||||
|
||||
/** The total size of all index blocks, not just the root level, in KB. */
|
||||
optional uint32 totalStaticIndexSizeKB = 13;
|
||||
optional uint32 total_static_index_size_KB = 13;
|
||||
|
||||
/**
|
||||
* The total size of all Bloom filter blocks, not just loaded into the
|
||||
* block cache, in KB.
|
||||
*/
|
||||
optional uint32 totalStaticBloomSizeKB = 14;
|
||||
optional uint32 total_static_bloom_size_KB = 14;
|
||||
|
||||
/** the most recent sequence Id from cache flush */
|
||||
optional uint64 completeSequenceId = 15;
|
||||
optional uint64 complete_sequence_id = 15;
|
||||
}
|
||||
|
||||
/* Server-level protobufs */
|
||||
|
||||
message ServerLoad {
|
||||
/** Number of requests since last report. */
|
||||
optional uint32 numberOfRequests = 1;
|
||||
optional uint32 number_of_requests = 1;
|
||||
|
||||
/** Total Number of requests from the start of the region server. */
|
||||
optional uint32 totalNumberOfRequests = 2;
|
||||
optional uint32 total_number_of_requests = 2;
|
||||
|
||||
/** the amount of used heap, in MB. */
|
||||
optional uint32 usedHeapMB = 3;
|
||||
optional uint32 used_heap_MB = 3;
|
||||
|
||||
/** the maximum allowable size of the heap, in MB. */
|
||||
optional uint32 maxHeapMB = 4;
|
||||
optional uint32 max_heap_MB = 4;
|
||||
|
||||
/** Information on the load of individual regions. */
|
||||
repeated RegionLoad regionLoads = 5;
|
||||
repeated RegionLoad region_loads = 5;
|
||||
|
||||
/**
|
||||
* Regionserver-level coprocessors, e.g., WALObserver implementations.
|
||||
|
@ -132,39 +132,38 @@ message ServerLoad {
|
|||
repeated Coprocessor coprocessors = 6;
|
||||
|
||||
/**
|
||||
* Time when incremental (non-total) counts began being calculated (e.g. numberOfRequests)
|
||||
* Time when incremental (non-total) counts began being calculated (e.g. number_of_requests)
|
||||
* time is measured as the difference, measured in milliseconds, between the current time
|
||||
* and midnight, January 1, 1970 UTC.
|
||||
*/
|
||||
optional uint64 reportStartTime = 7;
|
||||
optional uint64 report_start_time = 7;
|
||||
|
||||
/**
|
||||
* Time when report was generated.
|
||||
* time is measured as the difference, measured in milliseconds, between the current time
|
||||
* and midnight, January 1, 1970 UTC.
|
||||
*/
|
||||
optional uint64 reportEndTime = 8;
|
||||
optional uint64 report_end_time = 8;
|
||||
|
||||
/**
|
||||
* The port number that this region server is hosing an info server on.
|
||||
*/
|
||||
optional uint32 infoServerPort = 9;
|
||||
optional uint32 info_server_port = 9;
|
||||
}
|
||||
|
||||
message LiveServerInfo {
|
||||
required ServerName server = 1;
|
||||
required ServerLoad serverLoad = 2;
|
||||
required ServerLoad server_load = 2;
|
||||
}
|
||||
|
||||
|
||||
message ClusterStatus {
|
||||
optional HBaseVersionFileContent hbaseVersion = 1;
|
||||
repeated LiveServerInfo liveServers = 2;
|
||||
repeated ServerName deadServers = 3;
|
||||
repeated RegionInTransition regionsInTransition = 4;
|
||||
optional ClusterId clusterId = 5;
|
||||
repeated Coprocessor masterCoprocessors = 6;
|
||||
optional HBaseVersionFileContent hbase_version = 1;
|
||||
repeated LiveServerInfo live_servers = 2;
|
||||
repeated ServerName dead_servers = 3;
|
||||
repeated RegionInTransition regions_in_transition = 4;
|
||||
optional ClusterId cluster_id = 5;
|
||||
repeated Coprocessor master_coprocessors = 6;
|
||||
optional ServerName master = 7;
|
||||
repeated ServerName backupMasters = 8;
|
||||
optional bool balancerOn = 9;
|
||||
repeated ServerName backup_masters = 8;
|
||||
optional bool balancer_on = 9;
|
||||
}
|
||||
|
|
|
@ -28,7 +28,7 @@ option optimize_for = SPEED;
|
|||
|
||||
message Comparator {
|
||||
required string name = 1;
|
||||
optional bytes serializedComparator = 2;
|
||||
optional bytes serialized_comparator = 2;
|
||||
}
|
||||
|
||||
message ByteArrayComparable {
|
||||
|
@ -45,7 +45,7 @@ message BinaryPrefixComparator {
|
|||
|
||||
message BitComparator {
|
||||
required ByteArrayComparable comparable = 1;
|
||||
required BitwiseOp bitwiseOp = 2;
|
||||
required BitwiseOp bitwise_op = 2;
|
||||
|
||||
enum BitwiseOp {
|
||||
AND = 1;
|
||||
|
@ -59,7 +59,7 @@ message NullComparator {
|
|||
|
||||
message RegexStringComparator {
|
||||
required string pattern = 1;
|
||||
required int32 patternFlags = 2;
|
||||
required int32 pattern_flags = 2;
|
||||
required string charset = 3;
|
||||
}
|
||||
|
||||
|
|
|
@ -28,10 +28,10 @@ option optimize_for = SPEED;
|
|||
* so we can serialize exceptions.
|
||||
*/
|
||||
message StackTraceElementMessage {
|
||||
optional string declaringClass = 1;
|
||||
optional string methodName = 2;
|
||||
optional string fileName = 3;
|
||||
optional int32 lineNumber = 4;
|
||||
optional string declaring_class = 1;
|
||||
optional string method_name = 2;
|
||||
optional string file_name = 3;
|
||||
optional int32 line_number = 4;
|
||||
}
|
||||
|
||||
/**
|
||||
|
@ -41,9 +41,9 @@ message StackTraceElementMessage {
|
|||
* (which should be another protobuffed class).
|
||||
*/
|
||||
message GenericExceptionMessage {
|
||||
optional string className = 1;
|
||||
optional string class_name = 1;
|
||||
optional string message = 2;
|
||||
optional bytes errorInfo = 3;
|
||||
optional bytes error_info = 3;
|
||||
repeated StackTraceElementMessage trace = 4;
|
||||
}
|
||||
|
||||
|
@ -53,6 +53,5 @@ message GenericExceptionMessage {
|
|||
*/
|
||||
message ForeignExceptionMessage {
|
||||
optional string source = 1;
|
||||
optional GenericExceptionMessage genericException = 2;
|
||||
|
||||
}
|
||||
optional GenericExceptionMessage generic_exception = 2;
|
||||
}
|
||||
|
|
|
@ -29,7 +29,7 @@ import "Comparator.proto";
|
|||
|
||||
message Filter {
|
||||
required string name = 1;
|
||||
optional bytes serializedFilter = 2;
|
||||
optional bytes serialized_filter = 2;
|
||||
}
|
||||
|
||||
message ColumnCountGetFilter {
|
||||
|
@ -39,7 +39,7 @@ message ColumnCountGetFilter {
|
|||
message ColumnPaginationFilter {
|
||||
required int32 limit = 1;
|
||||
optional int32 offset = 2;
|
||||
optional bytes columnOffset = 3;
|
||||
optional bytes column_offset = 3;
|
||||
}
|
||||
|
||||
message ColumnPrefixFilter {
|
||||
|
@ -47,26 +47,26 @@ message ColumnPrefixFilter {
|
|||
}
|
||||
|
||||
message ColumnRangeFilter {
|
||||
optional bytes minColumn = 1;
|
||||
optional bool minColumnInclusive = 2;
|
||||
optional bytes maxColumn = 3;
|
||||
optional bool maxColumnInclusive = 4;
|
||||
optional bytes min_column = 1;
|
||||
optional bool min_column_inclusive = 2;
|
||||
optional bytes max_column = 3;
|
||||
optional bool max_column_inclusive = 4;
|
||||
}
|
||||
|
||||
message CompareFilter {
|
||||
required CompareType compareOp = 1;
|
||||
required CompareType compare_op = 1;
|
||||
optional Comparator comparator = 2;
|
||||
}
|
||||
|
||||
message DependentColumnFilter {
|
||||
required CompareFilter compareFilter = 1;
|
||||
optional bytes columnFamily = 2;
|
||||
optional bytes columnQualifier = 3;
|
||||
optional bool dropDependentColumn = 4;
|
||||
required CompareFilter compare_filter = 1;
|
||||
optional bytes column_family = 2;
|
||||
optional bytes column_qualifier = 3;
|
||||
optional bool drop_dependent_column = 4;
|
||||
}
|
||||
|
||||
message FamilyFilter {
|
||||
required CompareFilter compareFilter = 1;
|
||||
required CompareFilter compare_filter = 1;
|
||||
}
|
||||
|
||||
message FilterList {
|
||||
|
@ -91,23 +91,23 @@ message FirstKeyValueMatchingQualifiersFilter {
|
|||
}
|
||||
|
||||
message FuzzyRowFilter {
|
||||
repeated BytesBytesPair fuzzyKeysData = 1;
|
||||
repeated BytesBytesPair fuzzy_keys_data = 1;
|
||||
}
|
||||
|
||||
message InclusiveStopFilter {
|
||||
optional bytes stopRowKey = 1;
|
||||
optional bytes stop_row_key = 1;
|
||||
}
|
||||
|
||||
message KeyOnlyFilter {
|
||||
required bool lenAsVal = 1;
|
||||
required bool len_as_val = 1;
|
||||
}
|
||||
|
||||
message MultipleColumnPrefixFilter {
|
||||
repeated bytes sortedPrefixes = 1;
|
||||
repeated bytes sorted_prefixes = 1;
|
||||
}
|
||||
|
||||
message PageFilter {
|
||||
required int64 pageSize = 1;
|
||||
required int64 page_size = 1;
|
||||
}
|
||||
|
||||
message PrefixFilter {
|
||||
|
@ -115,7 +115,7 @@ message PrefixFilter {
|
|||
}
|
||||
|
||||
message QualifierFilter {
|
||||
required CompareFilter compareFilter = 1;
|
||||
required CompareFilter compare_filter = 1;
|
||||
}
|
||||
|
||||
message RandomRowFilter {
|
||||
|
@ -123,20 +123,20 @@ message RandomRowFilter {
|
|||
}
|
||||
|
||||
message RowFilter {
|
||||
required CompareFilter compareFilter = 1;
|
||||
required CompareFilter compare_filter = 1;
|
||||
}
|
||||
|
||||
message SingleColumnValueExcludeFilter {
|
||||
required SingleColumnValueFilter singleColumnValueFilter = 1;
|
||||
required SingleColumnValueFilter single_column_value_filter = 1;
|
||||
}
|
||||
|
||||
message SingleColumnValueFilter {
|
||||
optional bytes columnFamily = 1;
|
||||
optional bytes columnQualifier = 2;
|
||||
required CompareType compareOp = 3;
|
||||
optional bytes column_family = 1;
|
||||
optional bytes column_qualifier = 2;
|
||||
required CompareType compare_op = 3;
|
||||
required Comparator comparator = 4;
|
||||
optional bool filterIfMissing = 5;
|
||||
optional bool latestVersionOnly = 6;
|
||||
optional bool filter_if_missing = 5;
|
||||
optional bool latest_version_only = 6;
|
||||
}
|
||||
|
||||
message SkipFilter {
|
||||
|
@ -148,7 +148,7 @@ message TimestampsFilter {
|
|||
}
|
||||
|
||||
message ValueFilter {
|
||||
required CompareFilter compareFilter = 1;
|
||||
required CompareFilter compare_filter = 1;
|
||||
}
|
||||
|
||||
message WhileMatchFilter {
|
||||
|
|
|
@ -25,21 +25,21 @@ import "hbase.proto";
|
|||
|
||||
// Map of name/values
|
||||
message FileInfoProto {
|
||||
repeated BytesBytesPair mapEntry = 1;
|
||||
repeated BytesBytesPair map_entry = 1;
|
||||
}
|
||||
|
||||
// HFile file trailer
|
||||
message FileTrailerProto {
|
||||
optional uint64 fileInfoOffset = 1;
|
||||
optional uint64 loadOnOpenDataOffset = 2;
|
||||
optional uint64 uncompressedDataIndexSize = 3;
|
||||
optional uint64 totalUncompressedBytes = 4;
|
||||
optional uint32 dataIndexCount = 5;
|
||||
optional uint32 metaIndexCount = 6;
|
||||
optional uint64 entryCount = 7;
|
||||
optional uint32 numDataIndexLevels = 8;
|
||||
optional uint64 firstDataBlockOffset = 9;
|
||||
optional uint64 lastDataBlockOffset = 10;
|
||||
optional string comparatorClassName = 11;
|
||||
optional uint32 compressionCodec = 12;
|
||||
optional uint64 file_info_offset = 1;
|
||||
optional uint64 load_on_open_data_offset = 2;
|
||||
optional uint64 uncompressed_data_index_size = 3;
|
||||
optional uint64 total_uncompressed_bytes = 4;
|
||||
optional uint32 data_index_count = 5;
|
||||
optional uint32 meta_index_count = 6;
|
||||
optional uint64 entry_count = 7;
|
||||
optional uint32 num_data_index_levels = 8;
|
||||
optional uint64 first_data_block_offset = 9;
|
||||
optional uint64 last_data_block_offset = 10;
|
||||
optional string comparator_class_name = 11;
|
||||
optional uint32 compression_codec = 12;
|
||||
}
|
||||
|
|
|
@ -24,5 +24,5 @@ option java_generate_equals_and_hash = true;
|
|||
option optimize_for = SPEED;
|
||||
|
||||
message LoadBalancerState {
|
||||
optional bool balancerOn = 1;
|
||||
optional bool balancer_on = 1;
|
||||
}
|
||||
|
|
|
@ -28,11 +28,11 @@ message IsMasterRunningRequest {
|
|||
}
|
||||
|
||||
message IsMasterRunningResponse {
|
||||
required bool isMasterRunning = 1;
|
||||
required bool is_master_running = 1;
|
||||
}
|
||||
|
||||
service MasterService {
|
||||
/** return true if master is available */
|
||||
rpc isMasterRunning(IsMasterRunningRequest)
|
||||
rpc IsMasterRunning(IsMasterRunningRequest)
|
||||
returns(IsMasterRunningResponse);
|
||||
}
|
||||
|
|
|
@ -32,24 +32,24 @@ import "Client.proto";
|
|||
/* Column-level protobufs */
|
||||
|
||||
message AddColumnRequest {
|
||||
required bytes tableName = 1;
|
||||
required ColumnFamilySchema columnFamilies = 2;
|
||||
required bytes table_name = 1;
|
||||
required ColumnFamilySchema column_families = 2;
|
||||
}
|
||||
|
||||
message AddColumnResponse {
|
||||
}
|
||||
|
||||
message DeleteColumnRequest {
|
||||
required bytes tableName = 1;
|
||||
required bytes columnName = 2;
|
||||
required bytes table_name = 1;
|
||||
required bytes column_name = 2;
|
||||
}
|
||||
|
||||
message DeleteColumnResponse {
|
||||
}
|
||||
|
||||
message ModifyColumnRequest {
|
||||
required bytes tableName = 1;
|
||||
required ColumnFamilySchema columnFamilies = 2;
|
||||
required bytes table_name = 1;
|
||||
required ColumnFamilySchema column_families = 2;
|
||||
}
|
||||
|
||||
message ModifyColumnResponse {
|
||||
|
@ -59,7 +59,7 @@ message ModifyColumnResponse {
|
|||
|
||||
message MoveRegionRequest {
|
||||
required RegionSpecifier region = 1;
|
||||
optional ServerName destServerName = 2;
|
||||
optional ServerName dest_server_name = 2;
|
||||
}
|
||||
|
||||
message MoveRegionResponse {
|
||||
|
@ -69,8 +69,8 @@ message MoveRegionResponse {
|
|||
* Dispatch merging the specified regions.
|
||||
*/
|
||||
message DispatchMergingRegionsRequest {
|
||||
required RegionSpecifier regionA = 1;
|
||||
required RegionSpecifier regionB = 2;
|
||||
required RegionSpecifier region_a = 1;
|
||||
required RegionSpecifier region_b = 2;
|
||||
optional bool forcible = 3 [default = false];
|
||||
}
|
||||
|
||||
|
@ -102,37 +102,37 @@ message OfflineRegionResponse {
|
|||
/* Table-level protobufs */
|
||||
|
||||
message CreateTableRequest {
|
||||
required TableSchema tableSchema = 1;
|
||||
repeated bytes splitKeys = 2;
|
||||
required TableSchema table_schema = 1;
|
||||
repeated bytes split_keys = 2;
|
||||
}
|
||||
|
||||
message CreateTableResponse {
|
||||
}
|
||||
|
||||
message DeleteTableRequest {
|
||||
required bytes tableName = 1;
|
||||
required bytes table_name = 1;
|
||||
}
|
||||
|
||||
message DeleteTableResponse {
|
||||
}
|
||||
|
||||
message EnableTableRequest {
|
||||
required bytes tableName = 1;
|
||||
required bytes table_name = 1;
|
||||
}
|
||||
|
||||
message EnableTableResponse {
|
||||
}
|
||||
|
||||
message DisableTableRequest {
|
||||
required bytes tableName = 1;
|
||||
required bytes table_name = 1;
|
||||
}
|
||||
|
||||
message DisableTableResponse {
|
||||
}
|
||||
|
||||
message ModifyTableRequest {
|
||||
required bytes tableName = 1;
|
||||
required TableSchema tableSchema = 2;
|
||||
required bytes table_name = 1;
|
||||
required TableSchema table_schema = 2;
|
||||
}
|
||||
|
||||
message ModifyTableResponse {
|
||||
|
@ -157,7 +157,7 @@ message BalanceRequest {
|
|||
}
|
||||
|
||||
message BalanceResponse {
|
||||
required bool balancerRan = 1;
|
||||
required bool balancer_ran = 1;
|
||||
}
|
||||
|
||||
message SetBalancerRunningRequest {
|
||||
|
@ -166,14 +166,14 @@ message SetBalancerRunningRequest {
|
|||
}
|
||||
|
||||
message SetBalancerRunningResponse {
|
||||
optional bool prevBalanceValue = 1;
|
||||
optional bool prev_balance_value = 1;
|
||||
}
|
||||
|
||||
message CatalogScanRequest {
|
||||
}
|
||||
|
||||
message CatalogScanResponse {
|
||||
optional int32 scanResult = 1;
|
||||
optional int32 scan_result = 1;
|
||||
}
|
||||
|
||||
message EnableCatalogJanitorRequest {
|
||||
|
@ -181,7 +181,7 @@ message EnableCatalogJanitorRequest {
|
|||
}
|
||||
|
||||
message EnableCatalogJanitorResponse {
|
||||
optional bool prevValue = 1;
|
||||
optional bool prev_value = 1;
|
||||
}
|
||||
|
||||
message IsCatalogJanitorEnabledRequest {
|
||||
|
@ -196,7 +196,7 @@ message TakeSnapshotRequest{
|
|||
}
|
||||
|
||||
message TakeSnapshotResponse{
|
||||
required int64 expectedTimeout = 1;
|
||||
required int64 expected_timeout = 1;
|
||||
}
|
||||
|
||||
message ListSnapshotRequest{
|
||||
|
@ -242,36 +242,36 @@ message IsRestoreSnapshotDoneResponse {
|
|||
|
||||
service MasterAdminService {
|
||||
/** Adds a column to the specified table. */
|
||||
rpc addColumn(AddColumnRequest)
|
||||
rpc AddColumn(AddColumnRequest)
|
||||
returns(AddColumnResponse);
|
||||
|
||||
/** Deletes a column from the specified table. Table must be disabled. */
|
||||
rpc deleteColumn(DeleteColumnRequest)
|
||||
rpc DeleteColumn(DeleteColumnRequest)
|
||||
returns(DeleteColumnResponse);
|
||||
|
||||
/** Modifies an existing column on the specified table. */
|
||||
rpc modifyColumn(ModifyColumnRequest)
|
||||
rpc ModifyColumn(ModifyColumnRequest)
|
||||
returns(ModifyColumnResponse);
|
||||
|
||||
/** Move the region region to the destination server. */
|
||||
rpc moveRegion(MoveRegionRequest)
|
||||
rpc MoveRegion(MoveRegionRequest)
|
||||
returns(MoveRegionResponse);
|
||||
|
||||
/** Master dispatch merging the regions */
|
||||
rpc dispatchMergingRegions(DispatchMergingRegionsRequest)
|
||||
rpc DispatchMergingRegions(DispatchMergingRegionsRequest)
|
||||
returns(DispatchMergingRegionsResponse);
|
||||
|
||||
/** Assign a region to a server chosen at random. */
|
||||
rpc assignRegion(AssignRegionRequest)
|
||||
rpc AssignRegion(AssignRegionRequest)
|
||||
returns(AssignRegionResponse);
|
||||
|
||||
/**
|
||||
* Unassign a region from current hosting regionserver. Region will then be
|
||||
* assigned to a regionserver chosen at random. Region could be reassigned
|
||||
* back to the same server. Use moveRegion if you want
|
||||
* back to the same server. Use MoveRegion if you want
|
||||
* to control the region movement.
|
||||
*/
|
||||
rpc unassignRegion(UnassignRegionRequest)
|
||||
rpc UnassignRegion(UnassignRegionRequest)
|
||||
returns(UnassignRegionResponse);
|
||||
|
||||
/**
|
||||
|
@ -280,35 +280,35 @@ service MasterAdminService {
|
|||
* automatically reassign the region as in unassign. This is a special
|
||||
* method, and should only be used by experts or hbck.
|
||||
*/
|
||||
rpc offlineRegion(OfflineRegionRequest)
|
||||
rpc OfflineRegion(OfflineRegionRequest)
|
||||
returns(OfflineRegionResponse);
|
||||
|
||||
/** Deletes a table */
|
||||
rpc deleteTable(DeleteTableRequest)
|
||||
rpc DeleteTable(DeleteTableRequest)
|
||||
returns(DeleteTableResponse);
|
||||
|
||||
/** Puts the table on-line (only needed if table has been previously taken offline) */
|
||||
rpc enableTable(EnableTableRequest)
|
||||
rpc EnableTable(EnableTableRequest)
|
||||
returns(EnableTableResponse);
|
||||
|
||||
/** Take table offline */
|
||||
rpc disableTable(DisableTableRequest)
|
||||
rpc DisableTable(DisableTableRequest)
|
||||
returns(DisableTableResponse);
|
||||
|
||||
/** Modify a table's metadata */
|
||||
rpc modifyTable(ModifyTableRequest)
|
||||
rpc ModifyTable(ModifyTableRequest)
|
||||
returns(ModifyTableResponse);
|
||||
|
||||
/** Creates a new table asynchronously */
|
||||
rpc createTable(CreateTableRequest)
|
||||
rpc CreateTable(CreateTableRequest)
|
||||
returns(CreateTableResponse);
|
||||
|
||||
/** Shutdown an HBase cluster. */
|
||||
rpc shutdown(ShutdownRequest)
|
||||
rpc Shutdown(ShutdownRequest)
|
||||
returns(ShutdownResponse);
|
||||
|
||||
/** Stop HBase Master only. Does not shutdown the cluster. */
|
||||
rpc stopMaster(StopMasterRequest)
|
||||
rpc StopMaster(StopMasterRequest)
|
||||
returns(StopMasterResponse);
|
||||
|
||||
/**
|
||||
|
@ -316,72 +316,72 @@ service MasterAdminService {
|
|||
* go ahead and do the reassignments. Can NOT run for various reasons.
|
||||
* Check logs.
|
||||
*/
|
||||
rpc balance(BalanceRequest)
|
||||
rpc Balance(BalanceRequest)
|
||||
returns(BalanceResponse);
|
||||
|
||||
/**
|
||||
* Turn the load balancer on or off.
|
||||
* If synchronous is true, it waits until current balance() call, if outstanding, to return.
|
||||
*/
|
||||
rpc setBalancerRunning(SetBalancerRunningRequest)
|
||||
rpc SetBalancerRunning(SetBalancerRunningRequest)
|
||||
returns(SetBalancerRunningResponse);
|
||||
|
||||
/** Get a run of the catalog janitor */
|
||||
rpc runCatalogScan(CatalogScanRequest)
|
||||
rpc RunCatalogScan(CatalogScanRequest)
|
||||
returns(CatalogScanResponse);
|
||||
|
||||
/**
|
||||
* Enable the catalog janitor on or off.
|
||||
*/
|
||||
rpc enableCatalogJanitor(EnableCatalogJanitorRequest)
|
||||
rpc EnableCatalogJanitor(EnableCatalogJanitorRequest)
|
||||
returns(EnableCatalogJanitorResponse);
|
||||
|
||||
/**
|
||||
* Query whether the catalog janitor is enabled.
|
||||
*/
|
||||
rpc isCatalogJanitorEnabled(IsCatalogJanitorEnabledRequest)
|
||||
rpc IsCatalogJanitorEnabled(IsCatalogJanitorEnabledRequest)
|
||||
returns(IsCatalogJanitorEnabledResponse);
|
||||
|
||||
/**
|
||||
* Call a master coprocessor endpoint
|
||||
*/
|
||||
rpc execMasterService(CoprocessorServiceRequest)
|
||||
rpc ExecMasterService(CoprocessorServiceRequest)
|
||||
returns(CoprocessorServiceResponse);
|
||||
|
||||
/**
|
||||
* Create a snapshot for the given table.
|
||||
* @param snapshot description of the snapshot to take
|
||||
*/
|
||||
rpc snapshot(TakeSnapshotRequest) returns(TakeSnapshotResponse);
|
||||
rpc Snapshot(TakeSnapshotRequest) returns(TakeSnapshotResponse);
|
||||
|
||||
/**
|
||||
* List completed snapshots.
|
||||
* Returns a list of snapshot descriptors for completed snapshots
|
||||
*/
|
||||
rpc getCompletedSnapshots(ListSnapshotRequest) returns(ListSnapshotResponse);
|
||||
rpc GetCompletedSnapshots(ListSnapshotRequest) returns(ListSnapshotResponse);
|
||||
|
||||
/**
|
||||
* Delete an existing snapshot. This method can also be used to clean up an aborted snapshot.
|
||||
* @param snapshotName snapshot to delete
|
||||
*/
|
||||
rpc deleteSnapshot(DeleteSnapshotRequest) returns(DeleteSnapshotResponse);
|
||||
rpc DeleteSnapshot(DeleteSnapshotRequest) returns(DeleteSnapshotResponse);
|
||||
|
||||
/**
|
||||
* Determine if the snapshot is done yet.
|
||||
*/
|
||||
rpc isSnapshotDone(IsSnapshotDoneRequest) returns(IsSnapshotDoneResponse);
|
||||
rpc IsSnapshotDone(IsSnapshotDoneRequest) returns(IsSnapshotDoneResponse);
|
||||
|
||||
/**
|
||||
* Restore a snapshot
|
||||
* @param snapshot description of the snapshot to restore
|
||||
*/
|
||||
rpc restoreSnapshot(RestoreSnapshotRequest) returns(RestoreSnapshotResponse);
|
||||
rpc RestoreSnapshot(RestoreSnapshotRequest) returns(RestoreSnapshotResponse);
|
||||
|
||||
/**
|
||||
* Determine if the snapshot restore is done yet.
|
||||
*/
|
||||
rpc isRestoreSnapshotDone(IsRestoreSnapshotDoneRequest) returns(IsRestoreSnapshotDoneResponse);
|
||||
rpc IsRestoreSnapshotDone(IsRestoreSnapshotDoneRequest) returns(IsRestoreSnapshotDoneResponse);
|
||||
|
||||
/** return true if master is available */
|
||||
rpc isMasterRunning(IsMasterRunningRequest) returns(IsMasterRunningResponse);
|
||||
rpc IsMasterRunning(IsMasterRunningRequest) returns(IsMasterRunningResponse);
|
||||
}
|
||||
|
|
|
@ -29,42 +29,42 @@ import "hbase.proto";
|
|||
import "ClusterStatus.proto";
|
||||
|
||||
message GetSchemaAlterStatusRequest {
|
||||
required bytes tableName = 1;
|
||||
required bytes table_name = 1;
|
||||
}
|
||||
|
||||
message GetSchemaAlterStatusResponse {
|
||||
optional uint32 yetToUpdateRegions = 1;
|
||||
optional uint32 totalRegions = 2;
|
||||
optional uint32 yet_to_update_regions = 1;
|
||||
optional uint32 total_regions = 2;
|
||||
}
|
||||
|
||||
message GetTableDescriptorsRequest {
|
||||
repeated string tableNames = 1;
|
||||
repeated string table_names = 1;
|
||||
}
|
||||
|
||||
message GetTableDescriptorsResponse {
|
||||
repeated TableSchema tableSchema = 1;
|
||||
repeated TableSchema table_schema = 1;
|
||||
}
|
||||
|
||||
message GetClusterStatusRequest {
|
||||
}
|
||||
|
||||
message GetClusterStatusResponse {
|
||||
required ClusterStatus clusterStatus = 1;
|
||||
required ClusterStatus cluster_status = 1;
|
||||
}
|
||||
|
||||
service MasterMonitorService {
|
||||
/** Used by the client to get the number of regions that have received the updated schema */
|
||||
rpc getSchemaAlterStatus(GetSchemaAlterStatusRequest)
|
||||
rpc GetSchemaAlterStatus(GetSchemaAlterStatusRequest)
|
||||
returns(GetSchemaAlterStatusResponse);
|
||||
|
||||
/** Get list of TableDescriptors for requested tables. */
|
||||
rpc getTableDescriptors(GetTableDescriptorsRequest)
|
||||
rpc GetTableDescriptors(GetTableDescriptorsRequest)
|
||||
returns(GetTableDescriptorsResponse);
|
||||
|
||||
/** Return cluster status. */
|
||||
rpc getClusterStatus(GetClusterStatusRequest)
|
||||
rpc GetClusterStatus(GetClusterStatusRequest)
|
||||
returns(GetClusterStatusResponse);
|
||||
|
||||
/** return true if master is available */
|
||||
rpc isMasterRunning(IsMasterRunningRequest) returns(IsMasterRunningResponse);
|
||||
rpc IsMasterRunning(IsMasterRunningRequest) returns(IsMasterRunningResponse);
|
||||
}
|
||||
|
|
|
@ -23,13 +23,13 @@ option java_generic_services = true;
|
|||
option optimize_for = SPEED;
|
||||
|
||||
message MultiMutateRequest {
|
||||
repeated MutationProto mutationRequest = 1;
|
||||
repeated MutationProto mutation_request = 1;
|
||||
}
|
||||
|
||||
message MultiMutateResponse {
|
||||
}
|
||||
|
||||
service MultiRowMutationService {
|
||||
rpc mutateRows(MultiMutateRequest)
|
||||
rpc MutateRows(MultiMutateRequest)
|
||||
returns(MultiMutateResponse);
|
||||
}
|
||||
|
|
|
@ -72,20 +72,20 @@ option optimize_for = SPEED;
|
|||
|
||||
// User Information proto. Included in ConnectionHeader on connection setup
|
||||
message UserInformation {
|
||||
required string effectiveUser = 1;
|
||||
optional string realUser = 2;
|
||||
required string effective_user = 1;
|
||||
optional string real_user = 2;
|
||||
}
|
||||
|
||||
// This is sent on connection setup after the connection preamble is sent.
|
||||
message ConnectionHeader {
|
||||
optional UserInformation userInfo = 1;
|
||||
optional string serviceName = 2;
|
||||
optional UserInformation user_info = 1;
|
||||
optional string service_name = 2;
|
||||
// Cell block codec we will use sending over optional cell blocks. Server throws exception
|
||||
// if cannot deal.
|
||||
optional string cellBlockCodecClass = 3 [default = "org.apache.hadoop.hbase.codec.KeyValueCodec"];
|
||||
optional string cell_block_codec_class = 3 [default = "org.apache.hadoop.hbase.codec.KeyValueCodec"];
|
||||
// Compressor we will use if cell block is compressed. Server will throw exception if not supported.
|
||||
// Class must implement hadoop's CompressionCodec Interface
|
||||
optional string cellBlockCompressorClass = 4;
|
||||
optional string cell_block_compressor_class = 4;
|
||||
}
|
||||
|
||||
// Optional Cell block Message. Included in client RequestHeader
|
||||
|
@ -98,34 +98,34 @@ message CellBlockMeta {
|
|||
// the server side exception to the RPC client.
|
||||
message ExceptionResponse {
|
||||
// Class name of the exception thrown from the server
|
||||
optional string exceptionClassName = 1;
|
||||
optional string exception_class_name = 1;
|
||||
// Exception stack trace from the server side
|
||||
optional string stackTrace = 2;
|
||||
optional string stack_trace = 2;
|
||||
// Optional hostname. Filled in for some exceptions such as region moved
|
||||
// where exception gives clue on where the region may have moved.
|
||||
optional string hostname = 3;
|
||||
optional int32 port = 4;
|
||||
// Set if we are NOT to retry on receipt of this exception
|
||||
optional bool doNotRetry = 5;
|
||||
optional bool do_not_retry = 5;
|
||||
}
|
||||
|
||||
// Header sent making a request.
|
||||
message RequestHeader {
|
||||
// Monotonically increasing callId to keep track of RPC requests and their response
|
||||
optional uint32 callId = 1;
|
||||
optional RPCTInfo traceInfo = 2;
|
||||
optional string methodName = 3;
|
||||
// Monotonically increasing call_id to keep track of RPC requests and their response
|
||||
optional uint32 call_id = 1;
|
||||
optional RPCTInfo trace_info = 2;
|
||||
optional string method_name = 3;
|
||||
// If true, then a pb Message param follows.
|
||||
optional bool requestParam = 4;
|
||||
optional bool request_param = 4;
|
||||
// If present, then an encoded data block follows.
|
||||
optional CellBlockMeta cellBlockMeta = 5;
|
||||
optional CellBlockMeta cell_block_meta = 5;
|
||||
// TODO: Have client specify priority
|
||||
}
|
||||
|
||||
message ResponseHeader {
|
||||
optional uint32 callId = 1;
|
||||
optional uint32 call_id = 1;
|
||||
// If present, then request threw an exception and no response message (else we presume one)
|
||||
optional ExceptionResponse exception = 2;
|
||||
// If present, then an encoded data block follows.
|
||||
optional CellBlockMeta cellBlockMeta = 3;
|
||||
optional CellBlockMeta cell_block_meta = 3;
|
||||
}
|
||||
|
|
|
@ -32,10 +32,10 @@ message RegionServerStartupRequest {
|
|||
required uint32 port = 1;
|
||||
|
||||
/** This servers' startcode */
|
||||
required uint64 serverStartCode = 2;
|
||||
required uint64 server_start_code = 2;
|
||||
|
||||
/** Current time of the region server in ms */
|
||||
required uint64 serverCurrentTime = 3;
|
||||
required uint64 server_current_time = 3;
|
||||
}
|
||||
|
||||
message RegionServerStartupResponse {
|
||||
|
@ -44,7 +44,7 @@ message RegionServerStartupResponse {
|
|||
* hbase rootdir, the hostname to use creating the RegionServer ServerName,
|
||||
* etc
|
||||
*/
|
||||
repeated NameStringPair mapEntries = 1;
|
||||
repeated NameStringPair map_entries = 1;
|
||||
}
|
||||
|
||||
message RegionServerReportRequest {
|
||||
|
@ -62,7 +62,7 @@ message ReportRSFatalErrorRequest {
|
|||
required ServerName server = 1;
|
||||
|
||||
/** informative text to expose in the master logs and UI */
|
||||
required string errorMessage = 2;
|
||||
required string error_message = 2;
|
||||
}
|
||||
|
||||
message ReportRSFatalErrorResponse {
|
||||
|
@ -70,33 +70,33 @@ message ReportRSFatalErrorResponse {
|
|||
|
||||
message GetLastFlushedSequenceIdRequest {
|
||||
/** region name */
|
||||
required bytes regionName = 1;
|
||||
required bytes region_name = 1;
|
||||
}
|
||||
|
||||
message GetLastFlushedSequenceIdResponse {
|
||||
/** the last HLog sequence id flushed from MemStore to HFile for the region */
|
||||
required uint64 lastFlushedSequenceId = 1;
|
||||
required uint64 last_flushed_sequence_id = 1;
|
||||
}
|
||||
|
||||
service RegionServerStatusService {
|
||||
/** Called when a region server first starts. */
|
||||
rpc regionServerStartup(RegionServerStartupRequest)
|
||||
rpc RegionServerStartup(RegionServerStartupRequest)
|
||||
returns(RegionServerStartupResponse);
|
||||
|
||||
/** Called to report the load the RegionServer is under. */
|
||||
rpc regionServerReport(RegionServerReportRequest)
|
||||
rpc RegionServerReport(RegionServerReportRequest)
|
||||
returns(RegionServerReportResponse);
|
||||
|
||||
/**
|
||||
* Called by a region server to report a fatal error that is causing it to
|
||||
* abort.
|
||||
*/
|
||||
rpc reportRSFatalError(ReportRSFatalErrorRequest)
|
||||
rpc ReportRSFatalError(ReportRSFatalErrorRequest)
|
||||
returns(ReportRSFatalErrorResponse);
|
||||
|
||||
/** Called to get the sequence id of the last MemStore entry flushed to an
|
||||
* HFile for a specified region. Used by the region server to speed up
|
||||
* log splitting. */
|
||||
rpc getLastFlushedSequenceId(GetLastFlushedSequenceIdRequest)
|
||||
rpc GetLastFlushedSequenceId(GetLastFlushedSequenceIdRequest)
|
||||
returns(GetLastFlushedSequenceIdResponse);
|
||||
}
|
||||
|
|
|
@ -27,15 +27,15 @@ option java_generate_equals_and_hash = true;
|
|||
option optimize_for = SPEED;
|
||||
|
||||
message RowProcessorRequest {
|
||||
required string rowProcessorClassName = 1;
|
||||
optional string rowProcessorInitializerMessageName = 2;
|
||||
optional bytes rowProcessorInitializerMessage = 3;
|
||||
required string row_processor_class_name = 1;
|
||||
optional string row_processor_initializer_message_name = 2;
|
||||
optional bytes row_processor_initializer_message = 3;
|
||||
}
|
||||
|
||||
message RowProcessorResult {
|
||||
required bytes rowProcessorResult = 1;
|
||||
required bytes row_processor_result = 1;
|
||||
}
|
||||
|
||||
service RowProcessorService {
|
||||
rpc process (RowProcessorRequest) returns (RowProcessorResult);
|
||||
}
|
||||
rpc Process (RowProcessorRequest) returns (RowProcessorResult);
|
||||
}
|
||||
|
|
|
@ -26,10 +26,10 @@ import 'hbase.proto';
|
|||
import 'Client.proto';
|
||||
|
||||
message SecureBulkLoadHFilesRequest {
|
||||
repeated BulkLoadHFileRequest.FamilyPath familyPath = 1;
|
||||
optional bool assignSeqNum = 2;
|
||||
required DelegationTokenProto fsToken = 3;
|
||||
required string bulkToken = 4;
|
||||
repeated BulkLoadHFileRequest.FamilyPath family_path = 1;
|
||||
optional bool assign_seq_num = 2;
|
||||
required DelegationTokenProto fs_token = 3;
|
||||
required string bulk_token = 4;
|
||||
}
|
||||
|
||||
message SecureBulkLoadHFilesResponse {
|
||||
|
@ -44,15 +44,15 @@ message DelegationTokenProto {
|
|||
}
|
||||
|
||||
message PrepareBulkLoadRequest {
|
||||
required bytes tableName = 1;
|
||||
required bytes table_name = 1;
|
||||
}
|
||||
|
||||
message PrepareBulkLoadResponse {
|
||||
required string bulkToken = 1;
|
||||
required string bulk_token = 1;
|
||||
}
|
||||
|
||||
message CleanupBulkLoadRequest {
|
||||
required string bulkToken = 1;
|
||||
required string bulk_token = 1;
|
||||
|
||||
}
|
||||
|
||||
|
@ -60,12 +60,12 @@ message CleanupBulkLoadResponse {
|
|||
}
|
||||
|
||||
service SecureBulkLoadService {
|
||||
rpc prepareBulkLoad(PrepareBulkLoadRequest)
|
||||
rpc PrepareBulkLoad(PrepareBulkLoadRequest)
|
||||
returns (PrepareBulkLoadResponse);
|
||||
|
||||
rpc secureBulkLoadHFiles(SecureBulkLoadHFilesRequest)
|
||||
rpc SecureBulkLoadHFiles(SecureBulkLoadHFilesRequest)
|
||||
returns (SecureBulkLoadHFilesResponse);
|
||||
|
||||
rpc cleanupBulkLoad(CleanupBulkLoadRequest)
|
||||
rpc CleanupBulkLoad(CleanupBulkLoadRequest)
|
||||
returns (CleanupBulkLoadResponse);
|
||||
}
|
||||
|
|
|
@ -26,6 +26,6 @@ option optimize_for = SPEED;
|
|||
//the id of the current span when this message was sent, so we know
|
||||
//what span caused the new span we will create when this message is received.
|
||||
message RPCTInfo {
|
||||
optional int64 traceId = 1;
|
||||
optional int64 parentId = 2;
|
||||
optional int64 trace_id = 1;
|
||||
optional int64 parent_id = 2;
|
||||
}
|
||||
|
|
|
@ -24,21 +24,21 @@ option optimize_for = SPEED;
|
|||
import "hbase.proto";
|
||||
|
||||
message WALHeader {
|
||||
optional bool hasCompression = 1;
|
||||
optional bool has_compression = 1;
|
||||
}
|
||||
|
||||
// Protocol buffer version of HLogKey; see HLogKey comment, not really a key but WALEdit header for some KVs
|
||||
message WALKey {
|
||||
required bytes encodedRegionName = 1;
|
||||
required bytes tableName = 2;
|
||||
required uint64 logSequenceNumber = 3;
|
||||
required uint64 writeTime = 4;
|
||||
optional UUID clusterId = 5;
|
||||
required bytes encoded_region_name = 1;
|
||||
required bytes table_name = 2;
|
||||
required uint64 log_sequence_number = 3;
|
||||
required uint64 write_time = 4;
|
||||
optional UUID cluster_id = 5;
|
||||
|
||||
repeated FamilyScope scopes = 6;
|
||||
optional uint32 followingKvCount = 7;
|
||||
optional uint32 following_kv_count = 7;
|
||||
/*
|
||||
optional CustomEntryType customEntryType = 8;
|
||||
optional CustomEntryType custom_entry_type = 8;
|
||||
|
||||
enum CustomEntryType {
|
||||
COMPACTION = 0;
|
||||
|
@ -53,7 +53,7 @@ enum ScopeType {
|
|||
|
||||
message FamilyScope {
|
||||
required bytes family = 1;
|
||||
required ScopeType scopeType = 2;
|
||||
required ScopeType scope_type = 2;
|
||||
}
|
||||
|
||||
/**
|
||||
|
@ -67,12 +67,12 @@ message FamilyScope {
|
|||
* the * compaction should we fail the WAL write.
|
||||
*/
|
||||
message CompactionDescriptor {
|
||||
required bytes tableName = 1; // TODO: WALKey already stores these, might remove
|
||||
required bytes encodedRegionName = 2;
|
||||
required bytes familyName = 3;
|
||||
repeated string compactionInput = 4;
|
||||
repeated string compactionOutput = 5;
|
||||
required string storeHomeDir = 6;
|
||||
required bytes table_name = 1; // TODO: WALKey already stores these, might remove
|
||||
required bytes encoded_region_name = 2;
|
||||
required bytes family_name = 3;
|
||||
repeated string compaction_input = 4;
|
||||
repeated string compaction_output = 5;
|
||||
required string store_home_dir = 6;
|
||||
}
|
||||
|
||||
/**
|
||||
|
|
|
@ -36,7 +36,7 @@ message MetaRegionServer {
|
|||
// The major version of the rpc the server speaks. This is used so that
|
||||
// clients connecting to the cluster can have prior knowledge of what version
|
||||
// to send to a RegionServer. AsyncHBase will use this to detect versions.
|
||||
optional uint32 rpcVersion = 2;
|
||||
optional uint32 rpc_version = 2;
|
||||
}
|
||||
|
||||
/**
|
||||
|
@ -46,7 +46,7 @@ message Master {
|
|||
// The ServerName of the current Master
|
||||
required ServerName master = 1;
|
||||
// Major RPC version so that clients can know what version the master can accept.
|
||||
optional uint32 rpcVersion = 2;
|
||||
optional uint32 rpc_version = 2;
|
||||
}
|
||||
|
||||
/**
|
||||
|
@ -54,8 +54,8 @@ message Master {
|
|||
*/
|
||||
message ClusterUp {
|
||||
// If this znode is present, cluster is up. Currently
|
||||
// the data is cluster startDate.
|
||||
required string startDate = 1;
|
||||
// the data is cluster start_date.
|
||||
required string start_date = 1;
|
||||
}
|
||||
|
||||
/**
|
||||
|
@ -64,12 +64,12 @@ message ClusterUp {
|
|||
*/
|
||||
message RegionTransition {
|
||||
// Code for EventType gotten by doing o.a.h.h.EventHandler.EventType.getCode()
|
||||
required uint32 eventTypeCode = 1;
|
||||
required uint32 event_type_code = 1;
|
||||
// Full regionname in bytes
|
||||
required bytes regionName = 2;
|
||||
required uint64 createTime = 3;
|
||||
required bytes region_name = 2;
|
||||
required uint64 create_time = 3;
|
||||
// The region server where the transition will happen or is happening
|
||||
required ServerName serverName = 4;
|
||||
required ServerName server_name = 4;
|
||||
optional bytes payload = 5;
|
||||
}
|
||||
|
||||
|
@ -86,7 +86,7 @@ message SplitLogTask {
|
|||
ERR = 4;
|
||||
}
|
||||
required State state = 1;
|
||||
required ServerName serverName = 2;
|
||||
required ServerName server_name = 2;
|
||||
}
|
||||
|
||||
/**
|
||||
|
@ -110,7 +110,7 @@ message Table {
|
|||
* Used by replication. Holds a replication peer key.
|
||||
*/
|
||||
message ReplicationPeer {
|
||||
// clusterKey is the concatenation of the slave cluster's
|
||||
// clusterkey is the concatenation of the slave cluster's
|
||||
// hbase.zookeeper.quorum:hbase.zookeeper.property.clientPort:zookeeper.znode.parent
|
||||
required string clusterkey = 1;
|
||||
}
|
||||
|
@ -137,27 +137,27 @@ message ReplicationHLogPosition {
|
|||
* Used by replication. Used to lock a region server during failover.
|
||||
*/
|
||||
message ReplicationLock {
|
||||
required string lockOwner = 1;
|
||||
required string lock_owner = 1;
|
||||
}
|
||||
|
||||
/**
|
||||
* Metadata associated with a table lock in zookeeper
|
||||
*/
|
||||
message TableLock {
|
||||
optional bytes tableName = 1;
|
||||
optional ServerName lockOwner = 2;
|
||||
optional int64 threadId = 3;
|
||||
optional bool isShared = 4;
|
||||
optional bytes table_name = 1;
|
||||
optional ServerName lock_owner = 2;
|
||||
optional int64 thread_id = 3;
|
||||
optional bool is_shared = 4;
|
||||
optional string purpose = 5;
|
||||
optional int64 createTime = 6;
|
||||
optional int64 create_time = 6;
|
||||
}
|
||||
|
||||
/**
|
||||
* sequence Id of a store
|
||||
*/
|
||||
message StoreSequenceId {
|
||||
required bytes familyName = 1;
|
||||
required uint64 sequenceId = 2;
|
||||
required bytes family_name = 1;
|
||||
required uint64 sequence_id = 2;
|
||||
}
|
||||
|
||||
/**
|
||||
|
@ -165,6 +165,6 @@ message StoreSequenceId {
|
|||
* list sequence ids of the region's stores
|
||||
*/
|
||||
message RegionStoreSequenceIds {
|
||||
required uint64 lastFlushedSequenceId = 1;
|
||||
repeated StoreSequenceId storeSequenceId = 2;
|
||||
required uint64 last_flushed_sequence_id = 1;
|
||||
repeated StoreSequenceId store_sequence_id = 2;
|
||||
}
|
||||
|
|
|
@ -32,7 +32,7 @@ import "Cell.proto";
|
|||
message TableSchema {
|
||||
optional bytes name = 1;
|
||||
repeated BytesBytesPair attributes = 2;
|
||||
repeated ColumnFamilySchema columnFamilies = 3;
|
||||
repeated ColumnFamilySchema column_families = 3;
|
||||
repeated NameStringPair configuration = 4;
|
||||
}
|
||||
|
||||
|
@ -50,10 +50,10 @@ message ColumnFamilySchema {
|
|||
* Protocol buffer version of HRegionInfo.
|
||||
*/
|
||||
message RegionInfo {
|
||||
required uint64 regionId = 1;
|
||||
required bytes tableName = 2;
|
||||
optional bytes startKey = 3;
|
||||
optional bytes endKey = 4;
|
||||
required uint64 region_id = 1;
|
||||
required bytes table_name = 2;
|
||||
optional bytes start_key = 3;
|
||||
optional bytes end_key = 4;
|
||||
optional bool offline = 5;
|
||||
optional bool split = 6;
|
||||
optional bool recovering = 7;
|
||||
|
@ -63,7 +63,7 @@ message RegionInfo {
|
|||
* Protocol buffer for favored nodes
|
||||
*/
|
||||
message FavoredNodes {
|
||||
repeated ServerName favoredNode = 1;
|
||||
repeated ServerName favored_node = 1;
|
||||
}
|
||||
|
||||
/**
|
||||
|
@ -111,9 +111,9 @@ enum CompareType {
|
|||
* Protocol buffer version of ServerName
|
||||
*/
|
||||
message ServerName {
|
||||
required string hostName = 1;
|
||||
required string host_name = 1;
|
||||
optional uint32 port = 2;
|
||||
optional uint64 startCode = 3;
|
||||
optional uint64 start_code = 3;
|
||||
}
|
||||
|
||||
// Comment data structures
|
||||
|
@ -148,7 +148,7 @@ message NameInt64Pair {
|
|||
message SnapshotDescription {
|
||||
required string name = 1;
|
||||
optional string table = 2; // not needed for delete, but checked for in taking snapshot
|
||||
optional int64 creationTime = 3 [default = 0];
|
||||
optional int64 creation_time = 3 [default = 0];
|
||||
enum Type {
|
||||
DISABLED = 0;
|
||||
FLUSH = 1;
|
||||
|
@ -161,15 +161,15 @@ message EmptyMsg {
|
|||
}
|
||||
|
||||
message LongMsg {
|
||||
required int64 longMsg = 1;
|
||||
required int64 long_msg = 1;
|
||||
|
||||
}
|
||||
|
||||
message BigDecimalMsg {
|
||||
required bytes bigdecimalMsg = 1;
|
||||
required bytes bigdecimal_msg = 1;
|
||||
}
|
||||
|
||||
message UUID {
|
||||
required uint64 leastSigBits = 1;
|
||||
required uint64 mostSigBits = 2;
|
||||
required uint64 least_sig_bits = 1;
|
||||
required uint64 most_sig_bits = 2;
|
||||
}
|
||||
|
|
|
@ -133,7 +133,7 @@ public class TokenProvider implements AuthenticationProtos.AuthenticationService
|
|||
}
|
||||
|
||||
@Override
|
||||
public void whoami(RpcController controller, AuthenticationProtos.WhoAmIRequest request,
|
||||
public void whoAmI(RpcController controller, AuthenticationProtos.WhoAmIRequest request,
|
||||
RpcCallback<AuthenticationProtos.WhoAmIResponse> done) {
|
||||
User requestUser = RequestContext.getRequestUser();
|
||||
AuthenticationProtos.WhoAmIResponse.Builder response =
|
||||
|
|
|
@ -274,15 +274,15 @@ public class TestTokenAuthentication {
|
|||
}
|
||||
|
||||
@Override
|
||||
public AuthenticationProtos.WhoAmIResponse whoami(
|
||||
public AuthenticationProtos.WhoAmIResponse whoAmI(
|
||||
RpcController controller, AuthenticationProtos.WhoAmIRequest request)
|
||||
throws ServiceException {
|
||||
LOG.debug("whoami() request from "+RequestContext.getRequestUserName());
|
||||
LOG.debug("whoAmI() request from "+RequestContext.getRequestUserName());
|
||||
// ignore passed in controller -- it's always null
|
||||
ServerRpcController serverController = new ServerRpcController();
|
||||
BlockingRpcCallback<AuthenticationProtos.WhoAmIResponse> callback =
|
||||
new BlockingRpcCallback<AuthenticationProtos.WhoAmIResponse>();
|
||||
whoami(serverController, request, callback);
|
||||
whoAmI(serverController, request, callback);
|
||||
try {
|
||||
serverController.checkFailed();
|
||||
return callback.get();
|
||||
|
@ -383,7 +383,7 @@ public class TestTokenAuthentication {
|
|||
AuthenticationProtos.AuthenticationService.BlockingInterface stub =
|
||||
AuthenticationProtos.AuthenticationService.newBlockingStub(channel);
|
||||
AuthenticationProtos.WhoAmIResponse response =
|
||||
stub.whoami(null, AuthenticationProtos.WhoAmIRequest.getDefaultInstance());
|
||||
stub.whoAmI(null, AuthenticationProtos.WhoAmIRequest.getDefaultInstance());
|
||||
String myname = response.getUsername();
|
||||
assertEquals("testuser", myname);
|
||||
String authMethod = response.getAuthMethod();
|
||||
|
|
Loading…
Reference in New Issue