HBASE-8915 protobuf message style
git-svn-id: https://svn.apache.org/repos/asf/hbase/trunk@1501834 13f79535-47bb-0310-9956-ffa450edef68
This commit is contained in:
parent
8bf99de619
commit
457c928d35
|
@ -6422,11 +6422,11 @@ public final class AccessControlProtos {
|
||||||
"Permission\":\n\027CheckPermissionsRequest\022\037\n" +
|
"Permission\":\n\027CheckPermissionsRequest\022\037\n" +
|
||||||
"\npermission\030\001 \003(\0132\013.Permission\"\032\n\030CheckP" +
|
"\npermission\030\001 \003(\0132\013.Permission\"\032\n\030CheckP" +
|
||||||
"ermissionsResponse2\373\001\n\024AccessControlServ" +
|
"ermissionsResponse2\373\001\n\024AccessControlServ" +
|
||||||
"ice\022&\n\005grant\022\r.GrantRequest\032\016.GrantRespo",
|
"ice\022&\n\005Grant\022\r.GrantRequest\032\016.GrantRespo",
|
||||||
"nse\022)\n\006revoke\022\016.RevokeRequest\032\017.RevokeRe" +
|
"nse\022)\n\006Revoke\022\016.RevokeRequest\032\017.RevokeRe" +
|
||||||
"sponse\022G\n\022getUserPermissions\022\027.UserPermi" +
|
"sponse\022G\n\022GetUserPermissions\022\027.UserPermi" +
|
||||||
"ssionsRequest\032\030.UserPermissionsResponse\022" +
|
"ssionsRequest\032\030.UserPermissionsResponse\022" +
|
||||||
"G\n\020checkPermissions\022\030.CheckPermissionsRe" +
|
"G\n\020CheckPermissions\022\030.CheckPermissionsRe" +
|
||||||
"quest\032\031.CheckPermissionsResponseBI\n*org." +
|
"quest\032\031.CheckPermissionsResponseBI\n*org." +
|
||||||
"apache.hadoop.hbase.protobuf.generatedB\023" +
|
"apache.hadoop.hbase.protobuf.generatedB\023" +
|
||||||
"AccessControlProtosH\001\210\001\001\240\001\001"
|
"AccessControlProtosH\001\210\001\001\240\001\001"
|
||||||
|
|
|
@ -16,7 +16,7 @@ public final class AdminProtos {
|
||||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier getRegion();
|
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier getRegion();
|
||||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder getRegionOrBuilder();
|
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder getRegionOrBuilder();
|
||||||
|
|
||||||
// optional bool compactionState = 2;
|
// optional bool compaction_state = 2;
|
||||||
boolean hasCompactionState();
|
boolean hasCompactionState();
|
||||||
boolean getCompactionState();
|
boolean getCompactionState();
|
||||||
}
|
}
|
||||||
|
@ -62,8 +62,8 @@ public final class AdminProtos {
|
||||||
return region_;
|
return region_;
|
||||||
}
|
}
|
||||||
|
|
||||||
// optional bool compactionState = 2;
|
// optional bool compaction_state = 2;
|
||||||
public static final int COMPACTIONSTATE_FIELD_NUMBER = 2;
|
public static final int COMPACTION_STATE_FIELD_NUMBER = 2;
|
||||||
private boolean compactionState_;
|
private boolean compactionState_;
|
||||||
public boolean hasCompactionState() {
|
public boolean hasCompactionState() {
|
||||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||||
|
@ -166,7 +166,7 @@ public final class AdminProtos {
|
||||||
hash = (53 * hash) + getRegion().hashCode();
|
hash = (53 * hash) + getRegion().hashCode();
|
||||||
}
|
}
|
||||||
if (hasCompactionState()) {
|
if (hasCompactionState()) {
|
||||||
hash = (37 * hash) + COMPACTIONSTATE_FIELD_NUMBER;
|
hash = (37 * hash) + COMPACTION_STATE_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + hashBoolean(getCompactionState());
|
hash = (53 * hash) + hashBoolean(getCompactionState());
|
||||||
}
|
}
|
||||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||||
|
@ -515,7 +515,7 @@ public final class AdminProtos {
|
||||||
return regionBuilder_;
|
return regionBuilder_;
|
||||||
}
|
}
|
||||||
|
|
||||||
// optional bool compactionState = 2;
|
// optional bool compaction_state = 2;
|
||||||
private boolean compactionState_ ;
|
private boolean compactionState_ ;
|
||||||
public boolean hasCompactionState() {
|
public boolean hasCompactionState() {
|
||||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||||
|
@ -550,12 +550,12 @@ public final class AdminProtos {
|
||||||
public interface GetRegionInfoResponseOrBuilder
|
public interface GetRegionInfoResponseOrBuilder
|
||||||
extends com.google.protobuf.MessageOrBuilder {
|
extends com.google.protobuf.MessageOrBuilder {
|
||||||
|
|
||||||
// required .RegionInfo regionInfo = 1;
|
// required .RegionInfo region_info = 1;
|
||||||
boolean hasRegionInfo();
|
boolean hasRegionInfo();
|
||||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfo getRegionInfo();
|
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfo getRegionInfo();
|
||||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfoOrBuilder getRegionInfoOrBuilder();
|
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfoOrBuilder getRegionInfoOrBuilder();
|
||||||
|
|
||||||
// optional .GetRegionInfoResponse.CompactionState compactionState = 2;
|
// optional .GetRegionInfoResponse.CompactionState compaction_state = 2;
|
||||||
boolean hasCompactionState();
|
boolean hasCompactionState();
|
||||||
org.apache.hadoop.hbase.protobuf.generated.AdminProtos.GetRegionInfoResponse.CompactionState getCompactionState();
|
org.apache.hadoop.hbase.protobuf.generated.AdminProtos.GetRegionInfoResponse.CompactionState getCompactionState();
|
||||||
}
|
}
|
||||||
|
@ -663,8 +663,8 @@ public final class AdminProtos {
|
||||||
}
|
}
|
||||||
|
|
||||||
private int bitField0_;
|
private int bitField0_;
|
||||||
// required .RegionInfo regionInfo = 1;
|
// required .RegionInfo region_info = 1;
|
||||||
public static final int REGIONINFO_FIELD_NUMBER = 1;
|
public static final int REGION_INFO_FIELD_NUMBER = 1;
|
||||||
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfo regionInfo_;
|
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfo regionInfo_;
|
||||||
public boolean hasRegionInfo() {
|
public boolean hasRegionInfo() {
|
||||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||||
|
@ -676,8 +676,8 @@ public final class AdminProtos {
|
||||||
return regionInfo_;
|
return regionInfo_;
|
||||||
}
|
}
|
||||||
|
|
||||||
// optional .GetRegionInfoResponse.CompactionState compactionState = 2;
|
// optional .GetRegionInfoResponse.CompactionState compaction_state = 2;
|
||||||
public static final int COMPACTIONSTATE_FIELD_NUMBER = 2;
|
public static final int COMPACTION_STATE_FIELD_NUMBER = 2;
|
||||||
private org.apache.hadoop.hbase.protobuf.generated.AdminProtos.GetRegionInfoResponse.CompactionState compactionState_;
|
private org.apache.hadoop.hbase.protobuf.generated.AdminProtos.GetRegionInfoResponse.CompactionState compactionState_;
|
||||||
public boolean hasCompactionState() {
|
public boolean hasCompactionState() {
|
||||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||||
|
@ -776,11 +776,11 @@ public final class AdminProtos {
|
||||||
int hash = 41;
|
int hash = 41;
|
||||||
hash = (19 * hash) + getDescriptorForType().hashCode();
|
hash = (19 * hash) + getDescriptorForType().hashCode();
|
||||||
if (hasRegionInfo()) {
|
if (hasRegionInfo()) {
|
||||||
hash = (37 * hash) + REGIONINFO_FIELD_NUMBER;
|
hash = (37 * hash) + REGION_INFO_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + getRegionInfo().hashCode();
|
hash = (53 * hash) + getRegionInfo().hashCode();
|
||||||
}
|
}
|
||||||
if (hasCompactionState()) {
|
if (hasCompactionState()) {
|
||||||
hash = (37 * hash) + COMPACTIONSTATE_FIELD_NUMBER;
|
hash = (37 * hash) + COMPACTION_STATE_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + hashEnum(getCompactionState());
|
hash = (53 * hash) + hashEnum(getCompactionState());
|
||||||
}
|
}
|
||||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||||
|
@ -1045,7 +1045,7 @@ public final class AdminProtos {
|
||||||
|
|
||||||
private int bitField0_;
|
private int bitField0_;
|
||||||
|
|
||||||
// required .RegionInfo regionInfo = 1;
|
// required .RegionInfo region_info = 1;
|
||||||
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfo regionInfo_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfo.getDefaultInstance();
|
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfo regionInfo_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfo.getDefaultInstance();
|
||||||
private com.google.protobuf.SingleFieldBuilder<
|
private com.google.protobuf.SingleFieldBuilder<
|
||||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfo, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfo.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfoOrBuilder> regionInfoBuilder_;
|
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfo, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfo.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfoOrBuilder> regionInfoBuilder_;
|
||||||
|
@ -1135,7 +1135,7 @@ public final class AdminProtos {
|
||||||
return regionInfoBuilder_;
|
return regionInfoBuilder_;
|
||||||
}
|
}
|
||||||
|
|
||||||
// optional .GetRegionInfoResponse.CompactionState compactionState = 2;
|
// optional .GetRegionInfoResponse.CompactionState compaction_state = 2;
|
||||||
private org.apache.hadoop.hbase.protobuf.generated.AdminProtos.GetRegionInfoResponse.CompactionState compactionState_ = org.apache.hadoop.hbase.protobuf.generated.AdminProtos.GetRegionInfoResponse.CompactionState.NONE;
|
private org.apache.hadoop.hbase.protobuf.generated.AdminProtos.GetRegionInfoResponse.CompactionState compactionState_ = org.apache.hadoop.hbase.protobuf.generated.AdminProtos.GetRegionInfoResponse.CompactionState.NONE;
|
||||||
public boolean hasCompactionState() {
|
public boolean hasCompactionState() {
|
||||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||||
|
@ -1757,7 +1757,7 @@ public final class AdminProtos {
|
||||||
public interface GetStoreFileResponseOrBuilder
|
public interface GetStoreFileResponseOrBuilder
|
||||||
extends com.google.protobuf.MessageOrBuilder {
|
extends com.google.protobuf.MessageOrBuilder {
|
||||||
|
|
||||||
// repeated string storeFile = 1;
|
// repeated string store_file = 1;
|
||||||
java.util.List<String> getStoreFileList();
|
java.util.List<String> getStoreFileList();
|
||||||
int getStoreFileCount();
|
int getStoreFileCount();
|
||||||
String getStoreFile(int index);
|
String getStoreFile(int index);
|
||||||
|
@ -1790,8 +1790,8 @@ public final class AdminProtos {
|
||||||
return org.apache.hadoop.hbase.protobuf.generated.AdminProtos.internal_static_GetStoreFileResponse_fieldAccessorTable;
|
return org.apache.hadoop.hbase.protobuf.generated.AdminProtos.internal_static_GetStoreFileResponse_fieldAccessorTable;
|
||||||
}
|
}
|
||||||
|
|
||||||
// repeated string storeFile = 1;
|
// repeated string store_file = 1;
|
||||||
public static final int STOREFILE_FIELD_NUMBER = 1;
|
public static final int STORE_FILE_FIELD_NUMBER = 1;
|
||||||
private com.google.protobuf.LazyStringList storeFile_;
|
private com.google.protobuf.LazyStringList storeFile_;
|
||||||
public java.util.List<String>
|
public java.util.List<String>
|
||||||
getStoreFileList() {
|
getStoreFileList() {
|
||||||
|
@ -1875,7 +1875,7 @@ public final class AdminProtos {
|
||||||
int hash = 41;
|
int hash = 41;
|
||||||
hash = (19 * hash) + getDescriptorForType().hashCode();
|
hash = (19 * hash) + getDescriptorForType().hashCode();
|
||||||
if (getStoreFileCount() > 0) {
|
if (getStoreFileCount() > 0) {
|
||||||
hash = (37 * hash) + STOREFILE_FIELD_NUMBER;
|
hash = (37 * hash) + STORE_FILE_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + getStoreFileList().hashCode();
|
hash = (53 * hash) + getStoreFileList().hashCode();
|
||||||
}
|
}
|
||||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||||
|
@ -2106,7 +2106,7 @@ public final class AdminProtos {
|
||||||
|
|
||||||
private int bitField0_;
|
private int bitField0_;
|
||||||
|
|
||||||
// repeated string storeFile = 1;
|
// repeated string store_file = 1;
|
||||||
private com.google.protobuf.LazyStringList storeFile_ = com.google.protobuf.LazyStringArrayList.EMPTY;
|
private com.google.protobuf.LazyStringList storeFile_ = com.google.protobuf.LazyStringArrayList.EMPTY;
|
||||||
private void ensureStoreFileIsMutable() {
|
private void ensureStoreFileIsMutable() {
|
||||||
if (!((bitField0_ & 0x00000001) == 0x00000001)) {
|
if (!((bitField0_ & 0x00000001) == 0x00000001)) {
|
||||||
|
@ -2476,7 +2476,7 @@ public final class AdminProtos {
|
||||||
public interface GetOnlineRegionResponseOrBuilder
|
public interface GetOnlineRegionResponseOrBuilder
|
||||||
extends com.google.protobuf.MessageOrBuilder {
|
extends com.google.protobuf.MessageOrBuilder {
|
||||||
|
|
||||||
// repeated .RegionInfo regionInfo = 1;
|
// repeated .RegionInfo region_info = 1;
|
||||||
java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfo>
|
java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfo>
|
||||||
getRegionInfoList();
|
getRegionInfoList();
|
||||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfo getRegionInfo(int index);
|
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfo getRegionInfo(int index);
|
||||||
|
@ -2514,8 +2514,8 @@ public final class AdminProtos {
|
||||||
return org.apache.hadoop.hbase.protobuf.generated.AdminProtos.internal_static_GetOnlineRegionResponse_fieldAccessorTable;
|
return org.apache.hadoop.hbase.protobuf.generated.AdminProtos.internal_static_GetOnlineRegionResponse_fieldAccessorTable;
|
||||||
}
|
}
|
||||||
|
|
||||||
// repeated .RegionInfo regionInfo = 1;
|
// repeated .RegionInfo region_info = 1;
|
||||||
public static final int REGIONINFO_FIELD_NUMBER = 1;
|
public static final int REGION_INFO_FIELD_NUMBER = 1;
|
||||||
private java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfo> regionInfo_;
|
private java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfo> regionInfo_;
|
||||||
public java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfo> getRegionInfoList() {
|
public java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfo> getRegionInfoList() {
|
||||||
return regionInfo_;
|
return regionInfo_;
|
||||||
|
@ -2607,7 +2607,7 @@ public final class AdminProtos {
|
||||||
int hash = 41;
|
int hash = 41;
|
||||||
hash = (19 * hash) + getDescriptorForType().hashCode();
|
hash = (19 * hash) + getDescriptorForType().hashCode();
|
||||||
if (getRegionInfoCount() > 0) {
|
if (getRegionInfoCount() > 0) {
|
||||||
hash = (37 * hash) + REGIONINFO_FIELD_NUMBER;
|
hash = (37 * hash) + REGION_INFO_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + getRegionInfoList().hashCode();
|
hash = (53 * hash) + getRegionInfoList().hashCode();
|
||||||
}
|
}
|
||||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||||
|
@ -2869,7 +2869,7 @@ public final class AdminProtos {
|
||||||
|
|
||||||
private int bitField0_;
|
private int bitField0_;
|
||||||
|
|
||||||
// repeated .RegionInfo regionInfo = 1;
|
// repeated .RegionInfo region_info = 1;
|
||||||
private java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfo> regionInfo_ =
|
private java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfo> regionInfo_ =
|
||||||
java.util.Collections.emptyList();
|
java.util.Collections.emptyList();
|
||||||
private void ensureRegionInfoIsMutable() {
|
private void ensureRegionInfoIsMutable() {
|
||||||
|
@ -3069,7 +3069,7 @@ public final class AdminProtos {
|
||||||
public interface OpenRegionRequestOrBuilder
|
public interface OpenRegionRequestOrBuilder
|
||||||
extends com.google.protobuf.MessageOrBuilder {
|
extends com.google.protobuf.MessageOrBuilder {
|
||||||
|
|
||||||
// repeated .OpenRegionRequest.RegionOpenInfo openInfo = 1;
|
// repeated .OpenRegionRequest.RegionOpenInfo open_info = 1;
|
||||||
java.util.List<org.apache.hadoop.hbase.protobuf.generated.AdminProtos.OpenRegionRequest.RegionOpenInfo>
|
java.util.List<org.apache.hadoop.hbase.protobuf.generated.AdminProtos.OpenRegionRequest.RegionOpenInfo>
|
||||||
getOpenInfoList();
|
getOpenInfoList();
|
||||||
org.apache.hadoop.hbase.protobuf.generated.AdminProtos.OpenRegionRequest.RegionOpenInfo getOpenInfo(int index);
|
org.apache.hadoop.hbase.protobuf.generated.AdminProtos.OpenRegionRequest.RegionOpenInfo getOpenInfo(int index);
|
||||||
|
@ -3115,11 +3115,11 @@ public final class AdminProtos {
|
||||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfo getRegion();
|
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfo getRegion();
|
||||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfoOrBuilder getRegionOrBuilder();
|
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfoOrBuilder getRegionOrBuilder();
|
||||||
|
|
||||||
// optional uint32 versionOfOfflineNode = 2;
|
// optional uint32 version_of_offline_node = 2;
|
||||||
boolean hasVersionOfOfflineNode();
|
boolean hasVersionOfOfflineNode();
|
||||||
int getVersionOfOfflineNode();
|
int getVersionOfOfflineNode();
|
||||||
|
|
||||||
// repeated .ServerName favoredNodes = 3;
|
// repeated .ServerName favored_nodes = 3;
|
||||||
java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName>
|
java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName>
|
||||||
getFavoredNodesList();
|
getFavoredNodesList();
|
||||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName getFavoredNodes(int index);
|
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName getFavoredNodes(int index);
|
||||||
|
@ -3171,8 +3171,8 @@ public final class AdminProtos {
|
||||||
return region_;
|
return region_;
|
||||||
}
|
}
|
||||||
|
|
||||||
// optional uint32 versionOfOfflineNode = 2;
|
// optional uint32 version_of_offline_node = 2;
|
||||||
public static final int VERSIONOFOFFLINENODE_FIELD_NUMBER = 2;
|
public static final int VERSION_OF_OFFLINE_NODE_FIELD_NUMBER = 2;
|
||||||
private int versionOfOfflineNode_;
|
private int versionOfOfflineNode_;
|
||||||
public boolean hasVersionOfOfflineNode() {
|
public boolean hasVersionOfOfflineNode() {
|
||||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||||
|
@ -3181,8 +3181,8 @@ public final class AdminProtos {
|
||||||
return versionOfOfflineNode_;
|
return versionOfOfflineNode_;
|
||||||
}
|
}
|
||||||
|
|
||||||
// repeated .ServerName favoredNodes = 3;
|
// repeated .ServerName favored_nodes = 3;
|
||||||
public static final int FAVOREDNODES_FIELD_NUMBER = 3;
|
public static final int FAVORED_NODES_FIELD_NUMBER = 3;
|
||||||
private java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName> favoredNodes_;
|
private java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName> favoredNodes_;
|
||||||
public java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName> getFavoredNodesList() {
|
public java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName> getFavoredNodesList() {
|
||||||
return favoredNodes_;
|
return favoredNodes_;
|
||||||
|
@ -3312,11 +3312,11 @@ public final class AdminProtos {
|
||||||
hash = (53 * hash) + getRegion().hashCode();
|
hash = (53 * hash) + getRegion().hashCode();
|
||||||
}
|
}
|
||||||
if (hasVersionOfOfflineNode()) {
|
if (hasVersionOfOfflineNode()) {
|
||||||
hash = (37 * hash) + VERSIONOFOFFLINENODE_FIELD_NUMBER;
|
hash = (37 * hash) + VERSION_OF_OFFLINE_NODE_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + getVersionOfOfflineNode();
|
hash = (53 * hash) + getVersionOfOfflineNode();
|
||||||
}
|
}
|
||||||
if (getFavoredNodesCount() > 0) {
|
if (getFavoredNodesCount() > 0) {
|
||||||
hash = (37 * hash) + FAVOREDNODES_FIELD_NUMBER;
|
hash = (37 * hash) + FAVORED_NODES_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + getFavoredNodesList().hashCode();
|
hash = (53 * hash) + getFavoredNodesList().hashCode();
|
||||||
}
|
}
|
||||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||||
|
@ -3719,7 +3719,7 @@ public final class AdminProtos {
|
||||||
return regionBuilder_;
|
return regionBuilder_;
|
||||||
}
|
}
|
||||||
|
|
||||||
// optional uint32 versionOfOfflineNode = 2;
|
// optional uint32 version_of_offline_node = 2;
|
||||||
private int versionOfOfflineNode_ ;
|
private int versionOfOfflineNode_ ;
|
||||||
public boolean hasVersionOfOfflineNode() {
|
public boolean hasVersionOfOfflineNode() {
|
||||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||||
|
@ -3740,7 +3740,7 @@ public final class AdminProtos {
|
||||||
return this;
|
return this;
|
||||||
}
|
}
|
||||||
|
|
||||||
// repeated .ServerName favoredNodes = 3;
|
// repeated .ServerName favored_nodes = 3;
|
||||||
private java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName> favoredNodes_ =
|
private java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName> favoredNodes_ =
|
||||||
java.util.Collections.emptyList();
|
java.util.Collections.emptyList();
|
||||||
private void ensureFavoredNodesIsMutable() {
|
private void ensureFavoredNodesIsMutable() {
|
||||||
|
@ -3937,8 +3937,8 @@ public final class AdminProtos {
|
||||||
// @@protoc_insertion_point(class_scope:OpenRegionRequest.RegionOpenInfo)
|
// @@protoc_insertion_point(class_scope:OpenRegionRequest.RegionOpenInfo)
|
||||||
}
|
}
|
||||||
|
|
||||||
// repeated .OpenRegionRequest.RegionOpenInfo openInfo = 1;
|
// repeated .OpenRegionRequest.RegionOpenInfo open_info = 1;
|
||||||
public static final int OPENINFO_FIELD_NUMBER = 1;
|
public static final int OPEN_INFO_FIELD_NUMBER = 1;
|
||||||
private java.util.List<org.apache.hadoop.hbase.protobuf.generated.AdminProtos.OpenRegionRequest.RegionOpenInfo> openInfo_;
|
private java.util.List<org.apache.hadoop.hbase.protobuf.generated.AdminProtos.OpenRegionRequest.RegionOpenInfo> openInfo_;
|
||||||
public java.util.List<org.apache.hadoop.hbase.protobuf.generated.AdminProtos.OpenRegionRequest.RegionOpenInfo> getOpenInfoList() {
|
public java.util.List<org.apache.hadoop.hbase.protobuf.generated.AdminProtos.OpenRegionRequest.RegionOpenInfo> getOpenInfoList() {
|
||||||
return openInfo_;
|
return openInfo_;
|
||||||
|
@ -4030,7 +4030,7 @@ public final class AdminProtos {
|
||||||
int hash = 41;
|
int hash = 41;
|
||||||
hash = (19 * hash) + getDescriptorForType().hashCode();
|
hash = (19 * hash) + getDescriptorForType().hashCode();
|
||||||
if (getOpenInfoCount() > 0) {
|
if (getOpenInfoCount() > 0) {
|
||||||
hash = (37 * hash) + OPENINFO_FIELD_NUMBER;
|
hash = (37 * hash) + OPEN_INFO_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + getOpenInfoList().hashCode();
|
hash = (53 * hash) + getOpenInfoList().hashCode();
|
||||||
}
|
}
|
||||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||||
|
@ -4292,7 +4292,7 @@ public final class AdminProtos {
|
||||||
|
|
||||||
private int bitField0_;
|
private int bitField0_;
|
||||||
|
|
||||||
// repeated .OpenRegionRequest.RegionOpenInfo openInfo = 1;
|
// repeated .OpenRegionRequest.RegionOpenInfo open_info = 1;
|
||||||
private java.util.List<org.apache.hadoop.hbase.protobuf.generated.AdminProtos.OpenRegionRequest.RegionOpenInfo> openInfo_ =
|
private java.util.List<org.apache.hadoop.hbase.protobuf.generated.AdminProtos.OpenRegionRequest.RegionOpenInfo> openInfo_ =
|
||||||
java.util.Collections.emptyList();
|
java.util.Collections.emptyList();
|
||||||
private void ensureOpenInfoIsMutable() {
|
private void ensureOpenInfoIsMutable() {
|
||||||
|
@ -4492,7 +4492,7 @@ public final class AdminProtos {
|
||||||
public interface OpenRegionResponseOrBuilder
|
public interface OpenRegionResponseOrBuilder
|
||||||
extends com.google.protobuf.MessageOrBuilder {
|
extends com.google.protobuf.MessageOrBuilder {
|
||||||
|
|
||||||
// repeated .OpenRegionResponse.RegionOpeningState openingState = 1;
|
// repeated .OpenRegionResponse.RegionOpeningState opening_state = 1;
|
||||||
java.util.List<org.apache.hadoop.hbase.protobuf.generated.AdminProtos.OpenRegionResponse.RegionOpeningState> getOpeningStateList();
|
java.util.List<org.apache.hadoop.hbase.protobuf.generated.AdminProtos.OpenRegionResponse.RegionOpeningState> getOpeningStateList();
|
||||||
int getOpeningStateCount();
|
int getOpeningStateCount();
|
||||||
org.apache.hadoop.hbase.protobuf.generated.AdminProtos.OpenRegionResponse.RegionOpeningState getOpeningState(int index);
|
org.apache.hadoop.hbase.protobuf.generated.AdminProtos.OpenRegionResponse.RegionOpeningState getOpeningState(int index);
|
||||||
|
@ -4597,8 +4597,8 @@ public final class AdminProtos {
|
||||||
// @@protoc_insertion_point(enum_scope:OpenRegionResponse.RegionOpeningState)
|
// @@protoc_insertion_point(enum_scope:OpenRegionResponse.RegionOpeningState)
|
||||||
}
|
}
|
||||||
|
|
||||||
// repeated .OpenRegionResponse.RegionOpeningState openingState = 1;
|
// repeated .OpenRegionResponse.RegionOpeningState opening_state = 1;
|
||||||
public static final int OPENINGSTATE_FIELD_NUMBER = 1;
|
public static final int OPENING_STATE_FIELD_NUMBER = 1;
|
||||||
private java.util.List<org.apache.hadoop.hbase.protobuf.generated.AdminProtos.OpenRegionResponse.RegionOpeningState> openingState_;
|
private java.util.List<org.apache.hadoop.hbase.protobuf.generated.AdminProtos.OpenRegionResponse.RegionOpeningState> openingState_;
|
||||||
public java.util.List<org.apache.hadoop.hbase.protobuf.generated.AdminProtos.OpenRegionResponse.RegionOpeningState> getOpeningStateList() {
|
public java.util.List<org.apache.hadoop.hbase.protobuf.generated.AdminProtos.OpenRegionResponse.RegionOpeningState> getOpeningStateList() {
|
||||||
return openingState_;
|
return openingState_;
|
||||||
|
@ -4681,7 +4681,7 @@ public final class AdminProtos {
|
||||||
int hash = 41;
|
int hash = 41;
|
||||||
hash = (19 * hash) + getDescriptorForType().hashCode();
|
hash = (19 * hash) + getDescriptorForType().hashCode();
|
||||||
if (getOpeningStateCount() > 0) {
|
if (getOpeningStateCount() > 0) {
|
||||||
hash = (37 * hash) + OPENINGSTATE_FIELD_NUMBER;
|
hash = (37 * hash) + OPENING_STATE_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + hashEnumList(getOpeningStateList());
|
hash = (53 * hash) + hashEnumList(getOpeningStateList());
|
||||||
}
|
}
|
||||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||||
|
@ -4931,7 +4931,7 @@ public final class AdminProtos {
|
||||||
|
|
||||||
private int bitField0_;
|
private int bitField0_;
|
||||||
|
|
||||||
// repeated .OpenRegionResponse.RegionOpeningState openingState = 1;
|
// repeated .OpenRegionResponse.RegionOpeningState opening_state = 1;
|
||||||
private java.util.List<org.apache.hadoop.hbase.protobuf.generated.AdminProtos.OpenRegionResponse.RegionOpeningState> openingState_ =
|
private java.util.List<org.apache.hadoop.hbase.protobuf.generated.AdminProtos.OpenRegionResponse.RegionOpeningState> openingState_ =
|
||||||
java.util.Collections.emptyList();
|
java.util.Collections.emptyList();
|
||||||
private void ensureOpeningStateIsMutable() {
|
private void ensureOpeningStateIsMutable() {
|
||||||
|
@ -5001,15 +5001,15 @@ public final class AdminProtos {
|
||||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier getRegion();
|
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier getRegion();
|
||||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder getRegionOrBuilder();
|
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder getRegionOrBuilder();
|
||||||
|
|
||||||
// optional uint32 versionOfClosingNode = 2;
|
// optional uint32 version_of_closing_node = 2;
|
||||||
boolean hasVersionOfClosingNode();
|
boolean hasVersionOfClosingNode();
|
||||||
int getVersionOfClosingNode();
|
int getVersionOfClosingNode();
|
||||||
|
|
||||||
// optional bool transitionInZK = 3 [default = true];
|
// optional bool transition_in_ZK = 3 [default = true];
|
||||||
boolean hasTransitionInZK();
|
boolean hasTransitionInZK();
|
||||||
boolean getTransitionInZK();
|
boolean getTransitionInZK();
|
||||||
|
|
||||||
// optional .ServerName destinationServer = 4;
|
// optional .ServerName destination_server = 4;
|
||||||
boolean hasDestinationServer();
|
boolean hasDestinationServer();
|
||||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName getDestinationServer();
|
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName getDestinationServer();
|
||||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerNameOrBuilder getDestinationServerOrBuilder();
|
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerNameOrBuilder getDestinationServerOrBuilder();
|
||||||
|
@ -5056,8 +5056,8 @@ public final class AdminProtos {
|
||||||
return region_;
|
return region_;
|
||||||
}
|
}
|
||||||
|
|
||||||
// optional uint32 versionOfClosingNode = 2;
|
// optional uint32 version_of_closing_node = 2;
|
||||||
public static final int VERSIONOFCLOSINGNODE_FIELD_NUMBER = 2;
|
public static final int VERSION_OF_CLOSING_NODE_FIELD_NUMBER = 2;
|
||||||
private int versionOfClosingNode_;
|
private int versionOfClosingNode_;
|
||||||
public boolean hasVersionOfClosingNode() {
|
public boolean hasVersionOfClosingNode() {
|
||||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||||
|
@ -5066,8 +5066,8 @@ public final class AdminProtos {
|
||||||
return versionOfClosingNode_;
|
return versionOfClosingNode_;
|
||||||
}
|
}
|
||||||
|
|
||||||
// optional bool transitionInZK = 3 [default = true];
|
// optional bool transition_in_ZK = 3 [default = true];
|
||||||
public static final int TRANSITIONINZK_FIELD_NUMBER = 3;
|
public static final int TRANSITION_IN_ZK_FIELD_NUMBER = 3;
|
||||||
private boolean transitionInZK_;
|
private boolean transitionInZK_;
|
||||||
public boolean hasTransitionInZK() {
|
public boolean hasTransitionInZK() {
|
||||||
return ((bitField0_ & 0x00000004) == 0x00000004);
|
return ((bitField0_ & 0x00000004) == 0x00000004);
|
||||||
|
@ -5076,8 +5076,8 @@ public final class AdminProtos {
|
||||||
return transitionInZK_;
|
return transitionInZK_;
|
||||||
}
|
}
|
||||||
|
|
||||||
// optional .ServerName destinationServer = 4;
|
// optional .ServerName destination_server = 4;
|
||||||
public static final int DESTINATIONSERVER_FIELD_NUMBER = 4;
|
public static final int DESTINATION_SERVER_FIELD_NUMBER = 4;
|
||||||
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName destinationServer_;
|
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName destinationServer_;
|
||||||
public boolean hasDestinationServer() {
|
public boolean hasDestinationServer() {
|
||||||
return ((bitField0_ & 0x00000008) == 0x00000008);
|
return ((bitField0_ & 0x00000008) == 0x00000008);
|
||||||
|
@ -5215,15 +5215,15 @@ public final class AdminProtos {
|
||||||
hash = (53 * hash) + getRegion().hashCode();
|
hash = (53 * hash) + getRegion().hashCode();
|
||||||
}
|
}
|
||||||
if (hasVersionOfClosingNode()) {
|
if (hasVersionOfClosingNode()) {
|
||||||
hash = (37 * hash) + VERSIONOFCLOSINGNODE_FIELD_NUMBER;
|
hash = (37 * hash) + VERSION_OF_CLOSING_NODE_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + getVersionOfClosingNode();
|
hash = (53 * hash) + getVersionOfClosingNode();
|
||||||
}
|
}
|
||||||
if (hasTransitionInZK()) {
|
if (hasTransitionInZK()) {
|
||||||
hash = (37 * hash) + TRANSITIONINZK_FIELD_NUMBER;
|
hash = (37 * hash) + TRANSITION_IN_ZK_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + hashBoolean(getTransitionInZK());
|
hash = (53 * hash) + hashBoolean(getTransitionInZK());
|
||||||
}
|
}
|
||||||
if (hasDestinationServer()) {
|
if (hasDestinationServer()) {
|
||||||
hash = (37 * hash) + DESTINATIONSERVER_FIELD_NUMBER;
|
hash = (37 * hash) + DESTINATION_SERVER_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + getDestinationServer().hashCode();
|
hash = (53 * hash) + getDestinationServer().hashCode();
|
||||||
}
|
}
|
||||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||||
|
@ -5619,7 +5619,7 @@ public final class AdminProtos {
|
||||||
return regionBuilder_;
|
return regionBuilder_;
|
||||||
}
|
}
|
||||||
|
|
||||||
// optional uint32 versionOfClosingNode = 2;
|
// optional uint32 version_of_closing_node = 2;
|
||||||
private int versionOfClosingNode_ ;
|
private int versionOfClosingNode_ ;
|
||||||
public boolean hasVersionOfClosingNode() {
|
public boolean hasVersionOfClosingNode() {
|
||||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||||
|
@ -5640,7 +5640,7 @@ public final class AdminProtos {
|
||||||
return this;
|
return this;
|
||||||
}
|
}
|
||||||
|
|
||||||
// optional bool transitionInZK = 3 [default = true];
|
// optional bool transition_in_ZK = 3 [default = true];
|
||||||
private boolean transitionInZK_ = true;
|
private boolean transitionInZK_ = true;
|
||||||
public boolean hasTransitionInZK() {
|
public boolean hasTransitionInZK() {
|
||||||
return ((bitField0_ & 0x00000004) == 0x00000004);
|
return ((bitField0_ & 0x00000004) == 0x00000004);
|
||||||
|
@ -5661,7 +5661,7 @@ public final class AdminProtos {
|
||||||
return this;
|
return this;
|
||||||
}
|
}
|
||||||
|
|
||||||
// optional .ServerName destinationServer = 4;
|
// optional .ServerName destination_server = 4;
|
||||||
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName destinationServer_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.getDefaultInstance();
|
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName destinationServer_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.getDefaultInstance();
|
||||||
private com.google.protobuf.SingleFieldBuilder<
|
private com.google.protobuf.SingleFieldBuilder<
|
||||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerNameOrBuilder> destinationServerBuilder_;
|
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerNameOrBuilder> destinationServerBuilder_;
|
||||||
|
@ -6149,7 +6149,7 @@ public final class AdminProtos {
|
||||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier getRegion();
|
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier getRegion();
|
||||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder getRegionOrBuilder();
|
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder getRegionOrBuilder();
|
||||||
|
|
||||||
// optional uint64 ifOlderThanTs = 2;
|
// optional uint64 if_older_than_ts = 2;
|
||||||
boolean hasIfOlderThanTs();
|
boolean hasIfOlderThanTs();
|
||||||
long getIfOlderThanTs();
|
long getIfOlderThanTs();
|
||||||
}
|
}
|
||||||
|
@ -6195,8 +6195,8 @@ public final class AdminProtos {
|
||||||
return region_;
|
return region_;
|
||||||
}
|
}
|
||||||
|
|
||||||
// optional uint64 ifOlderThanTs = 2;
|
// optional uint64 if_older_than_ts = 2;
|
||||||
public static final int IFOLDERTHANTS_FIELD_NUMBER = 2;
|
public static final int IF_OLDER_THAN_TS_FIELD_NUMBER = 2;
|
||||||
private long ifOlderThanTs_;
|
private long ifOlderThanTs_;
|
||||||
public boolean hasIfOlderThanTs() {
|
public boolean hasIfOlderThanTs() {
|
||||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||||
|
@ -6299,7 +6299,7 @@ public final class AdminProtos {
|
||||||
hash = (53 * hash) + getRegion().hashCode();
|
hash = (53 * hash) + getRegion().hashCode();
|
||||||
}
|
}
|
||||||
if (hasIfOlderThanTs()) {
|
if (hasIfOlderThanTs()) {
|
||||||
hash = (37 * hash) + IFOLDERTHANTS_FIELD_NUMBER;
|
hash = (37 * hash) + IF_OLDER_THAN_TS_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + hashLong(getIfOlderThanTs());
|
hash = (53 * hash) + hashLong(getIfOlderThanTs());
|
||||||
}
|
}
|
||||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||||
|
@ -6648,7 +6648,7 @@ public final class AdminProtos {
|
||||||
return regionBuilder_;
|
return regionBuilder_;
|
||||||
}
|
}
|
||||||
|
|
||||||
// optional uint64 ifOlderThanTs = 2;
|
// optional uint64 if_older_than_ts = 2;
|
||||||
private long ifOlderThanTs_ ;
|
private long ifOlderThanTs_ ;
|
||||||
public boolean hasIfOlderThanTs() {
|
public boolean hasIfOlderThanTs() {
|
||||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||||
|
@ -6683,7 +6683,7 @@ public final class AdminProtos {
|
||||||
public interface FlushRegionResponseOrBuilder
|
public interface FlushRegionResponseOrBuilder
|
||||||
extends com.google.protobuf.MessageOrBuilder {
|
extends com.google.protobuf.MessageOrBuilder {
|
||||||
|
|
||||||
// required uint64 lastFlushTime = 1;
|
// required uint64 last_flush_time = 1;
|
||||||
boolean hasLastFlushTime();
|
boolean hasLastFlushTime();
|
||||||
long getLastFlushTime();
|
long getLastFlushTime();
|
||||||
|
|
||||||
|
@ -6720,8 +6720,8 @@ public final class AdminProtos {
|
||||||
}
|
}
|
||||||
|
|
||||||
private int bitField0_;
|
private int bitField0_;
|
||||||
// required uint64 lastFlushTime = 1;
|
// required uint64 last_flush_time = 1;
|
||||||
public static final int LASTFLUSHTIME_FIELD_NUMBER = 1;
|
public static final int LAST_FLUSH_TIME_FIELD_NUMBER = 1;
|
||||||
private long lastFlushTime_;
|
private long lastFlushTime_;
|
||||||
public boolean hasLastFlushTime() {
|
public boolean hasLastFlushTime() {
|
||||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||||
|
@ -6826,7 +6826,7 @@ public final class AdminProtos {
|
||||||
int hash = 41;
|
int hash = 41;
|
||||||
hash = (19 * hash) + getDescriptorForType().hashCode();
|
hash = (19 * hash) + getDescriptorForType().hashCode();
|
||||||
if (hasLastFlushTime()) {
|
if (hasLastFlushTime()) {
|
||||||
hash = (37 * hash) + LASTFLUSHTIME_FIELD_NUMBER;
|
hash = (37 * hash) + LAST_FLUSH_TIME_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + hashLong(getLastFlushTime());
|
hash = (53 * hash) + hashLong(getLastFlushTime());
|
||||||
}
|
}
|
||||||
if (hasFlushed()) {
|
if (hasFlushed()) {
|
||||||
|
@ -7072,7 +7072,7 @@ public final class AdminProtos {
|
||||||
|
|
||||||
private int bitField0_;
|
private int bitField0_;
|
||||||
|
|
||||||
// required uint64 lastFlushTime = 1;
|
// required uint64 last_flush_time = 1;
|
||||||
private long lastFlushTime_ ;
|
private long lastFlushTime_ ;
|
||||||
public boolean hasLastFlushTime() {
|
public boolean hasLastFlushTime() {
|
||||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||||
|
@ -7133,7 +7133,7 @@ public final class AdminProtos {
|
||||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier getRegion();
|
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier getRegion();
|
||||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder getRegionOrBuilder();
|
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder getRegionOrBuilder();
|
||||||
|
|
||||||
// optional bytes splitPoint = 2;
|
// optional bytes split_point = 2;
|
||||||
boolean hasSplitPoint();
|
boolean hasSplitPoint();
|
||||||
com.google.protobuf.ByteString getSplitPoint();
|
com.google.protobuf.ByteString getSplitPoint();
|
||||||
}
|
}
|
||||||
|
@ -7179,8 +7179,8 @@ public final class AdminProtos {
|
||||||
return region_;
|
return region_;
|
||||||
}
|
}
|
||||||
|
|
||||||
// optional bytes splitPoint = 2;
|
// optional bytes split_point = 2;
|
||||||
public static final int SPLITPOINT_FIELD_NUMBER = 2;
|
public static final int SPLIT_POINT_FIELD_NUMBER = 2;
|
||||||
private com.google.protobuf.ByteString splitPoint_;
|
private com.google.protobuf.ByteString splitPoint_;
|
||||||
public boolean hasSplitPoint() {
|
public boolean hasSplitPoint() {
|
||||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||||
|
@ -7283,7 +7283,7 @@ public final class AdminProtos {
|
||||||
hash = (53 * hash) + getRegion().hashCode();
|
hash = (53 * hash) + getRegion().hashCode();
|
||||||
}
|
}
|
||||||
if (hasSplitPoint()) {
|
if (hasSplitPoint()) {
|
||||||
hash = (37 * hash) + SPLITPOINT_FIELD_NUMBER;
|
hash = (37 * hash) + SPLIT_POINT_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + getSplitPoint().hashCode();
|
hash = (53 * hash) + getSplitPoint().hashCode();
|
||||||
}
|
}
|
||||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||||
|
@ -7632,7 +7632,7 @@ public final class AdminProtos {
|
||||||
return regionBuilder_;
|
return regionBuilder_;
|
||||||
}
|
}
|
||||||
|
|
||||||
// optional bytes splitPoint = 2;
|
// optional bytes split_point = 2;
|
||||||
private com.google.protobuf.ByteString splitPoint_ = com.google.protobuf.ByteString.EMPTY;
|
private com.google.protobuf.ByteString splitPoint_ = com.google.protobuf.ByteString.EMPTY;
|
||||||
public boolean hasSplitPoint() {
|
public boolean hasSplitPoint() {
|
||||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||||
|
@ -8878,12 +8878,12 @@ public final class AdminProtos {
|
||||||
public interface MergeRegionsRequestOrBuilder
|
public interface MergeRegionsRequestOrBuilder
|
||||||
extends com.google.protobuf.MessageOrBuilder {
|
extends com.google.protobuf.MessageOrBuilder {
|
||||||
|
|
||||||
// required .RegionSpecifier regionA = 1;
|
// required .RegionSpecifier region_a = 1;
|
||||||
boolean hasRegionA();
|
boolean hasRegionA();
|
||||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier getRegionA();
|
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier getRegionA();
|
||||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder getRegionAOrBuilder();
|
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder getRegionAOrBuilder();
|
||||||
|
|
||||||
// required .RegionSpecifier regionB = 2;
|
// required .RegionSpecifier region_b = 2;
|
||||||
boolean hasRegionB();
|
boolean hasRegionB();
|
||||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier getRegionB();
|
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier getRegionB();
|
||||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder getRegionBOrBuilder();
|
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder getRegionBOrBuilder();
|
||||||
|
@ -8921,8 +8921,8 @@ public final class AdminProtos {
|
||||||
}
|
}
|
||||||
|
|
||||||
private int bitField0_;
|
private int bitField0_;
|
||||||
// required .RegionSpecifier regionA = 1;
|
// required .RegionSpecifier region_a = 1;
|
||||||
public static final int REGIONA_FIELD_NUMBER = 1;
|
public static final int REGION_A_FIELD_NUMBER = 1;
|
||||||
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier regionA_;
|
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier regionA_;
|
||||||
public boolean hasRegionA() {
|
public boolean hasRegionA() {
|
||||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||||
|
@ -8934,8 +8934,8 @@ public final class AdminProtos {
|
||||||
return regionA_;
|
return regionA_;
|
||||||
}
|
}
|
||||||
|
|
||||||
// required .RegionSpecifier regionB = 2;
|
// required .RegionSpecifier region_b = 2;
|
||||||
public static final int REGIONB_FIELD_NUMBER = 2;
|
public static final int REGION_B_FIELD_NUMBER = 2;
|
||||||
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier regionB_;
|
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier regionB_;
|
||||||
public boolean hasRegionB() {
|
public boolean hasRegionB() {
|
||||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||||
|
@ -9068,11 +9068,11 @@ public final class AdminProtos {
|
||||||
int hash = 41;
|
int hash = 41;
|
||||||
hash = (19 * hash) + getDescriptorForType().hashCode();
|
hash = (19 * hash) + getDescriptorForType().hashCode();
|
||||||
if (hasRegionA()) {
|
if (hasRegionA()) {
|
||||||
hash = (37 * hash) + REGIONA_FIELD_NUMBER;
|
hash = (37 * hash) + REGION_A_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + getRegionA().hashCode();
|
hash = (53 * hash) + getRegionA().hashCode();
|
||||||
}
|
}
|
||||||
if (hasRegionB()) {
|
if (hasRegionB()) {
|
||||||
hash = (37 * hash) + REGIONB_FIELD_NUMBER;
|
hash = (37 * hash) + REGION_B_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + getRegionB().hashCode();
|
hash = (53 * hash) + getRegionB().hashCode();
|
||||||
}
|
}
|
||||||
if (hasForcible()) {
|
if (hasForcible()) {
|
||||||
|
@ -9370,7 +9370,7 @@ public final class AdminProtos {
|
||||||
|
|
||||||
private int bitField0_;
|
private int bitField0_;
|
||||||
|
|
||||||
// required .RegionSpecifier regionA = 1;
|
// required .RegionSpecifier region_a = 1;
|
||||||
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier regionA_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance();
|
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier regionA_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance();
|
||||||
private com.google.protobuf.SingleFieldBuilder<
|
private com.google.protobuf.SingleFieldBuilder<
|
||||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder> regionABuilder_;
|
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder> regionABuilder_;
|
||||||
|
@ -9460,7 +9460,7 @@ public final class AdminProtos {
|
||||||
return regionABuilder_;
|
return regionABuilder_;
|
||||||
}
|
}
|
||||||
|
|
||||||
// required .RegionSpecifier regionB = 2;
|
// required .RegionSpecifier region_b = 2;
|
||||||
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier regionB_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance();
|
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier regionB_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance();
|
||||||
private com.google.protobuf.SingleFieldBuilder<
|
private com.google.protobuf.SingleFieldBuilder<
|
||||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder> regionBBuilder_;
|
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder> regionBBuilder_;
|
||||||
|
@ -9890,12 +9890,12 @@ public final class AdminProtos {
|
||||||
org.apache.hadoop.hbase.protobuf.generated.WALProtos.WALKey getKey();
|
org.apache.hadoop.hbase.protobuf.generated.WALProtos.WALKey getKey();
|
||||||
org.apache.hadoop.hbase.protobuf.generated.WALProtos.WALKeyOrBuilder getKeyOrBuilder();
|
org.apache.hadoop.hbase.protobuf.generated.WALProtos.WALKeyOrBuilder getKeyOrBuilder();
|
||||||
|
|
||||||
// repeated bytes keyValueBytes = 2;
|
// repeated bytes key_value_bytes = 2;
|
||||||
java.util.List<com.google.protobuf.ByteString> getKeyValueBytesList();
|
java.util.List<com.google.protobuf.ByteString> getKeyValueBytesList();
|
||||||
int getKeyValueBytesCount();
|
int getKeyValueBytesCount();
|
||||||
com.google.protobuf.ByteString getKeyValueBytes(int index);
|
com.google.protobuf.ByteString getKeyValueBytes(int index);
|
||||||
|
|
||||||
// optional int32 associatedCellCount = 3;
|
// optional int32 associated_cell_count = 3;
|
||||||
boolean hasAssociatedCellCount();
|
boolean hasAssociatedCellCount();
|
||||||
int getAssociatedCellCount();
|
int getAssociatedCellCount();
|
||||||
}
|
}
|
||||||
|
@ -9941,8 +9941,8 @@ public final class AdminProtos {
|
||||||
return key_;
|
return key_;
|
||||||
}
|
}
|
||||||
|
|
||||||
// repeated bytes keyValueBytes = 2;
|
// repeated bytes key_value_bytes = 2;
|
||||||
public static final int KEYVALUEBYTES_FIELD_NUMBER = 2;
|
public static final int KEY_VALUE_BYTES_FIELD_NUMBER = 2;
|
||||||
private java.util.List<com.google.protobuf.ByteString> keyValueBytes_;
|
private java.util.List<com.google.protobuf.ByteString> keyValueBytes_;
|
||||||
public java.util.List<com.google.protobuf.ByteString>
|
public java.util.List<com.google.protobuf.ByteString>
|
||||||
getKeyValueBytesList() {
|
getKeyValueBytesList() {
|
||||||
|
@ -9955,8 +9955,8 @@ public final class AdminProtos {
|
||||||
return keyValueBytes_.get(index);
|
return keyValueBytes_.get(index);
|
||||||
}
|
}
|
||||||
|
|
||||||
// optional int32 associatedCellCount = 3;
|
// optional int32 associated_cell_count = 3;
|
||||||
public static final int ASSOCIATEDCELLCOUNT_FIELD_NUMBER = 3;
|
public static final int ASSOCIATED_CELL_COUNT_FIELD_NUMBER = 3;
|
||||||
private int associatedCellCount_;
|
private int associatedCellCount_;
|
||||||
public boolean hasAssociatedCellCount() {
|
public boolean hasAssociatedCellCount() {
|
||||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||||
|
@ -10074,11 +10074,11 @@ public final class AdminProtos {
|
||||||
hash = (53 * hash) + getKey().hashCode();
|
hash = (53 * hash) + getKey().hashCode();
|
||||||
}
|
}
|
||||||
if (getKeyValueBytesCount() > 0) {
|
if (getKeyValueBytesCount() > 0) {
|
||||||
hash = (37 * hash) + KEYVALUEBYTES_FIELD_NUMBER;
|
hash = (37 * hash) + KEY_VALUE_BYTES_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + getKeyValueBytesList().hashCode();
|
hash = (53 * hash) + getKeyValueBytesList().hashCode();
|
||||||
}
|
}
|
||||||
if (hasAssociatedCellCount()) {
|
if (hasAssociatedCellCount()) {
|
||||||
hash = (37 * hash) + ASSOCIATEDCELLCOUNT_FIELD_NUMBER;
|
hash = (37 * hash) + ASSOCIATED_CELL_COUNT_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + getAssociatedCellCount();
|
hash = (53 * hash) + getAssociatedCellCount();
|
||||||
}
|
}
|
||||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||||
|
@ -10449,7 +10449,7 @@ public final class AdminProtos {
|
||||||
return keyBuilder_;
|
return keyBuilder_;
|
||||||
}
|
}
|
||||||
|
|
||||||
// repeated bytes keyValueBytes = 2;
|
// repeated bytes key_value_bytes = 2;
|
||||||
private java.util.List<com.google.protobuf.ByteString> keyValueBytes_ = java.util.Collections.emptyList();;
|
private java.util.List<com.google.protobuf.ByteString> keyValueBytes_ = java.util.Collections.emptyList();;
|
||||||
private void ensureKeyValueBytesIsMutable() {
|
private void ensureKeyValueBytesIsMutable() {
|
||||||
if (!((bitField0_ & 0x00000002) == 0x00000002)) {
|
if (!((bitField0_ & 0x00000002) == 0x00000002)) {
|
||||||
|
@ -10500,7 +10500,7 @@ public final class AdminProtos {
|
||||||
return this;
|
return this;
|
||||||
}
|
}
|
||||||
|
|
||||||
// optional int32 associatedCellCount = 3;
|
// optional int32 associated_cell_count = 3;
|
||||||
private int associatedCellCount_ ;
|
private int associatedCellCount_ ;
|
||||||
public boolean hasAssociatedCellCount() {
|
public boolean hasAssociatedCellCount() {
|
||||||
return ((bitField0_ & 0x00000004) == 0x00000004);
|
return ((bitField0_ & 0x00000004) == 0x00000004);
|
||||||
|
@ -11728,7 +11728,7 @@ public final class AdminProtos {
|
||||||
public interface RollWALWriterResponseOrBuilder
|
public interface RollWALWriterResponseOrBuilder
|
||||||
extends com.google.protobuf.MessageOrBuilder {
|
extends com.google.protobuf.MessageOrBuilder {
|
||||||
|
|
||||||
// repeated bytes regionToFlush = 1;
|
// repeated bytes region_to_flush = 1;
|
||||||
java.util.List<com.google.protobuf.ByteString> getRegionToFlushList();
|
java.util.List<com.google.protobuf.ByteString> getRegionToFlushList();
|
||||||
int getRegionToFlushCount();
|
int getRegionToFlushCount();
|
||||||
com.google.protobuf.ByteString getRegionToFlush(int index);
|
com.google.protobuf.ByteString getRegionToFlush(int index);
|
||||||
|
@ -11761,8 +11761,8 @@ public final class AdminProtos {
|
||||||
return org.apache.hadoop.hbase.protobuf.generated.AdminProtos.internal_static_RollWALWriterResponse_fieldAccessorTable;
|
return org.apache.hadoop.hbase.protobuf.generated.AdminProtos.internal_static_RollWALWriterResponse_fieldAccessorTable;
|
||||||
}
|
}
|
||||||
|
|
||||||
// repeated bytes regionToFlush = 1;
|
// repeated bytes region_to_flush = 1;
|
||||||
public static final int REGIONTOFLUSH_FIELD_NUMBER = 1;
|
public static final int REGION_TO_FLUSH_FIELD_NUMBER = 1;
|
||||||
private java.util.List<com.google.protobuf.ByteString> regionToFlush_;
|
private java.util.List<com.google.protobuf.ByteString> regionToFlush_;
|
||||||
public java.util.List<com.google.protobuf.ByteString>
|
public java.util.List<com.google.protobuf.ByteString>
|
||||||
getRegionToFlushList() {
|
getRegionToFlushList() {
|
||||||
|
@ -11846,7 +11846,7 @@ public final class AdminProtos {
|
||||||
int hash = 41;
|
int hash = 41;
|
||||||
hash = (19 * hash) + getDescriptorForType().hashCode();
|
hash = (19 * hash) + getDescriptorForType().hashCode();
|
||||||
if (getRegionToFlushCount() > 0) {
|
if (getRegionToFlushCount() > 0) {
|
||||||
hash = (37 * hash) + REGIONTOFLUSH_FIELD_NUMBER;
|
hash = (37 * hash) + REGION_TO_FLUSH_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + getRegionToFlushList().hashCode();
|
hash = (53 * hash) + getRegionToFlushList().hashCode();
|
||||||
}
|
}
|
||||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||||
|
@ -12076,7 +12076,7 @@ public final class AdminProtos {
|
||||||
|
|
||||||
private int bitField0_;
|
private int bitField0_;
|
||||||
|
|
||||||
// repeated bytes regionToFlush = 1;
|
// repeated bytes region_to_flush = 1;
|
||||||
private java.util.List<com.google.protobuf.ByteString> regionToFlush_ = java.util.Collections.emptyList();;
|
private java.util.List<com.google.protobuf.ByteString> regionToFlush_ = java.util.Collections.emptyList();;
|
||||||
private void ensureRegionToFlushIsMutable() {
|
private void ensureRegionToFlushIsMutable() {
|
||||||
if (!((bitField0_ & 0x00000001) == 0x00000001)) {
|
if (!((bitField0_ & 0x00000001) == 0x00000001)) {
|
||||||
|
@ -13157,12 +13157,12 @@ public final class AdminProtos {
|
||||||
public interface ServerInfoOrBuilder
|
public interface ServerInfoOrBuilder
|
||||||
extends com.google.protobuf.MessageOrBuilder {
|
extends com.google.protobuf.MessageOrBuilder {
|
||||||
|
|
||||||
// required .ServerName serverName = 1;
|
// required .ServerName server_name = 1;
|
||||||
boolean hasServerName();
|
boolean hasServerName();
|
||||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName getServerName();
|
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName getServerName();
|
||||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerNameOrBuilder getServerNameOrBuilder();
|
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerNameOrBuilder getServerNameOrBuilder();
|
||||||
|
|
||||||
// optional uint32 webuiPort = 2;
|
// optional uint32 webui_port = 2;
|
||||||
boolean hasWebuiPort();
|
boolean hasWebuiPort();
|
||||||
int getWebuiPort();
|
int getWebuiPort();
|
||||||
}
|
}
|
||||||
|
@ -13195,8 +13195,8 @@ public final class AdminProtos {
|
||||||
}
|
}
|
||||||
|
|
||||||
private int bitField0_;
|
private int bitField0_;
|
||||||
// required .ServerName serverName = 1;
|
// required .ServerName server_name = 1;
|
||||||
public static final int SERVERNAME_FIELD_NUMBER = 1;
|
public static final int SERVER_NAME_FIELD_NUMBER = 1;
|
||||||
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName serverName_;
|
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName serverName_;
|
||||||
public boolean hasServerName() {
|
public boolean hasServerName() {
|
||||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||||
|
@ -13208,8 +13208,8 @@ public final class AdminProtos {
|
||||||
return serverName_;
|
return serverName_;
|
||||||
}
|
}
|
||||||
|
|
||||||
// optional uint32 webuiPort = 2;
|
// optional uint32 webui_port = 2;
|
||||||
public static final int WEBUIPORT_FIELD_NUMBER = 2;
|
public static final int WEBUI_PORT_FIELD_NUMBER = 2;
|
||||||
private int webuiPort_;
|
private int webuiPort_;
|
||||||
public boolean hasWebuiPort() {
|
public boolean hasWebuiPort() {
|
||||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||||
|
@ -13308,11 +13308,11 @@ public final class AdminProtos {
|
||||||
int hash = 41;
|
int hash = 41;
|
||||||
hash = (19 * hash) + getDescriptorForType().hashCode();
|
hash = (19 * hash) + getDescriptorForType().hashCode();
|
||||||
if (hasServerName()) {
|
if (hasServerName()) {
|
||||||
hash = (37 * hash) + SERVERNAME_FIELD_NUMBER;
|
hash = (37 * hash) + SERVER_NAME_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + getServerName().hashCode();
|
hash = (53 * hash) + getServerName().hashCode();
|
||||||
}
|
}
|
||||||
if (hasWebuiPort()) {
|
if (hasWebuiPort()) {
|
||||||
hash = (37 * hash) + WEBUIPORT_FIELD_NUMBER;
|
hash = (37 * hash) + WEBUI_PORT_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + getWebuiPort();
|
hash = (53 * hash) + getWebuiPort();
|
||||||
}
|
}
|
||||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||||
|
@ -13571,7 +13571,7 @@ public final class AdminProtos {
|
||||||
|
|
||||||
private int bitField0_;
|
private int bitField0_;
|
||||||
|
|
||||||
// required .ServerName serverName = 1;
|
// required .ServerName server_name = 1;
|
||||||
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName serverName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.getDefaultInstance();
|
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName serverName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.getDefaultInstance();
|
||||||
private com.google.protobuf.SingleFieldBuilder<
|
private com.google.protobuf.SingleFieldBuilder<
|
||||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerNameOrBuilder> serverNameBuilder_;
|
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerNameOrBuilder> serverNameBuilder_;
|
||||||
|
@ -13661,7 +13661,7 @@ public final class AdminProtos {
|
||||||
return serverNameBuilder_;
|
return serverNameBuilder_;
|
||||||
}
|
}
|
||||||
|
|
||||||
// optional uint32 webuiPort = 2;
|
// optional uint32 webui_port = 2;
|
||||||
private int webuiPort_ ;
|
private int webuiPort_ ;
|
||||||
public boolean hasWebuiPort() {
|
public boolean hasWebuiPort() {
|
||||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||||
|
@ -13696,7 +13696,7 @@ public final class AdminProtos {
|
||||||
public interface GetServerInfoResponseOrBuilder
|
public interface GetServerInfoResponseOrBuilder
|
||||||
extends com.google.protobuf.MessageOrBuilder {
|
extends com.google.protobuf.MessageOrBuilder {
|
||||||
|
|
||||||
// required .ServerInfo serverInfo = 1;
|
// required .ServerInfo server_info = 1;
|
||||||
boolean hasServerInfo();
|
boolean hasServerInfo();
|
||||||
org.apache.hadoop.hbase.protobuf.generated.AdminProtos.ServerInfo getServerInfo();
|
org.apache.hadoop.hbase.protobuf.generated.AdminProtos.ServerInfo getServerInfo();
|
||||||
org.apache.hadoop.hbase.protobuf.generated.AdminProtos.ServerInfoOrBuilder getServerInfoOrBuilder();
|
org.apache.hadoop.hbase.protobuf.generated.AdminProtos.ServerInfoOrBuilder getServerInfoOrBuilder();
|
||||||
|
@ -13730,8 +13730,8 @@ public final class AdminProtos {
|
||||||
}
|
}
|
||||||
|
|
||||||
private int bitField0_;
|
private int bitField0_;
|
||||||
// required .ServerInfo serverInfo = 1;
|
// required .ServerInfo server_info = 1;
|
||||||
public static final int SERVERINFO_FIELD_NUMBER = 1;
|
public static final int SERVER_INFO_FIELD_NUMBER = 1;
|
||||||
private org.apache.hadoop.hbase.protobuf.generated.AdminProtos.ServerInfo serverInfo_;
|
private org.apache.hadoop.hbase.protobuf.generated.AdminProtos.ServerInfo serverInfo_;
|
||||||
public boolean hasServerInfo() {
|
public boolean hasServerInfo() {
|
||||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||||
|
@ -13820,7 +13820,7 @@ public final class AdminProtos {
|
||||||
int hash = 41;
|
int hash = 41;
|
||||||
hash = (19 * hash) + getDescriptorForType().hashCode();
|
hash = (19 * hash) + getDescriptorForType().hashCode();
|
||||||
if (hasServerInfo()) {
|
if (hasServerInfo()) {
|
||||||
hash = (37 * hash) + SERVERINFO_FIELD_NUMBER;
|
hash = (37 * hash) + SERVER_INFO_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + getServerInfo().hashCode();
|
hash = (53 * hash) + getServerInfo().hashCode();
|
||||||
}
|
}
|
||||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||||
|
@ -14065,7 +14065,7 @@ public final class AdminProtos {
|
||||||
|
|
||||||
private int bitField0_;
|
private int bitField0_;
|
||||||
|
|
||||||
// required .ServerInfo serverInfo = 1;
|
// required .ServerInfo server_info = 1;
|
||||||
private org.apache.hadoop.hbase.protobuf.generated.AdminProtos.ServerInfo serverInfo_ = org.apache.hadoop.hbase.protobuf.generated.AdminProtos.ServerInfo.getDefaultInstance();
|
private org.apache.hadoop.hbase.protobuf.generated.AdminProtos.ServerInfo serverInfo_ = org.apache.hadoop.hbase.protobuf.generated.AdminProtos.ServerInfo.getDefaultInstance();
|
||||||
private com.google.protobuf.SingleFieldBuilder<
|
private com.google.protobuf.SingleFieldBuilder<
|
||||||
org.apache.hadoop.hbase.protobuf.generated.AdminProtos.ServerInfo, org.apache.hadoop.hbase.protobuf.generated.AdminProtos.ServerInfo.Builder, org.apache.hadoop.hbase.protobuf.generated.AdminProtos.ServerInfoOrBuilder> serverInfoBuilder_;
|
org.apache.hadoop.hbase.protobuf.generated.AdminProtos.ServerInfo, org.apache.hadoop.hbase.protobuf.generated.AdminProtos.ServerInfo.Builder, org.apache.hadoop.hbase.protobuf.generated.AdminProtos.ServerInfoOrBuilder> serverInfoBuilder_;
|
||||||
|
@ -15387,79 +15387,80 @@ public final class AdminProtos {
|
||||||
static {
|
static {
|
||||||
java.lang.String[] descriptorData = {
|
java.lang.String[] descriptorData = {
|
||||||
"\n\013Admin.proto\032\014Client.proto\032\013hbase.proto" +
|
"\n\013Admin.proto\032\014Client.proto\032\013hbase.proto" +
|
||||||
"\032\tWAL.proto\"Q\n\024GetRegionInfoRequest\022 \n\006r" +
|
"\032\tWAL.proto\"R\n\024GetRegionInfoRequest\022 \n\006r" +
|
||||||
"egion\030\001 \002(\0132\020.RegionSpecifier\022\027\n\017compact" +
|
"egion\030\001 \002(\0132\020.RegionSpecifier\022\030\n\020compact" +
|
||||||
"ionState\030\002 \001(\010\"\301\001\n\025GetRegionInfoResponse" +
|
"ion_state\030\002 \001(\010\"\303\001\n\025GetRegionInfoRespons" +
|
||||||
"\022\037\n\nregionInfo\030\001 \002(\0132\013.RegionInfo\022?\n\017com" +
|
"e\022 \n\013region_info\030\001 \002(\0132\013.RegionInfo\022@\n\020c" +
|
||||||
"pactionState\030\002 \001(\0162&.GetRegionInfoRespon" +
|
"ompaction_state\030\002 \001(\0162&.GetRegionInfoRes" +
|
||||||
"se.CompactionState\"F\n\017CompactionState\022\010\n" +
|
"ponse.CompactionState\"F\n\017CompactionState" +
|
||||||
"\004NONE\020\000\022\t\n\005MINOR\020\001\022\t\n\005MAJOR\020\002\022\023\n\017MAJOR_A" +
|
"\022\010\n\004NONE\020\000\022\t\n\005MINOR\020\001\022\t\n\005MAJOR\020\002\022\023\n\017MAJO" +
|
||||||
"ND_MINOR\020\003\"G\n\023GetStoreFileRequest\022 \n\006reg" +
|
"R_AND_MINOR\020\003\"G\n\023GetStoreFileRequest\022 \n\006" +
|
||||||
"ion\030\001 \002(\0132\020.RegionSpecifier\022\016\n\006family\030\002 ",
|
"region\030\001 \002(\0132\020.RegionSpecifier\022\016\n\006family",
|
||||||
"\003(\014\")\n\024GetStoreFileResponse\022\021\n\tstoreFile" +
|
"\030\002 \003(\014\"*\n\024GetStoreFileResponse\022\022\n\nstore_" +
|
||||||
"\030\001 \003(\t\"\030\n\026GetOnlineRegionRequest\":\n\027GetO" +
|
"file\030\001 \003(\t\"\030\n\026GetOnlineRegionRequest\";\n\027" +
|
||||||
"nlineRegionResponse\022\037\n\nregionInfo\030\001 \003(\0132" +
|
"GetOnlineRegionResponse\022 \n\013region_info\030\001" +
|
||||||
"\013.RegionInfo\"\270\001\n\021OpenRegionRequest\0223\n\010op" +
|
" \003(\0132\013.RegionInfo\"\275\001\n\021OpenRegionRequest\022" +
|
||||||
"enInfo\030\001 \003(\0132!.OpenRegionRequest.RegionO" +
|
"4\n\topen_info\030\001 \003(\0132!.OpenRegionRequest.R" +
|
||||||
"penInfo\032n\n\016RegionOpenInfo\022\033\n\006region\030\001 \002(" +
|
"egionOpenInfo\032r\n\016RegionOpenInfo\022\033\n\006regio" +
|
||||||
"\0132\013.RegionInfo\022\034\n\024versionOfOfflineNode\030\002" +
|
"n\030\001 \002(\0132\013.RegionInfo\022\037\n\027version_of_offli" +
|
||||||
" \001(\r\022!\n\014favoredNodes\030\003 \003(\0132\013.ServerName\"" +
|
"ne_node\030\002 \001(\r\022\"\n\rfavored_nodes\030\003 \003(\0132\013.S" +
|
||||||
"\234\001\n\022OpenRegionResponse\022<\n\014openingState\030\001" +
|
"erverName\"\235\001\n\022OpenRegionResponse\022=\n\ropen" +
|
||||||
" \003(\0162&.OpenRegionResponse.RegionOpeningS",
|
"ing_state\030\001 \003(\0162&.OpenRegionResponse.Reg",
|
||||||
"tate\"H\n\022RegionOpeningState\022\n\n\006OPENED\020\000\022\022" +
|
"ionOpeningState\"H\n\022RegionOpeningState\022\n\n" +
|
||||||
"\n\016ALREADY_OPENED\020\001\022\022\n\016FAILED_OPENING\020\002\"\232" +
|
"\006OPENED\020\000\022\022\n\016ALREADY_OPENED\020\001\022\022\n\016FAILED_" +
|
||||||
"\001\n\022CloseRegionRequest\022 \n\006region\030\001 \002(\0132\020." +
|
"OPENING\020\002\"\240\001\n\022CloseRegionRequest\022 \n\006regi" +
|
||||||
"RegionSpecifier\022\034\n\024versionOfClosingNode\030" +
|
"on\030\001 \002(\0132\020.RegionSpecifier\022\037\n\027version_of" +
|
||||||
"\002 \001(\r\022\034\n\016transitionInZK\030\003 \001(\010:\004true\022&\n\021d" +
|
"_closing_node\030\002 \001(\r\022\036\n\020transition_in_ZK\030" +
|
||||||
"estinationServer\030\004 \001(\0132\013.ServerName\"%\n\023C" +
|
"\003 \001(\010:\004true\022\'\n\022destination_server\030\004 \001(\0132" +
|
||||||
"loseRegionResponse\022\016\n\006closed\030\001 \002(\010\"M\n\022Fl" +
|
"\013.ServerName\"%\n\023CloseRegionResponse\022\016\n\006c" +
|
||||||
"ushRegionRequest\022 \n\006region\030\001 \002(\0132\020.Regio" +
|
"losed\030\001 \002(\010\"P\n\022FlushRegionRequest\022 \n\006reg" +
|
||||||
"nSpecifier\022\025\n\rifOlderThanTs\030\002 \001(\004\"=\n\023Flu" +
|
"ion\030\001 \002(\0132\020.RegionSpecifier\022\030\n\020if_older_" +
|
||||||
"shRegionResponse\022\025\n\rlastFlushTime\030\001 \002(\004\022",
|
"than_ts\030\002 \001(\004\"?\n\023FlushRegionResponse\022\027\n\017",
|
||||||
"\017\n\007flushed\030\002 \001(\010\"J\n\022SplitRegionRequest\022 " +
|
"last_flush_time\030\001 \002(\004\022\017\n\007flushed\030\002 \001(\010\"K" +
|
||||||
"\n\006region\030\001 \002(\0132\020.RegionSpecifier\022\022\n\nspli" +
|
"\n\022SplitRegionRequest\022 \n\006region\030\001 \002(\0132\020.R" +
|
||||||
"tPoint\030\002 \001(\014\"\025\n\023SplitRegionResponse\"W\n\024C" +
|
"egionSpecifier\022\023\n\013split_point\030\002 \001(\014\"\025\n\023S" +
|
||||||
"ompactRegionRequest\022 \n\006region\030\001 \002(\0132\020.Re" +
|
"plitRegionResponse\"W\n\024CompactRegionReque" +
|
||||||
"gionSpecifier\022\r\n\005major\030\002 \001(\010\022\016\n\006family\030\003" +
|
"st\022 \n\006region\030\001 \002(\0132\020.RegionSpecifier\022\r\n\005" +
|
||||||
" \001(\014\"\027\n\025CompactRegionResponse\"t\n\023MergeRe" +
|
"major\030\002 \001(\010\022\016\n\006family\030\003 \001(\014\"\027\n\025CompactRe" +
|
||||||
"gionsRequest\022!\n\007regionA\030\001 \002(\0132\020.RegionSp" +
|
"gionResponse\"v\n\023MergeRegionsRequest\022\"\n\010r" +
|
||||||
"ecifier\022!\n\007regionB\030\002 \002(\0132\020.RegionSpecifi" +
|
"egion_a\030\001 \002(\0132\020.RegionSpecifier\022\"\n\010regio" +
|
||||||
"er\022\027\n\010forcible\030\003 \001(\010:\005false\"\026\n\024MergeRegi" +
|
"n_b\030\002 \002(\0132\020.RegionSpecifier\022\027\n\010forcible\030" +
|
||||||
"onsResponse\"T\n\010WALEntry\022\024\n\003key\030\001 \002(\0132\007.W",
|
"\003 \001(\010:\005false\"\026\n\024MergeRegionsResponse\"X\n\010",
|
||||||
"ALKey\022\025\n\rkeyValueBytes\030\002 \003(\014\022\033\n\023associat" +
|
"WALEntry\022\024\n\003key\030\001 \002(\0132\007.WALKey\022\027\n\017key_va" +
|
||||||
"edCellCount\030\003 \001(\005\"4\n\030ReplicateWALEntryRe" +
|
"lue_bytes\030\002 \003(\014\022\035\n\025associated_cell_count" +
|
||||||
"quest\022\030\n\005entry\030\001 \003(\0132\t.WALEntry\"\033\n\031Repli" +
|
"\030\003 \001(\005\"4\n\030ReplicateWALEntryRequest\022\030\n\005en" +
|
||||||
"cateWALEntryResponse\"\026\n\024RollWALWriterReq" +
|
"try\030\001 \003(\0132\t.WALEntry\"\033\n\031ReplicateWALEntr" +
|
||||||
"uest\".\n\025RollWALWriterResponse\022\025\n\rregionT" +
|
"yResponse\"\026\n\024RollWALWriterRequest\"0\n\025Rol" +
|
||||||
"oFlush\030\001 \003(\014\"#\n\021StopServerRequest\022\016\n\006rea" +
|
"lWALWriterResponse\022\027\n\017region_to_flush\030\001 " +
|
||||||
"son\030\001 \002(\t\"\024\n\022StopServerResponse\"\026\n\024GetSe" +
|
"\003(\014\"#\n\021StopServerRequest\022\016\n\006reason\030\001 \002(\t" +
|
||||||
"rverInfoRequest\"@\n\nServerInfo\022\037\n\nserverN" +
|
"\"\024\n\022StopServerResponse\"\026\n\024GetServerInfoR" +
|
||||||
"ame\030\001 \002(\0132\013.ServerName\022\021\n\twebuiPort\030\002 \001(" +
|
"equest\"B\n\nServerInfo\022 \n\013server_name\030\001 \002(" +
|
||||||
"\r\"8\n\025GetServerInfoResponse\022\037\n\nserverInfo",
|
"\0132\013.ServerName\022\022\n\nwebui_port\030\002 \001(\r\"9\n\025Ge",
|
||||||
"\030\001 \002(\0132\013.ServerInfo2\337\006\n\014AdminService\022>\n\r" +
|
"tServerInfoResponse\022 \n\013server_info\030\001 \002(\013" +
|
||||||
"getRegionInfo\022\025.GetRegionInfoRequest\032\026.G" +
|
"2\013.ServerInfo2\337\006\n\014AdminService\022>\n\rGetReg" +
|
||||||
"etRegionInfoResponse\022;\n\014getStoreFile\022\024.G" +
|
"ionInfo\022\025.GetRegionInfoRequest\032\026.GetRegi" +
|
||||||
"etStoreFileRequest\032\025.GetStoreFileRespons" +
|
"onInfoResponse\022;\n\014GetStoreFile\022\024.GetStor" +
|
||||||
"e\022D\n\017getOnlineRegion\022\027.GetOnlineRegionRe" +
|
"eFileRequest\032\025.GetStoreFileResponse\022D\n\017G" +
|
||||||
"quest\032\030.GetOnlineRegionResponse\0225\n\nopenR" +
|
"etOnlineRegion\022\027.GetOnlineRegionRequest\032" +
|
||||||
"egion\022\022.OpenRegionRequest\032\023.OpenRegionRe" +
|
"\030.GetOnlineRegionResponse\0225\n\nOpenRegion\022" +
|
||||||
"sponse\0228\n\013closeRegion\022\023.CloseRegionReque" +
|
"\022.OpenRegionRequest\032\023.OpenRegionResponse" +
|
||||||
"st\032\024.CloseRegionResponse\0228\n\013flushRegion\022" +
|
"\0228\n\013CloseRegion\022\023.CloseRegionRequest\032\024.C" +
|
||||||
"\023.FlushRegionRequest\032\024.FlushRegionRespon",
|
"loseRegionResponse\0228\n\013FlushRegion\022\023.Flus",
|
||||||
"se\0228\n\013splitRegion\022\023.SplitRegionRequest\032\024" +
|
"hRegionRequest\032\024.FlushRegionResponse\0228\n\013" +
|
||||||
".SplitRegionResponse\022>\n\rcompactRegion\022\025." +
|
"SplitRegion\022\023.SplitRegionRequest\032\024.Split" +
|
||||||
"CompactRegionRequest\032\026.CompactRegionResp" +
|
"RegionResponse\022>\n\rCompactRegion\022\025.Compac" +
|
||||||
"onse\022;\n\014mergeRegions\022\024.MergeRegionsReque" +
|
"tRegionRequest\032\026.CompactRegionResponse\022;" +
|
||||||
"st\032\025.MergeRegionsResponse\022J\n\021replicateWA" +
|
"\n\014MergeRegions\022\024.MergeRegionsRequest\032\025.M" +
|
||||||
"LEntry\022\031.ReplicateWALEntryRequest\032\032.Repl" +
|
"ergeRegionsResponse\022J\n\021ReplicateWALEntry" +
|
||||||
"icateWALEntryResponse\022\'\n\006replay\022\r.MultiR" +
|
"\022\031.ReplicateWALEntryRequest\032\032.ReplicateW" +
|
||||||
"equest\032\016.MultiResponse\022>\n\rrollWALWriter\022" +
|
"ALEntryResponse\022\'\n\006Replay\022\r.MultiRequest" +
|
||||||
"\025.RollWALWriterRequest\032\026.RollWALWriterRe" +
|
"\032\016.MultiResponse\022>\n\rRollWALWriter\022\025.Roll" +
|
||||||
"sponse\022>\n\rgetServerInfo\022\025.GetServerInfoR",
|
"WALWriterRequest\032\026.RollWALWriterResponse",
|
||||||
"equest\032\026.GetServerInfoResponse\0225\n\nstopSe" +
|
"\022>\n\rGetServerInfo\022\025.GetServerInfoRequest" +
|
||||||
"rver\022\022.StopServerRequest\032\023.StopServerRes" +
|
"\032\026.GetServerInfoResponse\0225\n\nStopServer\022\022" +
|
||||||
"ponseBA\n*org.apache.hadoop.hbase.protobu" +
|
".StopServerRequest\032\023.StopServerResponseB" +
|
||||||
"f.generatedB\013AdminProtosH\001\210\001\001\240\001\001"
|
"A\n*org.apache.hadoop.hbase.protobuf.gene" +
|
||||||
|
"ratedB\013AdminProtosH\001\210\001\001\240\001\001"
|
||||||
};
|
};
|
||||||
com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
|
com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
|
||||||
new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() {
|
new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() {
|
||||||
|
|
|
@ -11,7 +11,7 @@ public final class AggregateProtos {
|
||||||
public interface AggregateArgumentOrBuilder
|
public interface AggregateArgumentOrBuilder
|
||||||
extends com.google.protobuf.MessageOrBuilder {
|
extends com.google.protobuf.MessageOrBuilder {
|
||||||
|
|
||||||
// required string interpreterClassName = 1;
|
// required string interpreter_class_name = 1;
|
||||||
boolean hasInterpreterClassName();
|
boolean hasInterpreterClassName();
|
||||||
String getInterpreterClassName();
|
String getInterpreterClassName();
|
||||||
|
|
||||||
|
@ -20,7 +20,7 @@ public final class AggregateProtos {
|
||||||
org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Scan getScan();
|
org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Scan getScan();
|
||||||
org.apache.hadoop.hbase.protobuf.generated.ClientProtos.ScanOrBuilder getScanOrBuilder();
|
org.apache.hadoop.hbase.protobuf.generated.ClientProtos.ScanOrBuilder getScanOrBuilder();
|
||||||
|
|
||||||
// optional bytes interpreterSpecificBytes = 3;
|
// optional bytes interpreter_specific_bytes = 3;
|
||||||
boolean hasInterpreterSpecificBytes();
|
boolean hasInterpreterSpecificBytes();
|
||||||
com.google.protobuf.ByteString getInterpreterSpecificBytes();
|
com.google.protobuf.ByteString getInterpreterSpecificBytes();
|
||||||
}
|
}
|
||||||
|
@ -53,8 +53,8 @@ public final class AggregateProtos {
|
||||||
}
|
}
|
||||||
|
|
||||||
private int bitField0_;
|
private int bitField0_;
|
||||||
// required string interpreterClassName = 1;
|
// required string interpreter_class_name = 1;
|
||||||
public static final int INTERPRETERCLASSNAME_FIELD_NUMBER = 1;
|
public static final int INTERPRETER_CLASS_NAME_FIELD_NUMBER = 1;
|
||||||
private java.lang.Object interpreterClassName_;
|
private java.lang.Object interpreterClassName_;
|
||||||
public boolean hasInterpreterClassName() {
|
public boolean hasInterpreterClassName() {
|
||||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||||
|
@ -98,8 +98,8 @@ public final class AggregateProtos {
|
||||||
return scan_;
|
return scan_;
|
||||||
}
|
}
|
||||||
|
|
||||||
// optional bytes interpreterSpecificBytes = 3;
|
// optional bytes interpreter_specific_bytes = 3;
|
||||||
public static final int INTERPRETERSPECIFICBYTES_FIELD_NUMBER = 3;
|
public static final int INTERPRETER_SPECIFIC_BYTES_FIELD_NUMBER = 3;
|
||||||
private com.google.protobuf.ByteString interpreterSpecificBytes_;
|
private com.google.protobuf.ByteString interpreterSpecificBytes_;
|
||||||
public boolean hasInterpreterSpecificBytes() {
|
public boolean hasInterpreterSpecificBytes() {
|
||||||
return ((bitField0_ & 0x00000004) == 0x00000004);
|
return ((bitField0_ & 0x00000004) == 0x00000004);
|
||||||
|
@ -215,7 +215,7 @@ public final class AggregateProtos {
|
||||||
int hash = 41;
|
int hash = 41;
|
||||||
hash = (19 * hash) + getDescriptorForType().hashCode();
|
hash = (19 * hash) + getDescriptorForType().hashCode();
|
||||||
if (hasInterpreterClassName()) {
|
if (hasInterpreterClassName()) {
|
||||||
hash = (37 * hash) + INTERPRETERCLASSNAME_FIELD_NUMBER;
|
hash = (37 * hash) + INTERPRETER_CLASS_NAME_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + getInterpreterClassName().hashCode();
|
hash = (53 * hash) + getInterpreterClassName().hashCode();
|
||||||
}
|
}
|
||||||
if (hasScan()) {
|
if (hasScan()) {
|
||||||
|
@ -223,7 +223,7 @@ public final class AggregateProtos {
|
||||||
hash = (53 * hash) + getScan().hashCode();
|
hash = (53 * hash) + getScan().hashCode();
|
||||||
}
|
}
|
||||||
if (hasInterpreterSpecificBytes()) {
|
if (hasInterpreterSpecificBytes()) {
|
||||||
hash = (37 * hash) + INTERPRETERSPECIFICBYTES_FIELD_NUMBER;
|
hash = (37 * hash) + INTERPRETER_SPECIFIC_BYTES_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + getInterpreterSpecificBytes().hashCode();
|
hash = (53 * hash) + getInterpreterSpecificBytes().hashCode();
|
||||||
}
|
}
|
||||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||||
|
@ -500,7 +500,7 @@ public final class AggregateProtos {
|
||||||
|
|
||||||
private int bitField0_;
|
private int bitField0_;
|
||||||
|
|
||||||
// required string interpreterClassName = 1;
|
// required string interpreter_class_name = 1;
|
||||||
private java.lang.Object interpreterClassName_ = "";
|
private java.lang.Object interpreterClassName_ = "";
|
||||||
public boolean hasInterpreterClassName() {
|
public boolean hasInterpreterClassName() {
|
||||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||||
|
@ -626,7 +626,7 @@ public final class AggregateProtos {
|
||||||
return scanBuilder_;
|
return scanBuilder_;
|
||||||
}
|
}
|
||||||
|
|
||||||
// optional bytes interpreterSpecificBytes = 3;
|
// optional bytes interpreter_specific_bytes = 3;
|
||||||
private com.google.protobuf.ByteString interpreterSpecificBytes_ = com.google.protobuf.ByteString.EMPTY;
|
private com.google.protobuf.ByteString interpreterSpecificBytes_ = com.google.protobuf.ByteString.EMPTY;
|
||||||
public boolean hasInterpreterSpecificBytes() {
|
public boolean hasInterpreterSpecificBytes() {
|
||||||
return ((bitField0_ & 0x00000004) == 0x00000004);
|
return ((bitField0_ & 0x00000004) == 0x00000004);
|
||||||
|
@ -664,12 +664,12 @@ public final class AggregateProtos {
|
||||||
public interface AggregateResponseOrBuilder
|
public interface AggregateResponseOrBuilder
|
||||||
extends com.google.protobuf.MessageOrBuilder {
|
extends com.google.protobuf.MessageOrBuilder {
|
||||||
|
|
||||||
// repeated bytes firstPart = 1;
|
// repeated bytes first_part = 1;
|
||||||
java.util.List<com.google.protobuf.ByteString> getFirstPartList();
|
java.util.List<com.google.protobuf.ByteString> getFirstPartList();
|
||||||
int getFirstPartCount();
|
int getFirstPartCount();
|
||||||
com.google.protobuf.ByteString getFirstPart(int index);
|
com.google.protobuf.ByteString getFirstPart(int index);
|
||||||
|
|
||||||
// optional bytes secondPart = 2;
|
// optional bytes second_part = 2;
|
||||||
boolean hasSecondPart();
|
boolean hasSecondPart();
|
||||||
com.google.protobuf.ByteString getSecondPart();
|
com.google.protobuf.ByteString getSecondPart();
|
||||||
}
|
}
|
||||||
|
@ -702,8 +702,8 @@ public final class AggregateProtos {
|
||||||
}
|
}
|
||||||
|
|
||||||
private int bitField0_;
|
private int bitField0_;
|
||||||
// repeated bytes firstPart = 1;
|
// repeated bytes first_part = 1;
|
||||||
public static final int FIRSTPART_FIELD_NUMBER = 1;
|
public static final int FIRST_PART_FIELD_NUMBER = 1;
|
||||||
private java.util.List<com.google.protobuf.ByteString> firstPart_;
|
private java.util.List<com.google.protobuf.ByteString> firstPart_;
|
||||||
public java.util.List<com.google.protobuf.ByteString>
|
public java.util.List<com.google.protobuf.ByteString>
|
||||||
getFirstPartList() {
|
getFirstPartList() {
|
||||||
|
@ -716,8 +716,8 @@ public final class AggregateProtos {
|
||||||
return firstPart_.get(index);
|
return firstPart_.get(index);
|
||||||
}
|
}
|
||||||
|
|
||||||
// optional bytes secondPart = 2;
|
// optional bytes second_part = 2;
|
||||||
public static final int SECONDPART_FIELD_NUMBER = 2;
|
public static final int SECOND_PART_FIELD_NUMBER = 2;
|
||||||
private com.google.protobuf.ByteString secondPart_;
|
private com.google.protobuf.ByteString secondPart_;
|
||||||
public boolean hasSecondPart() {
|
public boolean hasSecondPart() {
|
||||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||||
|
@ -810,11 +810,11 @@ public final class AggregateProtos {
|
||||||
int hash = 41;
|
int hash = 41;
|
||||||
hash = (19 * hash) + getDescriptorForType().hashCode();
|
hash = (19 * hash) + getDescriptorForType().hashCode();
|
||||||
if (getFirstPartCount() > 0) {
|
if (getFirstPartCount() > 0) {
|
||||||
hash = (37 * hash) + FIRSTPART_FIELD_NUMBER;
|
hash = (37 * hash) + FIRST_PART_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + getFirstPartList().hashCode();
|
hash = (53 * hash) + getFirstPartList().hashCode();
|
||||||
}
|
}
|
||||||
if (hasSecondPart()) {
|
if (hasSecondPart()) {
|
||||||
hash = (37 * hash) + SECONDPART_FIELD_NUMBER;
|
hash = (37 * hash) + SECOND_PART_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + getSecondPart().hashCode();
|
hash = (53 * hash) + getSecondPart().hashCode();
|
||||||
}
|
}
|
||||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||||
|
@ -1060,7 +1060,7 @@ public final class AggregateProtos {
|
||||||
|
|
||||||
private int bitField0_;
|
private int bitField0_;
|
||||||
|
|
||||||
// repeated bytes firstPart = 1;
|
// repeated bytes first_part = 1;
|
||||||
private java.util.List<com.google.protobuf.ByteString> firstPart_ = java.util.Collections.emptyList();;
|
private java.util.List<com.google.protobuf.ByteString> firstPart_ = java.util.Collections.emptyList();;
|
||||||
private void ensureFirstPartIsMutable() {
|
private void ensureFirstPartIsMutable() {
|
||||||
if (!((bitField0_ & 0x00000001) == 0x00000001)) {
|
if (!((bitField0_ & 0x00000001) == 0x00000001)) {
|
||||||
|
@ -1111,7 +1111,7 @@ public final class AggregateProtos {
|
||||||
return this;
|
return this;
|
||||||
}
|
}
|
||||||
|
|
||||||
// optional bytes secondPart = 2;
|
// optional bytes second_part = 2;
|
||||||
private com.google.protobuf.ByteString secondPart_ = com.google.protobuf.ByteString.EMPTY;
|
private com.google.protobuf.ByteString secondPart_ = com.google.protobuf.ByteString.EMPTY;
|
||||||
public boolean hasSecondPart() {
|
public boolean hasSecondPart() {
|
||||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||||
|
@ -1776,23 +1776,23 @@ public final class AggregateProtos {
|
||||||
descriptor;
|
descriptor;
|
||||||
static {
|
static {
|
||||||
java.lang.String[] descriptorData = {
|
java.lang.String[] descriptorData = {
|
||||||
"\n\017Aggregate.proto\032\014Client.proto\"h\n\021Aggre" +
|
"\n\017Aggregate.proto\032\014Client.proto\"l\n\021Aggre" +
|
||||||
"gateArgument\022\034\n\024interpreterClassName\030\001 \002" +
|
"gateArgument\022\036\n\026interpreter_class_name\030\001" +
|
||||||
"(\t\022\023\n\004scan\030\002 \002(\0132\005.Scan\022 \n\030interpreterSp" +
|
" \002(\t\022\023\n\004scan\030\002 \002(\0132\005.Scan\022\"\n\032interpreter" +
|
||||||
"ecificBytes\030\003 \001(\014\":\n\021AggregateResponse\022\021" +
|
"_specific_bytes\030\003 \001(\014\"<\n\021AggregateRespon" +
|
||||||
"\n\tfirstPart\030\001 \003(\014\022\022\n\nsecondPart\030\002 \001(\0142\366\002" +
|
"se\022\022\n\nfirst_part\030\001 \003(\014\022\023\n\013second_part\030\002 " +
|
||||||
"\n\020AggregateService\0220\n\006getMax\022\022.Aggregate" +
|
"\001(\0142\366\002\n\020AggregateService\0220\n\006GetMax\022\022.Agg" +
|
||||||
"Argument\032\022.AggregateResponse\0220\n\006getMin\022\022" +
|
"regateArgument\032\022.AggregateResponse\0220\n\006Ge" +
|
||||||
".AggregateArgument\032\022.AggregateResponse\0220" +
|
"tMin\022\022.AggregateArgument\032\022.AggregateResp" +
|
||||||
"\n\006getSum\022\022.AggregateArgument\032\022.Aggregate" +
|
"onse\0220\n\006GetSum\022\022.AggregateArgument\032\022.Agg" +
|
||||||
"Response\0223\n\tgetRowNum\022\022.AggregateArgumen",
|
"regateResponse\0223\n\tGetRowNum\022\022.AggregateA",
|
||||||
"t\032\022.AggregateResponse\0220\n\006getAvg\022\022.Aggreg" +
|
"rgument\032\022.AggregateResponse\0220\n\006GetAvg\022\022." +
|
||||||
"ateArgument\032\022.AggregateResponse\0220\n\006getSt" +
|
"AggregateArgument\032\022.AggregateResponse\0220\n" +
|
||||||
"d\022\022.AggregateArgument\032\022.AggregateRespons" +
|
"\006GetStd\022\022.AggregateArgument\032\022.AggregateR" +
|
||||||
"e\0223\n\tgetMedian\022\022.AggregateArgument\032\022.Agg" +
|
"esponse\0223\n\tGetMedian\022\022.AggregateArgument" +
|
||||||
"regateResponseBE\n*org.apache.hadoop.hbas" +
|
"\032\022.AggregateResponseBE\n*org.apache.hadoo" +
|
||||||
"e.protobuf.generatedB\017AggregateProtosH\001\210" +
|
"p.hbase.protobuf.generatedB\017AggregatePro" +
|
||||||
"\001\001\240\001\001"
|
"tosH\001\210\001\001\240\001\001"
|
||||||
};
|
};
|
||||||
com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
|
com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
|
||||||
new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() {
|
new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() {
|
||||||
|
|
|
@ -15,7 +15,7 @@ public final class AuthenticationProtos {
|
||||||
boolean hasId();
|
boolean hasId();
|
||||||
int getId();
|
int getId();
|
||||||
|
|
||||||
// required int64 expirationDate = 2;
|
// required int64 expiration_date = 2;
|
||||||
boolean hasExpirationDate();
|
boolean hasExpirationDate();
|
||||||
long getExpirationDate();
|
long getExpirationDate();
|
||||||
|
|
||||||
|
@ -62,8 +62,8 @@ public final class AuthenticationProtos {
|
||||||
return id_;
|
return id_;
|
||||||
}
|
}
|
||||||
|
|
||||||
// required int64 expirationDate = 2;
|
// required int64 expiration_date = 2;
|
||||||
public static final int EXPIRATIONDATE_FIELD_NUMBER = 2;
|
public static final int EXPIRATION_DATE_FIELD_NUMBER = 2;
|
||||||
private long expirationDate_;
|
private long expirationDate_;
|
||||||
public boolean hasExpirationDate() {
|
public boolean hasExpirationDate() {
|
||||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||||
|
@ -193,7 +193,7 @@ public final class AuthenticationProtos {
|
||||||
hash = (53 * hash) + getId();
|
hash = (53 * hash) + getId();
|
||||||
}
|
}
|
||||||
if (hasExpirationDate()) {
|
if (hasExpirationDate()) {
|
||||||
hash = (37 * hash) + EXPIRATIONDATE_FIELD_NUMBER;
|
hash = (37 * hash) + EXPIRATION_DATE_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + hashLong(getExpirationDate());
|
hash = (53 * hash) + hashLong(getExpirationDate());
|
||||||
}
|
}
|
||||||
if (hasKey()) {
|
if (hasKey()) {
|
||||||
|
@ -482,7 +482,7 @@ public final class AuthenticationProtos {
|
||||||
return this;
|
return this;
|
||||||
}
|
}
|
||||||
|
|
||||||
// required int64 expirationDate = 2;
|
// required int64 expiration_date = 2;
|
||||||
private long expirationDate_ ;
|
private long expirationDate_ ;
|
||||||
public boolean hasExpirationDate() {
|
public boolean hasExpirationDate() {
|
||||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||||
|
@ -549,19 +549,19 @@ public final class AuthenticationProtos {
|
||||||
boolean hasUsername();
|
boolean hasUsername();
|
||||||
com.google.protobuf.ByteString getUsername();
|
com.google.protobuf.ByteString getUsername();
|
||||||
|
|
||||||
// required int32 keyId = 3;
|
// required int32 key_id = 3;
|
||||||
boolean hasKeyId();
|
boolean hasKeyId();
|
||||||
int getKeyId();
|
int getKeyId();
|
||||||
|
|
||||||
// optional int64 issueDate = 4;
|
// optional int64 issue_date = 4;
|
||||||
boolean hasIssueDate();
|
boolean hasIssueDate();
|
||||||
long getIssueDate();
|
long getIssueDate();
|
||||||
|
|
||||||
// optional int64 expirationDate = 5;
|
// optional int64 expiration_date = 5;
|
||||||
boolean hasExpirationDate();
|
boolean hasExpirationDate();
|
||||||
long getExpirationDate();
|
long getExpirationDate();
|
||||||
|
|
||||||
// optional int64 sequenceNumber = 6;
|
// optional int64 sequence_number = 6;
|
||||||
boolean hasSequenceNumber();
|
boolean hasSequenceNumber();
|
||||||
long getSequenceNumber();
|
long getSequenceNumber();
|
||||||
}
|
}
|
||||||
|
@ -680,8 +680,8 @@ public final class AuthenticationProtos {
|
||||||
return username_;
|
return username_;
|
||||||
}
|
}
|
||||||
|
|
||||||
// required int32 keyId = 3;
|
// required int32 key_id = 3;
|
||||||
public static final int KEYID_FIELD_NUMBER = 3;
|
public static final int KEY_ID_FIELD_NUMBER = 3;
|
||||||
private int keyId_;
|
private int keyId_;
|
||||||
public boolean hasKeyId() {
|
public boolean hasKeyId() {
|
||||||
return ((bitField0_ & 0x00000004) == 0x00000004);
|
return ((bitField0_ & 0x00000004) == 0x00000004);
|
||||||
|
@ -690,8 +690,8 @@ public final class AuthenticationProtos {
|
||||||
return keyId_;
|
return keyId_;
|
||||||
}
|
}
|
||||||
|
|
||||||
// optional int64 issueDate = 4;
|
// optional int64 issue_date = 4;
|
||||||
public static final int ISSUEDATE_FIELD_NUMBER = 4;
|
public static final int ISSUE_DATE_FIELD_NUMBER = 4;
|
||||||
private long issueDate_;
|
private long issueDate_;
|
||||||
public boolean hasIssueDate() {
|
public boolean hasIssueDate() {
|
||||||
return ((bitField0_ & 0x00000008) == 0x00000008);
|
return ((bitField0_ & 0x00000008) == 0x00000008);
|
||||||
|
@ -700,8 +700,8 @@ public final class AuthenticationProtos {
|
||||||
return issueDate_;
|
return issueDate_;
|
||||||
}
|
}
|
||||||
|
|
||||||
// optional int64 expirationDate = 5;
|
// optional int64 expiration_date = 5;
|
||||||
public static final int EXPIRATIONDATE_FIELD_NUMBER = 5;
|
public static final int EXPIRATION_DATE_FIELD_NUMBER = 5;
|
||||||
private long expirationDate_;
|
private long expirationDate_;
|
||||||
public boolean hasExpirationDate() {
|
public boolean hasExpirationDate() {
|
||||||
return ((bitField0_ & 0x00000010) == 0x00000010);
|
return ((bitField0_ & 0x00000010) == 0x00000010);
|
||||||
|
@ -710,8 +710,8 @@ public final class AuthenticationProtos {
|
||||||
return expirationDate_;
|
return expirationDate_;
|
||||||
}
|
}
|
||||||
|
|
||||||
// optional int64 sequenceNumber = 6;
|
// optional int64 sequence_number = 6;
|
||||||
public static final int SEQUENCENUMBER_FIELD_NUMBER = 6;
|
public static final int SEQUENCE_NUMBER_FIELD_NUMBER = 6;
|
||||||
private long sequenceNumber_;
|
private long sequenceNumber_;
|
||||||
public boolean hasSequenceNumber() {
|
public boolean hasSequenceNumber() {
|
||||||
return ((bitField0_ & 0x00000020) == 0x00000020);
|
return ((bitField0_ & 0x00000020) == 0x00000020);
|
||||||
|
@ -874,19 +874,19 @@ public final class AuthenticationProtos {
|
||||||
hash = (53 * hash) + getUsername().hashCode();
|
hash = (53 * hash) + getUsername().hashCode();
|
||||||
}
|
}
|
||||||
if (hasKeyId()) {
|
if (hasKeyId()) {
|
||||||
hash = (37 * hash) + KEYID_FIELD_NUMBER;
|
hash = (37 * hash) + KEY_ID_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + getKeyId();
|
hash = (53 * hash) + getKeyId();
|
||||||
}
|
}
|
||||||
if (hasIssueDate()) {
|
if (hasIssueDate()) {
|
||||||
hash = (37 * hash) + ISSUEDATE_FIELD_NUMBER;
|
hash = (37 * hash) + ISSUE_DATE_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + hashLong(getIssueDate());
|
hash = (53 * hash) + hashLong(getIssueDate());
|
||||||
}
|
}
|
||||||
if (hasExpirationDate()) {
|
if (hasExpirationDate()) {
|
||||||
hash = (37 * hash) + EXPIRATIONDATE_FIELD_NUMBER;
|
hash = (37 * hash) + EXPIRATION_DATE_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + hashLong(getExpirationDate());
|
hash = (53 * hash) + hashLong(getExpirationDate());
|
||||||
}
|
}
|
||||||
if (hasSequenceNumber()) {
|
if (hasSequenceNumber()) {
|
||||||
hash = (37 * hash) + SEQUENCENUMBER_FIELD_NUMBER;
|
hash = (37 * hash) + SEQUENCE_NUMBER_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + hashLong(getSequenceNumber());
|
hash = (53 * hash) + hashLong(getSequenceNumber());
|
||||||
}
|
}
|
||||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||||
|
@ -1246,7 +1246,7 @@ public final class AuthenticationProtos {
|
||||||
return this;
|
return this;
|
||||||
}
|
}
|
||||||
|
|
||||||
// required int32 keyId = 3;
|
// required int32 key_id = 3;
|
||||||
private int keyId_ ;
|
private int keyId_ ;
|
||||||
public boolean hasKeyId() {
|
public boolean hasKeyId() {
|
||||||
return ((bitField0_ & 0x00000004) == 0x00000004);
|
return ((bitField0_ & 0x00000004) == 0x00000004);
|
||||||
|
@ -1267,7 +1267,7 @@ public final class AuthenticationProtos {
|
||||||
return this;
|
return this;
|
||||||
}
|
}
|
||||||
|
|
||||||
// optional int64 issueDate = 4;
|
// optional int64 issue_date = 4;
|
||||||
private long issueDate_ ;
|
private long issueDate_ ;
|
||||||
public boolean hasIssueDate() {
|
public boolean hasIssueDate() {
|
||||||
return ((bitField0_ & 0x00000008) == 0x00000008);
|
return ((bitField0_ & 0x00000008) == 0x00000008);
|
||||||
|
@ -1288,7 +1288,7 @@ public final class AuthenticationProtos {
|
||||||
return this;
|
return this;
|
||||||
}
|
}
|
||||||
|
|
||||||
// optional int64 expirationDate = 5;
|
// optional int64 expiration_date = 5;
|
||||||
private long expirationDate_ ;
|
private long expirationDate_ ;
|
||||||
public boolean hasExpirationDate() {
|
public boolean hasExpirationDate() {
|
||||||
return ((bitField0_ & 0x00000010) == 0x00000010);
|
return ((bitField0_ & 0x00000010) == 0x00000010);
|
||||||
|
@ -1309,7 +1309,7 @@ public final class AuthenticationProtos {
|
||||||
return this;
|
return this;
|
||||||
}
|
}
|
||||||
|
|
||||||
// optional int64 sequenceNumber = 6;
|
// optional int64 sequence_number = 6;
|
||||||
private long sequenceNumber_ ;
|
private long sequenceNumber_ ;
|
||||||
public boolean hasSequenceNumber() {
|
public boolean hasSequenceNumber() {
|
||||||
return ((bitField0_ & 0x00000020) == 0x00000020);
|
return ((bitField0_ & 0x00000020) == 0x00000020);
|
||||||
|
@ -2917,7 +2917,7 @@ public final class AuthenticationProtos {
|
||||||
boolean hasUsername();
|
boolean hasUsername();
|
||||||
String getUsername();
|
String getUsername();
|
||||||
|
|
||||||
// optional string authMethod = 2;
|
// optional string auth_method = 2;
|
||||||
boolean hasAuthMethod();
|
boolean hasAuthMethod();
|
||||||
String getAuthMethod();
|
String getAuthMethod();
|
||||||
}
|
}
|
||||||
|
@ -2982,8 +2982,8 @@ public final class AuthenticationProtos {
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// optional string authMethod = 2;
|
// optional string auth_method = 2;
|
||||||
public static final int AUTHMETHOD_FIELD_NUMBER = 2;
|
public static final int AUTH_METHOD_FIELD_NUMBER = 2;
|
||||||
private java.lang.Object authMethod_;
|
private java.lang.Object authMethod_;
|
||||||
public boolean hasAuthMethod() {
|
public boolean hasAuthMethod() {
|
||||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||||
|
@ -3100,7 +3100,7 @@ public final class AuthenticationProtos {
|
||||||
hash = (53 * hash) + getUsername().hashCode();
|
hash = (53 * hash) + getUsername().hashCode();
|
||||||
}
|
}
|
||||||
if (hasAuthMethod()) {
|
if (hasAuthMethod()) {
|
||||||
hash = (37 * hash) + AUTHMETHOD_FIELD_NUMBER;
|
hash = (37 * hash) + AUTH_METHOD_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + getAuthMethod().hashCode();
|
hash = (53 * hash) + getAuthMethod().hashCode();
|
||||||
}
|
}
|
||||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||||
|
@ -3374,7 +3374,7 @@ public final class AuthenticationProtos {
|
||||||
onChanged();
|
onChanged();
|
||||||
}
|
}
|
||||||
|
|
||||||
// optional string authMethod = 2;
|
// optional string auth_method = 2;
|
||||||
private java.lang.Object authMethod_ = "";
|
private java.lang.Object authMethod_ = "";
|
||||||
public boolean hasAuthMethod() {
|
public boolean hasAuthMethod() {
|
||||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||||
|
@ -3431,7 +3431,7 @@ public final class AuthenticationProtos {
|
||||||
org.apache.hadoop.hbase.protobuf.generated.AuthenticationProtos.TokenRequest request,
|
org.apache.hadoop.hbase.protobuf.generated.AuthenticationProtos.TokenRequest request,
|
||||||
com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.protobuf.generated.AuthenticationProtos.TokenResponse> done);
|
com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.protobuf.generated.AuthenticationProtos.TokenResponse> done);
|
||||||
|
|
||||||
public abstract void whoami(
|
public abstract void whoAmI(
|
||||||
com.google.protobuf.RpcController controller,
|
com.google.protobuf.RpcController controller,
|
||||||
org.apache.hadoop.hbase.protobuf.generated.AuthenticationProtos.WhoAmIRequest request,
|
org.apache.hadoop.hbase.protobuf.generated.AuthenticationProtos.WhoAmIRequest request,
|
||||||
com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.protobuf.generated.AuthenticationProtos.WhoAmIResponse> done);
|
com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.protobuf.generated.AuthenticationProtos.WhoAmIResponse> done);
|
||||||
|
@ -3450,11 +3450,11 @@ public final class AuthenticationProtos {
|
||||||
}
|
}
|
||||||
|
|
||||||
@java.lang.Override
|
@java.lang.Override
|
||||||
public void whoami(
|
public void whoAmI(
|
||||||
com.google.protobuf.RpcController controller,
|
com.google.protobuf.RpcController controller,
|
||||||
org.apache.hadoop.hbase.protobuf.generated.AuthenticationProtos.WhoAmIRequest request,
|
org.apache.hadoop.hbase.protobuf.generated.AuthenticationProtos.WhoAmIRequest request,
|
||||||
com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.protobuf.generated.AuthenticationProtos.WhoAmIResponse> done) {
|
com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.protobuf.generated.AuthenticationProtos.WhoAmIResponse> done) {
|
||||||
impl.whoami(controller, request, done);
|
impl.whoAmI(controller, request, done);
|
||||||
}
|
}
|
||||||
|
|
||||||
};
|
};
|
||||||
|
@ -3482,7 +3482,7 @@ public final class AuthenticationProtos {
|
||||||
case 0:
|
case 0:
|
||||||
return impl.getAuthenticationToken(controller, (org.apache.hadoop.hbase.protobuf.generated.AuthenticationProtos.TokenRequest)request);
|
return impl.getAuthenticationToken(controller, (org.apache.hadoop.hbase.protobuf.generated.AuthenticationProtos.TokenRequest)request);
|
||||||
case 1:
|
case 1:
|
||||||
return impl.whoami(controller, (org.apache.hadoop.hbase.protobuf.generated.AuthenticationProtos.WhoAmIRequest)request);
|
return impl.whoAmI(controller, (org.apache.hadoop.hbase.protobuf.generated.AuthenticationProtos.WhoAmIRequest)request);
|
||||||
default:
|
default:
|
||||||
throw new java.lang.AssertionError("Can't get here.");
|
throw new java.lang.AssertionError("Can't get here.");
|
||||||
}
|
}
|
||||||
|
@ -3532,7 +3532,7 @@ public final class AuthenticationProtos {
|
||||||
org.apache.hadoop.hbase.protobuf.generated.AuthenticationProtos.TokenRequest request,
|
org.apache.hadoop.hbase.protobuf.generated.AuthenticationProtos.TokenRequest request,
|
||||||
com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.protobuf.generated.AuthenticationProtos.TokenResponse> done);
|
com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.protobuf.generated.AuthenticationProtos.TokenResponse> done);
|
||||||
|
|
||||||
public abstract void whoami(
|
public abstract void whoAmI(
|
||||||
com.google.protobuf.RpcController controller,
|
com.google.protobuf.RpcController controller,
|
||||||
org.apache.hadoop.hbase.protobuf.generated.AuthenticationProtos.WhoAmIRequest request,
|
org.apache.hadoop.hbase.protobuf.generated.AuthenticationProtos.WhoAmIRequest request,
|
||||||
com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.protobuf.generated.AuthenticationProtos.WhoAmIResponse> done);
|
com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.protobuf.generated.AuthenticationProtos.WhoAmIResponse> done);
|
||||||
|
@ -3565,7 +3565,7 @@ public final class AuthenticationProtos {
|
||||||
done));
|
done));
|
||||||
return;
|
return;
|
||||||
case 1:
|
case 1:
|
||||||
this.whoami(controller, (org.apache.hadoop.hbase.protobuf.generated.AuthenticationProtos.WhoAmIRequest)request,
|
this.whoAmI(controller, (org.apache.hadoop.hbase.protobuf.generated.AuthenticationProtos.WhoAmIRequest)request,
|
||||||
com.google.protobuf.RpcUtil.<org.apache.hadoop.hbase.protobuf.generated.AuthenticationProtos.WhoAmIResponse>specializeCallback(
|
com.google.protobuf.RpcUtil.<org.apache.hadoop.hbase.protobuf.generated.AuthenticationProtos.WhoAmIResponse>specializeCallback(
|
||||||
done));
|
done));
|
||||||
return;
|
return;
|
||||||
|
@ -3641,7 +3641,7 @@ public final class AuthenticationProtos {
|
||||||
org.apache.hadoop.hbase.protobuf.generated.AuthenticationProtos.TokenResponse.getDefaultInstance()));
|
org.apache.hadoop.hbase.protobuf.generated.AuthenticationProtos.TokenResponse.getDefaultInstance()));
|
||||||
}
|
}
|
||||||
|
|
||||||
public void whoami(
|
public void whoAmI(
|
||||||
com.google.protobuf.RpcController controller,
|
com.google.protobuf.RpcController controller,
|
||||||
org.apache.hadoop.hbase.protobuf.generated.AuthenticationProtos.WhoAmIRequest request,
|
org.apache.hadoop.hbase.protobuf.generated.AuthenticationProtos.WhoAmIRequest request,
|
||||||
com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.protobuf.generated.AuthenticationProtos.WhoAmIResponse> done) {
|
com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.protobuf.generated.AuthenticationProtos.WhoAmIResponse> done) {
|
||||||
|
@ -3668,7 +3668,7 @@ public final class AuthenticationProtos {
|
||||||
org.apache.hadoop.hbase.protobuf.generated.AuthenticationProtos.TokenRequest request)
|
org.apache.hadoop.hbase.protobuf.generated.AuthenticationProtos.TokenRequest request)
|
||||||
throws com.google.protobuf.ServiceException;
|
throws com.google.protobuf.ServiceException;
|
||||||
|
|
||||||
public org.apache.hadoop.hbase.protobuf.generated.AuthenticationProtos.WhoAmIResponse whoami(
|
public org.apache.hadoop.hbase.protobuf.generated.AuthenticationProtos.WhoAmIResponse whoAmI(
|
||||||
com.google.protobuf.RpcController controller,
|
com.google.protobuf.RpcController controller,
|
||||||
org.apache.hadoop.hbase.protobuf.generated.AuthenticationProtos.WhoAmIRequest request)
|
org.apache.hadoop.hbase.protobuf.generated.AuthenticationProtos.WhoAmIRequest request)
|
||||||
throws com.google.protobuf.ServiceException;
|
throws com.google.protobuf.ServiceException;
|
||||||
|
@ -3693,7 +3693,7 @@ public final class AuthenticationProtos {
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
||||||
public org.apache.hadoop.hbase.protobuf.generated.AuthenticationProtos.WhoAmIResponse whoami(
|
public org.apache.hadoop.hbase.protobuf.generated.AuthenticationProtos.WhoAmIResponse whoAmI(
|
||||||
com.google.protobuf.RpcController controller,
|
com.google.protobuf.RpcController controller,
|
||||||
org.apache.hadoop.hbase.protobuf.generated.AuthenticationProtos.WhoAmIRequest request)
|
org.apache.hadoop.hbase.protobuf.generated.AuthenticationProtos.WhoAmIRequest request)
|
||||||
throws com.google.protobuf.ServiceException {
|
throws com.google.protobuf.ServiceException {
|
||||||
|
@ -3751,23 +3751,23 @@ public final class AuthenticationProtos {
|
||||||
descriptor;
|
descriptor;
|
||||||
static {
|
static {
|
||||||
java.lang.String[] descriptorData = {
|
java.lang.String[] descriptorData = {
|
||||||
"\n\024Authentication.proto\"D\n\021Authentication" +
|
"\n\024Authentication.proto\"E\n\021Authentication" +
|
||||||
"Key\022\n\n\002id\030\001 \002(\005\022\026\n\016expirationDate\030\002 \002(\003\022" +
|
"Key\022\n\n\002id\030\001 \002(\005\022\027\n\017expiration_date\030\002 \002(\003" +
|
||||||
"\013\n\003key\030\003 \002(\014\"\270\001\n\017TokenIdentifier\022#\n\004kind" +
|
"\022\013\n\003key\030\003 \002(\014\"\274\001\n\017TokenIdentifier\022#\n\004kin" +
|
||||||
"\030\001 \002(\0162\025.TokenIdentifier.Kind\022\020\n\010usernam" +
|
"d\030\001 \002(\0162\025.TokenIdentifier.Kind\022\020\n\010userna" +
|
||||||
"e\030\002 \002(\014\022\r\n\005keyId\030\003 \002(\005\022\021\n\tissueDate\030\004 \001(" +
|
"me\030\002 \002(\014\022\016\n\006key_id\030\003 \002(\005\022\022\n\nissue_date\030\004" +
|
||||||
"\003\022\026\n\016expirationDate\030\005 \001(\003\022\026\n\016sequenceNum" +
|
" \001(\003\022\027\n\017expiration_date\030\005 \001(\003\022\027\n\017sequenc" +
|
||||||
"ber\030\006 \001(\003\"\034\n\004Kind\022\024\n\020HBASE_AUTH_TOKEN\020\000\"" +
|
"e_number\030\006 \001(\003\"\034\n\004Kind\022\024\n\020HBASE_AUTH_TOK" +
|
||||||
">\n\005Token\022\022\n\nidentifier\030\001 \001(\014\022\020\n\010password" +
|
"EN\020\000\">\n\005Token\022\022\n\nidentifier\030\001 \001(\014\022\020\n\010pas" +
|
||||||
"\030\002 \001(\014\022\017\n\007service\030\003 \001(\014\"\016\n\014TokenRequest\"" +
|
"sword\030\002 \001(\014\022\017\n\007service\030\003 \001(\014\"\016\n\014TokenReq" +
|
||||||
"&\n\rTokenResponse\022\025\n\005token\030\001 \001(\0132\006.Token\"",
|
"uest\"&\n\rTokenResponse\022\025\n\005token\030\001 \001(\0132\006.T",
|
||||||
"\017\n\rWhoAmIRequest\"6\n\016WhoAmIResponse\022\020\n\010us" +
|
"oken\"\017\n\rWhoAmIRequest\"7\n\016WhoAmIResponse\022" +
|
||||||
"ername\030\001 \001(\t\022\022\n\nauthMethod\030\002 \001(\t2{\n\025Auth" +
|
"\020\n\010username\030\001 \001(\t\022\023\n\013auth_method\030\002 \001(\t2{" +
|
||||||
"enticationService\0227\n\026getAuthenticationTo" +
|
"\n\025AuthenticationService\0227\n\026GetAuthentica" +
|
||||||
"ken\022\r.TokenRequest\032\016.TokenResponse\022)\n\006wh" +
|
"tionToken\022\r.TokenRequest\032\016.TokenResponse" +
|
||||||
"oami\022\016.WhoAmIRequest\032\017.WhoAmIResponseBJ\n" +
|
"\022)\n\006WhoAmI\022\016.WhoAmIRequest\032\017.WhoAmIRespo" +
|
||||||
"*org.apache.hadoop.hbase.protobuf.genera" +
|
"nseBJ\n*org.apache.hadoop.hbase.protobuf." +
|
||||||
"tedB\024AuthenticationProtosH\001\210\001\001\240\001\001"
|
"generatedB\024AuthenticationProtosH\001\210\001\001\240\001\001"
|
||||||
};
|
};
|
||||||
com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
|
com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
|
||||||
new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() {
|
new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() {
|
||||||
|
|
|
@ -108,7 +108,7 @@ public final class CellProtos {
|
||||||
boolean hasTimestamp();
|
boolean hasTimestamp();
|
||||||
long getTimestamp();
|
long getTimestamp();
|
||||||
|
|
||||||
// optional .CellType cellType = 5;
|
// optional .CellType cell_type = 5;
|
||||||
boolean hasCellType();
|
boolean hasCellType();
|
||||||
org.apache.hadoop.hbase.protobuf.generated.CellProtos.CellType getCellType();
|
org.apache.hadoop.hbase.protobuf.generated.CellProtos.CellType getCellType();
|
||||||
|
|
||||||
|
@ -185,8 +185,8 @@ public final class CellProtos {
|
||||||
return timestamp_;
|
return timestamp_;
|
||||||
}
|
}
|
||||||
|
|
||||||
// optional .CellType cellType = 5;
|
// optional .CellType cell_type = 5;
|
||||||
public static final int CELLTYPE_FIELD_NUMBER = 5;
|
public static final int CELL_TYPE_FIELD_NUMBER = 5;
|
||||||
private org.apache.hadoop.hbase.protobuf.generated.CellProtos.CellType cellType_;
|
private org.apache.hadoop.hbase.protobuf.generated.CellProtos.CellType cellType_;
|
||||||
public boolean hasCellType() {
|
public boolean hasCellType() {
|
||||||
return ((bitField0_ & 0x00000010) == 0x00000010);
|
return ((bitField0_ & 0x00000010) == 0x00000010);
|
||||||
|
@ -355,7 +355,7 @@ public final class CellProtos {
|
||||||
hash = (53 * hash) + hashLong(getTimestamp());
|
hash = (53 * hash) + hashLong(getTimestamp());
|
||||||
}
|
}
|
||||||
if (hasCellType()) {
|
if (hasCellType()) {
|
||||||
hash = (37 * hash) + CELLTYPE_FIELD_NUMBER;
|
hash = (37 * hash) + CELL_TYPE_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + hashEnum(getCellType());
|
hash = (53 * hash) + hashEnum(getCellType());
|
||||||
}
|
}
|
||||||
if (hasValue()) {
|
if (hasValue()) {
|
||||||
|
@ -752,7 +752,7 @@ public final class CellProtos {
|
||||||
return this;
|
return this;
|
||||||
}
|
}
|
||||||
|
|
||||||
// optional .CellType cellType = 5;
|
// optional .CellType cell_type = 5;
|
||||||
private org.apache.hadoop.hbase.protobuf.generated.CellProtos.CellType cellType_ = org.apache.hadoop.hbase.protobuf.generated.CellProtos.CellType.MINIMUM;
|
private org.apache.hadoop.hbase.protobuf.generated.CellProtos.CellType cellType_ = org.apache.hadoop.hbase.protobuf.generated.CellProtos.CellType.MINIMUM;
|
||||||
public boolean hasCellType() {
|
public boolean hasCellType() {
|
||||||
return ((bitField0_ & 0x00000010) == 0x00000010);
|
return ((bitField0_ & 0x00000010) == 0x00000010);
|
||||||
|
@ -830,7 +830,7 @@ public final class CellProtos {
|
||||||
boolean hasTimestamp();
|
boolean hasTimestamp();
|
||||||
long getTimestamp();
|
long getTimestamp();
|
||||||
|
|
||||||
// optional .CellType keyType = 5;
|
// optional .CellType key_type = 5;
|
||||||
boolean hasKeyType();
|
boolean hasKeyType();
|
||||||
org.apache.hadoop.hbase.protobuf.generated.CellProtos.CellType getKeyType();
|
org.apache.hadoop.hbase.protobuf.generated.CellProtos.CellType getKeyType();
|
||||||
|
|
||||||
|
@ -907,8 +907,8 @@ public final class CellProtos {
|
||||||
return timestamp_;
|
return timestamp_;
|
||||||
}
|
}
|
||||||
|
|
||||||
// optional .CellType keyType = 5;
|
// optional .CellType key_type = 5;
|
||||||
public static final int KEYTYPE_FIELD_NUMBER = 5;
|
public static final int KEY_TYPE_FIELD_NUMBER = 5;
|
||||||
private org.apache.hadoop.hbase.protobuf.generated.CellProtos.CellType keyType_;
|
private org.apache.hadoop.hbase.protobuf.generated.CellProtos.CellType keyType_;
|
||||||
public boolean hasKeyType() {
|
public boolean hasKeyType() {
|
||||||
return ((bitField0_ & 0x00000010) == 0x00000010);
|
return ((bitField0_ & 0x00000010) == 0x00000010);
|
||||||
|
@ -1089,7 +1089,7 @@ public final class CellProtos {
|
||||||
hash = (53 * hash) + hashLong(getTimestamp());
|
hash = (53 * hash) + hashLong(getTimestamp());
|
||||||
}
|
}
|
||||||
if (hasKeyType()) {
|
if (hasKeyType()) {
|
||||||
hash = (37 * hash) + KEYTYPE_FIELD_NUMBER;
|
hash = (37 * hash) + KEY_TYPE_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + hashEnum(getKeyType());
|
hash = (53 * hash) + hashEnum(getKeyType());
|
||||||
}
|
}
|
||||||
if (hasValue()) {
|
if (hasValue()) {
|
||||||
|
@ -1498,7 +1498,7 @@ public final class CellProtos {
|
||||||
return this;
|
return this;
|
||||||
}
|
}
|
||||||
|
|
||||||
// optional .CellType keyType = 5;
|
// optional .CellType key_type = 5;
|
||||||
private org.apache.hadoop.hbase.protobuf.generated.CellProtos.CellType keyType_ = org.apache.hadoop.hbase.protobuf.generated.CellProtos.CellType.MINIMUM;
|
private org.apache.hadoop.hbase.protobuf.generated.CellProtos.CellType keyType_ = org.apache.hadoop.hbase.protobuf.generated.CellProtos.CellType.MINIMUM;
|
||||||
public boolean hasKeyType() {
|
public boolean hasKeyType() {
|
||||||
return ((bitField0_ & 0x00000010) == 0x00000010);
|
return ((bitField0_ & 0x00000010) == 0x00000010);
|
||||||
|
@ -1576,17 +1576,17 @@ public final class CellProtos {
|
||||||
descriptor;
|
descriptor;
|
||||||
static {
|
static {
|
||||||
java.lang.String[] descriptorData = {
|
java.lang.String[] descriptorData = {
|
||||||
"\n\nCell.proto\"u\n\004Cell\022\013\n\003row\030\001 \001(\014\022\016\n\006fam" +
|
"\n\nCell.proto\"v\n\004Cell\022\013\n\003row\030\001 \001(\014\022\016\n\006fam" +
|
||||||
"ily\030\002 \001(\014\022\021\n\tqualifier\030\003 \001(\014\022\021\n\ttimestam" +
|
"ily\030\002 \001(\014\022\021\n\tqualifier\030\003 \001(\014\022\021\n\ttimestam" +
|
||||||
"p\030\004 \001(\004\022\033\n\010cellType\030\005 \001(\0162\t.CellType\022\r\n\005" +
|
"p\030\004 \001(\004\022\034\n\tcell_type\030\005 \001(\0162\t.CellType\022\r\n" +
|
||||||
"value\030\006 \001(\014\"x\n\010KeyValue\022\013\n\003row\030\001 \002(\014\022\016\n\006" +
|
"\005value\030\006 \001(\014\"y\n\010KeyValue\022\013\n\003row\030\001 \002(\014\022\016\n" +
|
||||||
"family\030\002 \002(\014\022\021\n\tqualifier\030\003 \002(\014\022\021\n\ttimes" +
|
"\006family\030\002 \002(\014\022\021\n\tqualifier\030\003 \002(\014\022\021\n\ttime" +
|
||||||
"tamp\030\004 \001(\004\022\032\n\007keyType\030\005 \001(\0162\t.CellType\022\r" +
|
"stamp\030\004 \001(\004\022\033\n\010key_type\030\005 \001(\0162\t.CellType" +
|
||||||
"\n\005value\030\006 \001(\014*`\n\010CellType\022\013\n\007MINIMUM\020\000\022\007" +
|
"\022\r\n\005value\030\006 \001(\014*`\n\010CellType\022\013\n\007MINIMUM\020\000" +
|
||||||
"\n\003PUT\020\004\022\n\n\006DELETE\020\010\022\021\n\rDELETE_COLUMN\020\014\022\021" +
|
"\022\007\n\003PUT\020\004\022\n\n\006DELETE\020\010\022\021\n\rDELETE_COLUMN\020\014" +
|
||||||
"\n\rDELETE_FAMILY\020\016\022\014\n\007MAXIMUM\020\377\001B=\n*org.a" +
|
"\022\021\n\rDELETE_FAMILY\020\016\022\014\n\007MAXIMUM\020\377\001B=\n*org" +
|
||||||
"pache.hadoop.hbase.protobuf.generatedB\nC",
|
".apache.hadoop.hbase.protobuf.generatedB",
|
||||||
"ellProtosH\001\240\001\001"
|
"\nCellProtosH\001\240\001\001"
|
||||||
};
|
};
|
||||||
com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
|
com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
|
||||||
new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() {
|
new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() {
|
||||||
|
|
File diff suppressed because it is too large
Load Diff
|
@ -11,7 +11,7 @@ public final class ClusterIdProtos {
|
||||||
public interface ClusterIdOrBuilder
|
public interface ClusterIdOrBuilder
|
||||||
extends com.google.protobuf.MessageOrBuilder {
|
extends com.google.protobuf.MessageOrBuilder {
|
||||||
|
|
||||||
// required string clusterId = 1;
|
// required string cluster_id = 1;
|
||||||
boolean hasClusterId();
|
boolean hasClusterId();
|
||||||
String getClusterId();
|
String getClusterId();
|
||||||
}
|
}
|
||||||
|
@ -44,8 +44,8 @@ public final class ClusterIdProtos {
|
||||||
}
|
}
|
||||||
|
|
||||||
private int bitField0_;
|
private int bitField0_;
|
||||||
// required string clusterId = 1;
|
// required string cluster_id = 1;
|
||||||
public static final int CLUSTERID_FIELD_NUMBER = 1;
|
public static final int CLUSTER_ID_FIELD_NUMBER = 1;
|
||||||
private java.lang.Object clusterId_;
|
private java.lang.Object clusterId_;
|
||||||
public boolean hasClusterId() {
|
public boolean hasClusterId() {
|
||||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||||
|
@ -149,7 +149,7 @@ public final class ClusterIdProtos {
|
||||||
int hash = 41;
|
int hash = 41;
|
||||||
hash = (19 * hash) + getDescriptorForType().hashCode();
|
hash = (19 * hash) + getDescriptorForType().hashCode();
|
||||||
if (hasClusterId()) {
|
if (hasClusterId()) {
|
||||||
hash = (37 * hash) + CLUSTERID_FIELD_NUMBER;
|
hash = (37 * hash) + CLUSTER_ID_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + getClusterId().hashCode();
|
hash = (53 * hash) + getClusterId().hashCode();
|
||||||
}
|
}
|
||||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||||
|
@ -377,7 +377,7 @@ public final class ClusterIdProtos {
|
||||||
|
|
||||||
private int bitField0_;
|
private int bitField0_;
|
||||||
|
|
||||||
// required string clusterId = 1;
|
// required string cluster_id = 1;
|
||||||
private java.lang.Object clusterId_ = "";
|
private java.lang.Object clusterId_ = "";
|
||||||
public boolean hasClusterId() {
|
public boolean hasClusterId() {
|
||||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||||
|
@ -438,9 +438,9 @@ public final class ClusterIdProtos {
|
||||||
descriptor;
|
descriptor;
|
||||||
static {
|
static {
|
||||||
java.lang.String[] descriptorData = {
|
java.lang.String[] descriptorData = {
|
||||||
"\n\017ClusterId.proto\"\036\n\tClusterId\022\021\n\tcluste" +
|
"\n\017ClusterId.proto\"\037\n\tClusterId\022\022\n\ncluste" +
|
||||||
"rId\030\001 \002(\tBB\n*org.apache.hadoop.hbase.pro" +
|
"r_id\030\001 \002(\tBB\n*org.apache.hadoop.hbase.pr" +
|
||||||
"tobuf.generatedB\017ClusterIdProtosH\001\240\001\001"
|
"otobuf.generatedB\017ClusterIdProtosH\001\240\001\001"
|
||||||
};
|
};
|
||||||
com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
|
com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
|
||||||
new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() {
|
new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() {
|
||||||
|
|
File diff suppressed because it is too large
Load Diff
|
@ -15,7 +15,7 @@ public final class ComparatorProtos {
|
||||||
boolean hasName();
|
boolean hasName();
|
||||||
String getName();
|
String getName();
|
||||||
|
|
||||||
// optional bytes serializedComparator = 2;
|
// optional bytes serialized_comparator = 2;
|
||||||
boolean hasSerializedComparator();
|
boolean hasSerializedComparator();
|
||||||
com.google.protobuf.ByteString getSerializedComparator();
|
com.google.protobuf.ByteString getSerializedComparator();
|
||||||
}
|
}
|
||||||
|
@ -80,8 +80,8 @@ public final class ComparatorProtos {
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// optional bytes serializedComparator = 2;
|
// optional bytes serialized_comparator = 2;
|
||||||
public static final int SERIALIZEDCOMPARATOR_FIELD_NUMBER = 2;
|
public static final int SERIALIZED_COMPARATOR_FIELD_NUMBER = 2;
|
||||||
private com.google.protobuf.ByteString serializedComparator_;
|
private com.google.protobuf.ByteString serializedComparator_;
|
||||||
public boolean hasSerializedComparator() {
|
public boolean hasSerializedComparator() {
|
||||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||||
|
@ -180,7 +180,7 @@ public final class ComparatorProtos {
|
||||||
hash = (53 * hash) + getName().hashCode();
|
hash = (53 * hash) + getName().hashCode();
|
||||||
}
|
}
|
||||||
if (hasSerializedComparator()) {
|
if (hasSerializedComparator()) {
|
||||||
hash = (37 * hash) + SERIALIZEDCOMPARATOR_FIELD_NUMBER;
|
hash = (37 * hash) + SERIALIZED_COMPARATOR_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + getSerializedComparator().hashCode();
|
hash = (53 * hash) + getSerializedComparator().hashCode();
|
||||||
}
|
}
|
||||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||||
|
@ -458,7 +458,7 @@ public final class ComparatorProtos {
|
||||||
onChanged();
|
onChanged();
|
||||||
}
|
}
|
||||||
|
|
||||||
// optional bytes serializedComparator = 2;
|
// optional bytes serialized_comparator = 2;
|
||||||
private com.google.protobuf.ByteString serializedComparator_ = com.google.protobuf.ByteString.EMPTY;
|
private com.google.protobuf.ByteString serializedComparator_ = com.google.protobuf.ByteString.EMPTY;
|
||||||
public boolean hasSerializedComparator() {
|
public boolean hasSerializedComparator() {
|
||||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||||
|
@ -1805,7 +1805,7 @@ public final class ComparatorProtos {
|
||||||
org.apache.hadoop.hbase.protobuf.generated.ComparatorProtos.ByteArrayComparable getComparable();
|
org.apache.hadoop.hbase.protobuf.generated.ComparatorProtos.ByteArrayComparable getComparable();
|
||||||
org.apache.hadoop.hbase.protobuf.generated.ComparatorProtos.ByteArrayComparableOrBuilder getComparableOrBuilder();
|
org.apache.hadoop.hbase.protobuf.generated.ComparatorProtos.ByteArrayComparableOrBuilder getComparableOrBuilder();
|
||||||
|
|
||||||
// required .BitComparator.BitwiseOp bitwiseOp = 2;
|
// required .BitComparator.BitwiseOp bitwise_op = 2;
|
||||||
boolean hasBitwiseOp();
|
boolean hasBitwiseOp();
|
||||||
org.apache.hadoop.hbase.protobuf.generated.ComparatorProtos.BitComparator.BitwiseOp getBitwiseOp();
|
org.apache.hadoop.hbase.protobuf.generated.ComparatorProtos.BitComparator.BitwiseOp getBitwiseOp();
|
||||||
}
|
}
|
||||||
|
@ -1923,8 +1923,8 @@ public final class ComparatorProtos {
|
||||||
return comparable_;
|
return comparable_;
|
||||||
}
|
}
|
||||||
|
|
||||||
// required .BitComparator.BitwiseOp bitwiseOp = 2;
|
// required .BitComparator.BitwiseOp bitwise_op = 2;
|
||||||
public static final int BITWISEOP_FIELD_NUMBER = 2;
|
public static final int BITWISE_OP_FIELD_NUMBER = 2;
|
||||||
private org.apache.hadoop.hbase.protobuf.generated.ComparatorProtos.BitComparator.BitwiseOp bitwiseOp_;
|
private org.apache.hadoop.hbase.protobuf.generated.ComparatorProtos.BitComparator.BitwiseOp bitwiseOp_;
|
||||||
public boolean hasBitwiseOp() {
|
public boolean hasBitwiseOp() {
|
||||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||||
|
@ -2027,7 +2027,7 @@ public final class ComparatorProtos {
|
||||||
hash = (53 * hash) + getComparable().hashCode();
|
hash = (53 * hash) + getComparable().hashCode();
|
||||||
}
|
}
|
||||||
if (hasBitwiseOp()) {
|
if (hasBitwiseOp()) {
|
||||||
hash = (37 * hash) + BITWISEOP_FIELD_NUMBER;
|
hash = (37 * hash) + BITWISE_OP_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + hashEnum(getBitwiseOp());
|
hash = (53 * hash) + hashEnum(getBitwiseOp());
|
||||||
}
|
}
|
||||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||||
|
@ -2382,7 +2382,7 @@ public final class ComparatorProtos {
|
||||||
return comparableBuilder_;
|
return comparableBuilder_;
|
||||||
}
|
}
|
||||||
|
|
||||||
// required .BitComparator.BitwiseOp bitwiseOp = 2;
|
// required .BitComparator.BitwiseOp bitwise_op = 2;
|
||||||
private org.apache.hadoop.hbase.protobuf.generated.ComparatorProtos.BitComparator.BitwiseOp bitwiseOp_ = org.apache.hadoop.hbase.protobuf.generated.ComparatorProtos.BitComparator.BitwiseOp.AND;
|
private org.apache.hadoop.hbase.protobuf.generated.ComparatorProtos.BitComparator.BitwiseOp bitwiseOp_ = org.apache.hadoop.hbase.protobuf.generated.ComparatorProtos.BitComparator.BitwiseOp.AND;
|
||||||
public boolean hasBitwiseOp() {
|
public boolean hasBitwiseOp() {
|
||||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||||
|
@ -2724,7 +2724,7 @@ public final class ComparatorProtos {
|
||||||
boolean hasPattern();
|
boolean hasPattern();
|
||||||
String getPattern();
|
String getPattern();
|
||||||
|
|
||||||
// required int32 patternFlags = 2;
|
// required int32 pattern_flags = 2;
|
||||||
boolean hasPatternFlags();
|
boolean hasPatternFlags();
|
||||||
int getPatternFlags();
|
int getPatternFlags();
|
||||||
|
|
||||||
|
@ -2793,8 +2793,8 @@ public final class ComparatorProtos {
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// required int32 patternFlags = 2;
|
// required int32 pattern_flags = 2;
|
||||||
public static final int PATTERNFLAGS_FIELD_NUMBER = 2;
|
public static final int PATTERN_FLAGS_FIELD_NUMBER = 2;
|
||||||
private int patternFlags_;
|
private int patternFlags_;
|
||||||
public boolean hasPatternFlags() {
|
public boolean hasPatternFlags() {
|
||||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||||
|
@ -2946,7 +2946,7 @@ public final class ComparatorProtos {
|
||||||
hash = (53 * hash) + getPattern().hashCode();
|
hash = (53 * hash) + getPattern().hashCode();
|
||||||
}
|
}
|
||||||
if (hasPatternFlags()) {
|
if (hasPatternFlags()) {
|
||||||
hash = (37 * hash) + PATTERNFLAGS_FIELD_NUMBER;
|
hash = (37 * hash) + PATTERN_FLAGS_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + getPatternFlags();
|
hash = (53 * hash) + getPatternFlags();
|
||||||
}
|
}
|
||||||
if (hasCharset()) {
|
if (hasCharset()) {
|
||||||
|
@ -3250,7 +3250,7 @@ public final class ComparatorProtos {
|
||||||
onChanged();
|
onChanged();
|
||||||
}
|
}
|
||||||
|
|
||||||
// required int32 patternFlags = 2;
|
// required int32 pattern_flags = 2;
|
||||||
private int patternFlags_ ;
|
private int patternFlags_ ;
|
||||||
public boolean hasPatternFlags() {
|
public boolean hasPatternFlags() {
|
||||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||||
|
@ -3783,21 +3783,22 @@ public final class ComparatorProtos {
|
||||||
descriptor;
|
descriptor;
|
||||||
static {
|
static {
|
||||||
java.lang.String[] descriptorData = {
|
java.lang.String[] descriptorData = {
|
||||||
"\n\020Comparator.proto\"8\n\nComparator\022\014\n\004name" +
|
"\n\020Comparator.proto\"9\n\nComparator\022\014\n\004name" +
|
||||||
"\030\001 \002(\t\022\034\n\024serializedComparator\030\002 \001(\014\"$\n\023" +
|
"\030\001 \002(\t\022\035\n\025serialized_comparator\030\002 \001(\014\"$\n" +
|
||||||
"ByteArrayComparable\022\r\n\005value\030\001 \001(\014\"<\n\020Bi" +
|
"\023ByteArrayComparable\022\r\n\005value\030\001 \001(\014\"<\n\020B" +
|
||||||
"naryComparator\022(\n\ncomparable\030\001 \002(\0132\024.Byt" +
|
"inaryComparator\022(\n\ncomparable\030\001 \002(\0132\024.By" +
|
||||||
"eArrayComparable\"B\n\026BinaryPrefixComparat" +
|
"teArrayComparable\"B\n\026BinaryPrefixCompara" +
|
||||||
"or\022(\n\ncomparable\030\001 \002(\0132\024.ByteArrayCompar" +
|
"tor\022(\n\ncomparable\030\001 \002(\0132\024.ByteArrayCompa" +
|
||||||
"able\"\215\001\n\rBitComparator\022(\n\ncomparable\030\001 \002" +
|
"rable\"\216\001\n\rBitComparator\022(\n\ncomparable\030\001 " +
|
||||||
"(\0132\024.ByteArrayComparable\022+\n\tbitwiseOp\030\002 " +
|
"\002(\0132\024.ByteArrayComparable\022,\n\nbitwise_op\030" +
|
||||||
"\002(\0162\030.BitComparator.BitwiseOp\"%\n\tBitwise" +
|
"\002 \002(\0162\030.BitComparator.BitwiseOp\"%\n\tBitwi" +
|
||||||
"Op\022\007\n\003AND\020\001\022\006\n\002OR\020\002\022\007\n\003XOR\020\003\"\020\n\016NullComp",
|
"seOp\022\007\n\003AND\020\001\022\006\n\002OR\020\002\022\007\n\003XOR\020\003\"\020\n\016NullCo",
|
||||||
"arator\"O\n\025RegexStringComparator\022\017\n\007patte" +
|
"mparator\"P\n\025RegexStringComparator\022\017\n\007pat" +
|
||||||
"rn\030\001 \002(\t\022\024\n\014patternFlags\030\002 \002(\005\022\017\n\007charse" +
|
"tern\030\001 \002(\t\022\025\n\rpattern_flags\030\002 \002(\005\022\017\n\007cha" +
|
||||||
"t\030\003 \002(\t\"%\n\023SubstringComparator\022\016\n\006substr" +
|
"rset\030\003 \002(\t\"%\n\023SubstringComparator\022\016\n\006sub" +
|
||||||
"\030\001 \002(\tBF\n*org.apache.hadoop.hbase.protob" +
|
"str\030\001 \002(\tBF\n*org.apache.hadoop.hbase.pro" +
|
||||||
"uf.generatedB\020ComparatorProtosH\001\210\001\001\240\001\001"
|
"tobuf.generatedB\020ComparatorProtosH\001\210\001\001\240\001" +
|
||||||
|
"\001"
|
||||||
};
|
};
|
||||||
com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
|
com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
|
||||||
new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() {
|
new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() {
|
||||||
|
|
|
@ -11,19 +11,19 @@ public final class ErrorHandlingProtos {
|
||||||
public interface StackTraceElementMessageOrBuilder
|
public interface StackTraceElementMessageOrBuilder
|
||||||
extends com.google.protobuf.MessageOrBuilder {
|
extends com.google.protobuf.MessageOrBuilder {
|
||||||
|
|
||||||
// optional string declaringClass = 1;
|
// optional string declaring_class = 1;
|
||||||
boolean hasDeclaringClass();
|
boolean hasDeclaringClass();
|
||||||
String getDeclaringClass();
|
String getDeclaringClass();
|
||||||
|
|
||||||
// optional string methodName = 2;
|
// optional string method_name = 2;
|
||||||
boolean hasMethodName();
|
boolean hasMethodName();
|
||||||
String getMethodName();
|
String getMethodName();
|
||||||
|
|
||||||
// optional string fileName = 3;
|
// optional string file_name = 3;
|
||||||
boolean hasFileName();
|
boolean hasFileName();
|
||||||
String getFileName();
|
String getFileName();
|
||||||
|
|
||||||
// optional int32 lineNumber = 4;
|
// optional int32 line_number = 4;
|
||||||
boolean hasLineNumber();
|
boolean hasLineNumber();
|
||||||
int getLineNumber();
|
int getLineNumber();
|
||||||
}
|
}
|
||||||
|
@ -56,8 +56,8 @@ public final class ErrorHandlingProtos {
|
||||||
}
|
}
|
||||||
|
|
||||||
private int bitField0_;
|
private int bitField0_;
|
||||||
// optional string declaringClass = 1;
|
// optional string declaring_class = 1;
|
||||||
public static final int DECLARINGCLASS_FIELD_NUMBER = 1;
|
public static final int DECLARING_CLASS_FIELD_NUMBER = 1;
|
||||||
private java.lang.Object declaringClass_;
|
private java.lang.Object declaringClass_;
|
||||||
public boolean hasDeclaringClass() {
|
public boolean hasDeclaringClass() {
|
||||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||||
|
@ -88,8 +88,8 @@ public final class ErrorHandlingProtos {
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// optional string methodName = 2;
|
// optional string method_name = 2;
|
||||||
public static final int METHODNAME_FIELD_NUMBER = 2;
|
public static final int METHOD_NAME_FIELD_NUMBER = 2;
|
||||||
private java.lang.Object methodName_;
|
private java.lang.Object methodName_;
|
||||||
public boolean hasMethodName() {
|
public boolean hasMethodName() {
|
||||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||||
|
@ -120,8 +120,8 @@ public final class ErrorHandlingProtos {
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// optional string fileName = 3;
|
// optional string file_name = 3;
|
||||||
public static final int FILENAME_FIELD_NUMBER = 3;
|
public static final int FILE_NAME_FIELD_NUMBER = 3;
|
||||||
private java.lang.Object fileName_;
|
private java.lang.Object fileName_;
|
||||||
public boolean hasFileName() {
|
public boolean hasFileName() {
|
||||||
return ((bitField0_ & 0x00000004) == 0x00000004);
|
return ((bitField0_ & 0x00000004) == 0x00000004);
|
||||||
|
@ -152,8 +152,8 @@ public final class ErrorHandlingProtos {
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// optional int32 lineNumber = 4;
|
// optional int32 line_number = 4;
|
||||||
public static final int LINENUMBER_FIELD_NUMBER = 4;
|
public static final int LINE_NUMBER_FIELD_NUMBER = 4;
|
||||||
private int lineNumber_;
|
private int lineNumber_;
|
||||||
public boolean hasLineNumber() {
|
public boolean hasLineNumber() {
|
||||||
return ((bitField0_ & 0x00000008) == 0x00000008);
|
return ((bitField0_ & 0x00000008) == 0x00000008);
|
||||||
|
@ -270,19 +270,19 @@ public final class ErrorHandlingProtos {
|
||||||
int hash = 41;
|
int hash = 41;
|
||||||
hash = (19 * hash) + getDescriptorForType().hashCode();
|
hash = (19 * hash) + getDescriptorForType().hashCode();
|
||||||
if (hasDeclaringClass()) {
|
if (hasDeclaringClass()) {
|
||||||
hash = (37 * hash) + DECLARINGCLASS_FIELD_NUMBER;
|
hash = (37 * hash) + DECLARING_CLASS_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + getDeclaringClass().hashCode();
|
hash = (53 * hash) + getDeclaringClass().hashCode();
|
||||||
}
|
}
|
||||||
if (hasMethodName()) {
|
if (hasMethodName()) {
|
||||||
hash = (37 * hash) + METHODNAME_FIELD_NUMBER;
|
hash = (37 * hash) + METHOD_NAME_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + getMethodName().hashCode();
|
hash = (53 * hash) + getMethodName().hashCode();
|
||||||
}
|
}
|
||||||
if (hasFileName()) {
|
if (hasFileName()) {
|
||||||
hash = (37 * hash) + FILENAME_FIELD_NUMBER;
|
hash = (37 * hash) + FILE_NAME_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + getFileName().hashCode();
|
hash = (53 * hash) + getFileName().hashCode();
|
||||||
}
|
}
|
||||||
if (hasLineNumber()) {
|
if (hasLineNumber()) {
|
||||||
hash = (37 * hash) + LINENUMBER_FIELD_NUMBER;
|
hash = (37 * hash) + LINE_NUMBER_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + getLineNumber();
|
hash = (53 * hash) + getLineNumber();
|
||||||
}
|
}
|
||||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||||
|
@ -548,7 +548,7 @@ public final class ErrorHandlingProtos {
|
||||||
|
|
||||||
private int bitField0_;
|
private int bitField0_;
|
||||||
|
|
||||||
// optional string declaringClass = 1;
|
// optional string declaring_class = 1;
|
||||||
private java.lang.Object declaringClass_ = "";
|
private java.lang.Object declaringClass_ = "";
|
||||||
public boolean hasDeclaringClass() {
|
public boolean hasDeclaringClass() {
|
||||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||||
|
@ -584,7 +584,7 @@ public final class ErrorHandlingProtos {
|
||||||
onChanged();
|
onChanged();
|
||||||
}
|
}
|
||||||
|
|
||||||
// optional string methodName = 2;
|
// optional string method_name = 2;
|
||||||
private java.lang.Object methodName_ = "";
|
private java.lang.Object methodName_ = "";
|
||||||
public boolean hasMethodName() {
|
public boolean hasMethodName() {
|
||||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||||
|
@ -620,7 +620,7 @@ public final class ErrorHandlingProtos {
|
||||||
onChanged();
|
onChanged();
|
||||||
}
|
}
|
||||||
|
|
||||||
// optional string fileName = 3;
|
// optional string file_name = 3;
|
||||||
private java.lang.Object fileName_ = "";
|
private java.lang.Object fileName_ = "";
|
||||||
public boolean hasFileName() {
|
public boolean hasFileName() {
|
||||||
return ((bitField0_ & 0x00000004) == 0x00000004);
|
return ((bitField0_ & 0x00000004) == 0x00000004);
|
||||||
|
@ -656,7 +656,7 @@ public final class ErrorHandlingProtos {
|
||||||
onChanged();
|
onChanged();
|
||||||
}
|
}
|
||||||
|
|
||||||
// optional int32 lineNumber = 4;
|
// optional int32 line_number = 4;
|
||||||
private int lineNumber_ ;
|
private int lineNumber_ ;
|
||||||
public boolean hasLineNumber() {
|
public boolean hasLineNumber() {
|
||||||
return ((bitField0_ & 0x00000008) == 0x00000008);
|
return ((bitField0_ & 0x00000008) == 0x00000008);
|
||||||
|
@ -691,7 +691,7 @@ public final class ErrorHandlingProtos {
|
||||||
public interface GenericExceptionMessageOrBuilder
|
public interface GenericExceptionMessageOrBuilder
|
||||||
extends com.google.protobuf.MessageOrBuilder {
|
extends com.google.protobuf.MessageOrBuilder {
|
||||||
|
|
||||||
// optional string className = 1;
|
// optional string class_name = 1;
|
||||||
boolean hasClassName();
|
boolean hasClassName();
|
||||||
String getClassName();
|
String getClassName();
|
||||||
|
|
||||||
|
@ -699,7 +699,7 @@ public final class ErrorHandlingProtos {
|
||||||
boolean hasMessage();
|
boolean hasMessage();
|
||||||
String getMessage();
|
String getMessage();
|
||||||
|
|
||||||
// optional bytes errorInfo = 3;
|
// optional bytes error_info = 3;
|
||||||
boolean hasErrorInfo();
|
boolean hasErrorInfo();
|
||||||
com.google.protobuf.ByteString getErrorInfo();
|
com.google.protobuf.ByteString getErrorInfo();
|
||||||
|
|
||||||
|
@ -742,8 +742,8 @@ public final class ErrorHandlingProtos {
|
||||||
}
|
}
|
||||||
|
|
||||||
private int bitField0_;
|
private int bitField0_;
|
||||||
// optional string className = 1;
|
// optional string class_name = 1;
|
||||||
public static final int CLASSNAME_FIELD_NUMBER = 1;
|
public static final int CLASS_NAME_FIELD_NUMBER = 1;
|
||||||
private java.lang.Object className_;
|
private java.lang.Object className_;
|
||||||
public boolean hasClassName() {
|
public boolean hasClassName() {
|
||||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||||
|
@ -806,8 +806,8 @@ public final class ErrorHandlingProtos {
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// optional bytes errorInfo = 3;
|
// optional bytes error_info = 3;
|
||||||
public static final int ERRORINFO_FIELD_NUMBER = 3;
|
public static final int ERROR_INFO_FIELD_NUMBER = 3;
|
||||||
private com.google.protobuf.ByteString errorInfo_;
|
private com.google.protobuf.ByteString errorInfo_;
|
||||||
public boolean hasErrorInfo() {
|
public boolean hasErrorInfo() {
|
||||||
return ((bitField0_ & 0x00000004) == 0x00000004);
|
return ((bitField0_ & 0x00000004) == 0x00000004);
|
||||||
|
@ -942,7 +942,7 @@ public final class ErrorHandlingProtos {
|
||||||
int hash = 41;
|
int hash = 41;
|
||||||
hash = (19 * hash) + getDescriptorForType().hashCode();
|
hash = (19 * hash) + getDescriptorForType().hashCode();
|
||||||
if (hasClassName()) {
|
if (hasClassName()) {
|
||||||
hash = (37 * hash) + CLASSNAME_FIELD_NUMBER;
|
hash = (37 * hash) + CLASS_NAME_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + getClassName().hashCode();
|
hash = (53 * hash) + getClassName().hashCode();
|
||||||
}
|
}
|
||||||
if (hasMessage()) {
|
if (hasMessage()) {
|
||||||
|
@ -950,7 +950,7 @@ public final class ErrorHandlingProtos {
|
||||||
hash = (53 * hash) + getMessage().hashCode();
|
hash = (53 * hash) + getMessage().hashCode();
|
||||||
}
|
}
|
||||||
if (hasErrorInfo()) {
|
if (hasErrorInfo()) {
|
||||||
hash = (37 * hash) + ERRORINFO_FIELD_NUMBER;
|
hash = (37 * hash) + ERROR_INFO_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + getErrorInfo().hashCode();
|
hash = (53 * hash) + getErrorInfo().hashCode();
|
||||||
}
|
}
|
||||||
if (getTraceCount() > 0) {
|
if (getTraceCount() > 0) {
|
||||||
|
@ -1254,7 +1254,7 @@ public final class ErrorHandlingProtos {
|
||||||
|
|
||||||
private int bitField0_;
|
private int bitField0_;
|
||||||
|
|
||||||
// optional string className = 1;
|
// optional string class_name = 1;
|
||||||
private java.lang.Object className_ = "";
|
private java.lang.Object className_ = "";
|
||||||
public boolean hasClassName() {
|
public boolean hasClassName() {
|
||||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||||
|
@ -1326,7 +1326,7 @@ public final class ErrorHandlingProtos {
|
||||||
onChanged();
|
onChanged();
|
||||||
}
|
}
|
||||||
|
|
||||||
// optional bytes errorInfo = 3;
|
// optional bytes error_info = 3;
|
||||||
private com.google.protobuf.ByteString errorInfo_ = com.google.protobuf.ByteString.EMPTY;
|
private com.google.protobuf.ByteString errorInfo_ = com.google.protobuf.ByteString.EMPTY;
|
||||||
public boolean hasErrorInfo() {
|
public boolean hasErrorInfo() {
|
||||||
return ((bitField0_ & 0x00000004) == 0x00000004);
|
return ((bitField0_ & 0x00000004) == 0x00000004);
|
||||||
|
@ -1554,7 +1554,7 @@ public final class ErrorHandlingProtos {
|
||||||
boolean hasSource();
|
boolean hasSource();
|
||||||
String getSource();
|
String getSource();
|
||||||
|
|
||||||
// optional .GenericExceptionMessage genericException = 2;
|
// optional .GenericExceptionMessage generic_exception = 2;
|
||||||
boolean hasGenericException();
|
boolean hasGenericException();
|
||||||
org.apache.hadoop.hbase.protobuf.generated.ErrorHandlingProtos.GenericExceptionMessage getGenericException();
|
org.apache.hadoop.hbase.protobuf.generated.ErrorHandlingProtos.GenericExceptionMessage getGenericException();
|
||||||
org.apache.hadoop.hbase.protobuf.generated.ErrorHandlingProtos.GenericExceptionMessageOrBuilder getGenericExceptionOrBuilder();
|
org.apache.hadoop.hbase.protobuf.generated.ErrorHandlingProtos.GenericExceptionMessageOrBuilder getGenericExceptionOrBuilder();
|
||||||
|
@ -1620,8 +1620,8 @@ public final class ErrorHandlingProtos {
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// optional .GenericExceptionMessage genericException = 2;
|
// optional .GenericExceptionMessage generic_exception = 2;
|
||||||
public static final int GENERICEXCEPTION_FIELD_NUMBER = 2;
|
public static final int GENERIC_EXCEPTION_FIELD_NUMBER = 2;
|
||||||
private org.apache.hadoop.hbase.protobuf.generated.ErrorHandlingProtos.GenericExceptionMessage genericException_;
|
private org.apache.hadoop.hbase.protobuf.generated.ErrorHandlingProtos.GenericExceptionMessage genericException_;
|
||||||
public boolean hasGenericException() {
|
public boolean hasGenericException() {
|
||||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||||
|
@ -1719,7 +1719,7 @@ public final class ErrorHandlingProtos {
|
||||||
hash = (53 * hash) + getSource().hashCode();
|
hash = (53 * hash) + getSource().hashCode();
|
||||||
}
|
}
|
||||||
if (hasGenericException()) {
|
if (hasGenericException()) {
|
||||||
hash = (37 * hash) + GENERICEXCEPTION_FIELD_NUMBER;
|
hash = (37 * hash) + GENERIC_EXCEPTION_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + getGenericException().hashCode();
|
hash = (53 * hash) + getGenericException().hashCode();
|
||||||
}
|
}
|
||||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||||
|
@ -2006,7 +2006,7 @@ public final class ErrorHandlingProtos {
|
||||||
onChanged();
|
onChanged();
|
||||||
}
|
}
|
||||||
|
|
||||||
// optional .GenericExceptionMessage genericException = 2;
|
// optional .GenericExceptionMessage generic_exception = 2;
|
||||||
private org.apache.hadoop.hbase.protobuf.generated.ErrorHandlingProtos.GenericExceptionMessage genericException_ = org.apache.hadoop.hbase.protobuf.generated.ErrorHandlingProtos.GenericExceptionMessage.getDefaultInstance();
|
private org.apache.hadoop.hbase.protobuf.generated.ErrorHandlingProtos.GenericExceptionMessage genericException_ = org.apache.hadoop.hbase.protobuf.generated.ErrorHandlingProtos.GenericExceptionMessage.getDefaultInstance();
|
||||||
private com.google.protobuf.SingleFieldBuilder<
|
private com.google.protobuf.SingleFieldBuilder<
|
||||||
org.apache.hadoop.hbase.protobuf.generated.ErrorHandlingProtos.GenericExceptionMessage, org.apache.hadoop.hbase.protobuf.generated.ErrorHandlingProtos.GenericExceptionMessage.Builder, org.apache.hadoop.hbase.protobuf.generated.ErrorHandlingProtos.GenericExceptionMessageOrBuilder> genericExceptionBuilder_;
|
org.apache.hadoop.hbase.protobuf.generated.ErrorHandlingProtos.GenericExceptionMessage, org.apache.hadoop.hbase.protobuf.generated.ErrorHandlingProtos.GenericExceptionMessage.Builder, org.apache.hadoop.hbase.protobuf.generated.ErrorHandlingProtos.GenericExceptionMessageOrBuilder> genericExceptionBuilder_;
|
||||||
|
@ -2131,17 +2131,17 @@ public final class ErrorHandlingProtos {
|
||||||
descriptor;
|
descriptor;
|
||||||
static {
|
static {
|
||||||
java.lang.String[] descriptorData = {
|
java.lang.String[] descriptorData = {
|
||||||
"\n\023ErrorHandling.proto\"l\n\030StackTraceEleme" +
|
"\n\023ErrorHandling.proto\"p\n\030StackTraceEleme" +
|
||||||
"ntMessage\022\026\n\016declaringClass\030\001 \001(\t\022\022\n\nmet" +
|
"ntMessage\022\027\n\017declaring_class\030\001 \001(\t\022\023\n\013me" +
|
||||||
"hodName\030\002 \001(\t\022\020\n\010fileName\030\003 \001(\t\022\022\n\nlineN" +
|
"thod_name\030\002 \001(\t\022\021\n\tfile_name\030\003 \001(\t\022\023\n\013li" +
|
||||||
"umber\030\004 \001(\005\"z\n\027GenericExceptionMessage\022\021" +
|
"ne_number\030\004 \001(\005\"|\n\027GenericExceptionMessa" +
|
||||||
"\n\tclassName\030\001 \001(\t\022\017\n\007message\030\002 \001(\t\022\021\n\ter" +
|
"ge\022\022\n\nclass_name\030\001 \001(\t\022\017\n\007message\030\002 \001(\t\022" +
|
||||||
"rorInfo\030\003 \001(\014\022(\n\005trace\030\004 \003(\0132\031.StackTrac" +
|
"\022\n\nerror_info\030\003 \001(\014\022(\n\005trace\030\004 \003(\0132\031.Sta" +
|
||||||
"eElementMessage\"]\n\027ForeignExceptionMessa" +
|
"ckTraceElementMessage\"^\n\027ForeignExceptio" +
|
||||||
"ge\022\016\n\006source\030\001 \001(\t\0222\n\020genericException\030\002" +
|
"nMessage\022\016\n\006source\030\001 \001(\t\0223\n\021generic_exce" +
|
||||||
" \001(\0132\030.GenericExceptionMessageBF\n*org.ap" +
|
"ption\030\002 \001(\0132\030.GenericExceptionMessageBF\n" +
|
||||||
"ache.hadoop.hbase.protobuf.generatedB\023Er",
|
"*org.apache.hadoop.hbase.protobuf.genera",
|
||||||
"rorHandlingProtosH\001\240\001\001"
|
"tedB\023ErrorHandlingProtosH\001\240\001\001"
|
||||||
};
|
};
|
||||||
com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
|
com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
|
||||||
new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() {
|
new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() {
|
||||||
|
|
|
@ -15,7 +15,7 @@ public final class FilterProtos {
|
||||||
boolean hasName();
|
boolean hasName();
|
||||||
String getName();
|
String getName();
|
||||||
|
|
||||||
// optional bytes serializedFilter = 2;
|
// optional bytes serialized_filter = 2;
|
||||||
boolean hasSerializedFilter();
|
boolean hasSerializedFilter();
|
||||||
com.google.protobuf.ByteString getSerializedFilter();
|
com.google.protobuf.ByteString getSerializedFilter();
|
||||||
}
|
}
|
||||||
|
@ -80,8 +80,8 @@ public final class FilterProtos {
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// optional bytes serializedFilter = 2;
|
// optional bytes serialized_filter = 2;
|
||||||
public static final int SERIALIZEDFILTER_FIELD_NUMBER = 2;
|
public static final int SERIALIZED_FILTER_FIELD_NUMBER = 2;
|
||||||
private com.google.protobuf.ByteString serializedFilter_;
|
private com.google.protobuf.ByteString serializedFilter_;
|
||||||
public boolean hasSerializedFilter() {
|
public boolean hasSerializedFilter() {
|
||||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||||
|
@ -180,7 +180,7 @@ public final class FilterProtos {
|
||||||
hash = (53 * hash) + getName().hashCode();
|
hash = (53 * hash) + getName().hashCode();
|
||||||
}
|
}
|
||||||
if (hasSerializedFilter()) {
|
if (hasSerializedFilter()) {
|
||||||
hash = (37 * hash) + SERIALIZEDFILTER_FIELD_NUMBER;
|
hash = (37 * hash) + SERIALIZED_FILTER_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + getSerializedFilter().hashCode();
|
hash = (53 * hash) + getSerializedFilter().hashCode();
|
||||||
}
|
}
|
||||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||||
|
@ -458,7 +458,7 @@ public final class FilterProtos {
|
||||||
onChanged();
|
onChanged();
|
||||||
}
|
}
|
||||||
|
|
||||||
// optional bytes serializedFilter = 2;
|
// optional bytes serialized_filter = 2;
|
||||||
private com.google.protobuf.ByteString serializedFilter_ = com.google.protobuf.ByteString.EMPTY;
|
private com.google.protobuf.ByteString serializedFilter_ = com.google.protobuf.ByteString.EMPTY;
|
||||||
public boolean hasSerializedFilter() {
|
public boolean hasSerializedFilter() {
|
||||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||||
|
@ -883,7 +883,7 @@ public final class FilterProtos {
|
||||||
boolean hasOffset();
|
boolean hasOffset();
|
||||||
int getOffset();
|
int getOffset();
|
||||||
|
|
||||||
// optional bytes columnOffset = 3;
|
// optional bytes column_offset = 3;
|
||||||
boolean hasColumnOffset();
|
boolean hasColumnOffset();
|
||||||
com.google.protobuf.ByteString getColumnOffset();
|
com.google.protobuf.ByteString getColumnOffset();
|
||||||
}
|
}
|
||||||
|
@ -936,8 +936,8 @@ public final class FilterProtos {
|
||||||
return offset_;
|
return offset_;
|
||||||
}
|
}
|
||||||
|
|
||||||
// optional bytes columnOffset = 3;
|
// optional bytes column_offset = 3;
|
||||||
public static final int COLUMNOFFSET_FIELD_NUMBER = 3;
|
public static final int COLUMN_OFFSET_FIELD_NUMBER = 3;
|
||||||
private com.google.protobuf.ByteString columnOffset_;
|
private com.google.protobuf.ByteString columnOffset_;
|
||||||
public boolean hasColumnOffset() {
|
public boolean hasColumnOffset() {
|
||||||
return ((bitField0_ & 0x00000004) == 0x00000004);
|
return ((bitField0_ & 0x00000004) == 0x00000004);
|
||||||
|
@ -1053,7 +1053,7 @@ public final class FilterProtos {
|
||||||
hash = (53 * hash) + getOffset();
|
hash = (53 * hash) + getOffset();
|
||||||
}
|
}
|
||||||
if (hasColumnOffset()) {
|
if (hasColumnOffset()) {
|
||||||
hash = (37 * hash) + COLUMNOFFSET_FIELD_NUMBER;
|
hash = (37 * hash) + COLUMN_OFFSET_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + getColumnOffset().hashCode();
|
hash = (53 * hash) + getColumnOffset().hashCode();
|
||||||
}
|
}
|
||||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||||
|
@ -1351,7 +1351,7 @@ public final class FilterProtos {
|
||||||
return this;
|
return this;
|
||||||
}
|
}
|
||||||
|
|
||||||
// optional bytes columnOffset = 3;
|
// optional bytes column_offset = 3;
|
||||||
private com.google.protobuf.ByteString columnOffset_ = com.google.protobuf.ByteString.EMPTY;
|
private com.google.protobuf.ByteString columnOffset_ = com.google.protobuf.ByteString.EMPTY;
|
||||||
public boolean hasColumnOffset() {
|
public boolean hasColumnOffset() {
|
||||||
return ((bitField0_ & 0x00000004) == 0x00000004);
|
return ((bitField0_ & 0x00000004) == 0x00000004);
|
||||||
|
@ -1771,19 +1771,19 @@ public final class FilterProtos {
|
||||||
public interface ColumnRangeFilterOrBuilder
|
public interface ColumnRangeFilterOrBuilder
|
||||||
extends com.google.protobuf.MessageOrBuilder {
|
extends com.google.protobuf.MessageOrBuilder {
|
||||||
|
|
||||||
// optional bytes minColumn = 1;
|
// optional bytes min_column = 1;
|
||||||
boolean hasMinColumn();
|
boolean hasMinColumn();
|
||||||
com.google.protobuf.ByteString getMinColumn();
|
com.google.protobuf.ByteString getMinColumn();
|
||||||
|
|
||||||
// optional bool minColumnInclusive = 2;
|
// optional bool min_column_inclusive = 2;
|
||||||
boolean hasMinColumnInclusive();
|
boolean hasMinColumnInclusive();
|
||||||
boolean getMinColumnInclusive();
|
boolean getMinColumnInclusive();
|
||||||
|
|
||||||
// optional bytes maxColumn = 3;
|
// optional bytes max_column = 3;
|
||||||
boolean hasMaxColumn();
|
boolean hasMaxColumn();
|
||||||
com.google.protobuf.ByteString getMaxColumn();
|
com.google.protobuf.ByteString getMaxColumn();
|
||||||
|
|
||||||
// optional bool maxColumnInclusive = 4;
|
// optional bool max_column_inclusive = 4;
|
||||||
boolean hasMaxColumnInclusive();
|
boolean hasMaxColumnInclusive();
|
||||||
boolean getMaxColumnInclusive();
|
boolean getMaxColumnInclusive();
|
||||||
}
|
}
|
||||||
|
@ -1816,8 +1816,8 @@ public final class FilterProtos {
|
||||||
}
|
}
|
||||||
|
|
||||||
private int bitField0_;
|
private int bitField0_;
|
||||||
// optional bytes minColumn = 1;
|
// optional bytes min_column = 1;
|
||||||
public static final int MINCOLUMN_FIELD_NUMBER = 1;
|
public static final int MIN_COLUMN_FIELD_NUMBER = 1;
|
||||||
private com.google.protobuf.ByteString minColumn_;
|
private com.google.protobuf.ByteString minColumn_;
|
||||||
public boolean hasMinColumn() {
|
public boolean hasMinColumn() {
|
||||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||||
|
@ -1826,8 +1826,8 @@ public final class FilterProtos {
|
||||||
return minColumn_;
|
return minColumn_;
|
||||||
}
|
}
|
||||||
|
|
||||||
// optional bool minColumnInclusive = 2;
|
// optional bool min_column_inclusive = 2;
|
||||||
public static final int MINCOLUMNINCLUSIVE_FIELD_NUMBER = 2;
|
public static final int MIN_COLUMN_INCLUSIVE_FIELD_NUMBER = 2;
|
||||||
private boolean minColumnInclusive_;
|
private boolean minColumnInclusive_;
|
||||||
public boolean hasMinColumnInclusive() {
|
public boolean hasMinColumnInclusive() {
|
||||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||||
|
@ -1836,8 +1836,8 @@ public final class FilterProtos {
|
||||||
return minColumnInclusive_;
|
return minColumnInclusive_;
|
||||||
}
|
}
|
||||||
|
|
||||||
// optional bytes maxColumn = 3;
|
// optional bytes max_column = 3;
|
||||||
public static final int MAXCOLUMN_FIELD_NUMBER = 3;
|
public static final int MAX_COLUMN_FIELD_NUMBER = 3;
|
||||||
private com.google.protobuf.ByteString maxColumn_;
|
private com.google.protobuf.ByteString maxColumn_;
|
||||||
public boolean hasMaxColumn() {
|
public boolean hasMaxColumn() {
|
||||||
return ((bitField0_ & 0x00000004) == 0x00000004);
|
return ((bitField0_ & 0x00000004) == 0x00000004);
|
||||||
|
@ -1846,8 +1846,8 @@ public final class FilterProtos {
|
||||||
return maxColumn_;
|
return maxColumn_;
|
||||||
}
|
}
|
||||||
|
|
||||||
// optional bool maxColumnInclusive = 4;
|
// optional bool max_column_inclusive = 4;
|
||||||
public static final int MAXCOLUMNINCLUSIVE_FIELD_NUMBER = 4;
|
public static final int MAX_COLUMN_INCLUSIVE_FIELD_NUMBER = 4;
|
||||||
private boolean maxColumnInclusive_;
|
private boolean maxColumnInclusive_;
|
||||||
public boolean hasMaxColumnInclusive() {
|
public boolean hasMaxColumnInclusive() {
|
||||||
return ((bitField0_ & 0x00000008) == 0x00000008);
|
return ((bitField0_ & 0x00000008) == 0x00000008);
|
||||||
|
@ -1964,19 +1964,19 @@ public final class FilterProtos {
|
||||||
int hash = 41;
|
int hash = 41;
|
||||||
hash = (19 * hash) + getDescriptorForType().hashCode();
|
hash = (19 * hash) + getDescriptorForType().hashCode();
|
||||||
if (hasMinColumn()) {
|
if (hasMinColumn()) {
|
||||||
hash = (37 * hash) + MINCOLUMN_FIELD_NUMBER;
|
hash = (37 * hash) + MIN_COLUMN_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + getMinColumn().hashCode();
|
hash = (53 * hash) + getMinColumn().hashCode();
|
||||||
}
|
}
|
||||||
if (hasMinColumnInclusive()) {
|
if (hasMinColumnInclusive()) {
|
||||||
hash = (37 * hash) + MINCOLUMNINCLUSIVE_FIELD_NUMBER;
|
hash = (37 * hash) + MIN_COLUMN_INCLUSIVE_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + hashBoolean(getMinColumnInclusive());
|
hash = (53 * hash) + hashBoolean(getMinColumnInclusive());
|
||||||
}
|
}
|
||||||
if (hasMaxColumn()) {
|
if (hasMaxColumn()) {
|
||||||
hash = (37 * hash) + MAXCOLUMN_FIELD_NUMBER;
|
hash = (37 * hash) + MAX_COLUMN_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + getMaxColumn().hashCode();
|
hash = (53 * hash) + getMaxColumn().hashCode();
|
||||||
}
|
}
|
||||||
if (hasMaxColumnInclusive()) {
|
if (hasMaxColumnInclusive()) {
|
||||||
hash = (37 * hash) + MAXCOLUMNINCLUSIVE_FIELD_NUMBER;
|
hash = (37 * hash) + MAX_COLUMN_INCLUSIVE_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + hashBoolean(getMaxColumnInclusive());
|
hash = (53 * hash) + hashBoolean(getMaxColumnInclusive());
|
||||||
}
|
}
|
||||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||||
|
@ -2242,7 +2242,7 @@ public final class FilterProtos {
|
||||||
|
|
||||||
private int bitField0_;
|
private int bitField0_;
|
||||||
|
|
||||||
// optional bytes minColumn = 1;
|
// optional bytes min_column = 1;
|
||||||
private com.google.protobuf.ByteString minColumn_ = com.google.protobuf.ByteString.EMPTY;
|
private com.google.protobuf.ByteString minColumn_ = com.google.protobuf.ByteString.EMPTY;
|
||||||
public boolean hasMinColumn() {
|
public boolean hasMinColumn() {
|
||||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||||
|
@ -2266,7 +2266,7 @@ public final class FilterProtos {
|
||||||
return this;
|
return this;
|
||||||
}
|
}
|
||||||
|
|
||||||
// optional bool minColumnInclusive = 2;
|
// optional bool min_column_inclusive = 2;
|
||||||
private boolean minColumnInclusive_ ;
|
private boolean minColumnInclusive_ ;
|
||||||
public boolean hasMinColumnInclusive() {
|
public boolean hasMinColumnInclusive() {
|
||||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||||
|
@ -2287,7 +2287,7 @@ public final class FilterProtos {
|
||||||
return this;
|
return this;
|
||||||
}
|
}
|
||||||
|
|
||||||
// optional bytes maxColumn = 3;
|
// optional bytes max_column = 3;
|
||||||
private com.google.protobuf.ByteString maxColumn_ = com.google.protobuf.ByteString.EMPTY;
|
private com.google.protobuf.ByteString maxColumn_ = com.google.protobuf.ByteString.EMPTY;
|
||||||
public boolean hasMaxColumn() {
|
public boolean hasMaxColumn() {
|
||||||
return ((bitField0_ & 0x00000004) == 0x00000004);
|
return ((bitField0_ & 0x00000004) == 0x00000004);
|
||||||
|
@ -2311,7 +2311,7 @@ public final class FilterProtos {
|
||||||
return this;
|
return this;
|
||||||
}
|
}
|
||||||
|
|
||||||
// optional bool maxColumnInclusive = 4;
|
// optional bool max_column_inclusive = 4;
|
||||||
private boolean maxColumnInclusive_ ;
|
private boolean maxColumnInclusive_ ;
|
||||||
public boolean hasMaxColumnInclusive() {
|
public boolean hasMaxColumnInclusive() {
|
||||||
return ((bitField0_ & 0x00000008) == 0x00000008);
|
return ((bitField0_ & 0x00000008) == 0x00000008);
|
||||||
|
@ -2346,7 +2346,7 @@ public final class FilterProtos {
|
||||||
public interface CompareFilterOrBuilder
|
public interface CompareFilterOrBuilder
|
||||||
extends com.google.protobuf.MessageOrBuilder {
|
extends com.google.protobuf.MessageOrBuilder {
|
||||||
|
|
||||||
// required .CompareType compareOp = 1;
|
// required .CompareType compare_op = 1;
|
||||||
boolean hasCompareOp();
|
boolean hasCompareOp();
|
||||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.CompareType getCompareOp();
|
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.CompareType getCompareOp();
|
||||||
|
|
||||||
|
@ -2384,8 +2384,8 @@ public final class FilterProtos {
|
||||||
}
|
}
|
||||||
|
|
||||||
private int bitField0_;
|
private int bitField0_;
|
||||||
// required .CompareType compareOp = 1;
|
// required .CompareType compare_op = 1;
|
||||||
public static final int COMPAREOP_FIELD_NUMBER = 1;
|
public static final int COMPARE_OP_FIELD_NUMBER = 1;
|
||||||
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.CompareType compareOp_;
|
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.CompareType compareOp_;
|
||||||
public boolean hasCompareOp() {
|
public boolean hasCompareOp() {
|
||||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||||
|
@ -2499,7 +2499,7 @@ public final class FilterProtos {
|
||||||
int hash = 41;
|
int hash = 41;
|
||||||
hash = (19 * hash) + getDescriptorForType().hashCode();
|
hash = (19 * hash) + getDescriptorForType().hashCode();
|
||||||
if (hasCompareOp()) {
|
if (hasCompareOp()) {
|
||||||
hash = (37 * hash) + COMPAREOP_FIELD_NUMBER;
|
hash = (37 * hash) + COMPARE_OP_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + hashEnum(getCompareOp());
|
hash = (53 * hash) + hashEnum(getCompareOp());
|
||||||
}
|
}
|
||||||
if (hasComparator()) {
|
if (hasComparator()) {
|
||||||
|
@ -2770,7 +2770,7 @@ public final class FilterProtos {
|
||||||
|
|
||||||
private int bitField0_;
|
private int bitField0_;
|
||||||
|
|
||||||
// required .CompareType compareOp = 1;
|
// required .CompareType compare_op = 1;
|
||||||
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.CompareType compareOp_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.CompareType.LESS;
|
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.CompareType compareOp_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.CompareType.LESS;
|
||||||
public boolean hasCompareOp() {
|
public boolean hasCompareOp() {
|
||||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||||
|
@ -2898,20 +2898,20 @@ public final class FilterProtos {
|
||||||
public interface DependentColumnFilterOrBuilder
|
public interface DependentColumnFilterOrBuilder
|
||||||
extends com.google.protobuf.MessageOrBuilder {
|
extends com.google.protobuf.MessageOrBuilder {
|
||||||
|
|
||||||
// required .CompareFilter compareFilter = 1;
|
// required .CompareFilter compare_filter = 1;
|
||||||
boolean hasCompareFilter();
|
boolean hasCompareFilter();
|
||||||
org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter getCompareFilter();
|
org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter getCompareFilter();
|
||||||
org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilterOrBuilder getCompareFilterOrBuilder();
|
org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilterOrBuilder getCompareFilterOrBuilder();
|
||||||
|
|
||||||
// optional bytes columnFamily = 2;
|
// optional bytes column_family = 2;
|
||||||
boolean hasColumnFamily();
|
boolean hasColumnFamily();
|
||||||
com.google.protobuf.ByteString getColumnFamily();
|
com.google.protobuf.ByteString getColumnFamily();
|
||||||
|
|
||||||
// optional bytes columnQualifier = 3;
|
// optional bytes column_qualifier = 3;
|
||||||
boolean hasColumnQualifier();
|
boolean hasColumnQualifier();
|
||||||
com.google.protobuf.ByteString getColumnQualifier();
|
com.google.protobuf.ByteString getColumnQualifier();
|
||||||
|
|
||||||
// optional bool dropDependentColumn = 4;
|
// optional bool drop_dependent_column = 4;
|
||||||
boolean hasDropDependentColumn();
|
boolean hasDropDependentColumn();
|
||||||
boolean getDropDependentColumn();
|
boolean getDropDependentColumn();
|
||||||
}
|
}
|
||||||
|
@ -2944,8 +2944,8 @@ public final class FilterProtos {
|
||||||
}
|
}
|
||||||
|
|
||||||
private int bitField0_;
|
private int bitField0_;
|
||||||
// required .CompareFilter compareFilter = 1;
|
// required .CompareFilter compare_filter = 1;
|
||||||
public static final int COMPAREFILTER_FIELD_NUMBER = 1;
|
public static final int COMPARE_FILTER_FIELD_NUMBER = 1;
|
||||||
private org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter compareFilter_;
|
private org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter compareFilter_;
|
||||||
public boolean hasCompareFilter() {
|
public boolean hasCompareFilter() {
|
||||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||||
|
@ -2957,8 +2957,8 @@ public final class FilterProtos {
|
||||||
return compareFilter_;
|
return compareFilter_;
|
||||||
}
|
}
|
||||||
|
|
||||||
// optional bytes columnFamily = 2;
|
// optional bytes column_family = 2;
|
||||||
public static final int COLUMNFAMILY_FIELD_NUMBER = 2;
|
public static final int COLUMN_FAMILY_FIELD_NUMBER = 2;
|
||||||
private com.google.protobuf.ByteString columnFamily_;
|
private com.google.protobuf.ByteString columnFamily_;
|
||||||
public boolean hasColumnFamily() {
|
public boolean hasColumnFamily() {
|
||||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||||
|
@ -2967,8 +2967,8 @@ public final class FilterProtos {
|
||||||
return columnFamily_;
|
return columnFamily_;
|
||||||
}
|
}
|
||||||
|
|
||||||
// optional bytes columnQualifier = 3;
|
// optional bytes column_qualifier = 3;
|
||||||
public static final int COLUMNQUALIFIER_FIELD_NUMBER = 3;
|
public static final int COLUMN_QUALIFIER_FIELD_NUMBER = 3;
|
||||||
private com.google.protobuf.ByteString columnQualifier_;
|
private com.google.protobuf.ByteString columnQualifier_;
|
||||||
public boolean hasColumnQualifier() {
|
public boolean hasColumnQualifier() {
|
||||||
return ((bitField0_ & 0x00000004) == 0x00000004);
|
return ((bitField0_ & 0x00000004) == 0x00000004);
|
||||||
|
@ -2977,8 +2977,8 @@ public final class FilterProtos {
|
||||||
return columnQualifier_;
|
return columnQualifier_;
|
||||||
}
|
}
|
||||||
|
|
||||||
// optional bool dropDependentColumn = 4;
|
// optional bool drop_dependent_column = 4;
|
||||||
public static final int DROPDEPENDENTCOLUMN_FIELD_NUMBER = 4;
|
public static final int DROP_DEPENDENT_COLUMN_FIELD_NUMBER = 4;
|
||||||
private boolean dropDependentColumn_;
|
private boolean dropDependentColumn_;
|
||||||
public boolean hasDropDependentColumn() {
|
public boolean hasDropDependentColumn() {
|
||||||
return ((bitField0_ & 0x00000008) == 0x00000008);
|
return ((bitField0_ & 0x00000008) == 0x00000008);
|
||||||
|
@ -3103,19 +3103,19 @@ public final class FilterProtos {
|
||||||
int hash = 41;
|
int hash = 41;
|
||||||
hash = (19 * hash) + getDescriptorForType().hashCode();
|
hash = (19 * hash) + getDescriptorForType().hashCode();
|
||||||
if (hasCompareFilter()) {
|
if (hasCompareFilter()) {
|
||||||
hash = (37 * hash) + COMPAREFILTER_FIELD_NUMBER;
|
hash = (37 * hash) + COMPARE_FILTER_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + getCompareFilter().hashCode();
|
hash = (53 * hash) + getCompareFilter().hashCode();
|
||||||
}
|
}
|
||||||
if (hasColumnFamily()) {
|
if (hasColumnFamily()) {
|
||||||
hash = (37 * hash) + COLUMNFAMILY_FIELD_NUMBER;
|
hash = (37 * hash) + COLUMN_FAMILY_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + getColumnFamily().hashCode();
|
hash = (53 * hash) + getColumnFamily().hashCode();
|
||||||
}
|
}
|
||||||
if (hasColumnQualifier()) {
|
if (hasColumnQualifier()) {
|
||||||
hash = (37 * hash) + COLUMNQUALIFIER_FIELD_NUMBER;
|
hash = (37 * hash) + COLUMN_QUALIFIER_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + getColumnQualifier().hashCode();
|
hash = (53 * hash) + getColumnQualifier().hashCode();
|
||||||
}
|
}
|
||||||
if (hasDropDependentColumn()) {
|
if (hasDropDependentColumn()) {
|
||||||
hash = (37 * hash) + DROPDEPENDENTCOLUMN_FIELD_NUMBER;
|
hash = (37 * hash) + DROP_DEPENDENT_COLUMN_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + hashBoolean(getDropDependentColumn());
|
hash = (53 * hash) + hashBoolean(getDropDependentColumn());
|
||||||
}
|
}
|
||||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||||
|
@ -3402,7 +3402,7 @@ public final class FilterProtos {
|
||||||
|
|
||||||
private int bitField0_;
|
private int bitField0_;
|
||||||
|
|
||||||
// required .CompareFilter compareFilter = 1;
|
// required .CompareFilter compare_filter = 1;
|
||||||
private org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter compareFilter_ = org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter.getDefaultInstance();
|
private org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter compareFilter_ = org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter.getDefaultInstance();
|
||||||
private com.google.protobuf.SingleFieldBuilder<
|
private com.google.protobuf.SingleFieldBuilder<
|
||||||
org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter, org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter.Builder, org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilterOrBuilder> compareFilterBuilder_;
|
org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter, org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter.Builder, org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilterOrBuilder> compareFilterBuilder_;
|
||||||
|
@ -3492,7 +3492,7 @@ public final class FilterProtos {
|
||||||
return compareFilterBuilder_;
|
return compareFilterBuilder_;
|
||||||
}
|
}
|
||||||
|
|
||||||
// optional bytes columnFamily = 2;
|
// optional bytes column_family = 2;
|
||||||
private com.google.protobuf.ByteString columnFamily_ = com.google.protobuf.ByteString.EMPTY;
|
private com.google.protobuf.ByteString columnFamily_ = com.google.protobuf.ByteString.EMPTY;
|
||||||
public boolean hasColumnFamily() {
|
public boolean hasColumnFamily() {
|
||||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||||
|
@ -3516,7 +3516,7 @@ public final class FilterProtos {
|
||||||
return this;
|
return this;
|
||||||
}
|
}
|
||||||
|
|
||||||
// optional bytes columnQualifier = 3;
|
// optional bytes column_qualifier = 3;
|
||||||
private com.google.protobuf.ByteString columnQualifier_ = com.google.protobuf.ByteString.EMPTY;
|
private com.google.protobuf.ByteString columnQualifier_ = com.google.protobuf.ByteString.EMPTY;
|
||||||
public boolean hasColumnQualifier() {
|
public boolean hasColumnQualifier() {
|
||||||
return ((bitField0_ & 0x00000004) == 0x00000004);
|
return ((bitField0_ & 0x00000004) == 0x00000004);
|
||||||
|
@ -3540,7 +3540,7 @@ public final class FilterProtos {
|
||||||
return this;
|
return this;
|
||||||
}
|
}
|
||||||
|
|
||||||
// optional bool dropDependentColumn = 4;
|
// optional bool drop_dependent_column = 4;
|
||||||
private boolean dropDependentColumn_ ;
|
private boolean dropDependentColumn_ ;
|
||||||
public boolean hasDropDependentColumn() {
|
public boolean hasDropDependentColumn() {
|
||||||
return ((bitField0_ & 0x00000008) == 0x00000008);
|
return ((bitField0_ & 0x00000008) == 0x00000008);
|
||||||
|
@ -3575,7 +3575,7 @@ public final class FilterProtos {
|
||||||
public interface FamilyFilterOrBuilder
|
public interface FamilyFilterOrBuilder
|
||||||
extends com.google.protobuf.MessageOrBuilder {
|
extends com.google.protobuf.MessageOrBuilder {
|
||||||
|
|
||||||
// required .CompareFilter compareFilter = 1;
|
// required .CompareFilter compare_filter = 1;
|
||||||
boolean hasCompareFilter();
|
boolean hasCompareFilter();
|
||||||
org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter getCompareFilter();
|
org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter getCompareFilter();
|
||||||
org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilterOrBuilder getCompareFilterOrBuilder();
|
org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilterOrBuilder getCompareFilterOrBuilder();
|
||||||
|
@ -3609,8 +3609,8 @@ public final class FilterProtos {
|
||||||
}
|
}
|
||||||
|
|
||||||
private int bitField0_;
|
private int bitField0_;
|
||||||
// required .CompareFilter compareFilter = 1;
|
// required .CompareFilter compare_filter = 1;
|
||||||
public static final int COMPAREFILTER_FIELD_NUMBER = 1;
|
public static final int COMPARE_FILTER_FIELD_NUMBER = 1;
|
||||||
private org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter compareFilter_;
|
private org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter compareFilter_;
|
||||||
public boolean hasCompareFilter() {
|
public boolean hasCompareFilter() {
|
||||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||||
|
@ -3699,7 +3699,7 @@ public final class FilterProtos {
|
||||||
int hash = 41;
|
int hash = 41;
|
||||||
hash = (19 * hash) + getDescriptorForType().hashCode();
|
hash = (19 * hash) + getDescriptorForType().hashCode();
|
||||||
if (hasCompareFilter()) {
|
if (hasCompareFilter()) {
|
||||||
hash = (37 * hash) + COMPAREFILTER_FIELD_NUMBER;
|
hash = (37 * hash) + COMPARE_FILTER_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + getCompareFilter().hashCode();
|
hash = (53 * hash) + getCompareFilter().hashCode();
|
||||||
}
|
}
|
||||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||||
|
@ -3944,7 +3944,7 @@ public final class FilterProtos {
|
||||||
|
|
||||||
private int bitField0_;
|
private int bitField0_;
|
||||||
|
|
||||||
// required .CompareFilter compareFilter = 1;
|
// required .CompareFilter compare_filter = 1;
|
||||||
private org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter compareFilter_ = org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter.getDefaultInstance();
|
private org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter compareFilter_ = org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter.getDefaultInstance();
|
||||||
private com.google.protobuf.SingleFieldBuilder<
|
private com.google.protobuf.SingleFieldBuilder<
|
||||||
org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter, org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter.Builder, org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilterOrBuilder> compareFilterBuilder_;
|
org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter, org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter.Builder, org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilterOrBuilder> compareFilterBuilder_;
|
||||||
|
@ -5982,7 +5982,7 @@ public final class FilterProtos {
|
||||||
public interface FuzzyRowFilterOrBuilder
|
public interface FuzzyRowFilterOrBuilder
|
||||||
extends com.google.protobuf.MessageOrBuilder {
|
extends com.google.protobuf.MessageOrBuilder {
|
||||||
|
|
||||||
// repeated .BytesBytesPair fuzzyKeysData = 1;
|
// repeated .BytesBytesPair fuzzy_keys_data = 1;
|
||||||
java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair>
|
java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair>
|
||||||
getFuzzyKeysDataList();
|
getFuzzyKeysDataList();
|
||||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair getFuzzyKeysData(int index);
|
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair getFuzzyKeysData(int index);
|
||||||
|
@ -6020,8 +6020,8 @@ public final class FilterProtos {
|
||||||
return org.apache.hadoop.hbase.protobuf.generated.FilterProtos.internal_static_FuzzyRowFilter_fieldAccessorTable;
|
return org.apache.hadoop.hbase.protobuf.generated.FilterProtos.internal_static_FuzzyRowFilter_fieldAccessorTable;
|
||||||
}
|
}
|
||||||
|
|
||||||
// repeated .BytesBytesPair fuzzyKeysData = 1;
|
// repeated .BytesBytesPair fuzzy_keys_data = 1;
|
||||||
public static final int FUZZYKEYSDATA_FIELD_NUMBER = 1;
|
public static final int FUZZY_KEYS_DATA_FIELD_NUMBER = 1;
|
||||||
private java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair> fuzzyKeysData_;
|
private java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair> fuzzyKeysData_;
|
||||||
public java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair> getFuzzyKeysDataList() {
|
public java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair> getFuzzyKeysDataList() {
|
||||||
return fuzzyKeysData_;
|
return fuzzyKeysData_;
|
||||||
|
@ -6113,7 +6113,7 @@ public final class FilterProtos {
|
||||||
int hash = 41;
|
int hash = 41;
|
||||||
hash = (19 * hash) + getDescriptorForType().hashCode();
|
hash = (19 * hash) + getDescriptorForType().hashCode();
|
||||||
if (getFuzzyKeysDataCount() > 0) {
|
if (getFuzzyKeysDataCount() > 0) {
|
||||||
hash = (37 * hash) + FUZZYKEYSDATA_FIELD_NUMBER;
|
hash = (37 * hash) + FUZZY_KEYS_DATA_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + getFuzzyKeysDataList().hashCode();
|
hash = (53 * hash) + getFuzzyKeysDataList().hashCode();
|
||||||
}
|
}
|
||||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||||
|
@ -6375,7 +6375,7 @@ public final class FilterProtos {
|
||||||
|
|
||||||
private int bitField0_;
|
private int bitField0_;
|
||||||
|
|
||||||
// repeated .BytesBytesPair fuzzyKeysData = 1;
|
// repeated .BytesBytesPair fuzzy_keys_data = 1;
|
||||||
private java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair> fuzzyKeysData_ =
|
private java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair> fuzzyKeysData_ =
|
||||||
java.util.Collections.emptyList();
|
java.util.Collections.emptyList();
|
||||||
private void ensureFuzzyKeysDataIsMutable() {
|
private void ensureFuzzyKeysDataIsMutable() {
|
||||||
|
@ -6575,7 +6575,7 @@ public final class FilterProtos {
|
||||||
public interface InclusiveStopFilterOrBuilder
|
public interface InclusiveStopFilterOrBuilder
|
||||||
extends com.google.protobuf.MessageOrBuilder {
|
extends com.google.protobuf.MessageOrBuilder {
|
||||||
|
|
||||||
// optional bytes stopRowKey = 1;
|
// optional bytes stop_row_key = 1;
|
||||||
boolean hasStopRowKey();
|
boolean hasStopRowKey();
|
||||||
com.google.protobuf.ByteString getStopRowKey();
|
com.google.protobuf.ByteString getStopRowKey();
|
||||||
}
|
}
|
||||||
|
@ -6608,8 +6608,8 @@ public final class FilterProtos {
|
||||||
}
|
}
|
||||||
|
|
||||||
private int bitField0_;
|
private int bitField0_;
|
||||||
// optional bytes stopRowKey = 1;
|
// optional bytes stop_row_key = 1;
|
||||||
public static final int STOPROWKEY_FIELD_NUMBER = 1;
|
public static final int STOP_ROW_KEY_FIELD_NUMBER = 1;
|
||||||
private com.google.protobuf.ByteString stopRowKey_;
|
private com.google.protobuf.ByteString stopRowKey_;
|
||||||
public boolean hasStopRowKey() {
|
public boolean hasStopRowKey() {
|
||||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||||
|
@ -6687,7 +6687,7 @@ public final class FilterProtos {
|
||||||
int hash = 41;
|
int hash = 41;
|
||||||
hash = (19 * hash) + getDescriptorForType().hashCode();
|
hash = (19 * hash) + getDescriptorForType().hashCode();
|
||||||
if (hasStopRowKey()) {
|
if (hasStopRowKey()) {
|
||||||
hash = (37 * hash) + STOPROWKEY_FIELD_NUMBER;
|
hash = (37 * hash) + STOP_ROW_KEY_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + getStopRowKey().hashCode();
|
hash = (53 * hash) + getStopRowKey().hashCode();
|
||||||
}
|
}
|
||||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||||
|
@ -6911,7 +6911,7 @@ public final class FilterProtos {
|
||||||
|
|
||||||
private int bitField0_;
|
private int bitField0_;
|
||||||
|
|
||||||
// optional bytes stopRowKey = 1;
|
// optional bytes stop_row_key = 1;
|
||||||
private com.google.protobuf.ByteString stopRowKey_ = com.google.protobuf.ByteString.EMPTY;
|
private com.google.protobuf.ByteString stopRowKey_ = com.google.protobuf.ByteString.EMPTY;
|
||||||
public boolean hasStopRowKey() {
|
public boolean hasStopRowKey() {
|
||||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||||
|
@ -6949,7 +6949,7 @@ public final class FilterProtos {
|
||||||
public interface KeyOnlyFilterOrBuilder
|
public interface KeyOnlyFilterOrBuilder
|
||||||
extends com.google.protobuf.MessageOrBuilder {
|
extends com.google.protobuf.MessageOrBuilder {
|
||||||
|
|
||||||
// required bool lenAsVal = 1;
|
// required bool len_as_val = 1;
|
||||||
boolean hasLenAsVal();
|
boolean hasLenAsVal();
|
||||||
boolean getLenAsVal();
|
boolean getLenAsVal();
|
||||||
}
|
}
|
||||||
|
@ -6982,8 +6982,8 @@ public final class FilterProtos {
|
||||||
}
|
}
|
||||||
|
|
||||||
private int bitField0_;
|
private int bitField0_;
|
||||||
// required bool lenAsVal = 1;
|
// required bool len_as_val = 1;
|
||||||
public static final int LENASVAL_FIELD_NUMBER = 1;
|
public static final int LEN_AS_VAL_FIELD_NUMBER = 1;
|
||||||
private boolean lenAsVal_;
|
private boolean lenAsVal_;
|
||||||
public boolean hasLenAsVal() {
|
public boolean hasLenAsVal() {
|
||||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||||
|
@ -7065,7 +7065,7 @@ public final class FilterProtos {
|
||||||
int hash = 41;
|
int hash = 41;
|
||||||
hash = (19 * hash) + getDescriptorForType().hashCode();
|
hash = (19 * hash) + getDescriptorForType().hashCode();
|
||||||
if (hasLenAsVal()) {
|
if (hasLenAsVal()) {
|
||||||
hash = (37 * hash) + LENASVAL_FIELD_NUMBER;
|
hash = (37 * hash) + LEN_AS_VAL_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + hashBoolean(getLenAsVal());
|
hash = (53 * hash) + hashBoolean(getLenAsVal());
|
||||||
}
|
}
|
||||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||||
|
@ -7293,7 +7293,7 @@ public final class FilterProtos {
|
||||||
|
|
||||||
private int bitField0_;
|
private int bitField0_;
|
||||||
|
|
||||||
// required bool lenAsVal = 1;
|
// required bool len_as_val = 1;
|
||||||
private boolean lenAsVal_ ;
|
private boolean lenAsVal_ ;
|
||||||
public boolean hasLenAsVal() {
|
public boolean hasLenAsVal() {
|
||||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||||
|
@ -7328,7 +7328,7 @@ public final class FilterProtos {
|
||||||
public interface MultipleColumnPrefixFilterOrBuilder
|
public interface MultipleColumnPrefixFilterOrBuilder
|
||||||
extends com.google.protobuf.MessageOrBuilder {
|
extends com.google.protobuf.MessageOrBuilder {
|
||||||
|
|
||||||
// repeated bytes sortedPrefixes = 1;
|
// repeated bytes sorted_prefixes = 1;
|
||||||
java.util.List<com.google.protobuf.ByteString> getSortedPrefixesList();
|
java.util.List<com.google.protobuf.ByteString> getSortedPrefixesList();
|
||||||
int getSortedPrefixesCount();
|
int getSortedPrefixesCount();
|
||||||
com.google.protobuf.ByteString getSortedPrefixes(int index);
|
com.google.protobuf.ByteString getSortedPrefixes(int index);
|
||||||
|
@ -7361,8 +7361,8 @@ public final class FilterProtos {
|
||||||
return org.apache.hadoop.hbase.protobuf.generated.FilterProtos.internal_static_MultipleColumnPrefixFilter_fieldAccessorTable;
|
return org.apache.hadoop.hbase.protobuf.generated.FilterProtos.internal_static_MultipleColumnPrefixFilter_fieldAccessorTable;
|
||||||
}
|
}
|
||||||
|
|
||||||
// repeated bytes sortedPrefixes = 1;
|
// repeated bytes sorted_prefixes = 1;
|
||||||
public static final int SORTEDPREFIXES_FIELD_NUMBER = 1;
|
public static final int SORTED_PREFIXES_FIELD_NUMBER = 1;
|
||||||
private java.util.List<com.google.protobuf.ByteString> sortedPrefixes_;
|
private java.util.List<com.google.protobuf.ByteString> sortedPrefixes_;
|
||||||
public java.util.List<com.google.protobuf.ByteString>
|
public java.util.List<com.google.protobuf.ByteString>
|
||||||
getSortedPrefixesList() {
|
getSortedPrefixesList() {
|
||||||
|
@ -7446,7 +7446,7 @@ public final class FilterProtos {
|
||||||
int hash = 41;
|
int hash = 41;
|
||||||
hash = (19 * hash) + getDescriptorForType().hashCode();
|
hash = (19 * hash) + getDescriptorForType().hashCode();
|
||||||
if (getSortedPrefixesCount() > 0) {
|
if (getSortedPrefixesCount() > 0) {
|
||||||
hash = (37 * hash) + SORTEDPREFIXES_FIELD_NUMBER;
|
hash = (37 * hash) + SORTED_PREFIXES_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + getSortedPrefixesList().hashCode();
|
hash = (53 * hash) + getSortedPrefixesList().hashCode();
|
||||||
}
|
}
|
||||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||||
|
@ -7676,7 +7676,7 @@ public final class FilterProtos {
|
||||||
|
|
||||||
private int bitField0_;
|
private int bitField0_;
|
||||||
|
|
||||||
// repeated bytes sortedPrefixes = 1;
|
// repeated bytes sorted_prefixes = 1;
|
||||||
private java.util.List<com.google.protobuf.ByteString> sortedPrefixes_ = java.util.Collections.emptyList();;
|
private java.util.List<com.google.protobuf.ByteString> sortedPrefixes_ = java.util.Collections.emptyList();;
|
||||||
private void ensureSortedPrefixesIsMutable() {
|
private void ensureSortedPrefixesIsMutable() {
|
||||||
if (!((bitField0_ & 0x00000001) == 0x00000001)) {
|
if (!((bitField0_ & 0x00000001) == 0x00000001)) {
|
||||||
|
@ -7741,7 +7741,7 @@ public final class FilterProtos {
|
||||||
public interface PageFilterOrBuilder
|
public interface PageFilterOrBuilder
|
||||||
extends com.google.protobuf.MessageOrBuilder {
|
extends com.google.protobuf.MessageOrBuilder {
|
||||||
|
|
||||||
// required int64 pageSize = 1;
|
// required int64 page_size = 1;
|
||||||
boolean hasPageSize();
|
boolean hasPageSize();
|
||||||
long getPageSize();
|
long getPageSize();
|
||||||
}
|
}
|
||||||
|
@ -7774,8 +7774,8 @@ public final class FilterProtos {
|
||||||
}
|
}
|
||||||
|
|
||||||
private int bitField0_;
|
private int bitField0_;
|
||||||
// required int64 pageSize = 1;
|
// required int64 page_size = 1;
|
||||||
public static final int PAGESIZE_FIELD_NUMBER = 1;
|
public static final int PAGE_SIZE_FIELD_NUMBER = 1;
|
||||||
private long pageSize_;
|
private long pageSize_;
|
||||||
public boolean hasPageSize() {
|
public boolean hasPageSize() {
|
||||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||||
|
@ -7857,7 +7857,7 @@ public final class FilterProtos {
|
||||||
int hash = 41;
|
int hash = 41;
|
||||||
hash = (19 * hash) + getDescriptorForType().hashCode();
|
hash = (19 * hash) + getDescriptorForType().hashCode();
|
||||||
if (hasPageSize()) {
|
if (hasPageSize()) {
|
||||||
hash = (37 * hash) + PAGESIZE_FIELD_NUMBER;
|
hash = (37 * hash) + PAGE_SIZE_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + hashLong(getPageSize());
|
hash = (53 * hash) + hashLong(getPageSize());
|
||||||
}
|
}
|
||||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||||
|
@ -8085,7 +8085,7 @@ public final class FilterProtos {
|
||||||
|
|
||||||
private int bitField0_;
|
private int bitField0_;
|
||||||
|
|
||||||
// required int64 pageSize = 1;
|
// required int64 page_size = 1;
|
||||||
private long pageSize_ ;
|
private long pageSize_ ;
|
||||||
public boolean hasPageSize() {
|
public boolean hasPageSize() {
|
||||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||||
|
@ -8494,7 +8494,7 @@ public final class FilterProtos {
|
||||||
public interface QualifierFilterOrBuilder
|
public interface QualifierFilterOrBuilder
|
||||||
extends com.google.protobuf.MessageOrBuilder {
|
extends com.google.protobuf.MessageOrBuilder {
|
||||||
|
|
||||||
// required .CompareFilter compareFilter = 1;
|
// required .CompareFilter compare_filter = 1;
|
||||||
boolean hasCompareFilter();
|
boolean hasCompareFilter();
|
||||||
org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter getCompareFilter();
|
org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter getCompareFilter();
|
||||||
org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilterOrBuilder getCompareFilterOrBuilder();
|
org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilterOrBuilder getCompareFilterOrBuilder();
|
||||||
|
@ -8528,8 +8528,8 @@ public final class FilterProtos {
|
||||||
}
|
}
|
||||||
|
|
||||||
private int bitField0_;
|
private int bitField0_;
|
||||||
// required .CompareFilter compareFilter = 1;
|
// required .CompareFilter compare_filter = 1;
|
||||||
public static final int COMPAREFILTER_FIELD_NUMBER = 1;
|
public static final int COMPARE_FILTER_FIELD_NUMBER = 1;
|
||||||
private org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter compareFilter_;
|
private org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter compareFilter_;
|
||||||
public boolean hasCompareFilter() {
|
public boolean hasCompareFilter() {
|
||||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||||
|
@ -8618,7 +8618,7 @@ public final class FilterProtos {
|
||||||
int hash = 41;
|
int hash = 41;
|
||||||
hash = (19 * hash) + getDescriptorForType().hashCode();
|
hash = (19 * hash) + getDescriptorForType().hashCode();
|
||||||
if (hasCompareFilter()) {
|
if (hasCompareFilter()) {
|
||||||
hash = (37 * hash) + COMPAREFILTER_FIELD_NUMBER;
|
hash = (37 * hash) + COMPARE_FILTER_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + getCompareFilter().hashCode();
|
hash = (53 * hash) + getCompareFilter().hashCode();
|
||||||
}
|
}
|
||||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||||
|
@ -8863,7 +8863,7 @@ public final class FilterProtos {
|
||||||
|
|
||||||
private int bitField0_;
|
private int bitField0_;
|
||||||
|
|
||||||
// required .CompareFilter compareFilter = 1;
|
// required .CompareFilter compare_filter = 1;
|
||||||
private org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter compareFilter_ = org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter.getDefaultInstance();
|
private org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter compareFilter_ = org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter.getDefaultInstance();
|
||||||
private com.google.protobuf.SingleFieldBuilder<
|
private com.google.protobuf.SingleFieldBuilder<
|
||||||
org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter, org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter.Builder, org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilterOrBuilder> compareFilterBuilder_;
|
org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter, org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter.Builder, org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilterOrBuilder> compareFilterBuilder_;
|
||||||
|
@ -9346,7 +9346,7 @@ public final class FilterProtos {
|
||||||
public interface RowFilterOrBuilder
|
public interface RowFilterOrBuilder
|
||||||
extends com.google.protobuf.MessageOrBuilder {
|
extends com.google.protobuf.MessageOrBuilder {
|
||||||
|
|
||||||
// required .CompareFilter compareFilter = 1;
|
// required .CompareFilter compare_filter = 1;
|
||||||
boolean hasCompareFilter();
|
boolean hasCompareFilter();
|
||||||
org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter getCompareFilter();
|
org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter getCompareFilter();
|
||||||
org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilterOrBuilder getCompareFilterOrBuilder();
|
org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilterOrBuilder getCompareFilterOrBuilder();
|
||||||
|
@ -9380,8 +9380,8 @@ public final class FilterProtos {
|
||||||
}
|
}
|
||||||
|
|
||||||
private int bitField0_;
|
private int bitField0_;
|
||||||
// required .CompareFilter compareFilter = 1;
|
// required .CompareFilter compare_filter = 1;
|
||||||
public static final int COMPAREFILTER_FIELD_NUMBER = 1;
|
public static final int COMPARE_FILTER_FIELD_NUMBER = 1;
|
||||||
private org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter compareFilter_;
|
private org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter compareFilter_;
|
||||||
public boolean hasCompareFilter() {
|
public boolean hasCompareFilter() {
|
||||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||||
|
@ -9470,7 +9470,7 @@ public final class FilterProtos {
|
||||||
int hash = 41;
|
int hash = 41;
|
||||||
hash = (19 * hash) + getDescriptorForType().hashCode();
|
hash = (19 * hash) + getDescriptorForType().hashCode();
|
||||||
if (hasCompareFilter()) {
|
if (hasCompareFilter()) {
|
||||||
hash = (37 * hash) + COMPAREFILTER_FIELD_NUMBER;
|
hash = (37 * hash) + COMPARE_FILTER_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + getCompareFilter().hashCode();
|
hash = (53 * hash) + getCompareFilter().hashCode();
|
||||||
}
|
}
|
||||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||||
|
@ -9715,7 +9715,7 @@ public final class FilterProtos {
|
||||||
|
|
||||||
private int bitField0_;
|
private int bitField0_;
|
||||||
|
|
||||||
// required .CompareFilter compareFilter = 1;
|
// required .CompareFilter compare_filter = 1;
|
||||||
private org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter compareFilter_ = org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter.getDefaultInstance();
|
private org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter compareFilter_ = org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter.getDefaultInstance();
|
||||||
private com.google.protobuf.SingleFieldBuilder<
|
private com.google.protobuf.SingleFieldBuilder<
|
||||||
org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter, org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter.Builder, org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilterOrBuilder> compareFilterBuilder_;
|
org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter, org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter.Builder, org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilterOrBuilder> compareFilterBuilder_;
|
||||||
|
@ -9819,7 +9819,7 @@ public final class FilterProtos {
|
||||||
public interface SingleColumnValueExcludeFilterOrBuilder
|
public interface SingleColumnValueExcludeFilterOrBuilder
|
||||||
extends com.google.protobuf.MessageOrBuilder {
|
extends com.google.protobuf.MessageOrBuilder {
|
||||||
|
|
||||||
// required .SingleColumnValueFilter singleColumnValueFilter = 1;
|
// required .SingleColumnValueFilter single_column_value_filter = 1;
|
||||||
boolean hasSingleColumnValueFilter();
|
boolean hasSingleColumnValueFilter();
|
||||||
org.apache.hadoop.hbase.protobuf.generated.FilterProtos.SingleColumnValueFilter getSingleColumnValueFilter();
|
org.apache.hadoop.hbase.protobuf.generated.FilterProtos.SingleColumnValueFilter getSingleColumnValueFilter();
|
||||||
org.apache.hadoop.hbase.protobuf.generated.FilterProtos.SingleColumnValueFilterOrBuilder getSingleColumnValueFilterOrBuilder();
|
org.apache.hadoop.hbase.protobuf.generated.FilterProtos.SingleColumnValueFilterOrBuilder getSingleColumnValueFilterOrBuilder();
|
||||||
|
@ -9853,8 +9853,8 @@ public final class FilterProtos {
|
||||||
}
|
}
|
||||||
|
|
||||||
private int bitField0_;
|
private int bitField0_;
|
||||||
// required .SingleColumnValueFilter singleColumnValueFilter = 1;
|
// required .SingleColumnValueFilter single_column_value_filter = 1;
|
||||||
public static final int SINGLECOLUMNVALUEFILTER_FIELD_NUMBER = 1;
|
public static final int SINGLE_COLUMN_VALUE_FILTER_FIELD_NUMBER = 1;
|
||||||
private org.apache.hadoop.hbase.protobuf.generated.FilterProtos.SingleColumnValueFilter singleColumnValueFilter_;
|
private org.apache.hadoop.hbase.protobuf.generated.FilterProtos.SingleColumnValueFilter singleColumnValueFilter_;
|
||||||
public boolean hasSingleColumnValueFilter() {
|
public boolean hasSingleColumnValueFilter() {
|
||||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||||
|
@ -9943,7 +9943,7 @@ public final class FilterProtos {
|
||||||
int hash = 41;
|
int hash = 41;
|
||||||
hash = (19 * hash) + getDescriptorForType().hashCode();
|
hash = (19 * hash) + getDescriptorForType().hashCode();
|
||||||
if (hasSingleColumnValueFilter()) {
|
if (hasSingleColumnValueFilter()) {
|
||||||
hash = (37 * hash) + SINGLECOLUMNVALUEFILTER_FIELD_NUMBER;
|
hash = (37 * hash) + SINGLE_COLUMN_VALUE_FILTER_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + getSingleColumnValueFilter().hashCode();
|
hash = (53 * hash) + getSingleColumnValueFilter().hashCode();
|
||||||
}
|
}
|
||||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||||
|
@ -10188,7 +10188,7 @@ public final class FilterProtos {
|
||||||
|
|
||||||
private int bitField0_;
|
private int bitField0_;
|
||||||
|
|
||||||
// required .SingleColumnValueFilter singleColumnValueFilter = 1;
|
// required .SingleColumnValueFilter single_column_value_filter = 1;
|
||||||
private org.apache.hadoop.hbase.protobuf.generated.FilterProtos.SingleColumnValueFilter singleColumnValueFilter_ = org.apache.hadoop.hbase.protobuf.generated.FilterProtos.SingleColumnValueFilter.getDefaultInstance();
|
private org.apache.hadoop.hbase.protobuf.generated.FilterProtos.SingleColumnValueFilter singleColumnValueFilter_ = org.apache.hadoop.hbase.protobuf.generated.FilterProtos.SingleColumnValueFilter.getDefaultInstance();
|
||||||
private com.google.protobuf.SingleFieldBuilder<
|
private com.google.protobuf.SingleFieldBuilder<
|
||||||
org.apache.hadoop.hbase.protobuf.generated.FilterProtos.SingleColumnValueFilter, org.apache.hadoop.hbase.protobuf.generated.FilterProtos.SingleColumnValueFilter.Builder, org.apache.hadoop.hbase.protobuf.generated.FilterProtos.SingleColumnValueFilterOrBuilder> singleColumnValueFilterBuilder_;
|
org.apache.hadoop.hbase.protobuf.generated.FilterProtos.SingleColumnValueFilter, org.apache.hadoop.hbase.protobuf.generated.FilterProtos.SingleColumnValueFilter.Builder, org.apache.hadoop.hbase.protobuf.generated.FilterProtos.SingleColumnValueFilterOrBuilder> singleColumnValueFilterBuilder_;
|
||||||
|
@ -10292,15 +10292,15 @@ public final class FilterProtos {
|
||||||
public interface SingleColumnValueFilterOrBuilder
|
public interface SingleColumnValueFilterOrBuilder
|
||||||
extends com.google.protobuf.MessageOrBuilder {
|
extends com.google.protobuf.MessageOrBuilder {
|
||||||
|
|
||||||
// optional bytes columnFamily = 1;
|
// optional bytes column_family = 1;
|
||||||
boolean hasColumnFamily();
|
boolean hasColumnFamily();
|
||||||
com.google.protobuf.ByteString getColumnFamily();
|
com.google.protobuf.ByteString getColumnFamily();
|
||||||
|
|
||||||
// optional bytes columnQualifier = 2;
|
// optional bytes column_qualifier = 2;
|
||||||
boolean hasColumnQualifier();
|
boolean hasColumnQualifier();
|
||||||
com.google.protobuf.ByteString getColumnQualifier();
|
com.google.protobuf.ByteString getColumnQualifier();
|
||||||
|
|
||||||
// required .CompareType compareOp = 3;
|
// required .CompareType compare_op = 3;
|
||||||
boolean hasCompareOp();
|
boolean hasCompareOp();
|
||||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.CompareType getCompareOp();
|
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.CompareType getCompareOp();
|
||||||
|
|
||||||
|
@ -10309,11 +10309,11 @@ public final class FilterProtos {
|
||||||
org.apache.hadoop.hbase.protobuf.generated.ComparatorProtos.Comparator getComparator();
|
org.apache.hadoop.hbase.protobuf.generated.ComparatorProtos.Comparator getComparator();
|
||||||
org.apache.hadoop.hbase.protobuf.generated.ComparatorProtos.ComparatorOrBuilder getComparatorOrBuilder();
|
org.apache.hadoop.hbase.protobuf.generated.ComparatorProtos.ComparatorOrBuilder getComparatorOrBuilder();
|
||||||
|
|
||||||
// optional bool filterIfMissing = 5;
|
// optional bool filter_if_missing = 5;
|
||||||
boolean hasFilterIfMissing();
|
boolean hasFilterIfMissing();
|
||||||
boolean getFilterIfMissing();
|
boolean getFilterIfMissing();
|
||||||
|
|
||||||
// optional bool latestVersionOnly = 6;
|
// optional bool latest_version_only = 6;
|
||||||
boolean hasLatestVersionOnly();
|
boolean hasLatestVersionOnly();
|
||||||
boolean getLatestVersionOnly();
|
boolean getLatestVersionOnly();
|
||||||
}
|
}
|
||||||
|
@ -10346,8 +10346,8 @@ public final class FilterProtos {
|
||||||
}
|
}
|
||||||
|
|
||||||
private int bitField0_;
|
private int bitField0_;
|
||||||
// optional bytes columnFamily = 1;
|
// optional bytes column_family = 1;
|
||||||
public static final int COLUMNFAMILY_FIELD_NUMBER = 1;
|
public static final int COLUMN_FAMILY_FIELD_NUMBER = 1;
|
||||||
private com.google.protobuf.ByteString columnFamily_;
|
private com.google.protobuf.ByteString columnFamily_;
|
||||||
public boolean hasColumnFamily() {
|
public boolean hasColumnFamily() {
|
||||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||||
|
@ -10356,8 +10356,8 @@ public final class FilterProtos {
|
||||||
return columnFamily_;
|
return columnFamily_;
|
||||||
}
|
}
|
||||||
|
|
||||||
// optional bytes columnQualifier = 2;
|
// optional bytes column_qualifier = 2;
|
||||||
public static final int COLUMNQUALIFIER_FIELD_NUMBER = 2;
|
public static final int COLUMN_QUALIFIER_FIELD_NUMBER = 2;
|
||||||
private com.google.protobuf.ByteString columnQualifier_;
|
private com.google.protobuf.ByteString columnQualifier_;
|
||||||
public boolean hasColumnQualifier() {
|
public boolean hasColumnQualifier() {
|
||||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||||
|
@ -10366,8 +10366,8 @@ public final class FilterProtos {
|
||||||
return columnQualifier_;
|
return columnQualifier_;
|
||||||
}
|
}
|
||||||
|
|
||||||
// required .CompareType compareOp = 3;
|
// required .CompareType compare_op = 3;
|
||||||
public static final int COMPAREOP_FIELD_NUMBER = 3;
|
public static final int COMPARE_OP_FIELD_NUMBER = 3;
|
||||||
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.CompareType compareOp_;
|
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.CompareType compareOp_;
|
||||||
public boolean hasCompareOp() {
|
public boolean hasCompareOp() {
|
||||||
return ((bitField0_ & 0x00000004) == 0x00000004);
|
return ((bitField0_ & 0x00000004) == 0x00000004);
|
||||||
|
@ -10389,8 +10389,8 @@ public final class FilterProtos {
|
||||||
return comparator_;
|
return comparator_;
|
||||||
}
|
}
|
||||||
|
|
||||||
// optional bool filterIfMissing = 5;
|
// optional bool filter_if_missing = 5;
|
||||||
public static final int FILTERIFMISSING_FIELD_NUMBER = 5;
|
public static final int FILTER_IF_MISSING_FIELD_NUMBER = 5;
|
||||||
private boolean filterIfMissing_;
|
private boolean filterIfMissing_;
|
||||||
public boolean hasFilterIfMissing() {
|
public boolean hasFilterIfMissing() {
|
||||||
return ((bitField0_ & 0x00000010) == 0x00000010);
|
return ((bitField0_ & 0x00000010) == 0x00000010);
|
||||||
|
@ -10399,8 +10399,8 @@ public final class FilterProtos {
|
||||||
return filterIfMissing_;
|
return filterIfMissing_;
|
||||||
}
|
}
|
||||||
|
|
||||||
// optional bool latestVersionOnly = 6;
|
// optional bool latest_version_only = 6;
|
||||||
public static final int LATESTVERSIONONLY_FIELD_NUMBER = 6;
|
public static final int LATEST_VERSION_ONLY_FIELD_NUMBER = 6;
|
||||||
private boolean latestVersionOnly_;
|
private boolean latestVersionOnly_;
|
||||||
public boolean hasLatestVersionOnly() {
|
public boolean hasLatestVersionOnly() {
|
||||||
return ((bitField0_ & 0x00000020) == 0x00000020);
|
return ((bitField0_ & 0x00000020) == 0x00000020);
|
||||||
|
@ -10555,15 +10555,15 @@ public final class FilterProtos {
|
||||||
int hash = 41;
|
int hash = 41;
|
||||||
hash = (19 * hash) + getDescriptorForType().hashCode();
|
hash = (19 * hash) + getDescriptorForType().hashCode();
|
||||||
if (hasColumnFamily()) {
|
if (hasColumnFamily()) {
|
||||||
hash = (37 * hash) + COLUMNFAMILY_FIELD_NUMBER;
|
hash = (37 * hash) + COLUMN_FAMILY_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + getColumnFamily().hashCode();
|
hash = (53 * hash) + getColumnFamily().hashCode();
|
||||||
}
|
}
|
||||||
if (hasColumnQualifier()) {
|
if (hasColumnQualifier()) {
|
||||||
hash = (37 * hash) + COLUMNQUALIFIER_FIELD_NUMBER;
|
hash = (37 * hash) + COLUMN_QUALIFIER_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + getColumnQualifier().hashCode();
|
hash = (53 * hash) + getColumnQualifier().hashCode();
|
||||||
}
|
}
|
||||||
if (hasCompareOp()) {
|
if (hasCompareOp()) {
|
||||||
hash = (37 * hash) + COMPAREOP_FIELD_NUMBER;
|
hash = (37 * hash) + COMPARE_OP_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + hashEnum(getCompareOp());
|
hash = (53 * hash) + hashEnum(getCompareOp());
|
||||||
}
|
}
|
||||||
if (hasComparator()) {
|
if (hasComparator()) {
|
||||||
|
@ -10571,11 +10571,11 @@ public final class FilterProtos {
|
||||||
hash = (53 * hash) + getComparator().hashCode();
|
hash = (53 * hash) + getComparator().hashCode();
|
||||||
}
|
}
|
||||||
if (hasFilterIfMissing()) {
|
if (hasFilterIfMissing()) {
|
||||||
hash = (37 * hash) + FILTERIFMISSING_FIELD_NUMBER;
|
hash = (37 * hash) + FILTER_IF_MISSING_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + hashBoolean(getFilterIfMissing());
|
hash = (53 * hash) + hashBoolean(getFilterIfMissing());
|
||||||
}
|
}
|
||||||
if (hasLatestVersionOnly()) {
|
if (hasLatestVersionOnly()) {
|
||||||
hash = (37 * hash) + LATESTVERSIONONLY_FIELD_NUMBER;
|
hash = (37 * hash) + LATEST_VERSION_ONLY_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + hashBoolean(getLatestVersionOnly());
|
hash = (53 * hash) + hashBoolean(getLatestVersionOnly());
|
||||||
}
|
}
|
||||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||||
|
@ -10900,7 +10900,7 @@ public final class FilterProtos {
|
||||||
|
|
||||||
private int bitField0_;
|
private int bitField0_;
|
||||||
|
|
||||||
// optional bytes columnFamily = 1;
|
// optional bytes column_family = 1;
|
||||||
private com.google.protobuf.ByteString columnFamily_ = com.google.protobuf.ByteString.EMPTY;
|
private com.google.protobuf.ByteString columnFamily_ = com.google.protobuf.ByteString.EMPTY;
|
||||||
public boolean hasColumnFamily() {
|
public boolean hasColumnFamily() {
|
||||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||||
|
@ -10924,7 +10924,7 @@ public final class FilterProtos {
|
||||||
return this;
|
return this;
|
||||||
}
|
}
|
||||||
|
|
||||||
// optional bytes columnQualifier = 2;
|
// optional bytes column_qualifier = 2;
|
||||||
private com.google.protobuf.ByteString columnQualifier_ = com.google.protobuf.ByteString.EMPTY;
|
private com.google.protobuf.ByteString columnQualifier_ = com.google.protobuf.ByteString.EMPTY;
|
||||||
public boolean hasColumnQualifier() {
|
public boolean hasColumnQualifier() {
|
||||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||||
|
@ -10948,7 +10948,7 @@ public final class FilterProtos {
|
||||||
return this;
|
return this;
|
||||||
}
|
}
|
||||||
|
|
||||||
// required .CompareType compareOp = 3;
|
// required .CompareType compare_op = 3;
|
||||||
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.CompareType compareOp_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.CompareType.LESS;
|
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.CompareType compareOp_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.CompareType.LESS;
|
||||||
public boolean hasCompareOp() {
|
public boolean hasCompareOp() {
|
||||||
return ((bitField0_ & 0x00000004) == 0x00000004);
|
return ((bitField0_ & 0x00000004) == 0x00000004);
|
||||||
|
@ -11062,7 +11062,7 @@ public final class FilterProtos {
|
||||||
return comparatorBuilder_;
|
return comparatorBuilder_;
|
||||||
}
|
}
|
||||||
|
|
||||||
// optional bool filterIfMissing = 5;
|
// optional bool filter_if_missing = 5;
|
||||||
private boolean filterIfMissing_ ;
|
private boolean filterIfMissing_ ;
|
||||||
public boolean hasFilterIfMissing() {
|
public boolean hasFilterIfMissing() {
|
||||||
return ((bitField0_ & 0x00000010) == 0x00000010);
|
return ((bitField0_ & 0x00000010) == 0x00000010);
|
||||||
|
@ -11083,7 +11083,7 @@ public final class FilterProtos {
|
||||||
return this;
|
return this;
|
||||||
}
|
}
|
||||||
|
|
||||||
// optional bool latestVersionOnly = 6;
|
// optional bool latest_version_only = 6;
|
||||||
private boolean latestVersionOnly_ ;
|
private boolean latestVersionOnly_ ;
|
||||||
public boolean hasLatestVersionOnly() {
|
public boolean hasLatestVersionOnly() {
|
||||||
return ((bitField0_ & 0x00000020) == 0x00000020);
|
return ((bitField0_ & 0x00000020) == 0x00000020);
|
||||||
|
@ -12017,7 +12017,7 @@ public final class FilterProtos {
|
||||||
public interface ValueFilterOrBuilder
|
public interface ValueFilterOrBuilder
|
||||||
extends com.google.protobuf.MessageOrBuilder {
|
extends com.google.protobuf.MessageOrBuilder {
|
||||||
|
|
||||||
// required .CompareFilter compareFilter = 1;
|
// required .CompareFilter compare_filter = 1;
|
||||||
boolean hasCompareFilter();
|
boolean hasCompareFilter();
|
||||||
org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter getCompareFilter();
|
org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter getCompareFilter();
|
||||||
org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilterOrBuilder getCompareFilterOrBuilder();
|
org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilterOrBuilder getCompareFilterOrBuilder();
|
||||||
|
@ -12051,8 +12051,8 @@ public final class FilterProtos {
|
||||||
}
|
}
|
||||||
|
|
||||||
private int bitField0_;
|
private int bitField0_;
|
||||||
// required .CompareFilter compareFilter = 1;
|
// required .CompareFilter compare_filter = 1;
|
||||||
public static final int COMPAREFILTER_FIELD_NUMBER = 1;
|
public static final int COMPARE_FILTER_FIELD_NUMBER = 1;
|
||||||
private org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter compareFilter_;
|
private org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter compareFilter_;
|
||||||
public boolean hasCompareFilter() {
|
public boolean hasCompareFilter() {
|
||||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||||
|
@ -12141,7 +12141,7 @@ public final class FilterProtos {
|
||||||
int hash = 41;
|
int hash = 41;
|
||||||
hash = (19 * hash) + getDescriptorForType().hashCode();
|
hash = (19 * hash) + getDescriptorForType().hashCode();
|
||||||
if (hasCompareFilter()) {
|
if (hasCompareFilter()) {
|
||||||
hash = (37 * hash) + COMPAREFILTER_FIELD_NUMBER;
|
hash = (37 * hash) + COMPARE_FILTER_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + getCompareFilter().hashCode();
|
hash = (53 * hash) + getCompareFilter().hashCode();
|
||||||
}
|
}
|
||||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||||
|
@ -12386,7 +12386,7 @@ public final class FilterProtos {
|
||||||
|
|
||||||
private int bitField0_;
|
private int bitField0_;
|
||||||
|
|
||||||
// required .CompareFilter compareFilter = 1;
|
// required .CompareFilter compare_filter = 1;
|
||||||
private org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter compareFilter_ = org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter.getDefaultInstance();
|
private org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter compareFilter_ = org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter.getDefaultInstance();
|
||||||
private com.google.protobuf.SingleFieldBuilder<
|
private com.google.protobuf.SingleFieldBuilder<
|
||||||
org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter, org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter.Builder, org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilterOrBuilder> compareFilterBuilder_;
|
org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter, org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilter.Builder, org.apache.hadoop.hbase.protobuf.generated.FilterProtos.CompareFilterOrBuilder> compareFilterBuilder_;
|
||||||
|
@ -13105,51 +13105,52 @@ public final class FilterProtos {
|
||||||
static {
|
static {
|
||||||
java.lang.String[] descriptorData = {
|
java.lang.String[] descriptorData = {
|
||||||
"\n\014Filter.proto\032\013hbase.proto\032\020Comparator." +
|
"\n\014Filter.proto\032\013hbase.proto\032\020Comparator." +
|
||||||
"proto\"0\n\006Filter\022\014\n\004name\030\001 \002(\t\022\030\n\020seriali" +
|
"proto\"1\n\006Filter\022\014\n\004name\030\001 \002(\t\022\031\n\021seriali" +
|
||||||
"zedFilter\030\002 \001(\014\"%\n\024ColumnCountGetFilter\022" +
|
"zed_filter\030\002 \001(\014\"%\n\024ColumnCountGetFilter" +
|
||||||
"\r\n\005limit\030\001 \002(\005\"M\n\026ColumnPaginationFilter" +
|
"\022\r\n\005limit\030\001 \002(\005\"N\n\026ColumnPaginationFilte" +
|
||||||
"\022\r\n\005limit\030\001 \002(\005\022\016\n\006offset\030\002 \001(\005\022\024\n\014colum" +
|
"r\022\r\n\005limit\030\001 \002(\005\022\016\n\006offset\030\002 \001(\005\022\025\n\rcolu" +
|
||||||
"nOffset\030\003 \001(\014\"$\n\022ColumnPrefixFilter\022\016\n\006p" +
|
"mn_offset\030\003 \001(\014\"$\n\022ColumnPrefixFilter\022\016\n" +
|
||||||
"refix\030\001 \002(\014\"q\n\021ColumnRangeFilter\022\021\n\tminC" +
|
"\006prefix\030\001 \002(\014\"w\n\021ColumnRangeFilter\022\022\n\nmi" +
|
||||||
"olumn\030\001 \001(\014\022\032\n\022minColumnInclusive\030\002 \001(\010\022" +
|
"n_column\030\001 \001(\014\022\034\n\024min_column_inclusive\030\002" +
|
||||||
"\021\n\tmaxColumn\030\003 \001(\014\022\032\n\022maxColumnInclusive" +
|
" \001(\010\022\022\n\nmax_column\030\003 \001(\014\022\034\n\024max_column_i" +
|
||||||
"\030\004 \001(\010\"Q\n\rCompareFilter\022\037\n\tcompareOp\030\001 \002",
|
"nclusive\030\004 \001(\010\"R\n\rCompareFilter\022 \n\ncompa",
|
||||||
"(\0162\014.CompareType\022\037\n\ncomparator\030\002 \001(\0132\013.C" +
|
"re_op\030\001 \002(\0162\014.CompareType\022\037\n\ncomparator\030" +
|
||||||
"omparator\"\212\001\n\025DependentColumnFilter\022%\n\rc" +
|
"\002 \001(\0132\013.Comparator\"\217\001\n\025DependentColumnFi" +
|
||||||
"ompareFilter\030\001 \002(\0132\016.CompareFilter\022\024\n\014co" +
|
"lter\022&\n\016compare_filter\030\001 \002(\0132\016.CompareFi" +
|
||||||
"lumnFamily\030\002 \001(\014\022\027\n\017columnQualifier\030\003 \001(" +
|
"lter\022\025\n\rcolumn_family\030\002 \001(\014\022\030\n\020column_qu" +
|
||||||
"\014\022\033\n\023dropDependentColumn\030\004 \001(\010\"5\n\014Family" +
|
"alifier\030\003 \001(\014\022\035\n\025drop_dependent_column\030\004" +
|
||||||
"Filter\022%\n\rcompareFilter\030\001 \002(\0132\016.CompareF" +
|
" \001(\010\"6\n\014FamilyFilter\022&\n\016compare_filter\030\001" +
|
||||||
"ilter\"\200\001\n\nFilterList\022&\n\010operator\030\001 \002(\0162\024" +
|
" \002(\0132\016.CompareFilter\"\200\001\n\nFilterList\022&\n\010o" +
|
||||||
".FilterList.Operator\022\030\n\007filters\030\002 \003(\0132\007." +
|
"perator\030\001 \002(\0162\024.FilterList.Operator\022\030\n\007f" +
|
||||||
"Filter\"0\n\010Operator\022\021\n\rMUST_PASS_ALL\020\001\022\021\n" +
|
"ilters\030\002 \003(\0132\007.Filter\"0\n\010Operator\022\021\n\rMUS" +
|
||||||
"\rMUST_PASS_ONE\020\002\"(\n\rFilterWrapper\022\027\n\006fil",
|
"T_PASS_ALL\020\001\022\021\n\rMUST_PASS_ONE\020\002\"(\n\rFilte",
|
||||||
"ter\030\001 \002(\0132\007.Filter\"\024\n\022FirstKeyOnlyFilter" +
|
"rWrapper\022\027\n\006filter\030\001 \002(\0132\007.Filter\"\024\n\022Fir" +
|
||||||
"\";\n%FirstKeyValueMatchingQualifiersFilte" +
|
"stKeyOnlyFilter\";\n%FirstKeyValueMatching" +
|
||||||
"r\022\022\n\nqualifiers\030\001 \003(\014\"8\n\016FuzzyRowFilter\022" +
|
"QualifiersFilter\022\022\n\nqualifiers\030\001 \003(\014\":\n\016" +
|
||||||
"&\n\rfuzzyKeysData\030\001 \003(\0132\017.BytesBytesPair\"" +
|
"FuzzyRowFilter\022(\n\017fuzzy_keys_data\030\001 \003(\0132" +
|
||||||
")\n\023InclusiveStopFilter\022\022\n\nstopRowKey\030\001 \001" +
|
"\017.BytesBytesPair\"+\n\023InclusiveStopFilter\022" +
|
||||||
"(\014\"!\n\rKeyOnlyFilter\022\020\n\010lenAsVal\030\001 \002(\010\"4\n" +
|
"\024\n\014stop_row_key\030\001 \001(\014\"#\n\rKeyOnlyFilter\022\022" +
|
||||||
"\032MultipleColumnPrefixFilter\022\026\n\016sortedPre" +
|
"\n\nlen_as_val\030\001 \002(\010\"5\n\032MultipleColumnPref" +
|
||||||
"fixes\030\001 \003(\014\"\036\n\nPageFilter\022\020\n\010pageSize\030\001 " +
|
"ixFilter\022\027\n\017sorted_prefixes\030\001 \003(\014\"\037\n\nPag" +
|
||||||
"\002(\003\"\036\n\014PrefixFilter\022\016\n\006prefix\030\001 \001(\014\"8\n\017Q" +
|
"eFilter\022\021\n\tpage_size\030\001 \002(\003\"\036\n\014PrefixFilt" +
|
||||||
"ualifierFilter\022%\n\rcompareFilter\030\001 \002(\0132\016.",
|
"er\022\016\n\006prefix\030\001 \001(\014\"9\n\017QualifierFilter\022&\n",
|
||||||
"CompareFilter\"!\n\017RandomRowFilter\022\016\n\006chan" +
|
"\016compare_filter\030\001 \002(\0132\016.CompareFilter\"!\n" +
|
||||||
"ce\030\001 \002(\002\"2\n\tRowFilter\022%\n\rcompareFilter\030\001" +
|
"\017RandomRowFilter\022\016\n\006chance\030\001 \002(\002\"3\n\tRowF" +
|
||||||
" \002(\0132\016.CompareFilter\"[\n\036SingleColumnValu" +
|
"ilter\022&\n\016compare_filter\030\001 \002(\0132\016.CompareF" +
|
||||||
"eExcludeFilter\0229\n\027singleColumnValueFilte" +
|
"ilter\"^\n\036SingleColumnValueExcludeFilter\022" +
|
||||||
"r\030\001 \002(\0132\030.SingleColumnValueFilter\"\276\001\n\027Si" +
|
"<\n\032single_column_value_filter\030\001 \002(\0132\030.Si" +
|
||||||
"ngleColumnValueFilter\022\024\n\014columnFamily\030\001 " +
|
"ngleColumnValueFilter\"\305\001\n\027SingleColumnVa" +
|
||||||
"\001(\014\022\027\n\017columnQualifier\030\002 \001(\014\022\037\n\tcompareO" +
|
"lueFilter\022\025\n\rcolumn_family\030\001 \001(\014\022\030\n\020colu" +
|
||||||
"p\030\003 \002(\0162\014.CompareType\022\037\n\ncomparator\030\004 \002(" +
|
"mn_qualifier\030\002 \001(\014\022 \n\ncompare_op\030\003 \002(\0162\014" +
|
||||||
"\0132\013.Comparator\022\027\n\017filterIfMissing\030\005 \001(\010\022" +
|
".CompareType\022\037\n\ncomparator\030\004 \002(\0132\013.Compa" +
|
||||||
"\031\n\021latestVersionOnly\030\006 \001(\010\"%\n\nSkipFilter",
|
"rator\022\031\n\021filter_if_missing\030\005 \001(\010\022\033\n\023late",
|
||||||
"\022\027\n\006filter\030\001 \002(\0132\007.Filter\"*\n\020TimestampsF" +
|
"st_version_only\030\006 \001(\010\"%\n\nSkipFilter\022\027\n\006f" +
|
||||||
"ilter\022\026\n\ntimestamps\030\001 \003(\003B\002\020\001\"4\n\013ValueFi" +
|
"ilter\030\001 \002(\0132\007.Filter\"*\n\020TimestampsFilter" +
|
||||||
"lter\022%\n\rcompareFilter\030\001 \002(\0132\016.CompareFil" +
|
"\022\026\n\ntimestamps\030\001 \003(\003B\002\020\001\"5\n\013ValueFilter\022" +
|
||||||
"ter\"+\n\020WhileMatchFilter\022\027\n\006filter\030\001 \002(\0132" +
|
"&\n\016compare_filter\030\001 \002(\0132\016.CompareFilter\"" +
|
||||||
"\007.FilterBB\n*org.apache.hadoop.hbase.prot" +
|
"+\n\020WhileMatchFilter\022\027\n\006filter\030\001 \002(\0132\007.Fi" +
|
||||||
"obuf.generatedB\014FilterProtosH\001\210\001\001\240\001\001"
|
"lterBB\n*org.apache.hadoop.hbase.protobuf" +
|
||||||
|
".generatedB\014FilterProtosH\001\210\001\001\240\001\001"
|
||||||
};
|
};
|
||||||
com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
|
com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
|
||||||
new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() {
|
new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() {
|
||||||
|
|
|
@ -109,7 +109,7 @@ public final class HBaseProtos {
|
||||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPairOrBuilder getAttributesOrBuilder(
|
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPairOrBuilder getAttributesOrBuilder(
|
||||||
int index);
|
int index);
|
||||||
|
|
||||||
// repeated .ColumnFamilySchema columnFamilies = 3;
|
// repeated .ColumnFamilySchema column_families = 3;
|
||||||
java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema>
|
java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema>
|
||||||
getColumnFamiliesList();
|
getColumnFamiliesList();
|
||||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema getColumnFamilies(int index);
|
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema getColumnFamilies(int index);
|
||||||
|
@ -189,8 +189,8 @@ public final class HBaseProtos {
|
||||||
return attributes_.get(index);
|
return attributes_.get(index);
|
||||||
}
|
}
|
||||||
|
|
||||||
// repeated .ColumnFamilySchema columnFamilies = 3;
|
// repeated .ColumnFamilySchema column_families = 3;
|
||||||
public static final int COLUMNFAMILIES_FIELD_NUMBER = 3;
|
public static final int COLUMN_FAMILIES_FIELD_NUMBER = 3;
|
||||||
private java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema> columnFamilies_;
|
private java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema> columnFamilies_;
|
||||||
public java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema> getColumnFamiliesList() {
|
public java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema> getColumnFamiliesList() {
|
||||||
return columnFamilies_;
|
return columnFamilies_;
|
||||||
|
@ -356,7 +356,7 @@ public final class HBaseProtos {
|
||||||
hash = (53 * hash) + getAttributesList().hashCode();
|
hash = (53 * hash) + getAttributesList().hashCode();
|
||||||
}
|
}
|
||||||
if (getColumnFamiliesCount() > 0) {
|
if (getColumnFamiliesCount() > 0) {
|
||||||
hash = (37 * hash) + COLUMNFAMILIES_FIELD_NUMBER;
|
hash = (37 * hash) + COLUMN_FAMILIES_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + getColumnFamiliesList().hashCode();
|
hash = (53 * hash) + getColumnFamiliesList().hashCode();
|
||||||
}
|
}
|
||||||
if (getConfigurationCount() > 0) {
|
if (getConfigurationCount() > 0) {
|
||||||
|
@ -956,7 +956,7 @@ public final class HBaseProtos {
|
||||||
return attributesBuilder_;
|
return attributesBuilder_;
|
||||||
}
|
}
|
||||||
|
|
||||||
// repeated .ColumnFamilySchema columnFamilies = 3;
|
// repeated .ColumnFamilySchema column_families = 3;
|
||||||
private java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema> columnFamilies_ =
|
private java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema> columnFamilies_ =
|
||||||
java.util.Collections.emptyList();
|
java.util.Collections.emptyList();
|
||||||
private void ensureColumnFamiliesIsMutable() {
|
private void ensureColumnFamiliesIsMutable() {
|
||||||
|
@ -2306,19 +2306,19 @@ public final class HBaseProtos {
|
||||||
public interface RegionInfoOrBuilder
|
public interface RegionInfoOrBuilder
|
||||||
extends com.google.protobuf.MessageOrBuilder {
|
extends com.google.protobuf.MessageOrBuilder {
|
||||||
|
|
||||||
// required uint64 regionId = 1;
|
// required uint64 region_id = 1;
|
||||||
boolean hasRegionId();
|
boolean hasRegionId();
|
||||||
long getRegionId();
|
long getRegionId();
|
||||||
|
|
||||||
// required bytes tableName = 2;
|
// required bytes table_name = 2;
|
||||||
boolean hasTableName();
|
boolean hasTableName();
|
||||||
com.google.protobuf.ByteString getTableName();
|
com.google.protobuf.ByteString getTableName();
|
||||||
|
|
||||||
// optional bytes startKey = 3;
|
// optional bytes start_key = 3;
|
||||||
boolean hasStartKey();
|
boolean hasStartKey();
|
||||||
com.google.protobuf.ByteString getStartKey();
|
com.google.protobuf.ByteString getStartKey();
|
||||||
|
|
||||||
// optional bytes endKey = 4;
|
// optional bytes end_key = 4;
|
||||||
boolean hasEndKey();
|
boolean hasEndKey();
|
||||||
com.google.protobuf.ByteString getEndKey();
|
com.google.protobuf.ByteString getEndKey();
|
||||||
|
|
||||||
|
@ -2363,8 +2363,8 @@ public final class HBaseProtos {
|
||||||
}
|
}
|
||||||
|
|
||||||
private int bitField0_;
|
private int bitField0_;
|
||||||
// required uint64 regionId = 1;
|
// required uint64 region_id = 1;
|
||||||
public static final int REGIONID_FIELD_NUMBER = 1;
|
public static final int REGION_ID_FIELD_NUMBER = 1;
|
||||||
private long regionId_;
|
private long regionId_;
|
||||||
public boolean hasRegionId() {
|
public boolean hasRegionId() {
|
||||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||||
|
@ -2373,8 +2373,8 @@ public final class HBaseProtos {
|
||||||
return regionId_;
|
return regionId_;
|
||||||
}
|
}
|
||||||
|
|
||||||
// required bytes tableName = 2;
|
// required bytes table_name = 2;
|
||||||
public static final int TABLENAME_FIELD_NUMBER = 2;
|
public static final int TABLE_NAME_FIELD_NUMBER = 2;
|
||||||
private com.google.protobuf.ByteString tableName_;
|
private com.google.protobuf.ByteString tableName_;
|
||||||
public boolean hasTableName() {
|
public boolean hasTableName() {
|
||||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||||
|
@ -2383,8 +2383,8 @@ public final class HBaseProtos {
|
||||||
return tableName_;
|
return tableName_;
|
||||||
}
|
}
|
||||||
|
|
||||||
// optional bytes startKey = 3;
|
// optional bytes start_key = 3;
|
||||||
public static final int STARTKEY_FIELD_NUMBER = 3;
|
public static final int START_KEY_FIELD_NUMBER = 3;
|
||||||
private com.google.protobuf.ByteString startKey_;
|
private com.google.protobuf.ByteString startKey_;
|
||||||
public boolean hasStartKey() {
|
public boolean hasStartKey() {
|
||||||
return ((bitField0_ & 0x00000004) == 0x00000004);
|
return ((bitField0_ & 0x00000004) == 0x00000004);
|
||||||
|
@ -2393,8 +2393,8 @@ public final class HBaseProtos {
|
||||||
return startKey_;
|
return startKey_;
|
||||||
}
|
}
|
||||||
|
|
||||||
// optional bytes endKey = 4;
|
// optional bytes end_key = 4;
|
||||||
public static final int ENDKEY_FIELD_NUMBER = 4;
|
public static final int END_KEY_FIELD_NUMBER = 4;
|
||||||
private com.google.protobuf.ByteString endKey_;
|
private com.google.protobuf.ByteString endKey_;
|
||||||
public boolean hasEndKey() {
|
public boolean hasEndKey() {
|
||||||
return ((bitField0_ & 0x00000008) == 0x00000008);
|
return ((bitField0_ & 0x00000008) == 0x00000008);
|
||||||
|
@ -2588,19 +2588,19 @@ public final class HBaseProtos {
|
||||||
int hash = 41;
|
int hash = 41;
|
||||||
hash = (19 * hash) + getDescriptorForType().hashCode();
|
hash = (19 * hash) + getDescriptorForType().hashCode();
|
||||||
if (hasRegionId()) {
|
if (hasRegionId()) {
|
||||||
hash = (37 * hash) + REGIONID_FIELD_NUMBER;
|
hash = (37 * hash) + REGION_ID_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + hashLong(getRegionId());
|
hash = (53 * hash) + hashLong(getRegionId());
|
||||||
}
|
}
|
||||||
if (hasTableName()) {
|
if (hasTableName()) {
|
||||||
hash = (37 * hash) + TABLENAME_FIELD_NUMBER;
|
hash = (37 * hash) + TABLE_NAME_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + getTableName().hashCode();
|
hash = (53 * hash) + getTableName().hashCode();
|
||||||
}
|
}
|
||||||
if (hasStartKey()) {
|
if (hasStartKey()) {
|
||||||
hash = (37 * hash) + STARTKEY_FIELD_NUMBER;
|
hash = (37 * hash) + START_KEY_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + getStartKey().hashCode();
|
hash = (53 * hash) + getStartKey().hashCode();
|
||||||
}
|
}
|
||||||
if (hasEndKey()) {
|
if (hasEndKey()) {
|
||||||
hash = (37 * hash) + ENDKEY_FIELD_NUMBER;
|
hash = (37 * hash) + END_KEY_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + getEndKey().hashCode();
|
hash = (53 * hash) + getEndKey().hashCode();
|
||||||
}
|
}
|
||||||
if (hasOffline()) {
|
if (hasOffline()) {
|
||||||
|
@ -2928,7 +2928,7 @@ public final class HBaseProtos {
|
||||||
|
|
||||||
private int bitField0_;
|
private int bitField0_;
|
||||||
|
|
||||||
// required uint64 regionId = 1;
|
// required uint64 region_id = 1;
|
||||||
private long regionId_ ;
|
private long regionId_ ;
|
||||||
public boolean hasRegionId() {
|
public boolean hasRegionId() {
|
||||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||||
|
@ -2949,7 +2949,7 @@ public final class HBaseProtos {
|
||||||
return this;
|
return this;
|
||||||
}
|
}
|
||||||
|
|
||||||
// required bytes tableName = 2;
|
// required bytes table_name = 2;
|
||||||
private com.google.protobuf.ByteString tableName_ = com.google.protobuf.ByteString.EMPTY;
|
private com.google.protobuf.ByteString tableName_ = com.google.protobuf.ByteString.EMPTY;
|
||||||
public boolean hasTableName() {
|
public boolean hasTableName() {
|
||||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||||
|
@ -2973,7 +2973,7 @@ public final class HBaseProtos {
|
||||||
return this;
|
return this;
|
||||||
}
|
}
|
||||||
|
|
||||||
// optional bytes startKey = 3;
|
// optional bytes start_key = 3;
|
||||||
private com.google.protobuf.ByteString startKey_ = com.google.protobuf.ByteString.EMPTY;
|
private com.google.protobuf.ByteString startKey_ = com.google.protobuf.ByteString.EMPTY;
|
||||||
public boolean hasStartKey() {
|
public boolean hasStartKey() {
|
||||||
return ((bitField0_ & 0x00000004) == 0x00000004);
|
return ((bitField0_ & 0x00000004) == 0x00000004);
|
||||||
|
@ -2997,7 +2997,7 @@ public final class HBaseProtos {
|
||||||
return this;
|
return this;
|
||||||
}
|
}
|
||||||
|
|
||||||
// optional bytes endKey = 4;
|
// optional bytes end_key = 4;
|
||||||
private com.google.protobuf.ByteString endKey_ = com.google.protobuf.ByteString.EMPTY;
|
private com.google.protobuf.ByteString endKey_ = com.google.protobuf.ByteString.EMPTY;
|
||||||
public boolean hasEndKey() {
|
public boolean hasEndKey() {
|
||||||
return ((bitField0_ & 0x00000008) == 0x00000008);
|
return ((bitField0_ & 0x00000008) == 0x00000008);
|
||||||
|
@ -3098,7 +3098,7 @@ public final class HBaseProtos {
|
||||||
public interface FavoredNodesOrBuilder
|
public interface FavoredNodesOrBuilder
|
||||||
extends com.google.protobuf.MessageOrBuilder {
|
extends com.google.protobuf.MessageOrBuilder {
|
||||||
|
|
||||||
// repeated .ServerName favoredNode = 1;
|
// repeated .ServerName favored_node = 1;
|
||||||
java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName>
|
java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName>
|
||||||
getFavoredNodeList();
|
getFavoredNodeList();
|
||||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName getFavoredNode(int index);
|
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName getFavoredNode(int index);
|
||||||
|
@ -3136,8 +3136,8 @@ public final class HBaseProtos {
|
||||||
return org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.internal_static_FavoredNodes_fieldAccessorTable;
|
return org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.internal_static_FavoredNodes_fieldAccessorTable;
|
||||||
}
|
}
|
||||||
|
|
||||||
// repeated .ServerName favoredNode = 1;
|
// repeated .ServerName favored_node = 1;
|
||||||
public static final int FAVOREDNODE_FIELD_NUMBER = 1;
|
public static final int FAVORED_NODE_FIELD_NUMBER = 1;
|
||||||
private java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName> favoredNode_;
|
private java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName> favoredNode_;
|
||||||
public java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName> getFavoredNodeList() {
|
public java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName> getFavoredNodeList() {
|
||||||
return favoredNode_;
|
return favoredNode_;
|
||||||
|
@ -3229,7 +3229,7 @@ public final class HBaseProtos {
|
||||||
int hash = 41;
|
int hash = 41;
|
||||||
hash = (19 * hash) + getDescriptorForType().hashCode();
|
hash = (19 * hash) + getDescriptorForType().hashCode();
|
||||||
if (getFavoredNodeCount() > 0) {
|
if (getFavoredNodeCount() > 0) {
|
||||||
hash = (37 * hash) + FAVOREDNODE_FIELD_NUMBER;
|
hash = (37 * hash) + FAVORED_NODE_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + getFavoredNodeList().hashCode();
|
hash = (53 * hash) + getFavoredNodeList().hashCode();
|
||||||
}
|
}
|
||||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||||
|
@ -3491,7 +3491,7 @@ public final class HBaseProtos {
|
||||||
|
|
||||||
private int bitField0_;
|
private int bitField0_;
|
||||||
|
|
||||||
// repeated .ServerName favoredNode = 1;
|
// repeated .ServerName favored_node = 1;
|
||||||
private java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName> favoredNode_ =
|
private java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName> favoredNode_ =
|
||||||
java.util.Collections.emptyList();
|
java.util.Collections.emptyList();
|
||||||
private void ensureFavoredNodeIsMutable() {
|
private void ensureFavoredNodeIsMutable() {
|
||||||
|
@ -4662,7 +4662,7 @@ public final class HBaseProtos {
|
||||||
public interface ServerNameOrBuilder
|
public interface ServerNameOrBuilder
|
||||||
extends com.google.protobuf.MessageOrBuilder {
|
extends com.google.protobuf.MessageOrBuilder {
|
||||||
|
|
||||||
// required string hostName = 1;
|
// required string host_name = 1;
|
||||||
boolean hasHostName();
|
boolean hasHostName();
|
||||||
String getHostName();
|
String getHostName();
|
||||||
|
|
||||||
|
@ -4670,7 +4670,7 @@ public final class HBaseProtos {
|
||||||
boolean hasPort();
|
boolean hasPort();
|
||||||
int getPort();
|
int getPort();
|
||||||
|
|
||||||
// optional uint64 startCode = 3;
|
// optional uint64 start_code = 3;
|
||||||
boolean hasStartCode();
|
boolean hasStartCode();
|
||||||
long getStartCode();
|
long getStartCode();
|
||||||
}
|
}
|
||||||
|
@ -4703,8 +4703,8 @@ public final class HBaseProtos {
|
||||||
}
|
}
|
||||||
|
|
||||||
private int bitField0_;
|
private int bitField0_;
|
||||||
// required string hostName = 1;
|
// required string host_name = 1;
|
||||||
public static final int HOSTNAME_FIELD_NUMBER = 1;
|
public static final int HOST_NAME_FIELD_NUMBER = 1;
|
||||||
private java.lang.Object hostName_;
|
private java.lang.Object hostName_;
|
||||||
public boolean hasHostName() {
|
public boolean hasHostName() {
|
||||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||||
|
@ -4745,8 +4745,8 @@ public final class HBaseProtos {
|
||||||
return port_;
|
return port_;
|
||||||
}
|
}
|
||||||
|
|
||||||
// optional uint64 startCode = 3;
|
// optional uint64 start_code = 3;
|
||||||
public static final int STARTCODE_FIELD_NUMBER = 3;
|
public static final int START_CODE_FIELD_NUMBER = 3;
|
||||||
private long startCode_;
|
private long startCode_;
|
||||||
public boolean hasStartCode() {
|
public boolean hasStartCode() {
|
||||||
return ((bitField0_ & 0x00000004) == 0x00000004);
|
return ((bitField0_ & 0x00000004) == 0x00000004);
|
||||||
|
@ -4854,7 +4854,7 @@ public final class HBaseProtos {
|
||||||
int hash = 41;
|
int hash = 41;
|
||||||
hash = (19 * hash) + getDescriptorForType().hashCode();
|
hash = (19 * hash) + getDescriptorForType().hashCode();
|
||||||
if (hasHostName()) {
|
if (hasHostName()) {
|
||||||
hash = (37 * hash) + HOSTNAME_FIELD_NUMBER;
|
hash = (37 * hash) + HOST_NAME_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + getHostName().hashCode();
|
hash = (53 * hash) + getHostName().hashCode();
|
||||||
}
|
}
|
||||||
if (hasPort()) {
|
if (hasPort()) {
|
||||||
|
@ -4862,7 +4862,7 @@ public final class HBaseProtos {
|
||||||
hash = (53 * hash) + getPort();
|
hash = (53 * hash) + getPort();
|
||||||
}
|
}
|
||||||
if (hasStartCode()) {
|
if (hasStartCode()) {
|
||||||
hash = (37 * hash) + STARTCODE_FIELD_NUMBER;
|
hash = (37 * hash) + START_CODE_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + hashLong(getStartCode());
|
hash = (53 * hash) + hashLong(getStartCode());
|
||||||
}
|
}
|
||||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||||
|
@ -5118,7 +5118,7 @@ public final class HBaseProtos {
|
||||||
|
|
||||||
private int bitField0_;
|
private int bitField0_;
|
||||||
|
|
||||||
// required string hostName = 1;
|
// required string host_name = 1;
|
||||||
private java.lang.Object hostName_ = "";
|
private java.lang.Object hostName_ = "";
|
||||||
public boolean hasHostName() {
|
public boolean hasHostName() {
|
||||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||||
|
@ -5175,7 +5175,7 @@ public final class HBaseProtos {
|
||||||
return this;
|
return this;
|
||||||
}
|
}
|
||||||
|
|
||||||
// optional uint64 startCode = 3;
|
// optional uint64 start_code = 3;
|
||||||
private long startCode_ ;
|
private long startCode_ ;
|
||||||
public boolean hasStartCode() {
|
public boolean hasStartCode() {
|
||||||
return ((bitField0_ & 0x00000004) == 0x00000004);
|
return ((bitField0_ & 0x00000004) == 0x00000004);
|
||||||
|
@ -7579,7 +7579,7 @@ public final class HBaseProtos {
|
||||||
boolean hasTable();
|
boolean hasTable();
|
||||||
String getTable();
|
String getTable();
|
||||||
|
|
||||||
// optional int64 creationTime = 3 [default = 0];
|
// optional int64 creation_time = 3 [default = 0];
|
||||||
boolean hasCreationTime();
|
boolean hasCreationTime();
|
||||||
long getCreationTime();
|
long getCreationTime();
|
||||||
|
|
||||||
|
@ -7753,8 +7753,8 @@ public final class HBaseProtos {
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// optional int64 creationTime = 3 [default = 0];
|
// optional int64 creation_time = 3 [default = 0];
|
||||||
public static final int CREATIONTIME_FIELD_NUMBER = 3;
|
public static final int CREATION_TIME_FIELD_NUMBER = 3;
|
||||||
private long creationTime_;
|
private long creationTime_;
|
||||||
public boolean hasCreationTime() {
|
public boolean hasCreationTime() {
|
||||||
return ((bitField0_ & 0x00000004) == 0x00000004);
|
return ((bitField0_ & 0x00000004) == 0x00000004);
|
||||||
|
@ -7916,7 +7916,7 @@ public final class HBaseProtos {
|
||||||
hash = (53 * hash) + getTable().hashCode();
|
hash = (53 * hash) + getTable().hashCode();
|
||||||
}
|
}
|
||||||
if (hasCreationTime()) {
|
if (hasCreationTime()) {
|
||||||
hash = (37 * hash) + CREATIONTIME_FIELD_NUMBER;
|
hash = (37 * hash) + CREATION_TIME_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + hashLong(getCreationTime());
|
hash = (53 * hash) + hashLong(getCreationTime());
|
||||||
}
|
}
|
||||||
if (hasType()) {
|
if (hasType()) {
|
||||||
|
@ -8286,7 +8286,7 @@ public final class HBaseProtos {
|
||||||
onChanged();
|
onChanged();
|
||||||
}
|
}
|
||||||
|
|
||||||
// optional int64 creationTime = 3 [default = 0];
|
// optional int64 creation_time = 3 [default = 0];
|
||||||
private long creationTime_ ;
|
private long creationTime_ ;
|
||||||
public boolean hasCreationTime() {
|
public boolean hasCreationTime() {
|
||||||
return ((bitField0_ & 0x00000004) == 0x00000004);
|
return ((bitField0_ & 0x00000004) == 0x00000004);
|
||||||
|
@ -8666,7 +8666,7 @@ public final class HBaseProtos {
|
||||||
public interface LongMsgOrBuilder
|
public interface LongMsgOrBuilder
|
||||||
extends com.google.protobuf.MessageOrBuilder {
|
extends com.google.protobuf.MessageOrBuilder {
|
||||||
|
|
||||||
// required int64 longMsg = 1;
|
// required int64 long_msg = 1;
|
||||||
boolean hasLongMsg();
|
boolean hasLongMsg();
|
||||||
long getLongMsg();
|
long getLongMsg();
|
||||||
}
|
}
|
||||||
|
@ -8699,8 +8699,8 @@ public final class HBaseProtos {
|
||||||
}
|
}
|
||||||
|
|
||||||
private int bitField0_;
|
private int bitField0_;
|
||||||
// required int64 longMsg = 1;
|
// required int64 long_msg = 1;
|
||||||
public static final int LONGMSG_FIELD_NUMBER = 1;
|
public static final int LONG_MSG_FIELD_NUMBER = 1;
|
||||||
private long longMsg_;
|
private long longMsg_;
|
||||||
public boolean hasLongMsg() {
|
public boolean hasLongMsg() {
|
||||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||||
|
@ -8782,7 +8782,7 @@ public final class HBaseProtos {
|
||||||
int hash = 41;
|
int hash = 41;
|
||||||
hash = (19 * hash) + getDescriptorForType().hashCode();
|
hash = (19 * hash) + getDescriptorForType().hashCode();
|
||||||
if (hasLongMsg()) {
|
if (hasLongMsg()) {
|
||||||
hash = (37 * hash) + LONGMSG_FIELD_NUMBER;
|
hash = (37 * hash) + LONG_MSG_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + hashLong(getLongMsg());
|
hash = (53 * hash) + hashLong(getLongMsg());
|
||||||
}
|
}
|
||||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||||
|
@ -9010,7 +9010,7 @@ public final class HBaseProtos {
|
||||||
|
|
||||||
private int bitField0_;
|
private int bitField0_;
|
||||||
|
|
||||||
// required int64 longMsg = 1;
|
// required int64 long_msg = 1;
|
||||||
private long longMsg_ ;
|
private long longMsg_ ;
|
||||||
public boolean hasLongMsg() {
|
public boolean hasLongMsg() {
|
||||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||||
|
@ -9045,7 +9045,7 @@ public final class HBaseProtos {
|
||||||
public interface BigDecimalMsgOrBuilder
|
public interface BigDecimalMsgOrBuilder
|
||||||
extends com.google.protobuf.MessageOrBuilder {
|
extends com.google.protobuf.MessageOrBuilder {
|
||||||
|
|
||||||
// required bytes bigdecimalMsg = 1;
|
// required bytes bigdecimal_msg = 1;
|
||||||
boolean hasBigdecimalMsg();
|
boolean hasBigdecimalMsg();
|
||||||
com.google.protobuf.ByteString getBigdecimalMsg();
|
com.google.protobuf.ByteString getBigdecimalMsg();
|
||||||
}
|
}
|
||||||
|
@ -9078,8 +9078,8 @@ public final class HBaseProtos {
|
||||||
}
|
}
|
||||||
|
|
||||||
private int bitField0_;
|
private int bitField0_;
|
||||||
// required bytes bigdecimalMsg = 1;
|
// required bytes bigdecimal_msg = 1;
|
||||||
public static final int BIGDECIMALMSG_FIELD_NUMBER = 1;
|
public static final int BIGDECIMAL_MSG_FIELD_NUMBER = 1;
|
||||||
private com.google.protobuf.ByteString bigdecimalMsg_;
|
private com.google.protobuf.ByteString bigdecimalMsg_;
|
||||||
public boolean hasBigdecimalMsg() {
|
public boolean hasBigdecimalMsg() {
|
||||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||||
|
@ -9161,7 +9161,7 @@ public final class HBaseProtos {
|
||||||
int hash = 41;
|
int hash = 41;
|
||||||
hash = (19 * hash) + getDescriptorForType().hashCode();
|
hash = (19 * hash) + getDescriptorForType().hashCode();
|
||||||
if (hasBigdecimalMsg()) {
|
if (hasBigdecimalMsg()) {
|
||||||
hash = (37 * hash) + BIGDECIMALMSG_FIELD_NUMBER;
|
hash = (37 * hash) + BIGDECIMAL_MSG_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + getBigdecimalMsg().hashCode();
|
hash = (53 * hash) + getBigdecimalMsg().hashCode();
|
||||||
}
|
}
|
||||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||||
|
@ -9389,7 +9389,7 @@ public final class HBaseProtos {
|
||||||
|
|
||||||
private int bitField0_;
|
private int bitField0_;
|
||||||
|
|
||||||
// required bytes bigdecimalMsg = 1;
|
// required bytes bigdecimal_msg = 1;
|
||||||
private com.google.protobuf.ByteString bigdecimalMsg_ = com.google.protobuf.ByteString.EMPTY;
|
private com.google.protobuf.ByteString bigdecimalMsg_ = com.google.protobuf.ByteString.EMPTY;
|
||||||
public boolean hasBigdecimalMsg() {
|
public boolean hasBigdecimalMsg() {
|
||||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||||
|
@ -9427,11 +9427,11 @@ public final class HBaseProtos {
|
||||||
public interface UUIDOrBuilder
|
public interface UUIDOrBuilder
|
||||||
extends com.google.protobuf.MessageOrBuilder {
|
extends com.google.protobuf.MessageOrBuilder {
|
||||||
|
|
||||||
// required uint64 leastSigBits = 1;
|
// required uint64 least_sig_bits = 1;
|
||||||
boolean hasLeastSigBits();
|
boolean hasLeastSigBits();
|
||||||
long getLeastSigBits();
|
long getLeastSigBits();
|
||||||
|
|
||||||
// required uint64 mostSigBits = 2;
|
// required uint64 most_sig_bits = 2;
|
||||||
boolean hasMostSigBits();
|
boolean hasMostSigBits();
|
||||||
long getMostSigBits();
|
long getMostSigBits();
|
||||||
}
|
}
|
||||||
|
@ -9464,8 +9464,8 @@ public final class HBaseProtos {
|
||||||
}
|
}
|
||||||
|
|
||||||
private int bitField0_;
|
private int bitField0_;
|
||||||
// required uint64 leastSigBits = 1;
|
// required uint64 least_sig_bits = 1;
|
||||||
public static final int LEASTSIGBITS_FIELD_NUMBER = 1;
|
public static final int LEAST_SIG_BITS_FIELD_NUMBER = 1;
|
||||||
private long leastSigBits_;
|
private long leastSigBits_;
|
||||||
public boolean hasLeastSigBits() {
|
public boolean hasLeastSigBits() {
|
||||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||||
|
@ -9474,8 +9474,8 @@ public final class HBaseProtos {
|
||||||
return leastSigBits_;
|
return leastSigBits_;
|
||||||
}
|
}
|
||||||
|
|
||||||
// required uint64 mostSigBits = 2;
|
// required uint64 most_sig_bits = 2;
|
||||||
public static final int MOSTSIGBITS_FIELD_NUMBER = 2;
|
public static final int MOST_SIG_BITS_FIELD_NUMBER = 2;
|
||||||
private long mostSigBits_;
|
private long mostSigBits_;
|
||||||
public boolean hasMostSigBits() {
|
public boolean hasMostSigBits() {
|
||||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||||
|
@ -9574,11 +9574,11 @@ public final class HBaseProtos {
|
||||||
int hash = 41;
|
int hash = 41;
|
||||||
hash = (19 * hash) + getDescriptorForType().hashCode();
|
hash = (19 * hash) + getDescriptorForType().hashCode();
|
||||||
if (hasLeastSigBits()) {
|
if (hasLeastSigBits()) {
|
||||||
hash = (37 * hash) + LEASTSIGBITS_FIELD_NUMBER;
|
hash = (37 * hash) + LEAST_SIG_BITS_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + hashLong(getLeastSigBits());
|
hash = (53 * hash) + hashLong(getLeastSigBits());
|
||||||
}
|
}
|
||||||
if (hasMostSigBits()) {
|
if (hasMostSigBits()) {
|
||||||
hash = (37 * hash) + MOSTSIGBITS_FIELD_NUMBER;
|
hash = (37 * hash) + MOST_SIG_BITS_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + hashLong(getMostSigBits());
|
hash = (53 * hash) + hashLong(getMostSigBits());
|
||||||
}
|
}
|
||||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||||
|
@ -9824,7 +9824,7 @@ public final class HBaseProtos {
|
||||||
|
|
||||||
private int bitField0_;
|
private int bitField0_;
|
||||||
|
|
||||||
// required uint64 leastSigBits = 1;
|
// required uint64 least_sig_bits = 1;
|
||||||
private long leastSigBits_ ;
|
private long leastSigBits_ ;
|
||||||
public boolean hasLeastSigBits() {
|
public boolean hasLeastSigBits() {
|
||||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||||
|
@ -9845,7 +9845,7 @@ public final class HBaseProtos {
|
||||||
return this;
|
return this;
|
||||||
}
|
}
|
||||||
|
|
||||||
// required uint64 mostSigBits = 2;
|
// required uint64 most_sig_bits = 2;
|
||||||
private long mostSigBits_ ;
|
private long mostSigBits_ ;
|
||||||
public boolean hasMostSigBits() {
|
public boolean hasMostSigBits() {
|
||||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||||
|
@ -9971,42 +9971,43 @@ public final class HBaseProtos {
|
||||||
descriptor;
|
descriptor;
|
||||||
static {
|
static {
|
||||||
java.lang.String[] descriptorData = {
|
java.lang.String[] descriptorData = {
|
||||||
"\n\013hbase.proto\032\nCell.proto\"\225\001\n\013TableSchem" +
|
"\n\013hbase.proto\032\nCell.proto\"\226\001\n\013TableSchem" +
|
||||||
"a\022\014\n\004name\030\001 \001(\014\022#\n\nattributes\030\002 \003(\0132\017.By" +
|
"a\022\014\n\004name\030\001 \001(\014\022#\n\nattributes\030\002 \003(\0132\017.By" +
|
||||||
"tesBytesPair\022+\n\016columnFamilies\030\003 \003(\0132\023.C" +
|
"tesBytesPair\022,\n\017column_families\030\003 \003(\0132\023." +
|
||||||
"olumnFamilySchema\022&\n\rconfiguration\030\004 \003(\013" +
|
"ColumnFamilySchema\022&\n\rconfiguration\030\004 \003(" +
|
||||||
"2\017.NameStringPair\"o\n\022ColumnFamilySchema\022" +
|
"\0132\017.NameStringPair\"o\n\022ColumnFamilySchema" +
|
||||||
"\014\n\004name\030\001 \002(\014\022#\n\nattributes\030\002 \003(\0132\017.Byte" +
|
"\022\014\n\004name\030\001 \002(\014\022#\n\nattributes\030\002 \003(\0132\017.Byt" +
|
||||||
"sBytesPair\022&\n\rconfiguration\030\003 \003(\0132\017.Name" +
|
"esBytesPair\022&\n\rconfiguration\030\003 \003(\0132\017.Nam" +
|
||||||
"StringPair\"\207\001\n\nRegionInfo\022\020\n\010regionId\030\001 " +
|
"eStringPair\"\213\001\n\nRegionInfo\022\021\n\tregion_id\030" +
|
||||||
"\002(\004\022\021\n\ttableName\030\002 \002(\014\022\020\n\010startKey\030\003 \001(\014" +
|
"\001 \002(\004\022\022\n\ntable_name\030\002 \002(\014\022\021\n\tstart_key\030\003" +
|
||||||
"\022\016\n\006endKey\030\004 \001(\014\022\017\n\007offline\030\005 \001(\010\022\r\n\005spl",
|
" \001(\014\022\017\n\007end_key\030\004 \001(\014\022\017\n\007offline\030\005 \001(\010\022\r",
|
||||||
"it\030\006 \001(\010\022\022\n\nrecovering\030\007 \001(\010\"0\n\014FavoredN" +
|
"\n\005split\030\006 \001(\010\022\022\n\nrecovering\030\007 \001(\010\"1\n\014Fav" +
|
||||||
"odes\022 \n\013favoredNode\030\001 \003(\0132\013.ServerName\"\225" +
|
"oredNodes\022!\n\014favored_node\030\001 \003(\0132\013.Server" +
|
||||||
"\001\n\017RegionSpecifier\0222\n\004type\030\001 \002(\0162$.Regio" +
|
"Name\"\225\001\n\017RegionSpecifier\0222\n\004type\030\001 \002(\0162$" +
|
||||||
"nSpecifier.RegionSpecifierType\022\r\n\005value\030" +
|
".RegionSpecifier.RegionSpecifierType\022\r\n\005" +
|
||||||
"\002 \002(\014\"?\n\023RegionSpecifierType\022\017\n\013REGION_N" +
|
"value\030\002 \002(\014\"?\n\023RegionSpecifierType\022\017\n\013RE" +
|
||||||
"AME\020\001\022\027\n\023ENCODED_REGION_NAME\020\002\"%\n\tTimeRa" +
|
"GION_NAME\020\001\022\027\n\023ENCODED_REGION_NAME\020\002\"%\n\t" +
|
||||||
"nge\022\014\n\004from\030\001 \001(\004\022\n\n\002to\030\002 \001(\004\"?\n\nServerN" +
|
"TimeRange\022\014\n\004from\030\001 \001(\004\022\n\n\002to\030\002 \001(\004\"A\n\nS" +
|
||||||
"ame\022\020\n\010hostName\030\001 \002(\t\022\014\n\004port\030\002 \001(\r\022\021\n\ts" +
|
"erverName\022\021\n\thost_name\030\001 \002(\t\022\014\n\004port\030\002 \001" +
|
||||||
"tartCode\030\003 \001(\004\"\033\n\013Coprocessor\022\014\n\004name\030\001 " +
|
"(\r\022\022\n\nstart_code\030\003 \001(\004\"\033\n\013Coprocessor\022\014\n" +
|
||||||
"\002(\t\"-\n\016NameStringPair\022\014\n\004name\030\001 \002(\t\022\r\n\005v",
|
"\004name\030\001 \002(\t\"-\n\016NameStringPair\022\014\n\004name\030\001 ",
|
||||||
"alue\030\002 \002(\t\",\n\rNameBytesPair\022\014\n\004name\030\001 \002(" +
|
"\002(\t\022\r\n\005value\030\002 \002(\t\",\n\rNameBytesPair\022\014\n\004n" +
|
||||||
"\t\022\r\n\005value\030\002 \001(\014\"/\n\016BytesBytesPair\022\r\n\005fi" +
|
"ame\030\001 \002(\t\022\r\n\005value\030\002 \001(\014\"/\n\016BytesBytesPa" +
|
||||||
"rst\030\001 \002(\014\022\016\n\006second\030\002 \002(\014\",\n\rNameInt64Pa" +
|
"ir\022\r\n\005first\030\001 \002(\014\022\016\n\006second\030\002 \002(\014\",\n\rNam" +
|
||||||
"ir\022\014\n\004name\030\001 \001(\t\022\r\n\005value\030\002 \001(\003\"\255\001\n\023Snap" +
|
"eInt64Pair\022\014\n\004name\030\001 \001(\t\022\r\n\005value\030\002 \001(\003\"" +
|
||||||
"shotDescription\022\014\n\004name\030\001 \002(\t\022\r\n\005table\030\002" +
|
"\256\001\n\023SnapshotDescription\022\014\n\004name\030\001 \002(\t\022\r\n" +
|
||||||
" \001(\t\022\027\n\014creationTime\030\003 \001(\003:\0010\022.\n\004type\030\004 " +
|
"\005table\030\002 \001(\t\022\030\n\rcreation_time\030\003 \001(\003:\0010\022." +
|
||||||
"\001(\0162\031.SnapshotDescription.Type:\005FLUSH\022\017\n" +
|
"\n\004type\030\004 \001(\0162\031.SnapshotDescription.Type:" +
|
||||||
"\007version\030\005 \001(\005\"\037\n\004Type\022\014\n\010DISABLED\020\000\022\t\n\005" +
|
"\005FLUSH\022\017\n\007version\030\005 \001(\005\"\037\n\004Type\022\014\n\010DISAB" +
|
||||||
"FLUSH\020\001\"\n\n\010EmptyMsg\"\032\n\007LongMsg\022\017\n\007longMs" +
|
"LED\020\000\022\t\n\005FLUSH\020\001\"\n\n\010EmptyMsg\"\033\n\007LongMsg\022" +
|
||||||
"g\030\001 \002(\003\"&\n\rBigDecimalMsg\022\025\n\rbigdecimalMs",
|
"\020\n\010long_msg\030\001 \002(\003\"\'\n\rBigDecimalMsg\022\026\n\016bi",
|
||||||
"g\030\001 \002(\014\"1\n\004UUID\022\024\n\014leastSigBits\030\001 \002(\004\022\023\n" +
|
"gdecimal_msg\030\001 \002(\014\"5\n\004UUID\022\026\n\016least_sig_" +
|
||||||
"\013mostSigBits\030\002 \002(\004*r\n\013CompareType\022\010\n\004LES" +
|
"bits\030\001 \002(\004\022\025\n\rmost_sig_bits\030\002 \002(\004*r\n\013Com" +
|
||||||
"S\020\000\022\021\n\rLESS_OR_EQUAL\020\001\022\t\n\005EQUAL\020\002\022\r\n\tNOT" +
|
"pareType\022\010\n\004LESS\020\000\022\021\n\rLESS_OR_EQUAL\020\001\022\t\n" +
|
||||||
"_EQUAL\020\003\022\024\n\020GREATER_OR_EQUAL\020\004\022\013\n\007GREATE" +
|
"\005EQUAL\020\002\022\r\n\tNOT_EQUAL\020\003\022\024\n\020GREATER_OR_EQ" +
|
||||||
"R\020\005\022\t\n\005NO_OP\020\006B>\n*org.apache.hadoop.hbas" +
|
"UAL\020\004\022\013\n\007GREATER\020\005\022\t\n\005NO_OP\020\006B>\n*org.apa" +
|
||||||
"e.protobuf.generatedB\013HBaseProtosH\001\240\001\001"
|
"che.hadoop.hbase.protobuf.generatedB\013HBa" +
|
||||||
|
"seProtosH\001\240\001\001"
|
||||||
};
|
};
|
||||||
com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
|
com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
|
||||||
new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() {
|
new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() {
|
||||||
|
|
|
@ -11,7 +11,7 @@ public final class HFileProtos {
|
||||||
public interface FileInfoProtoOrBuilder
|
public interface FileInfoProtoOrBuilder
|
||||||
extends com.google.protobuf.MessageOrBuilder {
|
extends com.google.protobuf.MessageOrBuilder {
|
||||||
|
|
||||||
// repeated .BytesBytesPair mapEntry = 1;
|
// repeated .BytesBytesPair map_entry = 1;
|
||||||
java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair>
|
java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair>
|
||||||
getMapEntryList();
|
getMapEntryList();
|
||||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair getMapEntry(int index);
|
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair getMapEntry(int index);
|
||||||
|
@ -49,8 +49,8 @@ public final class HFileProtos {
|
||||||
return org.apache.hadoop.hbase.protobuf.generated.HFileProtos.internal_static_FileInfoProto_fieldAccessorTable;
|
return org.apache.hadoop.hbase.protobuf.generated.HFileProtos.internal_static_FileInfoProto_fieldAccessorTable;
|
||||||
}
|
}
|
||||||
|
|
||||||
// repeated .BytesBytesPair mapEntry = 1;
|
// repeated .BytesBytesPair map_entry = 1;
|
||||||
public static final int MAPENTRY_FIELD_NUMBER = 1;
|
public static final int MAP_ENTRY_FIELD_NUMBER = 1;
|
||||||
private java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair> mapEntry_;
|
private java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair> mapEntry_;
|
||||||
public java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair> getMapEntryList() {
|
public java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair> getMapEntryList() {
|
||||||
return mapEntry_;
|
return mapEntry_;
|
||||||
|
@ -142,7 +142,7 @@ public final class HFileProtos {
|
||||||
int hash = 41;
|
int hash = 41;
|
||||||
hash = (19 * hash) + getDescriptorForType().hashCode();
|
hash = (19 * hash) + getDescriptorForType().hashCode();
|
||||||
if (getMapEntryCount() > 0) {
|
if (getMapEntryCount() > 0) {
|
||||||
hash = (37 * hash) + MAPENTRY_FIELD_NUMBER;
|
hash = (37 * hash) + MAP_ENTRY_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + getMapEntryList().hashCode();
|
hash = (53 * hash) + getMapEntryList().hashCode();
|
||||||
}
|
}
|
||||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||||
|
@ -404,7 +404,7 @@ public final class HFileProtos {
|
||||||
|
|
||||||
private int bitField0_;
|
private int bitField0_;
|
||||||
|
|
||||||
// repeated .BytesBytesPair mapEntry = 1;
|
// repeated .BytesBytesPair map_entry = 1;
|
||||||
private java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair> mapEntry_ =
|
private java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair> mapEntry_ =
|
||||||
java.util.Collections.emptyList();
|
java.util.Collections.emptyList();
|
||||||
private void ensureMapEntryIsMutable() {
|
private void ensureMapEntryIsMutable() {
|
||||||
|
@ -604,51 +604,51 @@ public final class HFileProtos {
|
||||||
public interface FileTrailerProtoOrBuilder
|
public interface FileTrailerProtoOrBuilder
|
||||||
extends com.google.protobuf.MessageOrBuilder {
|
extends com.google.protobuf.MessageOrBuilder {
|
||||||
|
|
||||||
// optional uint64 fileInfoOffset = 1;
|
// optional uint64 file_info_offset = 1;
|
||||||
boolean hasFileInfoOffset();
|
boolean hasFileInfoOffset();
|
||||||
long getFileInfoOffset();
|
long getFileInfoOffset();
|
||||||
|
|
||||||
// optional uint64 loadOnOpenDataOffset = 2;
|
// optional uint64 load_on_open_data_offset = 2;
|
||||||
boolean hasLoadOnOpenDataOffset();
|
boolean hasLoadOnOpenDataOffset();
|
||||||
long getLoadOnOpenDataOffset();
|
long getLoadOnOpenDataOffset();
|
||||||
|
|
||||||
// optional uint64 uncompressedDataIndexSize = 3;
|
// optional uint64 uncompressed_data_index_size = 3;
|
||||||
boolean hasUncompressedDataIndexSize();
|
boolean hasUncompressedDataIndexSize();
|
||||||
long getUncompressedDataIndexSize();
|
long getUncompressedDataIndexSize();
|
||||||
|
|
||||||
// optional uint64 totalUncompressedBytes = 4;
|
// optional uint64 total_uncompressed_bytes = 4;
|
||||||
boolean hasTotalUncompressedBytes();
|
boolean hasTotalUncompressedBytes();
|
||||||
long getTotalUncompressedBytes();
|
long getTotalUncompressedBytes();
|
||||||
|
|
||||||
// optional uint32 dataIndexCount = 5;
|
// optional uint32 data_index_count = 5;
|
||||||
boolean hasDataIndexCount();
|
boolean hasDataIndexCount();
|
||||||
int getDataIndexCount();
|
int getDataIndexCount();
|
||||||
|
|
||||||
// optional uint32 metaIndexCount = 6;
|
// optional uint32 meta_index_count = 6;
|
||||||
boolean hasMetaIndexCount();
|
boolean hasMetaIndexCount();
|
||||||
int getMetaIndexCount();
|
int getMetaIndexCount();
|
||||||
|
|
||||||
// optional uint64 entryCount = 7;
|
// optional uint64 entry_count = 7;
|
||||||
boolean hasEntryCount();
|
boolean hasEntryCount();
|
||||||
long getEntryCount();
|
long getEntryCount();
|
||||||
|
|
||||||
// optional uint32 numDataIndexLevels = 8;
|
// optional uint32 num_data_index_levels = 8;
|
||||||
boolean hasNumDataIndexLevels();
|
boolean hasNumDataIndexLevels();
|
||||||
int getNumDataIndexLevels();
|
int getNumDataIndexLevels();
|
||||||
|
|
||||||
// optional uint64 firstDataBlockOffset = 9;
|
// optional uint64 first_data_block_offset = 9;
|
||||||
boolean hasFirstDataBlockOffset();
|
boolean hasFirstDataBlockOffset();
|
||||||
long getFirstDataBlockOffset();
|
long getFirstDataBlockOffset();
|
||||||
|
|
||||||
// optional uint64 lastDataBlockOffset = 10;
|
// optional uint64 last_data_block_offset = 10;
|
||||||
boolean hasLastDataBlockOffset();
|
boolean hasLastDataBlockOffset();
|
||||||
long getLastDataBlockOffset();
|
long getLastDataBlockOffset();
|
||||||
|
|
||||||
// optional string comparatorClassName = 11;
|
// optional string comparator_class_name = 11;
|
||||||
boolean hasComparatorClassName();
|
boolean hasComparatorClassName();
|
||||||
String getComparatorClassName();
|
String getComparatorClassName();
|
||||||
|
|
||||||
// optional uint32 compressionCodec = 12;
|
// optional uint32 compression_codec = 12;
|
||||||
boolean hasCompressionCodec();
|
boolean hasCompressionCodec();
|
||||||
int getCompressionCodec();
|
int getCompressionCodec();
|
||||||
}
|
}
|
||||||
|
@ -681,8 +681,8 @@ public final class HFileProtos {
|
||||||
}
|
}
|
||||||
|
|
||||||
private int bitField0_;
|
private int bitField0_;
|
||||||
// optional uint64 fileInfoOffset = 1;
|
// optional uint64 file_info_offset = 1;
|
||||||
public static final int FILEINFOOFFSET_FIELD_NUMBER = 1;
|
public static final int FILE_INFO_OFFSET_FIELD_NUMBER = 1;
|
||||||
private long fileInfoOffset_;
|
private long fileInfoOffset_;
|
||||||
public boolean hasFileInfoOffset() {
|
public boolean hasFileInfoOffset() {
|
||||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||||
|
@ -691,8 +691,8 @@ public final class HFileProtos {
|
||||||
return fileInfoOffset_;
|
return fileInfoOffset_;
|
||||||
}
|
}
|
||||||
|
|
||||||
// optional uint64 loadOnOpenDataOffset = 2;
|
// optional uint64 load_on_open_data_offset = 2;
|
||||||
public static final int LOADONOPENDATAOFFSET_FIELD_NUMBER = 2;
|
public static final int LOAD_ON_OPEN_DATA_OFFSET_FIELD_NUMBER = 2;
|
||||||
private long loadOnOpenDataOffset_;
|
private long loadOnOpenDataOffset_;
|
||||||
public boolean hasLoadOnOpenDataOffset() {
|
public boolean hasLoadOnOpenDataOffset() {
|
||||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||||
|
@ -701,8 +701,8 @@ public final class HFileProtos {
|
||||||
return loadOnOpenDataOffset_;
|
return loadOnOpenDataOffset_;
|
||||||
}
|
}
|
||||||
|
|
||||||
// optional uint64 uncompressedDataIndexSize = 3;
|
// optional uint64 uncompressed_data_index_size = 3;
|
||||||
public static final int UNCOMPRESSEDDATAINDEXSIZE_FIELD_NUMBER = 3;
|
public static final int UNCOMPRESSED_DATA_INDEX_SIZE_FIELD_NUMBER = 3;
|
||||||
private long uncompressedDataIndexSize_;
|
private long uncompressedDataIndexSize_;
|
||||||
public boolean hasUncompressedDataIndexSize() {
|
public boolean hasUncompressedDataIndexSize() {
|
||||||
return ((bitField0_ & 0x00000004) == 0x00000004);
|
return ((bitField0_ & 0x00000004) == 0x00000004);
|
||||||
|
@ -711,8 +711,8 @@ public final class HFileProtos {
|
||||||
return uncompressedDataIndexSize_;
|
return uncompressedDataIndexSize_;
|
||||||
}
|
}
|
||||||
|
|
||||||
// optional uint64 totalUncompressedBytes = 4;
|
// optional uint64 total_uncompressed_bytes = 4;
|
||||||
public static final int TOTALUNCOMPRESSEDBYTES_FIELD_NUMBER = 4;
|
public static final int TOTAL_UNCOMPRESSED_BYTES_FIELD_NUMBER = 4;
|
||||||
private long totalUncompressedBytes_;
|
private long totalUncompressedBytes_;
|
||||||
public boolean hasTotalUncompressedBytes() {
|
public boolean hasTotalUncompressedBytes() {
|
||||||
return ((bitField0_ & 0x00000008) == 0x00000008);
|
return ((bitField0_ & 0x00000008) == 0x00000008);
|
||||||
|
@ -721,8 +721,8 @@ public final class HFileProtos {
|
||||||
return totalUncompressedBytes_;
|
return totalUncompressedBytes_;
|
||||||
}
|
}
|
||||||
|
|
||||||
// optional uint32 dataIndexCount = 5;
|
// optional uint32 data_index_count = 5;
|
||||||
public static final int DATAINDEXCOUNT_FIELD_NUMBER = 5;
|
public static final int DATA_INDEX_COUNT_FIELD_NUMBER = 5;
|
||||||
private int dataIndexCount_;
|
private int dataIndexCount_;
|
||||||
public boolean hasDataIndexCount() {
|
public boolean hasDataIndexCount() {
|
||||||
return ((bitField0_ & 0x00000010) == 0x00000010);
|
return ((bitField0_ & 0x00000010) == 0x00000010);
|
||||||
|
@ -731,8 +731,8 @@ public final class HFileProtos {
|
||||||
return dataIndexCount_;
|
return dataIndexCount_;
|
||||||
}
|
}
|
||||||
|
|
||||||
// optional uint32 metaIndexCount = 6;
|
// optional uint32 meta_index_count = 6;
|
||||||
public static final int METAINDEXCOUNT_FIELD_NUMBER = 6;
|
public static final int META_INDEX_COUNT_FIELD_NUMBER = 6;
|
||||||
private int metaIndexCount_;
|
private int metaIndexCount_;
|
||||||
public boolean hasMetaIndexCount() {
|
public boolean hasMetaIndexCount() {
|
||||||
return ((bitField0_ & 0x00000020) == 0x00000020);
|
return ((bitField0_ & 0x00000020) == 0x00000020);
|
||||||
|
@ -741,8 +741,8 @@ public final class HFileProtos {
|
||||||
return metaIndexCount_;
|
return metaIndexCount_;
|
||||||
}
|
}
|
||||||
|
|
||||||
// optional uint64 entryCount = 7;
|
// optional uint64 entry_count = 7;
|
||||||
public static final int ENTRYCOUNT_FIELD_NUMBER = 7;
|
public static final int ENTRY_COUNT_FIELD_NUMBER = 7;
|
||||||
private long entryCount_;
|
private long entryCount_;
|
||||||
public boolean hasEntryCount() {
|
public boolean hasEntryCount() {
|
||||||
return ((bitField0_ & 0x00000040) == 0x00000040);
|
return ((bitField0_ & 0x00000040) == 0x00000040);
|
||||||
|
@ -751,8 +751,8 @@ public final class HFileProtos {
|
||||||
return entryCount_;
|
return entryCount_;
|
||||||
}
|
}
|
||||||
|
|
||||||
// optional uint32 numDataIndexLevels = 8;
|
// optional uint32 num_data_index_levels = 8;
|
||||||
public static final int NUMDATAINDEXLEVELS_FIELD_NUMBER = 8;
|
public static final int NUM_DATA_INDEX_LEVELS_FIELD_NUMBER = 8;
|
||||||
private int numDataIndexLevels_;
|
private int numDataIndexLevels_;
|
||||||
public boolean hasNumDataIndexLevels() {
|
public boolean hasNumDataIndexLevels() {
|
||||||
return ((bitField0_ & 0x00000080) == 0x00000080);
|
return ((bitField0_ & 0x00000080) == 0x00000080);
|
||||||
|
@ -761,8 +761,8 @@ public final class HFileProtos {
|
||||||
return numDataIndexLevels_;
|
return numDataIndexLevels_;
|
||||||
}
|
}
|
||||||
|
|
||||||
// optional uint64 firstDataBlockOffset = 9;
|
// optional uint64 first_data_block_offset = 9;
|
||||||
public static final int FIRSTDATABLOCKOFFSET_FIELD_NUMBER = 9;
|
public static final int FIRST_DATA_BLOCK_OFFSET_FIELD_NUMBER = 9;
|
||||||
private long firstDataBlockOffset_;
|
private long firstDataBlockOffset_;
|
||||||
public boolean hasFirstDataBlockOffset() {
|
public boolean hasFirstDataBlockOffset() {
|
||||||
return ((bitField0_ & 0x00000100) == 0x00000100);
|
return ((bitField0_ & 0x00000100) == 0x00000100);
|
||||||
|
@ -771,8 +771,8 @@ public final class HFileProtos {
|
||||||
return firstDataBlockOffset_;
|
return firstDataBlockOffset_;
|
||||||
}
|
}
|
||||||
|
|
||||||
// optional uint64 lastDataBlockOffset = 10;
|
// optional uint64 last_data_block_offset = 10;
|
||||||
public static final int LASTDATABLOCKOFFSET_FIELD_NUMBER = 10;
|
public static final int LAST_DATA_BLOCK_OFFSET_FIELD_NUMBER = 10;
|
||||||
private long lastDataBlockOffset_;
|
private long lastDataBlockOffset_;
|
||||||
public boolean hasLastDataBlockOffset() {
|
public boolean hasLastDataBlockOffset() {
|
||||||
return ((bitField0_ & 0x00000200) == 0x00000200);
|
return ((bitField0_ & 0x00000200) == 0x00000200);
|
||||||
|
@ -781,8 +781,8 @@ public final class HFileProtos {
|
||||||
return lastDataBlockOffset_;
|
return lastDataBlockOffset_;
|
||||||
}
|
}
|
||||||
|
|
||||||
// optional string comparatorClassName = 11;
|
// optional string comparator_class_name = 11;
|
||||||
public static final int COMPARATORCLASSNAME_FIELD_NUMBER = 11;
|
public static final int COMPARATOR_CLASS_NAME_FIELD_NUMBER = 11;
|
||||||
private java.lang.Object comparatorClassName_;
|
private java.lang.Object comparatorClassName_;
|
||||||
public boolean hasComparatorClassName() {
|
public boolean hasComparatorClassName() {
|
||||||
return ((bitField0_ & 0x00000400) == 0x00000400);
|
return ((bitField0_ & 0x00000400) == 0x00000400);
|
||||||
|
@ -813,8 +813,8 @@ public final class HFileProtos {
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// optional uint32 compressionCodec = 12;
|
// optional uint32 compression_codec = 12;
|
||||||
public static final int COMPRESSIONCODEC_FIELD_NUMBER = 12;
|
public static final int COMPRESSION_CODEC_FIELD_NUMBER = 12;
|
||||||
private int compressionCodec_;
|
private int compressionCodec_;
|
||||||
public boolean hasCompressionCodec() {
|
public boolean hasCompressionCodec() {
|
||||||
return ((bitField0_ & 0x00000800) == 0x00000800);
|
return ((bitField0_ & 0x00000800) == 0x00000800);
|
||||||
|
@ -1035,51 +1035,51 @@ public final class HFileProtos {
|
||||||
int hash = 41;
|
int hash = 41;
|
||||||
hash = (19 * hash) + getDescriptorForType().hashCode();
|
hash = (19 * hash) + getDescriptorForType().hashCode();
|
||||||
if (hasFileInfoOffset()) {
|
if (hasFileInfoOffset()) {
|
||||||
hash = (37 * hash) + FILEINFOOFFSET_FIELD_NUMBER;
|
hash = (37 * hash) + FILE_INFO_OFFSET_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + hashLong(getFileInfoOffset());
|
hash = (53 * hash) + hashLong(getFileInfoOffset());
|
||||||
}
|
}
|
||||||
if (hasLoadOnOpenDataOffset()) {
|
if (hasLoadOnOpenDataOffset()) {
|
||||||
hash = (37 * hash) + LOADONOPENDATAOFFSET_FIELD_NUMBER;
|
hash = (37 * hash) + LOAD_ON_OPEN_DATA_OFFSET_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + hashLong(getLoadOnOpenDataOffset());
|
hash = (53 * hash) + hashLong(getLoadOnOpenDataOffset());
|
||||||
}
|
}
|
||||||
if (hasUncompressedDataIndexSize()) {
|
if (hasUncompressedDataIndexSize()) {
|
||||||
hash = (37 * hash) + UNCOMPRESSEDDATAINDEXSIZE_FIELD_NUMBER;
|
hash = (37 * hash) + UNCOMPRESSED_DATA_INDEX_SIZE_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + hashLong(getUncompressedDataIndexSize());
|
hash = (53 * hash) + hashLong(getUncompressedDataIndexSize());
|
||||||
}
|
}
|
||||||
if (hasTotalUncompressedBytes()) {
|
if (hasTotalUncompressedBytes()) {
|
||||||
hash = (37 * hash) + TOTALUNCOMPRESSEDBYTES_FIELD_NUMBER;
|
hash = (37 * hash) + TOTAL_UNCOMPRESSED_BYTES_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + hashLong(getTotalUncompressedBytes());
|
hash = (53 * hash) + hashLong(getTotalUncompressedBytes());
|
||||||
}
|
}
|
||||||
if (hasDataIndexCount()) {
|
if (hasDataIndexCount()) {
|
||||||
hash = (37 * hash) + DATAINDEXCOUNT_FIELD_NUMBER;
|
hash = (37 * hash) + DATA_INDEX_COUNT_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + getDataIndexCount();
|
hash = (53 * hash) + getDataIndexCount();
|
||||||
}
|
}
|
||||||
if (hasMetaIndexCount()) {
|
if (hasMetaIndexCount()) {
|
||||||
hash = (37 * hash) + METAINDEXCOUNT_FIELD_NUMBER;
|
hash = (37 * hash) + META_INDEX_COUNT_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + getMetaIndexCount();
|
hash = (53 * hash) + getMetaIndexCount();
|
||||||
}
|
}
|
||||||
if (hasEntryCount()) {
|
if (hasEntryCount()) {
|
||||||
hash = (37 * hash) + ENTRYCOUNT_FIELD_NUMBER;
|
hash = (37 * hash) + ENTRY_COUNT_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + hashLong(getEntryCount());
|
hash = (53 * hash) + hashLong(getEntryCount());
|
||||||
}
|
}
|
||||||
if (hasNumDataIndexLevels()) {
|
if (hasNumDataIndexLevels()) {
|
||||||
hash = (37 * hash) + NUMDATAINDEXLEVELS_FIELD_NUMBER;
|
hash = (37 * hash) + NUM_DATA_INDEX_LEVELS_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + getNumDataIndexLevels();
|
hash = (53 * hash) + getNumDataIndexLevels();
|
||||||
}
|
}
|
||||||
if (hasFirstDataBlockOffset()) {
|
if (hasFirstDataBlockOffset()) {
|
||||||
hash = (37 * hash) + FIRSTDATABLOCKOFFSET_FIELD_NUMBER;
|
hash = (37 * hash) + FIRST_DATA_BLOCK_OFFSET_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + hashLong(getFirstDataBlockOffset());
|
hash = (53 * hash) + hashLong(getFirstDataBlockOffset());
|
||||||
}
|
}
|
||||||
if (hasLastDataBlockOffset()) {
|
if (hasLastDataBlockOffset()) {
|
||||||
hash = (37 * hash) + LASTDATABLOCKOFFSET_FIELD_NUMBER;
|
hash = (37 * hash) + LAST_DATA_BLOCK_OFFSET_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + hashLong(getLastDataBlockOffset());
|
hash = (53 * hash) + hashLong(getLastDataBlockOffset());
|
||||||
}
|
}
|
||||||
if (hasComparatorClassName()) {
|
if (hasComparatorClassName()) {
|
||||||
hash = (37 * hash) + COMPARATORCLASSNAME_FIELD_NUMBER;
|
hash = (37 * hash) + COMPARATOR_CLASS_NAME_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + getComparatorClassName().hashCode();
|
hash = (53 * hash) + getComparatorClassName().hashCode();
|
||||||
}
|
}
|
||||||
if (hasCompressionCodec()) {
|
if (hasCompressionCodec()) {
|
||||||
hash = (37 * hash) + COMPRESSIONCODEC_FIELD_NUMBER;
|
hash = (37 * hash) + COMPRESSION_CODEC_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + getCompressionCodec();
|
hash = (53 * hash) + getCompressionCodec();
|
||||||
}
|
}
|
||||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||||
|
@ -1457,7 +1457,7 @@ public final class HFileProtos {
|
||||||
|
|
||||||
private int bitField0_;
|
private int bitField0_;
|
||||||
|
|
||||||
// optional uint64 fileInfoOffset = 1;
|
// optional uint64 file_info_offset = 1;
|
||||||
private long fileInfoOffset_ ;
|
private long fileInfoOffset_ ;
|
||||||
public boolean hasFileInfoOffset() {
|
public boolean hasFileInfoOffset() {
|
||||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||||
|
@ -1478,7 +1478,7 @@ public final class HFileProtos {
|
||||||
return this;
|
return this;
|
||||||
}
|
}
|
||||||
|
|
||||||
// optional uint64 loadOnOpenDataOffset = 2;
|
// optional uint64 load_on_open_data_offset = 2;
|
||||||
private long loadOnOpenDataOffset_ ;
|
private long loadOnOpenDataOffset_ ;
|
||||||
public boolean hasLoadOnOpenDataOffset() {
|
public boolean hasLoadOnOpenDataOffset() {
|
||||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||||
|
@ -1499,7 +1499,7 @@ public final class HFileProtos {
|
||||||
return this;
|
return this;
|
||||||
}
|
}
|
||||||
|
|
||||||
// optional uint64 uncompressedDataIndexSize = 3;
|
// optional uint64 uncompressed_data_index_size = 3;
|
||||||
private long uncompressedDataIndexSize_ ;
|
private long uncompressedDataIndexSize_ ;
|
||||||
public boolean hasUncompressedDataIndexSize() {
|
public boolean hasUncompressedDataIndexSize() {
|
||||||
return ((bitField0_ & 0x00000004) == 0x00000004);
|
return ((bitField0_ & 0x00000004) == 0x00000004);
|
||||||
|
@ -1520,7 +1520,7 @@ public final class HFileProtos {
|
||||||
return this;
|
return this;
|
||||||
}
|
}
|
||||||
|
|
||||||
// optional uint64 totalUncompressedBytes = 4;
|
// optional uint64 total_uncompressed_bytes = 4;
|
||||||
private long totalUncompressedBytes_ ;
|
private long totalUncompressedBytes_ ;
|
||||||
public boolean hasTotalUncompressedBytes() {
|
public boolean hasTotalUncompressedBytes() {
|
||||||
return ((bitField0_ & 0x00000008) == 0x00000008);
|
return ((bitField0_ & 0x00000008) == 0x00000008);
|
||||||
|
@ -1541,7 +1541,7 @@ public final class HFileProtos {
|
||||||
return this;
|
return this;
|
||||||
}
|
}
|
||||||
|
|
||||||
// optional uint32 dataIndexCount = 5;
|
// optional uint32 data_index_count = 5;
|
||||||
private int dataIndexCount_ ;
|
private int dataIndexCount_ ;
|
||||||
public boolean hasDataIndexCount() {
|
public boolean hasDataIndexCount() {
|
||||||
return ((bitField0_ & 0x00000010) == 0x00000010);
|
return ((bitField0_ & 0x00000010) == 0x00000010);
|
||||||
|
@ -1562,7 +1562,7 @@ public final class HFileProtos {
|
||||||
return this;
|
return this;
|
||||||
}
|
}
|
||||||
|
|
||||||
// optional uint32 metaIndexCount = 6;
|
// optional uint32 meta_index_count = 6;
|
||||||
private int metaIndexCount_ ;
|
private int metaIndexCount_ ;
|
||||||
public boolean hasMetaIndexCount() {
|
public boolean hasMetaIndexCount() {
|
||||||
return ((bitField0_ & 0x00000020) == 0x00000020);
|
return ((bitField0_ & 0x00000020) == 0x00000020);
|
||||||
|
@ -1583,7 +1583,7 @@ public final class HFileProtos {
|
||||||
return this;
|
return this;
|
||||||
}
|
}
|
||||||
|
|
||||||
// optional uint64 entryCount = 7;
|
// optional uint64 entry_count = 7;
|
||||||
private long entryCount_ ;
|
private long entryCount_ ;
|
||||||
public boolean hasEntryCount() {
|
public boolean hasEntryCount() {
|
||||||
return ((bitField0_ & 0x00000040) == 0x00000040);
|
return ((bitField0_ & 0x00000040) == 0x00000040);
|
||||||
|
@ -1604,7 +1604,7 @@ public final class HFileProtos {
|
||||||
return this;
|
return this;
|
||||||
}
|
}
|
||||||
|
|
||||||
// optional uint32 numDataIndexLevels = 8;
|
// optional uint32 num_data_index_levels = 8;
|
||||||
private int numDataIndexLevels_ ;
|
private int numDataIndexLevels_ ;
|
||||||
public boolean hasNumDataIndexLevels() {
|
public boolean hasNumDataIndexLevels() {
|
||||||
return ((bitField0_ & 0x00000080) == 0x00000080);
|
return ((bitField0_ & 0x00000080) == 0x00000080);
|
||||||
|
@ -1625,7 +1625,7 @@ public final class HFileProtos {
|
||||||
return this;
|
return this;
|
||||||
}
|
}
|
||||||
|
|
||||||
// optional uint64 firstDataBlockOffset = 9;
|
// optional uint64 first_data_block_offset = 9;
|
||||||
private long firstDataBlockOffset_ ;
|
private long firstDataBlockOffset_ ;
|
||||||
public boolean hasFirstDataBlockOffset() {
|
public boolean hasFirstDataBlockOffset() {
|
||||||
return ((bitField0_ & 0x00000100) == 0x00000100);
|
return ((bitField0_ & 0x00000100) == 0x00000100);
|
||||||
|
@ -1646,7 +1646,7 @@ public final class HFileProtos {
|
||||||
return this;
|
return this;
|
||||||
}
|
}
|
||||||
|
|
||||||
// optional uint64 lastDataBlockOffset = 10;
|
// optional uint64 last_data_block_offset = 10;
|
||||||
private long lastDataBlockOffset_ ;
|
private long lastDataBlockOffset_ ;
|
||||||
public boolean hasLastDataBlockOffset() {
|
public boolean hasLastDataBlockOffset() {
|
||||||
return ((bitField0_ & 0x00000200) == 0x00000200);
|
return ((bitField0_ & 0x00000200) == 0x00000200);
|
||||||
|
@ -1667,7 +1667,7 @@ public final class HFileProtos {
|
||||||
return this;
|
return this;
|
||||||
}
|
}
|
||||||
|
|
||||||
// optional string comparatorClassName = 11;
|
// optional string comparator_class_name = 11;
|
||||||
private java.lang.Object comparatorClassName_ = "";
|
private java.lang.Object comparatorClassName_ = "";
|
||||||
public boolean hasComparatorClassName() {
|
public boolean hasComparatorClassName() {
|
||||||
return ((bitField0_ & 0x00000400) == 0x00000400);
|
return ((bitField0_ & 0x00000400) == 0x00000400);
|
||||||
|
@ -1703,7 +1703,7 @@ public final class HFileProtos {
|
||||||
onChanged();
|
onChanged();
|
||||||
}
|
}
|
||||||
|
|
||||||
// optional uint32 compressionCodec = 12;
|
// optional uint32 compression_codec = 12;
|
||||||
private int compressionCodec_ ;
|
private int compressionCodec_ ;
|
||||||
public boolean hasCompressionCodec() {
|
public boolean hasCompressionCodec() {
|
||||||
return ((bitField0_ & 0x00000800) == 0x00000800);
|
return ((bitField0_ & 0x00000800) == 0x00000800);
|
||||||
|
@ -1754,19 +1754,20 @@ public final class HFileProtos {
|
||||||
descriptor;
|
descriptor;
|
||||||
static {
|
static {
|
||||||
java.lang.String[] descriptorData = {
|
java.lang.String[] descriptorData = {
|
||||||
"\n\013HFile.proto\032\013hbase.proto\"2\n\rFileInfoPr" +
|
"\n\013HFile.proto\032\013hbase.proto\"3\n\rFileInfoPr" +
|
||||||
"oto\022!\n\010mapEntry\030\001 \003(\0132\017.BytesBytesPair\"\335" +
|
"oto\022\"\n\tmap_entry\030\001 \003(\0132\017.BytesBytesPair\"" +
|
||||||
"\002\n\020FileTrailerProto\022\026\n\016fileInfoOffset\030\001 " +
|
"\371\002\n\020FileTrailerProto\022\030\n\020file_info_offset" +
|
||||||
"\001(\004\022\034\n\024loadOnOpenDataOffset\030\002 \001(\004\022!\n\031unc" +
|
"\030\001 \001(\004\022 \n\030load_on_open_data_offset\030\002 \001(\004" +
|
||||||
"ompressedDataIndexSize\030\003 \001(\004\022\036\n\026totalUnc" +
|
"\022$\n\034uncompressed_data_index_size\030\003 \001(\004\022 " +
|
||||||
"ompressedBytes\030\004 \001(\004\022\026\n\016dataIndexCount\030\005" +
|
"\n\030total_uncompressed_bytes\030\004 \001(\004\022\030\n\020data" +
|
||||||
" \001(\r\022\026\n\016metaIndexCount\030\006 \001(\r\022\022\n\nentryCou" +
|
"_index_count\030\005 \001(\r\022\030\n\020meta_index_count\030\006" +
|
||||||
"nt\030\007 \001(\004\022\032\n\022numDataIndexLevels\030\010 \001(\r\022\034\n\024" +
|
" \001(\r\022\023\n\013entry_count\030\007 \001(\004\022\035\n\025num_data_in" +
|
||||||
"firstDataBlockOffset\030\t \001(\004\022\033\n\023lastDataBl" +
|
"dex_levels\030\010 \001(\r\022\037\n\027first_data_block_off" +
|
||||||
"ockOffset\030\n \001(\004\022\033\n\023comparatorClassName\030\013",
|
"set\030\t \001(\004\022\036\n\026last_data_block_offset\030\n \001(",
|
||||||
" \001(\t\022\030\n\020compressionCodec\030\014 \001(\rBA\n*org.ap" +
|
"\004\022\035\n\025comparator_class_name\030\013 \001(\t\022\031\n\021comp" +
|
||||||
"ache.hadoop.hbase.protobuf.generatedB\013HF" +
|
"ression_codec\030\014 \001(\rBA\n*org.apache.hadoop" +
|
||||||
"ileProtosH\001\210\001\001\240\001\001"
|
".hbase.protobuf.generatedB\013HFileProtosH\001" +
|
||||||
|
"\210\001\001\240\001\001"
|
||||||
};
|
};
|
||||||
com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
|
com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
|
||||||
new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() {
|
new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() {
|
||||||
|
|
|
@ -11,7 +11,7 @@ public final class LoadBalancerProtos {
|
||||||
public interface LoadBalancerStateOrBuilder
|
public interface LoadBalancerStateOrBuilder
|
||||||
extends com.google.protobuf.MessageOrBuilder {
|
extends com.google.protobuf.MessageOrBuilder {
|
||||||
|
|
||||||
// optional bool balancerOn = 1;
|
// optional bool balancer_on = 1;
|
||||||
boolean hasBalancerOn();
|
boolean hasBalancerOn();
|
||||||
boolean getBalancerOn();
|
boolean getBalancerOn();
|
||||||
}
|
}
|
||||||
|
@ -44,8 +44,8 @@ public final class LoadBalancerProtos {
|
||||||
}
|
}
|
||||||
|
|
||||||
private int bitField0_;
|
private int bitField0_;
|
||||||
// optional bool balancerOn = 1;
|
// optional bool balancer_on = 1;
|
||||||
public static final int BALANCERON_FIELD_NUMBER = 1;
|
public static final int BALANCER_ON_FIELD_NUMBER = 1;
|
||||||
private boolean balancerOn_;
|
private boolean balancerOn_;
|
||||||
public boolean hasBalancerOn() {
|
public boolean hasBalancerOn() {
|
||||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||||
|
@ -123,7 +123,7 @@ public final class LoadBalancerProtos {
|
||||||
int hash = 41;
|
int hash = 41;
|
||||||
hash = (19 * hash) + getDescriptorForType().hashCode();
|
hash = (19 * hash) + getDescriptorForType().hashCode();
|
||||||
if (hasBalancerOn()) {
|
if (hasBalancerOn()) {
|
||||||
hash = (37 * hash) + BALANCERON_FIELD_NUMBER;
|
hash = (37 * hash) + BALANCER_ON_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + hashBoolean(getBalancerOn());
|
hash = (53 * hash) + hashBoolean(getBalancerOn());
|
||||||
}
|
}
|
||||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||||
|
@ -347,7 +347,7 @@ public final class LoadBalancerProtos {
|
||||||
|
|
||||||
private int bitField0_;
|
private int bitField0_;
|
||||||
|
|
||||||
// optional bool balancerOn = 1;
|
// optional bool balancer_on = 1;
|
||||||
private boolean balancerOn_ ;
|
private boolean balancerOn_ ;
|
||||||
public boolean hasBalancerOn() {
|
public boolean hasBalancerOn() {
|
||||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||||
|
@ -393,10 +393,10 @@ public final class LoadBalancerProtos {
|
||||||
descriptor;
|
descriptor;
|
||||||
static {
|
static {
|
||||||
java.lang.String[] descriptorData = {
|
java.lang.String[] descriptorData = {
|
||||||
"\n\022LoadBalancer.proto\"\'\n\021LoadBalancerStat" +
|
"\n\022LoadBalancer.proto\"(\n\021LoadBalancerStat" +
|
||||||
"e\022\022\n\nbalancerOn\030\001 \001(\010BE\n*org.apache.hado" +
|
"e\022\023\n\013balancer_on\030\001 \001(\010BE\n*org.apache.had" +
|
||||||
"op.hbase.protobuf.generatedB\022LoadBalance" +
|
"oop.hbase.protobuf.generatedB\022LoadBalanc" +
|
||||||
"rProtosH\001\240\001\001"
|
"erProtosH\001\240\001\001"
|
||||||
};
|
};
|
||||||
com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
|
com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
|
||||||
new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() {
|
new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() {
|
||||||
|
|
|
@ -11,11 +11,11 @@ public final class MasterAdminProtos {
|
||||||
public interface AddColumnRequestOrBuilder
|
public interface AddColumnRequestOrBuilder
|
||||||
extends com.google.protobuf.MessageOrBuilder {
|
extends com.google.protobuf.MessageOrBuilder {
|
||||||
|
|
||||||
// required bytes tableName = 1;
|
// required bytes table_name = 1;
|
||||||
boolean hasTableName();
|
boolean hasTableName();
|
||||||
com.google.protobuf.ByteString getTableName();
|
com.google.protobuf.ByteString getTableName();
|
||||||
|
|
||||||
// required .ColumnFamilySchema columnFamilies = 2;
|
// required .ColumnFamilySchema column_families = 2;
|
||||||
boolean hasColumnFamilies();
|
boolean hasColumnFamilies();
|
||||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema getColumnFamilies();
|
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema getColumnFamilies();
|
||||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchemaOrBuilder getColumnFamiliesOrBuilder();
|
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchemaOrBuilder getColumnFamiliesOrBuilder();
|
||||||
|
@ -49,8 +49,8 @@ public final class MasterAdminProtos {
|
||||||
}
|
}
|
||||||
|
|
||||||
private int bitField0_;
|
private int bitField0_;
|
||||||
// required bytes tableName = 1;
|
// required bytes table_name = 1;
|
||||||
public static final int TABLENAME_FIELD_NUMBER = 1;
|
public static final int TABLE_NAME_FIELD_NUMBER = 1;
|
||||||
private com.google.protobuf.ByteString tableName_;
|
private com.google.protobuf.ByteString tableName_;
|
||||||
public boolean hasTableName() {
|
public boolean hasTableName() {
|
||||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||||
|
@ -59,8 +59,8 @@ public final class MasterAdminProtos {
|
||||||
return tableName_;
|
return tableName_;
|
||||||
}
|
}
|
||||||
|
|
||||||
// required .ColumnFamilySchema columnFamilies = 2;
|
// required .ColumnFamilySchema column_families = 2;
|
||||||
public static final int COLUMNFAMILIES_FIELD_NUMBER = 2;
|
public static final int COLUMN_FAMILIES_FIELD_NUMBER = 2;
|
||||||
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema columnFamilies_;
|
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema columnFamilies_;
|
||||||
public boolean hasColumnFamilies() {
|
public boolean hasColumnFamilies() {
|
||||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||||
|
@ -166,11 +166,11 @@ public final class MasterAdminProtos {
|
||||||
int hash = 41;
|
int hash = 41;
|
||||||
hash = (19 * hash) + getDescriptorForType().hashCode();
|
hash = (19 * hash) + getDescriptorForType().hashCode();
|
||||||
if (hasTableName()) {
|
if (hasTableName()) {
|
||||||
hash = (37 * hash) + TABLENAME_FIELD_NUMBER;
|
hash = (37 * hash) + TABLE_NAME_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + getTableName().hashCode();
|
hash = (53 * hash) + getTableName().hashCode();
|
||||||
}
|
}
|
||||||
if (hasColumnFamilies()) {
|
if (hasColumnFamilies()) {
|
||||||
hash = (37 * hash) + COLUMNFAMILIES_FIELD_NUMBER;
|
hash = (37 * hash) + COLUMN_FAMILIES_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + getColumnFamilies().hashCode();
|
hash = (53 * hash) + getColumnFamilies().hashCode();
|
||||||
}
|
}
|
||||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||||
|
@ -433,7 +433,7 @@ public final class MasterAdminProtos {
|
||||||
|
|
||||||
private int bitField0_;
|
private int bitField0_;
|
||||||
|
|
||||||
// required bytes tableName = 1;
|
// required bytes table_name = 1;
|
||||||
private com.google.protobuf.ByteString tableName_ = com.google.protobuf.ByteString.EMPTY;
|
private com.google.protobuf.ByteString tableName_ = com.google.protobuf.ByteString.EMPTY;
|
||||||
public boolean hasTableName() {
|
public boolean hasTableName() {
|
||||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||||
|
@ -457,7 +457,7 @@ public final class MasterAdminProtos {
|
||||||
return this;
|
return this;
|
||||||
}
|
}
|
||||||
|
|
||||||
// required .ColumnFamilySchema columnFamilies = 2;
|
// required .ColumnFamilySchema column_families = 2;
|
||||||
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema columnFamilies_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema.getDefaultInstance();
|
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema columnFamilies_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema.getDefaultInstance();
|
||||||
private com.google.protobuf.SingleFieldBuilder<
|
private com.google.protobuf.SingleFieldBuilder<
|
||||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchemaOrBuilder> columnFamiliesBuilder_;
|
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchemaOrBuilder> columnFamiliesBuilder_;
|
||||||
|
@ -861,11 +861,11 @@ public final class MasterAdminProtos {
|
||||||
public interface DeleteColumnRequestOrBuilder
|
public interface DeleteColumnRequestOrBuilder
|
||||||
extends com.google.protobuf.MessageOrBuilder {
|
extends com.google.protobuf.MessageOrBuilder {
|
||||||
|
|
||||||
// required bytes tableName = 1;
|
// required bytes table_name = 1;
|
||||||
boolean hasTableName();
|
boolean hasTableName();
|
||||||
com.google.protobuf.ByteString getTableName();
|
com.google.protobuf.ByteString getTableName();
|
||||||
|
|
||||||
// required bytes columnName = 2;
|
// required bytes column_name = 2;
|
||||||
boolean hasColumnName();
|
boolean hasColumnName();
|
||||||
com.google.protobuf.ByteString getColumnName();
|
com.google.protobuf.ByteString getColumnName();
|
||||||
}
|
}
|
||||||
|
@ -898,8 +898,8 @@ public final class MasterAdminProtos {
|
||||||
}
|
}
|
||||||
|
|
||||||
private int bitField0_;
|
private int bitField0_;
|
||||||
// required bytes tableName = 1;
|
// required bytes table_name = 1;
|
||||||
public static final int TABLENAME_FIELD_NUMBER = 1;
|
public static final int TABLE_NAME_FIELD_NUMBER = 1;
|
||||||
private com.google.protobuf.ByteString tableName_;
|
private com.google.protobuf.ByteString tableName_;
|
||||||
public boolean hasTableName() {
|
public boolean hasTableName() {
|
||||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||||
|
@ -908,8 +908,8 @@ public final class MasterAdminProtos {
|
||||||
return tableName_;
|
return tableName_;
|
||||||
}
|
}
|
||||||
|
|
||||||
// required bytes columnName = 2;
|
// required bytes column_name = 2;
|
||||||
public static final int COLUMNNAME_FIELD_NUMBER = 2;
|
public static final int COLUMN_NAME_FIELD_NUMBER = 2;
|
||||||
private com.google.protobuf.ByteString columnName_;
|
private com.google.protobuf.ByteString columnName_;
|
||||||
public boolean hasColumnName() {
|
public boolean hasColumnName() {
|
||||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||||
|
@ -1008,11 +1008,11 @@ public final class MasterAdminProtos {
|
||||||
int hash = 41;
|
int hash = 41;
|
||||||
hash = (19 * hash) + getDescriptorForType().hashCode();
|
hash = (19 * hash) + getDescriptorForType().hashCode();
|
||||||
if (hasTableName()) {
|
if (hasTableName()) {
|
||||||
hash = (37 * hash) + TABLENAME_FIELD_NUMBER;
|
hash = (37 * hash) + TABLE_NAME_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + getTableName().hashCode();
|
hash = (53 * hash) + getTableName().hashCode();
|
||||||
}
|
}
|
||||||
if (hasColumnName()) {
|
if (hasColumnName()) {
|
||||||
hash = (37 * hash) + COLUMNNAME_FIELD_NUMBER;
|
hash = (37 * hash) + COLUMN_NAME_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + getColumnName().hashCode();
|
hash = (53 * hash) + getColumnName().hashCode();
|
||||||
}
|
}
|
||||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||||
|
@ -1258,7 +1258,7 @@ public final class MasterAdminProtos {
|
||||||
|
|
||||||
private int bitField0_;
|
private int bitField0_;
|
||||||
|
|
||||||
// required bytes tableName = 1;
|
// required bytes table_name = 1;
|
||||||
private com.google.protobuf.ByteString tableName_ = com.google.protobuf.ByteString.EMPTY;
|
private com.google.protobuf.ByteString tableName_ = com.google.protobuf.ByteString.EMPTY;
|
||||||
public boolean hasTableName() {
|
public boolean hasTableName() {
|
||||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||||
|
@ -1282,7 +1282,7 @@ public final class MasterAdminProtos {
|
||||||
return this;
|
return this;
|
||||||
}
|
}
|
||||||
|
|
||||||
// required bytes columnName = 2;
|
// required bytes column_name = 2;
|
||||||
private com.google.protobuf.ByteString columnName_ = com.google.protobuf.ByteString.EMPTY;
|
private com.google.protobuf.ByteString columnName_ = com.google.protobuf.ByteString.EMPTY;
|
||||||
public boolean hasColumnName() {
|
public boolean hasColumnName() {
|
||||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||||
|
@ -1620,11 +1620,11 @@ public final class MasterAdminProtos {
|
||||||
public interface ModifyColumnRequestOrBuilder
|
public interface ModifyColumnRequestOrBuilder
|
||||||
extends com.google.protobuf.MessageOrBuilder {
|
extends com.google.protobuf.MessageOrBuilder {
|
||||||
|
|
||||||
// required bytes tableName = 1;
|
// required bytes table_name = 1;
|
||||||
boolean hasTableName();
|
boolean hasTableName();
|
||||||
com.google.protobuf.ByteString getTableName();
|
com.google.protobuf.ByteString getTableName();
|
||||||
|
|
||||||
// required .ColumnFamilySchema columnFamilies = 2;
|
// required .ColumnFamilySchema column_families = 2;
|
||||||
boolean hasColumnFamilies();
|
boolean hasColumnFamilies();
|
||||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema getColumnFamilies();
|
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema getColumnFamilies();
|
||||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchemaOrBuilder getColumnFamiliesOrBuilder();
|
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchemaOrBuilder getColumnFamiliesOrBuilder();
|
||||||
|
@ -1658,8 +1658,8 @@ public final class MasterAdminProtos {
|
||||||
}
|
}
|
||||||
|
|
||||||
private int bitField0_;
|
private int bitField0_;
|
||||||
// required bytes tableName = 1;
|
// required bytes table_name = 1;
|
||||||
public static final int TABLENAME_FIELD_NUMBER = 1;
|
public static final int TABLE_NAME_FIELD_NUMBER = 1;
|
||||||
private com.google.protobuf.ByteString tableName_;
|
private com.google.protobuf.ByteString tableName_;
|
||||||
public boolean hasTableName() {
|
public boolean hasTableName() {
|
||||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||||
|
@ -1668,8 +1668,8 @@ public final class MasterAdminProtos {
|
||||||
return tableName_;
|
return tableName_;
|
||||||
}
|
}
|
||||||
|
|
||||||
// required .ColumnFamilySchema columnFamilies = 2;
|
// required .ColumnFamilySchema column_families = 2;
|
||||||
public static final int COLUMNFAMILIES_FIELD_NUMBER = 2;
|
public static final int COLUMN_FAMILIES_FIELD_NUMBER = 2;
|
||||||
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema columnFamilies_;
|
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema columnFamilies_;
|
||||||
public boolean hasColumnFamilies() {
|
public boolean hasColumnFamilies() {
|
||||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||||
|
@ -1775,11 +1775,11 @@ public final class MasterAdminProtos {
|
||||||
int hash = 41;
|
int hash = 41;
|
||||||
hash = (19 * hash) + getDescriptorForType().hashCode();
|
hash = (19 * hash) + getDescriptorForType().hashCode();
|
||||||
if (hasTableName()) {
|
if (hasTableName()) {
|
||||||
hash = (37 * hash) + TABLENAME_FIELD_NUMBER;
|
hash = (37 * hash) + TABLE_NAME_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + getTableName().hashCode();
|
hash = (53 * hash) + getTableName().hashCode();
|
||||||
}
|
}
|
||||||
if (hasColumnFamilies()) {
|
if (hasColumnFamilies()) {
|
||||||
hash = (37 * hash) + COLUMNFAMILIES_FIELD_NUMBER;
|
hash = (37 * hash) + COLUMN_FAMILIES_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + getColumnFamilies().hashCode();
|
hash = (53 * hash) + getColumnFamilies().hashCode();
|
||||||
}
|
}
|
||||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||||
|
@ -2042,7 +2042,7 @@ public final class MasterAdminProtos {
|
||||||
|
|
||||||
private int bitField0_;
|
private int bitField0_;
|
||||||
|
|
||||||
// required bytes tableName = 1;
|
// required bytes table_name = 1;
|
||||||
private com.google.protobuf.ByteString tableName_ = com.google.protobuf.ByteString.EMPTY;
|
private com.google.protobuf.ByteString tableName_ = com.google.protobuf.ByteString.EMPTY;
|
||||||
public boolean hasTableName() {
|
public boolean hasTableName() {
|
||||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||||
|
@ -2066,7 +2066,7 @@ public final class MasterAdminProtos {
|
||||||
return this;
|
return this;
|
||||||
}
|
}
|
||||||
|
|
||||||
// required .ColumnFamilySchema columnFamilies = 2;
|
// required .ColumnFamilySchema column_families = 2;
|
||||||
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema columnFamilies_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema.getDefaultInstance();
|
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema columnFamilies_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema.getDefaultInstance();
|
||||||
private com.google.protobuf.SingleFieldBuilder<
|
private com.google.protobuf.SingleFieldBuilder<
|
||||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchemaOrBuilder> columnFamiliesBuilder_;
|
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchemaOrBuilder> columnFamiliesBuilder_;
|
||||||
|
@ -2475,7 +2475,7 @@ public final class MasterAdminProtos {
|
||||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier getRegion();
|
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier getRegion();
|
||||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder getRegionOrBuilder();
|
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder getRegionOrBuilder();
|
||||||
|
|
||||||
// optional .ServerName destServerName = 2;
|
// optional .ServerName dest_server_name = 2;
|
||||||
boolean hasDestServerName();
|
boolean hasDestServerName();
|
||||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName getDestServerName();
|
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName getDestServerName();
|
||||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerNameOrBuilder getDestServerNameOrBuilder();
|
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerNameOrBuilder getDestServerNameOrBuilder();
|
||||||
|
@ -2522,8 +2522,8 @@ public final class MasterAdminProtos {
|
||||||
return region_;
|
return region_;
|
||||||
}
|
}
|
||||||
|
|
||||||
// optional .ServerName destServerName = 2;
|
// optional .ServerName dest_server_name = 2;
|
||||||
public static final int DESTSERVERNAME_FIELD_NUMBER = 2;
|
public static final int DEST_SERVER_NAME_FIELD_NUMBER = 2;
|
||||||
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName destServerName_;
|
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName destServerName_;
|
||||||
public boolean hasDestServerName() {
|
public boolean hasDestServerName() {
|
||||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||||
|
@ -2635,7 +2635,7 @@ public final class MasterAdminProtos {
|
||||||
hash = (53 * hash) + getRegion().hashCode();
|
hash = (53 * hash) + getRegion().hashCode();
|
||||||
}
|
}
|
||||||
if (hasDestServerName()) {
|
if (hasDestServerName()) {
|
||||||
hash = (37 * hash) + DESTSERVERNAME_FIELD_NUMBER;
|
hash = (37 * hash) + DEST_SERVER_NAME_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + getDestServerName().hashCode();
|
hash = (53 * hash) + getDestServerName().hashCode();
|
||||||
}
|
}
|
||||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||||
|
@ -3003,7 +3003,7 @@ public final class MasterAdminProtos {
|
||||||
return regionBuilder_;
|
return regionBuilder_;
|
||||||
}
|
}
|
||||||
|
|
||||||
// optional .ServerName destServerName = 2;
|
// optional .ServerName dest_server_name = 2;
|
||||||
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName destServerName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.getDefaultInstance();
|
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName destServerName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.getDefaultInstance();
|
||||||
private com.google.protobuf.SingleFieldBuilder<
|
private com.google.protobuf.SingleFieldBuilder<
|
||||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerNameOrBuilder> destServerNameBuilder_;
|
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerNameOrBuilder> destServerNameBuilder_;
|
||||||
|
@ -3407,12 +3407,12 @@ public final class MasterAdminProtos {
|
||||||
public interface DispatchMergingRegionsRequestOrBuilder
|
public interface DispatchMergingRegionsRequestOrBuilder
|
||||||
extends com.google.protobuf.MessageOrBuilder {
|
extends com.google.protobuf.MessageOrBuilder {
|
||||||
|
|
||||||
// required .RegionSpecifier regionA = 1;
|
// required .RegionSpecifier region_a = 1;
|
||||||
boolean hasRegionA();
|
boolean hasRegionA();
|
||||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier getRegionA();
|
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier getRegionA();
|
||||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder getRegionAOrBuilder();
|
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder getRegionAOrBuilder();
|
||||||
|
|
||||||
// required .RegionSpecifier regionB = 2;
|
// required .RegionSpecifier region_b = 2;
|
||||||
boolean hasRegionB();
|
boolean hasRegionB();
|
||||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier getRegionB();
|
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier getRegionB();
|
||||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder getRegionBOrBuilder();
|
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder getRegionBOrBuilder();
|
||||||
|
@ -3450,8 +3450,8 @@ public final class MasterAdminProtos {
|
||||||
}
|
}
|
||||||
|
|
||||||
private int bitField0_;
|
private int bitField0_;
|
||||||
// required .RegionSpecifier regionA = 1;
|
// required .RegionSpecifier region_a = 1;
|
||||||
public static final int REGIONA_FIELD_NUMBER = 1;
|
public static final int REGION_A_FIELD_NUMBER = 1;
|
||||||
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier regionA_;
|
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier regionA_;
|
||||||
public boolean hasRegionA() {
|
public boolean hasRegionA() {
|
||||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||||
|
@ -3463,8 +3463,8 @@ public final class MasterAdminProtos {
|
||||||
return regionA_;
|
return regionA_;
|
||||||
}
|
}
|
||||||
|
|
||||||
// required .RegionSpecifier regionB = 2;
|
// required .RegionSpecifier region_b = 2;
|
||||||
public static final int REGIONB_FIELD_NUMBER = 2;
|
public static final int REGION_B_FIELD_NUMBER = 2;
|
||||||
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier regionB_;
|
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier regionB_;
|
||||||
public boolean hasRegionB() {
|
public boolean hasRegionB() {
|
||||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||||
|
@ -3597,11 +3597,11 @@ public final class MasterAdminProtos {
|
||||||
int hash = 41;
|
int hash = 41;
|
||||||
hash = (19 * hash) + getDescriptorForType().hashCode();
|
hash = (19 * hash) + getDescriptorForType().hashCode();
|
||||||
if (hasRegionA()) {
|
if (hasRegionA()) {
|
||||||
hash = (37 * hash) + REGIONA_FIELD_NUMBER;
|
hash = (37 * hash) + REGION_A_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + getRegionA().hashCode();
|
hash = (53 * hash) + getRegionA().hashCode();
|
||||||
}
|
}
|
||||||
if (hasRegionB()) {
|
if (hasRegionB()) {
|
||||||
hash = (37 * hash) + REGIONB_FIELD_NUMBER;
|
hash = (37 * hash) + REGION_B_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + getRegionB().hashCode();
|
hash = (53 * hash) + getRegionB().hashCode();
|
||||||
}
|
}
|
||||||
if (hasForcible()) {
|
if (hasForcible()) {
|
||||||
|
@ -3899,7 +3899,7 @@ public final class MasterAdminProtos {
|
||||||
|
|
||||||
private int bitField0_;
|
private int bitField0_;
|
||||||
|
|
||||||
// required .RegionSpecifier regionA = 1;
|
// required .RegionSpecifier region_a = 1;
|
||||||
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier regionA_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance();
|
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier regionA_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance();
|
||||||
private com.google.protobuf.SingleFieldBuilder<
|
private com.google.protobuf.SingleFieldBuilder<
|
||||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder> regionABuilder_;
|
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder> regionABuilder_;
|
||||||
|
@ -3989,7 +3989,7 @@ public final class MasterAdminProtos {
|
||||||
return regionABuilder_;
|
return regionABuilder_;
|
||||||
}
|
}
|
||||||
|
|
||||||
// required .RegionSpecifier regionB = 2;
|
// required .RegionSpecifier region_b = 2;
|
||||||
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier regionB_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance();
|
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier regionB_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance();
|
||||||
private com.google.protobuf.SingleFieldBuilder<
|
private com.google.protobuf.SingleFieldBuilder<
|
||||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder> regionBBuilder_;
|
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder> regionBBuilder_;
|
||||||
|
@ -6799,12 +6799,12 @@ public final class MasterAdminProtos {
|
||||||
public interface CreateTableRequestOrBuilder
|
public interface CreateTableRequestOrBuilder
|
||||||
extends com.google.protobuf.MessageOrBuilder {
|
extends com.google.protobuf.MessageOrBuilder {
|
||||||
|
|
||||||
// required .TableSchema tableSchema = 1;
|
// required .TableSchema table_schema = 1;
|
||||||
boolean hasTableSchema();
|
boolean hasTableSchema();
|
||||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema getTableSchema();
|
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema getTableSchema();
|
||||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchemaOrBuilder getTableSchemaOrBuilder();
|
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchemaOrBuilder getTableSchemaOrBuilder();
|
||||||
|
|
||||||
// repeated bytes splitKeys = 2;
|
// repeated bytes split_keys = 2;
|
||||||
java.util.List<com.google.protobuf.ByteString> getSplitKeysList();
|
java.util.List<com.google.protobuf.ByteString> getSplitKeysList();
|
||||||
int getSplitKeysCount();
|
int getSplitKeysCount();
|
||||||
com.google.protobuf.ByteString getSplitKeys(int index);
|
com.google.protobuf.ByteString getSplitKeys(int index);
|
||||||
|
@ -6838,8 +6838,8 @@ public final class MasterAdminProtos {
|
||||||
}
|
}
|
||||||
|
|
||||||
private int bitField0_;
|
private int bitField0_;
|
||||||
// required .TableSchema tableSchema = 1;
|
// required .TableSchema table_schema = 1;
|
||||||
public static final int TABLESCHEMA_FIELD_NUMBER = 1;
|
public static final int TABLE_SCHEMA_FIELD_NUMBER = 1;
|
||||||
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema tableSchema_;
|
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema tableSchema_;
|
||||||
public boolean hasTableSchema() {
|
public boolean hasTableSchema() {
|
||||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||||
|
@ -6851,8 +6851,8 @@ public final class MasterAdminProtos {
|
||||||
return tableSchema_;
|
return tableSchema_;
|
||||||
}
|
}
|
||||||
|
|
||||||
// repeated bytes splitKeys = 2;
|
// repeated bytes split_keys = 2;
|
||||||
public static final int SPLITKEYS_FIELD_NUMBER = 2;
|
public static final int SPLIT_KEYS_FIELD_NUMBER = 2;
|
||||||
private java.util.List<com.google.protobuf.ByteString> splitKeys_;
|
private java.util.List<com.google.protobuf.ByteString> splitKeys_;
|
||||||
public java.util.List<com.google.protobuf.ByteString>
|
public java.util.List<com.google.protobuf.ByteString>
|
||||||
getSplitKeysList() {
|
getSplitKeysList() {
|
||||||
|
@ -6957,11 +6957,11 @@ public final class MasterAdminProtos {
|
||||||
int hash = 41;
|
int hash = 41;
|
||||||
hash = (19 * hash) + getDescriptorForType().hashCode();
|
hash = (19 * hash) + getDescriptorForType().hashCode();
|
||||||
if (hasTableSchema()) {
|
if (hasTableSchema()) {
|
||||||
hash = (37 * hash) + TABLESCHEMA_FIELD_NUMBER;
|
hash = (37 * hash) + TABLE_SCHEMA_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + getTableSchema().hashCode();
|
hash = (53 * hash) + getTableSchema().hashCode();
|
||||||
}
|
}
|
||||||
if (getSplitKeysCount() > 0) {
|
if (getSplitKeysCount() > 0) {
|
||||||
hash = (37 * hash) + SPLITKEYS_FIELD_NUMBER;
|
hash = (37 * hash) + SPLIT_KEYS_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + getSplitKeysList().hashCode();
|
hash = (53 * hash) + getSplitKeysList().hashCode();
|
||||||
}
|
}
|
||||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||||
|
@ -7228,7 +7228,7 @@ public final class MasterAdminProtos {
|
||||||
|
|
||||||
private int bitField0_;
|
private int bitField0_;
|
||||||
|
|
||||||
// required .TableSchema tableSchema = 1;
|
// required .TableSchema table_schema = 1;
|
||||||
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema tableSchema_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema.getDefaultInstance();
|
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema tableSchema_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema.getDefaultInstance();
|
||||||
private com.google.protobuf.SingleFieldBuilder<
|
private com.google.protobuf.SingleFieldBuilder<
|
||||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchemaOrBuilder> tableSchemaBuilder_;
|
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchemaOrBuilder> tableSchemaBuilder_;
|
||||||
|
@ -7318,7 +7318,7 @@ public final class MasterAdminProtos {
|
||||||
return tableSchemaBuilder_;
|
return tableSchemaBuilder_;
|
||||||
}
|
}
|
||||||
|
|
||||||
// repeated bytes splitKeys = 2;
|
// repeated bytes split_keys = 2;
|
||||||
private java.util.List<com.google.protobuf.ByteString> splitKeys_ = java.util.Collections.emptyList();;
|
private java.util.List<com.google.protobuf.ByteString> splitKeys_ = java.util.Collections.emptyList();;
|
||||||
private void ensureSplitKeysIsMutable() {
|
private void ensureSplitKeysIsMutable() {
|
||||||
if (!((bitField0_ & 0x00000002) == 0x00000002)) {
|
if (!((bitField0_ & 0x00000002) == 0x00000002)) {
|
||||||
|
@ -7683,7 +7683,7 @@ public final class MasterAdminProtos {
|
||||||
public interface DeleteTableRequestOrBuilder
|
public interface DeleteTableRequestOrBuilder
|
||||||
extends com.google.protobuf.MessageOrBuilder {
|
extends com.google.protobuf.MessageOrBuilder {
|
||||||
|
|
||||||
// required bytes tableName = 1;
|
// required bytes table_name = 1;
|
||||||
boolean hasTableName();
|
boolean hasTableName();
|
||||||
com.google.protobuf.ByteString getTableName();
|
com.google.protobuf.ByteString getTableName();
|
||||||
}
|
}
|
||||||
|
@ -7716,8 +7716,8 @@ public final class MasterAdminProtos {
|
||||||
}
|
}
|
||||||
|
|
||||||
private int bitField0_;
|
private int bitField0_;
|
||||||
// required bytes tableName = 1;
|
// required bytes table_name = 1;
|
||||||
public static final int TABLENAME_FIELD_NUMBER = 1;
|
public static final int TABLE_NAME_FIELD_NUMBER = 1;
|
||||||
private com.google.protobuf.ByteString tableName_;
|
private com.google.protobuf.ByteString tableName_;
|
||||||
public boolean hasTableName() {
|
public boolean hasTableName() {
|
||||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||||
|
@ -7799,7 +7799,7 @@ public final class MasterAdminProtos {
|
||||||
int hash = 41;
|
int hash = 41;
|
||||||
hash = (19 * hash) + getDescriptorForType().hashCode();
|
hash = (19 * hash) + getDescriptorForType().hashCode();
|
||||||
if (hasTableName()) {
|
if (hasTableName()) {
|
||||||
hash = (37 * hash) + TABLENAME_FIELD_NUMBER;
|
hash = (37 * hash) + TABLE_NAME_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + getTableName().hashCode();
|
hash = (53 * hash) + getTableName().hashCode();
|
||||||
}
|
}
|
||||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||||
|
@ -8027,7 +8027,7 @@ public final class MasterAdminProtos {
|
||||||
|
|
||||||
private int bitField0_;
|
private int bitField0_;
|
||||||
|
|
||||||
// required bytes tableName = 1;
|
// required bytes table_name = 1;
|
||||||
private com.google.protobuf.ByteString tableName_ = com.google.protobuf.ByteString.EMPTY;
|
private com.google.protobuf.ByteString tableName_ = com.google.protobuf.ByteString.EMPTY;
|
||||||
public boolean hasTableName() {
|
public boolean hasTableName() {
|
||||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||||
|
@ -8365,7 +8365,7 @@ public final class MasterAdminProtos {
|
||||||
public interface EnableTableRequestOrBuilder
|
public interface EnableTableRequestOrBuilder
|
||||||
extends com.google.protobuf.MessageOrBuilder {
|
extends com.google.protobuf.MessageOrBuilder {
|
||||||
|
|
||||||
// required bytes tableName = 1;
|
// required bytes table_name = 1;
|
||||||
boolean hasTableName();
|
boolean hasTableName();
|
||||||
com.google.protobuf.ByteString getTableName();
|
com.google.protobuf.ByteString getTableName();
|
||||||
}
|
}
|
||||||
|
@ -8398,8 +8398,8 @@ public final class MasterAdminProtos {
|
||||||
}
|
}
|
||||||
|
|
||||||
private int bitField0_;
|
private int bitField0_;
|
||||||
// required bytes tableName = 1;
|
// required bytes table_name = 1;
|
||||||
public static final int TABLENAME_FIELD_NUMBER = 1;
|
public static final int TABLE_NAME_FIELD_NUMBER = 1;
|
||||||
private com.google.protobuf.ByteString tableName_;
|
private com.google.protobuf.ByteString tableName_;
|
||||||
public boolean hasTableName() {
|
public boolean hasTableName() {
|
||||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||||
|
@ -8481,7 +8481,7 @@ public final class MasterAdminProtos {
|
||||||
int hash = 41;
|
int hash = 41;
|
||||||
hash = (19 * hash) + getDescriptorForType().hashCode();
|
hash = (19 * hash) + getDescriptorForType().hashCode();
|
||||||
if (hasTableName()) {
|
if (hasTableName()) {
|
||||||
hash = (37 * hash) + TABLENAME_FIELD_NUMBER;
|
hash = (37 * hash) + TABLE_NAME_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + getTableName().hashCode();
|
hash = (53 * hash) + getTableName().hashCode();
|
||||||
}
|
}
|
||||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||||
|
@ -8709,7 +8709,7 @@ public final class MasterAdminProtos {
|
||||||
|
|
||||||
private int bitField0_;
|
private int bitField0_;
|
||||||
|
|
||||||
// required bytes tableName = 1;
|
// required bytes table_name = 1;
|
||||||
private com.google.protobuf.ByteString tableName_ = com.google.protobuf.ByteString.EMPTY;
|
private com.google.protobuf.ByteString tableName_ = com.google.protobuf.ByteString.EMPTY;
|
||||||
public boolean hasTableName() {
|
public boolean hasTableName() {
|
||||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||||
|
@ -9047,7 +9047,7 @@ public final class MasterAdminProtos {
|
||||||
public interface DisableTableRequestOrBuilder
|
public interface DisableTableRequestOrBuilder
|
||||||
extends com.google.protobuf.MessageOrBuilder {
|
extends com.google.protobuf.MessageOrBuilder {
|
||||||
|
|
||||||
// required bytes tableName = 1;
|
// required bytes table_name = 1;
|
||||||
boolean hasTableName();
|
boolean hasTableName();
|
||||||
com.google.protobuf.ByteString getTableName();
|
com.google.protobuf.ByteString getTableName();
|
||||||
}
|
}
|
||||||
|
@ -9080,8 +9080,8 @@ public final class MasterAdminProtos {
|
||||||
}
|
}
|
||||||
|
|
||||||
private int bitField0_;
|
private int bitField0_;
|
||||||
// required bytes tableName = 1;
|
// required bytes table_name = 1;
|
||||||
public static final int TABLENAME_FIELD_NUMBER = 1;
|
public static final int TABLE_NAME_FIELD_NUMBER = 1;
|
||||||
private com.google.protobuf.ByteString tableName_;
|
private com.google.protobuf.ByteString tableName_;
|
||||||
public boolean hasTableName() {
|
public boolean hasTableName() {
|
||||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||||
|
@ -9163,7 +9163,7 @@ public final class MasterAdminProtos {
|
||||||
int hash = 41;
|
int hash = 41;
|
||||||
hash = (19 * hash) + getDescriptorForType().hashCode();
|
hash = (19 * hash) + getDescriptorForType().hashCode();
|
||||||
if (hasTableName()) {
|
if (hasTableName()) {
|
||||||
hash = (37 * hash) + TABLENAME_FIELD_NUMBER;
|
hash = (37 * hash) + TABLE_NAME_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + getTableName().hashCode();
|
hash = (53 * hash) + getTableName().hashCode();
|
||||||
}
|
}
|
||||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||||
|
@ -9391,7 +9391,7 @@ public final class MasterAdminProtos {
|
||||||
|
|
||||||
private int bitField0_;
|
private int bitField0_;
|
||||||
|
|
||||||
// required bytes tableName = 1;
|
// required bytes table_name = 1;
|
||||||
private com.google.protobuf.ByteString tableName_ = com.google.protobuf.ByteString.EMPTY;
|
private com.google.protobuf.ByteString tableName_ = com.google.protobuf.ByteString.EMPTY;
|
||||||
public boolean hasTableName() {
|
public boolean hasTableName() {
|
||||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||||
|
@ -9729,11 +9729,11 @@ public final class MasterAdminProtos {
|
||||||
public interface ModifyTableRequestOrBuilder
|
public interface ModifyTableRequestOrBuilder
|
||||||
extends com.google.protobuf.MessageOrBuilder {
|
extends com.google.protobuf.MessageOrBuilder {
|
||||||
|
|
||||||
// required bytes tableName = 1;
|
// required bytes table_name = 1;
|
||||||
boolean hasTableName();
|
boolean hasTableName();
|
||||||
com.google.protobuf.ByteString getTableName();
|
com.google.protobuf.ByteString getTableName();
|
||||||
|
|
||||||
// required .TableSchema tableSchema = 2;
|
// required .TableSchema table_schema = 2;
|
||||||
boolean hasTableSchema();
|
boolean hasTableSchema();
|
||||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema getTableSchema();
|
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema getTableSchema();
|
||||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchemaOrBuilder getTableSchemaOrBuilder();
|
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchemaOrBuilder getTableSchemaOrBuilder();
|
||||||
|
@ -9767,8 +9767,8 @@ public final class MasterAdminProtos {
|
||||||
}
|
}
|
||||||
|
|
||||||
private int bitField0_;
|
private int bitField0_;
|
||||||
// required bytes tableName = 1;
|
// required bytes table_name = 1;
|
||||||
public static final int TABLENAME_FIELD_NUMBER = 1;
|
public static final int TABLE_NAME_FIELD_NUMBER = 1;
|
||||||
private com.google.protobuf.ByteString tableName_;
|
private com.google.protobuf.ByteString tableName_;
|
||||||
public boolean hasTableName() {
|
public boolean hasTableName() {
|
||||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||||
|
@ -9777,8 +9777,8 @@ public final class MasterAdminProtos {
|
||||||
return tableName_;
|
return tableName_;
|
||||||
}
|
}
|
||||||
|
|
||||||
// required .TableSchema tableSchema = 2;
|
// required .TableSchema table_schema = 2;
|
||||||
public static final int TABLESCHEMA_FIELD_NUMBER = 2;
|
public static final int TABLE_SCHEMA_FIELD_NUMBER = 2;
|
||||||
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema tableSchema_;
|
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema tableSchema_;
|
||||||
public boolean hasTableSchema() {
|
public boolean hasTableSchema() {
|
||||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||||
|
@ -9884,11 +9884,11 @@ public final class MasterAdminProtos {
|
||||||
int hash = 41;
|
int hash = 41;
|
||||||
hash = (19 * hash) + getDescriptorForType().hashCode();
|
hash = (19 * hash) + getDescriptorForType().hashCode();
|
||||||
if (hasTableName()) {
|
if (hasTableName()) {
|
||||||
hash = (37 * hash) + TABLENAME_FIELD_NUMBER;
|
hash = (37 * hash) + TABLE_NAME_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + getTableName().hashCode();
|
hash = (53 * hash) + getTableName().hashCode();
|
||||||
}
|
}
|
||||||
if (hasTableSchema()) {
|
if (hasTableSchema()) {
|
||||||
hash = (37 * hash) + TABLESCHEMA_FIELD_NUMBER;
|
hash = (37 * hash) + TABLE_SCHEMA_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + getTableSchema().hashCode();
|
hash = (53 * hash) + getTableSchema().hashCode();
|
||||||
}
|
}
|
||||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||||
|
@ -10151,7 +10151,7 @@ public final class MasterAdminProtos {
|
||||||
|
|
||||||
private int bitField0_;
|
private int bitField0_;
|
||||||
|
|
||||||
// required bytes tableName = 1;
|
// required bytes table_name = 1;
|
||||||
private com.google.protobuf.ByteString tableName_ = com.google.protobuf.ByteString.EMPTY;
|
private com.google.protobuf.ByteString tableName_ = com.google.protobuf.ByteString.EMPTY;
|
||||||
public boolean hasTableName() {
|
public boolean hasTableName() {
|
||||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||||
|
@ -10175,7 +10175,7 @@ public final class MasterAdminProtos {
|
||||||
return this;
|
return this;
|
||||||
}
|
}
|
||||||
|
|
||||||
// required .TableSchema tableSchema = 2;
|
// required .TableSchema table_schema = 2;
|
||||||
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema tableSchema_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema.getDefaultInstance();
|
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema tableSchema_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema.getDefaultInstance();
|
||||||
private com.google.protobuf.SingleFieldBuilder<
|
private com.google.protobuf.SingleFieldBuilder<
|
||||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchemaOrBuilder> tableSchemaBuilder_;
|
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchemaOrBuilder> tableSchemaBuilder_;
|
||||||
|
@ -12079,7 +12079,7 @@ public final class MasterAdminProtos {
|
||||||
public interface BalanceResponseOrBuilder
|
public interface BalanceResponseOrBuilder
|
||||||
extends com.google.protobuf.MessageOrBuilder {
|
extends com.google.protobuf.MessageOrBuilder {
|
||||||
|
|
||||||
// required bool balancerRan = 1;
|
// required bool balancer_ran = 1;
|
||||||
boolean hasBalancerRan();
|
boolean hasBalancerRan();
|
||||||
boolean getBalancerRan();
|
boolean getBalancerRan();
|
||||||
}
|
}
|
||||||
|
@ -12112,8 +12112,8 @@ public final class MasterAdminProtos {
|
||||||
}
|
}
|
||||||
|
|
||||||
private int bitField0_;
|
private int bitField0_;
|
||||||
// required bool balancerRan = 1;
|
// required bool balancer_ran = 1;
|
||||||
public static final int BALANCERRAN_FIELD_NUMBER = 1;
|
public static final int BALANCER_RAN_FIELD_NUMBER = 1;
|
||||||
private boolean balancerRan_;
|
private boolean balancerRan_;
|
||||||
public boolean hasBalancerRan() {
|
public boolean hasBalancerRan() {
|
||||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||||
|
@ -12195,7 +12195,7 @@ public final class MasterAdminProtos {
|
||||||
int hash = 41;
|
int hash = 41;
|
||||||
hash = (19 * hash) + getDescriptorForType().hashCode();
|
hash = (19 * hash) + getDescriptorForType().hashCode();
|
||||||
if (hasBalancerRan()) {
|
if (hasBalancerRan()) {
|
||||||
hash = (37 * hash) + BALANCERRAN_FIELD_NUMBER;
|
hash = (37 * hash) + BALANCER_RAN_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + hashBoolean(getBalancerRan());
|
hash = (53 * hash) + hashBoolean(getBalancerRan());
|
||||||
}
|
}
|
||||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||||
|
@ -12423,7 +12423,7 @@ public final class MasterAdminProtos {
|
||||||
|
|
||||||
private int bitField0_;
|
private int bitField0_;
|
||||||
|
|
||||||
// required bool balancerRan = 1;
|
// required bool balancer_ran = 1;
|
||||||
private boolean balancerRan_ ;
|
private boolean balancerRan_ ;
|
||||||
public boolean hasBalancerRan() {
|
public boolean hasBalancerRan() {
|
||||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||||
|
@ -12903,7 +12903,7 @@ public final class MasterAdminProtos {
|
||||||
public interface SetBalancerRunningResponseOrBuilder
|
public interface SetBalancerRunningResponseOrBuilder
|
||||||
extends com.google.protobuf.MessageOrBuilder {
|
extends com.google.protobuf.MessageOrBuilder {
|
||||||
|
|
||||||
// optional bool prevBalanceValue = 1;
|
// optional bool prev_balance_value = 1;
|
||||||
boolean hasPrevBalanceValue();
|
boolean hasPrevBalanceValue();
|
||||||
boolean getPrevBalanceValue();
|
boolean getPrevBalanceValue();
|
||||||
}
|
}
|
||||||
|
@ -12936,8 +12936,8 @@ public final class MasterAdminProtos {
|
||||||
}
|
}
|
||||||
|
|
||||||
private int bitField0_;
|
private int bitField0_;
|
||||||
// optional bool prevBalanceValue = 1;
|
// optional bool prev_balance_value = 1;
|
||||||
public static final int PREVBALANCEVALUE_FIELD_NUMBER = 1;
|
public static final int PREV_BALANCE_VALUE_FIELD_NUMBER = 1;
|
||||||
private boolean prevBalanceValue_;
|
private boolean prevBalanceValue_;
|
||||||
public boolean hasPrevBalanceValue() {
|
public boolean hasPrevBalanceValue() {
|
||||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||||
|
@ -13015,7 +13015,7 @@ public final class MasterAdminProtos {
|
||||||
int hash = 41;
|
int hash = 41;
|
||||||
hash = (19 * hash) + getDescriptorForType().hashCode();
|
hash = (19 * hash) + getDescriptorForType().hashCode();
|
||||||
if (hasPrevBalanceValue()) {
|
if (hasPrevBalanceValue()) {
|
||||||
hash = (37 * hash) + PREVBALANCEVALUE_FIELD_NUMBER;
|
hash = (37 * hash) + PREV_BALANCE_VALUE_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + hashBoolean(getPrevBalanceValue());
|
hash = (53 * hash) + hashBoolean(getPrevBalanceValue());
|
||||||
}
|
}
|
||||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||||
|
@ -13239,7 +13239,7 @@ public final class MasterAdminProtos {
|
||||||
|
|
||||||
private int bitField0_;
|
private int bitField0_;
|
||||||
|
|
||||||
// optional bool prevBalanceValue = 1;
|
// optional bool prev_balance_value = 1;
|
||||||
private boolean prevBalanceValue_ ;
|
private boolean prevBalanceValue_ ;
|
||||||
public boolean hasPrevBalanceValue() {
|
public boolean hasPrevBalanceValue() {
|
||||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||||
|
@ -13574,7 +13574,7 @@ public final class MasterAdminProtos {
|
||||||
public interface CatalogScanResponseOrBuilder
|
public interface CatalogScanResponseOrBuilder
|
||||||
extends com.google.protobuf.MessageOrBuilder {
|
extends com.google.protobuf.MessageOrBuilder {
|
||||||
|
|
||||||
// optional int32 scanResult = 1;
|
// optional int32 scan_result = 1;
|
||||||
boolean hasScanResult();
|
boolean hasScanResult();
|
||||||
int getScanResult();
|
int getScanResult();
|
||||||
}
|
}
|
||||||
|
@ -13607,8 +13607,8 @@ public final class MasterAdminProtos {
|
||||||
}
|
}
|
||||||
|
|
||||||
private int bitField0_;
|
private int bitField0_;
|
||||||
// optional int32 scanResult = 1;
|
// optional int32 scan_result = 1;
|
||||||
public static final int SCANRESULT_FIELD_NUMBER = 1;
|
public static final int SCAN_RESULT_FIELD_NUMBER = 1;
|
||||||
private int scanResult_;
|
private int scanResult_;
|
||||||
public boolean hasScanResult() {
|
public boolean hasScanResult() {
|
||||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||||
|
@ -13686,7 +13686,7 @@ public final class MasterAdminProtos {
|
||||||
int hash = 41;
|
int hash = 41;
|
||||||
hash = (19 * hash) + getDescriptorForType().hashCode();
|
hash = (19 * hash) + getDescriptorForType().hashCode();
|
||||||
if (hasScanResult()) {
|
if (hasScanResult()) {
|
||||||
hash = (37 * hash) + SCANRESULT_FIELD_NUMBER;
|
hash = (37 * hash) + SCAN_RESULT_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + getScanResult();
|
hash = (53 * hash) + getScanResult();
|
||||||
}
|
}
|
||||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||||
|
@ -13910,7 +13910,7 @@ public final class MasterAdminProtos {
|
||||||
|
|
||||||
private int bitField0_;
|
private int bitField0_;
|
||||||
|
|
||||||
// optional int32 scanResult = 1;
|
// optional int32 scan_result = 1;
|
||||||
private int scanResult_ ;
|
private int scanResult_ ;
|
||||||
public boolean hasScanResult() {
|
public boolean hasScanResult() {
|
||||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||||
|
@ -14324,7 +14324,7 @@ public final class MasterAdminProtos {
|
||||||
public interface EnableCatalogJanitorResponseOrBuilder
|
public interface EnableCatalogJanitorResponseOrBuilder
|
||||||
extends com.google.protobuf.MessageOrBuilder {
|
extends com.google.protobuf.MessageOrBuilder {
|
||||||
|
|
||||||
// optional bool prevValue = 1;
|
// optional bool prev_value = 1;
|
||||||
boolean hasPrevValue();
|
boolean hasPrevValue();
|
||||||
boolean getPrevValue();
|
boolean getPrevValue();
|
||||||
}
|
}
|
||||||
|
@ -14357,8 +14357,8 @@ public final class MasterAdminProtos {
|
||||||
}
|
}
|
||||||
|
|
||||||
private int bitField0_;
|
private int bitField0_;
|
||||||
// optional bool prevValue = 1;
|
// optional bool prev_value = 1;
|
||||||
public static final int PREVVALUE_FIELD_NUMBER = 1;
|
public static final int PREV_VALUE_FIELD_NUMBER = 1;
|
||||||
private boolean prevValue_;
|
private boolean prevValue_;
|
||||||
public boolean hasPrevValue() {
|
public boolean hasPrevValue() {
|
||||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||||
|
@ -14436,7 +14436,7 @@ public final class MasterAdminProtos {
|
||||||
int hash = 41;
|
int hash = 41;
|
||||||
hash = (19 * hash) + getDescriptorForType().hashCode();
|
hash = (19 * hash) + getDescriptorForType().hashCode();
|
||||||
if (hasPrevValue()) {
|
if (hasPrevValue()) {
|
||||||
hash = (37 * hash) + PREVVALUE_FIELD_NUMBER;
|
hash = (37 * hash) + PREV_VALUE_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + hashBoolean(getPrevValue());
|
hash = (53 * hash) + hashBoolean(getPrevValue());
|
||||||
}
|
}
|
||||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||||
|
@ -14660,7 +14660,7 @@ public final class MasterAdminProtos {
|
||||||
|
|
||||||
private int bitField0_;
|
private int bitField0_;
|
||||||
|
|
||||||
// optional bool prevValue = 1;
|
// optional bool prev_value = 1;
|
||||||
private boolean prevValue_ ;
|
private boolean prevValue_ ;
|
||||||
public boolean hasPrevValue() {
|
public boolean hasPrevValue() {
|
||||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||||
|
@ -15847,7 +15847,7 @@ public final class MasterAdminProtos {
|
||||||
public interface TakeSnapshotResponseOrBuilder
|
public interface TakeSnapshotResponseOrBuilder
|
||||||
extends com.google.protobuf.MessageOrBuilder {
|
extends com.google.protobuf.MessageOrBuilder {
|
||||||
|
|
||||||
// required int64 expectedTimeout = 1;
|
// required int64 expected_timeout = 1;
|
||||||
boolean hasExpectedTimeout();
|
boolean hasExpectedTimeout();
|
||||||
long getExpectedTimeout();
|
long getExpectedTimeout();
|
||||||
}
|
}
|
||||||
|
@ -15880,8 +15880,8 @@ public final class MasterAdminProtos {
|
||||||
}
|
}
|
||||||
|
|
||||||
private int bitField0_;
|
private int bitField0_;
|
||||||
// required int64 expectedTimeout = 1;
|
// required int64 expected_timeout = 1;
|
||||||
public static final int EXPECTEDTIMEOUT_FIELD_NUMBER = 1;
|
public static final int EXPECTED_TIMEOUT_FIELD_NUMBER = 1;
|
||||||
private long expectedTimeout_;
|
private long expectedTimeout_;
|
||||||
public boolean hasExpectedTimeout() {
|
public boolean hasExpectedTimeout() {
|
||||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||||
|
@ -15963,7 +15963,7 @@ public final class MasterAdminProtos {
|
||||||
int hash = 41;
|
int hash = 41;
|
||||||
hash = (19 * hash) + getDescriptorForType().hashCode();
|
hash = (19 * hash) + getDescriptorForType().hashCode();
|
||||||
if (hasExpectedTimeout()) {
|
if (hasExpectedTimeout()) {
|
||||||
hash = (37 * hash) + EXPECTEDTIMEOUT_FIELD_NUMBER;
|
hash = (37 * hash) + EXPECTED_TIMEOUT_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + hashLong(getExpectedTimeout());
|
hash = (53 * hash) + hashLong(getExpectedTimeout());
|
||||||
}
|
}
|
||||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||||
|
@ -16191,7 +16191,7 @@ public final class MasterAdminProtos {
|
||||||
|
|
||||||
private int bitField0_;
|
private int bitField0_;
|
||||||
|
|
||||||
// required int64 expectedTimeout = 1;
|
// required int64 expected_timeout = 1;
|
||||||
private long expectedTimeout_ ;
|
private long expectedTimeout_ ;
|
||||||
public boolean hasExpectedTimeout() {
|
public boolean hasExpectedTimeout() {
|
||||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||||
|
@ -22752,116 +22752,116 @@ public final class MasterAdminProtos {
|
||||||
static {
|
static {
|
||||||
java.lang.String[] descriptorData = {
|
java.lang.String[] descriptorData = {
|
||||||
"\n\021MasterAdmin.proto\032\014Master.proto\032\013hbase" +
|
"\n\021MasterAdmin.proto\032\014Master.proto\032\013hbase" +
|
||||||
".proto\032\014Client.proto\"R\n\020AddColumnRequest" +
|
".proto\032\014Client.proto\"T\n\020AddColumnRequest" +
|
||||||
"\022\021\n\ttableName\030\001 \002(\014\022+\n\016columnFamilies\030\002 " +
|
"\022\022\n\ntable_name\030\001 \002(\014\022,\n\017column_families\030" +
|
||||||
"\002(\0132\023.ColumnFamilySchema\"\023\n\021AddColumnRes" +
|
"\002 \002(\0132\023.ColumnFamilySchema\"\023\n\021AddColumnR" +
|
||||||
"ponse\"<\n\023DeleteColumnRequest\022\021\n\ttableNam" +
|
"esponse\">\n\023DeleteColumnRequest\022\022\n\ntable_" +
|
||||||
"e\030\001 \002(\014\022\022\n\ncolumnName\030\002 \002(\014\"\026\n\024DeleteCol" +
|
"name\030\001 \002(\014\022\023\n\013column_name\030\002 \002(\014\"\026\n\024Delet" +
|
||||||
"umnResponse\"U\n\023ModifyColumnRequest\022\021\n\tta" +
|
"eColumnResponse\"W\n\023ModifyColumnRequest\022\022" +
|
||||||
"bleName\030\001 \002(\014\022+\n\016columnFamilies\030\002 \002(\0132\023." +
|
"\n\ntable_name\030\001 \002(\014\022,\n\017column_families\030\002 " +
|
||||||
"ColumnFamilySchema\"\026\n\024ModifyColumnRespon" +
|
"\002(\0132\023.ColumnFamilySchema\"\026\n\024ModifyColumn" +
|
||||||
"se\"Z\n\021MoveRegionRequest\022 \n\006region\030\001 \002(\0132",
|
"Response\"\\\n\021MoveRegionRequest\022 \n\006region\030",
|
||||||
"\020.RegionSpecifier\022#\n\016destServerName\030\002 \001(" +
|
"\001 \002(\0132\020.RegionSpecifier\022%\n\020dest_server_n" +
|
||||||
"\0132\013.ServerName\"\024\n\022MoveRegionResponse\"~\n\035" +
|
"ame\030\002 \001(\0132\013.ServerName\"\024\n\022MoveRegionResp" +
|
||||||
"DispatchMergingRegionsRequest\022!\n\007regionA" +
|
"onse\"\200\001\n\035DispatchMergingRegionsRequest\022\"" +
|
||||||
"\030\001 \002(\0132\020.RegionSpecifier\022!\n\007regionB\030\002 \002(" +
|
"\n\010region_a\030\001 \002(\0132\020.RegionSpecifier\022\"\n\010re" +
|
||||||
"\0132\020.RegionSpecifier\022\027\n\010forcible\030\003 \001(\010:\005f" +
|
"gion_b\030\002 \002(\0132\020.RegionSpecifier\022\027\n\010forcib" +
|
||||||
"alse\" \n\036DispatchMergingRegionsResponse\"7" +
|
"le\030\003 \001(\010:\005false\" \n\036DispatchMergingRegion" +
|
||||||
"\n\023AssignRegionRequest\022 \n\006region\030\001 \002(\0132\020." +
|
"sResponse\"7\n\023AssignRegionRequest\022 \n\006regi" +
|
||||||
"RegionSpecifier\"\026\n\024AssignRegionResponse\"" +
|
"on\030\001 \002(\0132\020.RegionSpecifier\"\026\n\024AssignRegi" +
|
||||||
"O\n\025UnassignRegionRequest\022 \n\006region\030\001 \002(\013" +
|
"onResponse\"O\n\025UnassignRegionRequest\022 \n\006r" +
|
||||||
"2\020.RegionSpecifier\022\024\n\005force\030\002 \001(\010:\005false",
|
"egion\030\001 \002(\0132\020.RegionSpecifier\022\024\n\005force\030\002",
|
||||||
"\"\030\n\026UnassignRegionResponse\"8\n\024OfflineReg" +
|
" \001(\010:\005false\"\030\n\026UnassignRegionResponse\"8\n" +
|
||||||
"ionRequest\022 \n\006region\030\001 \002(\0132\020.RegionSpeci" +
|
"\024OfflineRegionRequest\022 \n\006region\030\001 \002(\0132\020." +
|
||||||
"fier\"\027\n\025OfflineRegionResponse\"J\n\022CreateT" +
|
"RegionSpecifier\"\027\n\025OfflineRegionResponse" +
|
||||||
"ableRequest\022!\n\013tableSchema\030\001 \002(\0132\014.Table" +
|
"\"L\n\022CreateTableRequest\022\"\n\014table_schema\030\001" +
|
||||||
"Schema\022\021\n\tsplitKeys\030\002 \003(\014\"\025\n\023CreateTable" +
|
" \002(\0132\014.TableSchema\022\022\n\nsplit_keys\030\002 \003(\014\"\025" +
|
||||||
"Response\"\'\n\022DeleteTableRequest\022\021\n\ttableN" +
|
"\n\023CreateTableResponse\"(\n\022DeleteTableRequ" +
|
||||||
"ame\030\001 \002(\014\"\025\n\023DeleteTableResponse\"\'\n\022Enab" +
|
"est\022\022\n\ntable_name\030\001 \002(\014\"\025\n\023DeleteTableRe" +
|
||||||
"leTableRequest\022\021\n\ttableName\030\001 \002(\014\"\025\n\023Ena" +
|
"sponse\"(\n\022EnableTableRequest\022\022\n\ntable_na" +
|
||||||
"bleTableResponse\"(\n\023DisableTableRequest\022" +
|
"me\030\001 \002(\014\"\025\n\023EnableTableResponse\")\n\023Disab" +
|
||||||
"\021\n\ttableName\030\001 \002(\014\"\026\n\024DisableTableRespon",
|
"leTableRequest\022\022\n\ntable_name\030\001 \002(\014\"\026\n\024Di",
|
||||||
"se\"J\n\022ModifyTableRequest\022\021\n\ttableName\030\001 " +
|
"sableTableResponse\"L\n\022ModifyTableRequest" +
|
||||||
"\002(\014\022!\n\013tableSchema\030\002 \002(\0132\014.TableSchema\"\025" +
|
"\022\022\n\ntable_name\030\001 \002(\014\022\"\n\014table_schema\030\002 \002" +
|
||||||
"\n\023ModifyTableResponse\"\021\n\017ShutdownRequest" +
|
"(\0132\014.TableSchema\"\025\n\023ModifyTableResponse\"" +
|
||||||
"\"\022\n\020ShutdownResponse\"\023\n\021StopMasterReques" +
|
"\021\n\017ShutdownRequest\"\022\n\020ShutdownResponse\"\023" +
|
||||||
"t\"\024\n\022StopMasterResponse\"\020\n\016BalanceReques" +
|
"\n\021StopMasterRequest\"\024\n\022StopMasterRespons" +
|
||||||
"t\"&\n\017BalanceResponse\022\023\n\013balancerRan\030\001 \002(" +
|
"e\"\020\n\016BalanceRequest\"\'\n\017BalanceResponse\022\024" +
|
||||||
"\010\"<\n\031SetBalancerRunningRequest\022\n\n\002on\030\001 \002" +
|
"\n\014balancer_ran\030\001 \002(\010\"<\n\031SetBalancerRunni" +
|
||||||
"(\010\022\023\n\013synchronous\030\002 \001(\010\"6\n\032SetBalancerRu" +
|
"ngRequest\022\n\n\002on\030\001 \002(\010\022\023\n\013synchronous\030\002 \001" +
|
||||||
"nningResponse\022\030\n\020prevBalanceValue\030\001 \001(\010\"" +
|
"(\010\"8\n\032SetBalancerRunningResponse\022\032\n\022prev" +
|
||||||
"\024\n\022CatalogScanRequest\")\n\023CatalogScanResp",
|
"_balance_value\030\001 \001(\010\"\024\n\022CatalogScanReque",
|
||||||
"onse\022\022\n\nscanResult\030\001 \001(\005\"-\n\033EnableCatalo" +
|
"st\"*\n\023CatalogScanResponse\022\023\n\013scan_result" +
|
||||||
"gJanitorRequest\022\016\n\006enable\030\001 \002(\010\"1\n\034Enabl" +
|
"\030\001 \001(\005\"-\n\033EnableCatalogJanitorRequest\022\016\n" +
|
||||||
"eCatalogJanitorResponse\022\021\n\tprevValue\030\001 \001" +
|
"\006enable\030\001 \002(\010\"2\n\034EnableCatalogJanitorRes" +
|
||||||
"(\010\" \n\036IsCatalogJanitorEnabledRequest\"0\n\037" +
|
"ponse\022\022\n\nprev_value\030\001 \001(\010\" \n\036IsCatalogJa" +
|
||||||
"IsCatalogJanitorEnabledResponse\022\r\n\005value" +
|
"nitorEnabledRequest\"0\n\037IsCatalogJanitorE" +
|
||||||
"\030\001 \002(\010\"=\n\023TakeSnapshotRequest\022&\n\010snapsho" +
|
"nabledResponse\022\r\n\005value\030\001 \002(\010\"=\n\023TakeSna" +
|
||||||
"t\030\001 \002(\0132\024.SnapshotDescription\"/\n\024TakeSna" +
|
"pshotRequest\022&\n\010snapshot\030\001 \002(\0132\024.Snapsho" +
|
||||||
"pshotResponse\022\027\n\017expectedTimeout\030\001 \002(\003\"\025" +
|
"tDescription\"0\n\024TakeSnapshotResponse\022\030\n\020" +
|
||||||
"\n\023ListSnapshotRequest\"?\n\024ListSnapshotRes" +
|
"expected_timeout\030\001 \002(\003\"\025\n\023ListSnapshotRe" +
|
||||||
"ponse\022\'\n\tsnapshots\030\001 \003(\0132\024.SnapshotDescr",
|
"quest\"?\n\024ListSnapshotResponse\022\'\n\tsnapsho",
|
||||||
"iption\"?\n\025DeleteSnapshotRequest\022&\n\010snaps" +
|
"ts\030\001 \003(\0132\024.SnapshotDescription\"?\n\025Delete" +
|
||||||
"hot\030\001 \002(\0132\024.SnapshotDescription\"\030\n\026Delet" +
|
"SnapshotRequest\022&\n\010snapshot\030\001 \002(\0132\024.Snap" +
|
||||||
"eSnapshotResponse\"@\n\026RestoreSnapshotRequ" +
|
"shotDescription\"\030\n\026DeleteSnapshotRespons" +
|
||||||
"est\022&\n\010snapshot\030\001 \002(\0132\024.SnapshotDescript" +
|
"e\"@\n\026RestoreSnapshotRequest\022&\n\010snapshot\030" +
|
||||||
"ion\"\031\n\027RestoreSnapshotResponse\"?\n\025IsSnap" +
|
"\001 \002(\0132\024.SnapshotDescription\"\031\n\027RestoreSn" +
|
||||||
"shotDoneRequest\022&\n\010snapshot\030\001 \001(\0132\024.Snap" +
|
"apshotResponse\"?\n\025IsSnapshotDoneRequest\022" +
|
||||||
"shotDescription\"U\n\026IsSnapshotDoneRespons" +
|
"&\n\010snapshot\030\001 \001(\0132\024.SnapshotDescription\"" +
|
||||||
"e\022\023\n\004done\030\001 \001(\010:\005false\022&\n\010snapshot\030\002 \001(\013" +
|
"U\n\026IsSnapshotDoneResponse\022\023\n\004done\030\001 \001(\010:" +
|
||||||
"2\024.SnapshotDescription\"F\n\034IsRestoreSnaps" +
|
"\005false\022&\n\010snapshot\030\002 \001(\0132\024.SnapshotDescr" +
|
||||||
"hotDoneRequest\022&\n\010snapshot\030\001 \001(\0132\024.Snaps",
|
"iption\"F\n\034IsRestoreSnapshotDoneRequest\022&",
|
||||||
"hotDescription\"3\n\035IsRestoreSnapshotDoneR" +
|
"\n\010snapshot\030\001 \001(\0132\024.SnapshotDescription\"3" +
|
||||||
"esponse\022\022\n\004done\030\001 \001(\010:\004true2\305\016\n\022MasterAd" +
|
"\n\035IsRestoreSnapshotDoneResponse\022\022\n\004done\030" +
|
||||||
"minService\0222\n\taddColumn\022\021.AddColumnReque" +
|
"\001 \001(\010:\004true2\305\016\n\022MasterAdminService\0222\n\tAd" +
|
||||||
"st\032\022.AddColumnResponse\022;\n\014deleteColumn\022\024" +
|
"dColumn\022\021.AddColumnRequest\032\022.AddColumnRe" +
|
||||||
".DeleteColumnRequest\032\025.DeleteColumnRespo" +
|
"sponse\022;\n\014DeleteColumn\022\024.DeleteColumnReq" +
|
||||||
"nse\022;\n\014modifyColumn\022\024.ModifyColumnReques" +
|
"uest\032\025.DeleteColumnResponse\022;\n\014ModifyCol" +
|
||||||
"t\032\025.ModifyColumnResponse\0225\n\nmoveRegion\022\022" +
|
"umn\022\024.ModifyColumnRequest\032\025.ModifyColumn" +
|
||||||
".MoveRegionRequest\032\023.MoveRegionResponse\022" +
|
"Response\0225\n\nMoveRegion\022\022.MoveRegionReque" +
|
||||||
"Y\n\026dispatchMergingRegions\022\036.DispatchMerg" +
|
"st\032\023.MoveRegionResponse\022Y\n\026DispatchMergi" +
|
||||||
"ingRegionsRequest\032\037.DispatchMergingRegio",
|
"ngRegions\022\036.DispatchMergingRegionsReques",
|
||||||
"nsResponse\022;\n\014assignRegion\022\024.AssignRegio" +
|
"t\032\037.DispatchMergingRegionsResponse\022;\n\014As" +
|
||||||
"nRequest\032\025.AssignRegionResponse\022A\n\016unass" +
|
"signRegion\022\024.AssignRegionRequest\032\025.Assig" +
|
||||||
"ignRegion\022\026.UnassignRegionRequest\032\027.Unas" +
|
"nRegionResponse\022A\n\016UnassignRegion\022\026.Unas" +
|
||||||
"signRegionResponse\022>\n\rofflineRegion\022\025.Of" +
|
"signRegionRequest\032\027.UnassignRegionRespon" +
|
||||||
"flineRegionRequest\032\026.OfflineRegionRespon" +
|
"se\022>\n\rOfflineRegion\022\025.OfflineRegionReque" +
|
||||||
"se\0228\n\013deleteTable\022\023.DeleteTableRequest\032\024" +
|
"st\032\026.OfflineRegionResponse\0228\n\013DeleteTabl" +
|
||||||
".DeleteTableResponse\0228\n\013enableTable\022\023.En" +
|
"e\022\023.DeleteTableRequest\032\024.DeleteTableResp" +
|
||||||
"ableTableRequest\032\024.EnableTableResponse\022;" +
|
"onse\0228\n\013EnableTable\022\023.EnableTableRequest" +
|
||||||
"\n\014disableTable\022\024.DisableTableRequest\032\025.D" +
|
"\032\024.EnableTableResponse\022;\n\014DisableTable\022\024" +
|
||||||
"isableTableResponse\0228\n\013modifyTable\022\023.Mod",
|
".DisableTableRequest\032\025.DisableTableRespo",
|
||||||
"ifyTableRequest\032\024.ModifyTableResponse\0228\n" +
|
"nse\0228\n\013ModifyTable\022\023.ModifyTableRequest\032" +
|
||||||
"\013createTable\022\023.CreateTableRequest\032\024.Crea" +
|
"\024.ModifyTableResponse\0228\n\013CreateTable\022\023.C" +
|
||||||
"teTableResponse\022/\n\010shutdown\022\020.ShutdownRe" +
|
"reateTableRequest\032\024.CreateTableResponse\022" +
|
||||||
"quest\032\021.ShutdownResponse\0225\n\nstopMaster\022\022" +
|
"/\n\010Shutdown\022\020.ShutdownRequest\032\021.Shutdown" +
|
||||||
".StopMasterRequest\032\023.StopMasterResponse\022" +
|
"Response\0225\n\nStopMaster\022\022.StopMasterReque" +
|
||||||
",\n\007balance\022\017.BalanceRequest\032\020.BalanceRes" +
|
"st\032\023.StopMasterResponse\022,\n\007Balance\022\017.Bal" +
|
||||||
"ponse\022M\n\022setBalancerRunning\022\032.SetBalance" +
|
"anceRequest\032\020.BalanceResponse\022M\n\022SetBala" +
|
||||||
"rRunningRequest\032\033.SetBalancerRunningResp" +
|
"ncerRunning\022\032.SetBalancerRunningRequest\032" +
|
||||||
"onse\022;\n\016runCatalogScan\022\023.CatalogScanRequ" +
|
"\033.SetBalancerRunningResponse\022;\n\016RunCatal" +
|
||||||
"est\032\024.CatalogScanResponse\022S\n\024enableCatal",
|
"ogScan\022\023.CatalogScanRequest\032\024.CatalogSca",
|
||||||
"ogJanitor\022\034.EnableCatalogJanitorRequest\032" +
|
"nResponse\022S\n\024EnableCatalogJanitor\022\034.Enab" +
|
||||||
"\035.EnableCatalogJanitorResponse\022\\\n\027isCata" +
|
"leCatalogJanitorRequest\032\035.EnableCatalogJ" +
|
||||||
"logJanitorEnabled\022\037.IsCatalogJanitorEnab" +
|
"anitorResponse\022\\\n\027IsCatalogJanitorEnable" +
|
||||||
"ledRequest\032 .IsCatalogJanitorEnabledResp" +
|
"d\022\037.IsCatalogJanitorEnabledRequest\032 .IsC" +
|
||||||
"onse\022L\n\021execMasterService\022\032.CoprocessorS" +
|
"atalogJanitorEnabledResponse\022L\n\021ExecMast" +
|
||||||
"erviceRequest\032\033.CoprocessorServiceRespon" +
|
"erService\022\032.CoprocessorServiceRequest\032\033." +
|
||||||
"se\0227\n\010snapshot\022\024.TakeSnapshotRequest\032\025.T" +
|
"CoprocessorServiceResponse\0227\n\010Snapshot\022\024" +
|
||||||
"akeSnapshotResponse\022D\n\025getCompletedSnaps" +
|
".TakeSnapshotRequest\032\025.TakeSnapshotRespo" +
|
||||||
"hots\022\024.ListSnapshotRequest\032\025.ListSnapsho" +
|
"nse\022D\n\025GetCompletedSnapshots\022\024.ListSnaps" +
|
||||||
"tResponse\022A\n\016deleteSnapshot\022\026.DeleteSnap",
|
"hotRequest\032\025.ListSnapshotResponse\022A\n\016Del",
|
||||||
"shotRequest\032\027.DeleteSnapshotResponse\022A\n\016" +
|
"eteSnapshot\022\026.DeleteSnapshotRequest\032\027.De" +
|
||||||
"isSnapshotDone\022\026.IsSnapshotDoneRequest\032\027" +
|
"leteSnapshotResponse\022A\n\016IsSnapshotDone\022\026" +
|
||||||
".IsSnapshotDoneResponse\022D\n\017restoreSnapsh" +
|
".IsSnapshotDoneRequest\032\027.IsSnapshotDoneR" +
|
||||||
"ot\022\027.RestoreSnapshotRequest\032\030.RestoreSna" +
|
"esponse\022D\n\017RestoreSnapshot\022\027.RestoreSnap" +
|
||||||
"pshotResponse\022V\n\025isRestoreSnapshotDone\022\035" +
|
"shotRequest\032\030.RestoreSnapshotResponse\022V\n" +
|
||||||
".IsRestoreSnapshotDoneRequest\032\036.IsRestor" +
|
"\025IsRestoreSnapshotDone\022\035.IsRestoreSnapsh" +
|
||||||
"eSnapshotDoneResponse\022D\n\017isMasterRunning" +
|
"otDoneRequest\032\036.IsRestoreSnapshotDoneRes" +
|
||||||
"\022\027.IsMasterRunningRequest\032\030.IsMasterRunn" +
|
"ponse\022D\n\017IsMasterRunning\022\027.IsMasterRunni" +
|
||||||
"ingResponseBG\n*org.apache.hadoop.hbase.p" +
|
"ngRequest\032\030.IsMasterRunningResponseBG\n*o" +
|
||||||
"rotobuf.generatedB\021MasterAdminProtosH\001\210\001",
|
"rg.apache.hadoop.hbase.protobuf.generate",
|
||||||
"\001\240\001\001"
|
"dB\021MasterAdminProtosH\001\210\001\001\240\001\001"
|
||||||
};
|
};
|
||||||
com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
|
com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
|
||||||
new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() {
|
new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() {
|
||||||
|
|
|
@ -11,7 +11,7 @@ public final class MasterMonitorProtos {
|
||||||
public interface GetSchemaAlterStatusRequestOrBuilder
|
public interface GetSchemaAlterStatusRequestOrBuilder
|
||||||
extends com.google.protobuf.MessageOrBuilder {
|
extends com.google.protobuf.MessageOrBuilder {
|
||||||
|
|
||||||
// required bytes tableName = 1;
|
// required bytes table_name = 1;
|
||||||
boolean hasTableName();
|
boolean hasTableName();
|
||||||
com.google.protobuf.ByteString getTableName();
|
com.google.protobuf.ByteString getTableName();
|
||||||
}
|
}
|
||||||
|
@ -44,8 +44,8 @@ public final class MasterMonitorProtos {
|
||||||
}
|
}
|
||||||
|
|
||||||
private int bitField0_;
|
private int bitField0_;
|
||||||
// required bytes tableName = 1;
|
// required bytes table_name = 1;
|
||||||
public static final int TABLENAME_FIELD_NUMBER = 1;
|
public static final int TABLE_NAME_FIELD_NUMBER = 1;
|
||||||
private com.google.protobuf.ByteString tableName_;
|
private com.google.protobuf.ByteString tableName_;
|
||||||
public boolean hasTableName() {
|
public boolean hasTableName() {
|
||||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||||
|
@ -127,7 +127,7 @@ public final class MasterMonitorProtos {
|
||||||
int hash = 41;
|
int hash = 41;
|
||||||
hash = (19 * hash) + getDescriptorForType().hashCode();
|
hash = (19 * hash) + getDescriptorForType().hashCode();
|
||||||
if (hasTableName()) {
|
if (hasTableName()) {
|
||||||
hash = (37 * hash) + TABLENAME_FIELD_NUMBER;
|
hash = (37 * hash) + TABLE_NAME_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + getTableName().hashCode();
|
hash = (53 * hash) + getTableName().hashCode();
|
||||||
}
|
}
|
||||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||||
|
@ -355,7 +355,7 @@ public final class MasterMonitorProtos {
|
||||||
|
|
||||||
private int bitField0_;
|
private int bitField0_;
|
||||||
|
|
||||||
// required bytes tableName = 1;
|
// required bytes table_name = 1;
|
||||||
private com.google.protobuf.ByteString tableName_ = com.google.protobuf.ByteString.EMPTY;
|
private com.google.protobuf.ByteString tableName_ = com.google.protobuf.ByteString.EMPTY;
|
||||||
public boolean hasTableName() {
|
public boolean hasTableName() {
|
||||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||||
|
@ -393,11 +393,11 @@ public final class MasterMonitorProtos {
|
||||||
public interface GetSchemaAlterStatusResponseOrBuilder
|
public interface GetSchemaAlterStatusResponseOrBuilder
|
||||||
extends com.google.protobuf.MessageOrBuilder {
|
extends com.google.protobuf.MessageOrBuilder {
|
||||||
|
|
||||||
// optional uint32 yetToUpdateRegions = 1;
|
// optional uint32 yet_to_update_regions = 1;
|
||||||
boolean hasYetToUpdateRegions();
|
boolean hasYetToUpdateRegions();
|
||||||
int getYetToUpdateRegions();
|
int getYetToUpdateRegions();
|
||||||
|
|
||||||
// optional uint32 totalRegions = 2;
|
// optional uint32 total_regions = 2;
|
||||||
boolean hasTotalRegions();
|
boolean hasTotalRegions();
|
||||||
int getTotalRegions();
|
int getTotalRegions();
|
||||||
}
|
}
|
||||||
|
@ -430,8 +430,8 @@ public final class MasterMonitorProtos {
|
||||||
}
|
}
|
||||||
|
|
||||||
private int bitField0_;
|
private int bitField0_;
|
||||||
// optional uint32 yetToUpdateRegions = 1;
|
// optional uint32 yet_to_update_regions = 1;
|
||||||
public static final int YETTOUPDATEREGIONS_FIELD_NUMBER = 1;
|
public static final int YET_TO_UPDATE_REGIONS_FIELD_NUMBER = 1;
|
||||||
private int yetToUpdateRegions_;
|
private int yetToUpdateRegions_;
|
||||||
public boolean hasYetToUpdateRegions() {
|
public boolean hasYetToUpdateRegions() {
|
||||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||||
|
@ -440,8 +440,8 @@ public final class MasterMonitorProtos {
|
||||||
return yetToUpdateRegions_;
|
return yetToUpdateRegions_;
|
||||||
}
|
}
|
||||||
|
|
||||||
// optional uint32 totalRegions = 2;
|
// optional uint32 total_regions = 2;
|
||||||
public static final int TOTALREGIONS_FIELD_NUMBER = 2;
|
public static final int TOTAL_REGIONS_FIELD_NUMBER = 2;
|
||||||
private int totalRegions_;
|
private int totalRegions_;
|
||||||
public boolean hasTotalRegions() {
|
public boolean hasTotalRegions() {
|
||||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||||
|
@ -532,11 +532,11 @@ public final class MasterMonitorProtos {
|
||||||
int hash = 41;
|
int hash = 41;
|
||||||
hash = (19 * hash) + getDescriptorForType().hashCode();
|
hash = (19 * hash) + getDescriptorForType().hashCode();
|
||||||
if (hasYetToUpdateRegions()) {
|
if (hasYetToUpdateRegions()) {
|
||||||
hash = (37 * hash) + YETTOUPDATEREGIONS_FIELD_NUMBER;
|
hash = (37 * hash) + YET_TO_UPDATE_REGIONS_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + getYetToUpdateRegions();
|
hash = (53 * hash) + getYetToUpdateRegions();
|
||||||
}
|
}
|
||||||
if (hasTotalRegions()) {
|
if (hasTotalRegions()) {
|
||||||
hash = (37 * hash) + TOTALREGIONS_FIELD_NUMBER;
|
hash = (37 * hash) + TOTAL_REGIONS_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + getTotalRegions();
|
hash = (53 * hash) + getTotalRegions();
|
||||||
}
|
}
|
||||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||||
|
@ -774,7 +774,7 @@ public final class MasterMonitorProtos {
|
||||||
|
|
||||||
private int bitField0_;
|
private int bitField0_;
|
||||||
|
|
||||||
// optional uint32 yetToUpdateRegions = 1;
|
// optional uint32 yet_to_update_regions = 1;
|
||||||
private int yetToUpdateRegions_ ;
|
private int yetToUpdateRegions_ ;
|
||||||
public boolean hasYetToUpdateRegions() {
|
public boolean hasYetToUpdateRegions() {
|
||||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||||
|
@ -795,7 +795,7 @@ public final class MasterMonitorProtos {
|
||||||
return this;
|
return this;
|
||||||
}
|
}
|
||||||
|
|
||||||
// optional uint32 totalRegions = 2;
|
// optional uint32 total_regions = 2;
|
||||||
private int totalRegions_ ;
|
private int totalRegions_ ;
|
||||||
public boolean hasTotalRegions() {
|
public boolean hasTotalRegions() {
|
||||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||||
|
@ -830,7 +830,7 @@ public final class MasterMonitorProtos {
|
||||||
public interface GetTableDescriptorsRequestOrBuilder
|
public interface GetTableDescriptorsRequestOrBuilder
|
||||||
extends com.google.protobuf.MessageOrBuilder {
|
extends com.google.protobuf.MessageOrBuilder {
|
||||||
|
|
||||||
// repeated string tableNames = 1;
|
// repeated string table_names = 1;
|
||||||
java.util.List<String> getTableNamesList();
|
java.util.List<String> getTableNamesList();
|
||||||
int getTableNamesCount();
|
int getTableNamesCount();
|
||||||
String getTableNames(int index);
|
String getTableNames(int index);
|
||||||
|
@ -863,8 +863,8 @@ public final class MasterMonitorProtos {
|
||||||
return org.apache.hadoop.hbase.protobuf.generated.MasterMonitorProtos.internal_static_GetTableDescriptorsRequest_fieldAccessorTable;
|
return org.apache.hadoop.hbase.protobuf.generated.MasterMonitorProtos.internal_static_GetTableDescriptorsRequest_fieldAccessorTable;
|
||||||
}
|
}
|
||||||
|
|
||||||
// repeated string tableNames = 1;
|
// repeated string table_names = 1;
|
||||||
public static final int TABLENAMES_FIELD_NUMBER = 1;
|
public static final int TABLE_NAMES_FIELD_NUMBER = 1;
|
||||||
private com.google.protobuf.LazyStringList tableNames_;
|
private com.google.protobuf.LazyStringList tableNames_;
|
||||||
public java.util.List<String>
|
public java.util.List<String>
|
||||||
getTableNamesList() {
|
getTableNamesList() {
|
||||||
|
@ -948,7 +948,7 @@ public final class MasterMonitorProtos {
|
||||||
int hash = 41;
|
int hash = 41;
|
||||||
hash = (19 * hash) + getDescriptorForType().hashCode();
|
hash = (19 * hash) + getDescriptorForType().hashCode();
|
||||||
if (getTableNamesCount() > 0) {
|
if (getTableNamesCount() > 0) {
|
||||||
hash = (37 * hash) + TABLENAMES_FIELD_NUMBER;
|
hash = (37 * hash) + TABLE_NAMES_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + getTableNamesList().hashCode();
|
hash = (53 * hash) + getTableNamesList().hashCode();
|
||||||
}
|
}
|
||||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||||
|
@ -1179,7 +1179,7 @@ public final class MasterMonitorProtos {
|
||||||
|
|
||||||
private int bitField0_;
|
private int bitField0_;
|
||||||
|
|
||||||
// repeated string tableNames = 1;
|
// repeated string table_names = 1;
|
||||||
private com.google.protobuf.LazyStringList tableNames_ = com.google.protobuf.LazyStringArrayList.EMPTY;
|
private com.google.protobuf.LazyStringList tableNames_ = com.google.protobuf.LazyStringArrayList.EMPTY;
|
||||||
private void ensureTableNamesIsMutable() {
|
private void ensureTableNamesIsMutable() {
|
||||||
if (!((bitField0_ & 0x00000001) == 0x00000001)) {
|
if (!((bitField0_ & 0x00000001) == 0x00000001)) {
|
||||||
|
@ -1249,7 +1249,7 @@ public final class MasterMonitorProtos {
|
||||||
public interface GetTableDescriptorsResponseOrBuilder
|
public interface GetTableDescriptorsResponseOrBuilder
|
||||||
extends com.google.protobuf.MessageOrBuilder {
|
extends com.google.protobuf.MessageOrBuilder {
|
||||||
|
|
||||||
// repeated .TableSchema tableSchema = 1;
|
// repeated .TableSchema table_schema = 1;
|
||||||
java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema>
|
java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema>
|
||||||
getTableSchemaList();
|
getTableSchemaList();
|
||||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema getTableSchema(int index);
|
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema getTableSchema(int index);
|
||||||
|
@ -1287,8 +1287,8 @@ public final class MasterMonitorProtos {
|
||||||
return org.apache.hadoop.hbase.protobuf.generated.MasterMonitorProtos.internal_static_GetTableDescriptorsResponse_fieldAccessorTable;
|
return org.apache.hadoop.hbase.protobuf.generated.MasterMonitorProtos.internal_static_GetTableDescriptorsResponse_fieldAccessorTable;
|
||||||
}
|
}
|
||||||
|
|
||||||
// repeated .TableSchema tableSchema = 1;
|
// repeated .TableSchema table_schema = 1;
|
||||||
public static final int TABLESCHEMA_FIELD_NUMBER = 1;
|
public static final int TABLE_SCHEMA_FIELD_NUMBER = 1;
|
||||||
private java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema> tableSchema_;
|
private java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema> tableSchema_;
|
||||||
public java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema> getTableSchemaList() {
|
public java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema> getTableSchemaList() {
|
||||||
return tableSchema_;
|
return tableSchema_;
|
||||||
|
@ -1380,7 +1380,7 @@ public final class MasterMonitorProtos {
|
||||||
int hash = 41;
|
int hash = 41;
|
||||||
hash = (19 * hash) + getDescriptorForType().hashCode();
|
hash = (19 * hash) + getDescriptorForType().hashCode();
|
||||||
if (getTableSchemaCount() > 0) {
|
if (getTableSchemaCount() > 0) {
|
||||||
hash = (37 * hash) + TABLESCHEMA_FIELD_NUMBER;
|
hash = (37 * hash) + TABLE_SCHEMA_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + getTableSchemaList().hashCode();
|
hash = (53 * hash) + getTableSchemaList().hashCode();
|
||||||
}
|
}
|
||||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||||
|
@ -1642,7 +1642,7 @@ public final class MasterMonitorProtos {
|
||||||
|
|
||||||
private int bitField0_;
|
private int bitField0_;
|
||||||
|
|
||||||
// repeated .TableSchema tableSchema = 1;
|
// repeated .TableSchema table_schema = 1;
|
||||||
private java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema> tableSchema_ =
|
private java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema> tableSchema_ =
|
||||||
java.util.Collections.emptyList();
|
java.util.Collections.emptyList();
|
||||||
private void ensureTableSchemaIsMutable() {
|
private void ensureTableSchemaIsMutable() {
|
||||||
|
@ -2142,7 +2142,7 @@ public final class MasterMonitorProtos {
|
||||||
public interface GetClusterStatusResponseOrBuilder
|
public interface GetClusterStatusResponseOrBuilder
|
||||||
extends com.google.protobuf.MessageOrBuilder {
|
extends com.google.protobuf.MessageOrBuilder {
|
||||||
|
|
||||||
// required .ClusterStatus clusterStatus = 1;
|
// required .ClusterStatus cluster_status = 1;
|
||||||
boolean hasClusterStatus();
|
boolean hasClusterStatus();
|
||||||
org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.ClusterStatus getClusterStatus();
|
org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.ClusterStatus getClusterStatus();
|
||||||
org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.ClusterStatusOrBuilder getClusterStatusOrBuilder();
|
org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.ClusterStatusOrBuilder getClusterStatusOrBuilder();
|
||||||
|
@ -2176,8 +2176,8 @@ public final class MasterMonitorProtos {
|
||||||
}
|
}
|
||||||
|
|
||||||
private int bitField0_;
|
private int bitField0_;
|
||||||
// required .ClusterStatus clusterStatus = 1;
|
// required .ClusterStatus cluster_status = 1;
|
||||||
public static final int CLUSTERSTATUS_FIELD_NUMBER = 1;
|
public static final int CLUSTER_STATUS_FIELD_NUMBER = 1;
|
||||||
private org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.ClusterStatus clusterStatus_;
|
private org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.ClusterStatus clusterStatus_;
|
||||||
public boolean hasClusterStatus() {
|
public boolean hasClusterStatus() {
|
||||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||||
|
@ -2266,7 +2266,7 @@ public final class MasterMonitorProtos {
|
||||||
int hash = 41;
|
int hash = 41;
|
||||||
hash = (19 * hash) + getDescriptorForType().hashCode();
|
hash = (19 * hash) + getDescriptorForType().hashCode();
|
||||||
if (hasClusterStatus()) {
|
if (hasClusterStatus()) {
|
||||||
hash = (37 * hash) + CLUSTERSTATUS_FIELD_NUMBER;
|
hash = (37 * hash) + CLUSTER_STATUS_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + getClusterStatus().hashCode();
|
hash = (53 * hash) + getClusterStatus().hashCode();
|
||||||
}
|
}
|
||||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||||
|
@ -2511,7 +2511,7 @@ public final class MasterMonitorProtos {
|
||||||
|
|
||||||
private int bitField0_;
|
private int bitField0_;
|
||||||
|
|
||||||
// required .ClusterStatus clusterStatus = 1;
|
// required .ClusterStatus cluster_status = 1;
|
||||||
private org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.ClusterStatus clusterStatus_ = org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.ClusterStatus.getDefaultInstance();
|
private org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.ClusterStatus clusterStatus_ = org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.ClusterStatus.getDefaultInstance();
|
||||||
private com.google.protobuf.SingleFieldBuilder<
|
private com.google.protobuf.SingleFieldBuilder<
|
||||||
org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.ClusterStatus, org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.ClusterStatus.Builder, org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.ClusterStatusOrBuilder> clusterStatusBuilder_;
|
org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.ClusterStatus, org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.ClusterStatus.Builder, org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.ClusterStatusOrBuilder> clusterStatusBuilder_;
|
||||||
|
@ -3068,26 +3068,26 @@ public final class MasterMonitorProtos {
|
||||||
static {
|
static {
|
||||||
java.lang.String[] descriptorData = {
|
java.lang.String[] descriptorData = {
|
||||||
"\n\023MasterMonitor.proto\032\014Master.proto\032\013hba" +
|
"\n\023MasterMonitor.proto\032\014Master.proto\032\013hba" +
|
||||||
"se.proto\032\023ClusterStatus.proto\"0\n\033GetSche" +
|
"se.proto\032\023ClusterStatus.proto\"1\n\033GetSche" +
|
||||||
"maAlterStatusRequest\022\021\n\ttableName\030\001 \002(\014\"" +
|
"maAlterStatusRequest\022\022\n\ntable_name\030\001 \002(\014" +
|
||||||
"P\n\034GetSchemaAlterStatusResponse\022\032\n\022yetTo" +
|
"\"T\n\034GetSchemaAlterStatusResponse\022\035\n\025yet_" +
|
||||||
"UpdateRegions\030\001 \001(\r\022\024\n\014totalRegions\030\002 \001(" +
|
"to_update_regions\030\001 \001(\r\022\025\n\rtotal_regions" +
|
||||||
"\r\"0\n\032GetTableDescriptorsRequest\022\022\n\ntable" +
|
"\030\002 \001(\r\"1\n\032GetTableDescriptorsRequest\022\023\n\013" +
|
||||||
"Names\030\001 \003(\t\"@\n\033GetTableDescriptorsRespon" +
|
"table_names\030\001 \003(\t\"A\n\033GetTableDescriptors" +
|
||||||
"se\022!\n\013tableSchema\030\001 \003(\0132\014.TableSchema\"\031\n" +
|
"Response\022\"\n\014table_schema\030\001 \003(\0132\014.TableSc" +
|
||||||
"\027GetClusterStatusRequest\"A\n\030GetClusterSt" +
|
"hema\"\031\n\027GetClusterStatusRequest\"B\n\030GetCl" +
|
||||||
"atusResponse\022%\n\rclusterStatus\030\001 \002(\0132\016.Cl",
|
"usterStatusResponse\022&\n\016cluster_status\030\001 ",
|
||||||
"usterStatus2\314\002\n\024MasterMonitorService\022S\n\024" +
|
"\002(\0132\016.ClusterStatus2\314\002\n\024MasterMonitorSer" +
|
||||||
"getSchemaAlterStatus\022\034.GetSchemaAlterSta" +
|
"vice\022S\n\024GetSchemaAlterStatus\022\034.GetSchema" +
|
||||||
"tusRequest\032\035.GetSchemaAlterStatusRespons" +
|
"AlterStatusRequest\032\035.GetSchemaAlterStatu" +
|
||||||
"e\022P\n\023getTableDescriptors\022\033.GetTableDescr" +
|
"sResponse\022P\n\023GetTableDescriptors\022\033.GetTa" +
|
||||||
"iptorsRequest\032\034.GetTableDescriptorsRespo" +
|
"bleDescriptorsRequest\032\034.GetTableDescript" +
|
||||||
"nse\022G\n\020getClusterStatus\022\030.GetClusterStat" +
|
"orsResponse\022G\n\020GetClusterStatus\022\030.GetClu" +
|
||||||
"usRequest\032\031.GetClusterStatusResponse\022D\n\017" +
|
"sterStatusRequest\032\031.GetClusterStatusResp" +
|
||||||
"isMasterRunning\022\027.IsMasterRunningRequest" +
|
"onse\022D\n\017IsMasterRunning\022\027.IsMasterRunnin" +
|
||||||
"\032\030.IsMasterRunningResponseBI\n*org.apache" +
|
"gRequest\032\030.IsMasterRunningResponseBI\n*or" +
|
||||||
".hadoop.hbase.protobuf.generatedB\023Master",
|
"g.apache.hadoop.hbase.protobuf.generated",
|
||||||
"MonitorProtosH\001\210\001\001\240\001\001"
|
"B\023MasterMonitorProtosH\001\210\001\001\240\001\001"
|
||||||
};
|
};
|
||||||
com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
|
com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
|
||||||
new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() {
|
new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() {
|
||||||
|
|
|
@ -311,7 +311,7 @@ public final class MasterProtos {
|
||||||
public interface IsMasterRunningResponseOrBuilder
|
public interface IsMasterRunningResponseOrBuilder
|
||||||
extends com.google.protobuf.MessageOrBuilder {
|
extends com.google.protobuf.MessageOrBuilder {
|
||||||
|
|
||||||
// required bool isMasterRunning = 1;
|
// required bool is_master_running = 1;
|
||||||
boolean hasIsMasterRunning();
|
boolean hasIsMasterRunning();
|
||||||
boolean getIsMasterRunning();
|
boolean getIsMasterRunning();
|
||||||
}
|
}
|
||||||
|
@ -344,8 +344,8 @@ public final class MasterProtos {
|
||||||
}
|
}
|
||||||
|
|
||||||
private int bitField0_;
|
private int bitField0_;
|
||||||
// required bool isMasterRunning = 1;
|
// required bool is_master_running = 1;
|
||||||
public static final int ISMASTERRUNNING_FIELD_NUMBER = 1;
|
public static final int IS_MASTER_RUNNING_FIELD_NUMBER = 1;
|
||||||
private boolean isMasterRunning_;
|
private boolean isMasterRunning_;
|
||||||
public boolean hasIsMasterRunning() {
|
public boolean hasIsMasterRunning() {
|
||||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||||
|
@ -427,7 +427,7 @@ public final class MasterProtos {
|
||||||
int hash = 41;
|
int hash = 41;
|
||||||
hash = (19 * hash) + getDescriptorForType().hashCode();
|
hash = (19 * hash) + getDescriptorForType().hashCode();
|
||||||
if (hasIsMasterRunning()) {
|
if (hasIsMasterRunning()) {
|
||||||
hash = (37 * hash) + ISMASTERRUNNING_FIELD_NUMBER;
|
hash = (37 * hash) + IS_MASTER_RUNNING_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + hashBoolean(getIsMasterRunning());
|
hash = (53 * hash) + hashBoolean(getIsMasterRunning());
|
||||||
}
|
}
|
||||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||||
|
@ -655,7 +655,7 @@ public final class MasterProtos {
|
||||||
|
|
||||||
private int bitField0_;
|
private int bitField0_;
|
||||||
|
|
||||||
// required bool isMasterRunning = 1;
|
// required bool is_master_running = 1;
|
||||||
private boolean isMasterRunning_ ;
|
private boolean isMasterRunning_ ;
|
||||||
public boolean hasIsMasterRunning() {
|
public boolean hasIsMasterRunning() {
|
||||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||||
|
@ -928,12 +928,12 @@ public final class MasterProtos {
|
||||||
static {
|
static {
|
||||||
java.lang.String[] descriptorData = {
|
java.lang.String[] descriptorData = {
|
||||||
"\n\014Master.proto\"\030\n\026IsMasterRunningRequest" +
|
"\n\014Master.proto\"\030\n\026IsMasterRunningRequest" +
|
||||||
"\"2\n\027IsMasterRunningResponse\022\027\n\017isMasterR" +
|
"\"4\n\027IsMasterRunningResponse\022\031\n\021is_master" +
|
||||||
"unning\030\001 \002(\0102U\n\rMasterService\022D\n\017isMaste" +
|
"_running\030\001 \002(\0102U\n\rMasterService\022D\n\017IsMas" +
|
||||||
"rRunning\022\027.IsMasterRunningRequest\032\030.IsMa" +
|
"terRunning\022\027.IsMasterRunningRequest\032\030.Is" +
|
||||||
"sterRunningResponseBB\n*org.apache.hadoop" +
|
"MasterRunningResponseBB\n*org.apache.hado" +
|
||||||
".hbase.protobuf.generatedB\014MasterProtosH" +
|
"op.hbase.protobuf.generatedB\014MasterProto" +
|
||||||
"\001\210\001\001\240\001\001"
|
"sH\001\210\001\001\240\001\001"
|
||||||
};
|
};
|
||||||
com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
|
com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
|
||||||
new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() {
|
new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() {
|
||||||
|
|
|
@ -11,7 +11,7 @@ public final class MultiRowMutation {
|
||||||
public interface MultiMutateRequestOrBuilder
|
public interface MultiMutateRequestOrBuilder
|
||||||
extends com.google.protobuf.MessageOrBuilder {
|
extends com.google.protobuf.MessageOrBuilder {
|
||||||
|
|
||||||
// repeated .MutationProto mutationRequest = 1;
|
// repeated .MutationProto mutation_request = 1;
|
||||||
java.util.List<org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MutationProto>
|
java.util.List<org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MutationProto>
|
||||||
getMutationRequestList();
|
getMutationRequestList();
|
||||||
org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MutationProto getMutationRequest(int index);
|
org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MutationProto getMutationRequest(int index);
|
||||||
|
@ -49,8 +49,8 @@ public final class MultiRowMutation {
|
||||||
return org.apache.hadoop.hbase.protobuf.generated.MultiRowMutation.internal_static_MultiMutateRequest_fieldAccessorTable;
|
return org.apache.hadoop.hbase.protobuf.generated.MultiRowMutation.internal_static_MultiMutateRequest_fieldAccessorTable;
|
||||||
}
|
}
|
||||||
|
|
||||||
// repeated .MutationProto mutationRequest = 1;
|
// repeated .MutationProto mutation_request = 1;
|
||||||
public static final int MUTATIONREQUEST_FIELD_NUMBER = 1;
|
public static final int MUTATION_REQUEST_FIELD_NUMBER = 1;
|
||||||
private java.util.List<org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MutationProto> mutationRequest_;
|
private java.util.List<org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MutationProto> mutationRequest_;
|
||||||
public java.util.List<org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MutationProto> getMutationRequestList() {
|
public java.util.List<org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MutationProto> getMutationRequestList() {
|
||||||
return mutationRequest_;
|
return mutationRequest_;
|
||||||
|
@ -142,7 +142,7 @@ public final class MultiRowMutation {
|
||||||
int hash = 41;
|
int hash = 41;
|
||||||
hash = (19 * hash) + getDescriptorForType().hashCode();
|
hash = (19 * hash) + getDescriptorForType().hashCode();
|
||||||
if (getMutationRequestCount() > 0) {
|
if (getMutationRequestCount() > 0) {
|
||||||
hash = (37 * hash) + MUTATIONREQUEST_FIELD_NUMBER;
|
hash = (37 * hash) + MUTATION_REQUEST_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + getMutationRequestList().hashCode();
|
hash = (53 * hash) + getMutationRequestList().hashCode();
|
||||||
}
|
}
|
||||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||||
|
@ -404,7 +404,7 @@ public final class MultiRowMutation {
|
||||||
|
|
||||||
private int bitField0_;
|
private int bitField0_;
|
||||||
|
|
||||||
// repeated .MutationProto mutationRequest = 1;
|
// repeated .MutationProto mutation_request = 1;
|
||||||
private java.util.List<org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MutationProto> mutationRequest_ =
|
private java.util.List<org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MutationProto> mutationRequest_ =
|
||||||
java.util.Collections.emptyList();
|
java.util.Collections.emptyList();
|
||||||
private void ensureMutationRequestIsMutable() {
|
private void ensureMutationRequestIsMutable() {
|
||||||
|
@ -1141,13 +1141,14 @@ public final class MultiRowMutation {
|
||||||
descriptor;
|
descriptor;
|
||||||
static {
|
static {
|
||||||
java.lang.String[] descriptorData = {
|
java.lang.String[] descriptorData = {
|
||||||
"\n\026MultiRowMutation.proto\032\014Client.proto\"=" +
|
"\n\026MultiRowMutation.proto\032\014Client.proto\">" +
|
||||||
"\n\022MultiMutateRequest\022\'\n\017mutationRequest\030" +
|
"\n\022MultiMutateRequest\022(\n\020mutation_request" +
|
||||||
"\001 \003(\0132\016.MutationProto\"\025\n\023MultiMutateResp" +
|
"\030\001 \003(\0132\016.MutationProto\"\025\n\023MultiMutateRes" +
|
||||||
"onse2R\n\027MultiRowMutationService\0227\n\nmutat" +
|
"ponse2R\n\027MultiRowMutationService\0227\n\nMuta" +
|
||||||
"eRows\022\023.MultiMutateRequest\032\024.MultiMutate" +
|
"teRows\022\023.MultiMutateRequest\032\024.MultiMutat" +
|
||||||
"ResponseBF\n*org.apache.hadoop.hbase.prot" +
|
"eResponseBF\n*org.apache.hadoop.hbase.pro" +
|
||||||
"obuf.generatedB\020MultiRowMutationH\001\210\001\001\240\001\001"
|
"tobuf.generatedB\020MultiRowMutationH\001\210\001\001\240\001" +
|
||||||
|
"\001"
|
||||||
};
|
};
|
||||||
com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
|
com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
|
||||||
new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() {
|
new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() {
|
||||||
|
|
|
@ -11,11 +11,11 @@ public final class RPCProtos {
|
||||||
public interface UserInformationOrBuilder
|
public interface UserInformationOrBuilder
|
||||||
extends com.google.protobuf.MessageOrBuilder {
|
extends com.google.protobuf.MessageOrBuilder {
|
||||||
|
|
||||||
// required string effectiveUser = 1;
|
// required string effective_user = 1;
|
||||||
boolean hasEffectiveUser();
|
boolean hasEffectiveUser();
|
||||||
String getEffectiveUser();
|
String getEffectiveUser();
|
||||||
|
|
||||||
// optional string realUser = 2;
|
// optional string real_user = 2;
|
||||||
boolean hasRealUser();
|
boolean hasRealUser();
|
||||||
String getRealUser();
|
String getRealUser();
|
||||||
}
|
}
|
||||||
|
@ -48,8 +48,8 @@ public final class RPCProtos {
|
||||||
}
|
}
|
||||||
|
|
||||||
private int bitField0_;
|
private int bitField0_;
|
||||||
// required string effectiveUser = 1;
|
// required string effective_user = 1;
|
||||||
public static final int EFFECTIVEUSER_FIELD_NUMBER = 1;
|
public static final int EFFECTIVE_USER_FIELD_NUMBER = 1;
|
||||||
private java.lang.Object effectiveUser_;
|
private java.lang.Object effectiveUser_;
|
||||||
public boolean hasEffectiveUser() {
|
public boolean hasEffectiveUser() {
|
||||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||||
|
@ -80,8 +80,8 @@ public final class RPCProtos {
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// optional string realUser = 2;
|
// optional string real_user = 2;
|
||||||
public static final int REALUSER_FIELD_NUMBER = 2;
|
public static final int REAL_USER_FIELD_NUMBER = 2;
|
||||||
private java.lang.Object realUser_;
|
private java.lang.Object realUser_;
|
||||||
public boolean hasRealUser() {
|
public boolean hasRealUser() {
|
||||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||||
|
@ -198,11 +198,11 @@ public final class RPCProtos {
|
||||||
int hash = 41;
|
int hash = 41;
|
||||||
hash = (19 * hash) + getDescriptorForType().hashCode();
|
hash = (19 * hash) + getDescriptorForType().hashCode();
|
||||||
if (hasEffectiveUser()) {
|
if (hasEffectiveUser()) {
|
||||||
hash = (37 * hash) + EFFECTIVEUSER_FIELD_NUMBER;
|
hash = (37 * hash) + EFFECTIVE_USER_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + getEffectiveUser().hashCode();
|
hash = (53 * hash) + getEffectiveUser().hashCode();
|
||||||
}
|
}
|
||||||
if (hasRealUser()) {
|
if (hasRealUser()) {
|
||||||
hash = (37 * hash) + REALUSER_FIELD_NUMBER;
|
hash = (37 * hash) + REAL_USER_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + getRealUser().hashCode();
|
hash = (53 * hash) + getRealUser().hashCode();
|
||||||
}
|
}
|
||||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||||
|
@ -444,7 +444,7 @@ public final class RPCProtos {
|
||||||
|
|
||||||
private int bitField0_;
|
private int bitField0_;
|
||||||
|
|
||||||
// required string effectiveUser = 1;
|
// required string effective_user = 1;
|
||||||
private java.lang.Object effectiveUser_ = "";
|
private java.lang.Object effectiveUser_ = "";
|
||||||
public boolean hasEffectiveUser() {
|
public boolean hasEffectiveUser() {
|
||||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||||
|
@ -480,7 +480,7 @@ public final class RPCProtos {
|
||||||
onChanged();
|
onChanged();
|
||||||
}
|
}
|
||||||
|
|
||||||
// optional string realUser = 2;
|
// optional string real_user = 2;
|
||||||
private java.lang.Object realUser_ = "";
|
private java.lang.Object realUser_ = "";
|
||||||
public boolean hasRealUser() {
|
public boolean hasRealUser() {
|
||||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||||
|
@ -530,20 +530,20 @@ public final class RPCProtos {
|
||||||
public interface ConnectionHeaderOrBuilder
|
public interface ConnectionHeaderOrBuilder
|
||||||
extends com.google.protobuf.MessageOrBuilder {
|
extends com.google.protobuf.MessageOrBuilder {
|
||||||
|
|
||||||
// optional .UserInformation userInfo = 1;
|
// optional .UserInformation user_info = 1;
|
||||||
boolean hasUserInfo();
|
boolean hasUserInfo();
|
||||||
org.apache.hadoop.hbase.protobuf.generated.RPCProtos.UserInformation getUserInfo();
|
org.apache.hadoop.hbase.protobuf.generated.RPCProtos.UserInformation getUserInfo();
|
||||||
org.apache.hadoop.hbase.protobuf.generated.RPCProtos.UserInformationOrBuilder getUserInfoOrBuilder();
|
org.apache.hadoop.hbase.protobuf.generated.RPCProtos.UserInformationOrBuilder getUserInfoOrBuilder();
|
||||||
|
|
||||||
// optional string serviceName = 2;
|
// optional string service_name = 2;
|
||||||
boolean hasServiceName();
|
boolean hasServiceName();
|
||||||
String getServiceName();
|
String getServiceName();
|
||||||
|
|
||||||
// optional string cellBlockCodecClass = 3 [default = "org.apache.hadoop.hbase.codec.KeyValueCodec"];
|
// optional string cell_block_codec_class = 3 [default = "org.apache.hadoop.hbase.codec.KeyValueCodec"];
|
||||||
boolean hasCellBlockCodecClass();
|
boolean hasCellBlockCodecClass();
|
||||||
String getCellBlockCodecClass();
|
String getCellBlockCodecClass();
|
||||||
|
|
||||||
// optional string cellBlockCompressorClass = 4;
|
// optional string cell_block_compressor_class = 4;
|
||||||
boolean hasCellBlockCompressorClass();
|
boolean hasCellBlockCompressorClass();
|
||||||
String getCellBlockCompressorClass();
|
String getCellBlockCompressorClass();
|
||||||
}
|
}
|
||||||
|
@ -576,8 +576,8 @@ public final class RPCProtos {
|
||||||
}
|
}
|
||||||
|
|
||||||
private int bitField0_;
|
private int bitField0_;
|
||||||
// optional .UserInformation userInfo = 1;
|
// optional .UserInformation user_info = 1;
|
||||||
public static final int USERINFO_FIELD_NUMBER = 1;
|
public static final int USER_INFO_FIELD_NUMBER = 1;
|
||||||
private org.apache.hadoop.hbase.protobuf.generated.RPCProtos.UserInformation userInfo_;
|
private org.apache.hadoop.hbase.protobuf.generated.RPCProtos.UserInformation userInfo_;
|
||||||
public boolean hasUserInfo() {
|
public boolean hasUserInfo() {
|
||||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||||
|
@ -589,8 +589,8 @@ public final class RPCProtos {
|
||||||
return userInfo_;
|
return userInfo_;
|
||||||
}
|
}
|
||||||
|
|
||||||
// optional string serviceName = 2;
|
// optional string service_name = 2;
|
||||||
public static final int SERVICENAME_FIELD_NUMBER = 2;
|
public static final int SERVICE_NAME_FIELD_NUMBER = 2;
|
||||||
private java.lang.Object serviceName_;
|
private java.lang.Object serviceName_;
|
||||||
public boolean hasServiceName() {
|
public boolean hasServiceName() {
|
||||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||||
|
@ -621,8 +621,8 @@ public final class RPCProtos {
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// optional string cellBlockCodecClass = 3 [default = "org.apache.hadoop.hbase.codec.KeyValueCodec"];
|
// optional string cell_block_codec_class = 3 [default = "org.apache.hadoop.hbase.codec.KeyValueCodec"];
|
||||||
public static final int CELLBLOCKCODECCLASS_FIELD_NUMBER = 3;
|
public static final int CELL_BLOCK_CODEC_CLASS_FIELD_NUMBER = 3;
|
||||||
private java.lang.Object cellBlockCodecClass_;
|
private java.lang.Object cellBlockCodecClass_;
|
||||||
public boolean hasCellBlockCodecClass() {
|
public boolean hasCellBlockCodecClass() {
|
||||||
return ((bitField0_ & 0x00000004) == 0x00000004);
|
return ((bitField0_ & 0x00000004) == 0x00000004);
|
||||||
|
@ -653,8 +653,8 @@ public final class RPCProtos {
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// optional string cellBlockCompressorClass = 4;
|
// optional string cell_block_compressor_class = 4;
|
||||||
public static final int CELLBLOCKCOMPRESSORCLASS_FIELD_NUMBER = 4;
|
public static final int CELL_BLOCK_COMPRESSOR_CLASS_FIELD_NUMBER = 4;
|
||||||
private java.lang.Object cellBlockCompressorClass_;
|
private java.lang.Object cellBlockCompressorClass_;
|
||||||
public boolean hasCellBlockCompressorClass() {
|
public boolean hasCellBlockCompressorClass() {
|
||||||
return ((bitField0_ & 0x00000008) == 0x00000008);
|
return ((bitField0_ & 0x00000008) == 0x00000008);
|
||||||
|
@ -799,19 +799,19 @@ public final class RPCProtos {
|
||||||
int hash = 41;
|
int hash = 41;
|
||||||
hash = (19 * hash) + getDescriptorForType().hashCode();
|
hash = (19 * hash) + getDescriptorForType().hashCode();
|
||||||
if (hasUserInfo()) {
|
if (hasUserInfo()) {
|
||||||
hash = (37 * hash) + USERINFO_FIELD_NUMBER;
|
hash = (37 * hash) + USER_INFO_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + getUserInfo().hashCode();
|
hash = (53 * hash) + getUserInfo().hashCode();
|
||||||
}
|
}
|
||||||
if (hasServiceName()) {
|
if (hasServiceName()) {
|
||||||
hash = (37 * hash) + SERVICENAME_FIELD_NUMBER;
|
hash = (37 * hash) + SERVICE_NAME_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + getServiceName().hashCode();
|
hash = (53 * hash) + getServiceName().hashCode();
|
||||||
}
|
}
|
||||||
if (hasCellBlockCodecClass()) {
|
if (hasCellBlockCodecClass()) {
|
||||||
hash = (37 * hash) + CELLBLOCKCODECCLASS_FIELD_NUMBER;
|
hash = (37 * hash) + CELL_BLOCK_CODEC_CLASS_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + getCellBlockCodecClass().hashCode();
|
hash = (53 * hash) + getCellBlockCodecClass().hashCode();
|
||||||
}
|
}
|
||||||
if (hasCellBlockCompressorClass()) {
|
if (hasCellBlockCompressorClass()) {
|
||||||
hash = (37 * hash) + CELLBLOCKCOMPRESSORCLASS_FIELD_NUMBER;
|
hash = (37 * hash) + CELL_BLOCK_COMPRESSOR_CLASS_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + getCellBlockCompressorClass().hashCode();
|
hash = (53 * hash) + getCellBlockCompressorClass().hashCode();
|
||||||
}
|
}
|
||||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||||
|
@ -1096,7 +1096,7 @@ public final class RPCProtos {
|
||||||
|
|
||||||
private int bitField0_;
|
private int bitField0_;
|
||||||
|
|
||||||
// optional .UserInformation userInfo = 1;
|
// optional .UserInformation user_info = 1;
|
||||||
private org.apache.hadoop.hbase.protobuf.generated.RPCProtos.UserInformation userInfo_ = org.apache.hadoop.hbase.protobuf.generated.RPCProtos.UserInformation.getDefaultInstance();
|
private org.apache.hadoop.hbase.protobuf.generated.RPCProtos.UserInformation userInfo_ = org.apache.hadoop.hbase.protobuf.generated.RPCProtos.UserInformation.getDefaultInstance();
|
||||||
private com.google.protobuf.SingleFieldBuilder<
|
private com.google.protobuf.SingleFieldBuilder<
|
||||||
org.apache.hadoop.hbase.protobuf.generated.RPCProtos.UserInformation, org.apache.hadoop.hbase.protobuf.generated.RPCProtos.UserInformation.Builder, org.apache.hadoop.hbase.protobuf.generated.RPCProtos.UserInformationOrBuilder> userInfoBuilder_;
|
org.apache.hadoop.hbase.protobuf.generated.RPCProtos.UserInformation, org.apache.hadoop.hbase.protobuf.generated.RPCProtos.UserInformation.Builder, org.apache.hadoop.hbase.protobuf.generated.RPCProtos.UserInformationOrBuilder> userInfoBuilder_;
|
||||||
|
@ -1186,7 +1186,7 @@ public final class RPCProtos {
|
||||||
return userInfoBuilder_;
|
return userInfoBuilder_;
|
||||||
}
|
}
|
||||||
|
|
||||||
// optional string serviceName = 2;
|
// optional string service_name = 2;
|
||||||
private java.lang.Object serviceName_ = "";
|
private java.lang.Object serviceName_ = "";
|
||||||
public boolean hasServiceName() {
|
public boolean hasServiceName() {
|
||||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||||
|
@ -1222,7 +1222,7 @@ public final class RPCProtos {
|
||||||
onChanged();
|
onChanged();
|
||||||
}
|
}
|
||||||
|
|
||||||
// optional string cellBlockCodecClass = 3 [default = "org.apache.hadoop.hbase.codec.KeyValueCodec"];
|
// optional string cell_block_codec_class = 3 [default = "org.apache.hadoop.hbase.codec.KeyValueCodec"];
|
||||||
private java.lang.Object cellBlockCodecClass_ = "org.apache.hadoop.hbase.codec.KeyValueCodec";
|
private java.lang.Object cellBlockCodecClass_ = "org.apache.hadoop.hbase.codec.KeyValueCodec";
|
||||||
public boolean hasCellBlockCodecClass() {
|
public boolean hasCellBlockCodecClass() {
|
||||||
return ((bitField0_ & 0x00000004) == 0x00000004);
|
return ((bitField0_ & 0x00000004) == 0x00000004);
|
||||||
|
@ -1258,7 +1258,7 @@ public final class RPCProtos {
|
||||||
onChanged();
|
onChanged();
|
||||||
}
|
}
|
||||||
|
|
||||||
// optional string cellBlockCompressorClass = 4;
|
// optional string cell_block_compressor_class = 4;
|
||||||
private java.lang.Object cellBlockCompressorClass_ = "";
|
private java.lang.Object cellBlockCompressorClass_ = "";
|
||||||
public boolean hasCellBlockCompressorClass() {
|
public boolean hasCellBlockCompressorClass() {
|
||||||
return ((bitField0_ & 0x00000008) == 0x00000008);
|
return ((bitField0_ & 0x00000008) == 0x00000008);
|
||||||
|
@ -1679,11 +1679,11 @@ public final class RPCProtos {
|
||||||
public interface ExceptionResponseOrBuilder
|
public interface ExceptionResponseOrBuilder
|
||||||
extends com.google.protobuf.MessageOrBuilder {
|
extends com.google.protobuf.MessageOrBuilder {
|
||||||
|
|
||||||
// optional string exceptionClassName = 1;
|
// optional string exception_class_name = 1;
|
||||||
boolean hasExceptionClassName();
|
boolean hasExceptionClassName();
|
||||||
String getExceptionClassName();
|
String getExceptionClassName();
|
||||||
|
|
||||||
// optional string stackTrace = 2;
|
// optional string stack_trace = 2;
|
||||||
boolean hasStackTrace();
|
boolean hasStackTrace();
|
||||||
String getStackTrace();
|
String getStackTrace();
|
||||||
|
|
||||||
|
@ -1695,7 +1695,7 @@ public final class RPCProtos {
|
||||||
boolean hasPort();
|
boolean hasPort();
|
||||||
int getPort();
|
int getPort();
|
||||||
|
|
||||||
// optional bool doNotRetry = 5;
|
// optional bool do_not_retry = 5;
|
||||||
boolean hasDoNotRetry();
|
boolean hasDoNotRetry();
|
||||||
boolean getDoNotRetry();
|
boolean getDoNotRetry();
|
||||||
}
|
}
|
||||||
|
@ -1728,8 +1728,8 @@ public final class RPCProtos {
|
||||||
}
|
}
|
||||||
|
|
||||||
private int bitField0_;
|
private int bitField0_;
|
||||||
// optional string exceptionClassName = 1;
|
// optional string exception_class_name = 1;
|
||||||
public static final int EXCEPTIONCLASSNAME_FIELD_NUMBER = 1;
|
public static final int EXCEPTION_CLASS_NAME_FIELD_NUMBER = 1;
|
||||||
private java.lang.Object exceptionClassName_;
|
private java.lang.Object exceptionClassName_;
|
||||||
public boolean hasExceptionClassName() {
|
public boolean hasExceptionClassName() {
|
||||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||||
|
@ -1760,8 +1760,8 @@ public final class RPCProtos {
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// optional string stackTrace = 2;
|
// optional string stack_trace = 2;
|
||||||
public static final int STACKTRACE_FIELD_NUMBER = 2;
|
public static final int STACK_TRACE_FIELD_NUMBER = 2;
|
||||||
private java.lang.Object stackTrace_;
|
private java.lang.Object stackTrace_;
|
||||||
public boolean hasStackTrace() {
|
public boolean hasStackTrace() {
|
||||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||||
|
@ -1834,8 +1834,8 @@ public final class RPCProtos {
|
||||||
return port_;
|
return port_;
|
||||||
}
|
}
|
||||||
|
|
||||||
// optional bool doNotRetry = 5;
|
// optional bool do_not_retry = 5;
|
||||||
public static final int DONOTRETRY_FIELD_NUMBER = 5;
|
public static final int DO_NOT_RETRY_FIELD_NUMBER = 5;
|
||||||
private boolean doNotRetry_;
|
private boolean doNotRetry_;
|
||||||
public boolean hasDoNotRetry() {
|
public boolean hasDoNotRetry() {
|
||||||
return ((bitField0_ & 0x00000010) == 0x00000010);
|
return ((bitField0_ & 0x00000010) == 0x00000010);
|
||||||
|
@ -1965,11 +1965,11 @@ public final class RPCProtos {
|
||||||
int hash = 41;
|
int hash = 41;
|
||||||
hash = (19 * hash) + getDescriptorForType().hashCode();
|
hash = (19 * hash) + getDescriptorForType().hashCode();
|
||||||
if (hasExceptionClassName()) {
|
if (hasExceptionClassName()) {
|
||||||
hash = (37 * hash) + EXCEPTIONCLASSNAME_FIELD_NUMBER;
|
hash = (37 * hash) + EXCEPTION_CLASS_NAME_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + getExceptionClassName().hashCode();
|
hash = (53 * hash) + getExceptionClassName().hashCode();
|
||||||
}
|
}
|
||||||
if (hasStackTrace()) {
|
if (hasStackTrace()) {
|
||||||
hash = (37 * hash) + STACKTRACE_FIELD_NUMBER;
|
hash = (37 * hash) + STACK_TRACE_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + getStackTrace().hashCode();
|
hash = (53 * hash) + getStackTrace().hashCode();
|
||||||
}
|
}
|
||||||
if (hasHostname()) {
|
if (hasHostname()) {
|
||||||
|
@ -1981,7 +1981,7 @@ public final class RPCProtos {
|
||||||
hash = (53 * hash) + getPort();
|
hash = (53 * hash) + getPort();
|
||||||
}
|
}
|
||||||
if (hasDoNotRetry()) {
|
if (hasDoNotRetry()) {
|
||||||
hash = (37 * hash) + DONOTRETRY_FIELD_NUMBER;
|
hash = (37 * hash) + DO_NOT_RETRY_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + hashBoolean(getDoNotRetry());
|
hash = (53 * hash) + hashBoolean(getDoNotRetry());
|
||||||
}
|
}
|
||||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||||
|
@ -2261,7 +2261,7 @@ public final class RPCProtos {
|
||||||
|
|
||||||
private int bitField0_;
|
private int bitField0_;
|
||||||
|
|
||||||
// optional string exceptionClassName = 1;
|
// optional string exception_class_name = 1;
|
||||||
private java.lang.Object exceptionClassName_ = "";
|
private java.lang.Object exceptionClassName_ = "";
|
||||||
public boolean hasExceptionClassName() {
|
public boolean hasExceptionClassName() {
|
||||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||||
|
@ -2297,7 +2297,7 @@ public final class RPCProtos {
|
||||||
onChanged();
|
onChanged();
|
||||||
}
|
}
|
||||||
|
|
||||||
// optional string stackTrace = 2;
|
// optional string stack_trace = 2;
|
||||||
private java.lang.Object stackTrace_ = "";
|
private java.lang.Object stackTrace_ = "";
|
||||||
public boolean hasStackTrace() {
|
public boolean hasStackTrace() {
|
||||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||||
|
@ -2390,7 +2390,7 @@ public final class RPCProtos {
|
||||||
return this;
|
return this;
|
||||||
}
|
}
|
||||||
|
|
||||||
// optional bool doNotRetry = 5;
|
// optional bool do_not_retry = 5;
|
||||||
private boolean doNotRetry_ ;
|
private boolean doNotRetry_ ;
|
||||||
public boolean hasDoNotRetry() {
|
public boolean hasDoNotRetry() {
|
||||||
return ((bitField0_ & 0x00000010) == 0x00000010);
|
return ((bitField0_ & 0x00000010) == 0x00000010);
|
||||||
|
@ -2425,24 +2425,24 @@ public final class RPCProtos {
|
||||||
public interface RequestHeaderOrBuilder
|
public interface RequestHeaderOrBuilder
|
||||||
extends com.google.protobuf.MessageOrBuilder {
|
extends com.google.protobuf.MessageOrBuilder {
|
||||||
|
|
||||||
// optional uint32 callId = 1;
|
// optional uint32 call_id = 1;
|
||||||
boolean hasCallId();
|
boolean hasCallId();
|
||||||
int getCallId();
|
int getCallId();
|
||||||
|
|
||||||
// optional .RPCTInfo traceInfo = 2;
|
// optional .RPCTInfo trace_info = 2;
|
||||||
boolean hasTraceInfo();
|
boolean hasTraceInfo();
|
||||||
org.apache.hadoop.hbase.protobuf.generated.Tracing.RPCTInfo getTraceInfo();
|
org.apache.hadoop.hbase.protobuf.generated.Tracing.RPCTInfo getTraceInfo();
|
||||||
org.apache.hadoop.hbase.protobuf.generated.Tracing.RPCTInfoOrBuilder getTraceInfoOrBuilder();
|
org.apache.hadoop.hbase.protobuf.generated.Tracing.RPCTInfoOrBuilder getTraceInfoOrBuilder();
|
||||||
|
|
||||||
// optional string methodName = 3;
|
// optional string method_name = 3;
|
||||||
boolean hasMethodName();
|
boolean hasMethodName();
|
||||||
String getMethodName();
|
String getMethodName();
|
||||||
|
|
||||||
// optional bool requestParam = 4;
|
// optional bool request_param = 4;
|
||||||
boolean hasRequestParam();
|
boolean hasRequestParam();
|
||||||
boolean getRequestParam();
|
boolean getRequestParam();
|
||||||
|
|
||||||
// optional .CellBlockMeta cellBlockMeta = 5;
|
// optional .CellBlockMeta cell_block_meta = 5;
|
||||||
boolean hasCellBlockMeta();
|
boolean hasCellBlockMeta();
|
||||||
org.apache.hadoop.hbase.protobuf.generated.RPCProtos.CellBlockMeta getCellBlockMeta();
|
org.apache.hadoop.hbase.protobuf.generated.RPCProtos.CellBlockMeta getCellBlockMeta();
|
||||||
org.apache.hadoop.hbase.protobuf.generated.RPCProtos.CellBlockMetaOrBuilder getCellBlockMetaOrBuilder();
|
org.apache.hadoop.hbase.protobuf.generated.RPCProtos.CellBlockMetaOrBuilder getCellBlockMetaOrBuilder();
|
||||||
|
@ -2476,8 +2476,8 @@ public final class RPCProtos {
|
||||||
}
|
}
|
||||||
|
|
||||||
private int bitField0_;
|
private int bitField0_;
|
||||||
// optional uint32 callId = 1;
|
// optional uint32 call_id = 1;
|
||||||
public static final int CALLID_FIELD_NUMBER = 1;
|
public static final int CALL_ID_FIELD_NUMBER = 1;
|
||||||
private int callId_;
|
private int callId_;
|
||||||
public boolean hasCallId() {
|
public boolean hasCallId() {
|
||||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||||
|
@ -2486,8 +2486,8 @@ public final class RPCProtos {
|
||||||
return callId_;
|
return callId_;
|
||||||
}
|
}
|
||||||
|
|
||||||
// optional .RPCTInfo traceInfo = 2;
|
// optional .RPCTInfo trace_info = 2;
|
||||||
public static final int TRACEINFO_FIELD_NUMBER = 2;
|
public static final int TRACE_INFO_FIELD_NUMBER = 2;
|
||||||
private org.apache.hadoop.hbase.protobuf.generated.Tracing.RPCTInfo traceInfo_;
|
private org.apache.hadoop.hbase.protobuf.generated.Tracing.RPCTInfo traceInfo_;
|
||||||
public boolean hasTraceInfo() {
|
public boolean hasTraceInfo() {
|
||||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||||
|
@ -2499,8 +2499,8 @@ public final class RPCProtos {
|
||||||
return traceInfo_;
|
return traceInfo_;
|
||||||
}
|
}
|
||||||
|
|
||||||
// optional string methodName = 3;
|
// optional string method_name = 3;
|
||||||
public static final int METHODNAME_FIELD_NUMBER = 3;
|
public static final int METHOD_NAME_FIELD_NUMBER = 3;
|
||||||
private java.lang.Object methodName_;
|
private java.lang.Object methodName_;
|
||||||
public boolean hasMethodName() {
|
public boolean hasMethodName() {
|
||||||
return ((bitField0_ & 0x00000004) == 0x00000004);
|
return ((bitField0_ & 0x00000004) == 0x00000004);
|
||||||
|
@ -2531,8 +2531,8 @@ public final class RPCProtos {
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// optional bool requestParam = 4;
|
// optional bool request_param = 4;
|
||||||
public static final int REQUESTPARAM_FIELD_NUMBER = 4;
|
public static final int REQUEST_PARAM_FIELD_NUMBER = 4;
|
||||||
private boolean requestParam_;
|
private boolean requestParam_;
|
||||||
public boolean hasRequestParam() {
|
public boolean hasRequestParam() {
|
||||||
return ((bitField0_ & 0x00000008) == 0x00000008);
|
return ((bitField0_ & 0x00000008) == 0x00000008);
|
||||||
|
@ -2541,8 +2541,8 @@ public final class RPCProtos {
|
||||||
return requestParam_;
|
return requestParam_;
|
||||||
}
|
}
|
||||||
|
|
||||||
// optional .CellBlockMeta cellBlockMeta = 5;
|
// optional .CellBlockMeta cell_block_meta = 5;
|
||||||
public static final int CELLBLOCKMETA_FIELD_NUMBER = 5;
|
public static final int CELL_BLOCK_META_FIELD_NUMBER = 5;
|
||||||
private org.apache.hadoop.hbase.protobuf.generated.RPCProtos.CellBlockMeta cellBlockMeta_;
|
private org.apache.hadoop.hbase.protobuf.generated.RPCProtos.CellBlockMeta cellBlockMeta_;
|
||||||
public boolean hasCellBlockMeta() {
|
public boolean hasCellBlockMeta() {
|
||||||
return ((bitField0_ & 0x00000010) == 0x00000010);
|
return ((bitField0_ & 0x00000010) == 0x00000010);
|
||||||
|
@ -2675,23 +2675,23 @@ public final class RPCProtos {
|
||||||
int hash = 41;
|
int hash = 41;
|
||||||
hash = (19 * hash) + getDescriptorForType().hashCode();
|
hash = (19 * hash) + getDescriptorForType().hashCode();
|
||||||
if (hasCallId()) {
|
if (hasCallId()) {
|
||||||
hash = (37 * hash) + CALLID_FIELD_NUMBER;
|
hash = (37 * hash) + CALL_ID_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + getCallId();
|
hash = (53 * hash) + getCallId();
|
||||||
}
|
}
|
||||||
if (hasTraceInfo()) {
|
if (hasTraceInfo()) {
|
||||||
hash = (37 * hash) + TRACEINFO_FIELD_NUMBER;
|
hash = (37 * hash) + TRACE_INFO_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + getTraceInfo().hashCode();
|
hash = (53 * hash) + getTraceInfo().hashCode();
|
||||||
}
|
}
|
||||||
if (hasMethodName()) {
|
if (hasMethodName()) {
|
||||||
hash = (37 * hash) + METHODNAME_FIELD_NUMBER;
|
hash = (37 * hash) + METHOD_NAME_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + getMethodName().hashCode();
|
hash = (53 * hash) + getMethodName().hashCode();
|
||||||
}
|
}
|
||||||
if (hasRequestParam()) {
|
if (hasRequestParam()) {
|
||||||
hash = (37 * hash) + REQUESTPARAM_FIELD_NUMBER;
|
hash = (37 * hash) + REQUEST_PARAM_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + hashBoolean(getRequestParam());
|
hash = (53 * hash) + hashBoolean(getRequestParam());
|
||||||
}
|
}
|
||||||
if (hasCellBlockMeta()) {
|
if (hasCellBlockMeta()) {
|
||||||
hash = (37 * hash) + CELLBLOCKMETA_FIELD_NUMBER;
|
hash = (37 * hash) + CELL_BLOCK_META_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + getCellBlockMeta().hashCode();
|
hash = (53 * hash) + getCellBlockMeta().hashCode();
|
||||||
}
|
}
|
||||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||||
|
@ -2997,7 +2997,7 @@ public final class RPCProtos {
|
||||||
|
|
||||||
private int bitField0_;
|
private int bitField0_;
|
||||||
|
|
||||||
// optional uint32 callId = 1;
|
// optional uint32 call_id = 1;
|
||||||
private int callId_ ;
|
private int callId_ ;
|
||||||
public boolean hasCallId() {
|
public boolean hasCallId() {
|
||||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||||
|
@ -3018,7 +3018,7 @@ public final class RPCProtos {
|
||||||
return this;
|
return this;
|
||||||
}
|
}
|
||||||
|
|
||||||
// optional .RPCTInfo traceInfo = 2;
|
// optional .RPCTInfo trace_info = 2;
|
||||||
private org.apache.hadoop.hbase.protobuf.generated.Tracing.RPCTInfo traceInfo_ = org.apache.hadoop.hbase.protobuf.generated.Tracing.RPCTInfo.getDefaultInstance();
|
private org.apache.hadoop.hbase.protobuf.generated.Tracing.RPCTInfo traceInfo_ = org.apache.hadoop.hbase.protobuf.generated.Tracing.RPCTInfo.getDefaultInstance();
|
||||||
private com.google.protobuf.SingleFieldBuilder<
|
private com.google.protobuf.SingleFieldBuilder<
|
||||||
org.apache.hadoop.hbase.protobuf.generated.Tracing.RPCTInfo, org.apache.hadoop.hbase.protobuf.generated.Tracing.RPCTInfo.Builder, org.apache.hadoop.hbase.protobuf.generated.Tracing.RPCTInfoOrBuilder> traceInfoBuilder_;
|
org.apache.hadoop.hbase.protobuf.generated.Tracing.RPCTInfo, org.apache.hadoop.hbase.protobuf.generated.Tracing.RPCTInfo.Builder, org.apache.hadoop.hbase.protobuf.generated.Tracing.RPCTInfoOrBuilder> traceInfoBuilder_;
|
||||||
|
@ -3108,7 +3108,7 @@ public final class RPCProtos {
|
||||||
return traceInfoBuilder_;
|
return traceInfoBuilder_;
|
||||||
}
|
}
|
||||||
|
|
||||||
// optional string methodName = 3;
|
// optional string method_name = 3;
|
||||||
private java.lang.Object methodName_ = "";
|
private java.lang.Object methodName_ = "";
|
||||||
public boolean hasMethodName() {
|
public boolean hasMethodName() {
|
||||||
return ((bitField0_ & 0x00000004) == 0x00000004);
|
return ((bitField0_ & 0x00000004) == 0x00000004);
|
||||||
|
@ -3144,7 +3144,7 @@ public final class RPCProtos {
|
||||||
onChanged();
|
onChanged();
|
||||||
}
|
}
|
||||||
|
|
||||||
// optional bool requestParam = 4;
|
// optional bool request_param = 4;
|
||||||
private boolean requestParam_ ;
|
private boolean requestParam_ ;
|
||||||
public boolean hasRequestParam() {
|
public boolean hasRequestParam() {
|
||||||
return ((bitField0_ & 0x00000008) == 0x00000008);
|
return ((bitField0_ & 0x00000008) == 0x00000008);
|
||||||
|
@ -3165,7 +3165,7 @@ public final class RPCProtos {
|
||||||
return this;
|
return this;
|
||||||
}
|
}
|
||||||
|
|
||||||
// optional .CellBlockMeta cellBlockMeta = 5;
|
// optional .CellBlockMeta cell_block_meta = 5;
|
||||||
private org.apache.hadoop.hbase.protobuf.generated.RPCProtos.CellBlockMeta cellBlockMeta_ = org.apache.hadoop.hbase.protobuf.generated.RPCProtos.CellBlockMeta.getDefaultInstance();
|
private org.apache.hadoop.hbase.protobuf.generated.RPCProtos.CellBlockMeta cellBlockMeta_ = org.apache.hadoop.hbase.protobuf.generated.RPCProtos.CellBlockMeta.getDefaultInstance();
|
||||||
private com.google.protobuf.SingleFieldBuilder<
|
private com.google.protobuf.SingleFieldBuilder<
|
||||||
org.apache.hadoop.hbase.protobuf.generated.RPCProtos.CellBlockMeta, org.apache.hadoop.hbase.protobuf.generated.RPCProtos.CellBlockMeta.Builder, org.apache.hadoop.hbase.protobuf.generated.RPCProtos.CellBlockMetaOrBuilder> cellBlockMetaBuilder_;
|
org.apache.hadoop.hbase.protobuf.generated.RPCProtos.CellBlockMeta, org.apache.hadoop.hbase.protobuf.generated.RPCProtos.CellBlockMeta.Builder, org.apache.hadoop.hbase.protobuf.generated.RPCProtos.CellBlockMetaOrBuilder> cellBlockMetaBuilder_;
|
||||||
|
@ -3269,7 +3269,7 @@ public final class RPCProtos {
|
||||||
public interface ResponseHeaderOrBuilder
|
public interface ResponseHeaderOrBuilder
|
||||||
extends com.google.protobuf.MessageOrBuilder {
|
extends com.google.protobuf.MessageOrBuilder {
|
||||||
|
|
||||||
// optional uint32 callId = 1;
|
// optional uint32 call_id = 1;
|
||||||
boolean hasCallId();
|
boolean hasCallId();
|
||||||
int getCallId();
|
int getCallId();
|
||||||
|
|
||||||
|
@ -3278,7 +3278,7 @@ public final class RPCProtos {
|
||||||
org.apache.hadoop.hbase.protobuf.generated.RPCProtos.ExceptionResponse getException();
|
org.apache.hadoop.hbase.protobuf.generated.RPCProtos.ExceptionResponse getException();
|
||||||
org.apache.hadoop.hbase.protobuf.generated.RPCProtos.ExceptionResponseOrBuilder getExceptionOrBuilder();
|
org.apache.hadoop.hbase.protobuf.generated.RPCProtos.ExceptionResponseOrBuilder getExceptionOrBuilder();
|
||||||
|
|
||||||
// optional .CellBlockMeta cellBlockMeta = 3;
|
// optional .CellBlockMeta cell_block_meta = 3;
|
||||||
boolean hasCellBlockMeta();
|
boolean hasCellBlockMeta();
|
||||||
org.apache.hadoop.hbase.protobuf.generated.RPCProtos.CellBlockMeta getCellBlockMeta();
|
org.apache.hadoop.hbase.protobuf.generated.RPCProtos.CellBlockMeta getCellBlockMeta();
|
||||||
org.apache.hadoop.hbase.protobuf.generated.RPCProtos.CellBlockMetaOrBuilder getCellBlockMetaOrBuilder();
|
org.apache.hadoop.hbase.protobuf.generated.RPCProtos.CellBlockMetaOrBuilder getCellBlockMetaOrBuilder();
|
||||||
|
@ -3312,8 +3312,8 @@ public final class RPCProtos {
|
||||||
}
|
}
|
||||||
|
|
||||||
private int bitField0_;
|
private int bitField0_;
|
||||||
// optional uint32 callId = 1;
|
// optional uint32 call_id = 1;
|
||||||
public static final int CALLID_FIELD_NUMBER = 1;
|
public static final int CALL_ID_FIELD_NUMBER = 1;
|
||||||
private int callId_;
|
private int callId_;
|
||||||
public boolean hasCallId() {
|
public boolean hasCallId() {
|
||||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||||
|
@ -3335,8 +3335,8 @@ public final class RPCProtos {
|
||||||
return exception_;
|
return exception_;
|
||||||
}
|
}
|
||||||
|
|
||||||
// optional .CellBlockMeta cellBlockMeta = 3;
|
// optional .CellBlockMeta cell_block_meta = 3;
|
||||||
public static final int CELLBLOCKMETA_FIELD_NUMBER = 3;
|
public static final int CELL_BLOCK_META_FIELD_NUMBER = 3;
|
||||||
private org.apache.hadoop.hbase.protobuf.generated.RPCProtos.CellBlockMeta cellBlockMeta_;
|
private org.apache.hadoop.hbase.protobuf.generated.RPCProtos.CellBlockMeta cellBlockMeta_;
|
||||||
public boolean hasCellBlockMeta() {
|
public boolean hasCellBlockMeta() {
|
||||||
return ((bitField0_ & 0x00000004) == 0x00000004);
|
return ((bitField0_ & 0x00000004) == 0x00000004);
|
||||||
|
@ -3443,7 +3443,7 @@ public final class RPCProtos {
|
||||||
int hash = 41;
|
int hash = 41;
|
||||||
hash = (19 * hash) + getDescriptorForType().hashCode();
|
hash = (19 * hash) + getDescriptorForType().hashCode();
|
||||||
if (hasCallId()) {
|
if (hasCallId()) {
|
||||||
hash = (37 * hash) + CALLID_FIELD_NUMBER;
|
hash = (37 * hash) + CALL_ID_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + getCallId();
|
hash = (53 * hash) + getCallId();
|
||||||
}
|
}
|
||||||
if (hasException()) {
|
if (hasException()) {
|
||||||
|
@ -3451,7 +3451,7 @@ public final class RPCProtos {
|
||||||
hash = (53 * hash) + getException().hashCode();
|
hash = (53 * hash) + getException().hashCode();
|
||||||
}
|
}
|
||||||
if (hasCellBlockMeta()) {
|
if (hasCellBlockMeta()) {
|
||||||
hash = (37 * hash) + CELLBLOCKMETA_FIELD_NUMBER;
|
hash = (37 * hash) + CELL_BLOCK_META_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + getCellBlockMeta().hashCode();
|
hash = (53 * hash) + getCellBlockMeta().hashCode();
|
||||||
}
|
}
|
||||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||||
|
@ -3729,7 +3729,7 @@ public final class RPCProtos {
|
||||||
|
|
||||||
private int bitField0_;
|
private int bitField0_;
|
||||||
|
|
||||||
// optional uint32 callId = 1;
|
// optional uint32 call_id = 1;
|
||||||
private int callId_ ;
|
private int callId_ ;
|
||||||
public boolean hasCallId() {
|
public boolean hasCallId() {
|
||||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||||
|
@ -3840,7 +3840,7 @@ public final class RPCProtos {
|
||||||
return exceptionBuilder_;
|
return exceptionBuilder_;
|
||||||
}
|
}
|
||||||
|
|
||||||
// optional .CellBlockMeta cellBlockMeta = 3;
|
// optional .CellBlockMeta cell_block_meta = 3;
|
||||||
private org.apache.hadoop.hbase.protobuf.generated.RPCProtos.CellBlockMeta cellBlockMeta_ = org.apache.hadoop.hbase.protobuf.generated.RPCProtos.CellBlockMeta.getDefaultInstance();
|
private org.apache.hadoop.hbase.protobuf.generated.RPCProtos.CellBlockMeta cellBlockMeta_ = org.apache.hadoop.hbase.protobuf.generated.RPCProtos.CellBlockMeta.getDefaultInstance();
|
||||||
private com.google.protobuf.SingleFieldBuilder<
|
private com.google.protobuf.SingleFieldBuilder<
|
||||||
org.apache.hadoop.hbase.protobuf.generated.RPCProtos.CellBlockMeta, org.apache.hadoop.hbase.protobuf.generated.RPCProtos.CellBlockMeta.Builder, org.apache.hadoop.hbase.protobuf.generated.RPCProtos.CellBlockMetaOrBuilder> cellBlockMetaBuilder_;
|
org.apache.hadoop.hbase.protobuf.generated.RPCProtos.CellBlockMeta, org.apache.hadoop.hbase.protobuf.generated.RPCProtos.CellBlockMeta.Builder, org.apache.hadoop.hbase.protobuf.generated.RPCProtos.CellBlockMetaOrBuilder> cellBlockMetaBuilder_;
|
||||||
|
@ -3981,24 +3981,25 @@ public final class RPCProtos {
|
||||||
static {
|
static {
|
||||||
java.lang.String[] descriptorData = {
|
java.lang.String[] descriptorData = {
|
||||||
"\n\tRPC.proto\032\rTracing.proto\032\013hbase.proto\"" +
|
"\n\tRPC.proto\032\rTracing.proto\032\013hbase.proto\"" +
|
||||||
":\n\017UserInformation\022\025\n\reffectiveUser\030\001 \002(" +
|
"<\n\017UserInformation\022\026\n\016effective_user\030\001 \002" +
|
||||||
"\t\022\020\n\010realUser\030\002 \001(\t\"\267\001\n\020ConnectionHeader" +
|
"(\t\022\021\n\treal_user\030\002 \001(\t\"\277\001\n\020ConnectionHead" +
|
||||||
"\022\"\n\010userInfo\030\001 \001(\0132\020.UserInformation\022\023\n\013" +
|
"er\022#\n\tuser_info\030\001 \001(\0132\020.UserInformation\022" +
|
||||||
"serviceName\030\002 \001(\t\022H\n\023cellBlockCodecClass" +
|
"\024\n\014service_name\030\002 \001(\t\022K\n\026cell_block_code" +
|
||||||
"\030\003 \001(\t:+org.apache.hadoop.hbase.codec.Ke" +
|
"c_class\030\003 \001(\t:+org.apache.hadoop.hbase.c" +
|
||||||
"yValueCodec\022 \n\030cellBlockCompressorClass\030" +
|
"odec.KeyValueCodec\022#\n\033cell_block_compres" +
|
||||||
"\004 \001(\t\"\037\n\rCellBlockMeta\022\016\n\006length\030\001 \001(\r\"w" +
|
"sor_class\030\004 \001(\t\"\037\n\rCellBlockMeta\022\016\n\006leng" +
|
||||||
"\n\021ExceptionResponse\022\032\n\022exceptionClassNam" +
|
"th\030\001 \001(\r\"|\n\021ExceptionResponse\022\034\n\024excepti" +
|
||||||
"e\030\001 \001(\t\022\022\n\nstackTrace\030\002 \001(\t\022\020\n\010hostname\030",
|
"on_class_name\030\001 \001(\t\022\023\n\013stack_trace\030\002 \001(\t",
|
||||||
"\003 \001(\t\022\014\n\004port\030\004 \001(\005\022\022\n\ndoNotRetry\030\005 \001(\010\"" +
|
"\022\020\n\010hostname\030\003 \001(\t\022\014\n\004port\030\004 \001(\005\022\024\n\014do_n" +
|
||||||
"\216\001\n\rRequestHeader\022\016\n\006callId\030\001 \001(\r\022\034\n\ttra" +
|
"ot_retry\030\005 \001(\010\"\224\001\n\rRequestHeader\022\017\n\007call" +
|
||||||
"ceInfo\030\002 \001(\0132\t.RPCTInfo\022\022\n\nmethodName\030\003 " +
|
"_id\030\001 \001(\r\022\035\n\ntrace_info\030\002 \001(\0132\t.RPCTInfo" +
|
||||||
"\001(\t\022\024\n\014requestParam\030\004 \001(\010\022%\n\rcellBlockMe" +
|
"\022\023\n\013method_name\030\003 \001(\t\022\025\n\rrequest_param\030\004" +
|
||||||
"ta\030\005 \001(\0132\016.CellBlockMeta\"n\n\016ResponseHead" +
|
" \001(\010\022\'\n\017cell_block_meta\030\005 \001(\0132\016.CellBloc" +
|
||||||
"er\022\016\n\006callId\030\001 \001(\r\022%\n\texception\030\002 \001(\0132\022." +
|
"kMeta\"q\n\016ResponseHeader\022\017\n\007call_id\030\001 \001(\r" +
|
||||||
"ExceptionResponse\022%\n\rcellBlockMeta\030\003 \001(\013" +
|
"\022%\n\texception\030\002 \001(\0132\022.ExceptionResponse\022" +
|
||||||
"2\016.CellBlockMetaB<\n*org.apache.hadoop.hb" +
|
"\'\n\017cell_block_meta\030\003 \001(\0132\016.CellBlockMeta" +
|
||||||
"ase.protobuf.generatedB\tRPCProtosH\001\240\001\001"
|
"B<\n*org.apache.hadoop.hbase.protobuf.gen" +
|
||||||
|
"eratedB\tRPCProtosH\001\240\001\001"
|
||||||
};
|
};
|
||||||
com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
|
com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
|
||||||
new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() {
|
new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() {
|
||||||
|
|
|
@ -15,11 +15,11 @@ public final class RegionServerStatusProtos {
|
||||||
boolean hasPort();
|
boolean hasPort();
|
||||||
int getPort();
|
int getPort();
|
||||||
|
|
||||||
// required uint64 serverStartCode = 2;
|
// required uint64 server_start_code = 2;
|
||||||
boolean hasServerStartCode();
|
boolean hasServerStartCode();
|
||||||
long getServerStartCode();
|
long getServerStartCode();
|
||||||
|
|
||||||
// required uint64 serverCurrentTime = 3;
|
// required uint64 server_current_time = 3;
|
||||||
boolean hasServerCurrentTime();
|
boolean hasServerCurrentTime();
|
||||||
long getServerCurrentTime();
|
long getServerCurrentTime();
|
||||||
}
|
}
|
||||||
|
@ -62,8 +62,8 @@ public final class RegionServerStatusProtos {
|
||||||
return port_;
|
return port_;
|
||||||
}
|
}
|
||||||
|
|
||||||
// required uint64 serverStartCode = 2;
|
// required uint64 server_start_code = 2;
|
||||||
public static final int SERVERSTARTCODE_FIELD_NUMBER = 2;
|
public static final int SERVER_START_CODE_FIELD_NUMBER = 2;
|
||||||
private long serverStartCode_;
|
private long serverStartCode_;
|
||||||
public boolean hasServerStartCode() {
|
public boolean hasServerStartCode() {
|
||||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||||
|
@ -72,8 +72,8 @@ public final class RegionServerStatusProtos {
|
||||||
return serverStartCode_;
|
return serverStartCode_;
|
||||||
}
|
}
|
||||||
|
|
||||||
// required uint64 serverCurrentTime = 3;
|
// required uint64 server_current_time = 3;
|
||||||
public static final int SERVERCURRENTTIME_FIELD_NUMBER = 3;
|
public static final int SERVER_CURRENT_TIME_FIELD_NUMBER = 3;
|
||||||
private long serverCurrentTime_;
|
private long serverCurrentTime_;
|
||||||
public boolean hasServerCurrentTime() {
|
public boolean hasServerCurrentTime() {
|
||||||
return ((bitField0_ & 0x00000004) == 0x00000004);
|
return ((bitField0_ & 0x00000004) == 0x00000004);
|
||||||
|
@ -193,11 +193,11 @@ public final class RegionServerStatusProtos {
|
||||||
hash = (53 * hash) + getPort();
|
hash = (53 * hash) + getPort();
|
||||||
}
|
}
|
||||||
if (hasServerStartCode()) {
|
if (hasServerStartCode()) {
|
||||||
hash = (37 * hash) + SERVERSTARTCODE_FIELD_NUMBER;
|
hash = (37 * hash) + SERVER_START_CODE_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + hashLong(getServerStartCode());
|
hash = (53 * hash) + hashLong(getServerStartCode());
|
||||||
}
|
}
|
||||||
if (hasServerCurrentTime()) {
|
if (hasServerCurrentTime()) {
|
||||||
hash = (37 * hash) + SERVERCURRENTTIME_FIELD_NUMBER;
|
hash = (37 * hash) + SERVER_CURRENT_TIME_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + hashLong(getServerCurrentTime());
|
hash = (53 * hash) + hashLong(getServerCurrentTime());
|
||||||
}
|
}
|
||||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||||
|
@ -482,7 +482,7 @@ public final class RegionServerStatusProtos {
|
||||||
return this;
|
return this;
|
||||||
}
|
}
|
||||||
|
|
||||||
// required uint64 serverStartCode = 2;
|
// required uint64 server_start_code = 2;
|
||||||
private long serverStartCode_ ;
|
private long serverStartCode_ ;
|
||||||
public boolean hasServerStartCode() {
|
public boolean hasServerStartCode() {
|
||||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||||
|
@ -503,7 +503,7 @@ public final class RegionServerStatusProtos {
|
||||||
return this;
|
return this;
|
||||||
}
|
}
|
||||||
|
|
||||||
// required uint64 serverCurrentTime = 3;
|
// required uint64 server_current_time = 3;
|
||||||
private long serverCurrentTime_ ;
|
private long serverCurrentTime_ ;
|
||||||
public boolean hasServerCurrentTime() {
|
public boolean hasServerCurrentTime() {
|
||||||
return ((bitField0_ & 0x00000004) == 0x00000004);
|
return ((bitField0_ & 0x00000004) == 0x00000004);
|
||||||
|
@ -538,7 +538,7 @@ public final class RegionServerStatusProtos {
|
||||||
public interface RegionServerStartupResponseOrBuilder
|
public interface RegionServerStartupResponseOrBuilder
|
||||||
extends com.google.protobuf.MessageOrBuilder {
|
extends com.google.protobuf.MessageOrBuilder {
|
||||||
|
|
||||||
// repeated .NameStringPair mapEntries = 1;
|
// repeated .NameStringPair map_entries = 1;
|
||||||
java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPair>
|
java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPair>
|
||||||
getMapEntriesList();
|
getMapEntriesList();
|
||||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPair getMapEntries(int index);
|
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPair getMapEntries(int index);
|
||||||
|
@ -576,8 +576,8 @@ public final class RegionServerStatusProtos {
|
||||||
return org.apache.hadoop.hbase.protobuf.generated.RegionServerStatusProtos.internal_static_RegionServerStartupResponse_fieldAccessorTable;
|
return org.apache.hadoop.hbase.protobuf.generated.RegionServerStatusProtos.internal_static_RegionServerStartupResponse_fieldAccessorTable;
|
||||||
}
|
}
|
||||||
|
|
||||||
// repeated .NameStringPair mapEntries = 1;
|
// repeated .NameStringPair map_entries = 1;
|
||||||
public static final int MAPENTRIES_FIELD_NUMBER = 1;
|
public static final int MAP_ENTRIES_FIELD_NUMBER = 1;
|
||||||
private java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPair> mapEntries_;
|
private java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPair> mapEntries_;
|
||||||
public java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPair> getMapEntriesList() {
|
public java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPair> getMapEntriesList() {
|
||||||
return mapEntries_;
|
return mapEntries_;
|
||||||
|
@ -669,7 +669,7 @@ public final class RegionServerStatusProtos {
|
||||||
int hash = 41;
|
int hash = 41;
|
||||||
hash = (19 * hash) + getDescriptorForType().hashCode();
|
hash = (19 * hash) + getDescriptorForType().hashCode();
|
||||||
if (getMapEntriesCount() > 0) {
|
if (getMapEntriesCount() > 0) {
|
||||||
hash = (37 * hash) + MAPENTRIES_FIELD_NUMBER;
|
hash = (37 * hash) + MAP_ENTRIES_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + getMapEntriesList().hashCode();
|
hash = (53 * hash) + getMapEntriesList().hashCode();
|
||||||
}
|
}
|
||||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||||
|
@ -931,7 +931,7 @@ public final class RegionServerStatusProtos {
|
||||||
|
|
||||||
private int bitField0_;
|
private int bitField0_;
|
||||||
|
|
||||||
// repeated .NameStringPair mapEntries = 1;
|
// repeated .NameStringPair map_entries = 1;
|
||||||
private java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPair> mapEntries_ =
|
private java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPair> mapEntries_ =
|
||||||
java.util.Collections.emptyList();
|
java.util.Collections.emptyList();
|
||||||
private void ensureMapEntriesIsMutable() {
|
private void ensureMapEntriesIsMutable() {
|
||||||
|
@ -2073,7 +2073,7 @@ public final class RegionServerStatusProtos {
|
||||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName getServer();
|
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName getServer();
|
||||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerNameOrBuilder getServerOrBuilder();
|
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerNameOrBuilder getServerOrBuilder();
|
||||||
|
|
||||||
// required string errorMessage = 2;
|
// required string error_message = 2;
|
||||||
boolean hasErrorMessage();
|
boolean hasErrorMessage();
|
||||||
String getErrorMessage();
|
String getErrorMessage();
|
||||||
}
|
}
|
||||||
|
@ -2119,8 +2119,8 @@ public final class RegionServerStatusProtos {
|
||||||
return server_;
|
return server_;
|
||||||
}
|
}
|
||||||
|
|
||||||
// required string errorMessage = 2;
|
// required string error_message = 2;
|
||||||
public static final int ERRORMESSAGE_FIELD_NUMBER = 2;
|
public static final int ERROR_MESSAGE_FIELD_NUMBER = 2;
|
||||||
private java.lang.Object errorMessage_;
|
private java.lang.Object errorMessage_;
|
||||||
public boolean hasErrorMessage() {
|
public boolean hasErrorMessage() {
|
||||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||||
|
@ -2249,7 +2249,7 @@ public final class RegionServerStatusProtos {
|
||||||
hash = (53 * hash) + getServer().hashCode();
|
hash = (53 * hash) + getServer().hashCode();
|
||||||
}
|
}
|
||||||
if (hasErrorMessage()) {
|
if (hasErrorMessage()) {
|
||||||
hash = (37 * hash) + ERRORMESSAGE_FIELD_NUMBER;
|
hash = (37 * hash) + ERROR_MESSAGE_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + getErrorMessage().hashCode();
|
hash = (53 * hash) + getErrorMessage().hashCode();
|
||||||
}
|
}
|
||||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||||
|
@ -2602,7 +2602,7 @@ public final class RegionServerStatusProtos {
|
||||||
return serverBuilder_;
|
return serverBuilder_;
|
||||||
}
|
}
|
||||||
|
|
||||||
// required string errorMessage = 2;
|
// required string error_message = 2;
|
||||||
private java.lang.Object errorMessage_ = "";
|
private java.lang.Object errorMessage_ = "";
|
||||||
public boolean hasErrorMessage() {
|
public boolean hasErrorMessage() {
|
||||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||||
|
@ -2952,7 +2952,7 @@ public final class RegionServerStatusProtos {
|
||||||
public interface GetLastFlushedSequenceIdRequestOrBuilder
|
public interface GetLastFlushedSequenceIdRequestOrBuilder
|
||||||
extends com.google.protobuf.MessageOrBuilder {
|
extends com.google.protobuf.MessageOrBuilder {
|
||||||
|
|
||||||
// required bytes regionName = 1;
|
// required bytes region_name = 1;
|
||||||
boolean hasRegionName();
|
boolean hasRegionName();
|
||||||
com.google.protobuf.ByteString getRegionName();
|
com.google.protobuf.ByteString getRegionName();
|
||||||
}
|
}
|
||||||
|
@ -2985,8 +2985,8 @@ public final class RegionServerStatusProtos {
|
||||||
}
|
}
|
||||||
|
|
||||||
private int bitField0_;
|
private int bitField0_;
|
||||||
// required bytes regionName = 1;
|
// required bytes region_name = 1;
|
||||||
public static final int REGIONNAME_FIELD_NUMBER = 1;
|
public static final int REGION_NAME_FIELD_NUMBER = 1;
|
||||||
private com.google.protobuf.ByteString regionName_;
|
private com.google.protobuf.ByteString regionName_;
|
||||||
public boolean hasRegionName() {
|
public boolean hasRegionName() {
|
||||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||||
|
@ -3068,7 +3068,7 @@ public final class RegionServerStatusProtos {
|
||||||
int hash = 41;
|
int hash = 41;
|
||||||
hash = (19 * hash) + getDescriptorForType().hashCode();
|
hash = (19 * hash) + getDescriptorForType().hashCode();
|
||||||
if (hasRegionName()) {
|
if (hasRegionName()) {
|
||||||
hash = (37 * hash) + REGIONNAME_FIELD_NUMBER;
|
hash = (37 * hash) + REGION_NAME_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + getRegionName().hashCode();
|
hash = (53 * hash) + getRegionName().hashCode();
|
||||||
}
|
}
|
||||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||||
|
@ -3296,7 +3296,7 @@ public final class RegionServerStatusProtos {
|
||||||
|
|
||||||
private int bitField0_;
|
private int bitField0_;
|
||||||
|
|
||||||
// required bytes regionName = 1;
|
// required bytes region_name = 1;
|
||||||
private com.google.protobuf.ByteString regionName_ = com.google.protobuf.ByteString.EMPTY;
|
private com.google.protobuf.ByteString regionName_ = com.google.protobuf.ByteString.EMPTY;
|
||||||
public boolean hasRegionName() {
|
public boolean hasRegionName() {
|
||||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||||
|
@ -3334,7 +3334,7 @@ public final class RegionServerStatusProtos {
|
||||||
public interface GetLastFlushedSequenceIdResponseOrBuilder
|
public interface GetLastFlushedSequenceIdResponseOrBuilder
|
||||||
extends com.google.protobuf.MessageOrBuilder {
|
extends com.google.protobuf.MessageOrBuilder {
|
||||||
|
|
||||||
// required uint64 lastFlushedSequenceId = 1;
|
// required uint64 last_flushed_sequence_id = 1;
|
||||||
boolean hasLastFlushedSequenceId();
|
boolean hasLastFlushedSequenceId();
|
||||||
long getLastFlushedSequenceId();
|
long getLastFlushedSequenceId();
|
||||||
}
|
}
|
||||||
|
@ -3367,8 +3367,8 @@ public final class RegionServerStatusProtos {
|
||||||
}
|
}
|
||||||
|
|
||||||
private int bitField0_;
|
private int bitField0_;
|
||||||
// required uint64 lastFlushedSequenceId = 1;
|
// required uint64 last_flushed_sequence_id = 1;
|
||||||
public static final int LASTFLUSHEDSEQUENCEID_FIELD_NUMBER = 1;
|
public static final int LAST_FLUSHED_SEQUENCE_ID_FIELD_NUMBER = 1;
|
||||||
private long lastFlushedSequenceId_;
|
private long lastFlushedSequenceId_;
|
||||||
public boolean hasLastFlushedSequenceId() {
|
public boolean hasLastFlushedSequenceId() {
|
||||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||||
|
@ -3450,7 +3450,7 @@ public final class RegionServerStatusProtos {
|
||||||
int hash = 41;
|
int hash = 41;
|
||||||
hash = (19 * hash) + getDescriptorForType().hashCode();
|
hash = (19 * hash) + getDescriptorForType().hashCode();
|
||||||
if (hasLastFlushedSequenceId()) {
|
if (hasLastFlushedSequenceId()) {
|
||||||
hash = (37 * hash) + LASTFLUSHEDSEQUENCEID_FIELD_NUMBER;
|
hash = (37 * hash) + LAST_FLUSHED_SEQUENCE_ID_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + hashLong(getLastFlushedSequenceId());
|
hash = (53 * hash) + hashLong(getLastFlushedSequenceId());
|
||||||
}
|
}
|
||||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||||
|
@ -3678,7 +3678,7 @@ public final class RegionServerStatusProtos {
|
||||||
|
|
||||||
private int bitField0_;
|
private int bitField0_;
|
||||||
|
|
||||||
// required uint64 lastFlushedSequenceId = 1;
|
// required uint64 last_flushed_sequence_id = 1;
|
||||||
private long lastFlushedSequenceId_ ;
|
private long lastFlushedSequenceId_ ;
|
||||||
public boolean hasLastFlushedSequenceId() {
|
public boolean hasLastFlushedSequenceId() {
|
||||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||||
|
@ -4176,31 +4176,31 @@ public final class RegionServerStatusProtos {
|
||||||
static {
|
static {
|
||||||
java.lang.String[] descriptorData = {
|
java.lang.String[] descriptorData = {
|
||||||
"\n\030RegionServerStatus.proto\032\013hbase.proto\032" +
|
"\n\030RegionServerStatus.proto\032\013hbase.proto\032" +
|
||||||
"\023ClusterStatus.proto\"^\n\032RegionServerStar" +
|
"\023ClusterStatus.proto\"b\n\032RegionServerStar" +
|
||||||
"tupRequest\022\014\n\004port\030\001 \002(\r\022\027\n\017serverStartC" +
|
"tupRequest\022\014\n\004port\030\001 \002(\r\022\031\n\021server_start" +
|
||||||
"ode\030\002 \002(\004\022\031\n\021serverCurrentTime\030\003 \002(\004\"B\n\033" +
|
"_code\030\002 \002(\004\022\033\n\023server_current_time\030\003 \002(\004" +
|
||||||
"RegionServerStartupResponse\022#\n\nmapEntrie" +
|
"\"C\n\033RegionServerStartupResponse\022$\n\013map_e" +
|
||||||
"s\030\001 \003(\0132\017.NameStringPair\"S\n\031RegionServer" +
|
"ntries\030\001 \003(\0132\017.NameStringPair\"S\n\031RegionS" +
|
||||||
"ReportRequest\022\033\n\006server\030\001 \002(\0132\013.ServerNa" +
|
"erverReportRequest\022\033\n\006server\030\001 \002(\0132\013.Ser" +
|
||||||
"me\022\031\n\004load\030\002 \001(\0132\013.ServerLoad\"\034\n\032RegionS" +
|
"verName\022\031\n\004load\030\002 \001(\0132\013.ServerLoad\"\034\n\032Re" +
|
||||||
"erverReportResponse\"N\n\031ReportRSFatalErro" +
|
"gionServerReportResponse\"O\n\031ReportRSFata" +
|
||||||
"rRequest\022\033\n\006server\030\001 \002(\0132\013.ServerName\022\024\n",
|
"lErrorRequest\022\033\n\006server\030\001 \002(\0132\013.ServerNa",
|
||||||
"\014errorMessage\030\002 \002(\t\"\034\n\032ReportRSFatalErro" +
|
"me\022\025\n\rerror_message\030\002 \002(\t\"\034\n\032ReportRSFat" +
|
||||||
"rResponse\"5\n\037GetLastFlushedSequenceIdReq" +
|
"alErrorResponse\"6\n\037GetLastFlushedSequenc" +
|
||||||
"uest\022\022\n\nregionName\030\001 \002(\014\"A\n GetLastFlush" +
|
"eIdRequest\022\023\n\013region_name\030\001 \002(\014\"D\n GetLa" +
|
||||||
"edSequenceIdResponse\022\035\n\025lastFlushedSeque" +
|
"stFlushedSequenceIdResponse\022 \n\030last_flus" +
|
||||||
"nceId\030\001 \002(\0042\354\002\n\031RegionServerStatusServic" +
|
"hed_sequence_id\030\001 \002(\0042\354\002\n\031RegionServerSt" +
|
||||||
"e\022P\n\023regionServerStartup\022\033.RegionServerS" +
|
"atusService\022P\n\023RegionServerStartup\022\033.Reg" +
|
||||||
"tartupRequest\032\034.RegionServerStartupRespo" +
|
"ionServerStartupRequest\032\034.RegionServerSt" +
|
||||||
"nse\022M\n\022regionServerReport\022\032.RegionServer" +
|
"artupResponse\022M\n\022RegionServerReport\022\032.Re" +
|
||||||
"ReportRequest\032\033.RegionServerReportRespon" +
|
"gionServerReportRequest\032\033.RegionServerRe" +
|
||||||
"se\022M\n\022reportRSFatalError\022\032.ReportRSFatal",
|
"portResponse\022M\n\022ReportRSFatalError\022\032.Rep",
|
||||||
"ErrorRequest\032\033.ReportRSFatalErrorRespons" +
|
"ortRSFatalErrorRequest\032\033.ReportRSFatalEr" +
|
||||||
"e\022_\n\030getLastFlushedSequenceId\022 .GetLastF" +
|
"rorResponse\022_\n\030GetLastFlushedSequenceId\022" +
|
||||||
"lushedSequenceIdRequest\032!.GetLastFlushed" +
|
" .GetLastFlushedSequenceIdRequest\032!.GetL" +
|
||||||
"SequenceIdResponseBN\n*org.apache.hadoop." +
|
"astFlushedSequenceIdResponseBN\n*org.apac" +
|
||||||
"hbase.protobuf.generatedB\030RegionServerSt" +
|
"he.hadoop.hbase.protobuf.generatedB\030Regi" +
|
||||||
"atusProtosH\001\210\001\001\240\001\001"
|
"onServerStatusProtosH\001\210\001\001\240\001\001"
|
||||||
};
|
};
|
||||||
com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
|
com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
|
||||||
new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() {
|
new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() {
|
||||||
|
|
|
@ -11,15 +11,15 @@ public final class RowProcessorProtos {
|
||||||
public interface RowProcessorRequestOrBuilder
|
public interface RowProcessorRequestOrBuilder
|
||||||
extends com.google.protobuf.MessageOrBuilder {
|
extends com.google.protobuf.MessageOrBuilder {
|
||||||
|
|
||||||
// required string rowProcessorClassName = 1;
|
// required string row_processor_class_name = 1;
|
||||||
boolean hasRowProcessorClassName();
|
boolean hasRowProcessorClassName();
|
||||||
String getRowProcessorClassName();
|
String getRowProcessorClassName();
|
||||||
|
|
||||||
// optional string rowProcessorInitializerMessageName = 2;
|
// optional string row_processor_initializer_message_name = 2;
|
||||||
boolean hasRowProcessorInitializerMessageName();
|
boolean hasRowProcessorInitializerMessageName();
|
||||||
String getRowProcessorInitializerMessageName();
|
String getRowProcessorInitializerMessageName();
|
||||||
|
|
||||||
// optional bytes rowProcessorInitializerMessage = 3;
|
// optional bytes row_processor_initializer_message = 3;
|
||||||
boolean hasRowProcessorInitializerMessage();
|
boolean hasRowProcessorInitializerMessage();
|
||||||
com.google.protobuf.ByteString getRowProcessorInitializerMessage();
|
com.google.protobuf.ByteString getRowProcessorInitializerMessage();
|
||||||
}
|
}
|
||||||
|
@ -52,8 +52,8 @@ public final class RowProcessorProtos {
|
||||||
}
|
}
|
||||||
|
|
||||||
private int bitField0_;
|
private int bitField0_;
|
||||||
// required string rowProcessorClassName = 1;
|
// required string row_processor_class_name = 1;
|
||||||
public static final int ROWPROCESSORCLASSNAME_FIELD_NUMBER = 1;
|
public static final int ROW_PROCESSOR_CLASS_NAME_FIELD_NUMBER = 1;
|
||||||
private java.lang.Object rowProcessorClassName_;
|
private java.lang.Object rowProcessorClassName_;
|
||||||
public boolean hasRowProcessorClassName() {
|
public boolean hasRowProcessorClassName() {
|
||||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||||
|
@ -84,8 +84,8 @@ public final class RowProcessorProtos {
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// optional string rowProcessorInitializerMessageName = 2;
|
// optional string row_processor_initializer_message_name = 2;
|
||||||
public static final int ROWPROCESSORINITIALIZERMESSAGENAME_FIELD_NUMBER = 2;
|
public static final int ROW_PROCESSOR_INITIALIZER_MESSAGE_NAME_FIELD_NUMBER = 2;
|
||||||
private java.lang.Object rowProcessorInitializerMessageName_;
|
private java.lang.Object rowProcessorInitializerMessageName_;
|
||||||
public boolean hasRowProcessorInitializerMessageName() {
|
public boolean hasRowProcessorInitializerMessageName() {
|
||||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||||
|
@ -116,8 +116,8 @@ public final class RowProcessorProtos {
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// optional bytes rowProcessorInitializerMessage = 3;
|
// optional bytes row_processor_initializer_message = 3;
|
||||||
public static final int ROWPROCESSORINITIALIZERMESSAGE_FIELD_NUMBER = 3;
|
public static final int ROW_PROCESSOR_INITIALIZER_MESSAGE_FIELD_NUMBER = 3;
|
||||||
private com.google.protobuf.ByteString rowProcessorInitializerMessage_;
|
private com.google.protobuf.ByteString rowProcessorInitializerMessage_;
|
||||||
public boolean hasRowProcessorInitializerMessage() {
|
public boolean hasRowProcessorInitializerMessage() {
|
||||||
return ((bitField0_ & 0x00000004) == 0x00000004);
|
return ((bitField0_ & 0x00000004) == 0x00000004);
|
||||||
|
@ -225,15 +225,15 @@ public final class RowProcessorProtos {
|
||||||
int hash = 41;
|
int hash = 41;
|
||||||
hash = (19 * hash) + getDescriptorForType().hashCode();
|
hash = (19 * hash) + getDescriptorForType().hashCode();
|
||||||
if (hasRowProcessorClassName()) {
|
if (hasRowProcessorClassName()) {
|
||||||
hash = (37 * hash) + ROWPROCESSORCLASSNAME_FIELD_NUMBER;
|
hash = (37 * hash) + ROW_PROCESSOR_CLASS_NAME_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + getRowProcessorClassName().hashCode();
|
hash = (53 * hash) + getRowProcessorClassName().hashCode();
|
||||||
}
|
}
|
||||||
if (hasRowProcessorInitializerMessageName()) {
|
if (hasRowProcessorInitializerMessageName()) {
|
||||||
hash = (37 * hash) + ROWPROCESSORINITIALIZERMESSAGENAME_FIELD_NUMBER;
|
hash = (37 * hash) + ROW_PROCESSOR_INITIALIZER_MESSAGE_NAME_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + getRowProcessorInitializerMessageName().hashCode();
|
hash = (53 * hash) + getRowProcessorInitializerMessageName().hashCode();
|
||||||
}
|
}
|
||||||
if (hasRowProcessorInitializerMessage()) {
|
if (hasRowProcessorInitializerMessage()) {
|
||||||
hash = (37 * hash) + ROWPROCESSORINITIALIZERMESSAGE_FIELD_NUMBER;
|
hash = (37 * hash) + ROW_PROCESSOR_INITIALIZER_MESSAGE_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + getRowProcessorInitializerMessage().hashCode();
|
hash = (53 * hash) + getRowProcessorInitializerMessage().hashCode();
|
||||||
}
|
}
|
||||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||||
|
@ -489,7 +489,7 @@ public final class RowProcessorProtos {
|
||||||
|
|
||||||
private int bitField0_;
|
private int bitField0_;
|
||||||
|
|
||||||
// required string rowProcessorClassName = 1;
|
// required string row_processor_class_name = 1;
|
||||||
private java.lang.Object rowProcessorClassName_ = "";
|
private java.lang.Object rowProcessorClassName_ = "";
|
||||||
public boolean hasRowProcessorClassName() {
|
public boolean hasRowProcessorClassName() {
|
||||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||||
|
@ -525,7 +525,7 @@ public final class RowProcessorProtos {
|
||||||
onChanged();
|
onChanged();
|
||||||
}
|
}
|
||||||
|
|
||||||
// optional string rowProcessorInitializerMessageName = 2;
|
// optional string row_processor_initializer_message_name = 2;
|
||||||
private java.lang.Object rowProcessorInitializerMessageName_ = "";
|
private java.lang.Object rowProcessorInitializerMessageName_ = "";
|
||||||
public boolean hasRowProcessorInitializerMessageName() {
|
public boolean hasRowProcessorInitializerMessageName() {
|
||||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||||
|
@ -561,7 +561,7 @@ public final class RowProcessorProtos {
|
||||||
onChanged();
|
onChanged();
|
||||||
}
|
}
|
||||||
|
|
||||||
// optional bytes rowProcessorInitializerMessage = 3;
|
// optional bytes row_processor_initializer_message = 3;
|
||||||
private com.google.protobuf.ByteString rowProcessorInitializerMessage_ = com.google.protobuf.ByteString.EMPTY;
|
private com.google.protobuf.ByteString rowProcessorInitializerMessage_ = com.google.protobuf.ByteString.EMPTY;
|
||||||
public boolean hasRowProcessorInitializerMessage() {
|
public boolean hasRowProcessorInitializerMessage() {
|
||||||
return ((bitField0_ & 0x00000004) == 0x00000004);
|
return ((bitField0_ & 0x00000004) == 0x00000004);
|
||||||
|
@ -599,7 +599,7 @@ public final class RowProcessorProtos {
|
||||||
public interface RowProcessorResultOrBuilder
|
public interface RowProcessorResultOrBuilder
|
||||||
extends com.google.protobuf.MessageOrBuilder {
|
extends com.google.protobuf.MessageOrBuilder {
|
||||||
|
|
||||||
// required bytes rowProcessorResult = 1;
|
// required bytes row_processor_result = 1;
|
||||||
boolean hasRowProcessorResult();
|
boolean hasRowProcessorResult();
|
||||||
com.google.protobuf.ByteString getRowProcessorResult();
|
com.google.protobuf.ByteString getRowProcessorResult();
|
||||||
}
|
}
|
||||||
|
@ -632,8 +632,8 @@ public final class RowProcessorProtos {
|
||||||
}
|
}
|
||||||
|
|
||||||
private int bitField0_;
|
private int bitField0_;
|
||||||
// required bytes rowProcessorResult = 1;
|
// required bytes row_processor_result = 1;
|
||||||
public static final int ROWPROCESSORRESULT_FIELD_NUMBER = 1;
|
public static final int ROW_PROCESSOR_RESULT_FIELD_NUMBER = 1;
|
||||||
private com.google.protobuf.ByteString rowProcessorResult_;
|
private com.google.protobuf.ByteString rowProcessorResult_;
|
||||||
public boolean hasRowProcessorResult() {
|
public boolean hasRowProcessorResult() {
|
||||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||||
|
@ -715,7 +715,7 @@ public final class RowProcessorProtos {
|
||||||
int hash = 41;
|
int hash = 41;
|
||||||
hash = (19 * hash) + getDescriptorForType().hashCode();
|
hash = (19 * hash) + getDescriptorForType().hashCode();
|
||||||
if (hasRowProcessorResult()) {
|
if (hasRowProcessorResult()) {
|
||||||
hash = (37 * hash) + ROWPROCESSORRESULT_FIELD_NUMBER;
|
hash = (37 * hash) + ROW_PROCESSOR_RESULT_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + getRowProcessorResult().hashCode();
|
hash = (53 * hash) + getRowProcessorResult().hashCode();
|
||||||
}
|
}
|
||||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||||
|
@ -943,7 +943,7 @@ public final class RowProcessorProtos {
|
||||||
|
|
||||||
private int bitField0_;
|
private int bitField0_;
|
||||||
|
|
||||||
// required bytes rowProcessorResult = 1;
|
// required bytes row_processor_result = 1;
|
||||||
private com.google.protobuf.ByteString rowProcessorResult_ = com.google.protobuf.ByteString.EMPTY;
|
private com.google.protobuf.ByteString rowProcessorResult_ = com.google.protobuf.ByteString.EMPTY;
|
||||||
public boolean hasRowProcessorResult() {
|
public boolean hasRowProcessorResult() {
|
||||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||||
|
@ -1218,15 +1218,16 @@ public final class RowProcessorProtos {
|
||||||
descriptor;
|
descriptor;
|
||||||
static {
|
static {
|
||||||
java.lang.String[] descriptorData = {
|
java.lang.String[] descriptorData = {
|
||||||
"\n\022RowProcessor.proto\"\210\001\n\023RowProcessorReq" +
|
"\n\022RowProcessor.proto\"\222\001\n\023RowProcessorReq" +
|
||||||
"uest\022\035\n\025rowProcessorClassName\030\001 \002(\t\022*\n\"r" +
|
"uest\022 \n\030row_processor_class_name\030\001 \002(\t\022." +
|
||||||
"owProcessorInitializerMessageName\030\002 \001(\t\022" +
|
"\n&row_processor_initializer_message_name" +
|
||||||
"&\n\036rowProcessorInitializerMessage\030\003 \001(\014\"" +
|
"\030\002 \001(\t\022)\n!row_processor_initializer_mess" +
|
||||||
"0\n\022RowProcessorResult\022\032\n\022rowProcessorRes" +
|
"age\030\003 \001(\014\"2\n\022RowProcessorResult\022\034\n\024row_p" +
|
||||||
"ult\030\001 \002(\0142K\n\023RowProcessorService\0224\n\007proc" +
|
"rocessor_result\030\001 \002(\0142K\n\023RowProcessorSer" +
|
||||||
"ess\022\024.RowProcessorRequest\032\023.RowProcessor" +
|
"vice\0224\n\007Process\022\024.RowProcessorRequest\032\023." +
|
||||||
"ResultBH\n*org.apache.hadoop.hbase.protob" +
|
"RowProcessorResultBH\n*org.apache.hadoop." +
|
||||||
"uf.generatedB\022RowProcessorProtosH\001\210\001\001\240\001\001"
|
"hbase.protobuf.generatedB\022RowProcessorPr" +
|
||||||
|
"otosH\001\210\001\001\240\001\001"
|
||||||
};
|
};
|
||||||
com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
|
com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
|
||||||
new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() {
|
new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() {
|
||||||
|
|
|
@ -11,7 +11,7 @@ public final class SecureBulkLoadProtos {
|
||||||
public interface SecureBulkLoadHFilesRequestOrBuilder
|
public interface SecureBulkLoadHFilesRequestOrBuilder
|
||||||
extends com.google.protobuf.MessageOrBuilder {
|
extends com.google.protobuf.MessageOrBuilder {
|
||||||
|
|
||||||
// repeated .BulkLoadHFileRequest.FamilyPath familyPath = 1;
|
// repeated .BulkLoadHFileRequest.FamilyPath family_path = 1;
|
||||||
java.util.List<org.apache.hadoop.hbase.protobuf.generated.ClientProtos.BulkLoadHFileRequest.FamilyPath>
|
java.util.List<org.apache.hadoop.hbase.protobuf.generated.ClientProtos.BulkLoadHFileRequest.FamilyPath>
|
||||||
getFamilyPathList();
|
getFamilyPathList();
|
||||||
org.apache.hadoop.hbase.protobuf.generated.ClientProtos.BulkLoadHFileRequest.FamilyPath getFamilyPath(int index);
|
org.apache.hadoop.hbase.protobuf.generated.ClientProtos.BulkLoadHFileRequest.FamilyPath getFamilyPath(int index);
|
||||||
|
@ -21,16 +21,16 @@ public final class SecureBulkLoadProtos {
|
||||||
org.apache.hadoop.hbase.protobuf.generated.ClientProtos.BulkLoadHFileRequest.FamilyPathOrBuilder getFamilyPathOrBuilder(
|
org.apache.hadoop.hbase.protobuf.generated.ClientProtos.BulkLoadHFileRequest.FamilyPathOrBuilder getFamilyPathOrBuilder(
|
||||||
int index);
|
int index);
|
||||||
|
|
||||||
// optional bool assignSeqNum = 2;
|
// optional bool assign_seq_num = 2;
|
||||||
boolean hasAssignSeqNum();
|
boolean hasAssignSeqNum();
|
||||||
boolean getAssignSeqNum();
|
boolean getAssignSeqNum();
|
||||||
|
|
||||||
// required .DelegationTokenProto fsToken = 3;
|
// required .DelegationTokenProto fs_token = 3;
|
||||||
boolean hasFsToken();
|
boolean hasFsToken();
|
||||||
org.apache.hadoop.hbase.protobuf.generated.SecureBulkLoadProtos.DelegationTokenProto getFsToken();
|
org.apache.hadoop.hbase.protobuf.generated.SecureBulkLoadProtos.DelegationTokenProto getFsToken();
|
||||||
org.apache.hadoop.hbase.protobuf.generated.SecureBulkLoadProtos.DelegationTokenProtoOrBuilder getFsTokenOrBuilder();
|
org.apache.hadoop.hbase.protobuf.generated.SecureBulkLoadProtos.DelegationTokenProtoOrBuilder getFsTokenOrBuilder();
|
||||||
|
|
||||||
// required string bulkToken = 4;
|
// required string bulk_token = 4;
|
||||||
boolean hasBulkToken();
|
boolean hasBulkToken();
|
||||||
String getBulkToken();
|
String getBulkToken();
|
||||||
}
|
}
|
||||||
|
@ -63,8 +63,8 @@ public final class SecureBulkLoadProtos {
|
||||||
}
|
}
|
||||||
|
|
||||||
private int bitField0_;
|
private int bitField0_;
|
||||||
// repeated .BulkLoadHFileRequest.FamilyPath familyPath = 1;
|
// repeated .BulkLoadHFileRequest.FamilyPath family_path = 1;
|
||||||
public static final int FAMILYPATH_FIELD_NUMBER = 1;
|
public static final int FAMILY_PATH_FIELD_NUMBER = 1;
|
||||||
private java.util.List<org.apache.hadoop.hbase.protobuf.generated.ClientProtos.BulkLoadHFileRequest.FamilyPath> familyPath_;
|
private java.util.List<org.apache.hadoop.hbase.protobuf.generated.ClientProtos.BulkLoadHFileRequest.FamilyPath> familyPath_;
|
||||||
public java.util.List<org.apache.hadoop.hbase.protobuf.generated.ClientProtos.BulkLoadHFileRequest.FamilyPath> getFamilyPathList() {
|
public java.util.List<org.apache.hadoop.hbase.protobuf.generated.ClientProtos.BulkLoadHFileRequest.FamilyPath> getFamilyPathList() {
|
||||||
return familyPath_;
|
return familyPath_;
|
||||||
|
@ -84,8 +84,8 @@ public final class SecureBulkLoadProtos {
|
||||||
return familyPath_.get(index);
|
return familyPath_.get(index);
|
||||||
}
|
}
|
||||||
|
|
||||||
// optional bool assignSeqNum = 2;
|
// optional bool assign_seq_num = 2;
|
||||||
public static final int ASSIGNSEQNUM_FIELD_NUMBER = 2;
|
public static final int ASSIGN_SEQ_NUM_FIELD_NUMBER = 2;
|
||||||
private boolean assignSeqNum_;
|
private boolean assignSeqNum_;
|
||||||
public boolean hasAssignSeqNum() {
|
public boolean hasAssignSeqNum() {
|
||||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||||
|
@ -94,8 +94,8 @@ public final class SecureBulkLoadProtos {
|
||||||
return assignSeqNum_;
|
return assignSeqNum_;
|
||||||
}
|
}
|
||||||
|
|
||||||
// required .DelegationTokenProto fsToken = 3;
|
// required .DelegationTokenProto fs_token = 3;
|
||||||
public static final int FSTOKEN_FIELD_NUMBER = 3;
|
public static final int FS_TOKEN_FIELD_NUMBER = 3;
|
||||||
private org.apache.hadoop.hbase.protobuf.generated.SecureBulkLoadProtos.DelegationTokenProto fsToken_;
|
private org.apache.hadoop.hbase.protobuf.generated.SecureBulkLoadProtos.DelegationTokenProto fsToken_;
|
||||||
public boolean hasFsToken() {
|
public boolean hasFsToken() {
|
||||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||||
|
@ -107,8 +107,8 @@ public final class SecureBulkLoadProtos {
|
||||||
return fsToken_;
|
return fsToken_;
|
||||||
}
|
}
|
||||||
|
|
||||||
// required string bulkToken = 4;
|
// required string bulk_token = 4;
|
||||||
public static final int BULKTOKEN_FIELD_NUMBER = 4;
|
public static final int BULK_TOKEN_FIELD_NUMBER = 4;
|
||||||
private java.lang.Object bulkToken_;
|
private java.lang.Object bulkToken_;
|
||||||
public boolean hasBulkToken() {
|
public boolean hasBulkToken() {
|
||||||
return ((bitField0_ & 0x00000004) == 0x00000004);
|
return ((bitField0_ & 0x00000004) == 0x00000004);
|
||||||
|
@ -258,19 +258,19 @@ public final class SecureBulkLoadProtos {
|
||||||
int hash = 41;
|
int hash = 41;
|
||||||
hash = (19 * hash) + getDescriptorForType().hashCode();
|
hash = (19 * hash) + getDescriptorForType().hashCode();
|
||||||
if (getFamilyPathCount() > 0) {
|
if (getFamilyPathCount() > 0) {
|
||||||
hash = (37 * hash) + FAMILYPATH_FIELD_NUMBER;
|
hash = (37 * hash) + FAMILY_PATH_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + getFamilyPathList().hashCode();
|
hash = (53 * hash) + getFamilyPathList().hashCode();
|
||||||
}
|
}
|
||||||
if (hasAssignSeqNum()) {
|
if (hasAssignSeqNum()) {
|
||||||
hash = (37 * hash) + ASSIGNSEQNUM_FIELD_NUMBER;
|
hash = (37 * hash) + ASSIGN_SEQ_NUM_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + hashBoolean(getAssignSeqNum());
|
hash = (53 * hash) + hashBoolean(getAssignSeqNum());
|
||||||
}
|
}
|
||||||
if (hasFsToken()) {
|
if (hasFsToken()) {
|
||||||
hash = (37 * hash) + FSTOKEN_FIELD_NUMBER;
|
hash = (37 * hash) + FS_TOKEN_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + getFsToken().hashCode();
|
hash = (53 * hash) + getFsToken().hashCode();
|
||||||
}
|
}
|
||||||
if (hasBulkToken()) {
|
if (hasBulkToken()) {
|
||||||
hash = (37 * hash) + BULKTOKEN_FIELD_NUMBER;
|
hash = (37 * hash) + BULK_TOKEN_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + getBulkToken().hashCode();
|
hash = (53 * hash) + getBulkToken().hashCode();
|
||||||
}
|
}
|
||||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||||
|
@ -597,7 +597,7 @@ public final class SecureBulkLoadProtos {
|
||||||
|
|
||||||
private int bitField0_;
|
private int bitField0_;
|
||||||
|
|
||||||
// repeated .BulkLoadHFileRequest.FamilyPath familyPath = 1;
|
// repeated .BulkLoadHFileRequest.FamilyPath family_path = 1;
|
||||||
private java.util.List<org.apache.hadoop.hbase.protobuf.generated.ClientProtos.BulkLoadHFileRequest.FamilyPath> familyPath_ =
|
private java.util.List<org.apache.hadoop.hbase.protobuf.generated.ClientProtos.BulkLoadHFileRequest.FamilyPath> familyPath_ =
|
||||||
java.util.Collections.emptyList();
|
java.util.Collections.emptyList();
|
||||||
private void ensureFamilyPathIsMutable() {
|
private void ensureFamilyPathIsMutable() {
|
||||||
|
@ -783,7 +783,7 @@ public final class SecureBulkLoadProtos {
|
||||||
return familyPathBuilder_;
|
return familyPathBuilder_;
|
||||||
}
|
}
|
||||||
|
|
||||||
// optional bool assignSeqNum = 2;
|
// optional bool assign_seq_num = 2;
|
||||||
private boolean assignSeqNum_ ;
|
private boolean assignSeqNum_ ;
|
||||||
public boolean hasAssignSeqNum() {
|
public boolean hasAssignSeqNum() {
|
||||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||||
|
@ -804,7 +804,7 @@ public final class SecureBulkLoadProtos {
|
||||||
return this;
|
return this;
|
||||||
}
|
}
|
||||||
|
|
||||||
// required .DelegationTokenProto fsToken = 3;
|
// required .DelegationTokenProto fs_token = 3;
|
||||||
private org.apache.hadoop.hbase.protobuf.generated.SecureBulkLoadProtos.DelegationTokenProto fsToken_ = org.apache.hadoop.hbase.protobuf.generated.SecureBulkLoadProtos.DelegationTokenProto.getDefaultInstance();
|
private org.apache.hadoop.hbase.protobuf.generated.SecureBulkLoadProtos.DelegationTokenProto fsToken_ = org.apache.hadoop.hbase.protobuf.generated.SecureBulkLoadProtos.DelegationTokenProto.getDefaultInstance();
|
||||||
private com.google.protobuf.SingleFieldBuilder<
|
private com.google.protobuf.SingleFieldBuilder<
|
||||||
org.apache.hadoop.hbase.protobuf.generated.SecureBulkLoadProtos.DelegationTokenProto, org.apache.hadoop.hbase.protobuf.generated.SecureBulkLoadProtos.DelegationTokenProto.Builder, org.apache.hadoop.hbase.protobuf.generated.SecureBulkLoadProtos.DelegationTokenProtoOrBuilder> fsTokenBuilder_;
|
org.apache.hadoop.hbase.protobuf.generated.SecureBulkLoadProtos.DelegationTokenProto, org.apache.hadoop.hbase.protobuf.generated.SecureBulkLoadProtos.DelegationTokenProto.Builder, org.apache.hadoop.hbase.protobuf.generated.SecureBulkLoadProtos.DelegationTokenProtoOrBuilder> fsTokenBuilder_;
|
||||||
|
@ -894,7 +894,7 @@ public final class SecureBulkLoadProtos {
|
||||||
return fsTokenBuilder_;
|
return fsTokenBuilder_;
|
||||||
}
|
}
|
||||||
|
|
||||||
// required string bulkToken = 4;
|
// required string bulk_token = 4;
|
||||||
private java.lang.Object bulkToken_ = "";
|
private java.lang.Object bulkToken_ = "";
|
||||||
public boolean hasBulkToken() {
|
public boolean hasBulkToken() {
|
||||||
return ((bitField0_ & 0x00000008) == 0x00000008);
|
return ((bitField0_ & 0x00000008) == 0x00000008);
|
||||||
|
@ -1972,7 +1972,7 @@ public final class SecureBulkLoadProtos {
|
||||||
public interface PrepareBulkLoadRequestOrBuilder
|
public interface PrepareBulkLoadRequestOrBuilder
|
||||||
extends com.google.protobuf.MessageOrBuilder {
|
extends com.google.protobuf.MessageOrBuilder {
|
||||||
|
|
||||||
// required bytes tableName = 1;
|
// required bytes table_name = 1;
|
||||||
boolean hasTableName();
|
boolean hasTableName();
|
||||||
com.google.protobuf.ByteString getTableName();
|
com.google.protobuf.ByteString getTableName();
|
||||||
}
|
}
|
||||||
|
@ -2005,8 +2005,8 @@ public final class SecureBulkLoadProtos {
|
||||||
}
|
}
|
||||||
|
|
||||||
private int bitField0_;
|
private int bitField0_;
|
||||||
// required bytes tableName = 1;
|
// required bytes table_name = 1;
|
||||||
public static final int TABLENAME_FIELD_NUMBER = 1;
|
public static final int TABLE_NAME_FIELD_NUMBER = 1;
|
||||||
private com.google.protobuf.ByteString tableName_;
|
private com.google.protobuf.ByteString tableName_;
|
||||||
public boolean hasTableName() {
|
public boolean hasTableName() {
|
||||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||||
|
@ -2088,7 +2088,7 @@ public final class SecureBulkLoadProtos {
|
||||||
int hash = 41;
|
int hash = 41;
|
||||||
hash = (19 * hash) + getDescriptorForType().hashCode();
|
hash = (19 * hash) + getDescriptorForType().hashCode();
|
||||||
if (hasTableName()) {
|
if (hasTableName()) {
|
||||||
hash = (37 * hash) + TABLENAME_FIELD_NUMBER;
|
hash = (37 * hash) + TABLE_NAME_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + getTableName().hashCode();
|
hash = (53 * hash) + getTableName().hashCode();
|
||||||
}
|
}
|
||||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||||
|
@ -2316,7 +2316,7 @@ public final class SecureBulkLoadProtos {
|
||||||
|
|
||||||
private int bitField0_;
|
private int bitField0_;
|
||||||
|
|
||||||
// required bytes tableName = 1;
|
// required bytes table_name = 1;
|
||||||
private com.google.protobuf.ByteString tableName_ = com.google.protobuf.ByteString.EMPTY;
|
private com.google.protobuf.ByteString tableName_ = com.google.protobuf.ByteString.EMPTY;
|
||||||
public boolean hasTableName() {
|
public boolean hasTableName() {
|
||||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||||
|
@ -2354,7 +2354,7 @@ public final class SecureBulkLoadProtos {
|
||||||
public interface PrepareBulkLoadResponseOrBuilder
|
public interface PrepareBulkLoadResponseOrBuilder
|
||||||
extends com.google.protobuf.MessageOrBuilder {
|
extends com.google.protobuf.MessageOrBuilder {
|
||||||
|
|
||||||
// required string bulkToken = 1;
|
// required string bulk_token = 1;
|
||||||
boolean hasBulkToken();
|
boolean hasBulkToken();
|
||||||
String getBulkToken();
|
String getBulkToken();
|
||||||
}
|
}
|
||||||
|
@ -2387,8 +2387,8 @@ public final class SecureBulkLoadProtos {
|
||||||
}
|
}
|
||||||
|
|
||||||
private int bitField0_;
|
private int bitField0_;
|
||||||
// required string bulkToken = 1;
|
// required string bulk_token = 1;
|
||||||
public static final int BULKTOKEN_FIELD_NUMBER = 1;
|
public static final int BULK_TOKEN_FIELD_NUMBER = 1;
|
||||||
private java.lang.Object bulkToken_;
|
private java.lang.Object bulkToken_;
|
||||||
public boolean hasBulkToken() {
|
public boolean hasBulkToken() {
|
||||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||||
|
@ -2492,7 +2492,7 @@ public final class SecureBulkLoadProtos {
|
||||||
int hash = 41;
|
int hash = 41;
|
||||||
hash = (19 * hash) + getDescriptorForType().hashCode();
|
hash = (19 * hash) + getDescriptorForType().hashCode();
|
||||||
if (hasBulkToken()) {
|
if (hasBulkToken()) {
|
||||||
hash = (37 * hash) + BULKTOKEN_FIELD_NUMBER;
|
hash = (37 * hash) + BULK_TOKEN_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + getBulkToken().hashCode();
|
hash = (53 * hash) + getBulkToken().hashCode();
|
||||||
}
|
}
|
||||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||||
|
@ -2720,7 +2720,7 @@ public final class SecureBulkLoadProtos {
|
||||||
|
|
||||||
private int bitField0_;
|
private int bitField0_;
|
||||||
|
|
||||||
// required string bulkToken = 1;
|
// required string bulk_token = 1;
|
||||||
private java.lang.Object bulkToken_ = "";
|
private java.lang.Object bulkToken_ = "";
|
||||||
public boolean hasBulkToken() {
|
public boolean hasBulkToken() {
|
||||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||||
|
@ -2770,7 +2770,7 @@ public final class SecureBulkLoadProtos {
|
||||||
public interface CleanupBulkLoadRequestOrBuilder
|
public interface CleanupBulkLoadRequestOrBuilder
|
||||||
extends com.google.protobuf.MessageOrBuilder {
|
extends com.google.protobuf.MessageOrBuilder {
|
||||||
|
|
||||||
// required string bulkToken = 1;
|
// required string bulk_token = 1;
|
||||||
boolean hasBulkToken();
|
boolean hasBulkToken();
|
||||||
String getBulkToken();
|
String getBulkToken();
|
||||||
}
|
}
|
||||||
|
@ -2803,8 +2803,8 @@ public final class SecureBulkLoadProtos {
|
||||||
}
|
}
|
||||||
|
|
||||||
private int bitField0_;
|
private int bitField0_;
|
||||||
// required string bulkToken = 1;
|
// required string bulk_token = 1;
|
||||||
public static final int BULKTOKEN_FIELD_NUMBER = 1;
|
public static final int BULK_TOKEN_FIELD_NUMBER = 1;
|
||||||
private java.lang.Object bulkToken_;
|
private java.lang.Object bulkToken_;
|
||||||
public boolean hasBulkToken() {
|
public boolean hasBulkToken() {
|
||||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||||
|
@ -2908,7 +2908,7 @@ public final class SecureBulkLoadProtos {
|
||||||
int hash = 41;
|
int hash = 41;
|
||||||
hash = (19 * hash) + getDescriptorForType().hashCode();
|
hash = (19 * hash) + getDescriptorForType().hashCode();
|
||||||
if (hasBulkToken()) {
|
if (hasBulkToken()) {
|
||||||
hash = (37 * hash) + BULKTOKEN_FIELD_NUMBER;
|
hash = (37 * hash) + BULK_TOKEN_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + getBulkToken().hashCode();
|
hash = (53 * hash) + getBulkToken().hashCode();
|
||||||
}
|
}
|
||||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||||
|
@ -3136,7 +3136,7 @@ public final class SecureBulkLoadProtos {
|
||||||
|
|
||||||
private int bitField0_;
|
private int bitField0_;
|
||||||
|
|
||||||
// required string bulkToken = 1;
|
// required string bulk_token = 1;
|
||||||
private java.lang.Object bulkToken_ = "";
|
private java.lang.Object bulkToken_ = "";
|
||||||
public boolean hasBulkToken() {
|
public boolean hasBulkToken() {
|
||||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||||
|
@ -3879,27 +3879,27 @@ public final class SecureBulkLoadProtos {
|
||||||
static {
|
static {
|
||||||
java.lang.String[] descriptorData = {
|
java.lang.String[] descriptorData = {
|
||||||
"\n\024SecureBulkLoad.proto\032\013hbase.proto\032\014Cli" +
|
"\n\024SecureBulkLoad.proto\032\013hbase.proto\032\014Cli" +
|
||||||
"ent.proto\"\244\001\n\033SecureBulkLoadHFilesReques" +
|
"ent.proto\"\251\001\n\033SecureBulkLoadHFilesReques" +
|
||||||
"t\0224\n\nfamilyPath\030\001 \003(\0132 .BulkLoadHFileReq" +
|
"t\0225\n\013family_path\030\001 \003(\0132 .BulkLoadHFileRe" +
|
||||||
"uest.FamilyPath\022\024\n\014assignSeqNum\030\002 \001(\010\022&\n" +
|
"quest.FamilyPath\022\026\n\016assign_seq_num\030\002 \001(\010" +
|
||||||
"\007fsToken\030\003 \002(\0132\025.DelegationTokenProto\022\021\n" +
|
"\022\'\n\010fs_token\030\003 \002(\0132\025.DelegationTokenProt" +
|
||||||
"\tbulkToken\030\004 \002(\t\".\n\034SecureBulkLoadHFiles" +
|
"o\022\022\n\nbulk_token\030\004 \002(\t\".\n\034SecureBulkLoadH" +
|
||||||
"Response\022\016\n\006loaded\030\001 \002(\010\"[\n\024DelegationTo" +
|
"FilesResponse\022\016\n\006loaded\030\001 \002(\010\"[\n\024Delegat" +
|
||||||
"kenProto\022\022\n\nidentifier\030\001 \001(\014\022\020\n\010password" +
|
"ionTokenProto\022\022\n\nidentifier\030\001 \001(\014\022\020\n\010pas" +
|
||||||
"\030\002 \001(\014\022\014\n\004kind\030\003 \001(\t\022\017\n\007service\030\004 \001(\t\"+\n" +
|
"sword\030\002 \001(\014\022\014\n\004kind\030\003 \001(\t\022\017\n\007service\030\004 \001" +
|
||||||
"\026PrepareBulkLoadRequest\022\021\n\ttableName\030\001 \002",
|
"(\t\",\n\026PrepareBulkLoadRequest\022\022\n\ntable_na",
|
||||||
"(\014\",\n\027PrepareBulkLoadResponse\022\021\n\tbulkTok" +
|
"me\030\001 \002(\014\"-\n\027PrepareBulkLoadResponse\022\022\n\nb" +
|
||||||
"en\030\001 \002(\t\"+\n\026CleanupBulkLoadRequest\022\021\n\tbu" +
|
"ulk_token\030\001 \002(\t\",\n\026CleanupBulkLoadReques" +
|
||||||
"lkToken\030\001 \002(\t\"\031\n\027CleanupBulkLoadResponse" +
|
"t\022\022\n\nbulk_token\030\001 \002(\t\"\031\n\027CleanupBulkLoad" +
|
||||||
"2\370\001\n\025SecureBulkLoadService\022D\n\017prepareBul" +
|
"Response2\370\001\n\025SecureBulkLoadService\022D\n\017Pr" +
|
||||||
"kLoad\022\027.PrepareBulkLoadRequest\032\030.Prepare" +
|
"epareBulkLoad\022\027.PrepareBulkLoadRequest\032\030" +
|
||||||
"BulkLoadResponse\022S\n\024secureBulkLoadHFiles" +
|
".PrepareBulkLoadResponse\022S\n\024SecureBulkLo" +
|
||||||
"\022\034.SecureBulkLoadHFilesRequest\032\035.SecureB" +
|
"adHFiles\022\034.SecureBulkLoadHFilesRequest\032\035" +
|
||||||
"ulkLoadHFilesResponse\022D\n\017cleanupBulkLoad" +
|
".SecureBulkLoadHFilesResponse\022D\n\017Cleanup" +
|
||||||
"\022\027.CleanupBulkLoadRequest\032\030.CleanupBulkL" +
|
"BulkLoad\022\027.CleanupBulkLoadRequest\032\030.Clea" +
|
||||||
"oadResponseBJ\n*org.apache.hadoop.hbase.p",
|
"nupBulkLoadResponseBJ\n*org.apache.hadoop",
|
||||||
"rotobuf.generatedB\024SecureBulkLoadProtosH" +
|
".hbase.protobuf.generatedB\024SecureBulkLoa" +
|
||||||
"\001\210\001\001\240\001\001"
|
"dProtosH\001\210\001\001\240\001\001"
|
||||||
};
|
};
|
||||||
com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
|
com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
|
||||||
new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() {
|
new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() {
|
||||||
|
|
|
@ -11,11 +11,11 @@ public final class Tracing {
|
||||||
public interface RPCTInfoOrBuilder
|
public interface RPCTInfoOrBuilder
|
||||||
extends com.google.protobuf.MessageOrBuilder {
|
extends com.google.protobuf.MessageOrBuilder {
|
||||||
|
|
||||||
// optional int64 traceId = 1;
|
// optional int64 trace_id = 1;
|
||||||
boolean hasTraceId();
|
boolean hasTraceId();
|
||||||
long getTraceId();
|
long getTraceId();
|
||||||
|
|
||||||
// optional int64 parentId = 2;
|
// optional int64 parent_id = 2;
|
||||||
boolean hasParentId();
|
boolean hasParentId();
|
||||||
long getParentId();
|
long getParentId();
|
||||||
}
|
}
|
||||||
|
@ -48,8 +48,8 @@ public final class Tracing {
|
||||||
}
|
}
|
||||||
|
|
||||||
private int bitField0_;
|
private int bitField0_;
|
||||||
// optional int64 traceId = 1;
|
// optional int64 trace_id = 1;
|
||||||
public static final int TRACEID_FIELD_NUMBER = 1;
|
public static final int TRACE_ID_FIELD_NUMBER = 1;
|
||||||
private long traceId_;
|
private long traceId_;
|
||||||
public boolean hasTraceId() {
|
public boolean hasTraceId() {
|
||||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||||
|
@ -58,8 +58,8 @@ public final class Tracing {
|
||||||
return traceId_;
|
return traceId_;
|
||||||
}
|
}
|
||||||
|
|
||||||
// optional int64 parentId = 2;
|
// optional int64 parent_id = 2;
|
||||||
public static final int PARENTID_FIELD_NUMBER = 2;
|
public static final int PARENT_ID_FIELD_NUMBER = 2;
|
||||||
private long parentId_;
|
private long parentId_;
|
||||||
public boolean hasParentId() {
|
public boolean hasParentId() {
|
||||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||||
|
@ -150,11 +150,11 @@ public final class Tracing {
|
||||||
int hash = 41;
|
int hash = 41;
|
||||||
hash = (19 * hash) + getDescriptorForType().hashCode();
|
hash = (19 * hash) + getDescriptorForType().hashCode();
|
||||||
if (hasTraceId()) {
|
if (hasTraceId()) {
|
||||||
hash = (37 * hash) + TRACEID_FIELD_NUMBER;
|
hash = (37 * hash) + TRACE_ID_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + hashLong(getTraceId());
|
hash = (53 * hash) + hashLong(getTraceId());
|
||||||
}
|
}
|
||||||
if (hasParentId()) {
|
if (hasParentId()) {
|
||||||
hash = (37 * hash) + PARENTID_FIELD_NUMBER;
|
hash = (37 * hash) + PARENT_ID_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + hashLong(getParentId());
|
hash = (53 * hash) + hashLong(getParentId());
|
||||||
}
|
}
|
||||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||||
|
@ -392,7 +392,7 @@ public final class Tracing {
|
||||||
|
|
||||||
private int bitField0_;
|
private int bitField0_;
|
||||||
|
|
||||||
// optional int64 traceId = 1;
|
// optional int64 trace_id = 1;
|
||||||
private long traceId_ ;
|
private long traceId_ ;
|
||||||
public boolean hasTraceId() {
|
public boolean hasTraceId() {
|
||||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||||
|
@ -413,7 +413,7 @@ public final class Tracing {
|
||||||
return this;
|
return this;
|
||||||
}
|
}
|
||||||
|
|
||||||
// optional int64 parentId = 2;
|
// optional int64 parent_id = 2;
|
||||||
private long parentId_ ;
|
private long parentId_ ;
|
||||||
public boolean hasParentId() {
|
public boolean hasParentId() {
|
||||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||||
|
@ -459,10 +459,10 @@ public final class Tracing {
|
||||||
descriptor;
|
descriptor;
|
||||||
static {
|
static {
|
||||||
java.lang.String[] descriptorData = {
|
java.lang.String[] descriptorData = {
|
||||||
"\n\rTracing.proto\"-\n\010RPCTInfo\022\017\n\007traceId\030\001" +
|
"\n\rTracing.proto\"/\n\010RPCTInfo\022\020\n\010trace_id\030" +
|
||||||
" \001(\003\022\020\n\010parentId\030\002 \001(\003B:\n*org.apache.had" +
|
"\001 \001(\003\022\021\n\tparent_id\030\002 \001(\003B:\n*org.apache.h" +
|
||||||
"oop.hbase.protobuf.generatedB\007TracingH\001\240" +
|
"adoop.hbase.protobuf.generatedB\007TracingH" +
|
||||||
"\001\001"
|
"\001\240\001\001"
|
||||||
};
|
};
|
||||||
com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
|
com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
|
||||||
new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() {
|
new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() {
|
||||||
|
|
|
@ -80,7 +80,7 @@ public final class WALProtos {
|
||||||
public interface WALHeaderOrBuilder
|
public interface WALHeaderOrBuilder
|
||||||
extends com.google.protobuf.MessageOrBuilder {
|
extends com.google.protobuf.MessageOrBuilder {
|
||||||
|
|
||||||
// optional bool hasCompression = 1;
|
// optional bool has_compression = 1;
|
||||||
boolean hasHasCompression();
|
boolean hasHasCompression();
|
||||||
boolean getHasCompression();
|
boolean getHasCompression();
|
||||||
}
|
}
|
||||||
|
@ -113,8 +113,8 @@ public final class WALProtos {
|
||||||
}
|
}
|
||||||
|
|
||||||
private int bitField0_;
|
private int bitField0_;
|
||||||
// optional bool hasCompression = 1;
|
// optional bool has_compression = 1;
|
||||||
public static final int HASCOMPRESSION_FIELD_NUMBER = 1;
|
public static final int HAS_COMPRESSION_FIELD_NUMBER = 1;
|
||||||
private boolean hasCompression_;
|
private boolean hasCompression_;
|
||||||
public boolean hasHasCompression() {
|
public boolean hasHasCompression() {
|
||||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||||
|
@ -192,7 +192,7 @@ public final class WALProtos {
|
||||||
int hash = 41;
|
int hash = 41;
|
||||||
hash = (19 * hash) + getDescriptorForType().hashCode();
|
hash = (19 * hash) + getDescriptorForType().hashCode();
|
||||||
if (hasHasCompression()) {
|
if (hasHasCompression()) {
|
||||||
hash = (37 * hash) + HASCOMPRESSION_FIELD_NUMBER;
|
hash = (37 * hash) + HAS_COMPRESSION_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + hashBoolean(getHasCompression());
|
hash = (53 * hash) + hashBoolean(getHasCompression());
|
||||||
}
|
}
|
||||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||||
|
@ -416,7 +416,7 @@ public final class WALProtos {
|
||||||
|
|
||||||
private int bitField0_;
|
private int bitField0_;
|
||||||
|
|
||||||
// optional bool hasCompression = 1;
|
// optional bool has_compression = 1;
|
||||||
private boolean hasCompression_ ;
|
private boolean hasCompression_ ;
|
||||||
public boolean hasHasCompression() {
|
public boolean hasHasCompression() {
|
||||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||||
|
@ -451,23 +451,23 @@ public final class WALProtos {
|
||||||
public interface WALKeyOrBuilder
|
public interface WALKeyOrBuilder
|
||||||
extends com.google.protobuf.MessageOrBuilder {
|
extends com.google.protobuf.MessageOrBuilder {
|
||||||
|
|
||||||
// required bytes encodedRegionName = 1;
|
// required bytes encoded_region_name = 1;
|
||||||
boolean hasEncodedRegionName();
|
boolean hasEncodedRegionName();
|
||||||
com.google.protobuf.ByteString getEncodedRegionName();
|
com.google.protobuf.ByteString getEncodedRegionName();
|
||||||
|
|
||||||
// required bytes tableName = 2;
|
// required bytes table_name = 2;
|
||||||
boolean hasTableName();
|
boolean hasTableName();
|
||||||
com.google.protobuf.ByteString getTableName();
|
com.google.protobuf.ByteString getTableName();
|
||||||
|
|
||||||
// required uint64 logSequenceNumber = 3;
|
// required uint64 log_sequence_number = 3;
|
||||||
boolean hasLogSequenceNumber();
|
boolean hasLogSequenceNumber();
|
||||||
long getLogSequenceNumber();
|
long getLogSequenceNumber();
|
||||||
|
|
||||||
// required uint64 writeTime = 4;
|
// required uint64 write_time = 4;
|
||||||
boolean hasWriteTime();
|
boolean hasWriteTime();
|
||||||
long getWriteTime();
|
long getWriteTime();
|
||||||
|
|
||||||
// optional .UUID clusterId = 5;
|
// optional .UUID cluster_id = 5;
|
||||||
boolean hasClusterId();
|
boolean hasClusterId();
|
||||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.UUID getClusterId();
|
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.UUID getClusterId();
|
||||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.UUIDOrBuilder getClusterIdOrBuilder();
|
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.UUIDOrBuilder getClusterIdOrBuilder();
|
||||||
|
@ -482,7 +482,7 @@ public final class WALProtos {
|
||||||
org.apache.hadoop.hbase.protobuf.generated.WALProtos.FamilyScopeOrBuilder getScopesOrBuilder(
|
org.apache.hadoop.hbase.protobuf.generated.WALProtos.FamilyScopeOrBuilder getScopesOrBuilder(
|
||||||
int index);
|
int index);
|
||||||
|
|
||||||
// optional uint32 followingKvCount = 7;
|
// optional uint32 following_kv_count = 7;
|
||||||
boolean hasFollowingKvCount();
|
boolean hasFollowingKvCount();
|
||||||
int getFollowingKvCount();
|
int getFollowingKvCount();
|
||||||
}
|
}
|
||||||
|
@ -515,8 +515,8 @@ public final class WALProtos {
|
||||||
}
|
}
|
||||||
|
|
||||||
private int bitField0_;
|
private int bitField0_;
|
||||||
// required bytes encodedRegionName = 1;
|
// required bytes encoded_region_name = 1;
|
||||||
public static final int ENCODEDREGIONNAME_FIELD_NUMBER = 1;
|
public static final int ENCODED_REGION_NAME_FIELD_NUMBER = 1;
|
||||||
private com.google.protobuf.ByteString encodedRegionName_;
|
private com.google.protobuf.ByteString encodedRegionName_;
|
||||||
public boolean hasEncodedRegionName() {
|
public boolean hasEncodedRegionName() {
|
||||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||||
|
@ -525,8 +525,8 @@ public final class WALProtos {
|
||||||
return encodedRegionName_;
|
return encodedRegionName_;
|
||||||
}
|
}
|
||||||
|
|
||||||
// required bytes tableName = 2;
|
// required bytes table_name = 2;
|
||||||
public static final int TABLENAME_FIELD_NUMBER = 2;
|
public static final int TABLE_NAME_FIELD_NUMBER = 2;
|
||||||
private com.google.protobuf.ByteString tableName_;
|
private com.google.protobuf.ByteString tableName_;
|
||||||
public boolean hasTableName() {
|
public boolean hasTableName() {
|
||||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||||
|
@ -535,8 +535,8 @@ public final class WALProtos {
|
||||||
return tableName_;
|
return tableName_;
|
||||||
}
|
}
|
||||||
|
|
||||||
// required uint64 logSequenceNumber = 3;
|
// required uint64 log_sequence_number = 3;
|
||||||
public static final int LOGSEQUENCENUMBER_FIELD_NUMBER = 3;
|
public static final int LOG_SEQUENCE_NUMBER_FIELD_NUMBER = 3;
|
||||||
private long logSequenceNumber_;
|
private long logSequenceNumber_;
|
||||||
public boolean hasLogSequenceNumber() {
|
public boolean hasLogSequenceNumber() {
|
||||||
return ((bitField0_ & 0x00000004) == 0x00000004);
|
return ((bitField0_ & 0x00000004) == 0x00000004);
|
||||||
|
@ -545,8 +545,8 @@ public final class WALProtos {
|
||||||
return logSequenceNumber_;
|
return logSequenceNumber_;
|
||||||
}
|
}
|
||||||
|
|
||||||
// required uint64 writeTime = 4;
|
// required uint64 write_time = 4;
|
||||||
public static final int WRITETIME_FIELD_NUMBER = 4;
|
public static final int WRITE_TIME_FIELD_NUMBER = 4;
|
||||||
private long writeTime_;
|
private long writeTime_;
|
||||||
public boolean hasWriteTime() {
|
public boolean hasWriteTime() {
|
||||||
return ((bitField0_ & 0x00000008) == 0x00000008);
|
return ((bitField0_ & 0x00000008) == 0x00000008);
|
||||||
|
@ -555,8 +555,8 @@ public final class WALProtos {
|
||||||
return writeTime_;
|
return writeTime_;
|
||||||
}
|
}
|
||||||
|
|
||||||
// optional .UUID clusterId = 5;
|
// optional .UUID cluster_id = 5;
|
||||||
public static final int CLUSTERID_FIELD_NUMBER = 5;
|
public static final int CLUSTER_ID_FIELD_NUMBER = 5;
|
||||||
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.UUID clusterId_;
|
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.UUID clusterId_;
|
||||||
public boolean hasClusterId() {
|
public boolean hasClusterId() {
|
||||||
return ((bitField0_ & 0x00000010) == 0x00000010);
|
return ((bitField0_ & 0x00000010) == 0x00000010);
|
||||||
|
@ -589,8 +589,8 @@ public final class WALProtos {
|
||||||
return scopes_.get(index);
|
return scopes_.get(index);
|
||||||
}
|
}
|
||||||
|
|
||||||
// optional uint32 followingKvCount = 7;
|
// optional uint32 following_kv_count = 7;
|
||||||
public static final int FOLLOWINGKVCOUNT_FIELD_NUMBER = 7;
|
public static final int FOLLOWING_KV_COUNT_FIELD_NUMBER = 7;
|
||||||
private int followingKvCount_;
|
private int followingKvCount_;
|
||||||
public boolean hasFollowingKvCount() {
|
public boolean hasFollowingKvCount() {
|
||||||
return ((bitField0_ & 0x00000020) == 0x00000020);
|
return ((bitField0_ & 0x00000020) == 0x00000020);
|
||||||
|
@ -771,23 +771,23 @@ public final class WALProtos {
|
||||||
int hash = 41;
|
int hash = 41;
|
||||||
hash = (19 * hash) + getDescriptorForType().hashCode();
|
hash = (19 * hash) + getDescriptorForType().hashCode();
|
||||||
if (hasEncodedRegionName()) {
|
if (hasEncodedRegionName()) {
|
||||||
hash = (37 * hash) + ENCODEDREGIONNAME_FIELD_NUMBER;
|
hash = (37 * hash) + ENCODED_REGION_NAME_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + getEncodedRegionName().hashCode();
|
hash = (53 * hash) + getEncodedRegionName().hashCode();
|
||||||
}
|
}
|
||||||
if (hasTableName()) {
|
if (hasTableName()) {
|
||||||
hash = (37 * hash) + TABLENAME_FIELD_NUMBER;
|
hash = (37 * hash) + TABLE_NAME_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + getTableName().hashCode();
|
hash = (53 * hash) + getTableName().hashCode();
|
||||||
}
|
}
|
||||||
if (hasLogSequenceNumber()) {
|
if (hasLogSequenceNumber()) {
|
||||||
hash = (37 * hash) + LOGSEQUENCENUMBER_FIELD_NUMBER;
|
hash = (37 * hash) + LOG_SEQUENCE_NUMBER_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + hashLong(getLogSequenceNumber());
|
hash = (53 * hash) + hashLong(getLogSequenceNumber());
|
||||||
}
|
}
|
||||||
if (hasWriteTime()) {
|
if (hasWriteTime()) {
|
||||||
hash = (37 * hash) + WRITETIME_FIELD_NUMBER;
|
hash = (37 * hash) + WRITE_TIME_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + hashLong(getWriteTime());
|
hash = (53 * hash) + hashLong(getWriteTime());
|
||||||
}
|
}
|
||||||
if (hasClusterId()) {
|
if (hasClusterId()) {
|
||||||
hash = (37 * hash) + CLUSTERID_FIELD_NUMBER;
|
hash = (37 * hash) + CLUSTER_ID_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + getClusterId().hashCode();
|
hash = (53 * hash) + getClusterId().hashCode();
|
||||||
}
|
}
|
||||||
if (getScopesCount() > 0) {
|
if (getScopesCount() > 0) {
|
||||||
|
@ -795,7 +795,7 @@ public final class WALProtos {
|
||||||
hash = (53 * hash) + getScopesList().hashCode();
|
hash = (53 * hash) + getScopesList().hashCode();
|
||||||
}
|
}
|
||||||
if (hasFollowingKvCount()) {
|
if (hasFollowingKvCount()) {
|
||||||
hash = (37 * hash) + FOLLOWINGKVCOUNT_FIELD_NUMBER;
|
hash = (37 * hash) + FOLLOWING_KV_COUNT_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + getFollowingKvCount();
|
hash = (53 * hash) + getFollowingKvCount();
|
||||||
}
|
}
|
||||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||||
|
@ -1178,7 +1178,7 @@ public final class WALProtos {
|
||||||
|
|
||||||
private int bitField0_;
|
private int bitField0_;
|
||||||
|
|
||||||
// required bytes encodedRegionName = 1;
|
// required bytes encoded_region_name = 1;
|
||||||
private com.google.protobuf.ByteString encodedRegionName_ = com.google.protobuf.ByteString.EMPTY;
|
private com.google.protobuf.ByteString encodedRegionName_ = com.google.protobuf.ByteString.EMPTY;
|
||||||
public boolean hasEncodedRegionName() {
|
public boolean hasEncodedRegionName() {
|
||||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||||
|
@ -1202,7 +1202,7 @@ public final class WALProtos {
|
||||||
return this;
|
return this;
|
||||||
}
|
}
|
||||||
|
|
||||||
// required bytes tableName = 2;
|
// required bytes table_name = 2;
|
||||||
private com.google.protobuf.ByteString tableName_ = com.google.protobuf.ByteString.EMPTY;
|
private com.google.protobuf.ByteString tableName_ = com.google.protobuf.ByteString.EMPTY;
|
||||||
public boolean hasTableName() {
|
public boolean hasTableName() {
|
||||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||||
|
@ -1226,7 +1226,7 @@ public final class WALProtos {
|
||||||
return this;
|
return this;
|
||||||
}
|
}
|
||||||
|
|
||||||
// required uint64 logSequenceNumber = 3;
|
// required uint64 log_sequence_number = 3;
|
||||||
private long logSequenceNumber_ ;
|
private long logSequenceNumber_ ;
|
||||||
public boolean hasLogSequenceNumber() {
|
public boolean hasLogSequenceNumber() {
|
||||||
return ((bitField0_ & 0x00000004) == 0x00000004);
|
return ((bitField0_ & 0x00000004) == 0x00000004);
|
||||||
|
@ -1247,7 +1247,7 @@ public final class WALProtos {
|
||||||
return this;
|
return this;
|
||||||
}
|
}
|
||||||
|
|
||||||
// required uint64 writeTime = 4;
|
// required uint64 write_time = 4;
|
||||||
private long writeTime_ ;
|
private long writeTime_ ;
|
||||||
public boolean hasWriteTime() {
|
public boolean hasWriteTime() {
|
||||||
return ((bitField0_ & 0x00000008) == 0x00000008);
|
return ((bitField0_ & 0x00000008) == 0x00000008);
|
||||||
|
@ -1268,7 +1268,7 @@ public final class WALProtos {
|
||||||
return this;
|
return this;
|
||||||
}
|
}
|
||||||
|
|
||||||
// optional .UUID clusterId = 5;
|
// optional .UUID cluster_id = 5;
|
||||||
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.UUID clusterId_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.UUID.getDefaultInstance();
|
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.UUID clusterId_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.UUID.getDefaultInstance();
|
||||||
private com.google.protobuf.SingleFieldBuilder<
|
private com.google.protobuf.SingleFieldBuilder<
|
||||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.UUID, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.UUID.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.UUIDOrBuilder> clusterIdBuilder_;
|
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.UUID, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.UUID.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.UUIDOrBuilder> clusterIdBuilder_;
|
||||||
|
@ -1544,7 +1544,7 @@ public final class WALProtos {
|
||||||
return scopesBuilder_;
|
return scopesBuilder_;
|
||||||
}
|
}
|
||||||
|
|
||||||
// optional uint32 followingKvCount = 7;
|
// optional uint32 following_kv_count = 7;
|
||||||
private int followingKvCount_ ;
|
private int followingKvCount_ ;
|
||||||
public boolean hasFollowingKvCount() {
|
public boolean hasFollowingKvCount() {
|
||||||
return ((bitField0_ & 0x00000040) == 0x00000040);
|
return ((bitField0_ & 0x00000040) == 0x00000040);
|
||||||
|
@ -1583,7 +1583,7 @@ public final class WALProtos {
|
||||||
boolean hasFamily();
|
boolean hasFamily();
|
||||||
com.google.protobuf.ByteString getFamily();
|
com.google.protobuf.ByteString getFamily();
|
||||||
|
|
||||||
// required .ScopeType scopeType = 2;
|
// required .ScopeType scope_type = 2;
|
||||||
boolean hasScopeType();
|
boolean hasScopeType();
|
||||||
org.apache.hadoop.hbase.protobuf.generated.WALProtos.ScopeType getScopeType();
|
org.apache.hadoop.hbase.protobuf.generated.WALProtos.ScopeType getScopeType();
|
||||||
}
|
}
|
||||||
|
@ -1626,8 +1626,8 @@ public final class WALProtos {
|
||||||
return family_;
|
return family_;
|
||||||
}
|
}
|
||||||
|
|
||||||
// required .ScopeType scopeType = 2;
|
// required .ScopeType scope_type = 2;
|
||||||
public static final int SCOPETYPE_FIELD_NUMBER = 2;
|
public static final int SCOPE_TYPE_FIELD_NUMBER = 2;
|
||||||
private org.apache.hadoop.hbase.protobuf.generated.WALProtos.ScopeType scopeType_;
|
private org.apache.hadoop.hbase.protobuf.generated.WALProtos.ScopeType scopeType_;
|
||||||
public boolean hasScopeType() {
|
public boolean hasScopeType() {
|
||||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||||
|
@ -1730,7 +1730,7 @@ public final class WALProtos {
|
||||||
hash = (53 * hash) + getFamily().hashCode();
|
hash = (53 * hash) + getFamily().hashCode();
|
||||||
}
|
}
|
||||||
if (hasScopeType()) {
|
if (hasScopeType()) {
|
||||||
hash = (37 * hash) + SCOPETYPE_FIELD_NUMBER;
|
hash = (37 * hash) + SCOPE_TYPE_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + hashEnum(getScopeType());
|
hash = (53 * hash) + hashEnum(getScopeType());
|
||||||
}
|
}
|
||||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||||
|
@ -2006,7 +2006,7 @@ public final class WALProtos {
|
||||||
return this;
|
return this;
|
||||||
}
|
}
|
||||||
|
|
||||||
// required .ScopeType scopeType = 2;
|
// required .ScopeType scope_type = 2;
|
||||||
private org.apache.hadoop.hbase.protobuf.generated.WALProtos.ScopeType scopeType_ = org.apache.hadoop.hbase.protobuf.generated.WALProtos.ScopeType.REPLICATION_SCOPE_LOCAL;
|
private org.apache.hadoop.hbase.protobuf.generated.WALProtos.ScopeType scopeType_ = org.apache.hadoop.hbase.protobuf.generated.WALProtos.ScopeType.REPLICATION_SCOPE_LOCAL;
|
||||||
public boolean hasScopeType() {
|
public boolean hasScopeType() {
|
||||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||||
|
@ -2044,29 +2044,29 @@ public final class WALProtos {
|
||||||
public interface CompactionDescriptorOrBuilder
|
public interface CompactionDescriptorOrBuilder
|
||||||
extends com.google.protobuf.MessageOrBuilder {
|
extends com.google.protobuf.MessageOrBuilder {
|
||||||
|
|
||||||
// required bytes tableName = 1;
|
// required bytes table_name = 1;
|
||||||
boolean hasTableName();
|
boolean hasTableName();
|
||||||
com.google.protobuf.ByteString getTableName();
|
com.google.protobuf.ByteString getTableName();
|
||||||
|
|
||||||
// required bytes encodedRegionName = 2;
|
// required bytes encoded_region_name = 2;
|
||||||
boolean hasEncodedRegionName();
|
boolean hasEncodedRegionName();
|
||||||
com.google.protobuf.ByteString getEncodedRegionName();
|
com.google.protobuf.ByteString getEncodedRegionName();
|
||||||
|
|
||||||
// required bytes familyName = 3;
|
// required bytes family_name = 3;
|
||||||
boolean hasFamilyName();
|
boolean hasFamilyName();
|
||||||
com.google.protobuf.ByteString getFamilyName();
|
com.google.protobuf.ByteString getFamilyName();
|
||||||
|
|
||||||
// repeated string compactionInput = 4;
|
// repeated string compaction_input = 4;
|
||||||
java.util.List<String> getCompactionInputList();
|
java.util.List<String> getCompactionInputList();
|
||||||
int getCompactionInputCount();
|
int getCompactionInputCount();
|
||||||
String getCompactionInput(int index);
|
String getCompactionInput(int index);
|
||||||
|
|
||||||
// repeated string compactionOutput = 5;
|
// repeated string compaction_output = 5;
|
||||||
java.util.List<String> getCompactionOutputList();
|
java.util.List<String> getCompactionOutputList();
|
||||||
int getCompactionOutputCount();
|
int getCompactionOutputCount();
|
||||||
String getCompactionOutput(int index);
|
String getCompactionOutput(int index);
|
||||||
|
|
||||||
// required string storeHomeDir = 6;
|
// required string store_home_dir = 6;
|
||||||
boolean hasStoreHomeDir();
|
boolean hasStoreHomeDir();
|
||||||
String getStoreHomeDir();
|
String getStoreHomeDir();
|
||||||
}
|
}
|
||||||
|
@ -2099,8 +2099,8 @@ public final class WALProtos {
|
||||||
}
|
}
|
||||||
|
|
||||||
private int bitField0_;
|
private int bitField0_;
|
||||||
// required bytes tableName = 1;
|
// required bytes table_name = 1;
|
||||||
public static final int TABLENAME_FIELD_NUMBER = 1;
|
public static final int TABLE_NAME_FIELD_NUMBER = 1;
|
||||||
private com.google.protobuf.ByteString tableName_;
|
private com.google.protobuf.ByteString tableName_;
|
||||||
public boolean hasTableName() {
|
public boolean hasTableName() {
|
||||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||||
|
@ -2109,8 +2109,8 @@ public final class WALProtos {
|
||||||
return tableName_;
|
return tableName_;
|
||||||
}
|
}
|
||||||
|
|
||||||
// required bytes encodedRegionName = 2;
|
// required bytes encoded_region_name = 2;
|
||||||
public static final int ENCODEDREGIONNAME_FIELD_NUMBER = 2;
|
public static final int ENCODED_REGION_NAME_FIELD_NUMBER = 2;
|
||||||
private com.google.protobuf.ByteString encodedRegionName_;
|
private com.google.protobuf.ByteString encodedRegionName_;
|
||||||
public boolean hasEncodedRegionName() {
|
public boolean hasEncodedRegionName() {
|
||||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||||
|
@ -2119,8 +2119,8 @@ public final class WALProtos {
|
||||||
return encodedRegionName_;
|
return encodedRegionName_;
|
||||||
}
|
}
|
||||||
|
|
||||||
// required bytes familyName = 3;
|
// required bytes family_name = 3;
|
||||||
public static final int FAMILYNAME_FIELD_NUMBER = 3;
|
public static final int FAMILY_NAME_FIELD_NUMBER = 3;
|
||||||
private com.google.protobuf.ByteString familyName_;
|
private com.google.protobuf.ByteString familyName_;
|
||||||
public boolean hasFamilyName() {
|
public boolean hasFamilyName() {
|
||||||
return ((bitField0_ & 0x00000004) == 0x00000004);
|
return ((bitField0_ & 0x00000004) == 0x00000004);
|
||||||
|
@ -2129,8 +2129,8 @@ public final class WALProtos {
|
||||||
return familyName_;
|
return familyName_;
|
||||||
}
|
}
|
||||||
|
|
||||||
// repeated string compactionInput = 4;
|
// repeated string compaction_input = 4;
|
||||||
public static final int COMPACTIONINPUT_FIELD_NUMBER = 4;
|
public static final int COMPACTION_INPUT_FIELD_NUMBER = 4;
|
||||||
private com.google.protobuf.LazyStringList compactionInput_;
|
private com.google.protobuf.LazyStringList compactionInput_;
|
||||||
public java.util.List<String>
|
public java.util.List<String>
|
||||||
getCompactionInputList() {
|
getCompactionInputList() {
|
||||||
|
@ -2143,8 +2143,8 @@ public final class WALProtos {
|
||||||
return compactionInput_.get(index);
|
return compactionInput_.get(index);
|
||||||
}
|
}
|
||||||
|
|
||||||
// repeated string compactionOutput = 5;
|
// repeated string compaction_output = 5;
|
||||||
public static final int COMPACTIONOUTPUT_FIELD_NUMBER = 5;
|
public static final int COMPACTION_OUTPUT_FIELD_NUMBER = 5;
|
||||||
private com.google.protobuf.LazyStringList compactionOutput_;
|
private com.google.protobuf.LazyStringList compactionOutput_;
|
||||||
public java.util.List<String>
|
public java.util.List<String>
|
||||||
getCompactionOutputList() {
|
getCompactionOutputList() {
|
||||||
|
@ -2157,8 +2157,8 @@ public final class WALProtos {
|
||||||
return compactionOutput_.get(index);
|
return compactionOutput_.get(index);
|
||||||
}
|
}
|
||||||
|
|
||||||
// required string storeHomeDir = 6;
|
// required string store_home_dir = 6;
|
||||||
public static final int STOREHOMEDIR_FIELD_NUMBER = 6;
|
public static final int STORE_HOME_DIR_FIELD_NUMBER = 6;
|
||||||
private java.lang.Object storeHomeDir_;
|
private java.lang.Object storeHomeDir_;
|
||||||
public boolean hasStoreHomeDir() {
|
public boolean hasStoreHomeDir() {
|
||||||
return ((bitField0_ & 0x00000008) == 0x00000008);
|
return ((bitField0_ & 0x00000008) == 0x00000008);
|
||||||
|
@ -2343,27 +2343,27 @@ public final class WALProtos {
|
||||||
int hash = 41;
|
int hash = 41;
|
||||||
hash = (19 * hash) + getDescriptorForType().hashCode();
|
hash = (19 * hash) + getDescriptorForType().hashCode();
|
||||||
if (hasTableName()) {
|
if (hasTableName()) {
|
||||||
hash = (37 * hash) + TABLENAME_FIELD_NUMBER;
|
hash = (37 * hash) + TABLE_NAME_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + getTableName().hashCode();
|
hash = (53 * hash) + getTableName().hashCode();
|
||||||
}
|
}
|
||||||
if (hasEncodedRegionName()) {
|
if (hasEncodedRegionName()) {
|
||||||
hash = (37 * hash) + ENCODEDREGIONNAME_FIELD_NUMBER;
|
hash = (37 * hash) + ENCODED_REGION_NAME_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + getEncodedRegionName().hashCode();
|
hash = (53 * hash) + getEncodedRegionName().hashCode();
|
||||||
}
|
}
|
||||||
if (hasFamilyName()) {
|
if (hasFamilyName()) {
|
||||||
hash = (37 * hash) + FAMILYNAME_FIELD_NUMBER;
|
hash = (37 * hash) + FAMILY_NAME_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + getFamilyName().hashCode();
|
hash = (53 * hash) + getFamilyName().hashCode();
|
||||||
}
|
}
|
||||||
if (getCompactionInputCount() > 0) {
|
if (getCompactionInputCount() > 0) {
|
||||||
hash = (37 * hash) + COMPACTIONINPUT_FIELD_NUMBER;
|
hash = (37 * hash) + COMPACTION_INPUT_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + getCompactionInputList().hashCode();
|
hash = (53 * hash) + getCompactionInputList().hashCode();
|
||||||
}
|
}
|
||||||
if (getCompactionOutputCount() > 0) {
|
if (getCompactionOutputCount() > 0) {
|
||||||
hash = (37 * hash) + COMPACTIONOUTPUT_FIELD_NUMBER;
|
hash = (37 * hash) + COMPACTION_OUTPUT_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + getCompactionOutputList().hashCode();
|
hash = (53 * hash) + getCompactionOutputList().hashCode();
|
||||||
}
|
}
|
||||||
if (hasStoreHomeDir()) {
|
if (hasStoreHomeDir()) {
|
||||||
hash = (37 * hash) + STOREHOMEDIR_FIELD_NUMBER;
|
hash = (37 * hash) + STORE_HOME_DIR_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + getStoreHomeDir().hashCode();
|
hash = (53 * hash) + getStoreHomeDir().hashCode();
|
||||||
}
|
}
|
||||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||||
|
@ -2691,7 +2691,7 @@ public final class WALProtos {
|
||||||
|
|
||||||
private int bitField0_;
|
private int bitField0_;
|
||||||
|
|
||||||
// required bytes tableName = 1;
|
// required bytes table_name = 1;
|
||||||
private com.google.protobuf.ByteString tableName_ = com.google.protobuf.ByteString.EMPTY;
|
private com.google.protobuf.ByteString tableName_ = com.google.protobuf.ByteString.EMPTY;
|
||||||
public boolean hasTableName() {
|
public boolean hasTableName() {
|
||||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||||
|
@ -2715,7 +2715,7 @@ public final class WALProtos {
|
||||||
return this;
|
return this;
|
||||||
}
|
}
|
||||||
|
|
||||||
// required bytes encodedRegionName = 2;
|
// required bytes encoded_region_name = 2;
|
||||||
private com.google.protobuf.ByteString encodedRegionName_ = com.google.protobuf.ByteString.EMPTY;
|
private com.google.protobuf.ByteString encodedRegionName_ = com.google.protobuf.ByteString.EMPTY;
|
||||||
public boolean hasEncodedRegionName() {
|
public boolean hasEncodedRegionName() {
|
||||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||||
|
@ -2739,7 +2739,7 @@ public final class WALProtos {
|
||||||
return this;
|
return this;
|
||||||
}
|
}
|
||||||
|
|
||||||
// required bytes familyName = 3;
|
// required bytes family_name = 3;
|
||||||
private com.google.protobuf.ByteString familyName_ = com.google.protobuf.ByteString.EMPTY;
|
private com.google.protobuf.ByteString familyName_ = com.google.protobuf.ByteString.EMPTY;
|
||||||
public boolean hasFamilyName() {
|
public boolean hasFamilyName() {
|
||||||
return ((bitField0_ & 0x00000004) == 0x00000004);
|
return ((bitField0_ & 0x00000004) == 0x00000004);
|
||||||
|
@ -2763,7 +2763,7 @@ public final class WALProtos {
|
||||||
return this;
|
return this;
|
||||||
}
|
}
|
||||||
|
|
||||||
// repeated string compactionInput = 4;
|
// repeated string compaction_input = 4;
|
||||||
private com.google.protobuf.LazyStringList compactionInput_ = com.google.protobuf.LazyStringArrayList.EMPTY;
|
private com.google.protobuf.LazyStringList compactionInput_ = com.google.protobuf.LazyStringArrayList.EMPTY;
|
||||||
private void ensureCompactionInputIsMutable() {
|
private void ensureCompactionInputIsMutable() {
|
||||||
if (!((bitField0_ & 0x00000008) == 0x00000008)) {
|
if (!((bitField0_ & 0x00000008) == 0x00000008)) {
|
||||||
|
@ -2819,7 +2819,7 @@ public final class WALProtos {
|
||||||
onChanged();
|
onChanged();
|
||||||
}
|
}
|
||||||
|
|
||||||
// repeated string compactionOutput = 5;
|
// repeated string compaction_output = 5;
|
||||||
private com.google.protobuf.LazyStringList compactionOutput_ = com.google.protobuf.LazyStringArrayList.EMPTY;
|
private com.google.protobuf.LazyStringList compactionOutput_ = com.google.protobuf.LazyStringArrayList.EMPTY;
|
||||||
private void ensureCompactionOutputIsMutable() {
|
private void ensureCompactionOutputIsMutable() {
|
||||||
if (!((bitField0_ & 0x00000010) == 0x00000010)) {
|
if (!((bitField0_ & 0x00000010) == 0x00000010)) {
|
||||||
|
@ -2875,7 +2875,7 @@ public final class WALProtos {
|
||||||
onChanged();
|
onChanged();
|
||||||
}
|
}
|
||||||
|
|
||||||
// required string storeHomeDir = 6;
|
// required string store_home_dir = 6;
|
||||||
private java.lang.Object storeHomeDir_ = "";
|
private java.lang.Object storeHomeDir_ = "";
|
||||||
public boolean hasStoreHomeDir() {
|
public boolean hasStoreHomeDir() {
|
||||||
return ((bitField0_ & 0x00000020) == 0x00000020);
|
return ((bitField0_ & 0x00000020) == 0x00000020);
|
||||||
|
@ -3256,22 +3256,23 @@ public final class WALProtos {
|
||||||
descriptor;
|
descriptor;
|
||||||
static {
|
static {
|
||||||
java.lang.String[] descriptorData = {
|
java.lang.String[] descriptorData = {
|
||||||
"\n\tWAL.proto\032\013hbase.proto\"#\n\tWALHeader\022\026\n" +
|
"\n\tWAL.proto\032\013hbase.proto\"$\n\tWALHeader\022\027\n" +
|
||||||
"\016hasCompression\030\001 \001(\010\"\266\001\n\006WALKey\022\031\n\021enco" +
|
"\017has_compression\030\001 \001(\010\"\277\001\n\006WALKey\022\033\n\023enc" +
|
||||||
"dedRegionName\030\001 \002(\014\022\021\n\ttableName\030\002 \002(\014\022\031" +
|
"oded_region_name\030\001 \002(\014\022\022\n\ntable_name\030\002 \002" +
|
||||||
"\n\021logSequenceNumber\030\003 \002(\004\022\021\n\twriteTime\030\004" +
|
"(\014\022\033\n\023log_sequence_number\030\003 \002(\004\022\022\n\nwrite" +
|
||||||
" \002(\004\022\030\n\tclusterId\030\005 \001(\0132\005.UUID\022\034\n\006scopes" +
|
"_time\030\004 \002(\004\022\031\n\ncluster_id\030\005 \001(\0132\005.UUID\022\034" +
|
||||||
"\030\006 \003(\0132\014.FamilyScope\022\030\n\020followingKvCount" +
|
"\n\006scopes\030\006 \003(\0132\014.FamilyScope\022\032\n\022followin" +
|
||||||
"\030\007 \001(\r\"<\n\013FamilyScope\022\016\n\006family\030\001 \002(\014\022\035\n" +
|
"g_kv_count\030\007 \001(\r\"=\n\013FamilyScope\022\016\n\006famil" +
|
||||||
"\tscopeType\030\002 \002(\0162\n.ScopeType\"\241\001\n\024Compact" +
|
"y\030\001 \002(\014\022\036\n\nscope_type\030\002 \002(\0162\n.ScopeType\"" +
|
||||||
"ionDescriptor\022\021\n\ttableName\030\001 \002(\014\022\031\n\021enco" +
|
"\251\001\n\024CompactionDescriptor\022\022\n\ntable_name\030\001" +
|
||||||
"dedRegionName\030\002 \002(\014\022\022\n\nfamilyName\030\003 \002(\014\022",
|
" \002(\014\022\033\n\023encoded_region_name\030\002 \002(\014\022\023\n\013fam",
|
||||||
"\027\n\017compactionInput\030\004 \003(\t\022\030\n\020compactionOu" +
|
"ily_name\030\003 \002(\014\022\030\n\020compaction_input\030\004 \003(\t" +
|
||||||
"tput\030\005 \003(\t\022\024\n\014storeHomeDir\030\006 \002(\t\"\014\n\nWALT" +
|
"\022\031\n\021compaction_output\030\005 \003(\t\022\026\n\016store_hom" +
|
||||||
"railer*F\n\tScopeType\022\033\n\027REPLICATION_SCOPE" +
|
"e_dir\030\006 \002(\t\"\014\n\nWALTrailer*F\n\tScopeType\022\033" +
|
||||||
"_LOCAL\020\000\022\034\n\030REPLICATION_SCOPE_GLOBAL\020\001B?" +
|
"\n\027REPLICATION_SCOPE_LOCAL\020\000\022\034\n\030REPLICATI" +
|
||||||
"\n*org.apache.hadoop.hbase.protobuf.gener" +
|
"ON_SCOPE_GLOBAL\020\001B?\n*org.apache.hadoop.h" +
|
||||||
"atedB\tWALProtosH\001\210\001\000\240\001\001"
|
"base.protobuf.generatedB\tWALProtosH\001\210\001\000\240" +
|
||||||
|
"\001\001"
|
||||||
};
|
};
|
||||||
com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
|
com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
|
||||||
new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() {
|
new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() {
|
||||||
|
|
|
@ -16,7 +16,7 @@ public final class ZooKeeperProtos {
|
||||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName getServer();
|
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName getServer();
|
||||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerNameOrBuilder getServerOrBuilder();
|
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerNameOrBuilder getServerOrBuilder();
|
||||||
|
|
||||||
// optional uint32 rpcVersion = 2;
|
// optional uint32 rpc_version = 2;
|
||||||
boolean hasRpcVersion();
|
boolean hasRpcVersion();
|
||||||
int getRpcVersion();
|
int getRpcVersion();
|
||||||
}
|
}
|
||||||
|
@ -62,8 +62,8 @@ public final class ZooKeeperProtos {
|
||||||
return server_;
|
return server_;
|
||||||
}
|
}
|
||||||
|
|
||||||
// optional uint32 rpcVersion = 2;
|
// optional uint32 rpc_version = 2;
|
||||||
public static final int RPCVERSION_FIELD_NUMBER = 2;
|
public static final int RPC_VERSION_FIELD_NUMBER = 2;
|
||||||
private int rpcVersion_;
|
private int rpcVersion_;
|
||||||
public boolean hasRpcVersion() {
|
public boolean hasRpcVersion() {
|
||||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||||
|
@ -166,7 +166,7 @@ public final class ZooKeeperProtos {
|
||||||
hash = (53 * hash) + getServer().hashCode();
|
hash = (53 * hash) + getServer().hashCode();
|
||||||
}
|
}
|
||||||
if (hasRpcVersion()) {
|
if (hasRpcVersion()) {
|
||||||
hash = (37 * hash) + RPCVERSION_FIELD_NUMBER;
|
hash = (37 * hash) + RPC_VERSION_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + getRpcVersion();
|
hash = (53 * hash) + getRpcVersion();
|
||||||
}
|
}
|
||||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||||
|
@ -515,7 +515,7 @@ public final class ZooKeeperProtos {
|
||||||
return serverBuilder_;
|
return serverBuilder_;
|
||||||
}
|
}
|
||||||
|
|
||||||
// optional uint32 rpcVersion = 2;
|
// optional uint32 rpc_version = 2;
|
||||||
private int rpcVersion_ ;
|
private int rpcVersion_ ;
|
||||||
public boolean hasRpcVersion() {
|
public boolean hasRpcVersion() {
|
||||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||||
|
@ -555,7 +555,7 @@ public final class ZooKeeperProtos {
|
||||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName getMaster();
|
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName getMaster();
|
||||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerNameOrBuilder getMasterOrBuilder();
|
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerNameOrBuilder getMasterOrBuilder();
|
||||||
|
|
||||||
// optional uint32 rpcVersion = 2;
|
// optional uint32 rpc_version = 2;
|
||||||
boolean hasRpcVersion();
|
boolean hasRpcVersion();
|
||||||
int getRpcVersion();
|
int getRpcVersion();
|
||||||
}
|
}
|
||||||
|
@ -601,8 +601,8 @@ public final class ZooKeeperProtos {
|
||||||
return master_;
|
return master_;
|
||||||
}
|
}
|
||||||
|
|
||||||
// optional uint32 rpcVersion = 2;
|
// optional uint32 rpc_version = 2;
|
||||||
public static final int RPCVERSION_FIELD_NUMBER = 2;
|
public static final int RPC_VERSION_FIELD_NUMBER = 2;
|
||||||
private int rpcVersion_;
|
private int rpcVersion_;
|
||||||
public boolean hasRpcVersion() {
|
public boolean hasRpcVersion() {
|
||||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||||
|
@ -705,7 +705,7 @@ public final class ZooKeeperProtos {
|
||||||
hash = (53 * hash) + getMaster().hashCode();
|
hash = (53 * hash) + getMaster().hashCode();
|
||||||
}
|
}
|
||||||
if (hasRpcVersion()) {
|
if (hasRpcVersion()) {
|
||||||
hash = (37 * hash) + RPCVERSION_FIELD_NUMBER;
|
hash = (37 * hash) + RPC_VERSION_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + getRpcVersion();
|
hash = (53 * hash) + getRpcVersion();
|
||||||
}
|
}
|
||||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||||
|
@ -1054,7 +1054,7 @@ public final class ZooKeeperProtos {
|
||||||
return masterBuilder_;
|
return masterBuilder_;
|
||||||
}
|
}
|
||||||
|
|
||||||
// optional uint32 rpcVersion = 2;
|
// optional uint32 rpc_version = 2;
|
||||||
private int rpcVersion_ ;
|
private int rpcVersion_ ;
|
||||||
public boolean hasRpcVersion() {
|
public boolean hasRpcVersion() {
|
||||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||||
|
@ -1089,7 +1089,7 @@ public final class ZooKeeperProtos {
|
||||||
public interface ClusterUpOrBuilder
|
public interface ClusterUpOrBuilder
|
||||||
extends com.google.protobuf.MessageOrBuilder {
|
extends com.google.protobuf.MessageOrBuilder {
|
||||||
|
|
||||||
// required string startDate = 1;
|
// required string start_date = 1;
|
||||||
boolean hasStartDate();
|
boolean hasStartDate();
|
||||||
String getStartDate();
|
String getStartDate();
|
||||||
}
|
}
|
||||||
|
@ -1122,8 +1122,8 @@ public final class ZooKeeperProtos {
|
||||||
}
|
}
|
||||||
|
|
||||||
private int bitField0_;
|
private int bitField0_;
|
||||||
// required string startDate = 1;
|
// required string start_date = 1;
|
||||||
public static final int STARTDATE_FIELD_NUMBER = 1;
|
public static final int START_DATE_FIELD_NUMBER = 1;
|
||||||
private java.lang.Object startDate_;
|
private java.lang.Object startDate_;
|
||||||
public boolean hasStartDate() {
|
public boolean hasStartDate() {
|
||||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||||
|
@ -1227,7 +1227,7 @@ public final class ZooKeeperProtos {
|
||||||
int hash = 41;
|
int hash = 41;
|
||||||
hash = (19 * hash) + getDescriptorForType().hashCode();
|
hash = (19 * hash) + getDescriptorForType().hashCode();
|
||||||
if (hasStartDate()) {
|
if (hasStartDate()) {
|
||||||
hash = (37 * hash) + STARTDATE_FIELD_NUMBER;
|
hash = (37 * hash) + START_DATE_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + getStartDate().hashCode();
|
hash = (53 * hash) + getStartDate().hashCode();
|
||||||
}
|
}
|
||||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||||
|
@ -1455,7 +1455,7 @@ public final class ZooKeeperProtos {
|
||||||
|
|
||||||
private int bitField0_;
|
private int bitField0_;
|
||||||
|
|
||||||
// required string startDate = 1;
|
// required string start_date = 1;
|
||||||
private java.lang.Object startDate_ = "";
|
private java.lang.Object startDate_ = "";
|
||||||
public boolean hasStartDate() {
|
public boolean hasStartDate() {
|
||||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||||
|
@ -1505,19 +1505,19 @@ public final class ZooKeeperProtos {
|
||||||
public interface RegionTransitionOrBuilder
|
public interface RegionTransitionOrBuilder
|
||||||
extends com.google.protobuf.MessageOrBuilder {
|
extends com.google.protobuf.MessageOrBuilder {
|
||||||
|
|
||||||
// required uint32 eventTypeCode = 1;
|
// required uint32 event_type_code = 1;
|
||||||
boolean hasEventTypeCode();
|
boolean hasEventTypeCode();
|
||||||
int getEventTypeCode();
|
int getEventTypeCode();
|
||||||
|
|
||||||
// required bytes regionName = 2;
|
// required bytes region_name = 2;
|
||||||
boolean hasRegionName();
|
boolean hasRegionName();
|
||||||
com.google.protobuf.ByteString getRegionName();
|
com.google.protobuf.ByteString getRegionName();
|
||||||
|
|
||||||
// required uint64 createTime = 3;
|
// required uint64 create_time = 3;
|
||||||
boolean hasCreateTime();
|
boolean hasCreateTime();
|
||||||
long getCreateTime();
|
long getCreateTime();
|
||||||
|
|
||||||
// required .ServerName serverName = 4;
|
// required .ServerName server_name = 4;
|
||||||
boolean hasServerName();
|
boolean hasServerName();
|
||||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName getServerName();
|
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName getServerName();
|
||||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerNameOrBuilder getServerNameOrBuilder();
|
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerNameOrBuilder getServerNameOrBuilder();
|
||||||
|
@ -1555,8 +1555,8 @@ public final class ZooKeeperProtos {
|
||||||
}
|
}
|
||||||
|
|
||||||
private int bitField0_;
|
private int bitField0_;
|
||||||
// required uint32 eventTypeCode = 1;
|
// required uint32 event_type_code = 1;
|
||||||
public static final int EVENTTYPECODE_FIELD_NUMBER = 1;
|
public static final int EVENT_TYPE_CODE_FIELD_NUMBER = 1;
|
||||||
private int eventTypeCode_;
|
private int eventTypeCode_;
|
||||||
public boolean hasEventTypeCode() {
|
public boolean hasEventTypeCode() {
|
||||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||||
|
@ -1565,8 +1565,8 @@ public final class ZooKeeperProtos {
|
||||||
return eventTypeCode_;
|
return eventTypeCode_;
|
||||||
}
|
}
|
||||||
|
|
||||||
// required bytes regionName = 2;
|
// required bytes region_name = 2;
|
||||||
public static final int REGIONNAME_FIELD_NUMBER = 2;
|
public static final int REGION_NAME_FIELD_NUMBER = 2;
|
||||||
private com.google.protobuf.ByteString regionName_;
|
private com.google.protobuf.ByteString regionName_;
|
||||||
public boolean hasRegionName() {
|
public boolean hasRegionName() {
|
||||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||||
|
@ -1575,8 +1575,8 @@ public final class ZooKeeperProtos {
|
||||||
return regionName_;
|
return regionName_;
|
||||||
}
|
}
|
||||||
|
|
||||||
// required uint64 createTime = 3;
|
// required uint64 create_time = 3;
|
||||||
public static final int CREATETIME_FIELD_NUMBER = 3;
|
public static final int CREATE_TIME_FIELD_NUMBER = 3;
|
||||||
private long createTime_;
|
private long createTime_;
|
||||||
public boolean hasCreateTime() {
|
public boolean hasCreateTime() {
|
||||||
return ((bitField0_ & 0x00000004) == 0x00000004);
|
return ((bitField0_ & 0x00000004) == 0x00000004);
|
||||||
|
@ -1585,8 +1585,8 @@ public final class ZooKeeperProtos {
|
||||||
return createTime_;
|
return createTime_;
|
||||||
}
|
}
|
||||||
|
|
||||||
// required .ServerName serverName = 4;
|
// required .ServerName server_name = 4;
|
||||||
public static final int SERVERNAME_FIELD_NUMBER = 4;
|
public static final int SERVER_NAME_FIELD_NUMBER = 4;
|
||||||
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName serverName_;
|
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName serverName_;
|
||||||
public boolean hasServerName() {
|
public boolean hasServerName() {
|
||||||
return ((bitField0_ & 0x00000008) == 0x00000008);
|
return ((bitField0_ & 0x00000008) == 0x00000008);
|
||||||
|
@ -1749,19 +1749,19 @@ public final class ZooKeeperProtos {
|
||||||
int hash = 41;
|
int hash = 41;
|
||||||
hash = (19 * hash) + getDescriptorForType().hashCode();
|
hash = (19 * hash) + getDescriptorForType().hashCode();
|
||||||
if (hasEventTypeCode()) {
|
if (hasEventTypeCode()) {
|
||||||
hash = (37 * hash) + EVENTTYPECODE_FIELD_NUMBER;
|
hash = (37 * hash) + EVENT_TYPE_CODE_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + getEventTypeCode();
|
hash = (53 * hash) + getEventTypeCode();
|
||||||
}
|
}
|
||||||
if (hasRegionName()) {
|
if (hasRegionName()) {
|
||||||
hash = (37 * hash) + REGIONNAME_FIELD_NUMBER;
|
hash = (37 * hash) + REGION_NAME_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + getRegionName().hashCode();
|
hash = (53 * hash) + getRegionName().hashCode();
|
||||||
}
|
}
|
||||||
if (hasCreateTime()) {
|
if (hasCreateTime()) {
|
||||||
hash = (37 * hash) + CREATETIME_FIELD_NUMBER;
|
hash = (37 * hash) + CREATE_TIME_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + hashLong(getCreateTime());
|
hash = (53 * hash) + hashLong(getCreateTime());
|
||||||
}
|
}
|
||||||
if (hasServerName()) {
|
if (hasServerName()) {
|
||||||
hash = (37 * hash) + SERVERNAME_FIELD_NUMBER;
|
hash = (37 * hash) + SERVER_NAME_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + getServerName().hashCode();
|
hash = (53 * hash) + getServerName().hashCode();
|
||||||
}
|
}
|
||||||
if (hasPayload()) {
|
if (hasPayload()) {
|
||||||
|
@ -2078,7 +2078,7 @@ public final class ZooKeeperProtos {
|
||||||
|
|
||||||
private int bitField0_;
|
private int bitField0_;
|
||||||
|
|
||||||
// required uint32 eventTypeCode = 1;
|
// required uint32 event_type_code = 1;
|
||||||
private int eventTypeCode_ ;
|
private int eventTypeCode_ ;
|
||||||
public boolean hasEventTypeCode() {
|
public boolean hasEventTypeCode() {
|
||||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||||
|
@ -2099,7 +2099,7 @@ public final class ZooKeeperProtos {
|
||||||
return this;
|
return this;
|
||||||
}
|
}
|
||||||
|
|
||||||
// required bytes regionName = 2;
|
// required bytes region_name = 2;
|
||||||
private com.google.protobuf.ByteString regionName_ = com.google.protobuf.ByteString.EMPTY;
|
private com.google.protobuf.ByteString regionName_ = com.google.protobuf.ByteString.EMPTY;
|
||||||
public boolean hasRegionName() {
|
public boolean hasRegionName() {
|
||||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||||
|
@ -2123,7 +2123,7 @@ public final class ZooKeeperProtos {
|
||||||
return this;
|
return this;
|
||||||
}
|
}
|
||||||
|
|
||||||
// required uint64 createTime = 3;
|
// required uint64 create_time = 3;
|
||||||
private long createTime_ ;
|
private long createTime_ ;
|
||||||
public boolean hasCreateTime() {
|
public boolean hasCreateTime() {
|
||||||
return ((bitField0_ & 0x00000004) == 0x00000004);
|
return ((bitField0_ & 0x00000004) == 0x00000004);
|
||||||
|
@ -2144,7 +2144,7 @@ public final class ZooKeeperProtos {
|
||||||
return this;
|
return this;
|
||||||
}
|
}
|
||||||
|
|
||||||
// required .ServerName serverName = 4;
|
// required .ServerName server_name = 4;
|
||||||
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName serverName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.getDefaultInstance();
|
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName serverName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.getDefaultInstance();
|
||||||
private com.google.protobuf.SingleFieldBuilder<
|
private com.google.protobuf.SingleFieldBuilder<
|
||||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerNameOrBuilder> serverNameBuilder_;
|
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerNameOrBuilder> serverNameBuilder_;
|
||||||
|
@ -2276,7 +2276,7 @@ public final class ZooKeeperProtos {
|
||||||
boolean hasState();
|
boolean hasState();
|
||||||
org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.SplitLogTask.State getState();
|
org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.SplitLogTask.State getState();
|
||||||
|
|
||||||
// required .ServerName serverName = 2;
|
// required .ServerName server_name = 2;
|
||||||
boolean hasServerName();
|
boolean hasServerName();
|
||||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName getServerName();
|
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName getServerName();
|
||||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerNameOrBuilder getServerNameOrBuilder();
|
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerNameOrBuilder getServerNameOrBuilder();
|
||||||
|
@ -2398,8 +2398,8 @@ public final class ZooKeeperProtos {
|
||||||
return state_;
|
return state_;
|
||||||
}
|
}
|
||||||
|
|
||||||
// required .ServerName serverName = 2;
|
// required .ServerName server_name = 2;
|
||||||
public static final int SERVERNAME_FIELD_NUMBER = 2;
|
public static final int SERVER_NAME_FIELD_NUMBER = 2;
|
||||||
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName serverName_;
|
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName serverName_;
|
||||||
public boolean hasServerName() {
|
public boolean hasServerName() {
|
||||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||||
|
@ -2509,7 +2509,7 @@ public final class ZooKeeperProtos {
|
||||||
hash = (53 * hash) + hashEnum(getState());
|
hash = (53 * hash) + hashEnum(getState());
|
||||||
}
|
}
|
||||||
if (hasServerName()) {
|
if (hasServerName()) {
|
||||||
hash = (37 * hash) + SERVERNAME_FIELD_NUMBER;
|
hash = (37 * hash) + SERVER_NAME_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + getServerName().hashCode();
|
hash = (53 * hash) + getServerName().hashCode();
|
||||||
}
|
}
|
||||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||||
|
@ -2802,7 +2802,7 @@ public final class ZooKeeperProtos {
|
||||||
return this;
|
return this;
|
||||||
}
|
}
|
||||||
|
|
||||||
// required .ServerName serverName = 2;
|
// required .ServerName server_name = 2;
|
||||||
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName serverName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.getDefaultInstance();
|
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName serverName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.getDefaultInstance();
|
||||||
private com.google.protobuf.SingleFieldBuilder<
|
private com.google.protobuf.SingleFieldBuilder<
|
||||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerNameOrBuilder> serverNameBuilder_;
|
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerNameOrBuilder> serverNameBuilder_;
|
||||||
|
@ -4621,7 +4621,7 @@ public final class ZooKeeperProtos {
|
||||||
public interface ReplicationLockOrBuilder
|
public interface ReplicationLockOrBuilder
|
||||||
extends com.google.protobuf.MessageOrBuilder {
|
extends com.google.protobuf.MessageOrBuilder {
|
||||||
|
|
||||||
// required string lockOwner = 1;
|
// required string lock_owner = 1;
|
||||||
boolean hasLockOwner();
|
boolean hasLockOwner();
|
||||||
String getLockOwner();
|
String getLockOwner();
|
||||||
}
|
}
|
||||||
|
@ -4654,8 +4654,8 @@ public final class ZooKeeperProtos {
|
||||||
}
|
}
|
||||||
|
|
||||||
private int bitField0_;
|
private int bitField0_;
|
||||||
// required string lockOwner = 1;
|
// required string lock_owner = 1;
|
||||||
public static final int LOCKOWNER_FIELD_NUMBER = 1;
|
public static final int LOCK_OWNER_FIELD_NUMBER = 1;
|
||||||
private java.lang.Object lockOwner_;
|
private java.lang.Object lockOwner_;
|
||||||
public boolean hasLockOwner() {
|
public boolean hasLockOwner() {
|
||||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||||
|
@ -4759,7 +4759,7 @@ public final class ZooKeeperProtos {
|
||||||
int hash = 41;
|
int hash = 41;
|
||||||
hash = (19 * hash) + getDescriptorForType().hashCode();
|
hash = (19 * hash) + getDescriptorForType().hashCode();
|
||||||
if (hasLockOwner()) {
|
if (hasLockOwner()) {
|
||||||
hash = (37 * hash) + LOCKOWNER_FIELD_NUMBER;
|
hash = (37 * hash) + LOCK_OWNER_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + getLockOwner().hashCode();
|
hash = (53 * hash) + getLockOwner().hashCode();
|
||||||
}
|
}
|
||||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||||
|
@ -4987,7 +4987,7 @@ public final class ZooKeeperProtos {
|
||||||
|
|
||||||
private int bitField0_;
|
private int bitField0_;
|
||||||
|
|
||||||
// required string lockOwner = 1;
|
// required string lock_owner = 1;
|
||||||
private java.lang.Object lockOwner_ = "";
|
private java.lang.Object lockOwner_ = "";
|
||||||
public boolean hasLockOwner() {
|
public boolean hasLockOwner() {
|
||||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||||
|
@ -5037,20 +5037,20 @@ public final class ZooKeeperProtos {
|
||||||
public interface TableLockOrBuilder
|
public interface TableLockOrBuilder
|
||||||
extends com.google.protobuf.MessageOrBuilder {
|
extends com.google.protobuf.MessageOrBuilder {
|
||||||
|
|
||||||
// optional bytes tableName = 1;
|
// optional bytes table_name = 1;
|
||||||
boolean hasTableName();
|
boolean hasTableName();
|
||||||
com.google.protobuf.ByteString getTableName();
|
com.google.protobuf.ByteString getTableName();
|
||||||
|
|
||||||
// optional .ServerName lockOwner = 2;
|
// optional .ServerName lock_owner = 2;
|
||||||
boolean hasLockOwner();
|
boolean hasLockOwner();
|
||||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName getLockOwner();
|
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName getLockOwner();
|
||||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerNameOrBuilder getLockOwnerOrBuilder();
|
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerNameOrBuilder getLockOwnerOrBuilder();
|
||||||
|
|
||||||
// optional int64 threadId = 3;
|
// optional int64 thread_id = 3;
|
||||||
boolean hasThreadId();
|
boolean hasThreadId();
|
||||||
long getThreadId();
|
long getThreadId();
|
||||||
|
|
||||||
// optional bool isShared = 4;
|
// optional bool is_shared = 4;
|
||||||
boolean hasIsShared();
|
boolean hasIsShared();
|
||||||
boolean getIsShared();
|
boolean getIsShared();
|
||||||
|
|
||||||
|
@ -5058,7 +5058,7 @@ public final class ZooKeeperProtos {
|
||||||
boolean hasPurpose();
|
boolean hasPurpose();
|
||||||
String getPurpose();
|
String getPurpose();
|
||||||
|
|
||||||
// optional int64 createTime = 6;
|
// optional int64 create_time = 6;
|
||||||
boolean hasCreateTime();
|
boolean hasCreateTime();
|
||||||
long getCreateTime();
|
long getCreateTime();
|
||||||
}
|
}
|
||||||
|
@ -5091,8 +5091,8 @@ public final class ZooKeeperProtos {
|
||||||
}
|
}
|
||||||
|
|
||||||
private int bitField0_;
|
private int bitField0_;
|
||||||
// optional bytes tableName = 1;
|
// optional bytes table_name = 1;
|
||||||
public static final int TABLENAME_FIELD_NUMBER = 1;
|
public static final int TABLE_NAME_FIELD_NUMBER = 1;
|
||||||
private com.google.protobuf.ByteString tableName_;
|
private com.google.protobuf.ByteString tableName_;
|
||||||
public boolean hasTableName() {
|
public boolean hasTableName() {
|
||||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||||
|
@ -5101,8 +5101,8 @@ public final class ZooKeeperProtos {
|
||||||
return tableName_;
|
return tableName_;
|
||||||
}
|
}
|
||||||
|
|
||||||
// optional .ServerName lockOwner = 2;
|
// optional .ServerName lock_owner = 2;
|
||||||
public static final int LOCKOWNER_FIELD_NUMBER = 2;
|
public static final int LOCK_OWNER_FIELD_NUMBER = 2;
|
||||||
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName lockOwner_;
|
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName lockOwner_;
|
||||||
public boolean hasLockOwner() {
|
public boolean hasLockOwner() {
|
||||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||||
|
@ -5114,8 +5114,8 @@ public final class ZooKeeperProtos {
|
||||||
return lockOwner_;
|
return lockOwner_;
|
||||||
}
|
}
|
||||||
|
|
||||||
// optional int64 threadId = 3;
|
// optional int64 thread_id = 3;
|
||||||
public static final int THREADID_FIELD_NUMBER = 3;
|
public static final int THREAD_ID_FIELD_NUMBER = 3;
|
||||||
private long threadId_;
|
private long threadId_;
|
||||||
public boolean hasThreadId() {
|
public boolean hasThreadId() {
|
||||||
return ((bitField0_ & 0x00000004) == 0x00000004);
|
return ((bitField0_ & 0x00000004) == 0x00000004);
|
||||||
|
@ -5124,8 +5124,8 @@ public final class ZooKeeperProtos {
|
||||||
return threadId_;
|
return threadId_;
|
||||||
}
|
}
|
||||||
|
|
||||||
// optional bool isShared = 4;
|
// optional bool is_shared = 4;
|
||||||
public static final int ISSHARED_FIELD_NUMBER = 4;
|
public static final int IS_SHARED_FIELD_NUMBER = 4;
|
||||||
private boolean isShared_;
|
private boolean isShared_;
|
||||||
public boolean hasIsShared() {
|
public boolean hasIsShared() {
|
||||||
return ((bitField0_ & 0x00000008) == 0x00000008);
|
return ((bitField0_ & 0x00000008) == 0x00000008);
|
||||||
|
@ -5166,8 +5166,8 @@ public final class ZooKeeperProtos {
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// optional int64 createTime = 6;
|
// optional int64 create_time = 6;
|
||||||
public static final int CREATETIME_FIELD_NUMBER = 6;
|
public static final int CREATE_TIME_FIELD_NUMBER = 6;
|
||||||
private long createTime_;
|
private long createTime_;
|
||||||
public boolean hasCreateTime() {
|
public boolean hasCreateTime() {
|
||||||
return ((bitField0_ & 0x00000020) == 0x00000020);
|
return ((bitField0_ & 0x00000020) == 0x00000020);
|
||||||
|
@ -5316,19 +5316,19 @@ public final class ZooKeeperProtos {
|
||||||
int hash = 41;
|
int hash = 41;
|
||||||
hash = (19 * hash) + getDescriptorForType().hashCode();
|
hash = (19 * hash) + getDescriptorForType().hashCode();
|
||||||
if (hasTableName()) {
|
if (hasTableName()) {
|
||||||
hash = (37 * hash) + TABLENAME_FIELD_NUMBER;
|
hash = (37 * hash) + TABLE_NAME_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + getTableName().hashCode();
|
hash = (53 * hash) + getTableName().hashCode();
|
||||||
}
|
}
|
||||||
if (hasLockOwner()) {
|
if (hasLockOwner()) {
|
||||||
hash = (37 * hash) + LOCKOWNER_FIELD_NUMBER;
|
hash = (37 * hash) + LOCK_OWNER_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + getLockOwner().hashCode();
|
hash = (53 * hash) + getLockOwner().hashCode();
|
||||||
}
|
}
|
||||||
if (hasThreadId()) {
|
if (hasThreadId()) {
|
||||||
hash = (37 * hash) + THREADID_FIELD_NUMBER;
|
hash = (37 * hash) + THREAD_ID_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + hashLong(getThreadId());
|
hash = (53 * hash) + hashLong(getThreadId());
|
||||||
}
|
}
|
||||||
if (hasIsShared()) {
|
if (hasIsShared()) {
|
||||||
hash = (37 * hash) + ISSHARED_FIELD_NUMBER;
|
hash = (37 * hash) + IS_SHARED_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + hashBoolean(getIsShared());
|
hash = (53 * hash) + hashBoolean(getIsShared());
|
||||||
}
|
}
|
||||||
if (hasPurpose()) {
|
if (hasPurpose()) {
|
||||||
|
@ -5336,7 +5336,7 @@ public final class ZooKeeperProtos {
|
||||||
hash = (53 * hash) + getPurpose().hashCode();
|
hash = (53 * hash) + getPurpose().hashCode();
|
||||||
}
|
}
|
||||||
if (hasCreateTime()) {
|
if (hasCreateTime()) {
|
||||||
hash = (37 * hash) + CREATETIME_FIELD_NUMBER;
|
hash = (37 * hash) + CREATE_TIME_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + hashLong(getCreateTime());
|
hash = (53 * hash) + hashLong(getCreateTime());
|
||||||
}
|
}
|
||||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||||
|
@ -5649,7 +5649,7 @@ public final class ZooKeeperProtos {
|
||||||
|
|
||||||
private int bitField0_;
|
private int bitField0_;
|
||||||
|
|
||||||
// optional bytes tableName = 1;
|
// optional bytes table_name = 1;
|
||||||
private com.google.protobuf.ByteString tableName_ = com.google.protobuf.ByteString.EMPTY;
|
private com.google.protobuf.ByteString tableName_ = com.google.protobuf.ByteString.EMPTY;
|
||||||
public boolean hasTableName() {
|
public boolean hasTableName() {
|
||||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||||
|
@ -5673,7 +5673,7 @@ public final class ZooKeeperProtos {
|
||||||
return this;
|
return this;
|
||||||
}
|
}
|
||||||
|
|
||||||
// optional .ServerName lockOwner = 2;
|
// optional .ServerName lock_owner = 2;
|
||||||
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName lockOwner_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.getDefaultInstance();
|
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName lockOwner_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.getDefaultInstance();
|
||||||
private com.google.protobuf.SingleFieldBuilder<
|
private com.google.protobuf.SingleFieldBuilder<
|
||||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerNameOrBuilder> lockOwnerBuilder_;
|
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerNameOrBuilder> lockOwnerBuilder_;
|
||||||
|
@ -5763,7 +5763,7 @@ public final class ZooKeeperProtos {
|
||||||
return lockOwnerBuilder_;
|
return lockOwnerBuilder_;
|
||||||
}
|
}
|
||||||
|
|
||||||
// optional int64 threadId = 3;
|
// optional int64 thread_id = 3;
|
||||||
private long threadId_ ;
|
private long threadId_ ;
|
||||||
public boolean hasThreadId() {
|
public boolean hasThreadId() {
|
||||||
return ((bitField0_ & 0x00000004) == 0x00000004);
|
return ((bitField0_ & 0x00000004) == 0x00000004);
|
||||||
|
@ -5784,7 +5784,7 @@ public final class ZooKeeperProtos {
|
||||||
return this;
|
return this;
|
||||||
}
|
}
|
||||||
|
|
||||||
// optional bool isShared = 4;
|
// optional bool is_shared = 4;
|
||||||
private boolean isShared_ ;
|
private boolean isShared_ ;
|
||||||
public boolean hasIsShared() {
|
public boolean hasIsShared() {
|
||||||
return ((bitField0_ & 0x00000008) == 0x00000008);
|
return ((bitField0_ & 0x00000008) == 0x00000008);
|
||||||
|
@ -5841,7 +5841,7 @@ public final class ZooKeeperProtos {
|
||||||
onChanged();
|
onChanged();
|
||||||
}
|
}
|
||||||
|
|
||||||
// optional int64 createTime = 6;
|
// optional int64 create_time = 6;
|
||||||
private long createTime_ ;
|
private long createTime_ ;
|
||||||
public boolean hasCreateTime() {
|
public boolean hasCreateTime() {
|
||||||
return ((bitField0_ & 0x00000020) == 0x00000020);
|
return ((bitField0_ & 0x00000020) == 0x00000020);
|
||||||
|
@ -5876,11 +5876,11 @@ public final class ZooKeeperProtos {
|
||||||
public interface StoreSequenceIdOrBuilder
|
public interface StoreSequenceIdOrBuilder
|
||||||
extends com.google.protobuf.MessageOrBuilder {
|
extends com.google.protobuf.MessageOrBuilder {
|
||||||
|
|
||||||
// required bytes familyName = 1;
|
// required bytes family_name = 1;
|
||||||
boolean hasFamilyName();
|
boolean hasFamilyName();
|
||||||
com.google.protobuf.ByteString getFamilyName();
|
com.google.protobuf.ByteString getFamilyName();
|
||||||
|
|
||||||
// required uint64 sequenceId = 2;
|
// required uint64 sequence_id = 2;
|
||||||
boolean hasSequenceId();
|
boolean hasSequenceId();
|
||||||
long getSequenceId();
|
long getSequenceId();
|
||||||
}
|
}
|
||||||
|
@ -5913,8 +5913,8 @@ public final class ZooKeeperProtos {
|
||||||
}
|
}
|
||||||
|
|
||||||
private int bitField0_;
|
private int bitField0_;
|
||||||
// required bytes familyName = 1;
|
// required bytes family_name = 1;
|
||||||
public static final int FAMILYNAME_FIELD_NUMBER = 1;
|
public static final int FAMILY_NAME_FIELD_NUMBER = 1;
|
||||||
private com.google.protobuf.ByteString familyName_;
|
private com.google.protobuf.ByteString familyName_;
|
||||||
public boolean hasFamilyName() {
|
public boolean hasFamilyName() {
|
||||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||||
|
@ -5923,8 +5923,8 @@ public final class ZooKeeperProtos {
|
||||||
return familyName_;
|
return familyName_;
|
||||||
}
|
}
|
||||||
|
|
||||||
// required uint64 sequenceId = 2;
|
// required uint64 sequence_id = 2;
|
||||||
public static final int SEQUENCEID_FIELD_NUMBER = 2;
|
public static final int SEQUENCE_ID_FIELD_NUMBER = 2;
|
||||||
private long sequenceId_;
|
private long sequenceId_;
|
||||||
public boolean hasSequenceId() {
|
public boolean hasSequenceId() {
|
||||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||||
|
@ -6023,11 +6023,11 @@ public final class ZooKeeperProtos {
|
||||||
int hash = 41;
|
int hash = 41;
|
||||||
hash = (19 * hash) + getDescriptorForType().hashCode();
|
hash = (19 * hash) + getDescriptorForType().hashCode();
|
||||||
if (hasFamilyName()) {
|
if (hasFamilyName()) {
|
||||||
hash = (37 * hash) + FAMILYNAME_FIELD_NUMBER;
|
hash = (37 * hash) + FAMILY_NAME_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + getFamilyName().hashCode();
|
hash = (53 * hash) + getFamilyName().hashCode();
|
||||||
}
|
}
|
||||||
if (hasSequenceId()) {
|
if (hasSequenceId()) {
|
||||||
hash = (37 * hash) + SEQUENCEID_FIELD_NUMBER;
|
hash = (37 * hash) + SEQUENCE_ID_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + hashLong(getSequenceId());
|
hash = (53 * hash) + hashLong(getSequenceId());
|
||||||
}
|
}
|
||||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||||
|
@ -6273,7 +6273,7 @@ public final class ZooKeeperProtos {
|
||||||
|
|
||||||
private int bitField0_;
|
private int bitField0_;
|
||||||
|
|
||||||
// required bytes familyName = 1;
|
// required bytes family_name = 1;
|
||||||
private com.google.protobuf.ByteString familyName_ = com.google.protobuf.ByteString.EMPTY;
|
private com.google.protobuf.ByteString familyName_ = com.google.protobuf.ByteString.EMPTY;
|
||||||
public boolean hasFamilyName() {
|
public boolean hasFamilyName() {
|
||||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||||
|
@ -6297,7 +6297,7 @@ public final class ZooKeeperProtos {
|
||||||
return this;
|
return this;
|
||||||
}
|
}
|
||||||
|
|
||||||
// required uint64 sequenceId = 2;
|
// required uint64 sequence_id = 2;
|
||||||
private long sequenceId_ ;
|
private long sequenceId_ ;
|
||||||
public boolean hasSequenceId() {
|
public boolean hasSequenceId() {
|
||||||
return ((bitField0_ & 0x00000002) == 0x00000002);
|
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||||
|
@ -6332,11 +6332,11 @@ public final class ZooKeeperProtos {
|
||||||
public interface RegionStoreSequenceIdsOrBuilder
|
public interface RegionStoreSequenceIdsOrBuilder
|
||||||
extends com.google.protobuf.MessageOrBuilder {
|
extends com.google.protobuf.MessageOrBuilder {
|
||||||
|
|
||||||
// required uint64 lastFlushedSequenceId = 1;
|
// required uint64 last_flushed_sequence_id = 1;
|
||||||
boolean hasLastFlushedSequenceId();
|
boolean hasLastFlushedSequenceId();
|
||||||
long getLastFlushedSequenceId();
|
long getLastFlushedSequenceId();
|
||||||
|
|
||||||
// repeated .StoreSequenceId storeSequenceId = 2;
|
// repeated .StoreSequenceId store_sequence_id = 2;
|
||||||
java.util.List<org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.StoreSequenceId>
|
java.util.List<org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.StoreSequenceId>
|
||||||
getStoreSequenceIdList();
|
getStoreSequenceIdList();
|
||||||
org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.StoreSequenceId getStoreSequenceId(int index);
|
org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.StoreSequenceId getStoreSequenceId(int index);
|
||||||
|
@ -6375,8 +6375,8 @@ public final class ZooKeeperProtos {
|
||||||
}
|
}
|
||||||
|
|
||||||
private int bitField0_;
|
private int bitField0_;
|
||||||
// required uint64 lastFlushedSequenceId = 1;
|
// required uint64 last_flushed_sequence_id = 1;
|
||||||
public static final int LASTFLUSHEDSEQUENCEID_FIELD_NUMBER = 1;
|
public static final int LAST_FLUSHED_SEQUENCE_ID_FIELD_NUMBER = 1;
|
||||||
private long lastFlushedSequenceId_;
|
private long lastFlushedSequenceId_;
|
||||||
public boolean hasLastFlushedSequenceId() {
|
public boolean hasLastFlushedSequenceId() {
|
||||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||||
|
@ -6385,8 +6385,8 @@ public final class ZooKeeperProtos {
|
||||||
return lastFlushedSequenceId_;
|
return lastFlushedSequenceId_;
|
||||||
}
|
}
|
||||||
|
|
||||||
// repeated .StoreSequenceId storeSequenceId = 2;
|
// repeated .StoreSequenceId store_sequence_id = 2;
|
||||||
public static final int STORESEQUENCEID_FIELD_NUMBER = 2;
|
public static final int STORE_SEQUENCE_ID_FIELD_NUMBER = 2;
|
||||||
private java.util.List<org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.StoreSequenceId> storeSequenceId_;
|
private java.util.List<org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.StoreSequenceId> storeSequenceId_;
|
||||||
public java.util.List<org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.StoreSequenceId> getStoreSequenceIdList() {
|
public java.util.List<org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.StoreSequenceId> getStoreSequenceIdList() {
|
||||||
return storeSequenceId_;
|
return storeSequenceId_;
|
||||||
|
@ -6495,11 +6495,11 @@ public final class ZooKeeperProtos {
|
||||||
int hash = 41;
|
int hash = 41;
|
||||||
hash = (19 * hash) + getDescriptorForType().hashCode();
|
hash = (19 * hash) + getDescriptorForType().hashCode();
|
||||||
if (hasLastFlushedSequenceId()) {
|
if (hasLastFlushedSequenceId()) {
|
||||||
hash = (37 * hash) + LASTFLUSHEDSEQUENCEID_FIELD_NUMBER;
|
hash = (37 * hash) + LAST_FLUSHED_SEQUENCE_ID_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + hashLong(getLastFlushedSequenceId());
|
hash = (53 * hash) + hashLong(getLastFlushedSequenceId());
|
||||||
}
|
}
|
||||||
if (getStoreSequenceIdCount() > 0) {
|
if (getStoreSequenceIdCount() > 0) {
|
||||||
hash = (37 * hash) + STORESEQUENCEID_FIELD_NUMBER;
|
hash = (37 * hash) + STORE_SEQUENCE_ID_FIELD_NUMBER;
|
||||||
hash = (53 * hash) + getStoreSequenceIdList().hashCode();
|
hash = (53 * hash) + getStoreSequenceIdList().hashCode();
|
||||||
}
|
}
|
||||||
hash = (29 * hash) + getUnknownFields().hashCode();
|
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||||
|
@ -6781,7 +6781,7 @@ public final class ZooKeeperProtos {
|
||||||
|
|
||||||
private int bitField0_;
|
private int bitField0_;
|
||||||
|
|
||||||
// required uint64 lastFlushedSequenceId = 1;
|
// required uint64 last_flushed_sequence_id = 1;
|
||||||
private long lastFlushedSequenceId_ ;
|
private long lastFlushedSequenceId_ ;
|
||||||
public boolean hasLastFlushedSequenceId() {
|
public boolean hasLastFlushedSequenceId() {
|
||||||
return ((bitField0_ & 0x00000001) == 0x00000001);
|
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||||
|
@ -6802,7 +6802,7 @@ public final class ZooKeeperProtos {
|
||||||
return this;
|
return this;
|
||||||
}
|
}
|
||||||
|
|
||||||
// repeated .StoreSequenceId storeSequenceId = 2;
|
// repeated .StoreSequenceId store_sequence_id = 2;
|
||||||
private java.util.List<org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.StoreSequenceId> storeSequenceId_ =
|
private java.util.List<org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.StoreSequenceId> storeSequenceId_ =
|
||||||
java.util.Collections.emptyList();
|
java.util.Collections.emptyList();
|
||||||
private void ensureStoreSequenceIdIsMutable() {
|
private void ensureStoreSequenceIdIsMutable() {
|
||||||
|
@ -7073,36 +7073,36 @@ public final class ZooKeeperProtos {
|
||||||
descriptor;
|
descriptor;
|
||||||
static {
|
static {
|
||||||
java.lang.String[] descriptorData = {
|
java.lang.String[] descriptorData = {
|
||||||
"\n\017ZooKeeper.proto\032\013hbase.proto\"C\n\020MetaRe" +
|
"\n\017ZooKeeper.proto\032\013hbase.proto\"D\n\020MetaRe" +
|
||||||
"gionServer\022\033\n\006server\030\001 \002(\0132\013.ServerName\022" +
|
"gionServer\022\033\n\006server\030\001 \002(\0132\013.ServerName\022" +
|
||||||
"\022\n\nrpcVersion\030\002 \001(\r\"9\n\006Master\022\033\n\006master\030" +
|
"\023\n\013rpc_version\030\002 \001(\r\":\n\006Master\022\033\n\006master" +
|
||||||
"\001 \002(\0132\013.ServerName\022\022\n\nrpcVersion\030\002 \001(\r\"\036" +
|
"\030\001 \002(\0132\013.ServerName\022\023\n\013rpc_version\030\002 \001(\r" +
|
||||||
"\n\tClusterUp\022\021\n\tstartDate\030\001 \002(\t\"\203\001\n\020Regio" +
|
"\"\037\n\tClusterUp\022\022\n\nstart_date\030\001 \002(\t\"\210\001\n\020Re" +
|
||||||
"nTransition\022\025\n\reventTypeCode\030\001 \002(\r\022\022\n\nre" +
|
"gionTransition\022\027\n\017event_type_code\030\001 \002(\r\022" +
|
||||||
"gionName\030\002 \002(\014\022\022\n\ncreateTime\030\003 \002(\004\022\037\n\nse" +
|
"\023\n\013region_name\030\002 \002(\014\022\023\n\013create_time\030\003 \002(" +
|
||||||
"rverName\030\004 \002(\0132\013.ServerName\022\017\n\007payload\030\005" +
|
"\004\022 \n\013server_name\030\004 \002(\0132\013.ServerName\022\017\n\007p" +
|
||||||
" \001(\014\"\230\001\n\014SplitLogTask\022\"\n\005state\030\001 \002(\0162\023.S" +
|
"ayload\030\005 \001(\014\"\231\001\n\014SplitLogTask\022\"\n\005state\030\001" +
|
||||||
"plitLogTask.State\022\037\n\nserverName\030\002 \002(\0132\013.",
|
" \002(\0162\023.SplitLogTask.State\022 \n\013server_name",
|
||||||
"ServerName\"C\n\005State\022\016\n\nUNASSIGNED\020\000\022\t\n\005O" +
|
"\030\002 \002(\0132\013.ServerName\"C\n\005State\022\016\n\nUNASSIGN" +
|
||||||
"WNED\020\001\022\014\n\010RESIGNED\020\002\022\010\n\004DONE\020\003\022\007\n\003ERR\020\004\"" +
|
"ED\020\000\022\t\n\005OWNED\020\001\022\014\n\010RESIGNED\020\002\022\010\n\004DONE\020\003\022" +
|
||||||
"n\n\005Table\022$\n\005state\030\001 \002(\0162\014.Table.State:\007E" +
|
"\007\n\003ERR\020\004\"n\n\005Table\022$\n\005state\030\001 \002(\0162\014.Table" +
|
||||||
"NABLED\"?\n\005State\022\013\n\007ENABLED\020\000\022\014\n\010DISABLED" +
|
".State:\007ENABLED\"?\n\005State\022\013\n\007ENABLED\020\000\022\014\n" +
|
||||||
"\020\001\022\r\n\tDISABLING\020\002\022\014\n\010ENABLING\020\003\"%\n\017Repli" +
|
"\010DISABLED\020\001\022\r\n\tDISABLING\020\002\022\014\n\010ENABLING\020\003" +
|
||||||
"cationPeer\022\022\n\nclusterkey\030\001 \002(\t\"^\n\020Replic" +
|
"\"%\n\017ReplicationPeer\022\022\n\nclusterkey\030\001 \002(\t\"" +
|
||||||
"ationState\022&\n\005state\030\001 \002(\0162\027.ReplicationS" +
|
"^\n\020ReplicationState\022&\n\005state\030\001 \002(\0162\027.Rep" +
|
||||||
"tate.State\"\"\n\005State\022\013\n\007ENABLED\020\000\022\014\n\010DISA" +
|
"licationState.State\"\"\n\005State\022\013\n\007ENABLED\020" +
|
||||||
"BLED\020\001\"+\n\027ReplicationHLogPosition\022\020\n\010pos" +
|
"\000\022\014\n\010DISABLED\020\001\"+\n\027ReplicationHLogPositi" +
|
||||||
"ition\030\001 \002(\003\"$\n\017ReplicationLock\022\021\n\tlockOw",
|
"on\022\020\n\010position\030\001 \002(\003\"%\n\017ReplicationLock\022",
|
||||||
"ner\030\001 \002(\t\"\207\001\n\tTableLock\022\021\n\ttableName\030\001 \001" +
|
"\022\n\nlock_owner\030\001 \002(\t\"\214\001\n\tTableLock\022\022\n\ntab" +
|
||||||
"(\014\022\036\n\tlockOwner\030\002 \001(\0132\013.ServerName\022\020\n\010th" +
|
"le_name\030\001 \001(\014\022\037\n\nlock_owner\030\002 \001(\0132\013.Serv" +
|
||||||
"readId\030\003 \001(\003\022\020\n\010isShared\030\004 \001(\010\022\017\n\007purpos" +
|
"erName\022\021\n\tthread_id\030\003 \001(\003\022\021\n\tis_shared\030\004" +
|
||||||
"e\030\005 \001(\t\022\022\n\ncreateTime\030\006 \001(\003\"9\n\017StoreSequ" +
|
" \001(\010\022\017\n\007purpose\030\005 \001(\t\022\023\n\013create_time\030\006 \001" +
|
||||||
"enceId\022\022\n\nfamilyName\030\001 \002(\014\022\022\n\nsequenceId" +
|
"(\003\";\n\017StoreSequenceId\022\023\n\013family_name\030\001 \002" +
|
||||||
"\030\002 \002(\004\"b\n\026RegionStoreSequenceIds\022\035\n\025last" +
|
"(\014\022\023\n\013sequence_id\030\002 \002(\004\"g\n\026RegionStoreSe" +
|
||||||
"FlushedSequenceId\030\001 \002(\004\022)\n\017storeSequence" +
|
"quenceIds\022 \n\030last_flushed_sequence_id\030\001 " +
|
||||||
"Id\030\002 \003(\0132\020.StoreSequenceIdBE\n*org.apache" +
|
"\002(\004\022+\n\021store_sequence_id\030\002 \003(\0132\020.StoreSe" +
|
||||||
".hadoop.hbase.protobuf.generatedB\017ZooKee" +
|
"quenceIdBE\n*org.apache.hadoop.hbase.prot" +
|
||||||
"perProtosH\001\210\001\001\240\001\001"
|
"obuf.generatedB\017ZooKeeperProtosH\001\210\001\001\240\001\001"
|
||||||
};
|
};
|
||||||
com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
|
com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
|
||||||
new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() {
|
new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() {
|
||||||
|
|
|
@ -85,15 +85,15 @@ message CheckPermissionsResponse {
|
||||||
}
|
}
|
||||||
|
|
||||||
service AccessControlService {
|
service AccessControlService {
|
||||||
rpc grant(GrantRequest)
|
rpc Grant(GrantRequest)
|
||||||
returns (GrantResponse);
|
returns (GrantResponse);
|
||||||
|
|
||||||
rpc revoke(RevokeRequest)
|
rpc Revoke(RevokeRequest)
|
||||||
returns (RevokeResponse);
|
returns (RevokeResponse);
|
||||||
|
|
||||||
rpc getUserPermissions(UserPermissionsRequest)
|
rpc GetUserPermissions(UserPermissionsRequest)
|
||||||
returns (UserPermissionsResponse);
|
returns (UserPermissionsResponse);
|
||||||
|
|
||||||
rpc checkPermissions(CheckPermissionsRequest)
|
rpc CheckPermissions(CheckPermissionsRequest)
|
||||||
returns (CheckPermissionsResponse);
|
returns (CheckPermissionsResponse);
|
||||||
}
|
}
|
||||||
|
|
|
@ -30,12 +30,12 @@ import "WAL.proto";
|
||||||
|
|
||||||
message GetRegionInfoRequest {
|
message GetRegionInfoRequest {
|
||||||
required RegionSpecifier region = 1;
|
required RegionSpecifier region = 1;
|
||||||
optional bool compactionState = 2;
|
optional bool compaction_state = 2;
|
||||||
}
|
}
|
||||||
|
|
||||||
message GetRegionInfoResponse {
|
message GetRegionInfoResponse {
|
||||||
required RegionInfo regionInfo = 1;
|
required RegionInfo region_info = 1;
|
||||||
optional CompactionState compactionState = 2;
|
optional CompactionState compaction_state = 2;
|
||||||
|
|
||||||
enum CompactionState {
|
enum CompactionState {
|
||||||
NONE = 0;
|
NONE = 0;
|
||||||
|
@ -55,28 +55,28 @@ message GetStoreFileRequest {
|
||||||
}
|
}
|
||||||
|
|
||||||
message GetStoreFileResponse {
|
message GetStoreFileResponse {
|
||||||
repeated string storeFile = 1;
|
repeated string store_file = 1;
|
||||||
}
|
}
|
||||||
|
|
||||||
message GetOnlineRegionRequest {
|
message GetOnlineRegionRequest {
|
||||||
}
|
}
|
||||||
|
|
||||||
message GetOnlineRegionResponse {
|
message GetOnlineRegionResponse {
|
||||||
repeated RegionInfo regionInfo = 1;
|
repeated RegionInfo region_info = 1;
|
||||||
}
|
}
|
||||||
|
|
||||||
message OpenRegionRequest {
|
message OpenRegionRequest {
|
||||||
repeated RegionOpenInfo openInfo = 1;
|
repeated RegionOpenInfo open_info = 1;
|
||||||
|
|
||||||
message RegionOpenInfo {
|
message RegionOpenInfo {
|
||||||
required RegionInfo region = 1;
|
required RegionInfo region = 1;
|
||||||
optional uint32 versionOfOfflineNode = 2;
|
optional uint32 version_of_offline_node = 2;
|
||||||
repeated ServerName favoredNodes = 3;
|
repeated ServerName favored_nodes = 3;
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
message OpenRegionResponse {
|
message OpenRegionResponse {
|
||||||
repeated RegionOpeningState openingState = 1;
|
repeated RegionOpeningState opening_state = 1;
|
||||||
|
|
||||||
enum RegionOpeningState {
|
enum RegionOpeningState {
|
||||||
OPENED = 0;
|
OPENED = 0;
|
||||||
|
@ -91,9 +91,9 @@ message OpenRegionResponse {
|
||||||
*/
|
*/
|
||||||
message CloseRegionRequest {
|
message CloseRegionRequest {
|
||||||
required RegionSpecifier region = 1;
|
required RegionSpecifier region = 1;
|
||||||
optional uint32 versionOfClosingNode = 2;
|
optional uint32 version_of_closing_node = 2;
|
||||||
optional bool transitionInZK = 3 [default = true];
|
optional bool transition_in_ZK = 3 [default = true];
|
||||||
optional ServerName destinationServer = 4;
|
optional ServerName destination_server = 4;
|
||||||
}
|
}
|
||||||
|
|
||||||
message CloseRegionResponse {
|
message CloseRegionResponse {
|
||||||
|
@ -107,11 +107,11 @@ message CloseRegionResponse {
|
||||||
*/
|
*/
|
||||||
message FlushRegionRequest {
|
message FlushRegionRequest {
|
||||||
required RegionSpecifier region = 1;
|
required RegionSpecifier region = 1;
|
||||||
optional uint64 ifOlderThanTs = 2;
|
optional uint64 if_older_than_ts = 2;
|
||||||
}
|
}
|
||||||
|
|
||||||
message FlushRegionResponse {
|
message FlushRegionResponse {
|
||||||
required uint64 lastFlushTime = 1;
|
required uint64 last_flush_time = 1;
|
||||||
optional bool flushed = 2;
|
optional bool flushed = 2;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -124,7 +124,7 @@ message FlushRegionResponse {
|
||||||
*/
|
*/
|
||||||
message SplitRegionRequest {
|
message SplitRegionRequest {
|
||||||
required RegionSpecifier region = 1;
|
required RegionSpecifier region = 1;
|
||||||
optional bytes splitPoint = 2;
|
optional bytes split_point = 2;
|
||||||
}
|
}
|
||||||
|
|
||||||
message SplitRegionResponse {
|
message SplitRegionResponse {
|
||||||
|
@ -150,8 +150,8 @@ message CompactRegionResponse {
|
||||||
* This method currently closes the regions and then merges them
|
* This method currently closes the regions and then merges them
|
||||||
*/
|
*/
|
||||||
message MergeRegionsRequest {
|
message MergeRegionsRequest {
|
||||||
required RegionSpecifier regionA = 1;
|
required RegionSpecifier region_a = 1;
|
||||||
required RegionSpecifier regionB = 2;
|
required RegionSpecifier region_b = 2;
|
||||||
optional bool forcible = 3 [default = false];
|
optional bool forcible = 3 [default = false];
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -163,10 +163,10 @@ message WALEntry {
|
||||||
required WALKey key = 1;
|
required WALKey key = 1;
|
||||||
// Following may be null if the KVs/Cells are carried along the side in a cellblock (See
|
// Following may be null if the KVs/Cells are carried along the side in a cellblock (See
|
||||||
// RPC for more on cellblocks). If Cells/KVs are in a cellblock, this next field is null
|
// RPC for more on cellblocks). If Cells/KVs are in a cellblock, this next field is null
|
||||||
// and associatedCellCount has count of Cells associated w/ this WALEntry
|
// and associated_cell_count has count of Cells associated w/ this WALEntry
|
||||||
repeated bytes keyValueBytes = 2;
|
repeated bytes key_value_bytes = 2;
|
||||||
// If Cell data is carried alongside in a cellblock, this is count of Cells in the cellblock.
|
// If Cell data is carried alongside in a cellblock, this is count of Cells in the cellblock.
|
||||||
optional int32 associatedCellCount = 3;
|
optional int32 associated_cell_count = 3;
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
|
@ -186,7 +186,7 @@ message RollWALWriterRequest {
|
||||||
|
|
||||||
message RollWALWriterResponse {
|
message RollWALWriterResponse {
|
||||||
// A list of encoded name of regions to flush
|
// A list of encoded name of regions to flush
|
||||||
repeated bytes regionToFlush = 1;
|
repeated bytes region_to_flush = 1;
|
||||||
}
|
}
|
||||||
|
|
||||||
message StopServerRequest {
|
message StopServerRequest {
|
||||||
|
@ -200,54 +200,54 @@ message GetServerInfoRequest {
|
||||||
}
|
}
|
||||||
|
|
||||||
message ServerInfo {
|
message ServerInfo {
|
||||||
required ServerName serverName = 1;
|
required ServerName server_name = 1;
|
||||||
optional uint32 webuiPort = 2;
|
optional uint32 webui_port = 2;
|
||||||
}
|
}
|
||||||
|
|
||||||
message GetServerInfoResponse {
|
message GetServerInfoResponse {
|
||||||
required ServerInfo serverInfo = 1;
|
required ServerInfo server_info = 1;
|
||||||
}
|
}
|
||||||
|
|
||||||
service AdminService {
|
service AdminService {
|
||||||
rpc getRegionInfo(GetRegionInfoRequest)
|
rpc GetRegionInfo(GetRegionInfoRequest)
|
||||||
returns(GetRegionInfoResponse);
|
returns(GetRegionInfoResponse);
|
||||||
|
|
||||||
rpc getStoreFile(GetStoreFileRequest)
|
rpc GetStoreFile(GetStoreFileRequest)
|
||||||
returns(GetStoreFileResponse);
|
returns(GetStoreFileResponse);
|
||||||
|
|
||||||
rpc getOnlineRegion(GetOnlineRegionRequest)
|
rpc GetOnlineRegion(GetOnlineRegionRequest)
|
||||||
returns(GetOnlineRegionResponse);
|
returns(GetOnlineRegionResponse);
|
||||||
|
|
||||||
rpc openRegion(OpenRegionRequest)
|
rpc OpenRegion(OpenRegionRequest)
|
||||||
returns(OpenRegionResponse);
|
returns(OpenRegionResponse);
|
||||||
|
|
||||||
rpc closeRegion(CloseRegionRequest)
|
rpc CloseRegion(CloseRegionRequest)
|
||||||
returns(CloseRegionResponse);
|
returns(CloseRegionResponse);
|
||||||
|
|
||||||
rpc flushRegion(FlushRegionRequest)
|
rpc FlushRegion(FlushRegionRequest)
|
||||||
returns(FlushRegionResponse);
|
returns(FlushRegionResponse);
|
||||||
|
|
||||||
rpc splitRegion(SplitRegionRequest)
|
rpc SplitRegion(SplitRegionRequest)
|
||||||
returns(SplitRegionResponse);
|
returns(SplitRegionResponse);
|
||||||
|
|
||||||
rpc compactRegion(CompactRegionRequest)
|
rpc CompactRegion(CompactRegionRequest)
|
||||||
returns(CompactRegionResponse);
|
returns(CompactRegionResponse);
|
||||||
|
|
||||||
rpc mergeRegions(MergeRegionsRequest)
|
rpc MergeRegions(MergeRegionsRequest)
|
||||||
returns(MergeRegionsResponse);
|
returns(MergeRegionsResponse);
|
||||||
|
|
||||||
rpc replicateWALEntry(ReplicateWALEntryRequest)
|
rpc ReplicateWALEntry(ReplicateWALEntryRequest)
|
||||||
returns(ReplicateWALEntryResponse);
|
returns(ReplicateWALEntryResponse);
|
||||||
|
|
||||||
rpc replay(MultiRequest)
|
rpc Replay(MultiRequest)
|
||||||
returns(MultiResponse);
|
returns(MultiResponse);
|
||||||
|
|
||||||
rpc rollWALWriter(RollWALWriterRequest)
|
rpc RollWALWriter(RollWALWriterRequest)
|
||||||
returns(RollWALWriterResponse);
|
returns(RollWALWriterResponse);
|
||||||
|
|
||||||
rpc getServerInfo(GetServerInfoRequest)
|
rpc GetServerInfo(GetServerInfoRequest)
|
||||||
returns(GetServerInfoResponse);
|
returns(GetServerInfoResponse);
|
||||||
|
|
||||||
rpc stopServer(StopServerRequest)
|
rpc StopServer(StopServerRequest)
|
||||||
returns(StopServerResponse);
|
returns(StopServerResponse);
|
||||||
}
|
}
|
||||||
|
|
|
@ -31,32 +31,32 @@ message AggregateArgument {
|
||||||
* (3) any bytes required to construct the ColumnInterpreter object
|
* (3) any bytes required to construct the ColumnInterpreter object
|
||||||
* properly
|
* properly
|
||||||
*/
|
*/
|
||||||
required string interpreterClassName = 1;
|
required string interpreter_class_name = 1;
|
||||||
required Scan scan = 2;
|
required Scan scan = 2;
|
||||||
optional bytes interpreterSpecificBytes = 3;
|
optional bytes interpreter_specific_bytes = 3;
|
||||||
}
|
}
|
||||||
|
|
||||||
message AggregateResponse {
|
message AggregateResponse {
|
||||||
/**
|
/**
|
||||||
* The AggregateService methods all have a response that either is a Pair
|
* The AggregateService methods all have a response that either is a Pair
|
||||||
* or a simple object. When it is a Pair both firstPart and secondPart
|
* or a simple object. When it is a Pair both first_part and second_part
|
||||||
* have defined values (and the secondPart is not present in the response
|
* have defined values (and the second_part is not present in the response
|
||||||
* when the response is not a pair). Refer to the AggregateImplementation
|
* when the response is not a pair). Refer to the AggregateImplementation
|
||||||
* class for an overview of the AggregateResponse object constructions.
|
* class for an overview of the AggregateResponse object constructions.
|
||||||
*/
|
*/
|
||||||
repeated bytes firstPart = 1;
|
repeated bytes first_part = 1;
|
||||||
optional bytes secondPart = 2;
|
optional bytes second_part = 2;
|
||||||
}
|
}
|
||||||
|
|
||||||
/** Refer to the AggregateImplementation class for an overview of the
|
/** Refer to the AggregateImplementation class for an overview of the
|
||||||
* AggregateService method implementations and their functionality.
|
* AggregateService method implementations and their functionality.
|
||||||
*/
|
*/
|
||||||
service AggregateService {
|
service AggregateService {
|
||||||
rpc getMax (AggregateArgument) returns (AggregateResponse);
|
rpc GetMax (AggregateArgument) returns (AggregateResponse);
|
||||||
rpc getMin (AggregateArgument) returns (AggregateResponse);
|
rpc GetMin (AggregateArgument) returns (AggregateResponse);
|
||||||
rpc getSum (AggregateArgument) returns (AggregateResponse);
|
rpc GetSum (AggregateArgument) returns (AggregateResponse);
|
||||||
rpc getRowNum (AggregateArgument) returns (AggregateResponse);
|
rpc GetRowNum (AggregateArgument) returns (AggregateResponse);
|
||||||
rpc getAvg (AggregateArgument) returns (AggregateResponse);
|
rpc GetAvg (AggregateArgument) returns (AggregateResponse);
|
||||||
rpc getStd (AggregateArgument) returns (AggregateResponse);
|
rpc GetStd (AggregateArgument) returns (AggregateResponse);
|
||||||
rpc getMedian (AggregateArgument) returns (AggregateResponse);
|
rpc GetMedian (AggregateArgument) returns (AggregateResponse);
|
||||||
}
|
}
|
|
@ -24,7 +24,7 @@ option optimize_for = SPEED;
|
||||||
|
|
||||||
message AuthenticationKey {
|
message AuthenticationKey {
|
||||||
required int32 id = 1;
|
required int32 id = 1;
|
||||||
required int64 expirationDate = 2;
|
required int64 expiration_date = 2;
|
||||||
required bytes key = 3;
|
required bytes key = 3;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -35,10 +35,10 @@ message TokenIdentifier {
|
||||||
}
|
}
|
||||||
required Kind kind = 1;
|
required Kind kind = 1;
|
||||||
required bytes username = 2;
|
required bytes username = 2;
|
||||||
required int32 keyId = 3;
|
required int32 key_id = 3;
|
||||||
optional int64 issueDate = 4;
|
optional int64 issue_date = 4;
|
||||||
optional int64 expirationDate = 5;
|
optional int64 expiration_date = 5;
|
||||||
optional int64 sequenceNumber = 6;
|
optional int64 sequence_number = 6;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
||||||
|
@ -67,15 +67,15 @@ message WhoAmIRequest {
|
||||||
|
|
||||||
message WhoAmIResponse {
|
message WhoAmIResponse {
|
||||||
optional string username = 1;
|
optional string username = 1;
|
||||||
optional string authMethod = 2;
|
optional string auth_method = 2;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
||||||
// RPC service
|
// RPC service
|
||||||
service AuthenticationService {
|
service AuthenticationService {
|
||||||
rpc getAuthenticationToken(TokenRequest)
|
rpc GetAuthenticationToken(TokenRequest)
|
||||||
returns (TokenResponse);
|
returns (TokenResponse);
|
||||||
|
|
||||||
rpc whoami(WhoAmIRequest)
|
rpc WhoAmI(WhoAmIRequest)
|
||||||
returns (WhoAmIResponse);
|
returns (WhoAmIResponse);
|
||||||
}
|
}
|
|
@ -46,7 +46,7 @@ message Cell {
|
||||||
optional bytes family = 2;
|
optional bytes family = 2;
|
||||||
optional bytes qualifier = 3;
|
optional bytes qualifier = 3;
|
||||||
optional uint64 timestamp = 4;
|
optional uint64 timestamp = 4;
|
||||||
optional CellType cellType = 5;
|
optional CellType cell_type = 5;
|
||||||
optional bytes value = 6;
|
optional bytes value = 6;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -59,6 +59,6 @@ message KeyValue {
|
||||||
required bytes family = 2;
|
required bytes family = 2;
|
||||||
required bytes qualifier = 3;
|
required bytes qualifier = 3;
|
||||||
optional uint64 timestamp = 4;
|
optional uint64 timestamp = 4;
|
||||||
optional CellType keyType = 5;
|
optional CellType key_type = 5;
|
||||||
optional bytes value = 6;
|
optional bytes value = 6;
|
||||||
}
|
}
|
||||||
|
|
|
@ -45,11 +45,11 @@ message Get {
|
||||||
repeated Column column = 2;
|
repeated Column column = 2;
|
||||||
repeated NameBytesPair attribute = 3;
|
repeated NameBytesPair attribute = 3;
|
||||||
optional Filter filter = 4;
|
optional Filter filter = 4;
|
||||||
optional TimeRange timeRange = 5;
|
optional TimeRange time_range = 5;
|
||||||
optional uint32 maxVersions = 6 [default = 1];
|
optional uint32 max_versions = 6 [default = 1];
|
||||||
optional bool cacheBlocks = 7 [default = true];
|
optional bool cache_blocks = 7 [default = true];
|
||||||
optional uint32 storeLimit = 8;
|
optional uint32 store_limit = 8;
|
||||||
optional uint32 storeOffset = 9;
|
optional uint32 store_offset = 9;
|
||||||
}
|
}
|
||||||
|
|
||||||
message Result {
|
message Result {
|
||||||
|
@ -60,18 +60,18 @@ message Result {
|
||||||
// not part of this protobuf message; they are passed alongside
|
// not part of this protobuf message; they are passed alongside
|
||||||
// and then this Message is just a placeholder with metadata.
|
// and then this Message is just a placeholder with metadata.
|
||||||
// The count is needed to know how many to peel off the block of Cells as
|
// The count is needed to know how many to peel off the block of Cells as
|
||||||
// ours. NOTE: This is different from the pb managed cellCount of the
|
// ours. NOTE: This is different from the pb managed cell_count of the
|
||||||
// 'cell' field above which is non-null when the cells are pb'd.
|
// 'cell' field above which is non-null when the cells are pb'd.
|
||||||
optional int32 associatedCellCount = 2;
|
optional int32 associated_cell_count = 2;
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* The get request. Perform a single Get operation.
|
* The get request. Perform a single Get operation.
|
||||||
* Unless existenceOnly is specified, return all the requested data
|
* Unless existence_only is specified, return all the requested data
|
||||||
* for the row that matches exactly, or the one that immediately
|
* for the row that matches exactly, or the one that immediately
|
||||||
* precedes it if closestRowBefore is specified.
|
* precedes it if closest_row_before is specified.
|
||||||
*
|
*
|
||||||
* If existenceOnly is set, only the existence will be returned.
|
* If existence_only is set, only the existence will be returned.
|
||||||
*/
|
*/
|
||||||
message GetRequest {
|
message GetRequest {
|
||||||
required RegionSpecifier region = 1;
|
required RegionSpecifier region = 1;
|
||||||
|
@ -79,12 +79,12 @@ message GetRequest {
|
||||||
|
|
||||||
// If the row to get doesn't exist, return the
|
// If the row to get doesn't exist, return the
|
||||||
// closest row before.
|
// closest row before.
|
||||||
optional bool closestRowBefore = 3;
|
optional bool closest_row_before = 3;
|
||||||
|
|
||||||
// The result isn't asked for, just check for
|
// The result isn't asked for, just check for
|
||||||
// the existence. If closestRowBefore specified,
|
// the existence. If closest_row_before specified,
|
||||||
// this will be ignored
|
// this will be ignored
|
||||||
optional bool existenceOnly = 4;
|
optional bool existence_only = 4;
|
||||||
}
|
}
|
||||||
|
|
||||||
message MultiGetRequest {
|
message MultiGetRequest {
|
||||||
|
@ -93,12 +93,12 @@ message MultiGetRequest {
|
||||||
|
|
||||||
// If the row to get doesn't exist, return the
|
// If the row to get doesn't exist, return the
|
||||||
// closest row before.
|
// closest row before.
|
||||||
optional bool closestRowBefore = 3;
|
optional bool closest_row_before = 3;
|
||||||
|
|
||||||
// The result isn't asked for, just check for
|
// The result isn't asked for, just check for
|
||||||
// the existence. If closestRowBefore specified,
|
// the existence. If closest_row_before specified,
|
||||||
// this will be ignored
|
// this will be ignored
|
||||||
optional bool existenceOnly = 4;
|
optional bool existence_only = 4;
|
||||||
}
|
}
|
||||||
|
|
||||||
message GetResponse {
|
message GetResponse {
|
||||||
|
@ -125,7 +125,7 @@ message Condition {
|
||||||
required bytes row = 1;
|
required bytes row = 1;
|
||||||
required bytes family = 2;
|
required bytes family = 2;
|
||||||
required bytes qualifier = 3;
|
required bytes qualifier = 3;
|
||||||
required CompareType compareType = 4;
|
required CompareType compare_type = 4;
|
||||||
required Comparator comparator = 5;
|
required Comparator comparator = 5;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -139,22 +139,22 @@ message Condition {
|
||||||
*/
|
*/
|
||||||
message MutationProto {
|
message MutationProto {
|
||||||
optional bytes row = 1;
|
optional bytes row = 1;
|
||||||
optional MutationType mutateType = 2;
|
optional MutationType mutate_type = 2;
|
||||||
repeated ColumnValue columnValue = 3;
|
repeated ColumnValue column_value = 3;
|
||||||
optional uint64 timestamp = 4;
|
optional uint64 timestamp = 4;
|
||||||
repeated NameBytesPair attribute = 5;
|
repeated NameBytesPair attribute = 5;
|
||||||
optional Durability durability = 6 [default = USE_DEFAULT];
|
optional Durability durability = 6 [default = USE_DEFAULT];
|
||||||
|
|
||||||
// For some mutations, a result may be returned, in which case,
|
// For some mutations, a result may be returned, in which case,
|
||||||
// time range can be specified for potential performance gain
|
// time range can be specified for potential performance gain
|
||||||
optional TimeRange timeRange = 7;
|
optional TimeRange time_range = 7;
|
||||||
// The below count is set when the associated cells are NOT
|
// The below count is set when the associated cells are NOT
|
||||||
// part of this protobuf message; they are passed alongside
|
// part of this protobuf message; they are passed alongside
|
||||||
// and then this Message is a placeholder with metadata. The
|
// and then this Message is a placeholder with metadata. The
|
||||||
// count is needed to know how many to peel off the block of Cells as
|
// count is needed to know how many to peel off the block of Cells as
|
||||||
// ours. NOTE: This is different from the pb managed cellCount of the
|
// ours. NOTE: This is different from the pb managed cell_count of the
|
||||||
// 'cell' field above which is non-null when the cells are pb'd.
|
// 'cell' field above which is non-null when the cells are pb'd.
|
||||||
optional int32 associatedCellCount = 8;
|
optional int32 associated_cell_count = 8;
|
||||||
|
|
||||||
enum Durability {
|
enum Durability {
|
||||||
USE_DEFAULT = 0;
|
USE_DEFAULT = 0;
|
||||||
|
@ -180,13 +180,13 @@ message MutationProto {
|
||||||
|
|
||||||
message ColumnValue {
|
message ColumnValue {
|
||||||
required bytes family = 1;
|
required bytes family = 1;
|
||||||
repeated QualifierValue qualifierValue = 2;
|
repeated QualifierValue qualifier_value = 2;
|
||||||
|
|
||||||
message QualifierValue {
|
message QualifierValue {
|
||||||
optional bytes qualifier = 1;
|
optional bytes qualifier = 1;
|
||||||
optional bytes value = 2;
|
optional bytes value = 2;
|
||||||
optional uint64 timestamp = 3;
|
optional uint64 timestamp = 3;
|
||||||
optional DeleteType deleteType = 4;
|
optional DeleteType delete_type = 4;
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
@ -225,18 +225,18 @@ message MutateResponse {
|
||||||
message Scan {
|
message Scan {
|
||||||
repeated Column column = 1;
|
repeated Column column = 1;
|
||||||
repeated NameBytesPair attribute = 2;
|
repeated NameBytesPair attribute = 2;
|
||||||
optional bytes startRow = 3;
|
optional bytes start_row = 3;
|
||||||
optional bytes stopRow = 4;
|
optional bytes stop_row = 4;
|
||||||
optional Filter filter = 5;
|
optional Filter filter = 5;
|
||||||
optional TimeRange timeRange = 6;
|
optional TimeRange time_range = 6;
|
||||||
optional uint32 maxVersions = 7 [default = 1];
|
optional uint32 max_versions = 7 [default = 1];
|
||||||
optional bool cacheBlocks = 8 [default = true];
|
optional bool cache_blocks = 8 [default = true];
|
||||||
optional uint32 batchSize = 9;
|
optional uint32 batch_size = 9;
|
||||||
optional uint64 maxResultSize = 10;
|
optional uint64 max_result_size = 10;
|
||||||
optional uint32 storeLimit = 11;
|
optional uint32 store_limit = 11;
|
||||||
optional uint32 storeOffset = 12;
|
optional uint32 store_offset = 12;
|
||||||
optional bool loadColumnFamiliesOnDemand = 13; /* DO NOT add defaults to loadColumnFamiliesOnDemand. */
|
optional bool load_column_families_on_demand = 13; /* DO NOT add defaults to load_column_families_on_demand. */
|
||||||
optional uint32 cachingCount = 14;
|
optional uint32 caching_count = 14;
|
||||||
optional bool prefetching = 15;
|
optional bool prefetching = 15;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -254,25 +254,25 @@ message Scan {
|
||||||
message ScanRequest {
|
message ScanRequest {
|
||||||
optional RegionSpecifier region = 1;
|
optional RegionSpecifier region = 1;
|
||||||
optional Scan scan = 2;
|
optional Scan scan = 2;
|
||||||
optional uint64 scannerId = 3;
|
optional uint64 scanner_id = 3;
|
||||||
optional uint32 numberOfRows = 4;
|
optional uint32 number_of_rows = 4;
|
||||||
optional bool closeScanner = 5;
|
optional bool close_scanner = 5;
|
||||||
optional uint64 nextCallSeq = 6;
|
optional uint64 next_call_seq = 6;
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* The scan response. If there are no more results, moreResults will
|
* The scan response. If there are no more results, more_results will
|
||||||
* be false. If it is not specified, it means there are more.
|
* be false. If it is not specified, it means there are more.
|
||||||
*/
|
*/
|
||||||
message ScanResponse {
|
message ScanResponse {
|
||||||
optional ResultCellMeta resultCellMeta = 1;
|
optional ResultCellMeta result_cell_meta = 1;
|
||||||
optional uint64 scannerId = 2;
|
optional uint64 scanner_id = 2;
|
||||||
optional bool moreResults = 3;
|
optional bool more_results = 3;
|
||||||
optional uint32 ttl = 4;
|
optional uint32 ttl = 4;
|
||||||
}
|
}
|
||||||
|
|
||||||
message ResultCellMeta {
|
message ResultCellMeta {
|
||||||
repeated uint32 cellsLength = 1;
|
repeated uint32 cells_length = 1;
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
|
@ -281,8 +281,8 @@ message ResultCellMeta {
|
||||||
*/
|
*/
|
||||||
message BulkLoadHFileRequest {
|
message BulkLoadHFileRequest {
|
||||||
required RegionSpecifier region = 1;
|
required RegionSpecifier region = 1;
|
||||||
repeated FamilyPath familyPath = 2;
|
repeated FamilyPath family_path = 2;
|
||||||
optional bool assignSeqNum = 3;
|
optional bool assign_seq_num = 3;
|
||||||
|
|
||||||
message FamilyPath {
|
message FamilyPath {
|
||||||
required bytes family = 1;
|
required bytes family = 1;
|
||||||
|
@ -296,8 +296,8 @@ message BulkLoadHFileResponse {
|
||||||
|
|
||||||
message CoprocessorServiceCall {
|
message CoprocessorServiceCall {
|
||||||
required bytes row = 1;
|
required bytes row = 1;
|
||||||
required string serviceName = 2;
|
required string service_name = 2;
|
||||||
required string methodName = 3;
|
required string method_name = 3;
|
||||||
required bytes request = 4;
|
required bytes request = 4;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -352,24 +352,24 @@ message MultiResponse {
|
||||||
|
|
||||||
|
|
||||||
service ClientService {
|
service ClientService {
|
||||||
rpc get(GetRequest)
|
rpc Get(GetRequest)
|
||||||
returns(GetResponse);
|
returns(GetResponse);
|
||||||
|
|
||||||
rpc multiGet(MultiGetRequest)
|
rpc MultiGet(MultiGetRequest)
|
||||||
returns(MultiGetResponse);
|
returns(MultiGetResponse);
|
||||||
|
|
||||||
rpc mutate(MutateRequest)
|
rpc Mutate(MutateRequest)
|
||||||
returns(MutateResponse);
|
returns(MutateResponse);
|
||||||
|
|
||||||
rpc scan(ScanRequest)
|
rpc Scan(ScanRequest)
|
||||||
returns(ScanResponse);
|
returns(ScanResponse);
|
||||||
|
|
||||||
rpc bulkLoadHFile(BulkLoadHFileRequest)
|
rpc BulkLoadHFile(BulkLoadHFileRequest)
|
||||||
returns(BulkLoadHFileResponse);
|
returns(BulkLoadHFileResponse);
|
||||||
|
|
||||||
rpc execService(CoprocessorServiceRequest)
|
rpc ExecService(CoprocessorServiceRequest)
|
||||||
returns(CoprocessorServiceResponse);
|
returns(CoprocessorServiceResponse);
|
||||||
|
|
||||||
rpc multi(MultiRequest)
|
rpc Multi(MultiRequest)
|
||||||
returns(MultiResponse);
|
returns(MultiResponse);
|
||||||
}
|
}
|
||||||
|
|
|
@ -29,5 +29,5 @@ option optimize_for = SPEED;
|
||||||
*/
|
*/
|
||||||
message ClusterId {
|
message ClusterId {
|
||||||
// This is the cluster id, a uuid as a String
|
// This is the cluster id, a uuid as a String
|
||||||
required string clusterId = 1;
|
required string cluster_id = 1;
|
||||||
}
|
}
|
||||||
|
|
|
@ -28,7 +28,7 @@ import "ClusterId.proto";
|
||||||
import "FS.proto";
|
import "FS.proto";
|
||||||
|
|
||||||
message RegionState {
|
message RegionState {
|
||||||
required RegionInfo regionInfo = 1;
|
required RegionInfo region_info = 1;
|
||||||
required State state = 2;
|
required State state = 2;
|
||||||
optional uint64 stamp = 3;
|
optional uint64 stamp = 3;
|
||||||
enum State {
|
enum State {
|
||||||
|
@ -50,12 +50,12 @@ message RegionState {
|
||||||
|
|
||||||
message RegionInTransition {
|
message RegionInTransition {
|
||||||
required RegionSpecifier spec = 1;
|
required RegionSpecifier spec = 1;
|
||||||
required RegionState regionState = 2;
|
required RegionState region_state = 2;
|
||||||
}
|
}
|
||||||
|
|
||||||
message RegionLoad {
|
message RegionLoad {
|
||||||
/** the region specifier */
|
/** the region specifier */
|
||||||
required RegionSpecifier regionSpecifier = 1;
|
required RegionSpecifier region_specifier = 1;
|
||||||
|
|
||||||
/** the number of stores for the region */
|
/** the number of stores for the region */
|
||||||
optional uint32 stores = 2;
|
optional uint32 stores = 2;
|
||||||
|
@ -64,65 +64,65 @@ message RegionLoad {
|
||||||
optional uint32 storefiles = 3;
|
optional uint32 storefiles = 3;
|
||||||
|
|
||||||
/** the total size of the store files for the region, uncompressed, in MB */
|
/** the total size of the store files for the region, uncompressed, in MB */
|
||||||
optional uint32 storeUncompressedSizeMB = 4;
|
optional uint32 store_uncompressed_size_MB = 4;
|
||||||
|
|
||||||
/** the current total size of the store files for the region, in MB */
|
/** the current total size of the store files for the region, in MB */
|
||||||
optional uint32 storefileSizeMB = 5;
|
optional uint32 storefile_size_MB = 5;
|
||||||
|
|
||||||
/** the current size of the memstore for the region, in MB */
|
/** the current size of the memstore for the region, in MB */
|
||||||
optional uint32 memstoreSizeMB = 6;
|
optional uint32 memstore_size_MB = 6;
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* The current total size of root-level store file indexes for the region,
|
* The current total size of root-level store file indexes for the region,
|
||||||
* in MB. The same as {@link #rootIndexSizeKB} but in MB.
|
* in MB. The same as {@link #rootIndexSizeKB} but in MB.
|
||||||
*/
|
*/
|
||||||
optional uint32 storefileIndexSizeMB = 7;
|
optional uint32 storefile_index_size_MB = 7;
|
||||||
|
|
||||||
/** the current total read requests made to region */
|
/** the current total read requests made to region */
|
||||||
optional uint64 readRequestsCount = 8;
|
optional uint64 read_requests_count = 8;
|
||||||
|
|
||||||
/** the current total write requests made to region */
|
/** the current total write requests made to region */
|
||||||
optional uint64 writeRequestsCount = 9;
|
optional uint64 write_requests_count = 9;
|
||||||
|
|
||||||
/** the total compacting key values in currently running compaction */
|
/** the total compacting key values in currently running compaction */
|
||||||
optional uint64 totalCompactingKVs = 10;
|
optional uint64 total_compacting_KVs = 10;
|
||||||
|
|
||||||
/** the completed count of key values in currently running compaction */
|
/** the completed count of key values in currently running compaction */
|
||||||
optional uint64 currentCompactedKVs = 11;
|
optional uint64 current_compacted_KVs = 11;
|
||||||
|
|
||||||
/** The current total size of root-level indexes for the region, in KB. */
|
/** The current total size of root-level indexes for the region, in KB. */
|
||||||
optional uint32 rootIndexSizeKB = 12;
|
optional uint32 root_index_size_KB = 12;
|
||||||
|
|
||||||
/** The total size of all index blocks, not just the root level, in KB. */
|
/** The total size of all index blocks, not just the root level, in KB. */
|
||||||
optional uint32 totalStaticIndexSizeKB = 13;
|
optional uint32 total_static_index_size_KB = 13;
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* The total size of all Bloom filter blocks, not just loaded into the
|
* The total size of all Bloom filter blocks, not just loaded into the
|
||||||
* block cache, in KB.
|
* block cache, in KB.
|
||||||
*/
|
*/
|
||||||
optional uint32 totalStaticBloomSizeKB = 14;
|
optional uint32 total_static_bloom_size_KB = 14;
|
||||||
|
|
||||||
/** the most recent sequence Id from cache flush */
|
/** the most recent sequence Id from cache flush */
|
||||||
optional uint64 completeSequenceId = 15;
|
optional uint64 complete_sequence_id = 15;
|
||||||
}
|
}
|
||||||
|
|
||||||
/* Server-level protobufs */
|
/* Server-level protobufs */
|
||||||
|
|
||||||
message ServerLoad {
|
message ServerLoad {
|
||||||
/** Number of requests since last report. */
|
/** Number of requests since last report. */
|
||||||
optional uint32 numberOfRequests = 1;
|
optional uint32 number_of_requests = 1;
|
||||||
|
|
||||||
/** Total Number of requests from the start of the region server. */
|
/** Total Number of requests from the start of the region server. */
|
||||||
optional uint32 totalNumberOfRequests = 2;
|
optional uint32 total_number_of_requests = 2;
|
||||||
|
|
||||||
/** the amount of used heap, in MB. */
|
/** the amount of used heap, in MB. */
|
||||||
optional uint32 usedHeapMB = 3;
|
optional uint32 used_heap_MB = 3;
|
||||||
|
|
||||||
/** the maximum allowable size of the heap, in MB. */
|
/** the maximum allowable size of the heap, in MB. */
|
||||||
optional uint32 maxHeapMB = 4;
|
optional uint32 max_heap_MB = 4;
|
||||||
|
|
||||||
/** Information on the load of individual regions. */
|
/** Information on the load of individual regions. */
|
||||||
repeated RegionLoad regionLoads = 5;
|
repeated RegionLoad region_loads = 5;
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Regionserver-level coprocessors, e.g., WALObserver implementations.
|
* Regionserver-level coprocessors, e.g., WALObserver implementations.
|
||||||
|
@ -132,39 +132,38 @@ message ServerLoad {
|
||||||
repeated Coprocessor coprocessors = 6;
|
repeated Coprocessor coprocessors = 6;
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Time when incremental (non-total) counts began being calculated (e.g. numberOfRequests)
|
* Time when incremental (non-total) counts began being calculated (e.g. number_of_requests)
|
||||||
* time is measured as the difference, measured in milliseconds, between the current time
|
* time is measured as the difference, measured in milliseconds, between the current time
|
||||||
* and midnight, January 1, 1970 UTC.
|
* and midnight, January 1, 1970 UTC.
|
||||||
*/
|
*/
|
||||||
optional uint64 reportStartTime = 7;
|
optional uint64 report_start_time = 7;
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Time when report was generated.
|
* Time when report was generated.
|
||||||
* time is measured as the difference, measured in milliseconds, between the current time
|
* time is measured as the difference, measured in milliseconds, between the current time
|
||||||
* and midnight, January 1, 1970 UTC.
|
* and midnight, January 1, 1970 UTC.
|
||||||
*/
|
*/
|
||||||
optional uint64 reportEndTime = 8;
|
optional uint64 report_end_time = 8;
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* The port number that this region server is hosing an info server on.
|
* The port number that this region server is hosing an info server on.
|
||||||
*/
|
*/
|
||||||
optional uint32 infoServerPort = 9;
|
optional uint32 info_server_port = 9;
|
||||||
}
|
}
|
||||||
|
|
||||||
message LiveServerInfo {
|
message LiveServerInfo {
|
||||||
required ServerName server = 1;
|
required ServerName server = 1;
|
||||||
required ServerLoad serverLoad = 2;
|
required ServerLoad server_load = 2;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
||||||
message ClusterStatus {
|
message ClusterStatus {
|
||||||
optional HBaseVersionFileContent hbaseVersion = 1;
|
optional HBaseVersionFileContent hbase_version = 1;
|
||||||
repeated LiveServerInfo liveServers = 2;
|
repeated LiveServerInfo live_servers = 2;
|
||||||
repeated ServerName deadServers = 3;
|
repeated ServerName dead_servers = 3;
|
||||||
repeated RegionInTransition regionsInTransition = 4;
|
repeated RegionInTransition regions_in_transition = 4;
|
||||||
optional ClusterId clusterId = 5;
|
optional ClusterId cluster_id = 5;
|
||||||
repeated Coprocessor masterCoprocessors = 6;
|
repeated Coprocessor master_coprocessors = 6;
|
||||||
optional ServerName master = 7;
|
optional ServerName master = 7;
|
||||||
repeated ServerName backupMasters = 8;
|
repeated ServerName backup_masters = 8;
|
||||||
optional bool balancerOn = 9;
|
optional bool balancer_on = 9;
|
||||||
}
|
}
|
||||||
|
|
|
@ -28,7 +28,7 @@ option optimize_for = SPEED;
|
||||||
|
|
||||||
message Comparator {
|
message Comparator {
|
||||||
required string name = 1;
|
required string name = 1;
|
||||||
optional bytes serializedComparator = 2;
|
optional bytes serialized_comparator = 2;
|
||||||
}
|
}
|
||||||
|
|
||||||
message ByteArrayComparable {
|
message ByteArrayComparable {
|
||||||
|
@ -45,7 +45,7 @@ message BinaryPrefixComparator {
|
||||||
|
|
||||||
message BitComparator {
|
message BitComparator {
|
||||||
required ByteArrayComparable comparable = 1;
|
required ByteArrayComparable comparable = 1;
|
||||||
required BitwiseOp bitwiseOp = 2;
|
required BitwiseOp bitwise_op = 2;
|
||||||
|
|
||||||
enum BitwiseOp {
|
enum BitwiseOp {
|
||||||
AND = 1;
|
AND = 1;
|
||||||
|
@ -59,7 +59,7 @@ message NullComparator {
|
||||||
|
|
||||||
message RegexStringComparator {
|
message RegexStringComparator {
|
||||||
required string pattern = 1;
|
required string pattern = 1;
|
||||||
required int32 patternFlags = 2;
|
required int32 pattern_flags = 2;
|
||||||
required string charset = 3;
|
required string charset = 3;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
@ -28,10 +28,10 @@ option optimize_for = SPEED;
|
||||||
* so we can serialize exceptions.
|
* so we can serialize exceptions.
|
||||||
*/
|
*/
|
||||||
message StackTraceElementMessage {
|
message StackTraceElementMessage {
|
||||||
optional string declaringClass = 1;
|
optional string declaring_class = 1;
|
||||||
optional string methodName = 2;
|
optional string method_name = 2;
|
||||||
optional string fileName = 3;
|
optional string file_name = 3;
|
||||||
optional int32 lineNumber = 4;
|
optional int32 line_number = 4;
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
|
@ -41,9 +41,9 @@ message StackTraceElementMessage {
|
||||||
* (which should be another protobuffed class).
|
* (which should be another protobuffed class).
|
||||||
*/
|
*/
|
||||||
message GenericExceptionMessage {
|
message GenericExceptionMessage {
|
||||||
optional string className = 1;
|
optional string class_name = 1;
|
||||||
optional string message = 2;
|
optional string message = 2;
|
||||||
optional bytes errorInfo = 3;
|
optional bytes error_info = 3;
|
||||||
repeated StackTraceElementMessage trace = 4;
|
repeated StackTraceElementMessage trace = 4;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -53,6 +53,5 @@ message GenericExceptionMessage {
|
||||||
*/
|
*/
|
||||||
message ForeignExceptionMessage {
|
message ForeignExceptionMessage {
|
||||||
optional string source = 1;
|
optional string source = 1;
|
||||||
optional GenericExceptionMessage genericException = 2;
|
optional GenericExceptionMessage generic_exception = 2;
|
||||||
|
|
||||||
}
|
}
|
|
@ -29,7 +29,7 @@ import "Comparator.proto";
|
||||||
|
|
||||||
message Filter {
|
message Filter {
|
||||||
required string name = 1;
|
required string name = 1;
|
||||||
optional bytes serializedFilter = 2;
|
optional bytes serialized_filter = 2;
|
||||||
}
|
}
|
||||||
|
|
||||||
message ColumnCountGetFilter {
|
message ColumnCountGetFilter {
|
||||||
|
@ -39,7 +39,7 @@ message ColumnCountGetFilter {
|
||||||
message ColumnPaginationFilter {
|
message ColumnPaginationFilter {
|
||||||
required int32 limit = 1;
|
required int32 limit = 1;
|
||||||
optional int32 offset = 2;
|
optional int32 offset = 2;
|
||||||
optional bytes columnOffset = 3;
|
optional bytes column_offset = 3;
|
||||||
}
|
}
|
||||||
|
|
||||||
message ColumnPrefixFilter {
|
message ColumnPrefixFilter {
|
||||||
|
@ -47,26 +47,26 @@ message ColumnPrefixFilter {
|
||||||
}
|
}
|
||||||
|
|
||||||
message ColumnRangeFilter {
|
message ColumnRangeFilter {
|
||||||
optional bytes minColumn = 1;
|
optional bytes min_column = 1;
|
||||||
optional bool minColumnInclusive = 2;
|
optional bool min_column_inclusive = 2;
|
||||||
optional bytes maxColumn = 3;
|
optional bytes max_column = 3;
|
||||||
optional bool maxColumnInclusive = 4;
|
optional bool max_column_inclusive = 4;
|
||||||
}
|
}
|
||||||
|
|
||||||
message CompareFilter {
|
message CompareFilter {
|
||||||
required CompareType compareOp = 1;
|
required CompareType compare_op = 1;
|
||||||
optional Comparator comparator = 2;
|
optional Comparator comparator = 2;
|
||||||
}
|
}
|
||||||
|
|
||||||
message DependentColumnFilter {
|
message DependentColumnFilter {
|
||||||
required CompareFilter compareFilter = 1;
|
required CompareFilter compare_filter = 1;
|
||||||
optional bytes columnFamily = 2;
|
optional bytes column_family = 2;
|
||||||
optional bytes columnQualifier = 3;
|
optional bytes column_qualifier = 3;
|
||||||
optional bool dropDependentColumn = 4;
|
optional bool drop_dependent_column = 4;
|
||||||
}
|
}
|
||||||
|
|
||||||
message FamilyFilter {
|
message FamilyFilter {
|
||||||
required CompareFilter compareFilter = 1;
|
required CompareFilter compare_filter = 1;
|
||||||
}
|
}
|
||||||
|
|
||||||
message FilterList {
|
message FilterList {
|
||||||
|
@ -91,23 +91,23 @@ message FirstKeyValueMatchingQualifiersFilter {
|
||||||
}
|
}
|
||||||
|
|
||||||
message FuzzyRowFilter {
|
message FuzzyRowFilter {
|
||||||
repeated BytesBytesPair fuzzyKeysData = 1;
|
repeated BytesBytesPair fuzzy_keys_data = 1;
|
||||||
}
|
}
|
||||||
|
|
||||||
message InclusiveStopFilter {
|
message InclusiveStopFilter {
|
||||||
optional bytes stopRowKey = 1;
|
optional bytes stop_row_key = 1;
|
||||||
}
|
}
|
||||||
|
|
||||||
message KeyOnlyFilter {
|
message KeyOnlyFilter {
|
||||||
required bool lenAsVal = 1;
|
required bool len_as_val = 1;
|
||||||
}
|
}
|
||||||
|
|
||||||
message MultipleColumnPrefixFilter {
|
message MultipleColumnPrefixFilter {
|
||||||
repeated bytes sortedPrefixes = 1;
|
repeated bytes sorted_prefixes = 1;
|
||||||
}
|
}
|
||||||
|
|
||||||
message PageFilter {
|
message PageFilter {
|
||||||
required int64 pageSize = 1;
|
required int64 page_size = 1;
|
||||||
}
|
}
|
||||||
|
|
||||||
message PrefixFilter {
|
message PrefixFilter {
|
||||||
|
@ -115,7 +115,7 @@ message PrefixFilter {
|
||||||
}
|
}
|
||||||
|
|
||||||
message QualifierFilter {
|
message QualifierFilter {
|
||||||
required CompareFilter compareFilter = 1;
|
required CompareFilter compare_filter = 1;
|
||||||
}
|
}
|
||||||
|
|
||||||
message RandomRowFilter {
|
message RandomRowFilter {
|
||||||
|
@ -123,20 +123,20 @@ message RandomRowFilter {
|
||||||
}
|
}
|
||||||
|
|
||||||
message RowFilter {
|
message RowFilter {
|
||||||
required CompareFilter compareFilter = 1;
|
required CompareFilter compare_filter = 1;
|
||||||
}
|
}
|
||||||
|
|
||||||
message SingleColumnValueExcludeFilter {
|
message SingleColumnValueExcludeFilter {
|
||||||
required SingleColumnValueFilter singleColumnValueFilter = 1;
|
required SingleColumnValueFilter single_column_value_filter = 1;
|
||||||
}
|
}
|
||||||
|
|
||||||
message SingleColumnValueFilter {
|
message SingleColumnValueFilter {
|
||||||
optional bytes columnFamily = 1;
|
optional bytes column_family = 1;
|
||||||
optional bytes columnQualifier = 2;
|
optional bytes column_qualifier = 2;
|
||||||
required CompareType compareOp = 3;
|
required CompareType compare_op = 3;
|
||||||
required Comparator comparator = 4;
|
required Comparator comparator = 4;
|
||||||
optional bool filterIfMissing = 5;
|
optional bool filter_if_missing = 5;
|
||||||
optional bool latestVersionOnly = 6;
|
optional bool latest_version_only = 6;
|
||||||
}
|
}
|
||||||
|
|
||||||
message SkipFilter {
|
message SkipFilter {
|
||||||
|
@ -148,7 +148,7 @@ message TimestampsFilter {
|
||||||
}
|
}
|
||||||
|
|
||||||
message ValueFilter {
|
message ValueFilter {
|
||||||
required CompareFilter compareFilter = 1;
|
required CompareFilter compare_filter = 1;
|
||||||
}
|
}
|
||||||
|
|
||||||
message WhileMatchFilter {
|
message WhileMatchFilter {
|
||||||
|
|
|
@ -25,21 +25,21 @@ import "hbase.proto";
|
||||||
|
|
||||||
// Map of name/values
|
// Map of name/values
|
||||||
message FileInfoProto {
|
message FileInfoProto {
|
||||||
repeated BytesBytesPair mapEntry = 1;
|
repeated BytesBytesPair map_entry = 1;
|
||||||
}
|
}
|
||||||
|
|
||||||
// HFile file trailer
|
// HFile file trailer
|
||||||
message FileTrailerProto {
|
message FileTrailerProto {
|
||||||
optional uint64 fileInfoOffset = 1;
|
optional uint64 file_info_offset = 1;
|
||||||
optional uint64 loadOnOpenDataOffset = 2;
|
optional uint64 load_on_open_data_offset = 2;
|
||||||
optional uint64 uncompressedDataIndexSize = 3;
|
optional uint64 uncompressed_data_index_size = 3;
|
||||||
optional uint64 totalUncompressedBytes = 4;
|
optional uint64 total_uncompressed_bytes = 4;
|
||||||
optional uint32 dataIndexCount = 5;
|
optional uint32 data_index_count = 5;
|
||||||
optional uint32 metaIndexCount = 6;
|
optional uint32 meta_index_count = 6;
|
||||||
optional uint64 entryCount = 7;
|
optional uint64 entry_count = 7;
|
||||||
optional uint32 numDataIndexLevels = 8;
|
optional uint32 num_data_index_levels = 8;
|
||||||
optional uint64 firstDataBlockOffset = 9;
|
optional uint64 first_data_block_offset = 9;
|
||||||
optional uint64 lastDataBlockOffset = 10;
|
optional uint64 last_data_block_offset = 10;
|
||||||
optional string comparatorClassName = 11;
|
optional string comparator_class_name = 11;
|
||||||
optional uint32 compressionCodec = 12;
|
optional uint32 compression_codec = 12;
|
||||||
}
|
}
|
||||||
|
|
|
@ -24,5 +24,5 @@ option java_generate_equals_and_hash = true;
|
||||||
option optimize_for = SPEED;
|
option optimize_for = SPEED;
|
||||||
|
|
||||||
message LoadBalancerState {
|
message LoadBalancerState {
|
||||||
optional bool balancerOn = 1;
|
optional bool balancer_on = 1;
|
||||||
}
|
}
|
||||||
|
|
|
@ -28,11 +28,11 @@ message IsMasterRunningRequest {
|
||||||
}
|
}
|
||||||
|
|
||||||
message IsMasterRunningResponse {
|
message IsMasterRunningResponse {
|
||||||
required bool isMasterRunning = 1;
|
required bool is_master_running = 1;
|
||||||
}
|
}
|
||||||
|
|
||||||
service MasterService {
|
service MasterService {
|
||||||
/** return true if master is available */
|
/** return true if master is available */
|
||||||
rpc isMasterRunning(IsMasterRunningRequest)
|
rpc IsMasterRunning(IsMasterRunningRequest)
|
||||||
returns(IsMasterRunningResponse);
|
returns(IsMasterRunningResponse);
|
||||||
}
|
}
|
||||||
|
|
|
@ -32,24 +32,24 @@ import "Client.proto";
|
||||||
/* Column-level protobufs */
|
/* Column-level protobufs */
|
||||||
|
|
||||||
message AddColumnRequest {
|
message AddColumnRequest {
|
||||||
required bytes tableName = 1;
|
required bytes table_name = 1;
|
||||||
required ColumnFamilySchema columnFamilies = 2;
|
required ColumnFamilySchema column_families = 2;
|
||||||
}
|
}
|
||||||
|
|
||||||
message AddColumnResponse {
|
message AddColumnResponse {
|
||||||
}
|
}
|
||||||
|
|
||||||
message DeleteColumnRequest {
|
message DeleteColumnRequest {
|
||||||
required bytes tableName = 1;
|
required bytes table_name = 1;
|
||||||
required bytes columnName = 2;
|
required bytes column_name = 2;
|
||||||
}
|
}
|
||||||
|
|
||||||
message DeleteColumnResponse {
|
message DeleteColumnResponse {
|
||||||
}
|
}
|
||||||
|
|
||||||
message ModifyColumnRequest {
|
message ModifyColumnRequest {
|
||||||
required bytes tableName = 1;
|
required bytes table_name = 1;
|
||||||
required ColumnFamilySchema columnFamilies = 2;
|
required ColumnFamilySchema column_families = 2;
|
||||||
}
|
}
|
||||||
|
|
||||||
message ModifyColumnResponse {
|
message ModifyColumnResponse {
|
||||||
|
@ -59,7 +59,7 @@ message ModifyColumnResponse {
|
||||||
|
|
||||||
message MoveRegionRequest {
|
message MoveRegionRequest {
|
||||||
required RegionSpecifier region = 1;
|
required RegionSpecifier region = 1;
|
||||||
optional ServerName destServerName = 2;
|
optional ServerName dest_server_name = 2;
|
||||||
}
|
}
|
||||||
|
|
||||||
message MoveRegionResponse {
|
message MoveRegionResponse {
|
||||||
|
@ -69,8 +69,8 @@ message MoveRegionResponse {
|
||||||
* Dispatch merging the specified regions.
|
* Dispatch merging the specified regions.
|
||||||
*/
|
*/
|
||||||
message DispatchMergingRegionsRequest {
|
message DispatchMergingRegionsRequest {
|
||||||
required RegionSpecifier regionA = 1;
|
required RegionSpecifier region_a = 1;
|
||||||
required RegionSpecifier regionB = 2;
|
required RegionSpecifier region_b = 2;
|
||||||
optional bool forcible = 3 [default = false];
|
optional bool forcible = 3 [default = false];
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -102,37 +102,37 @@ message OfflineRegionResponse {
|
||||||
/* Table-level protobufs */
|
/* Table-level protobufs */
|
||||||
|
|
||||||
message CreateTableRequest {
|
message CreateTableRequest {
|
||||||
required TableSchema tableSchema = 1;
|
required TableSchema table_schema = 1;
|
||||||
repeated bytes splitKeys = 2;
|
repeated bytes split_keys = 2;
|
||||||
}
|
}
|
||||||
|
|
||||||
message CreateTableResponse {
|
message CreateTableResponse {
|
||||||
}
|
}
|
||||||
|
|
||||||
message DeleteTableRequest {
|
message DeleteTableRequest {
|
||||||
required bytes tableName = 1;
|
required bytes table_name = 1;
|
||||||
}
|
}
|
||||||
|
|
||||||
message DeleteTableResponse {
|
message DeleteTableResponse {
|
||||||
}
|
}
|
||||||
|
|
||||||
message EnableTableRequest {
|
message EnableTableRequest {
|
||||||
required bytes tableName = 1;
|
required bytes table_name = 1;
|
||||||
}
|
}
|
||||||
|
|
||||||
message EnableTableResponse {
|
message EnableTableResponse {
|
||||||
}
|
}
|
||||||
|
|
||||||
message DisableTableRequest {
|
message DisableTableRequest {
|
||||||
required bytes tableName = 1;
|
required bytes table_name = 1;
|
||||||
}
|
}
|
||||||
|
|
||||||
message DisableTableResponse {
|
message DisableTableResponse {
|
||||||
}
|
}
|
||||||
|
|
||||||
message ModifyTableRequest {
|
message ModifyTableRequest {
|
||||||
required bytes tableName = 1;
|
required bytes table_name = 1;
|
||||||
required TableSchema tableSchema = 2;
|
required TableSchema table_schema = 2;
|
||||||
}
|
}
|
||||||
|
|
||||||
message ModifyTableResponse {
|
message ModifyTableResponse {
|
||||||
|
@ -157,7 +157,7 @@ message BalanceRequest {
|
||||||
}
|
}
|
||||||
|
|
||||||
message BalanceResponse {
|
message BalanceResponse {
|
||||||
required bool balancerRan = 1;
|
required bool balancer_ran = 1;
|
||||||
}
|
}
|
||||||
|
|
||||||
message SetBalancerRunningRequest {
|
message SetBalancerRunningRequest {
|
||||||
|
@ -166,14 +166,14 @@ message SetBalancerRunningRequest {
|
||||||
}
|
}
|
||||||
|
|
||||||
message SetBalancerRunningResponse {
|
message SetBalancerRunningResponse {
|
||||||
optional bool prevBalanceValue = 1;
|
optional bool prev_balance_value = 1;
|
||||||
}
|
}
|
||||||
|
|
||||||
message CatalogScanRequest {
|
message CatalogScanRequest {
|
||||||
}
|
}
|
||||||
|
|
||||||
message CatalogScanResponse {
|
message CatalogScanResponse {
|
||||||
optional int32 scanResult = 1;
|
optional int32 scan_result = 1;
|
||||||
}
|
}
|
||||||
|
|
||||||
message EnableCatalogJanitorRequest {
|
message EnableCatalogJanitorRequest {
|
||||||
|
@ -181,7 +181,7 @@ message EnableCatalogJanitorRequest {
|
||||||
}
|
}
|
||||||
|
|
||||||
message EnableCatalogJanitorResponse {
|
message EnableCatalogJanitorResponse {
|
||||||
optional bool prevValue = 1;
|
optional bool prev_value = 1;
|
||||||
}
|
}
|
||||||
|
|
||||||
message IsCatalogJanitorEnabledRequest {
|
message IsCatalogJanitorEnabledRequest {
|
||||||
|
@ -196,7 +196,7 @@ message TakeSnapshotRequest{
|
||||||
}
|
}
|
||||||
|
|
||||||
message TakeSnapshotResponse{
|
message TakeSnapshotResponse{
|
||||||
required int64 expectedTimeout = 1;
|
required int64 expected_timeout = 1;
|
||||||
}
|
}
|
||||||
|
|
||||||
message ListSnapshotRequest{
|
message ListSnapshotRequest{
|
||||||
|
@ -242,36 +242,36 @@ message IsRestoreSnapshotDoneResponse {
|
||||||
|
|
||||||
service MasterAdminService {
|
service MasterAdminService {
|
||||||
/** Adds a column to the specified table. */
|
/** Adds a column to the specified table. */
|
||||||
rpc addColumn(AddColumnRequest)
|
rpc AddColumn(AddColumnRequest)
|
||||||
returns(AddColumnResponse);
|
returns(AddColumnResponse);
|
||||||
|
|
||||||
/** Deletes a column from the specified table. Table must be disabled. */
|
/** Deletes a column from the specified table. Table must be disabled. */
|
||||||
rpc deleteColumn(DeleteColumnRequest)
|
rpc DeleteColumn(DeleteColumnRequest)
|
||||||
returns(DeleteColumnResponse);
|
returns(DeleteColumnResponse);
|
||||||
|
|
||||||
/** Modifies an existing column on the specified table. */
|
/** Modifies an existing column on the specified table. */
|
||||||
rpc modifyColumn(ModifyColumnRequest)
|
rpc ModifyColumn(ModifyColumnRequest)
|
||||||
returns(ModifyColumnResponse);
|
returns(ModifyColumnResponse);
|
||||||
|
|
||||||
/** Move the region region to the destination server. */
|
/** Move the region region to the destination server. */
|
||||||
rpc moveRegion(MoveRegionRequest)
|
rpc MoveRegion(MoveRegionRequest)
|
||||||
returns(MoveRegionResponse);
|
returns(MoveRegionResponse);
|
||||||
|
|
||||||
/** Master dispatch merging the regions */
|
/** Master dispatch merging the regions */
|
||||||
rpc dispatchMergingRegions(DispatchMergingRegionsRequest)
|
rpc DispatchMergingRegions(DispatchMergingRegionsRequest)
|
||||||
returns(DispatchMergingRegionsResponse);
|
returns(DispatchMergingRegionsResponse);
|
||||||
|
|
||||||
/** Assign a region to a server chosen at random. */
|
/** Assign a region to a server chosen at random. */
|
||||||
rpc assignRegion(AssignRegionRequest)
|
rpc AssignRegion(AssignRegionRequest)
|
||||||
returns(AssignRegionResponse);
|
returns(AssignRegionResponse);
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Unassign a region from current hosting regionserver. Region will then be
|
* Unassign a region from current hosting regionserver. Region will then be
|
||||||
* assigned to a regionserver chosen at random. Region could be reassigned
|
* assigned to a regionserver chosen at random. Region could be reassigned
|
||||||
* back to the same server. Use moveRegion if you want
|
* back to the same server. Use MoveRegion if you want
|
||||||
* to control the region movement.
|
* to control the region movement.
|
||||||
*/
|
*/
|
||||||
rpc unassignRegion(UnassignRegionRequest)
|
rpc UnassignRegion(UnassignRegionRequest)
|
||||||
returns(UnassignRegionResponse);
|
returns(UnassignRegionResponse);
|
||||||
|
|
||||||
/**
|
/**
|
||||||
|
@ -280,35 +280,35 @@ service MasterAdminService {
|
||||||
* automatically reassign the region as in unassign. This is a special
|
* automatically reassign the region as in unassign. This is a special
|
||||||
* method, and should only be used by experts or hbck.
|
* method, and should only be used by experts or hbck.
|
||||||
*/
|
*/
|
||||||
rpc offlineRegion(OfflineRegionRequest)
|
rpc OfflineRegion(OfflineRegionRequest)
|
||||||
returns(OfflineRegionResponse);
|
returns(OfflineRegionResponse);
|
||||||
|
|
||||||
/** Deletes a table */
|
/** Deletes a table */
|
||||||
rpc deleteTable(DeleteTableRequest)
|
rpc DeleteTable(DeleteTableRequest)
|
||||||
returns(DeleteTableResponse);
|
returns(DeleteTableResponse);
|
||||||
|
|
||||||
/** Puts the table on-line (only needed if table has been previously taken offline) */
|
/** Puts the table on-line (only needed if table has been previously taken offline) */
|
||||||
rpc enableTable(EnableTableRequest)
|
rpc EnableTable(EnableTableRequest)
|
||||||
returns(EnableTableResponse);
|
returns(EnableTableResponse);
|
||||||
|
|
||||||
/** Take table offline */
|
/** Take table offline */
|
||||||
rpc disableTable(DisableTableRequest)
|
rpc DisableTable(DisableTableRequest)
|
||||||
returns(DisableTableResponse);
|
returns(DisableTableResponse);
|
||||||
|
|
||||||
/** Modify a table's metadata */
|
/** Modify a table's metadata */
|
||||||
rpc modifyTable(ModifyTableRequest)
|
rpc ModifyTable(ModifyTableRequest)
|
||||||
returns(ModifyTableResponse);
|
returns(ModifyTableResponse);
|
||||||
|
|
||||||
/** Creates a new table asynchronously */
|
/** Creates a new table asynchronously */
|
||||||
rpc createTable(CreateTableRequest)
|
rpc CreateTable(CreateTableRequest)
|
||||||
returns(CreateTableResponse);
|
returns(CreateTableResponse);
|
||||||
|
|
||||||
/** Shutdown an HBase cluster. */
|
/** Shutdown an HBase cluster. */
|
||||||
rpc shutdown(ShutdownRequest)
|
rpc Shutdown(ShutdownRequest)
|
||||||
returns(ShutdownResponse);
|
returns(ShutdownResponse);
|
||||||
|
|
||||||
/** Stop HBase Master only. Does not shutdown the cluster. */
|
/** Stop HBase Master only. Does not shutdown the cluster. */
|
||||||
rpc stopMaster(StopMasterRequest)
|
rpc StopMaster(StopMasterRequest)
|
||||||
returns(StopMasterResponse);
|
returns(StopMasterResponse);
|
||||||
|
|
||||||
/**
|
/**
|
||||||
|
@ -316,72 +316,72 @@ service MasterAdminService {
|
||||||
* go ahead and do the reassignments. Can NOT run for various reasons.
|
* go ahead and do the reassignments. Can NOT run for various reasons.
|
||||||
* Check logs.
|
* Check logs.
|
||||||
*/
|
*/
|
||||||
rpc balance(BalanceRequest)
|
rpc Balance(BalanceRequest)
|
||||||
returns(BalanceResponse);
|
returns(BalanceResponse);
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Turn the load balancer on or off.
|
* Turn the load balancer on or off.
|
||||||
* If synchronous is true, it waits until current balance() call, if outstanding, to return.
|
* If synchronous is true, it waits until current balance() call, if outstanding, to return.
|
||||||
*/
|
*/
|
||||||
rpc setBalancerRunning(SetBalancerRunningRequest)
|
rpc SetBalancerRunning(SetBalancerRunningRequest)
|
||||||
returns(SetBalancerRunningResponse);
|
returns(SetBalancerRunningResponse);
|
||||||
|
|
||||||
/** Get a run of the catalog janitor */
|
/** Get a run of the catalog janitor */
|
||||||
rpc runCatalogScan(CatalogScanRequest)
|
rpc RunCatalogScan(CatalogScanRequest)
|
||||||
returns(CatalogScanResponse);
|
returns(CatalogScanResponse);
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Enable the catalog janitor on or off.
|
* Enable the catalog janitor on or off.
|
||||||
*/
|
*/
|
||||||
rpc enableCatalogJanitor(EnableCatalogJanitorRequest)
|
rpc EnableCatalogJanitor(EnableCatalogJanitorRequest)
|
||||||
returns(EnableCatalogJanitorResponse);
|
returns(EnableCatalogJanitorResponse);
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Query whether the catalog janitor is enabled.
|
* Query whether the catalog janitor is enabled.
|
||||||
*/
|
*/
|
||||||
rpc isCatalogJanitorEnabled(IsCatalogJanitorEnabledRequest)
|
rpc IsCatalogJanitorEnabled(IsCatalogJanitorEnabledRequest)
|
||||||
returns(IsCatalogJanitorEnabledResponse);
|
returns(IsCatalogJanitorEnabledResponse);
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Call a master coprocessor endpoint
|
* Call a master coprocessor endpoint
|
||||||
*/
|
*/
|
||||||
rpc execMasterService(CoprocessorServiceRequest)
|
rpc ExecMasterService(CoprocessorServiceRequest)
|
||||||
returns(CoprocessorServiceResponse);
|
returns(CoprocessorServiceResponse);
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Create a snapshot for the given table.
|
* Create a snapshot for the given table.
|
||||||
* @param snapshot description of the snapshot to take
|
* @param snapshot description of the snapshot to take
|
||||||
*/
|
*/
|
||||||
rpc snapshot(TakeSnapshotRequest) returns(TakeSnapshotResponse);
|
rpc Snapshot(TakeSnapshotRequest) returns(TakeSnapshotResponse);
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* List completed snapshots.
|
* List completed snapshots.
|
||||||
* Returns a list of snapshot descriptors for completed snapshots
|
* Returns a list of snapshot descriptors for completed snapshots
|
||||||
*/
|
*/
|
||||||
rpc getCompletedSnapshots(ListSnapshotRequest) returns(ListSnapshotResponse);
|
rpc GetCompletedSnapshots(ListSnapshotRequest) returns(ListSnapshotResponse);
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Delete an existing snapshot. This method can also be used to clean up an aborted snapshot.
|
* Delete an existing snapshot. This method can also be used to clean up an aborted snapshot.
|
||||||
* @param snapshotName snapshot to delete
|
* @param snapshotName snapshot to delete
|
||||||
*/
|
*/
|
||||||
rpc deleteSnapshot(DeleteSnapshotRequest) returns(DeleteSnapshotResponse);
|
rpc DeleteSnapshot(DeleteSnapshotRequest) returns(DeleteSnapshotResponse);
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Determine if the snapshot is done yet.
|
* Determine if the snapshot is done yet.
|
||||||
*/
|
*/
|
||||||
rpc isSnapshotDone(IsSnapshotDoneRequest) returns(IsSnapshotDoneResponse);
|
rpc IsSnapshotDone(IsSnapshotDoneRequest) returns(IsSnapshotDoneResponse);
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Restore a snapshot
|
* Restore a snapshot
|
||||||
* @param snapshot description of the snapshot to restore
|
* @param snapshot description of the snapshot to restore
|
||||||
*/
|
*/
|
||||||
rpc restoreSnapshot(RestoreSnapshotRequest) returns(RestoreSnapshotResponse);
|
rpc RestoreSnapshot(RestoreSnapshotRequest) returns(RestoreSnapshotResponse);
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Determine if the snapshot restore is done yet.
|
* Determine if the snapshot restore is done yet.
|
||||||
*/
|
*/
|
||||||
rpc isRestoreSnapshotDone(IsRestoreSnapshotDoneRequest) returns(IsRestoreSnapshotDoneResponse);
|
rpc IsRestoreSnapshotDone(IsRestoreSnapshotDoneRequest) returns(IsRestoreSnapshotDoneResponse);
|
||||||
|
|
||||||
/** return true if master is available */
|
/** return true if master is available */
|
||||||
rpc isMasterRunning(IsMasterRunningRequest) returns(IsMasterRunningResponse);
|
rpc IsMasterRunning(IsMasterRunningRequest) returns(IsMasterRunningResponse);
|
||||||
}
|
}
|
||||||
|
|
|
@ -29,42 +29,42 @@ import "hbase.proto";
|
||||||
import "ClusterStatus.proto";
|
import "ClusterStatus.proto";
|
||||||
|
|
||||||
message GetSchemaAlterStatusRequest {
|
message GetSchemaAlterStatusRequest {
|
||||||
required bytes tableName = 1;
|
required bytes table_name = 1;
|
||||||
}
|
}
|
||||||
|
|
||||||
message GetSchemaAlterStatusResponse {
|
message GetSchemaAlterStatusResponse {
|
||||||
optional uint32 yetToUpdateRegions = 1;
|
optional uint32 yet_to_update_regions = 1;
|
||||||
optional uint32 totalRegions = 2;
|
optional uint32 total_regions = 2;
|
||||||
}
|
}
|
||||||
|
|
||||||
message GetTableDescriptorsRequest {
|
message GetTableDescriptorsRequest {
|
||||||
repeated string tableNames = 1;
|
repeated string table_names = 1;
|
||||||
}
|
}
|
||||||
|
|
||||||
message GetTableDescriptorsResponse {
|
message GetTableDescriptorsResponse {
|
||||||
repeated TableSchema tableSchema = 1;
|
repeated TableSchema table_schema = 1;
|
||||||
}
|
}
|
||||||
|
|
||||||
message GetClusterStatusRequest {
|
message GetClusterStatusRequest {
|
||||||
}
|
}
|
||||||
|
|
||||||
message GetClusterStatusResponse {
|
message GetClusterStatusResponse {
|
||||||
required ClusterStatus clusterStatus = 1;
|
required ClusterStatus cluster_status = 1;
|
||||||
}
|
}
|
||||||
|
|
||||||
service MasterMonitorService {
|
service MasterMonitorService {
|
||||||
/** Used by the client to get the number of regions that have received the updated schema */
|
/** Used by the client to get the number of regions that have received the updated schema */
|
||||||
rpc getSchemaAlterStatus(GetSchemaAlterStatusRequest)
|
rpc GetSchemaAlterStatus(GetSchemaAlterStatusRequest)
|
||||||
returns(GetSchemaAlterStatusResponse);
|
returns(GetSchemaAlterStatusResponse);
|
||||||
|
|
||||||
/** Get list of TableDescriptors for requested tables. */
|
/** Get list of TableDescriptors for requested tables. */
|
||||||
rpc getTableDescriptors(GetTableDescriptorsRequest)
|
rpc GetTableDescriptors(GetTableDescriptorsRequest)
|
||||||
returns(GetTableDescriptorsResponse);
|
returns(GetTableDescriptorsResponse);
|
||||||
|
|
||||||
/** Return cluster status. */
|
/** Return cluster status. */
|
||||||
rpc getClusterStatus(GetClusterStatusRequest)
|
rpc GetClusterStatus(GetClusterStatusRequest)
|
||||||
returns(GetClusterStatusResponse);
|
returns(GetClusterStatusResponse);
|
||||||
|
|
||||||
/** return true if master is available */
|
/** return true if master is available */
|
||||||
rpc isMasterRunning(IsMasterRunningRequest) returns(IsMasterRunningResponse);
|
rpc IsMasterRunning(IsMasterRunningRequest) returns(IsMasterRunningResponse);
|
||||||
}
|
}
|
||||||
|
|
|
@ -23,13 +23,13 @@ option java_generic_services = true;
|
||||||
option optimize_for = SPEED;
|
option optimize_for = SPEED;
|
||||||
|
|
||||||
message MultiMutateRequest {
|
message MultiMutateRequest {
|
||||||
repeated MutationProto mutationRequest = 1;
|
repeated MutationProto mutation_request = 1;
|
||||||
}
|
}
|
||||||
|
|
||||||
message MultiMutateResponse {
|
message MultiMutateResponse {
|
||||||
}
|
}
|
||||||
|
|
||||||
service MultiRowMutationService {
|
service MultiRowMutationService {
|
||||||
rpc mutateRows(MultiMutateRequest)
|
rpc MutateRows(MultiMutateRequest)
|
||||||
returns(MultiMutateResponse);
|
returns(MultiMutateResponse);
|
||||||
}
|
}
|
||||||
|
|
|
@ -72,20 +72,20 @@ option optimize_for = SPEED;
|
||||||
|
|
||||||
// User Information proto. Included in ConnectionHeader on connection setup
|
// User Information proto. Included in ConnectionHeader on connection setup
|
||||||
message UserInformation {
|
message UserInformation {
|
||||||
required string effectiveUser = 1;
|
required string effective_user = 1;
|
||||||
optional string realUser = 2;
|
optional string real_user = 2;
|
||||||
}
|
}
|
||||||
|
|
||||||
// This is sent on connection setup after the connection preamble is sent.
|
// This is sent on connection setup after the connection preamble is sent.
|
||||||
message ConnectionHeader {
|
message ConnectionHeader {
|
||||||
optional UserInformation userInfo = 1;
|
optional UserInformation user_info = 1;
|
||||||
optional string serviceName = 2;
|
optional string service_name = 2;
|
||||||
// Cell block codec we will use sending over optional cell blocks. Server throws exception
|
// Cell block codec we will use sending over optional cell blocks. Server throws exception
|
||||||
// if cannot deal.
|
// if cannot deal.
|
||||||
optional string cellBlockCodecClass = 3 [default = "org.apache.hadoop.hbase.codec.KeyValueCodec"];
|
optional string cell_block_codec_class = 3 [default = "org.apache.hadoop.hbase.codec.KeyValueCodec"];
|
||||||
// Compressor we will use if cell block is compressed. Server will throw exception if not supported.
|
// Compressor we will use if cell block is compressed. Server will throw exception if not supported.
|
||||||
// Class must implement hadoop's CompressionCodec Interface
|
// Class must implement hadoop's CompressionCodec Interface
|
||||||
optional string cellBlockCompressorClass = 4;
|
optional string cell_block_compressor_class = 4;
|
||||||
}
|
}
|
||||||
|
|
||||||
// Optional Cell block Message. Included in client RequestHeader
|
// Optional Cell block Message. Included in client RequestHeader
|
||||||
|
@ -98,34 +98,34 @@ message CellBlockMeta {
|
||||||
// the server side exception to the RPC client.
|
// the server side exception to the RPC client.
|
||||||
message ExceptionResponse {
|
message ExceptionResponse {
|
||||||
// Class name of the exception thrown from the server
|
// Class name of the exception thrown from the server
|
||||||
optional string exceptionClassName = 1;
|
optional string exception_class_name = 1;
|
||||||
// Exception stack trace from the server side
|
// Exception stack trace from the server side
|
||||||
optional string stackTrace = 2;
|
optional string stack_trace = 2;
|
||||||
// Optional hostname. Filled in for some exceptions such as region moved
|
// Optional hostname. Filled in for some exceptions such as region moved
|
||||||
// where exception gives clue on where the region may have moved.
|
// where exception gives clue on where the region may have moved.
|
||||||
optional string hostname = 3;
|
optional string hostname = 3;
|
||||||
optional int32 port = 4;
|
optional int32 port = 4;
|
||||||
// Set if we are NOT to retry on receipt of this exception
|
// Set if we are NOT to retry on receipt of this exception
|
||||||
optional bool doNotRetry = 5;
|
optional bool do_not_retry = 5;
|
||||||
}
|
}
|
||||||
|
|
||||||
// Header sent making a request.
|
// Header sent making a request.
|
||||||
message RequestHeader {
|
message RequestHeader {
|
||||||
// Monotonically increasing callId to keep track of RPC requests and their response
|
// Monotonically increasing call_id to keep track of RPC requests and their response
|
||||||
optional uint32 callId = 1;
|
optional uint32 call_id = 1;
|
||||||
optional RPCTInfo traceInfo = 2;
|
optional RPCTInfo trace_info = 2;
|
||||||
optional string methodName = 3;
|
optional string method_name = 3;
|
||||||
// If true, then a pb Message param follows.
|
// If true, then a pb Message param follows.
|
||||||
optional bool requestParam = 4;
|
optional bool request_param = 4;
|
||||||
// If present, then an encoded data block follows.
|
// If present, then an encoded data block follows.
|
||||||
optional CellBlockMeta cellBlockMeta = 5;
|
optional CellBlockMeta cell_block_meta = 5;
|
||||||
// TODO: Have client specify priority
|
// TODO: Have client specify priority
|
||||||
}
|
}
|
||||||
|
|
||||||
message ResponseHeader {
|
message ResponseHeader {
|
||||||
optional uint32 callId = 1;
|
optional uint32 call_id = 1;
|
||||||
// If present, then request threw an exception and no response message (else we presume one)
|
// If present, then request threw an exception and no response message (else we presume one)
|
||||||
optional ExceptionResponse exception = 2;
|
optional ExceptionResponse exception = 2;
|
||||||
// If present, then an encoded data block follows.
|
// If present, then an encoded data block follows.
|
||||||
optional CellBlockMeta cellBlockMeta = 3;
|
optional CellBlockMeta cell_block_meta = 3;
|
||||||
}
|
}
|
||||||
|
|
|
@ -32,10 +32,10 @@ message RegionServerStartupRequest {
|
||||||
required uint32 port = 1;
|
required uint32 port = 1;
|
||||||
|
|
||||||
/** This servers' startcode */
|
/** This servers' startcode */
|
||||||
required uint64 serverStartCode = 2;
|
required uint64 server_start_code = 2;
|
||||||
|
|
||||||
/** Current time of the region server in ms */
|
/** Current time of the region server in ms */
|
||||||
required uint64 serverCurrentTime = 3;
|
required uint64 server_current_time = 3;
|
||||||
}
|
}
|
||||||
|
|
||||||
message RegionServerStartupResponse {
|
message RegionServerStartupResponse {
|
||||||
|
@ -44,7 +44,7 @@ message RegionServerStartupResponse {
|
||||||
* hbase rootdir, the hostname to use creating the RegionServer ServerName,
|
* hbase rootdir, the hostname to use creating the RegionServer ServerName,
|
||||||
* etc
|
* etc
|
||||||
*/
|
*/
|
||||||
repeated NameStringPair mapEntries = 1;
|
repeated NameStringPair map_entries = 1;
|
||||||
}
|
}
|
||||||
|
|
||||||
message RegionServerReportRequest {
|
message RegionServerReportRequest {
|
||||||
|
@ -62,7 +62,7 @@ message ReportRSFatalErrorRequest {
|
||||||
required ServerName server = 1;
|
required ServerName server = 1;
|
||||||
|
|
||||||
/** informative text to expose in the master logs and UI */
|
/** informative text to expose in the master logs and UI */
|
||||||
required string errorMessage = 2;
|
required string error_message = 2;
|
||||||
}
|
}
|
||||||
|
|
||||||
message ReportRSFatalErrorResponse {
|
message ReportRSFatalErrorResponse {
|
||||||
|
@ -70,33 +70,33 @@ message ReportRSFatalErrorResponse {
|
||||||
|
|
||||||
message GetLastFlushedSequenceIdRequest {
|
message GetLastFlushedSequenceIdRequest {
|
||||||
/** region name */
|
/** region name */
|
||||||
required bytes regionName = 1;
|
required bytes region_name = 1;
|
||||||
}
|
}
|
||||||
|
|
||||||
message GetLastFlushedSequenceIdResponse {
|
message GetLastFlushedSequenceIdResponse {
|
||||||
/** the last HLog sequence id flushed from MemStore to HFile for the region */
|
/** the last HLog sequence id flushed from MemStore to HFile for the region */
|
||||||
required uint64 lastFlushedSequenceId = 1;
|
required uint64 last_flushed_sequence_id = 1;
|
||||||
}
|
}
|
||||||
|
|
||||||
service RegionServerStatusService {
|
service RegionServerStatusService {
|
||||||
/** Called when a region server first starts. */
|
/** Called when a region server first starts. */
|
||||||
rpc regionServerStartup(RegionServerStartupRequest)
|
rpc RegionServerStartup(RegionServerStartupRequest)
|
||||||
returns(RegionServerStartupResponse);
|
returns(RegionServerStartupResponse);
|
||||||
|
|
||||||
/** Called to report the load the RegionServer is under. */
|
/** Called to report the load the RegionServer is under. */
|
||||||
rpc regionServerReport(RegionServerReportRequest)
|
rpc RegionServerReport(RegionServerReportRequest)
|
||||||
returns(RegionServerReportResponse);
|
returns(RegionServerReportResponse);
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Called by a region server to report a fatal error that is causing it to
|
* Called by a region server to report a fatal error that is causing it to
|
||||||
* abort.
|
* abort.
|
||||||
*/
|
*/
|
||||||
rpc reportRSFatalError(ReportRSFatalErrorRequest)
|
rpc ReportRSFatalError(ReportRSFatalErrorRequest)
|
||||||
returns(ReportRSFatalErrorResponse);
|
returns(ReportRSFatalErrorResponse);
|
||||||
|
|
||||||
/** Called to get the sequence id of the last MemStore entry flushed to an
|
/** Called to get the sequence id of the last MemStore entry flushed to an
|
||||||
* HFile for a specified region. Used by the region server to speed up
|
* HFile for a specified region. Used by the region server to speed up
|
||||||
* log splitting. */
|
* log splitting. */
|
||||||
rpc getLastFlushedSequenceId(GetLastFlushedSequenceIdRequest)
|
rpc GetLastFlushedSequenceId(GetLastFlushedSequenceIdRequest)
|
||||||
returns(GetLastFlushedSequenceIdResponse);
|
returns(GetLastFlushedSequenceIdResponse);
|
||||||
}
|
}
|
||||||
|
|
|
@ -27,15 +27,15 @@ option java_generate_equals_and_hash = true;
|
||||||
option optimize_for = SPEED;
|
option optimize_for = SPEED;
|
||||||
|
|
||||||
message RowProcessorRequest {
|
message RowProcessorRequest {
|
||||||
required string rowProcessorClassName = 1;
|
required string row_processor_class_name = 1;
|
||||||
optional string rowProcessorInitializerMessageName = 2;
|
optional string row_processor_initializer_message_name = 2;
|
||||||
optional bytes rowProcessorInitializerMessage = 3;
|
optional bytes row_processor_initializer_message = 3;
|
||||||
}
|
}
|
||||||
|
|
||||||
message RowProcessorResult {
|
message RowProcessorResult {
|
||||||
required bytes rowProcessorResult = 1;
|
required bytes row_processor_result = 1;
|
||||||
}
|
}
|
||||||
|
|
||||||
service RowProcessorService {
|
service RowProcessorService {
|
||||||
rpc process (RowProcessorRequest) returns (RowProcessorResult);
|
rpc Process (RowProcessorRequest) returns (RowProcessorResult);
|
||||||
}
|
}
|
|
@ -26,10 +26,10 @@ import 'hbase.proto';
|
||||||
import 'Client.proto';
|
import 'Client.proto';
|
||||||
|
|
||||||
message SecureBulkLoadHFilesRequest {
|
message SecureBulkLoadHFilesRequest {
|
||||||
repeated BulkLoadHFileRequest.FamilyPath familyPath = 1;
|
repeated BulkLoadHFileRequest.FamilyPath family_path = 1;
|
||||||
optional bool assignSeqNum = 2;
|
optional bool assign_seq_num = 2;
|
||||||
required DelegationTokenProto fsToken = 3;
|
required DelegationTokenProto fs_token = 3;
|
||||||
required string bulkToken = 4;
|
required string bulk_token = 4;
|
||||||
}
|
}
|
||||||
|
|
||||||
message SecureBulkLoadHFilesResponse {
|
message SecureBulkLoadHFilesResponse {
|
||||||
|
@ -44,15 +44,15 @@ message DelegationTokenProto {
|
||||||
}
|
}
|
||||||
|
|
||||||
message PrepareBulkLoadRequest {
|
message PrepareBulkLoadRequest {
|
||||||
required bytes tableName = 1;
|
required bytes table_name = 1;
|
||||||
}
|
}
|
||||||
|
|
||||||
message PrepareBulkLoadResponse {
|
message PrepareBulkLoadResponse {
|
||||||
required string bulkToken = 1;
|
required string bulk_token = 1;
|
||||||
}
|
}
|
||||||
|
|
||||||
message CleanupBulkLoadRequest {
|
message CleanupBulkLoadRequest {
|
||||||
required string bulkToken = 1;
|
required string bulk_token = 1;
|
||||||
|
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -60,12 +60,12 @@ message CleanupBulkLoadResponse {
|
||||||
}
|
}
|
||||||
|
|
||||||
service SecureBulkLoadService {
|
service SecureBulkLoadService {
|
||||||
rpc prepareBulkLoad(PrepareBulkLoadRequest)
|
rpc PrepareBulkLoad(PrepareBulkLoadRequest)
|
||||||
returns (PrepareBulkLoadResponse);
|
returns (PrepareBulkLoadResponse);
|
||||||
|
|
||||||
rpc secureBulkLoadHFiles(SecureBulkLoadHFilesRequest)
|
rpc SecureBulkLoadHFiles(SecureBulkLoadHFilesRequest)
|
||||||
returns (SecureBulkLoadHFilesResponse);
|
returns (SecureBulkLoadHFilesResponse);
|
||||||
|
|
||||||
rpc cleanupBulkLoad(CleanupBulkLoadRequest)
|
rpc CleanupBulkLoad(CleanupBulkLoadRequest)
|
||||||
returns (CleanupBulkLoadResponse);
|
returns (CleanupBulkLoadResponse);
|
||||||
}
|
}
|
||||||
|
|
|
@ -26,6 +26,6 @@ option optimize_for = SPEED;
|
||||||
//the id of the current span when this message was sent, so we know
|
//the id of the current span when this message was sent, so we know
|
||||||
//what span caused the new span we will create when this message is received.
|
//what span caused the new span we will create when this message is received.
|
||||||
message RPCTInfo {
|
message RPCTInfo {
|
||||||
optional int64 traceId = 1;
|
optional int64 trace_id = 1;
|
||||||
optional int64 parentId = 2;
|
optional int64 parent_id = 2;
|
||||||
}
|
}
|
||||||
|
|
|
@ -24,21 +24,21 @@ option optimize_for = SPEED;
|
||||||
import "hbase.proto";
|
import "hbase.proto";
|
||||||
|
|
||||||
message WALHeader {
|
message WALHeader {
|
||||||
optional bool hasCompression = 1;
|
optional bool has_compression = 1;
|
||||||
}
|
}
|
||||||
|
|
||||||
// Protocol buffer version of HLogKey; see HLogKey comment, not really a key but WALEdit header for some KVs
|
// Protocol buffer version of HLogKey; see HLogKey comment, not really a key but WALEdit header for some KVs
|
||||||
message WALKey {
|
message WALKey {
|
||||||
required bytes encodedRegionName = 1;
|
required bytes encoded_region_name = 1;
|
||||||
required bytes tableName = 2;
|
required bytes table_name = 2;
|
||||||
required uint64 logSequenceNumber = 3;
|
required uint64 log_sequence_number = 3;
|
||||||
required uint64 writeTime = 4;
|
required uint64 write_time = 4;
|
||||||
optional UUID clusterId = 5;
|
optional UUID cluster_id = 5;
|
||||||
|
|
||||||
repeated FamilyScope scopes = 6;
|
repeated FamilyScope scopes = 6;
|
||||||
optional uint32 followingKvCount = 7;
|
optional uint32 following_kv_count = 7;
|
||||||
/*
|
/*
|
||||||
optional CustomEntryType customEntryType = 8;
|
optional CustomEntryType custom_entry_type = 8;
|
||||||
|
|
||||||
enum CustomEntryType {
|
enum CustomEntryType {
|
||||||
COMPACTION = 0;
|
COMPACTION = 0;
|
||||||
|
@ -53,7 +53,7 @@ enum ScopeType {
|
||||||
|
|
||||||
message FamilyScope {
|
message FamilyScope {
|
||||||
required bytes family = 1;
|
required bytes family = 1;
|
||||||
required ScopeType scopeType = 2;
|
required ScopeType scope_type = 2;
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
|
@ -67,12 +67,12 @@ message FamilyScope {
|
||||||
* the * compaction should we fail the WAL write.
|
* the * compaction should we fail the WAL write.
|
||||||
*/
|
*/
|
||||||
message CompactionDescriptor {
|
message CompactionDescriptor {
|
||||||
required bytes tableName = 1; // TODO: WALKey already stores these, might remove
|
required bytes table_name = 1; // TODO: WALKey already stores these, might remove
|
||||||
required bytes encodedRegionName = 2;
|
required bytes encoded_region_name = 2;
|
||||||
required bytes familyName = 3;
|
required bytes family_name = 3;
|
||||||
repeated string compactionInput = 4;
|
repeated string compaction_input = 4;
|
||||||
repeated string compactionOutput = 5;
|
repeated string compaction_output = 5;
|
||||||
required string storeHomeDir = 6;
|
required string store_home_dir = 6;
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
|
|
|
@ -36,7 +36,7 @@ message MetaRegionServer {
|
||||||
// The major version of the rpc the server speaks. This is used so that
|
// The major version of the rpc the server speaks. This is used so that
|
||||||
// clients connecting to the cluster can have prior knowledge of what version
|
// clients connecting to the cluster can have prior knowledge of what version
|
||||||
// to send to a RegionServer. AsyncHBase will use this to detect versions.
|
// to send to a RegionServer. AsyncHBase will use this to detect versions.
|
||||||
optional uint32 rpcVersion = 2;
|
optional uint32 rpc_version = 2;
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
|
@ -46,7 +46,7 @@ message Master {
|
||||||
// The ServerName of the current Master
|
// The ServerName of the current Master
|
||||||
required ServerName master = 1;
|
required ServerName master = 1;
|
||||||
// Major RPC version so that clients can know what version the master can accept.
|
// Major RPC version so that clients can know what version the master can accept.
|
||||||
optional uint32 rpcVersion = 2;
|
optional uint32 rpc_version = 2;
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
|
@ -54,8 +54,8 @@ message Master {
|
||||||
*/
|
*/
|
||||||
message ClusterUp {
|
message ClusterUp {
|
||||||
// If this znode is present, cluster is up. Currently
|
// If this znode is present, cluster is up. Currently
|
||||||
// the data is cluster startDate.
|
// the data is cluster start_date.
|
||||||
required string startDate = 1;
|
required string start_date = 1;
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
|
@ -64,12 +64,12 @@ message ClusterUp {
|
||||||
*/
|
*/
|
||||||
message RegionTransition {
|
message RegionTransition {
|
||||||
// Code for EventType gotten by doing o.a.h.h.EventHandler.EventType.getCode()
|
// Code for EventType gotten by doing o.a.h.h.EventHandler.EventType.getCode()
|
||||||
required uint32 eventTypeCode = 1;
|
required uint32 event_type_code = 1;
|
||||||
// Full regionname in bytes
|
// Full regionname in bytes
|
||||||
required bytes regionName = 2;
|
required bytes region_name = 2;
|
||||||
required uint64 createTime = 3;
|
required uint64 create_time = 3;
|
||||||
// The region server where the transition will happen or is happening
|
// The region server where the transition will happen or is happening
|
||||||
required ServerName serverName = 4;
|
required ServerName server_name = 4;
|
||||||
optional bytes payload = 5;
|
optional bytes payload = 5;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -86,7 +86,7 @@ message SplitLogTask {
|
||||||
ERR = 4;
|
ERR = 4;
|
||||||
}
|
}
|
||||||
required State state = 1;
|
required State state = 1;
|
||||||
required ServerName serverName = 2;
|
required ServerName server_name = 2;
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
|
@ -110,7 +110,7 @@ message Table {
|
||||||
* Used by replication. Holds a replication peer key.
|
* Used by replication. Holds a replication peer key.
|
||||||
*/
|
*/
|
||||||
message ReplicationPeer {
|
message ReplicationPeer {
|
||||||
// clusterKey is the concatenation of the slave cluster's
|
// clusterkey is the concatenation of the slave cluster's
|
||||||
// hbase.zookeeper.quorum:hbase.zookeeper.property.clientPort:zookeeper.znode.parent
|
// hbase.zookeeper.quorum:hbase.zookeeper.property.clientPort:zookeeper.znode.parent
|
||||||
required string clusterkey = 1;
|
required string clusterkey = 1;
|
||||||
}
|
}
|
||||||
|
@ -137,27 +137,27 @@ message ReplicationHLogPosition {
|
||||||
* Used by replication. Used to lock a region server during failover.
|
* Used by replication. Used to lock a region server during failover.
|
||||||
*/
|
*/
|
||||||
message ReplicationLock {
|
message ReplicationLock {
|
||||||
required string lockOwner = 1;
|
required string lock_owner = 1;
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Metadata associated with a table lock in zookeeper
|
* Metadata associated with a table lock in zookeeper
|
||||||
*/
|
*/
|
||||||
message TableLock {
|
message TableLock {
|
||||||
optional bytes tableName = 1;
|
optional bytes table_name = 1;
|
||||||
optional ServerName lockOwner = 2;
|
optional ServerName lock_owner = 2;
|
||||||
optional int64 threadId = 3;
|
optional int64 thread_id = 3;
|
||||||
optional bool isShared = 4;
|
optional bool is_shared = 4;
|
||||||
optional string purpose = 5;
|
optional string purpose = 5;
|
||||||
optional int64 createTime = 6;
|
optional int64 create_time = 6;
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* sequence Id of a store
|
* sequence Id of a store
|
||||||
*/
|
*/
|
||||||
message StoreSequenceId {
|
message StoreSequenceId {
|
||||||
required bytes familyName = 1;
|
required bytes family_name = 1;
|
||||||
required uint64 sequenceId = 2;
|
required uint64 sequence_id = 2;
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
|
@ -165,6 +165,6 @@ message StoreSequenceId {
|
||||||
* list sequence ids of the region's stores
|
* list sequence ids of the region's stores
|
||||||
*/
|
*/
|
||||||
message RegionStoreSequenceIds {
|
message RegionStoreSequenceIds {
|
||||||
required uint64 lastFlushedSequenceId = 1;
|
required uint64 last_flushed_sequence_id = 1;
|
||||||
repeated StoreSequenceId storeSequenceId = 2;
|
repeated StoreSequenceId store_sequence_id = 2;
|
||||||
}
|
}
|
||||||
|
|
|
@ -32,7 +32,7 @@ import "Cell.proto";
|
||||||
message TableSchema {
|
message TableSchema {
|
||||||
optional bytes name = 1;
|
optional bytes name = 1;
|
||||||
repeated BytesBytesPair attributes = 2;
|
repeated BytesBytesPair attributes = 2;
|
||||||
repeated ColumnFamilySchema columnFamilies = 3;
|
repeated ColumnFamilySchema column_families = 3;
|
||||||
repeated NameStringPair configuration = 4;
|
repeated NameStringPair configuration = 4;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -50,10 +50,10 @@ message ColumnFamilySchema {
|
||||||
* Protocol buffer version of HRegionInfo.
|
* Protocol buffer version of HRegionInfo.
|
||||||
*/
|
*/
|
||||||
message RegionInfo {
|
message RegionInfo {
|
||||||
required uint64 regionId = 1;
|
required uint64 region_id = 1;
|
||||||
required bytes tableName = 2;
|
required bytes table_name = 2;
|
||||||
optional bytes startKey = 3;
|
optional bytes start_key = 3;
|
||||||
optional bytes endKey = 4;
|
optional bytes end_key = 4;
|
||||||
optional bool offline = 5;
|
optional bool offline = 5;
|
||||||
optional bool split = 6;
|
optional bool split = 6;
|
||||||
optional bool recovering = 7;
|
optional bool recovering = 7;
|
||||||
|
@ -63,7 +63,7 @@ message RegionInfo {
|
||||||
* Protocol buffer for favored nodes
|
* Protocol buffer for favored nodes
|
||||||
*/
|
*/
|
||||||
message FavoredNodes {
|
message FavoredNodes {
|
||||||
repeated ServerName favoredNode = 1;
|
repeated ServerName favored_node = 1;
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
|
@ -111,9 +111,9 @@ enum CompareType {
|
||||||
* Protocol buffer version of ServerName
|
* Protocol buffer version of ServerName
|
||||||
*/
|
*/
|
||||||
message ServerName {
|
message ServerName {
|
||||||
required string hostName = 1;
|
required string host_name = 1;
|
||||||
optional uint32 port = 2;
|
optional uint32 port = 2;
|
||||||
optional uint64 startCode = 3;
|
optional uint64 start_code = 3;
|
||||||
}
|
}
|
||||||
|
|
||||||
// Comment data structures
|
// Comment data structures
|
||||||
|
@ -148,7 +148,7 @@ message NameInt64Pair {
|
||||||
message SnapshotDescription {
|
message SnapshotDescription {
|
||||||
required string name = 1;
|
required string name = 1;
|
||||||
optional string table = 2; // not needed for delete, but checked for in taking snapshot
|
optional string table = 2; // not needed for delete, but checked for in taking snapshot
|
||||||
optional int64 creationTime = 3 [default = 0];
|
optional int64 creation_time = 3 [default = 0];
|
||||||
enum Type {
|
enum Type {
|
||||||
DISABLED = 0;
|
DISABLED = 0;
|
||||||
FLUSH = 1;
|
FLUSH = 1;
|
||||||
|
@ -161,15 +161,15 @@ message EmptyMsg {
|
||||||
}
|
}
|
||||||
|
|
||||||
message LongMsg {
|
message LongMsg {
|
||||||
required int64 longMsg = 1;
|
required int64 long_msg = 1;
|
||||||
|
|
||||||
}
|
}
|
||||||
|
|
||||||
message BigDecimalMsg {
|
message BigDecimalMsg {
|
||||||
required bytes bigdecimalMsg = 1;
|
required bytes bigdecimal_msg = 1;
|
||||||
}
|
}
|
||||||
|
|
||||||
message UUID {
|
message UUID {
|
||||||
required uint64 leastSigBits = 1;
|
required uint64 least_sig_bits = 1;
|
||||||
required uint64 mostSigBits = 2;
|
required uint64 most_sig_bits = 2;
|
||||||
}
|
}
|
||||||
|
|
|
@ -133,7 +133,7 @@ public class TokenProvider implements AuthenticationProtos.AuthenticationService
|
||||||
}
|
}
|
||||||
|
|
||||||
@Override
|
@Override
|
||||||
public void whoami(RpcController controller, AuthenticationProtos.WhoAmIRequest request,
|
public void whoAmI(RpcController controller, AuthenticationProtos.WhoAmIRequest request,
|
||||||
RpcCallback<AuthenticationProtos.WhoAmIResponse> done) {
|
RpcCallback<AuthenticationProtos.WhoAmIResponse> done) {
|
||||||
User requestUser = RequestContext.getRequestUser();
|
User requestUser = RequestContext.getRequestUser();
|
||||||
AuthenticationProtos.WhoAmIResponse.Builder response =
|
AuthenticationProtos.WhoAmIResponse.Builder response =
|
||||||
|
|
|
@ -274,15 +274,15 @@ public class TestTokenAuthentication {
|
||||||
}
|
}
|
||||||
|
|
||||||
@Override
|
@Override
|
||||||
public AuthenticationProtos.WhoAmIResponse whoami(
|
public AuthenticationProtos.WhoAmIResponse whoAmI(
|
||||||
RpcController controller, AuthenticationProtos.WhoAmIRequest request)
|
RpcController controller, AuthenticationProtos.WhoAmIRequest request)
|
||||||
throws ServiceException {
|
throws ServiceException {
|
||||||
LOG.debug("whoami() request from "+RequestContext.getRequestUserName());
|
LOG.debug("whoAmI() request from "+RequestContext.getRequestUserName());
|
||||||
// ignore passed in controller -- it's always null
|
// ignore passed in controller -- it's always null
|
||||||
ServerRpcController serverController = new ServerRpcController();
|
ServerRpcController serverController = new ServerRpcController();
|
||||||
BlockingRpcCallback<AuthenticationProtos.WhoAmIResponse> callback =
|
BlockingRpcCallback<AuthenticationProtos.WhoAmIResponse> callback =
|
||||||
new BlockingRpcCallback<AuthenticationProtos.WhoAmIResponse>();
|
new BlockingRpcCallback<AuthenticationProtos.WhoAmIResponse>();
|
||||||
whoami(serverController, request, callback);
|
whoAmI(serverController, request, callback);
|
||||||
try {
|
try {
|
||||||
serverController.checkFailed();
|
serverController.checkFailed();
|
||||||
return callback.get();
|
return callback.get();
|
||||||
|
@ -383,7 +383,7 @@ public class TestTokenAuthentication {
|
||||||
AuthenticationProtos.AuthenticationService.BlockingInterface stub =
|
AuthenticationProtos.AuthenticationService.BlockingInterface stub =
|
||||||
AuthenticationProtos.AuthenticationService.newBlockingStub(channel);
|
AuthenticationProtos.AuthenticationService.newBlockingStub(channel);
|
||||||
AuthenticationProtos.WhoAmIResponse response =
|
AuthenticationProtos.WhoAmIResponse response =
|
||||||
stub.whoami(null, AuthenticationProtos.WhoAmIRequest.getDefaultInstance());
|
stub.whoAmI(null, AuthenticationProtos.WhoAmIRequest.getDefaultInstance());
|
||||||
String myname = response.getUsername();
|
String myname = response.getUsername();
|
||||||
assertEquals("testuser", myname);
|
assertEquals("testuser", myname);
|
||||||
String authMethod = response.getAuthMethod();
|
String authMethod = response.getAuthMethod();
|
||||||
|
|
Loading…
Reference in New Issue