HBASE-18431 Mitigate compatibility concerns between branch-1.3 and branch-1.4
- Refactor TableName into its own proto module and fix up users - Move SnapshotDescription from Snapshot.proto back to HBase.proto - Restore FastLongHistogram and TestFastLongHistogram; deprecate FastLongHistogram - Move DeleteQueryTracker back to o.a.h.h.regionserver
This commit is contained in:
parent
6255dc7001
commit
3feb87b005
@ -43,9 +43,8 @@ import org.apache.hadoop.hbase.classification.InterfaceStability;
|
|||||||
import org.apache.hadoop.hbase.client.security.SecurityCapability;
|
import org.apache.hadoop.hbase.client.security.SecurityCapability;
|
||||||
import org.apache.hadoop.hbase.ipc.CoprocessorRpcChannel;
|
import org.apache.hadoop.hbase.ipc.CoprocessorRpcChannel;
|
||||||
import org.apache.hadoop.hbase.protobuf.generated.AdminProtos;
|
import org.apache.hadoop.hbase.protobuf.generated.AdminProtos;
|
||||||
import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos;
|
import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription;
|
||||||
import org.apache.hadoop.hbase.protobuf.generated.MasterProtos;
|
import org.apache.hadoop.hbase.protobuf.generated.MasterProtos;
|
||||||
import org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription;
|
|
||||||
import org.apache.hadoop.hbase.quotas.QuotaFilter;
|
import org.apache.hadoop.hbase.quotas.QuotaFilter;
|
||||||
import org.apache.hadoop.hbase.quotas.QuotaRetriever;
|
import org.apache.hadoop.hbase.quotas.QuotaRetriever;
|
||||||
import org.apache.hadoop.hbase.quotas.QuotaSettings;
|
import org.apache.hadoop.hbase.quotas.QuotaSettings;
|
||||||
|
@ -95,6 +95,7 @@ import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos;
|
|||||||
import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPair;
|
import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPair;
|
||||||
import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ProcedureDescription;
|
import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ProcedureDescription;
|
||||||
import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.RegionSpecifierType;
|
import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.RegionSpecifierType;
|
||||||
|
import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription;
|
||||||
import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema;
|
import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema;
|
||||||
import org.apache.hadoop.hbase.protobuf.generated.MasterProtos;
|
import org.apache.hadoop.hbase.protobuf.generated.MasterProtos;
|
||||||
import org.apache.hadoop.hbase.protobuf.generated.MasterProtos.AbortProcedureRequest;
|
import org.apache.hadoop.hbase.protobuf.generated.MasterProtos.AbortProcedureRequest;
|
||||||
@ -156,7 +157,7 @@ import org.apache.hadoop.hbase.protobuf.generated.MasterProtos.StopMasterRequest
|
|||||||
import org.apache.hadoop.hbase.protobuf.generated.MasterProtos.TruncateTableRequest;
|
import org.apache.hadoop.hbase.protobuf.generated.MasterProtos.TruncateTableRequest;
|
||||||
import org.apache.hadoop.hbase.protobuf.generated.MasterProtos.UnassignRegionRequest;
|
import org.apache.hadoop.hbase.protobuf.generated.MasterProtos.UnassignRegionRequest;
|
||||||
import org.apache.hadoop.hbase.protobuf.generated.ProcedureProtos;
|
import org.apache.hadoop.hbase.protobuf.generated.ProcedureProtos;
|
||||||
import org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription;
|
import org.apache.hadoop.hbase.protobuf.generated.TableProtos;
|
||||||
import org.apache.hadoop.hbase.quotas.QuotaFilter;
|
import org.apache.hadoop.hbase.quotas.QuotaFilter;
|
||||||
import org.apache.hadoop.hbase.quotas.QuotaRetriever;
|
import org.apache.hadoop.hbase.quotas.QuotaRetriever;
|
||||||
import org.apache.hadoop.hbase.quotas.QuotaSettings;
|
import org.apache.hadoop.hbase.quotas.QuotaSettings;
|
||||||
@ -3174,7 +3175,7 @@ public class HBaseAdmin implements Admin {
|
|||||||
public TableName[] call(int callTimeout) throws Exception {
|
public TableName[] call(int callTimeout) throws Exception {
|
||||||
HBaseRpcController controller = rpcControllerFactory.newController();
|
HBaseRpcController controller = rpcControllerFactory.newController();
|
||||||
controller.setCallTimeout(callTimeout);
|
controller.setCallTimeout(callTimeout);
|
||||||
List<HBaseProtos.TableName> tableNames =
|
List<TableProtos.TableName> tableNames =
|
||||||
master.listTableNamesByNamespace(controller, ListTableNamesByNamespaceRequest.
|
master.listTableNamesByNamespace(controller, ListTableNamesByNamespaceRequest.
|
||||||
newBuilder().setNamespaceName(name).build())
|
newBuilder().setNamespaceName(name).build())
|
||||||
.getTableNameList();
|
.getTableNameList();
|
||||||
|
@ -145,6 +145,7 @@ import org.apache.hadoop.hbase.protobuf.generated.MasterProtos.MasterService;
|
|||||||
import org.apache.hadoop.hbase.protobuf.generated.QuotaProtos;
|
import org.apache.hadoop.hbase.protobuf.generated.QuotaProtos;
|
||||||
import org.apache.hadoop.hbase.protobuf.generated.RegionServerStatusProtos.RegionServerReportRequest;
|
import org.apache.hadoop.hbase.protobuf.generated.RegionServerStatusProtos.RegionServerReportRequest;
|
||||||
import org.apache.hadoop.hbase.protobuf.generated.RegionServerStatusProtos.RegionServerStartupRequest;
|
import org.apache.hadoop.hbase.protobuf.generated.RegionServerStatusProtos.RegionServerStartupRequest;
|
||||||
|
import org.apache.hadoop.hbase.protobuf.generated.TableProtos;
|
||||||
import org.apache.hadoop.hbase.protobuf.generated.WALProtos;
|
import org.apache.hadoop.hbase.protobuf.generated.WALProtos;
|
||||||
import org.apache.hadoop.hbase.protobuf.generated.WALProtos.BulkLoadDescriptor;
|
import org.apache.hadoop.hbase.protobuf.generated.WALProtos.BulkLoadDescriptor;
|
||||||
import org.apache.hadoop.hbase.protobuf.generated.WALProtos.CompactionDescriptor;
|
import org.apache.hadoop.hbase.protobuf.generated.WALProtos.CompactionDescriptor;
|
||||||
@ -2975,18 +2976,18 @@ public final class ProtobufUtil {
|
|||||||
", type=" + proto.getMutateType().toString();
|
", type=" + proto.getMutateType().toString();
|
||||||
}
|
}
|
||||||
|
|
||||||
public static TableName toTableName(HBaseProtos.TableName tableNamePB) {
|
public static TableName toTableName(TableProtos.TableName tableNamePB) {
|
||||||
return TableName.valueOf(tableNamePB.getNamespace().asReadOnlyByteBuffer(),
|
return TableName.valueOf(tableNamePB.getNamespace().asReadOnlyByteBuffer(),
|
||||||
tableNamePB.getQualifier().asReadOnlyByteBuffer());
|
tableNamePB.getQualifier().asReadOnlyByteBuffer());
|
||||||
}
|
}
|
||||||
|
|
||||||
public static HBaseProtos.TableName toProtoTableName(TableName tableName) {
|
public static TableProtos.TableName toProtoTableName(TableName tableName) {
|
||||||
return HBaseProtos.TableName.newBuilder()
|
return TableProtos.TableName.newBuilder()
|
||||||
.setNamespace(ByteStringer.wrap(tableName.getNamespace()))
|
.setNamespace(ByteStringer.wrap(tableName.getNamespace()))
|
||||||
.setQualifier(ByteStringer.wrap(tableName.getQualifier())).build();
|
.setQualifier(ByteStringer.wrap(tableName.getQualifier())).build();
|
||||||
}
|
}
|
||||||
|
|
||||||
public static TableName[] getTableNameArray(List<HBaseProtos.TableName> tableNamesList) {
|
public static TableName[] getTableNameArray(List<TableProtos.TableName> tableNamesList) {
|
||||||
if (tableNamesList == null) {
|
if (tableNamesList == null) {
|
||||||
return new TableName[0];
|
return new TableName[0];
|
||||||
}
|
}
|
||||||
|
@ -21,7 +21,7 @@ package org.apache.hadoop.hbase.snapshot;
|
|||||||
|
|
||||||
import org.apache.hadoop.hbase.classification.InterfaceAudience;
|
import org.apache.hadoop.hbase.classification.InterfaceAudience;
|
||||||
import org.apache.hadoop.hbase.TableName;
|
import org.apache.hadoop.hbase.TableName;
|
||||||
import org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription;
|
import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription;
|
||||||
import org.apache.hadoop.hbase.util.Bytes;
|
import org.apache.hadoop.hbase.util.Bytes;
|
||||||
|
|
||||||
/**
|
/**
|
||||||
|
@ -19,7 +19,7 @@ package org.apache.hadoop.hbase.snapshot;
|
|||||||
|
|
||||||
import org.apache.hadoop.hbase.classification.InterfaceAudience;
|
import org.apache.hadoop.hbase.classification.InterfaceAudience;
|
||||||
import org.apache.hadoop.hbase.classification.InterfaceStability;
|
import org.apache.hadoop.hbase.classification.InterfaceStability;
|
||||||
import org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription;
|
import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription;
|
||||||
|
|
||||||
|
|
||||||
/**
|
/**
|
||||||
|
@ -20,7 +20,7 @@ package org.apache.hadoop.hbase.snapshot;
|
|||||||
import org.apache.hadoop.hbase.classification.InterfaceAudience;
|
import org.apache.hadoop.hbase.classification.InterfaceAudience;
|
||||||
import org.apache.hadoop.hbase.classification.InterfaceStability;
|
import org.apache.hadoop.hbase.classification.InterfaceStability;
|
||||||
import org.apache.hadoop.hbase.DoNotRetryIOException;
|
import org.apache.hadoop.hbase.DoNotRetryIOException;
|
||||||
import org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription;
|
import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription;
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* General exception base class for when a snapshot fails
|
* General exception base class for when a snapshot fails
|
||||||
|
@ -20,7 +20,7 @@ package org.apache.hadoop.hbase.snapshot;
|
|||||||
|
|
||||||
import org.apache.hadoop.hbase.classification.InterfaceAudience;
|
import org.apache.hadoop.hbase.classification.InterfaceAudience;
|
||||||
import org.apache.hadoop.hbase.classification.InterfaceStability;
|
import org.apache.hadoop.hbase.classification.InterfaceStability;
|
||||||
import org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription;
|
import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription;
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Thrown when a snapshot could not be restored due to a server-side error when restoring it.
|
* Thrown when a snapshot could not be restored due to a server-side error when restoring it.
|
||||||
|
@ -19,7 +19,7 @@ package org.apache.hadoop.hbase.snapshot;
|
|||||||
|
|
||||||
import org.apache.hadoop.hbase.classification.InterfaceAudience;
|
import org.apache.hadoop.hbase.classification.InterfaceAudience;
|
||||||
import org.apache.hadoop.hbase.classification.InterfaceStability;
|
import org.apache.hadoop.hbase.classification.InterfaceStability;
|
||||||
import org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription;
|
import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription;
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Thrown when a snapshot could not be created due to a server-side error when
|
* Thrown when a snapshot could not be created due to a server-side error when
|
||||||
|
@ -19,7 +19,7 @@ package org.apache.hadoop.hbase.snapshot;
|
|||||||
|
|
||||||
import org.apache.hadoop.hbase.classification.InterfaceAudience;
|
import org.apache.hadoop.hbase.classification.InterfaceAudience;
|
||||||
import org.apache.hadoop.hbase.classification.InterfaceStability;
|
import org.apache.hadoop.hbase.classification.InterfaceStability;
|
||||||
import org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription;
|
import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription;
|
||||||
|
|
||||||
|
|
||||||
/**
|
/**
|
||||||
|
@ -19,7 +19,7 @@ package org.apache.hadoop.hbase.snapshot;
|
|||||||
|
|
||||||
import org.apache.hadoop.hbase.classification.InterfaceAudience;
|
import org.apache.hadoop.hbase.classification.InterfaceAudience;
|
||||||
import org.apache.hadoop.hbase.classification.InterfaceStability;
|
import org.apache.hadoop.hbase.classification.InterfaceStability;
|
||||||
import org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription;
|
import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription;
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Thrown when a snapshot exists but should not
|
* Thrown when a snapshot exists but should not
|
||||||
|
@ -32,11 +32,11 @@ import org.apache.hadoop.hbase.HConstants;
|
|||||||
import org.apache.hadoop.hbase.TableName;
|
import org.apache.hadoop.hbase.TableName;
|
||||||
import org.apache.hadoop.hbase.ipc.HBaseRpcController;
|
import org.apache.hadoop.hbase.ipc.HBaseRpcController;
|
||||||
import org.apache.hadoop.hbase.ipc.RpcControllerFactory;
|
import org.apache.hadoop.hbase.ipc.RpcControllerFactory;
|
||||||
|
import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription;
|
||||||
import org.apache.hadoop.hbase.protobuf.generated.MasterProtos.IsSnapshotDoneRequest;
|
import org.apache.hadoop.hbase.protobuf.generated.MasterProtos.IsSnapshotDoneRequest;
|
||||||
import org.apache.hadoop.hbase.protobuf.generated.MasterProtos.IsSnapshotDoneResponse;
|
import org.apache.hadoop.hbase.protobuf.generated.MasterProtos.IsSnapshotDoneResponse;
|
||||||
import org.apache.hadoop.hbase.protobuf.generated.MasterProtos.SnapshotRequest;
|
import org.apache.hadoop.hbase.protobuf.generated.MasterProtos.SnapshotRequest;
|
||||||
import org.apache.hadoop.hbase.protobuf.generated.MasterProtos.SnapshotResponse;
|
import org.apache.hadoop.hbase.protobuf.generated.MasterProtos.SnapshotResponse;
|
||||||
import org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription;
|
|
||||||
import org.apache.hadoop.hbase.testclassification.SmallTests;
|
import org.apache.hadoop.hbase.testclassification.SmallTests;
|
||||||
import org.junit.Test;
|
import org.junit.Test;
|
||||||
import org.junit.experimental.categories.Category;
|
import org.junit.experimental.categories.Category;
|
||||||
|
@ -0,0 +1,319 @@
|
|||||||
|
/**
|
||||||
|
* Licensed to the Apache Software Foundation (ASF) under one
|
||||||
|
* or more contributor license agreements. See the NOTICE file
|
||||||
|
* distributed with this work for additional information
|
||||||
|
* regarding copyright ownership. The ASF licenses this file
|
||||||
|
* to you under the Apache License, Version 2.0 (the
|
||||||
|
* "License"); you may not use this file except in compliance
|
||||||
|
* with the License. You may obtain a copy of the License at
|
||||||
|
*
|
||||||
|
* http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
*
|
||||||
|
* Unless required by applicable law or agreed to in writing, software
|
||||||
|
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
* See the License for the specific language governing permissions and
|
||||||
|
* limitations under the License.
|
||||||
|
*/
|
||||||
|
package org.apache.hadoop.hbase.util;
|
||||||
|
|
||||||
|
import java.util.concurrent.atomic.AtomicBoolean;
|
||||||
|
import java.util.concurrent.atomic.AtomicLong;
|
||||||
|
|
||||||
|
import org.apache.hadoop.hbase.classification.InterfaceAudience;
|
||||||
|
import org.apache.hadoop.hbase.classification.InterfaceStability;
|
||||||
|
|
||||||
|
/**
|
||||||
|
* FastLongHistogram is a thread-safe class that estimate distribution of data and computes the
|
||||||
|
* quantiles.
|
||||||
|
*/
|
||||||
|
@Deprecated
|
||||||
|
@InterfaceAudience.Public
|
||||||
|
@InterfaceStability.Evolving
|
||||||
|
public class FastLongHistogram {
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Default number of bins.
|
||||||
|
*/
|
||||||
|
public static final int DEFAULT_NBINS = 255;
|
||||||
|
|
||||||
|
public static final double[] DEFAULT_QUANTILES =
|
||||||
|
new double[]{0.25, 0.5, 0.75, 0.90, 0.95, 0.98, 0.99, 0.999};
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Bins is a class containing a list of buckets(or bins) for estimation histogram of some data.
|
||||||
|
*/
|
||||||
|
private static class Bins {
|
||||||
|
private final Counter[] counts;
|
||||||
|
// inclusive
|
||||||
|
private final long binsMin;
|
||||||
|
// exclusive
|
||||||
|
private final long binsMax;
|
||||||
|
private final long bins10XMax;
|
||||||
|
private final AtomicLong min = new AtomicLong(Long.MAX_VALUE);
|
||||||
|
private final AtomicLong max = new AtomicLong(0L);
|
||||||
|
|
||||||
|
private final Counter count = new Counter(0);
|
||||||
|
private final Counter total = new Counter(0);
|
||||||
|
|
||||||
|
// set to true when any of data has been inserted to the Bins. It is set after the counts are
|
||||||
|
// updated.
|
||||||
|
private final AtomicBoolean hasData = new AtomicBoolean(false);
|
||||||
|
|
||||||
|
/**
|
||||||
|
* The constructor for creating a Bins without any prior data.
|
||||||
|
*/
|
||||||
|
public Bins(int numBins) {
|
||||||
|
counts = createCounters(numBins + 3);
|
||||||
|
this.binsMin = 1L;
|
||||||
|
|
||||||
|
// These two numbers are total guesses
|
||||||
|
// and should be treated as highly suspect.
|
||||||
|
this.binsMax = 1000;
|
||||||
|
this.bins10XMax = binsMax * 10;
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* The constructor for creating a Bins with last Bins.
|
||||||
|
*/
|
||||||
|
public Bins(Bins last, int numOfBins, double minQ, double maxQ) {
|
||||||
|
long[] values = last.getQuantiles(new double[] { minQ, maxQ });
|
||||||
|
long wd = values[1] - values[0] + 1;
|
||||||
|
// expand minQ and maxQ in two ends back assuming uniform distribution
|
||||||
|
this.binsMin = Math.max(0L, (long) (values[0] - wd * minQ));
|
||||||
|
long binsMax = (long) (values[1] + wd * (1 - maxQ)) + 1;
|
||||||
|
// make sure each of bins is at least of width 1
|
||||||
|
this.binsMax = Math.max(binsMax, this.binsMin + numOfBins);
|
||||||
|
this.bins10XMax = Math.max((long) (values[1] + (binsMax - 1) * 9), this.binsMax + 1);
|
||||||
|
|
||||||
|
this.counts = createCounters(numOfBins + 3);
|
||||||
|
}
|
||||||
|
|
||||||
|
private Counter[] createCounters(int num) {
|
||||||
|
Counter[] counters = new Counter[num];
|
||||||
|
for (int i = 0; i < num; i++) {
|
||||||
|
counters[i] = new Counter();
|
||||||
|
}
|
||||||
|
return counters;
|
||||||
|
}
|
||||||
|
|
||||||
|
private int getIndex(long value) {
|
||||||
|
if (value < this.binsMin) {
|
||||||
|
return 0;
|
||||||
|
} else if (value > this.bins10XMax) {
|
||||||
|
return this.counts.length - 1;
|
||||||
|
} else if (value >= this.binsMax) {
|
||||||
|
return this.counts.length - 2;
|
||||||
|
}
|
||||||
|
// compute the position
|
||||||
|
return 1 + (int) ((value - this.binsMin) * (this.counts.length - 3) /
|
||||||
|
(this.binsMax - this.binsMin));
|
||||||
|
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Adds a value to the histogram.
|
||||||
|
*/
|
||||||
|
public void add(long value, long count) {
|
||||||
|
if (value < 0) {
|
||||||
|
// The whole computation is completely thrown off if there are negative numbers
|
||||||
|
//
|
||||||
|
// Normally we would throw an IllegalArgumentException however this is the metrics
|
||||||
|
// system and it should be completely safe at all times.
|
||||||
|
// So silently throw it away.
|
||||||
|
return;
|
||||||
|
}
|
||||||
|
AtomicUtils.updateMin(min, value);
|
||||||
|
AtomicUtils.updateMax(max, value);
|
||||||
|
|
||||||
|
this.count.add(count);
|
||||||
|
this.total.add(value * count);
|
||||||
|
|
||||||
|
int pos = getIndex(value);
|
||||||
|
this.counts[pos].add(count);
|
||||||
|
|
||||||
|
// hasData needs to be updated as last
|
||||||
|
this.hasData.set(true);
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Computes the quantiles give the ratios.
|
||||||
|
*/
|
||||||
|
public long[] getQuantiles(double[] quantiles) {
|
||||||
|
if (!this.hasData.get()) {
|
||||||
|
// No data yet.
|
||||||
|
return new long[quantiles.length];
|
||||||
|
}
|
||||||
|
|
||||||
|
// Make a snapshot of lowerCounter, higherCounter and bins.counts to counts.
|
||||||
|
// This is not synchronized, but since the counter are accumulating, the result is a good
|
||||||
|
// estimation of a snapshot.
|
||||||
|
long[] counts = new long[this.counts.length];
|
||||||
|
long total = 0L;
|
||||||
|
for (int i = 0; i < this.counts.length; i++) {
|
||||||
|
counts[i] = this.counts[i].get();
|
||||||
|
total += counts[i];
|
||||||
|
}
|
||||||
|
|
||||||
|
int rIndex = 0;
|
||||||
|
double qCount = total * quantiles[0];
|
||||||
|
long cum = 0L;
|
||||||
|
|
||||||
|
long[] res = new long[quantiles.length];
|
||||||
|
countsLoop: for (int i = 0; i < counts.length; i++) {
|
||||||
|
// mn and mx define a value range
|
||||||
|
long mn, mx;
|
||||||
|
if (i == 0) {
|
||||||
|
mn = this.min.get();
|
||||||
|
mx = this.binsMin;
|
||||||
|
} else if (i == counts.length - 1) {
|
||||||
|
mn = this.bins10XMax;
|
||||||
|
mx = this.max.get();
|
||||||
|
} else if (i == counts.length - 2) {
|
||||||
|
mn = this.binsMax;
|
||||||
|
mx = this.bins10XMax;
|
||||||
|
} else {
|
||||||
|
mn = this.binsMin + (i - 1) * (this.binsMax - this.binsMin) / (this.counts.length - 3);
|
||||||
|
mx = this.binsMin + i * (this.binsMax - this.binsMin) / (this.counts.length - 3);
|
||||||
|
}
|
||||||
|
|
||||||
|
if (mx < this.min.get()) {
|
||||||
|
continue;
|
||||||
|
}
|
||||||
|
if (mn > this.max.get()) {
|
||||||
|
break;
|
||||||
|
}
|
||||||
|
mn = Math.max(mn, this.min.get());
|
||||||
|
mx = Math.min(mx, this.max.get());
|
||||||
|
|
||||||
|
// lastCum/cum are the corresponding counts to mn/mx
|
||||||
|
double lastCum = cum;
|
||||||
|
cum += counts[i];
|
||||||
|
|
||||||
|
// fill the results for qCount is within current range.
|
||||||
|
while (qCount <= cum) {
|
||||||
|
if (cum == lastCum) {
|
||||||
|
res[rIndex] = mn;
|
||||||
|
} else {
|
||||||
|
res[rIndex] = (long) ((qCount - lastCum) * (mx - mn) / (cum - lastCum) + mn);
|
||||||
|
}
|
||||||
|
|
||||||
|
// move to next quantile
|
||||||
|
rIndex++;
|
||||||
|
if (rIndex >= quantiles.length) {
|
||||||
|
break countsLoop;
|
||||||
|
}
|
||||||
|
qCount = total * quantiles[rIndex];
|
||||||
|
}
|
||||||
|
}
|
||||||
|
// In case quantiles contains values >= 100%
|
||||||
|
for (; rIndex < quantiles.length; rIndex++) {
|
||||||
|
res[rIndex] = this.max.get();
|
||||||
|
}
|
||||||
|
|
||||||
|
return res;
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
long getNumAtOrBelow(long val) {
|
||||||
|
final int targetIndex = getIndex(val);
|
||||||
|
long totalToCurrentIndex = 0;
|
||||||
|
for (int i = 0; i <= targetIndex; i++) {
|
||||||
|
totalToCurrentIndex += this.counts[i].get();
|
||||||
|
}
|
||||||
|
return totalToCurrentIndex;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// The bins counting values. It is replaced with a new one in calling of reset().
|
||||||
|
private volatile Bins bins;
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Constructor.
|
||||||
|
*/
|
||||||
|
public FastLongHistogram() {
|
||||||
|
this(DEFAULT_NBINS);
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Constructor.
|
||||||
|
* @param numOfBins the number of bins for the histogram. A larger value results in more precise
|
||||||
|
* results but with lower efficiency, and vice versus.
|
||||||
|
*/
|
||||||
|
public FastLongHistogram(int numOfBins) {
|
||||||
|
this.bins = new Bins(numOfBins);
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Constructor setting the bins assuming a uniform distribution within a range.
|
||||||
|
* @param numOfBins the number of bins for the histogram. A larger value results in more precise
|
||||||
|
* results but with lower efficiency, and vice versus.
|
||||||
|
* @param min lower bound of the region, inclusive.
|
||||||
|
* @param max higher bound of the region, inclusive.
|
||||||
|
*/
|
||||||
|
public FastLongHistogram(int numOfBins, long min, long max) {
|
||||||
|
this(numOfBins);
|
||||||
|
Bins bins = new Bins(numOfBins);
|
||||||
|
bins.add(min, 1);
|
||||||
|
bins.add(max, 1);
|
||||||
|
this.bins = new Bins(bins, numOfBins, 0.01, 0.999);
|
||||||
|
}
|
||||||
|
|
||||||
|
private FastLongHistogram(Bins bins) {
|
||||||
|
this.bins = bins;
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Adds a value to the histogram.
|
||||||
|
*/
|
||||||
|
public void add(long value, long count) {
|
||||||
|
this.bins.add(value, count);
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Computes the quantiles give the ratios.
|
||||||
|
*/
|
||||||
|
public long[] getQuantiles(double[] quantiles) {
|
||||||
|
return this.bins.getQuantiles(quantiles);
|
||||||
|
}
|
||||||
|
|
||||||
|
public long[] getQuantiles() {
|
||||||
|
return this.bins.getQuantiles(DEFAULT_QUANTILES);
|
||||||
|
}
|
||||||
|
|
||||||
|
public long getMin() {
|
||||||
|
long min = this.bins.min.get();
|
||||||
|
return min == Long.MAX_VALUE ? 0 : min; // in case it is not initialized
|
||||||
|
}
|
||||||
|
|
||||||
|
public long getMax() {
|
||||||
|
return this.bins.max.get();
|
||||||
|
}
|
||||||
|
|
||||||
|
public long getCount() {
|
||||||
|
return this.bins.count.get();
|
||||||
|
}
|
||||||
|
|
||||||
|
public long getMean() {
|
||||||
|
Bins bins = this.bins;
|
||||||
|
long count = bins.count.get();
|
||||||
|
long total = bins.total.get();
|
||||||
|
if (count == 0) {
|
||||||
|
return 0;
|
||||||
|
}
|
||||||
|
return total / count;
|
||||||
|
}
|
||||||
|
|
||||||
|
public long getNumAtOrBelow(long value) {
|
||||||
|
return this.bins.getNumAtOrBelow(value);
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Resets the histogram for new counting.
|
||||||
|
*/
|
||||||
|
public FastLongHistogram reset() {
|
||||||
|
Bins oldBins = this.bins;
|
||||||
|
this.bins = new Bins(this.bins, this.bins.counts.length - 3, 0.01, 0.99);
|
||||||
|
return new FastLongHistogram(oldBins);
|
||||||
|
}
|
||||||
|
}
|
@ -0,0 +1,132 @@
|
|||||||
|
/**
|
||||||
|
* Licensed to the Apache Software Foundation (ASF) under one
|
||||||
|
* or more contributor license agreements. See the NOTICE file
|
||||||
|
* distributed with this work for additional information
|
||||||
|
* regarding copyright ownership. The ASF licenses this file
|
||||||
|
* to you under the Apache License, Version 2.0 (the
|
||||||
|
* "License"); you may not use this file except in compliance
|
||||||
|
* with the License. You may obtain a copy of the License at
|
||||||
|
*
|
||||||
|
* http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
*
|
||||||
|
* Unless required by applicable law or agreed to in writing, software
|
||||||
|
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
* See the License for the specific language governing permissions and
|
||||||
|
* limitations under the License.
|
||||||
|
*/
|
||||||
|
package org.apache.hadoop.hbase.util;
|
||||||
|
|
||||||
|
import java.util.Arrays;
|
||||||
|
import java.util.Random;
|
||||||
|
|
||||||
|
|
||||||
|
import org.apache.hadoop.hbase.testclassification.SmallTests;
|
||||||
|
import org.junit.Assert;
|
||||||
|
import org.junit.Test;
|
||||||
|
import org.junit.experimental.categories.Category;
|
||||||
|
|
||||||
|
import static org.junit.Assert.assertEquals;
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Testcases for FastLongHistogram.
|
||||||
|
*/
|
||||||
|
@Category(SmallTests.class)
|
||||||
|
public class TestFastLongHistogram {
|
||||||
|
|
||||||
|
private static void doTestUniform(FastLongHistogram hist) {
|
||||||
|
long[] VALUES = { 0, 10, 20, 30, 40, 50 };
|
||||||
|
double[] qs = new double[VALUES.length];
|
||||||
|
for (int i = 0; i < qs.length; i++) {
|
||||||
|
qs[i] = (double) VALUES[i] / VALUES[VALUES.length - 1];
|
||||||
|
}
|
||||||
|
|
||||||
|
for (int i = 0; i < 10; i++) {
|
||||||
|
for (long v : VALUES) {
|
||||||
|
hist.add(v, 1);
|
||||||
|
}
|
||||||
|
long[] vals = hist.getQuantiles(qs);
|
||||||
|
System.out.println(Arrays.toString(vals));
|
||||||
|
for (int j = 0; j < qs.length; j++) {
|
||||||
|
Assert.assertTrue(j + "-th element org: " + VALUES[j] + ", act: " + vals[j],
|
||||||
|
Math.abs(vals[j] - VALUES[j]) <= 10);
|
||||||
|
}
|
||||||
|
hist.reset();
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
@Test
|
||||||
|
public void testUniform() {
|
||||||
|
FastLongHistogram hist = new FastLongHistogram(100, 0, 50);
|
||||||
|
doTestUniform(hist);
|
||||||
|
}
|
||||||
|
|
||||||
|
@Test
|
||||||
|
public void testAdaptionOfChange() {
|
||||||
|
// assumes the uniform distribution
|
||||||
|
FastLongHistogram hist = new FastLongHistogram(100, 0, 100);
|
||||||
|
|
||||||
|
Random rand = new Random();
|
||||||
|
|
||||||
|
for (int n = 0; n < 10; n++) {
|
||||||
|
for (int i = 0; i < 900; i++) {
|
||||||
|
hist.add(rand.nextInt(100), 1);
|
||||||
|
}
|
||||||
|
|
||||||
|
// add 10% outliers, this breaks the assumption, hope bin10xMax works
|
||||||
|
for (int i = 0; i < 100; i++) {
|
||||||
|
hist.add(1000 + rand.nextInt(100), 1);
|
||||||
|
}
|
||||||
|
|
||||||
|
long[] vals = hist.getQuantiles(new double[] { 0.25, 0.75, 0.95 });
|
||||||
|
System.out.println(Arrays.toString(vals));
|
||||||
|
if (n == 0) {
|
||||||
|
Assert.assertTrue("Out of possible value", vals[0] >= 0 && vals[0] <= 50);
|
||||||
|
Assert.assertTrue("Out of possible value", vals[1] >= 50 && vals[1] <= 100);
|
||||||
|
Assert.assertTrue("Out of possible value", vals[2] >= 900 && vals[2] <= 1100);
|
||||||
|
}
|
||||||
|
|
||||||
|
hist.reset();
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
@Test
|
||||||
|
public void testGetNumAtOrBelow() {
|
||||||
|
long[] VALUES = { 1, 10, 20, 30, 40, 50 };
|
||||||
|
|
||||||
|
FastLongHistogram h = new FastLongHistogram();
|
||||||
|
for (long v : VALUES) {
|
||||||
|
for (int i = 0; i < 100; i++) {
|
||||||
|
h.add(v, 1);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
h.add(Integer.MAX_VALUE, 1);
|
||||||
|
|
||||||
|
h.reset();
|
||||||
|
|
||||||
|
for (long v : VALUES) {
|
||||||
|
for (int i = 0; i < 100; i++) {
|
||||||
|
h.add(v, 1);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
// Add something way out there to make sure it doesn't throw off the counts.
|
||||||
|
h.add(Integer.MAX_VALUE, 1);
|
||||||
|
|
||||||
|
assertEquals(100, h.getNumAtOrBelow(1));
|
||||||
|
assertEquals(200, h.getNumAtOrBelow(11));
|
||||||
|
assertEquals(601, h.getNumAtOrBelow(Long.MAX_VALUE));
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
@Test
|
||||||
|
public void testSameValues() {
|
||||||
|
FastLongHistogram hist = new FastLongHistogram(100);
|
||||||
|
|
||||||
|
hist.add(50, 100);
|
||||||
|
|
||||||
|
hist.reset();
|
||||||
|
doTestUniform(hist);
|
||||||
|
}
|
||||||
|
}
|
@ -197,6 +197,7 @@
|
|||||||
<include>RowProcessor.proto</include>
|
<include>RowProcessor.proto</include>
|
||||||
<include>SecureBulkLoad.proto</include>
|
<include>SecureBulkLoad.proto</include>
|
||||||
<include>Snapshot.proto</include>
|
<include>Snapshot.proto</include>
|
||||||
|
<include>Table.proto</include>
|
||||||
<include>Tracing.proto</include>
|
<include>Tracing.proto</include>
|
||||||
<include>VisibilityLabels.proto</include>
|
<include>VisibilityLabels.proto</include>
|
||||||
<include>WAL.proto</include>
|
<include>WAL.proto</include>
|
||||||
|
@ -1285,11 +1285,11 @@ public final class AccessControlProtos {
|
|||||||
/**
|
/**
|
||||||
* <code>optional .hbase.pb.TableName table_name = 1;</code>
|
* <code>optional .hbase.pb.TableName table_name = 1;</code>
|
||||||
*/
|
*/
|
||||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName getTableName();
|
org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName getTableName();
|
||||||
/**
|
/**
|
||||||
* <code>optional .hbase.pb.TableName table_name = 1;</code>
|
* <code>optional .hbase.pb.TableName table_name = 1;</code>
|
||||||
*/
|
*/
|
||||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder();
|
org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder getTableNameOrBuilder();
|
||||||
|
|
||||||
// optional bytes family = 2;
|
// optional bytes family = 2;
|
||||||
/**
|
/**
|
||||||
@ -1377,11 +1377,11 @@ public final class AccessControlProtos {
|
|||||||
break;
|
break;
|
||||||
}
|
}
|
||||||
case 10: {
|
case 10: {
|
||||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder subBuilder = null;
|
org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder subBuilder = null;
|
||||||
if (((bitField0_ & 0x00000001) == 0x00000001)) {
|
if (((bitField0_ & 0x00000001) == 0x00000001)) {
|
||||||
subBuilder = tableName_.toBuilder();
|
subBuilder = tableName_.toBuilder();
|
||||||
}
|
}
|
||||||
tableName_ = input.readMessage(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.PARSER, extensionRegistry);
|
tableName_ = input.readMessage(org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.PARSER, extensionRegistry);
|
||||||
if (subBuilder != null) {
|
if (subBuilder != null) {
|
||||||
subBuilder.mergeFrom(tableName_);
|
subBuilder.mergeFrom(tableName_);
|
||||||
tableName_ = subBuilder.buildPartial();
|
tableName_ = subBuilder.buildPartial();
|
||||||
@ -1477,7 +1477,7 @@ public final class AccessControlProtos {
|
|||||||
private int bitField0_;
|
private int bitField0_;
|
||||||
// optional .hbase.pb.TableName table_name = 1;
|
// optional .hbase.pb.TableName table_name = 1;
|
||||||
public static final int TABLE_NAME_FIELD_NUMBER = 1;
|
public static final int TABLE_NAME_FIELD_NUMBER = 1;
|
||||||
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName tableName_;
|
private org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName tableName_;
|
||||||
/**
|
/**
|
||||||
* <code>optional .hbase.pb.TableName table_name = 1;</code>
|
* <code>optional .hbase.pb.TableName table_name = 1;</code>
|
||||||
*/
|
*/
|
||||||
@ -1487,13 +1487,13 @@ public final class AccessControlProtos {
|
|||||||
/**
|
/**
|
||||||
* <code>optional .hbase.pb.TableName table_name = 1;</code>
|
* <code>optional .hbase.pb.TableName table_name = 1;</code>
|
||||||
*/
|
*/
|
||||||
public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName getTableName() {
|
public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName getTableName() {
|
||||||
return tableName_;
|
return tableName_;
|
||||||
}
|
}
|
||||||
/**
|
/**
|
||||||
* <code>optional .hbase.pb.TableName table_name = 1;</code>
|
* <code>optional .hbase.pb.TableName table_name = 1;</code>
|
||||||
*/
|
*/
|
||||||
public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder() {
|
public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder getTableNameOrBuilder() {
|
||||||
return tableName_;
|
return tableName_;
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -1552,7 +1552,7 @@ public final class AccessControlProtos {
|
|||||||
}
|
}
|
||||||
|
|
||||||
private void initFields() {
|
private void initFields() {
|
||||||
tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance();
|
tableName_ = org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance();
|
||||||
family_ = com.google.protobuf.ByteString.EMPTY;
|
family_ = com.google.protobuf.ByteString.EMPTY;
|
||||||
qualifier_ = com.google.protobuf.ByteString.EMPTY;
|
qualifier_ = com.google.protobuf.ByteString.EMPTY;
|
||||||
action_ = java.util.Collections.emptyList();
|
action_ = java.util.Collections.emptyList();
|
||||||
@ -1797,7 +1797,7 @@ public final class AccessControlProtos {
|
|||||||
public Builder clear() {
|
public Builder clear() {
|
||||||
super.clear();
|
super.clear();
|
||||||
if (tableNameBuilder_ == null) {
|
if (tableNameBuilder_ == null) {
|
||||||
tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance();
|
tableName_ = org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance();
|
||||||
} else {
|
} else {
|
||||||
tableNameBuilder_.clear();
|
tableNameBuilder_.clear();
|
||||||
}
|
}
|
||||||
@ -1926,9 +1926,9 @@ public final class AccessControlProtos {
|
|||||||
private int bitField0_;
|
private int bitField0_;
|
||||||
|
|
||||||
// optional .hbase.pb.TableName table_name = 1;
|
// optional .hbase.pb.TableName table_name = 1;
|
||||||
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance();
|
private org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName tableName_ = org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance();
|
||||||
private com.google.protobuf.SingleFieldBuilder<
|
private com.google.protobuf.SingleFieldBuilder<
|
||||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder> tableNameBuilder_;
|
org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder> tableNameBuilder_;
|
||||||
/**
|
/**
|
||||||
* <code>optional .hbase.pb.TableName table_name = 1;</code>
|
* <code>optional .hbase.pb.TableName table_name = 1;</code>
|
||||||
*/
|
*/
|
||||||
@ -1938,7 +1938,7 @@ public final class AccessControlProtos {
|
|||||||
/**
|
/**
|
||||||
* <code>optional .hbase.pb.TableName table_name = 1;</code>
|
* <code>optional .hbase.pb.TableName table_name = 1;</code>
|
||||||
*/
|
*/
|
||||||
public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName getTableName() {
|
public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName getTableName() {
|
||||||
if (tableNameBuilder_ == null) {
|
if (tableNameBuilder_ == null) {
|
||||||
return tableName_;
|
return tableName_;
|
||||||
} else {
|
} else {
|
||||||
@ -1948,7 +1948,7 @@ public final class AccessControlProtos {
|
|||||||
/**
|
/**
|
||||||
* <code>optional .hbase.pb.TableName table_name = 1;</code>
|
* <code>optional .hbase.pb.TableName table_name = 1;</code>
|
||||||
*/
|
*/
|
||||||
public Builder setTableName(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName value) {
|
public Builder setTableName(org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName value) {
|
||||||
if (tableNameBuilder_ == null) {
|
if (tableNameBuilder_ == null) {
|
||||||
if (value == null) {
|
if (value == null) {
|
||||||
throw new NullPointerException();
|
throw new NullPointerException();
|
||||||
@ -1965,7 +1965,7 @@ public final class AccessControlProtos {
|
|||||||
* <code>optional .hbase.pb.TableName table_name = 1;</code>
|
* <code>optional .hbase.pb.TableName table_name = 1;</code>
|
||||||
*/
|
*/
|
||||||
public Builder setTableName(
|
public Builder setTableName(
|
||||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder builderForValue) {
|
org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder builderForValue) {
|
||||||
if (tableNameBuilder_ == null) {
|
if (tableNameBuilder_ == null) {
|
||||||
tableName_ = builderForValue.build();
|
tableName_ = builderForValue.build();
|
||||||
onChanged();
|
onChanged();
|
||||||
@ -1978,12 +1978,12 @@ public final class AccessControlProtos {
|
|||||||
/**
|
/**
|
||||||
* <code>optional .hbase.pb.TableName table_name = 1;</code>
|
* <code>optional .hbase.pb.TableName table_name = 1;</code>
|
||||||
*/
|
*/
|
||||||
public Builder mergeTableName(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName value) {
|
public Builder mergeTableName(org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName value) {
|
||||||
if (tableNameBuilder_ == null) {
|
if (tableNameBuilder_ == null) {
|
||||||
if (((bitField0_ & 0x00000001) == 0x00000001) &&
|
if (((bitField0_ & 0x00000001) == 0x00000001) &&
|
||||||
tableName_ != org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance()) {
|
tableName_ != org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance()) {
|
||||||
tableName_ =
|
tableName_ =
|
||||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.newBuilder(tableName_).mergeFrom(value).buildPartial();
|
org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.newBuilder(tableName_).mergeFrom(value).buildPartial();
|
||||||
} else {
|
} else {
|
||||||
tableName_ = value;
|
tableName_ = value;
|
||||||
}
|
}
|
||||||
@ -1999,7 +1999,7 @@ public final class AccessControlProtos {
|
|||||||
*/
|
*/
|
||||||
public Builder clearTableName() {
|
public Builder clearTableName() {
|
||||||
if (tableNameBuilder_ == null) {
|
if (tableNameBuilder_ == null) {
|
||||||
tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance();
|
tableName_ = org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance();
|
||||||
onChanged();
|
onChanged();
|
||||||
} else {
|
} else {
|
||||||
tableNameBuilder_.clear();
|
tableNameBuilder_.clear();
|
||||||
@ -2010,7 +2010,7 @@ public final class AccessControlProtos {
|
|||||||
/**
|
/**
|
||||||
* <code>optional .hbase.pb.TableName table_name = 1;</code>
|
* <code>optional .hbase.pb.TableName table_name = 1;</code>
|
||||||
*/
|
*/
|
||||||
public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder getTableNameBuilder() {
|
public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder getTableNameBuilder() {
|
||||||
bitField0_ |= 0x00000001;
|
bitField0_ |= 0x00000001;
|
||||||
onChanged();
|
onChanged();
|
||||||
return getTableNameFieldBuilder().getBuilder();
|
return getTableNameFieldBuilder().getBuilder();
|
||||||
@ -2018,7 +2018,7 @@ public final class AccessControlProtos {
|
|||||||
/**
|
/**
|
||||||
* <code>optional .hbase.pb.TableName table_name = 1;</code>
|
* <code>optional .hbase.pb.TableName table_name = 1;</code>
|
||||||
*/
|
*/
|
||||||
public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder() {
|
public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder getTableNameOrBuilder() {
|
||||||
if (tableNameBuilder_ != null) {
|
if (tableNameBuilder_ != null) {
|
||||||
return tableNameBuilder_.getMessageOrBuilder();
|
return tableNameBuilder_.getMessageOrBuilder();
|
||||||
} else {
|
} else {
|
||||||
@ -2029,11 +2029,11 @@ public final class AccessControlProtos {
|
|||||||
* <code>optional .hbase.pb.TableName table_name = 1;</code>
|
* <code>optional .hbase.pb.TableName table_name = 1;</code>
|
||||||
*/
|
*/
|
||||||
private com.google.protobuf.SingleFieldBuilder<
|
private com.google.protobuf.SingleFieldBuilder<
|
||||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder>
|
org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder>
|
||||||
getTableNameFieldBuilder() {
|
getTableNameFieldBuilder() {
|
||||||
if (tableNameBuilder_ == null) {
|
if (tableNameBuilder_ == null) {
|
||||||
tableNameBuilder_ = new com.google.protobuf.SingleFieldBuilder<
|
tableNameBuilder_ = new com.google.protobuf.SingleFieldBuilder<
|
||||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder>(
|
org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder>(
|
||||||
tableName_,
|
tableName_,
|
||||||
getParentForChildren(),
|
getParentForChildren(),
|
||||||
isClean());
|
isClean());
|
||||||
@ -7462,11 +7462,11 @@ public final class AccessControlProtos {
|
|||||||
/**
|
/**
|
||||||
* <code>optional .hbase.pb.TableName table_name = 2;</code>
|
* <code>optional .hbase.pb.TableName table_name = 2;</code>
|
||||||
*/
|
*/
|
||||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName getTableName();
|
org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName getTableName();
|
||||||
/**
|
/**
|
||||||
* <code>optional .hbase.pb.TableName table_name = 2;</code>
|
* <code>optional .hbase.pb.TableName table_name = 2;</code>
|
||||||
*/
|
*/
|
||||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder();
|
org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder getTableNameOrBuilder();
|
||||||
|
|
||||||
// optional bytes namespace_name = 3;
|
// optional bytes namespace_name = 3;
|
||||||
/**
|
/**
|
||||||
@ -7541,11 +7541,11 @@ public final class AccessControlProtos {
|
|||||||
break;
|
break;
|
||||||
}
|
}
|
||||||
case 18: {
|
case 18: {
|
||||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder subBuilder = null;
|
org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder subBuilder = null;
|
||||||
if (((bitField0_ & 0x00000002) == 0x00000002)) {
|
if (((bitField0_ & 0x00000002) == 0x00000002)) {
|
||||||
subBuilder = tableName_.toBuilder();
|
subBuilder = tableName_.toBuilder();
|
||||||
}
|
}
|
||||||
tableName_ = input.readMessage(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.PARSER, extensionRegistry);
|
tableName_ = input.readMessage(org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.PARSER, extensionRegistry);
|
||||||
if (subBuilder != null) {
|
if (subBuilder != null) {
|
||||||
subBuilder.mergeFrom(tableName_);
|
subBuilder.mergeFrom(tableName_);
|
||||||
tableName_ = subBuilder.buildPartial();
|
tableName_ = subBuilder.buildPartial();
|
||||||
@ -7616,7 +7616,7 @@ public final class AccessControlProtos {
|
|||||||
|
|
||||||
// optional .hbase.pb.TableName table_name = 2;
|
// optional .hbase.pb.TableName table_name = 2;
|
||||||
public static final int TABLE_NAME_FIELD_NUMBER = 2;
|
public static final int TABLE_NAME_FIELD_NUMBER = 2;
|
||||||
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName tableName_;
|
private org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName tableName_;
|
||||||
/**
|
/**
|
||||||
* <code>optional .hbase.pb.TableName table_name = 2;</code>
|
* <code>optional .hbase.pb.TableName table_name = 2;</code>
|
||||||
*/
|
*/
|
||||||
@ -7626,13 +7626,13 @@ public final class AccessControlProtos {
|
|||||||
/**
|
/**
|
||||||
* <code>optional .hbase.pb.TableName table_name = 2;</code>
|
* <code>optional .hbase.pb.TableName table_name = 2;</code>
|
||||||
*/
|
*/
|
||||||
public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName getTableName() {
|
public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName getTableName() {
|
||||||
return tableName_;
|
return tableName_;
|
||||||
}
|
}
|
||||||
/**
|
/**
|
||||||
* <code>optional .hbase.pb.TableName table_name = 2;</code>
|
* <code>optional .hbase.pb.TableName table_name = 2;</code>
|
||||||
*/
|
*/
|
||||||
public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder() {
|
public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder getTableNameOrBuilder() {
|
||||||
return tableName_;
|
return tableName_;
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -7654,7 +7654,7 @@ public final class AccessControlProtos {
|
|||||||
|
|
||||||
private void initFields() {
|
private void initFields() {
|
||||||
type_ = org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.Permission.Type.Global;
|
type_ = org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.Permission.Type.Global;
|
||||||
tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance();
|
tableName_ = org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance();
|
||||||
namespaceName_ = com.google.protobuf.ByteString.EMPTY;
|
namespaceName_ = com.google.protobuf.ByteString.EMPTY;
|
||||||
}
|
}
|
||||||
private byte memoizedIsInitialized = -1;
|
private byte memoizedIsInitialized = -1;
|
||||||
@ -7881,7 +7881,7 @@ public final class AccessControlProtos {
|
|||||||
type_ = org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.Permission.Type.Global;
|
type_ = org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.Permission.Type.Global;
|
||||||
bitField0_ = (bitField0_ & ~0x00000001);
|
bitField0_ = (bitField0_ & ~0x00000001);
|
||||||
if (tableNameBuilder_ == null) {
|
if (tableNameBuilder_ == null) {
|
||||||
tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance();
|
tableName_ = org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance();
|
||||||
} else {
|
} else {
|
||||||
tableNameBuilder_.clear();
|
tableNameBuilder_.clear();
|
||||||
}
|
}
|
||||||
@ -8027,9 +8027,9 @@ public final class AccessControlProtos {
|
|||||||
}
|
}
|
||||||
|
|
||||||
// optional .hbase.pb.TableName table_name = 2;
|
// optional .hbase.pb.TableName table_name = 2;
|
||||||
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance();
|
private org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName tableName_ = org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance();
|
||||||
private com.google.protobuf.SingleFieldBuilder<
|
private com.google.protobuf.SingleFieldBuilder<
|
||||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder> tableNameBuilder_;
|
org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder> tableNameBuilder_;
|
||||||
/**
|
/**
|
||||||
* <code>optional .hbase.pb.TableName table_name = 2;</code>
|
* <code>optional .hbase.pb.TableName table_name = 2;</code>
|
||||||
*/
|
*/
|
||||||
@ -8039,7 +8039,7 @@ public final class AccessControlProtos {
|
|||||||
/**
|
/**
|
||||||
* <code>optional .hbase.pb.TableName table_name = 2;</code>
|
* <code>optional .hbase.pb.TableName table_name = 2;</code>
|
||||||
*/
|
*/
|
||||||
public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName getTableName() {
|
public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName getTableName() {
|
||||||
if (tableNameBuilder_ == null) {
|
if (tableNameBuilder_ == null) {
|
||||||
return tableName_;
|
return tableName_;
|
||||||
} else {
|
} else {
|
||||||
@ -8049,7 +8049,7 @@ public final class AccessControlProtos {
|
|||||||
/**
|
/**
|
||||||
* <code>optional .hbase.pb.TableName table_name = 2;</code>
|
* <code>optional .hbase.pb.TableName table_name = 2;</code>
|
||||||
*/
|
*/
|
||||||
public Builder setTableName(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName value) {
|
public Builder setTableName(org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName value) {
|
||||||
if (tableNameBuilder_ == null) {
|
if (tableNameBuilder_ == null) {
|
||||||
if (value == null) {
|
if (value == null) {
|
||||||
throw new NullPointerException();
|
throw new NullPointerException();
|
||||||
@ -8066,7 +8066,7 @@ public final class AccessControlProtos {
|
|||||||
* <code>optional .hbase.pb.TableName table_name = 2;</code>
|
* <code>optional .hbase.pb.TableName table_name = 2;</code>
|
||||||
*/
|
*/
|
||||||
public Builder setTableName(
|
public Builder setTableName(
|
||||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder builderForValue) {
|
org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder builderForValue) {
|
||||||
if (tableNameBuilder_ == null) {
|
if (tableNameBuilder_ == null) {
|
||||||
tableName_ = builderForValue.build();
|
tableName_ = builderForValue.build();
|
||||||
onChanged();
|
onChanged();
|
||||||
@ -8079,12 +8079,12 @@ public final class AccessControlProtos {
|
|||||||
/**
|
/**
|
||||||
* <code>optional .hbase.pb.TableName table_name = 2;</code>
|
* <code>optional .hbase.pb.TableName table_name = 2;</code>
|
||||||
*/
|
*/
|
||||||
public Builder mergeTableName(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName value) {
|
public Builder mergeTableName(org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName value) {
|
||||||
if (tableNameBuilder_ == null) {
|
if (tableNameBuilder_ == null) {
|
||||||
if (((bitField0_ & 0x00000002) == 0x00000002) &&
|
if (((bitField0_ & 0x00000002) == 0x00000002) &&
|
||||||
tableName_ != org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance()) {
|
tableName_ != org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance()) {
|
||||||
tableName_ =
|
tableName_ =
|
||||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.newBuilder(tableName_).mergeFrom(value).buildPartial();
|
org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.newBuilder(tableName_).mergeFrom(value).buildPartial();
|
||||||
} else {
|
} else {
|
||||||
tableName_ = value;
|
tableName_ = value;
|
||||||
}
|
}
|
||||||
@ -8100,7 +8100,7 @@ public final class AccessControlProtos {
|
|||||||
*/
|
*/
|
||||||
public Builder clearTableName() {
|
public Builder clearTableName() {
|
||||||
if (tableNameBuilder_ == null) {
|
if (tableNameBuilder_ == null) {
|
||||||
tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance();
|
tableName_ = org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance();
|
||||||
onChanged();
|
onChanged();
|
||||||
} else {
|
} else {
|
||||||
tableNameBuilder_.clear();
|
tableNameBuilder_.clear();
|
||||||
@ -8111,7 +8111,7 @@ public final class AccessControlProtos {
|
|||||||
/**
|
/**
|
||||||
* <code>optional .hbase.pb.TableName table_name = 2;</code>
|
* <code>optional .hbase.pb.TableName table_name = 2;</code>
|
||||||
*/
|
*/
|
||||||
public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder getTableNameBuilder() {
|
public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder getTableNameBuilder() {
|
||||||
bitField0_ |= 0x00000002;
|
bitField0_ |= 0x00000002;
|
||||||
onChanged();
|
onChanged();
|
||||||
return getTableNameFieldBuilder().getBuilder();
|
return getTableNameFieldBuilder().getBuilder();
|
||||||
@ -8119,7 +8119,7 @@ public final class AccessControlProtos {
|
|||||||
/**
|
/**
|
||||||
* <code>optional .hbase.pb.TableName table_name = 2;</code>
|
* <code>optional .hbase.pb.TableName table_name = 2;</code>
|
||||||
*/
|
*/
|
||||||
public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder() {
|
public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder getTableNameOrBuilder() {
|
||||||
if (tableNameBuilder_ != null) {
|
if (tableNameBuilder_ != null) {
|
||||||
return tableNameBuilder_.getMessageOrBuilder();
|
return tableNameBuilder_.getMessageOrBuilder();
|
||||||
} else {
|
} else {
|
||||||
@ -8130,11 +8130,11 @@ public final class AccessControlProtos {
|
|||||||
* <code>optional .hbase.pb.TableName table_name = 2;</code>
|
* <code>optional .hbase.pb.TableName table_name = 2;</code>
|
||||||
*/
|
*/
|
||||||
private com.google.protobuf.SingleFieldBuilder<
|
private com.google.protobuf.SingleFieldBuilder<
|
||||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder>
|
org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder>
|
||||||
getTableNameFieldBuilder() {
|
getTableNameFieldBuilder() {
|
||||||
if (tableNameBuilder_ == null) {
|
if (tableNameBuilder_ == null) {
|
||||||
tableNameBuilder_ = new com.google.protobuf.SingleFieldBuilder<
|
tableNameBuilder_ = new com.google.protobuf.SingleFieldBuilder<
|
||||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder>(
|
org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder>(
|
||||||
tableName_,
|
tableName_,
|
||||||
getParentForChildren(),
|
getParentForChildren(),
|
||||||
isClean());
|
isClean());
|
||||||
@ -10499,7 +10499,7 @@ public final class AccessControlProtos {
|
|||||||
descriptor;
|
descriptor;
|
||||||
static {
|
static {
|
||||||
java.lang.String[] descriptorData = {
|
java.lang.String[] descriptorData = {
|
||||||
"\n\023AccessControl.proto\022\010hbase.pb\032\013HBase.p" +
|
"\n\023AccessControl.proto\022\010hbase.pb\032\013Table.p" +
|
||||||
"roto\"\314\002\n\nPermission\022\'\n\004type\030\001 \002(\0162\031.hbas" +
|
"roto\"\314\002\n\nPermission\022\'\n\004type\030\001 \002(\0162\031.hbas" +
|
||||||
"e.pb.Permission.Type\0225\n\021global_permissio" +
|
"e.pb.Permission.Type\0225\n\021global_permissio" +
|
||||||
"n\030\002 \001(\0132\032.hbase.pb.GlobalPermission\022;\n\024n" +
|
"n\030\002 \001(\0132\032.hbase.pb.GlobalPermission\022;\n\024n" +
|
||||||
@ -10648,7 +10648,7 @@ public final class AccessControlProtos {
|
|||||||
com.google.protobuf.Descriptors.FileDescriptor
|
com.google.protobuf.Descriptors.FileDescriptor
|
||||||
.internalBuildGeneratedFileFrom(descriptorData,
|
.internalBuildGeneratedFileFrom(descriptorData,
|
||||||
new com.google.protobuf.Descriptors.FileDescriptor[] {
|
new com.google.protobuf.Descriptors.FileDescriptor[] {
|
||||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.getDescriptor(),
|
org.apache.hadoop.hbase.protobuf.generated.TableProtos.getDescriptor(),
|
||||||
}, assigner);
|
}, assigner);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -22646,7 +22646,7 @@ public final class ClientProtos {
|
|||||||
* </pre>
|
* </pre>
|
||||||
*/
|
*/
|
||||||
private com.google.protobuf.SingleFieldBuilder<
|
private com.google.protobuf.SingleFieldBuilder<
|
||||||
org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Cursor, org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Cursor.Builder, org.apache.hadoop.hbase.protobuf.generated.ClientProtos.CursorOrBuilder>
|
org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Cursor, org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Cursor.Builder, org.apache.hadoop.hbase.protobuf.generated.ClientProtos.CursorOrBuilder>
|
||||||
getCursorFieldBuilder() {
|
getCursorFieldBuilder() {
|
||||||
if (cursorBuilder_ == null) {
|
if (cursorBuilder_ == null) {
|
||||||
cursorBuilder_ = new com.google.protobuf.SingleFieldBuilder<
|
cursorBuilder_ = new com.google.protobuf.SingleFieldBuilder<
|
||||||
|
File diff suppressed because it is too large
Load Diff
File diff suppressed because it is too large
Load Diff
File diff suppressed because it is too large
Load Diff
@ -2517,11 +2517,11 @@ public final class SecureBulkLoadProtos {
|
|||||||
/**
|
/**
|
||||||
* <code>required .hbase.pb.TableName table_name = 1;</code>
|
* <code>required .hbase.pb.TableName table_name = 1;</code>
|
||||||
*/
|
*/
|
||||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName getTableName();
|
org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName getTableName();
|
||||||
/**
|
/**
|
||||||
* <code>required .hbase.pb.TableName table_name = 1;</code>
|
* <code>required .hbase.pb.TableName table_name = 1;</code>
|
||||||
*/
|
*/
|
||||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder();
|
org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder getTableNameOrBuilder();
|
||||||
}
|
}
|
||||||
/**
|
/**
|
||||||
* Protobuf type {@code hbase.pb.PrepareBulkLoadRequest}
|
* Protobuf type {@code hbase.pb.PrepareBulkLoadRequest}
|
||||||
@ -2575,11 +2575,11 @@ public final class SecureBulkLoadProtos {
|
|||||||
break;
|
break;
|
||||||
}
|
}
|
||||||
case 10: {
|
case 10: {
|
||||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder subBuilder = null;
|
org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder subBuilder = null;
|
||||||
if (((bitField0_ & 0x00000001) == 0x00000001)) {
|
if (((bitField0_ & 0x00000001) == 0x00000001)) {
|
||||||
subBuilder = tableName_.toBuilder();
|
subBuilder = tableName_.toBuilder();
|
||||||
}
|
}
|
||||||
tableName_ = input.readMessage(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.PARSER, extensionRegistry);
|
tableName_ = input.readMessage(org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.PARSER, extensionRegistry);
|
||||||
if (subBuilder != null) {
|
if (subBuilder != null) {
|
||||||
subBuilder.mergeFrom(tableName_);
|
subBuilder.mergeFrom(tableName_);
|
||||||
tableName_ = subBuilder.buildPartial();
|
tableName_ = subBuilder.buildPartial();
|
||||||
@ -2629,7 +2629,7 @@ public final class SecureBulkLoadProtos {
|
|||||||
private int bitField0_;
|
private int bitField0_;
|
||||||
// required .hbase.pb.TableName table_name = 1;
|
// required .hbase.pb.TableName table_name = 1;
|
||||||
public static final int TABLE_NAME_FIELD_NUMBER = 1;
|
public static final int TABLE_NAME_FIELD_NUMBER = 1;
|
||||||
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName tableName_;
|
private org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName tableName_;
|
||||||
/**
|
/**
|
||||||
* <code>required .hbase.pb.TableName table_name = 1;</code>
|
* <code>required .hbase.pb.TableName table_name = 1;</code>
|
||||||
*/
|
*/
|
||||||
@ -2639,18 +2639,18 @@ public final class SecureBulkLoadProtos {
|
|||||||
/**
|
/**
|
||||||
* <code>required .hbase.pb.TableName table_name = 1;</code>
|
* <code>required .hbase.pb.TableName table_name = 1;</code>
|
||||||
*/
|
*/
|
||||||
public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName getTableName() {
|
public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName getTableName() {
|
||||||
return tableName_;
|
return tableName_;
|
||||||
}
|
}
|
||||||
/**
|
/**
|
||||||
* <code>required .hbase.pb.TableName table_name = 1;</code>
|
* <code>required .hbase.pb.TableName table_name = 1;</code>
|
||||||
*/
|
*/
|
||||||
public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder() {
|
public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder getTableNameOrBuilder() {
|
||||||
return tableName_;
|
return tableName_;
|
||||||
}
|
}
|
||||||
|
|
||||||
private void initFields() {
|
private void initFields() {
|
||||||
tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance();
|
tableName_ = org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance();
|
||||||
}
|
}
|
||||||
private byte memoizedIsInitialized = -1;
|
private byte memoizedIsInitialized = -1;
|
||||||
public final boolean isInitialized() {
|
public final boolean isInitialized() {
|
||||||
@ -2844,7 +2844,7 @@ public final class SecureBulkLoadProtos {
|
|||||||
public Builder clear() {
|
public Builder clear() {
|
||||||
super.clear();
|
super.clear();
|
||||||
if (tableNameBuilder_ == null) {
|
if (tableNameBuilder_ == null) {
|
||||||
tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance();
|
tableName_ = org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance();
|
||||||
} else {
|
} else {
|
||||||
tableNameBuilder_.clear();
|
tableNameBuilder_.clear();
|
||||||
}
|
}
|
||||||
@ -2940,9 +2940,9 @@ public final class SecureBulkLoadProtos {
|
|||||||
private int bitField0_;
|
private int bitField0_;
|
||||||
|
|
||||||
// required .hbase.pb.TableName table_name = 1;
|
// required .hbase.pb.TableName table_name = 1;
|
||||||
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance();
|
private org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName tableName_ = org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance();
|
||||||
private com.google.protobuf.SingleFieldBuilder<
|
private com.google.protobuf.SingleFieldBuilder<
|
||||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder> tableNameBuilder_;
|
org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder> tableNameBuilder_;
|
||||||
/**
|
/**
|
||||||
* <code>required .hbase.pb.TableName table_name = 1;</code>
|
* <code>required .hbase.pb.TableName table_name = 1;</code>
|
||||||
*/
|
*/
|
||||||
@ -2952,7 +2952,7 @@ public final class SecureBulkLoadProtos {
|
|||||||
/**
|
/**
|
||||||
* <code>required .hbase.pb.TableName table_name = 1;</code>
|
* <code>required .hbase.pb.TableName table_name = 1;</code>
|
||||||
*/
|
*/
|
||||||
public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName getTableName() {
|
public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName getTableName() {
|
||||||
if (tableNameBuilder_ == null) {
|
if (tableNameBuilder_ == null) {
|
||||||
return tableName_;
|
return tableName_;
|
||||||
} else {
|
} else {
|
||||||
@ -2962,7 +2962,7 @@ public final class SecureBulkLoadProtos {
|
|||||||
/**
|
/**
|
||||||
* <code>required .hbase.pb.TableName table_name = 1;</code>
|
* <code>required .hbase.pb.TableName table_name = 1;</code>
|
||||||
*/
|
*/
|
||||||
public Builder setTableName(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName value) {
|
public Builder setTableName(org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName value) {
|
||||||
if (tableNameBuilder_ == null) {
|
if (tableNameBuilder_ == null) {
|
||||||
if (value == null) {
|
if (value == null) {
|
||||||
throw new NullPointerException();
|
throw new NullPointerException();
|
||||||
@ -2979,7 +2979,7 @@ public final class SecureBulkLoadProtos {
|
|||||||
* <code>required .hbase.pb.TableName table_name = 1;</code>
|
* <code>required .hbase.pb.TableName table_name = 1;</code>
|
||||||
*/
|
*/
|
||||||
public Builder setTableName(
|
public Builder setTableName(
|
||||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder builderForValue) {
|
org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder builderForValue) {
|
||||||
if (tableNameBuilder_ == null) {
|
if (tableNameBuilder_ == null) {
|
||||||
tableName_ = builderForValue.build();
|
tableName_ = builderForValue.build();
|
||||||
onChanged();
|
onChanged();
|
||||||
@ -2992,12 +2992,12 @@ public final class SecureBulkLoadProtos {
|
|||||||
/**
|
/**
|
||||||
* <code>required .hbase.pb.TableName table_name = 1;</code>
|
* <code>required .hbase.pb.TableName table_name = 1;</code>
|
||||||
*/
|
*/
|
||||||
public Builder mergeTableName(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName value) {
|
public Builder mergeTableName(org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName value) {
|
||||||
if (tableNameBuilder_ == null) {
|
if (tableNameBuilder_ == null) {
|
||||||
if (((bitField0_ & 0x00000001) == 0x00000001) &&
|
if (((bitField0_ & 0x00000001) == 0x00000001) &&
|
||||||
tableName_ != org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance()) {
|
tableName_ != org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance()) {
|
||||||
tableName_ =
|
tableName_ =
|
||||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.newBuilder(tableName_).mergeFrom(value).buildPartial();
|
org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.newBuilder(tableName_).mergeFrom(value).buildPartial();
|
||||||
} else {
|
} else {
|
||||||
tableName_ = value;
|
tableName_ = value;
|
||||||
}
|
}
|
||||||
@ -3013,7 +3013,7 @@ public final class SecureBulkLoadProtos {
|
|||||||
*/
|
*/
|
||||||
public Builder clearTableName() {
|
public Builder clearTableName() {
|
||||||
if (tableNameBuilder_ == null) {
|
if (tableNameBuilder_ == null) {
|
||||||
tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance();
|
tableName_ = org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance();
|
||||||
onChanged();
|
onChanged();
|
||||||
} else {
|
} else {
|
||||||
tableNameBuilder_.clear();
|
tableNameBuilder_.clear();
|
||||||
@ -3024,7 +3024,7 @@ public final class SecureBulkLoadProtos {
|
|||||||
/**
|
/**
|
||||||
* <code>required .hbase.pb.TableName table_name = 1;</code>
|
* <code>required .hbase.pb.TableName table_name = 1;</code>
|
||||||
*/
|
*/
|
||||||
public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder getTableNameBuilder() {
|
public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder getTableNameBuilder() {
|
||||||
bitField0_ |= 0x00000001;
|
bitField0_ |= 0x00000001;
|
||||||
onChanged();
|
onChanged();
|
||||||
return getTableNameFieldBuilder().getBuilder();
|
return getTableNameFieldBuilder().getBuilder();
|
||||||
@ -3032,7 +3032,7 @@ public final class SecureBulkLoadProtos {
|
|||||||
/**
|
/**
|
||||||
* <code>required .hbase.pb.TableName table_name = 1;</code>
|
* <code>required .hbase.pb.TableName table_name = 1;</code>
|
||||||
*/
|
*/
|
||||||
public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder() {
|
public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder getTableNameOrBuilder() {
|
||||||
if (tableNameBuilder_ != null) {
|
if (tableNameBuilder_ != null) {
|
||||||
return tableNameBuilder_.getMessageOrBuilder();
|
return tableNameBuilder_.getMessageOrBuilder();
|
||||||
} else {
|
} else {
|
||||||
@ -3043,11 +3043,11 @@ public final class SecureBulkLoadProtos {
|
|||||||
* <code>required .hbase.pb.TableName table_name = 1;</code>
|
* <code>required .hbase.pb.TableName table_name = 1;</code>
|
||||||
*/
|
*/
|
||||||
private com.google.protobuf.SingleFieldBuilder<
|
private com.google.protobuf.SingleFieldBuilder<
|
||||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder>
|
org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder>
|
||||||
getTableNameFieldBuilder() {
|
getTableNameFieldBuilder() {
|
||||||
if (tableNameBuilder_ == null) {
|
if (tableNameBuilder_ == null) {
|
||||||
tableNameBuilder_ = new com.google.protobuf.SingleFieldBuilder<
|
tableNameBuilder_ = new com.google.protobuf.SingleFieldBuilder<
|
||||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder>(
|
org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder>(
|
||||||
tableName_,
|
tableName_,
|
||||||
getParentForChildren(),
|
getParentForChildren(),
|
||||||
isClean());
|
isClean());
|
||||||
@ -4857,30 +4857,31 @@ public final class SecureBulkLoadProtos {
|
|||||||
descriptor;
|
descriptor;
|
||||||
static {
|
static {
|
||||||
java.lang.String[] descriptorData = {
|
java.lang.String[] descriptorData = {
|
||||||
"\n\024SecureBulkLoad.proto\022\010hbase.pb\032\013HBase." +
|
"\n\024SecureBulkLoad.proto\022\010hbase.pb\032\013Table." +
|
||||||
"proto\032\014Client.proto\"\266\001\n\033SecureBulkLoadHF" +
|
"proto\032\013HBase.proto\032\014Client.proto\"\266\001\n\033Sec" +
|
||||||
"ilesRequest\022>\n\013family_path\030\001 \003(\0132).hbase" +
|
"ureBulkLoadHFilesRequest\022>\n\013family_path\030" +
|
||||||
".pb.BulkLoadHFileRequest.FamilyPath\022\026\n\016a" +
|
"\001 \003(\0132).hbase.pb.BulkLoadHFileRequest.Fa" +
|
||||||
"ssign_seq_num\030\002 \001(\010\022+\n\010fs_token\030\003 \002(\0132\031." +
|
"milyPath\022\026\n\016assign_seq_num\030\002 \001(\010\022+\n\010fs_t" +
|
||||||
"hbase.pb.DelegationToken\022\022\n\nbulk_token\030\004" +
|
"oken\030\003 \002(\0132\031.hbase.pb.DelegationToken\022\022\n" +
|
||||||
" \002(\t\".\n\034SecureBulkLoadHFilesResponse\022\016\n\006" +
|
"\nbulk_token\030\004 \002(\t\".\n\034SecureBulkLoadHFile" +
|
||||||
"loaded\030\001 \002(\010\"V\n\017DelegationToken\022\022\n\nident" +
|
"sResponse\022\016\n\006loaded\030\001 \002(\010\"V\n\017DelegationT" +
|
||||||
"ifier\030\001 \001(\014\022\020\n\010password\030\002 \001(\014\022\014\n\004kind\030\003 " +
|
"oken\022\022\n\nidentifier\030\001 \001(\014\022\020\n\010password\030\002 \001" +
|
||||||
"\001(\t\022\017\n\007service\030\004 \001(\t\"A\n\026PrepareBulkLoadR",
|
"(\014\022\014\n\004kind\030\003 \001(\t\022\017\n\007service\030\004 \001(\t\"A\n\026Pre",
|
||||||
"equest\022\'\n\ntable_name\030\001 \002(\0132\023.hbase.pb.Ta" +
|
"pareBulkLoadRequest\022\'\n\ntable_name\030\001 \002(\0132" +
|
||||||
"bleName\"-\n\027PrepareBulkLoadResponse\022\022\n\nbu" +
|
"\023.hbase.pb.TableName\"-\n\027PrepareBulkLoadR" +
|
||||||
"lk_token\030\001 \002(\t\",\n\026CleanupBulkLoadRequest" +
|
"esponse\022\022\n\nbulk_token\030\001 \002(\t\",\n\026CleanupBu" +
|
||||||
"\022\022\n\nbulk_token\030\001 \002(\t\"\031\n\027CleanupBulkLoadR" +
|
"lkLoadRequest\022\022\n\nbulk_token\030\001 \002(\t\"\031\n\027Cle" +
|
||||||
"esponse2\256\002\n\025SecureBulkLoadService\022V\n\017Pre" +
|
"anupBulkLoadResponse2\256\002\n\025SecureBulkLoadS" +
|
||||||
"pareBulkLoad\022 .hbase.pb.PrepareBulkLoadR" +
|
"ervice\022V\n\017PrepareBulkLoad\022 .hbase.pb.Pre" +
|
||||||
"equest\032!.hbase.pb.PrepareBulkLoadRespons" +
|
"pareBulkLoadRequest\032!.hbase.pb.PrepareBu" +
|
||||||
"e\022e\n\024SecureBulkLoadHFiles\022%.hbase.pb.Sec" +
|
"lkLoadResponse\022e\n\024SecureBulkLoadHFiles\022%" +
|
||||||
"ureBulkLoadHFilesRequest\032&.hbase.pb.Secu" +
|
".hbase.pb.SecureBulkLoadHFilesRequest\032&." +
|
||||||
"reBulkLoadHFilesResponse\022V\n\017CleanupBulkL",
|
"hbase.pb.SecureBulkLoadHFilesResponse\022V\n",
|
||||||
"oad\022 .hbase.pb.CleanupBulkLoadRequest\032!." +
|
"\017CleanupBulkLoad\022 .hbase.pb.CleanupBulkL" +
|
||||||
"hbase.pb.CleanupBulkLoadResponseBJ\n*org." +
|
"oadRequest\032!.hbase.pb.CleanupBulkLoadRes" +
|
||||||
"apache.hadoop.hbase.protobuf.generatedB\024" +
|
"ponseBJ\n*org.apache.hadoop.hbase.protobu" +
|
||||||
"SecureBulkLoadProtosH\001\210\001\001\240\001\001"
|
"f.generatedB\024SecureBulkLoadProtosH\001\210\001\001\240\001" +
|
||||||
|
"\001"
|
||||||
};
|
};
|
||||||
com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
|
com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
|
||||||
new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() {
|
new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() {
|
||||||
@ -4935,6 +4936,7 @@ public final class SecureBulkLoadProtos {
|
|||||||
com.google.protobuf.Descriptors.FileDescriptor
|
com.google.protobuf.Descriptors.FileDescriptor
|
||||||
.internalBuildGeneratedFileFrom(descriptorData,
|
.internalBuildGeneratedFileFrom(descriptorData,
|
||||||
new com.google.protobuf.Descriptors.FileDescriptor[] {
|
new com.google.protobuf.Descriptors.FileDescriptor[] {
|
||||||
|
org.apache.hadoop.hbase.protobuf.generated.TableProtos.getDescriptor(),
|
||||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.getDescriptor(),
|
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.getDescriptor(),
|
||||||
org.apache.hadoop.hbase.protobuf.generated.ClientProtos.getDescriptor(),
|
org.apache.hadoop.hbase.protobuf.generated.ClientProtos.getDescriptor(),
|
||||||
}, assigner);
|
}, assigner);
|
||||||
|
File diff suppressed because it is too large
Load Diff
@ -0,0 +1,607 @@
|
|||||||
|
// Generated by the protocol buffer compiler. DO NOT EDIT!
|
||||||
|
// source: Table.proto
|
||||||
|
|
||||||
|
package org.apache.hadoop.hbase.protobuf.generated;
|
||||||
|
|
||||||
|
public final class TableProtos {
|
||||||
|
private TableProtos() {}
|
||||||
|
public static void registerAllExtensions(
|
||||||
|
com.google.protobuf.ExtensionRegistry registry) {
|
||||||
|
}
|
||||||
|
public interface TableNameOrBuilder
|
||||||
|
extends com.google.protobuf.MessageOrBuilder {
|
||||||
|
|
||||||
|
// required bytes namespace = 1;
|
||||||
|
/**
|
||||||
|
* <code>required bytes namespace = 1;</code>
|
||||||
|
*/
|
||||||
|
boolean hasNamespace();
|
||||||
|
/**
|
||||||
|
* <code>required bytes namespace = 1;</code>
|
||||||
|
*/
|
||||||
|
com.google.protobuf.ByteString getNamespace();
|
||||||
|
|
||||||
|
// required bytes qualifier = 2;
|
||||||
|
/**
|
||||||
|
* <code>required bytes qualifier = 2;</code>
|
||||||
|
*/
|
||||||
|
boolean hasQualifier();
|
||||||
|
/**
|
||||||
|
* <code>required bytes qualifier = 2;</code>
|
||||||
|
*/
|
||||||
|
com.google.protobuf.ByteString getQualifier();
|
||||||
|
}
|
||||||
|
/**
|
||||||
|
* Protobuf type {@code hbase.pb.TableName}
|
||||||
|
*
|
||||||
|
* <pre>
|
||||||
|
**
|
||||||
|
* Table Name
|
||||||
|
* </pre>
|
||||||
|
*/
|
||||||
|
public static final class TableName extends
|
||||||
|
com.google.protobuf.GeneratedMessage
|
||||||
|
implements TableNameOrBuilder {
|
||||||
|
// Use TableName.newBuilder() to construct.
|
||||||
|
private TableName(com.google.protobuf.GeneratedMessage.Builder<?> builder) {
|
||||||
|
super(builder);
|
||||||
|
this.unknownFields = builder.getUnknownFields();
|
||||||
|
}
|
||||||
|
private TableName(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
|
||||||
|
|
||||||
|
private static final TableName defaultInstance;
|
||||||
|
public static TableName getDefaultInstance() {
|
||||||
|
return defaultInstance;
|
||||||
|
}
|
||||||
|
|
||||||
|
public TableName getDefaultInstanceForType() {
|
||||||
|
return defaultInstance;
|
||||||
|
}
|
||||||
|
|
||||||
|
private final com.google.protobuf.UnknownFieldSet unknownFields;
|
||||||
|
@java.lang.Override
|
||||||
|
public final com.google.protobuf.UnknownFieldSet
|
||||||
|
getUnknownFields() {
|
||||||
|
return this.unknownFields;
|
||||||
|
}
|
||||||
|
private TableName(
|
||||||
|
com.google.protobuf.CodedInputStream input,
|
||||||
|
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
|
||||||
|
throws com.google.protobuf.InvalidProtocolBufferException {
|
||||||
|
initFields();
|
||||||
|
int mutable_bitField0_ = 0;
|
||||||
|
com.google.protobuf.UnknownFieldSet.Builder unknownFields =
|
||||||
|
com.google.protobuf.UnknownFieldSet.newBuilder();
|
||||||
|
try {
|
||||||
|
boolean done = false;
|
||||||
|
while (!done) {
|
||||||
|
int tag = input.readTag();
|
||||||
|
switch (tag) {
|
||||||
|
case 0:
|
||||||
|
done = true;
|
||||||
|
break;
|
||||||
|
default: {
|
||||||
|
if (!parseUnknownField(input, unknownFields,
|
||||||
|
extensionRegistry, tag)) {
|
||||||
|
done = true;
|
||||||
|
}
|
||||||
|
break;
|
||||||
|
}
|
||||||
|
case 10: {
|
||||||
|
bitField0_ |= 0x00000001;
|
||||||
|
namespace_ = input.readBytes();
|
||||||
|
break;
|
||||||
|
}
|
||||||
|
case 18: {
|
||||||
|
bitField0_ |= 0x00000002;
|
||||||
|
qualifier_ = input.readBytes();
|
||||||
|
break;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
} catch (com.google.protobuf.InvalidProtocolBufferException e) {
|
||||||
|
throw e.setUnfinishedMessage(this);
|
||||||
|
} catch (java.io.IOException e) {
|
||||||
|
throw new com.google.protobuf.InvalidProtocolBufferException(
|
||||||
|
e.getMessage()).setUnfinishedMessage(this);
|
||||||
|
} finally {
|
||||||
|
this.unknownFields = unknownFields.build();
|
||||||
|
makeExtensionsImmutable();
|
||||||
|
}
|
||||||
|
}
|
||||||
|
public static final com.google.protobuf.Descriptors.Descriptor
|
||||||
|
getDescriptor() {
|
||||||
|
return org.apache.hadoop.hbase.protobuf.generated.TableProtos.internal_static_hbase_pb_TableName_descriptor;
|
||||||
|
}
|
||||||
|
|
||||||
|
protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
|
||||||
|
internalGetFieldAccessorTable() {
|
||||||
|
return org.apache.hadoop.hbase.protobuf.generated.TableProtos.internal_static_hbase_pb_TableName_fieldAccessorTable
|
||||||
|
.ensureFieldAccessorsInitialized(
|
||||||
|
org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.class, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder.class);
|
||||||
|
}
|
||||||
|
|
||||||
|
public static com.google.protobuf.Parser<TableName> PARSER =
|
||||||
|
new com.google.protobuf.AbstractParser<TableName>() {
|
||||||
|
public TableName parsePartialFrom(
|
||||||
|
com.google.protobuf.CodedInputStream input,
|
||||||
|
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
|
||||||
|
throws com.google.protobuf.InvalidProtocolBufferException {
|
||||||
|
return new TableName(input, extensionRegistry);
|
||||||
|
}
|
||||||
|
};
|
||||||
|
|
||||||
|
@java.lang.Override
|
||||||
|
public com.google.protobuf.Parser<TableName> getParserForType() {
|
||||||
|
return PARSER;
|
||||||
|
}
|
||||||
|
|
||||||
|
private int bitField0_;
|
||||||
|
// required bytes namespace = 1;
|
||||||
|
public static final int NAMESPACE_FIELD_NUMBER = 1;
|
||||||
|
private com.google.protobuf.ByteString namespace_;
|
||||||
|
/**
|
||||||
|
* <code>required bytes namespace = 1;</code>
|
||||||
|
*/
|
||||||
|
public boolean hasNamespace() {
|
||||||
|
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||||
|
}
|
||||||
|
/**
|
||||||
|
* <code>required bytes namespace = 1;</code>
|
||||||
|
*/
|
||||||
|
public com.google.protobuf.ByteString getNamespace() {
|
||||||
|
return namespace_;
|
||||||
|
}
|
||||||
|
|
||||||
|
// required bytes qualifier = 2;
|
||||||
|
public static final int QUALIFIER_FIELD_NUMBER = 2;
|
||||||
|
private com.google.protobuf.ByteString qualifier_;
|
||||||
|
/**
|
||||||
|
* <code>required bytes qualifier = 2;</code>
|
||||||
|
*/
|
||||||
|
public boolean hasQualifier() {
|
||||||
|
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||||
|
}
|
||||||
|
/**
|
||||||
|
* <code>required bytes qualifier = 2;</code>
|
||||||
|
*/
|
||||||
|
public com.google.protobuf.ByteString getQualifier() {
|
||||||
|
return qualifier_;
|
||||||
|
}
|
||||||
|
|
||||||
|
private void initFields() {
|
||||||
|
namespace_ = com.google.protobuf.ByteString.EMPTY;
|
||||||
|
qualifier_ = com.google.protobuf.ByteString.EMPTY;
|
||||||
|
}
|
||||||
|
private byte memoizedIsInitialized = -1;
|
||||||
|
public final boolean isInitialized() {
|
||||||
|
byte isInitialized = memoizedIsInitialized;
|
||||||
|
if (isInitialized != -1) return isInitialized == 1;
|
||||||
|
|
||||||
|
if (!hasNamespace()) {
|
||||||
|
memoizedIsInitialized = 0;
|
||||||
|
return false;
|
||||||
|
}
|
||||||
|
if (!hasQualifier()) {
|
||||||
|
memoizedIsInitialized = 0;
|
||||||
|
return false;
|
||||||
|
}
|
||||||
|
memoizedIsInitialized = 1;
|
||||||
|
return true;
|
||||||
|
}
|
||||||
|
|
||||||
|
public void writeTo(com.google.protobuf.CodedOutputStream output)
|
||||||
|
throws java.io.IOException {
|
||||||
|
getSerializedSize();
|
||||||
|
if (((bitField0_ & 0x00000001) == 0x00000001)) {
|
||||||
|
output.writeBytes(1, namespace_);
|
||||||
|
}
|
||||||
|
if (((bitField0_ & 0x00000002) == 0x00000002)) {
|
||||||
|
output.writeBytes(2, qualifier_);
|
||||||
|
}
|
||||||
|
getUnknownFields().writeTo(output);
|
||||||
|
}
|
||||||
|
|
||||||
|
private int memoizedSerializedSize = -1;
|
||||||
|
public int getSerializedSize() {
|
||||||
|
int size = memoizedSerializedSize;
|
||||||
|
if (size != -1) return size;
|
||||||
|
|
||||||
|
size = 0;
|
||||||
|
if (((bitField0_ & 0x00000001) == 0x00000001)) {
|
||||||
|
size += com.google.protobuf.CodedOutputStream
|
||||||
|
.computeBytesSize(1, namespace_);
|
||||||
|
}
|
||||||
|
if (((bitField0_ & 0x00000002) == 0x00000002)) {
|
||||||
|
size += com.google.protobuf.CodedOutputStream
|
||||||
|
.computeBytesSize(2, qualifier_);
|
||||||
|
}
|
||||||
|
size += getUnknownFields().getSerializedSize();
|
||||||
|
memoizedSerializedSize = size;
|
||||||
|
return size;
|
||||||
|
}
|
||||||
|
|
||||||
|
private static final long serialVersionUID = 0L;
|
||||||
|
@java.lang.Override
|
||||||
|
protected java.lang.Object writeReplace()
|
||||||
|
throws java.io.ObjectStreamException {
|
||||||
|
return super.writeReplace();
|
||||||
|
}
|
||||||
|
|
||||||
|
@java.lang.Override
|
||||||
|
public boolean equals(final java.lang.Object obj) {
|
||||||
|
if (obj == this) {
|
||||||
|
return true;
|
||||||
|
}
|
||||||
|
if (!(obj instanceof org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName)) {
|
||||||
|
return super.equals(obj);
|
||||||
|
}
|
||||||
|
org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName other = (org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName) obj;
|
||||||
|
|
||||||
|
boolean result = true;
|
||||||
|
result = result && (hasNamespace() == other.hasNamespace());
|
||||||
|
if (hasNamespace()) {
|
||||||
|
result = result && getNamespace()
|
||||||
|
.equals(other.getNamespace());
|
||||||
|
}
|
||||||
|
result = result && (hasQualifier() == other.hasQualifier());
|
||||||
|
if (hasQualifier()) {
|
||||||
|
result = result && getQualifier()
|
||||||
|
.equals(other.getQualifier());
|
||||||
|
}
|
||||||
|
result = result &&
|
||||||
|
getUnknownFields().equals(other.getUnknownFields());
|
||||||
|
return result;
|
||||||
|
}
|
||||||
|
|
||||||
|
private int memoizedHashCode = 0;
|
||||||
|
@java.lang.Override
|
||||||
|
public int hashCode() {
|
||||||
|
if (memoizedHashCode != 0) {
|
||||||
|
return memoizedHashCode;
|
||||||
|
}
|
||||||
|
int hash = 41;
|
||||||
|
hash = (19 * hash) + getDescriptorForType().hashCode();
|
||||||
|
if (hasNamespace()) {
|
||||||
|
hash = (37 * hash) + NAMESPACE_FIELD_NUMBER;
|
||||||
|
hash = (53 * hash) + getNamespace().hashCode();
|
||||||
|
}
|
||||||
|
if (hasQualifier()) {
|
||||||
|
hash = (37 * hash) + QUALIFIER_FIELD_NUMBER;
|
||||||
|
hash = (53 * hash) + getQualifier().hashCode();
|
||||||
|
}
|
||||||
|
hash = (29 * hash) + getUnknownFields().hashCode();
|
||||||
|
memoizedHashCode = hash;
|
||||||
|
return hash;
|
||||||
|
}
|
||||||
|
|
||||||
|
public static org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName parseFrom(
|
||||||
|
com.google.protobuf.ByteString data)
|
||||||
|
throws com.google.protobuf.InvalidProtocolBufferException {
|
||||||
|
return PARSER.parseFrom(data);
|
||||||
|
}
|
||||||
|
public static org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName parseFrom(
|
||||||
|
com.google.protobuf.ByteString data,
|
||||||
|
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
|
||||||
|
throws com.google.protobuf.InvalidProtocolBufferException {
|
||||||
|
return PARSER.parseFrom(data, extensionRegistry);
|
||||||
|
}
|
||||||
|
public static org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName parseFrom(byte[] data)
|
||||||
|
throws com.google.protobuf.InvalidProtocolBufferException {
|
||||||
|
return PARSER.parseFrom(data);
|
||||||
|
}
|
||||||
|
public static org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName parseFrom(
|
||||||
|
byte[] data,
|
||||||
|
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
|
||||||
|
throws com.google.protobuf.InvalidProtocolBufferException {
|
||||||
|
return PARSER.parseFrom(data, extensionRegistry);
|
||||||
|
}
|
||||||
|
public static org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName parseFrom(java.io.InputStream input)
|
||||||
|
throws java.io.IOException {
|
||||||
|
return PARSER.parseFrom(input);
|
||||||
|
}
|
||||||
|
public static org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName parseFrom(
|
||||||
|
java.io.InputStream input,
|
||||||
|
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
|
||||||
|
throws java.io.IOException {
|
||||||
|
return PARSER.parseFrom(input, extensionRegistry);
|
||||||
|
}
|
||||||
|
public static org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName parseDelimitedFrom(java.io.InputStream input)
|
||||||
|
throws java.io.IOException {
|
||||||
|
return PARSER.parseDelimitedFrom(input);
|
||||||
|
}
|
||||||
|
public static org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName parseDelimitedFrom(
|
||||||
|
java.io.InputStream input,
|
||||||
|
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
|
||||||
|
throws java.io.IOException {
|
||||||
|
return PARSER.parseDelimitedFrom(input, extensionRegistry);
|
||||||
|
}
|
||||||
|
public static org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName parseFrom(
|
||||||
|
com.google.protobuf.CodedInputStream input)
|
||||||
|
throws java.io.IOException {
|
||||||
|
return PARSER.parseFrom(input);
|
||||||
|
}
|
||||||
|
public static org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName parseFrom(
|
||||||
|
com.google.protobuf.CodedInputStream input,
|
||||||
|
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
|
||||||
|
throws java.io.IOException {
|
||||||
|
return PARSER.parseFrom(input, extensionRegistry);
|
||||||
|
}
|
||||||
|
|
||||||
|
public static Builder newBuilder() { return Builder.create(); }
|
||||||
|
public Builder newBuilderForType() { return newBuilder(); }
|
||||||
|
public static Builder newBuilder(org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName prototype) {
|
||||||
|
return newBuilder().mergeFrom(prototype);
|
||||||
|
}
|
||||||
|
public Builder toBuilder() { return newBuilder(this); }
|
||||||
|
|
||||||
|
@java.lang.Override
|
||||||
|
protected Builder newBuilderForType(
|
||||||
|
com.google.protobuf.GeneratedMessage.BuilderParent parent) {
|
||||||
|
Builder builder = new Builder(parent);
|
||||||
|
return builder;
|
||||||
|
}
|
||||||
|
/**
|
||||||
|
* Protobuf type {@code hbase.pb.TableName}
|
||||||
|
*
|
||||||
|
* <pre>
|
||||||
|
**
|
||||||
|
* Table Name
|
||||||
|
* </pre>
|
||||||
|
*/
|
||||||
|
public static final class Builder extends
|
||||||
|
com.google.protobuf.GeneratedMessage.Builder<Builder>
|
||||||
|
implements org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder {
|
||||||
|
public static final com.google.protobuf.Descriptors.Descriptor
|
||||||
|
getDescriptor() {
|
||||||
|
return org.apache.hadoop.hbase.protobuf.generated.TableProtos.internal_static_hbase_pb_TableName_descriptor;
|
||||||
|
}
|
||||||
|
|
||||||
|
protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
|
||||||
|
internalGetFieldAccessorTable() {
|
||||||
|
return org.apache.hadoop.hbase.protobuf.generated.TableProtos.internal_static_hbase_pb_TableName_fieldAccessorTable
|
||||||
|
.ensureFieldAccessorsInitialized(
|
||||||
|
org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.class, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder.class);
|
||||||
|
}
|
||||||
|
|
||||||
|
// Construct using org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.newBuilder()
|
||||||
|
private Builder() {
|
||||||
|
maybeForceBuilderInitialization();
|
||||||
|
}
|
||||||
|
|
||||||
|
private Builder(
|
||||||
|
com.google.protobuf.GeneratedMessage.BuilderParent parent) {
|
||||||
|
super(parent);
|
||||||
|
maybeForceBuilderInitialization();
|
||||||
|
}
|
||||||
|
private void maybeForceBuilderInitialization() {
|
||||||
|
if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
|
||||||
|
}
|
||||||
|
}
|
||||||
|
private static Builder create() {
|
||||||
|
return new Builder();
|
||||||
|
}
|
||||||
|
|
||||||
|
public Builder clear() {
|
||||||
|
super.clear();
|
||||||
|
namespace_ = com.google.protobuf.ByteString.EMPTY;
|
||||||
|
bitField0_ = (bitField0_ & ~0x00000001);
|
||||||
|
qualifier_ = com.google.protobuf.ByteString.EMPTY;
|
||||||
|
bitField0_ = (bitField0_ & ~0x00000002);
|
||||||
|
return this;
|
||||||
|
}
|
||||||
|
|
||||||
|
public Builder clone() {
|
||||||
|
return create().mergeFrom(buildPartial());
|
||||||
|
}
|
||||||
|
|
||||||
|
public com.google.protobuf.Descriptors.Descriptor
|
||||||
|
getDescriptorForType() {
|
||||||
|
return org.apache.hadoop.hbase.protobuf.generated.TableProtos.internal_static_hbase_pb_TableName_descriptor;
|
||||||
|
}
|
||||||
|
|
||||||
|
public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName getDefaultInstanceForType() {
|
||||||
|
return org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance();
|
||||||
|
}
|
||||||
|
|
||||||
|
public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName build() {
|
||||||
|
org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName result = buildPartial();
|
||||||
|
if (!result.isInitialized()) {
|
||||||
|
throw newUninitializedMessageException(result);
|
||||||
|
}
|
||||||
|
return result;
|
||||||
|
}
|
||||||
|
|
||||||
|
public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName buildPartial() {
|
||||||
|
org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName result = new org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName(this);
|
||||||
|
int from_bitField0_ = bitField0_;
|
||||||
|
int to_bitField0_ = 0;
|
||||||
|
if (((from_bitField0_ & 0x00000001) == 0x00000001)) {
|
||||||
|
to_bitField0_ |= 0x00000001;
|
||||||
|
}
|
||||||
|
result.namespace_ = namespace_;
|
||||||
|
if (((from_bitField0_ & 0x00000002) == 0x00000002)) {
|
||||||
|
to_bitField0_ |= 0x00000002;
|
||||||
|
}
|
||||||
|
result.qualifier_ = qualifier_;
|
||||||
|
result.bitField0_ = to_bitField0_;
|
||||||
|
onBuilt();
|
||||||
|
return result;
|
||||||
|
}
|
||||||
|
|
||||||
|
public Builder mergeFrom(com.google.protobuf.Message other) {
|
||||||
|
if (other instanceof org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName) {
|
||||||
|
return mergeFrom((org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName)other);
|
||||||
|
} else {
|
||||||
|
super.mergeFrom(other);
|
||||||
|
return this;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
public Builder mergeFrom(org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName other) {
|
||||||
|
if (other == org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance()) return this;
|
||||||
|
if (other.hasNamespace()) {
|
||||||
|
setNamespace(other.getNamespace());
|
||||||
|
}
|
||||||
|
if (other.hasQualifier()) {
|
||||||
|
setQualifier(other.getQualifier());
|
||||||
|
}
|
||||||
|
this.mergeUnknownFields(other.getUnknownFields());
|
||||||
|
return this;
|
||||||
|
}
|
||||||
|
|
||||||
|
public final boolean isInitialized() {
|
||||||
|
if (!hasNamespace()) {
|
||||||
|
|
||||||
|
return false;
|
||||||
|
}
|
||||||
|
if (!hasQualifier()) {
|
||||||
|
|
||||||
|
return false;
|
||||||
|
}
|
||||||
|
return true;
|
||||||
|
}
|
||||||
|
|
||||||
|
public Builder mergeFrom(
|
||||||
|
com.google.protobuf.CodedInputStream input,
|
||||||
|
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
|
||||||
|
throws java.io.IOException {
|
||||||
|
org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName parsedMessage = null;
|
||||||
|
try {
|
||||||
|
parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
|
||||||
|
} catch (com.google.protobuf.InvalidProtocolBufferException e) {
|
||||||
|
parsedMessage = (org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName) e.getUnfinishedMessage();
|
||||||
|
throw e;
|
||||||
|
} finally {
|
||||||
|
if (parsedMessage != null) {
|
||||||
|
mergeFrom(parsedMessage);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
return this;
|
||||||
|
}
|
||||||
|
private int bitField0_;
|
||||||
|
|
||||||
|
// required bytes namespace = 1;
|
||||||
|
private com.google.protobuf.ByteString namespace_ = com.google.protobuf.ByteString.EMPTY;
|
||||||
|
/**
|
||||||
|
* <code>required bytes namespace = 1;</code>
|
||||||
|
*/
|
||||||
|
public boolean hasNamespace() {
|
||||||
|
return ((bitField0_ & 0x00000001) == 0x00000001);
|
||||||
|
}
|
||||||
|
/**
|
||||||
|
* <code>required bytes namespace = 1;</code>
|
||||||
|
*/
|
||||||
|
public com.google.protobuf.ByteString getNamespace() {
|
||||||
|
return namespace_;
|
||||||
|
}
|
||||||
|
/**
|
||||||
|
* <code>required bytes namespace = 1;</code>
|
||||||
|
*/
|
||||||
|
public Builder setNamespace(com.google.protobuf.ByteString value) {
|
||||||
|
if (value == null) {
|
||||||
|
throw new NullPointerException();
|
||||||
|
}
|
||||||
|
bitField0_ |= 0x00000001;
|
||||||
|
namespace_ = value;
|
||||||
|
onChanged();
|
||||||
|
return this;
|
||||||
|
}
|
||||||
|
/**
|
||||||
|
* <code>required bytes namespace = 1;</code>
|
||||||
|
*/
|
||||||
|
public Builder clearNamespace() {
|
||||||
|
bitField0_ = (bitField0_ & ~0x00000001);
|
||||||
|
namespace_ = getDefaultInstance().getNamespace();
|
||||||
|
onChanged();
|
||||||
|
return this;
|
||||||
|
}
|
||||||
|
|
||||||
|
// required bytes qualifier = 2;
|
||||||
|
private com.google.protobuf.ByteString qualifier_ = com.google.protobuf.ByteString.EMPTY;
|
||||||
|
/**
|
||||||
|
* <code>required bytes qualifier = 2;</code>
|
||||||
|
*/
|
||||||
|
public boolean hasQualifier() {
|
||||||
|
return ((bitField0_ & 0x00000002) == 0x00000002);
|
||||||
|
}
|
||||||
|
/**
|
||||||
|
* <code>required bytes qualifier = 2;</code>
|
||||||
|
*/
|
||||||
|
public com.google.protobuf.ByteString getQualifier() {
|
||||||
|
return qualifier_;
|
||||||
|
}
|
||||||
|
/**
|
||||||
|
* <code>required bytes qualifier = 2;</code>
|
||||||
|
*/
|
||||||
|
public Builder setQualifier(com.google.protobuf.ByteString value) {
|
||||||
|
if (value == null) {
|
||||||
|
throw new NullPointerException();
|
||||||
|
}
|
||||||
|
bitField0_ |= 0x00000002;
|
||||||
|
qualifier_ = value;
|
||||||
|
onChanged();
|
||||||
|
return this;
|
||||||
|
}
|
||||||
|
/**
|
||||||
|
* <code>required bytes qualifier = 2;</code>
|
||||||
|
*/
|
||||||
|
public Builder clearQualifier() {
|
||||||
|
bitField0_ = (bitField0_ & ~0x00000002);
|
||||||
|
qualifier_ = getDefaultInstance().getQualifier();
|
||||||
|
onChanged();
|
||||||
|
return this;
|
||||||
|
}
|
||||||
|
|
||||||
|
// @@protoc_insertion_point(builder_scope:hbase.pb.TableName)
|
||||||
|
}
|
||||||
|
|
||||||
|
static {
|
||||||
|
defaultInstance = new TableName(true);
|
||||||
|
defaultInstance.initFields();
|
||||||
|
}
|
||||||
|
|
||||||
|
// @@protoc_insertion_point(class_scope:hbase.pb.TableName)
|
||||||
|
}
|
||||||
|
|
||||||
|
private static com.google.protobuf.Descriptors.Descriptor
|
||||||
|
internal_static_hbase_pb_TableName_descriptor;
|
||||||
|
private static
|
||||||
|
com.google.protobuf.GeneratedMessage.FieldAccessorTable
|
||||||
|
internal_static_hbase_pb_TableName_fieldAccessorTable;
|
||||||
|
|
||||||
|
public static com.google.protobuf.Descriptors.FileDescriptor
|
||||||
|
getDescriptor() {
|
||||||
|
return descriptor;
|
||||||
|
}
|
||||||
|
private static com.google.protobuf.Descriptors.FileDescriptor
|
||||||
|
descriptor;
|
||||||
|
static {
|
||||||
|
java.lang.String[] descriptorData = {
|
||||||
|
"\n\013Table.proto\022\010hbase.pb\"1\n\tTableName\022\021\n\t" +
|
||||||
|
"namespace\030\001 \002(\014\022\021\n\tqualifier\030\002 \002(\014B>\n*or" +
|
||||||
|
"g.apache.hadoop.hbase.protobuf.generated" +
|
||||||
|
"B\013TableProtosH\001\240\001\001"
|
||||||
|
};
|
||||||
|
com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
|
||||||
|
new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() {
|
||||||
|
public com.google.protobuf.ExtensionRegistry assignDescriptors(
|
||||||
|
com.google.protobuf.Descriptors.FileDescriptor root) {
|
||||||
|
descriptor = root;
|
||||||
|
internal_static_hbase_pb_TableName_descriptor =
|
||||||
|
getDescriptor().getMessageTypes().get(0);
|
||||||
|
internal_static_hbase_pb_TableName_fieldAccessorTable = new
|
||||||
|
com.google.protobuf.GeneratedMessage.FieldAccessorTable(
|
||||||
|
internal_static_hbase_pb_TableName_descriptor,
|
||||||
|
new java.lang.String[] { "Namespace", "Qualifier", });
|
||||||
|
return null;
|
||||||
|
}
|
||||||
|
};
|
||||||
|
com.google.protobuf.Descriptors.FileDescriptor
|
||||||
|
.internalBuildGeneratedFileFrom(descriptorData,
|
||||||
|
new com.google.protobuf.Descriptors.FileDescriptor[] {
|
||||||
|
}, assigner);
|
||||||
|
}
|
||||||
|
|
||||||
|
// @@protoc_insertion_point(outer_class_scope)
|
||||||
|
}
|
@ -8813,11 +8813,11 @@ public final class WALProtos {
|
|||||||
/**
|
/**
|
||||||
* <code>required .hbase.pb.TableName table_name = 1;</code>
|
* <code>required .hbase.pb.TableName table_name = 1;</code>
|
||||||
*/
|
*/
|
||||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName getTableName();
|
org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName getTableName();
|
||||||
/**
|
/**
|
||||||
* <code>required .hbase.pb.TableName table_name = 1;</code>
|
* <code>required .hbase.pb.TableName table_name = 1;</code>
|
||||||
*/
|
*/
|
||||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder();
|
org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder getTableNameOrBuilder();
|
||||||
|
|
||||||
// required bytes encoded_region_name = 2;
|
// required bytes encoded_region_name = 2;
|
||||||
/**
|
/**
|
||||||
@ -8921,11 +8921,11 @@ public final class WALProtos {
|
|||||||
break;
|
break;
|
||||||
}
|
}
|
||||||
case 10: {
|
case 10: {
|
||||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder subBuilder = null;
|
org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder subBuilder = null;
|
||||||
if (((bitField0_ & 0x00000001) == 0x00000001)) {
|
if (((bitField0_ & 0x00000001) == 0x00000001)) {
|
||||||
subBuilder = tableName_.toBuilder();
|
subBuilder = tableName_.toBuilder();
|
||||||
}
|
}
|
||||||
tableName_ = input.readMessage(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.PARSER, extensionRegistry);
|
tableName_ = input.readMessage(org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.PARSER, extensionRegistry);
|
||||||
if (subBuilder != null) {
|
if (subBuilder != null) {
|
||||||
subBuilder.mergeFrom(tableName_);
|
subBuilder.mergeFrom(tableName_);
|
||||||
tableName_ = subBuilder.buildPartial();
|
tableName_ = subBuilder.buildPartial();
|
||||||
@ -8996,7 +8996,7 @@ public final class WALProtos {
|
|||||||
private int bitField0_;
|
private int bitField0_;
|
||||||
// required .hbase.pb.TableName table_name = 1;
|
// required .hbase.pb.TableName table_name = 1;
|
||||||
public static final int TABLE_NAME_FIELD_NUMBER = 1;
|
public static final int TABLE_NAME_FIELD_NUMBER = 1;
|
||||||
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName tableName_;
|
private org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName tableName_;
|
||||||
/**
|
/**
|
||||||
* <code>required .hbase.pb.TableName table_name = 1;</code>
|
* <code>required .hbase.pb.TableName table_name = 1;</code>
|
||||||
*/
|
*/
|
||||||
@ -9006,13 +9006,13 @@ public final class WALProtos {
|
|||||||
/**
|
/**
|
||||||
* <code>required .hbase.pb.TableName table_name = 1;</code>
|
* <code>required .hbase.pb.TableName table_name = 1;</code>
|
||||||
*/
|
*/
|
||||||
public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName getTableName() {
|
public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName getTableName() {
|
||||||
return tableName_;
|
return tableName_;
|
||||||
}
|
}
|
||||||
/**
|
/**
|
||||||
* <code>required .hbase.pb.TableName table_name = 1;</code>
|
* <code>required .hbase.pb.TableName table_name = 1;</code>
|
||||||
*/
|
*/
|
||||||
public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder() {
|
public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder getTableNameOrBuilder() {
|
||||||
return tableName_;
|
return tableName_;
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -9085,7 +9085,7 @@ public final class WALProtos {
|
|||||||
}
|
}
|
||||||
|
|
||||||
private void initFields() {
|
private void initFields() {
|
||||||
tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance();
|
tableName_ = org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance();
|
||||||
encodedRegionName_ = com.google.protobuf.ByteString.EMPTY;
|
encodedRegionName_ = com.google.protobuf.ByteString.EMPTY;
|
||||||
stores_ = java.util.Collections.emptyList();
|
stores_ = java.util.Collections.emptyList();
|
||||||
bulkloadSeqNum_ = 0L;
|
bulkloadSeqNum_ = 0L;
|
||||||
@ -9347,7 +9347,7 @@ public final class WALProtos {
|
|||||||
public Builder clear() {
|
public Builder clear() {
|
||||||
super.clear();
|
super.clear();
|
||||||
if (tableNameBuilder_ == null) {
|
if (tableNameBuilder_ == null) {
|
||||||
tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance();
|
tableName_ = org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance();
|
||||||
} else {
|
} else {
|
||||||
tableNameBuilder_.clear();
|
tableNameBuilder_.clear();
|
||||||
}
|
}
|
||||||
@ -9516,9 +9516,9 @@ public final class WALProtos {
|
|||||||
private int bitField0_;
|
private int bitField0_;
|
||||||
|
|
||||||
// required .hbase.pb.TableName table_name = 1;
|
// required .hbase.pb.TableName table_name = 1;
|
||||||
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance();
|
private org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName tableName_ = org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance();
|
||||||
private com.google.protobuf.SingleFieldBuilder<
|
private com.google.protobuf.SingleFieldBuilder<
|
||||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder> tableNameBuilder_;
|
org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder> tableNameBuilder_;
|
||||||
/**
|
/**
|
||||||
* <code>required .hbase.pb.TableName table_name = 1;</code>
|
* <code>required .hbase.pb.TableName table_name = 1;</code>
|
||||||
*/
|
*/
|
||||||
@ -9528,7 +9528,7 @@ public final class WALProtos {
|
|||||||
/**
|
/**
|
||||||
* <code>required .hbase.pb.TableName table_name = 1;</code>
|
* <code>required .hbase.pb.TableName table_name = 1;</code>
|
||||||
*/
|
*/
|
||||||
public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName getTableName() {
|
public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName getTableName() {
|
||||||
if (tableNameBuilder_ == null) {
|
if (tableNameBuilder_ == null) {
|
||||||
return tableName_;
|
return tableName_;
|
||||||
} else {
|
} else {
|
||||||
@ -9538,7 +9538,7 @@ public final class WALProtos {
|
|||||||
/**
|
/**
|
||||||
* <code>required .hbase.pb.TableName table_name = 1;</code>
|
* <code>required .hbase.pb.TableName table_name = 1;</code>
|
||||||
*/
|
*/
|
||||||
public Builder setTableName(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName value) {
|
public Builder setTableName(org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName value) {
|
||||||
if (tableNameBuilder_ == null) {
|
if (tableNameBuilder_ == null) {
|
||||||
if (value == null) {
|
if (value == null) {
|
||||||
throw new NullPointerException();
|
throw new NullPointerException();
|
||||||
@ -9555,7 +9555,7 @@ public final class WALProtos {
|
|||||||
* <code>required .hbase.pb.TableName table_name = 1;</code>
|
* <code>required .hbase.pb.TableName table_name = 1;</code>
|
||||||
*/
|
*/
|
||||||
public Builder setTableName(
|
public Builder setTableName(
|
||||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder builderForValue) {
|
org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder builderForValue) {
|
||||||
if (tableNameBuilder_ == null) {
|
if (tableNameBuilder_ == null) {
|
||||||
tableName_ = builderForValue.build();
|
tableName_ = builderForValue.build();
|
||||||
onChanged();
|
onChanged();
|
||||||
@ -9568,12 +9568,12 @@ public final class WALProtos {
|
|||||||
/**
|
/**
|
||||||
* <code>required .hbase.pb.TableName table_name = 1;</code>
|
* <code>required .hbase.pb.TableName table_name = 1;</code>
|
||||||
*/
|
*/
|
||||||
public Builder mergeTableName(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName value) {
|
public Builder mergeTableName(org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName value) {
|
||||||
if (tableNameBuilder_ == null) {
|
if (tableNameBuilder_ == null) {
|
||||||
if (((bitField0_ & 0x00000001) == 0x00000001) &&
|
if (((bitField0_ & 0x00000001) == 0x00000001) &&
|
||||||
tableName_ != org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance()) {
|
tableName_ != org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance()) {
|
||||||
tableName_ =
|
tableName_ =
|
||||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.newBuilder(tableName_).mergeFrom(value).buildPartial();
|
org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.newBuilder(tableName_).mergeFrom(value).buildPartial();
|
||||||
} else {
|
} else {
|
||||||
tableName_ = value;
|
tableName_ = value;
|
||||||
}
|
}
|
||||||
@ -9589,7 +9589,7 @@ public final class WALProtos {
|
|||||||
*/
|
*/
|
||||||
public Builder clearTableName() {
|
public Builder clearTableName() {
|
||||||
if (tableNameBuilder_ == null) {
|
if (tableNameBuilder_ == null) {
|
||||||
tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance();
|
tableName_ = org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance();
|
||||||
onChanged();
|
onChanged();
|
||||||
} else {
|
} else {
|
||||||
tableNameBuilder_.clear();
|
tableNameBuilder_.clear();
|
||||||
@ -9600,7 +9600,7 @@ public final class WALProtos {
|
|||||||
/**
|
/**
|
||||||
* <code>required .hbase.pb.TableName table_name = 1;</code>
|
* <code>required .hbase.pb.TableName table_name = 1;</code>
|
||||||
*/
|
*/
|
||||||
public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder getTableNameBuilder() {
|
public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder getTableNameBuilder() {
|
||||||
bitField0_ |= 0x00000001;
|
bitField0_ |= 0x00000001;
|
||||||
onChanged();
|
onChanged();
|
||||||
return getTableNameFieldBuilder().getBuilder();
|
return getTableNameFieldBuilder().getBuilder();
|
||||||
@ -9608,7 +9608,7 @@ public final class WALProtos {
|
|||||||
/**
|
/**
|
||||||
* <code>required .hbase.pb.TableName table_name = 1;</code>
|
* <code>required .hbase.pb.TableName table_name = 1;</code>
|
||||||
*/
|
*/
|
||||||
public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder() {
|
public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder getTableNameOrBuilder() {
|
||||||
if (tableNameBuilder_ != null) {
|
if (tableNameBuilder_ != null) {
|
||||||
return tableNameBuilder_.getMessageOrBuilder();
|
return tableNameBuilder_.getMessageOrBuilder();
|
||||||
} else {
|
} else {
|
||||||
@ -9619,11 +9619,11 @@ public final class WALProtos {
|
|||||||
* <code>required .hbase.pb.TableName table_name = 1;</code>
|
* <code>required .hbase.pb.TableName table_name = 1;</code>
|
||||||
*/
|
*/
|
||||||
private com.google.protobuf.SingleFieldBuilder<
|
private com.google.protobuf.SingleFieldBuilder<
|
||||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder>
|
org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder>
|
||||||
getTableNameFieldBuilder() {
|
getTableNameFieldBuilder() {
|
||||||
if (tableNameBuilder_ == null) {
|
if (tableNameBuilder_ == null) {
|
||||||
tableNameBuilder_ = new com.google.protobuf.SingleFieldBuilder<
|
tableNameBuilder_ = new com.google.protobuf.SingleFieldBuilder<
|
||||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder>(
|
org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder>(
|
||||||
tableName_,
|
tableName_,
|
||||||
getParentForChildren(),
|
getParentForChildren(),
|
||||||
isClean());
|
isClean());
|
||||||
@ -11978,56 +11978,56 @@ public final class WALProtos {
|
|||||||
descriptor;
|
descriptor;
|
||||||
static {
|
static {
|
||||||
java.lang.String[] descriptorData = {
|
java.lang.String[] descriptorData = {
|
||||||
"\n\tWAL.proto\022\010hbase.pb\032\013HBase.proto\032\014Clie" +
|
"\n\tWAL.proto\022\010hbase.pb\032\013Table.proto\032\013HBas" +
|
||||||
"nt.proto\"\217\001\n\tWALHeader\022\027\n\017has_compressio" +
|
"e.proto\032\014Client.proto\"\217\001\n\tWALHeader\022\027\n\017h" +
|
||||||
"n\030\001 \001(\010\022\026\n\016encryption_key\030\002 \001(\014\022\033\n\023has_t" +
|
"as_compression\030\001 \001(\010\022\026\n\016encryption_key\030\002" +
|
||||||
"ag_compression\030\003 \001(\010\022\027\n\017writer_cls_name\030" +
|
" \001(\014\022\033\n\023has_tag_compression\030\003 \001(\010\022\027\n\017wri" +
|
||||||
"\004 \001(\t\022\033\n\023cell_codec_cls_name\030\005 \001(\t\"\273\002\n\006W" +
|
"ter_cls_name\030\004 \001(\t\022\033\n\023cell_codec_cls_nam" +
|
||||||
"ALKey\022\033\n\023encoded_region_name\030\001 \002(\014\022\022\n\nta" +
|
"e\030\005 \001(\t\"\273\002\n\006WALKey\022\033\n\023encoded_region_nam" +
|
||||||
"ble_name\030\002 \002(\014\022\033\n\023log_sequence_number\030\003 " +
|
"e\030\001 \002(\014\022\022\n\ntable_name\030\002 \002(\014\022\033\n\023log_seque" +
|
||||||
"\002(\004\022\022\n\nwrite_time\030\004 \002(\004\022&\n\ncluster_id\030\005 " +
|
"nce_number\030\003 \002(\004\022\022\n\nwrite_time\030\004 \002(\004\022&\n\n" +
|
||||||
"\001(\0132\016.hbase.pb.UUIDB\002\030\001\022%\n\006scopes\030\006 \003(\0132" +
|
"cluster_id\030\005 \001(\0132\016.hbase.pb.UUIDB\002\030\001\022%\n\006" +
|
||||||
"\025.hbase.pb.FamilyScope\022\032\n\022following_kv_c",
|
"scopes\030\006 \003(\0132\025.hbase.pb.FamilyScope\022\032\n\022f",
|
||||||
"ount\030\007 \001(\r\022#\n\013cluster_ids\030\010 \003(\0132\016.hbase." +
|
"ollowing_kv_count\030\007 \001(\r\022#\n\013cluster_ids\030\010" +
|
||||||
"pb.UUID\022\022\n\nnonceGroup\030\t \001(\004\022\r\n\005nonce\030\n \001" +
|
" \003(\0132\016.hbase.pb.UUID\022\022\n\nnonceGroup\030\t \001(\004" +
|
||||||
"(\004\022\034\n\024orig_sequence_number\030\013 \001(\004\"F\n\013Fami" +
|
"\022\r\n\005nonce\030\n \001(\004\022\034\n\024orig_sequence_number\030" +
|
||||||
"lyScope\022\016\n\006family\030\001 \002(\014\022\'\n\nscope_type\030\002 " +
|
"\013 \001(\004\"F\n\013FamilyScope\022\016\n\006family\030\001 \002(\014\022\'\n\n" +
|
||||||
"\002(\0162\023.hbase.pb.ScopeType\"\276\001\n\024CompactionD" +
|
"scope_type\030\002 \002(\0162\023.hbase.pb.ScopeType\"\276\001" +
|
||||||
"escriptor\022\022\n\ntable_name\030\001 \002(\014\022\033\n\023encoded" +
|
"\n\024CompactionDescriptor\022\022\n\ntable_name\030\001 \002" +
|
||||||
"_region_name\030\002 \002(\014\022\023\n\013family_name\030\003 \002(\014\022" +
|
"(\014\022\033\n\023encoded_region_name\030\002 \002(\014\022\023\n\013famil" +
|
||||||
"\030\n\020compaction_input\030\004 \003(\t\022\031\n\021compaction_" +
|
"y_name\030\003 \002(\014\022\030\n\020compaction_input\030\004 \003(\t\022\031" +
|
||||||
"output\030\005 \003(\t\022\026\n\016store_home_dir\030\006 \002(\t\022\023\n\013" +
|
"\n\021compaction_output\030\005 \003(\t\022\026\n\016store_home_" +
|
||||||
"region_name\030\007 \001(\014\"\244\003\n\017FlushDescriptor\0225\n",
|
"dir\030\006 \002(\t\022\023\n\013region_name\030\007 \001(\014\"\244\003\n\017Flush",
|
||||||
"\006action\030\001 \002(\0162%.hbase.pb.FlushDescriptor" +
|
"Descriptor\0225\n\006action\030\001 \002(\0162%.hbase.pb.Fl" +
|
||||||
".FlushAction\022\022\n\ntable_name\030\002 \002(\014\022\033\n\023enco" +
|
"ushDescriptor.FlushAction\022\022\n\ntable_name\030" +
|
||||||
"ded_region_name\030\003 \002(\014\022\035\n\025flush_sequence_" +
|
"\002 \002(\014\022\033\n\023encoded_region_name\030\003 \002(\014\022\035\n\025fl" +
|
||||||
"number\030\004 \001(\004\022E\n\rstore_flushes\030\005 \003(\0132..hb" +
|
"ush_sequence_number\030\004 \001(\004\022E\n\rstore_flush" +
|
||||||
"ase.pb.FlushDescriptor.StoreFlushDescrip" +
|
"es\030\005 \003(\0132..hbase.pb.FlushDescriptor.Stor" +
|
||||||
"tor\022\023\n\013region_name\030\006 \001(\014\032Y\n\024StoreFlushDe" +
|
"eFlushDescriptor\022\023\n\013region_name\030\006 \001(\014\032Y\n" +
|
||||||
"scriptor\022\023\n\013family_name\030\001 \002(\014\022\026\n\016store_h" +
|
"\024StoreFlushDescriptor\022\023\n\013family_name\030\001 \002" +
|
||||||
"ome_dir\030\002 \002(\t\022\024\n\014flush_output\030\003 \003(\t\"S\n\013F" +
|
"(\014\022\026\n\016store_home_dir\030\002 \002(\t\022\024\n\014flush_outp" +
|
||||||
"lushAction\022\017\n\013START_FLUSH\020\000\022\020\n\014COMMIT_FL" +
|
"ut\030\003 \003(\t\"S\n\013FlushAction\022\017\n\013START_FLUSH\020\000" +
|
||||||
"USH\020\001\022\017\n\013ABORT_FLUSH\020\002\022\020\n\014CANNOT_FLUSH\020\003",
|
"\022\020\n\014COMMIT_FLUSH\020\001\022\017\n\013ABORT_FLUSH\020\002\022\020\n\014C",
|
||||||
"\"q\n\017StoreDescriptor\022\023\n\013family_name\030\001 \002(\014" +
|
"ANNOT_FLUSH\020\003\"q\n\017StoreDescriptor\022\023\n\013fami" +
|
||||||
"\022\026\n\016store_home_dir\030\002 \002(\t\022\022\n\nstore_file\030\003" +
|
"ly_name\030\001 \002(\014\022\026\n\016store_home_dir\030\002 \002(\t\022\022\n" +
|
||||||
" \003(\t\022\035\n\025store_file_size_bytes\030\004 \001(\004\"\237\001\n\022" +
|
"\nstore_file\030\003 \003(\t\022\035\n\025store_file_size_byt" +
|
||||||
"BulkLoadDescriptor\022\'\n\ntable_name\030\001 \002(\0132\023" +
|
"es\030\004 \001(\004\"\237\001\n\022BulkLoadDescriptor\022\'\n\ntable" +
|
||||||
".hbase.pb.TableName\022\033\n\023encoded_region_na" +
|
"_name\030\001 \002(\0132\023.hbase.pb.TableName\022\033\n\023enco" +
|
||||||
"me\030\002 \002(\014\022)\n\006stores\030\003 \003(\0132\031.hbase.pb.Stor" +
|
"ded_region_name\030\002 \002(\014\022)\n\006stores\030\003 \003(\0132\031." +
|
||||||
"eDescriptor\022\030\n\020bulkload_seq_num\030\004 \002(\003\"\272\002" +
|
"hbase.pb.StoreDescriptor\022\030\n\020bulkload_seq" +
|
||||||
"\n\025RegionEventDescriptor\022=\n\nevent_type\030\001 " +
|
"_num\030\004 \002(\003\"\272\002\n\025RegionEventDescriptor\022=\n\n" +
|
||||||
"\002(\0162).hbase.pb.RegionEventDescriptor.Eve" +
|
"event_type\030\001 \002(\0162).hbase.pb.RegionEventD" +
|
||||||
"ntType\022\022\n\ntable_name\030\002 \002(\014\022\033\n\023encoded_re",
|
"escriptor.EventType\022\022\n\ntable_name\030\002 \002(\014\022",
|
||||||
"gion_name\030\003 \002(\014\022\033\n\023log_sequence_number\030\004" +
|
"\033\n\023encoded_region_name\030\003 \002(\014\022\033\n\023log_sequ" +
|
||||||
" \001(\004\022)\n\006stores\030\005 \003(\0132\031.hbase.pb.StoreDes" +
|
"ence_number\030\004 \001(\004\022)\n\006stores\030\005 \003(\0132\031.hbas" +
|
||||||
"criptor\022$\n\006server\030\006 \001(\0132\024.hbase.pb.Serve" +
|
"e.pb.StoreDescriptor\022$\n\006server\030\006 \001(\0132\024.h" +
|
||||||
"rName\022\023\n\013region_name\030\007 \001(\014\".\n\tEventType\022" +
|
"base.pb.ServerName\022\023\n\013region_name\030\007 \001(\014\"" +
|
||||||
"\017\n\013REGION_OPEN\020\000\022\020\n\014REGION_CLOSE\020\001\"\014\n\nWA" +
|
".\n\tEventType\022\017\n\013REGION_OPEN\020\000\022\020\n\014REGION_" +
|
||||||
"LTrailer*d\n\tScopeType\022\033\n\027REPLICATION_SCO" +
|
"CLOSE\020\001\"\014\n\nWALTrailer*d\n\tScopeType\022\033\n\027RE" +
|
||||||
"PE_LOCAL\020\000\022\034\n\030REPLICATION_SCOPE_GLOBAL\020\001" +
|
"PLICATION_SCOPE_LOCAL\020\000\022\034\n\030REPLICATION_S" +
|
||||||
"\022\034\n\030REPLICATION_SCOPE_SERIAL\020\002B?\n*org.ap" +
|
"COPE_GLOBAL\020\001\022\034\n\030REPLICATION_SCOPE_SERIA" +
|
||||||
"ache.hadoop.hbase.protobuf.generatedB\tWA" +
|
"L\020\002B?\n*org.apache.hadoop.hbase.protobuf." +
|
||||||
"LProtosH\001\210\001\000\240\001\001"
|
"generatedB\tWALProtosH\001\210\001\000\240\001\001"
|
||||||
};
|
};
|
||||||
com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
|
com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
|
||||||
new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() {
|
new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() {
|
||||||
@ -12100,6 +12100,7 @@ public final class WALProtos {
|
|||||||
com.google.protobuf.Descriptors.FileDescriptor
|
com.google.protobuf.Descriptors.FileDescriptor
|
||||||
.internalBuildGeneratedFileFrom(descriptorData,
|
.internalBuildGeneratedFileFrom(descriptorData,
|
||||||
new com.google.protobuf.Descriptors.FileDescriptor[] {
|
new com.google.protobuf.Descriptors.FileDescriptor[] {
|
||||||
|
org.apache.hadoop.hbase.protobuf.generated.TableProtos.getDescriptor(),
|
||||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.getDescriptor(),
|
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.getDescriptor(),
|
||||||
org.apache.hadoop.hbase.protobuf.generated.ClientProtos.getDescriptor(),
|
org.apache.hadoop.hbase.protobuf.generated.ClientProtos.getDescriptor(),
|
||||||
}, assigner);
|
}, assigner);
|
||||||
|
@ -5043,11 +5043,11 @@ public final class ZooKeeperProtos {
|
|||||||
/**
|
/**
|
||||||
* <code>optional .hbase.pb.TableName table_name = 1;</code>
|
* <code>optional .hbase.pb.TableName table_name = 1;</code>
|
||||||
*/
|
*/
|
||||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName getTableName();
|
org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName getTableName();
|
||||||
/**
|
/**
|
||||||
* <code>optional .hbase.pb.TableName table_name = 1;</code>
|
* <code>optional .hbase.pb.TableName table_name = 1;</code>
|
||||||
*/
|
*/
|
||||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder();
|
org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder getTableNameOrBuilder();
|
||||||
|
|
||||||
// repeated bytes families = 2;
|
// repeated bytes families = 2;
|
||||||
/**
|
/**
|
||||||
@ -5115,11 +5115,11 @@ public final class ZooKeeperProtos {
|
|||||||
break;
|
break;
|
||||||
}
|
}
|
||||||
case 10: {
|
case 10: {
|
||||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder subBuilder = null;
|
org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder subBuilder = null;
|
||||||
if (((bitField0_ & 0x00000001) == 0x00000001)) {
|
if (((bitField0_ & 0x00000001) == 0x00000001)) {
|
||||||
subBuilder = tableName_.toBuilder();
|
subBuilder = tableName_.toBuilder();
|
||||||
}
|
}
|
||||||
tableName_ = input.readMessage(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.PARSER, extensionRegistry);
|
tableName_ = input.readMessage(org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.PARSER, extensionRegistry);
|
||||||
if (subBuilder != null) {
|
if (subBuilder != null) {
|
||||||
subBuilder.mergeFrom(tableName_);
|
subBuilder.mergeFrom(tableName_);
|
||||||
tableName_ = subBuilder.buildPartial();
|
tableName_ = subBuilder.buildPartial();
|
||||||
@ -5180,7 +5180,7 @@ public final class ZooKeeperProtos {
|
|||||||
private int bitField0_;
|
private int bitField0_;
|
||||||
// optional .hbase.pb.TableName table_name = 1;
|
// optional .hbase.pb.TableName table_name = 1;
|
||||||
public static final int TABLE_NAME_FIELD_NUMBER = 1;
|
public static final int TABLE_NAME_FIELD_NUMBER = 1;
|
||||||
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName tableName_;
|
private org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName tableName_;
|
||||||
/**
|
/**
|
||||||
* <code>optional .hbase.pb.TableName table_name = 1;</code>
|
* <code>optional .hbase.pb.TableName table_name = 1;</code>
|
||||||
*/
|
*/
|
||||||
@ -5190,13 +5190,13 @@ public final class ZooKeeperProtos {
|
|||||||
/**
|
/**
|
||||||
* <code>optional .hbase.pb.TableName table_name = 1;</code>
|
* <code>optional .hbase.pb.TableName table_name = 1;</code>
|
||||||
*/
|
*/
|
||||||
public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName getTableName() {
|
public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName getTableName() {
|
||||||
return tableName_;
|
return tableName_;
|
||||||
}
|
}
|
||||||
/**
|
/**
|
||||||
* <code>optional .hbase.pb.TableName table_name = 1;</code>
|
* <code>optional .hbase.pb.TableName table_name = 1;</code>
|
||||||
*/
|
*/
|
||||||
public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder() {
|
public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder getTableNameOrBuilder() {
|
||||||
return tableName_;
|
return tableName_;
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -5224,7 +5224,7 @@ public final class ZooKeeperProtos {
|
|||||||
}
|
}
|
||||||
|
|
||||||
private void initFields() {
|
private void initFields() {
|
||||||
tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance();
|
tableName_ = org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance();
|
||||||
families_ = java.util.Collections.emptyList();
|
families_ = java.util.Collections.emptyList();
|
||||||
}
|
}
|
||||||
private byte memoizedIsInitialized = -1;
|
private byte memoizedIsInitialized = -1;
|
||||||
@ -5435,7 +5435,7 @@ public final class ZooKeeperProtos {
|
|||||||
public Builder clear() {
|
public Builder clear() {
|
||||||
super.clear();
|
super.clear();
|
||||||
if (tableNameBuilder_ == null) {
|
if (tableNameBuilder_ == null) {
|
||||||
tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance();
|
tableName_ = org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance();
|
||||||
} else {
|
} else {
|
||||||
tableNameBuilder_.clear();
|
tableNameBuilder_.clear();
|
||||||
}
|
}
|
||||||
@ -5546,9 +5546,9 @@ public final class ZooKeeperProtos {
|
|||||||
private int bitField0_;
|
private int bitField0_;
|
||||||
|
|
||||||
// optional .hbase.pb.TableName table_name = 1;
|
// optional .hbase.pb.TableName table_name = 1;
|
||||||
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance();
|
private org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName tableName_ = org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance();
|
||||||
private com.google.protobuf.SingleFieldBuilder<
|
private com.google.protobuf.SingleFieldBuilder<
|
||||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder> tableNameBuilder_;
|
org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder> tableNameBuilder_;
|
||||||
/**
|
/**
|
||||||
* <code>optional .hbase.pb.TableName table_name = 1;</code>
|
* <code>optional .hbase.pb.TableName table_name = 1;</code>
|
||||||
*/
|
*/
|
||||||
@ -5558,7 +5558,7 @@ public final class ZooKeeperProtos {
|
|||||||
/**
|
/**
|
||||||
* <code>optional .hbase.pb.TableName table_name = 1;</code>
|
* <code>optional .hbase.pb.TableName table_name = 1;</code>
|
||||||
*/
|
*/
|
||||||
public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName getTableName() {
|
public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName getTableName() {
|
||||||
if (tableNameBuilder_ == null) {
|
if (tableNameBuilder_ == null) {
|
||||||
return tableName_;
|
return tableName_;
|
||||||
} else {
|
} else {
|
||||||
@ -5568,7 +5568,7 @@ public final class ZooKeeperProtos {
|
|||||||
/**
|
/**
|
||||||
* <code>optional .hbase.pb.TableName table_name = 1;</code>
|
* <code>optional .hbase.pb.TableName table_name = 1;</code>
|
||||||
*/
|
*/
|
||||||
public Builder setTableName(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName value) {
|
public Builder setTableName(org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName value) {
|
||||||
if (tableNameBuilder_ == null) {
|
if (tableNameBuilder_ == null) {
|
||||||
if (value == null) {
|
if (value == null) {
|
||||||
throw new NullPointerException();
|
throw new NullPointerException();
|
||||||
@ -5585,7 +5585,7 @@ public final class ZooKeeperProtos {
|
|||||||
* <code>optional .hbase.pb.TableName table_name = 1;</code>
|
* <code>optional .hbase.pb.TableName table_name = 1;</code>
|
||||||
*/
|
*/
|
||||||
public Builder setTableName(
|
public Builder setTableName(
|
||||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder builderForValue) {
|
org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder builderForValue) {
|
||||||
if (tableNameBuilder_ == null) {
|
if (tableNameBuilder_ == null) {
|
||||||
tableName_ = builderForValue.build();
|
tableName_ = builderForValue.build();
|
||||||
onChanged();
|
onChanged();
|
||||||
@ -5598,12 +5598,12 @@ public final class ZooKeeperProtos {
|
|||||||
/**
|
/**
|
||||||
* <code>optional .hbase.pb.TableName table_name = 1;</code>
|
* <code>optional .hbase.pb.TableName table_name = 1;</code>
|
||||||
*/
|
*/
|
||||||
public Builder mergeTableName(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName value) {
|
public Builder mergeTableName(org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName value) {
|
||||||
if (tableNameBuilder_ == null) {
|
if (tableNameBuilder_ == null) {
|
||||||
if (((bitField0_ & 0x00000001) == 0x00000001) &&
|
if (((bitField0_ & 0x00000001) == 0x00000001) &&
|
||||||
tableName_ != org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance()) {
|
tableName_ != org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance()) {
|
||||||
tableName_ =
|
tableName_ =
|
||||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.newBuilder(tableName_).mergeFrom(value).buildPartial();
|
org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.newBuilder(tableName_).mergeFrom(value).buildPartial();
|
||||||
} else {
|
} else {
|
||||||
tableName_ = value;
|
tableName_ = value;
|
||||||
}
|
}
|
||||||
@ -5619,7 +5619,7 @@ public final class ZooKeeperProtos {
|
|||||||
*/
|
*/
|
||||||
public Builder clearTableName() {
|
public Builder clearTableName() {
|
||||||
if (tableNameBuilder_ == null) {
|
if (tableNameBuilder_ == null) {
|
||||||
tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance();
|
tableName_ = org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance();
|
||||||
onChanged();
|
onChanged();
|
||||||
} else {
|
} else {
|
||||||
tableNameBuilder_.clear();
|
tableNameBuilder_.clear();
|
||||||
@ -5630,7 +5630,7 @@ public final class ZooKeeperProtos {
|
|||||||
/**
|
/**
|
||||||
* <code>optional .hbase.pb.TableName table_name = 1;</code>
|
* <code>optional .hbase.pb.TableName table_name = 1;</code>
|
||||||
*/
|
*/
|
||||||
public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder getTableNameBuilder() {
|
public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder getTableNameBuilder() {
|
||||||
bitField0_ |= 0x00000001;
|
bitField0_ |= 0x00000001;
|
||||||
onChanged();
|
onChanged();
|
||||||
return getTableNameFieldBuilder().getBuilder();
|
return getTableNameFieldBuilder().getBuilder();
|
||||||
@ -5638,7 +5638,7 @@ public final class ZooKeeperProtos {
|
|||||||
/**
|
/**
|
||||||
* <code>optional .hbase.pb.TableName table_name = 1;</code>
|
* <code>optional .hbase.pb.TableName table_name = 1;</code>
|
||||||
*/
|
*/
|
||||||
public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder() {
|
public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder getTableNameOrBuilder() {
|
||||||
if (tableNameBuilder_ != null) {
|
if (tableNameBuilder_ != null) {
|
||||||
return tableNameBuilder_.getMessageOrBuilder();
|
return tableNameBuilder_.getMessageOrBuilder();
|
||||||
} else {
|
} else {
|
||||||
@ -5649,11 +5649,11 @@ public final class ZooKeeperProtos {
|
|||||||
* <code>optional .hbase.pb.TableName table_name = 1;</code>
|
* <code>optional .hbase.pb.TableName table_name = 1;</code>
|
||||||
*/
|
*/
|
||||||
private com.google.protobuf.SingleFieldBuilder<
|
private com.google.protobuf.SingleFieldBuilder<
|
||||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder>
|
org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder>
|
||||||
getTableNameFieldBuilder() {
|
getTableNameFieldBuilder() {
|
||||||
if (tableNameBuilder_ == null) {
|
if (tableNameBuilder_ == null) {
|
||||||
tableNameBuilder_ = new com.google.protobuf.SingleFieldBuilder<
|
tableNameBuilder_ = new com.google.protobuf.SingleFieldBuilder<
|
||||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder>(
|
org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder>(
|
||||||
tableName_,
|
tableName_,
|
||||||
getParentForChildren(),
|
getParentForChildren(),
|
||||||
isClean());
|
isClean());
|
||||||
@ -9260,11 +9260,11 @@ public final class ZooKeeperProtos {
|
|||||||
/**
|
/**
|
||||||
* <code>optional .hbase.pb.TableName table_name = 1;</code>
|
* <code>optional .hbase.pb.TableName table_name = 1;</code>
|
||||||
*/
|
*/
|
||||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName getTableName();
|
org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName getTableName();
|
||||||
/**
|
/**
|
||||||
* <code>optional .hbase.pb.TableName table_name = 1;</code>
|
* <code>optional .hbase.pb.TableName table_name = 1;</code>
|
||||||
*/
|
*/
|
||||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder();
|
org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder getTableNameOrBuilder();
|
||||||
|
|
||||||
// optional .hbase.pb.ServerName lock_owner = 2;
|
// optional .hbase.pb.ServerName lock_owner = 2;
|
||||||
/**
|
/**
|
||||||
@ -9382,11 +9382,11 @@ public final class ZooKeeperProtos {
|
|||||||
break;
|
break;
|
||||||
}
|
}
|
||||||
case 10: {
|
case 10: {
|
||||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder subBuilder = null;
|
org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder subBuilder = null;
|
||||||
if (((bitField0_ & 0x00000001) == 0x00000001)) {
|
if (((bitField0_ & 0x00000001) == 0x00000001)) {
|
||||||
subBuilder = tableName_.toBuilder();
|
subBuilder = tableName_.toBuilder();
|
||||||
}
|
}
|
||||||
tableName_ = input.readMessage(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.PARSER, extensionRegistry);
|
tableName_ = input.readMessage(org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.PARSER, extensionRegistry);
|
||||||
if (subBuilder != null) {
|
if (subBuilder != null) {
|
||||||
subBuilder.mergeFrom(tableName_);
|
subBuilder.mergeFrom(tableName_);
|
||||||
tableName_ = subBuilder.buildPartial();
|
tableName_ = subBuilder.buildPartial();
|
||||||
@ -9469,7 +9469,7 @@ public final class ZooKeeperProtos {
|
|||||||
private int bitField0_;
|
private int bitField0_;
|
||||||
// optional .hbase.pb.TableName table_name = 1;
|
// optional .hbase.pb.TableName table_name = 1;
|
||||||
public static final int TABLE_NAME_FIELD_NUMBER = 1;
|
public static final int TABLE_NAME_FIELD_NUMBER = 1;
|
||||||
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName tableName_;
|
private org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName tableName_;
|
||||||
/**
|
/**
|
||||||
* <code>optional .hbase.pb.TableName table_name = 1;</code>
|
* <code>optional .hbase.pb.TableName table_name = 1;</code>
|
||||||
*/
|
*/
|
||||||
@ -9479,13 +9479,13 @@ public final class ZooKeeperProtos {
|
|||||||
/**
|
/**
|
||||||
* <code>optional .hbase.pb.TableName table_name = 1;</code>
|
* <code>optional .hbase.pb.TableName table_name = 1;</code>
|
||||||
*/
|
*/
|
||||||
public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName getTableName() {
|
public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName getTableName() {
|
||||||
return tableName_;
|
return tableName_;
|
||||||
}
|
}
|
||||||
/**
|
/**
|
||||||
* <code>optional .hbase.pb.TableName table_name = 1;</code>
|
* <code>optional .hbase.pb.TableName table_name = 1;</code>
|
||||||
*/
|
*/
|
||||||
public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder() {
|
public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder getTableNameOrBuilder() {
|
||||||
return tableName_;
|
return tableName_;
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -9603,7 +9603,7 @@ public final class ZooKeeperProtos {
|
|||||||
}
|
}
|
||||||
|
|
||||||
private void initFields() {
|
private void initFields() {
|
||||||
tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance();
|
tableName_ = org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance();
|
||||||
lockOwner_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.getDefaultInstance();
|
lockOwner_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.getDefaultInstance();
|
||||||
threadId_ = 0L;
|
threadId_ = 0L;
|
||||||
isShared_ = false;
|
isShared_ = false;
|
||||||
@ -9892,7 +9892,7 @@ public final class ZooKeeperProtos {
|
|||||||
public Builder clear() {
|
public Builder clear() {
|
||||||
super.clear();
|
super.clear();
|
||||||
if (tableNameBuilder_ == null) {
|
if (tableNameBuilder_ == null) {
|
||||||
tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance();
|
tableName_ = org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance();
|
||||||
} else {
|
} else {
|
||||||
tableNameBuilder_.clear();
|
tableNameBuilder_.clear();
|
||||||
}
|
}
|
||||||
@ -10047,9 +10047,9 @@ public final class ZooKeeperProtos {
|
|||||||
private int bitField0_;
|
private int bitField0_;
|
||||||
|
|
||||||
// optional .hbase.pb.TableName table_name = 1;
|
// optional .hbase.pb.TableName table_name = 1;
|
||||||
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance();
|
private org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName tableName_ = org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance();
|
||||||
private com.google.protobuf.SingleFieldBuilder<
|
private com.google.protobuf.SingleFieldBuilder<
|
||||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder> tableNameBuilder_;
|
org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder> tableNameBuilder_;
|
||||||
/**
|
/**
|
||||||
* <code>optional .hbase.pb.TableName table_name = 1;</code>
|
* <code>optional .hbase.pb.TableName table_name = 1;</code>
|
||||||
*/
|
*/
|
||||||
@ -10059,7 +10059,7 @@ public final class ZooKeeperProtos {
|
|||||||
/**
|
/**
|
||||||
* <code>optional .hbase.pb.TableName table_name = 1;</code>
|
* <code>optional .hbase.pb.TableName table_name = 1;</code>
|
||||||
*/
|
*/
|
||||||
public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName getTableName() {
|
public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName getTableName() {
|
||||||
if (tableNameBuilder_ == null) {
|
if (tableNameBuilder_ == null) {
|
||||||
return tableName_;
|
return tableName_;
|
||||||
} else {
|
} else {
|
||||||
@ -10069,7 +10069,7 @@ public final class ZooKeeperProtos {
|
|||||||
/**
|
/**
|
||||||
* <code>optional .hbase.pb.TableName table_name = 1;</code>
|
* <code>optional .hbase.pb.TableName table_name = 1;</code>
|
||||||
*/
|
*/
|
||||||
public Builder setTableName(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName value) {
|
public Builder setTableName(org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName value) {
|
||||||
if (tableNameBuilder_ == null) {
|
if (tableNameBuilder_ == null) {
|
||||||
if (value == null) {
|
if (value == null) {
|
||||||
throw new NullPointerException();
|
throw new NullPointerException();
|
||||||
@ -10086,7 +10086,7 @@ public final class ZooKeeperProtos {
|
|||||||
* <code>optional .hbase.pb.TableName table_name = 1;</code>
|
* <code>optional .hbase.pb.TableName table_name = 1;</code>
|
||||||
*/
|
*/
|
||||||
public Builder setTableName(
|
public Builder setTableName(
|
||||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder builderForValue) {
|
org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder builderForValue) {
|
||||||
if (tableNameBuilder_ == null) {
|
if (tableNameBuilder_ == null) {
|
||||||
tableName_ = builderForValue.build();
|
tableName_ = builderForValue.build();
|
||||||
onChanged();
|
onChanged();
|
||||||
@ -10099,12 +10099,12 @@ public final class ZooKeeperProtos {
|
|||||||
/**
|
/**
|
||||||
* <code>optional .hbase.pb.TableName table_name = 1;</code>
|
* <code>optional .hbase.pb.TableName table_name = 1;</code>
|
||||||
*/
|
*/
|
||||||
public Builder mergeTableName(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName value) {
|
public Builder mergeTableName(org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName value) {
|
||||||
if (tableNameBuilder_ == null) {
|
if (tableNameBuilder_ == null) {
|
||||||
if (((bitField0_ & 0x00000001) == 0x00000001) &&
|
if (((bitField0_ & 0x00000001) == 0x00000001) &&
|
||||||
tableName_ != org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance()) {
|
tableName_ != org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance()) {
|
||||||
tableName_ =
|
tableName_ =
|
||||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.newBuilder(tableName_).mergeFrom(value).buildPartial();
|
org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.newBuilder(tableName_).mergeFrom(value).buildPartial();
|
||||||
} else {
|
} else {
|
||||||
tableName_ = value;
|
tableName_ = value;
|
||||||
}
|
}
|
||||||
@ -10120,7 +10120,7 @@ public final class ZooKeeperProtos {
|
|||||||
*/
|
*/
|
||||||
public Builder clearTableName() {
|
public Builder clearTableName() {
|
||||||
if (tableNameBuilder_ == null) {
|
if (tableNameBuilder_ == null) {
|
||||||
tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance();
|
tableName_ = org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance();
|
||||||
onChanged();
|
onChanged();
|
||||||
} else {
|
} else {
|
||||||
tableNameBuilder_.clear();
|
tableNameBuilder_.clear();
|
||||||
@ -10131,7 +10131,7 @@ public final class ZooKeeperProtos {
|
|||||||
/**
|
/**
|
||||||
* <code>optional .hbase.pb.TableName table_name = 1;</code>
|
* <code>optional .hbase.pb.TableName table_name = 1;</code>
|
||||||
*/
|
*/
|
||||||
public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder getTableNameBuilder() {
|
public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder getTableNameBuilder() {
|
||||||
bitField0_ |= 0x00000001;
|
bitField0_ |= 0x00000001;
|
||||||
onChanged();
|
onChanged();
|
||||||
return getTableNameFieldBuilder().getBuilder();
|
return getTableNameFieldBuilder().getBuilder();
|
||||||
@ -10139,7 +10139,7 @@ public final class ZooKeeperProtos {
|
|||||||
/**
|
/**
|
||||||
* <code>optional .hbase.pb.TableName table_name = 1;</code>
|
* <code>optional .hbase.pb.TableName table_name = 1;</code>
|
||||||
*/
|
*/
|
||||||
public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder() {
|
public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder getTableNameOrBuilder() {
|
||||||
if (tableNameBuilder_ != null) {
|
if (tableNameBuilder_ != null) {
|
||||||
return tableNameBuilder_.getMessageOrBuilder();
|
return tableNameBuilder_.getMessageOrBuilder();
|
||||||
} else {
|
} else {
|
||||||
@ -10150,11 +10150,11 @@ public final class ZooKeeperProtos {
|
|||||||
* <code>optional .hbase.pb.TableName table_name = 1;</code>
|
* <code>optional .hbase.pb.TableName table_name = 1;</code>
|
||||||
*/
|
*/
|
||||||
private com.google.protobuf.SingleFieldBuilder<
|
private com.google.protobuf.SingleFieldBuilder<
|
||||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder>
|
org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder>
|
||||||
getTableNameFieldBuilder() {
|
getTableNameFieldBuilder() {
|
||||||
if (tableNameBuilder_ == null) {
|
if (tableNameBuilder_ == null) {
|
||||||
tableNameBuilder_ = new com.google.protobuf.SingleFieldBuilder<
|
tableNameBuilder_ = new com.google.protobuf.SingleFieldBuilder<
|
||||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder>(
|
org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder>(
|
||||||
tableName_,
|
tableName_,
|
||||||
getParentForChildren(),
|
getParentForChildren(),
|
||||||
isClean());
|
isClean());
|
||||||
@ -10982,47 +10982,47 @@ public final class ZooKeeperProtos {
|
|||||||
descriptor;
|
descriptor;
|
||||||
static {
|
static {
|
||||||
java.lang.String[] descriptorData = {
|
java.lang.String[] descriptorData = {
|
||||||
"\n\017ZooKeeper.proto\022\010hbase.pb\032\013HBase.proto" +
|
"\n\017ZooKeeper.proto\022\010hbase.pb\032\013Table.proto" +
|
||||||
"\032\023ClusterStatus.proto\"y\n\020MetaRegionServe" +
|
"\032\013HBase.proto\032\023ClusterStatus.proto\"y\n\020Me" +
|
||||||
"r\022$\n\006server\030\001 \002(\0132\024.hbase.pb.ServerName\022" +
|
"taRegionServer\022$\n\006server\030\001 \002(\0132\024.hbase.p" +
|
||||||
"\023\n\013rpc_version\030\002 \001(\r\022*\n\005state\030\003 \001(\0162\033.hb" +
|
"b.ServerName\022\023\n\013rpc_version\030\002 \001(\r\022*\n\005sta" +
|
||||||
"ase.pb.RegionState.State\"V\n\006Master\022$\n\006ma" +
|
"te\030\003 \001(\0162\033.hbase.pb.RegionState.State\"V\n" +
|
||||||
"ster\030\001 \002(\0132\024.hbase.pb.ServerName\022\023\n\013rpc_" +
|
"\006Master\022$\n\006master\030\001 \002(\0132\024.hbase.pb.Serve" +
|
||||||
"version\030\002 \001(\r\022\021\n\tinfo_port\030\003 \001(\r\"\037\n\tClus" +
|
"rName\022\023\n\013rpc_version\030\002 \001(\r\022\021\n\tinfo_port\030" +
|
||||||
"terUp\022\022\n\nstart_date\030\001 \002(\t\"\221\001\n\020RegionTran" +
|
"\003 \001(\r\"\037\n\tClusterUp\022\022\n\nstart_date\030\001 \002(\t\"\221" +
|
||||||
"sition\022\027\n\017event_type_code\030\001 \002(\r\022\023\n\013regio" +
|
"\001\n\020RegionTransition\022\027\n\017event_type_code\030\001" +
|
||||||
"n_name\030\002 \002(\014\022\023\n\013create_time\030\003 \002(\004\022)\n\013ser",
|
" \002(\r\022\023\n\013region_name\030\002 \002(\014\022\023\n\013create_time",
|
||||||
"ver_name\030\004 \002(\0132\024.hbase.pb.ServerName\022\017\n\007" +
|
"\030\003 \002(\004\022)\n\013server_name\030\004 \002(\0132\024.hbase.pb.S" +
|
||||||
"payload\030\005 \001(\014\"\247\002\n\014SplitLogTask\022+\n\005state\030" +
|
"erverName\022\017\n\007payload\030\005 \001(\014\"\247\002\n\014SplitLogT" +
|
||||||
"\001 \002(\0162\034.hbase.pb.SplitLogTask.State\022)\n\013s" +
|
"ask\022+\n\005state\030\001 \002(\0162\034.hbase.pb.SplitLogTa" +
|
||||||
"erver_name\030\002 \002(\0132\024.hbase.pb.ServerName\022:" +
|
"sk.State\022)\n\013server_name\030\002 \002(\0132\024.hbase.pb" +
|
||||||
"\n\004mode\030\003 \001(\0162#.hbase.pb.SplitLogTask.Rec" +
|
".ServerName\022:\n\004mode\030\003 \001(\0162#.hbase.pb.Spl" +
|
||||||
"overyMode:\007UNKNOWN\"C\n\005State\022\016\n\nUNASSIGNE" +
|
"itLogTask.RecoveryMode:\007UNKNOWN\"C\n\005State" +
|
||||||
"D\020\000\022\t\n\005OWNED\020\001\022\014\n\010RESIGNED\020\002\022\010\n\004DONE\020\003\022\007" +
|
"\022\016\n\nUNASSIGNED\020\000\022\t\n\005OWNED\020\001\022\014\n\010RESIGNED\020" +
|
||||||
"\n\003ERR\020\004\">\n\014RecoveryMode\022\013\n\007UNKNOWN\020\000\022\021\n\r" +
|
"\002\022\010\n\004DONE\020\003\022\007\n\003ERR\020\004\">\n\014RecoveryMode\022\013\n\007" +
|
||||||
"LOG_SPLITTING\020\001\022\016\n\nLOG_REPLAY\020\002\"w\n\005Table" +
|
"UNKNOWN\020\000\022\021\n\rLOG_SPLITTING\020\001\022\016\n\nLOG_REPL" +
|
||||||
"\022-\n\005state\030\001 \002(\0162\025.hbase.pb.Table.State:\007",
|
"AY\020\002\"w\n\005Table\022-\n\005state\030\001 \002(\0162\025.hbase.pb.",
|
||||||
"ENABLED\"?\n\005State\022\013\n\007ENABLED\020\000\022\014\n\010DISABLE" +
|
"Table.State:\007ENABLED\"?\n\005State\022\013\n\007ENABLED" +
|
||||||
"D\020\001\022\r\n\tDISABLING\020\002\022\014\n\010ENABLING\020\003\"D\n\007Tabl" +
|
"\020\000\022\014\n\010DISABLED\020\001\022\r\n\tDISABLING\020\002\022\014\n\010ENABL" +
|
||||||
"eCF\022\'\n\ntable_name\030\001 \001(\0132\023.hbase.pb.Table" +
|
"ING\020\003\"D\n\007TableCF\022\'\n\ntable_name\030\001 \001(\0132\023.h" +
|
||||||
"Name\022\020\n\010families\030\002 \003(\014\"\330\001\n\017ReplicationPe" +
|
"base.pb.TableName\022\020\n\010families\030\002 \003(\014\"\330\001\n\017" +
|
||||||
"er\022\022\n\nclusterkey\030\001 \002(\t\022\037\n\027replicationEnd" +
|
"ReplicationPeer\022\022\n\nclusterkey\030\001 \002(\t\022\037\n\027r" +
|
||||||
"pointImpl\030\002 \001(\t\022&\n\004data\030\003 \003(\0132\030.hbase.pb" +
|
"eplicationEndpointImpl\030\002 \001(\t\022&\n\004data\030\003 \003" +
|
||||||
".BytesBytesPair\022/\n\rconfiguration\030\004 \003(\0132\030" +
|
"(\0132\030.hbase.pb.BytesBytesPair\022/\n\rconfigur" +
|
||||||
".hbase.pb.NameStringPair\022$\n\ttable_cfs\030\005 " +
|
"ation\030\004 \003(\0132\030.hbase.pb.NameStringPair\022$\n" +
|
||||||
"\003(\0132\021.hbase.pb.TableCF\022\021\n\tbandwidth\030\006 \001(" +
|
"\ttable_cfs\030\005 \003(\0132\021.hbase.pb.TableCF\022\021\n\tb" +
|
||||||
"\003\"g\n\020ReplicationState\022/\n\005state\030\001 \002(\0162 .h",
|
"andwidth\030\006 \001(\003\"g\n\020ReplicationState\022/\n\005st",
|
||||||
"base.pb.ReplicationState.State\"\"\n\005State\022" +
|
"ate\030\001 \002(\0162 .hbase.pb.ReplicationState.St" +
|
||||||
"\013\n\007ENABLED\020\000\022\014\n\010DISABLED\020\001\"+\n\027Replicatio" +
|
"ate\"\"\n\005State\022\013\n\007ENABLED\020\000\022\014\n\010DISABLED\020\001\"" +
|
||||||
"nHLogPosition\022\020\n\010position\030\001 \002(\003\"%\n\017Repli" +
|
"+\n\027ReplicationHLogPosition\022\020\n\010position\030\001" +
|
||||||
"cationLock\022\022\n\nlock_owner\030\001 \002(\t\"\252\001\n\tTable" +
|
" \002(\003\"%\n\017ReplicationLock\022\022\n\nlock_owner\030\001 " +
|
||||||
"Lock\022\'\n\ntable_name\030\001 \001(\0132\023.hbase.pb.Tabl" +
|
"\002(\t\"\252\001\n\tTableLock\022\'\n\ntable_name\030\001 \001(\0132\023." +
|
||||||
"eName\022(\n\nlock_owner\030\002 \001(\0132\024.hbase.pb.Ser" +
|
"hbase.pb.TableName\022(\n\nlock_owner\030\002 \001(\0132\024" +
|
||||||
"verName\022\021\n\tthread_id\030\003 \001(\003\022\021\n\tis_shared\030" +
|
".hbase.pb.ServerName\022\021\n\tthread_id\030\003 \001(\003\022" +
|
||||||
"\004 \001(\010\022\017\n\007purpose\030\005 \001(\t\022\023\n\013create_time\030\006 " +
|
"\021\n\tis_shared\030\004 \001(\010\022\017\n\007purpose\030\005 \001(\t\022\023\n\013c" +
|
||||||
"\001(\003\"\036\n\013SwitchState\022\017\n\007enabled\030\001 \001(\010BE\n*o" +
|
"reate_time\030\006 \001(\003\"\036\n\013SwitchState\022\017\n\007enabl" +
|
||||||
"rg.apache.hadoop.hbase.protobuf.generate",
|
"ed\030\001 \001(\010BE\n*org.apache.hadoop.hbase.prot",
|
||||||
"dB\017ZooKeeperProtosH\001\210\001\001\240\001\001"
|
"obuf.generatedB\017ZooKeeperProtosH\001\210\001\001\240\001\001"
|
||||||
};
|
};
|
||||||
com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
|
com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
|
||||||
new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() {
|
new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() {
|
||||||
@ -11113,6 +11113,7 @@ public final class ZooKeeperProtos {
|
|||||||
com.google.protobuf.Descriptors.FileDescriptor
|
com.google.protobuf.Descriptors.FileDescriptor
|
||||||
.internalBuildGeneratedFileFrom(descriptorData,
|
.internalBuildGeneratedFileFrom(descriptorData,
|
||||||
new com.google.protobuf.Descriptors.FileDescriptor[] {
|
new com.google.protobuf.Descriptors.FileDescriptor[] {
|
||||||
|
org.apache.hadoop.hbase.protobuf.generated.TableProtos.getDescriptor(),
|
||||||
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.getDescriptor(),
|
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.getDescriptor(),
|
||||||
org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.getDescriptor(),
|
org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.getDescriptor(),
|
||||||
}, assigner);
|
}, assigner);
|
||||||
|
@ -23,7 +23,7 @@ option java_generic_services = true;
|
|||||||
option java_generate_equals_and_hash = true;
|
option java_generate_equals_and_hash = true;
|
||||||
option optimize_for = SPEED;
|
option optimize_for = SPEED;
|
||||||
|
|
||||||
import "HBase.proto";
|
import "Table.proto";
|
||||||
|
|
||||||
message Permission {
|
message Permission {
|
||||||
enum Action {
|
enum Action {
|
||||||
|
@ -25,14 +25,8 @@ option java_generate_equals_and_hash = true;
|
|||||||
option optimize_for = SPEED;
|
option optimize_for = SPEED;
|
||||||
|
|
||||||
import "Cell.proto";
|
import "Cell.proto";
|
||||||
|
import "Table.proto";
|
||||||
/**
|
import "AccessControl.proto";
|
||||||
* Table Name
|
|
||||||
*/
|
|
||||||
message TableName {
|
|
||||||
required bytes namespace = 1;
|
|
||||||
required bytes qualifier = 2;
|
|
||||||
}
|
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Table Schema
|
* Table Schema
|
||||||
@ -221,3 +215,21 @@ message RegionServerInfo {
|
|||||||
optional int32 infoPort = 1;
|
optional int32 infoPort = 1;
|
||||||
optional VersionInfo version_info = 2;
|
optional VersionInfo version_info = 2;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Description of the snapshot to take
|
||||||
|
*/
|
||||||
|
message SnapshotDescription {
|
||||||
|
required string name = 1;
|
||||||
|
optional string table = 2; // not needed for delete, but checked for in taking snapshot
|
||||||
|
optional int64 creation_time = 3 [default = 0];
|
||||||
|
enum Type {
|
||||||
|
DISABLED = 0;
|
||||||
|
FLUSH = 1;
|
||||||
|
SKIPFLUSH = 2;
|
||||||
|
}
|
||||||
|
optional Type type = 4 [default = FLUSH];
|
||||||
|
optional int32 version = 5;
|
||||||
|
optional string owner = 6;
|
||||||
|
optional UsersAndPermissions users_and_permissions = 7;
|
||||||
|
}
|
||||||
|
@ -26,6 +26,7 @@ option java_generic_services = true;
|
|||||||
option java_generate_equals_and_hash = true;
|
option java_generate_equals_and_hash = true;
|
||||||
option optimize_for = SPEED;
|
option optimize_for = SPEED;
|
||||||
|
|
||||||
|
import "Table.proto";
|
||||||
import "HBase.proto";
|
import "HBase.proto";
|
||||||
import "Client.proto";
|
import "Client.proto";
|
||||||
import "ClusterStatus.proto";
|
import "ClusterStatus.proto";
|
||||||
|
@ -23,6 +23,7 @@ option java_generic_services = true;
|
|||||||
option java_generate_equals_and_hash = true;
|
option java_generate_equals_and_hash = true;
|
||||||
option optimize_for = SPEED;
|
option optimize_for = SPEED;
|
||||||
|
|
||||||
|
import "Table.proto";
|
||||||
import "HBase.proto";
|
import "HBase.proto";
|
||||||
import "RPC.proto";
|
import "RPC.proto";
|
||||||
|
|
||||||
|
@ -23,6 +23,7 @@ option java_generic_services = true;
|
|||||||
option java_generate_equals_and_hash = true;
|
option java_generate_equals_and_hash = true;
|
||||||
option optimize_for = SPEED;
|
option optimize_for = SPEED;
|
||||||
|
|
||||||
|
import 'Table.proto';
|
||||||
import 'HBase.proto';
|
import 'HBase.proto';
|
||||||
import 'Client.proto';
|
import 'Client.proto';
|
||||||
|
|
||||||
|
@ -23,28 +23,9 @@ option java_generic_services = true;
|
|||||||
option java_generate_equals_and_hash = true;
|
option java_generate_equals_and_hash = true;
|
||||||
option optimize_for = SPEED;
|
option optimize_for = SPEED;
|
||||||
|
|
||||||
import "AccessControl.proto";
|
|
||||||
import "FS.proto";
|
import "FS.proto";
|
||||||
import "HBase.proto";
|
import "HBase.proto";
|
||||||
|
|
||||||
/**
|
|
||||||
* Description of the snapshot to take
|
|
||||||
*/
|
|
||||||
message SnapshotDescription {
|
|
||||||
required string name = 1;
|
|
||||||
optional string table = 2; // not needed for delete, but checked for in taking snapshot
|
|
||||||
optional int64 creation_time = 3 [default = 0];
|
|
||||||
enum Type {
|
|
||||||
DISABLED = 0;
|
|
||||||
FLUSH = 1;
|
|
||||||
SKIPFLUSH = 2;
|
|
||||||
}
|
|
||||||
optional Type type = 4 [default = FLUSH];
|
|
||||||
optional int32 version = 5;
|
|
||||||
optional string owner = 6;
|
|
||||||
optional UsersAndPermissions users_and_permissions = 7;
|
|
||||||
}
|
|
||||||
|
|
||||||
message SnapshotFileInfo {
|
message SnapshotFileInfo {
|
||||||
enum Type {
|
enum Type {
|
||||||
HFILE = 1;
|
HFILE = 1;
|
||||||
|
33
hbase-protocol/src/main/protobuf/Table.proto
Normal file
33
hbase-protocol/src/main/protobuf/Table.proto
Normal file
@ -0,0 +1,33 @@
|
|||||||
|
/**
|
||||||
|
* Licensed to the Apache Software Foundation (ASF) under one
|
||||||
|
* or more contributor license agreements. See the NOTICE file
|
||||||
|
* distributed with this work for additional information
|
||||||
|
* regarding copyright ownership. The ASF licenses this file
|
||||||
|
* to you under the Apache License, Version 2.0 (the
|
||||||
|
* "License"); you may not use this file except in compliance
|
||||||
|
* with the License. You may obtain a copy of the License at
|
||||||
|
*
|
||||||
|
* http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
*
|
||||||
|
* Unless required by applicable law or agreed to in writing, software
|
||||||
|
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
* See the License for the specific language governing permissions and
|
||||||
|
* limitations under the License.
|
||||||
|
*/
|
||||||
|
|
||||||
|
// This file contains protocol buffers that are shared throughout HBase
|
||||||
|
package hbase.pb;
|
||||||
|
|
||||||
|
option java_package = "org.apache.hadoop.hbase.protobuf.generated";
|
||||||
|
option java_outer_classname = "TableProtos";
|
||||||
|
option java_generate_equals_and_hash = true;
|
||||||
|
option optimize_for = SPEED;
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Table Name
|
||||||
|
*/
|
||||||
|
message TableName {
|
||||||
|
required bytes namespace = 1;
|
||||||
|
required bytes qualifier = 2;
|
||||||
|
}
|
@ -23,6 +23,7 @@ option java_generic_services = false;
|
|||||||
option java_generate_equals_and_hash = true;
|
option java_generate_equals_and_hash = true;
|
||||||
option optimize_for = SPEED;
|
option optimize_for = SPEED;
|
||||||
|
|
||||||
|
import "Table.proto";
|
||||||
import "HBase.proto";
|
import "HBase.proto";
|
||||||
import "Client.proto";
|
import "Client.proto";
|
||||||
|
|
||||||
|
@ -26,6 +26,7 @@ option java_generic_services = true;
|
|||||||
option java_generate_equals_and_hash = true;
|
option java_generate_equals_and_hash = true;
|
||||||
option optimize_for = SPEED;
|
option optimize_for = SPEED;
|
||||||
|
|
||||||
|
import "Table.proto";
|
||||||
import "HBase.proto";
|
import "HBase.proto";
|
||||||
import "ClusterStatus.proto";
|
import "ClusterStatus.proto";
|
||||||
|
|
||||||
|
@ -49,7 +49,7 @@ org.apache.hadoop.hbase.HTableDescriptor;
|
|||||||
org.apache.hadoop.hbase.HBaseConfiguration;
|
org.apache.hadoop.hbase.HBaseConfiguration;
|
||||||
org.apache.hadoop.hbase.TableName;
|
org.apache.hadoop.hbase.TableName;
|
||||||
org.apache.hadoop.hbase.tool.Canary;
|
org.apache.hadoop.hbase.tool.Canary;
|
||||||
org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription;
|
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription;
|
||||||
org.apache.hadoop.hbase.master.DeadServer;
|
org.apache.hadoop.hbase.master.DeadServer;
|
||||||
org.apache.hadoop.hbase.protobuf.ProtobufUtil;
|
org.apache.hadoop.hbase.protobuf.ProtobufUtil;
|
||||||
org.apache.hadoop.hbase.security.visibility.VisibilityConstants;
|
org.apache.hadoop.hbase.security.visibility.VisibilityConstants;
|
||||||
|
@ -34,8 +34,8 @@ import org.apache.hadoop.hbase.client.Admin;
|
|||||||
import org.apache.hadoop.hbase.master.RegionPlan;
|
import org.apache.hadoop.hbase.master.RegionPlan;
|
||||||
import org.apache.hadoop.hbase.master.procedure.MasterProcedureEnv;
|
import org.apache.hadoop.hbase.master.procedure.MasterProcedureEnv;
|
||||||
import org.apache.hadoop.hbase.procedure2.ProcedureExecutor;
|
import org.apache.hadoop.hbase.procedure2.ProcedureExecutor;
|
||||||
|
import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription;
|
||||||
import org.apache.hadoop.hbase.protobuf.generated.QuotaProtos.Quotas;
|
import org.apache.hadoop.hbase.protobuf.generated.QuotaProtos.Quotas;
|
||||||
import org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription;
|
|
||||||
|
|
||||||
import java.io.IOException;
|
import java.io.IOException;
|
||||||
import java.util.List;
|
import java.util.List;
|
||||||
|
@ -34,8 +34,8 @@ import org.apache.hadoop.hbase.client.Admin;
|
|||||||
import org.apache.hadoop.hbase.master.RegionPlan;
|
import org.apache.hadoop.hbase.master.RegionPlan;
|
||||||
import org.apache.hadoop.hbase.master.procedure.MasterProcedureEnv;
|
import org.apache.hadoop.hbase.master.procedure.MasterProcedureEnv;
|
||||||
import org.apache.hadoop.hbase.procedure2.ProcedureExecutor;
|
import org.apache.hadoop.hbase.procedure2.ProcedureExecutor;
|
||||||
|
import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription;
|
||||||
import org.apache.hadoop.hbase.protobuf.generated.QuotaProtos.Quotas;
|
import org.apache.hadoop.hbase.protobuf.generated.QuotaProtos.Quotas;
|
||||||
import org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription;
|
|
||||||
|
|
||||||
import java.io.IOException;
|
import java.io.IOException;
|
||||||
import java.util.List;
|
import java.util.List;
|
||||||
|
@ -45,6 +45,7 @@ import org.apache.hadoop.hbase.filter.CompareFilter.CompareOp;
|
|||||||
import org.apache.hadoop.hbase.io.FSDataInputStreamWrapper;
|
import org.apache.hadoop.hbase.io.FSDataInputStreamWrapper;
|
||||||
import org.apache.hadoop.hbase.io.Reference;
|
import org.apache.hadoop.hbase.io.Reference;
|
||||||
import org.apache.hadoop.hbase.io.hfile.CacheConfig;
|
import org.apache.hadoop.hbase.io.hfile.CacheConfig;
|
||||||
|
import org.apache.hadoop.hbase.regionserver.DeleteTracker;
|
||||||
import org.apache.hadoop.hbase.regionserver.InternalScanner;
|
import org.apache.hadoop.hbase.regionserver.InternalScanner;
|
||||||
import org.apache.hadoop.hbase.regionserver.KeyValueScanner;
|
import org.apache.hadoop.hbase.regionserver.KeyValueScanner;
|
||||||
import org.apache.hadoop.hbase.regionserver.MiniBatchOperationInProgress;
|
import org.apache.hadoop.hbase.regionserver.MiniBatchOperationInProgress;
|
||||||
@ -56,7 +57,6 @@ import org.apache.hadoop.hbase.regionserver.Store;
|
|||||||
import org.apache.hadoop.hbase.regionserver.StoreFile;
|
import org.apache.hadoop.hbase.regionserver.StoreFile;
|
||||||
import org.apache.hadoop.hbase.regionserver.StoreFile.Reader;
|
import org.apache.hadoop.hbase.regionserver.StoreFile.Reader;
|
||||||
import org.apache.hadoop.hbase.regionserver.compactions.CompactionRequest;
|
import org.apache.hadoop.hbase.regionserver.compactions.CompactionRequest;
|
||||||
import org.apache.hadoop.hbase.regionserver.querymatcher.DeleteTracker;
|
|
||||||
import org.apache.hadoop.hbase.regionserver.wal.HLogKey;
|
import org.apache.hadoop.hbase.regionserver.wal.HLogKey;
|
||||||
import org.apache.hadoop.hbase.wal.WALKey;
|
import org.apache.hadoop.hbase.wal.WALKey;
|
||||||
import org.apache.hadoop.hbase.regionserver.wal.WALEdit;
|
import org.apache.hadoop.hbase.regionserver.wal.WALEdit;
|
||||||
|
@ -37,8 +37,8 @@ import org.apache.hadoop.hbase.client.Admin;
|
|||||||
import org.apache.hadoop.hbase.master.RegionPlan;
|
import org.apache.hadoop.hbase.master.RegionPlan;
|
||||||
import org.apache.hadoop.hbase.master.procedure.MasterProcedureEnv;
|
import org.apache.hadoop.hbase.master.procedure.MasterProcedureEnv;
|
||||||
import org.apache.hadoop.hbase.procedure2.ProcedureExecutor;
|
import org.apache.hadoop.hbase.procedure2.ProcedureExecutor;
|
||||||
|
import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription;
|
||||||
import org.apache.hadoop.hbase.protobuf.generated.QuotaProtos.Quotas;
|
import org.apache.hadoop.hbase.protobuf.generated.QuotaProtos.Quotas;
|
||||||
import org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription;
|
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Defines coprocessor hooks for interacting with operations on the
|
* Defines coprocessor hooks for interacting with operations on the
|
||||||
|
@ -45,6 +45,7 @@ import org.apache.hadoop.hbase.filter.CompareFilter.CompareOp;
|
|||||||
import org.apache.hadoop.hbase.io.FSDataInputStreamWrapper;
|
import org.apache.hadoop.hbase.io.FSDataInputStreamWrapper;
|
||||||
import org.apache.hadoop.hbase.io.Reference;
|
import org.apache.hadoop.hbase.io.Reference;
|
||||||
import org.apache.hadoop.hbase.io.hfile.CacheConfig;
|
import org.apache.hadoop.hbase.io.hfile.CacheConfig;
|
||||||
|
import org.apache.hadoop.hbase.regionserver.DeleteTracker;
|
||||||
import org.apache.hadoop.hbase.regionserver.Region;
|
import org.apache.hadoop.hbase.regionserver.Region;
|
||||||
import org.apache.hadoop.hbase.regionserver.Region.Operation;
|
import org.apache.hadoop.hbase.regionserver.Region.Operation;
|
||||||
import org.apache.hadoop.hbase.regionserver.InternalScanner;
|
import org.apache.hadoop.hbase.regionserver.InternalScanner;
|
||||||
@ -55,7 +56,6 @@ import org.apache.hadoop.hbase.regionserver.ScanType;
|
|||||||
import org.apache.hadoop.hbase.regionserver.Store;
|
import org.apache.hadoop.hbase.regionserver.Store;
|
||||||
import org.apache.hadoop.hbase.regionserver.StoreFile;
|
import org.apache.hadoop.hbase.regionserver.StoreFile;
|
||||||
import org.apache.hadoop.hbase.regionserver.compactions.CompactionRequest;
|
import org.apache.hadoop.hbase.regionserver.compactions.CompactionRequest;
|
||||||
import org.apache.hadoop.hbase.regionserver.querymatcher.DeleteTracker;
|
|
||||||
import org.apache.hadoop.hbase.regionserver.wal.HLogKey;
|
import org.apache.hadoop.hbase.regionserver.wal.HLogKey;
|
||||||
import org.apache.hadoop.hbase.wal.WALKey;
|
import org.apache.hadoop.hbase.wal.WALKey;
|
||||||
import org.apache.hadoop.hbase.regionserver.wal.WALEdit;
|
import org.apache.hadoop.hbase.regionserver.wal.WALEdit;
|
||||||
|
@ -36,8 +36,8 @@ import org.apache.hadoop.hbase.client.IsolationLevel;
|
|||||||
import org.apache.hadoop.hbase.client.Result;
|
import org.apache.hadoop.hbase.client.Result;
|
||||||
import org.apache.hadoop.hbase.client.Scan;
|
import org.apache.hadoop.hbase.client.Scan;
|
||||||
import org.apache.hadoop.hbase.io.ImmutableBytesWritable;
|
import org.apache.hadoop.hbase.io.ImmutableBytesWritable;
|
||||||
|
import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription;
|
||||||
import org.apache.hadoop.hbase.protobuf.generated.MapReduceProtos.TableSnapshotRegionSplit;
|
import org.apache.hadoop.hbase.protobuf.generated.MapReduceProtos.TableSnapshotRegionSplit;
|
||||||
import org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription;
|
|
||||||
import org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotRegionManifest;
|
import org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotRegionManifest;
|
||||||
import org.apache.hadoop.hbase.regionserver.HRegion;
|
import org.apache.hadoop.hbase.regionserver.HRegion;
|
||||||
import org.apache.hadoop.hbase.snapshot.RestoreSnapshotHelper;
|
import org.apache.hadoop.hbase.snapshot.RestoreSnapshotHelper;
|
||||||
|
@ -45,8 +45,8 @@ import org.apache.hadoop.hbase.coprocessor.ObserverContext;
|
|||||||
import org.apache.hadoop.hbase.master.procedure.MasterProcedureEnv;
|
import org.apache.hadoop.hbase.master.procedure.MasterProcedureEnv;
|
||||||
import org.apache.hadoop.hbase.metrics.MetricRegistry;
|
import org.apache.hadoop.hbase.metrics.MetricRegistry;
|
||||||
import org.apache.hadoop.hbase.procedure2.ProcedureExecutor;
|
import org.apache.hadoop.hbase.procedure2.ProcedureExecutor;
|
||||||
|
import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription;
|
||||||
import org.apache.hadoop.hbase.protobuf.generated.QuotaProtos.Quotas;
|
import org.apache.hadoop.hbase.protobuf.generated.QuotaProtos.Quotas;
|
||||||
import org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription;
|
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Provides the coprocessor framework and environment for master oriented
|
* Provides the coprocessor framework and environment for master oriented
|
||||||
|
@ -60,6 +60,7 @@ import org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.RegionStor
|
|||||||
import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPair;
|
import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPair;
|
||||||
import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ProcedureDescription;
|
import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ProcedureDescription;
|
||||||
import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.RegionSpecifierType;
|
import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.RegionSpecifierType;
|
||||||
|
import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription;
|
||||||
import org.apache.hadoop.hbase.protobuf.generated.MasterProtos.AbortProcedureRequest;
|
import org.apache.hadoop.hbase.protobuf.generated.MasterProtos.AbortProcedureRequest;
|
||||||
import org.apache.hadoop.hbase.protobuf.generated.MasterProtos.AbortProcedureResponse;
|
import org.apache.hadoop.hbase.protobuf.generated.MasterProtos.AbortProcedureResponse;
|
||||||
import org.apache.hadoop.hbase.protobuf.generated.MasterProtos.AddColumnRequest;
|
import org.apache.hadoop.hbase.protobuf.generated.MasterProtos.AddColumnRequest;
|
||||||
@ -185,7 +186,6 @@ import org.apache.hadoop.hbase.protobuf.generated.RegionServerStatusProtos.Repor
|
|||||||
import org.apache.hadoop.hbase.protobuf.generated.RegionServerStatusProtos.ReportRSFatalErrorResponse;
|
import org.apache.hadoop.hbase.protobuf.generated.RegionServerStatusProtos.ReportRSFatalErrorResponse;
|
||||||
import org.apache.hadoop.hbase.protobuf.generated.RegionServerStatusProtos.ReportRegionStateTransitionRequest;
|
import org.apache.hadoop.hbase.protobuf.generated.RegionServerStatusProtos.ReportRegionStateTransitionRequest;
|
||||||
import org.apache.hadoop.hbase.protobuf.generated.RegionServerStatusProtos.ReportRegionStateTransitionResponse;
|
import org.apache.hadoop.hbase.protobuf.generated.RegionServerStatusProtos.ReportRegionStateTransitionResponse;
|
||||||
import org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription;
|
|
||||||
import org.apache.hadoop.hbase.regionserver.RSRpcServices;
|
import org.apache.hadoop.hbase.regionserver.RSRpcServices;
|
||||||
import org.apache.hadoop.hbase.security.AccessDeniedException;
|
import org.apache.hadoop.hbase.security.AccessDeniedException;
|
||||||
import org.apache.hadoop.hbase.security.User;
|
import org.apache.hadoop.hbase.security.User;
|
||||||
@ -903,7 +903,7 @@ public class MasterRpcServices extends RSRpcServices
|
|||||||
List<TableName> tableNameList = null;
|
List<TableName> tableNameList = null;
|
||||||
if (req.getTableNamesCount() > 0) {
|
if (req.getTableNamesCount() > 0) {
|
||||||
tableNameList = new ArrayList<TableName>(req.getTableNamesCount());
|
tableNameList = new ArrayList<TableName>(req.getTableNamesCount());
|
||||||
for (HBaseProtos.TableName tableNamePB: req.getTableNamesList()) {
|
for (TableProtos.TableName tableNamePB: req.getTableNamesList()) {
|
||||||
tableNameList.add(ProtobufUtil.toTableName(tableNamePB));
|
tableNameList.add(ProtobufUtil.toTableName(tableNamePB));
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
@ -20,7 +20,7 @@ package org.apache.hadoop.hbase.master;
|
|||||||
import org.apache.hadoop.hbase.classification.InterfaceAudience;
|
import org.apache.hadoop.hbase.classification.InterfaceAudience;
|
||||||
import org.apache.hadoop.hbase.classification.InterfaceStability;
|
import org.apache.hadoop.hbase.classification.InterfaceStability;
|
||||||
import org.apache.hadoop.hbase.errorhandling.ForeignException;
|
import org.apache.hadoop.hbase.errorhandling.ForeignException;
|
||||||
import org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription;
|
import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription;
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Watch the current snapshot under process
|
* Watch the current snapshot under process
|
||||||
|
@ -41,7 +41,7 @@ import org.apache.hadoop.hbase.master.SnapshotSentinel;
|
|||||||
import org.apache.hadoop.hbase.master.handler.CreateTableHandler;
|
import org.apache.hadoop.hbase.master.handler.CreateTableHandler;
|
||||||
import org.apache.hadoop.hbase.monitoring.MonitoredTask;
|
import org.apache.hadoop.hbase.monitoring.MonitoredTask;
|
||||||
import org.apache.hadoop.hbase.monitoring.TaskMonitor;
|
import org.apache.hadoop.hbase.monitoring.TaskMonitor;
|
||||||
import org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription;
|
import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription;
|
||||||
import org.apache.hadoop.hbase.snapshot.ClientSnapshotDescriptionUtils;
|
import org.apache.hadoop.hbase.snapshot.ClientSnapshotDescriptionUtils;
|
||||||
import org.apache.hadoop.hbase.snapshot.RestoreSnapshotException;
|
import org.apache.hadoop.hbase.snapshot.RestoreSnapshotException;
|
||||||
import org.apache.hadoop.hbase.snapshot.RestoreSnapshotHelper;
|
import org.apache.hadoop.hbase.snapshot.RestoreSnapshotHelper;
|
||||||
|
@ -32,7 +32,7 @@ import org.apache.hadoop.hbase.ServerName;
|
|||||||
import org.apache.hadoop.hbase.client.RegionReplicaUtil;
|
import org.apache.hadoop.hbase.client.RegionReplicaUtil;
|
||||||
import org.apache.hadoop.hbase.errorhandling.ForeignException;
|
import org.apache.hadoop.hbase.errorhandling.ForeignException;
|
||||||
import org.apache.hadoop.hbase.master.MasterServices;
|
import org.apache.hadoop.hbase.master.MasterServices;
|
||||||
import org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription;
|
import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription;
|
||||||
import org.apache.hadoop.hbase.snapshot.ClientSnapshotDescriptionUtils;
|
import org.apache.hadoop.hbase.snapshot.ClientSnapshotDescriptionUtils;
|
||||||
import org.apache.hadoop.hbase.snapshot.SnapshotManifest;
|
import org.apache.hadoop.hbase.snapshot.SnapshotManifest;
|
||||||
import org.apache.hadoop.hbase.util.FSUtils;
|
import org.apache.hadoop.hbase.util.FSUtils;
|
||||||
|
@ -31,7 +31,7 @@ import org.apache.hadoop.hbase.errorhandling.ForeignException;
|
|||||||
import org.apache.hadoop.hbase.master.MasterServices;
|
import org.apache.hadoop.hbase.master.MasterServices;
|
||||||
import org.apache.hadoop.hbase.procedure.Procedure;
|
import org.apache.hadoop.hbase.procedure.Procedure;
|
||||||
import org.apache.hadoop.hbase.procedure.ProcedureCoordinator;
|
import org.apache.hadoop.hbase.procedure.ProcedureCoordinator;
|
||||||
import org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription;
|
import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription;
|
||||||
import org.apache.hadoop.hbase.snapshot.HBaseSnapshotException;
|
import org.apache.hadoop.hbase.snapshot.HBaseSnapshotException;
|
||||||
import org.apache.hadoop.hbase.util.Pair;
|
import org.apache.hadoop.hbase.util.Pair;
|
||||||
|
|
||||||
|
@ -34,7 +34,7 @@ import org.apache.hadoop.hbase.HTableDescriptor;
|
|||||||
import org.apache.hadoop.hbase.client.RegionReplicaUtil;
|
import org.apache.hadoop.hbase.client.RegionReplicaUtil;
|
||||||
import org.apache.hadoop.hbase.MetaTableAccessor;
|
import org.apache.hadoop.hbase.MetaTableAccessor;
|
||||||
import org.apache.hadoop.hbase.master.MasterServices;
|
import org.apache.hadoop.hbase.master.MasterServices;
|
||||||
import org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription;
|
import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription;
|
||||||
import org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotRegionManifest;
|
import org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotRegionManifest;
|
||||||
import org.apache.hadoop.hbase.snapshot.ClientSnapshotDescriptionUtils;
|
import org.apache.hadoop.hbase.snapshot.ClientSnapshotDescriptionUtils;
|
||||||
import org.apache.hadoop.hbase.snapshot.CorruptedSnapshotException;
|
import org.apache.hadoop.hbase.snapshot.CorruptedSnapshotException;
|
||||||
|
@ -46,7 +46,7 @@ import org.apache.hadoop.hbase.master.SnapshotSentinel;
|
|||||||
import org.apache.hadoop.hbase.master.handler.TableEventHandler;
|
import org.apache.hadoop.hbase.master.handler.TableEventHandler;
|
||||||
import org.apache.hadoop.hbase.monitoring.MonitoredTask;
|
import org.apache.hadoop.hbase.monitoring.MonitoredTask;
|
||||||
import org.apache.hadoop.hbase.monitoring.TaskMonitor;
|
import org.apache.hadoop.hbase.monitoring.TaskMonitor;
|
||||||
import org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription;
|
import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription;
|
||||||
import org.apache.hadoop.hbase.snapshot.ClientSnapshotDescriptionUtils;
|
import org.apache.hadoop.hbase.snapshot.ClientSnapshotDescriptionUtils;
|
||||||
import org.apache.hadoop.hbase.snapshot.RestoreSnapshotException;
|
import org.apache.hadoop.hbase.snapshot.RestoreSnapshotException;
|
||||||
import org.apache.hadoop.hbase.snapshot.RestoreSnapshotHelper;
|
import org.apache.hadoop.hbase.snapshot.RestoreSnapshotHelper;
|
||||||
|
@ -62,8 +62,8 @@ import org.apache.hadoop.hbase.procedure.ProcedureCoordinatorRpcs;
|
|||||||
import org.apache.hadoop.hbase.procedure.ZKProcedureCoordinatorRpcs;
|
import org.apache.hadoop.hbase.procedure.ZKProcedureCoordinatorRpcs;
|
||||||
import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPair;
|
import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPair;
|
||||||
import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ProcedureDescription;
|
import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ProcedureDescription;
|
||||||
import org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription;
|
import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription;
|
||||||
import org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription.Type;
|
import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription.Type;
|
||||||
import org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos;
|
import org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos;
|
||||||
import org.apache.hadoop.hbase.quotas.QuotaExceededException;
|
import org.apache.hadoop.hbase.quotas.QuotaExceededException;
|
||||||
import org.apache.hadoop.hbase.security.AccessDeniedException;
|
import org.apache.hadoop.hbase.security.AccessDeniedException;
|
||||||
|
@ -47,7 +47,7 @@ import org.apache.hadoop.hbase.master.TableLockManager;
|
|||||||
import org.apache.hadoop.hbase.master.TableLockManager.TableLock;
|
import org.apache.hadoop.hbase.master.TableLockManager.TableLock;
|
||||||
import org.apache.hadoop.hbase.monitoring.MonitoredTask;
|
import org.apache.hadoop.hbase.monitoring.MonitoredTask;
|
||||||
import org.apache.hadoop.hbase.monitoring.TaskMonitor;
|
import org.apache.hadoop.hbase.monitoring.TaskMonitor;
|
||||||
import org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription;
|
import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription;
|
||||||
import org.apache.hadoop.hbase.snapshot.ClientSnapshotDescriptionUtils;
|
import org.apache.hadoop.hbase.snapshot.ClientSnapshotDescriptionUtils;
|
||||||
import org.apache.hadoop.hbase.snapshot.SnapshotCreationException;
|
import org.apache.hadoop.hbase.snapshot.SnapshotCreationException;
|
||||||
import org.apache.hadoop.hbase.snapshot.SnapshotDescriptionUtils;
|
import org.apache.hadoop.hbase.snapshot.SnapshotDescriptionUtils;
|
||||||
|
@ -15,7 +15,7 @@
|
|||||||
* See the License for the specific language governing permissions and
|
* See the License for the specific language governing permissions and
|
||||||
* limitations under the License.
|
* limitations under the License.
|
||||||
*/
|
*/
|
||||||
package org.apache.hadoop.hbase.regionserver.querymatcher;
|
package org.apache.hadoop.hbase.regionserver;
|
||||||
|
|
||||||
import org.apache.hadoop.hbase.classification.InterfaceAudience;
|
import org.apache.hadoop.hbase.classification.InterfaceAudience;
|
||||||
import org.apache.hadoop.hbase.Cell;
|
import org.apache.hadoop.hbase.Cell;
|
@ -156,7 +156,7 @@ import org.apache.hadoop.hbase.protobuf.generated.ClientProtos;
|
|||||||
import org.apache.hadoop.hbase.protobuf.generated.ClientProtos.CoprocessorServiceCall;
|
import org.apache.hadoop.hbase.protobuf.generated.ClientProtos.CoprocessorServiceCall;
|
||||||
import org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.RegionLoad;
|
import org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.RegionLoad;
|
||||||
import org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.StoreSequenceId;
|
import org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.StoreSequenceId;
|
||||||
import org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription;
|
import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription;
|
||||||
import org.apache.hadoop.hbase.protobuf.generated.WALProtos;
|
import org.apache.hadoop.hbase.protobuf.generated.WALProtos;
|
||||||
import org.apache.hadoop.hbase.protobuf.generated.WALProtos.CompactionDescriptor;
|
import org.apache.hadoop.hbase.protobuf.generated.WALProtos.CompactionDescriptor;
|
||||||
import org.apache.hadoop.hbase.protobuf.generated.WALProtos.FlushDescriptor;
|
import org.apache.hadoop.hbase.protobuf.generated.WALProtos.FlushDescriptor;
|
||||||
|
@ -73,9 +73,9 @@ import org.apache.hadoop.hbase.io.Reference;
|
|||||||
import org.apache.hadoop.hbase.io.hfile.CacheConfig;
|
import org.apache.hadoop.hbase.io.hfile.CacheConfig;
|
||||||
import org.apache.hadoop.hbase.ipc.RpcServer;
|
import org.apache.hadoop.hbase.ipc.RpcServer;
|
||||||
import org.apache.hadoop.hbase.metrics.MetricRegistry;
|
import org.apache.hadoop.hbase.metrics.MetricRegistry;
|
||||||
|
import org.apache.hadoop.hbase.regionserver.DeleteTracker;
|
||||||
import org.apache.hadoop.hbase.regionserver.Region.Operation;
|
import org.apache.hadoop.hbase.regionserver.Region.Operation;
|
||||||
import org.apache.hadoop.hbase.regionserver.compactions.CompactionRequest;
|
import org.apache.hadoop.hbase.regionserver.compactions.CompactionRequest;
|
||||||
import org.apache.hadoop.hbase.regionserver.querymatcher.DeleteTracker;
|
|
||||||
import org.apache.hadoop.hbase.regionserver.wal.HLogKey;
|
import org.apache.hadoop.hbase.regionserver.wal.HLogKey;
|
||||||
import org.apache.hadoop.hbase.wal.WALKey;
|
import org.apache.hadoop.hbase.wal.WALKey;
|
||||||
import org.apache.hadoop.hbase.regionserver.wal.WALEdit;
|
import org.apache.hadoop.hbase.regionserver.wal.WALEdit;
|
||||||
|
@ -24,6 +24,7 @@ import org.apache.hadoop.hbase.HConstants;
|
|||||||
import org.apache.hadoop.hbase.KeepDeletedCells;
|
import org.apache.hadoop.hbase.KeepDeletedCells;
|
||||||
import org.apache.hadoop.hbase.classification.InterfaceAudience;
|
import org.apache.hadoop.hbase.classification.InterfaceAudience;
|
||||||
import org.apache.hadoop.hbase.filter.Filter;
|
import org.apache.hadoop.hbase.filter.Filter;
|
||||||
|
import org.apache.hadoop.hbase.regionserver.DeleteTracker;
|
||||||
import org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost;
|
import org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost;
|
||||||
import org.apache.hadoop.hbase.regionserver.ScanInfo;
|
import org.apache.hadoop.hbase.regionserver.ScanInfo;
|
||||||
import org.apache.hadoop.hbase.regionserver.ScanType;
|
import org.apache.hadoop.hbase.regionserver.ScanType;
|
||||||
|
@ -20,6 +20,7 @@ package org.apache.hadoop.hbase.regionserver.querymatcher;
|
|||||||
import org.apache.hadoop.hbase.Cell;
|
import org.apache.hadoop.hbase.Cell;
|
||||||
import org.apache.hadoop.hbase.KeepDeletedCells;
|
import org.apache.hadoop.hbase.KeepDeletedCells;
|
||||||
import org.apache.hadoop.hbase.classification.InterfaceAudience;
|
import org.apache.hadoop.hbase.classification.InterfaceAudience;
|
||||||
|
import org.apache.hadoop.hbase.regionserver.DeleteTracker;
|
||||||
import org.apache.hadoop.hbase.regionserver.ScanInfo;
|
import org.apache.hadoop.hbase.regionserver.ScanInfo;
|
||||||
|
|
||||||
/**
|
/**
|
||||||
|
@ -32,10 +32,11 @@ import org.apache.hadoop.hbase.client.Scan;
|
|||||||
import org.apache.hadoop.hbase.filter.Filter;
|
import org.apache.hadoop.hbase.filter.Filter;
|
||||||
import org.apache.hadoop.hbase.filter.Filter.ReturnCode;
|
import org.apache.hadoop.hbase.filter.Filter.ReturnCode;
|
||||||
import org.apache.hadoop.hbase.io.TimeRange;
|
import org.apache.hadoop.hbase.io.TimeRange;
|
||||||
|
import org.apache.hadoop.hbase.regionserver.DeleteTracker;
|
||||||
|
import org.apache.hadoop.hbase.regionserver.DeleteTracker.DeleteResult;
|
||||||
import org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost;
|
import org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost;
|
||||||
import org.apache.hadoop.hbase.regionserver.ScanInfo;
|
import org.apache.hadoop.hbase.regionserver.ScanInfo;
|
||||||
import org.apache.hadoop.hbase.regionserver.ScanType;
|
import org.apache.hadoop.hbase.regionserver.ScanType;
|
||||||
import org.apache.hadoop.hbase.regionserver.querymatcher.DeleteTracker.DeleteResult;
|
|
||||||
import org.apache.hadoop.hbase.util.Bytes;
|
import org.apache.hadoop.hbase.util.Bytes;
|
||||||
import org.apache.hadoop.hbase.util.EnvironmentEdgeManager;
|
import org.apache.hadoop.hbase.util.EnvironmentEdgeManager;
|
||||||
|
|
||||||
|
@ -22,6 +22,7 @@ import java.io.IOException;
|
|||||||
import org.apache.hadoop.hbase.Cell;
|
import org.apache.hadoop.hbase.Cell;
|
||||||
import org.apache.hadoop.hbase.CellUtil;
|
import org.apache.hadoop.hbase.CellUtil;
|
||||||
import org.apache.hadoop.hbase.classification.InterfaceAudience;
|
import org.apache.hadoop.hbase.classification.InterfaceAudience;
|
||||||
|
import org.apache.hadoop.hbase.regionserver.DeleteTracker;
|
||||||
import org.apache.hadoop.hbase.regionserver.ScanInfo;
|
import org.apache.hadoop.hbase.regionserver.ScanInfo;
|
||||||
|
|
||||||
/**
|
/**
|
||||||
|
@ -22,6 +22,7 @@ import java.io.IOException;
|
|||||||
import org.apache.hadoop.hbase.Cell;
|
import org.apache.hadoop.hbase.Cell;
|
||||||
import org.apache.hadoop.hbase.CellUtil;
|
import org.apache.hadoop.hbase.CellUtil;
|
||||||
import org.apache.hadoop.hbase.classification.InterfaceAudience;
|
import org.apache.hadoop.hbase.classification.InterfaceAudience;
|
||||||
|
import org.apache.hadoop.hbase.regionserver.DeleteTracker;
|
||||||
import org.apache.hadoop.hbase.regionserver.ScanInfo;
|
import org.apache.hadoop.hbase.regionserver.ScanInfo;
|
||||||
|
|
||||||
/**
|
/**
|
||||||
|
@ -24,6 +24,7 @@ import org.apache.hadoop.hbase.CellUtil;
|
|||||||
import org.apache.hadoop.hbase.KeepDeletedCells;
|
import org.apache.hadoop.hbase.KeepDeletedCells;
|
||||||
import org.apache.hadoop.hbase.classification.InterfaceAudience;
|
import org.apache.hadoop.hbase.classification.InterfaceAudience;
|
||||||
import org.apache.hadoop.hbase.client.Scan;
|
import org.apache.hadoop.hbase.client.Scan;
|
||||||
|
import org.apache.hadoop.hbase.regionserver.DeleteTracker;
|
||||||
import org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost;
|
import org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost;
|
||||||
import org.apache.hadoop.hbase.regionserver.ScanInfo;
|
import org.apache.hadoop.hbase.regionserver.ScanInfo;
|
||||||
|
|
||||||
|
@ -22,6 +22,7 @@ import java.util.SortedSet;
|
|||||||
import java.util.TreeSet;
|
import java.util.TreeSet;
|
||||||
|
|
||||||
import org.apache.hadoop.hbase.classification.InterfaceAudience;
|
import org.apache.hadoop.hbase.classification.InterfaceAudience;
|
||||||
|
import org.apache.hadoop.hbase.regionserver.DeleteTracker;
|
||||||
import org.apache.hadoop.hbase.Cell;
|
import org.apache.hadoop.hbase.Cell;
|
||||||
import org.apache.hadoop.hbase.KeyValue;
|
import org.apache.hadoop.hbase.KeyValue;
|
||||||
import org.apache.hadoop.hbase.util.Bytes;
|
import org.apache.hadoop.hbase.util.Bytes;
|
||||||
|
@ -26,10 +26,11 @@ import org.apache.hadoop.hbase.KeyValue.Type;
|
|||||||
import org.apache.hadoop.hbase.KeyValueUtil;
|
import org.apache.hadoop.hbase.KeyValueUtil;
|
||||||
import org.apache.hadoop.hbase.classification.InterfaceAudience;
|
import org.apache.hadoop.hbase.classification.InterfaceAudience;
|
||||||
import org.apache.hadoop.hbase.filter.Filter;
|
import org.apache.hadoop.hbase.filter.Filter;
|
||||||
|
import org.apache.hadoop.hbase.regionserver.DeleteTracker;
|
||||||
|
import org.apache.hadoop.hbase.regionserver.DeleteTracker.DeleteResult;
|
||||||
import org.apache.hadoop.hbase.regionserver.HStore;
|
import org.apache.hadoop.hbase.regionserver.HStore;
|
||||||
import org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost;
|
import org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost;
|
||||||
import org.apache.hadoop.hbase.regionserver.ScanInfo;
|
import org.apache.hadoop.hbase.regionserver.ScanInfo;
|
||||||
import org.apache.hadoop.hbase.regionserver.querymatcher.DeleteTracker.DeleteResult;
|
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* A query matcher that is specifically designed for the scan case.
|
* A query matcher that is specifically designed for the scan case.
|
||||||
|
@ -22,6 +22,7 @@ import java.io.IOException;
|
|||||||
import org.apache.hadoop.hbase.Cell;
|
import org.apache.hadoop.hbase.Cell;
|
||||||
import org.apache.hadoop.hbase.CellUtil;
|
import org.apache.hadoop.hbase.CellUtil;
|
||||||
import org.apache.hadoop.hbase.classification.InterfaceAudience;
|
import org.apache.hadoop.hbase.classification.InterfaceAudience;
|
||||||
|
import org.apache.hadoop.hbase.regionserver.DeleteTracker;
|
||||||
import org.apache.hadoop.hbase.regionserver.ScanInfo;
|
import org.apache.hadoop.hbase.regionserver.ScanInfo;
|
||||||
|
|
||||||
/**
|
/**
|
||||||
|
@ -30,7 +30,7 @@ import org.apache.hadoop.hbase.errorhandling.ForeignException;
|
|||||||
import org.apache.hadoop.hbase.errorhandling.ForeignExceptionDispatcher;
|
import org.apache.hadoop.hbase.errorhandling.ForeignExceptionDispatcher;
|
||||||
import org.apache.hadoop.hbase.procedure.ProcedureMember;
|
import org.apache.hadoop.hbase.procedure.ProcedureMember;
|
||||||
import org.apache.hadoop.hbase.procedure.Subprocedure;
|
import org.apache.hadoop.hbase.procedure.Subprocedure;
|
||||||
import org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription;
|
import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription;
|
||||||
import org.apache.hadoop.hbase.regionserver.HRegion;
|
import org.apache.hadoop.hbase.regionserver.HRegion;
|
||||||
import org.apache.hadoop.hbase.regionserver.Region;
|
import org.apache.hadoop.hbase.regionserver.Region;
|
||||||
import org.apache.hadoop.hbase.regionserver.Region.FlushResult;
|
import org.apache.hadoop.hbase.regionserver.Region.FlushResult;
|
||||||
|
@ -50,7 +50,7 @@ import org.apache.hadoop.hbase.procedure.RegionServerProcedureManager;
|
|||||||
import org.apache.hadoop.hbase.procedure.Subprocedure;
|
import org.apache.hadoop.hbase.procedure.Subprocedure;
|
||||||
import org.apache.hadoop.hbase.procedure.SubprocedureFactory;
|
import org.apache.hadoop.hbase.procedure.SubprocedureFactory;
|
||||||
import org.apache.hadoop.hbase.procedure.ZKProcedureMemberRpcs;
|
import org.apache.hadoop.hbase.procedure.ZKProcedureMemberRpcs;
|
||||||
import org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription;
|
import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription;
|
||||||
import org.apache.hadoop.hbase.regionserver.HRegionServer;
|
import org.apache.hadoop.hbase.regionserver.HRegionServer;
|
||||||
import org.apache.hadoop.hbase.regionserver.Region;
|
import org.apache.hadoop.hbase.regionserver.Region;
|
||||||
import org.apache.hadoop.hbase.regionserver.RegionServerServices;
|
import org.apache.hadoop.hbase.regionserver.RegionServerServices;
|
||||||
|
@ -91,10 +91,10 @@ import org.apache.hadoop.hbase.protobuf.ResponseConverter;
|
|||||||
import org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos;
|
import org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos;
|
||||||
import org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.AccessControlService;
|
import org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.AccessControlService;
|
||||||
import org.apache.hadoop.hbase.protobuf.generated.AdminProtos.WALEntry;
|
import org.apache.hadoop.hbase.protobuf.generated.AdminProtos.WALEntry;
|
||||||
|
import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription;
|
||||||
import org.apache.hadoop.hbase.protobuf.generated.QuotaProtos.Quotas;
|
import org.apache.hadoop.hbase.protobuf.generated.QuotaProtos.Quotas;
|
||||||
import org.apache.hadoop.hbase.protobuf.generated.SecureBulkLoadProtos.CleanupBulkLoadRequest;
|
import org.apache.hadoop.hbase.protobuf.generated.SecureBulkLoadProtos.CleanupBulkLoadRequest;
|
||||||
import org.apache.hadoop.hbase.protobuf.generated.SecureBulkLoadProtos.PrepareBulkLoadRequest;
|
import org.apache.hadoop.hbase.protobuf.generated.SecureBulkLoadProtos.PrepareBulkLoadRequest;
|
||||||
import org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription;
|
|
||||||
import org.apache.hadoop.hbase.regionserver.InternalScanner;
|
import org.apache.hadoop.hbase.regionserver.InternalScanner;
|
||||||
import org.apache.hadoop.hbase.regionserver.MiniBatchOperationInProgress;
|
import org.apache.hadoop.hbase.regionserver.MiniBatchOperationInProgress;
|
||||||
import org.apache.hadoop.hbase.regionserver.Region;
|
import org.apache.hadoop.hbase.regionserver.Region;
|
||||||
|
@ -90,13 +90,13 @@ import org.apache.hadoop.hbase.protobuf.generated.VisibilityLabelsProtos.Visibil
|
|||||||
import org.apache.hadoop.hbase.protobuf.generated.VisibilityLabelsProtos.VisibilityLabelsResponse;
|
import org.apache.hadoop.hbase.protobuf.generated.VisibilityLabelsProtos.VisibilityLabelsResponse;
|
||||||
import org.apache.hadoop.hbase.protobuf.generated.VisibilityLabelsProtos.VisibilityLabelsService;
|
import org.apache.hadoop.hbase.protobuf.generated.VisibilityLabelsProtos.VisibilityLabelsService;
|
||||||
import org.apache.hadoop.hbase.regionserver.BloomType;
|
import org.apache.hadoop.hbase.regionserver.BloomType;
|
||||||
|
import org.apache.hadoop.hbase.regionserver.DeleteTracker;
|
||||||
import org.apache.hadoop.hbase.regionserver.DisabledRegionSplitPolicy;
|
import org.apache.hadoop.hbase.regionserver.DisabledRegionSplitPolicy;
|
||||||
import org.apache.hadoop.hbase.regionserver.InternalScanner;
|
import org.apache.hadoop.hbase.regionserver.InternalScanner;
|
||||||
import org.apache.hadoop.hbase.regionserver.MiniBatchOperationInProgress;
|
import org.apache.hadoop.hbase.regionserver.MiniBatchOperationInProgress;
|
||||||
import org.apache.hadoop.hbase.regionserver.OperationStatus;
|
import org.apache.hadoop.hbase.regionserver.OperationStatus;
|
||||||
import org.apache.hadoop.hbase.regionserver.Region;
|
import org.apache.hadoop.hbase.regionserver.Region;
|
||||||
import org.apache.hadoop.hbase.regionserver.RegionScanner;
|
import org.apache.hadoop.hbase.regionserver.RegionScanner;
|
||||||
import org.apache.hadoop.hbase.regionserver.querymatcher.DeleteTracker;
|
|
||||||
import org.apache.hadoop.hbase.replication.ReplicationEndpoint;
|
import org.apache.hadoop.hbase.replication.ReplicationEndpoint;
|
||||||
import org.apache.hadoop.hbase.security.AccessDeniedException;
|
import org.apache.hadoop.hbase.security.AccessDeniedException;
|
||||||
import org.apache.hadoop.hbase.security.Superusers;
|
import org.apache.hadoop.hbase.security.Superusers;
|
||||||
|
@ -23,7 +23,7 @@ import org.apache.hadoop.hbase.TableName;
|
|||||||
import org.apache.hadoop.hbase.client.Admin;
|
import org.apache.hadoop.hbase.client.Admin;
|
||||||
import org.apache.hadoop.hbase.client.Connection;
|
import org.apache.hadoop.hbase.client.Connection;
|
||||||
import org.apache.hadoop.hbase.client.ConnectionFactory;
|
import org.apache.hadoop.hbase.client.ConnectionFactory;
|
||||||
import org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription;
|
import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription;
|
||||||
import org.apache.hadoop.hbase.util.AbstractHBaseTool;
|
import org.apache.hadoop.hbase.util.AbstractHBaseTool;
|
||||||
import java.util.Arrays;
|
import java.util.Arrays;
|
||||||
import java.util.Locale;
|
import java.util.Locale;
|
||||||
|
@ -53,7 +53,7 @@ import org.apache.hadoop.hbase.io.FileLink;
|
|||||||
import org.apache.hadoop.hbase.io.HFileLink;
|
import org.apache.hadoop.hbase.io.HFileLink;
|
||||||
import org.apache.hadoop.hbase.io.WALLink;
|
import org.apache.hadoop.hbase.io.WALLink;
|
||||||
import org.apache.hadoop.hbase.mapreduce.TableMapReduceUtil;
|
import org.apache.hadoop.hbase.mapreduce.TableMapReduceUtil;
|
||||||
import org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription;
|
import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription;
|
||||||
import org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotFileInfo;
|
import org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotFileInfo;
|
||||||
import org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotRegionManifest;
|
import org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotRegionManifest;
|
||||||
import org.apache.hadoop.hbase.util.FSUtils;
|
import org.apache.hadoop.hbase.util.FSUtils;
|
||||||
|
@ -54,7 +54,7 @@ import org.apache.hadoop.hbase.io.Reference;
|
|||||||
import org.apache.hadoop.hbase.monitoring.MonitoredTask;
|
import org.apache.hadoop.hbase.monitoring.MonitoredTask;
|
||||||
import org.apache.hadoop.hbase.monitoring.TaskMonitor;
|
import org.apache.hadoop.hbase.monitoring.TaskMonitor;
|
||||||
import org.apache.hadoop.hbase.protobuf.ProtobufUtil;
|
import org.apache.hadoop.hbase.protobuf.ProtobufUtil;
|
||||||
import org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription;
|
import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription;
|
||||||
import org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotRegionManifest;
|
import org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotRegionManifest;
|
||||||
import org.apache.hadoop.hbase.regionserver.HRegion;
|
import org.apache.hadoop.hbase.regionserver.HRegion;
|
||||||
import org.apache.hadoop.hbase.regionserver.HRegionFileSystem;
|
import org.apache.hadoop.hbase.regionserver.HRegionFileSystem;
|
||||||
|
@ -37,7 +37,7 @@ import org.apache.hadoop.hbase.client.Admin;
|
|||||||
import org.apache.hadoop.hbase.client.Connection;
|
import org.apache.hadoop.hbase.client.Connection;
|
||||||
import org.apache.hadoop.hbase.client.ConnectionFactory;
|
import org.apache.hadoop.hbase.client.ConnectionFactory;
|
||||||
import org.apache.hadoop.hbase.protobuf.ProtobufUtil;
|
import org.apache.hadoop.hbase.protobuf.ProtobufUtil;
|
||||||
import org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription;
|
import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription;
|
||||||
import org.apache.hadoop.hbase.security.User;
|
import org.apache.hadoop.hbase.security.User;
|
||||||
import org.apache.hadoop.hbase.security.access.AccessControlLists;
|
import org.apache.hadoop.hbase.security.access.AccessControlLists;
|
||||||
import org.apache.hadoop.hbase.security.access.TablePermission;
|
import org.apache.hadoop.hbase.security.access.TablePermission;
|
||||||
|
@ -43,7 +43,7 @@ import org.apache.hadoop.hbase.classification.InterfaceStability;
|
|||||||
import org.apache.hadoop.conf.Configured;
|
import org.apache.hadoop.conf.Configured;
|
||||||
import org.apache.hadoop.hbase.HRegionInfo;
|
import org.apache.hadoop.hbase.HRegionInfo;
|
||||||
import org.apache.hadoop.hbase.TableName;
|
import org.apache.hadoop.hbase.TableName;
|
||||||
import org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription;
|
import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription;
|
||||||
import org.apache.hadoop.util.StringUtils;
|
import org.apache.hadoop.util.StringUtils;
|
||||||
import org.apache.hadoop.util.Tool;
|
import org.apache.hadoop.util.Tool;
|
||||||
import org.apache.hadoop.util.ToolRunner;
|
import org.apache.hadoop.util.ToolRunner;
|
||||||
|
@ -43,8 +43,8 @@ import org.apache.hadoop.hbase.HRegionInfo;
|
|||||||
import org.apache.hadoop.hbase.HTableDescriptor;
|
import org.apache.hadoop.hbase.HTableDescriptor;
|
||||||
import org.apache.hadoop.hbase.classification.InterfaceAudience;
|
import org.apache.hadoop.hbase.classification.InterfaceAudience;
|
||||||
import org.apache.hadoop.hbase.errorhandling.ForeignExceptionSnare;
|
import org.apache.hadoop.hbase.errorhandling.ForeignExceptionSnare;
|
||||||
|
import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription;
|
||||||
import org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDataManifest;
|
import org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDataManifest;
|
||||||
import org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription;
|
|
||||||
import org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotRegionManifest;
|
import org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotRegionManifest;
|
||||||
import org.apache.hadoop.hbase.protobuf.ProtobufUtil;
|
import org.apache.hadoop.hbase.protobuf.ProtobufUtil;
|
||||||
import org.apache.hadoop.hbase.regionserver.HRegion;
|
import org.apache.hadoop.hbase.regionserver.HRegion;
|
||||||
|
@ -36,7 +36,7 @@ import org.apache.hadoop.fs.FileStatus;
|
|||||||
import org.apache.hadoop.fs.FileSystem;
|
import org.apache.hadoop.fs.FileSystem;
|
||||||
import org.apache.hadoop.fs.Path;
|
import org.apache.hadoop.fs.Path;
|
||||||
import org.apache.hadoop.hbase.HRegionInfo;
|
import org.apache.hadoop.hbase.HRegionInfo;
|
||||||
import org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription;
|
import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription;
|
||||||
import org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotRegionManifest;
|
import org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotRegionManifest;
|
||||||
import org.apache.hadoop.hbase.regionserver.HRegionFileSystem;
|
import org.apache.hadoop.hbase.regionserver.HRegionFileSystem;
|
||||||
import org.apache.hadoop.hbase.regionserver.StoreFileInfo;
|
import org.apache.hadoop.hbase.regionserver.StoreFileInfo;
|
||||||
|
@ -40,7 +40,7 @@ import org.apache.hadoop.fs.FileSystem;
|
|||||||
import org.apache.hadoop.fs.Path;
|
import org.apache.hadoop.fs.Path;
|
||||||
import org.apache.hadoop.fs.PathFilter;
|
import org.apache.hadoop.fs.PathFilter;
|
||||||
import org.apache.hadoop.hbase.HRegionInfo;
|
import org.apache.hadoop.hbase.HRegionInfo;
|
||||||
import org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription;
|
import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription;
|
||||||
import org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotRegionManifest;
|
import org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotRegionManifest;
|
||||||
import org.apache.hadoop.hbase.regionserver.StoreFileInfo;
|
import org.apache.hadoop.hbase.regionserver.StoreFileInfo;
|
||||||
import org.apache.hadoop.hbase.util.ByteStringer;
|
import org.apache.hadoop.hbase.util.ByteStringer;
|
||||||
|
@ -39,7 +39,7 @@ import org.apache.hadoop.fs.Path;
|
|||||||
import org.apache.hadoop.hbase.HRegionInfo;
|
import org.apache.hadoop.hbase.HRegionInfo;
|
||||||
import org.apache.hadoop.hbase.TableName;
|
import org.apache.hadoop.hbase.TableName;
|
||||||
import org.apache.hadoop.hbase.io.HFileLink;
|
import org.apache.hadoop.hbase.io.HFileLink;
|
||||||
import org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription;
|
import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription;
|
||||||
import org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotRegionManifest;
|
import org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotRegionManifest;
|
||||||
import org.apache.hadoop.hbase.regionserver.StoreFileInfo;
|
import org.apache.hadoop.hbase.regionserver.StoreFileInfo;
|
||||||
|
|
||||||
|
@ -24,7 +24,7 @@
|
|||||||
import="org.apache.hadoop.hbase.client.HConnectionManager"
|
import="org.apache.hadoop.hbase.client.HConnectionManager"
|
||||||
import="org.apache.hadoop.hbase.master.HMaster"
|
import="org.apache.hadoop.hbase.master.HMaster"
|
||||||
import="org.apache.hadoop.hbase.snapshot.SnapshotInfo"
|
import="org.apache.hadoop.hbase.snapshot.SnapshotInfo"
|
||||||
import="org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription"
|
import="org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription"
|
||||||
import="org.apache.hadoop.util.StringUtils"
|
import="org.apache.hadoop.util.StringUtils"
|
||||||
import="org.apache.hadoop.hbase.TableName"
|
import="org.apache.hadoop.hbase.TableName"
|
||||||
import="org.apache.hadoop.hbase.HBaseConfiguration" %>
|
import="org.apache.hadoop.hbase.HBaseConfiguration" %>
|
||||||
|
@ -27,7 +27,7 @@
|
|||||||
import="org.apache.hadoop.hbase.HBaseConfiguration"
|
import="org.apache.hadoop.hbase.HBaseConfiguration"
|
||||||
import="org.apache.hadoop.hbase.client.Admin"
|
import="org.apache.hadoop.hbase.client.Admin"
|
||||||
import="org.apache.hadoop.hbase.master.HMaster"
|
import="org.apache.hadoop.hbase.master.HMaster"
|
||||||
import="org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription"
|
import="org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription"
|
||||||
import="org.apache.hadoop.hbase.snapshot.SnapshotInfo"
|
import="org.apache.hadoop.hbase.snapshot.SnapshotInfo"
|
||||||
import="org.apache.hadoop.hbase.TableName"
|
import="org.apache.hadoop.hbase.TableName"
|
||||||
import="org.apache.hadoop.util.StringUtils" %>
|
import="org.apache.hadoop.util.StringUtils" %>
|
||||||
|
@ -34,7 +34,7 @@ import org.apache.hadoop.hbase.HTableDescriptor;
|
|||||||
import org.apache.hadoop.hbase.TableName;
|
import org.apache.hadoop.hbase.TableName;
|
||||||
import org.apache.hadoop.hbase.HBaseTestingUtility;
|
import org.apache.hadoop.hbase.HBaseTestingUtility;
|
||||||
import org.apache.hadoop.hbase.HConstants;
|
import org.apache.hadoop.hbase.HConstants;
|
||||||
import org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription;
|
import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription;
|
||||||
import org.apache.hadoop.hbase.testclassification.LargeTests;
|
import org.apache.hadoop.hbase.testclassification.LargeTests;
|
||||||
import org.apache.hadoop.hbase.TableNotFoundException;
|
import org.apache.hadoop.hbase.TableNotFoundException;
|
||||||
import org.apache.hadoop.hbase.master.snapshot.SnapshotManager;
|
import org.apache.hadoop.hbase.master.snapshot.SnapshotManager;
|
||||||
|
@ -59,10 +59,10 @@ import org.apache.hadoop.hbase.master.procedure.MasterProcedureEnv;
|
|||||||
import org.apache.hadoop.hbase.procedure2.ProcedureExecutor;
|
import org.apache.hadoop.hbase.procedure2.ProcedureExecutor;
|
||||||
import org.apache.hadoop.hbase.protobuf.ProtobufUtil;
|
import org.apache.hadoop.hbase.protobuf.ProtobufUtil;
|
||||||
import org.apache.hadoop.hbase.protobuf.RequestConverter;
|
import org.apache.hadoop.hbase.protobuf.RequestConverter;
|
||||||
|
import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription;
|
||||||
import org.apache.hadoop.hbase.protobuf.generated.MasterProtos.GetTableDescriptorsRequest;
|
import org.apache.hadoop.hbase.protobuf.generated.MasterProtos.GetTableDescriptorsRequest;
|
||||||
import org.apache.hadoop.hbase.protobuf.generated.MasterProtos.GetTableNamesRequest;
|
import org.apache.hadoop.hbase.protobuf.generated.MasterProtos.GetTableNamesRequest;
|
||||||
import org.apache.hadoop.hbase.protobuf.generated.QuotaProtos.Quotas;
|
import org.apache.hadoop.hbase.protobuf.generated.QuotaProtos.Quotas;
|
||||||
import org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription;
|
|
||||||
import org.apache.hadoop.hbase.regionserver.HRegionServer;
|
import org.apache.hadoop.hbase.regionserver.HRegionServer;
|
||||||
import org.apache.hadoop.hbase.testclassification.MediumTests;
|
import org.apache.hadoop.hbase.testclassification.MediumTests;
|
||||||
import org.apache.hadoop.hbase.util.Bytes;
|
import org.apache.hadoop.hbase.util.Bytes;
|
||||||
|
@ -34,7 +34,7 @@ import org.apache.hadoop.fs.Path;
|
|||||||
import org.apache.hadoop.hbase.HBaseTestingUtility;
|
import org.apache.hadoop.hbase.HBaseTestingUtility;
|
||||||
import org.apache.hadoop.hbase.HConstants;
|
import org.apache.hadoop.hbase.HConstants;
|
||||||
import org.apache.hadoop.hbase.HTableDescriptor;
|
import org.apache.hadoop.hbase.HTableDescriptor;
|
||||||
import org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription;
|
import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription;
|
||||||
import org.apache.hadoop.hbase.testclassification.MediumTests;
|
import org.apache.hadoop.hbase.testclassification.MediumTests;
|
||||||
import org.apache.hadoop.hbase.TableName;
|
import org.apache.hadoop.hbase.TableName;
|
||||||
import org.apache.hadoop.hbase.client.Admin;
|
import org.apache.hadoop.hbase.client.Admin;
|
||||||
|
@ -34,8 +34,8 @@ import org.apache.hadoop.hbase.HTableDescriptor;
|
|||||||
import org.apache.hadoop.hbase.testclassification.SmallTests;
|
import org.apache.hadoop.hbase.testclassification.SmallTests;
|
||||||
import org.apache.hadoop.hbase.TableName;
|
import org.apache.hadoop.hbase.TableName;
|
||||||
import org.apache.hadoop.hbase.exceptions.DeserializationException;
|
import org.apache.hadoop.hbase.exceptions.DeserializationException;
|
||||||
import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos;
|
|
||||||
import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfo;
|
import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfo;
|
||||||
|
import org.apache.hadoop.hbase.protobuf.generated.TableProtos;
|
||||||
import org.apache.hadoop.hbase.util.Bytes;
|
import org.apache.hadoop.hbase.util.Bytes;
|
||||||
import org.apache.hadoop.hbase.util.FSTableDescriptors;
|
import org.apache.hadoop.hbase.util.FSTableDescriptors;
|
||||||
import org.apache.hadoop.hbase.util.MD5Hash;
|
import org.apache.hadoop.hbase.util.MD5Hash;
|
||||||
@ -240,7 +240,7 @@ public class TestHRegionInfo {
|
|||||||
|
|
||||||
// test convert RegionInfo without replicaId
|
// test convert RegionInfo without replicaId
|
||||||
RegionInfo info = RegionInfo.newBuilder()
|
RegionInfo info = RegionInfo.newBuilder()
|
||||||
.setTableName(HBaseProtos.TableName.newBuilder()
|
.setTableName(TableProtos.TableName.newBuilder()
|
||||||
.setQualifier(ByteString.copyFrom(tableName.getQualifier()))
|
.setQualifier(ByteString.copyFrom(tableName.getQualifier()))
|
||||||
.setNamespace(ByteString.copyFrom(tableName.getNamespace()))
|
.setNamespace(ByteString.copyFrom(tableName.getNamespace()))
|
||||||
.build())
|
.build())
|
||||||
|
@ -22,7 +22,7 @@ package org.apache.hadoop.hbase.regionserver.querymatcher;
|
|||||||
import static org.junit.Assert.*;
|
import static org.junit.Assert.*;
|
||||||
|
|
||||||
import org.apache.hadoop.hbase.KeyValue;
|
import org.apache.hadoop.hbase.KeyValue;
|
||||||
import org.apache.hadoop.hbase.regionserver.querymatcher.DeleteTracker.DeleteResult;
|
import org.apache.hadoop.hbase.regionserver.DeleteTracker.DeleteResult;
|
||||||
import org.apache.hadoop.hbase.testclassification.RegionServerTests;
|
import org.apache.hadoop.hbase.testclassification.RegionServerTests;
|
||||||
import org.apache.hadoop.hbase.testclassification.SmallTests;
|
import org.apache.hadoop.hbase.testclassification.SmallTests;
|
||||||
import org.apache.hadoop.hbase.util.Bytes;
|
import org.apache.hadoop.hbase.util.Bytes;
|
||||||
|
@ -52,7 +52,7 @@ import org.apache.hadoop.hbase.HRegionLocation;
|
|||||||
import org.apache.hadoop.hbase.HTableDescriptor;
|
import org.apache.hadoop.hbase.HTableDescriptor;
|
||||||
import org.apache.hadoop.hbase.KeyValue;
|
import org.apache.hadoop.hbase.KeyValue;
|
||||||
import org.apache.hadoop.hbase.ProcedureInfo;
|
import org.apache.hadoop.hbase.ProcedureInfo;
|
||||||
import org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription;
|
import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription;
|
||||||
import org.apache.hadoop.hbase.security.Superusers;
|
import org.apache.hadoop.hbase.security.Superusers;
|
||||||
import org.apache.hadoop.hbase.testclassification.LargeTests;
|
import org.apache.hadoop.hbase.testclassification.LargeTests;
|
||||||
import org.apache.hadoop.hbase.MiniHBaseCluster;
|
import org.apache.hadoop.hbase.MiniHBaseCluster;
|
||||||
|
@ -57,7 +57,7 @@ import org.apache.hadoop.hbase.coprocessor.RegionServerCoprocessorEnvironment;
|
|||||||
import org.apache.hadoop.hbase.filter.BinaryComparator;
|
import org.apache.hadoop.hbase.filter.BinaryComparator;
|
||||||
import org.apache.hadoop.hbase.filter.CompareFilter;
|
import org.apache.hadoop.hbase.filter.CompareFilter;
|
||||||
import org.apache.hadoop.hbase.master.MasterCoprocessorHost;
|
import org.apache.hadoop.hbase.master.MasterCoprocessorHost;
|
||||||
import org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription;
|
import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription;
|
||||||
import org.apache.hadoop.hbase.regionserver.MiniBatchOperationInProgress;
|
import org.apache.hadoop.hbase.regionserver.MiniBatchOperationInProgress;
|
||||||
import org.apache.hadoop.hbase.regionserver.Region;
|
import org.apache.hadoop.hbase.regionserver.Region;
|
||||||
import org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost;
|
import org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost;
|
||||||
|
@ -59,7 +59,7 @@ import org.apache.hadoop.hbase.io.HFileLink;
|
|||||||
import org.apache.hadoop.hbase.master.HMaster;
|
import org.apache.hadoop.hbase.master.HMaster;
|
||||||
import org.apache.hadoop.hbase.master.MasterFileSystem;
|
import org.apache.hadoop.hbase.master.MasterFileSystem;
|
||||||
import org.apache.hadoop.hbase.protobuf.ProtobufUtil;
|
import org.apache.hadoop.hbase.protobuf.ProtobufUtil;
|
||||||
import org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription;
|
import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription;
|
||||||
import org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotRegionManifest;
|
import org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotRegionManifest;
|
||||||
import org.apache.hadoop.hbase.protobuf.generated.MasterProtos.IsSnapshotDoneRequest;
|
import org.apache.hadoop.hbase.protobuf.generated.MasterProtos.IsSnapshotDoneRequest;
|
||||||
import org.apache.hadoop.hbase.protobuf.generated.MasterProtos.IsSnapshotDoneResponse;
|
import org.apache.hadoop.hbase.protobuf.generated.MasterProtos.IsSnapshotDoneResponse;
|
||||||
|
@ -43,7 +43,7 @@ import org.apache.hadoop.hbase.client.Admin;
|
|||||||
import org.apache.hadoop.hbase.client.HTable;
|
import org.apache.hadoop.hbase.client.HTable;
|
||||||
import org.apache.hadoop.hbase.client.Table;
|
import org.apache.hadoop.hbase.client.Table;
|
||||||
import org.apache.hadoop.hbase.master.snapshot.SnapshotManager;
|
import org.apache.hadoop.hbase.master.snapshot.SnapshotManager;
|
||||||
import org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription;
|
import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription;
|
||||||
import org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotFileInfo;
|
import org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotFileInfo;
|
||||||
import org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotRegionManifest;
|
import org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotRegionManifest;
|
||||||
import org.apache.hadoop.hbase.snapshot.SnapshotTestingUtils.SnapshotMock;
|
import org.apache.hadoop.hbase.snapshot.SnapshotTestingUtils.SnapshotMock;
|
||||||
|
@ -43,7 +43,7 @@ import org.apache.hadoop.hbase.client.Admin;
|
|||||||
import org.apache.hadoop.hbase.client.Table;
|
import org.apache.hadoop.hbase.client.Table;
|
||||||
import org.apache.hadoop.hbase.master.HMaster;
|
import org.apache.hadoop.hbase.master.HMaster;
|
||||||
import org.apache.hadoop.hbase.master.snapshot.SnapshotManager;
|
import org.apache.hadoop.hbase.master.snapshot.SnapshotManager;
|
||||||
import org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription;
|
import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription;
|
||||||
import org.apache.hadoop.hbase.regionserver.ConstantSizeRegionSplitPolicy;
|
import org.apache.hadoop.hbase.regionserver.ConstantSizeRegionSplitPolicy;
|
||||||
import org.apache.hadoop.hbase.testclassification.LargeTests;
|
import org.apache.hadoop.hbase.testclassification.LargeTests;
|
||||||
import org.apache.hadoop.hbase.util.Bytes;
|
import org.apache.hadoop.hbase.util.Bytes;
|
||||||
|
@ -27,6 +27,7 @@ import org.apache.hadoop.hbase.HTableDescriptor;
|
|||||||
import org.apache.hadoop.hbase.TableName;
|
import org.apache.hadoop.hbase.TableName;
|
||||||
import org.apache.hadoop.hbase.client.Table;
|
import org.apache.hadoop.hbase.client.Table;
|
||||||
import org.apache.hadoop.hbase.errorhandling.ForeignExceptionDispatcher;
|
import org.apache.hadoop.hbase.errorhandling.ForeignExceptionDispatcher;
|
||||||
|
import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos;
|
||||||
import org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos;
|
import org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos;
|
||||||
import org.apache.hadoop.hbase.regionserver.ConstantSizeRegionSplitPolicy;
|
import org.apache.hadoop.hbase.regionserver.ConstantSizeRegionSplitPolicy;
|
||||||
import org.apache.hadoop.hbase.regionserver.HRegion;
|
import org.apache.hadoop.hbase.regionserver.HRegion;
|
||||||
@ -108,10 +109,9 @@ public class TestRegionSnapshotTask {
|
|||||||
|
|
||||||
List<HRegion> hRegions = TEST_UTIL.getHBaseCluster().getRegions(tableName);
|
List<HRegion> hRegions = TEST_UTIL.getHBaseCluster().getRegions(tableName);
|
||||||
|
|
||||||
final SnapshotProtos.SnapshotDescription snapshot =
|
final HBaseProtos.SnapshotDescription snapshot = HBaseProtos.SnapshotDescription.newBuilder()
|
||||||
SnapshotProtos.SnapshotDescription.newBuilder()
|
|
||||||
.setTable(tableName.getNameAsString())
|
.setTable(tableName.getNameAsString())
|
||||||
.setType(SnapshotProtos.SnapshotDescription.Type.FLUSH)
|
.setType(HBaseProtos.SnapshotDescription.Type.FLUSH)
|
||||||
.setName("test_table_snapshot")
|
.setName("test_table_snapshot")
|
||||||
.setVersion(SnapshotManifestV2.DESCRIPTOR_VERSION)
|
.setVersion(SnapshotManifestV2.DESCRIPTOR_VERSION)
|
||||||
.build();
|
.build();
|
||||||
@ -161,7 +161,7 @@ public class TestRegionSnapshotTask {
|
|||||||
SnapshotReferenceUtil.verifySnapshot(conf, fs, manifest);
|
SnapshotReferenceUtil.verifySnapshot(conf, fs, manifest);
|
||||||
}
|
}
|
||||||
|
|
||||||
private void addRegionToSnapshot(SnapshotProtos.SnapshotDescription snapshot,
|
private void addRegionToSnapshot(HBaseProtos.SnapshotDescription snapshot,
|
||||||
HRegion region, SnapshotManifest manifest) throws Exception {
|
HRegion region, SnapshotManifest manifest) throws Exception {
|
||||||
LOG.info("Adding region to snapshot: " + region.getRegionInfo().getRegionNameAsString());
|
LOG.info("Adding region to snapshot: " + region.getRegionInfo().getRegionNameAsString());
|
||||||
Path workingDir = SnapshotDescriptionUtils.getWorkingSnapshotDir(snapshot, rootDir);
|
Path workingDir = SnapshotDescriptionUtils.getWorkingSnapshotDir(snapshot, rootDir);
|
||||||
@ -171,7 +171,7 @@ public class TestRegionSnapshotTask {
|
|||||||
}
|
}
|
||||||
|
|
||||||
private SnapshotManifest.RegionVisitor createRegionVisitorWithDelay(
|
private SnapshotManifest.RegionVisitor createRegionVisitorWithDelay(
|
||||||
SnapshotProtos.SnapshotDescription desc, Path workingDir) {
|
HBaseProtos.SnapshotDescription desc, Path workingDir) {
|
||||||
return new SnapshotManifestV2.ManifestBuilder(conf, fs, workingDir) {
|
return new SnapshotManifestV2.ManifestBuilder(conf, fs, workingDir) {
|
||||||
@Override
|
@Override
|
||||||
public void storeFile(final SnapshotProtos.SnapshotRegionManifest.Builder region,
|
public void storeFile(final SnapshotProtos.SnapshotRegionManifest.Builder region,
|
||||||
|
@ -23,7 +23,7 @@ import org.apache.commons.logging.Log;
|
|||||||
import org.apache.commons.logging.LogFactory;
|
import org.apache.commons.logging.LogFactory;
|
||||||
import org.apache.hadoop.hbase.HBaseTestingUtility;
|
import org.apache.hadoop.hbase.HBaseTestingUtility;
|
||||||
import org.apache.hadoop.hbase.HConstants;
|
import org.apache.hadoop.hbase.HConstants;
|
||||||
import org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription;
|
import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription;
|
||||||
import org.apache.hadoop.hbase.testclassification.LargeTests;
|
import org.apache.hadoop.hbase.testclassification.LargeTests;
|
||||||
import org.apache.hadoop.hbase.TableName;
|
import org.apache.hadoop.hbase.TableName;
|
||||||
import org.apache.hadoop.hbase.client.Admin;
|
import org.apache.hadoop.hbase.client.Admin;
|
||||||
|
@ -31,7 +31,7 @@ import org.apache.hadoop.fs.Path;
|
|||||||
import org.apache.hadoop.hbase.HBaseTestingUtility;
|
import org.apache.hadoop.hbase.HBaseTestingUtility;
|
||||||
import org.apache.hadoop.hbase.HConstants;
|
import org.apache.hadoop.hbase.HConstants;
|
||||||
import org.apache.hadoop.hbase.HTableDescriptor;
|
import org.apache.hadoop.hbase.HTableDescriptor;
|
||||||
import org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription;
|
import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription;
|
||||||
import org.apache.hadoop.hbase.testclassification.SmallTests;
|
import org.apache.hadoop.hbase.testclassification.SmallTests;
|
||||||
import org.apache.hadoop.hbase.errorhandling.ForeignExceptionDispatcher;
|
import org.apache.hadoop.hbase.errorhandling.ForeignExceptionDispatcher;
|
||||||
import org.apache.hadoop.hbase.io.HFileLink;
|
import org.apache.hadoop.hbase.io.HFileLink;
|
||||||
|
@ -32,7 +32,7 @@ import org.apache.hadoop.hbase.coprocessor.BaseMasterObserver;
|
|||||||
import org.apache.hadoop.hbase.coprocessor.CoprocessorHost;
|
import org.apache.hadoop.hbase.coprocessor.CoprocessorHost;
|
||||||
import org.apache.hadoop.hbase.coprocessor.MasterCoprocessorEnvironment;
|
import org.apache.hadoop.hbase.coprocessor.MasterCoprocessorEnvironment;
|
||||||
import org.apache.hadoop.hbase.coprocessor.ObserverContext;
|
import org.apache.hadoop.hbase.coprocessor.ObserverContext;
|
||||||
import org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription;
|
import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription;
|
||||||
import org.apache.hadoop.hbase.testclassification.MediumTests;
|
import org.apache.hadoop.hbase.testclassification.MediumTests;
|
||||||
import org.apache.hadoop.hbase.util.TestTableName;
|
import org.apache.hadoop.hbase.util.TestTableName;
|
||||||
import org.junit.After;
|
import org.junit.After;
|
||||||
|
@ -29,7 +29,7 @@ import org.apache.hadoop.fs.FileSystem;
|
|||||||
import org.apache.hadoop.fs.Path;
|
import org.apache.hadoop.fs.Path;
|
||||||
import org.apache.hadoop.hbase.HBaseTestingUtility;
|
import org.apache.hadoop.hbase.HBaseTestingUtility;
|
||||||
import org.apache.hadoop.hbase.HConstants;
|
import org.apache.hadoop.hbase.HConstants;
|
||||||
import org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription;
|
import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription;
|
||||||
import org.apache.hadoop.hbase.testclassification.MediumTests;
|
import org.apache.hadoop.hbase.testclassification.MediumTests;
|
||||||
import org.apache.hadoop.hbase.util.EnvironmentEdgeManagerTestHelper;
|
import org.apache.hadoop.hbase.util.EnvironmentEdgeManagerTestHelper;
|
||||||
import org.junit.After;
|
import org.junit.After;
|
||||||
|
@ -27,8 +27,8 @@ import org.apache.hadoop.hbase.HColumnDescriptor;
|
|||||||
import org.apache.hadoop.hbase.HRegionInfo;
|
import org.apache.hadoop.hbase.HRegionInfo;
|
||||||
import org.apache.hadoop.hbase.TableName;
|
import org.apache.hadoop.hbase.TableName;
|
||||||
import org.apache.hadoop.hbase.HBaseTestingUtility;
|
import org.apache.hadoop.hbase.HBaseTestingUtility;
|
||||||
|
import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription;
|
||||||
import org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDataManifest;
|
import org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDataManifest;
|
||||||
import org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription;
|
|
||||||
import org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotRegionManifest;
|
import org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotRegionManifest;
|
||||||
import org.apache.hadoop.hbase.testclassification.MasterTests;
|
import org.apache.hadoop.hbase.testclassification.MasterTests;
|
||||||
import org.apache.hadoop.hbase.testclassification.SmallTests;
|
import org.apache.hadoop.hbase.testclassification.SmallTests;
|
||||||
|
Loading…
x
Reference in New Issue
Block a user